TY - GEN
T1 - Triple-View Feature Learning for Medical Image Segmentation
AU - Wang, Ziyang
AU - Voiculescu, Irina
PY - 2022/9/15
Y1 - 2022/9/15
N2 - Deep learning models, e.g. supervised Encoder-Decoder style networks, exhibit promising performance in medical image segmentation, but come with a high labelling cost. We propose TriSegNet, a semi-supervised semantic segmentation framework. It uses triple-view feature learning on a limited amount of labelled data and a large amount of unlabeled data. The triple-view architecture consists of three pixel-level classifiers and a low-level shared-weight learning module. The model is first initialized with labelled data. Label processing, including data perturbation, confidence label voting and unconfident label detection for annotation, enables the model to train on labelled and unlabeled data simultaneously. The confidence of each model gets improved through the other two views of the feature learning. This process is repeated until each model reaches the same confidence level as its counterparts. This strategy enables triple-view learning of generic medical image datasets. Bespoke overlap-based and boundary-based loss functions are tailored to the different stages of the training. The segmentation results are evaluated on four publicly available benchmark datasets including Ultrasound, CT, MRI, and Histology images. Repeated experiments demonstrate the effectiveness of the proposed network compared against other semi-supervised algorithms, across a large set of evaluation measures.
AB - Deep learning models, e.g. supervised Encoder-Decoder style networks, exhibit promising performance in medical image segmentation, but come with a high labelling cost. We propose TriSegNet, a semi-supervised semantic segmentation framework. It uses triple-view feature learning on a limited amount of labelled data and a large amount of unlabeled data. The triple-view architecture consists of three pixel-level classifiers and a low-level shared-weight learning module. The model is first initialized with labelled data. Label processing, including data perturbation, confidence label voting and unconfident label detection for annotation, enables the model to train on labelled and unlabeled data simultaneously. The confidence of each model gets improved through the other two views of the feature learning. This process is repeated until each model reaches the same confidence level as its counterparts. This strategy enables triple-view learning of generic medical image datasets. Bespoke overlap-based and boundary-based loss functions are tailored to the different stages of the training. The segmentation results are evaluated on four publicly available benchmark datasets including Ultrasound, CT, MRI, and Histology images. Repeated experiments demonstrate the effectiveness of the proposed network compared against other semi-supervised algorithms, across a large set of evaluation measures.
UR - https://link.springer.com/chapter/10.1007/978-3-031-16876-5_5
UR - http://www.scopus.com/inward/record.url?scp=85138787169&partnerID=8YFLogxK
U2 - 10.1007/978-3-031-16876-5_5
DO - 10.1007/978-3-031-16876-5_5
M3 - Conference publication
AN - SCOPUS:85138787169
SN - 9783031168758
T3 - Lecture Notes in Computer Science (LNCS)
SP - 42
EP - 54
BT - Resource-Efficient Medical Image Analysis
A2 - Xu, Xinxing
A2 - Li, Xiaomeng
A2 - Mahapatra, Dwarikanath
A2 - Cheng, Li
A2 - Petitjean, Caroline
A2 - Fu, Huazhu
PB - Springer
T2 - 1st International Workshop on Resource-Efficient Medical Image Analysis, REMIA 2022, held in conjunction with the 25th International Conference on Medical Imaging and Computer-Assisted Intervention, MICCAI 2022
Y2 - 22 September 2022 through 22 September 2022
ER -