Abstract
Conventional deep learning methods have shown promising results in the medical domain when trained on accurate ground truth data. Pragmatically, due to constraints like lack of time or annotator inexperience, the ground truth data obtained from clinical environments may not always be impeccably accurate. In this paper, we investigate whether the presence of noise in ground truth data can be mitigated. We propose an innovative and efficient approach that addresses the challenge posed by noise in segmentation labels. Our method consists of four key components within a deep learning framework. First, we introduce a Vision Transformer-based modified encoder combined with a convolution-based decoder for the segmentation network, capitalizing on the recent success of self-attention mechanisms. Second, we consider a public CT spine segmentation dataset and devise a preprocessing step to generate (and even exaggerate) noisy labels, simulating real-world clinical situations. Third, to counteract the influence of noisy labels, we incorporate an adaptive denoising learning strategy (ADL) into the network training. Finally, we demonstrate through experimental results that the proposed method achieves noise-robust performance, outperforming existing baseline segmentation methods across multiple evaluation metrics.
| Original language | English |
|---|---|
| Article number | 7966 |
| Number of pages | 13 |
| Journal | Applied Sciences (Switzerland) |
| Volume | 13 |
| Issue number | 13 |
| DOIs | |
| Publication status | Published - 7 Jul 2023 |
Bibliographical note
Publisher Copyright:© 2023 by the authors.
Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Data Access Statement
The dataset used in this study is public available at http://spineweb.digitalimaginggroup.ca/Keywords
- computed tomography
- image segmentation
- noisy label
- Vision Transformer