Next Article in Journal
The Accuracy of Three-Dimensional Soft Tissue Simulation in Orthognathic Surgery—A Systematic Review
Previous Article in Journal
Bayesian Networks in the Management of Hospital Admissions: A Comparison between Explainable AI and Black Box AI during the Pandemic
Previous Article in Special Issue
Elevating Chest X-ray Image Super-Resolution with Residual Network Enhancement
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semi-Supervised Medical Image Segmentation Based on Deep Consistent Collaborative Learning

College of Information Engineering, Dalian University, Dalian 116622, China
*
Author to whom correspondence should be addressed.
J. Imaging 2024, 10(5), 118; https://doi.org/10.3390/jimaging10050118
Submission received: 25 March 2024 / Revised: 2 May 2024 / Accepted: 10 May 2024 / Published: 14 May 2024

Abstract

:
In the realm of medical image analysis, the cost associated with acquiring accurately labeled data is prohibitively high. To address the issue of label scarcity, semi-supervised learning methods are employed, utilizing unlabeled data alongside a limited set of labeled data. This paper presents a novel semi-supervised medical segmentation framework, DCCLNet (deep consistency collaborative learning UNet), grounded in deep consistent co-learning. The framework synergistically integrates consistency learning from feature and input perturbations, coupled with collaborative training between CNN (convolutional neural networks) and ViT (vision transformer), to capitalize on the learning advantages offered by these two distinct paradigms. Feature perturbation involves the application of auxiliary decoders with varied feature disturbances to the main CNN backbone, enhancing the robustness of the CNN backbone through consistency constraints generated by the auxiliary and main decoders. Input perturbation employs an MT (mean teacher) architecture wherein the main network serves as the student model guided by a teacher model subjected to input perturbations. Collaborative training aims to improve the accuracy of the main networks by encouraging mutual learning between the CNN and ViT. Experiments conducted on publicly available datasets for ACDC (automated cardiac diagnosis challenge) and Prostate datasets yielded Dice coefficients of 0.890 and 0.812, respectively. Additionally, comprehensive ablation studies were performed to demonstrate the effectiveness of each methodological contribution in this study.

1. Introduction

Image segmentation algorithms based on supervised learning have demonstrated outstanding performance in a wide range of tasks [1,2,3], benefiting from their reliance on large amounts of high-quality, pixel-level annotated data, enabling models to learn precise segmentation decisions. However, the substantial demand for annotated data is not always feasible in practical applications, especially in the field of medical image segmentation. Due to the complexity of medical images and the high requirement for annotation accuracy, annotation tasks must be manually performed by experienced experts, which is not only time-consuming and labor-intensive but also costly. Therefore, in the field of medical image segmentation, the reality often involves only a small amount of annotated data available for use, while there is a relatively large amount of unlabeled data. Many semi-supervised segmentation algorithms have limited learning capabilities for the features of unlabeled data under conditions of a small amount of annotated data, resulting in low image segmentation accuracy. In this situation, effectively utilizing limited annotated data for training becomes a challenge in medical image segmentation.
An effective solution is to adopt semi-supervised learning methods. Currently, semi-supervised methods proposed based on deep learning mainly include consistency learning [4], co-training [5], self-training [6], adversarial learning [7], entropy minimization [8], and other methods. Among these methods, consistency regularization and co-training are widely used.
The consistency learning method is based on the assumption that the output should not change significantly when the input data are slightly perturbed. This method encourages the model to produce similar outputs for similar inputs, thereby enhancing the model’s robustness. According to the research in [9], the co-training framework is established on the premise that each data sample has two independent views, and each view can make predictions independently, with one view being redundant to the other. Simultaneously, this framework encourages the model to make consistent predictions for these two views. Initially, a segmentation model is separately trained for each view on the labeled data, and then the predictions of these two models for the unlabeled data are gradually introduced into the training set for continued training. The difference between co-training and consistency learning lies in that co-training utilizes pseudo-labels generated from unlabeled data to guide the training of models other than itself, while consistency learning forces the model’s output to remain consistent under different types of perturbations.
To address the challenges of low segmentation accuracy and inadequate generalization ability in the medical domain under conditions of scarce labels, this chapter proposes a semi-supervised medical image segmentation model named DCCLNet (deep consistency collaborative learning UNet) based on deep consistent co-training learning. This model combines the performance of consistency learning with both feature and input perturbations, and effectively leverages the advantages of learning different paradigms in co-training. The consistency learning strategy is primarily manifested by adding different auxiliary decoders on top of the backbone network for auxiliary learning, while employing a teacher–student architecture composed of a teacher model and the backbone network. The teacher model guides the learning of the backbone network, with the backbone network acting as the student model. Meanwhile, the co-training framework comprises two different network architectures, CNN (convolutional neural networks) and ViT (vision transformer), trained simultaneously and mutually learning from each other. Overall, the main contributions of this chapter are as follows:
(1)
Proposed a semi-supervised segmentation model named DCCLNet based on deep consistent co-training learning. Inspired by the CCT (cross-consistency training) semi-supervised method [10], this model adds different feature perturbations to the output of the backbone network’s CNN encoder, which are then inputted into auxiliary decoders. This encourages consistency between the outputs of the main decoder and the auxiliary decoder, thereby enhancing the robustness of the backbone network CNN.
(2)
Added a teacher model to form an MT (mean teacher) architecture [11] with the backbone network. Data with input perturbations are inputted into the teacher model, and a consistency constraint is imposed between the predictions of the teacher model and the backbone network to guide the training of the backbone network further, thereby improving the robustness and accuracy of the backbone network CNN. Moreover, the parameters of the teacher model are obtained from the backbone network, effectively reducing computational complexity.
(3)
Utilized the backbone network CNN and ViT to form a co-training architecture, where CNN can better capture local features, and ViT can better capture long-range dependencies. By simultaneously training from the perspectives of two different network architectures and learning pseudo-labels generated from each other’s predictions, the accuracy of the backbone network CNN can be improved.
(4)
Conducted experiments on the ACDC dataset [12] and Prostate dataset [13] under different percentages of labeled data. Experimental results demonstrate that DCCLNet outperforms other semi-supervised segmentation methods on the ACDC dataset and Prostate dataset in terms of segmentation effectiveness.
The subsequent structure of the paper is as follows: Section 2 reviews the research on semi-supervised medical image segmentation methods and prior studies related to the model we propose. Section 3 provides an in-depth introduction to the various modules of the proposed model. Section 4 and Section 5 delve into a thorough discussion of the experimental results. Finally, Section 6 concludes with considerations for future directions of this research.

2. Related Works

2.1. Semi-Supervised Medical Image Segmentation

Recently, the importance of semi-supervised learning in the field of medical image analysis has been increasing, and the use of semi-supervised semantic segmentation methods has become increasingly widespread, which is an excellent method for reducing the cost of image annotation. For example, Bai et al. [14] proposed a semi-supervised method for cardiac MR image segmentation. The training regimen alternates between labeled and unlabeled datasets, with pseudo-labels generated for the unlabeled data to facilitate segmentation of the originally labeled data. Xu et al. [15] proposed a semi-supervised segmentation method tailored for transrectal ultrasound (TRUS) images. The method leverages shadow consistency, which encompasses both shadow enhancement and shadow loss, to replicate the effects of low image quality and shadow artifacts commonly found in medical imaging. Shadow enhancement is achieved by superimposing synthetically generated shadow artifacts onto the input images, whereas shadow loss induces neural node degradation based on prior understanding of shadow artifacts, directly impacting the feature map. Zhu et al. [16] introduced an asymmetric multi-modal deep co-training framework for semi-supervised medical image segmentation, which includes two segmentation networks and two image translation networks designed for cross-modality image translation. Labeled data from one modality are utilized to train the segmentation network of the other modality, with consistency constraints applied to minimize the discrepancy between the two segmentation networks. Zhao et al. [17] presented a semi-supervised segmentation method featuring an information self-integration architecture for 3D whole-brain segmentation. The method employs various transformation strategies for unlabeled data, with the averaged prediction outcomes serving as pseudo-labels for the unlabeled data. Wu et al. [18] introduced a semi-supervised polyp segmentation model that leverages two segmentation networks to learn from each other. The approach utilizes a discriminator to generate confidence maps and incorporates an auxiliary discriminator to aid the primary discriminator, which may face performance limitations due to the scarcity of labeled data. Hou et al. [19] proposed a semi-supervised medical image segmentation method that employs a Leaking GAN (generative adversarial network) to contaminate the discriminator by leaking information from the generator, thereby promoting better learning of the discriminator. Wu et al. [20] introduced a semi-supervised segmentation method that augments the inter-class distance within the feature space by means of feature gradient map regularization. Additionally, the method employs class consistency constraints to alleviate the impact of noise interference on pseudo-labels. Wu et al. [21] presented a semi-supervised nuclei segmentation method that enhances the intra-class compactness and inter-class separability of features by aligning the feature maps of a teacher model with those of a student network, sampling from cross-image patches and pixel-level images. Miao et al. [22] proposed a semi-supervised medical image segmentation method with self-correcting collaborative training, which employs a self-correcting module to enhance the accuracy of model predictions on unlabeled data and further improves the model’s learning capability using pixel-wise contrastive learning loss. Shen et al. [23] proposed a UCMT (uncertainty-guided collaborative mean teacher) model for semi-supervised medical image segmentation, where sub-networks learn from unlabeled data through collaborative training, and uncertainty estimation strategies are applied to predictions of the teacher network, allowing sub-networks to learn from regions of high confidence.

2.2. Consistent Learning

The consistency learning method imposes consistency on the predictions of input images under different perturbations, where perturbations should not alter the network’s predictions. Various perturbations are employed in consistency learning methods, which can roughly be categorized into input perturbations and feature map perturbations.
Input perturbation refers to applying certain perturbations to the input samples, such as Gaussian noise, random rotation, flipping, and contrast changes, encouraging segmentation networks to maintain consistency in predictions for unlabeled data under different perturbations. The Π-model proposed by Sajjadi et al. [24] and the MT (Mean Teacher) method proposed by Tarvianen et al. are similar approaches. The former applies two versions of data perturbation to samples and adds a consistency constraint between their predictions, while the latter improves upon the former by obtaining the weights of the student model through training, and the weights of the teacher model are continually updated through exponential moving average (EMA) calculation based on the weights of the student model, imposing consistency constraints between the predictions of the teacher and student. The ICT (Interpolation Consistency Training) [25] method introduces a training approach based on interpolation consistency, which interpolates unlabeled data according to the idea of mix-up [26]. Mix-up involves applying a simple linear transformation to input data. ICT calculates consistency loss between predictions at interpolated pixel points of unlabeled data and the predictions of those points themselves, training the network through consistency constraints between them. Huang et al. [27] proposed a CC-Net based on complementary consistency training for semi-supervised left atrial image segmentation. The method utilizes two auxiliary models and introduces model-level perturbations, enhancing the learning capability of the main model through the imposition of consistency constraints between the auxiliary and main models. Although CC-Net facilitates ample assistance for the backbone network’s learning through two auxiliary models with model perturbations, employing identical encoder and decoder structures may potentially diminish the efficacy of co-training.
Feature perturbation involves perturbations added to the feature maps after convolution of input samples. The CCT (Cross Consistency Training) method proposes a cross-consistency approach, where unlabeled data are input into a shared encoder, and different feature perturbations are added to the output of the shared encoder, which are then inputted into auxiliary decoders. The network model is trained by imposing consistency loss between the predictions of the auxiliary decoder and the main decoder. The RD (Regularized Dropout) [28] method introduces a simple consistency training strategy called R-Drop, which enforces consistency between the outputs of different sub-models through dropout. Li et al. [29] introduced a semi-supervised segmentation approach for COVID-19 based on an uncertainty-guided dual consistency learning network (UDC-Net). UDC-Net incorporated seven types of feature perturbations for seven additional auxiliary decoders, necessitating that these seven predictions remain consistent with the main decoder. The feature-level perturbations included feature noise, feature dropout, object masking, context masking, guidance loss, intermediate boosting, and random loss. This method is very comprehensive in its use of feature perturbations but overlooks that input perturbations can also mislead the encoder’s feature learning during the encoding process.

2.3. Co-Training

The co-training method involves constructing two or more network models capable of representing independent perspectives. Typically, different network architectures and specialized training methods are used to build co-training models. The CPS (Cross Consistency Training) [30] method proposes a cross-pseudo-supervised training approach, employing two networks that are structurally consistent but differ in initialization state. It introduces consistency constraints to encourage these two networks to produce the same output for the same input sample. CPS generates labels from the predictions of one network and uses them as pseudo-labels for the predictions of the other network, and vice versa, allowing these two networks to learn from each other’s perspective. The CTCT (Cross Teaching between CNN and Transformer) [31] method further proposes mutual teaching between two different networks utilizing CNN and transformer architectures, enabling them to observe data from two completely different perspectives and learn from each other, thus compensating for the limitations of learning with a single CNN network architecture. The DCT (Deep Co-Training) [32] method introduces a deep co-learning approach, using multiple deep neural networks to observe data from multiple perspectives and impose constraints by leveraging differences generated by adversarial samples. Wang et al. [33] proposed a semi-supervised medical image segmentation framework, consisting of a feature learning module composed of CNN and ViT, and a feature-guidance module composed of a ViT teacher model. The feature learning module combines the advantages of CNN and ViT through dual-view co-training. The feature guidance module averages network weights using an MT architecture and considers the output of the teacher model as the final prediction. The framework employed two different network architectures, and to further guide the ViT, the framework did not consider the potential negative effects of feature perturbations. Huang et al. [34] developed a semi-supervised dual cooperative network (SD-Net) for liver CT segmentation, which adopts two cooperatively trained models. They used adaptive mask refinement to refine the differences between predictions for labeled data and utilized a dynamic pseudo-label-generation strategy for unlabeled data to select better predictions as pseudo-labels. This method utilizes two networks to collaborate and imposes constraints on the uncertain regions predicted by both networks. However, both network architectures are CNN-based, lacking the advantage of ViT in capturing long-range dependencies.

3. Method

3.1. The Overall Structure of the Model

The overall model architecture of DCCLNet is shown in Figure 1. The model is primarily based on the backbone network UNet [35], with three auxiliary decoders added on the basis of the main decoder of the UNet backbone network, each applying different feature perturbations. Additionally, a teacher model with input perturbation is added, which shares the same UNet network structure as the UNet backbone network, along with Swin-UNet [36].
In semi-supervised segmentation, images are typically divided into two parts: labeled images and unlabeled images. A dataset with N labeled images is represented as x l , y l L , while a dataset with M unlabeled images is represented as x u U .
The labeled data are only received by the backbone networks UNet and Swin-UNet, while the unlabeled data are received by all networks in the framework. For the labeled data, the predictions generated by DCCLNet are compared only to the ground truth labels and continually learned from them. For the unlabeled data, DCCLNet guides the backbone network UNet with different levels of perturbation by adding auxiliary decoders with feature perturbation and a teacher model with input perturbation. Consistency constraints are imposed by the differences between their predictions and those of the backbone network UNet. Additionally, DCCLNet employs a collaborative training approach using two different network architectures, UNet and Swin-UNet, to learn from each other.

3.2. Auxiliary Decoder Assist

Inspired by the CCT semi-supervised method, three different types of feature perturbations are added to the output of the backbone network encoder, which are then respectively inputted into the auxiliary decoders. Based on the principle of consistency learning, perturbing the feature maps of the decoder input should not change the output predictions. By imposing consistency between the predictions of the backbone network and the auxiliary decoders, the generalization ability of the backbone network UNet is enhanced.
As shown in Figure 2, the decoder structure of the auxiliary decoder is the same as that of the backbone network UNet, and they share an encoder. Here, P1, P2, and P3 represent three different types of feature perturbations. Unlabeled data are input into the shared UNet encoder, whereas the main decoder receives the direct output of the encoder, and the auxiliary decoders respectively add these three types of feature perturbations to the output feature maps of the encoder. This enables the auxiliary decoders to assist the backbone network in learning from feature perturbations.

3.2.1. Characteristic Perturbation

Feature Noise: Uniformly sample a noise tensor N~U(-3,3) with the same size as the feature map h, then multiply this noise tensor with h, injecting the noise into h, resulting in the perturbed result h = h N + h . The injected noise is proportional to each activation.
Feature Dropout: Uniformly sample a threshold value γ~U(0.6,0.9), perform channel-wise summation and normalization on the feature map h to obtain h , generate a mask F d r o p = h < γ , and use this mask to add perturbation to h, resulting in h = F d r o p h , effectively masking the most active regions of h by 10% to 40%.
Dropout: Use Dropout [37] for feature map h generation as random perturbations.

3.2.2. Characteristic Perturbation Loss

For the unlabeled dataset x u U , let the output of the backbone network be represented as f U x u , and the output of the auxiliary decoder Aux be represented as f A i x u , i = 1 , 2 , 3 . The difference between the predictions of the backbone network UNet and the auxiliary decoder Aux is constrained by the unsupervised loss L u n sup u a constructed by mean squared error (MSE):
L u n sup u a = 1 3 × i = 1 3 M S E f U x u , f A i x u

3.3. Teacher Model Guidance

Addition of the teacher model forms the MT architecture with the backbone network UNet, where the backbone network acts as the student model. Perturbations are added to the inputs of the teacher model based on the principle that perturbations should not alter the model’s output. Consistency constraints are imposed between the predictions of the backbone network and the teacher model, further enhancing the robustness of the backbone network.
As shown in Figure 3, both the teacher model and the backbone network are UNet networks. The parameters of the teacher model are directly obtained from the backbone network UNet through exponential moving average (EMA), effectively reducing the computational cost of the model. Unlabeled data are directly input into the backbone network UNet, while after adding noise, it is input into the teacher model. This enables the teacher model to guide the backbone network in learning from input perturbations.

3.3.1. Teacher Model Parameter Update

The study in [11] indicates that averaging model weights during training often results in models that are more accurate than those directly using the final weights. Therefore, the EMA method is utilized to update the parameters of the teacher model. Let θs denote the parameters of the student model, and the update calculation for the parameters θt of the teacher model is as follows:
θ t = λ θ t 1 + 1 λ θ s
where λ is the decay coefficient used to control the updating rate of EMA, and t denotes the updating round.

3.3.2. Input Disturbance Loss

For the unlabeled dataset x u U , let the data with added input perturbation be denoted as x u n o i , and the output of the teacher model for the unlabeled data is represented as f T x u n o i . The discrepancy between the predictions of the main network and the teacher model is computed through the unsupervised loss L u n sup u t , which is based on the mean square error.
L u n sup u t = M S E f U x u , f T x u n o i

3.4. CNN and ViT Collaborative Training

3.4.1. Collaborative Training Process

In reference [31], a semi-supervised segmentation method was proposed, which involves the cross-teaching of UNet and Swin-UNet, yielding impressive segmentation results. Consequently, this paper utilizes a collaborative training framework composed of the CNN-based UNet and the Transformer-based Swin-UNet, which are trained concurrently. Through the observation of data by two different network architectures and mutual learning during the training process, the advantages of each are leveraged to compensate for shortcomings, effectively enhancing the segmentation accuracy of the main network UNet.
As illustrated in Figure 4, when inputting labeled data, the predictions of the main network UNet and Swin-UNet are constrained by the ground truth labels. When inputting unlabeled data, the labels generated by Swin-UNet are utilized to guide the predictions of UNet. Simultaneously, the labels generated by UNet are used to guide the predictions of Swin-UNet. Thus, the objective is achieved where UNet and Swin-UNet mutually learn and progress together.

3.4.2. Co-Training Loss

For the labeled dataset x l , y l L , let the outputs of the main network UNet be denoted as f U x l and the outputs of Swin-UNet be denoted as f S U x l . The differences between them and the ground truth labels are calculated using the supervised losses constructed by the cross-entropy loss function (CE) and the Dice loss function, denoted as L sup u and L sup s u , respectively:
L sup u = 1 2 C E f U x l , y l + D i c e f U x l , y l
L sup s u = 1 2 C E f S U x l , y l + D i c e f S U x l , y l
For the unlabeled dataset x u U , let the predictions of the main network UNet be denoted as f U x u with generated labels y U x u = arg max f U x u , and the predictions of Swin-UNet be denoted as f S U x u with generated labels y S U x u = arg max f S U x u . On one hand, the predictions f U x u of the main network and the labels y S U x u generated by Swin-UNet are used to compute the loss. On the other hand, the predictions f S U x u of Swin-UNet and the labels y U x u generated by the main network are used to compute the loss. The differences between them are calculated by the unsupervised loss L u n sup u s u :
L u n sup u s u = D i c e f U x u , y S U x u + D i c e f S U x u , y U x u

3.5. Overall Loss Function

The loss of the semi-supervised framework consists of the supervised loss L sup and the unsupervised loss L u n sup . In the overall framework of DCCLNet proposed in this paper, the training of the auxiliary decoder, the main network UNet, the teacher model, and Swin-UNet are conducted simultaneously. Therefore, the overall loss function L t o t a l of DCCLNet is composed of all the aforementioned loss functions:
L u n sup u s u = D i c e f U x u , y S U x u + D i c e f S U x u , y U x u
In this equation, λ represents the weight of the unsupervised loss. Different weights are used for consistency learning and collaborative training in this paper, where λ 1 is the weight factor for the consistency loss of the main UNet network, λ 1 = 0.01 × e 5 × 1 t / t max 2 ; λ 2 is the weight factor for the collaborative training loss of the main UNet network, λ 2 = u × λ 1 ; and u is a scaling factor. t denotes the current iteration round, updated every 150 iterations and gradually increased during training [38], allowing the model to focus on labeled images during initialization and gradually shift the focus to unlabeled images. The overall loss L t o t a l is the final training objective of this framework.

4. Experiments

4.1. Data Preparation

ACDC dataset: The semi-supervised segmentation method proposed in this paper is evaluated on the MICCAI 2017 ACDC (Automated Cardiac Diagnosis Challenge) dataset to demonstrate its effectiveness and compare it with other baseline methods. The ACDC dataset consists of 200 short-axis cardiac cine MRI scans from 100 patients, each containing four segmentation classes: left ventricle, right ventricle, left ventricular myocardium and background. Among the 100 patients, 70 were selected as training samples, 10 as validation samples, and 20 as testing samples.
PROMISE12 dataset: Similar experiments to those on the ACDC dataset are conducted on the PROMISE12 (Prostate MR Image Segmentation Challenge) dataset. The PROMISE12 dataset comprises 50 MRI scans from patients, with 35 patients used for training, 5 for validation, and 10 for testing, each containing two segmentation classes: prostate and background.

4.2. Experimental Setup

To ensure fair comparison among all methods, all comparative methods and ablation experiments in this study were conducted under the same conditions. The experiments were performed using an NVIDIA GeForce RTX 2080 Ti GPU (NVIDIA, Dalian, China) with retraining in PyTorch 3.9. The datasets were normalized and resized to 224 × 224, and data augmentation techniques such as random rotation and flipping were applied. An SGD (stochastic gradient descent) optimizer was used during training with an initial learning rate of 0.01, and the Poly learning rate policy was employed to update the learning rate continuously. The training process consisted of 30,000 iterations, with a batch size of 16, including 8 labeled data and 8 unlabeled data per batch. The scaling factor u of DCCLNet was set to 15. The optimal weights of the backbone network in DCCLNet were used to generate the final predictions.

4.3. Evaluation Index

This paper evaluates the performance of the proposed model using three standard evaluation metrics: Dice similarity coefficient (DSC), 95% Hausdorff distance (HD95), and average symmetric surface distance (ASD). The formulas for calculation are as follows:
D S C = 2 T P F N + 2 T P + F P
DSC mainly measures the similarity between two segmentation maps, where TP is true positive, FP is false positive, TN is true negative, and FN is false negative. DSC ranges from [0, 1], with values closer to 1 indicating better segmentation results.
H D 95 = max t 2 max t 1 T 2 P 2 × 0.95
HD represents the maximum distance measure between the surface point sets of the predicted result and the ground truth label. Since the Hausdorff distance metric is affected by noisy points, the 95th percentile of the Hausdorff distance is used to measure the segmentation result, where T is the boundary of the label region and P is the boundary of the segmented region predicted by the model.
A S D = 1 S A + S B s A S A d s A , S B + s B S B d s B , S A
ASD represents the average distance between the surface point sets of the predicted result and the ground truth label, where S(A) represents the surface voxels in set A, and d(v, S(A)) represents the shortest distance from any voxel to S(A); HD95 and ASD measure the segmentation model’s ability to predict edges, with smaller values indicating that the segmented edges are closer to the edges of the ground truth label.

4.4. Comparative Experimental

4.4.1. ACDC Dataset Comparative Experimental Analysis

As shown in Table 1, this paper utilized two different percentages of labeled datasets from the ACDC training set, namely 5% and 10%, and conducted experiments with other semi-supervised segmentation methods for these two scenarios. The proposed framework DCCLNet achieved better performance on most metrics.
When using 5% labeled samples, consistency-based methods like CCT, MT, and ICT performed poorly on all three metrics. This is because these three consistency methods only employ a single network architecture. CCT solely assists the main network in learning from unlabeled data by adding auxiliary decoders with feature perturbations; MT also only utilizes a teacher model with the same architecture to guide the student model, where the teacher model’s input is perturbed; ICT uses interpolation to train on perturbed data. UAMT, building upon MT, incorporates uncertainty estimation to select more reliable pseudo-labels for guiding the student model. However, when the labeled data volume is 5%, the available labeled data for training after selection is too scarce, resulting in inferior performance compared to MT. URPC generates multi-scale predictions through a pyramid framework, to which consistency loss and uncertainty correction are applied. EM improves pseudo-label quality by reducing the entropy of pseudo-labels, but due to the scarcity of labels and a single network, the effect is poor. DAN also performs poorly as adversarial training is not applicable in this scenario. CPS supervises the pseudo-labels of each other through two networks with the same architecture, which enhances label quality with acceptable performance. DCT increases adversarial sample loss by adding noise to generate adversarial samples, but the accuracy of labels corresponding to adversarial samples is insufficient due to the use of a single network for label generation. RD trains the network using only a regularization loss, resulting in the worst performance. CTCT exhibits excellent performance, employing cross-teaching between CNN and Transformer, effectively improving segmentation accuracy by mutual learning from two different nework architectures. Compared to CTCT, the proposed method DCCLNet shows signifcant improvements in all metrics. Specifically, DSC, HD95, and ASD increased by 7.8%, 1.5 mm, and −1.0 mm, respectively, indicating that DCCLNet can maintain good performance with fewer labeled data.
When using 10% of the samples, CCT and CPS performed relatively well, indicating that the potential of cross-pseudo-supervision and feature perturbation methods can be better utilized with more labeled data. CTCT still outperformed other semi-supervised segmentation methods without ViT. Compared to CTCT, DCCLNet showed improvements in DSC, HD95, and ASD by 2.4%, −2.3 mm, and −0.1 mm, respectively, indicating that DCCLNet learns more effectively from unlabeled data, and the feature perturbation and input perturbation added in the framework also achieved certain effects. In Figure 5, the final results of all methods trained and tested with 10% labeled samples are visualized. It can be seen from Figure 5 that DCCLNet’s segmentation performance has notable improvements in both boundary and overall accuracy compared to other methods.

4.4.2. Prostate Dataset Comparative Experimental Analysis

As shown in Table 2, this paper utilized two different percentages of labeled datasets from the Prostate training set, namely 10% and 20%. On this dataset, DCCLNet achieved the best results in terms of DSC, HD95, and ASD.
When using 10% labeled samples, it is evident that the segmentation performance of the majority of semi-supervised segmentation methods is poor, even when segmenting only the prostate organ. This indicates that when labeled data is extremely scarce, the segmentation results from single-network architectures are subpar. URPC and CPS methods performed extremely poorly, with DSC values even below 0.35. The segmentation results of the CTCT method were relatively good, suggesting that even with very few labeled data, decent results can be obtained through the use of CNN and ViT, two drastically different learning paradigms. Compared to CTCT, DCCLNet showed improvements in all metrics, with DSC, HD95, and ASD values increasing by 2.9%, −7.6 mm, and −2.0 mm, respectively. This demonstrates that the deep integration of consistency regularization methods and CNN-ViT collaborative training methods can enhance model performance. Figure 6 visualizes the segmentation results of the DCCLNet method and other semi-supervised methods on the Prostate dataset.

4.5. Ablation Experiment

The ablation study results of DCCLNet on the ACDC dataset using 10% labeled data are presented in Table 3, while those on the Prostate dataset with 20% labeled data are shown in Table 4. As DCCLNet employs the UNet network as its backbone, the experimental results of UNet are used as the baseline to assess the effectiveness of the proposed improvements in this study.
For the results of the ablation experiments on these two datasets, the study first investigated the segmentation performance when only a single component was added to UNet, namely UNet + Aux, UNet + Tea, and UNet + ViT. All three architectures exhibited better experimental results compared to the baseline, indicating that adding feature perturbation or input perturbation, as well as using different network architectures for co-training, can enhance the learning of the backbone network for unlabeled data. Subsequently, experiments were conducted by combining any two components, all of which yielded excellent results, indicating that both the combination of perturbations and the combination of perturbations with co-training can further improve the learning ability of the backbone network. However, it was also observed that the improvement achieved by the co-training architecture of CNN with ViT was significant, outperforming the direct addition of perturbations. This is because medical image perturbations lead to significant data variations, and a single UNet network architecture is insufficient to learn enough knowledge for correction. Finally, applying all components to the backbone network yielded better results than all previous scenarios.
The experimental data in Table 3 demonstrate that for the DSC evaluation metric, DCCLNet outperformed the baseline by 11.2 percentage points, and for the HD95 and ASD evaluation metrics, it improved by −7.8 mm and −1.4 mm, respectively. Compared to the framework with only perturbations added (UNet + Aux + Tea), DCCLNet improved by 2.4 percentage points in DSC score and by −2.3 mm and −0.1 mm in HD95 and ASD distances, respectively. Compared to the framework with only co-training architecture used (UNet + ViT), DCCLNet improved by 6.4 percentage points in DSC score and by −7.9 mm and −1.1 mm in HD95 and ASD distances, respectively.
The experimental data in Table 4 indicate that for the DSC evaluation metric, DCCLNet outperforms the baseline by 24.9 percentage points, while for the HD95 and ASD evaluation metrics, it improves by −76 mm and −18.2 mm, respectively. Compared to the framework that only adds perturbations (UNet + Aux + Tea), DCCLNet achieves an increase of 11.7 percentage points in DSC score, and improvements of −13.9 mm and −3.7 mm in HD95 and ASD distances, respectively. Compared to the framework using only collaborative training (UNet + ViT), DCCLNet shows an increase of 2.9 percentage points in DSC score, with improvements of −7.6 mm and −2.0 mm in HD95 and ASD distances, respectively.

5. Discussion

In response to the challenges and issues in current semi-supervised medical image segmentation tasks, this paper proposes a semi-supervised medical image segmentation model called DCCLNet based on deep consistent co-training learning. The model combines the performance of consistency learning with feature perturbation and input perturbation, effectively leveraging the advantages of learning from different paradigms in co-training. When applied to various medical datasets, the segmentation results are outstanding even with limited annotated data. Although this study has made some progress, there are also shortcomings.
Firstly, although the semi-supervised learning method proposed in this study can fully learn the feature information of unlabeled data and enhance network generalization ability, there is still room for improvement in segmentation performance compared to supervised learning algorithms. Therefore, future research will focus on further developing semi-supervised segmentation algorithms to make their performance approach that of fully supervised methods.
Secondly, this study introduced feature perturbation and input perturbation from the perspective of perturbations to help the backbone network learn knowledge from unlabeled data and enhance its robustness. Additionally, from the standpoint of co-training, the combination of CNN and ViT architectures mutually learned from each other, effectively improving the accuracy of the backbone network. Although the improved semi-supervised segmentation model further enhances the accuracy of medical image segmentation, it also requires more computational resources. Future work could focus on more efficiently combining CNN and ViT with fewer computational requirements to improve model performance and reduce parameter count.
Moreover, this study involves setting numerous hyperparameters, including learning rates and weights of unsupervised losses, which need to be explored through continuous experimentation. Hence, future research directions could explore more automated hyperparameter optimization strategies.

6. Conclusions

Supervised image segmentation algorithms have demonstrated outstanding performance in a wide range of tasks, owing to their reliance on large amounts of high-quality, pixel-level annotated data, which allows models to learn precise segmentation decisions. However, this heavy reliance on annotated data is not always feasible in practical applications, especially in the field of medical image segmentation. Due to the complexity of medical images and the high requirement for annotation accuracy, the annotation task must be manually performed by experienced experts, which is not only time-consuming but also costly. Therefore, in medical image segmentation, the reality often involves having only a small amount of labeled data available, while there is a relatively large amount of unlabeled data. Many semi-supervised segmentation algorithms have limited ability to learn features from unlabeled data under the condition of a small amount of labeled data, leading to low accuracy in image segmentation.
To address the above issues, a semi-supervised medical image segmentation model based on deep consistency collaborative learning, named DCCLNet, is proposed. This model combines the consistency learning performance of feature perturbation and input perturbation and effectively utilizes the advantages of different paradigms learning in collaborative training. The consistency learning strategy is mainly reflected in the addition of different auxiliary decoders on the basis of the main network, enabling the auxiliary main network to learn. Meanwhile, the MT architecture, consisting of a teacher model and the main network, is used, with the teacher model guiding the main network’s learning, whereas the main network acts as the student model. The collaborative training framework consists of two different network architectures, CNN and ViT, which are trained simultaneously and learn from each other.

Author Contributions

The authors acknowledge their involvement in the paper in the following capacities: X.Z. and W.W. contributed to the study’s conception and design. X.Z. and W.W. conducted the analysis and interpretation of the results. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by The National Natural Science Foundation of China (Grant No. 61971424).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The ACDC dataset and Prostate dataset are publicly available at https://github.com/HiLab-git/SSL4MIS (accessed on 10 October 2022).

Acknowledgments

We would like to thank the editors and reviewers for their valuable opinions and suggestions that improved this research.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, J.; Wei, L.; Wang, L.; Zhou, Q.; Zhu, L.; Qin, J. Boundary-aware transformers for skin lesion segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2021: 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Proceedings, Part I 24. pp. 206–216. [Google Scholar]
  2. Hatamizadeh, A.; Tang, Y.; Nath, V.; Yang, D.; Myronenko, A.; Landman, B.; Roth, H.R.; Xu, D. Unetr: Transformers for 3D medical image segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 574–584. [Google Scholar]
  3. Xu, G.; Zhang, X.; He, X.; Wu, X. Levit-unet: Make faster encoders with transformer for medical image segmentation. In Proceedings of the Chinese Conference on Pattern Recognition and Computer Vision (PRCV), Xiamen, China, 13–15 October 2023; pp. 42–53. [Google Scholar]
  4. Su, J.; Luo, Z.; Lian, S.; Lin, D.; Li, S. Consistency learning with dynamic weighting and class-agnostic regularization for semi-supervised medical image segmentation. Biomed. Signal Process. Control 2024, 90, 105902. [Google Scholar] [CrossRef]
  5. Wang, P.; Peng, J.; Pedersoli, M.; Zhou, Y.; Zhang, C.; Desrosiers, C. Self-paced and self-consistent co-training for semi-supervised image segmentation. Med. Image Anal. 2021, 73, 102146. [Google Scholar] [CrossRef] [PubMed]
  6. Chaitanya, K.; Erdil, E.; Karani, N.; Konukoglu, E. Local contrastive loss with pseudo-label based self-training for semi-supervised medical image segmentation. Med. Image Anal. 2023, 87, 102792. [Google Scholar] [CrossRef] [PubMed]
  7. Peiris, H.; Hayat, M.; Chen, Z.; Egan, G.; Harandi, M. Uncertainty-guided dual-views for semi-supervised volumetric medical image segmentation. Nat. Mach. Intell. 2023, 5, 724–738. [Google Scholar] [CrossRef]
  8. Yang, L.; Qi, L.; Feng, L.; Zhang, W.; Shi, Y. Revisiting weak-to-strong consistency in semi-supervised semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7236–7246. [Google Scholar]
  9. Blum, A.; Mitchell, T. Combining labeled and unlabeled data with co-training. In Proceedings of the Eleventh Annual Conference on Computational Learning Theory, Madison, WI, USA, 24–26 July 1998; pp. 92–100. [Google Scholar]
  10. Ouali, Y.; Hudelot, C.; Tami, M. Semi-supervised semantic segmentation with cross-consistency training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12674–12684. [Google Scholar]
  11. Tarvainen, A.; Valpola, H. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  12. Bernard, O.; Lalande, A.; Zotti, C.; Cervenansky, F.; Yang, X.; Heng, P.-A.; Cetin, I.; Lekadir, K.; Camara, O.; Ballester, M.A.G. Deep learning techniques for automatic MRI cardiac multi-structures segmentation and diagnosis: Is the problem solved? IEEE Trans. Med. Imaging 2018, 37, 2514–2525. [Google Scholar] [CrossRef] [PubMed]
  13. Litjens, G.; Toth, R.; Van De Ven, W.; Hoeks, C.; Kerkstra, S.; Van Ginneken, B.; Vincent, G.; Guillard, G.; Birbeck, N.; Zhang, J. Evaluation of prostate segmentation algorithms for MRI: The PROMISE12 challenge. Med. Image Anal. 2014, 18, 359–373. [Google Scholar] [CrossRef] [PubMed]
  14. Bai, W.; Oktay, O.; Sinclair, M.; Suzuki, H.; Rajchl, M.; Tarroni, G.; Glocker, B.; King, A.; Matthews, P.M.; Rueckert, D. Semi-supervised learning for network-based cardiac MR image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2017: 20th International Conference, Quebec City, QC, Canada, 11–13 September 2017; Proceedings, Part II 20. pp. 253–260. [Google Scholar]
  15. Xu, X.; Sanford, T.; Turkbey, B.; Xu, S.; Wood, B.J.; Yan, P. Shadow-consistent semi-supervised learning for prostate ultrasound segmentation. IEEE Trans. Med. Imaging 2021, 41, 1331–1345. [Google Scholar] [CrossRef] [PubMed]
  16. Zhu, L.; Yang, K.; Zhang, M.; Chan, L.L.; Ng, T.K.; Ooi, B.C. Semi-supervised unpaired multi-modal learning for label-efficient medical image segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2021: 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Proceedings, Part II 24. pp. 394–404. [Google Scholar]
  17. Zhao, Y.-X.; Zhang, Y.-M.; Song, M.; Liu, C.-L. Multi-view semi-supervised 3D whole brain segmentation with a self-ensemble network. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2019: 22nd International Conference, Shenzhen, China, 13–17 October 2019; Proceedings, Part III 22. pp. 256–265. [Google Scholar]
  18. Wu, H.; Chen, G.; Wen, Z.; Qin, J. Collaborative and adversarial learning of focused and dispersive representations for semi-supervised polyp segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 3489–3498. [Google Scholar]
  19. Hou, J.; Ding, X.; Deng, J.D. Semi-supervised semantic segmentation of vessel images using leaking perturbations. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 2625–2634. [Google Scholar]
  20. Wu, J.; Fan, H.; Zhang, X.; Lin, S.; Li, Z. Semi-supervised semantic segmentation via entropy minimization. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021; pp. 1–6. [Google Scholar]
  21. Wu, H.; Wang, Z.; Song, Y.; Yang, L.; Qin, J. Cross-patch dense contrastive learning for semi-supervised segmentation of cellular nuclei in histopathologic images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11666–11675. [Google Scholar]
  22. Miao, J.; Zhou, S.-P.; Zhou, G.-Q.; Wang, K.-N.; Yang, M.; Zhou, S.; Chen, Y. SC-SSL: Self-correcting Collaborative and Contrastive Co-training Model for Semi-Supervised Medical Image Segmentation. IEEE Trans. Med. Imaging 2023, 43, 1347–1364. [Google Scholar] [CrossRef] [PubMed]
  23. Shen, Z.; Cao, P.; Yang, H.; Liu, X.; Yang, J.; Zaiane, O.R. Co-training with high-confidence pseudo labels for semi-supervised medical image segmentation. arXiv 2023, arXiv:2301.04465. [Google Scholar]
  24. Sajjadi, M.; Javanmardi, M.; Tasdizen, T. Regularization with stochastic transformations and perturbations for deep semi-supervised learning. In Proceedings of the Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  25. Verma, V.; Kawaguchi, K.; Lamb, A.; Kannala, J.; Solin, A.; Bengio, Y.; Lopez-Paz, D. Interpolation consistency training for semi-supervised learning. Neural Netw. 2022, 145, 90–106. [Google Scholar] [CrossRef] [PubMed]
  26. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar]
  27. Huang, H.; Chen, Z.; Chen, C.; Lu, M.; Zou, Y. Complementary consistency semi-supervised learning for 3D left atrial image segmentation. Comput. Biol. Med. 2023, 165, 107368. [Google Scholar] [CrossRef] [PubMed]
  28. Wu, L.; Li, J.; Wang, Y.; Meng, Q.; Qin, T.; Chen, W.; Zhang, M.; Liu, T.-Y. R-drop: Regularized dropout for neural networks. Adv. Neural Inf. Process. Syst. 2021, 34, 10890–10905. [Google Scholar]
  29. Li, Y.; Luo, L.; Lin, H.; Chen, H.; Heng, P.-A. Dual-consistency semi-supervised learning with uncertainty quantification for COVID-19 lesion segmentation from CT images. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2021: 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Proceedings, Part II 24. pp. 199–209. [Google Scholar]
  30. Chen, X.; Yuan, Y.; Zeng, G.; Wang, J. Semi-supervised semantic segmentation with cross pseudo supervision. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 2613–2622. [Google Scholar]
  31. Luo, X.; Hu, M.; Song, T.; Wang, G.; Zhang, S. Semi-supervised medical image segmentation via cross teaching between cnn and transformer. In Proceedings of the International Conference on Medical Imaging with Deep Learning, Zurich, Switzerland, 6–8 July 2022; pp. 820–833. [Google Scholar]
  32. Peng, J.; Estrada, G.; Pedersoli, M.; Desrosiers, C. Deep co-training for semi-supervised image segmentation. Pattern Recognit. 2020, 107, 107269. [Google Scholar] [CrossRef]
  33. Wang, Z.; Li, T.; Zheng, J.-Q.; Huang, B. When cnn meet with vit: Towards semi-supervised learning for multi-class medical image semantic segmentation. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 424–441. [Google Scholar]
  34. Huang, S.; Luo, J.; Ou, Y.; Pang, Y.; Nie, X.; Zhang, G. Sd-net: A semi-supervised double-cooperative network for liver segmentation from computed tomography (CT) images. J. Cancer Res. Clin. Oncol. 2024, 150, 79. [Google Scholar] [CrossRef]
  35. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. pp. 234–241. [Google Scholar]
  36. Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-unet: Unet-like pure transformer for medical image segmentation. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 205–218. [Google Scholar]
  37. Tompson, J.; Goroshin, R.; Jain, A.; LeCun, Y.; Bregler, C. Efficient object localization using convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Quebec City, QC, Canada, 27–30 September 2015; pp. 648–656. [Google Scholar]
  38. Laine, S.; Aila, T. Temporal ensembling for semi-supervised learning. arXiv 2016, arXiv:1610.02242. [Google Scholar]
  39. Vu, T.-H.; Jain, H.; Bucher, M.; Cord, M.; Pérez, P. Advent: Adversarial entropy minimization for domain adaptation in semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2517–2526. [Google Scholar]
  40. Liu, J.; Desrosiers, C.; Zhou, Y. Semi-supervised medical image segmentation using cross-model pseudo-supervision with shape awareness and local context constraints. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Singapore, 18–22 September 2022; pp. 140–150. [Google Scholar]
  41. Luo, X.; Liao, W.; Chen, J.; Song, T.; Chen, Y.; Zhang, S.; Chen, N.; Wang, G.; Zhang, S. Efficient semi-supervised gross target volume of nasopharyngeal carcinoma segmentation via uncertainty rectified pyramid consistency. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2021: 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Proceedings, Part II 24. pp. 318–329. [Google Scholar]
Figure 1. Overall framework DCCLNet. where P1, P2, P3 are three different characteristic perturbations.
Figure 1. Overall framework DCCLNet. where P1, P2, P3 are three different characteristic perturbations.
Jimaging 10 00118 g001
Figure 2. Auxiliary decoder and backbone network UNet.
Figure 2. Auxiliary decoder and backbone network UNet.
Jimaging 10 00118 g002
Figure 3. The framework of MT (mean teacher) (EMA: exponential moving average).
Figure 3. The framework of MT (mean teacher) (EMA: exponential moving average).
Jimaging 10 00118 g003
Figure 4. UNet and Swin-UNet collaborative training framework.
Figure 4. UNet and Swin-UNet collaborative training framework.
Jimaging 10 00118 g004
Figure 5. Visualization results of different methods on the ACDC test set. In the image, red denotes the right ventricle, blue denotes the myocardium, and green denotes the left ventricle.
Figure 5. Visualization results of different methods on the ACDC test set. In the image, red denotes the right ventricle, blue denotes the myocardium, and green denotes the left ventricle.
Jimaging 10 00118 g005
Figure 6. Visualized results of different methods on the PROMISE12 test set. In the image, yellow represents the prostate, while the rest constitutes the background.
Figure 6. Visualized results of different methods on the PROMISE12 test set. In the image, yellow represents the prostate, while the rest constitutes the background.
Jimaging 10 00118 g006
Table 1. Different segmentation algorithms compare results on ACDC 5% and 10% labeled data.
Table 1. Different segmentation algorithms compare results on ACDC 5% and 10% labeled data.
Labeled
Data
MethodRVMyoRVMean
DSCHD95ASDDSCHD95ASDDSCHD95ASDDSCHD95ASD
5%MT [11]0.42560.116.80.58632.47.20.62546.311.50.54546.311.5
3/70EM [39]0.41540.511.50.57320.44.60.67125.86.20.55329.37.4
DAN [24]0.49244.617.80.52737.69.60.60138.08.00.54040.111.8
UAMT [40]0.41744.014.20.55729.56.80.61331.17.30.52934.99.4
ICT [21]0.43629.211.20.57321.65.50.62325.27.20.54425.38.0
URPC [41]0.38738.915.50.44125.77.20.54532.911.70.45832.511.5
CPS [25]0.42533.88.50.56920.24.60.65323.13.50.54925.76.2
CCT [10]0.46734.412.00.53918.94.70.63921.15.80.54824.87.5
DCT [27]0.37440.312.70.49422.56.00.55325.37.40.47320.41.7
RD [23]0.37636.011.80.43723.75.40.50126.26.30.43828.67.8
CTCT [26]0.67717.65.10.64212.63.00.75014.13.40.69014.73.9
Ours0.73414.83.50.73815.63.00.8329.32.10.76813.22.9
10%MT0.79115.52.70.76433.34.80.83220.03.90.79622.93.8
7/70EM0.7433.91.10.7987.81.40.84911.02.00.7977.61.5
DAN0.7998.81.40.7956.31.10.84511.62.10.8138.91.5
UAMT0.7728.31.30.79611.51.80.84915.72.70.80611.82.0
ICT0.8155.11.10.80910.71.60.85016.52.80.82510.81.8
URPC0.8178.71.90.8128.31,40.88611.72.30.8389.61.9
CPS0.8313.90.80.8266.61.30.87113.12.30.8437.91.5
CCT0.8375.10.90.8206.41.20.87811.31.80.8457.61.3
DCT0.7575.91.30.76236.15.80.85517.82.60.79219.93.2
RD0.8146.61.30.8107.41.20.86911.02.00.8318.41.5
CTCT0.8615.01.10.8416.31.00.89513.51.80.8668.31.3
Ours0.8886.81.30.8614.81.00.9216.51.40.8906.01.2
Table 2. Compare the results of different segmentation algorithms on Prostate 10% and 20% labeled data.
Table 2. Compare the results of different segmentation algorithms on Prostate 10% and 20% labeled data.
Labeled DataMethodMean
DSCHD95ASD
10% 4/35MT0.42494.525.7
EM0.49185.322.2
DAN0.56896.623.5
UAMT0.49091.226.7
ICT0.62365.916.2
URPC0.31765.124.5
CPS0.32460.015.8
CCT0.40957.321.4
DCT0.41085.624.3
RD0.43257.622.3
CTCT0.76425.27.8
Ours0.79221.27.3
20% 7/35MT0.63535.111.6
EM0.62041.613.2
DAN0.69564.913.5
UAMT0.63930.210.7
ICT0.73426.59.2
URPC0.64235.312.7
CPS0.60247.113.6
CCT0.57287.721.9
DCT0.65936.112.1
RD0.63339.612.8
CTCT0.78326.98.4
Ours0.81219.36.4
Table 3. Comparison of ACDC 10% labeled data network structure ablation results.
Table 3. Comparison of ACDC 10% labeled data network structure ablation results.
Method RV Myo LV Mean
DSCHD95ASDDSCHD95ASDDSCHD95ASDDSCHD95ASD
UNet0.67316.33.60.78510.01.70.87413.82.60.77813.82.6
UNet + Aux0.79115.52.70.76433.34.80.83220.03.90.79622.93.8
UNet + Tea0.8375.11.90.8206.41.20.87811.31.80.8457.61.3
UNet + ViT0.8615.01.10.8416.31.00.89513.51.80.8668.31.3
UNet + Tea + Aux0.80110.12.00.80810.21.70.87121.33.20.82613.92.3
UNet + Tea + ViT0.8707.41.50.8577.11.20.91311.72.00.8808.71.6
UNet + Aux + ViT0.8806.81.40.8604.11.00.9109.41.60.8846.81.3
DCCLNet0.8886.81.30.8614.81.00.9216.51.40.8906.01.2
Table 4. Comparison of Prostate 20% labeled data network structure ablation results.
Table 4. Comparison of Prostate 20% labeled data network structure ablation results.
MethodMean
DSCHD95ASD
UNet0.56395.324.6
UNet + Aux0.57287.721.9
UNet + Tea0.63535.111.6
UNet + ViT0.78326.98.4
UNet + Tea + Aux0.69533.210.1
UNet + Tea + ViT0.79424.67.9
UNet + Aux + ViT0.80721.07.2
DCCLNet0.81219.36.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, X.; Wang, W. Semi-Supervised Medical Image Segmentation Based on Deep Consistent Collaborative Learning. J. Imaging 2024, 10, 118. https://doi.org/10.3390/jimaging10050118

AMA Style

Zhao X, Wang W. Semi-Supervised Medical Image Segmentation Based on Deep Consistent Collaborative Learning. Journal of Imaging. 2024; 10(5):118. https://doi.org/10.3390/jimaging10050118

Chicago/Turabian Style

Zhao, Xin, and Wenqi Wang. 2024. "Semi-Supervised Medical Image Segmentation Based on Deep Consistent Collaborative Learning" Journal of Imaging 10, no. 5: 118. https://doi.org/10.3390/jimaging10050118

APA Style

Zhao, X., & Wang, W. (2024). Semi-Supervised Medical Image Segmentation Based on Deep Consistent Collaborative Learning. Journal of Imaging, 10(5), 118. https://doi.org/10.3390/jimaging10050118

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop