A Robust Semi-Supervised Brain Tumor MRI Classification Network for Data-Constrained Clinical Environments
Abstract
1. Introduction
- Labelled data is utilized to train a custom convolutional network model and a pretrained ResNet50 model as baseline supervised classifiers.
- A confidence-based iterative pseudo-labeling strategy is applied to unlabelled data, and semi-supervised training of SSPLNet is performed.
- For the final classification, a dense layer network is designed for feature fusion.
- A comprehensive comparative study is conducted by varying key parameters to evaluate model performance across different labelled–unlabelled distributions.
- To evaluate the improvement offered by the proposed semi-supervised approach, we conducted a paired t-test comparing the accuracy values of supervised (baseline) and semi-supervised learning across all labelled–unlabelled splits.
2. Related Work
3. Methodology
- 1.
- Supervised Loss: For labelled data, the standard cross-entropy loss is applied:
- 2.
- Confidence-Guided Pseudo-Labelling: For unlabelled data , pseudo-labels are generated only if the model’s softmax confidence exceeds a thresholdThe filtered set of pseudo-labelled samples are given as follows:
- 3.
- Unsupervised Pseudo-Label Loss: The pseudo-labels are treated as ground truth with cross-entropy loss:
- 4.
- Total Loss Function: The total training objective combines the supervised and unsupervised components:
- 5.
- Iterative Rejection Strategy: An iterative filtering mechanism is applied:
- -
- At each iteration , pseudo-labels are re-evaluated based on confidence.
- -
- Rejected samples (with low confidence) are excluded in future iterations.
- -
- The process repeats until convergence or max iterations .
Algorithm 1: SSPLNet—Semi-Supervised Pseudo-Labelling Network |
Input: Labelled dataset , Unlabelled dataset. , Neural network model , Confidence threshold , Loss weights: for pseudo-label loss, Max iterations for pseudo-label refinement Output: Trained model with updated parameters, Final labelled set including selected pseudo-labelled samples
|
4. Experiments
4.1. Dataset
4.2. Parameter Settings
4.3. Evaluation Parameters
4.4. SSPLNet: Fusing ResNet50 and Custom CNN
4.4.1. Input Modalities
4.4.2. Architecture Details
Pseudo-Labelling Framework with Dual Network Architecture
Deep Hybrid Model Design: ResNet50 and Custom CNN Integration
5. Results and Analysis
5.1. Pseudo-Labeling Strategy Using Custom CNN Model and Resnet50 Model
- Statistical Comparison of Supervised vs. Semi-Supervised Performance
- (a)
- In case of Custom CNN
- (b)
- In case of ResNet50
5.2. Semi-Supervised Pseudo-Labeling Network (SSPLNet)
5.3. Ablation Study: Effect of Classification Using Individual Models
5.4. Comparison with State-of-the-Art Techniques
5.5. Statistical Evaluation and Robustness Analysis of SSPLNet Performance
5.5.1. Confidence Interval Analysis of Model Performance
- = accuracy proportion (e.g., 0.9687 for 96.87%),
- = test sample size (1311),
- = 1.96 (critical value for 95% CI).
5.5.2. Quantifying Pseudo-Labelling Impact Through Effect Size for Custom CNN Only
5.5.3. Statistical Significance of Performance Improvements
- = cases incorrect in ResNet152V2 but correct in SSPLNet (30),
- = cases correct in ResNet152V2 but incorrect in SSPLNet (15).
- Results: = 4.36 (with Yates’ continuity correction), p-value = 0.037.
- Effect: SSPLNet made 15 fewer errors than ResNet152V2 in their disagreement cases.
5.5.4. Robustness Analysis via Bootstrapping
- Methodology
- For each bootstrap iteration i, the following steps occur:
- o
- Randomly sample 1311 test cases with replacement.
- o
- Calculate the model accuracy .
- o
- Store the accuracy value.
- The bootstrap distribution was then used to compute the following:
- Median accuracy,
- 95% percentile confidence interval (2.5th to 97.5th percentiles),
- Standard error: .
- Key Findings
- Narrow confidence intervals (CI width = 0.70%) indicate high stability.
- Minimal difference between median (98.21%) and mean (98.19%) suggests symmetric distribution.
- Class-specific CIs remain tight, with meningioma showing slightly wider bounds.
5.6. Additional Experiments with Pima Indians Diabetes Dataset
6. Discussion
7. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
MRI | Magnetic Resonance Imaging |
CNN | Convolutional Neural Network |
SSPLNet | Semi-Supervised Pseudo-Labeling Network |
SSL | Semi-Supervised Learning |
DNN | Deep Neural Network |
Acc | Accuracy |
References
- Sultan, H.H.; Salem, N.M.; Al-Atabany, W. Multi-classification of brain tumor images using deep neural network. IEEE Access 2019, 7, 19915–19923. [Google Scholar] [CrossRef]
- Zahoor, M.M.; Qureshi, S.A.; Bibi, S.; Khan, S.H.; Khan, A.; Ghafoor, U.; Bhutta, M.R. A new deep hybrid boosted and ensemble learning-based brain tumor analysis using MRI. Sensors 2022, 22, 2726. [Google Scholar] [CrossRef]
- Younis, A.; Qiang, L.; Nyatega, C.O.; Adamu, M.J.; Kawuwa, H.B. Brain tumor analysis using deep learning and VGG-16 ensembling learning approaches. Appl. Sci. 2022, 12, 7282. [Google Scholar] [CrossRef]
- Mahmud, M.I.; Mamun, M.; Abdelgawad, A. A deep analysis of brain tumor detection from MRI images using deep learning networks. Algorithms 2023, 16, 176. [Google Scholar] [CrossRef]
- Nassar, S.E.; Yasser, I.; Amer, H.M. A robust MRI based brain tumor classification via a hybrid deep learning technique. J. Supercomput. 2024, 80, 2403–2427. [Google Scholar] [CrossRef]
- ZainEldin, H.; Gamel, S.A.; El-Kenawy, E.-S.M.; Alharbi, A.H.; Khafaga, D.S.; Ibrahim, A.; Talaat, F.M. Brain tumor detection and classification using deep learning and sine-cosine fitness grey wolf optimization. Bioengineering 2023, 10, 18. [Google Scholar] [CrossRef]
- Abdusalomov, A.B.; Mukhiddinov, M.; Whangbo, T.K. Brain tumor detection based on deep learning approaches and magnetic resonance imaging. Cancers 2023, 15, 4172. [Google Scholar] [CrossRef]
- Mostafa, A.M.; Zakariah, M.; Aldakheel, E.A. Brain tumor segmentation using deep learning on MRI images. Diagnostics 2023, 13, 1562. [Google Scholar] [CrossRef] [PubMed]
- Mruthyunjaya, M.S.K. A brain tumor identification using convolution neural network and fully convolution neural network. In MATEC Web of Conferences; EDP Sciences: Les Ulis, France, 2024; Volume 392, p. 01130. [Google Scholar]
- Nadeem, M.B.; Ali, A.; Aziz, M.W.; Ghani, M.U.; Mustafa, G.; Farooq, A.B. Automated brain tumor detection via transfer learning techniques. J. Comput. Biomed. Inform. 2024, 7, 1. [Google Scholar]
- Oliver, A.; Odena, A.; Raffel, C.; Cubuk, E.D.; Goodfellow, I.J. Realistic evaluation of deep semi-supervised learning algorithms. arXiv 2018, arXiv:1804.09170. [Google Scholar]
- Chapelle, O.; Chi, M.; Zien, A. A continuation method for semi-supervised SVMs. In Proceedings of the 23rd International Conference on Machine Learning (ICML), Pittsburgh, PA, USA, 25–29 June 2006; pp. 185–192. [Google Scholar]
- Chapelle, O.; Schölkopf, B.; Zien, A. Semi-Supervised Learning, 1st ed.; MIT Press: Cambridge, MA, USA, 2006. [Google Scholar]
- Triguero, I.; García, S.; Herrera, F. Self-labelled techniques for semi-supervised learning: Taxonomy, software and empirical study. Knowl. Inf. Syst. 2015, 42, 245–284. [Google Scholar] [CrossRef]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Zeiler, M.D.; Fergus, R. Visualizing and understanding convolutional networks. In Proceedings of the European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014; pp. 818–833. [Google Scholar]
- Bengio, Y.; Simard, P.; Frasconi, P. Learning long-term dependencies with gradient descent is difficult. IEEE Trans. Neural Netw. 1994, 5, 157–166. [Google Scholar] [CrossRef]
- Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the 13th International Conference on Artificial Intelligence and Statistics (AISTATS), Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on ImageNet classification. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
- He, K.; Sun, J. Convolutional neural networks at constrained time cost. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 5353–5360. [Google Scholar]
- Ng, A. Sparse autoencoder. CS294A Lecture Notes 2011. Volume 72, pp. 1–19. Available online: https://web.stanford.edu/class/cs294a/sparseAutoencoder_2011new.pdf (accessed on 15 June 2025).
- Vincent, P.; Larochelle, H.; Lajoie, I.; Bengio, Y.; Manzagol, P.-A. Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion. J. Mach. Learn. Res. 2010, 11, 3371–3408. [Google Scholar]
- Radford, A.; Metz, L.; Chintala, S. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
- Mishra, A.; Jha, R.; Bhattacharjee, V. SSCLNet: A self-supervised contrastive loss-based pre-trained network for brain MRI classification. IEEE Access 2023, 11. [Google Scholar] [CrossRef]
- Mishra, A.; Bhattacharjee, V. Applying semi-supervised learning on human activity recognition data. In Proceedings of the International Conference on IoT and Blockchain Technology (ICIBT), Bhubaneswar, India, 26–28 May 2022; pp. 1–6. [Google Scholar]
- Kingma, D.P.; Welling, M. Auto-encoding variational Bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
- Xu, J.; Xiang, L.; Liu, Q.; Gilmore, H.; Wu, J.; Tang, J.; Madabhushi, A. Stacked sparse autoencoder (SSAE) for nuclei detection on breast cancer histopathology images. IEEE Trans. Med. Imaging 2016, 35, 119–130. [Google Scholar] [CrossRef] [PubMed]
- Chang, H.; Han, J.; Zhong, C.; Snijders, A.M.; Mao, J.-H. Unsupervised transfer learning via multi-scale convolutional sparse coding for biomedical applications. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 1182–1194. [Google Scholar] [CrossRef] [PubMed]
- Hou, L.; Nguyen, V.; Kanevsky, A.B.; Samaras, D.; Kurc, T.M.; Zhao, T.; Gupta, R.R.; Gao, Y.; Chen, W.; Foran, D. Sparse autoencoder for unsupervised nucleus detection and representation in histopathology images. Pattern Recognit. 2019, 86, 188–200. [Google Scholar] [CrossRef] [PubMed]
- Hu, B.; Tang, Y.; Chang, E.I.-C.; Fan, Y.; Lai, M.; Xu, Y. Unsupervised learning for cell-level visual representation in histopathology images with generative adversarial networks. IEEE J. Biomed. Health Inform. 2019, 23, 1316–1328. [Google Scholar] [CrossRef]
- Atha, Z.; Chaki, J. SSBTCNet: Semi-Supervised Brain Tumor Classification Network. IEEE Access 2023, 11. [Google Scholar] [CrossRef]
- Mijwil, M.M.; Doshi, R.; Hiran, K.K.; Unogwu, O.J.; Bala, I. MobileNetV1-based deep learning model for accurate brain tumor classification. Mesopotamian J. Comput. Sci. 2023, 2023, 32–41. [Google Scholar]
- Asif, S.; Zhao, M.; Tang, F.; Zhu, Y. An enhanced deep learning method for multi-class brain tumor classification using deep transfer learning. Multimed. Tools Appl. 2023, 82, 31709–31736. [Google Scholar] [CrossRef]
- Rezk, N.G.; Alshathri, S.; Sayed, A.; Hemdan, E.E.-D.; El-Behery, H. Secure Hybrid Deep Learning for MRI-Based Brain Tumor Detection in Smart Medical IoT Systems. Diagnostics 2025, 15, 639. [Google Scholar] [CrossRef]
- Ali, R.R.; Yaacob, N.M.; Alqaryouti, M.H.; Sadeq, A.E.; Doheir, M.; Iqtait, M.; Rachmawanto, E.H.; Sari, C.A.; Yaacob, S.S. Learning Architecture for Brain Tumor Classification Based on Deep Convolutional Neural Network: Classic and ResNet50. Diagnostics 2025, 15, 624. [Google Scholar] [CrossRef]
- Gupta, S.C.; Vijayvargiya, S.; Bhattacharjee, V. Role of feature diversity in the performance of hybrid models—An investigation of brain tumor classification from brain MRI scans. Diagnostics 2025, 15, 1863. [Google Scholar] [CrossRef]
- Jiao, R.; Zhang, Y.; Ding, L.; Cai, R.; Zhang, J. Learning with limited annotations: A survey on deep semi-supervised learning for medical image segmentation. Comput. Biol. Med. 2024, 202, 105. [Google Scholar] [CrossRef]
- Chen, Y. Semi-supervised and unsupervised deep visual learning: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 1632–1653. [Google Scholar] [CrossRef] [PubMed]
- Lee, J.; Cho, S. Semi-supervised image classification with Grad-CAM consistency. arXiv 2021, arXiv:2108.13673. Available online: https://arxiv.org/abs/2108.13673 (accessed on 16 February 2025).
- Gupta, S.C.; Goel, N. Predictive modeling and analytics for diabetes using hyperparameter tuned machine learning techniques. Procedia Comput. Sci. 2023, 218, 1257–1269. [Google Scholar] [CrossRef]
- Chang, V.; Bailey, J.; Xu, Q.A.; Sun, Z. Pima Indians diabetes mellitus classification based on machine learning (ML) algorithms. Neural Comput. Appl. 2023, 35, 16157–16173. [Google Scholar] [CrossRef] [PubMed]
- Tigga, N.P.; Garg, S. Prediction of type 2 diabetes using machine learning classification methods. Procedia Comput. Sci. 2020, 167, 706–716. [Google Scholar] [CrossRef]
- Brain Tumor MRI Dataset. Available online: https://www.kaggle.com/datasets/masoudnickparvar/brain-tumor-mri-dataset/data (accessed on 16 February 2025).
- Bishnu, P.S.; Bhattacherjee, V. Data Analysis: Using Statistics and Probability with R Language; PHI Learning: Delhi, India, 2019; pp. 1–592. ISBN 9789387472655. [Google Scholar]
- Efron, B.; Tibshirani, R.J. An Introduction to the Bootstrap; Chapman & Hall/CRC: Boca Raton, FL, USA, 1993. [Google Scholar]
- Pima-Indians-Diabetes. Available online: https://www.openml.org/search?type=data&status=active&id=43582&sort=runs (accessed on 10 June 2025).
S.No. | Labelled: Unlabelled Ratio (L%/U% of Training Data) | Supervised Acc (%) of Custom CNN Trained with L% Data) | Supervised Acc (%) of Resnet50 Trained with L% Data) | Semi-Supervised Acc (%) of Custom CNN Trained with L% Labelled + Pseudo-Labelled Subset of U% Unlabelled Data) | Semi-Supervised Acc (%) of Resnet50 Trained with L% Labelled + Pseudo-Labeled Subset of U% Unlabelled Data |
---|---|---|---|---|---|
1 | 90/10 | 98.81 | 99.08 | 99.08 | 98.70 |
2 | 80/20 | 98.09 | 98.62 | 98.24 | 98.62 |
3 | 70/30 | 97.88 | 97.63 | 98.15 | 98.01 |
4 | 60/40 | 96.56 | 97.77 | 97.17 | 97.71 |
5 | 50/50 | 95.19 | 96.26 | 96.10 | 96.26 |
6 | 40/60 | 94.93 | 95.19 | 95.34 | 97.02 |
7 | 30/70 | 92.4 | 94.05 | 93.21 | 94.81 |
8 | 20/80 | 88.9 | 92.75 | 90.08 | 94.43 |
9 | 10/90 | 84.74 | 84.50 | 86.04 | 86.57 |
S.No. | Split | Threshold (1) | Rejected (1) | Threshold (2) | Rejected (2) | Threshold (3) | Rejected (3) |
---|---|---|---|---|---|---|---|
1 | 90/10 | 0.98 | 10 | – | – | – | – |
2 | 80/20 | 0.98 | 40 | – | – | – | – |
3 | 70/30 | 0.98 | 70 | – | – | – | – |
4 | 60/40 | 0.98 | 92 | 0.95 | 19 | – | – |
5 | 50/50 | 0.98 | 117 | 0.95 | 21 | – | – |
6 | 40/60 | 0.98 | 212 | 0.95 | 39 | – | – |
7 | 30/70 | 0.98 | 278 | 0.95 | 62 | 0.90 | 20 |
8 | 20/80 | 0.98 | 485 | 0.95 | 89 | 0.90 | 13 |
9 | 10/90 | 0.98 | 1118 | 0.95 | 188 | 0.90 | 58 |
Split | Labelled | Unlabelled | Iterations (Threshold: Rejected) | Remarks |
---|---|---|---|---|
90–10 | 5143 | 569 | 0.98: 10 | No further iteration as retraining 10 samples had negligible effect. |
80–20 | 4571 | 1141 | 0.98: 40 | No further iteration as retraining 40 samples had negligible effect. |
70–30 | 4000 | 1712 | 0.98: 70 | No further iteration as retraining 70 samples had negligible effect. |
60–40 | 3429 | 2283 | 0.98: 92, 0.95: 19 | Two iterations were used due to moderate sample rejection. |
50–50 | 2858 | 2854 | 0.98: 117, 0.95: 21 | Similar trend with moderate rejections requiring two iterations. |
40–60 | 2286 | 3426 | 0.98: 212, 0.95: 39 | Two iterations were used to capture confident samples. |
30–70 | 1716 | 3999 | 0.98: 278, 0.95: 62, 0.90: 20 | Three iterations were used to adapt to large-scale rejection. |
20–80 | 1144 | 4568 | 0.98: 485, 0.95: 89, 0.90: 13 | Three iterations effectively filtered low-confidence predictions. |
10–90 | 573 | 5139 | 0.98: 1118, 0.95: 188, 0.90: 58 | Most aggressive use of three iterations to achieve convergence. |
S.No. | Labelled: Unlabelled Ratio (L/U) (Labeled + Pseudo-Labeled Merged Data) | Semi-Supervised Acc (%) Trained with L% Labelled + Pseudo-Labeled Subset of U% Unlabelled Data | Semi-Supervised Ac (%) Trained with L% Labelled + Pseudo-Labeled Subset of U% Unlabelled Data | |
---|---|---|---|---|
(Check Points) | ResNet50 | Custom CNN | Concatenated Model | |
1 | 90/10 | 98.70 | 99.08 | 98.39 |
2 | 80/20 | 98.62 | 98.24 | 98.48 |
3 | 70/30 | 98.01 | 98.15 | 98.77 |
4 | 60/40 | 97.71 | 97.17 | 98.17 |
5 | 50/50 | 96.26 | 96.10 | 96.03 |
6 | 40/60 | 97.02 | 95.34 | 96.87 |
7 | 30/70 | 94.81 | 93.21 | 95.18 |
8 | 20/80 | 94.43 | 90.08 | 88.48 |
9 | 10/90 | 86.57 | 86.04 | 90.08 |
S. No | Split | Initial Threshold | Rejected Samples (Iter 1) | Threshold (Iter 2) | Rejected Samples (Iter 2) | Threshold (Iter 3) | Rejected Samples (Iter 3) | Total Iterations |
---|---|---|---|---|---|---|---|---|
1. | 90/10 | 0.98 | 1 | - | - | - | - | 1 |
2. | 80/20 | 0.98 | 5 | - | - | - | - | 1 |
3. | 70/30 | 0.98 | 6 | - | - | - | - | 1 |
4. | 60/40 | 0.98 | 16 | - | - | - | - | 1 |
5. | 50/50 | 0.98 | 24 | - | - | - | - | 1 |
6. | 40/60 | 0.98 | 39 | - | - | - | - | 1 |
7. | 30/70 | 0.98 | 59 | - | - | - | - | 1 |
8. | 20/80 | 0.98 | 116 | 0.98 | 20 | - | - | 2 |
9. | 10/90 | 0.98 | 1150 | 0.98 | 236 | 0.98 | 89 | 3 |
Split | Class Name | Precision | Recall | F1-Score | Accuracy |
---|---|---|---|---|---|
90/10 | G | 0.99 | 0.97 | 0.98 | 0.98 |
M | 0.96 | 0.97 | 0.97 | ||
P | 0.99 | 0.99 | 0.99 | ||
N | 0.99 | 1.00 | 1.00 | ||
80/20 | G | 0.98 | 0.98 | 0.98 | 0.98 |
M | 0.97 | 0.97 | 0.97 | ||
P | 0.98 | 0.99 | 0.98 | ||
N | 1.00 | 1.00 | 1.00 | ||
70/30 | G | 0.98 | 0.99 | 0.99 | 0.99 |
M | 0.98 | 0.96 | 0.97 | ||
P | 0.99 | 1.00 | 0.99 | ||
N | 0.99 | 1.00 | 0.99 | ||
60/40 | G | 0.99 | 0.98 | 0.99 | 0.98 |
M | 0.97 | 0.95 | 0.96 | ||
P | 0.99 | 0.99 | 0.99 | ||
N | 0.97 | 0.99 | 0.98 | ||
50/50 | G | 0.98 | 0.96 | 0.97 | 0.96 |
M | 0.95 | 0.89 | 0.92 | ||
P | 0.98 | 0.99 | 0.99 | ||
N | 0.94 | 1.00 | 0.97 | ||
40/60 | G | 0.97 | 0.97 | 0.97 | 0.97 |
M | 0.96 | 0.92 | 0.94 | ||
P | 0.98 | 0.99 | 0.98 | ||
N | 0.97 | 0.99 | 0.98 | ||
30/70 | G | 0.96 | 0.95 | 0.95 | 0.95 |
M | 0.93 | 0.87 | 0.90 | ||
P | 0.98 | 0.99 | 0.98 | ||
N | 0.94 | 0.99 | 0.96 | ||
20/80 | G | 0.97 | 0.81 | 0.88 | 0.88 |
M | 0.83 | 0.74 | 0.78 | ||
P | 0.89 | 0.99 | 0.93 | ||
N | 0.87 | 0.97 | 0.92 | ||
10/90 | G | 0.95 | 0.94 | 0.94 | 0.90 |
M | 0.90 | 0.72 | 0.80 | ||
P | 0.90 | 0.99 | 0.94 | ||
N | 0.87 | 0.94 | 0.90 |
S. No | Labelled–Unlabelled Ratio | Supervised Accuracy | Semi-Supervised Accuracy | ||
---|---|---|---|---|---|
ResNet50 | Custom CNN | ResNet50 | Custom CNN | ||
1 | 90/10 | 99.08 | 98.81 | 98.70 | 99.08 |
2 | 80/20 | 98.62 | 98.09 | 98.62 | 98.24 |
3 | 70/30 | 97.63 | 97.88 | 98.01 | 98.15 |
4 | 60/40 | 97.77 | 96.56 | 97.71 | 97.17 |
5 | 50/50 | 96.26 | 95.19 | 96.26 | 96.10 |
6 | 40/60 | 95.19 | 94.93 | 97.02 | 95.34 |
7 | 30/70 | 94.05 | 92.4 | 94.81 | 93.21 |
8 | 20/80 | 92.75 | 88.9 | 94.43 | 90.08 |
9 | 10/90 | 84.50 | 84.74 | 86.57 | 86.04 |
S. No | Split | Initial Threshold | Rejected Samples (Iter 1) | Threshold (Iter 2) | Rejected Samples (Iter 2) | Threshold (Iter 3) | Rejected Samples (Iter 3) | Total Iterations |
---|---|---|---|---|---|---|---|---|
1. | 90/10 | 0.98 | 0 | - | - | - | - | 1 |
2. | 80/20 | 0.98 | 11 | - | - | - | - | 1 |
3. | 70/30 | 0.98 | 5 | - | - | - | - | 1 |
4. | 60/40 | 0.98 | 15 | - | - | - | - | 1 |
5. | 50/50 | 0.98 | 14 | - | - | - | - | 1 |
6. | 40/60 | 0.98 | 35 | - | - | - | - | 1 |
7. | 30/70 | 0.98 | 72 | 0.98 | 8 | - | - | 2 |
8. | 20/80 | 0.98 | 117 | 0.98 | 13 | - | - | 2 |
9. | 10/90 | 0.98 | 720 | 0.98 | 87 | 0.98 | 26 | 3 |
Ref. and Technique | Precision | Recall | Specificity | F1-Score | Accuracy |
---|---|---|---|---|---|
MobileNet [33] | 94 | 97 | 96.1 | 95 | 97 |
XceptionNet [34] | 94.98 | 95.41 | 97.82 | 95.63 | 95.63 |
DenseNet201 [34] | 94.36 | 95.05 | 97.60 | 95.10 | 95.10 |
DensetNet121 [34] | 92.87 | 93.26 | 97.10 | 93.33 | 93.98 |
ResNet152V2 [34] | 95.23 | 96.01 | 98.14 | 96.65 | 96.65 |
InceptionResNetV2 [34] | 91.94 | 91.57 | 96.23 | 92.02 | 92.02 |
SSBTCNet (45% Unlabelled) [32] | - | 91.8 | 97.3 | - | 96.5 |
Proposed SSPLNet (50% Unlabelled) | 96.25 | 96 | - | 96.25 | 96.03 |
Proposed SSPLNet (40% Unlabelled) | 98 * | 97.75 * | - | 98 * | 98.17 * |
Split (L/U) | Accuracy (%) | 95% CI Lower | 95% CI Upper | Standard Error |
---|---|---|---|---|
90/10 | 98.39 | 97.71 | 99.07 | 0.00348 |
40/60 | 96.87 | 95.93 | 97.81 | 0.00481 |
10/90 | 90.08 | 88.46 | 91.70 | 0.00825 |
Split (L/U) | Supervised Mean (SD) | SSL Mean (SD) | Cohen’s d | Magnitude |
---|---|---|---|---|
90/10 | 98.81 (0.19) | 99.08 (0.12) | 1.42 | Very Large |
50/50 | 95.19 (1.03) | 96.10 (0.92) | 0.93 | Large |
10/90 | 84.74 (1.42) | 86.04 (1.58) | 0.87 | Large |
ResNet152v2 Correct | ResNet152v2 Incorrect | Total | |
---|---|---|---|
SSPLNet Correct | 1265 | 30 | 1295 |
SSPLNet Incorrect | 15 | 1 | 16 |
Total | 1280 | 31 | 1311 |
Metric | ResNet152V2 | SSPLNet | Statistical Test | Value | p-Value |
---|---|---|---|---|---|
Misclassification Rate | 2.36% | 1.22% | McNemar’s | 4.36 | 0.037 |
Agreement Rate | 96.6% | Cohen’s k | 0.89 | <0.001 |
Metric | Value | 95% CI | SE (Bootstrap) |
---|---|---|---|
Overall Accuracy | 98.21% | [97.82%, 98.52%] | 0.18% |
Glioma Class | 98.05% | [97.31%, 98.67%] | 0.35% |
Meningioma Class | 97.88% | [96.94%, 98.52%] | 0.41% |
Dataset Split Ratio: (Labelled%/Unlabelled%) | 90/10 | 80/20 | 70/30 | 60/40 | 50/50 |
---|---|---|---|---|---|
Accuracy (%) | 74.68 | 77.27 | 75.97 | 74.68 | 75.97 |
S.No. | References | Dataset Split Ratio: Labeled%/Unlabelled% | Accuracy (%) |
---|---|---|---|
1. | Gupta et al. [41] | 100% Labelled | 80.52 |
2. | Chang et al. [42] | 100% labelled | 75.0 |
3. | Tigga et al. [43] | 100% labelled | 79.97 |
4. | Proposed work | 80%/20% split | 77.27 |
5. | Proposed work | 50%/50% split | 75.97 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Gupta, S.C.; Bhattacharjee, V.; Vijayvargiya, S.; Bishnu, P.S.; Oraon, R.; Majhi, R. A Robust Semi-Supervised Brain Tumor MRI Classification Network for Data-Constrained Clinical Environments. Diagnostics 2025, 15, 2485. https://doi.org/10.3390/diagnostics15192485
Gupta SC, Bhattacharjee V, Vijayvargiya S, Bishnu PS, Oraon R, Majhi R. A Robust Semi-Supervised Brain Tumor MRI Classification Network for Data-Constrained Clinical Environments. Diagnostics. 2025; 15(19):2485. https://doi.org/10.3390/diagnostics15192485
Chicago/Turabian StyleGupta, Subhash Chand, Vandana Bhattacharjee, Shripal Vijayvargiya, Partha Sarathi Bishnu, Raushan Oraon, and Rajendra Majhi. 2025. "A Robust Semi-Supervised Brain Tumor MRI Classification Network for Data-Constrained Clinical Environments" Diagnostics 15, no. 19: 2485. https://doi.org/10.3390/diagnostics15192485
APA StyleGupta, S. C., Bhattacharjee, V., Vijayvargiya, S., Bishnu, P. S., Oraon, R., & Majhi, R. (2025). A Robust Semi-Supervised Brain Tumor MRI Classification Network for Data-Constrained Clinical Environments. Diagnostics, 15(19), 2485. https://doi.org/10.3390/diagnostics15192485