Scattering-Based Self-Supervised Learning for Label-Efficient Cardiac Image Segmentation
Abstract
1. Introduction
- We propose a scattering-based self-supervised learning (SSL) framework that integrates both Wavelet Scattering Networks (WSNs) and Parametric Scattering Networks (PSNs) into the initial stages of the SSL encoder. By replacing the early convolutional layers with scattering-based front-ends, the proposed design reduces the number of learnable parameters while preserving the expressive capacity of the deep encoder.
- We introduce a stability-aware representation learning strategy that explicitly exploits the translation invariance and deformation stability properties of scattering transforms within an SSL pipeline. By embedding these mathematically grounded priors into self-supervised pretraining, the proposed approach yields a more robust and data-efficient initialization for downstream cardiac image segmentation, particularly under limited supervision.
- Through extensive experiments on two cardiac imaging datasets with different modalities (cardiac cine MRI and cardiac CT), we demonstrate that the proposed scattering-based SSL framework consistently improves segmentation performance over random initialization and standard SSL baselines in label-scarce regimes. Among the evaluated variants, the PSN-based approach exhibits particularly strong performance in low-label settings, highlighting the benefit of combining structured scattering priors with a small number of learnable wavelet parameters for cardiac image segmentation.
2. Related Works
2.1. Self-Supervised Learning
2.2. Self-Supervised Learning for Semantic Segmentation
2.3. Wavelet Scattering Networks (WSNs)
3. Methods
3.1. Bootstrap Your Own Latent (BYOL)
3.2. Scattering-Based Feature Extraction
3.3. Parametric Scattering Networks (PSNs)
3.4. Image Segmentation with U-Net
3.5. Proposed Method
4. Experiments and Results
4.1. Dataset
4.2. Experimental Setup
4.2.1. Self-Supervised Pretraining
4.2.2. Segmentation Fine-Tuning
4.3. Evaluation Metrics
4.4. Results
4.4.1. Quantitative Results
4.4.2. Qualitative Results
5. Discussion
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015—Conference Track Proceedings, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
- Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar] [CrossRef]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the 29th Annual Conference on Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 7–12 December 2015; Volume 28. [Google Scholar]
- Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
- Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
- Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
- Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef]
- Jin, C.; Guo, Z.; Lin, Y.; Luo, L.; Chen, H. Label-efficient deep learning in medical image analysis: Challenges and future directions. arXiv 2023, arXiv:2303.12484. [Google Scholar] [CrossRef]
- Rayed, M.E.; Islam, S.M.S.; Niha, S.I.; Jim, J.R.; Kabir, M.M.; Mridha, M.F. Deep Learning for Medical Image Segmentation: State-of-the-Art Advancements and Challenges. Inform. Med. Unlocked 2024, 47, 101504. [Google Scholar] [CrossRef]
- Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
- Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 3320–3328. [Google Scholar]
- Oquab, M.; Bottou, L.; Laptev, I.; Sivic, J. Learning and Transferring Mid-Level Image Representations Using Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1717–1724. [Google Scholar]
- Zoetmulder, R.; Gavves, E.; Caan, M.; Marquering, H. Domain- and task-specific transfer learning for medical segmentation tasks. Comput. Methods Programs Biomed. 2021, 214, 106539. [Google Scholar] [CrossRef]
- Tendle, A.; Hasan, M.R. A study of the generalizability of self-supervised representations. Mach. Learn. Appl. 2021, 6, 100124. [Google Scholar] [CrossRef]
- Caron, M.; Bojanowski, P.; Joulin, A.; Douze, M. Deep Clustering for Unsupervised Learning of Visual Features. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 132–149. [Google Scholar]
- Noroozi, M.; Favaro, P. Unsupervised learning of visual representations by solving jigsaw puzzles. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 69–84. [Google Scholar]
- Gidaris, S.; Singh, P.; Komodakis, N. Unsupervised Representation Learning by Predicting Image Rotations. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
- Pathak, D.; Krahenbuhl, P.; Donahue, J.; Darrell, T.; Efros, A.A. Context encoders: Feature learning by inpainting. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2536–2544. [Google Scholar]
- Doersch, C.; Gupta, A.; Efros, A.A. Unsupervised visual representation learning by context prediction. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1422–1430. [Google Scholar]
- Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, Virtual, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
- Oquab, M.; Darcet, T.; Moutakanni, T.; Vo, H.; Szafraniec, M.; Khalidov, V.; Fernandez, P.; Haziza, D.; Massa, F.; El-Nouby, A.; et al. Dinov2: Learning robust visual features without supervision. arXiv 2023, arXiv:2304.07193. [Google Scholar]
- Caron, M.; Touvron, H.; Misra, I.; Jégou, H.; Mairal, J.; Bojanowski, P.; Joulin, A. Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 9650–9660. [Google Scholar]
- Zbontar, J.; Jing, L.; Misra, I.; LeCun, Y.; Deny, S. Barlow twins: Self-supervised learning via redundancy reduction. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 12310–12320. [Google Scholar]
- Grill, J.B.; Strub, F.; Altché, F.; Tallec, C.; Richemond, P.; Buchatskaya, E.; Doersch, C.; Avila Pires, B.; Guo, Z.; Gheshlaghi Azar, M.; et al. Bootstrap your own latent-a new approach to self-supervised learning. Adv. Neural Inf. Process. Syst. 2020, 33, 21271–21284. [Google Scholar]
- He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9729–9738. [Google Scholar]
- Bruna, J.; Mallat, S. Invariant scattering convolution networks. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1872–1886. [Google Scholar] [CrossRef] [PubMed]
- Jing, L.; Tian, Y. Self-Supervised Visual Feature Learning With Deep Neural Networks: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 4037–4058. [Google Scholar] [CrossRef]
- Ohri, K.; Kumar, M. Review on self-supervised image recognition using deep neural networks. Knowl.-Based Syst. 2021, 224, 107090. [Google Scholar] [CrossRef]
- Liu, X.; Zhang, F.; Hou, Z.; Mian, L.; Wang, Z.; Zhang, J.; Tang, J. Self-Supervised Learning: Generative or Contrastive. IEEE Trans. Knowl. Data Eng. 2021, 35, 857–876. [Google Scholar] [CrossRef]
- Albelwi, S. Survey on self-supervised learning: Auxiliary pretext tasks and contrastive learning methods in imaging. Entropy 2022, 24, 551. [Google Scholar] [CrossRef] [PubMed]
- Wu, Z.; Xiong, Y.; Yu, S.X.; Lin, D. Unsupervised feature learning via non-parametric instance discrimination. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3733–3742. [Google Scholar]
- Chen, X.; Fan, H.; Girshick, R.; He, K. Improved Baselines with Momentum Contrastive Learning. arXiv 2020, arXiv:2003.04297. [Google Scholar] [CrossRef]
- Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
- Hadsell, R.; Chopra, S.; LeCun, Y. Dimensionality reduction by learning an invariant mapping. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; Volume 2, pp. 1735–1742. [Google Scholar]
- Chen, X.; He, K. Exploring simple siamese representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 15750–15758. [Google Scholar]
- Wong, B.; Yi, M.Y. Premix: Label-Efficient Multiple Instance Learning via Non-Contrastive Pre-Training and Feature Mixing. arXiv 2024, arXiv:2408.01162. [Google Scholar]
- Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef]
- Hidy, G.; Bakos, B.; Lukács, A. Enhancing Pretraining Efficiency for Medical Image Segmentation via Transferability Metrics. arXiv 2024, arXiv:2410.18677. [Google Scholar] [CrossRef]
- Rani, V.; Kumar, M.; Gupta, A.; Sachdeva, M.; Mittal, A.; Kumar, K. Self-supervised learning for medical image analysis: A comprehensive review. Evol. Syst. 2024, 15, 1607–1633. [Google Scholar] [CrossRef]
- Zeng, D.; Wu, Y.; Hu, X.; Xu, X.; Yuan, H.; Huang, M.; Zhuang, J.; Hu, J.; Shi, Y. Positional contrastive learning for volumetric medical image segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Proceedings, Part II 24. pp. 221–230. [Google Scholar]
- Punn, N.S.; Agarwal, S. BT-Unet: A self-supervised learning framework for biomedical image segmentation using barlow twins with U-net models. Mach. Learn. 2022, 111, 4585–4600. [Google Scholar] [CrossRef]
- Kalapos, A.; Gyires-Tóth, B. Self-Supervised Pretraining for 2D Medical Image Segmentation. In Proceedings of the Computer Vision—ECCV 2022 Workshops, Tel Aviv, Israel, 23–27 October 2022; pp. 472–484. [Google Scholar]
- Wang, X.; Zhang, R.; Shen, C.; Kong, T.; Li, L. Dense contrastive learning for self-supervised visual pre-training. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 10–25 June 2021; pp. 3024–3033. [Google Scholar]
- Xie, Z.; Lin, Y.; Zhang, Z.; Cao, Y.; Lin, S.; Hu, H. Propagate Yourself: Exploring Pixel-Level Consistency for Unsupervised Visual Representation Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 16684–16693. [Google Scholar]
- Chaitanya, K.; Erdil, E.; Karani, N.; Konukoglu, E. Contrastive learning of global and local features for medical image segmentation with limited annotations. Adv. Neural Inf. Process. Syst. 2020, 33, 12546–12558. [Google Scholar]
- Mallat, S. Group Invariant Scattering. Commun. Pure Appl. Math. 2012, 65, 1331–1398. [Google Scholar] [CrossRef]
- Sifre, L.; Mallat, S. Rotation, Scaling and Deformation Invariant Scattering for Texture Discrimination. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 1233–1240. [Google Scholar]
- Leterme, H.; Polisano, K.; Perrier, V.; Alahari, K. On the Shift Invariance of Max Pooling Feature Maps in Convolutional Neural Networks. arXiv 2022, arXiv:2209.11740. [Google Scholar]
- Oyallon, E.; Belilovsky, E.; Zagoruyko, S. Scaling the Scattering Transform: Deep Hybrid Networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5618–5627. [Google Scholar]
- Oyallon, E.; Zagoruyko, S.; Huang, G.; Komodakis, N.; Lacoste-Julien, S.; Blaschko, M.B.; Belilovsky, E. Scattering Networks for Hybrid Representation Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 2208–2221. [Google Scholar] [CrossRef] [PubMed]
- Cotter, F.; Kingsbury, N. A Learnable ScatterNet: Locally Invariant Convolutional Layers. In Proceedings of the IEEE International Conference on Image Processing, Taipei, Taiwan, 22–25 September 2019; pp. 350–354. [Google Scholar]
- Gauthier, S.; Thérien, B.; Alsene-Racicot, L.; Chaudhary, M.; Rish, I.; Belilovsky, E.; Eickenberg, M.; Wolf, G. Parametric scattering networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5749–5758. [Google Scholar]
- Kinakh, V.; Taran, O.; Voloshynovskiy, S. ScatSimCLR: Self-Supervised Contrastive Learning with Pretext Task Regularization for Small-Scale Datasets. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Virtual, 11–17 October 2021; pp. 1098–1106. [Google Scholar]
- Bernard, O.; Lalande, A.; Zotti, C.; Cervenansky, F.; Yang, X.; Heng, P.A.; Cetin, I.; Lekadir, K.; Camara, O.; Ballester, M.A.G.; et al. Deep learning techniques for automatic MRI cardiac multi-structures segmentation and diagnosis: Is the problem solved? IEEE Trans. Med. Imaging 2018, 37, 2514–2525. [Google Scholar] [CrossRef]
- Xu, X.; Wang, T.; Shi, Y.; Yuan, H.; Jia, Q.; Huang, M.; Zhuang, J. Whole heart and great vessel segmentation in congenital heart disease using deep neural networks and graph matching. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer International Publishing: Cham, Switzerland, 2019; pp. 477–485. [Google Scholar]
- Janik, A.; Dodd, J.; Ifrim, G.; Sankaran, K.; Curran, K. Interpretability of a Deep Learning Model in the Application of Cardiac MRI Segmentation with an ACDC Challenge Dataset. In Proceedings of the Medical Imaging 2021: Image Processing, Virtual Event, 15–20 February 2021; pp. 861–872. [Google Scholar]
- Andreux, M.; Angles, T.; Exarchakis, G.; Leonarduzzi, R.; Rochette, G.; Thiry, L.; Zarka, J.; Mallat, S.; Andén, J.; Belilovsky, E.; et al. Kymatio: Scattering transforms in python. J. Mach. Learn. Res. 2020, 21, 1–6. [Google Scholar]
- You, Y.; Gitman, I.; Ginsburg, B. Large batch training of convolutional networks. arXiv 2017, arXiv:1708.03888. [Google Scholar] [CrossRef]
- Yagis, E.; Atnafu, S.W.; de Herrera, A.G.S.; Marzi, C.; Scheda, R.; Giannelli, M.; Tessa, C.; Citi, L.; Diciotti, S. Effect of data leakage in brain MRI classification using 2D convolutional neural networks. Sci. Rep. 2021, 11, 22544. [Google Scholar] [CrossRef] [PubMed]
- Li, H.; Ouyang, Y.; Wan, X. Self-Supervised Alignment Learning for Medical Image Segmentation. In Proceedings of the IEEE International Symposium on Biomedical Imaging, Athens, Greece, 27–30 May 2024; pp. 1–5. [Google Scholar]
- Seince, M.; Le Folgoc, L.; Facury de Souza, L.A.; Angelini, E. Dense Self-Supervised Learning for Medical Image Segmentation. arXiv 2024, arXiv:2407.20395. [Google Scholar] [CrossRef]
- Miron, R.; Moisii, C.; Dinu, S.; Breaban, M.E. Evaluating volumetric and slice-based approaches for COVID-19 detection in chest CTs. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 529–536. [Google Scholar]
- Ammar, A.; Bouattane, O.; Youssfi, M. Automatic cardiac cine MRI segmentation and heart disease classification. Comput. Med. Imaging Graph. 2021, 88, 101864. [Google Scholar] [CrossRef] [PubMed]
- Wen, Y.; Chen, L.; Deng, Y.; Zhou, C. Rethinking pre-training on medical imaging. J. Vis. Commun. Image Represent. 2021, 78, 103145. [Google Scholar] [CrossRef]
- Isensee, F.; Jaeger, P.F.; Kohl, S.A.A.; Petersen, J.; Maier-Hein, K.H. nnU-Net: A Self-Configuring Method for Deep Learning-Based Biomedical Image Segmentation. Nat. Methods 2021, 18, 203–211. [Google Scholar] [CrossRef] [PubMed]
- Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
- Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Deci-sion Support, Granada, Spain, 20 September 2018; pp. 3–11. [Google Scholar]







| Pretraining | M = 5 | M = 10 | M = 20 | M = 40 | M = 80 |
|---|---|---|---|---|---|
| Random Init. | 0.2949 ± 0.1364 | 0.6054 ± 0.0941 | 0.7955 ± 0.0291 | 0.8586 ± 0.0206 | 0.8891 ± 0.0075 |
| ImageNet | 0.5763 ± 0.1491 | 0.7993 ± 0.0280 | 0.8579 ± 0.0092 | 0.8825 ± 0.0056 | 0.8989 ± 0.0037 |
| BYOL | 0.7857 ± 0.0263 | 0.8435 ± 0.0097 | 0.8728 ± 0.0039 | 0.8863 ± 0.0039 | 0.8986 ± 0.0026 |
| BYOL + WSN | 0.8172 ± 0.0172 | 0.8596 ± 0.0107 | 0.8830 ± 0.0051 | 0.8971 ± 0.0053 | 0.9082 ± 0.0020 |
| BYOL + PSN | 0.8273 ± 0.0138 | 0.8646 ± 0.0096 | 0.8871 ± 0.0071 | 0.9010 ± 0.0045 | 0.9111 ± 0.0022 |
| Pretraining | M = 5 | M = 10 | M = 20 | M = 40 | M = 80 |
|---|---|---|---|---|---|
| Random Init | 0.4449 ± 0.1543 | 0.7008 ± 0.1115 | 0.8688 ± 0.0232 | 0.9141 ± 0.0118 | 0.9316 ± 0.0044 |
| ImageNet | 0.7679 ± 0.0655 | 0.8657 ± 0.0241 | 0.9094 ± 0.0091 | 0.9256 ± 0.0059 | 0.9370 ± 0.0034 |
| BYOL | 0.8725 ± 0.0274 | 0.9066 ± 0.0060 | 0.9235 ± 0.0057 | 0.9316 ± 0.0040 | 0.9380 ± 0.0023 |
| BYOL + WSN | 0.9003 ± 0.0116 | 0.9186 ± 0.0098 | 0.9311 ± 0.0043 | 0.9382 ± 0.0044 | 0.9450 ± 0.0022 |
| BYOL + PSN | 0.9041 ± 0.0096 | 0.9227 ± 0.0067 | 0.9336 ± 0.0059 | 0.9420 ± 0.0031 | 0.9472 ± 0.0011 |
| Pretraining | M = 5 | M = 10 | M = 20 | M = 40 | M = 80 |
|---|---|---|---|---|---|
| Random Init | 0.2872 ± 0.1979 | 0.5960 ± 0.1128 | 0.7957 ± 0.0296 | 0.8560 ± 0.0241 | 0.8921 ± 0.0091 |
| ImageNet | 0.5720 ± 0.2460 | 0.8023 ± 0.0376 | 0.8592 ± 0.0099 | 0.8858 ± 0.0055 | 0.9030 ± 0.0042 |
| BYOL | 0.7509 ± 0.0391 | 0.8244 ± 0.0190 | 0.8664 ± 0.0062 | 0.8825 ± 0.0082 | 0.8995 ± 0.0035 |
| BYOL + WSN | 0.7688 ± 0.0330 | 0.8392 ± 0.0167 | 0.8710 ± 0.0091 | 0.8915 ± 0.0102 | 0.9062 ± 0.0028 |
| BYOL + PSN | 0.7879 ± 0.0272 | 0.8432 ± 0.0175 | 0.8771 ± 0.0118 | 0.8958 ± 0.0088 | 0.9097 ± 0.0046 |
| Pretraining | M = 5 | M = 10 | M = 20 | M = 40 | M = 80 |
|---|---|---|---|---|---|
| Random Init | 0.1526 ± 0.1766 | 0.5192 ± 0.0841 | 0.7220 ± 0.0418 | 0.8057 ± 0.0270 | 0.8437 ± 0.0104 |
| ImageNet | 0.3890 ± 0.2831 | 0.7300 ± 0.0333 | 0.8052 ± 0.0138 | 0.8363 ± 0.0088 | 0.8567 ± 0.0055 |
| BYOL | 0.7338 ± 0.0306 | 0.7996 ± 0.0077 | 0.8284 ± 0.0058 | 0.8448 ± 0.0047 | 0.8584 ± 0.0039 |
| BYOL + WSN | 0.7824 ± 0.0155 | 0.8210 ± 0.0117 | 0.8469 ± 0.0051 | 0.8616 ± 0.0044 | 0.8734 ± 0.0037 |
| BYOL + PSN | 0.7898 ± 0.0120 | 0.8278 ± 0.0122 | 0.8507 ± 0.0081 | 0.8652 ± 0.0051 | 0.8763 ± 0.0023 |
| Pretraining | M = 2 | M = 5 | M = 10 | M = 20 | M = 40 |
|---|---|---|---|---|---|
| Random Init | 0.2182 ± 0.0748 | 0.2971 ± 0.0564 | 0.5149 ± 0.0392 | 0.5764 ± 0.0158 | 0.6395 ± 0.0064 |
| ImageNet | 0.3027 ± 0.0696 | 0.4567 ± 0.0372 | 0.5901 ± 0.0207 | 0.6372 ± 0.0116 | 0.6982 ± 0.0037 |
| BYOL | 0.3493 ± 0.0523 | 0.4692 ± 0.0785 | 0.5842 ± 0.0178 | 0.6440 ± 0.0118 | 0.6904 ± 0.0048 |
| BYOL + WSN | 0.3286 ± 0.0688 | 0.4733 ± 0.0795 | 0.5654 ± 0.0571 | 0.6498 ± 0.0079 | 0.6959 ± 0.0073 |
| BYOL + PSN | 0.3547 ± 0.0637 | 0.4673 ± 0.0707 | 0.5940 ± 0.0130 | 0.6451 ± 0.0159 | 0.6873 ± 0.0090 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Alasu, S.; Talu, M.F. Scattering-Based Self-Supervised Learning for Label-Efficient Cardiac Image Segmentation. Electronics 2026, 15, 506. https://doi.org/10.3390/electronics15030506
Alasu S, Talu MF. Scattering-Based Self-Supervised Learning for Label-Efficient Cardiac Image Segmentation. Electronics. 2026; 15(3):506. https://doi.org/10.3390/electronics15030506
Chicago/Turabian StyleAlasu, Serdar, and Muhammed Fatih Talu. 2026. "Scattering-Based Self-Supervised Learning for Label-Efficient Cardiac Image Segmentation" Electronics 15, no. 3: 506. https://doi.org/10.3390/electronics15030506
APA StyleAlasu, S., & Talu, M. F. (2026). Scattering-Based Self-Supervised Learning for Label-Efficient Cardiac Image Segmentation. Electronics, 15(3), 506. https://doi.org/10.3390/electronics15030506
