CHARMS: A CNN-Transformer Hybrid with Attention Regularization for MRI Super-Resolution
Highlights
- CHARMS, a lightweight CNN-Transformer hybrid (~1.9 M parameters, ~30 GFLOPs), outperforms state-of-the-art lightweight MRI super-resolution models (EDSR, PAN, W2AMSN-S, and FMEN) by 0.1–0.6 dB PSNR and up to 1% SSIM at ×2/×4 upscaling while reducing inference time by ~40%.
- With cross-field fine-tuning on only twenty subjects, CHARMS upgrades clinical 3T MRI to near-7T quality, yielding ~6 dB PSNR and 0.12 SSIM gains over native 3T scans across T1w/T2w contrasts.
- Near-real-time performance (~11 ms/slice enabling ~1.6–1.9 s processing per 3D brain volume on RTX 4090) and small model size enable practical deployment in clinical workstations, online reconstruction pipelines, and resource-constrained environments including low-field and portable MRI scanners.
- Superior fidelity–efficiency balance paves the way for shorter scan times, reduced motion artifacts, and 7T-like diagnostic quality from standard 3T systems without additional hardware.
Abstract
1. Introduction
2. Related Work
2.1. CNN-Based MRI Super-Resolution
2.2. Attention Mechanisms in SR
2.3. Transformer and Hybrid Architectures
2.4. Diffusion Models and Emerging Trends
3. Materials and Methods
3.1. CHARMS Framework
3.2. Reverse Residual Attention Fusion (RRAF) Block
3.3. Pixel–Channel Attention (PCA) Module
3.4. Transformer Module with MDDTA and GDDFN
3.5. High-Frequency Information Refinement (HFIR)
3.6. Datasets and Preprocessing
3.7. Training Protocol and Comparative Models
3.8. Cross-Field Adaptation and Evaluation Procedure
4. Results
4.1. Benchmark Performance
4.2. Ablation Study
4.3. Cross-Field Validation Using Paired 3T/7T Datasets
5. Discussion
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Appendix A. Glossary of Key Acronyms
| CHARMS | Full Name | Brief Description |
| CHARMS | CNN-Transformer Hybrid with Attention Regularization for MRI Super-Resolution | The proposed lightweight model for MRI super-resolution, combining CNN and Transformer elements with attention regularization. |
| RRAF | Reverse Residual Attention Fusion | Backbone module for hierarchical local feature extraction, integrating residual learning and attention. |
| RLFE | Residual Local Feature Extraction | Units within RRAF blocks, consisting of convolutions, ReLU activations, and ESA for feature encoding. |
| ESA | Enhanced Spatial Attention | Spatial attention operator that highlights high-frequency regions using dilated convolutions. |
| PCA | Pixel–Channel Attention | Mechanism merging pixel- and channel-level attention for fine-grained feature recalibration. |
| MDDTA | Multi-Depthwise Dilated Transformer Attention | Transformer block for efficient long-range dependency modeling with linear complexity. |
| GDDFN | Gated Depthwise Dilated Feed-Forward Network | Feed-forward component in the Transformer module, enhancing nonlinearity via gated convolutions. |
| THFIR | High-Frequency Information Refinement | Refinement module post-upsampling to restore high-frequency details and suppress artifacts. |
References
- Li, Y.; Sixou, B.; Peyrin, F. A Review of the Deep Learning Methods for Medical Images Super Resolution Problems. IRBM 2021, 42, 120–133. [Google Scholar] [CrossRef]
- Yang, H.; Wang, Z.; Liu, X.; Li, C.; Xin, J.; Wang, Z. Deep learning in medical image super resolution: A review. Appl. Intell. 2023, 53, 20891–20916. [Google Scholar] [CrossRef]
- Ji, Z.; Zou, B.; Kui, X.; Liu, J.; Zhao, W.; Zhu, C.; Dai, P.; Dai, Y. Deep learning-based magnetic resonance image super-resolution: A survey. Neural Comput. Appl. 2024, 36, 12725–12752. [Google Scholar] [CrossRef]
- Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a Deep Convolutional Network for Image Super-Resolution. In Computer Vision—ECCV 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2014; pp. 184–199. [Google Scholar]
- Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
- Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced Deep Residual Networks for Single Image Super-Resolution. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 1132–1140. [Google Scholar]
- Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image Super-Resolution Using Very Deep Residual Channel Attention Networks. In Computer Vision—ECCV 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2018; pp. 294–310. [Google Scholar]
- Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar] [CrossRef]
- Zhang, X.; Zhou, X.; Lin, M.; Sun, J.; Recognition, P. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6848–6856. [Google Scholar]
- Mehta, S.; Rastegari, M.; Caspi, A.; Shapiro, L.; Hajishirzi, H. ESPNet: Efficient Spatial Pyramid of Dilated Convolutions for Semantic Segmentation; Springer International Publishing: Cham, Switzerland, 2018; pp. 561–580. [Google Scholar]
- Wang, Z.; Xie, X.; Yang, J.; Song, X. RA-Net: Reverse attention for generalizing residual learning. Sci. Rep. 2024, 14, 12771. [Google Scholar] [CrossRef] [PubMed]
- Zhao, H.; Kong, X.; He, J.; Qiao, Y.; Dong, C. Efficient Image Super-Resolution Using Pixel Attention. In Proceedings of the ECCV Workshops, Glasgow, UK, 23–28 August 2020. [Google Scholar]
- Liu, J.; Zhang, W.; Tang, Y.; Tang, J.; Wu, G. Residual Feature Aggregation Network for Image Super-Resolution. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 2359–2368. [Google Scholar]
- Shen, L.; Qin, F.; Zhang, Y.; Wang, Q.; Hu, B.; Zhao, W. MATNet: MRI Super-Resolution with Multiple Attention Mechanisms. In Proceedings of the 2023 International Conference on New Trends in Computational Intelligence (NTCI), Qingdao, China, 3–5 November 2023; Volume 1, pp. 171–180. [Google Scholar]
- Wood, D.A.; Kafiabadi, S.; Busaidi, A.A.; Guilhem, E.; Montvila, A.; Lynch, J.; Townend, M.; Agarwal, S.; Mazumder, A.; Barker, G.J.; et al. Accurate brain-age models for routine clinical MRI examinations. Neuroimage 2022, 249, 118871. [Google Scholar] [CrossRef] [PubMed]
- Glasser, M.F.; Smith, S.M.; Marcus, D.S.; Andersson, J.L.; Auerbach, E.J.; Behrens, T.E.; Coalson, T.S.; Harms, M.P.; Jenkinson, M.; Moeller, S.; et al. The Human Connectome Project’s neuroimaging approach. Nat. Neurosci. 2016, 19, 1175–1187. [Google Scholar] [CrossRef]
- Chu, L.; Ma, B.; Dong, X.; He, Y.; Che, T.; Zeng, D.; Zhang, Z.; Li, S. A paired dataset of multi-modal MRI at 3 Tesla and 7 Tesla with manual hippocampal subfield segmentations. Sci. Data 2025, 12, 260. [Google Scholar] [CrossRef]
- Chen, X.; Qu, L.; Xie, Y.; Ahmad, S.; Yap, P.T. A paired dataset of T1- and T2-weighted MRI at 3 Tesla and 7 Tesla. Sci. Data 2023, 10, 489. [Google Scholar] [CrossRef]
- Chen, J.; Sasaki, H.; Lai, H.; Su, Y.; Liu, J.; Wu, Y.; Zhovmer, A.; Combs, C.A.; Rey-Suarez, I.; Chang, H.Y.; et al. Three-dimensional residual channel attention networks denoise and sharpen fluorescence microscopy image volumes. Nat. Methods 2021, 18, 678–687. [Google Scholar] [CrossRef]
- Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.-S.J.A. CBAM: Convolutional Block Attention Module. arXiv 2018, arXiv:1807.06521. [Google Scholar] [CrossRef]
- Liu, J.; Tang, J.; Wu, G. Residual Feature Distillation Network for Lightweight Image Super-Resolution; Springer International Publishing: Cham, Switzerland, 2020; pp. 41–55. [Google Scholar]
- Liu, J.-J.; Hou, Q.; Cheng, M.-M.; Wang, C.; Feng, J. Improving Convolutional Networks with Self-Calibrated Convolutions. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10093–10102. [Google Scholar]
- Vaswani, A.; Shazeer, N.M.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All you Need. In Proceedings of the Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
- Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
- Li, G.; Lv, J.; Tian, Y.; Dou, Q.; Wang, C.; Xu, C.; Qin, J. Transformer-empowered Multi-scale Contextual Matching and Aggregation for Multi-contrast MRI Super-resolution. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 20604–20613. [Google Scholar]
- Mahapatra, D.; Ge, Z. MR Image Super Resolution by Combining Feature Disentanglement CNNs and Vision Transformers. In Proceedings of the 5th International Conference on Medical Imaging with Deep Learning, Zurich, Switzerland, 6–8 July 2022; Ender, K., Bjoern, M., Archana, V., Christian, B., Qi, D., Shadi, A., Eds.; PMLR: Proceedings of Machine Learning Research: Cambridge, MA, USA, 2022; Volume 172, pp. 858–878. [Google Scholar]
- Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H. Restormer. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, 18–24 June 2022; IEEE Computer Society: Washington, DC, USA, 2022; pp. 5718–5729. [Google Scholar]
- Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. SwinIR: Image Restoration Using Swin Transformer. In Proceedings of the ICCVW, Montreal, BC, Canada, 11–17 October 2021; IEEE: New York, NY, USA, 2021; pp. 1833–1844. [Google Scholar]
- Liu, Y.; Zhang, M.; Zhang, W.; Hou, B.; Liu, D.; Lian, H.; Jiang, B. Flexible Alignment Super-Resolution Network for Multi-Contrast MRI. arXiv 2022, arXiv:2210.03460. [Google Scholar]
- Wang, J.; Zou, Y.; Wu, H. Image super-resolution method based on attention aggregation hierarchy feature. Vis. Comput. 2023, 40, 2655–2666. [Google Scholar] [CrossRef]
- Lyu, J.; Li, G.; Wang, C.; Cai, Q.; Dou, Q.; Zhang, D.; Qin, J. Multicontrast MRI Super-Resolution via Transformer-Empowered Multiscale Contextual Matching and Aggregation. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 12004–12014. [Google Scholar] [CrossRef]
- Huang, S.; Liu, X.; Tan, T.; Hu, M.; Wei, X.; Chen, T.; Sheng, B. TransMRSR: Transformer-based self-distilled generative prior for brain MRI super-resolution. Vis. Comput. 2023, 39, 3647–3659. [Google Scholar] [CrossRef]
- He, K.; Gan, C.; Li, Z.; Rekik, I.; Yin, Z.; Ji, W.; Gao, Y.; Wang, Q.; Zhang, J.; Shen, D. Transformers in medical image analysis. Intell. Med. 2023, 3, 59–78. [Google Scholar] [CrossRef]
- Meng, F.; Guo, Y.; He, W.; Xu, Z. Ultra-Low Field Magnetic Resonance Image Enhancement based on Deep-Learning Method. In Proceedings of the 2023 2nd International Conference on Health Big Data and Intelligent Healthcare (ICHIH), Zhuhai, China, 27–29 October 2023; pp. 39–42. [Google Scholar]
- Bischoff, L.M.; Peeters, J.M.; Weinhold, L.; Krausewitz, P.; Ellinger, J.; Katemann, C.; Isaak, A.; Weber, O.M.; Kuetting, D.; Attenberger, U.; et al. Deep Learning Super-Resolution Reconstruction for Fast and Motion-Robust T2-weighted Prostate MRI. Radiology 2023, 308, e230427. [Google Scholar] [CrossRef]
- Yang, G.; Zhang, L.; Zhou, M.; Liu, A.; Chen, X.; Xiong, Z.; Wu, F. Model-Guided Multi-Contrast Deep Unfolding Network for MRI Super-resolution Reconstruction. In Proceedings of the 30th ACM International Conference on Multimedia, Lisbon, Portugal, 10–14 October 2022; pp. 3974–3982. [Google Scholar]
- Wang, H.; Hu, X.; Zhao, X.; Zhang, Y. Wide Weighted Attention Multi-Scale Network for Accurate MR Image Super-Resolution. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 962–975. [Google Scholar] [CrossRef]
- Zhao, C.; Dewey, B.E.; Pham, D.L.; Calabresi, P.A.; Reich, D.S.; Prince, J.L. SMORE: A Self-Supervised Anti-Aliasing and Super-Resolution Algorithm for MRI Using Deep Learning. IEEE Trans. Med. Imaging 2021, 40, 805–817. [Google Scholar] [CrossRef]
- Rakotonirina, N.C.; Rasoanaivo, A. ESRGAN+: Further Improving Enhanced Super-Resolution Generative Adversarial Network. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 3637–3641. [Google Scholar]
- Li, Y.; Yang, M.; Bian, T.; Wu, H. MRI Super-Resolution Analysis via MRISR: Deep Learning for Low-Field Imaging. Information 2024, 15, 655. [Google Scholar] [CrossRef]
- Wang, R.; Cao, Z.; He, Y.; Liu, J.; Shi, F.; Shen, D. Clinical Brain MRI Super-Resolution with 2D Slice-Wise Diffusion Model; Springer Nature: Cham, Switzerland, 2025; pp. 166–176. [Google Scholar]
- Zhao, K.; Pang, K.; Hung, A.L.Y.; Zheng, H.; Yan, R.; Sung, K. MRI Super-Resolution With Partial Diffusion Models. IEEE Trans. Med. Imaging 2025, 44, 1194–1205. [Google Scholar] [CrossRef] [PubMed]
- Safari, M.; Wang, S.; Eidex, Z.; Li, Q.; Qiu, R.L.J.; Middlebrooks, E.H.; Yu, D.S.; Yang, X. MRI super-resolution reconstruction using efficient diffusion probabilistic model with residual shifting. Phys. Med. Biol. 2025, 70, 125008. [Google Scholar] [CrossRef]
- Chen, W.; Wu, S.; Wang, S.; Li, Z.; Yang, J.; Yao, H.; Tian, Q.; Song, X. Multi-contrast image super-resolution with deformable attention and neighborhood-based feature aggregation (DANCE): Applications in anatomic and metabolic MRI. Med. Image Anal. 2025, 99, 103359. [Google Scholar] [CrossRef] [PubMed]
- Yoon, D.; Myong, Y.; Kim, Y.G.; Sim, Y.; Cho, M.; Oh, B.M.; Kim, S. Latent diffusion model-based MRI superresolution enhances mild cognitive impairment prognostication and Alzheimer’s disease classification. Neuroimage 2024, 296, 120663. [Google Scholar] [CrossRef]
- Georgescu, M.-I.; Ionescu, R.T.; Miron, A.-I.; Savencu, O.; Ristea, N.-C.; Verga, N.; Khan, F.S. Multimodal Multi-Head Convolutional Attention with Various Kernel Sizes for Medical Image Super-Resolution. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; IEEE: New York, NY, USA, 2023. [Google Scholar]
- Wang, Y.; Hu, H.; Yu, S.; Yang, Y.; Guo, Y.; Song, X.; Chen, F.; Liu, Q. A unified hybrid transformer for joint MRI sequences super-resolution and missing data imputation. Phys. Med. Biol. 2023, 68, 135006. [Google Scholar] [CrossRef] [PubMed]
- Zhang, M.; Xia, C.; Tang, J.; Yao, L.; Hu, N.; Li, J.; Peng, W.; Hu, S.; Ye, Z.; Zhang, X.; et al. Evaluation of high-resolution pituitary dynamic contrast-enhanced MRI using deep learning-based compressed sensing and super-resolution reconstruction. Eur. Radiol. 2025, 35, 5922–5934. [Google Scholar] [CrossRef]
- Xiao, H.; Yang, Z.; Liu, T.; Liu, S.; Huang, X.; Dai, J. Deep learning for medical imaging super-resolution: A comprehensive review. Neurocomputing 2025, 630, 129667. [Google Scholar] [CrossRef]
- Kravchenko, D.; Isaak, A.; Mesropyan, N.; Peeters, J.M.; Kuetting, D.; Pieper, C.C.; Katemann, C.; Attenberger, U.; Emrich, T.; Varga-Szemes, A.; et al. Deep learning super-resolution reconstruction for fast and high-quality cine cardiovascular magnetic resonance. Eur. Radiol. 2025, 35, 2877–2887. [Google Scholar] [CrossRef]
- Chatterjee, S.; Sciarra, A.; Dünnwald, M.; Ashoka, A.B.T.; Vasudeva, M.G.C.; Saravanan, S.; Sambandham, V.T.; Tummala, P.; Oeltze-Jafra, S.; Speck, O.; et al. Beyond Nyquist: A Comparative Analysis of 3D Deep Learning Models Enhancing MRI Resolution. J. Imaging 2024, 10, 207. [Google Scholar] [CrossRef]
- Yang, Q.; Zhang, Y.; Chandler, D.M.; Farias, M.C.Q. SSRT: Intra- and cross-view attention for stereo image super-resolution. Multimed. Tools Appl. 2025, 84, 22917–22945. [Google Scholar] [CrossRef]
- Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar]
- Singh, V.; Ramnath, K.; Mittal, A. Refining high-frequencies for sharper super-resolution and deblurring. Comput. Vis. Image Underst. 2020, 199, 103034. [Google Scholar] [CrossRef]
- Pham, C.-H.; Tor-Díez, C.; Meunier, H.; Bednarek, N.; Fablet, R.; Passat, N.; Rousseau, F. Multiscale brain MRI super-resolution using deep 3D convolutional networks. Comput. Med. Imaging Graph. 2019, 77, 101647. [Google Scholar] [CrossRef]
- Kong, F.; Li, M.; Liu, S.; Liu, D.; He, J.; Bai, Y.; Chen, F.; Fu, L. Residual Local Feature Network for Efficient Super-Resolution. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA, 19–20 June 2022; pp. 765–775. [Google Scholar]
- Tustison, N.J.; Avants, B.B.; Cook, P.A.; Zheng, Y.; Egan, A.; Yushkevich, P.A.; Gee, J.C. N4ITK: Improved N3 bias correction. IEEE Trans. Med. Imaging 2010, 29, 1310–1320. [Google Scholar] [CrossRef] [PubMed]
- Hoopes, A.; Mora, J.S.; Dalca, A.V.; Fischl, B.; Hoffmann, M. SynthStrip: Skull-stripping for any brain image. Neuroimage 2022, 260, 119474. [Google Scholar] [CrossRef] [PubMed]
- Loshchilov, I.; Hutter, F. Fixing Weight Decay Regularization in Adam. arXiv 2017, arXiv:1711.05101. [Google Scholar]
- Duchi, J.C.; Hazan, E.; Singer, Y. Adaptive Subgradient Methods for Online Learning and Stochastic Optimization. J. Mach. Learn. Res. 2011, 12, 2121–2159. [Google Scholar]
- Smith, L.N. Cyclical Learning Rates for Training Neural Networks. In Proceedings of the 2017 IEEE Winter Conference on Applications of Computer Vision (WACV), Santa Rosa, CA, USA, 24–31 March 2017; pp. 464–472. [Google Scholar]
- Masters, D.; Luschi, C. Revisiting Small Batch Training for Deep Neural Networks. arXiv 2018, arXiv:1804.07612. [Google Scholar] [CrossRef]
- Bengio, Y. Practical Recommendations for Gradient-Based Training of Deep Architectures. In Proceedings of the Neural Networks, Shenyang, China, 11–14 July 2012. [Google Scholar]
- Zou, W.; Ye, T.; Zheng, W.; Zhang, Y.; Chen, L.; Wu, Y. Self-Calibrated Efficient Transformer for Lightweight Super-Resolution. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA, 19–20 June 2022; pp. 929–938. [Google Scholar]
- Du, Z.; Liu, D.; Liu, J.; Tang, J.; Wu, G.; Fu, L. Fast and Memory-Efficient Network Towards Efficient Image Super-Resolution. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA, 19–20 June 2022; pp. 852–861. [Google Scholar]
- Horé, A.; Ziou, D. Image Quality Metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar]
- Zhou, W.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
- Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How Transferable Are Features in Deep Neural Networks? In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
- Tajbakhsh, N.; Shin, J.Y.; Gurudu, S.R.; Hurst, R.T.; Kendall, C.B.; Gotway, M.B.; Liang, J. Convolutional Neural Networks for Medical Image Analysis: Full Training or Fine Tuning? IEEE Trans. Med. Imaging 2016, 35, 1299–1312. [Google Scholar] [CrossRef]
- Zhang, Y.; Li, K.; Li, K.; Fu, Y.R. MR Image Super-Resolution with Squeeze and Excitation Reasoning Attention Network. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; pp. 13420–13429. [Google Scholar]
- Feng, C.-M.; Yan, Y.; Fu, H.; Chen, L.; Xu, Y. Task Transformer Network for Joint MRI Reconstruction and Super-Resolution. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2021; de Bruijne, M., Cattin, P.C., Cotin, S., Padoy, N., Speidel, S., Zheng, Y., Essert, C., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2021; pp. 307–317. [Google Scholar]
- Georgescu, M.-I.; Ionescu, R.T.; Verga, N. Convolutional Neural Networks with Intermediate Loss for 3D Super-Resolution of CT and MRI Scans. IEEE Access 2020, 8, 49112–49124. [Google Scholar] [CrossRef]
- Zhao, X.; Zhang, Y.; Zhang, T.; Zou, X. Channel Splitting Network for Single MR Image Super-Resolution. IEEE Trans. Image Process. 2019, 28, 5649–5662. [Google Scholar] [CrossRef] [PubMed]
- Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual Dense Network for Image Super-Resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2472–2481. [Google Scholar]
- Hui, Z.; Wang, X.; Gao, X. Fast and Accurate Single Image Super-Resolution via Information Distillation Network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 723–731. [Google Scholar]
- Dong, C.; Loy, C.C.; He, K.; Tang, X. Image Super-Resolution Using Deep Convolutional Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 295–307. [Google Scholar] [CrossRef] [PubMed]
- Iglesias, J.E.; Billot, B.; Balbastre, Y.; Magdamo, C.; Arnold, S.E.; Das, S.; Edlow, B.L.; Alexander, D.C.; Golland, P.; Fischl, B. SynthSR: A public AI tool to turn heterogeneous clinical brain scans into high-resolution T1-weighted images for 3D morphometry. Sci. Adv. 2023, 9, eadd3607. [Google Scholar] [CrossRef]
- Wang, X.; Zhang, S.; Lin, Y.; Lyu, Y.; Zhang, J. Pixel attention convolutional network for image super-resolution. Neural Comput. Appl. 2022, 35, 8589–8599. [Google Scholar] [CrossRef]
- Forigua, C.; Escobar, M.; Arbelaez, P. SuperFormer: Volumetric Transformer Architectures for MRI Super-Resolution. In Simulation and Synthesis in Medical Imaging; Springer: Berlin/Heidelberg, Germany, 2022; Volume Lecture Notes in Computer Science. [Google Scholar]
- Song, Z. REHRSeg: Unleashing the Power of Self-Supervised Super-Resolution for Resource-Efficient 3D MRI Segmentation. Neurocomputing 2025, 624, 129425. [Google Scholar] [CrossRef]
- Schlereth, M.; Schillinger, M.; Breininger, K. Faster, Self-Supervised Super-Resolution for Anisotropic Multi-View MRI Using a Sparse Coordinate Loss. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2025, Daejeon, Republic of Korea, 23–27 September 2025. [Google Scholar]
- Zhang, H.; Zhang, Y.; Wu, Q.; Wu, J.; Zhen, Z.; Shi, F.; Yuan, J.; Wei, H.; Liu, C.; Zhang, Y. Self-Supervised Arbitrary Scale Super-Resolution Framework for Anisotropic MRI. In Proceedings of the 2023 IEEE 20th International Symposium on Biomedical Imaging (ISBI), Cartagena, Colombia, 18–21 April 2023; pp. 1–5. [Google Scholar]
- Wang, X. Inter-Slice Super-Resolution of Magnetic Resonance Images by Pre-Training and Self-Supervised Fine-Tuning. In Proceedings of the 2024 IEEE International Symposium on Biomedical Imaging (ISBI), Athens, Greece, 27–30 May 2024. [Google Scholar]
- Gundogdu, B. Self-Supervised Multi-Contrast Super-Resolution for Diffusion-Weighted Prostate MRI. Magn. Reson. Med. 2024, 92, 319–331. [Google Scholar] [CrossRef] [PubMed]
- Chen, H. Low-Res Leads the Way: Improving Generalization for Super-Resolution by Self-Supervised Learning. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 25857–25867. [Google Scholar]
- Benisty, R. SIMPLE: Simultaneous Multi-Plane Self-Supervised Learning for Isotropic MRI Restoration from Anisotropic Data. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Marrakesh, Morocco, 6–10 October 2024. [Google Scholar]
- Remedios, S.W. Self-Supervised Super-Resolution for Anisotropic MR Images with and Without Slice Gap; Wang, L., Dou, Q., Fletcher, P.T., Speidel, S., Li, S., Eds.; Springer: Berlin/Heidelberg, Germany, 2023; pp. 118–128. [Google Scholar]
- Mayo, P. Physics informed guided diffusion for accelerated multi-parametric MRI reconstruction. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Daejeon, Republic of Korea, 23–27 September 2025. [Google Scholar]
- Zhang, L.; Nie, J.; Wei, W.; Zhang, Y. Unsupervised Test-Time Adaptation Learning for Effective Hyperspectral Image Super-Resolution With Unknown Degeneration. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 5008–5025. [Google Scholar] [CrossRef]
- Wang, Z. One for Multiple: Physics-informed Synthetic Data Boosts Generalizable Deep Learning for Fast MRI Reconstruction. Med. Image Anal. 2025, 103, 103616. [Google Scholar] [CrossRef]
- Ren, Y.; Liu, W.; Zhou, Z.; Hu, P. PI-GNN: Physics-Informed Graph Neural Network for Super-Resolution of 4D Flow MRI. In Proceedings of the 2024 IEEE International Symposium on Biomedical Imaging (ISBI), Athens, Greece, 27–30 May 2024. [Google Scholar]
- Jacobs, L. Generalizable synthetic MRI with physics-informed convolutional networks. Med. Phys. 2023, 51, 3348–3359. [Google Scholar] [CrossRef]
- Ferdian, E.; Marlevi, D.; Schollenberger, J.; Aristova, M.; Edelman, E.R. Cerebrovascular super-resolution 4D Flow MRI: Sequential combination of resolution enhancement by deep learning and physics-informed image processing. Med. Image Anal. 2023, 88, 102831. [Google Scholar] [CrossRef]
- Donnay, C.; Okar, S.V.; Tsagkas, C.; Gaitán, M.I.; Poorman, M.; Reich, D.S.; Nair, G. Super resolution using sparse sampling at portable ultra-low field MR. Front. Neurol. 2024, 15, 1330203. [Google Scholar] [CrossRef] [PubMed]
- Man, C.; Lau, V.; Su, S.; Zhao, Y.; Xiao, L.; Ding, Y.; Leung, G.K.K.; Leong, A.T.L.; Wu, E.X. Deep learning enabled fast 3D brain MRI at 0.055 tesla. Sci. Adv. 2023, 9, eadi9327. [Google Scholar] [CrossRef]
- Lau, V.; Xiao, L.; Zhao, Y.; Su, S.; Ding, Y.; Man, C.; Wang, X.; Tsang, A.; Cao, P.; Lau, G.K.K.; et al. Pushing the limits of low-cost ultra-low-field MRI by dual-acquisition deep learning 3D superresolution. Magn. Reson. Med. 2023, 90, 400–416. [Google Scholar] [CrossRef] [PubMed]
- Islam, K.T.; Zhong, S.; Zakavi, P.; Rawluk, N.; Hill, M.; Spreeuwers, L.; Mateen, F.; Sher, I.; McCreary, C.; Frayne, R. Improving portable low-field MRI image quality through image-to-image translation using paired low- and high-field images. Sci. Rep. 2023, 13, 20848. [Google Scholar] [CrossRef] [PubMed]
- Iglesias, J.E.; Schleicher, R.; Laguna, S.; Billot, B.; Schaefer, P.; McKaig, B.; Goldstein, J.N.; Sheth, K.N.; Rosen, M.S.; Kimberly, W.T. Quantitative brain morphometry of portable low-field-strength MRI using super-resolution machine learning. Radiology 2023, 306, e220522. [Google Scholar] [CrossRef]
- Samarasinghe, D.; Wickramasinghe, D.; Wijerathne, T.; Meedeniya, D.; Yogarajah, P. Brain Tumour Segmentation and Edge Detection Using Self-Supervised Learning. Int. J. Online Biomed. Eng. (IJOE) 2025, 21, 127–141. [Google Scholar] [CrossRef]









| Dataset | Source | Subjects | Demographics | Contrasts | Resolution (mm3) |
|---|---|---|---|---|---|
| HCP-YA | [16] | 1206 HC | 22–36 years m/f = 507/699 | T1w | 0.7 × 0.7 × 0.7 |
| T2w | 0.7 × 0.7 × 0.7 | ||||
| IXI | [15] | 563 HC | 20–86 years m/f = 250/313 | T1w | 0.94 × 0.94 × 1.2 |
| T2w | 0.94 × 0.94 × 1.2 | ||||
| PTT1 | [17] | 20 HC | 18–25 years m/f = 10/10 | T1w | 1 × 1×1, 0.7 × 0.7 × 0.7 |
| T2w | 0.9 × 0.9 × 1.9, 0.4 × 0.4 × 1 | ||||
| PTT2 | [18] | 10 HC | 25–41 years m/f = 7/3 | T1w | 0.8 × 0.8 × 0.8, 0.7 × 0.7 × 0.7 |
| T2w | 0.8 × 0.8 × 0.8, 0.7 × 0.7 × 0.7 |
| Model | Parameters (Million) | Train (h) | PSNR (IXIT1w) | SSIM (IXIT1w) | PSNR (IXI-T2w) | SSIM (IXI-T2w) | PSNR (HCP-T1w) | SSIM (HCP-T1w) |
|---|---|---|---|---|---|---|---|---|
| Bicubic | — | 31.68 ± 0.55 | 0.935 ± 0.028 | 32.51 ± 0.58 | 0.946 ± 0.025 | 40.19 ± 0.48 | 0.9874 ± 0.020 | |
| SRCNN | 0.82 | 0.7 | 33.86 ± 0.42 | 0.961 ± 0.018 | 35.33 ± 0.45 | 0.969 ± 0.016 | 43.80 ± 0.38 | 0.9935 ± 0.014 |
| VDSR | 0.37 | 3.0 | 35.43 ± 0.35 | 0.968 ± 0.015 | 36.74 ± 0.38 | 0.963 ± 0.013 | 44.77 ± 0.32 | 0.9936 ± 0.011 |
| EDSR | 1.36 | 4.7 | 36.13 ± 0.28 | 0.971 ± 0.012 | 38.07 ± 0.30 | 0.979 ± 0.010 | 46.11 ± 0.25 | 0.9957 ± 0.010 |
| PAN | 0.78 | 5.0 | 36.38 ± 0.27 | 0.970 ± 0.013 | 37.79 ± 0.27 | 0.978 ± 0.011 | 45.71 ± 0.21 | 0.9954 ± 0.008 |
| W2AMSN-S | 11.37 | 14 | 37.12 ± 0.22 | 0.972 ± 0.009 | 38.30 ± 0.26 | 0.980 ± 0.008 | 46.43 ± 0.20 | 0.9961 ± 0.007 |
| FMEN | 3.80 | 11 | 37.22 ± 0.16 | 0.973 ± 0.008 | 38.40 ± 0.17 | 0.981 ± 0.007 | 46.58 ± 0.14 | 0.9963 ± 0.007 |
| CHARMS | 1.74 | 10 | 37.79 ± 0.11 | 0.973 ± 0.008 | 38.56 ± 0.11 | 0.981 ± 0.006 | 46.58 ± 0.10 | 0.9963 ± 0.006 |
| Model | Parameters (Million) | Train (h) | PSNR (IXI-T1w) | SSIM (IXI-T1w) | PSNR (IXI-T2W) | SSIM (IXI-T2W) | PSNR (HCP-T1w) | SSIM (HCP-T1w) |
|---|---|---|---|---|---|---|---|---|
| Bicubic | — | 26.17 ± 0.58 | 0.786 ± 0.030 | 26.92 ± 0.58 | 0.826 ± 0.025 | 31.57 ± 0.48 | 0.924 ± 0.024 | |
| SRCNN | 0.82 | 0.5 | 29.18 ± 0.50 | 0.888 ± 0.028 | 30.63 ± 0.50 | 0.873 ± 0.022 | 33.96 ± 0.39 | 0.948 ± 0.021 |
| VDSR | 0.37 | 1.5 | 30.49 ± 0.45 | 0.909 ± 0.020 | 32.44 ± 0.46 | 0.890 ± 0.020 | 34.60 ± 0.36 | 0.954 ± 0.018 |
| EDSR | 1.51 | 2.3 | 31.48 ± 0.36 | 0.931 ± 0.018 | 32.53 ± 0.38 | 0.948 ± 0.012 | 36.10 ± 0.33 | 0.965 ± 0.011 |
| PAN | 0.92 | 2.7 | 31.76 ± 0.28 | 0.926 ± 0.016 | 32.27 ± 0.28 | 0.944 ± 0.011 | 35.90 ± 0.28 | 0.964 ± 0.009 |
| W2AMSN-S | 11.41 | 9.0 | 32.61 ± 0.22 | 0.939 ± 0.016 | 32.76 ± 0.11 | 0.953 ± 0.009 | 36.56 ± 0.20 | 0.968 ± 0.008 |
| FMEN | 3.95 | 7.0 | 32.72 ± 0.20 | 0.944 ± 0.010 | 32.92 ± 0.20 | 0.956 ± 0.007 | 36.81 ± 0.18 | 0.969 ± 0.007 |
| CHARMS | 1.89 | 7.0 | 33.27 ± 0.14 | 0.945 ± 0.010 | 32.97 ± 0.15 | 0.956 ± 0.007 | 36.65 ± 0.11 | 0.969 ± 0.007 |
| Model | Baseline | +CS | +CS +PCA | +CS +PCA +Transformer | Full Model | |
|---|---|---|---|---|---|---|
| 2× | Parameters (MB) | 1.46 | 1.49 | 1.69 | 1.72 | 1.74 |
| PSNR/SSIM (IXI-T1w) | 35.12/0.967 | 36.19/0.973 | 36.27/0.973 | 36.28/0.973 | 36.29/0.973 | |
| PSNR/SSIM (IXI-T2w) | 37.38/0.973 | 38.43/0.979 | 38.50/0.981 | 38.51/0.981 | 38.56/0.981 | |
| PSNR/SSIM (HCP-T1w) | 45.63/0.989 | 46.44/0.995 | 46.56/0.996 | 46.56/0.996 | 46.58/0.996 | |
| 4× | Size of Parameters (MB) | 1.61 | 1.64 | 1.84 | 1.87 | 1.89 |
| PSNR/SSIM (IXI-T1w) | 28.66/0.869 | 29.37/0.889 | 29.42/0.890 | 29.44/0.890 | 29.47/0.891 | |
| PSNR/SSIM (IXI-T2w) | 29.56/0.896 | 30.39/0.909 | 30.41/0.916 | 30.44/0.916 | 30.47/0.916 | |
| PSNR/SSIM (HCP-T1w) | 35.81/0.946 | 36.53/0.967 | 36.60/0.968 | 36.63/0.968 | 36.65/0.969 | |
| Subject | 3T (T1w) | SR (T1w) | 3T (T2w) | SR (T2w) |
|---|---|---|---|---|
| 1 | 27.310 | 33.330 | 27.359 | 28.475 |
| 2 | 28.639 | 34.660 | 32.265 | 38.286 |
| 3 | 27.927 | 33.947 | 31.938 | 37.958 |
| 4 | 26.724 | 32.744 | 30.251 | 36.271 |
| 5 | 27.174 | 33.195 | 30.546 | 36.566 |
| 6 | 27.752 | 33.773 | 31.982 | 38.003 |
| 7 | 27.717 | 33.737 | 31.956 | 37.976 |
| 8 | 27.622 | 33.643 | 30.472 | 36.493 |
| 9 | 28.394 | 34.415 | 31.208 | 37.229 |
| 10 | 27.213 | 33.234 | 31.952 | 37.973 |
| Mean | 27.647 | 33.668 | 30.993 | 36.523 |
| Std | 0.579 | 0.579 | 1.476 | 2.923 |
| Subject | 3T (T1w) | SR (T1w) | 3T (T2w) | SR (T2w) |
|---|---|---|---|---|
| 1 | 0.821 | 0.938 | 0.726 | 0.727 |
| 2 | 0.870 | 0.954 | 0.872 | 0.948 |
| 3 | 0.830 | 0.945 | 0.851 | 0.933 |
| 4 | 0.807 | 0.933 | 0.783 | 0.890 |
| 5 | 0.832 | 0.947 | 0.822 | 0.907 |
| 6 | 0.844 | 0.949 | 0.829 | 0.920 |
| 7 | 0.816 | 0.944 | 0.816 | 0.914 |
| 8 | 0.836 | 0.949 | 0.817 | 0.911 |
| 9 | 0.854 | 0.951 | 0.840 | 0.925 |
| 10 | 0.818 | 0.939 | 0.825 | 0.914 |
| Mean | 0.833 | 0.945 | 0.818 | 0.899 |
| Std | 0.019 | 0.007 | 0.040 | 0.062 |
| Subject | 3T (T1w) | 7T (T1w) | SR (T1w) | 3T (T2w) | 3T (T2w) | SR (T2w) |
|---|---|---|---|---|---|---|
| 1 | 8.007 | 9.013 | 8.557 | 4.301 | 5.983 | 5.481 |
| 2 | 8.887 | 10.792 | 10.7 | 5.695 | 7.331 | 6.689 |
| 3 | 8.478 | 10.368 | 9.628 | 5.353 | 6.501 | 5.956 |
| 4 | 7.451 | 7.458 | 7.588 | 4.119 | 5.071 | 4.542 |
| 5 | 7.995 | 7.576 | 7.963 | 4.133 | 5.497 | 5.105 |
| 6 | 8.311 | 10.153 | 9.375 | 5.147 | 6.368 | 5.791 |
| 7 | 8.159 | 9.771 | 8.879 | 4.961 | 6.285 | 5.779 |
| 8 | 8.055 | 9.277 | 8.642 | 4.652 | 6.174 | 5.725 |
| 9 | 8.616 | 10.729 | 10.301 | 5.466 | 6.888 | 6.522 |
| 10 | 8.001 | 8.791 | 8.305 | 4.192 | 5.878 | 5.389 |
| Mean | 8.196 | 9.393 | 8.994 | 4.802 | 6.198 | 5.697 |
| Std | 0.401 | 1.201 | 1.003 | 0.601 | 0.652 | 0.631 |
| Subject | 3T (T1w) | 7T (T1w) | SR (T1w) | 3T (T2w) | 3T (T2w) | SR (T2w) |
|---|---|---|---|---|---|---|
| 1 | 1.916 | 1.915 | 2.081 | 1.551 | 2.823 | 2.357 |
| 2 | 3.193 | 3.537 | 3.46 | 2.755 | 4.676 | 3.646 |
| 3 | 2.559 | 2.557 | 2.492 | 2.431 | 3.321 | 2.520 |
| 4 | 1.234 | 1.451 | 1.205 | 1.042 | 1.623 | 1.578 |
| 5 | 1.393 | 1.679 | 1.307 | 1.298 | 1.722 | 1.742 |
| 6 | 2.476 | 2.305 | 2.42 | 2.015 | 2.185 | 2.603 |
| 7 | 1.953 | 2.285 | 2.21 | 1.986 | 3.082 | 3.398 |
| 8 | 1.923 | 2.011 | 2.173 | 1.842 | 3.001 | 2.715 |
| 9 | 2.607 | 3.421 | 2.718 | 2.549 | 4.174 | 3.416 |
| 10 | 1.775 | 1.81 | 2.025 | 1.539 | 1.642 | 2.332 |
| Mean | 2.103 | 2.297 | 2.209 | 1.901 | 2.825 | 2.631 |
| Std | 0.601 | 0.702 | 0.652 | 0.559 | 1.201 | 1.066 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Li, X.; Sun, H.; Li, T.-Q. CHARMS: A CNN-Transformer Hybrid with Attention Regularization for MRI Super-Resolution. Sensors 2026, 26, 738. https://doi.org/10.3390/s26020738
Li X, Sun H, Li T-Q. CHARMS: A CNN-Transformer Hybrid with Attention Regularization for MRI Super-Resolution. Sensors. 2026; 26(2):738. https://doi.org/10.3390/s26020738
Chicago/Turabian StyleLi, Xia, Haicheng Sun, and Tie-Qiang Li. 2026. "CHARMS: A CNN-Transformer Hybrid with Attention Regularization for MRI Super-Resolution" Sensors 26, no. 2: 738. https://doi.org/10.3390/s26020738
APA StyleLi, X., Sun, H., & Li, T.-Q. (2026). CHARMS: A CNN-Transformer Hybrid with Attention Regularization for MRI Super-Resolution. Sensors, 26(2), 738. https://doi.org/10.3390/s26020738

