Adjustable Complexity Transformer Architecture for Image Denoising
Abstract
1. Introduction
2. Related Works
3. Proposed Method
3.1. Overall Architecture
3.2. Attention with Adjustable Complexity and Transformer
3.3. Distinctions from Other Efficiency-Focused Mechanisms
4. Experimental Results
4.1. Experimental Setup
4.2. Denoising Performance
4.3. Complexity Analysis
4.4. Ablation Study
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Elad, M.; Kawar, B.; Vaksman, G. Image denoising: The deep learning revolution and beyond—A survey paper. SIAM J. Imaging Sci. 2023, 16, 1594–1654. [Google Scholar] [CrossRef]
- Jiang, B.; Li, J.; Lu, Y.; Cai, Q.; Song, H.; Lu, G. Efficient image denoising using deep learning: A brief survey. Inf. Fusion 2025, 118, 103013. [Google Scholar] [CrossRef]
- Mao, J.; Sun, L.; Chen, J.; Yu, S. Overview of research on digital image denoising methods. Sensors 2025, 25, 2615. [Google Scholar] [CrossRef] [PubMed]
- Buades, A.; Coll, B.; Morel, J.M. A review of image denoising algorithms, with a new one. Multiscale Model. Simul. 2005, 4, 490–530. [Google Scholar] [CrossRef]
- Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3-D transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef] [PubMed]
- Chang, L.W.; Liao, J.R. Improving non-local means image denoising by correlation correction. Multidimens. Syst. Signal Process. 2023, 34, 147–162. [Google Scholar] [CrossRef]
- Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a gaussian denoiser: Residual learning of deep CNN for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef]
- Plotz, T.; Roth, S. Benchmarking denoising algorithms with real photographs. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1586–1595. [Google Scholar] [CrossRef]
- Abdelhamed, A.; Lin, S.; Brown, M.S. A high-quality denoising dataset for smartphone cameras. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1692–1700. [Google Scholar] [CrossRef]
- Abdelhamed, A.; Timofte, R.; Brown, M.S.; Yu, S.; Park, B.; Jeong, J.; Jung, S.-W.; Kim, D.-W.; Chung, J.-R.; Liu, J.; et al. NTIRE 2019 challenge on real image denoising: Methods and results. In IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–17 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–14. [Google Scholar] [CrossRef]
- Mao, X.; Shen, C.; Yang, Y.B. Image restoration using very deep convolutional encoder-decoder networks with symmetric skip connections. In 30th Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; Curran Associates Inc.: Red Hook, NY, USA, 2016; pp. 2802–2810. [Google Scholar] [CrossRef]
- Chen, Y.; Pock, T. Trainable nonlinear reaction diffusion: A flexible framework for fast and effective image restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1256–1272. [Google Scholar] [CrossRef]
- Plotz, T.; Roth, S. Neural nearest neighbors networks. In 32nd Conference on Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; Curran Associates Inc.: Red Hook, NY, USA, 2018; pp. 1095–1106. [Google Scholar] [CrossRef]
- Zhang, Y.; Li, K.; Li, K.; Zhong, B.; Fu, Y. Residual non-local attention networks for image restoration. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019; pp. 1–18. [Google Scholar] [CrossRef]
- Tian, C.; Xu, Y.; Li, Z.; Zuo, W.; Fei, L.; Liu, H. Attention-guided CNN for image denoising. Neural Netw. 2020, 124, 117–129. [Google Scholar] [CrossRef]
- Mei, Y.; Fan, Y.; Zhang, Y.; Yu, J.; Zhou, Y.; Liu, D.; Fu, Y.; Huang, T.S.; Shi, H. Pyramid attention network for image restoration. Int. J. Comput. Vis. 2023, 131, 3207–3225. [Google Scholar] [CrossRef]
- Pan, Y.; Ren, C.; Wu, X.; Huang, J.; He, X. Real image denoising via guided residual estimation and noise correction. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 1994–2000. [Google Scholar] [CrossRef]
- Qin, M.; Ren, C.; Yang, H.; He, X.; Wang, Z. Blind image denoising via deep unfolding network with degradation information guidance. IEEE Trans. Circuits Syst. II Express Briefs 2023, 70, 3179–3183. [Google Scholar] [CrossRef]
- Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H.; Shao, L. Learning enriched features for fast image restoration and enhancement. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 1934–1948. [Google Scholar] [CrossRef]
- Wu, W.; Liu, S.; Xia, Y.; Zhan, Y. Dual residual attention network for image denoising. Pattern Recognit. 2024, 149, 110291. [Google Scholar] [CrossRef]
- Zafar, A.; Aftab, D.; Qureshi, R.; Fan, X.; Chen, P.; Wu, J. Single stage adaptive multi-attention network for image restoration. IEEE Trans. Image Process. 2024, 33, 2924–2935. [Google Scholar] [CrossRef] [PubMed]
- Ge, X.; Zhu, Y.; Qi, L.; Hu, Y.; Sun, J.; Zhang, Y. Enhancing border learning for better image denoising. Mathematics 2025, 13, 1119. [Google Scholar] [CrossRef]
- Chen, L.; Chu, X.; Zhang, X.; Sun, J. Simple baselines for image restoration. In 17th European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; Springer: Cham, Switzerland, 2022; pp. 17–33. [Google Scholar] [CrossRef]
- Zamir, S.W.; Arora, A.; Khan, S.H.; Hayat, M.; Khan, F.S.; Yang, M.H. Restormer: Efficient transformer for high-resolution image restoration. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA; IEEE: Piscataway, NJ, USA, 2022; pp. 5728–5739. [Google Scholar] [CrossRef]
- Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. SwinIR: Image restoration using swin transformer. In IEEE/CVF International Conference on Computer Vision Workshops, Montreal, QC, Canada, 11–17 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1833–1844. [Google Scholar] [CrossRef]
- Fan, C.M.; Liu, T.J.; Liu, K.H. SUNet: Swin transformer unet for image denoising. In IEEE International Symposium on Circuits and Systems, Austin, Texas, USA, 28 May–1 June 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 2333–2337. [Google Scholar] [CrossRef]
- Li, J.; Cheng, B.; Chen, Y.; Gao, G.; Shi, J.; Zeng, T. EWT: Efficient wavelet-transformerfor single image denoising. Neural Netw. 2024, 177, 106378. [Google Scholar] [CrossRef]
- Song, M.; Wang, W.; Zhao, Y. A dynamic network with transformer for image denoising. Electronics 2024, 13, 1676. [Google Scholar] [CrossRef]
- Li, J.; Zhang, Z.; Zuo, W. Rethinking transformer-based blind-spot network for self-supervised image denoising. Proc. AAAI Conf. Artif. Intell. 2025, 39, 4788–4796. [Google Scholar] [CrossRef]
- Yao, M.; Xu, R.; Guan, Y.; Huang, J.; Xiong, Z. Neural degradation representation learning for all-in-one image restoration. IEEE Trans. Image Process. 2024, 33, 5408–5423. [Google Scholar] [CrossRef] [PubMed]
- Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a fast and flexible solution for CNN based image denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef]
- El Helou, M.; Susstrunk, S. Blind universal Bayesian image denoising with gaussian noise level learning. IEEE Trans. Image Process. 2020, 29, 4885–4897. [Google Scholar] [CrossRef] [PubMed]
- Du, Y.; Han, G.; Tan, Y.; Xiao, C.; He, S. Blind image denoising via dynamic dual learning. IEEE Trans. Multimed. 2021, 23, 2139–2152. [Google Scholar] [CrossRef]
- Zhang, Y.; Li, K.; Li, K.; Sun, G.; Kong, Y.; Fu, Y. Accurate and fast image denoising via attention guided scaling. IEEE Trans. Image Process. 2021, 30, 6255–6265. [Google Scholar] [CrossRef]
- Mou, C.; Zhang, J.; Fan, X.; Liu, H.; Wang, R. COLA-Net: Collaborative attention network for image restoration. IEEE Trans. Multimed. 2022, 24, 1366–1377. [Google Scholar] [CrossRef]
- Guo, S.; Yan, Z.; Zhang, K.; Zuo, W.; Zhang, L. Toward convolutional blind denoising of real photographs. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1712–1722. [Google Scholar] [CrossRef]
- Yue, Z.; Yong, H.e.; Zhao, Q.; Meng, D.; Ozair, S.; Zhang, L. Variational denoising network: Toward blind noise modeling and removal. In 32rd Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Curran Associates Inc.: Red Hook, NY, USA, 2019; pp. 1690–1701. [Google Scholar]
- Zhou, Y.; Jiao, J.; Huang, H.; Wang, Y. When AWGN-based denoiser meets real noises. Proc. AAAI Conf. Artif. Intell. 2020, 34, 13074–13081. [Google Scholar] [CrossRef]
- Tian, C.; Xu, Y.; Zuo, W. Image denoising using deep CNN with batch renormalization. Neural Netw. 2020, 121, 461–473. [Google Scholar] [CrossRef]
- Jang, Y.I.; Kim, Y.; Cho, N.I. Dual path denoising network for real photographic noise. IEEE Signal Process. Lett. 2020, 27, 860–864. [Google Scholar] [CrossRef]
- Zamir, S.W.; Arora, A.; Khan, S.H.; Munawar, H.; Khan, F.S.; Yang, M.H.; Shao, L. Learning enriched features for real image restoration and enhancement. In European Conference on Computer Vision—ECVA; Springer: Cham, Switzerland, 2020; pp. 492–511. [Google Scholar]
- Chen, H.; Wang, Y.; Guo, T.; Xu, C.; Deng, Y.; Liu, Z.; Ma, S.; Xu, C.; Xu, C.; Gao, W. Pre-trained image processing transformer. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 12299–12310. [Google Scholar] [CrossRef]
- Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. In Proceedings of the 9th International Conference on Learning Representations (ICLR), Virtual, 3–7 May 2021; pp. 1–21. [Google Scholar]
- Zhang, J.; Zhang, Y.; Gu, J.; Dong, J.; Kong, L.; Yang, X. Xformer: Hybrid X-shaped transformer for image denoising. In Proceedings of the 12th International Conference on Learning Representations (ICLR), Vienna, Austria, 7–11 May 2024; pp. 1–13. [Google Scholar]
- Wang, Z.; Cun, X.; Bao, J.; Zhou, W.; Liu, J.; Li, H. Uformer: A general u-shaped transformer for image restoration. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 17662–17672. [Google Scholar] [CrossRef]
- Xiao, J.; Fu, X.; Zhou, M.; Liu, H.; Zha, Z.J. Random shuffle transformer for image restoration. In 40th International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023; PMLR: Cambridge, MA, USA, 2023; pp. 38039–38058. [Google Scholar]
- Zhang, J.; Zhang, Y.; Gu, J.; Zhang, Y.; Kong, L.; Yuan, X. Accurate image restoration with attention retractable transformer. In Proceedings of the 11th International Conference on Learning Representations (ICLR), Kigali, Rwanda, 1–5 May 2023; pp. 1–13. [Google Scholar]
- Liu, K.; Du, X.; Liu, S.; Zheng, Y.; Wu, X.; Jin, C. DDT: Dual-branch deformable transformer for image denoising. In IEEE International Conference on Multimedia and Expo (ICME), Brisbane, Australia, 10–14 July 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 2765–2770. [Google Scholar] [CrossRef]
- Zhao, H.; Gou, Y.; Li, B.; Peng, D.; Lv, J.; Peng, X. Comprehensive and delicate: An efficient transformer for image restoration. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 14122–14132. [Google Scholar] [CrossRef]
- Jiang, X.; Zhang, X.; Gao, N.; Deng, Y. When fast Fourier transform meets transformer for image restoration. In 18th European Conference on Computer Vision (ECCV), Milano, Italy, 29 September–4 October 2024; Springer: Cham, Switzerland, 2024; pp. 381–402. [Google Scholar] [CrossRef]
- Tian, C.; Zheng, M.; Lin, C.W.; Li, Z.; Zhang, D. Heterogeneous window transformer for image denoising. IEEE Trans. Syst. Man, Cybern. Syst. 2024, 54, 6621–6632. [Google Scholar] [CrossRef]
- Zhang, Q.; Zhang, J.; Xu, Y.; Tao, D. Vision transformer with quadrangle attention. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 3608–3624. [Google Scholar] [CrossRef]
- Timofte, R.; Agustsson, E.; Gool, L.V.; Yang, M.H.; Zhang, L.; Lim, B.; Son, S. NTIRE 2017 challenge on single image super-resolution: Methods and results. In IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1110–1121. [Google Scholar] [CrossRef]
- Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced deep residual networks for single image super-resolution. In IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1132–1140. [Google Scholar] [CrossRef]
- Ma, K.; Duanmu, Z.; Wu, Q.; Wang, Z.; Yong, H.; Li, H.; Zhang, L. Waterloo exploration database: New challenges for image quality assessment models. IEEE Trans. Image Process. 2017, 26, 1004–1016. [Google Scholar] [CrossRef] [PubMed]
- Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological sstatistics. In 8th IEEE International Conference on Computer Vision (ICCV), Vancouver, BC, Canada, 7–14 July 2001; IEEE: Piscataway, NJ, USA, 2001; Volume 2, pp. 416–423. [Google Scholar] [CrossRef]
- Franzen, R. Kodak Lossless True Color Image Suite. Available online: http://r0k.us/graphics/kodak/ (accessed on 26 July 2025).
- Zhang, L.; Wu, X.; Buades, A.; Li, X. Color demosaicking by local directional interpolation and nonlocal adaptive thresholding. J. Electron. Imaging 2011, 20, 023016. [Google Scholar] [CrossRef]
- Huang, J.B.; Singh, A.; Ahuja, N. Single image super-resolution from transformed self-exemplars. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 5197–5206. [Google Scholar] [CrossRef]
- Mou, C.; Zhang, J.; Wu, Z. Dynamic attentive graph learning for image restoration. In IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 4308–4317. [Google Scholar] [CrossRef]
- Ren, C.; He, X.; Wang, C.; Zhao, Z. Adaptive consistency prior based deep network for image denoising. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 8592–8602. [Google Scholar] [CrossRef]
- Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H.; Shao, L. Multi-stage progressive image restoration. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 14821–14831. [Google Scholar] [CrossRef]
- Huang, J.; Liu, X.; Pan, Y.; He, X.; Ren, C. CasapuNet: Channel affine self-attention- based progressively updated network for real image denoising. IEEE Trans. Ind. Inform. 2023, 19, 9145–9156. [Google Scholar] [CrossRef]
- Wu, W.; Lv, G.; Duan, Y.; Liang, P.; Zhang, Y.; Xia, Y. Dual convolutional neural network with attention for image blind denoising. Multimed. Syst. 2024, 30, 263. [Google Scholar] [CrossRef]
- Wu, W.; Ge, A.; Lv, G.; Xia, Y.; Zhang, Y.; Xiong, W. Two-stage progressive residual dense attention network for image denoising. arXiv 2024. [Google Scholar] [CrossRef]
- Yue, Z.; Yong, H.; Zhao, Q.; Zhang, L.; Meng, D.; Wong, K.Y.K. Deep variational network toward blind image restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 7011–7026. [Google Scholar] [CrossRef] [PubMed]
- Brateanu, A.; Balmez, R.; Avram, A.; Orhei, C. AKDT: Adaptive kernel dilation transformer for effective image denoising. In Proceedings of the International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISAPP), Porto, Portugal, 26–28 February 2025; pp. 418–425. [Google Scholar]
- Wu, W.; Liao, S.; Lv, G.; Liang, P.; Zhang, Y. Image blind denoising using dual convolutional neural network with skip connection. Signal Process. Image Commun. 2025, 138, 117365. [Google Scholar] [CrossRef]
- Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2012, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]






| Category | Strategy | Method Examples |
|---|---|---|
| Traditional | NSS-based | BM3D [5] |
| Early CNN | Basic residual network | DnCNN [7] |
| Noise-level map assistance | FFDNet [31] | |
| Multi-Path | Noise estimation | GrencNet [17] |
| BDUNet [18] | ||
| BUIFD [32] | ||
| CBDNet [36] | ||
| VDN [37] | ||
| PD [38] | ||
| Multi-Path | Convolutional strategies | MIRNet-v2 [19] |
| BRDNet [39] | ||
| DPN [40] | ||
| MIRNet [41] | ||
| Attention | Residual non-local | RNAN [14] |
| Attention-guided | ADNet [15] | |
| Pyramid attention | PANet [16] | |
| Collaborative attention | COLA-Net [35] | |
| Dynamic dual learning | DualBDNet [33] | |
| Attention-guided scaling | AGS [34] | |
| Transformer | Pixel-wise window attention | SwinIR [25,26] |
| Channel attention | Restormer [24] | |
| Wavelet | EWT [27] | |
| Neural degradation representation | NDR [30] | |
| Multi-scale U-Net | Uformer [45] | |
| Randomly shuffling | ShuffleFormer [46] | |
| Interleaved pixels | ART [47] | |
| Dual-branch deformable | DDT [48] | |
| Superpixels | CODE [49] | |
| Fast Fourier transform | SFHformer [50] | |
| Heterogeneous window | HWformer [51] | |
| Quadrangle attention | QFormer [52] |
| In Each Transformer | |||
|---|---|---|---|
| Block Name | # of Transformers | # of Heads | # of Channels |
| Cascaded Transformer Block 1 | 4 | 1 | 48 |
| Cascaded Transformer Block 2 | 6 | 2 | 96 |
| Cascaded Transformer Block 3 | 6 | 4 | 192 |
| Cascaded Transformer Block 4 | 8 | 8 | 384 |
| Cascaded Transformer Block 5 | 6 | 4 | 192 |
| Cascaded Transformer Block 6 | 6 | 2 | 96 |
| Cascaded Transformer Block 7 | 4 | 1 | 48 |
| Cascaded Transformer Block 8 | 4 | 1 | 48 |
| Method | PSNR | SSIM |
|---|---|---|
| DAGL (2021) [60] | 38.87 | 0.906 |
| DeamNet (2021) [61] | 38.87 | 0.906 |
| GrencNet (2023) [17] | 39.45 | 0.912 |
| CasaPuNet (2023) [63] | 39.50 | 0.910 |
| TSP-RDANet (2024) [65] | 39.51 | 0.911 |
| VIRNet (2024) [66] | 39.62 | 0.912 |
| AKDT (2025) [67] | 39.62 | 0.912 |
| MPRNet (2021) [62] | 39.63 | 0.913 |
| Restormer (2022) [24] | 39.93 | 0.915 |
| Ours | 39.65 | 0.913 |
| Dataset | CBSD68 | Kodak | |||||
| Method | |||||||
| DnCNN (2017) [7] | 33.89 | 31.24 | 27.94 | 34.60 | 32.14 | 28.95 | |
| FFDNet (2017) [31] | 33.88 | 31.22 | 27.97 | 34.75 | 32.25 | 29.11 | |
| BUIFD (2020) [32] | 33.89 | 31.24 | 27.98 | 34.69 | 32.22 | 29.10 | |
| DCBDNet (2025) [68] | 34.01 | 31.40 | 28.21 | 34.89 | 32.47 | 29.41 | |
| DCANet (2024) [64] | 34.05 | 31.44 | 28.27 | 34.98 | 32.56 | 29.51 | |
| TSP-RDANet (2024) [65] | 34.14 | 31.51 | 28.32 | 35.10 | 32.66 | 29.60 | |
| DRANet (2024) [20] | 34.17 | 31.55 | 28.35 | 35.15 | 32.71 | 29.65 | |
| Restormer (2022) [24] | 34.39 | 31.78 | 28.60 | 35.44 | 33.02 | 30.00 | |
| Ours | 34.29 | 31.66 | 28.46 | 35.30 | 32.86 | 29.81 | |
| SwinIR (2021) [25] | 34.41 | 31.78 | 28.56 | 35.46 | 33.01 | 29.95 | |
| Dataset | McMaster | Urban100 | |||||
| Method | |||||||
| DnCNN (2017) [7] | 33.45 | 31.52 | 28.61 | 32.98 | 30.81 | 27.59 | |
| FFDNet (2017) [31] | 34.65 | 32.35 | 29.18 | 33.83 | 31.40 | 28.05 | |
| BUIFD (2020) [32] | 34.18 | 32.03 | 29.00 | 33.67 | 31.31 | 28.00 | |
| DCBDNet (2025) [68] | 34.76 | 32.56 | 29.54 | 34.05 | 31.77 | 28.53 | |
| DCANet (2024) [64] | 34.83 | 32.62 | 29.59 | 34.17 | 31.90 | 28.76 | |
| TSP-RDANet (2024) [65] | 35.06 | 32.80 | 29.73 | 34.43 | 32.14 | 28.99 | |
| DRANet (2024) [20] | 35.09 | 32.84 | 29.77 | 34.49 | 32.23 | 29.09 | |
| Restormer (2022) [24] | 35.55 | 33.31 | 30.29 | 35.06 | 32.91 | 30.16 | |
| Ours | 35.34 | 33.08 | 30.01 | 34.60 | 32.45 | 29.42 | |
| SwinIR (2021) [25] | 35.61 | 33.32 | 30.20 | 35.16 | 32.93 | 29.86 | |
| Dataset | CBSD68 | Kodak | |||||
| Method | |||||||
| DnCNN (2017) [7] | 0.932 | 0.887 | 0.793 | 0.923 | 0.879 | 0.790 | |
| FFDNet (2017) [31] | 0.932 | 0.886 | 0.792 | 0.924 | 0.881 | 0.794 | |
| BUIFD (2020) [32] | 0.931 | 0.886 | 0.793 | 0.922 | 0.879 | 0.793 | |
| DCBDNet (2025) [68] | 0.934 | 0.891 | 0.804 | 0.927 | 0.887 | 0.807 | |
| DCANet (2024) [64] | 0.935 | 0.892 | 0.807 | 0.928 | 0.888 | 0.812 | |
| TSP-RDANet (2024) [65] | 0.935 | 0.893 | 0.808 | 0.929 | 0.890 | 0.813 | |
| DRANet (2024) [20] | 0.936 | 0.893 | 0.808 | 0.929 | 0.891 | 0.814 | |
| Restormer (2022) [24] | 0.938 | 0.898 | 0.817 | 0.933 | 0.896 | 0.824 | |
| Ours | 0.937 | 0.896 | 0.813 | 0.931 | 0.894 | 0.820 | |
| SwinIR (2021) [25] | 0.939 | 0.898 | 0.816 | 0.933 | 0.896 | 0.823 | |
| Dataset | McMaster | Urban100 | |||||
| Method | |||||||
| DnCNN (2017) [7] | 0.907 | 0.873 | 0.799 | 0.934 | 0.904 | 0.835 | |
| FFDNet (2017) [31] | 0.925 | 0.889 | 0.816 | 0.944 | 0.914 | 0.850 | |
| BUIFD (2020) [32] | 0.918 | 0.882 | 0.811 | 0.940 | 0.910 | 0.843 | |
| DCBDNet (2025) [68] | 0.926 | 0.895 | 0.831 | 0.946 | 0.920 | 0.862 | |
| DCANet (2024) [64] | 0.927 | 0.896 | 0.833 | 0.947 | 0.922 | 0.867 | |
| TSP-RDANet (2024) [65] | 0.931 | 0.900 | 0.838 | 0.949 | 0.925 | 0.872 | |
| DRANet (2024) [20] | 0.930 | 0.900 | 0.838 | 0.949 | 0.925 | 0.873 | |
| Restormer (2022) [24] | 0.937 | 0.909 | 0.854 | 0.954 | 0.933 | 0.892 | |
| Ours | 0.935 | 0.905 | 0.847 | 0951 | 0.929 | 0.881 | |
| SwinIR (2021) [25] | 0.937 | 0.909 | 0.851 | 0.954 | 0.933 | 0.888 | |
| Model | FLOPs | Parameters |
|---|---|---|
| FFDNet (2017) [31] | 13.978 G | 0.852 M |
| DnCNN (2017) [7] | 43.873 G | 0.668 M |
| DCBDNet (2025) [68] | 49.173 G | 1.013 M |
| AKDT (2025) [67] | 60.152 G | 11.424 M |
| DCANet (2024) [64] | 75.543 G | 1.389 M |
| BUIFD (2020) [32] | 78.428 G | 1.196 M |
| Ours | 120.536 G | 22.866 M |
| DeamNet (2021) [61] | 146.369 G | 1.876 M |
| Restormer (2022) [24] | 155.894 G | 26.097 M |
| VIRNet (2024) [66] | 159.488 G | 15.404 M |
| CasaPuNet (2023) [63] | 240.635 G | 2.411 M |
| DAGL (2021) [60] | 273.386 G | 5.717 M |
| DRANet (2024) [20] | 589.428 G | 1.617 M |
| GrencNet (2023) [17] | 611.020 G | 2.019 M |
| TSP-RDANet (2024) [65] | 678.123 G | 2.848 M |
| SwinIR (2021) [25] | 808.369 G | 11.456 M |
| MPRNet (2021) [62] | 1393.831 G | 15.741 M |
| Kernel | Stride Factor F | ||||
| 128 | 256 | 512 | 1024 | 2048 | |
| PSNR | 26.62 | 29.08 | 30.10 | 30.39 | 30.39 |
| SSIM | 0.718 | 0.810 | 0.841 | 0.848 | 0.848 |
| Stride | Kernel Size and | ||||
| Kernel | 3 | 5 | 7 | 9 | 11 |
| PSNR | 30.38 | 30.38 | 30.39 | 30.38 | 30.38 |
| SSIM | 0.847 | 0.843 | 0.848 | 0.847 | 0.847 |
| Stride | Kernel Size | ||||
| Kernel | 3 | 5 | 7 | 9 | 11 |
| PSNR | 30.34 | 30.39 | 30.38 | 30.38 | 30.38 |
| SSIM | 0.847 | 0.848 | 0.847 | 0.847 | 0.847 |
| Kernel | Stride Factor F | ||||
| 64 | 128 | 256 | 512 | 1024 | |
| PSNR | 26.71 | 36.35 | 38.84 | 38.84 | 38.84 |
| SSIM | 0.850 | 0.937 | 0.953 | 0.953 | 0.953 |
| Stride | Kernel Size and | ||||
| Kernel | 5 | 7 | 9 | 11 | 13 |
| PSNR | 38.81 | 38.82 | 38.82 | 38.84 | 38.8 |
| SSIM | 0.952 | 0.952 | 0.952 | 0.953 | 0.952 |
| Stride | Kernel Size | ||||
| Kernel | 3 | 5 | 7 | 9 | 11 |
| PSNR | 38.76 | 38.81 | 38.80 | 38.84 | 38.79 |
| SSIM | 0.952 | 0.952 | 0.952 | 0.953 | 0.952 |
| Test Situation | Gaussian | Real | |||
|---|---|---|---|---|---|
| Configuration | PSNR | SSIM | PSNR | SSIM | |
| Without Pooling | 30.13 | 0.836 | 38.26 | 0.948 | |
| Max Pooling | 30.28 | 0.845 | 38.64 | 0.951 | |
| Avg Pooling After Conv | 30.30 | 0.845 | 38.63 | 0.951 | |
| Final (Avg Pooling Before Conv) | 30.39 | 0.848 | 38.84 | 0.953 | |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Liao, J.-R.; Lin, W.; Chang, L.-W. Adjustable Complexity Transformer Architecture for Image Denoising. Signals 2026, 7, 33. https://doi.org/10.3390/signals7020033
Liao J-R, Lin W, Chang L-W. Adjustable Complexity Transformer Architecture for Image Denoising. Signals. 2026; 7(2):33. https://doi.org/10.3390/signals7020033
Chicago/Turabian StyleLiao, Jan-Ray, Wen Lin, and Li-Wen Chang. 2026. "Adjustable Complexity Transformer Architecture for Image Denoising" Signals 7, no. 2: 33. https://doi.org/10.3390/signals7020033
APA StyleLiao, J.-R., Lin, W., & Chang, L.-W. (2026). Adjustable Complexity Transformer Architecture for Image Denoising. Signals, 7(2), 33. https://doi.org/10.3390/signals7020033

