PID-NET: A Novel Parallel Image-Dehazing Network
Abstract
:1. Introduction
- Novel parallel architecture: This paper proposes PID-Net, a Parallel Image-Dehazing Network, leveraging a ViT for global context and a multi-layer CNN for local detail features. This synergistic approach enhances dehazing capabilities by harnessing the complementary strengths of a CNN’s localized feature representation and a ViT’s holistic scene understanding capabilities.
- Lightweight Attention Module: This paper introduces a down-sampling method in ViT’s multi-head attention to reduce model complexity. A deep convolutional network compensates for local detail loss, maintaining performance while reducing parameters.
- Redundant Feature Filtering Module: This paper proposes a Redundant Feature Filtering Module to effectively remove redundant features like noise and haze, enabling deeper feature extraction and more accurate image restoration.
- Extensive experimental validation: This paper validates PID-Net’s effectiveness on multiple public datasets, demonstrating superior performance compared to other dehazing models and through ablation studies.
2. Related Work
2.1. Prior-Based Dehazing Methods
2.2. CNN-Based Dehazing Methods
2.3. Transformer-Based Dehazing Methods
3. Overall Architecture
3.1. Parallel Feature Extraction Module (PFEM)
3.2. Lightweight Attention Module (LAM)
3.3. Redundant Feature Filtering Module (RFFM)
3.4. Loss Functions
4. Experiments
4.1. Experimental Setup
4.2. Datasets
- DENSE-HAZE: This dataset carefully collected and organized 55 pairs of real haze images and their corresponding smog-free images. The dehaze scene with uniform distribution of haze was simulated.
- NH-HAZE: This dataset comprised images depicting non-homogeneous atmospheric conditions, including 55 paired samples of haze scenes from the real world, corresponding to clear images of the ground. The hazy images were collected under the real hazy conditions created by the professional haze machine, and the haze distribution was uneven, offering the advantage of high fidelity in simulating real-world haze conditions.
- RESIDE-IN and RESIDE-OUT: The RESIDE-IN and RESIDE-OUT datasets were derived from the RESIDE-6K dataset [34], which includes both synthetic and real-world images intended to simulate various weather conditions. This paper selected two scenes from this dataset, indoor and outdoor, which together had a total of 6000 image pairs.
4.3. Comparative Methods and Evaluation Metrics
4.4. Experimental Results
4.5. Ablation Study
- Base-LAM: After replacing the standard multi-head attention mechanism with LAM, the number of parameters decreased from 30.7 M to 28.6 M and FLOPs were reduced from 24.35 G to 18.71 G, while PSNR and SSIM improved from 12.527 and 0.462 to 13.576 and 0.509, respectively. This demonstrates that the LAM module not only reduces the computational cost but also improves the performance of the model through the CNN compensation mechanism, realizing the double optimization of efficiency and accuracy.
- Base+Multi-CNN: Incorporating a multi-scale convolutional network into the Base model led to a significant improvement in the PSNR and SSIM, increasing from 13.576 and 0.509 to 17.609 and 0.633, respectively. This validates the effectiveness of a multi-scale CNN in capturing local haze features and enhancing detail restoration. The parallel processing of local and global features significantly improved dehazing performance, albeit with a corresponding increase in parameters and FLOPs.
- Base+Multi-CNN+RFFM: With the addition of the RFFM, the PSNR and SSIM further increased to 18.625 and 0.675, respectively. This indicates that the RFFM effectively filters out redundant features, generating cleaner dehazed images. By suppressing noise and haze-related information, this module enables the model to focus on learning more discriminative dehazing features.
- Ours: The best performance was achieved by integrating LAM, Multi-CNN, and RFFM into the Base model, with the PSNR and SSIM reaching 20.358 and 0.745, respectively. Analysis of the model’s parameters revealed that, despite several enhancements, the number of parameters had only increased slightly and the FLOPs had decreased instead compared with the Base model. The reason was that the down-sampling operation reduced the size of the features in the middle layer, thereby reducing the computational load of the model. This result further proves the efficiency provided by incorporating LAM.
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Guo, Q.; Feng, X.; Xue, P.; Sun, S.; Li, X. Dual-domain multi-scale feature extraction for image dehazing. Multimed. Syst. 2025, 31, 1–15. [Google Scholar] [CrossRef]
- He, K.; Sun, J.; Tang, X. Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 2341–2353. [Google Scholar] [PubMed]
- Zhu, Q.; Mai, J.; Shao, L. A fast single image haze removal algorithm using color attenuation prior. IEEE Trans. Image Process. 2015, 24, 3522–3533. [Google Scholar] [PubMed]
- Zhou, Y.; Chen, Z.; Li, P.; Song, H.; Chen, C.P.; Sheng, B. FSAD-Net: Feedback spatial attention dehazing network. IEEE Trans. Neural Netw. Learn. Syst. 2022, 34, 7719–7733. [Google Scholar] [CrossRef]
- Zhang, D.; Wang, Y.; Meng, L.; Yan, J.; Qin, C. Adaptive critic design for safety-optimal FTC of unknown nonlinear systems with asymmetric constrained-input. ISA Trans. 2024, 155, 309–318. [Google Scholar] [CrossRef]
- Zhao, X.; Xu, F.; Liu, Z. TransDehaze: Transformer-enhanced texture attention for end-to-end single image dehaze. Vis. Comput. 2024, 41, 1621–1635. [Google Scholar] [CrossRef]
- Chen, D.; He, M.; Fan, Q.; Liao, J.; Zhang, L.; Hou, D.; Yuan, L.; Hua, G. Gated context aggregation network for image dehazing and deraining. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa Village, HI, USA, 7–11 January 2019; pp. 1375–1383. [Google Scholar]
- Zhang, S.; He, F. DRCDN: Learning deep residual convolutional dehazing networks. Vis. Comput. 2020, 36, 1797–1808. [Google Scholar] [CrossRef]
- Chen, Z.; He, Z.; Lu, Z.M. DEA-Net: Single image dehazing based on detail-enhanced convolution and content-guided attention. IEEE Trans. Image Process. 2024, 33, 1002–1015. [Google Scholar] [CrossRef]
- Song, Y.; He, Z.; Qian, H.; Du, X. Vision transformers for single image dehazing. IEEE Trans. Image Process. 2023, 32, 1927–1941. [Google Scholar] [CrossRef]
- Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
- Guo, C.L.; Yan, Q.; Anwar, S.; Cong, R.; Ren, W.; Li, C. Image dehazing transformer with transmission-aware 3D position embedding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5812–5820. [Google Scholar]
- Ling, P.; Chen, H.; Tan, X.; Jin, Y.; Chen, E. Single image dehazing using saturation line prior. IEEE Trans. Image Process. 2023, 32, 3238–3253. [Google Scholar] [CrossRef]
- Cui, Z.; Wang, N.; Su, Y.; Zhang, W.; Lan, Y.; Li, A. ECANet: Enhanced context aggregation network for single image dehazing. Signal Image Video Process. 2023, 17, 471–479. [Google Scholar] [CrossRef]
- Zhang, D.; Hao, X.; Wang, D.; Qin, C.; Zhao, B.; Liang, L.; Liu, W. An efficient lightweight convolutional neural network for industrial surface defect detection. Artif. Intell. Rev. 2023, 56, 10651–10677. [Google Scholar] [CrossRef]
- Qu, Y.; Chen, Y.; Huang, J.; Xie, Y. Enhanced pix2pix dehazing network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8160–8168. [Google Scholar]
- Luo, P.; Xiao, G.; Gao, X.; Wu, S. LKD-Net: Large kernel convolution network for single image dehazing. In Proceedings of the 2023 IEEE International Conference on Multimedia and Expo (ICME), Brisbane, Australia, 10–14 July 2023; pp. 1601–1606. [Google Scholar]
- Yi, W.; Dong, L.; Liu, M.; Hui, M.; Kong, L.; Zhao, Y. SID-Net: Single image dehazing network using adversarial and contrastive learning. Multimed. Tools Appl. 2024, 83, 71619–71638. [Google Scholar] [CrossRef]
- Gao, G.; Cao, J.; Bao, C.; Hao, Q.; Ma, A.; Li, G. A novel transformer-based attention network for image dehazing. Sensors 2022, 22, 3428. [Google Scholar] [CrossRef]
- Yang, Y.; Zhang, H.; Wu, X.; Liang, X. MSTFDN: Multi-scale transformer fusion dehazing network. Appl. Intell. 2023, 53, 5951–5962. [Google Scholar] [CrossRef]
- Song, T.; Fan, S.; Li, P.; Jin, J.; Jin, G.; Fan, L. Learning an effective transformer for remote sensing satellite image dehazing. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar] [CrossRef]
- Qiu, Y.; Zhang, K.; Wang, C.; Luo, W.; Li, H.; Jin, Z. Mb-taylorformer: Multi-branch efficient transformer expanded by taylor formula for image dehazing. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 12802–12813. [Google Scholar]
- Wang, Y.; Yan, X.; Wang, F.L.; Xie, H.; Yang, W.; Zhang, X.P.; Qin, J.; Wei, M. UCL-Dehaze: Toward real-world image dehazing via unsupervised contrastive learning. IEEE Trans. Image Process. 2024, 33, 1361–1374. [Google Scholar] [CrossRef]
- Xu, J.; Chen, Z.X.; Luo, H.; Lu, Z.M. An efficient dehazing algorithm based on the fusion of transformer and convolutional neural network. Sensors 2022, 23, 43. [Google Scholar] [CrossRef]
- Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
- Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
- Zhang, T.; Li, L.; Zhou, Y.; Liu, W.; Qian, C.; Hwang, J.N.; Ji, X. Cas-vit: Convolutional additive self-attention vision transformers for efficient mobile applications. arXiv 2024, arXiv:2408.03703. [Google Scholar]
- Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
- Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part II 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 694–711. [Google Scholar]
- Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
- Yu, Y.; Liu, H.; Fu, M.; Chen, J.; Wang, X.; Wang, K. A two-branch neural network for non-homogeneous dehazing via ensemble learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 193–202. [Google Scholar]
- Zhao, H.; Gallo, O.; Frosio, I.; Kautz, J. Loss functions for image restoration with neural networks. IEEE Trans. Comput. Imaging 2016, 3, 47–57. [Google Scholar] [CrossRef]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- Li, B.; Ren, W.; Fu, D.; Tao, D.; Feng, D.; Zeng, W.; Wang, Z. Benchmarking single-image dehazing and beyond. IEEE Trans. Image Process. 2018, 28, 492–505. [Google Scholar] [CrossRef]
- Ancuti, C.O.; Ancuti, C.; Timofte, R. NH-HAZE: An image dehazing benchmark with non-homogeneous hazy and haze-free images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 444–445. [Google Scholar]
- Ancuti, C.O.; Ancuti, C.; Sbert, M.; Timofte, R. Dense-haze: A benchmark for image dehazing with dense-haze and haze-free images. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 1014–1018. [Google Scholar]
- Yang, Y.; Wang, C.; Liu, R.; Zhang, L.; Guo, X.; Tao, D. Self-augmented unpaired image dehazing via density and depth decomposition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–22 June 2022; pp. 2037–2046. [Google Scholar]
- Zheng, Y.; Zhan, J.; He, S.; Dong, J.; Du, Y. Curricular contrastive regularization for physics-aware single image dehazing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 5785–5794. [Google Scholar]
- Kumar, M.R.; Linslal, C.; Pillai, V.M.; Krishna, S.S. Color image encryption and decryption based on jigsaw transform employed at the input plane of a double random phase encoding system. In Proceedings of the International Congress on Ultra Modern Telecommunications and Control Systems, Moscow, Russia, 18–20 October 2010; pp. 860–862. [Google Scholar]
- Singh, P.; Agrawal, B.; Chaurasiya, R.K.; Acharya, B. Low-area and high-speed hardware architectures of KLEIN lightweight block cipher for image encryption. J. Electron. Imaging 2023, 32, 013012. [Google Scholar] [CrossRef]
- Kolokoltsev, O.; Gómez-Arista, I.; Treviño-Palacios, C.G.; Qureshi, N.; Mejia-Uriarte, E.V. Swept source OCT beyond the coherence length limit. IEEE J. Sel. Top. Quantum Electron. 2016, 22, 222–227. [Google Scholar] [CrossRef]
Parameter | Value |
---|---|
Learning rate | 0.0001 |
Batch size | 8 |
Epoch | 300 |
Optimization algorithm | Adam |
Train size | 0.8 |
Test size | 0.1 |
Validation size | 0.1 |
Methods | DENSE-HAZE | NH-HAZE | RESIDE-OUT | RESIDE-IN | Average | Overhead | ||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
PSNR | SSIM | PSNR | SSIM | PSNR | SSIM | PSNR | SSIM | PSNR | SSIM | # Param. (M) | # FLOPs (G) | |
GCANet | 14.536 | 0.418 | 16.563 | 0.526 | 23.821 | 0.759 | 29.591 | 0.790 | 21.128 | 0.623 | 0.7 | 0.279 |
EPDN | 16.531 | 0.461 | 15.037 | 0.531 | 25.015 | 0.730 | 29.853 | 0.866 | 21.609 | 0.647 | 17.38 | 4.92 |
D4 | 11.607 | 0.393 | 12.419 | 0.553 | 23.361 | 0.691 | 28.618 | 0.861 | 19.002 | 0.625 | 10.7 | 2.51 |
SLP | 9.727 | 0.433 | 11.031 | 0.521 | 23.597 | 0.706 | 28.909 | 0.808 | 18.320 | 0.617 | - | - |
C2PNet | 13.726 | 0.371 | 12.837 | 0.595 | 24.808 | 0.713 | 30.005 | 0.854 | 20.344 | 0.633 | 7.17 | 21.72 |
DEA-Net | 17.290 | 0.526 | 18.802 | 0.660 | 24.831 | 0.716 | 30.231 | 0.904 | 22.789 | 0.702 | 3.65 | 33.18 |
TransDehaze | 18.354 | 0.559 | 18.390 | 0.621 | 25.026 | 0.750 | 30.314 | 0.912 | 23.021 | 0.711 | 30.62 | 28.51 |
Ours | 20.162 | 0.538 | 20.401 | 0.665 | 25.808 | 0.782 | 30.916 | 0.915 | 24.322 | 0.725 | 32.5 | 23.97 |
Net Name | PSNR | SSIM | # Param. (M) | # FLOPs (G) |
---|---|---|---|---|
Base | 12.527 | 0.462 | 30.7 | 24.35 |
Base-LAM | 13.576 | 0.509 | 28.6 | 18.71 |
Base+Multi-CNN | 17.609 | 0.633 | 36.7 | 26.93 |
Base+Multi-CNN+RFFM | 18.625 | 0.675 | 37.3 | 28.16 |
The complete framework proposed (Ours) | 20.358 | 0.745 | 32.5 | 23.97 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liu, W.; Zhou, Y.; Zhang, D.; Qin, Y. PID-NET: A Novel Parallel Image-Dehazing Network. Electronics 2025, 14, 1906. https://doi.org/10.3390/electronics14101906
Liu W, Zhou Y, Zhang D, Qin Y. PID-NET: A Novel Parallel Image-Dehazing Network. Electronics. 2025; 14(10):1906. https://doi.org/10.3390/electronics14101906
Chicago/Turabian StyleLiu, Wei, Yi Zhou, Dehua Zhang, and Yi Qin. 2025. "PID-NET: A Novel Parallel Image-Dehazing Network" Electronics 14, no. 10: 1906. https://doi.org/10.3390/electronics14101906
APA StyleLiu, W., Zhou, Y., Zhang, D., & Qin, Y. (2025). PID-NET: A Novel Parallel Image-Dehazing Network. Electronics, 14(10), 1906. https://doi.org/10.3390/electronics14101906