Object Detection Method Based on Polarimetric Features and PFOD-Net Under Adverse Weather Conditions
Abstract
1. Introduction
- Polarimetric Transformation and Feature Enhancement (PTFE) module: In response to the variations in light intensity across different polarization angles, an input processing module incorporating prior knowledge from the polarimetric domain was designed. Through multi-scale polarization channel fusion and weight-sharing mechanisms, this module not only accelerates network convergence but also achieves efficient image feature enhancement and extraction with minimal parametric overhead.
- Dynamic Coordinate Attention Spatial Pyramid Pooling (DCASPP) module: To address feature redundancy in the original SPPF (Spatial Pyramid Pooling—Fast) during the later stages of the network, we integrated a dual-branch dynamic pooling structure with a Coordinate Attention mechanism, implemented in parallel with the traditional max-pooling layer. This refinement maintains the robustness of the original max-pooling layer while significantly enhancing the spatial pyramid’s capacity for multi-scale information fusion and effective feature extraction through average pooling and Coordinate Attention recalibration.
- Furthermore, GhostNet was introduced into the network to achieve further lightweighting. Comparative experiments across various lightweight architectures validated the effectiveness of GhostNet within PFOD-Net (Polarization Feature Object Detection Network). By replacing feature extraction modules in local or global network components with GhostNet, the results demonstrate that utilizing GhostNet in the backbone effectively balances detection accuracy and inference speed.
2. Related Works
2.1. YOLOv8 Object Detection Method
2.2. Polarimetric Imaging Methods
2.3. Attention Mechanisms
- SE Module [18]—Hu et al. proposed the SE attention module. The basic idea is to obtain channel-level statistics through global average pooling and use fully connected layers for feature compression and recalibration, thereby achieving adaptive adjustment of weights for each channel.
- ECA Module [19]—Wang et al. further reduced parameter redundancy based on SE and proposed the ECA module. ECA uses 1D convolution instead of fully connected layers, effectively avoiding the introduction of excessive parameters and significantly reducing computational complexity while maintaining performance improvements.
- CBAM [20]—Sanghyun extended the attention mechanism to the joint modeling of channel and spatial dimensions and proposed CBAM. CBAM first weights the channel features and then generates a spatial attention map, focusing on the importance of both features and spatial positions simultaneously, making the network more precise in feature selection and target localization.
- CA (Coordinate Attention) [21]—Hou et al. proposed the CA module. It encodes positional information along the horizontal and vertical directions, respectively, through direction-decomposed global pooling operations and embeds this information into the generation process of channel attention. This captures long-range dependencies while preserving precise positional information. The CA module significantly enhances the network’s representation of target structure and spatial layout while remaining lightweight, making it especially suitable for tasks that require simultaneous attention to global and local features.
2.4. Polarimetric Object Detection Datasets
3. Methods
3.1. Overall Model Architecture
- PTFE—We introduced the PTFE module to enhance the model’s feature representation capability in the shallow layers.
- DCASPP—We redesigned the receptive field module by developing the DCASPP.
- Lightweight Backbone Integration—By replacing specific standard convolutions in the backbone network with Ghost convolutions, which require fewer parameters, we achieved a 22.4% reduction in model size at the cost of only a 0.7% decrease in precision.
3.2. PTFE Module
- a.
- Hardware simplification—Eliminates the need for additional 135° polarizers, reducing system complexity.
- b.
- Framework compatibility—Three-channel data structurally aligns with conventional RGB imaging, enabling seamless integration with mainstream vision frameworks.
3.3. DCASPP Module
3.4. Lightweight Convolutions
4. Results
4.1. Implementation Details
4.1.1. Experimental Setup
4.1.2. Training Strategy
4.1.3. Loss Function and Post-Processing
4.1.4. Data Augmentation
4.2. Experiment Datasets
4.3. Evaluation Indicators
4.4. Validation of the Proposed Algorithm
4.4.1. Comparison of Different Image Sources
4.4.2. Comparison Between the Proposed Algorithm and SOTA Methods
4.4.3. Comparison of Different Lightweight Models
4.5. Ablation Experiments
4.6. Grad-CAM Visualization
4.7. Comparison of Detection Performance
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
| AoP | Angle of Polarization |
| DCASPP | Dynamic Coordinate Attention Spatial Pyramid Pooling |
| DoLP | Degree of Linear Polarization |
| FPS | Frames Per Second |
| IoU | Intersection over Union |
| mAP | mean Average Precision |
| PFOD-Net | Polarization Feature Object Detection Network |
| PTFE | Polarimetric Transformation and Feature Enhancement |
| RGB | Red, Green, Blue |
| SOTA | State-of-the-art |
| SPPF | Spatial Pyramid Pooling—Fast |
References
- Yang, K.; Liu, F.; Liang, S.; Xiang, M.; Han, P.; Liu, J.; Dong, X.; Wei, Y.; Wang, B.; Shimizu, K.; et al. Data-Driven Polarimetric Imaging: A Review. Opto-Electron. Sci. 2024, 3, 230042. [Google Scholar] [CrossRef]
- Varghese, R.; Sambath, M. YOLOv8: A Novel Object Detection Algorithm with Enhanced Performance and Robustness. In Proceedings of the 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS), Chennai, India, 18–19 April 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–6. [Google Scholar] [CrossRef]
- Bian, Y.F.; Luo, D.X.; Zhang, M.L. A Review of FPGA Accelerated Computing Methods for YOLO Models. In Proceedings of the 2024 4th International Conference on Internet of Things and Machine Learning (IoTML 2024), Nanchang, China, 9–11 August 2024; p. 9. [Google Scholar] [CrossRef]
- Nafaa, S.; Ashour, K.; Mohamed, R.; Essam, H.; Emad, D.; Elhenawy, M.; Ashqar, H.I.; Hassan, A.A.; Alhadidi, T.I. Advancing Roadway Sign Detection with YOLO Models and Transfer Learning. In Proceedings of the 2024 IEEE 3rd International Conference on Computing and Machine Intelligence (ICMI), Mt Pleasant, MI, USA, 13–14 April 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–4. [Google Scholar] [CrossRef]
- Wang, H.; Shang, J.; Wang, X.; Zhang, Q.; Wang, X.; Li, J.; Wang, Y. RSW-YOLO: A Vehicle Detection Model for Urban UAV Remote Sensing Images. Sensors 2025, 25, 4335. [Google Scholar] [CrossRef] [PubMed]
- Blin, R.; Ainouz, S.; Canu, S.; Meriaudeau, F. Road Scenes Analysis in Adverse Weather Conditions by Polarization-Encoded Images and Adapted Deep Learning. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 27–32. [Google Scholar] [CrossRef]
- Blin, R.; Ainouz, S.; Canu, S.; Meriaudeau, F. A New Multimodal RGB and Polarimetric Image Dataset for Road Scenes Analysis. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 867–876. [Google Scholar] [CrossRef]
- Li, Z.; Zhang, A.; Jiang, Y.; Li, G.; Wang, D.; Wang, W.; Shi, L.; Ji, T.; Liu, F.; Chen, Y. Research, Application, and Progress of Optical Polarization Imaging Technology. Infrared Laser Eng. 2023, 52, 20220808. [Google Scholar] [CrossRef]
- Liu, J.Y.; Li, S.T.; Dian, R.; Song, Z. DT-F Transformer: Dual transpose fusion transformer for polarization image fusion. Inf. Fusion 2024, 106, 102274. [Google Scholar] [CrossRef]
- Shen, Y.; Liu, X.C.; Wang, S.; Huang, F. Real-Time Detection of Low-Altitude Camouflaged Targets Based on Polarization Encoded Images. Acta Armamentarii 2024, 45, 1374–1383. [Google Scholar] [CrossRef]
- Hu, H.F.; Fei, X.T.; Shen, L.H.; Li, X.B. Underwater Image Recovery under Non-Uniform Illumination Based on Polarimetric Imaging. Acta Opt. Sin. 2025, 45, 0629001. [Google Scholar] [CrossRef]
- Tan, A.; Guo, T.; Zhao, Y.; Wang, Y.; Li, X. Object Detection Based on Polarization Image Fusion and Grouped Convolutional Attention Network. Vis. Comput. 2024, 40, 3199–3215. [Google Scholar] [CrossRef]
- Sun, R.; Sun, X.; Chen, F.; Song, Q.; Pan, H. Polarimetric Imaging Detection Using a Convolutional Neural Network with Three-Dimensional and Two-Dimensional Convolutional Layers. Appl. Opt. 2020, 59, 151. [Google Scholar] [CrossRef] [PubMed]
- Huang, F.; Zheng, J.; Liu, X.; Shen, Y.; Chen, J. Polarization of Road Target Detection under Complex Weather Conditions. Sci. Rep. 2024, 14, 30348. [Google Scholar] [CrossRef] [PubMed]
- Dey, J.; Anandan, P.; Rajagopal, S.; Mani, M. Improved Target Detection with YOLOv8 for GAN Augmented Polarimetric Images Using MIRNet Denoising Model. IEEE Access 2024, 12, 166885–166910. [Google Scholar] [CrossRef]
- Jocher, G.; Stoken, A.; Borovec, J.; NanoCode012; ChristopherSTAN; Liu, C.Y.; Laughing; Hogan, A.; lorenzomammana; tkianai; et al. ultralytics/yolov5, v3.0; Zenodo: Geneva, Switzerland, 2020. [CrossRef]
- Zhang, J.C.; Wu, C.Y.; Luo, Y.D.; Li, C.G.; Jiang, N.; Song, Y.C. Research Status and Prospects on Super-Resolution Imaging Technology for Division-of-Focal-Plane Polarimeters (Invited). Infrared Laser Eng. 2025, 54, 20240165. [Google Scholar] [CrossRef]
- Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 31st IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar] [CrossRef]
- Wang, Q.L.; Wu, B.G.; Zhu, P.F.; Li, P.H.; Zuo, W.M.; Hu, Q.H. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 11531–11539. [Google Scholar] [CrossRef]
- Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Computer Vision—ECCV 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2018; Volume 11211, pp. 3–19. [Google Scholar] [CrossRef]
- Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 13708–13717. [Google Scholar] [CrossRef]
- Wang, X.; Ding, J.; Zhang, Z.; Xu, J.; Gao, J. IPNet: Polarization-Based Camouflaged Object Detection via Dual-Flow Network. Eng. Appl. Artif. Intell. 2024, 127, 107303. [Google Scholar] [CrossRef]
- Zhao, Y.Q.; Li, N.; Zhang, P.; Yao, J.X.; Pan, Q. Infrared Polarization Perception and Intelligent Processing. Infrared Laser Eng. 2018, 47, 1102001. [Google Scholar] [CrossRef]
- Liu, Y.; Shi, H.D.; Jiang, H.L.; Li, Y.C.; Wang, C.; Liu, Z.; Li, G.L. Infrared Polarization Properties of Targets with Rough Surface. Chin. Opt. 2020, 13, 459–471. Available online: https://kns.cnki.net/kcms2/article/abstract?v=dKcr_PZ1zcuZNhwQkbLvOM2LvJh5iFdkDrzcSt5lbNd5FpagAWWyV30apjvhtGd4D_8aYinswsVmrsQrMVzle9VorHlVdlaEXL2Gi8keUSkPLv7oGQecMFGtyekQI5NziVGv82f5pGGKTlr45gEzUte1KfmUCOqC33Zy906dsU2Y3qIxtSI6Cg6ML1lFzPWX&uniplatform=NZKPT&language=CHS (accessed on 6 January 2026).
- Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar] [CrossRef]
- Zhang, X.; Zhou, X.; Lin, M.; Sun, J. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 6848–6856. [Google Scholar] [CrossRef]
- Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C.; Xu, C. GhostNet: More Features From Cheap Operations. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1577–1586. [Google Scholar] [CrossRef]
- Chen, J.; Kao, S.; He, H.; Zhuo, W.; Wen, S.; Lee, C.-H.; Chan, S.-H.G. Run, Don’t Walk: Chasing Higher FLOPS for Faster Neural Networks. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 12021–12031. [Google Scholar] [CrossRef]
- Khanam, R.; Hussain, M. YOLOv11: An Overview of the Key Architectural Enhancements. arXiv 2024, arXiv:2410.17725v1. [Google Scholar] [CrossRef]
- Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. Int. J. Comput. Vis. 2020, 128, 336–359. [Google Scholar] [CrossRef]







| Class | Train | Val |
|---|---|---|
| Insects | 453 | 107 |
| Animals | 276 | 70 |
| Products | 230 | 55 |
| Camouflage Net | 39 | 10 |
| Detection Method | Imaging Method | mAP@0.5 on the Validation Set (%) | mAP@0.5 on the Test Set (%) | Params (M) | FPS |
|---|---|---|---|---|---|
| YOLOv5n | RGB | 75.2 | 41.5 | 2.51 | 232 |
| Polarization | 70.4 | 66.8 | |||
| YOLOv8n | RGB | 75.7 | 33.5 | 3.01 | 222 |
| Polarization | 72.4 | 71.4 | |||
| YOLOv11n | RGB | 76.0 | 39.6 | 2.62 | 227 |
| Polarization | 70.2 | 72.5 | |||
| Faster R-CNN | RGB | 77.1 | 42.1 | 41 | 65 |
| Polarization | 69.8 | 70.2 |
| Method | Class | mAP@0.5 (%) | Params (M) | FPS | GFlops |
|---|---|---|---|---|---|
| YOLOv5n | Car | 73.5 | 2.51 | 232 | 4.2G |
| Person | 60.1 | ||||
| All | 66.8 | ||||
| YOLOv8n | Car | 72.9 | 3.01 | 222 | 8.1G |
| Person | 70.0 | ||||
| All | 71.4 | ||||
| YOLOv11n | Car | 74.6 | 2.62 | 227 | 6.3G |
| Person | 70.4 | ||||
| All | 72.5 | ||||
| Faster R-CNN | Car | 70.5 | 41 | 65 | 208G |
| Person | 69.9 | ||||
| All | 70.2 | ||||
| GCAnet | Car | 90.0 | 29.8 | 145 | 12.8G |
| Person | 63.0 | ||||
| All | 76.4 | ||||
| PFOD-Net (ours) | Car | 77.4 | 2.96 | 192 | 8.0G |
| Person | 86 | ||||
| All | 81.7 |
| Method | mAP@0.5 (%) | Params (M) | FPS |
|---|---|---|---|
| YOLOv5n | 64.4 | 2.51 | 232 |
| YOLOv8n | 68.8 | 3.01 | 222 |
| YOLOv11n | 70.1 | 2.62 | 227 |
| GCAnet | 73.8 | 29.8 | 145 |
| PFOD-Net (ours) | 79.0 | 2.96 | 192 |
| Method | mAP@0.5 (%) | Params (M) | FPS |
|---|---|---|---|
| PFOD-Net | 81.7 | 2.96 | 192 |
| PFOD-Net + MobileNet | 71.7 | 2.11 | 227 |
| PFOD-Net + ShuffleNet | 70.1 | 2.06 | 238 |
| PFOD-Net + GhostNet | 81.0 | 2.25 | 210 |
| PFOD-Net + FasterNet | 66.8 | 1.41 | 256 |
| Model Number | Polarized Image | PTFE | DCASPP | GhostNet | mAP@0.5 (%) | Params (M) | FPS |
|---|---|---|---|---|---|---|---|
| 1 | × | × | × | × | 33.5 | 3.01 | 222 |
| 2 | √ | × | × | × | 71.4 | 3.01 | 222 |
| 3 | √ | √ | × | × | 79.3 | 3.01 | 192 |
| 4 | √ | √ | √ | × | 81.7 | 2.96 | 192 |
| 5 | √ | √ | √ | √ | 81.0 | 2.25 | 210 |
| Model Number | Polarized Image | PTFE | DCASPP | GhostNet | mAP@0.5 (%) | Params (M) | FPS |
|---|---|---|---|---|---|---|---|
| 1 | √ | × | × | × | 68.8 | 3.01 | 222 |
| 2 | √ | √ | × | × | 72.0 | 3.01 | 192 |
| 3 | √ | √ | √ | × | 79.0 | 2.96 | 192 |
| 4 | √ | √ | √ | √ | 77.6 | 2.25 | 210 |
| Model Number | mAP@0.5 (%) | Params (M) | FPS | GFlops |
|---|---|---|---|---|
| 1 | 81.7 | 3.01 | 192 | 8.0G |
| 2 | 81.0 | 2.25 | 210 | 6.3G |
| 3 | 65.1 | 1.66 | 277 | 5.6G |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Li, X.; Li, W.; Yan, X.; Wang, W.; Bu, F. Object Detection Method Based on Polarimetric Features and PFOD-Net Under Adverse Weather Conditions. Appl. Sci. 2026, 16, 1698. https://doi.org/10.3390/app16041698
Li X, Li W, Yan X, Wang W, Bu F. Object Detection Method Based on Polarimetric Features and PFOD-Net Under Adverse Weather Conditions. Applied Sciences. 2026; 16(4):1698. https://doi.org/10.3390/app16041698
Chicago/Turabian StyleLi, Xingtao, Wenjuan Li, Xiaoyao Yan, Weifeng Wang, and Fan Bu. 2026. "Object Detection Method Based on Polarimetric Features and PFOD-Net Under Adverse Weather Conditions" Applied Sciences 16, no. 4: 1698. https://doi.org/10.3390/app16041698
APA StyleLi, X., Li, W., Yan, X., Wang, W., & Bu, F. (2026). Object Detection Method Based on Polarimetric Features and PFOD-Net Under Adverse Weather Conditions. Applied Sciences, 16(4), 1698. https://doi.org/10.3390/app16041698
