Integrating Frequency-Spatial Features for Energy-Efficient OPGW Target Recognition in UAV-Assisted Mobile Monitoring
Abstract
1. Introduction
- MC-GAP for Small-Target Enhancement: We develop an MC-GAP module that aggregates features from different receptive fields via multi-scale convolutions, followed by concatenation, fusion, and global average pooling augmented with spectrally motivated features. By jointly exploiting spatial and frequency cues in a multi-scale manner, MC-GAP strengthens the representation of fine textures and global context for small OPGW targets, leading to notably improved detection accuracy under complex backgrounds.
- Hybrid Gating for Frequency–Spatial Feature Balancing: We design a hybrid gating mechanism that combines global learnable scalars (, ) with spatial-adaptive gate maps to dynamically weight frequency-enhanced and spatial-enhanced features. The global scalars provide coarse-grained balance control, while the gate maps enable pixel-wise adaptivity. Together with residual connections that preserve the original feature information, this scheme enriches feature diversity and robustness, and alleviates the limitations of conventional convolutional layers constrained by single receptive fields and limited feature expressiveness.
2. Related Work
2.1. Transmission Line Inspection Technologies
2.2. Learning-Based Object Detection
3. OPGW-DETR with Frequency-Selective Spatial Feature Enhancement for UAV Inspection Images
3.1. Multi-Scale Convolution-Based Global Average Pooling (MC-GAP)
3.1.1. Multi-Scale Parallel Convolution Processing
- Step 1: Parallel Multi-scale Convolution. Three convolutional layers with kernel sizes , , and are applied in parallel to extract multi-scale spatial features. All convolutions include Batch Normalization and GELU activation:The convolution captures point-wise information, the convolution aggregates local fine textures enhancing the cable’s fibrous details, and the convolution integrates semantic and textural continuity over a larger spatial range.
- Step 2: Channel Concatenation. The three branches are concatenated along the channel dimension:yielding .
- Step 3: Feature Fusion. A convolution is used to fuse and reduce the dimensionality back to D channels:satisfying .
3.1.2. Approximate Extraction of Low-Frequency Information
3.2. Hybrid Gating Mechanism and Residual Connection
- Spatial Adaptivity: The map varies across spatial locations, enabling the model to emphasize frequency features in regions where global structural information is crucial while preserving spatial details where fine textures dominate;
- Global Balance: The scalar parameters and provide coarse-grained control over the relative importance of frequency versus spatial pathways across the entire feature map, allowing the model to learn task-specific optimal weighting strategies.
4. Numerical Results
4.1. Experimental Setup
4.2. Detection Performance and Efficiency Comparison
- Low-complexity regime (<50 GFLOPs): Here, D-FINE variants achieve higher mAP per GFLOP. Our model’s architectural components (e.g., MC-GAP, hybrid gating) incur fixed overhead that cannot be amortized at very low compute budgets, resulting in suboptimal efficiency.
- Medium-complexity regime (50–150 GFLOPs): OPGW-DETR demonstrates clear superiority. With sufficient capacity, the feature enhancement and attention mechanisms effectively capture discriminative cues from small, cluttered targets, yielding up to +3.9% mAP gain over baselines at comparable GFLOPs.
- High-complexity regime (>150 GFLOPs): Performance gains saturate. Over-parameterized M/X variants show signs of overfitting and degraded gradient flow, limiting further improvements despite increased computation.
4.3. Performance Gains from MC-GAP and Hybrid Gating
- Low-level features () require substantial frequency-selective enhancement () to amplify fine-grained textures and edge information. The relatively balanced ratio indicates that low-level representations benefit nearly equally from both domains. The high absolute values () suggest that both frequency and spatial enhancements are necessary to compensate for the inherently limited semantic discriminability of low-level features.
- Mid-level features () exhibit a pronounced shift toward spatial-domain dominance. While maintaining a high , the substantial reduction in indicates that mid-level semantic abstractions are susceptible to frequency-selective perturbations. At this level, features encode partially hierarchical patterns and compositional structures, relying on spatial coherence rather than high-frequency details. Excessive frequency enhancement () may introduce artifacts that compromise semantic information acquisition.
- High-level features () demonstrate an extreme spatial bias, with frequency enhancement nearly eliminated. The minimal value reflects that high-level semantic representations exhibit minimal dependence on spectrally motivated features while being particularly sensitive to high-frequency noise therein. Notably, also falls below 1.0, indicating that high-level features obtained from the pretrained backbone already possess sufficient representational capacity, requiring only conservative spatial refinement.
4.4. Analysis of Accuracy Improvement and Loss Stabilization in OPGW-DETR
4.5. Frequency-Selective Analysis of Feature Enhancement
- Low-frequency energy (0–) increases from 93.0% to 97.4% (+4.7% relative);
- High-frequency energy (50–100% Nyquist) drops from 2.2% to 0.8% (–63.6%);
- The spectral centroid shifts leftward from 2.33 to 0.88 (–62.3%);
- The 90% cumulative energy threshold moves from 6 to 1 (83.3% reduction), as visualized in Figure 12.
5. Discussion on UAV Images Detection with OPGW-DETR
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Song, J.; Sun, K.; Wu, D.; Cheng, F.; Xu, Z. Fault Analysis and Diagnosis Method for Intelligent Maintenance of OPGW Optical Cable in Power Systems. In Proceedings of the InInternational Conference on Applications and Techniques in Cyber Security and Intelligence, Fuyang, China, 19–21 June 2021; pp. 186–191. [Google Scholar]
- Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. In Proceedings of the European Conference on Computer Vision, Online, 23–28 August 2020; pp. 213–229. [Google Scholar]
- Peng, Y.; Li, H.; Wu, P.; Zhang, Y.; Sun, X.; Wu, F. D-FINE: Redefine Regression Task in DETRs as Fine-grained Distribution Refinement. arXiv 2024, arXiv:2410.13842. [Google Scholar] [CrossRef]
- Alhassan, A.B.; Zhang, X.; Shen, H.; Xu, H. Power transmission line inspection robots: A review, trends and challenges for future research. Int. J. Electr. Power Energy Syst. 2020, 118, 105862. [Google Scholar] [CrossRef]
- Ahmed, F.; Mohanta, J.C.; Keshari, A. Power Transmission Line Inspections: Methods, Challenges, Current Status and Usage of Unmanned Aerial Systems. J. Intell. Robot. Syst. 2024, 110, 1–25. [Google Scholar] [CrossRef]
- Ferbin, F.J.; Meganathan, L.; Esha, M.; Thao, N.G.M.; Doss, A.S.A. Power Transmission Line Inspection Using Unmanned Aerial Vehicle—A Review. In Proceedings of the 2023 Innovations in Power and Advanced Computing Technologies (i-PACT), Kuala Lumpur, Malaysia, 8–10 December 2023; pp. 1–5. [Google Scholar]
- Dlamini, T.; Vilakati, S. Remote and Rural Connectivity: Infrastructure and Resource Sharing Principles. arXiv 2021, arXiv:2110.01910. [Google Scholar] [CrossRef]
- Taheri, T.; Nilsson, R.; van de Beek, J. The Potential of Massive-MIMO on TV Towers for Cellular Coverage Extension. Wirel. Commun. Mob. Comput. 2021, 2021, 8164367. [Google Scholar] [CrossRef]
- Nguyen, M.D.; Ho, T.M.; Le, L.B.; Girard, A. UAV Placement and Bandwidth Allocation for UAV Based Wireless Networks. In Proceedings of the 2019 IEEE Global Communications Conference (GLOBECOM), Waikoloa, HI, USA, 9–13 December 2019; pp. 1–6. [Google Scholar] [CrossRef]
- Yang, Z.; Pan, C.; Wang, K.; Shikh-Bahaei, M. Energy Efficient Resource Allocation in UAV-Enabled Mobile Edge Computing Networks. IEEE Trans. Wirel. Commun. 2019, 18, 4576–4589. [Google Scholar] [CrossRef]
- Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of Deep Learning: Concepts, CNN Architectures, Challenges, Applications, Future Directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
- Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
- Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
- Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar] [CrossRef]
- Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
- Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
- Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8759–8768. [Google Scholar]
- Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2020; pp. 10778–10787. [Google Scholar]
- Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable Transformers for End-to-End Object Detection. arXiv 2021, arXiv:2010.04159. [Google Scholar] [CrossRef]
- Meng, D.; Chen, X.; Fan, Z.; Zeng, G.; Li, H.; Yuan, Y.; Sun, L.; Wang, J. Conditional DETR for Fast Training Convergence. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 10–17 October 2021; pp. 3631–3640. [Google Scholar]
- Liu, S.; Li, F.; Zhang, H.; Yang, X.; Qi, X.; Su, H.; Zhu, J.; Zhang, L. DAB-DETR: Dynamic Anchor Boxes are Better Queries for DETR. arXiv 2022, arXiv:2201.12329. [Google Scholar] [CrossRef]
- Wang, Y.; Zhang, X.; Yang, T.; Sun, J. Anchor DETR: Query Design for Transformer-Based Object Detection. arXiv 2022, arXiv:2109.07107. [Google Scholar] [CrossRef]
- Yao, Z.; Ai, J.; Li, B.; Zhang, C. Efficient DETR: Improving End-to-End Object Detector with Dense Prior. arXiv 2021, arXiv:2104.01318. [Google Scholar] [CrossRef]
- Li, F.; Zhang, H.; Liu, S.; Guo, J.; Ni, L.M.; Zhang, L. DN-DETR: Accelerate DETR Training by Introducing Query DeNoising. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 13619–13627. [Google Scholar]
- Zhang, H.; Li, F.; Liu, S.; Zhang, L.; Su, H.; Zhu, J.; Ni, L.M.; Shum, H.Y. DINO: DETR with Improved DeNoising Anchor Boxes for End-to-End Object Detection. arXiv 2022, arXiv:2203.03605. [Google Scholar] [CrossRef]
- Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. DETRs Beat YOLOs on Real-time Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 16965–16974. [Google Scholar]
- Wei, J.; Shen, W.; Hu, T.; Liu, Q.; Yu, J.; Huang, J. Dynamic-DETR: Dynamic Perception for RT -DETR. In Proceedings of the 2024 6th International Conference on Robotics and Computer Vision (ICRCV 2024), Wuxi, China, 20–22 September 2024; pp. 28–32. [Google Scholar]
- Zhang, H.; Liu, K.; Gan, Z.; Zhu, G.N. UAV-DETR: Efficient End-to-End Object Detection for Unmanned Aerial Vehicle Imagery. arXiv 2025, arXiv:2501.01855. [Google Scholar] [CrossRef]
- Huang, S.; Lu, Z.; Cun, X.; Yu, Y.; Zhou, X.; Shen, X. DEIM: DETR with Improved Matching for Fast Convergence. In Proceedings of the Computer Vision and Pattern Recognition Conference, Suzhou, China, 10–13 November 2025; pp. 15162–15171. [Google Scholar]
- Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar] [CrossRef]
- Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. arXiv 2018, arXiv:1807.06521. [Google Scholar] [CrossRef]
- Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. arXiv 2020, arXiv:1910.03151. [Google Scholar] [CrossRef]
- Gao, S.H.; Cheng, M.M.; Zhao, K.; Zhang, X.Y.; Yang, M.H.; Torr, P. Res2Net: A New Multi-Scale Backbone Architecture. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 652–662. [Google Scholar] [CrossRef]
- Cao, S.; Long, S.; Zhu, W.; Liao, F.; Yuan, Z.; Guan, X. Multiple Layers Global Average Pooling Fusion. Procedia Comput. Sci. 2023, 153, 3–10. [Google Scholar] [CrossRef]
- Brigham, E.O.; Morrow, R.E. The fast Fourier transform. IEEE Spectr. 1967, 4, 63–70. [Google Scholar] [CrossRef]
- Lin, M.; Chen, Q.; Yan, S. Network in Network. arXiv 2014, arXiv:1312.4400. [Google Scholar] [CrossRef]
- Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. arXiv 2019, arXiv:1711.05101. [Google Scholar] [CrossRef]
- Wang, C.Y.; Yeh, I.H.; Liao, H.Y.M. YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; pp. 1–21. [Google Scholar]
- Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. YOLOv10: Real-Time End-to-End Object Detection. Adv. Neural Inf. Process. Syst. (NeurIPS) 2024, 37, 107984–108011. [Google Scholar]
- Jiang, B.; Luo, R.; Mao, J.; Xiao, T.; Jiang, Y. Acquisition of Localization Confidence for Accurate Object Detection. In Proceedings of the Computer Vision—ECCV 2018; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2018; Volume 11218, pp. 784–799. [Google Scholar] [CrossRef]
- Küppers, F.; Kronenberger, J.; Shantia, A.; Haselhoff, A. Multivariate Confidence Calibration for Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 1322–1330. [Google Scholar] [CrossRef]
- Aliferis, C.; Simon, G. Overfitting, Underfitting and General Model Overconfidence and Under-Performance Pitfalls and Best Practices in Machine Learning and AI. In Artificial Intelligence and Machine Learning in Health Care and Medical Sciences; Health Informatics; Springer: Cham, Switzerland, 2024; pp. 235–268. [Google Scholar] [CrossRef]
- Hekler, A.; Kuhn, L.; Buettner, F. Beyond Overconfidence: Foundation Models Redefine Calibration in Deep Neural Networks. arXiv 2025, arXiv:2506.09593. [Google Scholar] [CrossRef]













| Category ID | Name | Training Set | Validation Set | Test Set | Total Instances |
|---|---|---|---|---|---|
| 0 | O1 | 1323 (34.9%) | 448 (35.4%) | 451 (35.0%) | 2222 |
| 1 | D1 | 1251 (33.0%) | 413 (32.6%) | 423 (32.8%) | 2087 |
| 3 | O2 | 681 (18.0%) | 226 (17.8%) | 236 (18.3%) | 1143 |
| 4 | D2 | 533 (14.1%) | 180 (14.2%) | 179 (13.9%) | 892 |
| Total | — | 3788 | 1267 | 1289 | 6344 |
| Statistic | Width (px) | Height (px) | Area (px2) | Aspect Ratio |
|---|---|---|---|---|
| Minimum | 2.1 | 1.8 | 7.4 | 0.08 |
| 5th Percentile | 12.4 | 3.1 | 48.2 | 1.87 |
| 25th Percentile | 18.3 | 4.2 | 89.3 | 2.47 |
| Median (50%) | 45.7 | 7.8 | 428.6 | 5.86 |
| 75th Percentile | 98.4 | 15.2 | 1847.3 | 12.31 |
| 95th Percentile | 234.6 | 38.9 | 8562.1 | 28.45 |
| Maximum | 578.2 | 162.4 | 52,387.9 | 89.73 |
| Mean | 72.8 | 12.6 | 1986.4 | 9.87 |
| Std. Dev. | 84.3 | 14.9 | 4123.7 | 11.24 |
| CV (%) | 115.8% | 118.3% | 207.6% | 113.9% |
| Relative Size | Area Range | Instance Count | Percentage | Description |
|---|---|---|---|---|
| Tiny Targets | <0.1% | 1582 | 24.9% | Near resolution limit |
| Small Targets | – | 3124 | 49.2% | Primary detection objects |
| Medium Targets | 0.5%–2.0% | 1398 | 22.0% | Relatively easy to detect |
| Large Targets | >2.0% | 240 | 3.8% | Close-range captures |
| Image Region | X-Range | Y-Range | Count | Percentage |
|---|---|---|---|---|
| Center Region | 0.33–0.67 | 0.33–0.67 | 2847 | 44.9% |
| Upper Center | 0.33–0.67 | 0.00–0.33 | 523 | 8.2% |
| Lower Center | 0.33–0.67 | 0.67–1.00 | 687 | 10.8% |
| Left Center | 0.00–0.33 | 0.33–0.67 | 412 | 6.5% |
| Right Center | 0.67–1.00 | 0.33–0.67 | 398 | 6.3% |
| Corner Regions | — | — | 1477 | 23.3% |
| Boxes per Image | Image Count | Percentage | Scene Characteristics |
|---|---|---|---|
| 1 box | 1287 | 41.0% | Single conductor/long distance/occlusion |
| 2 boxes | 1024 | 32.6% | Typical dual-wire configuration |
| 3 boxes | 537 | 17.1% | Multi-span segments/complex towers |
| 4 boxes | 198 | 6.3% | Tower connection points/multi-circuit |
| ≥5 boxes | 91 | 2.9% | Extremely complex scenarios |
| Model Scale | Backbone Size | Returned Stages | D | ||
|---|---|---|---|---|---|
| N | B0 | , | (512, 1024, –) | (16, 32, –) | 128 |
| S | B0 | , , | (256, 512, 1024) | (8, 16, 32) | 256 |
| M | B2 | , , | (384, 768, 1536) | (8, 16, 32) | 256 |
| X | B5 | , , | (512, 1024, 2048) | (8, 16, 32) | 384 |
| Model | GPU Power (W) | Avg. Memory (MB) | GPU Utilization (%) |
|---|---|---|---|
| Peak/Idle | Peak/Idle | ||
| RT-DETR-R50 | 280/153 | 15,130 | 98/53 |
| RT-DETR-R101 | 280/158 | 18,252 | 100/55 |
| RT-DETR-X | 282/169 | 18,138 | 100/57 |
| UAV-DETR-Ev2 | 278/168 | 10,544 | 100/42 |
| UAV-DETR-R18 | 246/151 | 9044 | 100/39 |
| UAV-DETR-R50 | 259/159 | 15,526 | 100/43 |
| YOLOv9-T | 175/110 | 3006 | 47/22 |
| YOLOv9-S | 225/116 | 3688 | 62/39 |
| YOLOv10-S | 231/118 | 4422 | 77/50 |
| YOLOv10-N | 200/115 | 3510 | 57/42 |
| D-FINE-N | 233/68 | 6408 | 72/23 |
| D-FINE-S | 246/66 | 10,524 | 79/17 |
| OPGW-DETR-N (Ours) | 265/66 | 7584 | 78/27 |
| OPGW-DETR-S (Ours) | 268/66 | 18,376 | 99/32 |
| Model | InputSize | mAP50 | mAP50∼95 | Param(M) | GFLOPs |
|---|---|---|---|---|---|
| RT-DETR-R50 [28] | 640 × 640 | 70.9 | 45.2 | 40.8 | 130.5 |
| RT-DETR-R101 [28] | 640 × 640 | 72.8 | 46.7 | 58.9 | 191.4 |
| RT-DETR-X [28] | 640 × 640 | 74.1 | 46.3 | 65.5 | 222.5 |
| UAV-DETR-Ev2 [30] | 640 × 640 | 71.2 | 43.0 | 12.6 | 44.0 |
| UAV-DETR-R18 [30] | 640 × 640 | 72.6 | 44.7 | 20.5 | 73.9 |
| UAV-DETR-R50 [30] | 640 × 640 | 75.4 | 47.1 | 43.4 | 166.4 |
| YOLOv9-T [40] | 640 × 640 | 68.5 | 41.7 | 1.9 | 7.9 |
| YOLOv9-S [40] | 640 × 640 | 72.8 | 46.8 | 6.9 | 27.4 |
| YOLOv10-S [41] | 640 × 640 | 69.6 | 44.3 | 7.7 | 24.8 |
| YOLOv10-N [41] | 640 × 640 | 64.0 | 39.0 | 2.6 | 8.4 |
| OPGW-DETR-N (Ours) | 640 × 640 | 71.3 | 42.2 | 4.8 | 9.7 |
| OPGW-DETR-S (Ours) | 640 × 640 | 78.5 | 50.6 | 17.2 | 68.9 |
| OPGW-DETR- (Ours) | 640 × 640 | 78.3 | 49.8 | 25.8 | 100.4 |
| OPGW-DETR- (Ours) | 640 × 640 | 77.3 | 48.8 | 75.6 | 301.3 |
| Baseline | MC-GAP | Gate | Params(M) | GFLOPs | mAP50 | mAP50∼95 |
|---|---|---|---|---|---|---|
| ✓ | 9.7 | 24.8 | 74.6 | 46.3 | ||
| ✓ | ✓ | 17.1 | 67.8 | 78.4 | 49.4 | |
| ✓ | ✓ | ✓ | 17.2 | 68.9 | 78.5 | 50.6 |
| Features | Level | / | ||
|---|---|---|---|---|
| Low | ||||
| Mid | ||||
| High |
| Metric | Projected | UAV-Enhanced | Relative |
|---|---|---|---|
| Low-freq Energy (%) | 93.0% | 97.4% | +4.7% |
| 90% Energy Cutoff () | 6 | 1 | –83.3% |
| High-freq Energy (%) | 2.2% | 0.8% | –63.6% |
| Spectral Centroid () | 2.33 | 0.88 | –62.3% |
| Spectral Spread () | 6.82 | 4.28 | –37.3% |
| Mid-freq Energy (%) | 4.8% | 1.8% | –62.8% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Huang, L.; Ren, X.; Qu, D.; Li, L.; Xu, J. Integrating Frequency-Spatial Features for Energy-Efficient OPGW Target Recognition in UAV-Assisted Mobile Monitoring. Sensors 2026, 26, 506. https://doi.org/10.3390/s26020506
Huang L, Ren X, Qu D, Li L, Xu J. Integrating Frequency-Spatial Features for Energy-Efficient OPGW Target Recognition in UAV-Assisted Mobile Monitoring. Sensors. 2026; 26(2):506. https://doi.org/10.3390/s26020506
Chicago/Turabian StyleHuang, Lin, Xubin Ren, Daiming Qu, Lanhua Li, and Jing Xu. 2026. "Integrating Frequency-Spatial Features for Energy-Efficient OPGW Target Recognition in UAV-Assisted Mobile Monitoring" Sensors 26, no. 2: 506. https://doi.org/10.3390/s26020506
APA StyleHuang, L., Ren, X., Qu, D., Li, L., & Xu, J. (2026). Integrating Frequency-Spatial Features for Energy-Efficient OPGW Target Recognition in UAV-Assisted Mobile Monitoring. Sensors, 26(2), 506. https://doi.org/10.3390/s26020506

