Next Article in Journal
Field Explosives Detectors—Current Status and Development Prospects
Previous Article in Journal
Towards Optimal Sensor Placement for Cybersecurity: An Extensible Model for Defensive Cybersecurity Sensor Placement Evaluation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unmanned Airborne Target Detection Method with Multi-Branch Convolution and Attention-Improved C2F Module

School of Physics and Electronics, Hunan Normal University, Changsha 410081, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(19), 6023; https://doi.org/10.3390/s25196023
Submission received: 16 July 2025 / Revised: 21 August 2025 / Accepted: 24 September 2025 / Published: 1 October 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

In this paper, a target detection network algorithm based on a multi-branch convolution and attention improvement Cross-Stage Partial-Fusion Bottleneck with Two Convolutions (C2F) module is proposed for the difficult task of detecting small targets in unmanned aerial vehicles. A C2F module method consisting of fusing partial convolutional (PConv) layers was designed to improve the speed and efficiency of extracting features, and a method consisting of combining multi-scale feature fusion with a channel space attention mechanism was applied in the neck network. An FA-Block module was designed to improve feature fusion and attention to small targets’ features; this design increases the size of the miniscule target layer, allowing richer feature information about the small targets to be retained. Finally, the lightweight up-sampling operator Content-Aware ReAssembly of Features was used to replace the original up-sampling method to expand the network’s sensory field. Experimental tests were conducted on a self-complied mountain pedestrian dataset and the public VisDrone dataset. Compared with the base algorithm, the improved algorithm improved the mAP50, mAP50-95, P-value, and R-value by 2.8%, 3.5%, 2.3%, and 0.2%, respectively, on the Mountain Pedestrian dataset and the mAP50, mAP50-95, P-value, and R-value by 9.2%, 6.4%, 7.7%, and 7.6%, respectively, on the VisDrone dataset.

1. Introduction

With the rapid development of unmanned aerial vehicle (UAV) technology, the application of UAVs in emergency rescue [1] and ecological protection [2] is becoming increasingly widespread. Traditional search-and-rescue methods are often limited by problems such as complex terrain, insufficient manpower, and slow response times, which make it difficult to meet the demand for efficient rescue. However, through the advantageous combination of UAV technology with target detection algorithms, new solutions for mountain searches have been found. With advantages such as flexibility, mobility, and wide coverage, drones can quickly access difficult-to-reach areas. Additionally, by integrating advanced target detection technology, drones can analyze image and video data in real time to efficiently locate missing people.
However, the widespread application of this technology is still hindered by certain challenges: the target personnel in the unmanned end view in mountainous environments are usually small and in front of complex backgrounds. The proportion of small target size is less than 0.1% (such as pedestrians with 20 × 20 pixels), These types of targets lose information in deep networks due to multiple downsampling. The accuracy and robustness of small-target detection cannot meet the actual requirements. Deep learning methods used in target detection algorithms constitute a current research hotspot in unmanned airborne target detection owing to their capacity for high-precision feature extraction and real-time detection. The target detection model based on deep learning is mainly divided into single- and two-stage models according to the process. The two-stage model divides the detection process into two stages: generating candidate regions, followed by classification and bounding box regression on the candidate regions. Representative algorithms include the Region-Convolutional Neural Network [3] (R-CNN), Fast R-CNN [4], Faster R-CNN [5], and Mask R-CNN [6]. However, their two-stage detection processes lead to high computational complexity, slow inference, and numerous model parameters. These characteristics present two-stage detection algorithms with many challenges when deployed on UAV devices with limited arithmetic power. Single-phase target detection algorithms, such as RetinaNet [7], SSD [8], and the You Only Look Once (YOLO) series algorithms [9,10,11,12,13,14,15,16,17,18,19,20], are very suitable for deployment in real applications, especially in arithmetic devices such as UAVs.
To address the difficulty of small-target detection, in 2014, a Google research team introduced the Inception network architecture in GoogLeNet [21], the core idea of which is to improve the feature extraction ability of the model through multi-scale convolution and a parallel structure while controlling the amount of computation and the number of parameters. In 2018, Songtao Liu et al. [22] proposed the Receptive Field Block (RFB) module, which is designed to improve a model’s ability to detect multi-scale targets by simulating the receptive field mechanism of the human visual system. Through the introduction of the attention mechanism, the target detection model can capture the important regions and features in an image more efficiently, thus improving detection accuracy, especially in complex scenes (e.g., those involving weather and background interference). In 2018, Sanghyun Woo et al. [23] proposed the spatial attention mechanism, and Jie Hu et al. [24] proposed the channel attention mechanism. In 2022, Wang L. et al. [25] proposed an improved deep-learning-based SSD model that enhances small-target feature extraction in shallow networks by introducing an improved Inception network. In 2023, S. Tang et al. [26] proposed an improved YOLOv5 model called HIC-YOLOv5 that enhances small-target feature extraction in shallow networks by adding additional small-target-specific prediction heads, introducing involution blocks to enhance the channel information of the feature mAP, and applying the CBAM attention mechanism to improve small-target-detection performance. In 2024, Zhang M. et al. [27] designed an improved YOLOv8 [17] network specialized for detecting small targets in underwater images.
In recent years, there have been many small object detection and recognition algorithms based on YOLO, ranging from YOLOv7 to YOLOv11 [28,29,30,31,32,33,34,35]. With the continuous improvement of algorithms, the accuracy of detection and recognition has also increased. for example, Chen, J.G. et al. [30] conducted research on using improved YOLOv7 to achieve small object detection. They are effective for simple scenes such as wide fields, city roads, etc., but still perform poorly when applied to scenes with complex terrains, such as those common in mountain manhunts. As the number of network layers deepens and increases, under convolution and upsampling, the feature information pertaining to the target in the image inevitably reduces and is lost, resulting in the omission of small targets, in turn leading to a reduction in the detection accuracy of the final model.
To address these small-target detection difficulties, we designed a target detection network algorithm with multibranch convolution and attention combined with an improved C2F module to improve small-target detection accuracy for mountainous scenes from the viewpoint of a UAV.

2. Materials and Methods

2.1. Target Detection Network Based on Multi-Branch Convolution with an Attention-Improved C2F Module

The structure of the target detection network based on multi-branch convolution with an attention improvement C2F module, shown in Figure 1, is divided into three parts. The first part is the backbone network, whose main function is to extract the input features. The second part is the neck network, which is responsible for fusing the features of different scales sent from the backbone network in layers to improve the characterization of the target features of different sizes. The third part is the detection head, which is responsible for outputting the class probabilities and bounding box coordinates of the objects of different scales.

2.2. Backbone Incorporating Partial Convolution

To cope with the limited arithmetic power of UAVs, the C2F module in the backbone network reference FasterNet network structure was improved to enhance the speed and efficiency of feature extraction in the backbone network.
The core idea of the FasterNet network [36] is to reduce the amount of computation through partial convolution (PConv) while ensuring feature extraction ability. Standard convolution, which is used in the C2F module in a traditional network, is a convolution operation performed on all channels of the input feature mAP to generate the output feature mAP. Each output channel is the weighted sum of all the channels of the input feature mAP, which performs the convolution operation on only some of the channels of the input feature mAP, while the other channels are directly passed to the next layer, thereby reducing the amount of computation while retaining some of the original feature information.
Assuming that the shape of the input network feature mAP is ( C i n , H , W ) —which assumes the shape ( C o u t , C i n , K , K ) , where C i n is the number of input channels, C o u t is the number of output channels, K is the size of the convolution kernel, and H and W are the height and width of the feature mAP, respectively—and the shape of the output feature mAP is ( C o u t , H , W ) , where H and W are the height and width of the output feature mAP, PConv selects the first C p channel for the convolution operation, and the remaining C i n C p channels are passed directly. The shape of the convolution kernel is ( C o u t , C p , K , K ) . The ratio of the computational complexity of the standard convolution to that of PConv is as follows:
O ( C i n × C o u t × K × K × H × W ) O ( C p × C o u t × K × K × H × W ) ,
Because C p < C i n , network computation is significantly reduced. In this study, to improve the efficiency of the network, PConv was applied to the C2F modules to design the C2F-F module, whose structure is shown in Figure 2. The original standard convolution was changed to PConv, which utilizes the characteristics of PConv partial convolution to reduce complexity and improve network computation efficiency. Table 1 shows that after the C2F-F module was employed, the number of parameters reduced by 10.46%, and the required floating-point operations reduced by 0.6 GFLPs, with an increase in the number of network layers, proving that the module can reduce the number of network parameters and improve the efficiency of network computation.

2.3. FA-Block Based on the Design of a Multi-Scale Feature Fusion and Attention Mechanism

Aiming to address the neck network’s difficulty detecting small targets, we combined multiscale feature fusion with channel and spatial attention mechanisms in the neck network to design the FA-Block. This block expands the sensory field of the network and enhances its ability to extract small-target features through multiscale feature fusion and an attention mechanism. Meanwhile, we replaced the bottleneck structure in the C2F module with the FA-Block block and designed the C2F-FA block to replace the original C2F module to enhance feature fusion capacity and the ability to focus on small-target features.
The FA-Block block combines multibranch convolution with the attention mechanism to enhance the model’s feature fusion capacity and detection accuracy, as shown in Figure 3. The multibranch convolution component has three branches with dilation rates of 1, 2, and 3.
The base branch (d = 1) is a 5 × 5 standard convolution that captures local details, the middle-range branch (d = 2) is an equivalent 9 × 9 receptive field that extracts medium-range semantic information, and the remote branch (d = 3) is an equivalent 13 × 13 receptive field that models long-range spatial dependencies.
The output features of each branch were spliced through channels to form a multiscale feature tensor to help the network capture contextual information at different scales. Subsequently, the outputs of the three branches were spliced, and channel and spatial attention were applied.
Finally, the double-attention processed features were convolved by 1 × 1 to realize feature fusion and residual concatenation. This design is intended to enhance the diversity and robustness of the features, and it is hoped that this module will yield good results when dealing with targets at different scales. Output formula is
Output = C o n c a t C o n v 1 × 1 ( x ) , C o n v 3 × 3 ( x ) , C o n v 5 × 5 ( x ) , M a x P o o l 3 × 3 ( x ) ,
In this study, a C2F-FA module was designed by replacing the bottleneck blocks in the C2F module with FA-Blocks, as shown in Figure 4. Each bottleneck in the C2F structure was replaced with an FA-Block with an attention mechanism and multi-branch convolution, and the number was reduced to one, thereby enhancing the expressiveness of the model and balancing the network’s complexity throughout the feature extraction process.

2.4. Addition of a Tiny-Target Detection Layer

Small targets have fewer pixels, are easily overwhelmed by background targets and background noise, and exhibit weak feature expression; conventional down-sampling will result in the loss of small-target features or their confusion with background noise. In this study, we designed a tiny-target layer with the following dimensions: 160 × 160 × 128.
The change in the size of the feature mAP after the improvement is shown in Figure 5 and Table 2. T3 has fewer convolutions, and the feature mAP is larger, a state more conducive to small-target recognition. In this study, optimizing the network structure enabled an increase in the tiny-target layer, which focuses on extracting small-scale features, allowing the network to better meet the current requirements of UAV small-target detection.
When the dimensions of the input image are 640 × 640, the 160 × 160 layer corresponds to 4-fold down-sampling with the following number of pixels:
640 4 × 640 4 = 25 , 600
The number of pixels in the original 8-fold down-sampling procedure is as follows:
640 8 × 640 8 = 6400
A comparison of the formulas reveals that the improved tiny-target layer significantly increases the number of pixels for small targets.

2.5. Improvements to the Up-Sampling Methodology

To improve UAVs’ recognition performance with respect to small targets, we improved the neck of the base network and adopted the lightweight up-sampling operator CARAFE to replace the original nearest-neighbor interpolation up-sampling operator. As shown in Figure 6 [37], the CARAFE operator consists of an up-sampling kernel prediction module and a feature reorganization module.
There are two steps in the CARAFE module. The first step is predicting a reorganization kernel based on the content of each target location, and the second step is reorganizing the features with the predicted kernel.
In the up-sampling kernel prediction module, assuming that the up-sampling multiplicity of the network is σ and the shape of the output feature mAP X of the previous layer is H × W × C, the module compresses the number of channels of the feature mAP to C m via a 1 × 1 convolution operation and then predicts the up-sampling kernel with a convolutional layer of k e × k e , resulting in an up-sampling kernel with the following shape: σ H × σ W × k u p 2 . Finally, this up-sampling kernel is normalized using the Softmax pair in order to ensure that the weights of the convolution kernel sum to 1.
In the feature recombination module, region-centered k u p × k u p at each position of the output feature mAP is extracted, and the recombined kernel W t is subjected to dot-product computation to enhance the semantic information of the feature mAP.
The optimized shallow network focuses on capturing detailed information to accurately locate the target, whereas the deep network focuses more on understanding contextual semantic information for inference. Applying CARAFE to the neck network feature pyramid FPN can enhance the deep network’s characterization ability, thereby endowing the fused features with a richer expressive capacity.
This method was developed in combination with the special needs of mountainous scenes. Optimized the network structure for mountain small target recognition, improving the robustness of the model in special scenarios. Although there are some limitations in the design of the network structure, we have actively explored a network structure suitable for mountain unmanned aerial vehicle search missions.

3. Results

3.1. Dataset and Experimental Setup

In the experiments, we used the mountain pedestrian dataset, a target detection dataset consisting of images taken from a UAV’s viewpoint collected and produced in this study. The dataset contains two types of detection objects: people and occluded pedestrians. We used a total of 7464 datasets, and the images were randomly assigned in a ratio of 8:1:1 for the training, validation, and test sets.

3.2. Environmental Configuration and Evaluation Indicators

The parameter settings for the experimental platform are shown in Table 3.
Using the PyTorch 2.5.0 framework on an Ubuntu 20.04 server, we conducted the experiments using a server configuration consisting of an Intel(R) Xeon(R) Gold 6338 CPU operating @ 2.00 GHz (Intel, Santa Clara, CA, USA), 1 T of RAM, and four Nvidia RTX 4090 24 G GPUs (Nvidia, Santa Clara, CA, USA).
Environment configuration: Python version 3.8, torch version 2.4.1, CUDA version 12.1, and YOLO version 8.0.138.
The dimensions of each image were 640 × 640, the number of training rounds was 150, and the single-image input batch was set to 16. The remaining experimental parameters were set according to the system’s default settings. Precision (P), recall (R), and mean average precision (mAP) were used as experimental evaluation metrics, and two specific versions of mAP50 and mAP50-95 were used for mean precision. Equations (5), (6), (7), and (8) are the formulae for P, R, mAP50, and mAP50-95, respectively.
P = T P / ( T P + F P ) ,
R = T P / ( T P + F N ) ,
m A P = 1 m i = 1 m A P i I o U = 0.5 ,
m A P = 1 m i = 1 m A P i I o U = 0.5 : 0.05 : 0.95 ,

3.3. Ablation Experiment

To verify the performance impact each module had on the network, we performed ablation experiments using the mountain pedestrian dataset, using YOLOV8n as the base network.
Table 4 shows the performance impact of each module of the improved algorithm on the network with respect to the mountain pedestrian dataset. In comparison to the performance of the base network, each module improved the model. Still effective when the modules acted together, the improved network improved the mAP50, mAP50-95, P-value, and R-value by 2.8%, 3.5%, 2.3%, and 0.2%, respectively. The detection of small targets was further improved, demonstrating the effectiveness of the target detection network incorporating multi-branch convolution and attention combined with an improved C2F module.
To verify the generalization ability and robustness of each module of the improved network, we also used the public dataset VisDrone2019, as shown in Table 5. In comparison to the base network, the mAP50, mAP50-95, P-value, and R-value improved by 9.2%, 6.4%, 7.7%, and 7.6%, respectively. The experimental results show that even when applied to the public dataset VisDrone2019, the modules in the target detection network incorporating multibranch convolution and attention combined with the improved C2F module are still effective, demonstrating the robustness of the improved network algorithm.

3.4. Comparison of the Results for Different Algorithms

To verify the model’s performance, we compared it with several other mainstream target detection models, including the YOLOV5n, YOLOV5s, YOLOV7n, YOLOV8n, and YOLOv11n models, which are more advanced target detection models.
YOLOv11 is the latest iteration version of the YOLO model developed by the Ultralytics team. YOLOv11 has introduced architectural improvements to optimize feature extraction and processing on the solid foundation of YOLOv8, which has resulted in better performance of YOLOv11 in small object processing compared to YOLOv8. Therefore, this article also includes YOLOv11 in the comparison method.
Table 6 presents the quantitative results obtained with respect to the mountain pedestrian dataset. The mAP50, mAP50-95, P-value, and R-value of the improved algorithm model in this study were 85.5%, 46.1%, 80.9%, and 82.9%, respectively, constituting the best figures among all the models, and the side-by-side comparisons verify the effectiveness of the model designed in this study.
Table 7 shows the test results obtained for the various algorithms with respect to VisDrone2019. Compared with the best-performing algorithm model in regard to VisDrone2019, the mAP50, mAP50-95, P, and R values of the model designed in this study are 3.4%, 4.8%, 2%, and 0.9% better, respectively. These results prove that our model can effectively and accurately detect small target objects.

3.5. Visualization Results

To more concretely prove the effectiveness of our improved C2F modular target detection network based on the combination of multi-branch convolution and attention for small-target detection, pictures were randomly selected from the mountain pedestrian dataset and used to test the improved algorithmic model and other mainstream detection algorithmic models, respectively. Local magnified details of the original image are shown in Figure 7, in which square 1 is a vehicle, which was not a detection target, whereas squares 2 to 5 are occluded pedestrian targets.
Figure 8 shows the results of comparing the detection effectiveness of the other algorithmic models with that of the improved algorithmic model. Column a is the original image, while columns b-g correspond to the detection results for the different models. The first row shows that the YOLOv8n and YOLOv11n models misrecognized vehicles as occluded pedestrians; the second, third, and fifth rows show that the YOLOv5n, YOLOv5s, and YOLOv7n models failed to detect occluded people. In contrast, the improved model accurately recognized all occluded pedestrians and did not misidentify the vehicle as an occluded person, demonstrating its superior reasoning ability for small targets and higher detection accuracy relative to the other algorithmic models, comprehensively proving its effectiveness in small-target detection.

4. Conclusions

In this paper, we address the current problems in UAV detection attributable to the limited arithmetic ability of UAVs and the inability of mainstream detection algorithms to effectively deal with small targets. Referring to the partial convolutional (PConv) layers in the FasterNet network in the backbone network, we designed the C2F-F module to utilize the characteristics of partial convolutions to improve the network’s computational efficiency. In the neck network, the FasterNet Block was improved by combining multi-scale feature fusion with channel and spatial attention mechanisms, and the FA-Block block was designed. The bottleneck structure in the C2F block was replaced with the FA-Block, and the C2F-FA block was designed to replace the original C2F block to improve the network’s feature fusion capacity and ability to focus on small-target features. A tiny 160 × 160 target layer was added to improve the extraction of small-scale features, thus enabling the network to better meet the current requirements of UAV small-target detection. Finally, CARAFE was applied to the neck network feature pyramid FPN to enhance the network’s deep characterization capacity, thus endowing the fused features with richer expressive capabilities.
In the ablation experiments, the target detection network designed in this study based on multi-branch convolution and attention combined with the improved C2F module exhibited improved experimental metrics in comparison to the original network on the self-built dataset and the publicly available dataset VisDrone dataset, proving its effectiveness in detecting small targets. In the experiments comparing five mainstream detection algorithms on different datasets, all the metrics for our algorithm were superior to those of the other algorithms. In the visualization comparison of the detection results, the improved algorithm model accurately identified all the obscured pedestrians and did not misidentify the vehicles as obscured people, showing its superior reasoning ability and detection accuracy for small targets relative to the other models, concretely proving its effectiveness in the field of small-target detection. This study reveals the effectiveness of the model designed in this study and its good robustness.
In addition to the influence of terrain, adverse weather conditions also have a significant impact on the recognition of small targets, which requires the use of multiple sensors and multi-source image fusion to solve. This is the next research direction.

Author Contributions

F.Q. and W.T. were responsible for the experiment and the writing of this manuscript. H.T. provided academic advice and general support during the research. Y.C. supervised the researchers and revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the College Students’ Innovative Entrepreneurial Training Plan Program (University Level) (No.2024132).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data supporting this article have been uploaded at https://pan.baidu.com/s/1X5b7qtDRcDnk2nyw2rXoRg (accessed on 23 September 2025). The extraction code is hv33. links of the Code: https://github.com/hunnutangweiwei/-multi-branch-convolution-and-attention-improved-C2F-module (accessed on 23 September 2025).

Acknowledgments

All classmates and junior students in the laboratory provided varying degrees of assistance in the same research direction.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Asignacion, A.; Satoshi, S. Historical and Current Landscapes of Autonomous Quadrotor Control: An Early-Career Researchers’ Guide. Drones 2024, 8, 72. [Google Scholar] [CrossRef]
  2. Calamoneri, T.; Corò, F.; Mancini, S. Management of a post-disaster emergency scenario through unmanned aerial vehicles: Multi-Depot Multi-Trip Vehicle Routing with Total Completion Time Minimization. Expert Syst. Appl. 2024, 251, 123766–123778. [Google Scholar] [CrossRef]
  3. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  4. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  5. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef]
  6. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 42, 386–397. [Google Scholar] [CrossRef] [PubMed]
  7. Lin, T.Y.; Goyal, P.; Girshick, R.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 42, 318–327. [Google Scholar] [CrossRef]
  8. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. Eur. Conf. Comput. Vis. 2016, 9905, 21–37. [Google Scholar]
  9. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  10. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  11. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar] [CrossRef]
  12. Mao, Q.C.; Sun, H.M.; Liu, Y.B.; Jia, R.-S. Mini-YOLOv3: Real-Time Object Detector for Embedded Applications. IEEE Access 2019, 7, 133529–133538. [Google Scholar] [CrossRef]
  13. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar] [CrossRef]
  14. Singha, S.; Aydin, B. Automated Drone Detection Using YOLOv4. Drones 2021, 5, 95. [Google Scholar] [CrossRef]
  15. Khanam, R.; Hussain, M. What is YOLOv5: A deep look into the internal features of the popular object detector. arXiv 2024, arXiv:2407.20892. [Google Scholar] [CrossRef]
  16. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2022, arXiv:2207.02696. [Google Scholar]
  17. Varghese, R.; Sambath, M. YOLOv8: A Novel Object Detection Algorithm with Enhanced Performance and Robustness. In Proceedings of the 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS), Chennai, India, 18–19 April 2024. [Google Scholar]
  18. Wang, C.Y.; Yeh, I.H.; Liao, H.Y.M. YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information. arXiv 2024, arXiv:2402.13616. [Google Scholar] [CrossRef]
  19. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J. YOLOv10: Real-Time End-to-End Object Detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
  20. Khanam, R.; Hussain, M. YOLOv11: An Overview of the Key Architectural Enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar] [CrossRef]
  21. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  22. Liu, S.; Huang, D.; Wang, Y. Receptive Field Block Net for Accurate and Fast Object Detection. Lect. Notes Comput. Sci. 2018, 11215, 385–400. [Google Scholar]
  23. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. Lect. Notes Comput. Sci. 2018, 11211, 3–19. [Google Scholar]
  24. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, Salt Lake City, UT, USA, 18–23 June 2018; Volume 99, pp. 7132–7141. [Google Scholar]
  25. Wang, L.G.; Shou, L.Y.; Alyami, H.; Laghari, A.A.; Rashid, M.; Almotiri, J.; Alyamani, H.J.; Alturise, F. A novel deep learning-based single shot multibox detector model for object detection in optical remote sensing images. Geosci. Data J. 2022, 11, 2049–6060. [Google Scholar] [CrossRef]
  26. Tang, S.; Zhang, S.; Fang, Y. HIC-YOLOv5: Improved YOLOv5 for Small Object Detection. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 6614–6619. [Google Scholar]
  27. Zhang, M.; Wang, Z.; Song, W.; Zhao, D.; Zhao, H. Efficient Small-Object Detection in Underwater Images Using the Enhanced YOLOv8 Network. Appl. Sci. 2024, 14, 1095. [Google Scholar] [CrossRef]
  28. Liu, Y.F.; Zhang, D.C.; Guo, C. GL-YOLOv5: An Improved Lightweight Non-Dimensional Attention Algorithm Based on YOLOv5. Comput. Mater. Contin. 2024, 81, 3281–3299. [Google Scholar] [CrossRef]
  29. Kang, Z.P.; Liao, Y.R.; Du, S.H.; Li, H.; Li, Z. SE-CBAM-YOLOv7: An Improved Lightweight Attention Mechanism-Based YOLOv7 for Real-Time Detection of Small Aircraft Targets in Microsatellite Remote Sensing Imaging. Aerospace 2024, 11, 605. [Google Scholar] [CrossRef]
  30. Chen, J.; Wen, R.; Ma, L. Small object detection model for UAV aerial image based on YOLOv7. Signal Image Video Process. 2024, 18, 2695–2707. [Google Scholar] [CrossRef]
  31. Shi, Y.; Duan, Z.; Qing, H.H.; Zhao, L.; Wang, F.; Yuwen, X. YOLOv9s-Pear: A Lightweight YOLOv9s-Based Improved Model for Young Red Pear Small-Target Recognition. Agronomy 2024, 14, 2086. [Google Scholar] [CrossRef]
  32. Xu, K.Y.; Song, C.T.; Xie, Y.; Pan, L.; Gan, X.; Huang, G. RMT-YOLOv9s: An Infrared Small Target Detection Method Based on UAV Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2024, 21, 7002205. [Google Scholar] [CrossRef]
  33. Zhang, Q.Y.; Wang, X.; Shi, H.; Wang, K.; Tian, Y.; Xu, Z.; Zhang, Y.; Jia, G. BRA-YOLOv10: UAV Small Target Detection Based on YOLOv10. Drones 2025, 9, 159. [Google Scholar] [CrossRef]
  34. Wang, C.C.; Han, Y.Q.; Yang, C.G.; Wu, M.; Chen, Z.; Yun, L.; Jin, X. CF-YOLO for small target detection in drone imagery based on YOLOv11 algorithm. Sci. Rep. 2025, 15, 16741. [Google Scholar] [CrossRef]
  35. Zhang, Z.; Zhang, Z.Y.; Li, G.; Xia, C. ZZ-YOLOv11: A Lightweight Vehicle Detection Model Based on Improved YOLOv11. Sensors 2025, 25, 3399. [Google Scholar] [CrossRef]
  36. Chen, J.; Kao, S.H.; He, H.; Zhuo, W.; Wen, S.; Lee, C.-H.; Chan, S.-H.G. Run, Don’t Walk: Chasing Higher FLOPS for Faster Neural Networks. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–24 June 2023; pp. 12021–12031. [Google Scholar]
  37. Wang, J.Q.; Chen, K.; Xu, R.; Liu, Z.; Loy, C.C.; Lin, D. CARAFE: Content-Aware Reassembly of Features. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, South Korea, 27 October–2 November 2019; pp. 3007–3016. [Google Scholar]
Figure 1. The structure of the target detection network based on multi-branch convolution and attention combined with an improved C2F module.
Figure 1. The structure of the target detection network based on multi-branch convolution and attention combined with an improved C2F module.
Sensors 25 06023 g001
Figure 2. The structure of the C2F-F module.
Figure 2. The structure of the C2F-F module.
Sensors 25 06023 g002
Figure 3. Structure of the FA-Block module.
Figure 3. Structure of the FA-Block module.
Sensors 25 06023 g003
Figure 4. Structure of the C2F-FA module.
Figure 4. Structure of the C2F-FA module.
Sensors 25 06023 g004
Figure 5. Change in the dimensions of the improved network feature mAP.
Figure 5. Change in the dimensions of the improved network feature mAP.
Sensors 25 06023 g005
Figure 6. Structure of the CARAFE module.
Figure 6. Structure of the CARAFE module.
Sensors 25 06023 g006
Figure 7. Enlarged view of the local details of the test pictures.
Figure 7. Enlarged view of the local details of the test pictures.
Sensors 25 06023 g007
Figure 8. Comparison of the detection results for the original, YOLOv5n, YOLOv5s, YOLOv7n, YOLOv8n, YOLOv11n, and improved algorithmic models.
Figure 8. Comparison of the detection results for the original, YOLOv5n, YOLOv5s, YOLOv7n, YOLOv8n, YOLOv11n, and improved algorithmic models.
Sensors 25 06023 g008
Table 1. A comparison of the structures and computational complexity of the models.
Table 1. A comparison of the structures and computational complexity of the models.
LayersParametersGFLOPsFPS
Base2253,157,2008.911
C2F-F2312,826,9608.38.72
Table 2. Change in the size of the improved network feature mAP.
Table 2. Change in the size of the improved network feature mAP.
Network StageName of Feature mAPDimensions
(W × H × Aisle)
Operation Description
BackboneInput640 × 640 × 3Original input
BackboneF1160 × 160 × 128Generated via convolution operation on input
BackboneF280 × 80 × 256Generated via F1 convolution operation
BackboneF340 × 40 × 512Generated via F2 convolution operation
NeckU180 × 80 × 512F3 is up-sampled
NeckC180 × 80 × 256Generated by concatenating U1 and F2
NeckU2160 × 160 × 256C1 is up-sampled
HeadT3160 × 160 × 128U2 is spliced with F1 and fed into the detector head
HeadT280 × 80 × 256C1 is spliced with T3 and then generated and fed into the detection head
HeadT140 × 40 × 512F3 and T2 are spliced, generated, and fed into the detector head
Table 3. Configuration of the experimental environment.
Table 3. Configuration of the experimental environment.
ProjectVersion
Operating SystemUbuntu 20.04
CPUIntel Xeon Gold 6338
GPUNvidia RTX 4090 24 G
CompilerPyCharm
Algorithmic FrameworkPytorch-2.4.1 + Cuda12.1
Programming LanguagesPython3.8
YOLO Versionv8.0.138
Table 4. The performance impact of each module on the network. ‘—’ Represents that this module has not been added,‘√’ Represents that this module has been added.
Table 4. The performance impact of each module on the network. ‘—’ Represents that this module has not been added,‘√’ Represents that this module has been added.
C2F-FC2F-FATinyCARAFEmAP50/%mAP50-95/%P/%R/%
82.742.678.682.7
82.843.078.780.6
83.143.278.380.1
84.545.079.681.4
83.042.979.080.4
85.546.180.982.9
Table 5. The effectiveness of each module with respect to the VisDrone2019 dataset. ‘—’ Represents that this module has not been added,‘√’ Represents that this module has been added.
Table 5. The effectiveness of each module with respect to the VisDrone2019 dataset. ‘—’ Represents that this module has not been added,‘√’ Represents that this module has been added.
C2F-FC2F-FATinyCARAFEmAP50/%mAP50-95/%P/%R/%
32.118.343.232.6
32.318.742.732.3
38.222.649.637.3
35.220.644.535.5
33.019.042.533.4
41.324.750.940.2
Table 6. The test results for the different algorithms when applied to the self-constructed dataset.
Table 6. The test results for the different algorithms when applied to the self-constructed dataset.
ModelmAP50/%mAP50-95/%P/%R/%
YOLOv5n77.134.575.775.0
YOLOv5s82.140.078.979.1
YOLOv7n72.832.171.373.2
YOLOv8n82.742.678.682.7
YOLOv11n82.342.177.780.1
Ours85.546.180.982.9
Table 7. Test results obtained for the different algorithms on the VisDrone2019 dataset.
Table 7. Test results obtained for the different algorithms on the VisDrone2019 dataset.
ModelmAP50/%mAP50-95/%P/%R/%
YOLOv5n25.313.035.527.7
YOLOv5s32.918.244.533.2
YOLOv7n37.919.948.939.3
YOLOv8n32.118.343.232.6
YOLOv11n34.219.945.133.8
Ours41.324.750.940.2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qin, F.; Tang, W.; Tian, H.; Chen, Y. Unmanned Airborne Target Detection Method with Multi-Branch Convolution and Attention-Improved C2F Module. Sensors 2025, 25, 6023. https://doi.org/10.3390/s25196023

AMA Style

Qin F, Tang W, Tian H, Chen Y. Unmanned Airborne Target Detection Method with Multi-Branch Convolution and Attention-Improved C2F Module. Sensors. 2025; 25(19):6023. https://doi.org/10.3390/s25196023

Chicago/Turabian Style

Qin, Fangyuan, Weiwei Tang, Haishan Tian, and Yuyu Chen. 2025. "Unmanned Airborne Target Detection Method with Multi-Branch Convolution and Attention-Improved C2F Module" Sensors 25, no. 19: 6023. https://doi.org/10.3390/s25196023

APA Style

Qin, F., Tang, W., Tian, H., & Chen, Y. (2025). Unmanned Airborne Target Detection Method with Multi-Branch Convolution and Attention-Improved C2F Module. Sensors, 25(19), 6023. https://doi.org/10.3390/s25196023

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop