Next Article in Journal
A Computationally Efficient MUSIC Algorithm with an Enhanced DOA Estimation Performance for a Crossed-Dipole Array
Previous Article in Journal
Wearable Device for Continuous and Real-Time Monitoring of Human Sweat Sodium
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MEP-YOLOv5s: Small-Target Detection Model for Unmanned Aerial Vehicle-Captured Images

1
Guangxi Key Laboratory of Functional Information Materials and Intelligent Information Processing, Nanning Normal University, Nanning 530001, China
2
Guangxi Geographical Indication Crops Research Center of Big Data Mining and Experimental Engineering Technology, Nanning Normal University, Nanning 530001, China
3
Guangxi Key Laboratory of Earth Surface Processes and Intelligent Simulation, Nanning Normal University, Nanning 530001, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2025, 25(11), 3468; https://doi.org/10.3390/s25113468
Submission received: 18 May 2025 / Revised: 27 May 2025 / Accepted: 29 May 2025 / Published: 30 May 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

Due to complex backgrounds, significant scale variations of targets, and dense distributions of small objects in Unmanned Aerial Vehicle (UAV) aerial images, traditional object detection algorithms face challenges in adapting to such scenarios. This article introduces a drone detection model, MEP-YOLOv5s, which optimizes the Backbone, Neck layer, and C3 module based on YOLOv5s, and combines effective attention mechanisms to improve the training efficiency of the model by replacing the traditional CIoU loss (Complete Intersection over Union) with MPDIoU (Minimum Point Distance-based Intersection over Union) loss. This model demonstrates an excellent performance in handling typical drone detection scenarios, especially for small and dense objects. To holistically balance the detection accuracy and inference efficiency, we propose a Comprehensive Performance Indicator (CPI), which evaluates the model performance by considering both accuracy and efficiency. Evaluations on the VisDrone2019 dataset demonstrate that MEP-YOLOv5s achieves a 3.3% improvement in precision (P), a 20.9% increase in mAP@0.5, and a 19.86% gain in the CPI ( α = 0.5) compared with the baseline model. Additional experiments on the NWPU VHR-10 dataset confirm that MEP-YOLOv5s outperforms the existing state-of-the-art methods, offering a robust solution for UAV-based small object detection with enhanced feature extraction and attention-driven adaptability.

1. Introduction

With the rapid development of drone technology, it has been gradually applied in various fields, including military and civilian applications. In the field of environmental monitoring, Wang et al. designed ODTDS, a real-time drone-based detection system, for target recognition in complex environments [1]. In the field of dynamics, Sudhanshu Shekhar Jha et al. detected and identified various artificial targets placed in complex imaging geometries with different target backgrounds in different application scenarios, based on drone-based hyperspectral imaging [2]. Rohan et al. proposed a method for detecting and tracking drone targets, including both moving and stationary targets [3]. In the field of smart agriculture, Liao et al. proposed an AI-based agricultural drone to improve the accuracy and efficiency of pesticide spraying and enhance the effectiveness of plant protection [4]. In the field of intelligent inspection, Li et al. designed a drone-based autonomous inspection system based on target detection [5]. In the field of traffic monitoring, Ángel et al. proposed a multi-object tracking (MOT) algorithm for traffic monitoring using drones equipped with optical and thermal cameras [6]. In these applications, drone target detection technology plays a crucial role. Drone target detection refers to the use of high-definition cameras or other sensors mounted on drones to identify, locate, and track specific targets in images or videos using image processing technology and computer vision algorithms. Traditional object detection approaches, which are predominantly based on manually engineered feature extraction techniques [7,8] and handcrafted features [9], demonstrate substantial limitations when applied to aerial imagery. These constraints manifest in several critical aspects: First, the characteristically minute pixel dimensions of objects in aerial images pose significant challenges for conventional handcrafted features to reliably extract discriminative patterns from low-resolution targets, consequently leading to markedly elevated missed detection rates; Second, the distinctive overhead perspective inherent to aerial imaging frequently results in densely clustered objects and severe occlusion scenarios, wherein traditional non-maximum suppression (NMS) algorithms tend to produce erroneous suppression outcomes. Third, the inherent variability in UAV flight altitudes induces pronounced scale variations among targets, fundamentally limiting the effectiveness of fixed-scale sliding window detection paradigms. Furthermore, the complex background artifacts commonly encountered in aerial images—including shadows and regions with homogeneous textures—substantially increase the propensity for false positives in traditional detection frameworks. These compounded challenges underscore the pressing need for novel detection methodologies that can simultaneously address the unique demands of aerial imagery while maintaining computational efficiency, thereby presenting both significant research opportunities and practical applications in the field.
With the rise of deep learning technology, especially the successful application of convolutional neural networks (CNNs) in the field of image recognition, drone target detection technology has ushered in new development opportunities. At present, mainstream target detection algorithms include two-stage detection algorithms based on deep learning (such as the R-CNN series [10], Mask R-CNN [11], and Cascade R-CNN [12]) and one-stage detection algorithms (such as the YOLO series [13,14,15,16], SSD [17] and RetinaNet [18]). These algorithms have achieved great success in the field of natural image detection, but their detection performance significantly decreases when faced with aerial images.
The rest of this paper is organized as follows: Section 2 reviews the relevant literature on drone target detection; Section 3 elaborates on the proposed MEP-YOLOv5s model and introduces the principles of its feature extraction module; Section 4 verifies the superiority of MEP-YOLOv5s through a series of ablation and comparative experiments; and Section 5 summarizes the work of this paper and looks forward to the future research directions.

2. Materials and Methods

Many traditional drone target detection methods rely on digital signal analysis methods, such as audio processing and radar signal interpretation. However, these methods exhibit limited accuracy in complex scenarios involving numerous or overlapping targets, and their detection stability degrades significantly in complex environments. Consequently, there has been a growing trend among researchers to adopt deep learning algorithms for extracting object detection features from drone aerial imagery. L Shaodan et al. [19] proposed a segmentation-based detection model (RiceblastSegMask) for rice blast disease detection and resistance assessment, which enables the effective quantification of disease infection levels. H Liang et al. [20] developed a drone-based low-altitude remote sensing system for multi-category concrete bridge damage detection. This system employs a Swin Transformer-based backbone network combined with a multi-scale attention pyramid network and lightweight residual global attention network (LRGA-Net), achieving significant performance improvements in speed and accuracy. B Zhao et al. [21] introduced a cluster analysis-based anchor box optimization strategy to enhance the Faster R-CNN framework’s performance in drone-based marine search and rescue operations. Z Yang et al. [22] presented an effective drone detection framework for infrared imagery. The framework first determines continuous coupling neural network (CCNN) parameters through image statistical features (standard deviation, mean value, etc.), followed by iterative pixel grouping, morphological operations (expansion and erosion), and minimum bounding rectangle extraction to produce final detection results. H Li et al. [23] proposed an improved Mask R-CNN instance segmentation method integrated with hyperspectral sensor technology, enabling the early stage identification of pine wood nematode infections in individual trees. D Wu et al. [24] introduced FSNet, which enhances local and global diversity in forest fire imagery through YOCO data augmentation. The model incorporates a Transformer-based EBBlock module that leverages cross-group interactions to comprehensively extract fire and smoke features while suppressing the background interference. Additionally, FSNet employs a feature pyramid architecture to fuse outputs from four hierarchical stages, significantly mitigating small-target feature loss.
Compared with two-stage drone target detection methods, lightweight single-stage models with a high precision have gained significant traction in specific domains while maintaining competitive detection accuracy. The YOLO series represents a prominent single-stage detection approach [14], with ongoing research focusing on balancing accuracy and efficiency for drone-based applications. J Guo et al. [25] developed a drone aerial detection algorithm based on YOLOv5s, incorporating a more accurate small target detection (MASD) mechanism and multi-scale feature fusion (MCF) path to enhance the feature representation. Y Xu et al. [26] optimized YOLOv5s anchor boxes using K-Means++ clustering and introduced the EIoU loss function to replace the CIoU loss; CIoU enhances DIoU by adding an aspect ratio consistency penalty to jointly optimize the overlap area, center distance, and aspect ratio alignment. EIoU further refines this by decomposing aspect ratio loss into independent width and height components and explicitly modeling center distance with edge length differences. Through finer geometric constraints, EIoU achieves better small-object detection and scale variation handling with improved computational efficiency, while demonstrating an improved detection performance through real-world flight tests. Y Dong et al. [27] designed a lightweight YOLOv4-based drone detection model, enhancing small target detection through improved feature fusion, target augmentation, and candidate box refinement. X Su et al. [28] proposed ASFF Small, an adaptive feature fusion method combining YOLOv3-SPP with ASFF, which is tailored for drone image characteristics. H Tang et al. [29] presented a radar-optical fusion detection method for marine UAVs, effectively addressing the challenges posed by closely spaced overlapping targets in radar imagery.
Contemporary target detection algorithms exhibit distinct technical evolution paths. Two-stage frameworks primarily focus on optimizing feature extraction networks through mechanisms such as multi-scale feature fusion and attention module enhancements. These methods demonstrate a superior performance in complex scenarios (e.g., achieving 46.2% mAP on the COCO dataset [12]) but introduce significant computational overheads due to their hierarchical processing architecture, resulting in an approximately 40% higher inference latency compared with single-stage models (e.g., the delay gap between Faster R-CNN and RetinaNet reaches 41.6% [30]). In contrast, single-stage detectors emphasize module-level innovations, including dynamic sampling strategies, nonlinear activation functions (e.g., Swish [31]), and cross-layer feature interaction mechanisms. However, these improvements often increase the model complexity, with typical single-stage architectures exceeding 45M parameters (e.g., YOLOv5x [15]), complicating deployment on resource-constrained drone platforms.
These technical challenges manifest across three dimensions: (1) Perceptual limitations—drone-acquired images frequently suffer from sub-1080p resolution, high vegetation coverage (>70%), and minute target sizes (<0.05% pixel area); (2) Dynamic environmental interference—flight-induced posture changes and abrupt lighting variations compromise image quality; and (3) Bottlenecks in engineering implementation—algorithms must balance real-time performance and accuracy under hardware constraints. Key limitations include the drone platform’s computing resources (e.g., CPU/GPU computing power, memory bandwidth), power consumption budget, cooling capacity (primarily passive cooling), and payload weight, all of which directly restrict the deployment of complex detection models.
To address these issues, this paper proposes MEP-YOLOv5s—an innovative algorithm for small-target detection in drone aerial imagery.

3. Improved Algorithm

3.1. Model Structure

This study proposes enhancements to YOLOv5 tailored for the distinct detection challenges that are inherent in drone-captured aerial imagery. To address these challenges, we identify three critical components within the YOLOv5 architecture requiring optimization: feature extraction, loss function formulation, and detection head design. The proposed MEP-YOLOv5s algorithm implements three key modifications:
  • Architectural Redesign: The original C3 module is replaced with a redesigned C2f module integrated with an Exponential Moving Average (EMA [32]) attention mechanism, forming the C2f_EMA module. This configuration optimizes the receptive field geometry to enhance the detection precision for smaller targets.
  • Loss Function Innovation: We replace the traditional CIoU loss function with the improved Modified Probabilistic Distance-IoU (MPDIoU [33]) loss function to enhance the convergence behavior.
  • Detection Head Augmentation: A dedicated small-target detection layer is integrated into the prediction head to strengthen the network’s capability in resolving densely packed minute objects.
The resulting optimized network architecture, termed MEP-YOLOv5s, is visualized in Figure 1.

3.2. Feature Converged Network Architecture

3.2.1. Attention-Based Receptive Field Feature Retrieval Module

In the YOLOv5 architecture, the C3 module is extensively utilized, whereas the YOLOv8 network incorporates the C2f module. These modules represent distinct structural innovations within their respective frameworks, differing significantly in both functional design and operational principles. The C2f module integrates design philosophies from both the C3 module and ELAN, prioritizing lightweight construction while enhancing the gradient flow information richness. Specifically, it achieves increased model depth and complexity through augmented convolutional layers and connectivity patterns, thereby improving the gradient information capture capabilities. By contrast, the C3 module adopts the split extraction strategy of CSPNet [34], and is integrated with a residual architecture. It comprises two primary components: the CBS main branch gradient module for feature extraction and the Bottleneck module for optimizing computational efficiency by reducing the parametric complexity. The architectural configurations of the C2f and C3 modules are visualized in Figure 2 and Figure 3, respectively.
As illustrated in the figure, while the C2f module retains the convolutional neural network architecture (functioning as a local feature extractor), the attention mechanism’s strength resides in its global receptive field, enabling the capture of long-range feature dependencies. In this work, we propose C2f_EMA by substituting the C3 module with the C2f module and integrating the EMA attention mechanism within the Neck component (Figure 4).
Assuming that the output tensor of the previous module connected to the C2F_EMA module is X R C × H × W , it first passes through the CBS (Convolution-BatchNorm-Silu) module, and then goes through convolution, batch normalization, and the activation function in turn, as shown in Formulas (1)–(3).
C o n v ( X ) = W c o n v X + b c o n v ,
B N X = γ · C o n v X μ σ 2 + ϵ + β ,
S i l u X = X · σ B N X ,
Among them, the convolution kernel is W c o n v , the bias is b c o n v , and the number of output channels is C . Here, μ and σ 2 are the mean and variance of the batch data, respectively. γ and β are trainable parameters, ϵ is a small constant used to prevent division by zero, and σ ( X ) is the Sigmoid function. The output from the CBS module is X C B S R C × H × W .
Then, after the EMA attention mechanism, X E M A = E M A ( X C B S ) . The EMA module outputs the same dimension through multi-scale pooling and the attention weight calculation (see the decomposition below). The Split module splits the EMA output along the channel dimension into two parts:
X S p l i t 1 , X S p l i t 2 = c h u n k X E M A , 2 ,
The output is X S p l i t 1 R C 2 × H × W , X S p l i t 2 R C 2 × H × W . Next, multiple Bottleneck modules are used to progressively extract the features. Each Bottleneck module consists of multiple convolutional layers and can be configured to use a shortcut connection (residual connection). For the i cycle ( i = 1, …, n):
X B o t t l e n e c k ( i ) = X S p l i t 2   if   i = 1 , B o t t l e n e c k ( X B o t t l e n e c k ( i 1 ) )   if   i > 1 . ,
B o t t l e n e c k X = X + C o n v 1 C o n v 1 X if   shortcut = True   and   c 1 = c 2 , C o n v 1 C o n v 1 X otherwise .   ,
Among them, the first Conv1: Conv1 (c, c, k = 3) (the number of channels is unchanged), and the second Conv1: Conv1Conv1 (c, c, k = 3). If residual connections are enabled (shortcut = True), the input X is superimposed. The output X B o t t l e n e c k = R C 2 × H × W .
After n Bottleneck processing, we obtain n feature maps X B o t t l e n e c k ( 1 ) , .... X B o t t l e n e c k ( n ) , as well as the X S p l i t 2 parts from the initial Split. In the C2f_EMA module, the Concat operation is used to concatenate the feature maps or Bottleneck outputs after the Split along the channel dimension. The concatenation process:
X C o n c a t = C o n c a t X S p l i t 1 , X B o t t l e n e c k ( 1 ) , X B o t t l e n e c k ( 2 ) , . . . . , X B o t t l e n e c k ( n ) ,
The output dimension is X C o n c a k = R C 2 × ( n + 1 ) × H × W . The overall process formula is:
X o u t = C B S C o n c a t X S p l i t 1 , X B o t t l e n e c k ( 1 ) , X B o t t l e n e c k ( 2 ) , . . . . , X B o t t l e n e c k ( n ) ,
The EMA [32] module’s architecture is detailed in Figure 5 [32]. To circumnavigate dimensionality reduction issues that are inherent in standard convolutions, the module initially restructures partial channel dimensions into batch dimensions. This is followed by the fusion of output feature maps from two parallel sub-networks, achieved through a local cross-channel interaction mechanism and an innovative cross-spatial learning paradigm. A multi-scale parallel sub-network architecture is implemented to facilitate the collaborative modeling of both short-range and long-range contextual relationships. The implementation workflow proceeds as follows:
  • Feature Encoding and Dimension Preservation:
Inspired by the CA attention mechanism, two encoded features are concatenated along the vertical spatial dimension and processed through a shared 1 × 1 convolution to maintain dimensional consistency. Input tensor: X C B S R C × H × W (output from the CBS module), divided into three groups, each with channel number: c g = C G . Then the first two groups are horizontally and vertically pooled, respectively:
X h = A d a p t i v e A v g P o o l 2 d H , 1 X C B S , X W = A d a p t i v e A v g P o o l 2 d 1 , W X C B S ,
2.
Attention Decomposition and Distribution:
The 1 × 1 convolution output is decomposed into independent vectors, with a 2D binary distribution over linear convolution approximated via a nonlinear Sigmoid activation. The specific process is as follows:
h w = C o n v 1 c o n c a t X h , X W ,
Among them, Conv1 is a 1 × 1 convolution, which is used to fuse the horizontal and vertical information.
X h , X w = s p l i t h w , H , W ,
Then it is weighted by Sigmoid:
X w e i g h t e d = G r o u p N o r m X G r o u p · σ X h , σ X w ,
X G r o u p is the initial grouping tensor, σ ( x ) is the Sigmoid function, and GroupNorm normalizes each group of features.
3.
Dual-Path Attention Aggregation:
(1)
1 × 1 Branch: Intra-group channel attention is aggregated through element-wise multiplication to achieve diversified cross-channel interactions.
(2)
3 × 3 Branch: Local cross-channel interaction information is captured via convolution, augmenting the feature space expressiveness. For the 3 × 3 branch, the 3 × 3 convolution operation is first performed as X C o n v 3 , followed by cross-channel interaction attention mechanism fusion, and finally the combined output is performed. The specific process is as follows:
(1)
Calculate the global weight of the weighted features X w e i g h t e d and X C o n v 3 respectively:
Q 1 = A d a p t i v e A v g P o o l 2 d 1,1 X w e i g h t e d ,
Q 2 = A d a p t i v e A v g P o o l 2 d 1,1 X C o n v 3 ,
A 1 = S i g m o i d r e s h a p e Q 1 , 1 , c g ,
A 2 = S i g m o i d r e s h a p e Q 2 , 1 , c g ,
(2)
Cross attention fusion:
w e i g h t s = σ M a t M u l ( A 1 , Q 2 ) + M a t M u l ( A 2 , Q 1 ) ,
(3)
Adjust the weight shape and act on the grouped features:
X O U T = X G r o u p · r e s h a p e w e i g h t s , 1 , c g ,
4.
Spatial Attention Weighting:
A Sigmoid function generates spatial attention weights, with intra-group output feature maps aggregated to model the pixel-level correlations and emphasize the global contextual significance.
This design paradigm significantly enhances the feature representation capabilities and improves the detection performance across multi-scale targets. By mitigating redundant feature interactions and optimizing multi-scale feature fusion, the EMA mechanism reduces computational complexity while maintaining the advantages of conventional attention mechanisms. This efficient attention framework addresses the limitations of excessive memory consumption and computational overhead associated with traditional approaches.

3.2.2. The Improvement of the Loss Function

The loss function measures the discrepancy between the predicted and ground truth values. A smaller loss value indicates superior model performance. The selection of the loss function is model-specific and is primarily applied during the training phase. The optimization process, which minimizes the loss value to align predictions with the ground truths, hinges on the appropriate selection of the loss function. The selection of the loss function is critical to the model performance. In the YOLOv5s model, the total loss consists of three parts: classification loss, confidence loss, and bounding box loss. Specifically, CIoU loss [35] is employed in bounding box regression, whereas Binary Cross-Entropy (BCE) loss [36] is utilized in confidence and classification tasks. The BCE loss and CIoU loss are calculated as follows:
B C E   l o s s = 1 N × n = 1 N y n × log x n + 1 y n × log 1 x n ,
where y n represents the true class, usually taking a value of 0 or 1, x n represents the predicted confidence or target probability obtained using the Sigmoid function, and N is the number of positive and negative samples.
C I o U   l o s s = 1 I o U A , B + ρ 2 A c t r , B c t r C 2 + α ν ,
ν = 4 π 2 ( a r c t a n ω g t h g t a r c t a n ω h ) 2 ,
α = ν ( 1 I o U ) + ν ,
where IoU represents the Intersection over Union of the areas of the actual and predicted bounding boxes; A and B represent the actual and predicted bounding boxes, respectively; and C represents the length of the diagonal of the smallest rectangle that encloses both A and B. A c t r and B c t r represent the center points of boxes A and B, respectively; α is a positive trade-off parameter; and ν measures the consistency of the aspect ratio. ρ 2 ( A c t r , B c t r ) represents the squared Euclidean distance between A c t r and B c t r , and ω g t , h g t ,   ω , and h represent the widths and heights of the actual and predicted bounding boxes, respectively.
From Equations (20)–(22), it is evident that when the aspect ratios of the predicted and ground-truth bounding boxes match, the aspect ratio penalty term becomes zero ( ν = 0 ). This behavior introduces notable theoretical limitations. By examining the gradients of width (w) and height (h) with respect to the aspect ratio parameter ( ν ) in the CIoU, we observe that these gradients are opposites. This indicates that w and h cannot be adjusted simultaneously, which is a critical limitation. As shown in Figure 6, the CIoU metric yields identical values for both cases (a) and (b). To mitigate this issue, we introduce MPDIoU [33], a bounding box similarity metric based on the minimum point distance, as an extension of CIoU. This metric optimizes bounding box regression by minimizing the distance between the top-left and bottom-right coordinates of the predicted and ground-truth boxes. The MPDIoU calculation is formally defined in Algorithm 1. As shown in Figure 6, MPDIoU distinguishes between cases (a) and (b).
Algorithm 1 outlines the computational process for MPDIoU:
Algorithm 1. Intersection over Union with Minimum Points Distance.
Input: Two arbitrary convex shapes: A , B S R n , width and height of input image: w ,   h
1: For A and B , ( x 1 A , y 1 A ) , ( x 2 A , y 2 A ) denote the top-left and bottom-right point coordinates of A ,   ( x 1 B , y 1 B ) , ( x 2 B , y 2 B ) denote the top-left and bottom-right point coordinates of B.
2: d 1 2 = ( x 1 B x 1 A ) 2 + ( y 1 B y 1 A ) 2 , d 2 2 = ( x 2 B x 2 A ) 2 + ( y 2 B y 2 A ) 2
3: M P D I o U = A B A B d 1 2 w 2 + h 2 d 2 2 w 2 + h 2
The Euclidean distance between the top-left vertices of A and B is denoted as d 1 2 , and the Euclidean distance between the bottom-right vertices of A and B is denoted as d 2 2 . The loss function of MPDIoU can be formulated as:
L M P D I o u = 1 M P D I o u = 1 A B A B + d 1 2 w 2 + h 2 + d 2 2 w 2 + h 2 ,
From the preceding formulation, it is evident that MPDIoU quantifies positional and dimensional discrepancies by directly computing the distance between the vertex coordinates of the predicted and ground-truth bounding boxes. As shown in Figure 6, the MPDIoU values differ between cases (a) and (b), while CIOU remains identical. This metric enables effective differentiation between scenarios where the predicted bounding box is inside versus outside the ground truth bounding box, even when maintaining the same aspect ratio. This characteristic ensures accurate bounding box regression and helps to reduce redundant predictions. The loss function depends solely on vertex coordinate distances, with the gradient computation focused exclusively on coordinate offsets, thereby eliminating mutual dependencies from the aspect ratio parameters. This enables the independent adjustment of width and height via coordinate optimization, mitigating gradient conflicts. As shown in Equation (24), all the factors considered in existing loss functions—such as the area of the non-overlapping regions, distance between the center points, width discrepancy, and height discrepancy—can be determined through the coordinates of the top-left and bottom-right corners. This implies that MPDIoU not only incorporates these evaluation dimensions but also streamlines the computational process by using corner coordinates as the sole input for regression.
C = ( m a x ( x 2 g t , x 2 p r d ) m i n ( x 1 g t , x 1 p r d ) ) * ( m a x ( y 2 g t , y 2 p r d ) m i n ( y 1 g t , y 1 p r d ) ) x C g t = x 1 g t + x 2 g t 2 , y C g t = y 1 g t + y 2 g t 2 , x C p r d = x 1 p r d + x 2 p r d 2 , y C p r d = y 1 p r d + y 2 p r d 2 w g t = x 2 g t x 1 g t , h g t = h 2 g t h 1 g t , w p r d = x 2 p r d x 1 p r d , h p r d = h 2 p r d h 1 p r d ,
where the four points ( x 1 g t , y 1 g t ), ( x 2 g t , x 2 g t ), ( x 1 p r d , y 1 p r d ), and ( x 2 p r d , y 2 p r d ) represent the coordinates of the top-left and bottom-right points between the actual and predicted bounding boxes, respectively. C represents the area of the smallest rectangle that encloses both boxes A and B. The points ( x C g t , y C g t ) and ( x C p r d , y C p r d ) represent the coordinates of the center points of the actual and predicted bounding boxes, respectively. w g t , h g t , w p r d , and h p r d represent the widths and heights of the actual and predicted bounding boxes, respectively.

3.2.3. Small Target Detection Layer

As illustrated in Figure 1, the newly introduced P2 detection branch (highlighted by the dashed box) is tailored for the detection of extremely small objects. Through a cross-level feature fusion mechanism, this branch efficiently integrates low-level features derived from shallow convolutional layers. These features preserve critical spatial structural details, including edge contours and texture patterns.
H I = H i n p u t 2 I ( I = n u m b e r   o f   l a y e r s )
However, as demonstrated by the feature map resolution formula in Equation (8), successive convolutional and pooling operations progressively reduce the feature map resolution with increasing network depth. This resolution loss has dual consequences: (1) deep features may sacrifice the geometric details of small objects (e.g., diminutive pedestrians and vehicles); and (2) channel attention mechanisms may overemphasize large-object semantics, exacerbating false/missed detections. By fusing the shallow features, the P2 branch preserves object geometric integrity, enhances bounding box regression precision for objects < 16 × 16 pixels, and consequently improves the overall detection performance. Each branch is shown in Table 1.

4. Experiment and Result Analysis

4.1. Experimental Setup

The hardware environment configuration used in the experiment is as follows: the central processing unit (CPU) is Intel(R)Core (TM)i5-12500T, and the graphics processing unit (GPU) is NVIDIA GeForceGTX3090 (equipped with 24 GB of video memory). The software environment includes CUDA11.2 and Python3.8. During the experiment, unified training parameters were used, and the model performance was further optimized through fine-tuning. The specific parameter settings are detailed in Table 2.

4.1.1. Experimental Datasets

The experiments employed the VisDrone2019 benchmark—a large-scale drone-captured dataset specifically designed for small-object detection—publicly released by research teams, including Tianjin University. This dataset comprises 6471 training samples, 548 validation samples, and 1610 testing samples, all annotated across 10 distinct object categories. Figure 7 displays representative samples from the dataset, featuring ground-truth annotations of four fundamental transportation categories: pedestrians, cars, motorcycles, and vans. The visual examples demonstrate typical object distribution patterns and scale variations in urban surveillance scenarios. Figure 8 represents the class and bounding box information for each detected object, where Figure 8a shows the data volume of the training set and the number of each category; Figure 8b illustrates the size and quantity of the bounding boxes; Figure 8c depicts the position of the center point relative to the entire image; and Figure 8d presents the aspect ratio of the target relative to the entire image.

4.1.2. Experimental Evaluation Metrics

This study adopts Precision (P), Recall (R), F1 score, mean average precision (mAP@0.5), and mAP@0.5:0.95 as the evaluation metrics. Among these, mAP@0.5 is designated as the primary evaluation criterion. The definitions are elaborated as follows:
  • Precision (P) is defined as the proportion of true positive samples among those predicted as positive by the model. The calculation formula is:
    P = T P T P + F P ,
    where TP represents True Positive, i.e., the number of samples correctly predicted as positive by the model, and FP represents False Positive, i.e., the number of samples incorrectly predicted as positive by the model.
  • Recall (R) represents the proportion of true positive samples that are correctly predicted, and the calculation formula is:
    R = T P T P + F N ,
    where FN represents False Negative, i.e., the number of samples incorrectly predicted as negative by the model.
  • F1 score (also known as the F-score or F-measure) is the harmonic mean of Precision and Recall, serving as a balanced evaluation metric that combines both a model’s precision and recall capabilities, and the calculation formula is:
    F 1 = 2 1 P + 1 R = 2 × P × R P + R ,
    where P and R, respectively, denote Precision and Recall.
  • The Average Precision (AP) for a single class is calculated by interpolating and integrating the precision–recall curve, and the formula is:
    A P = 0 1 P ( R ) d R ,
    where P (R) represents the precision at recall rate R.
  • The mean Average Precision (mAP) is the average of AP across all the classes, and the calculation formula is:
    m A P = 1 N i = 1 N A P i ,
    where N represents the total number of classes, and A P i represents the average precision of the i-th class.
Specifically, mAP@0.5 denotes the mean average precision calculated with an Intersection-over-Union (IoU) threshold of 0.5, while mAP@0.5:0.95 represents the average mAP across ten discrete IoU thresholds (0.5 to 0.95, in 0.05 increments) to systematically assess the model performance under varying localization precision requirements.
To address the computational constraints of deploying drone-based systems on mobile devices, we introduce a Comprehensive Performance Indicator (CPI).
C P I = α × m A P @ 0.5 + 1 α × 1 G F L O P S × 100 ,
As can be seen from Equation (31), the CPI takes into account both detection accuracy and inference speed. Here, α is the weighting coefficient, α 0,1 , with a default value of 0.5. When α > 0.5, greater emphasis is placed on the detection accuracy; when α < 0.5, greater emphasis is placed on the inference speed.

4.2. Contrast Experiment

4.2.1. Results of the MEP-YOLOv5s Experiments

To quantitatively validate the performance enhancement of the proposed model, a comparative analysis was conducted between MEP-YOLOv5s and the baseline YOLOv5s architecture. Figure 9 and Figure 10 present the training results of the two models on the VisDrone-2019 dataset. During the model training, the dynamic evolution of loss values is a key indicator of learning efficiency. The experimental results show that MEP-YOLOv5s demonstrates significant advantages over the baseline YOLOv5s model. Specifically, MEP-YOLOv5s exhibits a steeper initial decline in localization loss (Box Loss), indicating a faster convergence to a high-precision solution space in the bounding box regression tasks. This confirms the improved architecture’s enhanced spatial feature parsing capability, particularly in geometric constraint modeling. The quantitative analysis of confidence loss (Objectness Loss) shows that MEP-YOLOv5s reduces background–foreground classification uncertainty through multi-scale feature fusion, maintaining consistently lower loss values than the baseline and demonstrating more robust target presence judgment. However, when observing the baseline model, it is found that it first shows a straight upward trend and then declines. This indicates that the objectness loss is overfitting. At this point, the obj or obj_pw hyperparameters should be reduced to decrease the contribution of the objectness loss to the total loss. The classification loss (Classification Loss) trajectory further supports this, with MEP-YOLOv5s achieving early and sustained loss reduction, attributed to its attention enhancement module that boosts the category discrimination confidence via cross-channel feature calibration.
Precision (Precision) dynamics, measuring positive detection accuracy, reveal the improved generalization capability of MEP-YOLOv5s. While the baseline YOLOv5s shows stable but fluctuating precision, MEP-YOLOv5s achieves significant late-stage precision gains, especially in complex scenarios, indicating superior false positive suppression. This late-stage surge is linked to its progressive feature refinement mechanism, which dynamically adjusts feature pyramid contributions to enhance the final detection box quality.
Recall (Recall), assessing instance coverage, highlights differences in object retrieval between models. Despite fluctuations, MEP-YOLOv5s periodically surpasses the baseline, indicating enhanced recall in challenging samples while maintaining precision, which is valuable in dense or small-object detection.
In the detection quality assessment, MEP-YOLOv5s shows clear advantages in mAP@0.5 (IoU threshold 0.5), with a 4.2% higher mAP than the baseline, indicating better detection box screening under relaxed standards via optimized NMS. Under a stricter mAP@0.5:0.95 evaluation, MEP-YOLOv5s still achieves higher scores, validating the multi-scale feature enhancement module’s robustness in complex scenarios.
A systematic comparison confirms the comprehensive improvements in localization accuracy, classification confidence, and false positive control of MEP-YOLOv5s through architectural innovations. Its training loss curves, precision–recall balance, and multi-scale detection performance collectively demonstrate its strong adaptability to diverse detection scenarios. For practical deployment, MEP-YOLOv5s enhances the detection reliability in complex environments while preserving the real-time performance of YOLO, making it suitable for accuracy-critical applications like autonomous driving and intelligent surveillance.
Quantitative evaluations reveal that MEP-YOLOv5s outperforms the baseline model across all metrics: (1) achieving a 2.5% absolute gain in small-object detection P (Precision), (2) demonstrating a 19.5% relative improvement in R (Recall), (3) increasing mAP@0.5 by 20.1%, and (4) showing a 19.07% absolute enhancement in the CPI (Comprehensive Performance Index). The detailed results are documented in Table 3.

4.2.2. Effect Diagram

Figure 11 and Figure 12 present the detection results of the YOLOv5s model and the MEP-YOLOv5s model under both daytime and nighttime conditions. The schematic diagrams in these figures demonstrate a higher level of detail, reflecting the algorithmic advancements achieved. A comparison of the two figures reveals that the MEP-YOLOv5s model exhibits certain advantages in detecting small targets. This enhancement effectively mitigates the challenges that are typically associated with missed detections and false alarms, thereby improving the overall detection accuracy of the network.
Upon closer observation, it can be noted that during the nighttime, some small-pixel targets exhibit confidence scores below 0.4. Even our improved model struggles to avoid this issue, which is attributed to the degradation in image resolution caused by the reduced ambient lighting. To address this problem in future work, integrating infrared images for detection presents a promising direction.

4.2.3. Ablation Experiment

Ablation studies were conducted to rigorously validate the efficacy of individual algorithmic enhancements. Based on the YOLOv5s baseline, systematic evaluations were performed to quantify the performance contributions of each modified component within the MEP-YOLOv5s architecture. As detailed in Table 4, modular components were incrementally integrated while maintaining consistent optimization strategies and hyperparameter configurations. This controlled approach enabled the precise observation of metric variations across: (1) Precision (P), (2) Recall (R), (3) F1 (F1 score), (3) mAP@0.5, (4) mAP@0.5:0.95, and (5) the Comprehensive Performance Index (CPI).
Three architectural modifications were implemented: (1) Incorporation of a 3× downsampling small-object detection head into the original YOLOv5s architecture; (2) substitution of the Backbone’s C3 module with C2f, and replacement of the Neck’s C3 module with C2f-EMA; and (3) adoption of the MPDIoU loss function. The empirical findings demonstrate that: (1) The added P2 detection head yielded an 18.2-percentage-point absolute improvement in mAP@0.5, confirming that size-adaptive anchor boxes significantly mitigate missed detections due to scale discrepancies; (2) The C2f-EMA module produced an additional 18.75-percentage-point enhancement in mAP@0.5, substantiating the attention mechanism’s capability to augment feature discriminability; (3) The implementation of MPDIoU resulted in a 4% precision reduction compared with the previous configuration. This is primarily attributed to differences in the gradient characteristics between MPDIoU and CIOU, where the original learning rate led to training instability. Additionally, the weighting between MPDIoU and the other loss components (classification loss and confidence loss) requires rebalancing. However, under the same computational budget, it achieved relative improvements of 3.3% in Recall and 1.83% in mAP@0.5, demonstrating the effectiveness of cross-layer feature fusion in enhancing small-object detection capability.

4.3. Comparison with the Other Models

To evaluate the methodological contributions of MEP-YOLOv5s, we conducted comparative analyses against seven state-of-the-art object detection frameworks—Faster-RCNN, YOLOv3, YOLOv5s, YOLOv5m, YOLOv5L, YOLOv7, and YOLOv8—under identical experimental protocols. As presented in Table 5, comprehensive evaluations on the VisDrone-2019 benchmark demonstrate the superior performance of MEP-YOLOv5s across both precision (P) and recall (R) metrics. Notably, the model achieved absolute mAP@0.5 gains of 12.94, 17.72, 20.9, 24.30, 17.72, 11.25, and 8.53 percentage points over the comparative baselines, respectively, with the most substantial enhancement (20.9 percentage points) observed against YOLOv5s. Furthermore, MEP-YOLOv5s achieved the highest Comprehensive Performance Indicator (CPI,   α = 0.5) score of 22.69, surpassing all the comparative models in this metric.

4.4. Comparative Experiments on the Other Datasets

To validate the robustness and practical applicability of the proposed methodology, the MEP-YOLOv5s model underwent testing on the NWPU VHR-10 dataset. This dataset, originally published by Northwestern Polytechnical University, comprises 800 high-resolution remote-sensing images spanning ten distinct categories of geospatial objects. To assess the model’s generalization capability, a comparative analysis was conducted between MEP-YOLOv5s and alternative one-stage detection frameworks utilizing the identical NWPU VHR-10 dataset. The experimental configuration maintained consistent training parameters and evaluation metrics as previously outlined, with the detailed findings presented in Table 6.
With an mAP@0.5 of 90%, the MEP-YOLOv5s model demonstrated significant performance superiority over the established benchmark models, including the YOLOv5s baseline architecture and larger-scale variants, such as YOLOv5m and YOLOv5l. Specifically, the model achieved mAP@0.5 values that were 0.020, 0.0089, and 0.0192 higher than those of the comparable YOLOv5m, YOLOv5l, and YOLOv5s frameworks, respectively. This quantitative enhancement underscores the model’s capacity to maintain superior detection accuracy and generalization competence across diverse datasets.
Furthermore, within the UAV-captured scene dataset, the Comprehensive Performance Indicator (CPI) of the MEP-YOLOv5s framework exceeded that of the equivalently scaled models, thereby providing further empirical validation of the proposed enhancement strategies. Systematic ablation studies and comparative experiments confirmed that each innovative module introduced in this work—specifically the C2f_EMA mechanism and MPDIoU loss function—contributed positively to optimizing the YOLOv5 architecture. These findings collectively substantiate the operational viability of deploying the MEP-YOLOv5s model on UAV platforms for the efficient execution of multi-category object detection tasks.

5. Summary and Expectation

This paper proposes an efficient and practical object detection model based on YOLOv5, specifically addressing the challenges that are inherent in small object detection within UAV aerial imagery. To enhance the detection accuracy while maintaining real-time UAV detection capabilities, three effective improvement strategies are presented. First, to mitigate the difficulties in identifying small-scale objects in UAV-captured images, an additional detection layer was incorporated into the feature fusion network, thereby augmenting the model’s sensitivity to fine-grained features. Second, we developed a novel feature extraction module, termed C2f_EMA, which replaces the conventional C3 module with a modified C2f architecture and integrates the Exponential Moving Average (EMA) attention mechanism in the Neck component. This innovation demonstrably enhances the model’s capability to detect distant and diminutive objects. Third, we introduce the MPDIoU loss function to simultaneously optimize localization precision and classification accuracy. Concurrently, a novel evaluation metric, termed CPI, was designed to harmonize detection accuracy with inference speed. The experimental results across two benchmark datasets—VisDrone2019 and NWPU VHR-10—demonstrate that our proposed MEP-YOLOv5s model achieves a superior performance compared with the baseline YOLOv5, with significant improvements observed in the precision (P), mAP@0.5, and CPI metrics. These advancements underscore the model’s deployment potential in UAV-based object detection applications.
Future research directions will focus on advancing UAV aerial object detection toward lightweight and intelligent paradigms. By integrating advanced technologies, such as multi-source data fusion, self-supervised learning frameworks, and edge computing architectures, it is anticipated that real-time detection systems with enhanced accuracy and reduced latency can be realized in complex operational scenarios. Such developments will provide robust technical support for applications in smart cities, environmental monitoring, and disaster-response missions.

Author Contributions

Conceptualization, S.Z. (Shengbang Zhou), S.Z. (Song Zhang), and C.L.; methodology, S.Z. (Shengbang Zhou), S.Z. (Song Zhang), C.L. and S.L.; software, S.Z. (Shengbang Zhou) and S.Z. (Song Zhang); validation, S.Z. (Shengbang Zhou) and D.C.; formal analysis, S.Z. (Shengbang Zhou) and D.C.; investigation, S.Z. (Shengbang Zhou), S.Z. (Song Zhang), and C.L.; resources, S.Z. (Shengbang Zhou) and S.L.; data curation, S.Z. (Shengbang Zhou) and C.L.; writing—original draft preparation, S.Z. (Shengbang Zhou) and D.C.; writing—review and editing, S.Z. (Shengbang Zhou) and S.Z. (Song Zhang); visualization, S.Z. (Shengbang Zhou), S.Z. (Song Zhang), and C.L.; supervision, S.Z. (Shengbang Zhou), S.Z. (Song Zhang), C.L., S.L. and D.C.; project administration, S.Z. (Shengbang Zhou), S.Z. (Song Zhang), S.L. and D.C.; funding acquisition, S.Z. (Shengbang Zhou), C.L. and S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Guangxi Key Technologies R&D Program (No.AB241484046) and the Guangxi Science and Technology Program (No. GUIKEAB23075177).

Data Availability Statement

The publicly accessible VisDrone-2019 and the NWPU VHR-10 datasets are available for download from the following links: https://github.com/VisDrone/VisDrone-Dataset (accessed on 20 March 2025) and https://github.com/Gaoshuaikun/NWPU-VHR-10 (accessed on 10 November 2024), respectively. Publicly available web data were used in this study and can be accessed without restrictions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, Q.; Gu, J.; Huang, H. A resource-efficient online target detection system with autonomous drone-assisted IoT. IEEE Internet Things J. 2022, 9, 13755–13766. [Google Scholar] [CrossRef]
  2. Jha, S.S.; Nidamanuri, R.R. Dynamics of target detection using drone based hyperspectral imagery. In Proceedings of the International Conference on Unmanned Aerial System in Geomatics, Roorkee, India, 6–7 April 2019; Springer International Publishing: Cham, Switzerland, 2019; pp. 91–97. [Google Scholar]
  3. Rohan, A.; Rabah, M.; Kim, S.H. Convolutional neural network-based real-time object detection and tracking for parrot AR drone 2. IEEE Access 2019, 7, 69575–69584. [Google Scholar] [CrossRef]
  4. Liao, C.; Huang, J.; Zhou, F. Design of six-rotor drone based on target detection for intelligent agriculture. In Algorithms and Architectures for Parallel Processing: 20th International Conference, ICA3PP 2020, New York City, NY, USA, October 2–4, 2020, Proceedings, Part III; Springer International Publishing: Cham, Switzerland, 2020; pp. 270–281. [Google Scholar]
  5. Li, Z.; Wang, Q.; Zhang, T. UAV high-voltage power transmission line autonomous correction inspection system based on object detection. IEEE Sens. J. 2023, 23, 10215–10230. [Google Scholar] [CrossRef]
  6. García-Fernández, Á.F.; Xiao, J. Trajectory Poisson multi-Bernoulli mixture filter for traffic monitoring using a drone. IEEE Trans. Veh. Technol. 2023, 73, 402–413. [Google Scholar] [CrossRef]
  7. Viola, P.; Jones, M. Rapid object detection using a boosted cascade of simple features. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2001, Kauai, HI, USA, 8–14 December 2001; Volume 1, p. I. [Google Scholar]
  8. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  9. Felzenszwalb, P.; McAllester, D.; Ramanan, D. A discriminatively trained, multiscale, deformable part model. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  10. Zhuang, L.; Xu, Y.; Ni, B. Pedestrian Detection Using ACF Based Fast R-CNN. In Digital TV and Wireless Multimedia Communication: 14th International Forum, IFTC 2017, Shanghai, China, November 8–9, 2017, Revised Selected Papers; Springer: Singapore, 2018; Volume 815, p. 172. [Google Scholar]
  11. He, K.; Gkioxari, G.; Dollár, P. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  12. Cai, Z.; Vasconcelos, N. Cascade r-cnn. Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6154–6162. [Google Scholar]
  13. Hussain, M. YOLO-v1 to YOLO-v8, the rise of YOLO and its complementary nature toward digital manufacturing and industrial defect detection. Machines 2023, 11, 677. [Google Scholar] [CrossRef]
  14. Chen, C.; Zheng, Z.; Xu, T. Yolo-based uav technology: A review of the research and its applications. Drones 2023, 7, 190. [Google Scholar] [CrossRef]
  15. YOLOv5: An Open-Source Object Detection Algorithm. Available online: https://github.com/ultralytics/yolov5 (accessed on 15 May 2024).
  16. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar]
  17. Liu, W.; Anguelov, D.; Erhan, D. Ssd: Single shot multibox detector. In Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016, Proceedings, Part I; Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  18. Wang, Y.; Wang, C.; Zhang, H. Automatic ship detection based on RetinaNet using multi-resolution Gaofen-3 imagery. Remote Sens. 2019, 11, 531. [Google Scholar] [CrossRef]
  19. Shaodan, L.; Yue, Y.; Jiayi, L. Application of UAV-based imaging and deep learning in assessment of rice blast resistance. Rice Sci. 2023, 30, 652–660. [Google Scholar] [CrossRef]
  20. Liang, H.; Lee, S.C.; Seo, S. UAV-based low altitude remote sensing for concrete bridge multi-category damage automatic detection system. Drones 2023, 7, 386. [Google Scholar] [CrossRef]
  21. Zhao, B.; Song, R. Enhancing two-stage object detection models via data-driven anchor box optimization in UAV-based maritime SAR. Sci. Rep. 2024, 14, 4765. [Google Scholar] [CrossRef]
  22. Yang, Z.; Lian, J.; Liu, J. Infrared UAV target detection based on continuous-coupled neural network. Micromachines 2023, 14, 2113. [Google Scholar] [CrossRef] [PubMed]
  23. Li, H.; Chen, L.; Yao, Z. Intelligent identification of pine wilt disease infected individual trees using UAV-based hyperspectral imagery. Remote Sens. 2023, 15, 3295. [Google Scholar] [CrossRef]
  24. Wu, D.; Qian, Z.; Wu, D. FSNet: Enhancing Forest-Fire and Smoke Detection with an Advanced UAV-Based Network. Forests 2024, 15, 787. [Google Scholar] [CrossRef]
  25. Guo, J.; Liu, X.; Bi, L. Un-yolov5s: A uav-based aerial photography detection algorithm. Sensors 2023, 23, 5907. [Google Scholar] [CrossRef]
  26. Xu, Y.; Liu, Y.; Li, H. A Deep Learning Approach of Intrusion Detection and Tracking with UAV-Based 360° Camera and 3-Axis Gimbal. Drones 2024, 8, 68. [Google Scholar] [CrossRef]
  27. Dong, Y.; Ma, Y.; Li, Y. High-precision real-time UAV target recognition based on improved YOLOv4. Comput. Commun. 2023, 206, 124–132. [Google Scholar] [CrossRef]
  28. Su, X.; Hu, J.; Chen, L. Research on real-time dense small target detection algorithm of UAV based on YOLOv3-SPP. J. Braz. Soc. Mech. Sci. Eng. 2023, 45, 488. [Google Scholar] [CrossRef]
  29. Tang, H.; Xiong, W.; Dong, K. Radar-optical fusion detection of UAV based on improved YOLOv7-tiny. Meas. Sci. Technol. 2024, 35, 085110. [Google Scholar] [CrossRef]
  30. Chen, K.; Wang, J.; Pang, J. MMDetection: Open mmlab detection toolbox and benchmark. arXiv 2019, arXiv:1906.07155. [Google Scholar]
  31. Ramachandran, P.; Zoph, B.; Le, Q.V. Swish: A self-gated activation function. arXiv 2017, arXiv:1710.05941. [Google Scholar]
  32. Ouyang, D.; He, S.; Zhang, G. Efficient multi-scale attention module with cross-spatial learning. In Proceedings of the ICASSP 2023—2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar]
  33. Siliang, M.; Yong, X. Mpdiou: A loss for efficient and accurate bounding box regression. arXiv 2023, arXiv:2307.07662. [Google Scholar]
  34. Wang, C.-Y.; Liao, H.-Y.M.; Wu, Y.-H.; Chen, P.-Y.; Hsieh, J.-W.; Yeh, I.-H. CSPNet: A new backbone that can enhance learning capability of CNN. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 390–391. [Google Scholar]
  35. Zheng, Z.; Wang, P.; Liu, W. Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12993–13000. [Google Scholar]
  36. Guo, C.; Chen, X.; Chen, Y. Multi-stage attentive network for motion deblurring via binary cross-entropy loss. Entropy 2022, 24, 1414. [Google Scholar] [CrossRef]
Figure 1. The MEP-YOLOv5s model structure.
Figure 1. The MEP-YOLOv5s model structure.
Sensors 25 03468 g001
Figure 2. The C2f module structure diagram.
Figure 2. The C2f module structure diagram.
Sensors 25 03468 g002
Figure 3. The C3 module structure diagram.
Figure 3. The C3 module structure diagram.
Sensors 25 03468 g003
Figure 4. C2f_EMA structure diagram.
Figure 4. C2f_EMA structure diagram.
Sensors 25 03468 g004
Figure 5. Overall structure of the EMA module.
Figure 5. Overall structure of the EMA module.
Sensors 25 03468 g005
Figure 6. Two cases with different bounding box regression results. Green boxes indicate the true bounding boxes, and red boxes indicate the predicted bounding boxes. The image has a length of 6 and a width of 5. (a) Calculate the loss values of CIoU and MPDIoU when a 4 × 4 predicted bounding box fully contains a 2 × 2 ground truth bounding box. (b) Calculate the loss values of CIoU and MPDIoU when a 1 × 1 predicted bounding box is fully contained within a 2 × 2 ground truth bounding box.
Figure 6. Two cases with different bounding box regression results. Green boxes indicate the true bounding boxes, and red boxes indicate the predicted bounding boxes. The image has a length of 6 and a width of 5. (a) Calculate the loss values of CIoU and MPDIoU when a 4 × 4 predicted bounding box fully contains a 2 × 2 ground truth bounding box. (b) Calculate the loss values of CIoU and MPDIoU when a 1 × 1 predicted bounding box is fully contained within a 2 × 2 ground truth bounding box.
Sensors 25 03468 g006
Figure 7. Randomly selected images from the VisDrone-2019 dataset.
Figure 7. Randomly selected images from the VisDrone-2019 dataset.
Sensors 25 03468 g007
Figure 8. The class and bounding box information for each detected object. (a) Number of instances per category. (b) Bounding box dimensions and quantities. (c) Center point coordinates of the bounding boxes. (d) Length and width of the bounding boxes.
Figure 8. The class and bounding box information for each detected object. (a) Number of instances per category. (b) Bounding box dimensions and quantities. (c) Center point coordinates of the bounding boxes. (d) Length and width of the bounding boxes.
Sensors 25 03468 g008
Figure 9. YOLOv5s plot of the training results.
Figure 9. YOLOv5s plot of the training results.
Sensors 25 03468 g009
Figure 10. The MEP-YOLOv5s training results plot.
Figure 10. The MEP-YOLOv5s training results plot.
Sensors 25 03468 g010
Figure 11. Test result diagram of YOLOv5s.
Figure 11. Test result diagram of YOLOv5s.
Sensors 25 03468 g011
Figure 12. Test result diagram of MEP-YOLOv5s.
Figure 12. Test result diagram of MEP-YOLOv5s.
Sensors 25 03468 g012
Table 1. The anchor point setting for each detection branch.
Table 1. The anchor point setting for each detection branch.
Detection BranchAnchor Frame Configuration
P2(4,5), (8,10), (22,18)
P3(10,13), (16,30), (33,23)
P4(30,61), (62,45), (59,119)
P5(116,90), (156,198), (373,326)
Table 2. Experimental parameters.
Table 2. Experimental parameters.
Parameter NameParameter Setting
batch size16
learning rate0.01
size of the image640 × 640
number of iterations300
network depth, network width0.8, 1
Table 3. Comparison of training results.
Table 3. Comparison of training results.
ModelPrecision (%)Recall (%)F1
(%)
mAP@0.5
(%)
mAP@0.5:0.95
(%)
GFLOPSCPI
α = 0.25 α = 0.5 α = 0.75
YOLOv5s53.937.844.4336.820.593.110.0018.9327.86
MEP-YOLOv5s55.345.249.7444.225.7112.411.7122.5433.37
Table 4. Ablation test results.
Table 4. Ablation test results.
ModelPrecision (%)Recall (%)F1
(%)
mAP@0.5
(%)
mAP@0.5:0.95
(%)
GFLOPSCPI
α = 0.25 α = 0.5 α = 0.75
YOLOv5s53.238.144.4036.820.593.110.0018.9327.86
+P25743.949.5943.524.8104.111.5922.2332.86
+C2f_EMA57.344.349.9643.725.3112.411.5922.2932.99
+MPDIOU5545.950.0344.525.9112.411.7922.6933.59
Table 5. Comparison of MEP-YOLOv5s and other models on the VisDrone-2019 dataset.
Table 5. Comparison of MEP-YOLOv5s and other models on the VisDrone-2019 dataset.
ModelPrecision (%)Recall
(%)
F1
(%)
mAP@0.5
(%)
mAP@0.5:0.95
(%)
GFLOPSCPI
α = 0.25 α = 0.5 α = 0.75
Faster-RCNN---39.423251.410.1419.8929.64
YOLOv351.83944.4937.820.8154.79.9319.2228.51
YOLO5s53.238.144.4036.820.593.110.0018.9327.86
YOLOv5m51.937.143.2635.819.950.710.4218.8827.34
YOLOv5L54.138.845.1937.821.5114.110.1029.3328.56
YOLOv751.641.345.874022.8103.310.7220.4830.24
YOLOv851.540.145.094124.4165.710.7020.8030.90
MEP-YOLOv5s5545.950.0344.525.9112.411.7922.6933.59
Table 6. Comparing the experimental results of MEP-YOLOv5s and other models on the NWPU VHR-10 dataset.
Table 6. Comparing the experimental results of MEP-YOLOv5s and other models on the NWPU VHR-10 dataset.
ModelPrecision (%)Recall (%)F1
(%)
mAP@0.5
(%)
mAP@0.5:0.95
(%)
GFLOPSCPI
α = 0.25 α = 0.5 α
Faster-RCNN---74.647.6251.418.9437.4956.04
YOLOv394.983.588.8389.760.9154.722.9045.1767.43
YOLOv5s95.983.589.2788.359.693.122.8844.6866.49
YOLOv5m93.683.488.2088.259.250.723.5245.0866.64
YOLOv5L93.684.488.7689.260.4114.722.9545.0367.11
YOLOv794.682.388.0289.656103.323.1245.2867.44
YOLOv890.48285.9986.757.9165.722.1243.6565.17
MEP-YOLOv5s93.084.288.3890.059.1112.423.1645.4467.72
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, S.; Zhang, S.; Li, C.; Liu, S.; Chen, D. MEP-YOLOv5s: Small-Target Detection Model for Unmanned Aerial Vehicle-Captured Images. Sensors 2025, 25, 3468. https://doi.org/10.3390/s25113468

AMA Style

Zhou S, Zhang S, Li C, Liu S, Chen D. MEP-YOLOv5s: Small-Target Detection Model for Unmanned Aerial Vehicle-Captured Images. Sensors. 2025; 25(11):3468. https://doi.org/10.3390/s25113468

Chicago/Turabian Style

Zhou, Shengbang, Song Zhang, Chuanqi Li, Shutian Liu, and Dong Chen. 2025. "MEP-YOLOv5s: Small-Target Detection Model for Unmanned Aerial Vehicle-Captured Images" Sensors 25, no. 11: 3468. https://doi.org/10.3390/s25113468

APA Style

Zhou, S., Zhang, S., Li, C., Liu, S., & Chen, D. (2025). MEP-YOLOv5s: Small-Target Detection Model for Unmanned Aerial Vehicle-Captured Images. Sensors, 25(11), 3468. https://doi.org/10.3390/s25113468

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop