Next Article in Journal
Wearable Sensor Analysis of Movement Biomechanics and Lateralization in Dart Throwing
Previous Article in Journal
Rotary Panoramic and Full-Depth-of-Field Imaging System for Pipeline Inspection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Small-Target Detection Algorithm Based on STDA-YOLOv8

School of Mechanical and Electronic Engineering, Nanjing Forestry University, Nanjing 210037, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2025, 25(9), 2861; https://doi.org/10.3390/s25092861
Submission received: 13 March 2025 / Revised: 17 April 2025 / Accepted: 29 April 2025 / Published: 30 April 2025
(This article belongs to the Section Sensor Networks)

Abstract

:
Due to the inherent limitations of detection networks and the imbalance in training data, small-target detection has always been a challenging issue in the field of target detection. To address the issues of false positives and missed detections in small-target detection scenarios, a new algorithm based on STDA-YOLOv8 is proposed for small-target detection. A novel network architecture for small-target detection is designed, incorporating a Contextual Augmentation Module (CAM) and a Feature Refinement Module (FRM) to enhance the detection performance for small targets. The CAM introduces multi-scale dilated convolutions, where convolutional kernels with different dilation rates capture contextual information from various receptive fields, enabling more accurate extraction of small-target features. The FRM performs adaptive feature fusion in both channel and spatial dimensions, significantly improving the detection precision for small targets. Addressing the issue of a significant disparity in the number of annotations between small and larger objects in existing classic public datasets, a new data augmentation method called Copy–Reduce–Paste is introduced. Ablation and comparative experiments conducted on the proposed STDA-YOLOv8 model demonstrate that on the VisDrone dataset, its accuracy improved by 5.3% compared to YOLOv8, reaching 93.5%; on the PASCAL VOC dataset, its accuracy increased by 5.7% compared to YOLOv8, achieving 94.2%, outperforming current mainstream target detection models and small-target detection algorithms like QueryDet, effectively enhancing small-target detection capabilities.

1. Introduction

Small-target detection is a critical research direction in the field of target detection, aimed at precisely identifying objects of interest in small-target scenarios and determining their categories, shapes, positions, and other information. This provides perceptual support for various applications such as obstacle avoidance, interaction, and motion adjustment. Small-target detection has broad applicability and can be utilized in multiple areas, including intelligent surveillance, robot navigation, industrial automation, etc. By capturing distant small object information, systems can gain longer decision-making time, thereby better avoiding potential risks [1]. Therefore, researching efficient small-target detection methods holds significant application value and research importance.
Since Ross G. et al. [2] introduced the use of regions with CNN features (RCNNs) for target detection, deep learning-based target detection algorithms have evolved. Deep learning target detection algorithms are divided into two-stage and single-stage algorithms. Common two-stage algorithms include R-CNN [3], SPPNet [4], Fast RCNN [5], Faster RCNN [6], Feature Pyramid Networks (FPNs) [7], etc., comprising two separate network models: a Region Proposal Network (RPN) that generates candidate bounding boxes and a network that classifies these boxes and performs bounding box regression. Two-stage algorithms utilize RPN for precise candidate box generation, reducing false positive rates and achieving higher accuracy; however, they require computation across two stages, making them slower. Thus, two-stage algorithms are suitable for large targets and complex scenes, leaving much room for improvement in detecting small objects. Common single-stage algorithms like YOLO [8], SSD [9], RetinaNet [10], etc., contain a single network model capable of simultaneously estimating object positions and classifying them, offering faster speeds. In recent years, the YOLOv8 algorithm has been applied in the target-detection field due to its high accuracy and real-time detection advantages, yet it still exhibits high false positive and miss rates in small-target localization and detection.
The poor performance of small-target detection is mainly due to the inherent limitations of networks and imbalanced training data [11]. To obtain accurate semantic information, new detection algorithms incorporate increasing numbers of pooling and downsampling operations, causing the gradual loss of features from small objects with fewer pixels during iterations, which limits the detection performance of small targets [12]. FPN connects low-resolution, high-semantic information from higher layers and high-resolution, low-semantic information from lower layers through a top-down lateral connection, enriching the semantic information at all scales and alleviating the issue of information loss to some extent. Zhang F. et al. [13] introduced a high-resolution attention mechanism into FPN to mine small-target information. PANet [14] added a bottom-up fusion path based on FPN, allowing effective information transfer from lower to upper layers. ASFF [15] adaptively learns spatial weights at different scales, addressing the inconsistency of candidate boxes across different scale feature maps. NAS-FPN [16] and BiFPN [17] search for an effective block within FPN, repeatedly stacking it to elastically control the FPN size and improve efficiency. These structures significantly enhance the multi-scale representation capability of the network, but directly fusing information from different scales can cause semantic conflicts, limiting multi-scale feature expression and submerging small object information amidst conflicting information. Additionally, in existing classic public datasets, the number of small target annotations is far less than that of larger targets [18]. During training, the network continuously converges towards larger targets, leading to poor performance on small targets.
To address the issues of network limitations and imbalanced training data in small-target detection, this paper proposes an improved YOLOv8 small-target detection algorithm, STDA-YOLOv8 (small-target detection algorithm based on YOLOv8), which enhances the detection accuracy of small targets while ensuring real-time detection. The main contributions of this paper are as follows:
(1)
The STDA-YOLOv8 algorithm is proposed to significantly improve the performance of small-target detection by combining the Context Augmentation Module (CAM) and Feature Refinement Module (FRM).
(2)
A new data enhancement method (Copy–Reduce–Paste) is proposed, which effectively solves the problem of an insufficient number of small targets in the training data and makes the training process more balanced.
(3)
The experiments show that the mAP of the STDA-YOLOv8 model on the VisDrone dataset and the PASCAL VOC dataset reaches 93.5% and 94.2%, respectively, which is 5.3% and 5.7% higher, respectively, than that of the original YOLOv8, which reflects a significant performance advantage, and there is no significant increase in the model complexity.

2. Related Work

2.1. Target Detection Based on Deep Learning

The main purpose of target detection is to identify and locate objects in images or videos, which can be considered a regression problem. Before the advent of deep learning, traditional target detection algorithms relied on hand-crafted features, which had weak generalization capabilities and performed poorly in complex scenes.
In 2014, the introduction of the two-stage target detection algorithm R-CNN [3] ushered in the era of deep learning for target detection. In 2015, Joseph Redmon and others proposed YOLOv1 [8], which predicts the position and category of objects in an image through single forward propagation. YOLOv2 [19] improved upon YOLOv1 by introducing batch normalization, using high-resolution images for training, and adopting a fully convolutional network structure. YOLOv3 [20] employed a deeper Darknet-53 as its backbone and introduced FPN to enhance the detection ability for objects of different scales. To further improve speed and accuracy, Alexey Bochkovskiy and others proposed YOLOv4 [21], incorporating advanced techniques such as CSPNet, PANet, and SAM for more efficient feature extraction and target detection. YOLOv5 further enhanced the algorithm’s speed and accuracy while simplifying the code structure for easier use and deployment.
Wang J. et al. [22] proposed an attention-improved YOLOv5 target detection algorithm, UTD-Yolov5, which replaces the original backbone with a two-stage cascaded CSP and introduces a visual channel attention mechanism module SE. It also designed a random anchor similarity calculation method, WBF, and iterative refinement mechanisms, enabling more flexible detection and accurate feature capture. Feng Jiangfan et al. [23] presented a lightweight solution that includes a cross-stage partial bottleneck transformer (CSP BoT) module and an angular classification prediction branch. They also combined the prediction head network with an adaptive spatial feature fusion block (ASFF-Head) to further improve multi-scale feature maps, accommodating spatial variations in prediction uncertainty. Chunshan W. et al. [24] modified the BottleNeck structure to enhance the model’s multi-scale feature capture capability, embedded the SimAM attention mechanism module to focus on key features without increasing parameter count, and added Vision Transformer components to improve global feature perception within images. Zhai Xianyi et al. [25] introduced a Multi-Scale Retinex with Color Restoration (MSRCR) algorithm for image processing to enhance contrast, added a Convolutional Block Attention Module (CBAM) to improve recognition precision and efficiency, and incorporated a detect layer into the YOLOv5s head network to precisely identify small targets. Jiangjie X. et al. [26] integrated an OCAM (Object Convolution Attention Module) into the feature extraction layer of the YOLOv5 network. This module builds contextual relationships between features to enhance their relevance, establishing long-range dependencies between channel features and positional features, effectively improving detection performance.
YOLOv6 introduced the new RepVGG architecture with an anchor-free detector, enhancing the model’s adaptability for various industrial applications. YOLOv7 proposed several architectural changes and a series of free packages, optimizing model structure re-parameterization and dynamic label assignment issues to improve accuracy while maintaining real-time performance. YOLOv8 introduced new features and improvements to further enhance performance and flexibility, including a new backbone network, a new anchor-free detection head, and a new loss function, aiding in higher detection accuracy suitable for real-time target detection tasks in diverse application fields [27]. Recent studies such as Ashraf et al. [28] analyzed the performance of the YOLO series of models under degraded imaging conditions, and Varghese et al. [29] and Chai et al. [30] proposed an improved version of YOLOv8, which was optimized for model performance and small-target detection, respectively, and further promoted the application of the YOLO algorithm to real complex scenes.
As a landmark framework in the field of deep learning-based target detection, the YOLO series has continuously evolved to meet the complex and diverse demands of detection tasks [31]. However, due to the small pixel size of small objects, extracting effective information is challenging, posing significant difficulties and challenges for small-target detection. Among the current YOLO versions, YOLOv8 better maintains detection accuracy while improving detection speed. Therefore, this paper adopts YOLOv8 as the baseline framework.

2.2. Multi-Scale Feature Fusion

Multi-scale feature fusion enhances the understanding of image and video content by capturing feature information at different scales and levels, providing a solid foundation for applications in image processing, computer vision, and deep learning.
MSGNet [32] leverages frequency domain analysis and adaptive graph convolution to capture inter-series correlations across multiple temporal scales, demonstrating good generalization capabilities. CEDNet [33], with its simplified cascade encoder–decoder network designed for dense prediction tasks, performs multi-scale feature fusion in the decoder. The DAU-FI Net architecture [34], primarily aimed at semantic segmentation of imbalanced datasets, integrates a multi-scale spatial channel attention mechanism and feature injection to improve accuracy. Its core consists of a multi-scale depthwise separable convolution block and a spatial-channel compression and excitation (scSE) attention unit, modeling channel and spatial region dependencies within feature maps. DAU-FI Net utilizes additive attention gates to optimize segmentation and expand the feature space. The Multi-Scale Dilated Attention (MSDA) module [35] models local and sparse patch interactions, constructing a multi-scale dilated transformer (DilateFormer) that excels in visual tasks. The Concentrated Feature Pyramid (CFP) [36] optimizes features through global explicit supervision. CFP employs a lightweight MLP to capture global dependencies and focuses on image corner regions through a learnable visual center mechanism. This method adjusts shallow features using information extracted from deeper layers, achieving a more comprehensive and distinctive feature representation that enhances the performance of YOLOv5 and YOLOX baselines in target detection.
Despite significant progress in target detection through multi-scale feature fusion methods, there are still limitations in small-target detection tasks. For example, the FPN structure [7], which fuses high-level semantic information into low-resolution features through a top-down approach, lacks rich contextual information for small objects during processing, leading to insufficient expression of small object features and limited detection performance. Although PANet [14] improves bidirectional feature flow through a bottom-up pathway, semantic inconsistencies between features of different scales persist, especially as small objects can be overwhelmed by stronger high-level semantic information.
To address these issues, this paper proposes a Contextual Augmentation Module (CAM) and a Feature Refinement Module (FRM). The CAM captures contextual information from different receptive fields through multi-scale dilated convolution, compensating for the FPN’s shortcomings in acquiring contextual information, particularly suitable for extracting small-object features. The FRM resolves conflicts between features of different scales by adaptively fusing features along channel and spatial dimensions, ensuring that small-object features are not overshadowed during fusion. Consequently, our method effectively improves the accuracy and reliability of small-target detection.

3. Methodology

3.1. YOLOv8 Model

YOLOv8 builds upon the foundational network architecture concepts of the YOLO series, introducing various improvements and innovations. Its network structure primarily consists of three components: the backbone, neck, and head [29].
The backbone serves as the foundation of the model, adopting a structure similar to CSPDarknet, which includes the CBS module, C2f module, and SPPF module, responsible for feature extraction from input images. The C2f module ensures that YOLOv8 remains lightweight while capturing richer gradient flow information. The SPPF module replaces the SPP module in YOLOv5, reducing execution time by half while retaining the functionality of transforming feature maps of any size into fixed-size feature vectors.
The neck, positioned between the backbone and head, is composed of FPN and PAN structures, tasked with feature fusion and enhancement. As illustrated in Figure 1, compared to YOLOv5 and YOLOv6, YOLOv8 replaces the C3 module and RepBlock module with C2f and removes the 1 × 1 convolution before upsampling, directly feeding features from different stages of the backbone into the upsampling operation.
The detection head is the decision-making component of target detection, employing a decoupled head approach that separately extracts target position and category information through different network branches. These branches are trained independently before fusion to produce the final detection results. This method effectively reduces the number of parameters and computational complexity, enhancing the model’s generalization capability and robustness.

3.2. YOLOv8 Model Enhancements

Applying the YOLOv8 model to small-target detection in the field of autonomous driving still yields relatively poor performance on road surfaces [30]. To address this issue, this paper introduces a new small-target detection network structure—STDA-YOLOv8, which is integrated into the fundamental framework of YOLOv8, as shown in Figure 2.
This includes a new small-target detection network structure, as shown in Figure 3. This structure combines the Contextual Augmentation Module and the Feature Refinement Module, enhancing small-target detection performance through multi-scale dilated convolution and adaptive feature fusion.
The CAM enhances the contextual information needed for small-target detection through multi-scale dilated convolution. Compared to the traditional FPN, the CAM obtains richer contextual information using convolutional kernels with different dilation rates, especially effective at capturing detailed features of small objects at low resolutions. This improves the model’s detection of small objects without significantly increasing computational complexity.
The FRM addresses conflicts between features of different scales by adaptively fusing features along both channel and spatial dimensions. Compared to attention mechanisms like the SE module, the FRM better preserves the effective information of multi-scale features, particularly preventing small-object features from being overshadowed during fusion. This innovative design makes the FRM more suitable forsmall-target detection tasks, significantly improving detection accuracy.

3.2.1. CAM Module

In order to enrich the contextual information required for small-target detection, the CAM (Contextual Augmentation Module) is proposed, where “A” stands for augmentation, which reflects the core idea of the module to improve the detection performance by enhancing the contextual information of small targets. The CAM is applied to the neck structure of YOLOv8, where different scale feature maps (C3, C4, C5, respectively) extracted from the backbone network are used as inputs to the CAM in order to generate augmented multi-scale feature maps F1, F2, F3 (corresponding to different scale levels). Figure 4 shows the specific module structure of the CAM as an example when inputting the feature maps of layer C5 of the backbone network, which supplements the contextual information required for small-target detection by introducing multi-scale expansion convolution, using convolution kernels with different dilation rates to obtain contextual information from different receptive fields and injecting this information into the various layers of the FPN.
Dilation convolution, also known as null convolution, expands the convolution’s sense field by introducing intervals (dilation rates) in the convolution kernel without increasing the number of parameters or computational complexity. The specific calculation formula is:
y ( i , j ) = m = 0 k 1 [ n = 0 k 1 [ x ( i + r · m , j + r · n ) · w ( m , n ) ) ] ]
where r is the dilation rate and k is the convolution kernel size.
The advantage of dilated convolution is to increase the sensory field to obtain more contextual information without losing spatial resolution, which is especially important for capturing detailed features of small targets. Since small targets have fewer pixels and are easily ignored, we can efficiently extract rich contextual information from different spatial scales through multi-scale dilated convolution to enhance the recognition of small targets.
Taking convolutional kernel dilation rates of 1, 3, and 5 as examples, the formula for obtaining contextual information through multi-scale dilated convolution is as follows, and the calculation for other dilation rates follows the same principle:
C A M o u t p u t = C o n v 3 3 ( R a t e = 1 ) + C o n v 3 3 ( R a t e = 3 ) + C o n v 3 3 ( R a t e = 5 )
where Rate represents the dilation rate, and Formula (2) indicates that multi-scale contextual information is acquired through convolution kernels with different dilation rates and then performs weighted fusion of this information.
Common feature fusion methods include weighted fusion and concatenation fusion, as shown in Figure 5.
Figure 5a,b represent weighted fusion and concatenation fusion, respectively, adding feature maps directly on the spatial and channel dimensions. Concatenation fusion connects two features directly, concatenating them along the input feature’s dimensionality. Weighted fusion employs a parallel strategy, combining two feature vectors into a complex vector. The formula for the weighted fusion method is as follows:
F u s i o n w e i g h t e d = i = 1 n α i + F i
where α i is the weighting coefficient and F i is the feature map at different scales.
In Figure 4, the fusion module employs an adaptive fusion method. Assuming the input size is (N, C, H, W), through convolution, fully connected layers, and Softmax operations, it obtains spatially adaptive weights of size (N, 3, H, W). The three channels correspond to the three inputs, and contextual information is integrated into the output by calculating a weighted sum. The formula for the adaptive fusion method is as follows:
F u s i o n a d a p t i v e = S o f t m a x ( C o n v 1 1 ( C o n c a t ( F 1 , F 2 , F 3 ) ) )
S o f t m a x ( x ) = e x p ( x i ) / i e x p ( x i )
where Concat denotes the concatenation of multi-scale feature maps, C o n v 1 1 denotes dimensionality reduction by 1 1 convolution, and Softmax ensures that smaller values have lower probabilities and are not directly discarded, generating adaptive fusion weights.

3.2.2. FRM

To address the issue of redundant and conflicting information arising from direct fusion of features at different scales, which can reduce multi-scale expression capabilities, this paper introduces a Feature Refinement Module. Figure 6 shows the feature maps at different levels within the FRM. The FRM performs adaptive feature fusion along both channel and spatial dimensions, eliminating conflicting information between features of different scales to prevent small-object features from being overshadowed. This significantly enhances the fusion effect, ensuring that the feature information of small targets is preserved and reinforced, substantially improving the detection accuracy for small objects.
The FRM mainly consists of two branch modules, namely the Channel Attention Module and the Spatial Attention Module, which generate adaptive weights along the channel and spatial dimensions, respectively, enabling the network to focus on more important information. The channel attention mechanism emphasizes the relationships between channels by applying weights across them, helping the network learn the inter-channel relationships and extract significant channel information. Specifically, the input feature map is first compressed along the spatial dimension to obtain a spatial information representation containing global features of the image. Then, accurate global features are obtained through adaptive average pooling and adaptive max pooling. The formula for generating channel attention is as follows:
K m x , y = a m · X ( 1 , m ) x , y + b m · X ( 2 , m ) x , y + c m · X ( 3 , m ) x , y
[ a m , b m , c m ] = σ [ A P ( F ) + M P ( F ) ]
σ ( x ) = 1 / ( 1 + e x p ( x ) )
where K m x , y denotes the output vector of the m t h layer at position (x,y); F denotes the features generated by the concatenation operation; a m , b m , c m denotes the adaptive weights of size 1 1 1 , which are generated by average pooling (AP) and maximum pooling (MP) of the feature maps followed by the σ (sigmoid) operation.
The Spatial Attention Module operates on the spatial dimension of the feature map, capturing spatial information at different scales to extract relevant spatial locations. This is achieved by generating relative weights for each location across channels using Softmax. The formula for generating spatial attention is as follows:
ϕ m x , y = c = 1 3 k , x , y ( μ m c , x , y · X ( 1 , m ) k , x , y + υ m c , x , y · X ( 2 , m ) k , x , y + η m c , x , y · X ( 3 , m ) k , x , y )
[ μ m , υ m , η m ] = S o f t m a x ( F )
where ϕ m x , y denotes the output vector of the m t h layer at position (x,y) and μ m c , x , y , υ m c , x , y , η m c , x , y denote the spatial attention weight relative to the m t h layer, which is obtained by normalizing the feature maps using Softmax.
In summary, the generation formula for the total output of the FRM is as follows:
p m = K m + ϕ m
The features of each layer of the FRM are fused together under the guidance of adaptive weights, and the final output of the whole network is p 1 , p 2 , p 3 .
To summarize:
(1)
Unlike traditional FPN and PANet, the CAM proposed in this paper effectively solves the problem of insufficient capturing of small-target contexts by traditional structures through multi-scale expansion convolution;
(2)
The FRM performs adaptive fusion in both channel and spatial dimensions, which solves the problem of scale feature conflict in traditional feature fusion methods;
(3)
The theoretical contribution lies not only in the innovation of module structure but also in the proposed data enhancement method (Copy–Reduce–Paste) which effectively solves the training data imbalance problem with clear theoretical and practical application implications.

4. Results and Analysis

4.1. Experimental Dataset and Experimental Setup

4.1.1. Experimental Dataset

Experiments were conducted on the VisDrone dataset. The VisDrone dataset consists of 288 video segments, 261,908 frame images, and 10,209 static images, covering multiple aspects including locations (14 different cities in China), environments (urban and rural), objects (pedestrians, vehicles, bicycles, etc.), and densities (sparse and crowded scenes). It includes manually annotated bounding boxes for over 2.6 million targets such as pedestrians, cars, bicycles, and tricycles. To validate the applicability of STDA-YOLOv8, we also tested it on the PASCAL VOC dataset to ensure its effectiveness in small-target detection. These datasets were divided into training and testing sets at a ratio of 7:3. The training set was used to train the model, while the test set was used to detect the model’s detection accuracy and evaluate its generalization ability.

4.1.2. Data Enhancement Method: Copy–Reduce–Paste (CRP)

In existing classic public datasets, the number of small-target annotations is significantly less than that of larger targets, causing the network to continuously converge towards larger targets during training and resulting in poor performance on small targets. To address this issue, this paper proposes a novel data augmentation method, called “Copy–Reduce–Paste” data augmentation. This method involves copying larger targets, shrinking their size, and then pasting them back into the original image to increase the contribution of small objects to the loss function during training, ensuring balanced training. In this study, a bounding box area larger than 64 × 64 pixels (corresponding to the input image size of 640 × 640 pixels) is defined as a ‘larger’ object, and a bounding box area smaller than 32 × 32 pixels is defined as a ‘small’ object.
For specific implementation, this paper first identifies larger targets with an area larger than 64 × 64 pixels in the original image, copies these targets and reduces them by a random proportion (usually 30% to 50% of the original size), and then randomly selects a blank area in the image for pasting. In order to prevent the pasted targets from overlapping with the existing targets, this paper implements a simple verification step: if there is any intersection between the randomly selected location and the location of the existing targets in the original image, a new blank location is randomly selected again to be pasted until the no-overlap condition is satisfied.
To further validate the effectiveness of this data augmentation method, we conducted a comparative analysis of the datasets before and after enhancement and evaluated the number of small targets and model performance under different scene types in detail. Table 1 shows the changes in the number of small targets and model performance metrics such as precision, recall, and mean average precision (mAP) before and after data augmentation on the VisDrone and PASCAL VOC datasets across different scenes, where ‘Quantity’ represents the quantity of small targets, ‘Proportion’ represents the proportion of small targets and large targets, ‘before’ indicates the situation before data augmentation, and ‘after’ indicates the situation after data augmentation.
From the table, it is evident that data augmentation effectively increased the proportion of small targets in the training data, significantly improving the model’s precision, recall, and mAP across different scenes. The performance enhancement was particularly notable in sparse and crowded scenarios.

4.1.3. Experimental Setup

To ensure the accuracy of the experiments, all experiments were conducted on a Windows 11 operating system. The GPU platform used was an NVIDIA RTX 2060, and the hardware configuration remained consistent. The deep learning framework employed was Pytorch 2.2.2, with CUDA version 12.1. The programming language used was Python, version 3.9.19. Specific training parameter settings are detailed in Table 2.

4.2. Evaluation Metrics

Precision (P), recall (R) and mean average precision (mAP) are selected as the metrics to evaluate the model’s detection performance.
Precision measures the proportion of correctly predicted positive samples out of the total number of samples, while recall measures the proportion of correctly predicted positive samples out of the total labeled samples. The formulas are as follows:
P = T P / ( T P + F P )
R = T P / ( T P + F N )
where TP (true positives), FP (false positives) and FN (false negatives) represent the number of correctly detected targets, falsely detected backgrounds, and falsely detected targets as backgrounds, respectively.
Mean average precision takes into account the average precision (AP) of each category as well as the number of categories (C). The formula is as follows:
m A P = i = 1 C A P i / C

4.3. Results and Comparison

4.3.1. Ablation Study

To verify the effectiveness of each module in the model, this paper conducted ablation studies by adding modules one by one. Based on the original YOLOv8 model, we incorporated the CAM and the FRM. P, R, mAP, parameters, floating-point operations, and model size were used as evaluation metrics. The results of the ablation experiments are shown in Table 3 below.
It can be observed that the modules designed in this paper have improved the model’s detection performance for small targets to a certain extent. Building on the foundation of YOLOv8, adding CAMs with different dilation rates allows the model to capture features at various scales and levels, thereby enhancing the detection performance for small targets. Using smaller dilation rates (such as [1, 2, 3]) can maintain the precision of details and improve the detection rate of small objects, especially when the background is simple; using larger dilation rates (such as [2, 4, 6]) enhances the recognition ability of small targets in more complex backgrounds by increasing the coverage of the receptive field. The combination experiments of different dilation rates comprehensively demonstrate the effectiveness and flexibility of the CAM and its performance changes under different configurations, optimizing the model to find the best configuration for small-target-detection scenarios.
On top of adding the CAM, further improvements were made to YOLOv8 with the FRM, resulting in further enhancements in precision, recall, and m A P 0.5 . Additionally, both the number of parameters and computational load decreased, indicating that the FRM not only reduced the model’s computational complexity and parameter count but also strengthened the learning capability of multi-scale features to some extent.
In some experimental data, while the number of parameters decreased, the model size increased. This could be due to reducing the parameter count while introducing non-parameterized structures or modules, leading to an increase in overall storage requirements for the model. For example, adding the FRM on top of CAMs with dilation rates of [1, 3, 5] reduced the parameter count, but this module might include more feature processing operations that require storing additional information within the model, slightly increasing the overall model size. Therefore, when improving and optimizing a model, it is essential to consider the specific impacts of each module on the number of parameters and model size to achieve optimal performance and resource utilization.
The experimental results show that when introducing three sets of CAMs with different dilation rates along with the FRM, the resulting STDA-YOLOv8 model achieved a 15.5% improvement in precision, a 12.6% improvement in recall, and a 6% improvement in m A P 0.5 compared to YOLOv8 on the VisDrone dataset; on the PASCAL VOC dataset, there was a 11.5% improvement in precision, a 10.1% improvement in recall, and a 6.4% improvement in m A P 0.5 . Additionally, both the model’s parameter count and size were smaller than those of the baseline model, mainly due to the synergistic effect of the two modules. By reducing redundant features and optimizing multi-scale feature fusion, the complexity of the model was reduced. This optimization not only reduced the computational load but also decreased the model’s storage footprint, making the model more efficient.
In summary, the proposed STDA-YOLOv8 algorithm exhibits outstanding performance in small-target detection.

4.3.2. Comparative Experiment

In order to evaluate the detection effect of the improved model, performance comparison experiments are conducted between STDA-YOLOv8 and the mainstream target detection models YOLOv5, YOLOv7, YOLOv8, the small-target detection algorithm QueryDet, and the target detection algorithms in the studies [37,38].The experiments are conducted using the same equipment, environment, and dataset for training and testing, and the results of the comparison experiments are shown in Table 4.
As can be seen from Table 4, the improved STDA-YOLOv8 model has an mAP metric of 93.5% on the VisDrone dataset, which is a 5.3% improvement compared to the original YOLOv8; a 10.5% improvement in recall; and a 12% improvement in precision. On the PASCAL VOC dataset, the mAP metric of STDA-YOLOv8 is 94.2%, which improves 5.7% compared to YOLOv8, and it improves 8.5% in recall and 9.3% in precision. The above results fully demonstrate that STDA-YOLOv8 outperforms YOLOv5, YOLOv7, YOLOv8, QueryDet, and the recently proposed YOLO-Z [37] and One-Shot PCB [38] methods, demonstrating significant performance advantages in small-target detection tasks. In addition, the number of parameters of this paper’s method is only 2.6 × 10 6 , and the computational complexity is 7.5 G F L O P s , which are lower than the above comparative models, reflecting that the algorithm effectively reduces the complexity and resource requirements of the model while improving the performance and takes into account the requirements of high efficiency and real time, which makes it suitable for a wider range of practical application scenarios.
Figure 7 shows a comparison of some sample detection results on the VisDrone dataset. Among them, the original YOLOv8 algorithm is used in left side of figure, and it can be observed that there are obvious false positives, false negatives, and serious overlapping bounding boxes in the white magnified box region, and in right side of figure, these false positives, false negatives, and serious overlapping bounding boxes are significantly reduced and the overlapping bounding boxes are effectively improved in the same white zoomed-in box region, which shows that the proposed algorithm STDA-YOLOv8 has obvious advantages in the accuracy and robustness of small-target detection.
To verify the applicability of STDA-YOLOv8, we also conducted testing on the PASCAL VOC dataset to ensure the algorithm’s effectiveness in small-target detection. Figure 8 shows a comparison of some sample detection results on the PASCAL VOC dataset. Among them, left side of Figure 8 shows the detection results of the original YOLOv8 algorithm, with multiple obvious misdetections and omissions in the blue zoomed-box region and insufficient recognition accuracy of small targets; right side of Figure 8 shows the detection results of the STDAYOLOv8 algorithm, with a significant reduction in misdetections and omissions in the same region and a significant increase in the detection accuracy and stability of small targets, which verifies the effectiveness of this paper’s method in a real environment.

4.3.3. Evaluation of Computational Efficiency on Different Hardware Platforms

To comprehensively evaluate the computational efficiency of the proposed STDA-YOLOv8 model, experiments were conducted on different hardware platforms, including both CPU and GPU platforms. The running time and memory usage of the model were measured on these platforms. The specific hardware platforms used are as follows:
(1)
CPU platform: Intel i7-12700K with 16 GB of RAM;
(2)
GPU platforms: NVIDIA RTX 2060 with 8 GB of VRAM; NVIDIA RTX 3080 with 10 GB of VRAM.
The evaluation of computational efficiency on different hardware platforms is shown in Table 5.
When running on the CPU, the inference time of STDA-YOLOv8 increased by approximately 90 ms compared to the original YOLOv8. This is because STDA-YOLOv8 introduces more convolution operations (such as multi-scale expansion convolution and feature fusion modules), increasing the computational complexity. The memory usage also increased slightly, mainly because the CAM and the FRM need to store more intermediate features.
On the GPU platform, the increase in inference time of STDA-YOLOv8 compared to the original YOLOv8 was smaller. For the RTX 2060, the inference time increased by about 10 ms, while for the RTX 3080, the increase was about 8 ms. This indicates that the parallel acceleration of the GPU can effectively mitigate the consumption brought by additional convolution operations to a certain extent. The VRAM usage also increased slightly, but the increase was relatively small, indicating that the model optimization design achieved a good balance in VRAM usage.
In summary, although the improved STDA-YOLOv8 has a slight increase in inference time and memory usage, its detection accuracy and recall rate have significantly improved compared to the original model. Such a design makes a certain trade-off between accuracy and efficiency. Meanwhile, on high-performance GPU platforms (such as RTX 3080), STDA-YOLOv8 can maintain lower inference time and moderate VRAM usage, making it more applicable in practice, especially in tasks requiring high-precision detection of small targets.

5. Summary and Outlook

To address the issues of false positives and missed detections in small-target detection caused by network limitations and imbalanced training data, this paper proposes an improved YOLOv8 algorithm for small-target detection, named STDA-YOLOv8. This algorithm introduces a new small-target detection network structure into the YOLOv8 model, which combines a Contextual Augmentation Module (CAM) and a Feature Refinement Module (FRM). The CAM incorporates multi-scale expansion convolutions, with different dilation rates to capture contextual information from various receptive fields. The FRM performs adaptive feature fusion in both channel and spatial dimensions, eliminating conflicting information between features of different scales to prevent the features of small objects from being overwhelmed. Additionally, a new data augmentation method called Copy–Reduce–Paste is proposed, increasing the contribution of small objects to the loss function during training and ensuring balanced training. Ablation experiments on the VisDrone dataset and the PASCAL VOC dataset verified the effectiveness of the above improvements, and comparative experiments showed that the STDA-YOLOv8 algorithm outperformed mainstream algorithms in small-target detection.
The study by Ashraf [28] shows that there is degradation in the performance of the YOLO series models in bad weather environments, and thus how to improve the model performance through more advanced data enhancement techniques and more robust feature extraction modules will become an important research direction. Target detection in degraded imaging environments (e.g., bad weather, low light, etc.) will also be further explored. Efforts will also continue to improve the model’s network structure, such as introducing lightweight Transformers to simplify the model while maintaining detection performance or adding Convolutional Block Attention Modules to enhance recognition accuracy and efficiency. Furthermore, additional suitable data augmentation methods for small-target detection will be explored, such as adversarial data augmentation or methods based on generative adversarial networks, to address the issue of imbalanced training data.

Author Contributions

Conceptualization, C.L.; methodology, C.L.; software, C.L.; validation, C.L., S.J. and X.C.; formal analysis, C.L.; investigation, S.J.; resources, S.J.; data curation, X.C.; writing—original draft preparation, C.L.; writing—review and editing, S.J.; visualization, C.L.; supervision, S.J.; project administration, C.L.; funding acquisition, S.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the 2023 Jiangsu Graduate Student Research Innovation Program under the project titled ’Research on Outdoor Mobile Robot Path Planning Method Based on Visual Perception’ [No. KYCX23_1140]. It is also supported by The National Special Research Fund for Non-profit Sector [201404402-03].

Institutional Review Board Statement

Not application.

Informed Consent Statement

Not application.

Data Availability Statement

The original data presented in the study are openly available in VisDrone dataset at “https://github.com/VisDrone/VisDrone-Dataset, accessed on 13 March 2025” and PASCAL VOC dataset at “http://host.robots.ox.ac.uk/pascal/VOC/, accessed on 13 March 2025”.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Sui, J.; Chen, D.; Zheng, X.; Wang, H. A New Algorithm for Small Target Detection From the Perspective of Unmanned Aerial Vehicles. IEEE Access 2024, 12, 29690–29697. [Google Scholar] [CrossRef]
  2. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Region-Based Convolutional Networks for Accurate Object Detection and Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 142–158. [Google Scholar] [CrossRef] [PubMed]
  3. Uijlings, J.R.; Van De Sande, K.E.; Gevers, T.; Smeulders, A.W. Selective Search for Object Recognition. Int. Comput. Vis. 2013, 104, 154–171. [Google Scholar] [CrossRef]
  4. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef]
  5. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar] [CrossRef]
  6. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  7. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar] [CrossRef]
  8. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the Computer Vision & Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; IEEE: Piscataway, NJ, USA, 2016. [Google Scholar] [CrossRef]
  9. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Lecture Notes in Computer Science; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer: Cham, Switzerland, 2016; Volume 9905. [Google Scholar] [CrossRef]
  10. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar] [CrossRef]
  11. Kisantal, M.; Wojna, Z.; Murawski, J.; Naruniec, J.; Cho, K. Augmentation for small object detection. arXiv 2019, arXiv:1902.07296. [Google Scholar] [CrossRef]
  12. Liu, Y.; Sun, P.; Wergeles, N.; Shang, Y. A Survey and Performance Evaluation of Deep Learning Methods for Small Object Detection. Expert Syst. Appl. 2021, 172, 114602. [Google Scholar] [CrossRef]
  13. Zhang, F.; Jiao, L.; Li, L.; Liu, F.; Liu, X. MultiResolution Attention Extractor for Small Object Detection. arXiv 2020, arXiv:2006.05941. [Google Scholar]
  14. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8759–8768. [Google Scholar] [CrossRef]
  15. Liu, S.; Huang, D.; Wang, Y. Learning Spatial Fusion for Single-Shot Object Detection. arXiv 2019, arXiv:1911.09516. [Google Scholar]
  16. Ghiasi, G.; Lin, T.-Y.; Le, Q.V. NAS-FPN: Learning Scalable Feature Pyramid Architecture for Object Detection. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 7029–7038. [Google Scholar] [CrossRef]
  17. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10778–10787. [Google Scholar] [CrossRef]
  18. Chen, Y.; Zhang, P.; Li, Z.; Li, Y.; Zhang, X.; Meng, G.; Xiang, S.; Sun, J.; Jia, J. Stitcher: Feedback-driven Data Provider for Object Detection. arXiv 2020, arXiv:2004.12432. [Google Scholar] [CrossRef]
  19. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar] [CrossRef]
  20. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018. [Google Scholar] [CrossRef]
  21. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020. [Google Scholar] [CrossRef]
  22. Wang, J.; Yu, N. UTD-Yolov5: A Real-time Underwater Targets Detection Method based on Attention Improved YOLOv5. arXiv 2022, arXiv:2207.00837. [Google Scholar]
  23. Feng, J.; Wang, J.; Qin, R. Lightweight detection network for arbitrary-oriented vehicles in UAV imagery via precise positional information encoding and bidirectional feature fusion. Int. J. Remote Sens. 2023, 44, 4529–4558. [Google Scholar] [CrossRef]
  24. Wang, C.; Sun, W.; Wu, H.; Zhao, C.; Teng, G.; Yang, Y.; Du, P. A Low-Altitude Remote Sensing Inspection Method on Rural Living Environments Based on a Modified YOLOv5s-ViT. Remote Sens. 2022, 14, 4784. [Google Scholar] [CrossRef]
  25. Zhai, X.; Wei, H.; He, Y.; Shang, Y.; Liu, C. Underwater Sea Cucumber Identification Based on Improved YOLOv5. Appl. Sci. 2022, 12, 9105. [Google Scholar] [CrossRef]
  26. Xu, J.; Zou, Y.; Tan, Y.; Yu, Z. Chip Pad Inspection Method Based on an Improved YOLOv5 Algorithm. Sensors 2022, 22, 6685. [Google Scholar] [CrossRef]
  27. Sohan, M.; Sai Ram, T.; Rami Reddy, C.V. A Review on YOLOv8 and Its Advancements. In Proceedings of the International Conference on Data Intelligence and Cognitive Informatics, Tirunelveli, India, 18–20 November 2024; Jacob, I.J., Piramuthu, S., Falkowski-Gilski, P., Eds.; Springer: Singapore, 2024. [Google Scholar] [CrossRef]
  28. Ashraf, I.; Hur, S.; Kim, G.; Park, Y. Analyzing Performance of YOLOx for Detecting Vehicles in Bad Weather Conditions. Sensors 2024, 24, 522. [Google Scholar] [CrossRef]
  29. Varghese, R.; Sambath, M. YOLOv8: A Novel Object Detection Algorithm with Enhanced Performance and Robustness. In Proceedings of the 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS), Chennai, India, 18–19 April 2024; pp. 1–6. [Google Scholar] [CrossRef]
  30. Chai, W.; Han, D.; Zhou, H.; Wang, S.; Zhou, F. FDW-YOLOv8: A Lightweight Unmanned Aerial Vehicle Small Target Detection Algorithm Based on Enhanced YOLOv8. In Proceedings of the 2024 IEEE International Workshop on Radio Frequency and Antenna Technologies (iWRF&AT), Shenzhen, China, 31 May–3 June 2024; pp. 368–373. [Google Scholar] [CrossRef]
  31. Hussain, M. YOLOv1 to v8: Unveiling Each Variant–A Comprehensive Review of YOLO. IEEE Access 2024, 12, 42816–42833. [Google Scholar] [CrossRef]
  32. Cai, W.; Liang, Y.; Liu, X.; Feng, J.; Wu, Y. MSGNet: Learning Multi-Scale Inter-Series Correlations for Multivariate Time Series Forecasting. arXiv 2023, arXiv:2401.00423. [Google Scholar] [CrossRef]
  33. Zeng, T.; Wu, B.; Zhou, J.; Davidson, I.; Ji, S. Recurrent Encoder-Decoder Networks for Time-Varying Dense Prediction. In Proceedings of the 2017 IEEE International Conference on Data Mining (ICDM), New Orleans, LA, USA, 18–21 November 2017; pp. 1165–1170. [Google Scholar] [CrossRef]
  34. Alshawi, R.; Hoque, T.; Ferdaus, M.; Abdelguerfi, M.; Niles, K.N.; Prathak, K.; Tom, J.; Klein, J.; Mousa, M.; Lopez, J.J. Dual Attention U-Net with Feature Infusion: Pushing the Boundaries of Multiclass Defect Segmentation. arXiv 2023, arXiv:2312.14053. [Google Scholar]
  35. Jiao, J.; Tang, Y.M.; Lin, K.Y.; Gao, Y.; Ma, A.J.; Wang, Y.; Zheng, W.S. DilateFormer: Multi-Scale Dilated Transformer for Visual Recognition. IEEE Trans. Multimed. 2023, 25, 8906–8919. [Google Scholar] [CrossRef]
  36. Quan, Y.; Zhang, D.; Zhang, L.; Tang, J. Centralized Feature Pyramid for Object Detection. IEEE Trans. Image Process. 2023, 32, 4341–4354. [Google Scholar] [CrossRef] [PubMed]
  37. Benjumea, A.; Teeti, I.; Cuzzolin, F.; Bradley, A. YOLO-Z: Improving small object detection in YOLOv5 for autonomous vehicles. arXiv 2021, arXiv:2112.11798. [Google Scholar] [CrossRef]
  38. Spadaro, G.; Vetrano, G.; Penna, B.; Serena, A.; Fiandrotti, A. Towards One-Shot PCB Component Detection with YOLO. In Image Analysis and Processing-ICIAP 2023 Workshops, Proceeding of the 22nd International Conference on Image Analysis and Processing, Udine, Italy, 11–15 September 2023; Springer: Cham, Switzerland, 2024; Volume 14365, pp. 51–61. [Google Scholar]
Figure 1. The neck structures of various YOLO algorithms.
Figure 1. The neck structures of various YOLO algorithms.
Sensors 25 02861 g001
Figure 2. STDA-YOLOv8 general structure.
Figure 2. STDA-YOLOv8 general structure.
Sensors 25 02861 g002
Figure 3. Structure of small-target detection network.
Figure 3. Structure of small-target detection network.
Sensors 25 02861 g003
Figure 4. CAM structure.
Figure 4. CAM structure.
Sensors 25 02861 g004
Figure 5. Feature fusion method.
Figure 5. Feature fusion method.
Sensors 25 02861 g005
Figure 6. FRM structure.
Figure 6. FRM structure.
Sensors 25 02861 g006
Figure 7. Plot of some of the detection results on the VisDrone dataset.
Figure 7. Plot of some of the detection results on the VisDrone dataset.
Sensors 25 02861 g007
Figure 8. Plot of some of the detection results on the PASCAL VOC dataset.
Figure 8. Plot of some of the detection results on the PASCAL VOC dataset.
Sensors 25 02861 g008
Table 1. Performance comparison table before and after data enhancement.
Table 1. Performance comparison table before and after data enhancement.
DatasetScene TypeQuantityQuantityProportionProportionP/%P/%R/%R/% mAP 0.5 /% mAP 0.5 /%
(Before) (After) (Before) (After) (Before) (After) (Before) (After) (Before) (After)
VisDroneUrban scene2004001:51:277.682.183.385.388.289.6
VisDroneRural scene1503001:41:279.884.284.286.88990
VisDroneSparse scene1202401:61:38186.584.788.489.891.3
VisDroneCrowded scene1803601:51:280.285.983.587.690.591.9
PASCAL VOCUniversal scene2004001:31:1.582.386.885.488.59092.2
Overall average 85017001:4.61:2.180.285.184.287.389.591
Table 2. Training parameter settings.
Table 2. Training parameter settings.
ParameterValue
Epochs600
Patience30
Batch8
Imgsz8
Workers4
OptimizerSGD
Weight_decay0.0005
Irf0.05
Momentum0.937
Warmuop_momentum0.8
Close_mosaic10
Patience50
Table 3. Ablation study results.
Table 3. Ablation study results.
YOLOv8CAMCAMCAMFRMVisDroneVisDroneVisDronePASCAL VOCPASCAL VOCPASCAL VOCParams/ GFLOP s Model Size/
(1, 3, 5) (2, 4, 6) (1, 2, 3) P/% R/% mAP 0.5 /% P/% R/% mAP 0.5 /% 10 6 MB
77.683.388.2818488.53.18.16.2
82.185.389.685.58790.32.57.46.3
83.58790.286.887.5912.67.66.4
84.286.89086.587.291.22.57.56.4
85.98891.387.988.5922.67.66.3
87.289.29288.589.592.82.67.56.3
8788.991.888.28992.52.57.46.2
88.69192.889.590.593.22.67.56.2
88.390.792.589.19093.32.67.56.2
89.693.893.590.392.594.22.67.56.1
Table 4. Results of comparative experiments.
Table 4. Results of comparative experiments.
Model (ModelVisDroneVisDroneVisDronePASCAL VOCPASCAL VOCPASCAL VOCParams/ 10 6 GFLOP s
Size: Small) P/% R/% mAP 0.5 /% P/% R/% mAP 0.5 /%
YOLOv577.878.684.380.983.485.52.47.5
QueryDet72.178.282.580.582.683.84.110.8
YOLOv776.883.285.680.783.185.85.711.3
YOLOv877.683.388.2818488.53.18.1
literatures [37]80.585.285.782.385.488.94.39.7
literatures [38]78.982.183.981.283.686.74.710.2
STDA-YOLOv889.693.893.590.392.594.22.67.5
Table 5. Computational efficiency evaluation.
Table 5. Computational efficiency evaluation.
Hardware PlatformModelAverage Reasoning Time (ms)Memory Usage (MB)GPU Video Memory Usage (MB)
Intel i7-12700KYOLOv84202900
Intel i7-12700KSTDA-YOLOv85103100
NVIDIA RTX 2060YOLOv838 2200
NVIDIA RTX 2060STDA-YOLOv848 2600
NVIDIA RTX 3080YOLOv821 1600
NVIDIA RTX 3080STDA-YOLOv829 1800
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, C.; Jiang, S.; Cao, X. Small-Target Detection Algorithm Based on STDA-YOLOv8. Sensors 2025, 25, 2861. https://doi.org/10.3390/s25092861

AMA Style

Li C, Jiang S, Cao X. Small-Target Detection Algorithm Based on STDA-YOLOv8. Sensors. 2025; 25(9):2861. https://doi.org/10.3390/s25092861

Chicago/Turabian Style

Li, Cun, Shuhai Jiang, and Xunan Cao. 2025. "Small-Target Detection Algorithm Based on STDA-YOLOv8" Sensors 25, no. 9: 2861. https://doi.org/10.3390/s25092861

APA Style

Li, C., Jiang, S., & Cao, X. (2025). Small-Target Detection Algorithm Based on STDA-YOLOv8. Sensors, 25(9), 2861. https://doi.org/10.3390/s25092861

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop