Next Article in Journal
Integrating Multi-Source Data for Aviation Noise Prediction: A Hybrid CNN–BiLSTM–Attention Model Approach
Previous Article in Journal
Language-Driven Cross-Attention for Visible–Infrared Image Fusion Using CLIP
Previous Article in Special Issue
Review of State Estimation Methods for Autonomous Ground Vehicles: Perspectives on Estimation Objects, Vehicle Characteristics, and Key Algorithms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent Vehicle Target Detection Algorithm Based on Multiscale Features

1
School of Automotive Engineering, Shandong Jiaotong University, Jinan 250357, China
2
Department of Environmental Economics and Sustainability, Budapest University of Technology and Economics, 1111 Budapest, Hungary
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(16), 5084; https://doi.org/10.3390/s25165084
Submission received: 4 July 2025 / Revised: 12 August 2025 / Accepted: 12 August 2025 / Published: 15 August 2025

Abstract

To address the issues of false detections and missed detections in object detection for intelligent driving scenarios, this study focuses on optimizing the YOLOv10 algorithm to reduce model complexity while enhancing detection accuracy. The method involves three key improvements. First, it involves the design of multi-scale flexible convolution (MSFC), which can capture multi-scale information simultaneously, thereby reducing network stacking and computational load. Second, it reconstructs the neck network structure by incorporating Shallow Auxiliary Fusion (SAF) and Advanced Auxiliary Fusion (AAF), enabling better capture of multi-scale features of objects. Third, it improves the detection head through the combination of multi-scale convolution and channel adaptive attention mechanism, enhancing the diversity and accuracy of feature extraction. Results show that the improved YOLOv10 model has a size of 13.4 MB, meaning a reduction of 11.8%, and that the detection accuracy mAP@0.5 reaches 93.0%, outperforming mainstream models in comprehensive performance. This work provides a detection framework for intelligent driving scenarios, balancing accuracy and model size.

1. Introduction

Object detection is the cornerstone of intelligent driving systems [1], enabling vehicles to perceive obstacles, pedestrians, and traffic signs around them in real time and respond accordingly. However, the complex road environment, characterized by different object scales and severe occlusion, poses significant challenges to the existing models [2]. This often leads to incorrect detection, missed detection, and low deployment efficiency on resource-limited in-vehicle hardware.
Traditional object detection methods rely on manual feature extraction [3], which cannot capture deep semantic information and limits performance in complex scenarios. The emergence of deep learning has completely transformed this field. Single-stage detectors like YOLO [4] and SSD [5] have become leaders due to the balance they strike between speed and accuracy. Since its launch in 2016, YOLO has undergone rapid iteration, continuously optimizing accuracy and latency in practical applications. According to recent research results, many applications for YOLOv10 have been explored in object detection. Du Hong et al. [6] proposed a YOLOV10-Vehicle object detection algorithm. A new attention mechanism module, WT-PSA, was designed to enhance the model’s attention to vehicle targets in complex weather conditions. Chen Haixiu et al. [7] proposed an improved pedestrian small-target detection algorithm based on YOLOv10n. This algorithm integrates adaptive point movement and convolutional gating mechanisms into the C2f module, significantly enhancing its ability to capture local features and improving feature screening. Junhui Mei et al. [8] proposed a lightweight object detection algorithm, BGF-YOLOv10, which combines variants of BoTNet, C2f, and C3 in the backbone network to enhance detection performance regarding small objects. However, even the latest version still struggles with small targets and partially occluded vehicles in heavy traffic and has difficulties with multi-scale object representation and occlusion processing.
Against this backdrop, the limitations of current methods in intelligent driving scenarios are evident:
(1)
Environmental complexity degrades feature quality, increasing detection errors.
(2)
Extreme scale variations—from distant cyclists to nearby trucks—strain fixed-receptive-field convolutions, reducing localization accuracy.
(3)
Severe occlusions obscure critical features, leading to missed or misclassified targets.
To address these issues, this study proposes targeted improvements to YOLOv10s, aiming to enhance detection precision while reducing model complexity. The core contributions are threefold:
(1)
A multi-scale flexible convolution (MSFC) to dynamically adapt to varying feature scales, reducing computational overhead.
(2)
A reconstructed neck network integrating Shallow Auxiliary Fusion (SAF) and Advanced Auxiliary Fusion (AAF) to optimize multi-scale feature interaction.
(3)
A SEAM, combining multi-scale convolutions and channel attention, to boost feature extraction robustness.
These innovations collectively mitigate the limitations of existing models in complex traffic scenarios, paving the way for more-reliable intelligent driving perception systems.
The structure of this article is as follows: Section 1 introduces the background, limitations of existing methods, and core contributions of the study. Section 2 reviews related work on YOLO series evolution and YOLOv10 research. Section 3 details the improved YOLOv10 target detection algorithm and its key components. Section 4 details the training and validation of the model, as well as the comparison and ablation experiment to verify the effectiveness of the algorithm in this paper. Section 5 discusses the study’s results, limitations, and future directions. And Section 6 draws conclusions to summarize the whole paper.

2. Related Work

2.1. Target Detection Method

The application of object detection technology in the field of intelligent driving needs to simultaneously meet the three core requirements of real-time performance, multi-scale adaptability, and anti-interference ability [9]. At present, the mainstream methods can be classified into two-stage detectors, single-stage detectors, and anchor-free detectors, etc. Each type of method has specific advantages and limitations in intelligent driving scenarios. Two-stage detectors (such as Faster R-CNN [10] and Mask R-CNN [11]) have relatively high accuracy, but they have slow reasoning speed and large computational load, making it difficult to adapt to real-time scenarios. Single-stage detectors (such as SSD [5] and Retina Net [12]) are more efficient, but they have problems such as weak detection of small targets and poor adaptability to occluded scenes. Anchor-free methods (such as Center Net [13] and FCOS [14]) simplify the network but are sensitive to scale changes and prone to false detections in dense scenarios. Although lightweight models (such as Mobile Net-SSD [15]) are compatible with the hardware, their accuracy loss is significant in complex road conditions. The YOLO series, as a representative of single-stage detectors, is widely used in intelligent driving.

2.2. Evolution of YYOLO Series in Intelligent Driving

YOLO has become a dominant choice for real-time object detection in intelligent driving due to its efficiency. YOLOv1 [16] pioneered the grid-based single-stage approach but struggled with small target detection. Subsequent iterations addressed this: YOLOv3 [17] introduced multi-scale prediction via feature pyramids, improving small target recall; YOLOv5 [18] optimized backbone and neck structures, reducing model size while enhancing accuracy; YOLOv7 [19] integrated trainable bag-of-freebies techniques, boosting performance on traffic datasets like BDD100k. YOLOv10 [20], the latest iteration, achieves state-of-the-art speed–accuracy tradeoffs, provides higher detection accuracy while maintaining lower latency compared to existing real-time end-to-end detectors, and represents the latest advancement in real-time target detection.

2.3. Research on the Object Detection Algorithm of YOLOv10

YOLOv10 is a next-generation real-time end-to-end target detection model proposed in May 2024 by a team from Tsinghua University. Compared with previous generations of YOLO detection algorithms, YOLOv10 achieves significant computational accuracy tradeoffs across model sizes [21], but its fixed convolution kernels and conventional neck design still restrict adaptability in complex road environments. This paper makes targeted improvements. It designs multi-scale flexible convolution (MSFC), reconstructs the neck network structure to include Shallow Auxiliary Fusion (SAF) and Advanced Auxiliary Fusion (AAF), and introduces the SEAM module to enhance the focus of the detection head. While maintaining a lightweight model size (13.4 MB), it enhances the detection robustness in complex traffic scenarios.

3. Methodology

3.1. Overall Architecture of Improved YOLOv10

Due to the limited performance of current intelligent vehicles’ on-board computers, higher requirements are placed on the deployed model, i.e., minimizing model size while ensuring higher accuracy and real-time performance. Therefore, in this paper, YOLOv10s is used as the base target detection algorithm to meet the initial requirements of hardware conditions and detection performance of intelligent vehicles. However, the base model still has some room for improvement in terms of model complexity, accuracy of small and occluded targets. Aiming at the above problems. In this paper, YOLOv10s is improved by designing multi-scale flexible convolution (MSFC), reconstructing the neck network structure to include Shallow Auxiliary Fusion (SAF) and Advanced Auxiliary Fusion (AAF), and improving the detection head (SEAM) that combines multi-scale convolution and a channel adaptive attention mechanism. In Figure 1, the network structure diagram of the improved YOLOv10s is presented to better meet the requirements of intelligent vehicle target detection and provide more-accurate detection information for subsequent target tracking.

3.2. Design of Multi-Scale Flexible Convolution

Traditional convolutional layers usually use a fixed convolutional kernel size (e.g., 3 × 3 or 5 × 5). The limitations of this fixed sense field mean the convolution operation is not flexible enough to face different scales of features. When the size of the input image or the feature size difference is large, the traditional convolution needs to be stacked several times in order to capture different sizes of the features, which increases the amount of computation. In addition, in the face of complex traffic conditions, the feature size varies greatly, and the same convolutional kernel cannot adapt to features of different sizes, resulting in unsatisfactory detection results.
In response to this issue, this paper proposes MSFC, whose mathematical expression is shown in Equation (1).
y = σ ( W 1 C o n v 3 × 3 ( x 1 ) + W 2 C o n v 5 × 5 ( x 2 ) + W 0 C o n v 1 × 1 ( x ) )
where x = x 1 x 2 ; W1, W2, and W3 are learnable fusion weights that satisfy W1 + W2 + W3 = 1.
The structural schematic diagram of multi-scale flexible convolution is shown in Figure 2. A schematic diagram of the structure of multi-scale flexible convolution, which uses multiple different scales to process input features, can simultaneously capture information from different scales and fuse it with 1 × 1 convolution to fully integrate multi-scale information, reduce network stacking, and lower the computational load. The input is a feature map of size H × W with channel number C. After the segmentation operation, two feature maps with halved channel numbers are obtained and named f1 and f2, respectively. Two convolutions of different scales were performed on f1 and then concatenated together. Through a 1 × 1 convolutional layer, the features of different scales were integrated to obtain the same result as the f1 features, resulting in an output of size H × W with channel number C.
As a key feature fusion module in YOLOv10, the original C2f [22] excels in balancing feature extraction efficiency and representational capability through its dual-path structure. Although MSFC can extract multi-scale features more effectively, it lacks layer-by-layer feature fusion and hierarchical information transfer. To compensate for this, this paper chooses to combine it with the improved C2f module to leverage the advantages of both and optimize the information flow and transfer through multiple feature fusion, so as to enhance the model’s representation capability and performance.
In this paper, the lightweight Bottleneck_MSFC module is first designed to replace the Bottleneck module in the YOLOv10s network, and its structure is shown in Figure 3a. On this basis, this paper improves the C2f module by using MSFC-Bottleneck instead of the original Bottleneck, and designs the C2f-Bottleneck module, whose structure is shown in Figure 3b.

3.3. Design of Neck Network

The main goal of Feature Pyramid Network (FPN) is to solve the problem of information loss and resolution mismatch in target detection and segmentation at different scales. At first, it was proposed by Lin et al. [23] in 2016 and applied in the Faster R-CNN detector; then, based on this, Liu et al. [24] proposed PAN structure to solve the problem of information loss and incompleteness that may exist in the feature fusion process of FPN. Yang et al. [25] proposed a novel structure of Multi-Branch Auxiliary FPN (MAFPN). MAFPN achieves a richer feature fusion by introducing the two key modules of Superficial Assisted Fusion (SAF) and Advanced Assisted Fusion (AAF). By introducing these two key modules, MAFPN realizes richer feature interactions and fusion and effectively improves the detection performance of the model for targets at different scales.

3.3.1. Superficial Assisted Fusion (SAF)

The SAF module is a key component in MAFPN for preserving shallow spatial information. The “shallow features” referred to in this article mean the feature maps in neural networks that are close to the input layer. It features higher spatial resolution and richer detailed information. However, the semantic information is relatively weak and is mainly used for target positioning. “Deep features” refer to the feature maps close to the output layer, which have relatively low spatial resolution but rich semantic information and are mainly used for target classification. In the task of object detection, the shallow features provided by the shallow network contain rich edge information and detailed features, which are crucial for the precise localization of small targets. Traditional feature fusion methods often fail to preserve such shallow information during the fusion process. To overcome this problem, MAFPN designed the SAF module, incorporating shallow information as an auxiliary branch into the deep network. As shown in the schematic diagram of the SAF module structure in Figure 4, the SAF module is spliced with feature maps of different resolutions (e.g., Pn−1, Pn, Pn+1). In the feature pyramid structure, the current layer Pn is taken as the reference. Pn−1 represents the upper layer (belonging to shallow features), and Pn+1 represents the lower layer (belonging to deep features). Pn−1 and Pn+1 are subjected to downsampling and upsampling operations, respectively, to ensure the spatial alignment of the feature map. In addition, this paper uses 1 × 1 convolution to control the number of channels of shallow information, making its proportion in the concatenation operation smaller and thus not affecting subsequent learning. By using the SAF module, MAFPN can retain more shallow spatial information in deep networks, significantly improving the detection accuracy of small targets and enhancing the spatial representation of the network.

3.3.2. Advanced Assisted Fusion (AAF)

The AAF module is a key component in MAFPN used to further enhance the feature–level information interaction. In the target detection task, accurate classification relies on the coarse-grained information provided by the deep network, while accurate localization requires the detailed information provided by the shallow network. In order to realize the full fusion of shallow and deep information, MAFPN designs the AAF module to improve the model’s detection performance for medium-sized targets through the integration of multi-scale information. For specific implementation, the schematic structure of the AAF module is shown in Figure 5. AAF is achieved through multi-source feature aggregation, and the specific formula is shown in Equation (2).
P i A A F = C o n v 1 × 1 ( C o n c a t ( P i , P i 1 , P i + 1 , P i s i b l i n g ) )
where and represent downsampling and upsampling, respectively.
AAF performs multi-scale information integration in the deep layer of MAFPN, involving the aggregation of information from the shallow high-resolution layer ( P n + 1 ) the shallow low-resolution layer ( P n 1 ), the sibling shallow layer ( P n ), and the previous layer ( P n 1 ).
In this way, the final output layer is able to fuse information from four different layers at the same time, which significantly improves the model’s ability to detect medium-sized targets. The AAF also employs a 1 × 1 convolution to regulate the influence of each layer on the output results, ensuring that the model is able to capture rich information from features at different scales. With the AAF module, the MAFPN is able to better capture the multi-scale features of the target, which improves the detection accuracy of medium-sized targets and enhances the feature representation capability of the model.

3.4. Detection Head Improvements

Conventional convolutional operations and feature fusion approaches may not be able to handle multi-scale features efficiently, thus limiting the model’s ability to perceive objects of different sizes. Existing detection heads usually do not have the flexibility to weight different channels in the feature map, which affects the performance of the model in complex scenes. To compensate for these shortcomings, this paper introduces an improved scheme based on SEAM [26] (separated and enhancement attention module), which aims to enhance the feature extraction capability and robustness of the model by introducing multi-scale convolution and channel attention mechanisms.
As shown in the schematic structure of SEAM in Figure 6, the core idea behind the SEAM module is to combine multi-scale convolution and a channel adaptive attention mechanism to enhance the diversity and accuracy of feature extraction. SEAM employs multi-scale depth-separable convolution, as shown in Equation (3).
F = Concat   ( DSConv 3 × 3 ( x ) ,   DSConv 5 × 5 ( x ) ,   DSConv 7 × 7 ( x ) )
In this module, local features at different scales are extracted by first applying convolution operations to the input feature map using convolution kernels of different sizes. In this way, the model is able to capture feature information about the object from multiple scales of view, especially when dealing with targets of different sizes or with complex backgrounds, which can significantly improve the detection accuracy. Subsequently, SEAM decomposes the convolution operation into depth-wise convolution and pointwise convolution using depth-wise separable volume accumulation. While maintaining the efficiency of feature extraction, the number of parameters is reduced, thereby effectively lowering the computational overhead while ensuring performance.
In addition, SEAM introduces a channel attention mechanism that utilizes an adaptive approach to weight the importance of different channels. Specifically, the output of each channel is weighted through global average pooling and fully connected layers to adjust the contribution of each channel in the feature graph. This channel attention mechanism enables the network to pay more attention to features that are helpful for target recognition during the learning process, while suppressing the influence of noise and irrelevant features, thus improving the accuracy and robustness of target detection.

4. Experimentation and Analysis

4.1. Dataset Construction

This study adopted a systematic data construction strategy, with the aim of creating a specialized vehicle inspection dataset. The original training data was derived from the fusion of the BDD100k [27] (Berkeley DeepDrive 100k) dataset with rich traffic scene semantics and the high-resolution street view dataset Cityscapes [28]. In view of the original annotation characteristics of the BDD100k dataset, we first implemented semantic annotation space reconstruction. We then performed annotation merging operations on the five types of transportation vehicles in the original annotations, namely “car”, “bus”, “truck”, and “motorcycle”, and uniformly marked them as included in the “car” category. Through Python script automated batch processing, multi-scale vehicles in a single image can obtain unified semantic identifiers. To address the limitations of single-dataset scenarios, left-view RGB images from the Cityscapes dataset were introduced for cross-dataset fusion. The format conversion of bounding box annotations was then completed using the LabelImg tool. A total of 3218 urban scene images containing multi-angle vehicle entities were selected. After supplementation, the total number of the database expanded to 12,367, representing an increase of 35.1% compared to the original BDD100k vehicle-related data volume. The comparison of the number of datasets is shown in Figure 7, which compares the number of datasets before and after processing.
To guarantee the generalization performance of the model, the data was divided by random sampling: 20% of the data was reserved for the test set, and the remaining 80% was split into the training set and validation set by 8:2.

4.2. Model Training and Experimental Validation

In order to test the performance of the proposed algorithm, the above expanded dataset was used as the experimental dataset for training and validation, and the hardware configuration parameters, software environment parameters, and experimental parameters of the experiment are shown in Table 1 (Experimental Hardware Configuration Parameters), Table 2 (Experimental Software Environment Parameters), and Table 3 (Experimental Parameter Settings), respectively.
In order to verify the performance of the improvement algorithms proposed in this paper and evaluate the effect of different modules and the interaction between different modules, we designed eight groups of ablation experiments to verify the effectiveness of each improvement strategy using YOLOv10s as a baseline model, named A–H, and the meanings represented by the names of each serial number are shown in Table 4 (Meaning of each serial number name in the ablation experiment). “√” indicates that this module exists in this group of experiments. Among them, A represents the original YOLOv10s, and B to H are the combined schemes incorporating MSFC, MAFPN, and SEAM modules.
From the results of the ablation experiments in Table 5, it can be seen that the experimental group B, which adds the MSFC module to the baseline model, shows a large reduction in both the model computation and the number of parameters, which proves the effectiveness of the model in lightweighting; the experimental group C, which adds the MAFPN structure, shows the same reduction in the number of parameters, the size of the model is decreased by 2 MB, and the accuracies, mAP50 and mAP50-95, are both improved, indicating that the network structure can realize the functions of accuracy improvement and lightweighting at the same time. Experimental group D adds SEAM to the detection head, which causes a greater improvement in accuracy but, at the same time, brings about a significant increase in model size. Group E simultaneously introduces multi-scale flexible convolution (MSFC) and an improved neck network (MAFPN). The experimental results show that the model parameters (6.40 M) and computational load (20.7GFlOPs) are significantly lower than those of the baseline model A. The model size was compressed to 13.4 MB, and the mAP50 reached 91.7%. The reduced parameters of MSFC provide computational redundancy space for the feature fusion of MAFPN. The hierarchical fusion mechanism of MAFPN makes up for the deficiency of MSFC in deep feature transfer. Eventually, alongside a 19% reduction in parameters, mAP50 improved by 0.6% compared to the baseline. Group F added MSFC and an improved detection head (SEAM) on the basis of the baseline model, increasing mAP50 to 93.2%. However, the model parameters of 13.00 M and the computational load of 21.1GFlOPs have increased. The lightweight feature of MSFC partially offsets the increase in parameters brought by SEAM. The model size is 15.8 MB, only 1.1 MB larger than the baseline, while maintaining a high real-time performance of 139FPS. However, due to the depth-wise separable convolution and attention mechanism of the SEAM module, additional computational branches are still introduced. This led to the total number of parameters being higher than that of group B, which only contained MSFC, verifying the suggestion that improvements in accuracy come at the cost of a moderate increase in parameters. Group G, combining MAFPN and SEAM, achieved a mAP50 of 93.5%, making it the group with the highest accuracy among all combinations except group H. However, the model parameters of 12.13 M and the computational load of 21.8GFlOPs are still relatively high. The enhancement of medium target features by the AAF module complements the multi-scale convolution of SEAM. MAFPN enhances the spatial correlation of features, while SEAM strengthens the discriminability of channel dimensions. The two work together to increase mAP50-95 by 1.3% compared to the baseline. However, due to the non-introduction of the lightweight design of MSFC, the model parameters are still higher than those of group E. This indicates that although the combination of feature fusion and attention mechanisms can maximize accuracy, the computational cost needs to be balanced through lightweight modules such as MSFC.
After integrating MSFC and MAFPN, group H effectively mitigates the larger parameter count and model size introduced by SEAM and achieves a balance between accuracy and lightness, with mAP50 and mAP50-95 improved by 1.9 and 1.7 percentage points, and arithmetic and model sizes reduced by 3.5 and 2.3 percentage points, respectively.
In order to show the advantages of this paper’s algorithm in comparison with other mainstream algorithms of the same volume, seven groups of comparison experiments were conducted for this paper, and the results of the experiments are shown in Table 6. In the comparison of mAP50 and mAP50-95 evaluation metrics, the algorithm proposed in this paper shows an advantage in the seven sets of comparison experiments; its results are 1.4 and 0.5 percentage points higher than that of yolov10m in the higher level, respectively. In terms of the number of parameters, that of this paper’s algorithm is higher than the baseline model due to the improvement of the detector head on the basis of yolov10s, which has increased the parameter number by a certain amount, but it is still lower than that of yolov10m. In terms of the number of operations, that of this paper’s algorithm is higher than the baseline model because of the improvement in detection head based on yolov10s and yolov10m. In terms of computing power, that of this paper’s algorithm is 3.5GFlops lower than that of the baseline model, and the model size is the smallest except for yolov7-tiny. Although the FPS is reduced after the improvement, the real-time performance can still be maintained at a high level. The comparison of experimental results shows that the improved algorithm proposed in this paper, compared with other mainstream models of the same level, shows certain advantages in most of the evaluation indicators.
In order to visually display the results of the comparison experiment, seven sets of comparison experiments and mAP comparison diagrams were drawn. Figure 8 (Comparative experimental loss results comparison chart) shows the loss results of seven groups of comparison experiments. When the experiments are all converged, the algorithm in this paper can obtain the lowest loss results; Figure 9 (Comparative experimental mAP@0.5 results comparison chart) shows the comparison results of the evaluation index mAP@0.5; Figure 10 (Comparative experimental mAP@0.5:0.95 results comparison chart) shows the results of mAP@0.5:0.95, and the comparison charts show that the algorithm (pink curve) in this paper has an advantage over mainstream algorithms.

5. Discussion

In this study, the improved YOLOv10 model addresses the key challenges in intelligent driving target detection by introducing MSFC, MAFPN, and SEAM. Experimental results prove that these improvements effectively balance detection accuracy and model size, and the algorithm outperforms mainstream baselines such as YOLOv10s, YOLOv9s, and YOLOv8s (93.0%) regarding mAP@0.5, while reducing the model size to 13.4 MB. These results are consistent with the core assumption that optimizing multi-scale feature extraction, neck network information flow, and the adaptability of the detection head can enhance performance in complex traffic scenarios.
However, this study also has limitations. Firstly, the datasets from BDD100k and Cityscape mainly focus on urban road scenarios, which may limit the application of the results under extreme conditions (such as off-road environments, heavy fog or snow, and other bad weather), where the degradation of sensor data is more obvious. Secondly, although the model size was compressed, the FPS dropped to 128, which may not fully meet the ultra-low latency requirements of real-time autonomous driving systems, especially for high-speed decision-making tasks. Thirdly, although SEAM improves accuracy, it introduces additional computational overhead, and its efficiency in dynamic traffic with rapidly changing target densities still needs to be studied.
Future research can focus on two aspects: First, expanding the dataset to include different scenarios (such as rural roads, bad weather) will enhance the robustness of the model. Future work will also explore lightweight focus mechanisms or model quantization techniques to further enhance inference speed without sacrificing accuracy and to ensure compatibility with in-vehicle edge computing devices.

6. Conclusions

To address issues of false detections, missed detections, and low deployment efficiency in intelligent vehicle target detection within complex traffic scenarios, this study optimized YOLOv10s to balance detection accuracy and model lightweighting. Key improvements include designing multi-scale flexible convolution (MSFC) to synchronously extract multi-scale features; reducing computational costs and enhancing representation of targets across scales; reconstructing the neck network with Shallow Auxiliary Fusion (SAF) and Advanced Auxiliary Fusion (AAF) for richer feature fusion; improving small and medium target detection; and proposing SEAM, which integrates multi-scale convolution and channel adaptive attention, to enhance feature extraction robustness. Experimental results on the fused BDD100k and Cityscapes dataset show that the improved model achieves a mAP@0.5 of 93.0%, outperforming mainstream models like YOLOv10s and YOLOv9s, while compressing the model size to 13.4 MB (an 11.8% reduction from the baseline). This work provides a high-performance solution for intelligent driving, laying the groundwork for in-vehicle embedded deployment. Future research will focus on expanding datasets to include diverse scenarios to enhance model robustness, as well as exploring lightweight attention mechanisms or model quantization to further improve inference speed, ensuring compatibility with in-vehicle edge computing devices.

Author Contributions

Methodology and writing—original draft preparation, A.L.; formal analysis and investigation, X.N.; resources, M.Z.; data curation, J.C.; validation, G.X. All authors have read and agreed to the published version of the manuscript.

Funding

This project is supported by Shandong Province Science and Technology oriented Small and Medium Enterprises Enhancement Project (Grant No. 2023 TSGC0288); Jinan 2023 talent development special fund research leader studio project (Grant No. 202333067); Shandong Provincial Higher Educational Youth Innovation Science and Technology Program (Grant No. 2019KJB019); Major science and technology innovation project in Shandong Province (Grant No. 2022CXGC020706); and Ministry of industry and information technology manufacturing high quality development project, China (Grant No. 2023ZY02002).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Acknowledgments

We thank all the authors for their contributions to the writing of this article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lu, H.; Huang, J.; Zhang, M.; Shangguan, W. Night driving target detection algorithm based on YOLOv8n improvement study. J. Hunan Univ. (Nat. Sci. Ed.) 2025, 52, 59–68. [Google Scholar]
  2. Zhou, J.; Chen, X.; Gao, J.; Tang, Y.Q. Traffic participants of the infrared target under complex traffic environment lightweight detection model. J. Saf. Environ. 2025, 25, 1400–1411. [Google Scholar]
  3. Zhang, M.; Zhang, Z. Research on Vehicle Target Detection Method Based on Improved YOLOv8. Appl. Sci. 2025, 15, 5546. [Google Scholar] [CrossRef]
  4. Wen, S.; Li, L.; Ren, W. A Lightweight and Effective YOLO Model for Infrared Small Object Detection. Int. J. Pattern Recognit. Artif. Intell. 2025, 39, 32–48. [Google Scholar] [CrossRef]
  5. Zhang, H.; Yang, R.; Deng, X.; Li, F.; Zhong, M. Infrared Detection Algorithm for Defect Targets Based on Improved SSD. Laser Infrared 2024, 54, 1885–1893. [Google Scholar]
  6. Du, H.; Gu, C.; Zhang, X.; Liu, G.T.; Li, X.X.; Yang, Z.L. Vehicle Target Detection Method in Complex Weather Conditions Based on YOLOv10-vehicle Algorithm. J. Jilin Univ. (Eng. Technol. Ed.) 2025, 1–10. [Google Scholar] [CrossRef]
  7. Chen, H.; Chen, Z.; Ning, X.; Wang, S.Y.; Lu, Y.S.; Lu, H.T. Improved Detection Method for Dense Pedestrian Small Targets by YOLOv10n. Small Micro Comput. Syst. 2025, 48, 1–12. [Google Scholar]
  8. Mei, J.; Zhu, W. BGF-YOLOv10: Small Object Detection Algorithm from Unmanned Aerial Vehicle Perspective Based on Improved YOLOv10. Sensors 2024, 24, 6911. [Google Scholar] [CrossRef]
  9. Del-Coco, M.; Carcagnì, P.; Oliver, S.T.; Iskandaryan, D.; Leo, M. The Role of AI in Smart Mobility: A Comprehensive Survey. Electronics 2025, 14, 1801. [Google Scholar] [CrossRef]
  10. Wen, J.; Wang, Y.; Li, J.; Zhang, Y. Faster R-CNN Object Detection Algorithm Based on Multi-head Self-Attention Mechanism. Mod. Electron. Technol. 2024, 47, 8–16. [Google Scholar]
  11. Lai, S.; Shi, F.; Liao, J.; Zhou, X.L.; Zhao, J. Street Tree Object Detection Based on Mask R-CNN. Radio Eng. 2022, 52, 2263–2270. [Google Scholar]
  12. Liu, Y.; Wang, X.; Song, J.; Liu, X.Y.; Pu, J.P. Lightweight Aerial Target Detection Method Based on Improved RetinaNet. Fire Command Control 2024, 49, 76–82+89. [Google Scholar]
  13. Liu, X.; Huang, J.; Yang, Y.; Li, J. Improved Remote Sensing Rotating Target Detection Based on CenterNet. Remote Sens. Technol. Appl. 2023, 38, 1081–1091. [Google Scholar]
  14. Chen, J.; Cheng, M.; Xu, Z. Improved FCOS Object Detection Algorithm. Comput. Sci. 2022, 49, 467–472. [Google Scholar]
  15. Li, Y.; Xiong, X.; Xin, W.; Huang, J.; Hao, H. MobileNetV3-CenterNet: A Target Recognition Method for Avoiding Missed Detection Effectively Based on a Lightweight Network. J. Beijing Inst. Technol. 2023, 32, 82–94. [Google Scholar]
  16. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  17. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar] [CrossRef]
  18. Qu, Z.; Gao, L.Y.; Wang, S.Y.; Yin, H.N.; Yi, T.M. An improved YOLOv5 method for large objects detection with multi-scale feature cross-layer fusion network. Image Vis. Comput. 2022, 125, 104518. [Google Scholar] [CrossRef]
  19. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar]
  20. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J. Yolov10: Real-time end-to-end object detection. Adv. Neural Inf. Process. Syst. 2024, 37, 107984–108011. [Google Scholar]
  21. Yang, Y.; Zhang, X. Unmanned Aerial Vehicle Small Target Detection Algorithm Improved by YOLOv10n. Comput. Appl. Softw. 2025, 1–9. Available online: https://www.cnki.com.cn/Article/CJFDTotal-JYRJ20250713001.htm (accessed on 5 August 2005).
  22. Chen, G.; Liu, S.J.; Sun, Y.J.; Ji, G.P.; Wu, Y.F.; Zhou, T. Camouflaged Object Detection via Context-Aware Cross-Level Fusion. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 6981–6993. [Google Scholar] [CrossRef]
  23. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  24. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8759–8768. [Google Scholar]
  25. Yang, Z.; Guan, Q.; Zhao, K.; Yang, J.; Xu, X.; Long, H.; Tang, Y. Multi-branch Auxiliary Fusion YOLO with Re-parameterization Heterogeneous Convolutional for Accurate Object Detection. In Proceedings of the Pattern Recognition and Computer Vision, Urumqi, China, 18–20 October 2024; Springer Nature: Singapore, 2025; pp. 492–505. [Google Scholar]
  26. Yu, Z.; Huang, H.; Chen, W.; Su, Y.; Liu, Y.; Wang, X. YOLO-FaceV2: A scale and occlusion aware face detector. Pattern Recognit. 2024, 155, 110714. [Google Scholar] [CrossRef]
  27. Yu, F.; Chen, H.; Wang, X.; Xian, W.; Chen, Y.; Liu, F.; Madhavan, V.; Darrell, T. BDD100K: A Diverse Driving Dataset for Heterogeneous Multitask Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 2636–2645. [Google Scholar]
  28. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The Cityscapes Dataset for Semantic Urban Scene Understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar]
Figure 1. Improved YOLOv10s network structure diagram.
Figure 1. Improved YOLOv10s network structure diagram.
Sensors 25 05084 g001
Figure 2. Schematic diagram of multi-scale flexible convolution structure.
Figure 2. Schematic diagram of multi-scale flexible convolution structure.
Sensors 25 05084 g002
Figure 3. MSFC structure embedded in other modules. (a) Bottleneck_MSFC, (b) C2f_MSFC.
Figure 3. MSFC structure embedded in other modules. (a) Bottleneck_MSFC, (b) C2f_MSFC.
Sensors 25 05084 g003
Figure 4. Schematic diagram of SAF module structure.
Figure 4. Schematic diagram of SAF module structure.
Sensors 25 05084 g004
Figure 5. Schematic diagram of AAF module structure.
Figure 5. Schematic diagram of AAF module structure.
Sensors 25 05084 g005
Figure 6. Schematic diagram of SEAM module structure.
Figure 6. Schematic diagram of SEAM module structure.
Sensors 25 05084 g006
Figure 7. Comparison of dataset quantities before and after processing.
Figure 7. Comparison of dataset quantities before and after processing.
Sensors 25 05084 g007
Figure 8. Comparison of the loss results of the comparison experiment.
Figure 8. Comparison of the loss results of the comparison experiment.
Sensors 25 05084 g008
Figure 9. Comparison of the mAP@0.5 results of the comparison experiment.
Figure 9. Comparison of the mAP@0.5 results of the comparison experiment.
Sensors 25 05084 g009
Figure 10. Comparison of the mAP@0.5:0.95 results of the comparison experiment.
Figure 10. Comparison of the mAP@0.5:0.95 results of the comparison experiment.
Sensors 25 05084 g010
Table 1. Experimental hardware configuration parameters.
Table 1. Experimental hardware configuration parameters.
ParametersConfiguration
CPUIntel i5-13600KF CPU@ 3.5 GHz
RAM32G
GPUNVIDIA GeForce RTX 4070Ti SUPER
Display Memory16G
Table 2. Experimental software environment parameters.
Table 2. Experimental software environment parameters.
Virtualized EnvironmentAnaconda 3.0
LanguagePython 3.9
CUDA 12.3
Deep Learning FramePyTorch 1.12.0
OpenCV4.9.0
OSWindows 10
Table 3. Experimental parameter settings.
Table 3. Experimental parameter settings.
ParametersConfiguration
Epochs200
Batch size8
Learning rate0.0001
Image size640 × 640
Table 4. Meaning of each serial number name in the ablation experiment.
Table 4. Meaning of each serial number name in the ablation experiment.
LabelYOLOv10sMSFCMAFPNSEAM
A
B
C
D
E
F
G
H (ours)
Table 5. Results of ablation experiments.
Table 5. Results of ablation experiments.
ExperimentsmAP50 (%)mAP50-95 (%)P (%)R (%)FlOPs (G)Params (M)FPS
(f/s)
Model Size (MB)
A91.169.291.283.124.48.0427015.7
B91.469.190.882.921.07.8422716.9
C91.969.992.282.721.76.9824414.5
D93.871.293.485.622.313.2615127.3
E91.769.991.482.820.76.4019613.4
F93.270.590.182.221.113.0013915.8
G93.570.590.483.121.812.1313515.5
H (ours)9370.992.482.820.911.5612813.4
Table 6. Comparison of experimental results.
Table 6. Comparison of experimental results.
DetectorsmAP50 (%)mAP50-95 (%)P (%)R (%)Params (M)FlOPs
(G)
Model Size (MB)FPS
(f/s)
yolov10s91.169.291.283.18.0424.415.7270
yolov10m91.670.591.983.215.3163.431.9250
yolov9s91.670.391.383.17.1726.715.2222
yolov8s91.369.290.384.111.122.521.4238
yolov7-tiny90.569.091.183.16.013.012.3276
yolov5s91.169.292.881.39.1123.818.5256
ours9370.992.482.811.5620.913.4128
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, A.; Ning, X.; Zöldy, M.; Chen, J.; Xu, G. Intelligent Vehicle Target Detection Algorithm Based on Multiscale Features. Sensors 2025, 25, 5084. https://doi.org/10.3390/s25165084

AMA Style

Li A, Ning X, Zöldy M, Chen J, Xu G. Intelligent Vehicle Target Detection Algorithm Based on Multiscale Features. Sensors. 2025; 25(16):5084. https://doi.org/10.3390/s25165084

Chicago/Turabian Style

Li, Aijuan, Xiangsen Ning, Máté Zöldy, Jiaqi Chen, and Guangpeng Xu. 2025. "Intelligent Vehicle Target Detection Algorithm Based on Multiscale Features" Sensors 25, no. 16: 5084. https://doi.org/10.3390/s25165084

APA Style

Li, A., Ning, X., Zöldy, M., Chen, J., & Xu, G. (2025). Intelligent Vehicle Target Detection Algorithm Based on Multiscale Features. Sensors, 25(16), 5084. https://doi.org/10.3390/s25165084

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop