Next Article in Journal
In-Shoe Foot Temperature Patterns During Lying, Sitting and Standing Postures: Baseline Data from Healthy Individuals
Previous Article in Journal
When Lie Groups Meet Hyperspectral Images: Equivariant Manifold Network for Few-Shot HSI Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

YOLOv10-Intrusion: An Improved YOLOv10-Based Algorithm for Vehicle Area Intrusion Detection

School of Software, Nanjing University of Information Science and Technology, Nanjing 210044, China
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(7), 2118; https://doi.org/10.3390/s26072118
Submission received: 5 March 2026 / Revised: 26 March 2026 / Accepted: 27 March 2026 / Published: 29 March 2026
(This article belongs to the Section Intelligent Sensors)

Abstract

In intelligent transportation systems and urban traffic management, accurate vehicle area intrusion detection based on surveillance imagery plays a critical role in ensuring road safety and operational efficiency. However, under real-world road surveillance conditions characterized by complex backgrounds, varying illumination, occlusion, and scale variations, mainstream detection algorithms often suffer from high false detection and missed detection rates, limiting their reliability and practical deployment. To address these challenges, this paper proposes YOLOv10-Intrusion, a high-precision vehicle area intrusion detection framework based on an improved version of YOLOv10s. The proposed algorithm incorporates Omni-Dimensional Dynamic Convolution (ODConv) and a custom-designed RCS_M module to enhance feature extraction and fine-grained recognition capability. In addition, a Bidirectional Feature Pyramid Network (BiFPN) is employed to optimize multi-scale feature fusion at the neck level. These improvements collectively reduce false detections and missed detections while improving model recall and mean Average Precision (mAP). Furthermore, the Wise-IoU (WIoU) loss function replaces the original Complete IoU (CIoU) loss to accelerate convergence and stabilize bounding box regression under complex surveillance conditions. A dedicated vehicle area intrusion dataset is constructed from real-world road surveillance footage, covering five vehicle categories across diverse road environments and lighting conditions. Experimental results demonstrate that, compared with the baseline YOLOv10s, YOLOv10-Intrusion achieves improvements of 1.5, 3.3, 3.6, and 2.8 percentage points in Precision, Recall, mAP@0.5, and mAP@0.5:0.95, respectively, and outperforms other mainstream detection algorithms in vehicle area intrusion detection tasks.

1. Introduction

With the rapid advancement of deep learning technology, intelligent perception systems have found widespread applications across transportation management, public safety, and infrastructure monitoring [1,2,3]. In urban road management, law enforcement agencies face mounting pressure to replace manual patrols with automated, real-time surveillance solutions capable of detecting traffic violations accurately and efficiently. Vehicle area intrusion—wherein vehicles illegally enter restricted zones such as pedestrian walkways, bicycle lanes, or construction site boundaries—poses significant risks to public safety and presents a persistent challenge in modern road governance [4,5].
Object detection algorithms based on deep learning have demonstrated powerful feature representation and generalization capabilities, making them the dominant paradigm in intelligent traffic surveillance [3]. Among these, the YOLO (You Only Look Once) series of single-stage detectors has achieved widespread adoption due to its favorable balance of speed and accuracy, enabling real-time detection on edge devices [2,6]. Various YOLO-based methods have been proposed for area intrusion detection across different application domains. For instance, Hu et al. [4] proposed MSIA-YOLOv8 for railway obstacle intrusion detection, achieving 97.3% mAP with enhanced robustness in adverse lighting conditions. Zhang et al. [5] integrated YOLOv5-based detection with track area extraction to realize level-based warning for railway obstacles. Dong et al. [7] introduced FPE-YOLO, a YOLOv8-based system achieving 99.14% mAP@0.5 for third-party pipeline intrusion detection under challenging small-object conditions. However, these works primarily address domain-specific intrusion scenarios and rarely tackle the complexity of multi-class vehicle identification under dense traffic, occlusion, and varying illumination conditions.
In the vehicle detection domain, deep learning-based classification and localization methods have been extensively studied. Alnfiai [8] proposed a Lightning Search Algorithm with Deep Transfer Learning for vehicle classification in intelligent transportation systems, demonstrating the importance of accurate multi-class vehicle recognition. Liu et al. [1] developed TIE-LR, a multi-module framework for vehicle identification and tracking from low-resolution CCTV footage using self-supervised learning and distilled YOLO models. Despite significant progress, existing vehicle detection algorithms frequently exhibit high false detection and missed detection rates when applied to practical road surveillance scenarios, particularly when facing visually similar vehicle categories (e.g., trucks and muck trucks), dense overlapping targets, and degraded imaging conditions such as motion blur, night-time low light, or transmission artifacts [6,9].
Multi-domain intrusion detection benchmarks have further revealed the limitations of current approaches. Han et al. [10] introduced the MMID-Bench benchmark, systematically evaluating multi-domain, multi-category intrusion detection algorithms under foggy, rainy, and night-time conditions, highlighting the performance degradation of conventional detectors in adverse environments. Similarly, Huang et al. [11] demonstrated that spatio-temporal feature fusion is critical for accurate intrusion detection in unmanned aerial vehicle (UAV) surveillance scenarios. These findings underscore the need for detection algorithms specifically designed for the practical demands of vehicle area intrusion enforcement, including high precision, strong recall, and robust performance across diverse real-world conditions.
The surveillance scenes monitored by the platform where the proposed algorithm is deployed include multiple corner sections leading to construction sites. In such sections, large vehicles frequently enter the camera view at close range and then gradually move away, resulting in considerable variations in detection scales. Large vehicles appearing at close range to the camera are often incorrectly detected as multiple distinct vehicles. Furthermore, the target road sections covered by the algorithm mainly consist of highways with heavy traffic and field paths with complex backgrounds. In the surveillance view, vehicles in high traffic sections easily occlude each other. At night, strong vehicle headlights cause overexposure, and under rainy, snowy, or hazy weather conditions, images suffer from blurring and color distortion. These issues tend to destroy the integrity of the visual features of vehicle targets, making it difficult for the backbone network of the original YOLO algorithm to extract sufficiently effective and discriminative features.
To address these challenges, this paper proposes YOLOv10-Intrusion, an improved object detection algorithm based on YOLOv10s, tailored for vehicle area intrusion detection in real-world road surveillance environments. The main contributions of this paper are as follows:
  • The RCS_M module, a self-designed backbone component, replaces the C2f structure by combining channel-shuffled reparameterized convolution (RCS) with a one-shot aggregation (OSA) strategy and substituting the RepVGG block with MobileOne Block. This design reduces training parameters while enhancing cross-channel information interaction and local feature representation. An additional RCS layer is appended after the second RCS stage to compensate for the accuracy reduction associated with the lighter parameterization, achieving improved precision and mean average precision on the vehicle intrusion dataset.
  • Omni-Dimensional Dynamic Convolution (ODConv) is introduced into the neck C2f module (C2f_OD) to replace standard static convolution. By generating four complementary attention weights across spatial, input-channel, output-channel, and kernel dimensions simultaneously, ODConv adaptively highlights discriminative features of visually similar vehicle categories (e.g., trucks versus muck trucks), improving fine-grained recognition capability and suppressing background interference.
  • BiFPN (Bidirectional Feature Pyramid Network) is incorporated into the neck to replace the original Path Aggregation Network (PAN)/Feature Pyramid Network (FPN) fusion path. In conjunction, WIoU (Wise-IoU) replaces the Complete IoU (CIoU) loss function in the detection head. BiFPN constructs bidirectional feature flow with learnable fusion weights, adaptively balancing shallow detail and deep semantic features to improve recall in dense and occluded traffic scenes. WIoU dynamically assigns gradient gains based on anchor quality, concentrating regression optimization on normal-quality anchors to accelerate convergence and alleviate the impact of class imbalance.
  • A high-quality vehicle area intrusion dataset is constructed from real-world road surveillance footage, covering five vehicle categories across six monitored sections with diverse angles, lighting conditions, and time periods. Field validation across all six monitored sections under both daytime and nighttime conditions confirms the practical effectiveness of the proposed algorithm.
The remainder of this paper is organized as follows. Section 2 reviews related work on vehicle detection, area intrusion detection, and YOLO-based architectural improvements. Section 3 presents the proposed YOLOv10-Intrusion algorithm, covering the technical workflow and all architectural module designs. Section 4 describes the experimental setup and reports results, including ablation studies, comparative experiments, Generalization Experiments, visualization analysis, and field validation. Finally, Section 5 concludes the paper and outlines future research directions.

2. Related Work

2.1. Vehicle Detection in Traffic Surveillance

Vehicle detection and classification are fundamental tasks in intelligent transportation systems, with broad applications in traffic monitoring, law enforcement, and autonomous driving [3,12]. Early YOLO-based approaches, such as the O-YOLO-v2 model proposed by Han et al. [13] for tiny vehicle detection, demonstrated that multi-scale feature fusion with residual modules significantly improves detection accuracy for small and distant vehicles. Alnfiai [8] proposed a Lightning Search Algorithm with Deep Transfer Learning (LSADTL-VCITS) combining YOLOv5 with Capsule Networks for multi-class vehicle classification, while Liu et al. [1] developed the TIE-LR framework for vehicle identification from low-resolution surveillance footage using self-supervised SimCLR representations, achieving a 17-fold speed improvement over conventional YOLOv8x-based methods. Chaman et al. [14] conducted a comprehensive benchmarking study of YOLOv8 through YOLOv12 in Advanced Driver Assistance Systems (ADAS) scenarios, demonstrating clear generational improvements with YOLOv12 achieving the highest accuracy (mAP@50–95 = 82.2%) across traffic sign recognition, pedestrian detection, and vehicle detection tasks.
A persistent challenge in vehicle detection is maintaining robust performance under degraded image conditions and complex environmental factors. Wang et al. [15] proposed MC-YOLO, integrating MobileNetV2 and CBAM attention for nighttime vehicle detection, achieving 92.75% precision on the BDD100K dataset. Luo et al. [16] introduced IR-YOLO for infrared pedestrian-vehicle detection, employing a dedicated small-target detection head and Focal Generalized IoU (GIoU) loss to achieve 90.1% mAP under low-visibility conditions. Kang et al. [17] proposed YOLO-FA with type-1 fuzzy attention to reduce detection uncertainty in rainy and nighttime scenarios, achieving 8.1% AP50 improvement on the UA-DETRAC dataset. Duan et al. [18] designed DD-YOLO with dual-channel feature extraction and a hybrid pooling pyramid module for blurred vehicle detection, improving mAP@0.5 by 2.7% on KITTI. Du et al. [19] proposed MLE-YOLO based on YOLOv11, combining a Multi-Stage Partial Transformer Module (M-SPTM) with a mixed aggregation network for vehicle and pedestrian detection in adverse weather, achieving 3.0% mAP improvement with 15.8% fewer parameters. Wang et al. [20] proposed Hybrid-YOLO integrating Mamba-based state space modeling and transformer-driven global attention with multi-scale feature fusion, achieving 90.11% mAP@0.5 at 66.3 FPS on the KITTI benchmark.
Dense traffic scenes with frequent occlusion and visually similar vehicles pose additional detection challenges. Liu et al. [21] addressed long-distance, truncation, and occlusion challenges through a Visual Attention Module (VAM) and Feature Reconstruction Module (FRM) inspired by human visual perception, achieving state-of-the-art performance on KITTI. Li et al. [22] proposed YOLO-CCS with coordinate attention and C2f modules for vehicle detection, improving mAP50 by 3.2% over YOLOv5s. Pan et al. [23] proposed LVD-YOLO with EfficientNetv2 backbone and bidirectional feature pyramid for lightweight vehicle detection, reducing FLOPs by 64.6% while maintaining competitive accuracy. Alahdal et al. [24] evaluated multiple YOLO versions in autonomous vehicle environments, highlighting the importance of early detection for diverse road objects. Tang et al. [2] proposed YOLO-Fusion integrating infrared and visible-light images through FusionAttention and Dynamic Fusion modules for multimodal intelligent transportation scenarios. While these works demonstrate the breadth of YOLO-based vehicle detection research, they do not specifically address the multi-class vehicle area intrusion detection scenario under dense, occluded road surveillance conditions with enforcement requirements for high precision and recall.

2.2. Area Intrusion Detection

Area intrusion detection aims to identify objects or events that violate spatial boundaries, with applications ranging from railway safety and pipeline monitoring to wildlife protection and road law enforcement. Hu et al. [4] proposed MSIA-YOLOv8, an improved YOLOv8-based detector for railway obstacle intrusion, incorporating multi-scale feature extraction and Frequency Domain Aggregation and Enhancement (FDAE) modules to improve detection robustness under low-light and adverse weather conditions, achieving 97.3% mAP at 137 FPS on a railway dataset. Zhang et al. [5] developed an early warning system integrating YOLOv5-based detection with track area extraction and obstacle severity classification, providing tiered warning outputs for railway intrusion scenarios. Tang et al. [9] introduced YOLO-RCNN, a hybrid framework combining YOLO foreground detection and region of interest (ROI)-aligned RCNN classification for real-time railroad crossing surveillance, achieving 54.15% mAP on a custom railroad crossing dataset and demonstrating deployment feasibility on edge hardware via TensorRT optimization.
For pipeline and infrastructure protection, Dong et al. [7] proposed FPE-YOLO, an enhanced YOLOv8 system for real-time small-object intrusion detection in oil and gas pipeline rights-of-way, employing bio-inspired foveal attention, cross-level fusion, and single-level enhancement modules to achieve 99.14% mAP@0.5 with only a 19.5% increase in computational complexity. Hu et al. [25] proposed a convolutional neural network-based infrared target intrusion detection algorithm combining static target mode analysis (LBP texture) with dynamic multi-frame correlation detection, achieving superior performance in military and intelligent warning applications. Sodhro et al. [6] conducted a systematic real-time evaluation of YOLOv5 and YOLOv8 for human intrusion detection across diverse environmental conditions (luminance variation, indoor/outdoor, simulated weather), revealing that YOLOv8 achieves 99.1% outdoor confidence while adverse weather significantly reduces its effectiveness. Balakrishnan et al. [26] proposed Deep-Track, a field surveillance system combining VGG16-based animal classification with Deep-SORT tracking for real-time wildlife intrusion detection in human-fringe areas, achieving 92.19% accuracy.
Multi-domain benchmarking of intrusion detection algorithms has highlighted performance degradation under challenging environmental conditions. Han et al. [10] introduced MMID-Bench, a comprehensive benchmark for multi-domain multi-category intrusion detection, proposing the MMID-YOLO framework with unsupervised domain adaptation (I-DANN) and diffusion model-based augmentation to improve generalization across normal, foggy, rainy, and night-time city intrusion scenarios. Huang et al. [11] developed a Fused-ConvLSTM-based railway intrusion detection method for UAV surveillance, effectively capturing spatio-temporal features for detecting unknown intruder types in complex aerial scenes with multiple railway regions. Despite this progress, most existing intrusion detection methods focus on specific object categories (persons, animals, or obstacles) and do not address the multi-class vehicle area intrusion scenario under complex road surveillance conditions.

2.3. YOLO Architecture Improvements for Detection Tasks

The YOLO series has undergone continuous architectural evolution to improve detection accuracy and efficiency for diverse application scenarios. In backbone and feature extraction design, channel shuffling and reparameterized convolution have emerged as effective strategies for strengthening cross-channel feature interaction. Lu et al. [27] proposed IMV-YOLO incorporating channel shuffle and one-shot aggregation (OSA) strategies for infrared multi-angle vehicle detection under adverse weather, achieving improvements of 4.5% and 7.7% in mAP50 and mAP50-95 over YOLOv11. Li et al. [28] proposed MBS-YOLO with a C2f-Pu module and Multi-Branch Feature Pyramid Network (MB-FPN) for anti-drone detection, reducing parameters by 21.2% while improving mAP@0.5 by 4.5% on Det-Fly. Lu et al. [29] improved YOLOv5 for underground sewer defect detection by incorporating selective kernel attention and a bidirectional cascade feature fusion structure, achieving 4.5% mAP improvement at 69.9 FPS. Fan et al. [30] proposed LUD-YOLO for UAV small object detection, introducing a multi-scale feature fusion mode with upsampling in FPN and dynamic sparse attention mechanisms, demonstrating superior performance on VisDrone2019 and UAVDT datasets. Zhang et al. [31] proposed NOC-YOLO based on YOLOv10n for small-target vehicle detection in aerial infrared images, achieving 79.5% mAP50 on the DroneVehicle dataset by incorporating attention mechanisms and multi-scale feature fusion.
In feature pyramid design, bidirectional and adaptive multi-scale fusion strategies have demonstrated consistent improvements in handling targets of varying scales and densities. Song et al. [32] proposed MEB-YOLO combining BiFPN with ECA attention for complex traffic road vehicle detection on the UA-DETRAC benchmark, demonstrating that bidirectional feature fusion significantly improves multi-scale vehicle detection performance. Liu et al. [33] proposed PV-YOLO replacing PANet with BiFPN in the neck for pedestrian and vehicle detection on BDD100K and KITTI, achieving higher detection accuracy than YOLOv8n with lower computational complexity. Nan et al. [34] proposed MS-YOLO-DLKA combining multi-scale feature extraction (MS-Block) and large convolutional kernels (D-LKA) for LiDAR-camera joint railway obstacle detection, achieving 91% mAP on a custom railway dataset. Chaurasia and Patro [35] introduced YOLO-CSL with a novel Channel and Spatial Attention (CSA) module for rotational object detection in satellite and aerial imagery.
For loss function optimization, improved IoU-based losses have shown consistent benefits in bounding box regression stability, particularly for datasets with class imbalance and dense, occluded targets. Zhou et al. [36] proposed MP-YOLO for dense vehicle detection on the DAIR-V2X dataset, replacing CIoU with WIoU to effectively address the high-overlap characteristics of road targets and improve AP50 by 4.7%. Tahir et al. [37] incorporated WIoU v3 into PV3M-YOLO for UAV-based pedestrian and vehicle detection on VisDrone2019, demonstrating that WIoU’s dynamic gradient allocation reduces the influence of low-quality anchors and improves mAP@0.5 by 3.9% over the baseline. These developments in feature extraction, multi-scale fusion, and loss function design collectively motivate and provide the methodological foundation for the architectural improvements proposed in YOLOv10-Intrusion, which are specifically designed for the demands of vehicle area intrusion detection in real-world road surveillance environments.

3. Materials and Methods

3.1. Technical Workflow

The overall technical workflow of the proposed system is illustrated in Figure 1. The system consists of two main components: the vehicle area intrusion dataset construction and the YOLOv10-Intrusion detection algorithm. Live video streams are acquired from surveillance cameras via the RTSP (Real-Time Streaming Protocol), and OpenCV is used to extract frames at a configured sampling rate and to delineate detection zones within each frame. The YOLOv10-Intrusion algorithm is then applied to classify and localize vehicles within the designated area. The detection model underwent multiple rounds of field testing before operational deployment.

3.2. YOLOv10 Architecture Overview

The YOLO series represents the dominant paradigm in single-stage object detection, offering superior detection accuracy and inference speed for real-time applications. To eliminate the inference latency introduced by Non-Maximum Suppression (NMS) post-processing, YOLOv10 employs a consistent dual-assignment training strategy that achieves NMS-free inference while simultaneously optimizing model efficiency and detection accuracy. YOLOv10 further incorporates a partial self-attention module that substantially enhances feature extraction and multi-scale representation capability, making it particularly suitable as a lightweight baseline for resource-constrained real-time detection tasks.
YOLOv11 introduced the C3k2 module to replace the C2f module in the backbone and neck, decomposing the original large convolution into two smaller convolutions to reduce processing time. A Cross-Stage Partial Spatial Attention (C2PSA) module was appended after the Spatial Pyramid Pooling Fast (SPPF) module to enhance spatial attention and improve detection accuracy in regions of interest. YOLOv12 further extends this by introducing area attention, which partitions feature maps into multiple non-overlapping regions to maintain large receptive fields while avoiding the computational overhead of complex window partitioning operations, effectively integrating attention mechanisms into the YOLO framework. However, the relatively constrained parameter capacity of YOLOv12 limits its representational power for complex multi-scale feature patterns.
Considering the lightweight design philosophy, excellent detection performance, and NMS-free inference pipeline of YOLOv10s, it is selected as the baseline model for vehicle area intrusion detection in this paper. Recent works have further validated YOLOv10 as an effective baseline for vehicle detection; for instance, Zhang et al. [31] demonstrated that YOLOv10n achieves competitive performance for aerial infrared small-target vehicle detection, confirming the model’s adaptability to challenging real-world scenarios.

3.3. YOLOv10-Intrusion Architecture

The proposed YOLOv10-Intrusion algorithm adopts YOLOv10s as the baseline. As illustrated in Figure 2, the backbone C2f modules are replaced with the proposed RCS_M module, while standard convolution in the neck C2f module is replaced with ODConv to form the C2f_OD module. BiFPN is introduced in the neck to optimize multi-scale feature fusion, and the CIoU loss function in the detection head is replaced with WIoU. Modified modules are highlighted in dark color in the figure.
The RCS_M module, C2f_OD module, BiFPN feature fusion structure, and WIoU loss function used in YOLOv10s form a complete collaborative mechanism that covers feature extraction, feature enhancement, multi-scale feature fusion, and bounding box regression, jointly solving the challenging problems of large target scale changes, insufficient feature extraction capabilities, and class imbalance in vehicle area intrusion detection. The RCS_M module in the backbone enhances local features and facilitates cross-channel information interaction through structural reparameterization, channel shuffle, and one-shot aggregation (OSA), while introducing only a small increase in parameters and computational cost. Compared with the original C2f module, RCS_M provides more complete and discriminative features for subsequent modules, thereby solving the difficulty of feature extraction in complex environments. In the neck network, the C2f_OD module integrated with ODConv adaptively highlights critical vehicle structures, including contours and textures, via multi-dimensional dynamic attention; suppresses background interference; and makes the basic features output by RCS_M more distinguishable. BiFPN enables bidirectional delivery and weighted fusion of deep semantic information and shallow detailed information, allowing the fine-grained features extracted by the preceding two modules to be fully utilized across all scales. This makes the model more adaptable to large variations in target scale, ranging from close-range large trucks to distant small cars, and reduces false recognition. Finally, the WIoU loss dynamically allocates gradients according to anchor quality, reduces interference from low-quality anchors, accelerates model convergence, and mitigates class imbalance. In this way, the high-quality features refined by the backbone and neck can be more stably converted into accurate detection results. When deployed on the intelligent enforcement platform, the improved model effectively detects overlapping vehicles and significantly reduces false recognition of close-range large vehicles compared with the original YOLOv10s. Thus, it is highly suitable for the target road sections monitored by the platform.

3.3.1. RCS_M Module

In road surveillance scenarios with high traffic density, mutual occlusion between vehicles frequently results in incomplete or partially visible feature regions. Furthermore, large vehicles such as trucks and muck trucks share highly similar appearance characteristics, leading to frequent inter-class misclassification. To address these problems, the backbone structure is improved by introducing the self-designed RCS_M module to replace the C2f structure, strengthening cross-channel information exchange and local feature representation capability.
The RCS (Reparameterized Channel Shuffle) module maintains a multi-branch structure during training, consisting of a 1 × 1 dimensionality-reduction convolution, a 3 × 3 convolution, and an identity connection. Channel shuffling has been shown to be an effective strategy for strengthening cross-channel feature interaction in vehicle detection backbones [27]; combined with the multi-branch structure, it significantly enhances cross-channel information flow during training. The training-time output of the RCS module is expressed as:
y R C S = W 1 × 1 x + W 3 × 3 x + x
where W 1 × 1 and W 3 × 3 denote the 1 × 1 and 3 × 3 convolutional kernels, respectively, ∗ denotes the convolution operation, and x is the input feature map. During inference, structural reparameterization merges all branches into a single equivalent 3 × 3 convolution:
y R C S = W e q x , W e q = W 1 × 1 + W 3 × 3 + W i d
where W 1 × 1 is the zero-padded 3 × 3 equivalent of the 1 × 1 kernel and W i d is the identity mapping expressed as a 3 × 3 kernel, reducing computational overhead while preserving multi-branch feature learning capability and improving feature representation in dense and occluded regions. The structure of the RCS module is shown in Figure 3.
The principle of Channel Shuffle is to unfold the input feature map into a matrix, then transpose the resulting matrix, and finally flatten the transposed channels to complete the final channel shuffle. The specific operation is shown in Figure 4. Each piece of data is shuffled through channels and then aggregated using Concat, followed by a RepVGG operation.
OSA (One-Shot Aggregation) is a complementary module that aggregates all intermediate features through a single global operation at the module output, maintaining cross-channel information exchange while improving network efficiency relative to dense connectivity strategies. Input features from the preceding layer are first dimensionality-reduced by the RepVGG block and then routed through two parallel paths: a direct skip connection and a sequential processing path through two stacked RCS modules that extract hierarchical features at different depths. The outputs of both paths are merged via channel shuffling, producing representations with stronger cross-scale interaction capability.
Although the original RCS-OSA module yields remarkable accuracy gains for the model, it suffers from excessive parameters. In addition, the features produced by two RCS layers still result in misidentification in practical deployment. Inspired by the lightweight MobileOne network, we revise the RCS-OSA module by replacing its RepVGG block with the MobileOne Block. In the RCS_M module, input features first go through channel dimension reduction via the MobileOne Block and are then split into two branches: a shortcut branch and an RCS processing branch. The RCS branch adopts three consecutive RCS operations to mine deeper-level features, laying a richer feature foundation for the subsequent channel shuffle. The features processed by the RCS modules are then fused with the shortcut features from the MobileOne Block via channel shuffle, which breaks information barriers between channels, enables free feature flow across different groups, and allows the model to capture more comprehensive representations. Finally, the OSA (One-Shot Aggregation) strategy is adopted to globally aggregate the shuffled features in one step, outputting enhanced multi-scale vehicle features. Compared with the original RCS-OSA module, RCS_M utilizes the MobileOne Block instead of the RepVGG block for channel downsampling. Its depthwise separable convolutions reduce both parameters and computational cost. Meanwhile, an additional third RCS layer is appended to further mine deep features and compensate for the accuracy loss caused by the MobileOne Block. This operation introduces no significant extra computation or parameter overhead, thus preserving the overall lightweight advantage of the model.
When integrated into the backbone network, the improved RCS_M module reduces the parameters by 1.7 M and improves mAP@0.5 by 0.8 percentage points relative to the original RCS-OSA module. In practical deployment, the RCS_M module correctly identifies vehicles that were misclassified by the original algorithm, leading to a significant reduction in large-vehicle misidentification in surveillance scenarios. The improvement from RCS-OSA to RCS_M is illustrated in Figure 5.

3.3.2. ODConv Module (C2f_OD)

Road surveillance environments involve multiple vehicle categories with complex and fine-grained appearance features. The fixed weights of standard convolutions are incapable of dynamically adjusting their responses to varying feature distributions, limiting the model’s ability to distinguish visually similar vehicle categories. To address this limitation, Omni-Dimensional Dynamic Convolution (ODConv) is introduced to replace standard convolution in the neck C2f module, optimizing the feature extraction process.
ODConv generates four complementary attention weights across spatial, input-channel, output-channel, and kernel dimensions simultaneously. This multi-dimensional dynamic adjustment enables the model to highlight discriminative features such as vehicle contour, texture channels, and spatial regions while suppressing irrelevant background features. An overview of the ODConv module is illustrated in Figure 6. The ODConv formulation is given in Equations (3) and (4):
y = i = 1 n β i x
β i = α s i α c i α f i α w i W i
where x denotes the input feature; y denotes the output feature; β i represents the dynamic weight of the i-th convolutional kernel; ⊙ denotes element-wise multiplication along different dimensions of the kernel space; W i is the i-th convolutional kernel; and α s i , α c i , α f i , and α w i represent attention weights along the spatial, input-channel, output-channel, and kernel dimensions, respectively.
In the attention generation process, ODConv first applies Global Average Pooling (GAP) to each input channel to produce a compact channel-wise descriptor g R C :
g c = 1 H × W h = 1 H w = 1 W x c ( h , w )
where H and W are the spatial dimensions of the input feature and x c ( h , w ) is the feature value at position ( h , w ) in channel c. A fully connected (FC) layer subsequently reduces dimensionality and performs global feature combination, followed by ReLU activation. The spatial attention weight α s i is then generated via Sigmoid activation, while the input-channel, output-channel, and kernel attention weights are generated via Softmax:
α s i = σ f s ( g )
α c i = softmax f c ( g ) , α f i = softmax f f ( g ) , α w i = softmax f w ( g )
where f s , f c , f f , f w are the corresponding FC projection layers and σ ( · ) denotes the Sigmoid function. The convolutional kernels W i are then combined with their respective attention weights through element-wise multiplication, and the weighted results are summed to produce the dynamically adjusted output feature y.
The four complementary attention weights α s i , α c i , α f i , and α w i enable ODConv to differentially emphasize vehicle-discriminative features: highlighting vehicle contour and texture channels for compact cars, adapting kernels to distinctive shape characteristics of large vehicles (muck trucks versus trucks), focusing attention on vehicle-surrounding regions, and suppressing road background interference, thereby improving detection accuracy for partially occluded vehicles. By replacing standard convolution in the C2f module with ODConv, the improved C2f_OD module (shown in Figure 7) enables the network to adaptively focus on key vehicle regions during feature extraction.

3.3.3. BiFPN Feature Pyramid

Traditional YOLO models employ PAN/FPN structures that perform feature transfer through simple upsampling, downsampling, and element-wise addition, without adaptive adjustment of feature scale contributions. In road surveillance environments with dense and mutually occluded vehicles, this design is prone to losing critical information across scales. To resolve this issue, the Bidirectional Feature Pyramid Network (BiFPN) [32,33] is introduced to optimize the neck network’s feature fusion process.
BiFPN extends conventional multi-scale feature fusion by breaking unidirectional feature flow constraints and constructing bidirectional feature fusion paths. Through skip connections, high-level semantic information is propagated downward to enrich shallow feature maps with contextual awareness, while low-level spatial detail features are propagated upward to reinforce the localization capability of higher-level representations. BiFPN further introduces learnable fusion weights to adaptively calibrate the contribution of each feature scale. The weighted top-down fusion at intermediate node P l t d is formulated as:
P l t d = w 1 · P l + w 2 · Resize ( P l + 1 ) w 1 + w 2 + ϵ
The output node P l o u t is subsequently obtained by fusing the original node, the top-down intermediate, and the upsampled lower-level output:
P l o u t = w 1 · P l + w 2 · P l t d + w 3 · Resize ( P l 1 o u t ) w 1 + w 2 + w 3 + ϵ
where w i and w i are learnable non-negative fusion weights, P l denotes the feature map at pyramid level l, and ϵ = 10 4 is a small constant for numerical stability. Compared with FPN and PAN structures, the BiFPN-augmented neck significantly improves model recall and mean average precision and reduces the missed detection rate for partially obscured targets in real-world road environments. A structural comparison of FPN, PAN, and BiFPN is shown in Figure 8.

3.3.4. WIoU Loss Function

IoU (Intersection over Union) is a standard metric for evaluating the localization accuracy of object detection predictions. The basic IoU loss function L IoU is defined as:
IoU = P G P G
L IoU = 1 IoU
where P denotes the predicted bounding box and G denotes the ground-truth bounding box. The IoU loss is zero when boxes perfectly overlap and one when they are completely disjoint. However, the original IoU loss suffers from two key limitations: it provides no gradient information when predicted and ground-truth boxes do not overlap, and identical IoU values may correspond to different relative box configurations, limiting its discriminative power for precise localization.
To overcome these shortcomings, CIoU extends IoU by incorporating center-point distance and aspect ratio consistency penalties:
L CIoU = 1 IoU + ρ 2 ( b , b g ) c 2 + α v
where ρ 2 ( b , b g ) is the squared Euclidean distance between the centers of the predicted box b and ground-truth box b g , c is the diagonal length of the minimum enclosing box, and the aspect ratio consistency term v and its trade-off coefficient α are:
v = 4 π 2 arctan w g h g arctan w h 2 , α = v ( 1 IoU ) + v
where w, h and w g , h g are the width and height of the predicted and ground-truth boxes, respectively. Although CIoU provides richer geometric supervision than plain IoU, during field testing on the monitored road sections, the original CIoU loss was found to produce suboptimal gradient weight assignments, leading to bounding box drift and redundant box generation that degraded localization precision. The Wise-IoU (WIoU) loss function [36,37] dynamically adjusts loss weights and gradient allocation, assigning higher weights to overlapping or occluded vehicle samples to strengthen bounding box regression optimization for these challenging cases. Its gradient gain allocation strategy also alleviates gradient explosion and reduces the generation of invalid gradients without incurring additional computational cost. The WIoUv1 formulation is given in Equations (14) and (15):
L WIoUv 1 = R WIoU · L IoU
R WIoU = exp x x g 2 + y y g 2 W s 2 + H s 2 *
where W s and H s are the width and height of the minimum enclosing box, the superscript ∗ denotes detachment from the computation graph to prevent R WIoU from generating gradients that impede convergence, ( x , y ) is the center coordinate of the predicted box, and ( x g , y g ) is the center coordinate of the ground-truth box. WIoUv1 constructs distance attention through R WIoU and L IoU , enhancing the importance of average-quality predicted boxes while weakening geometric penalties when predicted and ground-truth boxes are highly overlapping.
WIoU employs an outlier degree variable β to measure anchor box quality. Smaller outlier degree indicates higher anchor quality. To concentrate bounding box regression on normal-quality anchors, WIoU adopts a gradient gain assignment strategy that assigns smaller gains to both low and high outlier degree anchors, reducing the negative influence of low-quality samples on model training. The final WIoU loss function is:
L WIoU = β δ · α β δ · R WIoU · L IoU
where δ is a hyperparameter used to dynamically regulate the bounding box regression optimization direction and gradient distribution in the detection task. In the initial phase of dataset construction, the vehicle area intrusion dataset suffered from significant class imbalance, with muck trucks and vans substantially underrepresented compared to other categories. WIoU’s robustness to low-quality samples, combined with its dynamic gradient gain allocation mechanism, partially alleviates this class imbalance problem.

4. Experiments

4.1. Dataset

The construction of the vehicle area intrusion dataset proceeded through three stages of iterative collection and refinement. In the initial stage, 3000 passenger car images sourced from public datasets and 500 muck truck images captured at adjacent construction sites constituted the preliminary dataset, primarily used to evaluate the algorithm’s baseline detection capability for small and large vehicle categories. In the subsequent collection phase, an additional 2000 images covering van, truck, and tricycle categories were gathered from roads and pedestrian areas. To address class imbalance caused by the disproportionate number of car images, approximately 60% of low-quality (duplicate, blurred) car images were filtered, and the WIoU loss function was introduced concurrently. Data augmentation including random cropping was applied to underrepresented van and tricycle samples. In the final refinement stage, category-specific supplementation was performed for truck and muck truck data, including close-range and long-range examples to enrich vehicle detail and small-target representation. After the online platform began operation, background images without any target vehicles and images with transmission artifacts (i.e., frame corruption caused by signal interference) were added as negative samples to suppress false alerts.
The complete dataset encompasses five vehicle categories: Car, Van, Muck Car, Truck, and Tricycle, covering multiple road environments including urban roads, pedestrian walkways, rural paths, and construction site entrances. A total of 6530 images were collected, of which 80% were captured by on-site photography and surveillance cameras, and 20% were selected from online public datasets. The dataset encompasses vehicles captured across diverse viewing angles and under varying illumination conditions across the full 24-h daily cycle. Data augmentation including Mosaic, random cropping, contrast adjustment, and geometric transformation was applied to selected subsets. Annotation was performed using Roboflow, yielding a total of 12,460 labeled vehicle instances. The dataset is partitioned into 5674 training images, 756 validation images, and 100 test images supplemented by several video segments for qualitative evaluation. As the proposed algorithm is specifically designed for real-time vehicle area intrusion detection in continuous surveillance video streams, the 100 images in the test set are only used for preliminary quantitative evaluation. The primary assessment of model detection performance is conducted on real surveillance video streams collected from the target road sections under monitoring. Field validation is performed using over 24 h of real-world monitoring footage, which is highly consistent with the actual deployment scenario of the model. Category-level instance counts are provided in Table 1. Representative sample images from the dataset are shown in Figure 9.
Compared with generic vehicle datasets, the proposed dataset possesses several distinctive characteristics. The collection encompasses near-full-angle vehicle images together with a diverse set of hard examples including mutually occluded vehicles and motion-blurred high-speed vehicles. The dataset spans six monitored road sections ranging from rural grass-bordered paths to urban roads, with 24-h temporal coverage that provides complex background variation for improving cross-scenario generalization. Moreover, the dataset undergoes continuous iterative refinement driven by real-world deployment observations, incorporating newly discovered failure modes such as rain and snow conditions and transmission artifact frames, which provides robust data support for ongoing algorithm improvement.

4.2. Experimental Setup

4.2.1. Implementation Details

All training and testing experiments were conducted on the same workstation running Windows 11, with all models trained for 300 epochs using a batch size of 16, an initial learning rate of 0.01 with cosine annealing decay, and an input image resolution of 640 × 640 . The software and hardware configuration is listed in Table 2.
The input data were resized to 640 × 640 pixels, and the Stochastic Gradient Descent (SGD) algorithm was employed as the optimizer. The performance of the network model ceased to improve after 200 training epochs; thus, the number of training epochs was set to 200 in this paper. The experimental parameters is listed in Table 3.

4.2.2. Comparison Methods

To comprehensively evaluate the proposed method, YOLOv10-Intrusion is compared against the following representative baseline detectors. Faster R-CNN [38] is a classical two-stage detector that uses a Region Proposal Network (RPN) to generate candidate regions followed by an ROI-aligned classification branch. SSD (Single Shot MultiBox Detector) [39] is a one-stage multi-scale anchor-based detector that directly regresses class scores and box offsets from multiple feature map levels. YOLOv8s [40] is the small-scale variant of YOLOv8, adopting a C2f backbone and a decoupled detection head with anchor-free regression. YOLOv11s [41] introduces C3k2 and C2PSA modules for enhanced multi-scale spatial attention. YOLOv12s [42] incorporates an area attention mechanism that partitions feature maps into non-overlapping regions for improved large receptive field modeling. RT-DETR-L is an advanced transformer detector that achieves fast, accurate real-time object detection through dynamic sparse attention and efficient end-to-end decoding. Deformable DETR extends the DETR framework by introducing deformable attention mechanisms, which adaptively sample informative regions from feature maps to alleviate the high computational cost of standard self-attention [43,44]. This design enables Deformable DETR to effectively handle dense object detection and complex scenes, providing a strong transformer-based baseline for comparison. YOLO-LCR [45] is a YOLO-based detector incorporating the Reparameterized Channel Shuffle (RSC) mechanism, serving as a direct architectural reference for the proposed RCS_M module. TSA-YOLO [46] is a YOLO variant specifically designed for vehicle detection under variable illumination conditions, providing a domain-targeted baseline for nighttime and adverse-lighting scenarios. All comparison models were trained and evaluated under identical experimental configurations on the vehicle area intrusion dataset.

4.2.3. Evaluation Metrics

The main evaluation metrics adopted in this paper are precision (P), recall (R), mAP@0.5, and mAP@0.5:0.95, defined as follows:
Precision (P) measures the proportion of correctly predicted positive samples among all samples predicted as positive:
P = T P T P + F P
Precision directly quantifies the proportion of valid enforcement alerts among all triggered alerts, reflecting whether intrusion warnings issued by the platform correspond to actual violations. High precision is essential for the practical usability of the detection system.
Recall (R) measures the proportion of correctly detected positive samples among all actual positive samples:
R = T P T P + F N
where T P denotes true positives, F P denotes false positives, and F N denotes false negatives. In vehicle area intrusion detection, recall quantifies the algorithm’s capacity to detect all genuine intrusion events without omission. High recall is a prerequisite for reliable enforcement monitoring, particularly under challenging conditions such as target occlusion, nighttime low illumination, and adverse weather.
Mean average precision (mAP) reflects the average detection accuracy across all categories. The AP for each category is computed as the area under the precision-recall (PR) curve:
AP = P d R
mAP = 1 K i = 1 K AP i
where K is the total number of categories (5 in this dataset). mAP@0.5 uses a fixed IoU threshold of 0.5. mAP@0.5:0.95 provides a more stringent localization evaluation by averaging mAP across ten uniformly spaced IoU thresholds:
mAP @ 0.5 : 0.95 = 1 10 t { 0.50 , 0.55 , , 0.95 } mAP ( t )
The F 1 score combines precision and recall into a single balanced indicator of detection performance:
F 1 = 2 · P · R P + R
A high F 1 score reflects a favorable trade-off between false alarms and missed detections, which is particularly relevant for enforcement applications that demand both precision and recall.

4.3. Results and Analysis

4.3.1. Comparison Experiments

To validate the superiority of YOLOv10-Intrusion for vehicle area intrusion detection, comparison experiments were conducted against the seven baseline detectors described in Section 4.2.2 under identical experimental settings. Results are shown in Table 4.
YOLOv10-Intrusion achieves mAP@0.5 values that are 15.4, 13.3, 4.5, 3.8, 4.2, 5.1, and 8.8 percentage points higher than Faster R-CNN, SSD, YOLOv8s, YOLOv11s, YOLOv12s, RTDETR-L and Deformable DETR respectively, with mAP@0.5:0.95, precision, and recall all outperforming compared methods. The algorithm also outperforms TSA-YOLO and YOLO-LCR, confirming that the proposed combination of RCS_M, ODConv, BiFPN, and WIoU yields consistent and complementary improvements for the vehicle area intrusion detection task, meeting the demands of real-time, high-precision enforcement monitoring under real-world road surveillance conditions.

4.3.2. Generalization Experiment

To further verify the effectiveness of the proposed algorithm, we conducted generalization experiments on three public datasets: KITTI, VOC2007, and COCO. We selected 3000 images from KITTI, 2500 images from VOC2007, and 2800 images from COCO for the generalization experiments. Since the target categories in this study are car, van, truck, muck car, and tricycle, we only retained images containing the above five vehicle categories from the public datasets and removed the corresponding annotations of irrelevant classes. The experimental environment and parameter settings were kept consistent with those described above. Results are shown in Table 5.
For model detection performance, on the KITTI dataset, YOLOv10-Intrusion improves the mAP@0.5 by 2.4 percentage points and mAP@0.5:0.95 by 1.8 percentage points compared with YOLOv10s. On the VOC2007 dataset, YOLOv10-Intrusion increases mAP@0.5 by 1.7 percentage points and mAP@0.5:0.95 by 2.6 percentage points. On the COCO dataset, YOLOv10-Intrusion boosts mAP@0.5 by 5.4 percentage points and mAP@0.5:0.95 by 3.8 percentage points. The experimental results demonstrate that the proposed algorithm achieves superior detection performance on various public datasets and exhibits strong generalization ability, effectively reducing false detection and missed detection rates.

4.3.3. Ablation Study

YOLOv10-Intrusion incorporates four architectural improvements over the baseline YOLOv10s: the RCS_M module in the backbone, the C2f_OD module replacing the original C2f in the neck, BiFPN for neck-level feature fusion, and WIoU replacing CIoU in the detection head. To evaluate the individual contribution of each modification and assess the effectiveness of the overall design, ablation experiments were conducted on the custom dataset under consistent hardware, software, and training configurations across all experiment groups. Results are shown in Table 6.
Experiment 1 shows that the C2f_OD module adaptively weights visual features and focuses on the detailed and discriminative regions of vehicles, enabling the model to recognize hard samples that are difficult to detect using the original model. This brings a 1.3% improvement in recall and enhances the fine-grained recognition capability in practical applications, which better meets the expected requirements of law enforcement departments. Experiment 2 demonstrates that the RCS_M module improves the model performance comprehensively. Compared with the original C2f module, the RCS_M module extracts more complete feature information and enhances cross-channel information interaction in the backbone network, providing more discriminative features for subsequent modules. Consequently, the overall performance of the model is improved, especially in terms of mAP@0.5, which is increased by 2.3 percentage points. Experiment 3 indicates that the BiFPN module complements shallow detailed information and deep semantic information through bidirectional weighted feature fusion, which improves the detection accuracy of the model under scenarios of drastic scale variation and dense occlusion. Specifically, the recall and mAP@0.5 are increased by 1.6 and 0.9 percentage points, respectively. Experiment 4 confirms that WIoU loss function reduces the interference caused by background clutter and low-quality anchors by dynamically adjusting gradient gains, thus improving the bounding box regression accuracy. This leads to a 2.5% improvement in recall and accelerates model convergence. Experiments 5–8 demonstrate that combining multiple modules yields further performance gains across all metrics. The synergistic effects of RCS_M and ODConv in feature enhancement, combined with the multi-scale fusion capability of BiFPN, prove particularly effective in dense, heavily occluded scenarios by improving localization accuracy for previously difficult vehicle instances.

4.3.4. Classification Accuracy Evaluation

The confusion matrices of YOLOv10s and YOLOv10-intrusion generated on the vehicle area intrusion detection dataset are shown in Figure 10.
As can be seen from Figure 10a, for YOLOv10s, the classification accuracies of car, van, muck car, truck, and tricycle are 85%, 75%, 83%, 81%, and 85%, respectively. As shown in Figure 10b, for YOLOv10-intrusion, the classification accuracies of car, van, muck car, truck, and tricycle reach 87%, 78%, 86%, 83%, and 85%, respectively. Compared with YOLOv10s, the accuracies of the first four categories are improved by 2%, 3%, 3%, and 2%, respectively. In addition, van has a 9% probability of being misclassified as car, which is 2% lower than that of YOLOv10s. Meanwhile, car no longer has a 1% probability of being misclassified as truck, representing a 1% reduction compared with YOLOv10s. In summary, the YOLOv10-intrusion model can effectively improve classification accuracy and alleviate the problems of misclassification and missed detection.

4.3.5. Visualization Analysis

As illustrated in Figure 11, YOLOv10-Intrusion consistently outperforms the baseline YOLOv10s across all evaluation metrics: precision improves by 1.5 percentage points, recall by 3.3 percentage points, mAP@0.5 by 3.6 percentage points, and mAP@0.5:0.95 by 2.8 percentage points.
To provide a more direct visual illustration of the detection improvements, representative road section scenes comparing baseline and improved model outputs are shown in Figure 12. Columns from left to right correspond to the original image, the YOLOv10s detection result, and the YOLOv10-Intrusion detection result. Row 1 presents a close-range large vehicle detection scenario; Rows 2 and 3 present scenes with dense target occlusion.
In Scene 1, YOLOv10s misidentified the rear of a large vehicle as a passenger car. In Scenes 2 and 3, YOLOv10s failed to detect occluded vehicles, causing missed detections. The improved model reduces inter-class misclassification of medium and large vehicle categories while achieving higher detection confidence for occluded instances, thereby enhancing detection stability and reducing missed detections in the vehicle area intrusion detection task.

4.3.6. Field Validation

Following preliminary online evaluation, field tests were conducted to validate detection performance under real-world conditions. The test procedure involved physically driving a test vehicle along the monitored road sections to simulate illegal intrusion events. Each test session covered three stages: simulating camera orientation perturbations to assess robustness to environmental disturbances; traversing the legal driving zone to verify the absence of false alerts; and entering the designated intrusion zone to confirm real-time detection and alert triggering. Field test scenarios are shown in Figure 13.
The first two test sections were located on high-traffic main roads with cameras primarily capturing vehicle side profiles. When the test vehicle entered the pedestrian path from a distance, no alert was initially triggered because only the front section had entered the detection zone—an issue resolved after dataset refinement. The third and fourth sections were low-traffic forest roads with frontal camera perspectives, where immediate detection was achieved even for small vehicle front intrusions. The fifth and sixth sections were construction site entrances with rear-focused cameras, where alerts were rapidly triggered after vehicle entry. All six sections achieved satisfactory detection performance in daytime evaluation. Nighttime tests conducted with large vehicles successfully replicated all daytime scenarios, achieving complete detection coverage across all simulated intrusion events.
Three rounds of iterative dataset refinement were performed. In the first round, additional side-profile images and targeted augmentation were incorporated to address underrepresentation of small vehicle side views, resolving the false negatives at the first two sections. In the second round, ODConv was introduced and low-quality large vehicle images were filtered to address nighttime truck/muck-truck misclassification; recall consequently improved by 1.3 percentage points, and the detection threshold was raised from 0.5 to 0.6 without significant missed detections. In the third round, representative background and transmission artifact images were added to suppress false alerts, improving precision by 0.6 percentage points. Figure 14 compares detection results before and after dataset improvement.
Before improvement, detected vehicle confidence scores were low and one headlight-occluded vehicle was missed. After refinement, more vehicles were detected with higher confidence scores, demonstrating the effectiveness of the iterative dataset improvement in reducing missed detections under challenging nighttime conditions.

5. Conclusions

This paper proposes YOLOv10-Intrusion, an improved object detection algorithm based on YOLOv10s for vehicle area intrusion detection in real-world road surveillance environments. To address prevalent false detection and missed detection challenges, a dedicated vehicle area intrusion dataset was constructed from field surveillance footage, covering five vehicle categories across diverse road environments and lighting conditions.
The proposed algorithm integrates four complementary architectural improvements. The self-designed RCS_M module replaces the backbone C2f structure to enhance cross-channel information exchange, local feature representation, and cross-scale interaction capability. ODConv (C2f_OD) replaces standard convolution in the neck C2f module to strengthen fine-grained recognition of visually similar vehicle categories. BiFPN replaces the original neck feature fusion structure to adaptively balance multi-scale feature contributions and improve recall in dense, occluded scenarios. WIoU replaces the CIoU loss function to improve bounding box regression stability and convergence speed while alleviating the adverse effects of class imbalance.
Experimental results and field testing demonstrate the capacity of YOLOv10-Intrusion to accurately identify illegal vehicles in real-world surveillance environments, meeting practical enforcement requirements. Compared with the baseline YOLOv10s, the proposed algorithm achieves improvements of 1.5, 3.3, 3.6, and 2.8 percentage points in precision, recall, mAP@0.5, and mAP@0.5:0.95, respectively, and outperforms other mainstream object detection algorithms in terms of both model accuracy evaluation metrics and actual deployment effectiveness.
Future work will explore model compression techniques (pruning and knowledge distillation) to reduce parameter count and improve deployment efficiency under limited computational resources. Additionally, extending the platform to support multi-camera joint detection and incorporating nighttime-enhanced imaging modalities will further improve intrusion detection robustness under adverse environmental conditions.

Author Contributions

Conceptualization, C.J. and F.K.; Methodology, C.J. and F.K.; Software, C.J.; Validation, C.J. and F.K.; Formal analysis, C.J.; Investigation, C.J.; Resources, F.K.; Data curation, C.J.; Writing—original draft preparation, C.J.; Writing—review and editing, C.J. and F.K.; Visualization, C.J.; Supervision, F.K.; Project administration, F.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available at https://github.com/MrFENGXINGYU/vehicle-intrusion-det (accessed on 4 March 2026).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Liu, D.; Gu, Y.; Zhang, H.; Khojastehpour, M.; Han, L.D. TIE-LR: A multi-module deep learning framework for traffic monitoring on low-resolution videos. Eng. Appl. Artif. Intell. 2026, 163, 113112. [Google Scholar] [CrossRef]
  2. Tang, J.; Ye, C.; Zhou, X.; Xu, L. YOLO-Fusion and Internet of Things: Advancing object detection in smart transportation. Alex. Eng. J. 2024, 107, 1–12. [Google Scholar] [CrossRef]
  3. Boroujeni, S.P.H.; Mehrabi, N.; Alzorgan, H.; Fazeli, M.; Razi, A. All you need for object detection: From pixels, points, and prompts to Next-Gen fusion and multimodal LLMs/VLMs in autonomous vehicles. Image Vis. Comput. 2026, 169, 105944. [Google Scholar] [CrossRef]
  4. Hu, T.; Gao, F.; Zhou, F. Railway obstacle intrusion detection and risk assessment based on MSIA-YOLOv8 and DALNet. Expert Syst. Appl. 2026, 299, 130132. [Google Scholar] [CrossRef]
  5. Zhang, Z.; Chen, P.; Huang, Y.; Dai, L.; Xu, F.; Hu, H. Railway obstacle intrusion warning mechanism integrating YOLO-based detection and risk assessment. J. Ind. Inf. Integr. 2024, 38, 100571. [Google Scholar] [CrossRef]
  6. Sodhro, A.H.; Kannam, S.; Jensen, M. Real-time efficiency of YOLOv5 and YOLOv8 in human intrusion detection across diverse environments and recommendation. Internet Things 2025, 33, 101707. [Google Scholar] [CrossRef]
  7. Dong, S.; Li, Y.; Wang, W.; Hu, Q.; Zhang, R.; Wang, X. FPE-YOLO: Real-Time Small-Object Detection for Third-Party Intrusion in Oil and Gas Pipeline Integrity Monitoring. J. Pipeline Sci. Eng. 2026; In Press. [CrossRef]
  8. Alnfiai, M.M. Lightning Search Algorithm with Deep Transfer Learning-Based Vehicle Classification. Comput. Mater. Contin. 2022, 74, 6505–6521. [Google Scholar] [CrossRef]
  9. Tang, Y.; Qian, Y. Real-time railroad crossing surveillance with edge computing. Eng. Appl. Artif. Intell. 2026, 165, 113338. [Google Scholar] [CrossRef]
  10. Han, F.; Ye, P.; She, C.; Duan, S.; Wang, L.; Liu, D. MMID-Bench: A Comprehensive Benchmark for Multi-Domain Multi-Category Intrusion Detection. IEEE Trans. Intell. Veh. 2024, 9, 6419–6435. [Google Scholar] [CrossRef]
  11. Huang, H.; Zhao, G.; Bo, Y.; Yu, J.; Liang, L.; Yang, Y.; Ou, K. Railway intrusion detection based on refined spatial and temporal features for UAV surveillance scene. Measurement 2023, 211, 112602. [Google Scholar] [CrossRef]
  12. Omodaratan, B.; Jamali, A.; Wiley, T.; Al-Saadi, Z.; Mallipeddi, R.; Asadi, E.; Asadi, H.; Sadeghian, R.; Sareh, S.; Khayyam, H. Advances in You Only Look Once (YOLO) algorithms for lane and object detection in autonomous vehicles. Eng. Appl. Artif. Intell. 2026, 168, 113893. [Google Scholar] [CrossRef]
  13. Han, X.; Chang, J.; Wang, K. Real-time object detection based on YOLO-v2 for tiny vehicle object. Procedia Comput. Sci. 2021, 183, 61–72. [Google Scholar] [CrossRef]
  14. Chaman, M.; Maliki, A.E.; Dahou, H.; Hadjoudja, A. Benchmarking YOLO-based deep learning models for real-time object detection in hybrid ADAS and intelligent transportation systems. Results Eng. 2026, 29, 108942. [Google Scholar] [CrossRef]
  15. Wang, X.; Hao, X.; Wang, K. MC-YOLO-Based Lightweight Detection Method for Nighttime Vehicle Images in a Semantic Web-Based Video Surveillance System. Int. J. Semant. Web Inf. Syst. 2023, 19, 18. [Google Scholar] [CrossRef]
  16. Luo, X.; Zhu, H.; Zhang, Z. IR-YOLO: Real-Time Infrared Vehicle and Pedestrian Detection. Comput. Mater. Contin. 2024, 78, 2667–2687. [Google Scholar] [CrossRef]
  17. Kang, L.; Lu, Z.; Meng, L.; Gao, Z. YOLO-FA: Type-1 fuzzy attention based YOLO detector for vehicle detection. Expert Syst. Appl. 2024, 237, 121209. [Google Scholar] [CrossRef]
  18. Duan, C.; Guo, Y.; Duan, X.; Li, G.; Sheng, B. DD-YOLO: A dual-channel dual-path YOLO network for target detection of blurred vehicles. Digit. Signal Process. 2026, 168, 105565. [Google Scholar] [CrossRef]
  19. Du, D.; Bi, M.; Xie, Y.; Liu, Y.; Qi, G.; Guo, Y. MLE-YOLO: A lightweight and robust vehicle and pedestrian detector for adverse weather in autonomous driving. Digit. Signal Process. 2026, 168, 105628. [Google Scholar] [CrossRef]
  20. Wang, H.; Chaw, J.K.; Mat Daud, M.; Shi, L.; Huang, N.; Ting, T.T.; Pu, L. Hybrid-YOLO: Lightweight Mamba-Transformer Hybrid with multi-scale fusion for real-world traffic detection. ICT Express 2026, 12, 214–222. [Google Scholar] [CrossRef]
  21. Liu, W.; Qiao, X.; Zhao, C.; Deng, T.; Yan, F. VP-YOLO: A human visual perception-inspired robust vehicle-pedestrian detection model for complex traffic scenarios. Expert Syst. Appl. 2025, 274, 126837. [Google Scholar] [CrossRef]
  22. Li, Y.; Zhang, M.; Zhang, C.; Liang, H.; Li, P.; Zhang, W. YOLO-CCS: Vehicle detection algorithm based on coordinate attention mechanism. Digit. Signal Process. 2024, 153, 104632. [Google Scholar] [CrossRef]
  23. Pan, H.; Guan, S.; Zhao, X. LVD-YOLO: An efficient lightweight vehicle detection model for intelligent transportation systems. Image Vis. Comput. 2024, 151, 105276. [Google Scholar] [CrossRef]
  24. Alahdal, N.M.; Abukhodair, F.; Meftah, L.H.; Cherif, A. Real-time Object Detection in Autonomous Vehicles with YOLO. Procedia Comput. Sci. 2024, 246, 2792–2801. [Google Scholar] [CrossRef]
  25. Hu, X.; Wang, X.; Yang, X.; Wang, D.; Zhang, P.; Xiao, Y. An infrared target intrusion detection method based on feature fusion and enhancement. Def. Technol. 2020, 16, 737–746. [Google Scholar] [CrossRef]
  26. Balakrishnan, P.; Anny Leema, A.; Kiruba B, G.G.; Gupta, A.; Aryan, R. Deep-track: A real-time animal detection and monitoring system for mitigating human-wildlife conflict in fringe areas. J. Nat. Conserv. 2025, 88, 127063. [Google Scholar] [CrossRef]
  27. Lu, S.; Zhao, H.; Zhang, E.; Zhao, Y.; Zhang, Y.; Zhang, Z. IMV-YOLO: Infrared multi-angle vehicle real-time detection network based YOLOv11 for adverse weather conditions. Int. J. Intell. Comput. Cybern. 2025, 18, 731–758. [Google Scholar] [CrossRef]
  28. Li, M.; Li, H.; Zhu, J.; Zhang, X. Visual Detection Algorithms for Counter-UAV in Low-Altitude Air Defense. Comput. Mater. Contin. 2026, 86, 32. [Google Scholar] [CrossRef]
  29. Lu, J.; Song, W.; Zhang, Y.; Yin, X.; Zhao, S. Real-time defect detection in underground sewage pipelines using an improved YOLOv5 model. Autom. Constr. 2025, 173, 106068. [Google Scholar] [CrossRef]
  30. Fan, Q.; Li, Y.; Deveci, M.; Zhong, K.; Kadry, S. LUD-YOLO: A novel lightweight object detection network for unmanned aerial vehicle. Inf. Sci. 2025, 686, 121366. [Google Scholar] [CrossRef]
  31. Zhang, Y.; Dai, Z.; Pan, C.; Zhang, G.; Xu, J. NOC-YOLO: An exploration to enhance small-target vehicle detection accuracy in aerial infrared images. Infrared Phys. Technol. 2025, 149, 105905. [Google Scholar] [CrossRef]
  32. Song, Y.; Hong, S.; Hu, C.; He, P.; Tao, L.; Tie, Z.; Ding, C. MEB-YOLO: An Efficient Vehicle Detection Method in Complex Traffic Road Scenes. Comput. Mater. Contin. 2023, 75, 5761–5784. [Google Scholar] [CrossRef]
  33. Liu, Y.; Huang, Z.; Song, Q.; Bai, K. PV-YOLO: A lightweight pedestrian and vehicle detection model based on improved YOLOv8. Digit. Signal Process. 2025, 156, 104857. [Google Scholar] [CrossRef]
  34. Nan, Z.; Liu, W.; Zhu, G.; Zhao, H.; Xia, W.; Lin, X.; Yang, Y. LiDAR-Camera joint obstacle detection algorithm for railway track area. Expert Syst. Appl. 2025, 275, 127089. [Google Scholar] [CrossRef]
  35. Chaurasia, D.; Patro, B. Detection of objects in satellite and aerial imagery using channel and spatially attentive YOLO-CSL for surveillance. Image Vis. Comput. 2024, 147, 105070. [Google Scholar] [CrossRef]
  36. Zhou, W.; Wang, J.; Meng, X.; Wang, J.; Song, Y.; Liu, Z. MP-YOLO: Multidimensional feature fusion based layer adaptive pruning YOLO for dense vehicle object detection algorithm. J. Vis. Commun. Image Represent. 2025, 112, 104560. [Google Scholar] [CrossRef]
  37. Tahir, N.U.A.; Kuang, L.; Sinishaw, M.L.; Asim, M. PV3M-YOLO: A triple attention-enhanced model for detecting pedestrians and vehicles in UAV-enabled smart transport networks. J. Vis. Commun. Image Represent. 2026, 115, 104701. [Google Scholar] [CrossRef]
  38. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2015; Volume 28. [Google Scholar]
  39. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  40. Jocher, G.; Chaurasia, A.; Qiu, J. Ultralytics YOLOv8. Version 8.0.0, AGPL-3.0 License. 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 10 January 2023).
  41. Jocher, G.; Qiu, J. Ultralytics YOLO11. Version 11.0.0, AGPL-3.0 License. 2024. Available online: https://github.com/ultralytics/ultralytics (accessed on 30 September 2024).
  42. Tian, Y.; Ye, Q.; Doermann, D. YOLOv12: Attention-Centric Real-Time Object Detectors. arXiv 2025, arXiv:2502.12524. [Google Scholar]
  43. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable Transformers for End-to-End Object Detection. arXiv 2021, arXiv:2010.04159. [Google Scholar]
  44. Kong, Y.; Shang, X.; Jia, S. Drone-DETR: Efficient Small Object Detection for Remote Sensing Image Using Enhanced RT-DETR Model. Sensors 2024, 24, 5496. [Google Scholar] [CrossRef]
  45. Moussaoui, H.; El Akkad, N.; Benslimane, M.; El-Shafai, W.; Baihan, A.; Hewage, C.; Rathore, R.S. Enhancing automated vehicle identification by integrating YOLO v8 and OCR techniques for high-precision license plate detection and recognition. Sci. Rep. 2024, 14, 14389. [Google Scholar] [CrossRef] [PubMed]
  46. Han, X.J.; Qu, Z.; Xia, S.F. Teacher-student adversarial YOLO for domain adaptive detection in traffic scenes under adverse weather. Displays 2025, 92, 103289. [Google Scholar] [CrossRef]
Figure 1. Technical workflow of the vehicle area intrusion detection system.
Figure 1. Technical workflow of the vehicle area intrusion detection system.
Sensors 26 02118 g001
Figure 2. Architecture of YOLOv10-Intrusion. Modified modules are highlighted in dark color.
Figure 2. Architecture of YOLOv10-Intrusion. Modified modules are highlighted in dark color.
Sensors 26 02118 g002
Figure 3. Structure of the RCS module.
Figure 3. Structure of the RCS module.
Sensors 26 02118 g003
Figure 4. Channel Shuffle.
Figure 4. Channel Shuffle.
Sensors 26 02118 g004
Figure 5. Comparison of RCS-OSA and the proposed RCS_M module.
Figure 5. Comparison of RCS-OSA and the proposed RCS_M module.
Sensors 26 02118 g005
Figure 6. Overview of the ODConv module with four complementary attention dimensions.
Figure 6. Overview of the ODConv module with four complementary attention dimensions.
Sensors 26 02118 g006
Figure 7. Structure of the C2f_OD module with ODConv replacing standard convolution.
Figure 7. Structure of the C2f_OD module with ODConv replacing standard convolution.
Sensors 26 02118 g007
Figure 8. Structural comparison of FPN, PAN, and BiFPN.
Figure 8. Structural comparison of FPN, PAN, and BiFPN.
Sensors 26 02118 g008
Figure 9. Sample images from the vehicle area intrusion dataset. (The text in the picture represents the time and location of the shooting).
Figure 9. Sample images from the vehicle area intrusion dataset. (The text in the picture represents the time and location of the shooting).
Sensors 26 02118 g009
Figure 10. Confusion Matrices of Different Models on the Self-Built Dataset YOLOv10s.
Figure 10. Confusion Matrices of Different Models on the Self-Built Dataset YOLOv10s.
Sensors 26 02118 g010
Figure 11. Performance metric comparison between YOLOv10-Intrusion and baseline YOLOv10s.
Figure 11. Performance metric comparison between YOLOv10-Intrusion and baseline YOLOv10s.
Sensors 26 02118 g011
Figure 12. Detection result comparison before and after model improvement. (The text in the picture represents the time and location of the shooting).
Figure 12. Detection result comparison before and after model improvement. (The text in the picture represents the time and location of the shooting).
Sensors 26 02118 g012
Figure 13. Field tests on monitored road sections.
Figure 13. Field tests on monitored road sections.
Sensors 26 02118 g013
Figure 14. Comparison of nighttime and no-target misdetection before and after dataset improvement. (The text in the picture represents the time and location of the shooting).
Figure 14. Comparison of nighttime and no-target misdetection before and after dataset improvement. (The text in the picture represents the time and location of the shooting).
Sensors 26 02118 g014
Table 1. Category distribution of the vehicle area intrusion dataset.
Table 1. Category distribution of the vehicle area intrusion dataset.
CategoryInstance Count
Car7413
Van603
Muck Car1594
Truck1886
Tricycle964
Total12,460
Table 2. Experimental software and hardware configuration.
Table 2. Experimental software and hardware configuration.
ConfigurationSpecification
Operating SystemWindows 11
Programming LanguagePython 3.9
Deep Learning FrameworkPyTorch 2.0.1
GPU AccelerationCUDA 11.3
CPUIntel Xeon Platinum 8350C
System RAM56 GB
GPUNVIDIA RTX 3090 (24 GB VRAM)
Table 3. Experimental parameters.
Table 3. Experimental parameters.
ParameterParameter Value
Epoch200
BatchSize8
Ir00.01
Momentum0.937
Weight decay0.0005
Table 4. Performance comparison of different algorithms on the vehicle area intrusion dataset.
Table 4. Performance comparison of different algorithms on the vehicle area intrusion dataset.
ModelP (%)R (%)mAP50 (%)mAP50:95 (%)Params (M)GFLOPsFPS
YOLOv10-Intrusion (Ours)89.278.686.669.711.529.9123
Faster R-CNN78.865.271.259.1137.0370.232
SSD79.163.173.359.324.8275.457
YOLOv8s84.371.982.165.811.128.6127
YOLOv11s87.175.182.867.39.421.5144
YOLOv12s86.374.782.467.19.119.7148
RTDETR-L84.576.681.563.031.0108.376
Deformable DETR80.672.177.861.239.897.683
YOLO-LCR82.170.281.164.412.440.490
TSA-YOLO84.871.982.667.717.455.683
Table 5. Generalization experiment.
Table 5. Generalization experiment.
DatasetModelmAP50 (%)mAP50:95 (%)Params (M)GFLOPsFPS
KITTIYOLOv10s83.860.48.124.6122
KITTIYOLOv10-intrusion86.062.211.529.9116
VOC2007YOLOv10s82.359.28.124.6107
VOC2007YOLOv10-intrusion84.161.811.529.9102
COCOYOLOv10s55.636.48.124.697
COCOYOLOv10-intrusion61.040.211.529.993
Table 6. Ablation experiment results on the vehicle area intrusion dataset. (× indicates that the module has not been used, ✓ indicates the use of the module).
Table 6. Ablation experiment results on the vehicle area intrusion dataset. (× indicates that the module has not been used, ✓ indicates the use of the module).
Exp.C2f_ODRCS_MBiFPNWIoUP (%)R (%)mAP50 (%)mAP50:95 (%)Params (M)GFLOPsFPS
0××××87.775.383.066.98.124.6142
1 86.776.683.067.08.625.8134
2 87.675.585.368.610.627.9128
3 88.476.983.967.58.525.4137
4 87.377.783.667.48.124.6142
5 87.976.685.067.811.529.9123
6 89.776.285.668.911.028.7130
7 88.677.385.468.511.129.1126
889.278.686.669.711.529.9123
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jie, C.; Ke, F. YOLOv10-Intrusion: An Improved YOLOv10-Based Algorithm for Vehicle Area Intrusion Detection. Sensors 2026, 26, 2118. https://doi.org/10.3390/s26072118

AMA Style

Jie C, Ke F. YOLOv10-Intrusion: An Improved YOLOv10-Based Algorithm for Vehicle Area Intrusion Detection. Sensors. 2026; 26(7):2118. https://doi.org/10.3390/s26072118

Chicago/Turabian Style

Jie, Chuanyue, and Fuyang Ke. 2026. "YOLOv10-Intrusion: An Improved YOLOv10-Based Algorithm for Vehicle Area Intrusion Detection" Sensors 26, no. 7: 2118. https://doi.org/10.3390/s26072118

APA Style

Jie, C., & Ke, F. (2026). YOLOv10-Intrusion: An Improved YOLOv10-Based Algorithm for Vehicle Area Intrusion Detection. Sensors, 26(7), 2118. https://doi.org/10.3390/s26072118

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop