Next Article in Journal
Enhancing Deep-Learning Classification for Remote Motor Imagery Rehabilitation Using Multi-Subject Transfer Learning in IoT Environment
Previous Article in Journal
Research Progress of MEMS Gas Sensors: A Comprehensive Review of Sensing Materials
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

YOLO-BOS: An Emerging Approach for Vehicle Detection with a Novel BRSA Mechanism

1
College of Electrical Engineering, Henan University of Technology, Zhengzhou 450001, China
2
School of Artificial Intelligence and Big Data, Henan University of Technology, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(24), 8126; https://doi.org/10.3390/s24248126
Submission received: 18 November 2024 / Revised: 14 December 2024 / Accepted: 18 December 2024 / Published: 19 December 2024
(This article belongs to the Section Vehicular Sensing)

Abstract

:
In intelligent transportation systems, accurate vehicle target recognition within road scenarios is crucial for achieving intelligent traffic management. Addressing the challenges posed by complex environments and severe vehicle occlusion in such scenarios, this paper proposes a novel vehicle-detection method, YOLO-BOS. First, to bolster the feature-extraction capabilities of the backbone network, we propose a novel Bi-level Routing Spatial Attention (BRSA) mechanism, which selectively filters features based on task requirements and adjusts the importance of spatial locations to more accurately enhance relevant features. Second, we incorporate Omni-directional Dynamic Convolution (ODConv) into the head network, which is capable of simultaneously learning complementary attention across the four dimensions of the kernel space, therefore facilitating the capture of multifaceted features from the input data. Lastly, we introduce Shape-IOU, a new loss function that significantly enhances the accuracy and robustness of detection results for vehicles of varying sizes. Experimental evaluations conducted on the UA-DETRAC dataset demonstrate that our model achieves improvements of 4.7 and 4.4 percentage points in mAP@0.5 and mAP@0.5:0.95, respectively, compared to the baseline model. Furthermore, comparative experiments on the SODA10M dataset corroborate the superiority of our method in terms of precision and accuracy.

1. Introduction

Currently, the acceleration of urbanization processes and the increase in vehicle ownership have exerted immense pressure on transportation systems, with traffic congestion emerging as a severe issue confronting cities. Prolonged periods of traffic congestion not only affect the efficiency of citizens’ travel and the quality of their lives but also exacerbate air pollution and energy consumption [1]. According to statistics released by TOMTOM in 2023, commuters in more than 10 global cities experience annual delays exceeding 100 h due to traffic congestion, with the problem expected to worsen. Consequently, the development of intelligent transportation systems (ITSs) [2] has become an urgent priority to alleviate traffic congestion and enhance transportation efficiency. Accurate identification of on-road vehicles, pedestrians, and other targets is a critical task within ITS and forms the foundation for achieving traffic intelligence.
However, in urban road traffic scenarios, the prevalence of occlusion and overlap phenomena significantly impacts the regression localization accuracy of vehicle detection and the classification results of various vehicle types, leading to frequent false detections and missed detections. To address these issues, this paper proposes YOLO-BOS, a novel vehicle-detection method based on a series of improvements to YOLOv8 [3], which effectively enhances the detection accuracy of small targets in complex road scenarios while maintaining a balance between accuracy and speed. The key improvements presented in this paper are as follows:
  • To enhance the feature-extraction capabilities of the backbone network, we propose a novel Bi-level Routing Spatial Attention (BRSA) mechanism, which integrates the Bi-level Routing Attention (BRA) mechanism with the Spatial Attention Mechanism (SAM). This BRSA mechanism selectively filters features and dynamically adjusts the weights of spatial positions based on task requirements, facilitating focused attention on capturing intricate and subtle vehicle-related details and, therefore, enhancing the network’s ability to recognize salient features in complex backgrounds.
  • In the head network, we integrate the Omni-directional Dynamic Convolution (ODConv), which dynamically learns complementary attention across four dimensions of the kernel space, enabling multi-dimensional information extraction. This enhances the convolutional neural network’s adaptability to variations in input data features, providing more flexible convolutional operations for handling complex and diverse data and thus better accommodating the intricate and changing conditions of road targets.
  • To enhance the accuracy and robustness of vehicle target detection across various sizes, we introduce a novel loss function termed Shape-IOU. Building upon the traditional IOU, this loss function accounts for the impact of inherent bounding box attributes, namely shape and scale, on bounding box regression, therefore improving the localization precision of detected vehicles.
This article is structured into five sections. To understand the innovation of this research, Section 2 summarizes prior studies in the domain. Subsequently, Section 3 delineates the enhanced network architecture designed for more precise vehicle detection. In Section 4, the experimental setup and analysis are presented to evaluate the performance of the detection network. Finally, Section 5 consolidates the primary findings of the research and outlines directions for future investigation.

2. Related Work

2.1. Object Detection

Object detection is a critical task in computer vision aimed at identifying and locating objects of interest within images or videos. With the emergence of GPU technologies, deep learning-based object detection methods have progressively supplanted traditional approaches, becoming predominant. These methods are broadly categorized into two main types: two-stage and one-stage algorithms. Two-stage algorithms utilize a region proposal network (RPN) to propose candidate areas, followed by classification and regression processes. They typically feature complex network architectures and moderate detection speeds. Early two-stage detection algorithms employed selective search techniques, such as R-CNN [4], SPPNet [5], and Faster R-CNN [6]. Recent advancements in two-stage algorithms incorporate feature pyramid network (FPN) structures, which integrate spatial information from both high and low layers, as demonstrated in Mask R-CNN [7] and Libra R-CNN [8]. In contrast, one-stage detection algorithms employ simpler network architectures and achieve faster detection speeds, making them suitable for applications in scenarios like complex road scenes. Consequently, the pursuit of research on one-stage vehicle-detection algorithms holds paramount significance. The YOLO series of algorithms constitutes a paradigmatic example of one-stage detection networks, originally introduced by Joseph et al. in 2016 [9]. Following its inception, a multitude of research teams have embarked on investigations based on the YOLO framework, propelling its continual evolution and refinement. The sequential unveiling of YOLOv4 and YOLOv5 marked significant milestones in performance enhancement, solidifying the series’ prominent position within the realm of one-stage detection. Capitalizing on this robust foundation, subsequent iterations have undergone meticulous optimization. YOLOv8 refined the feature-extraction capabilities of YOLOv5 through the adoption of a decoupled head architecture and an anchor-free mechanism. YOLOv9, built upon the framework of YOLOv7, incorporated Programmable Gradient Information (PGI) alongside a Generalized Efficient Layer Aggregation Network (GELAN) structure. YOLOv10 innovatively eliminated the necessity for Non-Maximum Suppression (NMS) while introducing a sophisticated dual-label assignment strategy. As of the present, the YOLO series has advanced to its eleventh iteration, YOLO11.

2.2. Vehicle Detection

Currently, an extensive body of research has delved into developing algorithms for vehicle detection. Within the realm of two-stage detection frameworks, Zhang et al. [10] have augmented Faster R-CNN by leveraging inter-frame differences and spatiotemporal context, effectively mitigating the issue of missed detections in vehicle recognition tasks. Xu et al. [11] have further bolstered the accuracy of nighttime vehicle detection by integrating deformable convolution and soft-NMS within the Faster R-CNN architecture. Othmani et al. [12], on the other hand, have amalgamated the strengths of three distinct networks to enhance Faster R-CNN, yielding notable gains in both speed and accuracy. In the domain of one-stage vehicle target detection, Guo et al. [13] have significantly improved the accuracy of vehicle detection in remote sensing imagery by seamlessly incorporating an adaptive fusion module and a sub-pixel convolution module into YOLOv3. Zhang et al. [14] have addressed the challenge of detecting small targets by introducing bounding boxes with orientation angles, therefore enhancing detection precision. Additionally, Zhang et al. [15] have integrated a coordinate attention mechanism into YOLOv4, empowering the model to dynamically adjust weights and focus on salient regions within images. To optimize performance further, Wang et al. [16] and Chen et al. [17] have independently replaced the backbone network of YOLOv4 with CSPDarknet53-dcn and MobileNetv2, respectively, achieving notable improvements in both inference speed and detection accuracy. More recently, Miao et al. [18] presented the YOLO-VSF model by incorporating methodologies such as the VGG16 network and the SENet attention mechanism. The research endeavors surrounding YOLOv5 primarily focus on two pivotal directions: lightweight design [19,20,21] and enhanced small object detection [22,23,24]. Moreover, He et al. [25] have introduced the ShuffYOLOX model, a novel approach that leverages the ShuffDet and ECA modules within the YOLOX framework. This integration has yielded notable improvements. On the other hand, Gao et al. [26] have devised two innovative lightweight modules, DG and DS, which have achieved a remarkable 51% reduction in the number of YOLOX model parameters while maintaining its performance, showcasing the ongoing advancements in optimizing object detection models for efficiency and accuracy. Regarding YOLOv7, Wang et al. [27] have made seminal contributions by incorporating the Ghost module alongside improvements to the feature pyramid and loss function, culminating in the YOLOv7-GBW model. Concurrently, Zhang et al. [28] have reimagined the YOLOv7 backbone by reconstructing it with the Res3Unit structure and augmenting it with the ACmix module, significantly boosting the model’s precision in detecting minute objects within images. Similarly, Li et al. [29] presented the MST-YOLOv8 model through the integration of the C2f-MLCA structure and the ST-P2Neck structure. Meanwhile, Liu et al. [30] introduced a more streamlined vehicle-detection model by refining the detection head and upsampling architecture of YOLOv8. In recent years, object detection models leveraging the Transformer architecture, particularly Real-Time DETR (RT-DETR), have been widely adopted for vehicle detection owing to their superior real-time performance. Liu et al. [31] made further contributions by proposing the MDFD2-DETR model, which incorporates innovative feature de-redundancy modules and hybrid positional encoding strategies. Additionally, Jin et al. [32] developed a lightweight algorithm grounded in RT-DETR, specifically tailored to enhance the detection of distant vehicle logos.

3. Methods

3.1. Network Architecture

In complex scenarios of vehicle detection, including intersections and congested roads during peak traffic periods, existing detection models encounter substantial challenges. Due to mutual occlusion between vehicles, which results in incomplete visibility of vehicle bodies and consequently reduces their proportion in the image, feature information is more prone to being lost in shallow network layers. This often results in a heightened incidence of false positives and missed detections during target identification. Hence, it is crucial for detection models to bolster their capacity to extract feature information and augment the richness of valid data in order to enhance detection accuracy.
To augment the model’s capacity for extracting feature information, this paper innovatively introduces the YOLO-BOS model, with its network architecture depicted in Figure 1. First, to enhance the capability of extracting salient foreground object features, we introduce an innovative BRSA attention mechanism strategically inserted prior to the SPPF module in the backbone network. By utilizing a bi-level routing approach and adjusting the weights of various positions, this mechanism reinforces useful features, enabling flexible computational allocation and content attention. This, in turn, enhances the ability to identify and capture salient features amid complex backgrounds. Additionally, we modify the C2F module in the head network using ODConv, which expands the model’s receptive field and perception range, facilitating the capture of more useful features, particularly for small-scale objects. Furthermore, during the training process, we introduce the Shape-IOU loss function, which takes into account the impact of the shape and size of bounding boxes on bounding box regression. This results in more accurate bounding box predictions for objects of various shapes, enhancing the regression performance and subsequently improving the overall detection accuracy.

3.2. BRSA Mechanism

In the backbone network, a novel BRSA attention mechanism is proposed, which is based on the BRA mechanism and capable of selectively filtering features according to task demands while leveraging SAM attention to adjust the significance of spatial positions for more accurate enhancement of useful features such as the shape, color, and texture of the target vehicle, therefore augmenting the model’s capacity to identify and capture salient features in complex backgrounds. Figure 2 illustrates the working process of BRSA: Initially, query ( Q ), key ( K ), value ( V ) and the adjacent matrix A r are obtained through linear mapping. Subsequently, a top-k routing attention mechanism is applied to A r to generate the routing index matrix I r , which is then utilized to aggregate K and V into K g and V g . Then, token-to-token attention is applied to Q , K g , and V g to produce O. Finally, max pooling ( F m a x ) and average pooling ( F a v g ) operations are performed to extract the final feature representation.

3.2.1. Bi-Level Routing Attention

The Bi-level Routing Attention (BRA) mechanism represents a dynamic sparse attention mechanism capable of adjusting the distribution of attention dynamically based on input content, therefore conserving computational time and resources. Its core concept bifurcates the attention mechanism into two phases: the first phase filters out most irrelevant key-value pairs in coarse regions, retaining only a small portion of routing areas, therefore reducing the interference of redundant information. Subsequently, in the second phase, token-to-token attention is applied within the preserved routing areas to identify the most relevant regions. The operational process is depicted in Figure 3.
Initially, the input feature map is partitioned into regions, with the given two-dimensional feature map X R H × W × C divided into S × S non-overlapping regions. This division enables the transformation of X into X r R S 2 × H W S 2 × C , therefore facilitating the linear mapping of Q , K , V R S 2 × H W S 2 × C , specifically:
Q = X r W q
K = X r W k
V = X r W v
where W q , W k , W v represent the projection weights for Q , K , V , respectively.
To determine which areas each given subregion should focus on, average pooling is utilized to compute the regional features, yielding Q r , K r R S 2 × C . Subsequently, the region-to-region affinity adjacency matrix A r is obtained through the following formula:
A r = Q r K r T
Each element within the adjacency matrix signifies the semantic relevance between two regions. Subsequently, by employing the top-k algorithm, the top-k connections in each region are preserved, effectively eliminating the portions with weaker associations:
I r = t o p I n d e x A r
Here, I r denotes the routing index matrix, in which the i t h row contains the K indices that are most strongly correlated with the i t h region.
The subsequent phase involves the collection of keys and values for each query within the region and the computation of token-to-token attention. Specifically, for each query in the region, tensors for keys and values are collected as follows:
K g = g a t h e r K , I r
V g = g a t h e r V , I r
In Equations (6) and (7), the gather function extracts data from the input tensor based on the index matrix, generating a new, combined tensor. Attention is then focused on the collected information:
O = A t t e n t i o n Q , K g , V g + L C E V
Herein, the L C E function aims to parameterize V through deep convolution, with the kernel size set to 5.
Through Bi-level Routing Attention, the model can achieve more flexible computation allocation and content awareness, with selective focusing aiding in the capture of complex details and subtle features pertinent to vehicles, therefore enhancing the efficacy of vehicle object detection.

3.2.2. Spatial Attention Mechanism

The Spatial Attention Module (SAM) focuses on the location information of the target, which can adjust the importance of spatial locations to more accurately enhance useful features. The operational process of the SAM is illustrated in Figure 4.
Initially, it generates two feature maps of H × W × 1 dimension by applying max pooling and average pooling operations on the channel dimension to process the input feature maps obtained from the BRA module. These feature maps, concurrently capturing maximum and average values, empower the model with a more holistic comprehension of spatial distribution. Subsequently, these feature maps undergo concatenation, which is followed by a k × k convolution layer to extract intricate spatial features and formulate a spatial attention weight map. This map is then subjected to a sigmoid normalization function, ensuring its values lie within the range of [0, 1], thus yielding the output feature map of SAM. Ultimately, this refined output is element-wise multiplied with the original feature map, restoring its original H × W × C dimensionality while emphasizing salient spatial regions. The computational formulation underlying this SAM process is outlined as follows:
M s F = σ f k × k M a x P o o l F ; A v g P o o l F
where f k × k denotes the convolution operation.

3.3. ODConv

In this work, we incorporate the ODConv after the C2F module in the head network. This particular convolution has the capability to adaptively adjust its parameters based on the input image, and it computes four distinct types of attention across all four dimensions of the kernel space—namely, the convolution kernel dimension, spatial dimension, input channel dimension, and output channel dimension—in a parallel fashion. By assigning different attention weights to each convolution kernel, this approach enhances the model’s capability to acquire input information and capture contextual cues effectively. Its definition is given by Equation (10) as follows:
y = α ω 1 α s 1 α c 1 α f 1 W 1 + + α ω n α s n α c n α f n W n x
where α ω i R denotes the attention weight in the convolution kernel dimension, α s i R k × k represents the attention weight along the spatial dimension, α c i R c i n signifies the attention weight across the input channel dimension, and   α f i R c o u t indicates the attention weight along the output channel dimension. The attention weights in these four dimensions are computed by the multi-head attention module π i x The symbol ⊙ denotes the element-wise multiplication operation along different dimensions of the kernel space.
The schematic diagram of ODConv is presented in Figure 5. Initially, the input feature map X undergoes processing through a global average pooling (GAP) layer to diminish its feature dimensions. Subsequently, it is passed through a fully connected (FC) layer followed by a ReLU activation module. The processed features are then distributed into four parallel branches, each comprising a fully connected layer and a sigmoid function, to derive four distinct types of attention weights: α ω i , α s i , α c i , α f i . These attention weights are individually multiplied with the convolution kernels, enabling adaptive adjustments to the shape and weights of the kernels based on the content of the input features. Ultimately, the refined output feature map Y is obtained.
More specifically, Figure 6 depicts the process of multiplying the four types of dimensional attention in the kernel space by the convolution kernels. This procedure involves adjusting the weights of the convolution along the dimensions of the convolution kernel, spatial dimension, input channel dimension, and output channel dimension, respectively. These four attention mechanisms complement and cooperate, enabling the model to exert different dimensional influences on the input data across various dimensions. This enhances the convolutional neural network’s adaptability to variations in input data features and provides more flexible convolutional operations for processing complex and diverse data, therefore better accommodating the intricate and ever-changing conditions of road targets.

3.4. Shape-IOU

In urban traffic-dense scenarios, the shape and size of bounding boxes for road vehicles vary significantly due to factors such as occlusion and distance. The Shape-IOU loss function not only takes into account the geometric relationship between the predicted box and the ground truth box but also comprehensively considers the shape and scale factors of the bounding boxes [33]. Therefore, to achieve more accurate bounding box regression, we adopt the Shape-IOU loss function as the loss function for bounding box regression.
As illustrated in Figure 7, the shape and size of the bounding boxes themselves will affect the IOU value. Concretely, in Figure 7a, all bounding box regression samples exhibit the same deviation, with a shape deviation of 0. Conversely, in Figure 7b, all bounding box regression samples possess the same shape deviation, specifically with a deviation of 0. Within each set of figures, it is observed that samples A and B share the same shape, while samples C and D share another identical shape. Samples A and C are of the same size, and samples B and D are also of the same size, which is larger than that of A and C. It is noteworthy that for bounding box regression samples of equal size, despite having the same deviation or shape deviation, their IOU values can still vary. Furthermore, deviations or shape deviations along the shorter side of the bounding box exert a more significant impact on the IOU value. In the case of bounding box regression samples of the same shape, the IOU values of smaller-sized samples are more prone to being influenced by the shape of the ground truth (GT) box.
Therefore, taking into account the influence of the size and shape of the bounding box on its regression, the calculation method for Shape-IOU is defined as follows:
I O U = B B g t B B g t
ω ω = 2 × ω g t s c a l e ω g t s c a l e + h g t s c a l e
h h = 2 × h g t s c a l e ω g t s c a l e + h g t s c a l e
d i s t a n c e s h a p e = h h × x c x c g t 2 c 2 + ω ω × y c y c g t 2 c 2
Ω s h a p e = t = ω , h 1 e ω t 4
ω ω = h h × ω ω g t max ω , ω g t
ω h = ω ω × h h g t max h , h g t
where s c a l e denotes the scaling factor related to the size of the targets in the dataset, ranging from 0 to 1.5, while ω , h , ω g t , and h g t signify the width and height of the predicted and the ground truth bounding boxes, respectively. ω g t s c a l e and h g t s c a l e represent ω g t and h g t adjusted by the scaling factor. x c , y c , x c g t , and y c g t , respectively, denote the center coordinates of the predicted bounding box and the ground truth bounding box. c stands for the length of the diagonal of the minimum bounding rectangle enclosing both the predicted bounding box and the ground truth bounding box. ω ω and h h , respectively, indicate the weighting coefficients in the horizontal and vertical directions, with their values being related to the shape of the ground truth bounding box. The d i s t a n c e s h a p e represents the distance loss function while Ω s h a p e denotes the shape loss function. The corresponding bounding box regression loss, denoted as L S h a p e I O U , is formulated as follows:
L S h a p e I O U = 1 I O U + d i s t a n c e s h a p e + 0.5 × Ω s h a p e

4. Experiments

4.1. Datasets

The primary dataset selected for the experimental section of this paper is the UA-DETRAC dataset, which was collected from 24 road overpasses located in Beijing and Tianjin, China. This comprehensive dataset comprises over 10 h of video footage and encompasses more than 1.21 million object bounding boxes. The videos were captured with a frame rate of 25 fps and a resolution of 960 × 540 pixels. Each image in the dataset contains multiple vehicles of different types, exhibiting varying degrees of occlusion and truncation [34]. Furthermore, the dataset is annotated with four distinct label categories, namely car, bus, vans, and others. Considering that the dataset contains multiple continuous time-series videos and the differences in vehicles and background changes between adjacent video frames are not significant, directly utilizing it for object detection model training may result in data redundancy. Therefore, to enhance the model’s generalization ability and reduce training time, one frame is extracted from every ten frames. Subsequently, the processed dataset is divided into a training set and a validation set at a ratio of 8:2, yielding 8639 images for training and 2231 images for validation. The specific distribution of each category, as depicted in Figure 8a, reveals a clear dominance of the car category with a substantial count of 72,125 instances. This is followed by a gradual decline in the number of buses, vans, and other categories, exhibiting a pronounced long-tail effect.
In addition, to validate the extensive applicability of the proposed model, this study also performs supplementary verification on the SODA10M dataset. This dataset, jointly released by Huawei Noah’s Ark Lab and Sun Yat-sen University [35], encompasses road scene images from more than 30 cities across China. It consists of 20,000 labeled images with a resolution of 1920 × 1080 pixels, primarily classified into six categories: Pedestrian, Cyclist, Car, Truck, Tram, and Tricycle. In this experiment, 10,000 images are randomly selected from the dataset and divided into training, validation, and testing sets in a ratio of 8:1:1. Figure 8b illustrates the specific quantities of six label categories, where the number of cars is the highest, whereas the number of tricycles is the lowest. This distribution characteristic aligns with the popularity and inherent features of current transportation modes.

4.2. Experimental Settings

To ensure the validity of the experiments, all experiments presented in this work were conducted in a consistent operating environment, with specific configurations detailed in Table 1. In pursuit of achieving the best training outcomes, the models were trained for 100 epochs on the UA-DETRAC dataset and for 200 epochs on the SODA10M dataset. For both datasets, mosaic data augmentation was disabled during the final 10 epochs. The remaining training parameters were kept consistent across all experiments and are provided in Table 2 for reference.

4.3. Evaluation Metrics

In vehicle detection tasks, the primary metrics used to evaluate model performance include Precision (P), Recall (R), Mean Average Precision (mAP), and Frames Per Second (FPS). In this experiment, mAP@0.5 and mAP@0.5:0.95 serve as the main reference indices for performance, with their values positively correlated with the model’s effectiveness. Specifically, mAP@0.5 refers to the average of the Average Precision (AP) for each category when the IOU is set at 0.5, while mAP@0.5:0.95 denotes the average AP for each category when the IOU threshold ranges from 0.5 to 0.95 in increments of 0.05. The corresponding computational formulas are as follows:
P = T P T P + F P
R = T P T P + F N
A P = 0 1 P R   d r
m A P = 1 N i = 1 N A P i
where TP denotes the number of correctly detected items, FP represents the number of vehicles incorrectly predicted, FN is the number of actual vehicles not detected, AP signifies the area enclosed by the PR curve, and N denotes the total number of categories in the dataset.

4.4. Impact of Different Modules on Model Performance

4.4.1. Effect Verification of BRSA Mechanism

To validate the effectiveness of the newly proposed BRSA attention mechanism, we integrated it alongside CBAM, EMA, and SE attention mechanisms at an identical location within the backbone network, precisely preceding the SPPF module. Comprehensive experiments were performed utilizing the UA-DETRAC dataset, and the resultant P, R, and mAP curves are depicted in Figure 9. The results unequivocally demonstrate that BRSA outperforms all other mechanisms across all evaluated curves.
Table 3 presents a detailed summary of the specific performance metrics observed after incorporating various attention mechanisms into the model. While the EMA and SE mechanisms exhibit faster detection speeds, they compromise on detection accuracy. Conversely, BRSA achieves an impressive mAP@0.5 of 58.8% and a noteworthy mAP@0.5:0.95 of 42.6%, outperforming all other evaluated attention mechanisms and demonstrating the most superior performance. This underscores the superior ability of our proposed attention mechanism to identify and capture salient features in complex backgrounds.

4.4.2. The Impact of the Quantity and Placement of ODConv on Model Performance

In this research, we substituted two conventional convolution layers within the head network with ODConv modules, aiming to augment the convolutional neural network’s capacity to accommodate variations in the characteristics of input data, ultimately enhancing the precision of vehicle detection. Considering that the YOLOv8 network encompasses numerous convolution modules, we undertook a series of experiments to systematically analyze the influence of varying quantities and placements of ODConv modules on model performance.
Table 4 exhibits the influence of different quantities of ODConv modules in the backbone on model performance. It is evident that as the quantity of ODConv convolutions increases, the layer of the model escalates accordingly, accompanied by an improvement in the mAP@0.5 metric. However, when all convolutions in the backbone network are replaced with ODConv, the resulting model becomes excessively complex, leading to overfitting the training data and compromising its ability to generalize to new data, ultimately resulting in a decline in accuracy. As a result, we have chosen to replace only two standard convolutions in the head network with ODConv modules.
Table 5 delineates the effects of substituting standard convolutions with ODConv modules at various positions on model performance. In Group 1, a single convolution was replaced in both the head network and the backbone network. In Group 2, two convolutions within the backbone network were substituted. Group 3 encompassed the replacement of two convolutions in the head network. Group 4 signifies a combined experimental setup that integrates the model from Group 2 with the BRSA attention mechanism. Finally, Group 5 represents a combined experimental configuration that merges the model from Group 3 with the BRSA mechanism. From the experimental results, it can be observed that when replacing the convolutions in the backbone network, the model’s P improves, but the R decreases concurrently. Although both Group 2 and Group 3 exhibit comparably impressive performance in terms of mAP@0.5, Group 5, which combines the model from Group 3 with the BRSA attention mechanism, demonstrates superior performance. This could be attributed to the redundancy of the backbone network model in Group 4, which fails to provide additional valuable feature information. Consequently, we ultimately select the experimental configuration of Group 5, where two regular convolutions in the head network are replaced by the ODConv module.

4.4.3. The Impact of Different Loss Functions on Model Performance

To evaluate the impact of different loss functions on model performance, ablation experiments were conducted using DIOU, EIOU, GIOU, SIOU, WIOU, Shape-IOU, and the baseline model. As shown in Table 6, the model’s GFLOPs remained unaffected by the substitution of loss functions. Notably, the EIOU loss function exhibited consistently poor performance across all metrics, potentially due to the need for refinement in the formulation of the Focal Loss term incorporated within EIOU. Additionally, while the SIOU loss function demonstrated an improvement in P, it suffered from a reduced R, thus compromising the overall detection effectiveness. Conversely, the model utilizing the Shape-IOU loss function achieved the highest performance on the dataset, with the highest mAP@0.5 and mAP@0.5:0.95 values, along with faster processing speeds. This robustly validates the effectiveness of incorporating the inherent attributes of bounding boxes in enhancing regression performance.

4.5. Ablation Studies

To validate the effectiveness of the various methods proposed in this paper, an ablation study was conducted, which comprised a total of 8 groups. To ensure fairness across all experimental groups, the input images, training hyperparameters, and other experimental configurations were kept consistent throughout the experimental process. The comprehensive results of this ablation study are presented in Table 7, which indicates that the baseline YOLOv8n model exhibits a rather modest detection performance, marked by an mAP@0.5 of 56.5%. By individually integrating the BRSA and ODConv, as well as modifying the loss function within the model, we observe distinct improvements in P, R, and mAP. Notably, although the inclusion of the BRSA module introduces an increase in computational complexity, it concurrently yields a substantial enhancement in mAP@0.5 by 2.3 percentage points. When these modules are paired, we see further improvements in mAP@0.5, with gains of 2.8, 2.6, and 3.8 percentage points, respectively, over the baseline YOLOv8n. Remarkably, when all three enhancements are combined, P surges by 5.3 percentage points, R increases modestly by 0.7 percentage points, and both mAP@0.5 and mAP@0.5:0.95 exhibit significant improvements of 4.7 and 4.4 percentage points, respectively. This underscores the profound enhancement in detection capabilities achieved by our model. In conclusion, the ablation study results conclusively validate the efficacy of each of the proposed modules.

4.6. Comparisons with Other Methods

Table 8 presents a comparison of evaluation metrics across various categories in the UA-DETRAC dataset before and after model improvement. As can be observed from the table, the improved model demonstrates enhancements of varying degrees across all categories, particularly with the mAP for the “other” category increasing from 29.6% to 44.3%, representing an improvement of over 40%.
To evaluate the performance of the improved model, we conducted a comparison with a range of YOLO models as well as the latest end-to-end object detection model, RT-DETR [36]. The performance of these models was assessed using various metrics, including mAP@0.5, mAP@0.5:0.95, GFLOPs, and Params. The experimental results are presented in Table 9. Under the metrics of mAP@0.5 and mAP@0.50:0.95, YOLO-BOS exhibited significantly higher performance compared to the other models, surpassing the second-best method, YOLOv5m, by 3.9 percentage points and 3.8 percentage points, respectively. Although the P metric of YOLO-BOS was slightly lower than that of the YOLOv5m, YOLOv7, and YOLOv7-CHS [37] models, the Params of these three models were substantially larger than that of YOLO-BOS. In terms of the FLOPs metric, YOLO-BOS only exceeded YOLOv5s by 2.7G, yet it achieved a 6.2 percentage point improvement in the mAP metric. Furthermore, compared to the RT-DETR model, our model demonstrates significant advantages in both accuracy and speed. The experimental results substantiate that enhancing the feature-extraction capability of the model can notably improve its detection accuracy.
To further demonstrate the generalization capability of our model, we conducted a series of comparative experiments on the SODA10M dataset. Table 10 presents a comparison of evaluation metrics across various categories in this dataset before and after model improvement. Similarly, for the “tricycle” category, which has the smallest proportion of samples, the mAP@0.5 improved by 16.1%. This indicates that our model possesses strong learning capabilities. As shown in Table 11, YOLO-BOS demonstrates significant improvements in both mAP@0.5 and mAP@0.5:0.95 metrics compared to YOLOv8n. Furthermore, it outperforms the YOLOv5s and YOLOv7 series in terms of both detection performance and model size.
For a more intuitive evaluation, we conducted a visual comparison of detection performance between our model and other detection models under varying lighting conditions and degrees of occlusion using scenes from the SODA10M dataset, as illustrated in Figure 10. The figure clearly indicates that YOLOv5s and YOLOv7 models suffer from elevated miss detection rates. Although YOLOv8n detects most targets, its performance deteriorates in complex scenarios. YOLO11 exhibits good performance across various scenarios, yet its detection efficacy in complex environments still lags behind that of YOLO-BOS, with instances of misdetection observed. This demonstrates that YOLO-BOS, with the three improvements proposed in this paper, significantly enhances feature-extraction capabilities and is less affected by factors such as lighting and occlusion, making it well-suited to various complex scenarios in urban road traffic.

5. Discussion

This paper proposes a novel YOLO-BOS vehicle-detection model to address the issues of missed detections, false detections, and reduced accuracy in vehicle detection caused by occlusion in dense scenarios. First, a novel BRSA attention mechanism is proposed, which can selectively filter features according to task requirements and utilize the SAM attention to adjust the importance of spatial positions, therefore more accurately enhancing useful features and improving the quality of feature extraction from input images. Second, the ODConv is integrated into the head network, which dynamically adjusts the size and weights of the convolution kernel across four dimensions of the kernel space, enabling better adaptation to the complex and varying conditions of road targets. Additionally, Shape-IOU is employed as the loss function instead of CIOU, which not only considers the geometric relationship between the predicted box and the ground truth box but also comprehensively takes into account the shape and scale factors of the bounding boxes, therefore enhancing the accuracy and robustness of vehicle target detection across different sizes. The feasibility of each improvement is demonstrated through ablation experiments, and comparative experiments are conducted on two different datasets to prove the generalization capability of the model. Although the proposed model has achieved notable improvements in detection accuracy, there is still considerable room for optimization in terms of computational complexity and inference speed. In our future work, we will focus on deeply optimizing the computational complexity and inference speed of the model while maintaining its detection accuracy.

Author Contributions

Conceptualization, L.Z.; methodology, L.F.; validation, X.J.; formal analysis, B.C.; investigation, X.Z.; resources, L.Z.; data curation, J.J.; writing—original draft preparation, L.F.; writing—review and editing, L.Z.; funding acquisition, L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant Nos. 61473114 and 62106068), the Henan Province Science and Technology Research Projects (Grant No. 242102211040), the Key Scientific Research Projects of Higher Education Institutions in Henan Province (Grant Nos. 24A520014 and 24B520006), and the Open Fund of Key Laboratory of Grain Information Processing and Control (Henan University of Technology), Ministry of Education (Grant Nos. KFJJ2024013 and KFJJ2023015).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, P.; Zhang, R.; Sun, S.; Gao, M.; Zheng, B.; Zhang, D.; Zhang, Y.; Carmichael, G.R.; Zhang, H. Aggravated Air Pollution and Health Burden Due to Traffic Congestion in Urban China. Atmos. Chem. Phys. 2023, 23, 2983–2996. [Google Scholar] [CrossRef]
  2. Yuan, T.; da Rocha Neto, W.; Rothenberg, C.E.; Obraczka, K.; Barakat, C.; Turletti, T. Machine Learning for Next-Generation Intelligent Transportation Systems: A survey. Trans. Emerg. Telecommun. Technol. 2022, 33, e4427. [Google Scholar] [CrossRef]
  3. Terven, J.; Córdova-Esparza, D.-M.; Romero-González, J.-A. A Comprehensive Review of Yolo Architectures in Computer Vision: From Yolov1 to Yolov8 and Yolo-NAS. Mach. Learn. Knowl. Extr. 2023, 5, 1680–1716. [Google Scholar] [CrossRef]
  4. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  5. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef]
  6. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef]
  7. He, K.; Gkioxari, G.; Dollar, P.; Girshick, R. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  8. Pang, J.; Chen, K.; Shi, J.; Feng, H.; Ouyang, W.; Lin, D. Libra R-CNN: Towards Balanced Learning for Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  9. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  10. Zhang, H.; Shao, F.; Chu, W.; Dai, J.; Li, X.; Zhang, X.; Gong, C. Faster R-CNN Based on Frame Difference and Spatiotemporal Context for Vehicle Detection. Signal Image Video Process. 2024, 18, 7013–7027. [Google Scholar] [CrossRef]
  11. Xu, Y.; Chu, K.; Zhang, J. Nighttime Vehicle Detection Algorithm Based on Improved Faster-RCNN. IEEE Access 2023, 12, 19299–19306. [Google Scholar] [CrossRef]
  12. Othmani, M.J.M.T. A Vehicle Detection and Tracking Method for Traffic Video Based on Faster R-CNN. Multimed. Tools Appl. 2022, 81, 28347–28365. [Google Scholar] [CrossRef]
  13. Guo, D.; Wang, Y.; Zhu, S.; Li, X. A Vehicle Detection Method Based on An Improved U-YOLO Network for High-Resolution Remote-Sensing Images. Sustainability 2023, 15, 10397. [Google Scholar] [CrossRef]
  14. Zhang, Q.; Han, Z.; Zhang, Y. Vehicle Target Detection Method Based on Improved YOLO V3 Network Model. PeerJ Comput. Sci 2023, 9, e1673. [Google Scholar] [CrossRef]
  15. Zhang, Y.; Zhou, A.; Zhao, F.; Wu, H. A Lightweight Vehicle-pedestrian Detection Algorithm Based on Attention Mechanism in Traffic Scenarios. Sensors 2022, 22, 8480. [Google Scholar] [CrossRef]
  16. Wang, P.; Wang, X.; Liu, Y.; Song, J. Research on Road Object Detection Model Based on YOLOv4 of Autonomous Vehicle. IEEE Access 2024, 12, 8198–8206. [Google Scholar] [CrossRef]
  17. Chen, C.; Wang, C.; Liu, B.; He, C.; Cong, L.; Wan, S. Edge Intelligence Empowered Vehicle Detection and Image Segmentation for Autonomous Vehicles. IEEE Trans. Intell. Transp. Syst. 2023, 24, 13023–13034. [Google Scholar] [CrossRef]
  18. Miao, J.; Gong, S.; Deng, Y.; Liang, H.; Li, J.; Qi, H.; Zhang, M. YOLO-VSF: An Improved YOLO Model by Incorporating Attention Mechanism for Object Detection in Traffic Scenes. J. Shanghai Jiao Tong Univ. (Sci.) 2024, 1–14. [Google Scholar] [CrossRef]
  19. Li, H.; Zhuang, X.; Bao, S.; Chen, J.; Yang, C. SCD-YOLO: A Lightweight Vehicle Target Detection Method Based on Improved YOLOv5n. J. Electron. Imaging 2024, 33, 023041. [Google Scholar] [CrossRef]
  20. Wang, Y.; Xu, S.; Wang, P.; Li, K.; Song, Z.; Zheng, Q.; Li, Y.; He, Q. Lightweight Vehicle Detection Based on Improved YOLOv5s. Sensors 2024, 24, 1182. [Google Scholar] [CrossRef]
  21. Wang, H.; Xu, Y.; He, Y.; Cai, Y.; Chen, L.; Li, Y.; Sotelo, M.A.; Li, Z. YOLOv5-Fog: A Multiobjective Visual Detection Algorithm for Fog Driving Scenes Based on Improved YOLOv5. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  22. Ren, J.; Yang, J.; Zhang, W.; Cai, K. RBS-YOLO: A Vehicle Detection Algorithm Based on Multi-scale Feature Extraction. Signal Image Video Process. 2024, 18, 3421–3430. [Google Scholar] [CrossRef]
  23. Li, Y.; Zhang, M.; Zhang, C.; Liang, H.; Li, P.; Zhang, W. YOLO-CCS: Vehicle Detection Algorithm Based on Coordinate Attention Mechanism. Digit. Signal Process. 2024, 153, 104632. [Google Scholar] [CrossRef]
  24. Wang, Z.; Men, S.; Bai, Y.; Yuan, Y.; Wang, J.; Wang, K.; Zhang, L. Improved Small Object Detection Algorithm CRL-YOLOv5. Sensors 2024, 24, 6437. [Google Scholar] [CrossRef] [PubMed]
  25. He, Q.; Xu, A.; Ye, Z.; Zhou, W.; Cai, T. Object Detection Based on Lightweight YOLOX for Autonomous Driving. Sensors 2023, 23, 7596. [Google Scholar] [CrossRef]
  26. Gao, X.; Yu, A.; Tan, J.; Gao, X.; Zeng, X.; Wu, C. GSD-YOLOX: Lightweight and More Accurate Object Detection Models. J. Vis. Commun. Image Represent. 2024, 98, 104009. [Google Scholar] [CrossRef]
  27. Wang, W.; Chen, J.; Huang, Z.; Yuan, H.; Li, P.; Jiang, X.; Wang, X.; Zhong, C.; Lin, Q. Improved YOLOv7-Based Algorithm for Detecting Foreign Objects on the Roof of a Subway Vehicle. Sensors 2023, 23, 9440. [Google Scholar] [CrossRef]
  28. Zhang, Y.; Sun, Y.; Wang, Z.; Jiang, Y. YOLOv7-RAR for Urban Vehicle Detection. Sensors 2023, 23, 1801. [Google Scholar] [CrossRef] [PubMed]
  29. Li, M.; Liu, X.; Chen, S.; Yang, L.; Du, Q.; Han, Z.; Wang, J. MST-YOLO: Small Object Detection Model for Autonomous Driving. Sensors 2024, 24, 7347. [Google Scholar] [CrossRef] [PubMed]
  30. Liu, X.; Wang, Y.; Yu, D.; Yuan, Z. YOLOv8-FDD: A Real-Time Vehicle Detection Method based on Improved YOLOv8. IEEE Access 2024, 12, 136280–136296. [Google Scholar] [CrossRef]
  31. Liu, J.; Yang, D.; Feng, T.; Fu, J. MDFD2-DETR: A Real-Time Complex Road Object Detection Model Based on Multi-Domain Feature Decomposition and De-Redundancy. IEEE Trans. Intell. Veh. 2024. [Google Scholar] [CrossRef]
  32. Jin, M.; Zhang, J. Research on Microscale Vehicle Logo Detection Based on Real-Time DEtection TRansformer (RT-DETR). Sensors 2024, 24, 6987. [Google Scholar] [CrossRef]
  33. Zhang, H.; Zhang, S. Shape-IoU: More Accurate Metric Considering Bounding Box Shape and Scale. arXiv 2023, arXiv:2312.17663. [Google Scholar]
  34. Wen, L.; Du, D.; Cai, Z.; Lei, Z.; Chang, M.-C.; Qi, H.; Lim, J.; Yang, M.-H.; Lyu, S. UA-DETRAC: A New Benchmark and Protocol for Multi-object Detection and Tracking. Comput. Vis. Image Underst. 2020, 193, 102907. [Google Scholar] [CrossRef]
  35. Han, J.; Liang, X.; Xu, H.; Chen, K.; Hong, L.; Mao, J.; Ye, C.; Zhang, W.; Li, Z.; Liang, X. SODA10M: A Large-Scale 2D Self/Semi-Supervised Object Detection Dataset for Autonomous Driving. arXiv 2021, arXiv:2106.11118. [Google Scholar]
  36. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. Detrs Beat Yolos on Real-Time Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024. [Google Scholar]
  37. Zhao, L.; Yun, Q.; Yuan, F.; Ren, X.; Jin, J.; Zhu, X. YOLOv7-CHS: An Emerging Model for Underwater Object Detection. J. Mar. Sci. Eng. 2023, 11, 1949. [Google Scholar] [CrossRef]
Figure 1. Structure diagram of the YOLO-BOS network.
Figure 1. Structure diagram of the YOLO-BOS network.
Sensors 24 08126 g001
Figure 2. Structure diagram of the BRSA module.
Figure 2. Structure diagram of the BRSA module.
Sensors 24 08126 g002
Figure 3. Structure diagram of the BRA module.
Figure 3. Structure diagram of the BRA module.
Sensors 24 08126 g003
Figure 4. Structure diagram of the SAM module.
Figure 4. Structure diagram of the SAM module.
Sensors 24 08126 g004
Figure 5. The schematic diagram of ODConv.
Figure 5. The schematic diagram of ODConv.
Sensors 24 08126 g005
Figure 6. The four types of dimensional attention in the ODConv.
Figure 6. The four types of dimensional attention in the ODConv.
Sensors 24 08126 g006
Figure 7. An illustration of regression affected by the shape and size of the bounding box themselves. (a) All bounding box regression samples possess the same deviation; (b) All bounding box regression samples possess the same shape deviation.
Figure 7. An illustration of regression affected by the shape and size of the bounding box themselves. (a) All bounding box regression samples possess the same deviation; (b) All bounding box regression samples possess the same shape deviation.
Sensors 24 08126 g007
Figure 8. The distribution of instances for the datasets. (a) UA-DETRAC; (b) SODA10M.
Figure 8. The distribution of instances for the datasets. (a) UA-DETRAC; (b) SODA10M.
Sensors 24 08126 g008
Figure 9. Experimental results of the comparison of different attention mechanisms.
Figure 9. Experimental results of the comparison of different attention mechanisms.
Sensors 24 08126 g009
Figure 10. Detection results on the SODA10M dataset. The detection performance of different detection models on the SODA10M dataset.
Figure 10. Detection results on the SODA10M dataset. The detection performance of different detection models on the SODA10M dataset.
Sensors 24 08126 g010
Table 1. Configuration settings for the experimental environment.
Table 1. Configuration settings for the experimental environment.
TypeConfiguration
CPUE5–2678 v3@2.50 GHz
GPUNVIDIA GeForce RTX 2080 Ti
Operating systemWindows 10
RAM128 GB
CUDA11.6
Programming languagePython 3.8
Deep learning architecturePytorch 1.13.1
Table 2. Configuration of parameters in the experimental setup.
Table 2. Configuration of parameters in the experimental setup.
ParametersValue
Image size640 × 640
Batch size32
OptimizerSGD
Initial learning rate0.01
Momentum0.937
Weight decay 5 × 10 4
Workers8
Table 3. Experimental results on different attention mechanisms.
Table 3. Experimental results on different attention mechanisms.
ModulemAP@0.5 (%)mAP@0.5:0.95 (%)LayersFPS
CBAM55.841.1180217.39
EMA57.341.3176285.71
SE56.941.5175285.71
BRSA58.842.6183263.16
Table 4. The impact of the quantity of ODConv on model performance.
Table 4. The impact of the quantity of ODConv on model performance.
QuantityP (%)R (%)mAP@0.5 (%)mAP@0.5:0.95 (%)Layers
059.257.756.540.6168
161.855.356.840.4183
257.457.457.641.6193
360.352.854.239.9204
Table 5. The impact of the placement of ODConv on model performance.
Table 5. The impact of the placement of ODConv on model performance.
GroupP (%)R (%)mAP@0.5 (%)mAP@0.5:0.95 (%)Layers
164.253.656.041.0194
264.957.157.541.8194
357.457.457.641.6193
463.557.158.643.0209
566.854.359.343.5208
Table 6. The impact of different loss functions on model performance.
Table 6. The impact of different loss functions on model performance.
Loss FunctionFLOPs (G)P (%)R (%)mAP@0.5 (%)mAP@0.5:0.95 (%)
Baseline8.159.257.756.540.6
DIOU8.160.955.656.040.6
EIOU8.155.354.154.839.5
GIOU8.165.551.956.740.6
SIOU8.167.151.757.040.6
WIOU8.155.659.356.040.2
Shape-IOU8.161.656.257.640.8
Table 7. Ablation experiments. In columns 2 through 4, the “√” sign indicates that the module in line 1 was added to the YOLOv8 network.
Table 7. Ablation experiments. In columns 2 through 4, the “√” sign indicates that the module in line 1 was added to the YOLOv8 network.
GroupBRSAODConvShape-IOUP (%)R (%)mAP@0.5 (%)mAP@0.5:0.95 (%)FLOPs (G)FPS
1 59.257.756.540.68.1312.50
2 65.655.858.842.618.7263.16
3 57.457.457.641.68.1277.78
4 61.656.257.640.88.1303.03
5 66.854.359.3 (+2.8)43.518.5285.71
6 60.661.160.3 (+3.8)43.718.7285.71
7 59.340.959.1 (+2.6)40.98.1277.78
864.558.461.2 (+4.7)45.018.5263.16
Table 8. Model detection results for different categories in the UA-DETRAC dataset.
Table 8. Model detection results for different categories in the UA-DETRAC dataset.
CategoryYOLOv8nOurs
P (%)R (%)mAP@0.5 (%)mAP@0.5:0.95 (%)P (%)R (%)mAP@0.5 (%)mAP@0.5:0.95 (%)
car70.570.774.153.268.871.774.353.0
bus71.172.072.352.874.572.974.955.8
vans53.252.450.138.351.952.951.339.1
others41.935.629.618.162.835.844.332.1
all59.257.756.540.664.558.461.245.0
Table 9. Comparisons of different object detectors on the UA-DETRAC dataset.
Table 9. Comparisons of different object detectors on the UA-DETRAC dataset.
ModelP (%)R (%)mAP@0.5 (%)mAP@0.5:0.95 (%)Params (M)FLOPs (G)
YOLOv5s59.355.555.039.27.0215.8
YOLOv5m70.653.357.341.220.8747.9
YOLOv782.051.953.441.736.50103.2
YOLOv7-CHS77.1448.152.636.033.5640.3
YOLOv8n59.257.756.540.63.008.1
YOLO1161.851.755.940.22.586.3
RT-DETR51.549.945.932.732.8108.0
Ours64.558.461.245.03.2718.5
Table 10. Model detection results for different categories in the SODA10M dataset.
Table 10. Model detection results for different categories in the SODA10M dataset.
CategoryYOLOv8nOurs
P (%)R (%)mAP@0.5 (%)mAP@0.5:0.95 (%)P (%)R (%)mAP@0.5 (%)mAP@0.5:0.95 (%)
pedestrian69.228.839.518.557.032.739.117.7
cyclist73.053.663.736.869.655.961.934.8
car85.176.884.361.878.378.483.059.3
truck77.766.374.152.176.370.176.051.8
tram71.663.969.552.172.064.970.452.8
tricycle60.231.234.325.169.440.150.430.4
all72.853.460.940.170.457.063.542.1
Table 11. Comparisons of different object detectors on the SODA10M dataset.
Table 11. Comparisons of different object detectors on the SODA10M dataset.
ModelP (%)R (%)mAP@0.5 (%)mAP@0.5:0.95 (%)Params (M)FLOPs (G)
YOLOv5s79.348.656.635.77.0315.8
YOLOv778.748.154.835.336.5103.2
YOLOv7-tiny65.935.340.325.86.0213.1
YOLOv8n70.853.460.940.13.08.1
YOLO1175.653.862.042.22.586.3
RT-DETR73.350.555.335.632.0108.0
Ours70.457.063.542.13.2718.5
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, L.; Fu, L.; Jia, X.; Cui, B.; Zhu, X.; Jin, J. YOLO-BOS: An Emerging Approach for Vehicle Detection with a Novel BRSA Mechanism. Sensors 2024, 24, 8126. https://doi.org/10.3390/s24248126

AMA Style

Zhao L, Fu L, Jia X, Cui B, Zhu X, Jin J. YOLO-BOS: An Emerging Approach for Vehicle Detection with a Novel BRSA Mechanism. Sensors. 2024; 24(24):8126. https://doi.org/10.3390/s24248126

Chicago/Turabian Style

Zhao, Liang, Lulu Fu, Xin Jia, Beibei Cui, Xianchao Zhu, and Junwei Jin. 2024. "YOLO-BOS: An Emerging Approach for Vehicle Detection with a Novel BRSA Mechanism" Sensors 24, no. 24: 8126. https://doi.org/10.3390/s24248126

APA Style

Zhao, L., Fu, L., Jia, X., Cui, B., Zhu, X., & Jin, J. (2024). YOLO-BOS: An Emerging Approach for Vehicle Detection with a Novel BRSA Mechanism. Sensors, 24(24), 8126. https://doi.org/10.3390/s24248126

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop