Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (27)

Search Parameters:
Keywords = UA_DETRAC

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
22 pages, 12020 KiB  
Article
TFF-Net: A Feature Fusion Graph Neural Network-Based Vehicle Type Recognition Approach for Low-Light Conditions
by Huizhi Xu, Wenting Tan, Yamei Li and Yue Tian
Sensors 2025, 25(12), 3613; https://doi.org/10.3390/s25123613 - 9 Jun 2025
Viewed by 661
Abstract
Accurate vehicle type recognition in low-light environments remains a critical challenge for intelligent transportation systems (ITSs). To address the performance degradation caused by insufficient lighting, complex backgrounds, and light interference, this paper proposes a Twin-Stream Feature Fusion Graph Neural Network (TFF-Net) model. The [...] Read more.
Accurate vehicle type recognition in low-light environments remains a critical challenge for intelligent transportation systems (ITSs). To address the performance degradation caused by insufficient lighting, complex backgrounds, and light interference, this paper proposes a Twin-Stream Feature Fusion Graph Neural Network (TFF-Net) model. The model employs multi-scale convolutional operations combined with an Efficient Channel Attention (ECA) module to extract discriminative local features, while independent convolutional layers capture hierarchical global representations. These features are mapped as nodes to construct fully connected graph structures. Hybrid graph neural networks (GNNs) process the graph structures and model spatial dependencies and semantic associations. TFF-Net enhances the representation of features by fusing local details and global context information from the output of GNNs. To further improve its robustness, we propose an Adaptive Weighted Fusion-Bagging (AWF-Bagging) algorithm, which dynamically assigns weights to base classifiers based on their F1 scores. TFF-Net also includes dynamic feature weighting and label smoothing techniques for solving the category imbalance problem. Finally, the proposed TFF-Net is integrated into YOLOv11n (a lightweight real-time object detector) with an improved adaptive loss function. For experimental validation in low-light scenarios, we constructed the low-light vehicle dataset VDD-Light based on the public dataset UA-DETRAC. Experimental results demonstrate that our model achieves 2.6% and 2.2% improvements in mAP50 and mAP50-95 metrics over the baseline model. Compared to mainstream models and methods, the proposed model shows excellent performance and practical deployment potential. Full article
(This article belongs to the Section Vehicular Sensing)
Show Figures

Figure 1

30 pages, 20203 KiB  
Article
Multi-Feature Fusion Method Based on Adaptive Dilation Convolution for Small-Object Detection
by Lin Cao, Jin Wu, Zongmin Zhao, Chong Fu and Dongfeng Wang
Sensors 2025, 25(10), 3182; https://doi.org/10.3390/s25103182 - 18 May 2025
Cited by 1 | Viewed by 581
Abstract
This paper addresses the challenge of small-object detection in traffic surveillance by proposing a hybrid network architecture that combines attention mechanisms with convolutional layers. The network introduces an innovative attention mechanism into the YOLOv8 backbone, which effectively enhances the detection accuracy and robustness [...] Read more.
This paper addresses the challenge of small-object detection in traffic surveillance by proposing a hybrid network architecture that combines attention mechanisms with convolutional layers. The network introduces an innovative attention mechanism into the YOLOv8 backbone, which effectively enhances the detection accuracy and robustness of small objects through fine-grained and coarse-grained attention routing on feature maps. During the feature fusion stage, we employ adaptive dilated convolution, which dynamically adjusts the dilation rate spatially based on frequency components. This adaptive convolution kernel helps preserve the details of small objects while strengthening their feature representation. It also expands the receptive field, which is beneficial for capturing contextual information and the overall features of small objects. Our method demonstrates an improvement in Average Precision (AP) by 1% on the UA-DETRAC-test dataset and 3% on the VisDrone-test dataset when compared to state-of-the-art methods. The experiments indicate that the new architecture achieves significant performance improvements across various evaluation metrics. To fully leverage the potential of our approach, we conducted extended research on radar–camera systems. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

19 pages, 8036 KiB  
Article
Research on Vehicle Target Detection Method Based on Improved YOLOv8
by Mengchen Zhang and Zhenyou Zhang
Appl. Sci. 2025, 15(10), 5546; https://doi.org/10.3390/app15105546 - 15 May 2025
Viewed by 662
Abstract
To improve the performance of vehicle target detection in complex traffic environments and solve the problem that it is difficult to make a lightweight detection model, this paper proposes a lightweight vehicle detection model based on enhanced You Only Look Once v8. This [...] Read more.
To improve the performance of vehicle target detection in complex traffic environments and solve the problem that it is difficult to make a lightweight detection model, this paper proposes a lightweight vehicle detection model based on enhanced You Only Look Once v8. This method improves the feature extraction aggregation network by introducing an Adaptive Downsampling module, which can dynamically adjust the downsampling method, thereby increasing the model’s attention to key features, especially for small objects and occluded objects, while maintaining a lightweight structure, effectively reducing the model complexity while improving detection accuracy. A Lightweight Shared Convolution Detection Head was designed. By designing a shared convolution layer through group normalization, the detection head of the original model was improved, which can reduce redundant calculations and parameters and enhance the ability of global information fusion between feature maps, thereby achieving the purpose of improving computational efficiency. When tested in the KITTI 2D and UA-DETRAC datasets, the mAP of the proposed model was improved by 1.1% and 2.0%, respectively, the FPS was improved by 12% and 11%, respectively, the number of parameters was reduced by 33%, and the FLOPs were reduced by 28%. Full article
(This article belongs to the Special Issue AI in Object Detection)
Show Figures

Figure 1

19 pages, 29576 KiB  
Article
Vehicle Detection in Videos Leveraging Multi-Scale Feature and Memory Information
by Yanni Yang and Shengnan Lu
Electronics 2025, 14(10), 2009; https://doi.org/10.3390/electronics14102009 - 15 May 2025
Viewed by 457
Abstract
Vehicle detection in videos is a critical task in traffic monitoring. Existing vehicle detection tasks commonly use static detectors. Since video frames are processed as discrete static images, static detectors neglect the temporal information of vehicles when detecting vehicles in videos, leading to [...] Read more.
Vehicle detection in videos is a critical task in traffic monitoring. Existing vehicle detection tasks commonly use static detectors. Since video frames are processed as discrete static images, static detectors neglect the temporal information of vehicles when detecting vehicles in videos, leading to a reduction in detection accuracy. To address the above shortcoming, this paper improves the detection performance by introducing a video vehicle detection method that combines multi-scale features with memory information. We design a Multi-scale Feature Generation Network (MFGN) to improve the detector’s self-adaptation ability to vehicle scales. MFGN generates features with two scales and predefines multi-scale anchors for each feature scale. Based on MFGN, we propose a Memory-based Multi-scale Feature Aggregation Network (MMFAN), which aggregates historical features with current features through two parallel memory networks. The multi-scale feature and memory based method enhances the features of each frame in two perspectives, thus enhancing the vehicle detection accuracy. On the commonly adopted vehicle detection dataset UA-DETRAC, the mAP of our method is 7.4% higher compared to its static detector. The proposed approach is further validated on the well-known ImageNet VID benchmark. It demonstrates comparable performance with the memory-driven state-of-the-art frameworks. Full article
Show Figures

Figure 1

20 pages, 5228 KiB  
Article
YOLO-BOS: An Emerging Approach for Vehicle Detection with a Novel BRSA Mechanism
by Liang Zhao, Lulu Fu, Xin Jia, Beibei Cui, Xianchao Zhu and Junwei Jin
Sensors 2024, 24(24), 8126; https://doi.org/10.3390/s24248126 - 19 Dec 2024
Cited by 4 | Viewed by 1131
Abstract
In intelligent transportation systems, accurate vehicle target recognition within road scenarios is crucial for achieving intelligent traffic management. Addressing the challenges posed by complex environments and severe vehicle occlusion in such scenarios, this paper proposes a novel vehicle-detection method, YOLO-BOS. First, to bolster [...] Read more.
In intelligent transportation systems, accurate vehicle target recognition within road scenarios is crucial for achieving intelligent traffic management. Addressing the challenges posed by complex environments and severe vehicle occlusion in such scenarios, this paper proposes a novel vehicle-detection method, YOLO-BOS. First, to bolster the feature-extraction capabilities of the backbone network, we propose a novel Bi-level Routing Spatial Attention (BRSA) mechanism, which selectively filters features based on task requirements and adjusts the importance of spatial locations to more accurately enhance relevant features. Second, we incorporate Omni-directional Dynamic Convolution (ODConv) into the head network, which is capable of simultaneously learning complementary attention across the four dimensions of the kernel space, therefore facilitating the capture of multifaceted features from the input data. Lastly, we introduce Shape-IOU, a new loss function that significantly enhances the accuracy and robustness of detection results for vehicles of varying sizes. Experimental evaluations conducted on the UA-DETRAC dataset demonstrate that our model achieves improvements of 4.7 and 4.4 percentage points in mAP@0.5 and mAP@0.5:0.95, respectively, compared to the baseline model. Furthermore, comparative experiments on the SODA10M dataset corroborate the superiority of our method in terms of precision and accuracy. Full article
(This article belongs to the Section Vehicular Sensing)
Show Figures

Figure 1

24 pages, 4899 KiB  
Article
Enhancing YOLOv8’s Performance in Complex Traffic Scenarios: Optimization Design for Handling Long-Distance Dependencies and Complex Feature Relationships
by Bingyu Li, Qiao Meng, Xin Li, Zhijie Wang, Xin Liu and Siyuan Kong
Electronics 2024, 13(22), 4411; https://doi.org/10.3390/electronics13224411 - 11 Nov 2024
Cited by 2 | Viewed by 1868
Abstract
In recent years, the field of deep learning and computer vision has increasingly focused on the problem of vehicle target detection, becoming the forefront of many technological innovations. YOLOv8, as an efficient vehicle target detection model, has achieved good results in many scenarios. [...] Read more.
In recent years, the field of deep learning and computer vision has increasingly focused on the problem of vehicle target detection, becoming the forefront of many technological innovations. YOLOv8, as an efficient vehicle target detection model, has achieved good results in many scenarios. However, when faced with complex traffic scenarios, such as occluded targets, small target detection, changes in lighting, and variable weather conditions, YOLOv8 still has insufficient detection accuracy and robustness. To address these issues, this paper delves into the optimization strategies of YOLOv8 in the field of vehicle target detection, focusing on the EMA module in the backbone part and replacing the original SPPF module with focal modulation technology, all of which effectively improved the model’s performance. At the same time, modifications to the head part were approached with caution to avoid unnecessary interference with the original design. The experiment used the UA-DETRAC dataset, which contains a variety of traffic scenarios, a rich variety of vehicle types, and complex dynamic environments, making it suitable for evaluating and validating the performance of traffic monitoring systems. The 5-fold cross-validation method was used to ensure the reliability and comprehensiveness of the evaluation results. The final results showed that the improved model’s precision rate increased from 0.859 to 0.961, the recall rate from 0.83 to 0.908, and the mAP50 from 0.881 to 0.962. Meanwhile, the optimized YOLOv8 model demonstrated strong robustness in terms of detection accuracy and the ability to adapt to complex environments. Full article
(This article belongs to the Special Issue Applications of Artificial Intelligence in Image and Video Processing)
Show Figures

Figure 1

16 pages, 4399 KiB  
Article
Lightweight Vehicle Detection Based on Mamba_ViT
by Ze Song, Yuhai Wang, Shuobo Xu, Peng Wang and Lele Liu
Sensors 2024, 24(22), 7138; https://doi.org/10.3390/s24227138 - 6 Nov 2024
Viewed by 1275
Abstract
Vehicle detection algorithms are essential for intelligent traffic management and autonomous driving systems. Current vehicle detection algorithms largely rely on deep learning techniques, enabling the automatic extraction of vehicle image features through convolutional neural networks (CNNs). However, in real traffic scenarios, relying only [...] Read more.
Vehicle detection algorithms are essential for intelligent traffic management and autonomous driving systems. Current vehicle detection algorithms largely rely on deep learning techniques, enabling the automatic extraction of vehicle image features through convolutional neural networks (CNNs). However, in real traffic scenarios, relying only on a single feature extraction unit makes it difficult to fully understand the vehicle information in the traffic scenario, thus affecting the vehicle detection effect. To address this issue, we propose a lightweight vehicle detection algorithm based on Mamba_ViT. First, we introduce a new feature extraction architecture (Mamba_ViT) that separates shallow and deep features and processes them independently to obtain a more complete contextual representation, ensuring comprehensive and accurate feature extraction. Additionally, a multi-scale feature fusion mechanism is employed to enhance the integration of shallow and deep features, leading to the development of a vehicle detection algorithm named Mamba_ViT_YOLO. The experimental results on the UA-DETRAC dataset show that our proposed algorithm improves mAP@50 by 3.2% compared to the latest YOLOv8 algorithm, while using only 60% of the model parameters. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

19 pages, 4078 KiB  
Article
A Robust Multi-Camera Vehicle Tracking Algorithm in Highway Scenarios Using Deep Learning
by Menghao Li, Miao Liu, Weiwei Zhang, Wenfeng Guo, Enqing Chen and Cheng Zhang
Appl. Sci. 2024, 14(16), 7071; https://doi.org/10.3390/app14167071 - 12 Aug 2024
Cited by 2 | Viewed by 2332
Abstract
In intelligent traffic monitoring systems, the significant distance between cameras and their non-overlapping fields of view leads to several issues. These include incomplete tracking results from individual cameras, difficulty in matching targets across multiple cameras, and the complexity of inferring the global trajectory [...] Read more.
In intelligent traffic monitoring systems, the significant distance between cameras and their non-overlapping fields of view leads to several issues. These include incomplete tracking results from individual cameras, difficulty in matching targets across multiple cameras, and the complexity of inferring the global trajectory of a target. In response to the challenges above, a deep learning-based vehicle tracking algorithm called FairMOT-MCVT is proposed. This algorithm con-siders the vehicles’ characteristics as rigid targets from a roadside perspective. Firstly, a Block-Efficient module is designed to enhance the network’s ability to capture and characterize image features across different layers by integrating a multi-branch structure and depth-separable convolutions. Secondly, the Multi-scale Dilated Attention (MSDA) module is introduced to improve the feature extraction capability and computational efficiency by combining multi-scale feature fusion and attention mechanisms. Finally, a joint loss function is crafted to better distinguish between vehicles with similar appearances by combining the trajectory smoothing loss and velocity consistency loss, thereby considering both position and velocity continuity during the optimization process. The proposed method was evaluated on the public UA-DETRAC dataset, which comprises 1210 video sequences and over 140,000 frames captured under various weather and lighting conditions. The experimental results demonstrate that the FairMOT-MCVT algorithm significantly enhances multi-target tracking accuracy (MOTA) to 79.0, IDF1 to 84.5, and FPS to 29.03, surpassing the performance of previous algorithms. Additionally, this algorithm expands the detection range and reduces the deployment cost of roadside equipment, effectively meeting the practical application requirements. Full article
(This article belongs to the Special Issue Unmanned Vehicle and Industrial Sensors for Internet of Everything)
Show Figures

Figure 1

18 pages, 3242 KiB  
Article
Multi-Object Vehicle Detection and Tracking Algorithm Based on Improved YOLOv8 and ByteTrack
by Longxiang You, Yajun Chen, Ci Xiao, Chaoyue Sun and Rongzhen Li
Electronics 2024, 13(15), 3033; https://doi.org/10.3390/electronics13153033 - 1 Aug 2024
Cited by 8 | Viewed by 6143
Abstract
Vehicle detection and tracking technology plays a crucial role in Intelligent Transportation Systems. However, due to factors such as complex scenarios, diverse scales, and occlusions, issues like false detections, missed detections, and identity switches frequently occur. To address these problems, this paper proposes [...] Read more.
Vehicle detection and tracking technology plays a crucial role in Intelligent Transportation Systems. However, due to factors such as complex scenarios, diverse scales, and occlusions, issues like false detections, missed detections, and identity switches frequently occur. To address these problems, this paper proposes a multi-object vehicle detection and tracking algorithm based on CDS-YOLOv8 and improved ByteTrack. For vehicle detection, the Context-Guided (CG) module is introduced during the downsampling process to enhance feature extraction capabilities in complex scenarios. The Dilated Reparam Block (DRB) is reconstructed to tackle multi-scale issues, and Soft-NMS replaces the traditional NMS to improve performance in densely populated vehicle scenarios. For vehicle tracking, the state vector and covariance matrix of the Kalman filter are improved to better handle the nonlinear movement of vehicles, and Gaussian Smoothed Interpolation (GSI) is introduced to fill in trajectory gaps caused by detection misses. Experiments conducted on the UA-DETRAC dataset show that the improved algorithm increases detection performance, with mAP@0.5 and mAP@0.5:0.95 improving by 9% and 8.8%, respectively. In terms of tracking performance, mMOTA improves by 6.7%. Additionally, comparative experiments with mainstream detection and two-stage tracking algorithms demonstrate the superior performance of the proposed algorithm. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

20 pages, 3178 KiB  
Article
Deep Efficient Data Association for Multi-Object Tracking: Augmented with SSIM-Based Ambiguity Elimination
by Aswathy Prasannakumar and Deepak Mishra
J. Imaging 2024, 10(7), 171; https://doi.org/10.3390/jimaging10070171 - 16 Jul 2024
Cited by 4 | Viewed by 2156
Abstract
Recently, to address the multiple object tracking (MOT) problem, we harnessed the power of deep learning-based methods. The tracking-by-detection approach to multiple object tracking (MOT) involves two primary steps: object detection and data association. In the first step, objects of interest are detected [...] Read more.
Recently, to address the multiple object tracking (MOT) problem, we harnessed the power of deep learning-based methods. The tracking-by-detection approach to multiple object tracking (MOT) involves two primary steps: object detection and data association. In the first step, objects of interest are detected in each frame of a video. The second step establishes the correspondence between these detected objects across different frames to track their trajectories. This paper proposes an efficient and unified data association method that utilizes a deep feature association network (deepFAN) to learn the associations. Additionally, the Structural Similarity Index Metric (SSIM) is employed to address uncertainties in the data association, complementing the deep feature association network. These combined association computations effectively link the current detections with the previous tracks, enhancing the overall tracking performance. To evaluate the efficiency of the proposed MOT framework, we conducted a comprehensive analysis of the popular MOT datasets, such as the MOT challenge and UA-DETRAC. The results showed that our technique performed substantially better than the current state-of-the-art methods in terms of standard MOT metrics. Full article
(This article belongs to the Section Computer Vision and Pattern Recognition)
Show Figures

Figure 1

16 pages, 5634 KiB  
Article
HeMoDU: High-Efficiency Multi-Object Detection Algorithm for Unmanned Aerial Vehicles on Urban Roads
by Hanyi Shi, Ningzhi Wang, Xinyao Xu, Yue Qian, Lingbin Zeng and Yi Zhu
Sensors 2024, 24(13), 4045; https://doi.org/10.3390/s24134045 - 21 Jun 2024
Cited by 4 | Viewed by 1857
Abstract
Unmanned aerial vehicle (UAV)-based object detection methods are widely used in traffic detection due to their high flexibility and extensive coverage. In recent years, with the increasing complexity of the urban road environment, UAV object detection algorithms based on deep learning have gradually [...] Read more.
Unmanned aerial vehicle (UAV)-based object detection methods are widely used in traffic detection due to their high flexibility and extensive coverage. In recent years, with the increasing complexity of the urban road environment, UAV object detection algorithms based on deep learning have gradually become a research hotspot. However, how to further improve algorithmic efficiency in response to the numerous and rapidly changing road elements, and thus achieve high-speed and accurate road object detection, remains a challenging issue. Given this context, this paper proposes the high-efficiency multi-object detection algorithm for UAVs (HeMoDU). HeMoDU reconstructs a state-of-the-art, deep-learning-based object detection model and optimizes several aspects to improve computational efficiency and detection accuracy. To validate the performance of HeMoDU in urban road environments, this paper uses the public urban road datasets VisDrone2019 and UA-DETRAC for evaluation. The experimental results show that the HeMoDU model effectively improves the speed and accuracy of UAV object detection. Full article
(This article belongs to the Special Issue 6G and Blockchain for Advanced Future Applications)
Show Figures

Figure 1

19 pages, 2999 KiB  
Article
Novel Deep Learning Domain Adaptation Approach for Object Detection Using Semi-Self Building Dataset and Modified YOLOv4
by Ahmed Gomaa and Ahmad Abdalrazik
World Electr. Veh. J. 2024, 15(6), 255; https://doi.org/10.3390/wevj15060255 - 12 Jun 2024
Cited by 50 | Viewed by 3756
Abstract
Moving object detection is a vital research area that plays an essential role in intelligent transportation systems (ITSs) and various applications in computer vision. Recently, researchers have utilized convolutional neural networks (CNNs) to develop new techniques in object detection and recognition. However, with [...] Read more.
Moving object detection is a vital research area that plays an essential role in intelligent transportation systems (ITSs) and various applications in computer vision. Recently, researchers have utilized convolutional neural networks (CNNs) to develop new techniques in object detection and recognition. However, with the increasing number of machine learning strategies used for object detection, there has been a growing need for large datasets with accurate ground truth used for the training, usually demanding their manual labeling. Moreover, most of these deep strategies are supervised and only applicable for specific scenes with large computational resources needed. Alternatively, other object detection techniques such as classical background subtraction need low computational resources and can be used with general scenes. In this paper, we propose a new a reliable semi-automatic method that combines a modified version of the detection-based CNN You Only Look Once V4 (YOLOv4) technique and background subtraction technique to perform an unsupervised object detection for surveillance videos. In this proposed strategy, background subtraction-based low-rank decomposition is applied firstly to extract the moving objects. Then, a clustering method is adopted to refine the background subtraction (BS) result. Finally, the refined results are used to fine-tune the modified YOLO v4 before using it in the detection and classification of objects. The main contribution of this work is a new detection framework that overcomes manual labeling and creates an automatic labeler that can replace manual labeling using motion information to supply labeled training data (background and foreground) directly from the detection video. Extensive experiments using real-world object monitoring benchmarks indicate that the suggested framework obtains a considerable increase in mAP compared to state-of-the-art results on both the CDnet 2014 and UA-DETRAC datasets. Full article
(This article belongs to the Special Issue Electric Vehicle Autonomous Driving Based on Image Recognition)
Show Figures

Figure 1

16 pages, 9027 KiB  
Article
DV3-IBi_YOLOv5s: A Lightweight Backbone Network and Multiscale Neck Network Vehicle Detection Algorithm
by Liu Wang, Lijuan Shi, Jian Zhao, Chen Yang, Haixia Li, Yaodong Jia and Haiyan Wang
Sensors 2024, 24(12), 3791; https://doi.org/10.3390/s24123791 - 11 Jun 2024
Cited by 3 | Viewed by 1104
Abstract
Vehicle detection is a research direction in the field of target detection and is widely used in intelligent transportation, automatic driving, urban planning, and other fields. To balance the high-speed advantage of lightweight networks and the high-precision advantage of multiscale networks, a vehicle [...] Read more.
Vehicle detection is a research direction in the field of target detection and is widely used in intelligent transportation, automatic driving, urban planning, and other fields. To balance the high-speed advantage of lightweight networks and the high-precision advantage of multiscale networks, a vehicle detection algorithm based on a lightweight backbone network and a multiscale neck network is proposed. The mobile NetV3 lightweight network based on deep separable convolution is used as the backbone network to improve the speed of vehicle detection. The icbam attention mechanism module is used to strengthen the processing of the vehicle feature information detected by the backbone network to enrich the input information of the neck network. The bifpn and icbam attention mechanism modules are integrated into the neck network to improve the detection accuracy of vehicles of different sizes and categories. A vehicle detection experiment on the Ua-Detrac dataset verifies that the proposed algorithm can effectively balance vehicle detection accuracy and speed. The detection accuracy is 71.19%, the number of parameters is 3.8 MB, and the detection speed is 120.02 fps, which meets the actual requirements of the parameter quantity, detection speed, and accuracy of the vehicle detection algorithm embedded in the mobile device. Full article
(This article belongs to the Section Sensor Networks)
Show Figures

Figure 1

15 pages, 12495 KiB  
Article
An Approach to Incorporating Implicit Knowledge in Object Detection Models
by Wenbo Peng and Jinjie Huang
Appl. Sci. 2024, 14(5), 2124; https://doi.org/10.3390/app14052124 - 4 Mar 2024
Cited by 2 | Viewed by 1372
Abstract
Current object detection methods typically focus on addressing the distribution discrepancies between source and target domains. However, solely concentrating on this aspect may lead to overlooking the inherent limitations of the samples themselves. This study proposes a method to integrate implicit knowledge into [...] Read more.
Current object detection methods typically focus on addressing the distribution discrepancies between source and target domains. However, solely concentrating on this aspect may lead to overlooking the inherent limitations of the samples themselves. This study proposes a method to integrate implicit knowledge into object detection models, aiming to enhance the models’ effectiveness in identifying target features within images. We analyze the sources of information loss in object detection models, treating this loss as a form of implicit knowledge and modeling it in the form of dictionaries. We explore potentially effective ways of integrating latent knowledge into the models and then apply it to object detection models. The models demonstrate a 1% and 0.8% improvement in mean average precision(mAP) in the UA-DETRAC and KITTI datasets, respectively. The results indicate that the proposed method can effectively enhance the relevant metrics of object detection models without significantly increasing the parameter or computational overhead. By excavating and utilizing implicit knowledge, we enhance the performance and efficiency of the models, offering new perspectives and methods for addressing challenges in practical applications. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

16 pages, 5162 KiB  
Article
Transfer Learning-Based YOLOv3 Model for Road Dense Object Detection
by Chunhua Zhu, Jiarui Liang and Fei Zhou
Information 2023, 14(10), 560; https://doi.org/10.3390/info14100560 - 12 Oct 2023
Cited by 6 | Viewed by 2688
Abstract
Stemming from the overlap of objects and undertraining due to few samples, road dense object detection is confronted with poor object identification performance and the inability to recognize edge objects. Based on this, one transfer learning-based YOLOv3 approach for identifying dense objects on [...] Read more.
Stemming from the overlap of objects and undertraining due to few samples, road dense object detection is confronted with poor object identification performance and the inability to recognize edge objects. Based on this, one transfer learning-based YOLOv3 approach for identifying dense objects on the road has been proposed. Firstly, the Darknet-53 network structure is adopted to obtain a pre-trained YOLOv3 model. Then, the transfer training is introduced as the output layer for the special dataset of 2000 images containing vehicles. In the proposed model, one random function is adapted to initialize and optimize the weights of the transfer training model, which is separately designed from the pre-trained YOLOv3. The object detection classifier replaces the fully connected layer, which further improves the detection effect. The reduced size of the network model can further reduce the training and detection time. As a result, it can be better applied to actual scenarios. The experimental results demonstrate that the object detection accuracy of the presented approach is 87.75% for the Pascal VOC 2007 dataset, which is superior to the traditional YOLOv3 and the YOLOv5 by 4% and 0.59%, respectively. Additionally, the test was carried out using UA-DETRAC, a public road vehicle detection dataset. The object detection accuracy of the presented approach reaches 79.23% in detecting images, which is 4.13% better than the traditional YOLOv3, and compared with the existing relatively new object detection algorithm YOLOv5, the detection accuracy is 1.36% better. Moreover, the detection speed of the proposed YOLOv3 method reaches 31.2 Fps/s in detecting images, which is 7.6 Fps/s faster than the traditional YOLOv3, and compared with the existing new object detection algorithm YOLOv7, the speed is 1.5 Fps/s faster. The proposed YOLOv3 performs 67.36 Bn of floating point operations per second in detecting video, which is obviously less than the traditional YOLOv3 and the newer object detection algorithm YOLOv5. Full article
(This article belongs to the Topic Lightweight Deep Neural Networks for Video Analytics)
Show Figures

Figure 1

Back to TopTop