Next Article in Journal
Leakage Detection Method of Coal Mine Power Supply System Based on 3AC
Previous Article in Journal
Research on the Settlement Patterns of Tunnel-Surrounding Rock Under Groundwater Conditions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Vulnerable Road User Detection for Roadside-Assisted Safety Protection: A Comprehensive Survey

Center for Environmental Research and Technology (CE-CERT), Bourns College of Engineering, University of California, Riverside, CA 92507, USA
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(7), 3797; https://doi.org/10.3390/app15073797
Submission received: 7 March 2025 / Revised: 24 March 2025 / Accepted: 28 March 2025 / Published: 30 March 2025
(This article belongs to the Special Issue Computer Vision of Edge AI on Automobile)

Abstract

:
In recent years, the safety of vulnerable road users (VRUs), such as pedestrians, cyclists, and micro-mobility users, has become an increasingly significant concern in urban transportation systems worldwide. Reliable and accurate detection of VRUs is essential for effective safety protection. This survey explores the techniques and methodologies used to detect VRUs, ranging from conventional methods to state-of-the-art (SOTA) approaches, with a primary focus on infrastructure-based detection. This study synthesizes findings from recent research papers and technical reports, emphasizing sensor modalities such as cameras, LiDAR, and RADAR. Furthermore, the survey examines benchmark datasets used to train and evaluate VRU detection models. Alongside innovative detection models and sufficient datasets, key challenges and emerging trends in algorithm development and dataset collection are also discussed. This comprehensive overview aims to provide insights into current advancements and inform the development of robust and reliable roadside detection systems to enhance the safety and efficiency of VRUs in modern transportation systems.

1. Introduction

1.1. Motivation

The world faces critical challenges in safeguarding vulnerable road users (VRUs), including pedestrians, cyclists, and micro-mobility users, as evidenced by concerning statistics. In the US, 2021 National Highway Traffic Safety Administration (NHTSA) data showed that pedestrian fatalities are up by 13% and bicyclists up by 5% [1]. Taking California for an example, in 2020 alone, the state recorded 1013 pedestrian fatalities, alongside 136 cyclist fatalities, according to data from the California Office of Traffic Safety (OTS) [2]. Moreover, the California Highway Patrol (CHP) reported that 1043 pedestrians and 153 bicyclists were killed during the same period [3]. These figures underscore the urgent need to address those factors contributing to road user vulnerability, such as distracted driving, and unsafe interactions at intersections. In 2021, there has been a rise in pedestrian fatalities, which increased by 9.4% from 1013 in 2020 to 1108 in 2021 [2].
According to the definition of Federal Highway Administration (FHWA), a VRU could be a pedestrian, a bicyclist, an e-cyclist, a person using other conveyance like a scooter or skateboard, or a highway worker on foot in a work zone, but this does not include motorcyclists [1]. To protect small and active objects like these in a complex traffic system is a major challenge. With the advancement in computer vision technology, more advanced detection methods have been proposed continuously over the past decades, using either 2D or 3D object detection. From the perspective of traffic management departments, the increasing availability of cost-effective and highly accurate sensors in recent years presents an opportunity to enhance real-time detection and protection of vulnerable road users, in addition to their traditional use for monitoring traffic flow. Usually, roadside sensors are mounted on the poles at the intersections and middle blocks with fixed locations and angles, which is different from onboard sensors’ operating conditions. With well-developed sensors and corresponding detection models, VRUs can be detected accurately, which brings the following benefits: (1) VRU safety can be assured as much as possible; (2) drivers could be aware of the situation by receiving the warning messages in advance if the vehicle/smartphone is connected to the roadside unit (RSU); and (3) the intersection could be smarter by optimizing the priority and phase duration for different road users, especially VRUs.
Roadside detection systems have made significant strides in practical applications to safeguard Vulnerable Road Users (VRUs). Key examples include the Intelligent Roadside Unit (IRSU) developed by the University of Sydney and Cohda Wireless, which integrates camera and LiDAR sensors for precise VRU detection [4], and Commsignia’s V2X units [5], which combine cameras, LiDAR, and RADAR to enhance traffic services and VRU protection. Ouster’s BlueCity solution [6], implemented in Chattanooga, utilizes LiDAR-based traffic management to optimize signal timings and improve safety by processing real-time data [7]. These systems showcase the effectiveness of roadside detection technologies in improving VRU safety and traffic management. Additionally, the National Highway Traffic Safety Administration (NHTSA) is advancing connected vehicle systems, underscoring the continued importance of such technologies in creating safer roadways for VRUs, reducing accidents, and optimizing traffic management globally.
Unlike previous surveys that focus solely on vehicle or pedestrian detection, this paper provides a comprehensive review of VRU detection models and datasets from a roadside perspective, which is an area that has not been systematically explored in detail. This survey will explicitly answer the following questions: (1) What sensors should be used for roadside VRU detection? (2) What models should be used for each type of sensor? (3) How does it work to fuse multiple sensors? (4) Which dataset is suitable for a specific use case? Additionally, we highlight emerging challenges and opportunities, such as refining the definition of VRU subclasses and addressing the limitations of current datasets. By systematically analyzing roadside VRU detection from multiple perspectives—including sensors, models, and datasets—this work serves as a reference for researchers and practitioners developing advanced roadside safety systems.

1.2. Organization

The rest of this paper is organized as follows: sensors, detection models, datasets, and discussions are introduced one by one, respectively. In Section 2, multiple roadside sensors that have been used for traffic surveillance or object detection are listed. In Section 3, different detection approaches by sensor node multiplicity and sensor modality, as well as sensor fusion methods are introduced. Specifically, given the lack of infrastructure-based VRU detection models and the insufficiency of real-world data collected from roadside sensors, this paper will include VRU detection models developed using either roadside or onboard data. Section 4 discusses the infrastructure-based datasets, besides the vehicle–infrastructure cooperative datasets, which include data from both roadside and onboard sources. In Section 5, the challenges and opportunities are discussed in detail, which may guide the algorithm development/data collection. Section 6 concludes this review (the structure of this paper is depicted in Figure 1).

2. Typical Roadside Sensors

To develop a holistic understanding of the overall effectiveness of various sensors employed for VRU perception in transportation systems, Table 1 offers a condensed overview of those commonly utilized in traffic monitoring at intersections. Each type of sensor used alone possesses distinct capabilities and advantages/disadvantages relative to different scenarios and applications. For practical applications, cost and durability are also emphasized in the table. In the following content, camera will be used to indicate a monocular camera.
The commonly used infrastructure-based sensors are summarized in Table 1, with each offering distinct advantages. Cameras provide rich 2D information at a relatively low cost, making them the most widely used for roadside sensing. LiDAR, known for its precise 3D depth information and stability across various weather conditions, has gained traction in recent years. RADAR, with its ability to detect objects at long range and measure speed accurately, is often used in traffic monitoring but is less common for detailed VRU detection due to its lower spatial resolution.
Compared to previous studies, which primarily focus on onboard sensors for VRU detection, this paper emphasizes the selection of roadside sensors, where factors such as a wider field of view, environmental robustness, and long-term deployment feasibility are critical. While previous research has extensively explored camera-based roadside sensing, recent advancements in LiDAR technology have led to its increasing adoption, particularly in sensor fusion setups to enhance detection accuracy. Given these trends, this paper primarily focuses on cameras and LiDAR, as they are the most widely used and effective sensors for roadside VRU detection.
Upon the data collection with different sensors under real-time traffic conditions, the downstream data processing module will be the detection. Thus, along with this, the papers regarding the detection models for VRU will be discussed.

3. VRU Detection Models

In the traffic domain, with the progress of autonomous driving, an increasing number of algorithms and datasets for vehicle detection have been proposed, and vehicle detection accuracy has become increasingly accurate. In comparison, VRU detection has received less attention from scholars. Some generic detection algorithms include pedestrians and bicyclists in the categories of recognized objects, but their recognition rates are generally low. For example, VoxelNet’s [8] detection performances in car, pedestrian, and cyclist are 81.97%, 57.86%, and 67.17% in average precision on the easy KITTI [9] validation set. At the same time, research focusing on VRU detection models and datasets has not made systematic progress, which makes the summary of the relevant literature on VRU detection even more important. The papers on the detection models we reviewed are all related to VRU detection. Some are generic detection models with other VRUs in the categories of recognized objects, and some are specifically focusing on VRU detection. The ratio of detection models we reviewed belonging to the two categories is shown in Figure 2. Specifically, the VRU detection models are developed for various explicit VRUs, including pedestrians, cyclists, skateboarders, e-scooter users, wheelchair users, stroller users, and even pedestrians using umbrellas, which are scattered and not systematic [10,11,12,13,14,15,16]. In general, the VRU detection models utilized rather conventional or simple methods implemented on a specific VRU category without progressive progress.
VRUs are generally small in size, have inconsistent speed distribution, and are very dynamic. The inconsistent speed distribution means similar-sized scooters, bicycles, and skateboards move rather fast; however, pedestrians and wheelchairs move slower. The elderly, especially, move much more slowly. Dynamic changes mean that VRUs do not necessarily travel strictly in prescribed lane areas or along smooth predictable trajectories; they may rush out of blind spots at any time, and children may run across pedestrian areas. Thus, the different characteristics of VRUs make them vulnerable in open traffic environments, highlighting the importance of accurate detection and classification of VRU subclasses. Various challenging scenarios and the lack of the number and variety of VRUs in general datasets all increase the difficulty of the detection model development.
In the following introduction to VRU detection, existing studies are categorized into single-sensor and multi-sensor approaches. Multi-sensor studies are further classified based on sensor modality and fusion strategies, as shown in Figure 3. To provide a comprehensive overview, this paper presents a summary Table 2 highlighting VRU-related detection models, offering readers a clear understanding of the key methodologies and trends in the field.

3.1. Single-Sensor

The single-sensor detection methods mean there is only a single sensor used for detection, such as a single camera or a single LiDAR.

3.1.1. Camera-Based Detection

The camera is the most commonly used sensor for detecting VRUs. Thus, the camera-based detection method is one of the earliest and most mature algorithm types. Two decades ago, Viola and Jones [72] introduced a method using the Haar-like feature and Adaboost algorithm for real-time detection of human faces that surpassed all contemporaneous algorithms in both real-time performance and detection accuracy, without imposing any constraints. Dalal and Triggs [73] presented the Histogram of Oriented Gradients (HOG) feature descriptor, which significantly improved upon the scale-invariant feature transform and shape context. This detector has become a cornerstone in various object detection systems and has found extensive use in practical applications. For example, HOG was used on pedestrian detection, and the miss rate was under 0.01 at 10-6 False Positives Per Window (FPPW) on the original MIT pedestrian database [73,74]. The Deformable Part-based Model (DPM) is another classical 2D detector, which is the SOTA method of traditional methods and is commonly used in VRU detection [13]. In general, HOG and DPM are suitable for detecting non-rigid objects like pedestrians, with tolerance to the pose and view variations.
With the development of deep learning (DL), convolutional neural networks (CNNs) [75] began to be widespread by extracting the features with data-driven methods instead of manual feature engineering. In addition, Girshick et al. [76] introduced Regions with CNN features (R-CNN) for object detection, making the detection in an image with multiple objects possible by generating region proposals. Concurrently, He et al. [77] developed Spatial Pyramid Pooling Networks (SPPNets), which achieved feature representation independent of image size and operated 20 times faster than R-CNN while maintaining accuracy. Later, many additional fast and accurate 2D image detection algorithms were proposed. Fast R-CNN [78] is over 200 times faster than R-CNN. Faster R-CNN [17] is the first end-to-end and near real-time DL detector, which is also the start of anchor-based detection models. Feature Pyramid Networks (FPNs) [20] utilized a pyramid-like structure to detect different-scale objects in the same image, which can be combined with Faster R-CNN. You Only Look Once (YOLO) [79] is the first one-stage DL detector with extremely fast speed. As of now, it has been updated to YOLO v9 [80]. Single Shot MultiBox Detector (SSD) [18] is also a one-stage detector with significantly improved accuracy, but is not suitable for small object detection. Another simple dense one-stage detector is RetinaNet [19], which addresses the detection of rare or small class objects with focal loss function. In addition, there are some anchor-free methods, such as CornerNet [25] and CenterNet [26]. In recent years, Transformers [81] incorporating attention mechanisms have emerged as the predominant trend in most object perception tasks. Vision Transformer (ViT) [27], DEtection TRansformer (DETR) [28], Swin Transformer [29], and Pyramid Vision Transformer (PVT) [30] are the milestone Transformer-based 2D detection models. The trend of 2D detection models is becoming simpler in structure with fewer steps without many manual definitions and introducing new backbone networks.
Regarding the existing VRU detection research, researchers mainly utilized the CNN-based 2D detectors, fine-tuned them, and trained them with specific class datasets to detect different VRUs, not limited to pedestrians. Garcia-Venegas et al. [10] compared the performance of Faster R-CNN [17], SSD [18] and Region-based Fully Convolutional Network (R-FCN) [82] with different feature extractors in cyclist detection. As a result, the SSD was chosen as the detector because of its higher Frames Per Second (FPS) for real-time implementation. Additionally, this work utilized the Kalman filter to enhance the robustness against false negative detections and object occlusions. Another work also used Faster R-CNN employing transfer learning from diverse pre-trained architectures to detect pedestrians with infrastructure-based cameras [83], which proposed a solution for the occlusion challenge for onboard detection. Zhou et al. [84] refined the FPN with multi-scale feature fusion and attention mechanisms, which showed better performance on crowded pedestrian detection; however, the limitation of computational efficiency still existed. Mammeri et al. [85] explored the application of CNN-based algorithms for identifying VRUs like pedestrians, motorcyclists, and cyclists from the roadside. In the experiment, the detection performance of the large-scale object was better than the small-scale object. For example, the detection performance of small objects using the YOLOv3 [21] and Faster R-CNN was worse by about 5% in mAP, compared with the detection performance of large objects. Sharma et al. [86] proposed a rare open-source vulnerable pedestrian dataset BGVP, and trained and tested five pre-trained models, including YOLOv4 [22], YOLOv5 [23], YOLOX [24], Faster R-CNN, and EfficientDet [87], where YOLOv4 and YOLOX performed the best. In general, these methods are still limited by low computational efficiency, small size, and heavy occlusion of the objects.
Similar to 2D detection, camera-based 3D detection methods also follows the trend from CNN-based methods to Transformer-based methods. FCOS3D [31], based on CNNs, generates 3D bounding boxes from monocular images by utilizing depth estimation. In contrast, DETR3D [32] and PETR [33] adopt Transformer-based models, combining features from multiple views to improve detection precision. In recent years, the Bird’s Eye View (BEV) feature space was preferred because it contains rich spatial information within the global view. BEVDepth [34] improved depth estimation by adding depth supervision, a camera-aware module, and refinement. BEVHeight [35] predicted height to the ground instead of depth, enabling distance-agnostic optimization. The Calibration-free BEV Representation (CBR) [36] eliminated the need for calibration, but at the cost of a drop in AP3D|R40. Complementary-BEV (CoBEV) [37] combined depth and height with a two-stage feature selection and BEV distillation, showing superior performance.
Specifically, BEVHeight, CBR, and CoBEV were all developed based on roadside cameras. Thus, BEV-based 3D detection methods using cameras could become a new trend for VRU detection, as roadside cameras provide rich depth and height information, and many intersections are already equipped with them, reducing the need for upgrades. Regarding the variation in roadside cameras with different pitch angles and focal lengths, Jinrang et al. [88] proposed UniMono, which addresses this 3D detection challenge by predicting an independent normalized depth that is unaffected by these variations, making the algorithm adaptable to different camera settings.
In addition to the commonly detected VRUs, such as pedestrians and cyclists, some subclasses, like pedestrians using umbrellas, have also attracted attention from a small number of researchers. Shimbo et al. [13] proposed the Parts Selective DPM to detect pedestrians carrying umbrellas, aiming to address the umbrella occlusion issue in rainy weather. Huang et al. [89] employed a cascade decision tree to detect wheelchairs for a healthcare system. Gilroy et al. [11] focused on e-scooter detection in densely populated urban areas to avoid misclassifying e-scooters as pedestrians. Yang et al. [16] used YOLOv5 [23] to detect strollers. Sharma et al. [86] collected a VRU dataset and categorized VRUs into children, elderly individuals, and persons with disabilities. These rare subclasses of VRUs are often overlooked, yet they represent very vulnerable groups within the transportation system. The definition of VRU subclasses and corresponding detection methods remains insufficient. An explicit definition of VRU subclasses would greatly benefit research in this domain.

3.1.2. LiDAR-Based Detection

LiDAR is a remote sensing technique employing pulsed laser light to measure distances at various angles relative to the sensor (i.e., spherical coordinates of detections in the LiDAR frame of reference). LiDAR can offer detailed 3D point cloud data to accurately determine the 3D positions of traffic objects [90]. The detection methods proposed using LiDAR could be categorized into conventional methods and deep learning methods.
Conventional methods often use some statistical techniques to process the point cloud data. At the early stage, researchers use manually selected features of point clouds as input to classification models, which belong to non-deep learning methods. Limited by the price and synchronization issues related to using multiple LiDAR sensors, many researchers start with one LiDAR to detect road users at the intersection. A skateboarder detection method using a 16-channel LiDAR was proposed by Wu et al. [15]. They utilized 3D Density Statistic Filtering (3D-DSF) [91] to filter the background, and Density-Based Spatial Clustering of Applications with Noise (DBSCAN) [92] to cluster the points. With a two-step classification method, the clustered object is first classified as a vehicle or non-vehicle object. If it is a non-vehicle object, the second step will be deployed with the object’s dynamic features including the average speed in its historical trajectory and the standard deviation of the speed. Thus, the non-vehicle objects were classified as skateboarders, cyclists, and pedestrians. Because of the similar speeds of skateboarders and cyclists, these two may be misclassified using this method. Classical machine learning techniques such as K-Nearest Neighbors (KNN) [93], Support Vector Machine (SVM) [94], Random Forest (RF) [95], and Naïve Bayes (NB) [96] are compared in the two-step classification tasks. RF performed best in both steps under their own datasets. To further optimize the detection results of pedestrians using the same LiDAR, researchers developed efficient detecting methods according to different distances [97]. The detecting area was divided into three sub-areas in concentric circles according to the distance to the LiDAR, and DBSCAN was deployed with different radii and minimum numbers of points in different subareas. Taking the number of points in a cluster, 2D distance to LiDAR, and direction of the clustered points’ distribution as object features, a Back Propagation Artificial Neural Network (BPANN) [98] was used to classify the vehicles and pedestrians. Based on previous research, Song et al. [99] used 3D-DSF and DBSCAN to filter the background and cluster the points, respectively. Then, the overall performance of SVM, RF, BPANN, and Probabilistic Neural Network (PNN) [100] were compared, using measurements of a 16-channel LiDAR as inputs. SVM performed best, benefiting from its better nonlinear regression ability of the Gaussian kernel. In general, conventional methods have the limitation of scene adaptability; different methods may be suitable for different data with different hand-crafted features.
With the rapid advancement of deep learning, some researchers have begun leveraging it for 3D VRU detection due to its strong regression and classification capabilities, enabling the prediction of bounding boxes and object classes through feature learning. From the perspective of generic 3D object detection, the detection models can be divided into front view (FV)-based, bird’s eye view (BEV)-based, voxel-based, and point-based categories. The different categories represent different feature-extracting ways. FV-based and BEV-based methods aim to project the point clouds to the 2D plane and utilize a rather mature 2D detection method to detect the 3D objects, such as VeloFCN [38] and PIXOR [39]. In recent years, voxel-based and point-based methods have become the mainstream because of the richer information encoded. Voxel-based methods divide the 3D space into a structured grid of voxels and use CNNs to extract features, as seen in VoxelNet [8] and SECOND [40]. A variant, pillar-based methods, segment point clouds into pillars by extending voxel boundaries in the z-direction, as demonstrated in PointPillars [41]. Point-based methods directly process raw, unstructured 3D point clouds, using networks like PointNet [101] or PointRCNN [44] to capture fine-grained local and global features. For VRU detection, different feature-extracting categories have their characteristics, which are analyzed in Table 3 to show their pros and cons for VRU detection.
Specifically, in DL-based roadside VRU detection methods using a single LiDAR sensor, researchers have primarily enhanced conventional approaches by replacing one of the modules with a deep learning model, while maintaining the original pipeline. In the early stages, researchers simply replaced the final classification step with DL and attempted to transform the data and features from 3D to 2D. For example, in [103], after background filtering, clustering, and feature extraction, a lightweight Visual Geometry Group Network (VGGNet) was used to classify objects with hand-crafted features as inputs.
With the VGGNet and a rich set of features, the model outperformed the Random Forest (RF) by 14.4% and 11.9% in detecting pedestrians and cyclists, respectively, and surpassed the Support Vector Machine (SVM) by 10.4% and 14.3% in the same categories. Similarly, Zhang et al. [104] proposed a Global Context Network (GCNet) prototype, consisting of three stages: gridding, clustering, and classification. In the gridding stage, the LiDAR point clouds were transformed into a grid format in a plane, enabling the use of a CNN for classifying clustered objects, much like image classification in the 2D plane.
Recently, researchers have shifted towards using end-to-end deep learning networks for roadside VRU detection with LiDAR, replacing traditional clustering, feature extraction, and classification steps. Some studies focused on adapting mature models developed for onboard applications to roadside scenarios. Blanch et al. [105] analyzed key milestones of LiDAR-based detection for onboard systems and compared their performance in roadside surveillance, including SECOND [40], CenterPoint [42] with voxel encoding, CenterPoint with pillar encoding, and PillarNet [43]. Compared to anchor-based methods, center-based methods showed better performance with lower computational demands. Among the methods tested, CenterPoint using voxel representation with specific hyperparameters (distinct from onboard setups) achieved the best performance for pedestrian detection in video surveillance scenarios. Meanwhile, fewer researchers have focused on developing dedicated roadside 3D object detection models, including for VRU objects. Taking the Feature Enhancement Network (FecNet) as an example, FecNet [102] is a pillar-based feature enhancement cascade network designed for roadside object detection with LiDAR. Compared to conventional 3D detection methods, FecNet emphasizes the importance of foreground features with a foreground attention branch to reduce the similarity between objects and the background. Furthermore, multi-level feature fusion strengthens the model’s ability to detect objects at different scales, which is critical for roadside sensing. Based on their dataset, FecNet achieved a mean Average Precision (mAP) of 86.32%, and demonstrated an excellent trade-off between accuracy and efficiency with the smallest model size, outperforming models like PointPillars, SECOND, PointRCNN [44], PV-RCNN [45], PV-RCNN++ [46], and PillarNet. Shi et al. [106] combined voxel-based encoding, center-based proposals, and a Transformer-based detector in an innovative approach for roadside sensing purposes. Tested on the DAIR-V2X-I [107] roadside detection dataset, their method outperformed all other models in cyclist detection, including PointPillars, SECOND, Object DGCNN [108], and FSD [109]. In general, models specifically developed for roadside detection using LiDAR remain limited, and there is insufficient focus on VRU detection.

3.2. Multi-Sensor

Unlike single-sensor, multi-sensor techniques use several sensors for detection at the same time, which is usually called sensor fusion. In this paper, we define two kinds of classification methods for sensor fusion. One is classified by modal, which includes single-modal and multi-modal, corresponding to the type of sensors in sensor fusion. The other one is classified by fusion scheme, including early fusion, intermediate fusion, and late fusion.

3.2.1. Classification by Modality

With the development of the perception research domain, multiple sensors have been utilized to detect the VRUs in traffic systems. In the early stage, some researchers utilized the combination of three low-cost sensors, i.e., Doppler microwave RADAR sensors, passive infrared (PIR), and geomagnetic sensors at the roadside to perceive if there is a pedestrian or vehicle passing [14]. With the cost of sensors going down, increasing numbers of advanced accurate sensors are being utilized at intersections. Different sensors have different advantages in detecting VRUs, which is explained in Table 1. Cameras can provide texture and color, but their detection results are significantly affected by the weather and illumination. RADAR is suitable for long-range detection and operates effectively in all weather conditions but has low resolution. LiDARs can provide depth information under various environmental conditions but with a lower density of detail than camera imagery. Infrared cameras can be effectively used to detect VRUs but can be affected by temperature. Thus, multiple sensors should be combined to detect objects cooperatively, which will fully tap the potentials of different sensors. The detection methods based on sensor modality can be classified into single-modal and multi-modal. For single-modal methods, we focus on the camera-only and the LiDAR-only methods, while for multi-modal approaches, combining the camera with LiDAR or combining the camera with RADAR are prevalent setups.
  • Single-Modal
    Camera only: Multiple cameras can provide a wider field of view (FOV) and more accurate results in VRU detection, particularly in addressing issues related to illumination and occlusion. Inspired by research in the multiple-camera multi-object tracking (MOT) domain [110], cameras from different directions can be fused to detect and track objects. In this approach, objects are first detected in each camera view, and then their movement is tracked across multiple cameras, which constitutes the fusion in the temporal aspect. This paper, however, focuses only on the detection aspect, meaning that the primary goal of using multiple cameras from different directions is to solve issues of illumination and occlusion to ensure VRUs are detected, rather than performing re-identification and tracking. In this context, detections from different cameras are used to complement one another. Moreover, camera fusion can enhance the detection performance. For example, in InfraDet3D [111], a fusion method for the 3D detection results from two roadside cameras was proposed as one of the baseline approaches. In this method, the detections from two cameras are transformed to the same reference frame and matched within a threshold distance. The merging process involved combining matched detections by choosing the central position and yaw vector of the identified object from the nearest sensor. The dimensions of the merged detections were determined by averaging the measurements from both detectors. In general, camera-only fusion with a post-processing technique is designed to complement the detection results from different cameras and improve the detection accuracy of objects detected simultaneously, thus enhancing the safety of VRUs with greater confidence.
    In recent years, deep learning-based feature fusion methods for multi-view 3D object detection, such as DETR3D [32] and PETR [33], have been proposed using onboard-collected images. However, their potential performance in roadside detection remains underexplored.
    LiDAR only: Multiple LiDARs can improve detection accuracy by registering data from different viewpoints to obtain a higher-quality point cloud or by combining detections through late fusion. Meissner et al. [112] established a real-time pedestrian detection network of eight multilayer laser scanners in the KoPER project [113]. The collected data from different sensors are simply registered together. Then, a real-time adaptive foreground extractor was proposed using the Mixture of Gaussian (MoG) [114], and then an improved DBSCAN [112] was achieved by projecting the points to a 2D plane and dividing the clustering area with cells to reduce computation. Limited by the fixed range and the required minimum number of clusters in DBSCAN, when it comes to two pedestrians being very close, it is very challenging to separate them. In InfraDet3D [111], different fusion methods of two LiDARs were proposed as the baseline methods, including early fusion and late fusion. Early fusion of a LiDAR-only setup is to merge numerous point cloud scans generated by various LiDAR sensors at time step t to create a unified and densely populated point cloud, which used Fast Global Registration [115] to register the merged point cloud in the initial transformation and point-to-point Iterative Closest Point (ICP) [116] for the refinement. A late fusion baseline method of LiDAR was also talked about in the paper, which continued to use the same late fusion approach mentioned in camera-only fusion. In this work, pedestrians and cyclists were considered in the detection but not as the main objects. Although not much attention is given to the VRUs, the two fusion methods proposed are very straightforward and computationally cheap, which gives some inspiration for roadside real-time VRU sensing. Additionally, feature-level fusion using multiple LiDARs has been explored. For example, PillarGrid [117] extracts features from two different LiDARs in a pillar-wise manner and fuses them in a grid-wise fashion to enhance detection performance and extend the detection range.
  • Multi-Modal
    Regarding multi-modal sensor fusion, various sensor combinations can be used for detection, including but not limited to cameras and LiDAR or cameras and RADAR. However, this paper primarily focuses on these two common combinations for VRU detection, with other multi-modal fusion methods briefly introduced.
    Camera and LiDAR: The fusion of camera and LiDAR sensors is one of the most common multi-modal sensor combinations. Cameras provide rich features that are beneficial for classification and detection during the daytime, while LiDAR can detect objects at night and offer more accurate point cloud data without distortion. In recent years, many promising fusion methods using cameras and LiDAR have been proposed. Generic fusion methods are listed in Table 4, with fusion strategies categorized based on pairs of sensor-derived feature types. Each category combines specific types of features from LiDAR and camera sensors; for example, FV/BEV and feature map represent features derived from LiDAR’s FV/BEV and camera’s feature map.
    In roadside sensor fusion applications, methods have evolved from assigning distinct functions to individual sensors for detection to integrating detections or features from multiple sensors. Shan et al. [4] designed an infrastructure-based roadside unit (IRSU) with two cameras, one LiDAR, and a Cohda Wireless RSU to detect pedestrians in the blind spots of connected and autonomous vehicles (CAVs), providing the CAVs with collision avoidance data. In this setup, the cameras were used for object detection and classification, while LiDAR was used to generate 3D dimensions and orientations. InfraDet3D [111] proposed a camera–LiDAR fusion method where LiDAR detection results were first transformed into the camera coordinate system, and then the detections from both sensors were synchronized based on the distance between the center points of the detected objects. In this method, the camera’s detection served as a supplement to LiDAR detection, employing a late fusion strategy. In CoopDet3D [118], Zimmer et al. extracted features with roadside data from a camera and LiDAR. The features were transformed into a bird’s eye view (BEV) representation. This method outperformed InfraDet3D on the TUMTraf dataset. VBRFusion [119] was proposed for VRU detection at medium to long ranges, based on the DAIR-V2X dataset [107]. VBRFusion enhanced the encoding of point and image feature mapping, as well as the fusion of voxel features, to generate high-dimensional voxel representations. In general, roadside fusion methods are gradually shifting from conventional function fusion approaches to feature fusion methods, aligning with generic LiDAR and camera sensor fusion techniques.
    Camera and RADAR: The combination of a camera and RADAR as a sensing unit is a cost-effective solution for detection. Similar to LiDAR-based detection models, this category includes both conventional methods and deep learning-based methods for fusing data from cameras and RADAR. Conventional methods often employ weighted averaging and clustering techniques, while deep learning approaches leverage neural networks to improve detection accuracy and robustness [120]. Bai et al. [121] developed a fusion method that capitalizes on the strengths of both cameras and RADAR to detect road users, including VRUs. RADAR detection results provide the distance and velocity of the target by measuring the time delay and phase shift of the echo signal. The camera detects the target’s angle and classification using YOLO V4 [22]. After calibrating the coordinates of both sensors, the results from the two sensors are fused using the nearest-neighbor correlation method. Similarly, in RODNet, RGB images are used to classify and localize objects, while transformed radio frequency images are utilized only for object localization [122]. Liu et al. [123] proposed methods for both single-scan and multi-scan fusion using Dempster’s combination rule to enhance evidence fusion performance. They also employed pre-calibration and the nearest-neighbor correlation method to associate detections from the camera and RADAR. Furthermore, the authors fused the detection results from multiple scans by associating detections from adjacent scans, which improved overall performance. With camera and RADAR fusion, challenging scenarios, particularly those captured under poor lighting conditions, can be detected more accurately.
    Regarding the DL-based methods, CenterFusion [67] associates RADAR point clouds with image detections, concatenates features from both the camera and RADAR, and recalculates the depth, rotation, and velocity of the objects. CRAFT [68] introduced the Soft-Polar Association and Spatio-Contextual Fusion Transformer to facilitate information sharing between the camera and RADAR. In recent years, inspired by BEV-based 3D detection methods with cameras, BEV-based feature maps have also been generated in camera and RADAR fusion, leveraging RADAR’s precise occupancy information. RCBEV4d [69] extracted spatial–temporal features from RADAR and applied point fusion and ROI fusion with transformed BEV image features. CRN [70] achieved comparable performance to LiDAR detectors on the nuScenes dataset, outperforming them at long distances. RCBEVDet [71] introduced a dual-stream RADAR backbone to capture both local RADAR features and global information, along with a RADAR Cross-Section (RCS)-aware BEV encoder that scatters point features across multiple BEV pixels rather than just one. This method outperformed previous approaches on the nuScenes dataset. In general, RADAR serves as a complementary information source to camera-based 3D detection methods, providing more accurate depth information, especially in adverse weather conditions.
    Others: Another camera combination involves the fusion of RGB cameras and infrared (IR) cameras, which can significantly improve detection performance, particularly under poor visibility conditions such as adverse weather or nighttime [124,125]. Pei et al. [126] also proposed a multi-spectral pedestrian detector that combines visible-optical (VIS) and infrared (IR) images. This method achieves state-of-the-art (SOTA) performance on the KAIST dataset [127]. When the RGB and infrared cameras share the same field of view (FOV), an image fusion technique can be employed to create a single combined image, making it easier to distinguish VRUs from the background [128].
    MMPedestron [129] is a novel multi-modal model for pedestrian detection that consists of a unified multi-modal encoder and a detection head. It can fuse data from five modalities (RGB, IR, Depth, LiDAR, and Event) and is extendable to include additional modalities. The data sources can also span multiple domains, such as traffic surveillance, autonomous driving, and robotics. Trained on a unified dataset composed of five public datasets with different modalities, MMPedestron surpasses the existing SOTA models for each individual dataset. This demonstrates the potential of a unified generalist model for multi-sensor perception in pedestrian detection, outperforming sensor fusion models tailored to specific sensor modalities.

3.2.2. Classification by Fusion Scheme

Regarding the phase of sensor fusion [130], as depicted in Figure 4, a multi-sensor perception system can be categorized into three main classes: (1) Early fusion (EF) involves the fusion of raw data during the pre-processing phase. (2) Intermediate fusion (IF) entails the fusion of features during the feature extraction stage. (3) Late fusion (LF) involves the fusion of perception outcomes during the post-processing phase. Each fusion scheme has its own set of advantages and drawbacks. For instance, EF and IF typically require more computational resources and more complex model structures to efficiently extract features and achieve higher accuracy. In contrast, LF generally employs conventional rule-based methods to improve final performance with greater efficiency, making it suitable for real-time applications. Therefore, different scenarios may require different fusion schemes to meet specific objectives. Moreover, various fusion schemes may be applied simultaneously within a detection model to leverage the strengths of each approach.
The following section highlights some major studies classified by fusion scheme:
  • Early Fusion
    Sharing the raw sensor data with other sensors to extend the perceptual range or point cloud density (multiple LiDARs) and enhance detection accuracy is a logical approach. In line with this strategy, the raw data collected from various sensors is transformed into a unified coordinate system to facilitate subsequent processing [131]. However, this kind of strategy is inevitably sensitive to some issues such as sensor calibration, data synchronization, and communication capacity, which are critical considerations in real-time implementations [132]. For example, with a limited communication bandwidth, it may be viable to transmit limited-resolution image data, but it might not be practical to transmit LiDAR point cloud data in real-time. For instance, a 64-beam Velodyne LiDAR operating at 10 Hz can produce approximately 20 MB of data per second [9].
  • Intermediate Fusion
    The fundamental principle behind IF can be succinctly described as employing deeply extracted features for fusion at intermediate stages within the perception pipeline. IF relies on hidden features primarily derived from deep neural networks, which exhibit greater robustness compared to the raw sensor data utilized in early fusion. Moreover, feature-based fusion approaches typically employ a single detector to produce object perception outcomes, thereby eliminating the necessity of consolidating multiple proposals as required in LF. For example, MVX-Net utilizes the IF strategy to fuse features from 2D images and 3D points or voxels, and it is processed by VoxelNet [8]. Bai et al. also proposed the PillarGrid to fuse the extracted features of two LiDARs to improve the detection accuracy [117].
  • Late Fusion
    Late fusion also adopts a natural cooperative paradigm for perception, wherein perception outcomes are independently generated and subsequently fused. Unlike EF, although LF still requires a relative positioning for merging perception results, its robustness to calibration errors and synchronization issues is significantly improved. This is primarily because object-level fusion can be established based on spatial and temporal constraints. Rauch et al. [133] employed the Extended Kalman Filter (EKF) to coherently align the shared bounding box proposals, leveraging spatiotemporal constraints. Moreover, techniques such as Non-Maximum Suppression (NMS) [134] and other machine learning-driven proposal refining methods find extensive applications in LF methodologies for object perception [135].
Thus, different fusion schemes have their advantages and disadvantages. Early fusion and late fusion are both straightforward, but early fusion requires accurate calibration and sufficient communication bandwidth, and late fusion may result in limited detection accuracy. Deep fusion is a balanced and relatively well-performing method because of its low communication requirements and high detection accuracy.

4. Datasets

In the first stages of research, because of the lack of public infrastructure-based VRU datasets, some researchers chose to utilize some existing vehicle-based datasets including pedestrians or cyclists to evaluate their algorithms, because these datasets are well labeled and widely accepted. In [105], the authors listed some datasets developed for autonomous driving such as KITTI [9], Waymo [136], and nuScenes [137], and analyzed which one is more suitable for the LiDAR-assisted pedestrian detection for video surveillance. The nuScenes dataset was chosen because of its abundant annotations, like bounding boxes, orientations, and velocities of pedestrians. Also, it has relatively low spatial resolution and high temporal resolution, which aligns with the requirements of roadside setup in real-world situations. By integrating continuous sweeps from nuScenes, the range of the distances to the sensor is changeable and controllable for the specific detection task, e.g., VRU detection at intersections.
In this paper, we have reviewed the existing VRU detection datasets, including single-sensor datasets and multi-sensor datasets. Within single-sensor datasets, we include both the vehicle-based datasets and infrastructure-based datasets because of the limited amount of the latter. Within multi-sensor datasets, we further divide them into infrastructure-based datasets and vehicle–infrastructure cooperative datasets.

4.1. Single-Sensor Datasets

Single-sensor datasets are primarily based on cameras, and these datasets are listed in Table 5. The table only includes datasets that focus on pedestrian detection in traffic-related scenes, and some older datasets are excluded. In the object class attribute, VRU object classes are bolded, and the “Ignore Region” refers to detected humans that do not need to be matched, such as those truncated by image boundaries or depicted as fake humans on posters.
There are several factors that single-sensor datasets should consider. The image resolution and the sufficiency of both images and annotations can directly impact detection performance. The sufficiency of VRU classes is particularly important for in-depth research on VRU detection, such as detecting children or wheelchair users, which is critical for protecting these specific VRU categories. The variety of scenes is essential to ensure the successful detection of VRUs under different seasons, weather conditions, and lighting, especially some safety critical scenes. Detailed annotations for occlusion and truncation can help researchers develop more robust detectors. Lastly, orientation labels are valuable as they allow researchers to predict both the bounding box and the orientation simultaneously, which are crucial pieces of information for object tracking and state prediction of VRUs.

4.2. Multi-Sensor Datasets

Regarding multi-sensor datasets, both infrastructure-based and cooperative datasets will be discussed, as both types include infrastructure components, and the shared information from the vehicle side can help detect VRUs from the roadside.
In chronological order, the details of various infrastructure-based datasets are shown in Table 6. KoPER [113] is one of the earliest roadside datasets, collected using 14 eight-layer laser scanners and two monochrome Charged-Coupled Device (CCD) cameras. The sensors were mounted at a height of over 5 m and were tuned to reduce occlusion. The VRUs labeled in this dataset include pedestrians and cyclists. BAAI-VANJEE [144] is a dataset collected using a 32-line LiDAR and four cameras, mounted at 4.5 m, with partial synchronization of the data from different sensors. The VRUs in this dataset comprise pedestrians, bicycles, and tricycles. DAIR-V2X-I [107] is a roadside dataset collected across several intersections, where at least one pair of cameras and LiDAR sensors was installed at each location. The data from different sensors were fully synchronized, and the VRU objects in this dataset include cyclists, tricyclists, barrows, and pedestrians. Zimmer et al. proposed the TUMTraf Intersection Dataset, which includes pedestrians and bicycles as VRUs, along with complex driving maneuvers such as left/right turns, overtaking, and U-turns [118], which are safety critical scenes. In LLVIP, visual and infrared cameras were used to collect pedestrian and cyclist data in low-light scenes [125]. In TUMTraf Event, a visual camera and event-based camera were used to collect roadside detection data, especially for motion blur and low-light scenes [145]. Recently, the MMPD multi-modal dataset was released, which incorporates five public datasets with different fusion modalities and scenarios [129].
The cooperative datasets contain data from both vehicles and infrastructure, as summarized in Table 7. These datasets benefit roadside detection by sharing information from vehicles. A widely used dataset is Rope3D [146], which uses one LiDAR on a vehicle and several cameras on the roadside. Within Rope3D, various setups were explored by collecting data with different camera mounting heights, pitch angles, and focal lengths. Yu et al. [107] released the DAIR-V2X-C, which also provided cooperative tracking annotations. The TUMTraf V2X cooperative perception dataset [147] provided some corner cases such as tailgating, overtaking, U-turns, traffic violations, and occlusion scenarios; however, the duration of each case may be too short for training a robust model. An SOTA vehicle-to-everything (V2X) cooperative detection dataset is V2X-Real [148]. Based on the ego agent type, the authors propose four types of datasets: vehicle-centric, infrastructure-centric, vehicle-to-vehicle (V2V), and infrastructure-to-infrastructure (I2I), each tailored for specific research tasks. This dataset contains the most abundant frames and annotations and includes a scooter category not recorded by other datasets. Another highlight is that the number of pedestrians exceeds the number of vehicles, making it well suited for VRU detection.
Additionally, the simulated dataset V2X-SIM is also mentioned [149]. However, its limitation is that it only includes one type of VRU: pedestrians. Simulated datasets are particularly useful for generating corner cases, such as accidents and occlusions involving VRUs, which are difficult and hazardous to capture in real-world settings. In conclusion, a comprehensive dataset that covers a wide range of VRU categories and includes diverse scenarios, including corner cases, is still lacking.

5. Discussion

5.1. Detection Models

5.1.1. Challenges

The following are challenges found in the detection models:
  • Real-time Processing
    As GPU computing performance continues to advance, researchers are developing more complex models and exploring the fusion of diverse data across various modalities and scenarios. The goal is to train robust models with strong generalization capabilities that are not constrained by a single modality or scenario. However, in practical roadside assistance systems, deploying large-scale multi-modal models can be challenging due to limitations in computing performance and communication latency between sensors and the edge computer. Therefore, developing lightweight models that offer robust performance and high tolerance to variations presents a significant research direction in road-assisted detection systems.
  • Non-rigid Small 3D Object Detection
    Given the small size of VRUs, accurately detecting them in 3D is inherently more challenging due to the limited information collected by sensors. Relying on a single sensor may not provide sufficient data, making sensor fusion a more effective solution to address the missing features and enhance detection accuracy.
    Considering the non-rigid characteristic of the VRUs, such as walking, running, sitting and standing, it will be more challenging to detect them compared with rigid objects like vehicles. Compared with 2D detection, there is much less research focusing on this issue in 3D detection with LiDAR, which may be the result of sparse points on VRU itself and most LiDAR’s application in rigid object detection such as for vehicles. However, with the advancement of LiDAR resolution and the ability to capture more detailed VRU gestures, this area presents promising opportunities for future research.
  • Domain Generalization from Onboard to Roadside
    In real-world applications, many pre-trained models are directly used for practical purposes, such as roadside surveillance. However, most pre-trained detection models that use point clouds as input are primarily designed and trained for onboard applications. A key question arises: how adaptable are these pre-trained onboard detection models to roadside detection tasks? Typically, when a LiDAR-based detection model trained on onboard datasets is directly applied to roadside data, performance degradation often occurs. This issue falls under domain generalization, where the challenge is ensuring the model performs well despite domain differences between development and deployment data, with minimal additional investment.

5.1.2. Trends and Future Research

The following are potential future trends in related research:
  • Low-cost Solution for Comparative Resilience to LiDAR
    An increasing number of studies are exploring multi-camera 3D detection combined with RADAR as an affordable alternative to LiDAR. RADAR provides robust performance under varying lighting conditions and is effective for depth measurement, while cameras deliver high resolution and rich semantic information. Additionally, RADAR demonstrates strong weather robustness, velocity estimation, and long-range detection, offering several advantages over LiDAR [68]. Although RADAR has limitations, such as sparsity, inaccurate measurements, and ambiguity, researchers have proposed RADAR noise filtering modules to enhance detection accuracy [150,151]. Notably, CRAFT [68] achieves a 41.1% mAP on the nuScenes test dataset, with 46.2% AP for pedestrians and 31.0% AP for bicycles. However, none of the existing methods have been specifically tailored for VRU detection. As a result, roadside VRU detection using camera and RADAR fusion remains largely unexplored, representing a promising area for future research.
  • Unified Generalist Model
    Recently, some researchers have attempted to develop unified detection models that integrate multiple modalities and domains to enhance robustness across various settings. For example, MMPedestron [129] fused data from five modalities across domains, ranging from traffic surveillance to robotics. As a result, it outperforms existing state-of-the-art models on individual datasets. In this context, variations in roadside data, such as tilt angle, mounting height, and resolution, may not pose limitations for a unified model trained with diverse data.
  • Tracking by Detection
    Recently, a new trend has been combining the detection and tracking module as an end-to-end method to solve the detection and tracking at the same time, which solves VRU occlusion issues while saving the training process and avoiding the complex data association process.

5.2. Datasets

5.2.1. Challenges

The following are challenges found in the related datasets:
  • Unclear Definition of VRU Subclasses
    Existing datasets lack a variety of VRUs, and the VRU subclasses within these datasets are inconsistent, which increases the difficulty of developing detection models. As a result, a clear definition of VRU subclasses is missing, and some rare VRUs do not receive sufficient attention. To facilitate dataset collection, annotation, and the development of corresponding detection models, we propose a standardized classification for most VRU categories, building on previous classifications found in various datasets and the Intersection Safety Challenge hosted by the U.S. Department of Transportation [152]. The detailed VRU classifications are presented in Table 8.
  • Insufficient Critical VRU Scenarios
    Most perception datasets focus on detecting main road users, such as vehicles, while often overlooking critical corner cases, such as high-risk interactions between vehicles and pedestrians [153]. This creates a strong need for real-world or synthetic datasets that adequately represent critical VRU scenarios. Although simulated datasets can generate challenging situations like accidents and occlusions, a comprehensive dataset covering diverse VRU categories and corner cases is still lacking.

5.2.2. Trends and Future Research

The following is the potential future trend in related research:
  • Sensor Placement Testing
    A quick and precise evaluation of roadside sensor placement before installation is crucial for effective data collection or monitoring. Sensor placement without proper evaluation can be time-consuming and may make it difficult to identify the optimal viewpoint. This is especially critical for VRU detection, where it is essential to select locations that cover sidewalks and crossings while minimizing occlusion. Therefore, evaluating the influence of sensor placement on the view and performance of baseline detection models is necessary to ensure optimal functionality.
    For instance, Cai et al. (2023) [154] conducted a simulation study comparing various LiDAR placements across multiple intersection scenarios. This approach enabled the evaluation of different placements alongside various detection algorithms, using real-world mirror scenarios for comparison. This study offers valuable insights into sensor placement testing, highlighting how small changes in sensor positioning can significantly impact VRU detection accuracy. Proper sensor placement is thus a key factor in optimizing detection performance and minimizing false negatives, making it an essential aspect of system design.
Table 8. Detailed classification of VRUs.
Table 8. Detailed classification of VRUs.
Classes *SubclassesReference
Pedestrian
  • Child
  • Adult
  • Elderly
[73,83,85,86,97,105,112,129,131,138,139,141,155,156,157,158,159]
Wheelchair
  • Manual Wheelchair
  • Motorized Wheelchair
[12]
Bicycle
  • Manual Bicycle
  • Motorized Bicycle
[10]
Non-Motorized Device
  • Cane
  • Stroller
  • Walker
  • Crutches
  • Umbrella
  • Cardboard Box
  • Luggage
[13,16,89]
Scooter
  • Manual Scooter
  • Motorized Scooter
[11]
SkateboardN.A.[15]
* The items listed refer to pedestrians or individuals using devices, such as the pedestrian using a wheelchair.

6. Conclusions

This paper presents a comprehensive review of VRU detection for roadside-assisted safety, covering various roadside sensors, detection models, and datasets. The review follows a progression from generic techniques to specific VRU applications, analyzing existing models and datasets while highlighting challenges and emerging trends in the field. We aim to provide researchers with a concise overview of roadside VRU detection, fostering inspiration for future advancements. However, several limitations must be acknowledged. The article selection may be biased toward studies from more accessible or well-known databases. Additionally, this review primarily focuses on LiDAR and cameras, potentially overlooking other emerging technologies. The consulted databases may also have missed relevant studies, particularly those in non-English publications or published after the review period. These factors could impact the completeness and generalizability of the findings. Future research should broaden inclusion criteria and explore a wider range of sensor technologies.

Author Contributions

Conceptualization, G.W.; information collection, Z.Z.; formal analysis, Z.Z.; investigation, Z.Z. and C.W.; writing—original draft preparation, Z.Z.; writing—review and editing, Z.Z., C.W., G.W. and M.J.B.; visualization, Z.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by California Department of Transportation (65A1057).

Acknowledgments

The contents of this paper only reflect the views of the authors, who are responsible for the facts presented herein. The contents do not necessarily reflect the official views of California Department of Transportation.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Sun, C.; Brown, H.; Edara, P.; Stilley, J.; Reneker, J. Vulnerable Road User (VRU) Safety Assessment; Technical Report; Department of Transportation, Construction and Materials Division: Jefferson City, MO, USA, 2023. [Google Scholar]
  2. California Traffic Safety Quick Stats; California Office of Traffic Safety: Elk Grove, CA, USA, 2021.
  3. California Quick Crash Facts; California Highway Patrol: Sacramento, CA, USA, 2020.
  4. Shan, M.; Narula, K.; Wong, Y.F.; Worrall, S.; Khan, M.; Alexander, P.; Nebot, E. Demonstrations of Cooperative Perception: Safety and Robustness in Connected and Automated Vehicle Operations. Sensors 2020, 21, 200. [Google Scholar] [CrossRef] [PubMed]
  5. Commsignia. Vulnerable Road User (VRU) Detection Case Study. n.d. Available online: https://www.commsignia.com/casestudy/vru (accessed on 23 March 2025).
  6. Ouster. BlueCity Solution. 2025. Available online: https://ouster.com/products/software/bluecity (accessed on 23 March 2025).
  7. Ouster. Ouster BlueCity to Power Largest Lidar-Enabled Smart Traffic. 2025. Available online: https://investors.ouster.com/news-releases/news-release-details/ouster-bluecity-power-largest-lidar-enabled-smart-traffic (accessed on 23 March 2025).
  8. Zhou, Y.; Tuzel, O. Voxelnet: End-to-end learning for point cloud based 3d object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4490–4499. [Google Scholar]
  9. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? the kitti vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar]
  10. García-Venegas, M.; Mercado-Ravell, D.A.; Pinedo-Sánchez, L.A.; Carballo-Monsivais, C.A. On the safety of vulnerable road users by cyclist detection and tracking. Mach. Vis. Appl. 2021, 32, 109. [Google Scholar] [CrossRef]
  11. Gilroy, S.; Mullins, D.; Jones, E.; Parsi, A.; Glavin, M. E-Scooter Rider detection and classification in dense urban environments. Results Eng. 2022, 16, 100677. [Google Scholar] [CrossRef]
  12. Beyer, L.; Hermans, A.; Leibe, B. DROW: Real-Time Deep Learning-Based Wheelchair Detection in 2-D Range Data. IEEE Robot. Autom. Lett. 2017, 2, 585–592. [Google Scholar] [CrossRef]
  13. Shimbo, Y.; Kawanishi, Y.; Deguchi, D.; Ide, I.; Murase, H. Parts Selective DPM for detection of pedestrians possessing an umbrella. In Proceedings of the 2016 IEEE Intelligent Vehicles Symposium (IV), Gothenburg, Sweden, 19–22 June 2016; pp. 972–977. [Google Scholar] [CrossRef]
  14. Wang, H.; Li, C.; Zhang, Y.; Liu, Z.; Hui, Y.; Mao, G. A scheme on pedestrian detection using multi-sensor data fusion for smart roads. In Proceedings of the 2020 IEEE 91st Vehicular Technology Conference (VTC2020-Spring), Victoria, BC, Canada, 18 November–16 December 2020; pp. 1–5. [Google Scholar]
  15. Wu, J.; Xu, H.; Yue, R.; Tian, Z.; Tian, Y.; Tian, Y. An automatic skateboarder detection method with roadside LiDAR data. J. Transp. Saf. Secur. 2021, 13, 298–317. [Google Scholar] [CrossRef]
  16. Yang, J.; Tong, Q.; Zhong, Y.; Li, Q. Improved YOLOv5 for stroller and luggage detection. In Proceedings of the 2023 4th International Conference on Computer Engineering and Application (ICCEA), Hangzhou, China, 7–9 April 2023; pp. 252–257. [Google Scholar]
  17. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the 29th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; Volume 1, pp. 91–99. [Google Scholar]
  18. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Computer Vision—ECCV 2016; Lecture Notes in Computer Science; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; Volume 9905, pp. 21–37. [Google Scholar] [CrossRef]
  19. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  20. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  21. Farhadi, A.; Redmon, J. Yolov3: An incremental improvement. In Proceedings of the Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; Springer: Berlin/Heidelberg, Germany, 2018; Volume 1804, pp. 1–6. [Google Scholar]
  22. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  23. Ultralytics. YOLOv5: A State-of-the-Art Real-Time Object Detection System. 2021. Available online: https://docs.ultralytics.com (accessed on 23 March 2025).
  24. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. YOLOX: Exceeding YOLO Series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  25. Law, H.; Deng, J. Cornernet: Detecting objects as paired keypoints. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 734–750. [Google Scholar]
  26. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. CenterNet: Keypoint Triplets for Object Detection. arXiv 2019, arXiv:1904.08189. [Google Scholar]
  27. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. arXiv 2021, arXiv:2010.11929. [Google Scholar]
  28. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. In Computer Vision—ECCV 2020; Lecture Notes in Computer Science; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Springer International Publishing: Cham, Switzerland, 2020; Volume 12346, pp. 213–229. [Google Scholar] [CrossRef]
  29. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Online, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  30. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Online, 11–17 October 2021; pp. 568–578. [Google Scholar]
  31. Wang, T.; Zhu, X.; Pang, J.; Lin, D. Fcos3d: Fully convolutional one-stage monocular 3d object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Online, 11–17 October 2021; pp. 913–922. [Google Scholar]
  32. Wang, Y.; Guizilini, V.C.; Zhang, T.; Wang, Y.; Zhao, H.; Solomon, J. Detr3d: 3d object detection from multi-view images via 3d-to-2d queries. In Proceedings of the Conference on Robot Learning, PMLR, London, UK, 8–11 November 2022; pp. 180–191. [Google Scholar]
  33. Liu, Y.; Wang, T.; Zhang, X.; Sun, J. Petr: Position embedding transformation for multi-view 3d object detection. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 531–548. [Google Scholar]
  34. Li, Y.; Ge, Z.; Yu, G.; Yang, J.; Wang, Z.; Shi, Y.; Sun, J.; Li, Z. Bevdepth: Acquisition of reliable depth for multi-view 3d object detection. Proc. AAAI Conf. Artif. Intell. 2023, 37, 1477–1485. [Google Scholar]
  35. Yang, L.; Yu, K.; Tang, T.; Li, J.; Yuan, K.; Wang, L.; Zhang, X.; Chen, P. Bevheight: A robust framework for vision-based roadside 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 21611–21620. [Google Scholar]
  36. Fan, S.; Wang, Z.; Huo, X.; Wang, Y.; Liu, J. Calibration-free bev representation for infrastructure perception. In Proceedings of the 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Detroit, MI, USA, 1–5 October 2023; pp. 9008–9013. [Google Scholar]
  37. Shi, H.; Pang, C.; Zhang, J.; Yang, K.; Wu, Y.; Ni, H.; Lin, Y.; Stiefelhagen, R.; Wang, K. CoBEV: Elevating Roadside 3D Object Detection with Depth and Height Complementarity. arXiv 2023, arXiv:2310.02815. [Google Scholar] [CrossRef] [PubMed]
  38. Li, B.; Zhang, T.; Xia, T. Vehicle detection from 3d lidar using fully convolutional network. arXiv 2016, arXiv:1608.07916. [Google Scholar]
  39. Yang, B.; Luo, W.; Urtasun, R. Pixor: Real-time 3d object detection from point clouds. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7652–7660. [Google Scholar]
  40. Yan, Y.; Mao, Y.; Li, B. SECOND: Sparsely Embedded Convolutional Detection. Sensors 2018, 18, 3337. [Google Scholar] [CrossRef] [PubMed]
  41. Lang, A.H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; Beijbom, O. Pointpillars: Fast encoders for object detection from point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 12697–12705. [Google Scholar]
  42. Yin, T.; Zhou, X.; Krahenbuhl, P. Center-based 3d object detection and tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11784–11793. [Google Scholar]
  43. Shi, G.; Li, R.; Ma, C. PillarNet: Real-Time and High-Performance Pillar-Based 3D Object Detection. In Computer Vision—ECCV 2022; Lecture Notes in Computer Science; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Springer Nature: Cham, Switzerland, 2022; Volume 13670, pp. 35–52. [Google Scholar] [CrossRef]
  44. Shi, S.; Wang, X.; Li, H. Pointrcnn: 3d object proposal generation and detection from point cloud. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 770–779. [Google Scholar]
  45. Shi, S.; Guo, C.; Jiang, L.; Wang, Z.; Shi, J.; Wang, X.; Li, H. Pv-rcnn: Point-voxel feature set abstraction for 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10529–10538. [Google Scholar]
  46. Shi, S.; Jiang, L.; Deng, J.; Wang, Z.; Guo, C.; Shi, J.; Wang, X.; Li, H. PV-RCNN++: Point-Voxel Feature Set Abstraction with Local Vector Representation for 3D Object Detection. Int. J. Comput. Vis. 2023, 131, 531–551. [Google Scholar] [CrossRef]
  47. Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3d object detection network for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1907–1915. [Google Scholar]
  48. Ku, J.; Mozifian, M.; Lee, J.; Harakeh, A.; Waslander, S.L. Joint 3d proposal generation and object detection from view aggregation. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1–8. [Google Scholar]
  49. Liang, M.; Yang, B.; Wang, S.; Urtasun, R. Deep continuous fusion for multi-sensor 3d object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 641–656. [Google Scholar]
  50. Lu, H.; Chen, X.; Zhang, G.; Zhou, Q.; Ma, Y.; Zhao, Y. SCANet: Spatial-channel attention network for 3D object detection. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 1992–1996. [Google Scholar]
  51. Krispel, G.; Opitz, M.; Waltner, G.; Possegger, H.; Bischof, H. Fuseseg: Lidar point cloud segmentation fusing multi-modal data. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass, CO, USA, 1–5 March 2020; pp. 1874–1883. [Google Scholar]
  52. Liu, Z.; Tang, H.; Amini, A.; Yang, X.; Mao, H.; Rus, D.L.; Han, S. Bevfusion: Multi-task multi-sensor fusion with unified bird’s-eye view representation. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 2774–2781. [Google Scholar]
  53. Qiao, D.; Zulkernine, F. CoBEVFusion: Cooperative Perception with LiDAR-Camera Bird’s-Eye View Fusion. arXiv 2023, arXiv:2310.06008. [Google Scholar]
  54. Kim, M.; Kim, G.; Jin, K.H.; Choi, S. Broadbev: Collaborative lidar-camera fusion for broad-sighted bird’s eye view map construction. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 11125–11132. [Google Scholar]
  55. Sindagi, V.A.; Zhou, Y.; Tuzel, O. Mvx-net: Multimodal voxelnet for 3d object detection. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 7276–7282. [Google Scholar]
  56. Yoo, J.H.; Kim, Y.; Kim, J.; Choi, J.W. 3d-cvf: Generating joint camera and lidar features using cross-view spatial feature fusion for 3d object detection. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XXVII 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 720–736. [Google Scholar]
  57. Zhu, H.; Deng, J.; Zhang, Y.; Ji, J.; Mao, Q.; Li, H.; Zhang, Y. Vpfnet: Improving 3d object detection with virtual point based lidar and stereo data fusion. IEEE Trans. Multimed. 2022, 25, 5291–5304. [Google Scholar] [CrossRef]
  58. Wang, C.; Ma, C.; Zhu, M.; Yang, X. Pointaugmenting: Cross-modal augmentation for 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11794–11803. [Google Scholar]
  59. Xu, D.; Anguelov, D.; Jain, A. Pointfusion: Deep sensor fusion for 3d bounding box estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 244–253. [Google Scholar]
  60. Xie, L.; Xiang, C.; Yu, Z.; Xu, G.; Yang, Z.; Cai, D.; He, X. PI-RCNN: An efficient multi-sensor 3D object detector with point-based attentive cont-conv fusion module. Proc. AAAI Conf. Artif. Intell. 2020, 34, 12460–12467. [Google Scholar] [CrossRef]
  61. Huang, T.; Liu, Z.; Chen, X.; Bai, X. Epnet: Enhancing point features with image semantics for 3d object detection. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XV 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 35–52. [Google Scholar]
  62. Jiao, Y.; Jie, Z.; Chen, S.; Chen, J.; Ma, L.; Jiang, Y.G. Msmdfusion: Fusing lidar and camera at multiple scales with multi-depth seeds for 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 21643–21652. [Google Scholar]
  63. Vora, S.; Lang, A.H.; Helou, B.; Beijbom, O. Pointpainting: Sequential fusion for 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4604–4612. [Google Scholar]
  64. Li, Y.; Yu, A.W.; Meng, T.; Caine, B.; Ngiam, J.; Peng, D.; Shen, J.; Lu, Y.; Zhou, D.; Le, Q.V.; et al. Deepfusion: Lidar-camera deep fusion for multi-modal 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 17182–17191. [Google Scholar]
  65. Bai, X.; Hu, Z.; Zhu, X.; Huang, Q.; Chen, Y.; Fu, H.; Tai, C.L. Transfusion: Robust lidar-camera fusion for 3d object detection with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 1090–1099. [Google Scholar]
  66. Qin, Y.; Wang, C.; Kang, Z.; Ma, N.; Li, Z.; Zhang, R. SupFusion: Supervised LiDAR-camera fusion for 3D object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 22014–22024. [Google Scholar]
  67. Nabati, R.; Qi, H. Centerfusion: Center-based radar and camera fusion for 3d object detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2021; pp. 1527–1536. [Google Scholar]
  68. Kim, Y.; Kim, S.; Choi, J.W.; Kum, D. Craft: Camera-radar 3d object detection with spatio-contextual fusion transformer. Proc. AAAI Conf. Artif. Intell. 2023, 37, 1160–1168. [Google Scholar] [CrossRef]
  69. Zhou, T.; Chen, J.; Shi, Y.; Jiang, K.; Yang, M.; Yang, D. Bridging the view disparity between radar and camera features for multi-modal fusion 3d object detection. IEEE Trans. Intell. Veh. 2023, 8, 1523–1535. [Google Scholar] [CrossRef]
  70. Kim, Y.; Shin, J.; Kim, S.; Lee, I.J.; Choi, J.W.; Kum, D. Crn: Camera radar net for accurate, robust, efficient 3d perception. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 17615–17626. [Google Scholar]
  71. Lin, Z.; Liu, Z.; Xia, Z.; Wang, X.; Wang, Y.; Qi, S.; Dong, Y.; Dong, N.; Zhang, L.; Zhu, C. RCBEVDet: Radar-camera Fusion in Bird’s Eye View for 3D Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 14928–14937. [Google Scholar]
  72. Viola, P.; Jones, M.J. Robust real-time face detection. Int. J. Comput. Vis. 2004, 57, 137–154. [Google Scholar] [CrossRef]
  73. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  74. Papageorgiou, C.; Poggio, T. A trainable system for object detection. Int. J. Comput. Vis. 2000, 38, 15–33. [Google Scholar] [CrossRef]
  75. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the 26th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; Volume 1, pp. 1097–1105. [Google Scholar]
  76. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Region-based convolutional networks for accurate object detection and segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 142–158. [Google Scholar] [CrossRef] [PubMed]
  77. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef]
  78. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  79. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  80. Wang, C.Y.; Yeh, I.H.; Liao, H.Y.M. YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information. arXiv 2024, arXiv:2402.13616. [Google Scholar]
  81. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31th International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
  82. Dai, J.; Li, Y.; He, K.; Sun, J. R-FCN: Object detection via region-based fully convolutional networks. In Proceedings of the 30th International Conference on Neural Information Processing Systems, NIPS’16, Barcelona, Spain, 5–10 December 2016; pp. 379–387. [Google Scholar]
  83. Tarchoun, B.; Jegham, I.; Khalifa, A.B.; Alouani, I.; Mahjoub, M.A. Deep cnn-based pedestrian detection for intelligent infrastructure. In Proceedings of the 2020 5th International Conference on Advanced Technologies for Signal and Image Processing (ATSIP), Sousse, Tunisia, 2–5 September 2020; pp. 1–6. [Google Scholar]
  84. Zhou, X.; Zhang, L. SA-FPN: An effective feature pyramid network for crowded human detection. Appl. Intell. 2022, 52, 12556–12568. [Google Scholar] [CrossRef]
  85. Mammeri, A.; Siddiqui, A.J.; Zhao, Y.; Pekilis, B. Vulnerable road users detection based on convolutional neural networks. In Proceedings of the 2020 International Symposium on Networks, Computers and Communications (ISNCC), Montreal, QC, Canada, 20–22 October 2020; pp. 1–6. [Google Scholar]
  86. Sharma, D.; Hade, T.; Tian, Q. Comparison Of Deep Object Detectors On A New Vulnerable Pedestrian Dataset. arXiv 2024, arXiv:2212.06218. [Google Scholar]
  87. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  88. Jinrang, J.; Li, Z.; Shi, Y. MonoUNI: A unified vehicle and infrastructure-side monocular 3d object detection network with sufficient depth clues. In Proceedings of the 37th International Conference on Neural Information Processing Systems, New Orleans, LA, USA, 10–16 December 2023; pp. 11703–11715. [Google Scholar]
  89. Huang, C.R.; Chung, P.C.; Lin, K.W.; Tseng, S.C. Wheelchair Detection Using Cascaded Decision Tree. IEEE Trans. Inf. Technol. Biomed. 2010, 14, 292–300. [Google Scholar] [CrossRef]
  90. Arnold, E.; Al-Jarrah, O.Y.; Dianati, M.; Fallah, S.; Oxtoby, D.; Mouzakitis, A. A survey on 3d object detection methods for autonomous driving applications. IEEE Trans. Intell. Transp. Syst. 2019, 20, 3782–3795. [Google Scholar] [CrossRef]
  91. Sun, Y.; Xu, H.; Wu, J.; Zheng, J.; Dietrich, K.M. 3-D data processing to extract vehicle trajectories from roadside LiDAR data. Transp. Res. Rec. 2018, 2672, 14–22. [Google Scholar] [CrossRef]
  92. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A density-based algorithm for discovering clusters in large spatial databases with noise. kdd 1996, 96, 226–231. [Google Scholar]
  93. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  94. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  95. Ho, T.K. Random decision forests. In Proceedings of the 3rd International Conference on Document Analysis and Recognition, Montreal, QC, Canada, 14–16 August 1995; Volume 1, pp. 278–282. [Google Scholar]
  96. Kononenko, I. Comparison of inductive and naive Bayesian learning approaches to automatic knowledge acquisition. Curr. Trends Knowl. Acquis. 1990, 8, 190. [Google Scholar]
  97. Zhao, J.; Xu, H.; Liu, H.; Wu, J.; Zheng, Y.; Wu, D. Detection and tracking of pedestrians and vehicles using roadside LiDAR sensors. Transp. Res. Part C: Emerg. Technol. 2019, 100, 68–87. [Google Scholar] [CrossRef]
  98. Hecht-Nielsen, R. Theory of the backpropagation neural network. In Proceedings of the International 1989 Joint Conference on Neural Networks, Washington, DC, USA, 16–18 October 1989; Volume 1, pp. 593–605. [Google Scholar] [CrossRef]
  99. Song, Y.; Tian, J.; Li, T.; Sun, R.; Zhang, H.; Wu, J.; Song, X. Road-Users Classification Utilizing Roadside Light Detection and Ranging Data; Technical Report, SAE Technical Paper; Society of Automotive Engineers: Warrendale, PA, USA, 2020. [Google Scholar]
  100. Specht, D.F. Probabilistic neural networks. Neural Netw. 1990, 3, 109–118. [Google Scholar] [CrossRef]
  101. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  102. Gong, Z.; Wang, Z.; Yu, G.; Liu, W.; Yang, S.; Zhou, B. FecNet: A Feature Enhancement and Cascade Network for Object Detection Using Roadside LiDAR. IEEE Sens. J. 2023, 23, 23780–23791. [Google Scholar] [CrossRef]
  103. Xu, Y.; Liu, W.; Qi, Y.; Hu, Y.; Zhang, W. A Common Traffic Object Recognition Method Based on Roadside LiDAR. In Proceedings of the 2022 IEEE 7th International Conference on Intelligent Transportation Engineering (ICITE), Beijing, China, 11–13 November 2022; pp. 436–441. [Google Scholar]
  104. Zhang, L.; Zheng, J.; Sun, R.; Tao, Y. Gc-net: Gridding and clustering for traffic object detection with roadside lidar. IEEE Intell. Syst. 2020, 36, 104–113. [Google Scholar]
  105. Blanch, M.R.; Li, Z.; Escalera, S.; Nasrollahi, K. LiDAR-Assisted 3D Human Detection for Video Surveillance. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 4–8 January 2024; pp. 123–131. [Google Scholar]
  106. Shi, H.; Hou, D.; Li, X. Center-Aware 3D Object Detection with Attention Mechanism Based on Roadside LiDAR. Sustainability 2023, 15, 2628. [Google Scholar] [CrossRef]
  107. Yu, H.; Luo, Y.; Shu, M.; Huo, Y.; Yang, Z.; Shi, Y.; Guo, Z.; Li, H.; Hu, X.; Yuan, J. Dair-v2x: A large-scale dataset for vehicle-infrastructure cooperative 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 21361–21370. [Google Scholar]
  108. Wang, Y.; Solomon, J.M. Object dgcnn: 3d object detection using dynamic graphs. Adv. Neural Inf. Process. Syst. 2021, 34, 20745–20758. [Google Scholar]
  109. Fan, L.; Wang, F.; Wang, N.; Zhang, Z.X. Fully sparse 3d object detection. Adv. Neural Inf. Process. Syst. 2022, 35, 351–363. [Google Scholar]
  110. Amosa, T.I.; Sebastian, P.; Izhar, L.I.; Ibrahim, O.; Ayinla, L.S.; Bahashwan, A.A.; Bala, A.; Samaila, Y.A. Multi-camera multi-object tracking: A review of current trends and future advances. Neurocomputing 2023, 552, 126558. [Google Scholar] [CrossRef]
  111. Zimmer, W.; Birkner, J.; Brucker, M.; Nguyen, H.T.; Petrovski, S.; Wang, B.; Knoll, A.C. Infradet3d: Multi-modal 3d object detection based on roadside infrastructure camera and lidar sensors. In Proceedings of the 2023 IEEE Intelligent Vehicles Symposium (IV), Anchorage, AK, USA, 4–7 June 2023; pp. 1–8. [Google Scholar]
  112. Meissner, D.; Reuter, S.; Dietmayer, K. Real-time detection and tracking of pedestrians at intersections using a network of laserscanners. In Proceedings of the 2012 IEEE Intelligent Vehicles Symposium, Madrid, Spain, 3–7 June 2012; pp. 630–635. [Google Scholar]
  113. Strigel, E.; Meissner, D.; Seeliger, F.; Wilking, B.; Dietmayer, K. The ko-per intersection laserscanner and video dataset. In Proceedings of the 17th International IEEE Conference on Intelligent Transportation Systems (ITSC), Qingdao, China, 8–11 October 2014; pp. 1900–1901. [Google Scholar]
  114. Stauffer, C.; Grimson, W.E.L. Adaptive background mixture models for real-time tracking. In Proceedings of the 1999 IEEE computer society Conference on Computer Vision and pattern recognition (Cat. No PR00149), Fort Collins, CO, USA, 23–25 June 1999; Volume 2, pp. 246–252. [Google Scholar]
  115. Zhou, Q.Y.; Park, J.; Koltun, V. Fast Global Registration. In Computer Vision—ECCV 2016; Lecture Notes in Computer Science; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; Volume 9906, pp. 766–782. [Google Scholar] [CrossRef]
  116. Besl, P.J.; McKay, N.D. Method for registration of 3-D shapes. In Sensor Fusion IV: Control Paradigms and Data Structures; SPIE: Bellingham, WA, USA, 1992; Volume 1611, pp. 586–606. [Google Scholar]
  117. Bai, Z.; Wu, G.; Barth, M.J.; Liu, Y.; Sisbot, E.A.; Oguchi, K. Pillargrid: Deep learning-based cooperative perception for 3d object detection from onboard-roadside lidar. In Proceedings of the 2022 IEEE 25th International Conference on Intelligent Transportation Systems (ITSC), Macau, China, 8–12 October 2022; pp. 1743–1749. [Google Scholar]
  118. Zimmer, W.; Creß, C.; Nguyen, H.T.; Knoll, A.C. Tumtraf intersection dataset: All you need for urban 3d camera-lidar roadside perception. In Proceedings of the 2023 IEEE 26th International Conference on Intelligent Transportation Systems (ITSC), Bilbao, Spain, 24–28 September 2023; pp. 1030–1037. [Google Scholar]
  119. Yao, J.; Zhou, J.; Wang, Y.; Gao, Z.; Hu, W. Infrastructure-assisted 3D detection networks based on camera-lidar early fusion strategy. Neurocomputing 2024, 600, 128180. [Google Scholar] [CrossRef]
  120. Zhou, Y.; Dong, Y.; Hou, F.; Wu, J. Review on millimeter-wave radar and camera fusion technology. Sustainability 2022, 14, 5114. [Google Scholar] [CrossRef]
  121. Bai, J.; Li, S.; Zhang, H.; Huang, L.; Wang, P. Robust Target Detection and Tracking Algorithm Based on Roadside Radar and Camera. Sensors 2021, 21, 1116. [Google Scholar] [CrossRef]
  122. Wang, Y.; Jiang, Z.; Li, Y.; Hwang, J.N.; Xing, G.; Liu, H. RODNet: A Real-Time Radar Object Detection Network Cross-Supervised by Camera-Radar Fused Object 3D Localization. IEEE J. Sel. Top. Signal Process. 2021, 15, 954–967. [Google Scholar] [CrossRef]
  123. Liu, P.; Yu, G.; Wang, Z.; Zhou, B.; Chen, P. Object classification based on enhanced evidence theory: Radar–vision fusion approach for roadside application. IEEE Trans. Instrum. Meas. 2022, 71, 5006412. [Google Scholar] [CrossRef]
  124. Xu, D.; Ouyang, W.; Ricci, E.; Wang, X.; Sebe, N. Learning cross-modal deep representations for robust pedestrian detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5363–5371. [Google Scholar]
  125. Jia, X.; Zhu, C.; Li, M.; Tang, W.; Zhou, W. LLVIP: A visible-infrared paired dataset for low-light vision. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Online, 11–17 October 2021; pp. 3496–3504. [Google Scholar]
  126. Pei, D.; Jing, M.; Liu, H.; Sun, F.; Jiang, L. A fast RetinaNet fusion framework for multi-spectral pedestrian detection. Infrared Phys. Technol. 2020, 105, 103178. [Google Scholar] [CrossRef]
  127. Choi, Y.; Kim, N.; Hwang, S.; Park, K.; Yoon, J.S.; An, K.; Kweon, I.S. KAIST multi-spectral day/night data set for autonomous and assisted driving. IEEE Trans. Intell. Transp. Syst. 2018, 19, 934–948. [Google Scholar] [CrossRef]
  128. Zhang, H.; Xu, H.; Tian, X.; Jiang, J.; Ma, J. Image fusion meets deep learning: A survey and perspective. Inf. Fusion 2021, 76, 323–336. [Google Scholar] [CrossRef]
  129. Zhang, Y.; Zeng, W.; Jin, S.; Qian, C.; Luo, P.; Liu, W. When Pedestrian Detection Meets Multi-Modal Learning: Generalist Model and Benchmark Dataset. arXiv 2024, arXiv:2407.10125. [Google Scholar]
  130. Bai, Z.; Wu, G.; Barth, M.J.; Liu, Y.; Sisbot, E.A.; Oguchi, K.; Huang, Z. A survey and framework of cooperative perception: From heterogeneous singleton to hierarchical cooperation. IEEE Trans. Intell. Transp. Syst. 2024, 25, 15191–15209. [Google Scholar]
  131. Eshel, R.; Moses, Y. Homography based multiple camera detection and tracking of people in a dense crowd. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  132. Wang, X. Intelligent multi-camera video surveillance: A review. Pattern Recognit. Lett. 2013, 34, 3–19. [Google Scholar] [CrossRef]
  133. Rauch, A.; Klanner, F.; Rasshofer, R.; Dietmayer, K. Car2x-based perception in a high-level fusion architecture for cooperative perception systems. In Proceedings of the 2012 IEEE Intelligent Vehicles Symposium, Madrid, Spain, 3–7 June 2012; pp. 270–275. [Google Scholar]
  134. Neubeck, A.; Van Gool, L. Efficient non-maximum suppression. In Proceedings of the 18th International Conference on Pattern Recognition (ICPR’06), Hong Kong, China, 20–24 August 2006; Volume 3, pp. 850–855. [Google Scholar]
  135. Arnold, E.; Dianati, M.; de Temple, R.; Fallah, S. Cooperative perception for 3D object detection in driving scenarios using infrastructure sensors. IEEE Trans. Intell. Transp. Syst. 2020, 23, 1852–1864. [Google Scholar]
  136. Sun, P.; Kretzschmar, H.; Dotiwalla, X.; Chouard, A.; Patnaik, V.; Tsui, P.; Guo, J.; Zhou, Y.; Chai, Y.; Caine, B. Scalability in perception for autonomous driving: Waymo open dataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2446–2454. [Google Scholar]
  137. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. nuscenes: A multimodal dataset for autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11621–11631. [Google Scholar]
  138. Dollar, P.; Wojek, C.; Schiele, B.; Perona, P. Pedestrian Detection: An Evaluation of the State of the Art. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 743–761. [Google Scholar] [CrossRef] [PubMed]
  139. Zhang, S.; Benenson, R.; Schiele, B. CityPersons: A Diverse Dataset for Pedestrian Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4457–4465. [Google Scholar] [CrossRef]
  140. Braun, M.; Krebs, S.; Flohr, F.; Gavrila, D.M. EuroCity Persons: A Novel Benchmark for Person Detection in Traffic Scenes. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 1844–1861. [Google Scholar] [CrossRef]
  141. Neumann, L.; Karg, M.; Zhang, S.; Scharfenberger, C.; Piegert, E.; Mistr, S.; Prokofyeva, O.; Thiel, R.; Vedaldi, A.; Zisserman, A.; et al. NightOwls: A Pedestrians at Night Dataset. In Proceedings of the Computer Vision—ACCV 2018, Perth, Australia, 2–6 December 2018; Jawahar, C.V., Li, H., Mori, G., Schindler, K., Eds.; Springer: Cham, Switzerland, 2019; pp. 691–705. [Google Scholar] [CrossRef]
  142. Pang, Y.; Cao, J.; Li, Y.; Xie, J.; Sun, H.; Gong, J. TJU-DHD: A Diverse High-Resolution Dataset for Object Detection. arXiv 2020, arXiv:2011.09170. [Google Scholar] [CrossRef] [PubMed]
  143. Shah, A.; Lamare, J.B.; Anh, T.N.; Hauptmann, A. CADP: A Novel Dataset for CCTV Traffic Camera based Accident Analysis. arXiv 2018, arXiv:1809.05782. [Google Scholar]
  144. Yongqiang, D.; Dengjiang, W.; Gang, C.; Bing, M.; Xijia, G.; Yajun, W.; Jianchao, L.; Yanming, F.; Juanjuan, L. BAAI-VANJEE Roadside Dataset: Towards the Connected Automated Vehicle Highway technologies in Challenging Environments of China. arXiv 2021, arXiv:2105.14370. [Google Scholar]
  145. Creß, C.; Zimmer, W.; Purschke, N.; Doan, B.N.; Kirchner, S.; Lakshminarasimhan, V.; Strand, L.; Knoll, A.C. TUMTraf event: Calibration and fusion resulting in a dataset for roadside event-based and RGB cameras. IEEE Trans. Intell. Veh. 2024, 9, 5186–5203. [Google Scholar]
  146. Ye, X.; Shu, M.; Li, H.; Shi, Y.; Li, Y.; Wang, G.; Tan, X.; Ding, E. Rope3d: The roadside perception dataset for autonomous driving and monocular 3d object detection task. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 21341–21350. [Google Scholar]
  147. Zimmer, W.; Wardana, G.A.; Sritharan, S.; Zhou, X.; Song, R.; Knoll, A.C. TUMTraf V2X Cooperative Perception Dataset. arXiv 2024, arXiv:2403.01316. [Google Scholar]
  148. Xiang, H.; Zheng, Z.; Xia, X.; Xu, R.; Gao, L.; Zhou, Z.; Han, X.; Ji, X.; Li, M.; Meng, Z.; et al. V2X-Real: A Largs-Scale Dataset for Vehicle-to-Everything Cooperative Perception. arXiv 2024, arXiv:2403.16034. [Google Scholar]
  149. Li, Y.; Ma, D.; An, Z.; Wang, Z.; Zhong, Y.; Chen, S.; Feng, C. V2X-Sim: Multi-agent collaborative perception dataset and benchmark for autonomous driving. IEEE Robot. Autom. Lett. 2022, 7, 10914–10921. [Google Scholar] [CrossRef]
  150. Long, Y.; Morris, D.; Liu, X.; Castro, M.; Chakravarty, P.; Narayanan, P. Radar-camera pixel depth association for depth completion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 12507–12516. [Google Scholar]
  151. Lin, J.T.; Dai, D.; Van Gool, L. Depth estimation from monocular images and sparse radar data. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 25–29 October 2020; pp. 10233–10240. [Google Scholar]
  152. U.S. DOT. ITS—Intersection Safety Challenge; U.S. Department of Transportation: Washington, DC, USA, 2024.
  153. Yang, L.; Liu, S.; Feng, S.; Wang, H.; Zhao, X.; Qu, G.; Fang, S. Generation of critical pedestrian scenarios for autonomous vehicle testing. Accid. Anal. Prev. 2025, 214, 107962. [Google Scholar] [CrossRef] [PubMed]
  154. Cai, X.; Jiang, W.; Xu, R.; Zhao, W.; Ma, J.; Liu, S.; Li, Y. Analyzing infrastructure lidar placement with realistic lidar simulation library. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 5581–5587. [Google Scholar]
  155. Shao, S.; Zhao, Z.; Li, B.; Xiao, T.; Yu, G.; Zhang, X.; Sun, J. CrowdHuman: A Benchmark for Detecting Human in a Crowd. arXiv 2018, arXiv:1805.00123. [Google Scholar]
  156. Hwang, S.; Park, J.; Kim, N.; Choi, Y.; Kweon, I.S. Multispectral pedestrian detection: Benchmark dataset and baseline. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1037–1045. [Google Scholar] [CrossRef]
  157. Ojala, R.; Vepsäläinen, J.; Hanhirova, J.; Hirvisalo, V.; Tammi, K. Novel convolutional neural network-based roadside unit for accurate pedestrian localisation. IEEE Trans. Intell. Transp. Syst. 2019, 21, 3756–3765. [Google Scholar]
  158. Cong, P.; Zhu, X.; Qiao, F.; Ren, Y.; Peng, X.; Hou, Y.; Xu, L.; Yang, R.; Manocha, D.; Ma, Y. STCrowd: A Multimodal Dataset for Pedestrian Perception in Crowded Scenes. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 19576–19585. [Google Scholar] [CrossRef]
  159. Song, H.; Choi, I.K.; Ko, M.S.; Bae, J.; Kwak, S.; Yoo, J. Vulnerable pedestrian detection and tracking using deep learning. In Proceedings of the 2018 International Conference on Electronics, Information, and Communication (ICEIC), Honolulu, HI, USA, 24–27 January 2018; pp. 1–2. [Google Scholar] [CrossRef]
Figure 1. Organization of this review.
Figure 1. Organization of this review.
Applsci 15 03797 g001
Figure 2. Ratio of detection models.
Figure 2. Ratio of detection models.
Applsci 15 03797 g002
Figure 3. Matrix of sensor and modal (this figure was designed using resources from Flaticon.com).
Figure 3. Matrix of sensor and modal (this figure was designed using resources from Flaticon.com).
Applsci 15 03797 g003
Figure 4. The fusion scheme.
Figure 4. The fusion scheme.
Applsci 15 03797 g004
Table 1. Detection Sensors.
Table 1. Detection Sensors.
SensorsAdvantagesDisadvantagesCostDurability
Monocular CameraRich Information (Signs and Objects)Strong Light Dependence; Poor Environmental Adaptability; Low Ranging Accuracy; Low Privacy; Narrow Field of View (FOV); Limited 3D InformationLowModerate
LiDARSimple Structure; Low Power Consumption; Strong Environmental Adaptability; Extensive FOV; High Accuracy; High Privacy; Rich 3D InformationInsufficient Ability to Read Signs and Distinguish ColorsHighModerate to Low (moving parts)
RADARLong Detection Distance; High Accuracy in Object Position and Speed; Strong Environmental Adaptability; High PrivacyInsufficient Accuracy in Object Detection and Classification; Insufficient Ability to Read Signs and Distinguish ColorsModerateHigh
Depth CameraAccurate Depth Perception; Moderate Light DependenceInsufficient Range and Resolution in Depth; Poor Environmental AdaptabilityModerateModerate
Infrared Camera (IR)Low Light Dependence; Strong Environmental AdaptabilityNarrow FOV; Insufficient Ability to Read Signs and Distinguish ColorsModerateModerate
Fisheye CameraExtensive FOV; Rich Information (Signs and Objects)Low PrivacyModerateModerate to Low (lens scratches)
Event-based CameraRobustness to Motion Blur; Low Power Consumption; High Dynamic Range; Minimal Data RateLimited Spatial Resolution; Complex Data ProcessingHighHigh
Table 2. Overview of generic deep learning (DL)-based vulnerable road user (VRU) detection methods.
Table 2. Overview of generic deep learning (DL)-based vulnerable road user (VRU) detection methods.
MethodsYearDimensionsSensors
2D3DCameraLiDARRADAR
Faster R-CNN [17]2015
SSD [18]2016
RetinaNet [19]2017
FPN [20]2017
YOLO series [21,22,23,24]-
CornerNet [25]2018
CenterNet [26]2019
ViT [27]2021
DETR [28]2020
Swin Transformer [29]2021
PVT [30]2021
FCOS3D [31]2021
DETR3D [32]2022
PETR [33]2022
BEVDepth [34]2023
BEVHeight [35]2023
CBR [36]2023
CoBEV [37]2023
VeloFCN [38]2016
PIXOR [39]2018
VoxelNet [8]2018
SECOND [40]2018
PointPillars [41]2019
CenterPoint [42]2021
PillarNet [43]2022
PointRCNN [44]2019
PV-RCNN [45]2020
PV-RCNN++ [46]2023
MV3D [47]2017
AVOD [48]2018
Cont Fuse [49]2018
SACNet [50]2019
FuseSeg [51]2020
BEVFusion [52]2023
CoBEVFusion [53]2023
BroadBEV [54]2024
MVXNet [55]2019
3D-CVF [56]2020
VPF-Net [57]2022
PointAugmenting [58]2021
PointFusion [59]2018
PI-RCNN [60]2020
EPNet [61]2020
MSMDFusion [62]2023
PointPainting [63]2020
DeepFusion [64]2022
TransFusion [65]2022
SupFusion [66]2023
CenterFusion [67]2021
CRAFT [68]2023
RCBEV4d [69]2023
CRN [70]2023
RCBEVDet [71]2024
Table 3. Comparison of different LiDAR detection methods by category.
Table 3. Comparison of different LiDAR detection methods by category.
Category Based on Feature Extraction MethodMethodsProsCons
Front view (FV)-basedVeloFCN [38]Good for detecting VRUs in foreground of the scenario.Struggles with occlusions and side or distant VRUs.
Bird’s eye view (BEV)-basedPIXOR [39]Comprehensive view of surroundings, improved occlusion handling.Loss of fine-grained details in crowded environments.
Voxel-basedVoxelNet [8],
SECOND [40],
PointPillars [41],
CenterPoint [42],
FecNet [102],
PillarNet [43]
Effective for large-scale, structured objects.Struggles with small or non-rigid objects, such as VRUs.
Point-basedPointRCNN [44],
PV-RCNN [45],
PV-RCNN++ [46]
Best for small, non-rigid objects, better occlusion handling.Computationally expensive, slower inference.
Table 4. Comparison of different generic camera and LiDAR fusion methods by category.
Table 4. Comparison of different generic camera and LiDAR fusion methods by category.
Category Based on Feature Extraction MethodMethodsComments
FV/BEV- and feature map-basedMV3D [47], AVOD [48], Cont Fuse [49], SACNet [50], FuseSeg [51], BEVFusion [52], CoBEVFusion [53], BroadBEV [54]BEV provides spatial context for VRUs, and camera feature maps enhance object classification; FV may not be as robust as BEV for VRU detection, especially for detecting VRUs that might be occluded or outside the direct front-view.
Voxel- and feature map-basedMVXNet [55], 3D-CVF [56], VPF-Net [57], PointAugmenting [58]Voxel grid provides depth but lacks high resolution for fine VRU detection. Combined with feature maps, detection improves but may struggle in cluttered environments.
Point- and feature map-basedPointFusion [59], PI-RCNN [60], EPNet [61], MSMDFusion [62]Precise 3D spatial data from LiDAR combined with camera’s semantic information helps detect VRUs, even in occlusions or complex environments.
Point- and mask-basedPointPainting [63]Combines precise depth information with semantic segmentation for accurate VRU detection, especially in occlusions or partial visibility scenarios.
Deep feature-basedDeepFusion [64], TransFusion [65], SupFusion [66]Advanced fusion of high-level features from both sensors leads to the best detection of VRUs, especially in dynamic or complex scenes.
Table 5. Infrastructure-based single-sensor datasets.
Table 5. Infrastructure-based single-sensor datasets.
DatasetsData SourceObject ClassContent
Caltech [138]
  • 1× camera
    -
    Resolution: 640 × 480
  • Pedestrian
  • Ignore Region
  • 250K frames of images
  • 350K person annotations and 126K ignore region annotations
  • With occlusion labels
CityPersons [139]
  • Cameras
  • Pedestrian
  • Rider
  • Sitting and Other Person
  • Ignore Region
  • 5K frames of images
  • 35K person annotations and 13K ignore region annotations
  • Data collected from different cities and seasons with reasonable distribution in occlusion ratio
EuroCity Persons [140]
  • 1× camera
    -
    Resolution: 1920 × 1024
  • Pedestrian
  • Rider
  • Ignore Region
  • 47K frames of images
  • 218K person annotations, 19K rider annotations, 95K ignore region annotations, and 211K body orientation annotations
  • Data collected from 31 cities of 12 European countries
  • Including day and night, dry and wet scenarios
NightOwls [141]
  • 1× camera
    -
    Resolution: 1024 × 640
  • Pedestrian
  • Cyclist
  • Motorcyclist
  • Ignore Region
  • 279K frames of images
  • 42K annotations with height and pose information
  • Focusing on the night scenario
  • Including rain and snow weather conditions
TJU-DHD [142]
  • 1× camera (Traffic Scene)
    -
    Resolution: 1624 × 1200
  • 1× camera (Campus Scene)
    -
    Resolution: 2560 × 1440
  • Pedestrian
  • Rider
  • Car
  • Van
  • Truck
  • 60K/55K frames of images in traffic/campus
  • 331K/377K annotations in traffic/campus with truncation levels
  • With full-body and visible-body annotations
  • Including variances in illumination, scene, weather, and season
CADP [143]
  • YouTube videos of traffic accidents
  • Pedestrian
  • Two-wheeler
  • Three-wheeler
  • Car
  • Bus
  • Others
  • 518K frames of images
  • 1416 accident annotations
BGVP [86]
  • Images from public domains (mainly the Internet)
    -
    Median Resolution: 600 × 408
  • Children w/o Disability
  • Elderly w/o Disability
  • With Disability
  • Non-vulnerable
  • 2000 frames of images
  • 5932 annotations: 1646 in children w/o disability, 815 in elderly w/o disability, 942 in with disability, and 2529 in non-vulnerable
Table 6. Infrastructure-based multi-sensor datasets.
Table 6. Infrastructure-based multi-sensor datasets.
DatasetsData SourceSynchronizationObject ClassContent
BAAI-VANJEE [144]
  • 1× 32-layer LiDAR
  • 4× cameras
20% synchronized
  • Pedestrian
  • Cyclist
  • Tricyclist
  • Motorcyclist
  • Car
  • Van
  • Cargo
  • Bus
  • Truck
  • Semi-trailing tractor
  • Special vehicle (e.g., ambulance.)
  • Roadblock
  • 2500 frames of LiDAR data, 5000 frames of images
  • 74K 3D object annotations and 105K 2D object annotations
DAIR-V2X-I [107]
  • 1× 300-layer LiDAR
  • 1× camera
Yes
  • Pedestrian
  • Cyclist
  • Tricyclist
  • Barrow
  • Motorcyclist
  • Car
  • Bus
  • Van
  • Truck
  • Traffic Cone
  • 10,084 frames of LiDAR data, 10,084 frames of images
  • 493K 3D object annotations
TUMTraf Intersection Dataset [118]
  • 2× 64-layer LiDARs
  • 2× cameras
Yes
  • Pedestrian
  • Cyclist
  • Motorcyclist
  • Car
  • Bus
  • Van
  • Truck
  • Trailer
  • Emergency Vehicle
  • Other
  • 4800 frames of LiDAR data, 4800 frames of images
  • 57.4K 3D object annotations with track IDs
LLVIP [125]
  • 1× binocular camera with visible and infrared images
Yes
  • Pedestrian
  • Cyclist
  • 30K frames of images
  • Focusing on low-light conditions
TUMTraf Event [145]
  • 1× camera
  • 1× event-based camera
Yes
  • Pedestrian
  • Cyclist
  • Motorcyclist
  • Car
  • Bus
  • Truck
  • Trailer
  • 4111 frames of data
  • 50K 2D object annotations
MMPD [129]
  • Multi-modal data from existing public datasets
  • Diverse modality combinations: RGB+IR, RGB+Depth, RGB+LiDAR, RGB+Event
Yes
  • Pedestrian
  • 260K frames of data
  • Covering various scenarios: surveillance, automobile, robotics, and outdoor and in-door environments
Table 7. Cooperative multi-sensor datasets.
Table 7. Cooperative multi-sensor datasets.
DatasetsData SourceSynchronizationObject ClassContent
Rope3D [146]
  • Infrastructure
    -
    Cameras
  • Vehicle
    -
    1× 40/300-layer LiDAR
Yes
  • Pedestrian
  • Cyclist
  • Tricyclist
  • Barrow
  • Motorcyclist
  • Car
  • Bus
  • Van
  • Truck
  • Traffic cone
  • Triangle plate
  • Unknown-unmovable
  • Unknown-movable
  • 50K frames of images
  • 1.5 M 3D object annotations and 670K 2D object annotations for objects not scanned by the laser
DAIR-V2X-C [107]
  • Infrastructure
    -
    1× 300-layer LiDAR
    -
    1× camera
  • Vehicle
    -
    1× 40-layer LiDAR
    -
    1× camera
Yes
  • Pedestrian
  • Cyclist
  • Tricyclist
  • Barrow
  • Motorcyclist
  • Car
  • Bus
  • Van
  • Truck
  • Traffic cone
  • 38,845 frames of LiDAR data, 38,845 frames of images
  • 464K 3D object annotations
TUMTraf V2X Cooperative Perception Dataset [147]
  • Infrastructure
    -
    1× 64-layer LiDAR
    -
    4× cameras
  • Vehicle
    -
    1× 32-layer LiDAR
    -
    1× camera
Yes
  • Pedestrian
  • Cyclist
  • Motorcyclist
  • Car
  • Van
  • Bus
  • Truck
  • Trailer
  • 2000 frames of LiDAR data, 5000 frames of images
  • 30K 3D bounding boxes with track IDs
V2X-Real [148]
  • Infrastructure
    -
    1× 64/128-layer LiDAR
    -
    2× cameras
  • Vehicle
    -
    1× 128-layer LiDAR
    -
    4× cameras
Yes
  • Pedestrian
  • Scooter
  • Cyclist
  • Motorcyclist
  • Car
  • Van
  • Bus
  • Truck
  • Box truck
  • Barrier
  • 33K frames of LiDAR data, 171K frames of images
  • 1.2 M 3D object annotations
V2X-SIM [149]
  • Infrastructure
    -
    1× LiDAR
    -
    1× semantic LiDAR
    -
    4× cameras
    -
    1× BEV semantic segmentation camera
  • Vehicle
    -
    1× LiDAR
    -
    1× semantic LiDAR
    -
    6× RGB/depth/semantic segmentation cameras
    -
    1× BEV semantic segmentation camera
Yes
  • Pedestrian
  • Vehicle
  • Sidewalk
  • Terrain road
  • Building
  • Vegetation
  • 10K frames of data
  • 47.2K 3D object annotations
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Z.; Wei, C.; Wu, G.; Barth, M.J. Vulnerable Road User Detection for Roadside-Assisted Safety Protection: A Comprehensive Survey. Appl. Sci. 2025, 15, 3797. https://doi.org/10.3390/app15073797

AMA Style

Zhang Z, Wei C, Wu G, Barth MJ. Vulnerable Road User Detection for Roadside-Assisted Safety Protection: A Comprehensive Survey. Applied Sciences. 2025; 15(7):3797. https://doi.org/10.3390/app15073797

Chicago/Turabian Style

Zhang, Ziyan, Chuheng Wei, Guoyuan Wu, and Matthew J. Barth. 2025. "Vulnerable Road User Detection for Roadside-Assisted Safety Protection: A Comprehensive Survey" Applied Sciences 15, no. 7: 3797. https://doi.org/10.3390/app15073797

APA Style

Zhang, Z., Wei, C., Wu, G., & Barth, M. J. (2025). Vulnerable Road User Detection for Roadside-Assisted Safety Protection: A Comprehensive Survey. Applied Sciences, 15(7), 3797. https://doi.org/10.3390/app15073797

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop