Next Article in Journal
Monitoring Distracted Driving Behaviours with Smartphones: An Extended Systematic Literature Review
Next Article in Special Issue
Real-Time Three-Dimensional Pedestrian Localization System Using Smartphones
Previous Article in Journal
Rapidly Determine the Maximum Power Point in the Parallel Configuration of the Photovoltaic System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning and Geometry Flow Vector Using Estimating Vehicle Cuboid Technology in a Monovision Environment

1
Department of AI and Big Data, Soonchunhyang University, 22 Soonchunhyang-ro, Asan 31538, Republic of Korea
2
Department of Civil and Environmental Engineering, Korea Advanced Institute of Science and Technology, 291 Daehak-ro, Yuseong-gu, Daejeon 34141, Republic of Korea
3
Department of Software, Sangmyung University, Cheonan 31066, Republic of Korea
4
School of Artificial Intelligence Convergence, Hallym University, Chuncheon 24252, Republic of Korea
*
Authors to whom correspondence should be addressed.
The authors corresponded and supervised equally.
Sensors 2023, 23(17), 7504; https://doi.org/10.3390/s23177504
Submission received: 10 July 2023 / Revised: 24 July 2023 / Accepted: 11 August 2023 / Published: 29 August 2023
(This article belongs to the Special Issue AI and Sensors in Smart Cities)

Abstract

:
This study introduces a novel model for accurately estimating the cuboid of a road vehicle using a monovision sensor and road geometry information. By leveraging object detection models and core vectors, the proposed model overcomes the limitations of multi-sensor setups and provides a cost-effective solution. The model demonstrates promising results in accurately estimating cuboids by utilizing the magnitudes of core vectors and considering the average ratio of distances. This research contributes to the field of intelligent transportation by offering a practical and efficient approach to 3D bounding box estimation using monovision sensors. We validated feasibility and applicability are through real-world road images captured by CCTV cameras.

1. Introduction

The rapid progress in information and communication technology (ICT) has prompted cities worldwide to embrace the concept of smart cities, leveraging intelligent platforms to enhance various aspects of urban life [1,2,3,4,5]. In particular, the field of intelligent transportation has witnessed significant transformations, aiming to automatically gather crucial information about vehicle positions, speeds, and types using vision sensors [6,7]. These data play a pivotal role in enabling intelligent signal control [8], analyzing traffic risk behavior [1,9], predicting traffic states [10], and more [3,11,12]. Consequently, there is a growing emphasis on the utilization of vision sensors to detect, track, and identify road users.
Meanwhile, object detection stands as a fundamental challenge in the realm of computer vision [13]. Over the past few years, extensive research has been conducted in 2D object detection using deep learning models such as convolutional neural networks (CNNs), which are widely used deep learning models, and their applications [14,15,16,17]. Nowadays, the You Only Look Once (YOLO) series models are frequently used to detect objects in images [13,18]. For example, in 2016, Redmon et al. pioneered a unique convolutional network that simultaneously predicts multiple bounding boxes and their associated class probabilities, enhancing performance speed [19]. Fast forward to 2018, the YOLOv3 model was developed by building on YOLO’s foundation, addressing the challenge of detecting small-scale objects and achieving a balance between efficiency and accuracy [18]. Bochkovskiy et al., in an effort to optimize the performance of high frames-per-second (FPS) videos, engineered the YOLOv4 model, introducing improvements to the backbone, activation function, and loss function [20]. More recently, YOLOv7 has been launched, dramatically improving detection accuracy without escalating the inference cost for real-time object detection. In parallel, Tian et al. proposed a revolutionary framework, FCOS, that eliminates the need for proposals and anchors, significantly cutting down the number of design parameters [21,22].
Recent focus has shifted towards a more fundamental challenge in the field: the accurate estimation of 3D bounding boxes for vehicles within the surrounding environment. Ensuring the precise perception of spatial dimensions is crucial for a multitude of applications, including object detection, tracking, and collision avoidance. Traditionally, the estimation of 3D bounding boxes has relied on multi-sensor setups, such as the integration of multiple RGB cameras or the fusion of camera and Light Detection and Ranging (LiDAR) sensors [23]. Significant literature has explored the utilization of multiple RGB cameras [24], RGB-D cameras (such as Microsoft Kinect) [25], and sensor fusion techniques [26] for accurate 3D bounding box recognition. For example, Li et al. made significant contributions with their introduction of a 3D object detection model termed Stereo R-CNN, a model that is based on stereovision techniques [24]. Striving for superior performance, Ku et al. advanced a distinctive model named AVOD. This model taps into the potential of LiDAR point clouds and RGB images to undertake a variety of tasks, including 3D localization, orientation estimation, and category classification [26]. Building on the versatility of RGB-D images, Deng et al. utilized RGB-Depth Images for object detection. Their approach aims to identify object locations within the 3D environment and gauge their physical dimensions and pose, proving effective even in situations where objects overlap [25].
However, these configurations come with significant costs, calibration complexities, maintenance requirements, and the absence of large-scale annotated datasets tailored for multiple sensing-based approaches, similar to the impact of ImageNet in 2D object detection [27]. Even though the expenditure associated with a stereo camera is not as hefty as that of LiDAR, it still presents challenges, such as the complexity of calibration and the need for extensive work on stereo matching to minimize disparity errors [28]. The maintenance costs for LiDAR and stereo cameras are considerably higher compared to Closed-Circuit Television (CCTV) cameras [29]. As a result, there has been a rising demand for research on 3D object detection using a single-vision sensor. Researchers have explored pioneering approaches to address 3D object detection using monocular cameras. For instance, Li and his team introduced Mono3D, a system that leverages a mono-camera to employ segmentation, context, and location cues for precise 3D object detection [29]. Similarly, Chabot and colleagues proposed a two-step approach that predicts 2D boxes, vehicle geometry, and visibility attributes first, followed by recovering 3D orientations and locations [30]. Mousavian and his team advanced a simplified network architecture for 3D object prediction, utilizing projective geometry constraints for 3D object bounding box detection and employing deep CNN for object orientation and size regression [31].
Despite the promising performance of current research based on monocular cameras, a significant constraint remains: the need for a high-quality labeled dataset specifically for model training. The model’s performance relies heavily on the quality of the 3D labeled dataset. Therefore, it is crucial to explore more efficient and low computational complexity methods for handling 3D object detection. In summary, while multi-sensor setups offer advantages in certain scenarios, they also come with various challenges and costs. The emerging research on 3D object detection using monocular cameras has yielded promising results, but the need for high-quality labeled datasets remains a critical aspect to be addressed for further improvements. Efficient and low computational complexity techniques are essential to enhance the performance of 3D object detection using single vision sensors.
In this study, we propose a novel model named “Vehiclectron” for accurately estimating the 3D bounding boxes (cuboids) of vehicles on roads. Our model leverages a monovision sensor, specifically a CCTV camera, along with road geometry information. To begin, the proposed model detects vehicles in video footage using object detection models. Subsequently, three core vectors are extracted: the flow vector, orthogonal vector, and perpendicular vector. The flow vector represents the front direction of the vehicle, which is determined by considering the directions of traffic flow and road geometry. The orthogonal vector is perpendicular to the flow vector and serves as the basis for the cuboid orthogonal to the vehicle. Lastly, the perpendicular vector points vertically downwards within the vehicle in the virtual z-axis of the 2D image.
By utilizing these three core vectors, the directions of the eight edges in the vehicle’s cuboid can be determined. To estimate the magnitudes of these edges, we employ the average ratio of distances between the center points of the 2D bounding box for each object and the corresponding vertex in the 3D bounding box (ground truth). Consequently, the 3D bounding box of the vehicles in the 2D image can be accurately estimated using only the monovision sensor.
Our proposed Vehiclectron model presents a valuable contribution as it overcomes the limitations of traditional approaches that rely on multi-sensor setups and demonstrates the potential for accurate 3D bounding box (cuboid) estimation using a monovision sensor and road geometry information. By harnessing the capabilities of a monovision sensor such as CCTV, this model offers a cost-effective and easily deployable solution for estimating the 3D bounding boxes of vehicles. We confirm the feasibility and applicability of the proposed model by applying it to vehicles in actual road images from CCTVs of multiple roads.

2. Materials and Methods

2.1. Data Sources

For our experiment, we obtained road vehicle images from the AI-Hub platform [32], which serves as a data repository for sharing datasets with model developers who train their artificial intelligence models. This platform allows users to access publicly available data, including text (such as corpus data and machine reading comprehension data) as well as vision data (such as video, image, eye gaze, and traffic data). In this particular study, we utilized approximately 117,926 images containing around 724,196 road vehicles in Bucheon City, Republic of Korea. Among these vehicles, there were approximately 62,332 trucks, 40,104 buses, 607,030 general cars, 17,730 bikes, and 1143 unknown objects. An unknown object refers to an object that does not fall within the predefined categories of vehicle types. This could be because the object belongs to a special category such as excavator or other unique vehicles. It could also be due to the object being relatively small in size or appearing blurry or overlapped in the image or video frames, often a result of being located at a significant distance from the camera.
The utilized dataset includes annotations for cuboids, consisting of eight pairs of x–y coordinates, as depicted in Figure 1. Figure 1a,b represent the original image and the annotated image, respectively. In Figure 1b, the yellow box denotes the region of interest (RoI) used to capture the traffic flow, which is essential for calculating the flow vector. The annotated data are provided as XML, and consist of various image metadata, such as the image size, file name, and location ID. Additionally, there are eight pairs of x–y coordinates for cuboids, V c u b o i d , in the local image plane, namely P t l 1 x t l 1 , y t l 1 , P b l 1 x b l 1 , y b l 1 , P t r 1 x t r 1 , y t r 1 , P b r 1 x b r 1 , y b r 1 , P t l 2 x t l 2 , y t l 2 , P b l 2 x b l 2 , y b l 2 , P t r 2 x t r 2 , y t r 2 , P b r 2 x b r 2 , y b r 2 , as illustrated in Figure 2. These coordinates correspond to the top-left and bottom-right positions of the cuboid, and the class of the vehicle type is also represented. Specifically, ‘xtl1’ and ‘ytl1’ refer to the top-left coordinates on the bottom side of the cuboid, while ‘xbr2’ and ‘ybr2’ represent the bottom-right coordinates on the top side of the cuboid.

2.2. Proposed Methods

The proposed model consists of three primary components: (1) object detection, (2) core vector extraction, and (3) cuboid estimation. In the object detection stage, deep learning-based algorithms such as the YOLO-series or faster/mask R-CNN models are employed to detect various objects, particularly road vehicles. Subsequently, utilizing the center points extracted from the 2D bounding boxes obtained through object detection, three core vectors are derived to estimate the cuboid of the road vehicle. Finally, the cuboid is estimated based on the outcomes of the object detection model and core vector extraction. This proposed model comprises these interconnected stages to achieve the accurate estimation of the cuboids of road vehicles.

2.2.1. Object Detection Using Deep Learning Model

Our primary objective at this stage is to identify road vehicles in the 2D space of the image, particularly the center points within the 2D bounding boxes, which are crucial for estimating the corresponding 3D bounding boxes. To accomplish this, we utilize YOLOv7 [13], the latest iteration in the YOLO series, which incorporates significant improvements to the detector heads and backbone network. YOLOv7 benefits from the trainable bag-of-freebies techniques, which enhances real-time object detection accuracy without increasing the computational cost. Furthermore, this model offers the advantages of reducing parameter and computation requirements. These enhancements make this highly effective and accurate, making it a valuable tool for object detection in our proposed system. Additionally, its streamlined and lightweight architecture enables seamless integration with edge devices, such as CCTV systems, to enable real-time detection capabilities.
Hence, we employ the YOLOv7 network to detect road vehicles in our proposed model, and the outputs include the specific class of the detected object, and bounding box information, B ^ b o x , such as the x and y coordinates of the object’s center point (measured in pixels). The information of the estimated bounding box is used to obtain the core vectors and size of cuboids in our proposed system.

2.2.2. Core Vector Extraction

In this section, we describe the process to extract three core vectors (flow vector, orthogonal vector, and perpendicular vector) to estimate the cuboid of a road vehicle. Note that each core vector is extracted at the site level. In the cuboid estimation step, we can obtain the complete cuboid from the 2D bounding box in the object detection model and the three core vectors.
Flow vector extraction The flow vector is decided by taking into account the directions of traffic flow and road geometry. The underlying assumption is that vehicles generally adhere to the lanes on the road. Therefore, within the designated region of interest (RoI), we detect the lanes and extract the primary vector that represents the front direction of the vehicle. This is achieved through the utilization of the Hough transform [33,34,35], a widely used feature extraction technique in the fields of computer vision and pattern recognition. The Hough transform algorithm operates by converting the Cartesian coordinates ( x , y ) of image points into a parameter space known as the Hough space, denoted by ( ρ , θ ) [36,37]. Within the Hough space, each image point corresponds to a curve or sinusoid, which represents a potential shape or pattern. Prior to applying the Hough transform, an initial step of edge detection is typically performed to identify pertinent features such as lines or curves. In our experiment, our focus is specifically on detecting the road lanes, allowing us to isolate the desired features for detection purposes. The Hough transform functions as a voting process, where each point belonging to potential patterns contributes votes to all the possible patterns passing through that point. These votes are accumulated in an accumulator array, referred to as bins, and the pattern with the highest vote count is recognized as the desired pattern. As a result of the Hough transform, we are able to obtain multiple lanes within the RoI, and by calculating the average of the obtained lane vectors, we can determine the flow vector at that specific location. For visual reference, Figure 3a,c,e show the directional edges of lanes using the edge detection and Hough transform algorithms. A set of the directional edges in spot i is notated as D f i = d f , 1 i ,   d f , 2 i , , d f , n i , where n is the number of directional edges in the given spot. The flow vector f i is only directional information, and not the magnitude of the vector, which is the mean of D f i . Figure 3b,d,f show the corresponding directions of the flow vectors f i in each spot.
Estimation of orthogonal and perpendicular vectors Following the extraction of the flow vector at each site, we proceed to estimate the orthogonal and perpendicular vectors. The orthogonal vector is perpendicular to the flow vector and acts as the foundation for the cuboid orthogonal to the vehicle. In contrast, the perpendicular vector extends vertically downwards within the vehicle along the virtual z-axis of the 2D image. These vectors play a crucial role in determining the spatial dimensions and orientation of the vehicle cuboid. Figure 4 illustrates the examples of the extracted vectors for constructing the cuboid; flow vector ( f ), orthogonal vector ( o ), and perpendicular vector ( p ).
In general, vector consists of the direction and magnitude; o = θ o ,   o and p = θ p ,   p , so we first calculate the directions of o and p . For this, we compute the average angles, representing the directions, between the acquired f i and the edges within the ground truth cuboid B c u b o i d in spot i . θ o i , is calculated by f i and a set of the horizontal edge vectors in cuboid j in spot i , denoted as H i ,   j = P t l 1 i , j P t r 1 i , j ,   P t l 2 i , j P t r 2 i , j ,   P b l 1 i , j P b r 1 i , j ,   P b l 2 i , j P b r 2 i , j , as follows:
θ o i = j = 1 T g H i , j f i   T
g H i , j f i   = P i , j H i ,   j arccos P i , j · f i / P i , j × f i H i ,   j
where T is the number of total vehicles in spot i . Similarly, θ p i is calculated with f i and a set of the vertical edge vectors in cuboid j in spot i , denoted as V i ,   j = P b l 1 i , j P t l 1 i , j ,   P b r 1 i , j P t r 1 i , j ,   P b l 2 i , j P t l 2 i , j ,   P b r 2 i , j P t r 2 i , j , as follows:
θ p i = j = 1 T g V i , j f i   T
g V i , j f i   = P i , j V i ,   j arccos P i , j · f i / P i , j × f i V i ,   j
Next, the magnitude of o i and p i are obtained by each ratio of distances between center point m ^ in B ^ b o x and eight vertices in B c u b o i d and a length of diagonal line in B ^ b o x , which assumes that the size of the bounding box is proportional to the size of the cuboid. Thus, by utilizing them, we can estimate the size of the cuboid when the objects are detected in the object detection model. Combining these distance ratios with the core vectors allows us to estimate the complete form of the cuboid. In detail, we compute the ratios of distances from the center point to each vertex considering the specific class of road vehicle, denoted as c . It is important to note that different vehicle types have varying sizes; hence, there is differentiation in computations based on the vehicle class. The ratios of distances by cuboid’s vertex and class in spot i are calculated as follows:
r c i = d c i l c i ;   d c i = j = 1 T c d i s t m ^ i ,   c ,   P * i ,   c |   P * B c u b o i d
where l c i is the average length of the diagonal line in the bounding box. As a result, we can obtain o and p by using d c i . Figure 5 depicts the overall process for estimating the vehicle’s cuboid with flow vectors, B ^ b o x and B c u b o i d .

3. Experiments and Results

3.1. Experimental Design

This section presents the experimental design used to evaluate the proposed model for estimating the cuboid of road vehicles using an object detection algorithm and core vector calculation. The study consists of two main experiments with the following objectives:
(1)
To detect road vehicles in video footage and determine their 2D bounding boxes using various object detection models. The aim is to select the most optimal object detection model for this task.
(2)
To evaluate the accuracy of the estimated cuboid using core vectors by 3D IoU (Intersection on Union) as the evaluation metric.
The experimental design involves conducting these three experiments to achieve the specified objectives. The results obtained from each experiment will contribute to the overall evaluation of the proposed model.
Meanwhile, a typical IoU is a metric used to assess the accuracy of object detection, instance segmentation, and semantic segmentation algorithms in computer vision tasks in 2D image [38]. It quantifies the spatial similarity between a predicted region and a ground truth region, providing a measure of localization and segmentation accuracy. IoU is defined as the ratio of the area of intersection between the predicted and ground truth regions to the area of their union. The resulting value ranges from 0 to 1, where 0 indicates no overlap between the regions, and 1 signifies a perfect match. The IoU serves as an evaluation metric, with higher values indicating better algorithm performance in terms of localization and segmentation accuracy. Lower IoU values, however, suggest poorer performance. By comparing the predicted and ground truth regions, IoU provides a quantitative measure of how well an algorithm is performing, enabling researchers and practitioners to assess and improve the quality of their models. The formula is as follows:
I o U = A r e a   o f   I n t e r s e c t i o n A r e a   o f   U n i o n
In this study, we have employed the 3D IoU as an evaluation metric for assessing the accuracy of the estimated cuboid. The 3D IoU extends the concept of IoU from 2D images to a 3D perspective. It takes into account the IoUs of all sides of the cuboid. Let I be a set of IoUs in all sides of the cuboid, where the cardinality of the set is 6 for a cuboid. The 3D IoU is calculated as the average of the IoUs in all sides of the cuboid, using the following formula:
3 D   I o U = 1 I I o U * I I o U *
By averaging the IoUs of all sides, the 3D IoU provides a comprehensive measure of the spatial similarity between the estimated cuboid and the ground truth, enabling an accurate assessment of the localization and segmentation performance in a 3D context.

3.2. Results

This section focuses on validating the performance of the object detection model used, YOLOv7, and the results of cuboid estimation. First, we utilized the COCO-pretrained model and trained it on 70% of the available data, reserving the remaining 30% for testing purposes in our experiment. The dimensions of the input image are 1920 × 1080 pixels, which are subsequently resized to 640 × 640 pixels to serve as the input shape for the YOLOv7 model. The evaluation metrics employed are accuracy and mIoU between the ground truth and the estimated bounding boxes by YOLOv7. Accuracy is determined by comparing the number of vehicles in the ground truth with the estimated results. Additionally, mIoU are computed based on the information provided by the ground truth and the estimated bounding boxes for each class.
The results of this experiment are presented in Table 1, indicating an average detection accuracy of approximately 0.98 and mIoU of about 0.871. These values demonstrate the effectiveness of the YOLOv7 object detection model in accurately identifying road vehicles in the video footage.
Figure 6a–d show the sample images illustrating the results of the object detection model in the experiment. This model successfully detects the majority of road vehicles, such as general vehicles (annotated as ‘car’), bus, truck, and bike (annotated as ‘motorcycle’). Upon examining the images, it becomes evident that the object detection model effectively identifies road vehicles. Nevertheless, in some instances, certain objects might be partially or completely obscured by other objects present in the scene. This occlusion can result in the model failing to detect the entire extent of the obscured vehicle.
Next, we proceeded with the validation of the results obtained from the cuboid estimation. To evaluate the accuracy of the estimated cuboids, we employed the 3D IoU as the evaluation metric, with an average 3D IoU value of 0.89. Figure 7 illustrates some examples of the estimated cuboids in various locations. The visual inspection suggests a close match between the estimated cuboids and the actual objects.
These findings indicate that the estimated cuboids align well with the ground truth, as indicated by the high 3D IoU values and visual agreement was observed.

3.3. Discussion

The proposed model aims to estimate the cuboid of road vehicles in a 3D context using a vision-based object detection model and core vectors. The YOLOv7 model, a general object detection model, is utilized to detect road vehicles in 2D images based on their classes. Additionally, core vectors are derived from the road geometry, assuming that traffic flow is parallel to the lane direction. The flow vector is obtained through the edge detection and Hough transform algorithms, while orthogonal and perpendicular vectors are extracted based on the vehicle’s location and type. These core vectors contain directional information only. The size of the cuboid, specifically the lengths of its edges, is calculated using the magnitudes of these vectors, derived from the ratios of the distances between the diagonal line in bounding box and the center point obtained from object detection and each vertex in the annotated cuboid. By utilizing these core vectors, the complete cuboid of road vehicles can be drawn.
Through a series of experiments, the effectiveness of the system is evaluated both quantitatively and qualitatively. The first experiment focuses on validating the object detection model used, demonstrating its effective detection of road vehicles in images. The second experiment involves estimating the actual cuboid using the detected bounding box information and flow vector. The results are validated using the 3D IoU as the evaluation metric.
It is worth noting that the proposed model has certain limitations, such as its dependence on the geometric information of the site and the object detection model. For example, in Figure 7b,d, some bikes and trucks were not detected. However, to the best of our knowledge, this model represents a pioneering attempt in utilizing a monovision sensor alone to estimate cuboids in a 3D context. Furthermore, the proposed system can be applied to stationary video sensors, such as CCTVs, for traffic flow recognition and management, where precise information regarding vehicle coordinates and sizes is required.

4. Conclusions

In this study, we proposed a model named “Vehiclectron” for estimating the cuboid of road vehicles in a 3D context by combining a vision-based object detection model and core vectors. The utilization of the YOLOv7 object detection model proved effective in accurately detecting road vehicles in 2D images. The core vectors, derived from road geometry and flow information, played a crucial role in estimating the cuboid by providing essential directional information. Our experimental results, validated using the 3D Intersection over Union (IoU) metric, demonstrated the effectiveness of the proposed model with an average 3D IoU value of 0.89. Furthermore, we conducted practical implementation and application of the proposed model using various road vehicles in different real-world sites in Bucheon City, Republic of Korea. This validation confirmed the feasibility and applicability of our model.
However, since the proposed model estimates the cuboids of the road vehicles by using the 2D bounding box and core vectors from road geometry, there are some limitations that the 3D cuboid estimation has a dependency on the results from the object detection model (YOLOv7 in our case) in terms of accuracy and inference time. The objects may be partially or entirely obscured by other nearby objects, leading to reduced detection accuracy. Additionally, the detection of small-sized vehicles located far from the CCTV camera remains a challenge. The reduced resolution and distance from the camera can hinder the model’s ability to accurately recognize and classify such vehicles. However, it is essential to consider that the target vehicles in our model are those approaching the camera, not those moving away in other lanes. Consequently, these vehicles may eventually become more visible and detectable as they come closer to the camera, and their size in the image increases.
From the perspective of inference time, our model’s performance is commendable in GPU environments, with processing time, including inference time and core vector computing time, typically below 0.1 s per frame. While this is a respectable frame rate, it falls short of real-time processing capabilities. To address the limitation in achieving true real-time performance, further optimizations and hardware considerations may be necessary. However, it is important to note that our proposed model utilizes a single vision sensor, eliminating the need for data processing from multiple sensors and avoids the time-consuming calibration process required by multi-camera systems for cuboid estimation. While multi-camera systems may achieve higher accuracy in certain cases, they often come with the trade-off of increased computational complexity and inference time. In contrast, our proposed model strikes a balance between accuracy and inference speed, making it well-suited for dynamic and time-sensitive applications.
Overall, our proposed model represents a novel approach to estimating cuboids using a monovision sensor in 2D images combined with road geometry and core vectors. We believe that, in the future, this model holds potential for providing precise information in the field of intelligent traffic recognition and control. Ongoing research and development efforts are focused on advancing this model to further enhance its capabilities and broaden its application scope.

Author Contributions

Conceptualization by B.N. and T.L.; methodology by B.N.; formal analysis by S.L.; resources by B.N. and T.L.; data curation B.N.; writing—original draft by B.N.; writing—review and editing, T.L.; supervision by S.L. and T.J.; funding acquisition by T.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT, No. 2021R1A2C2095181).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing not applicable.

Acknowledgments

This work was supported by the Soonchunhyang University Research Fund.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Balfaqih, M.; Alharbi, S.A. Associated Information and Communication Technologies Challenges of Smart City Development. Sustainability 2022, 14, 16240. [Google Scholar] [CrossRef]
  2. Leem, Y.; Han, H.; Lee, S.H. Sejong smart city: On the road to be a city of the future. In Lecture Notes in Geoinformation and Cartography; Springer: Berlin/Heidelberg, Germany, 2019; pp. 17–33. [Google Scholar] [CrossRef]
  3. Bawany, N.Z.; Shamsi, J.A. Smart City Architecture: Vision and Challenges. 2015. Available online: www.thesai.org (accessed on 26 July 2023).
  4. Telang, S.; Chel, A.; Nemade, A.; Kaushik, G. Intelligent transport system for a smart city. In Studies in Systems, Decision and Control; Springer Science and Business Media Deutschland GmbH: Frankfurt, Germany, 2021; pp. 171–187. [Google Scholar] [CrossRef]
  5. Wang, J.; Jiang, C.; Zhang, K.; Quek, T.Q.S.; Ren, Y.; Hanzo, L. Vehicular Sensing Networks in a Smart City: Principles, Technologies and Applications. IEEE Wirel. Commun. 2017, 25, 122–132. [Google Scholar] [CrossRef]
  6. Wang, Z.; Huang, J.; Xiong, N.N.; Zhou, X.; Lin, X.; Ward, T.L. A Robust Vehicle Detection Scheme for Intelligent Traffic Surveillance Systems in Smart Cities. IEEE Access 2020, 8, 139299–139312. [Google Scholar] [CrossRef]
  7. Rathore, M.M.; Son, H.; Ahmad, A.; Paul, A. Real-time video processing for traffic control in smart city using Hadoop ecosystem with GPUs. Soft Comput. 2017, 22, 1533–1544. [Google Scholar] [CrossRef]
  8. Razavi, M.; Hamidkhani, M.; Sadeghi, R. Smart traffic light scheduling in smart city using image and video processing. In Proceedings of the 2019 3rd International Conference on Internet of Things and Applications (IoT), Isfahan, Iran, 17–18 April 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–4. [Google Scholar]
  9. Dai, S.; Li, L.; Li, Z. Modeling Vehicle Interactions via Modified LSTM Models for Trajectory Prediction. IEEE Access 2019, 7, 38287–38296. [Google Scholar] [CrossRef]
  10. Barthélemy, J.; Verstaevel, N.; Forehead, H.; Perez, P. Edge-Computing Video Analytics for Real-Time Traffic Monitoring in a Smart City. Sensors 2019, 19, 2048. [Google Scholar] [CrossRef] [PubMed]
  11. Sevillano, X.; Màrmol, E.; Fernandez-Arguedas, V. Towards Smart Traffic Management Systems: Vacant On-Street Parking Spot Detection Based on Video Analytics. In Proceedings of the 17th International Conference on Information Fusion (FUSION), Salamanca, Spain, 7–10 July 2014. [Google Scholar]
  12. Gagliardi, G.; Lupia, M.; Cario, G.; Tedesco, F.; Gaccio, F.C.; Scudo, F.L.; Casavola, A. Advanced Adaptive Street Lighting Systems for Smart Cities. Smart Cities 2020, 3, 1495–1512. [Google Scholar] [CrossRef]
  13. Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M. YOLOv7: Trainable Bag-of-Freebies Sets New State-of-the-Art for Real-Time Object Detectors. 2022. Available online: http://arxiv.org/abs/2207.02696 (accessed on 26 July 2023).
  14. Rahman, M.M.; Chakma, S.; Raza, D.M.; Akter, S.; Sattar, A. Real-Time Object Detection using Machine Learning. In 2021 12th International Conference on Computing Communication and Networking Technologies, ICCCNT 2021; Institute of Electrical and Electronics Engineers Inc.: New York, NY, USA, 2021. [Google Scholar] [CrossRef]
  15. Tang, C.; Feng, Y.; Yang, X.; Zheng, C.; Zhou, Y. The object detection based on deep learning. In 2017 4th International Conference on Information Science and Control Engineering, ICISCE 2017; Institute of Electrical and Electronics Engineers Inc.: New York, NY, USA, 2017; pp. 723–728. [Google Scholar] [CrossRef]
  16. Kimutai, G.; Ngenzi, A.; Said, R.N.; Kiprop, A.; Förster, A. An Optimum Tea Fermentation Detection Model Based on Deep Convolutional Neural Networks. Data 2020, 5, 44. [Google Scholar] [CrossRef]
  17. Massaro, A.; Birardi, G.; Manca, F.; Marin, C.; Birardi, V.; Giannone, D.; Galiano, A.M. Innovative DSS for intelligent monitoring and urban square design approaches: A case of study. Sustain. Cities Soc. 2020, 65, 102653. [Google Scholar] [CrossRef]
  18. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. 2018. Available online: http://arxiv.org/abs/1804.02767 (accessed on 26 July 2023).
  19. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. Available online: http://pjreddie.com/yolo/ (accessed on 26 July 2023).
  20. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. 2020. Available online: http://arxiv.org/abs/2004.10934 (accessed on 26 July 2023).
  21. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: Fully Convolutional One-Stage Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  22. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: A Simple and Strong Anchor-free Object Detector. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 1922–1933. [Google Scholar] [CrossRef] [PubMed]
  23. Arnold, E.; Al-Jarrah, O.Y.; Dianati, M.; Fallah, S.; Oxtoby, D.; Mouzakitis, A. A Survey on 3D Object Detection Methods for Autonomous Driving Applications. IEEE Trans. Intell. Transp. Syst. 2019, 20, 3782–3795. [Google Scholar] [CrossRef]
  24. Li, P.; Chen, X.; Shen, S. Stereo R-CNN based 3D Object Detection for Autonomous Driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  25. Longin, Z.D.; Latecki, J. Amodal Detection of 3D Objects: Inferring 3D Bounding Boxes from 2D Ones in RGB-Depth Images. Available online: https://github.com/phoenixnn/ (accessed on 26 July 2023).
  26. Ku, J.; Mozifian, M.; Lee, J.; Harakeh, A.; Waslander, S.L. Joint 3d proposal generation and object detection from view aggregation. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–8. [Google Scholar]
  27. Ma, X.; Ouyang, W.; Simonelli, A.; Ricci, E. 3D Object Detection from Images for Autonomous Driving: A Survey. 2022. Available online: http://arxiv.org/abs/2202.02980 (accessed on 26 July 2023).
  28. Fayyaz, M.A.B.; Alexoulis-Chrysovergis, A.C.; Southgate, M.J.; Johnson, C. A review of the technological developments for interlocking at level crossing. In Proceedings of the Institution of Mechanical Engineers, Part F: Journal of Rail and Rapid Transit; SAGE Publications Ltd.: Newbury Park, CA, USA, 2021; Volume 235, pp. 529–539. [Google Scholar] [CrossRef]
  29. Chen, X.; Kundu, K.; Zhang, Z.; Ma, H.; Fidler, S.; Urtasun, R. Monocular 3D Object Detection for Autonomous Driving. In Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  30. Chabot, F.; Chaouch, M.; Rabarisoa, J.; Teulière, C.; Chateau, T. Deep MANTA: A Coarse-to-Fine Many-Task Network for Joint 2D and 3D Vehicle Analysis from Monocular Image. 2017. Available online: http://arxiv.org/abs/1703.07570 (accessed on 26 July 2023).
  31. Mousavian, A.; Anguelov, D.; Flynn, J.; Kosecka, J. 3D Bounding Box Estimation Using Deep Learning and Geometry. 2016. Available online: http://arxiv.org/abs/1612.00496 (accessed on 26 July 2023).
  32. AI-Hub Platform. Available online: https://www.aihub.or.kr/aihubdata/data/view.do?currMenu=115&topMenu=100 (accessed on 26 July 2023).
  33. Newman, W.; Duda, R.O.; Hart, P.E. Graphics and Use of the Hough Transformation To Detect Lines and Curves in Pictures. Commun. ACM 1972, 15, 11–15. [Google Scholar]
  34. Aggarwal, N.; Karl, W. Line detection in images through regularized hough transform. IEEE Trans. Image Process. 2006, 15, 582–591. [Google Scholar] [CrossRef] [PubMed]
  35. Conservatoire national des arts et métiers (France). IEEE Systems, and Institute of Electrical and Electronics Engineers. In Proceedings of the 2019 6th International Conference on Control, Decision and Information Technologies (CoDIT’19)), Le Cnam, Paris, France, 23–26 April 2019. [Google Scholar]
  36. Duan, D.; Xie, M.; Mo, Q.; Han, Z.; Wan, Y. An improved Hough transform for line detection. In Proceedings of the 2010 International Conference on Computer Application and System Modeling (ICCASM 2010), Taiyuan, China, 22–24 October 2010; IEEE: Piscataway, NJ, USA, 2010; Volume 2, p. V2-354. [Google Scholar]
  37. Zhao, K.; Han, Q.; Zhang, C.B.; Xu, J.; Cheng, M.M. Deep Hough Transform for Semantic Line Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 4793–4806. [Google Scholar] [CrossRef] [PubMed]
  38. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection over Union: A Metric and A Loss for Bounding Box Regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
Figure 1. Examples of used (a) original image; and (b) 3D bounding box-annotated image.
Figure 1. Examples of used (a) original image; and (b) 3D bounding box-annotated image.
Sensors 23 07504 g001
Figure 2. Notation of x–y coordinates in V c u b o i d .
Figure 2. Notation of x–y coordinates in V c u b o i d .
Sensors 23 07504 g002
Figure 3. Examples of the extracted lanes (a,c,e) and flow vectors (b,d,f), respectively.
Figure 3. Examples of the extracted lanes (a,c,e) and flow vectors (b,d,f), respectively.
Sensors 23 07504 g003aSensors 23 07504 g003b
Figure 4. Representation of flow vector ( f ), orthogonal vector ( o ), and perpendicular vector ( p ).
Figure 4. Representation of flow vector ( f ), orthogonal vector ( o ), and perpendicular vector ( p ).
Sensors 23 07504 g004
Figure 5. Overall process for estimating cuboid using flow vectors, B ^ b o x and B c u b o i d .
Figure 5. Overall process for estimating cuboid using flow vectors, B ^ b o x and B c u b o i d .
Sensors 23 07504 g005
Figure 6. (ad) Results of object detection model in our experiment using YOLOv7.
Figure 6. (ad) Results of object detection model in our experiment using YOLOv7.
Sensors 23 07504 g006
Figure 7. (ad) Results the estimated cuboids in some different sites.
Figure 7. (ad) Results the estimated cuboids in some different sites.
Sensors 23 07504 g007
Table 1. The performance of the object detection model. # means “the number of”.
Table 1. The performance of the object detection model. # means “the number of”.
Class# of Objects in TestAccuracymIoU
Ground truthGeneral vehicle182,109--
Bus12,031--
Trucks18,699--
Bikes5319--
Object
detection model (YOLOv7)
General vehicle179,3770.9850.871
Bus11,4290.9500.910
Trucks18,0250.9640.870
Bikes45840.8620.774
Average0.9780.871
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Noh, B.; Lin, T.; Lee, S.; Jeong, T. Deep Learning and Geometry Flow Vector Using Estimating Vehicle Cuboid Technology in a Monovision Environment. Sensors 2023, 23, 7504. https://doi.org/10.3390/s23177504

AMA Style

Noh B, Lin T, Lee S, Jeong T. Deep Learning and Geometry Flow Vector Using Estimating Vehicle Cuboid Technology in a Monovision Environment. Sensors. 2023; 23(17):7504. https://doi.org/10.3390/s23177504

Chicago/Turabian Style

Noh, Byeongjoon, Tengfeng Lin, Sungju Lee, and Taikyeong Jeong. 2023. "Deep Learning and Geometry Flow Vector Using Estimating Vehicle Cuboid Technology in a Monovision Environment" Sensors 23, no. 17: 7504. https://doi.org/10.3390/s23177504

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop