Next Article in Journal
Cellulose Nanopaper Cross-Linked Amino Graphene/Polyaniline Sensors to Detect CO2 Gas at Room Temperature
Next Article in Special Issue
Waterfall Atrous Spatial Pooling Architecture for Efficient Semantic Segmentation
Previous Article in Journal
A Classification Method for Select Defects in Power Transformers Based on the Acoustic Signals
Previous Article in Special Issue
Real-Time Vehicle-Detection Method in Bird-View Unmanned-Aerial-Vehicle Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Vehicular Traffic Congestion Classification by Visual Features and Deep Learning Approaches: A Comparison

Dipartimento di Informatica, Università degli Studi di Bari Aldo Moro, 70125 Bari, Italy
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(23), 5213; https://doi.org/10.3390/s19235213
Submission received: 17 October 2019 / Revised: 24 November 2019 / Accepted: 26 November 2019 / Published: 28 November 2019

Abstract

:
Automatic traffic flow classification is useful to reveal road congestions and accidents. Nowadays, roads and highways are equipped with a huge amount of surveillance cameras, which can be used for real-time vehicle identification, and thus providing traffic flow estimation. This research provides a comparative analysis of state-of-the-art object detectors, visual features, and classification models useful to implement traffic state estimations. More specifically, three different object detectors are compared to identify vehicles. Four machine learning techniques are successively employed to explore five visual features for classification aims. These classic machine learning approaches are compared with the deep learning techniques. This research demonstrates that, when methods and resources are properly implemented and tested, results are very encouraging for both methods, but the deep learning method is the most accurately performing one reaching an accuracy of 99.9% for binary traffic state classification and 98.6% for multiclass classification.

1. Introduction

Urban roads and highways have, nowadays, plenty of surveillance cameras initially installed for various security reasons. Traffic videos coming from these cameras can be used to estimate the traffic state, to automatically identify congestions, accidents, and infractions, and thus helping the transport management to face critical aspects of the mobility. At the same time, this information can also be used to plan the mid and long-term roads mobility strategy. This is a clear example of a smart city application having also a strong impact on citizens’ security [1].
Studies dealing with traffic state estimation by videos adopt a common processing pipeline, which includes the following:
  • The pre-processing of the video frames to highlight the useful elements (vehicles) and hide the unnecessary ones (background, etc.);
  • The extraction of visual features able to describe the traffic state (e.g., number of vehicles, speed, etc.);
  • One or more methods to classify the traffic state.
It is difficult to identify and select the best algorithms to be adopted because, often, systems reported in many studies are different at many stages, as well as adopt different datasets and testing conditions. This research provides a brief review of the most used techniques and reports an extended and systematic experimental comparison under common set-up conditions, thus highlighting strengths and weaknesses of each approach and supporting interested readers in the most profitable choices. More specifically:
  • Haar Cascade, You Only Look Once (YOLO), the Single Shot MultiBox Detector (SSD), and Mask R-Convolutional Neural Networks (R-CNN) are adopted and compared for vehicle detection;
  • Results provided by the two most accurately performing detectors, among the aforementioned, are used in a comparative schema to evaluate a set of visual features able to characterize the traffic state. These features were fed to four different classic classifiers, thus highlighting the most accurately performing one;
  • Results obtained with classic approaches (previous point of this list) were compared to the use of deep learning techniques.
The article is organized as follows: Section 2 describes related studies, Section 3 presents the object detectors candidate to identify vehicles, the visual features useful to characterize a traffic video frame, and classifiers. The video datasets, evaluation metrics, along with experimental results are shown in Section 4. Section 5 presents conclusions and future researches.

2. Related Studies

Among the different methods that can be adopted to provide traffic flow (congestion) estimation, surveillance cameras play a crucial role. These systems can be installed without interfering with road infrastructures. Moreover, a large plethora of retrofit solutions are available in many cases, or systems have been already installed for some initial different aim. In any case, surveillance cameras can supply real-time information. The estimation of the traffic can be provided to users and police patrols to help in departures planning and congestion avoiding. Road panels or integrated vehicular monitors can also be used to reach the aim [2].
One of the first steps within the pipeline is vehicle identification [3]. Vehicles can be identified using features such as colors, pixels, and edges along with some machine learning algorithms. More specifically, detectors able to locate and classify objects in video frames exploiting visual features must be considered. State-of-the-art features are SURF (Speeded Up Robust Features) and bag-of-features [4], Haar Features [5], Edge Features [6], Shape Features [7], and Oriented Gradients Histograms [8]. Approaches based on visual features and machine learning models greatly benefited in efficiency with the introduction Convolutional Neural Networks (CNN) [9].
Choudhury et al. [10] proposed a system based on Dynamic Bayesian Network (DBN). It was tested on three videos after extracting visual features such as the number of moving vehicles [11]. Other examples of vehicle detector are in [12] and in [13] where features to characterize the traffic density are extracted from video footage. Li et al. [14] exploited the texture difference between congestion and unobstructed images, while similar approaches were used to recognize tow-away road signs from videos [15].
Vehicular traffic behavior can be also revealed by observable motion [16], in fact, it can be used to determine the number of vehicles performing the same action (e.g., by exploiting trajectory clustering for scene description). Drones and radio-controlled model aircrafts are exploited in the works of Liu et al. [17] and Gao et al. [18] to shoot live traffic flow videos and to study roads conditions. Shen and He [19] analyzed the vehicular behaviors at traffic bottlenecks and their movement to verify the decision-making process. In the research of Avinash et al. [20], Multiple Linear Regression (MLR) Technique was adopted to comprehend the factors influencing pedestrian safety. Koganti et al. [21] adopted Lane Distribution Factor (LDF) to describe the distribution of vehicular traffic across the roadway. Thomas et al. [22] presented a perceptual video summarization technique on a stack of videos to solve accident detection.
Real-time traffic videos have been used to observe temporal state of vehicles on a pedestrian crossing lane by using several image processing algorithms, connected components, and ray-casting technique [23]. Lin and Wang [24] implement a Vehicular Digital Video Recorder System able to support an online real-time navigator and an offline data viewer: The system was able to consider data security about vehicular parameters and to depict the instant status of vehicles.

3. Methods and Materials

Two approaches are compared in this research: The first one relies on visual features evaluated from traffic videos through computer vision algorithms using state-of-the-art object detectors and classifiers, the latter considers deep learning models able to automatically extract features from videos needed for the final classification.

3.1. Object Detectors

Four different object detectors have been explored: Haar Cascade, You Only Look Once (YOLO), Single Shot MultiBox Detector (SSD), and Mask R-CNN.
The Haar Cascade object detector, originally developed by Viola and Jones [25], relies on a set of visual features able to exploit rectangular regions at a specific location identifying pixel intensities and differences between the regions. The term “Cascade” refers to the decision sequencing, in fact the algorithm constructs a “strong” classifier as a combination of weighted weak classifiers using a boosting technique. A search window moves through the whole image to cover all the pieces and, at the same time, it is resized to different scales to find objects of different sizes. The detector requires a set of positive and negative samples. Haar features are extracted in the test phase and then compared with those used in the training phase, in order to obtain a positive or negative identification. Haar Cascade has been successfully used for Vehicle Detection [26], also to evaluate a traffic congestion index [27].
The YOLO (You Only Look Once) object detector consists of a CNN called Darknet [28] with an architecture made by 24 convolutional layers working as feature extractors and 2 dense layers for the prediction. Successively, YOLO v2 introduced anchors as a set of boxes used to predict the bounding boxes, while in YOLO v3 the prediction is performed at different scales. The algorithm looks at the image only once and splits it into an NxN grid where each cell predicts a fixed number of bounding boxes to associate an object to the supposed class providing a confidence score. Many of the bounding boxes have a very low confidence. Therefore, they can be filtered by applying a threshold. YOLO is a fast algorithm because it requires only one image processing, at the same time accuracy decreases when two different objects are very close to each other. YOLO has been successfully used for Vehicle Detection in [29] where the focal loss is exploited and validated on the BIT-Vehicle dataset and in [30] where, on the same dataset, a mean Average Precision (mAP) of 94.78% is reported.
Convolutional Neural Network (CNN) is also at the basis of the Single Shot MultiBox Detector (SSD), producing a set of fixed-size bounding, a confidence score is provided representing the probability that the object in the box belongs to a specific class. The CNN is constituted by several layers that progressively decrease in size so that objects can be predicted on several scales. SSD has been successfully used for Vehicle Detection in [12] while Sreekumar et al. [31] developed a multi-algorithm method for the real-time traffic pattern generation and Asvadi et al. [32] exploited a Dense Reflection Map (DRM) inputted to a Deep Convolutional Neural Network for the vehicle detection.
Mask R-CNN is the evolution of R-CNNs [33]. The original R-CNN detector used selective search to extract a huge number of candidate (proposal) regions (2000) from the input image. The algorithm first generates many candidate regions by performing initial sub-segmentation. Then, a greedy approach combines similar (adjacent) regions in larger ones. The selective search and the greedy approach result in very low computing [34]. R-CNN developed by Microsoft Research addresses this speed computation problem. One single model is used to extract features from each region, predict the belonging class, and compute the box coordinates. This is performed on a filtered image, which uses the low rank approximation with Singular Value Decomposition (SVD) of the original image. Fast R-CNN is 10× faster than R-CNN. Faster R-CNN [34] improves Fast R-CNN by using an additional network, called Region Proposal Network, in place of the selective search (originally derived from R-CNN) used for the generation of the regions of interest. This increases the prediction speed by about 10× [35]. Mask R-CNN is built on top of Faster R-CNN and, in addition to Faster R-CNN, it provides also the object segmentation. The mask of the segmented object could be used for inferring the valid shape of the classified object.
It is important to note that, while SSD used pre-trained weights on the Pascal Visual Object Classes (VOC) dataset [36], YOLO v3 and Mask R-CNN used pre-trained weights on the Coco dataset [37].

3.2. Visual Features

The output produced by a specific detector can be exploited to evaluate the traffic state and its density. In this research, five visual descriptors were considered: Total Vehicles, Traffic Velocity, Traffic Flow, Road Occupancy, and the Texture Feature.
The Total Vehicles feature is the number of bounding boxes provided by the vehicle detector and evaluated for each video frame [10,13,20,21,24,38,39,40,41,42,43].
The Traffic Velocity (average speed of all vehicles in the frame) is evaluated by tracking the vehicular bounding boxes centers and calculating the Euclidean distance with the corresponding position in the next frame [31]. The distance of each vehicle has been normalized according to the fps (frame per second) rate thus finding the individual speed and the global average speed (the sum of all vehicles velocities divided by their number). The last parameter can be considered as an estimation of the global traffic velocity [12,13,42,44,45].
The third visual feature is Traffic Flow, it is calculated considering the difference between the incoming and the outgoing traffic, respectively evaluated as the number of vehicles entering the camera field of view and those leaving it [15,39,40,42,45].
Background suppression and morphological transformations (i.e., opening and closing) can be used to isolate vehicle shapes: This process returns the Road Occupancy providing a relationship between the flat road surface (white pixels in Figure 1) and vehicles (black pixels in Figure 1). To gain this result, frames are converted to grayscale. Successively, pixel values are subtracted from the next frame highlighting non-modified areas (background) and modified ones (foreground): Pixels belonging to vehicles shape are those related to changes in the scene. A threshold operator (i.e., Regional Minimum) is adopted to distinguish flat areas from areas occupied by vehicles which result from the ratio between black and white pixels dynamically changing according to the number of vehicles [9,12,16,41,42,44,45].
The Texture Feature was calculated according to the Gray Level Co-occurrence Matrix (GLCM) method [14,45,46,47]. This parameter is typically used to estimate vehicles’ density by exploiting the corresponding ‘energy’ and ‘entropy’ values. More specifically, energy reveals whether the texture is thick (high value) or thin, while the entropy expresses how the elements are distributed (uniformly featuring a high value or not) [48]. The value of energy is inversely proportional with vehicle density, and the value of entropy is proportional with vehicle density. In other words, the gray histogram of image should be distributed uniformly and the texture of the image should be thin in case there are many vehicles in a frame. Therefore, the energy feature value should be small, and the entropy feature value should be big.

3.3. Machine Learning Classifiers

In this research three classifiers were considered and compared: k-Nearest Neighbors, Support Vector Machine, and Random Forest.
The K-Nearest Neighbors (KNN) is used for classification and regression tasks exploiting sample vectors in a multi-dimensional feature space. K is a user-defined parameter referred to the number of class labels: the unknown input vector is classified assigning it the “nearest class” among those already known. Many distance measures can be considered as, for example, the Euclidean or the Manhattan one. K-NN has been applied to this specific field in [38] and [49].
Support Vector Machine (SVM) maps feature vectors of two different classes within a hyperspace and searches for the best separating hyperplane taking into account only a reduced set of the initial amount of examples (called support vectors), which are those difficult to be classified. According to data distribution, different separating hyperplanes (kernel) can be considered. In this research the linear kernel and a Gaussian Radial Basis (rbf) function were considered as in [50] where the traffic congestion is classified through a comparison between AlexNet+SVM and VGGNet+SVM.
The Random Forest (RF) classifier relies on a bagging method to build several “base learners”, usually decision trees. The base learners are successively combined to provide the final classification. RF repeatedly selects a bootstrap sample from the training set. It selects a random subset of features, and then fits a decision tree to this sample. Due to this randomness, the bias of the forest increases, but, due to averaging, its variance also decreases. In extremely randomized trees (ET), randomness is taken a step further by also completely randomizing the cut-point choice while splitting each node of a tree. This allows the variance to be reduced a little more at the expense of a slight increase in bias. In this research, 100 decision trees were used as base learners for RF and ET.

3.4. Deep Learning Classification Models

He et al. [51] proposed a residual deep learning framework for image classification, where layers were redrafted in order to learn residual functions with respect to the input layer. The proposed ResNET had 34 layers that follow the same pattern while performing 3 × 3 convolutions with a fixed feature map dimension, the input is bypassed every two convolutions. Moreover, the width and height dimensions remain constant for the entire layer thus reducing the complexity per layer of the network. The output is a binary classification (congested/not congested). The ResNET has been also re-trained in [52] on the Shaanxi Province dataset.
Kurniawan et al. [53] used two convolutional layers, a max pooling layer, and a fully connected layer where the first two layers are convolute with 3 × 3 filters and 32 feature maps, the third one is a 2 × 2 max pooling layer used for down-sampling, and the last one is a fully connected layer with 128 neurons. Rectified Linear Units (ReLU) activation function has been exploited in both the convolutional and fully connected layers while a sigmoid activation function has been used for the output layer [45,50].

4. Experiments and Discussion

4.1. Video Datasets

Different datasets, used for different aims, were adopted in this research. The GRAM Road Traffic Monitoring (RTM) is generally used for vehicle detection and it was adopted here to evaluate and compare performance of object (vehicles) detectors [54]. Trafficdb contains annotations related to the state of the traffic and it is here used to compare classification techniques [55,56].

4.1.1. GRAM RTM

The Road-Traffic Monitoring [54] is a dataset specifically used for vehicular detection in a traffic environment. It consists of three video sequences from which individual frames were labeled with bounding box around vehicles. The “M-30” video includes 7520 frames recorded on a sunny day with a Nikon Coolpix L20 camera having a resolution of 640 × 480 pixels at 30 fps. The second video “M-30-HD” includes 9390 frames recorded in the same place of the previous video but on a cloudy day at a higher resolution (1280 × 720 pixels at 30 fps using a Nikon DX3100 camera). The last “Urban1” video contains 23,435 frames in low resolution (480 × 320 pixels at 25 fps). This dataset offers the possibility to evaluate vehicles detectors under different working conditions.
From each video of the dataset, the ground-truth consists in bounding boxes around all vehicles per each frame. Information about the acquisition properties are provided, together with pixel masks useful to extract region of interests (ROI) and decrease the computational load of subsequent processing phases (Figure 2).

4.1.2. Trafficdb

Trafficdb dataset is a state-of-the-art dataset used for vehicular traffic state classification since it is provided with specific annotations [55,56]. It is constituted by 254 videos acquired between 8 April and 8 June 2004 on Seattle (USA) highway segments. The ground-truth includes three classes (Figure 3): ‘Heavy’—very congested traffic, ‘Medium’—low vehicular flow, and ‘Light’—normal travel velocity.

4.2. Evaluation Metrics

Vehicle bounding boxes provided by the object detector can be compared to the real vehicle annotation highlighting the most accurately performing one. It is quite clear that the result of this phase has a strong impact on all the subsequent stages within the processing pipeline.
The metric Correct detections refers to correctly detected vehicles for each frame and it is measured as the pixel intersection between the original ground-truth bounding box area (named G) and the predicted one (named P). The Jaccard Index or Intersection over Union (1) was considered.
J ( G , P ) = | G P | | G P | .
The second performance metric here considered is the Computing time evaluated by adding the processing time for each video frame at each iteration. It is useful to select the most suitable detector for a specific problem (real-time, on site, off-line, etc.).
The accuracy was considered and evaluated as follow:
accuracy = (TP + TN)/(TP + TN + FP + FN),
where:
• TP: the true positive samples, example of class X and classified by the system as X;
• TN: the true negative samples, object not of class X and not classified by the system as X;
• FP: the false positive samples, object not of class X but classified by the system as X;
• FN: the false negative samples, object of class X, but not classified by the system as X;
Experiments were performed on a System featuring Ubuntu 18.04 as Operating System, AMD Ryzhen threadripper 1920x with 12 cores, Nvidia Titan RTX 24 GB RAM, 64 GB RAM DDR4.

4.3. Vehicle Detectors Evaluation

Table 1 reports results of the object detectors on the Road-Traffic Monitoring GRAM dataset.
The experimental phase pointed out that the Haar Cascade detector is the fastest one on each dataset. However, it provides good accuracy only on M-30-HD (75%).
The lowest processing time is achieved on the ‘Urban1’ video due to the low frame resolution that also impacts on the correctly identified vehicles, with 40% of accuracy also due to incorrect multiple detections.
The Single Shot MultiBox Detector (SSD) is the slowest object detector. The experiment on the "M-30" video reported an accuracy of 22%, which is the lowest between the four solutions. On the "M-30-HD" video, the computational time is much heavier (11 to 14 s) with frequent peaks between 14 and 17 s, the reported accuracy reaches 70%.
YOLO detector offers the best compromise with acceptable execution times and very good performances on all datasets.
Mask R-CNN exhibits very discordant results in terms of accuracy. In particular, the poor performance on Urban1 dataset is probably due to the poor image quality: JPEG compression kneaded colors, and thus cheated the Region Proposal Network.
The most accurately performing detector, considering the processing time, is Haar Cascade, while YOLO represents the compromise between time resources and detection accuracy.

4.4. Traffic State Classification: Visual Features and Machine Learning Classification

Results obtained in the previous section clearly report YOLO and Haar Cascade as the most accurately performing vehicles detectors in terms of accuracy and processing time. For this reason, they have been chosen to support the visual features extraction to build the input vector for vehicle classifiers.
The first experimental session involves the five visual features calculated using the best two selected object detectors with the different machine learning classifiers seen in Section 3.3. A 10-fold cross validation setup was adopted to minimize the effect of variance when choosing the training and testing examples. Classification results on the Trafficdb dataset are reported in Table 2. Visual features were extracted on a sampling rate of 30 frames.
Table 2 shows that YOLO detector combined with the Random Forest is the most accurately performing solution with an accuracy of 84%. The confusion matrix provided by this solution is reported in Table 3. Due to the unbalance of the classes in Trafficdb (the ‘Heavy’ class instances are about four times the other two), the classification results are provided in a normalized form.

4.5. Traffic State Classification: Deep Learning

The deep learning models described in Section 3.4 were implemented and re-trained on the Trafficdb video dataset in a 10-fold cross validation setup. The ResNet [51] and the deep network architecture proposed in [52] were originally tested by respective authors on a two class (Heavy vs. Light) traffic state classification. To perform similar tests, at a first stage, samples labeled as ‘Medium’ were removed from the Trafficdb: results are shown in Table 4.
Finally, to compare results to the cases if the previous section, the two deep learning architectures were extended to perform traffic state classification on three classes. In this case, the best performance was reached by the ResNet [51] with an accuracy of 98.61% (results are in Table 5).

5. Conclusions

A pipeline to develop state-of-the-art traffic state classification systems from videos has been presented in this research. The pipeline is made up of three main steps: vehicle detection, feature extraction, and classification. Several state-of-the-art approaches have been considered and compared. A preliminary comparison between object detectors, performed on the GRAM Road Traffic Monitoring video dataset, has pointed out that YOLO v3 can be used for real-time vehicle detector exhibiting a detection accuracy of over 80%.
For the traffic state classification, two different approaches have been studied and tested on the Trafficdb video dataset. The first approach relies on visual features calculated through computer vision techniques and machine learning classifiers while the second one exploits deep learning able to embed the features extraction when training the model on the annotated dataset. In the classic approach (visual features and machine learning classifiers), the Random Forest has gained 84% of accuracy while the deep learning approach has reached an accuracy of over 98% with the same experimental setup, and thus showing a noticeable increase of +14% in the results.
The problem here considered is obviously complex, and the provided results need further improvement as for example: refinements of the object detection algorithms, validation on more traffic datasets. The last aspect is non-trivial because different road settings (country road, city road, road junction and crossroad, double lane roads, etc.) and weather conditions (rainy nights, fog, snow, gusts of wind, and so on) could have significant impact on systems.

Author Contributions

Conceptualization, D.I. and G.P.; methodology, D.I.; software, F.B. and V.D.; validation, D.I. and G.P.; investigation, D.I. and G.P.; resources, D.I.; data curation, F.B. and V.D.; writing—original draft preparation, F.B. and D.I.; writing—review and editing, F.B., D.I., and V.D.; supervision, D.I.; project administration, D.I.; funding acquisition, D.I.

Funding

This research is within the “Metrominuto Advanced Social Games” project funded by POR Puglia FESR-FSE 2014–2020—Fondo Europeo Sviluppo Regionale—Asse I—Azione 1.4—Sub-Azione 1.4.b—Avviso pubblico “Innolabs”.

Conflicts of Interest

The authors declare no conflict of interest.

Code

References

  1. Ministero Delle Infrastrutture e dei Trasporti, “I Sistemi di Trasporto Intelligenti (ITS)”. 2016. Available online: mit.gov.it/documentazione/sistemi-trasporto-intelligenti (accessed on 22 April 2019).
  2. Wan, J.; Yuan, Y.; Wang, Q. Traffic congestion analysis: A new Perspective. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 1398–1402. [Google Scholar]
  3. Shi, X.; Shan, Z.; Zhao, N. Learning for an aesthetic model for estimating the traffic state in the traffic video. Neurocomputing 2016, 181, 29–37. [Google Scholar] [CrossRef]
  4. Nguyen, H.N.; Krishnakumari, P.; Vu, H.L.; Van Lint, H. Traffic COngestion pattern classification using multi-class SVM. In Proceedings of the IEEE 19th International Conference on Intelligent Transportation Systems (ITSC), Rio de Janeiro, Brazil, 1–4 November 2016; pp. 1059–1064. [Google Scholar]
  5. Viola, P.A.; Jones, M.J.; Snow, D. Detecting pedestrians using patterns of motion and appearance. Int. J. Comput. Vis. 2005, 63, 153–161. [Google Scholar] [CrossRef]
  6. Wu, B.; Nevatia, R. Detection of multiple, partially occluded humans in a single image by bayesian combination of edgelet part detectors. In Proceedings of the 10th IEEE International Conference on Computer Vision (ICCV 2005), Beijing, China, 17–21 October 2005; pp. 90–97. [Google Scholar]
  7. Sabzmeydani, P.; Mori, G. Detecting pedestrians by learning shapelet features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar]
  8. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  9. Wali, S.B.; Abdullah, M.A.; Hannan, M.A.; Hussain, A.; Samad, S.A.; Ker, P.J.; Mansor, M.B. Vision-Based Traffic Sign Detection and Recognition Systems: Current Trends and Challenges. Sensors 2019, 19, 2093. [Google Scholar] [CrossRef] [PubMed]
  10. Chaudhary, S.; Indu, S.; Chaudhury, S. Video-based road traffic monitoring and prediction using dynamic Bayesian networks. IET Intell. Transp. Syst. 2018, 12, 169–176. [Google Scholar] [CrossRef]
  11. Choudhury, S.; Chattopadhyay, S.P.; Hazra, T.K. Vehicle detection and counting using haar feature based classifier. In Proceedings of the 8th Annual Industrial Automation and Electromechanical Engineering Conference (IEMECON), Bangkok, Thailand, 16–18 August 2017; pp. 106–109. [Google Scholar]
  12. Qiong, W.U.; Sheng-bin, L. Single Shot MultiBox Detector for Vehicles and Pedestrians Detection and Classification. In DEStech Transactions on Engineering and Technology Research; DEStech Publications, Inc.: Lancaster, PA, USA, 2018. [Google Scholar]
  13. Dailey, X.D.; Pumin, S.; Cathey, F.W. An algorithm to estimate mean traffic speed using uncalibrated cameras. IEEE Trans. Intell. Transp. Syst. 2000, 1, 98–107. [Google Scholar] [CrossRef]
  14. Balducci, F.; Impedovo, D.; Pirlo, G. Detection and validation of tow-away road sign licenses through deep learning methods. Sensors 2018, 18, 4147. [Google Scholar] [CrossRef] [PubMed]
  15. Morris, B.; Trivedi, M. Understanding vehicular traffic behavior from video: A survey of unsupervised approaches. J. Electron. Imaging 2013, 22, 041113. [Google Scholar] [CrossRef]
  16. Liu, S.; Zheng, Y.; Luo, H.; Duan, S.; Wang, H. Vehicle trajectory observation based on traffic video provided by radio-controlled model aircraft. In Proceedings of the 4th International Conference on Transportation Engineering, Chengdu, China, 19–20 October 2013; pp. 122–128. [Google Scholar]
  17. Gao, H.; Kong, S.; Zhou, S.; Lv, F.; Chen, Q. Automatic extraction of multi-vehicle trajectory based on traffic videotaping from quadcopter model. Appl. Mech. Mater. 2014, 552, 232–239. [Google Scholar] [CrossRef]
  18. Shen, X.; He, Z. Analysis of vehicular behavior at bottlenecks considering lateral separation. Smart Innov. Syst. Technol. 2017, 53, 169–185. [Google Scholar]
  19. Avinash, C.; Jiten, S.; Shriniwas, A.; Gaurang, J.; Manoranjan, P. Evaluation of pedestrian safety margin at mid-block crosswalks in India. Saf. Sci. 2018, 119, 188–198. [Google Scholar] [CrossRef]
  20. Koganti, S.; Raja, K.; Sajja, S.; Narendra, M.S. A study on volume, speed and lane distribution of mixed traffic flow by using video graphic technique. Int. J. Eng. Technol. (UAE) 2018, 7, 59–62. [Google Scholar] [CrossRef]
  21. Thomas, S.; Gupta, S.; Subramanian, V. Event detection on roads using perceptual video summarization. IEEE Trans. Intell. Transp. Syst. 2018, 19, 2944–2954. [Google Scholar] [CrossRef]
  22. Goma, J.C.d.; Ammuyutan, L.A.B.; Capulong, H.L.S.; Naranjo, K.P.; Devaraj, M. Vehicular obstruction detection in the zebra lane using computer vision. In Proceedings of the IEEE 6th International Conference on Industrial Engineering and Applications (ICIEA), Tokyo, Japan, 12–15 April 2019; pp. 362–366. [Google Scholar]
  23. Lin, C.; Wang, M. An implementation of a vehicular digital video recorder system. In Proceedings of the IEEE/ACM Int’l Conference on Green Computing and Communications Int’l Conference on Cyber, Physical and Social Computing, Hangzhou, China, 18–20 December 2010; pp. 907–911. [Google Scholar]
  24. Viola, P.; Jones, M. Robust Real-Time Object Detection. Int. J. Comput. Vis. 2001, 57, 137–154. [Google Scholar] [CrossRef]
  25. Dubey, A.; Rane, S. Implementation of an intelligent traffic control system and real time traffic statistics broadcasting. In Proceedings of the International conference of Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India, 20–22 April 2017; pp. 33–37. [Google Scholar]
  26. Lam, C.; Gao, H.; Ng, B. A real-time traffic congestion detection system using on-line images. In Proceedings of the IEEE 17th International Conference on Communication Technology (ICCT), Chengdu, Chian, 27–30 October 2017; pp. 1548–1552. [Google Scholar]
  27. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  28. Wu, Z.; Sang, J.; Zhang, Q.; Xiang, H.; Cai, B.; Xia, X. Multi-Scale Vehicle Detection for Foreground-Background Class Imbalance with Improved YOLOv2. Sensors 2019, 19, 3336. [Google Scholar] [CrossRef] [PubMed]
  29. Sang, J.; Wu, Z.; Guo, P.; Hu, H.; Xiang, H.; Zhang, Q.; Cai, B. An Improved YOLOv2 for Vehicle Detection. Sensors 2018, 18, 4272. [Google Scholar] [CrossRef]
  30. Sreekumar, U.K.; Devaraj, R.; Li, Q.; Liu, K. TPCAM: Real-time traffic pattern collection and analysis model based on deep learning. In Proceedings of the IEEE SmartWorld, Ubiquitous Intelligence Computing, San Francisco, CA, USA, 4–8 August 2017; pp. 1–4. [Google Scholar]
  31. Asvadi, A.; Garrote, L.; Premebida, C.; Peixoto, P.; Nunes, U.J. Real-time deep convnet-based vehicle detection using 3d-lidar reflection intensity data. In ROBOT 2017: Third Iberian Robotics Conference; Springer International Publishing: Cham, Switzerland, 2017; pp. 475–486. [Google Scholar]
  32. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  33. Girshick, R. Fast R-CNN Object Detection with Caffe; Microsoft Research. arXiv 2015, arXiv:1504.08083v2. [Google Scholar]
  34. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  35. Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The Pascal Visual Object Classes (VOC) Challenge. Int. J. Comput. Vis. 2009, 88, 303–338. [Google Scholar] [CrossRef]
  36. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Zitnick, C.L. Microsoft coco: Common objects in context. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]
  37. Asmaa, O.; Mokhtar, K.; Abdelaziz, O. Road traffic density estimation using microscopic and macroscopic parameters. Image Vis. Comput. 2013, 31, 887–894. [Google Scholar] [CrossRef]
  38. Zhu, F. A video-based traffic congestion monitoring system using adaptive background subtraction. In Proceedings of the Second International Symposium on Electronic Commerce and Security, Nanchang, China, 22–24 May 2009; pp. 73–77. [Google Scholar]
  39. Zhu, F.; Li, L. An optimized video-based traffic congestion monitoring system. In Proceedings of the Third International Conference on Knowledge Discovery and Data Mining, Phuket, Thailand, 9–10 January 2010; pp. 150–153. [Google Scholar]
  40. Kanungo, A.; Sharma, A.; Singla, C. Smart traffic lights switching and traffic density calculation using video processing. In Proceedings of the 2014 Recent Advances in Engineering and Computational Sciences (RAECS), Chandigarh, India, 6–8 March 2014; pp. 1–6. [Google Scholar]
  41. Perkasa, O.; Widyantoro, D.H. Video-based system development for automatic traffic monitoring. In Proceedings of the International Conference on Electrical Engineering and Computer Science (ICEECS), Kuta, Indonesia, 24–25 November 2014; pp. 240–244. [Google Scholar]
  42. Eamthanakul, B.; Ketcham, M.; Chumuang, N. The traffic congestion investigating system by image processing from cctv camera. In Proceedings of the International Conference on Digital Arts, Media and Technology (ICDAMT), Kuta, Indonesia, 24–25 November 2017; pp. 240–245. [Google Scholar]
  43. Xun, F.; Yang, X.; Xie, Y.; Wang, L. Congestion detection of urban intersections based on surveillance video. In Proceedings of the 18th International Symposium on Communications and Information Technologies (ISCIT), Bangkok, Thailand, 26–28 September 2018; pp. 495–498. [Google Scholar]
  44. Ke, X.; Shi, L.; Guo, W.; Chen, D. Multi-dimensional traffic congestion detection based on fusion of visual features and convolutional neural network. IEEE Trans. Intell. Transp. Syst. 2019, 20, 2157–2170. [Google Scholar] [CrossRef]
  45. Li, W.; Dai, H.Y. Real-time road congestion detection based on image texture analysis. Procedia Eng. 2016, 137, 196–201. [Google Scholar]
  46. Haralick, R.M.; Shanmugnm, K.; Dinstein, I. Textural features for image classification. IEEE Trans. Syst. Man. Cybern. 1973, 3, 610–621. [Google Scholar] [CrossRef]
  47. Xu, Y. Crowd density estimation using texture analysis and learning. In Proceedings of the IEEE International Conference on Robotics and Biomimetics, Kunming, China, 17–20 December 2006; pp. 214–219. [Google Scholar]
  48. Lozano, A.; Manfredi, G.; Nieddu, L. An algorithm for the recognition of levels of congestion in road traffic problems. Math. Comput. Simul. 2009, 79, 1926–1934. [Google Scholar] [CrossRef]
  49. Wang, P.; Li, L.; Jin, Y.; Wang, G. Detection of unwanted traffic congestion based on existing surveillance system using in freeway via a CNN-architecture trafficnet. In Proceedings of the 13th IEEE Conference on Industrial Electronics and Applications (ICIEA), Wuhan, China, 31 May–2 June 2018; pp. 1134–1139. [Google Scholar]
  50. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  51. Wang, P.; Hao, W.; Sun, Z.; Wang, S.; Tan, E.; Li, L.; Jin, Y. Regional detection of traffic congestion using in a large-scale surveillance system via deep residual TrafficNet. IEEE Access 2018, 6, 68910–68919. [Google Scholar] [CrossRef]
  52. Kurniawan, J.; Syahra, S.G.; Dewa, C.K. Traffic Congestion Detection: Learning from CCTV Monitoring Images using Convolutional Neural Network. Procedia Comput. Sci. 2018, 144, 291–297. [Google Scholar] [CrossRef]
  53. Guerrero-Gomez-Olmedo, R.; Lopez-Sastre, R.J.; Maldonado-Bascon, S.; Fernandez-Caballero, A. Vehicle tracking by simultaneous detection and viewpoint estimation. In Proceedings of the IWINAC 2013, Part II, LNCS 7931, Mallorca, Spain, 10–14 June 2013; pp. 306–316. [Google Scholar]
  54. Chan, B.; Vasconcelos, N. Probabilistic kernels for the classification of auto-regressive visual processes. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; pp. 846–851. [Google Scholar]
  55. Chan, V.N.; Antoni, B. Classification and retrieval of traffic video using auto-regressive stochastic processes. In Proceedings of the IEEE Intelligent Vehicles Symposium, Las Vegas, NV, USA, 6–8 June 2005; pp. 771–776. [Google Scholar]
  56. Snijders, C.; Matzat, U.; Reips, U.D. Big Data: Big gaps of knowledge in the field of Internet. Int. J. Internet Sci. 2012, 7, 1–5. [Google Scholar]
Figure 1. The morphological operator applied to the frame of the vehicular traffic video.
Figure 1. The morphological operator applied to the frame of the vehicular traffic video.
Sensors 19 05213 g001
Figure 2. GRAM dataset: (left) a pixel mask that highlights the Region of Interest about vehicles present, annotated (and subsequently trackable) in each video frame (right).
Figure 2. GRAM dataset: (left) a pixel mask that highlights the Region of Interest about vehicles present, annotated (and subsequently trackable) in each video frame (right).
Sensors 19 05213 g002
Figure 3. Three images from Trafficdb that depict a traffic state classified as Light, Medium, and Heavy.
Figure 3. Three images from Trafficdb that depict a traffic state classified as Light, Medium, and Heavy.
Sensors 19 05213 g003
Table 1. Performance comparison according to the processing time and vehicle detection accuracy of the four object detectors on the three videos in the GRAM dataset.
Table 1. Performance comparison according to the processing time and vehicle detection accuracy of the four object detectors on the three videos in the GRAM dataset.
M-30M-30-HDUrban1
Haar CascadeTime [s]0.08–0.130.3–0.440.02–0.06
Accuracy43%75%40%
SSDTime [s]4–711–142.6–5.6
Accuracy22%70%69%
YOLO v3Time [s]1.0–1.81.0–1.81.0–1.8
Accuracy82%86%91%
Mask R-CNNTime [s]2.4–3.02.4–3.02.4–3.0
Accuracy89%91%46%
Table 2. Traffic state classification accuracy on the Trafficdb dataset.
Table 2. Traffic state classification accuracy on the Trafficdb dataset.
KNNSVM (Linear)SVM (rbf)Random Forest
YOLO v30.81 ± 0.100.78 ± 0.120.79 ± 0.160.84 ± 0.13
Haar Cascade0.66 ± 0.210.64 ± 0.110.64 ± 0.080.68 ± 0.21
Table 3. Normalized confusion matrix of the traffic state classification reached by the Random Forest classifier on the Trafficdb video dataset.
Table 3. Normalized confusion matrix of the traffic state classification reached by the Random Forest classifier on the Trafficdb video dataset.
Random ForestLight (Pred)Medium (Pred)Heavy (Pred)
Light0.940.020.04
Medium0.420.470.11
Heavy0.200.110.68
Table 4. Normalized confusion matrix about the binary traffic state classification performed by the deep neural network of Kurniawan et al. [53] on the Trafficdb video dataset.
Table 4. Normalized confusion matrix about the binary traffic state classification performed by the deep neural network of Kurniawan et al. [53] on the Trafficdb video dataset.
Deep Learning Architecture [44]Light (Pred)Heavy (Pred)
Light0.9950.004
Heavy01
Table 5. Normalized confusion matrix about the multiclass traffic state classification performed by the deep neural network of Kurniawan et al. [53] on the Trafficdb video dataset.
Table 5. Normalized confusion matrix about the multiclass traffic state classification performed by the deep neural network of Kurniawan et al. [53] on the Trafficdb video dataset.
Deep Learning Architecture [43]Light (Pred)Medium (Pred)Heavy (Pred)
Light0.9970.0030.
Medium0.0040.9720.025
Heavy0.0.0400.959

Share and Cite

MDPI and ACS Style

Impedovo, D.; Balducci, F.; Dentamaro, V.; Pirlo, G. Vehicular Traffic Congestion Classification by Visual Features and Deep Learning Approaches: A Comparison. Sensors 2019, 19, 5213. https://doi.org/10.3390/s19235213

AMA Style

Impedovo D, Balducci F, Dentamaro V, Pirlo G. Vehicular Traffic Congestion Classification by Visual Features and Deep Learning Approaches: A Comparison. Sensors. 2019; 19(23):5213. https://doi.org/10.3390/s19235213

Chicago/Turabian Style

Impedovo, Donato, Fabrizio Balducci, Vincenzo Dentamaro, and Giuseppe Pirlo. 2019. "Vehicular Traffic Congestion Classification by Visual Features and Deep Learning Approaches: A Comparison" Sensors 19, no. 23: 5213. https://doi.org/10.3390/s19235213

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop