You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

18 March 2021

A Robust Road Vanishing Point Detection Adapted to the Real-world Driving Scenes

,
,
and
1
Department of Information and Communication Engineering, Yeungnam University, Gyeongsan 38544, Korea
2
Department of Electrical Engineering, Yeungnam University, Gyeongsan 38544, Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Collection Computer Vision Based Smart Sensing

Abstract

Vanishing point (VP) provides extremely useful information related to roads in driving scenes for advanced driver assistance systems (ADAS) and autonomous vehicles. Existing VP detection methods for driving scenes still have not achieved sufficiently high accuracy and robustness to apply for real-world driving scenes. This paper proposes a robust motion-based road VP detection method to compensate for the deficiencies. For such purposes, three main processing steps often used in the existing road VP detection methods are carefully examined. Based on the analysis, stable motion detection, stationary point-based motion vector selection, and angle-based RANSAC (RANdom SAmple Consensus) voting are proposed. A ground-truth driving dataset including various objects and illuminations is used to verify the robustness and real-time capability of the proposed method. The experimental results show that the proposed method outperforms the existing motion-based and edge-based road VP detection methods for various illumination conditioned driving scenes.

1. Introduction

If a set of lines that are parallel in the 3D object space is projected onto a 2D image plane of a camera under perspective view, their images will form a set of lines that intersect at a point. The intersection point is called a vanishing point (VP) [1]. Generally, the number of VPs in an image depends on the number of sets of parallel lines in the 3D object space that are projected on the image. Among a lot of VPs presented in every frame of the driving videos, the most special VP is the intersection point of lanes or road boundaries. Hereafter, the VP is called R-VP (road VP) [2] in the rest of this paper.
R-VPs provide important information regarding driving scenes for ADAS (advanced driver assistance systems) and autonomous vehicles. For example, camera rotations [3], depth distance estimations, an instantaneous moving direction of a host vehicle, drivable areas of roads, etc., can be determined by using R-VPs.
Recently, deep learning (DL) has achieved a lot of impressive successes. Particularly, CNN (convolutional neural networks) has obtained great results in image classification, image recognition, etc. Some studies have applied DL to R-VP detection. CNN is used to train R-VP detectors in [4,5,6]. Regression ResNet-34 is used as an R-VP detector in [7]. A combination of CNN and heatmap regression is applied to R-VP detection, where features get from a modified version of HRNet [8]. These DL-based R-VP detectors detect R-VPs for limited driving situations. Moreover, the supervised learning method in general is the data-driven learning method that needs a huge size of training datasets with the exact labels in real-world driving scenes. As can be seen in [4,5,6,7,8], the number of images in the training datasets is limited. It is still difficult to realize a sophisticated DL-based R-VP detection that is suitable for driving scenes in practice.
Currently, geometrical feature-based R-VP detection methods for driving scenes are still being studied and utilized. Recently, the number of geometrical feature-based methods for R-VP detection increased significantly. These recent existing methods can be classified into four groups based on types of geometrical features used in the methods, such as line segment-based [9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29], edge-based [30,31,32,33], motion-based [34,35,36,37], and texture-based methods [38,39,40,41,42,43,44,45,46,47,48,49,50,51,52].
Most geometrical feature-based R-VP detection methods include three main processing steps: Step (1) Line segments are extracted from various features in driving images. Step (2) The line segments that tend to converge at R-VP locations are selected. Step (3) Intersections of the selected line segments are used to vote R-VPs through a voting process.
Optical flows of image points are generated by the movement of an observer that is a camera of a host vehicle in the R-VP detection problem. The convergence point of the optical flows of stationary image points is called the focus of expansion (FOE) of the host vehicle [53]. The great advantage of the FOE is that the FOE forms from the self-movements of the host vehicle. In other words, the FOE formation of a host vehicle does not depend on geometrical feature extractions from lanes, road boundaries, and other lines. Furthermore, if a host vehicle moves in parallel with road boundaries and the roads are almost flat, its FOE will coincide with R-VP [2]. That is why FOEs of host vehicles can be used to detect efficiently R-VPs for both structured roads and unstructured roads in the existing motion-based R-VP detection methods. This paper carefully examines the main processing steps of the existing motion-based R-VP detection methods. Some disadvantages in these processing steps are analyzed and considered to design our proposed method. In this paper, a new motion-based R-VP detection method is proposed to overcome the shortcomings of the existing motion-based methods.
The novel contributions of this paper are outlined below.
  • Stable motion vector detection is proposed to detect the stable vectors from several consecutive frames.
  • Stationary point-based motion vector selection is proposed to keep useful motion vectors and to reduce unhelpful ones for the R-VP detection.
  • Angle-based RANSAC voting is proposed to reduce efficiently the contributions of the outliers.
  • The proposed R-VP detection method that consists of stable motion detection, stationary point-based motion vector selection, and angle-based RANSAC (RANdom SAmple Consensus) voting achieves high accuracy and robustness for real-world driving scenes.
The next section discusses related research work. Section 3 explains the proposed method in detail. Section 4 shows experimental results. We conclude in Section 5.

3. Proposed Method

The proposed method also has the same main processing steps as shown in Figure 1. However, the three processing steps are proposed to improve the robustness and real-time ability of the R-VP detection. The flowchart of the proposed R-VP detection method is shown in Figure 3. The following sub-sections describe three processing steps in detail.
Figure 3. Flowchart of the proposed R-VP detection method.

3.1. Stable Motion Vector Detection

In contrast to most existing motion-based R-VP detection methods that obtain motion vectors from two consecutive frames, our motion vector detection estimates stable motion vectors over several consecutive frames. From a given first frame f t 0 at a time t 0 , corner points are detected by using the Shi-Tomasi corner detector [54]. All corner points form an initial set, denoted as P t 0 = { p t 0 ( 1 ) , p t 0 ( 2 ) , , p t 0 ( n t 0 ) } , where n t 0 is the number of corner points. Every point in the initial set is tracked over the successive frames f t 0 + k ( k = 1 , 2 , 3 , ) by using the Kanade-Lucas optical flow method [55]. Any point in the initial set that is not tracked in the current frame f t 0 + k is deleted from the set P t 0 . Otherwise, the successive tracked points are updated into the tracked set P t 0 + k . Even though the set is denoted as P t 0 + k , only one set is used and updated for a different time t 0 + k . The initial point p t 0 ( i ) P t 0 and its corresponding tracked point p t 0 + k ( i ) P t 0 + k have the same index i and indicate the pixel-wise coordinates.
Detail procedures to detect stable motion vectors are as follows. If the corner points are not tracked in the current frame f t 0 + k or have a very short displacement (smaller than a predefined distance T D ) between the previous frame f t 0 + k 1 and the current frame f t 0 + k , the points in the initial set are deleted. The number of remaining corner points at the frame f t 0 + k decreases to n t 0 + k ( n t 0 + k 1 ) , then the initial set and tracked set are now updated as P t 0 = { p t 0 ( 1 ) , p t 0 ( 2 ) , , p t 0 ( n t 0 + k ) } and P t 0 + k = { p t 0 + k ( 1 ) ,   p t 0 + k ( 2 ) , , p t 0 + k ( n t 0 + k ) } , respectively. The lifetime of tracked points depends on how fast or slow the movement of the host vehicle is. The corner point tracking process performs iteratively until the number of the remaining points n t 0 + k is less than a pre-defined threshold number T N . Whenever the number of remaining points is less than T N , new corner points are detected in the actual frame and appended to the initial set. The new detected corner points and remaining old ones keep a sufficient number of points to estimate stable motion vectors.
The motion vector set V t 0 + k = { v t 0 + k ( 1 ) , v t 0 + k ( 2 ) , , v t 0 + k ( n t 0 + k ) } is obtained from every remaining point in the tracked set and its corresponding point in the initial set. The length of each motion vector v t 0 + k ( i ) is the distance between a point p t 0 ( i ) and its corresponding point p t 0 + k ( i ) at the current frame f t 0 + k . Figure 2b shows that the estimated motion vectors of the proposed method are much longer than the ones taken from two consecutive frames. Compared to Figure 2a, most redundant and short motion vectors are eliminated in Figure 2b. As long motion vectors appear over many successive frames, they are considered to be stable motion vectors. Besides, many long motion vectors coincide with lanes and road boundaries. The stable motion vectors are very helpful to improve the performance of R-VP detection.

3.2. Stationary Point-Based Motion Vector Selection

In general, there are three different types of motion vectors in the driving scenes when a camera of a host vehicle moves. The first is motion vectors that are generated from the approaching of stationary objects such as lanes, road boundaries, trees, traffic signs, buildings, etc. Two other motion vectors occur from other moving objects. The second type of vector is generated from other accelerating, overtaking, and land changing vehicles. The third one occurs from on-coming or decelerating vehicles. Figure 4a shows an example of the three types of motion vectors. In this Figure, the numbers 1, 2, and 3 indicate motion vectors from the stationary background, an accelerating vehicle, and a decelerating vehicle, respectively. Type 2 and type 3 motion vectors are numerous and might generate many incorrect R-VPs. Moreover, if the number of moving objects increases, their appearances will affect strongly the accuracy of R-VP detection. To achieve high performance of the R-VP detection, it is important to reduce the motion vectors type 2 and type 3 as much as possible, maintaining type 1.
Figure 4. (a) Three types of motion vectors in a driving scene, where motion vectors types 1, 2, and 3 come from the stationary background, accelerating, and decelerating vehicles, respectively. The motion vectors (b) when applying motion vector type 2 reduction sub-step and (c) when applying both motion vector type 2 and 3 reduction sub-steps.
In the following sub-sections, the proposed stationary point-based motion vector selection method is introduced to reduce two different types of motion vectors respectively, based on the experimental observations.

3.2.1. Reduction of Motion Vector Type 2

It is observed that a current tracked point p t 0 + k ( i ) belonging to motion vector type 2 is closer to the image center when k increases. Then, the motion vector set V t 0 + k at the time t 0 + k is updated to maintain motion vectors of which the head points p t 0 + k ( i ) get farther away from the image center p c as follows;
V t 0 + k = { v t 0 + k ( i )   |   p t 0 + k ( i ) p c > ( p t 0 ( i ) p c ) , v t 0 + k ( i ) V t 0 + k }
The above constraint does not guarantee to eliminate all motion vector type 2. It is observed that the motion vector with a completely different direction has equal or very similar p t 0 ( i ) p c and p t 0 + k ( i ) p c . For this case, an extended actual head point denoted as p e x t _ t 0 + k ( i ) is calculated using Equation (2) and applied to Equation (1) to eliminate efficiently the motion vector type 2.
p e x t _ t 0 + k ( i ) = p t 0 + k ( i ) + l v t 0 + k ( i ) v t 0 + k ( i )
where p t 0 + k ( i ) is extended l pixels along the direction of the corresponding motion vector v t 0 + k ( i ) . Figure 4b shows an example of motion vector type 2 reduction where the left side white vehicle-related motion vectors are almost eliminated.

3.2.2. Reduction of Motion Vector Type 3

The motion vector type 3 comes from on-coming or decelerating vehicles. As both motion vector types 1 and 3 have similar directions and lengths, it is difficult to distinguish type 3 from type 1. For such reasons, the region where the type 3 motion occurs frequently is excluded in [34]. The angle between every motion vector and image horizon is used to exclude the region. Motion vectors having an angle in the interval [ 0 0 , 30 0 ] are excluded in [34]. Our motion vector type 3 reduction uses the angle criterion, but a smaller angle interval is applied to prevent from excluding type 1. Motion vectors with angles in the interval [ 0 0 , 10 0 ] are excluded from the current vector set. A vector length-based reduction is proposed to reduce the remaining type 3, based on the experimental observation that type 3 tends to be shorter than type 1. A certain percentage of the longest motion vectors are kept in the current vector set V t 0 + k and others are excluded. In summary, the vectors that have small angles with image horizon or short lengths are eliminated from the current vector set. Figure 4c shows that the right-side truck-related motion vectors are reduced.

3.3. Angle-Based RANSAC Voting Method

In this section, we propose an angle-based RANSAC voting method to overcome the disadvantage of the existing perpendicular distance-based metrics as mentioned in Section 2.3 (Figure 5a). As shown in Figure 5b, the intersection point (indicated by the blue dot) calculated from randomly selected two vectors (indicated by black lines) in the set V t 0 + k is hypothesis R-VP p H . The angle θ ( i ) between two vectors v t 0 + k ( i ) and u t 0 + k ( i ) is calculated as given in Equation (3) and used for voting. u t 0 + k ( i ) is a vector pointing in the direction from point p t 0 + k ( i ) to the hypothesis p H .
θ ( i ) = arccos ( u t 0 + k ( i ) v t 0 + k ( i ) u t 0 + k ( i )   v t 0 + k ( i ) )
Figure 5. Examples of (a) perpendicular distance metric and (b) the proposed angle-based metric between vectors and hypothesis R-VP, assuming that hypothesis R-VP p H is obtained from randomly selected two vectors.
As the angle θ ( i ) is smaller when the vector v t 0 + k ( i ) orients closer to the hypothesis, it can be used as a good metric for voting. As shown in Figure 5b, given two vectors v t 0 + k ( i ) and v t 0 + k ( j ) (indicated by the red arrows) have different angles θ ( i ) and θ ( j ) , unlike the perpendicular distance metric described in Section 2.3 and Figure 5a. Even when two motion vectors move further along their respective directions, the vector v t 0 + k ( i ) orienting closer to the hypothesis has still a smaller angle than another v t 0 + k ( j ) . This is the reason why the angle-based metric is helpful to achieve higher performance of the R-VP detection compared to the perpendicular distance-based one. The angle metric is used in the proposed angle-based RANSAC voting method.
A score function is designed by using the exponential function so that a vector with a smaller angle contributes more to the voting. The score S ( i ) is calculated on every motion vector v t 0 + k ( i ) V t 0 + k for the given hypothesis. A motion vector is regarded as an inlier if it has an angle θ ( i ) that is smaller than the predefined T θ .
S ( i ) = { e | θ ( i ) | θ ( i ) < T θ 0 o t h e r w i s e

4. Experiments

For the simulations, the proposed R-VP detection is tested on Jiqing Expressway dataset [56] that includes various driving conditions such as straight roads, slightly curved roads, shadows, illuminations, occlusions, moving objects, etc. The dataset is composed of 32 video clips with FHD (1920 × 1080) dimensions. Each video clip has 5393 frames, then the total number of frames is 172,576 frames (=32 clips × 5393 frames). The coordinates of points on the lanes are fully provided in the dataset. Intersection points of the lanes are calculated and used as ground-truth R-VPs.
Two existing R-VP detection methods are also tested for comparison with the proposed method. The motion-based method [34] is implemented as faithfully as possible, which consists mainly of two successive frames-based motion vector detection, region-based motion vector selection, and the perpendicular distance metric-based RANSAC voting. Maximum 500 motion vectors are detected from every two successive frames and used for the voting. The edge-based R-VP method [30] using Canny edge detector and probabilistic Hough Transform is also implemented as faithfully as possible. For the proposed method, 500 corner points are detected in the initial frame ( n t 0 = 500 ) and tracked. Whenever the number of the tracked points falls under T N = 400 , new 500 corner points will be detected and appended to the initial set. In the stable motion vector detection step, a tracked point is kept when its displacement between two successive frames is bigger than T D = 2 . The parameter T θ = 45 0 is used to distinguish inliers and outliers.
The performances are evaluated by using the normalized Euclidean distance between the detected R-VP’s coordinate ( x E , y E ) and the ground truth one ( x G , y G ) . The normalized error d i s t is given by [34].
d i s t = ( x E x G ) 2 + ( y E y G ) 2 D
Here D is the length of the diagonal of the image. The error d i s t ranges over [ 0 , 1 ) . When d i s t is closer to zero, the position of the detected R-VP is closer to that of ground-truth R-VP.
Table 1 shows the accuracy of the proposed method and the other two methods, where the average of the normalized error with its standard deviation is shown. The processing time in milliseconds is also presented for the comparison of computational complexity. To evaluate the validity of the stationary point-based motion vector selection step, the proposed method is tested in four different cases; One is the case of not using motion vector type 2 reduction and type 3 reduction. Two others are the cases of applying either type 2 or type 3 reduction. The last one is applying both reductions.
Table 1. Performance comparisons of the proposed, the motion-based [33] and the edge-based [29] methods in terms of the average distance between the detected and ground truth R-VPs, its standard deviation, and processing time.
As shown in Table 1, the proposed motion-based R-VP detection, which applies both motion vector type 2 reduction and type 3 reduction, has the highest accuracy (the smallest average error) and the smallest standard deviation. The proposed method has outstanding processing time. The proposed method achieves the average of error 0.0038549 (8.48 pixels in distance) and the standard deviation of 0.0073061 (16.08 pixels in distance). The achieved values are 3.86 and 28.58 times smaller than those of the existing motion-based method, respectively. In particular, the proposed method with a smaller RANSAC iteration number of 45 is the fastest (62.57 ms), maintaining high accuracy similar to the case of iteration number 900. Even in the case of not using the motion vector selection step, the proposed method shows higher accuracy than the two existing methods in terms of both average error and standard deviation. This means that the proposed angle-based RANSAC voting works very well. Simulation results show that the proposed stationary point-based motion vector selection is also efficient to improve the R-VP detection performance. In particular, motion vector type 2 reduction contributes to the improvements slightly more than type 3 reduction. Figure 6 shows the distribution of the normalized errors for the proposed and two existing methods. The normalized errors of the proposed method distribute closer to zero, compared to the others. The performances of the proposed method are also evaluated under three different illumination environments as in Table 2. The video dataset is divided into bright days (17 video clips), slightly dark days (5 video clips), and in-tunnel (10 video clips). The performances are evaluated separately. As shown in the table, the average error and the standard deviation are very similar in three different illuminations. This demonstrates the robustness of the proposed method.
Figure 6. Distribution of normalized error of the proposed (blue), existing motion-based (green), and edge-based (red) R-VP detection methods.
Table 2. The performances of the proposed method with reduction of type 2 and type 3 under three different illumination conditions of driving scenes such as bright days, slightly dark days, and in-tunnels.
To evaluate the performances according to the feature extractions, the proposed R-VP detections using SIFT (scale-invariant feature transform) and SURF (speeded-up robust features) features instead of the Shi-Tomasi feature are tested and compared. Table 3 shows that both SIFT and SURF have higher accuracies compared to the two existing methods, but the accuracies are lower than those of using the Shi-Tomasi feature. As mentioned in [54], the Shi-Tomasi feature was mainly designed for tracking, whereas SIFT [57] was mainly designed for image matching and SURF [58] is just a speeded-up version of SIFT. Additionally, Table 3 also shows that SIFT and SURF require very high computational costs. Therefore, SIFT and SURF are not efficient features for motion vector estimation in the R-VP detection problem.
Table 3. The performances of the proposed method with reduction of type 2 and type 3 using scale-invariant feature transform (SIFT) and speeded-up robust features (SURF) features.
We also test the proposed method using the perpendicular distance metric-based RANSAC [59] instead of the proposed angle-based RANSAC. Results in Table 4 show that RANSAC using the angle-based metric is better than that of using the perpendicular distance metric.
Table 4. The performances of the proposed method with reduction of type 2 and type 3 using perpendicular distance metric-based RANSAC.
Some examples of the proposed motion-based R-VP detection are shown in Figure 7, where ground-truth and detected R-VPs are indicated as green and red dots, respectively. The motion vectors are also represented by red lines. Examples show that the proposed method works well in various illumination conditions such as bright and slightly dark daytime, tunnel, and nighttime. An extra test is performed for a nighttime scene as the nighttime video is not included in the ground-truth dataset. Note that ground-truth R-VP is not represented in Figure 7d. We observe that the detected R-VP overlaps visually the intersection point of the road boundaries. The red and green dots in Figure 7a–c almost coincide. Two examples of nighttime R-VP detection are shown in Figure 8, where video frames extracted from “ZJU Day and Night Driving Dataset” [60,61] are used without ground-truth R-VPs. White lines are road boundaries drawing by hand and the red circle is detected R-VP. The red circle and the intersection of road boundaries are also almost coincided. Such coincidences demonstrate the accuracy of the proposed method.
Figure 7. Some examples of the proposed motion-based R-VP detection in various illumination conditions such as (a) bright and (b) slightly dark daytime, (c) tunnel, and (d) nighttime. Note that ground-truth R-VP is not indicated in (d).
Figure 8. Detected R-VPs (indicated by red circles) and road boundaries drawing by hand (by white lines) under nighttime conditions using the proposed method applying reduction of type 2 and type 3, using (a) RANSAC iteration 900 and (b) iteration 45.
In general, we observed that the proposed method can detect R-VP efficiently when host vehicles move in parallel with road boundaries and road surfaces are almost flat as mentioned in Section 1. The proposed method sometimes fails to detect correctly R-VP in cases of uneven (or bumpy) and curved roads.

5. Conclusions

In this paper, we proposed a motion-based R-VP detection method. Three main processing steps often used in the existing R-VP detection methods are carefully examined. Based on the analysis, we propose stable motion detection, stationary point-based motion vector selection, and angle-based RANSAC voting, which are successfully applied to improve the performances of the main three steps.
Through the simulations, we show that the proposed stable motion detection and angle-based RANSAC voting contribute considerably to the improvement of R-VP detection performance. In particular, the proposed angle-based RANSAC voting works much better than conventional perpendicular distance-based voting. The proposed motion-based R-VP detection that combines all three proposed three-step algorithms outperforms the existing motion-based and edge-based R-VP detection methods. The simulation results show that the proposed method is applicable for nighttime, as well as various illumination conditioned driving scenes. It is also shown that the standard deviation of the detected R-VP position is the smallest, compared to others. The performances are very similar in three different illumination conditions. This means the proposed method is robust as well as accurate compared to the existing R-VP detection methods. From the viewpoint of computational complexity, the proposed can be realized in real-time. It is possible to apply selectively the proposed three-step algorithms depending on the system resources. Even though the proposed method is not evaluated on the unstructured road driving video because of the lack of the ground truth datasets, it is expected that it works in a stable on the unstructured road driving.
The proposed R-VP detection method plays an important role to improve the performances of various applications such as online camera calibration, detection of drivable areas, and driving direction of the host vehicle.

Author Contributions

Conceptualization, C.N.K.; methodology, C.N.K.; software, C.N.K.; validation, C.N.K.; formal analysis, C.N.K.; investigation, C.N.K.; resources, C.N.K.; data curation, C.N.K. and J.H.P.; writing—original draft preparation, C.N.K.; writing—review and editing, C.N.K. and H.-Y.J.; visualization, C.N.K. and Y.C.; supervision, J.H.P.; project administration, H.-Y.J.; funding acquisition, H.-Y.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by Basic Science Research Programs of the Ministry of Education (grant NRF-2018R1A2B6005105), in part by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (number 2019R1A5A8080290), and in part by Ministry of Science, ICT (MSIT), Korea, under the Information Technology Research Center (ITRC) support program (IITP-2020-2016-0-00313) supervised by the IITP(Institute for Information and communications Technology Planning and Evaluation).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hearn, D.; Baker, M.P. Chapter 12: Three-Dimensional Viewing. In Computer Graphics C Version, 2nd ed.; Pearson: Noida, India, 1997; pp. 446–447. [Google Scholar]
  2. Guo, X.; Li, Q.; Sun, C. A new road tracking method based on heading direction detection. J. Automob. Eng. 2018, 233, 232–248. [Google Scholar] [CrossRef]
  3. Orghidan, R.; Salvi, J.; Gordan, M.; Orza, B. Camera calibration using two or three vanishing points. In Proceedings of the Federated Conference on Computer Science and Information Systems (FedCSIS), Wrocław, Poland, 9–12 September 2012; pp. 123–130. [Google Scholar]
  4. Chang, C.; Zhao, J.; Itti, L. DeepVP: Deep Learning for Vanishing Point Detection on 1 Million Street View Images. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018; pp. 4496–4503. [Google Scholar]
  5. Xingxin, L.; Liqiang, Z.; Zujun, Y.; Yanqin, W. Adaptive auxiliary input extraction based on vanishing point detection for distant object detection in high-resolution railway scene. In Proceedings of the 14th IEEE International Conference on Electronic Measurement & Instruments (ICEMI), Changsha, China, 1–3 November 2019; pp. 522–527. [Google Scholar]
  6. Zeng, Z.; Wu, M.; Zeng, W.; Fu, C.W. Deep Recognition of Vanishing-Point-Constrained Building Planes in Urban Street Views. IEEE Trans. Image Process. 2020, 29, 5912–5923. [Google Scholar] [CrossRef]
  7. Han-Soo, C.; Keunhoi, A.; Myungjoo, K. Regression with residual neural network for vanishing point detection. Image Vis. Comput. J. Books 2019, 91, 103797. [Google Scholar]
  8. Liu, Y.B.; Zeng, M.; Meng, Q.H. Unstructured Road Vanishing Point Detection Using Convolutional Neural Networks and Heatmap Regression. IEEE Trans. Instrum. Meas. 2020, 70, 1–8. [Google Scholar] [CrossRef]
  9. Wang, Y.; Lu, X.; Ling, Z.; Yang, Y.; Zhang, Z.; Wang, K. A Method to Calibrate Vehicle-Mounted Cameras under Urban Traffic Scenes. IEEE Trans. Intell. Transp. Syst. 2015, 16, 3270–3279. [Google Scholar] [CrossRef]
  10. Yoo, J.H.; Lee, S.; Park, S.; Kim, D.H. A Robust Lane Detection Method Based on Vanishing Point Estimation Using the Relevance of Line Segments. IEEE Trans. Intell. Transp. Syst. 2017, 18, 3254–3266. [Google Scholar] [CrossRef]
  11. Ding, W.; Li, Y. Efficient vanishing point detection method in complex urban road environments. IET Comput. Vis. 2015, 9, 549–558. [Google Scholar] [CrossRef]
  12. Wang, E.; Sun, A.; Li, Y.; Hou, X.; Zhu, Y. Fast vanishing point detection method based on road border region estimation. IET Image Process. 2018, 12, 361–373. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Su, Y.; Yang, J.; Ponce, J.; Kong, H. When Dijkstra Meets Vanishing Point: A Stereo Vision Approach for Road Detection. IEEE Trans. Image Process. 2018, 27, 2176–2188. [Google Scholar] [CrossRef]
  14. Ding, W.; Li, Y.; Liu, H. Efficient vanishing point detection method in unstructured road environments based on dark channel prior. IET Comput. Vis. 2016, 10, 852–860. [Google Scholar] [CrossRef]
  15. Hwang, H.J.; Yoon, G.; Yoon, S.M. Optimized Clustering Scheme-Based Robust Vanishing Point Detection. IEEE Trans. Intell. Transp. Syst. 2020, 21, 199–208. [Google Scholar] [CrossRef]
  16. VanQuang, N.; Heungsuk, K.; SeoChang, J.; Kwangsuck, B. A Study on Real-Time Detection Method of Lane and Vehicle for Lane Change Assistant System Using Vision System on Highway. Eng. Sci. Technol. Int. J. 2018, 21, 822–833. [Google Scholar]
  17. Kim, J. Efficient Vanishing Point Detection for Driving Assistance Based on Visual Saliency Map and Image Segmentation from a Vehicle Black-Box Camera. Symmetry 2019, 11, 1492. [Google Scholar] [CrossRef]
  18. Pei, L.; Liu, K.; Zou, D.; Li, T.; Wu, Q.; Zhu, Y.; Li, Y.; He, Z.; Chen, Y.; Sartori, D. IVPR: An Instant Visual Place Recognition Approach Based on Structural Lines in Manhattan World. IEEE Trans. Instrum. Meas. 2020, 69, 4173–4187. [Google Scholar] [CrossRef]
  19. Li, B.; Guo, Y.; Zhou, J.; Cai, Y.; Xiao, J.; Zeng, W. Lane Detection and Road Surface Reconstruction Based on Multiple Vanishing Point & Symposia. In Proceedings of the IEEE Intelligent Vehicles Symposium (IV), Changshu, Suzhou, 30 June–1 July 2018; pp. 209–214. [Google Scholar]
  20. Suttorp, T.; Bucher, T. Robust vanishing point estimation for driver assistance. In Proceedings of the IEEE Intelligent Transportation Systems Conference, Toronto, ON, Canada, 17–20 September 2006; pp. 1550–1555. [Google Scholar]
  21. Young-Woo, S.; Ragnunathan, R. Detection and tracking of the vanishing point on a horizon for automotive applications. In Proceedings of the 6th IEEE IROS Workshop on Planning, Perception and Navigation for Intelligent Vehicles (PPNIV-14), Chicago, IL, USA, 14–18 September 2014. [Google Scholar]
  22. Kroeger, T.; Dai, D.; Timofte, R.; Van Gool, L. Discovery of Sets of Mutually Orthogonal Vanishing Points in Videos. In Proceedings of the 2015 IEEE Winter Applications and Computer Vision Workshops, Waikoloa, HI, USA, 6–9 January 2015; pp. 63–70. [Google Scholar]
  23. Ju-Han, Y.; Dong-Hwan, K.; Sung-Kee, P. A new lane detection method based on vanishing point estimation with probabilistic voting. In Proceedings of the IEEE International Conference on Consumer Electronics (ICCE), Berlin, Germany, 6–9 September 2015; pp. 204–205. [Google Scholar]
  24. Young-woo, S.; Ragunathan, R. Use of a monocular camera to analyze a ground vehicle’s lateral movements for reliable autonomous city driving. In Proceedings of the IEEE IROS Workshop on Planning, Perception and Navigation for Intelligent Vehicles, Tokyo, Japan, 3–7 November 2013. [Google Scholar]
  25. Tian, Y.; Cui, W.; Li, X.; Chen, L. A Robust Lane Detection Method Based on Vanishing Point Estimation. Procedia Comput. Sci. 2018, 131, 354–360. [Google Scholar]
  26. Piao, J.; Shin, H. Robust hypothesis generation method using binary blob analysis for multi-lane detection. IET Image Process. 2017, 11, 1210–1218. [Google Scholar] [CrossRef]
  27. Yong, L.; Weili, D.; XuGuang, Z.; Zhaojie, J. Road detection algorithm for Autonomous Navigation Systems based on dark channel prior and vanishing point in complex road scenes. Robot. Auton. Syst. 2016, 85, 1–11. [Google Scholar]
  28. Jiaming, H.; Zhong, Y.; Guoxiong, H.; Ting, F.; Hao, X. Robust and efficient vanishing point detection in unstructured road scenes for assistive navigation. Sens. Rev. 2018, 39, 137–146. [Google Scholar]
  29. Bazin, J.; Pollefeys, M. 3-line RANSAC for orthogonal vanishing point detection. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Vilamoura-Algarve, Portugal, 7–12 October 2012; pp. 4282–4287. [Google Scholar]
  30. Yoon-Young, M.; Zong-Woo, G.; Gi-Tae, H. Vanishing point detection for self-driving car using harmony search algorithm. Swarm Evol. Comput. 2018, 41, 111–119. [Google Scholar]
  31. Su, Y.; Zhang, Y.; Alvarez, J.M.; Kong, H. An illumination-invariant nonparametric model for urban road detection using monocular camera and single-line lidar. In Proceedings of the IEEE International Conference on Robotics and Biomimetics (ROBIO), Macau, China, 5–8 December 2017; pp. 68–73. [Google Scholar]
  32. Mistry, V.H.; Makwana, R.M. Computationally efficient vanishing point detection algorithm based road segmentation in road images. In Proceedings of the IEEE International Conference on Advances in Electronics, Communication and Computer Technology (ICAECCT), New York, NY, USA, 2–3 December 2016; pp. 103–110. [Google Scholar]
  33. Wang, Z.; Qu, W.; Kamata, S. Vehicle detection from onboard camera using patch decided vanishing point. In Proceedings of the 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Shanghai, China, 14–16 October 2017; pp. 1–7. [Google Scholar]
  34. Yu, Z.; Zhu, L. Roust Vanishing Point Detection Based on the Combination of Edge and Optical Flow. In Proceedings of the 4th Asia-Pacific Conference on Intelligent Robot Systems (ACIRS), Nagoya, Japan, 13–15 July 2019; pp. 184–188. [Google Scholar]
  35. Yu, Z.; Zhu, L.; Lu, G. An Improved Phase Correlation Method for Stop Detection of Autonomous Driving. IEEE Access 2020, 8, 77972–77986. [Google Scholar] [CrossRef]
  36. Jang, J.; Jo, Y.; Shin, M.; Paik, J. Camera Orientation Estimation Using Motion-Based Vanishing Point Detection for Advanced Driver-Assistance Systems. IEEE Trans. Intell. Transp. Syst. 2020, 1–11. [Google Scholar] [CrossRef]
  37. Jo, Y.; Jang, J.; Paik, J. Camera orientation estimation using motion based vanishing point detection for automatic driving assistance system. In Proceedings of the IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 12–14 January 2018; pp. 1–2. [Google Scholar]
  38. Yang, W.; Fang, B.; Tang, Y.Y. Fast and Accurate Vanishing Point Detection and Its Application in Inverse Perspective Mapping of Structured Road. IEEE Trans. Syst. Man Cybern. Syst. 2018, 48, 755–766. [Google Scholar] [CrossRef]
  39. Yang, G.; Wang, Y.; Yang, J.; Lu, Z. Fast and Robust Vanishing Point Detection Using Contourlet Texture Detector for Unstructured Road. IEEE Access 2019, 7, 139358–139367. [Google Scholar] [CrossRef]
  40. Moghadam, P.; Dong, J.F. Road direction detection based on vanishing-point tracking. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Algarve, Portugal, 7–12 October 2012; pp. 1553–1560. [Google Scholar]
  41. Hui, K.; Jean-Yves, A.; Jean, P. Vanishing point detection for road detection. In Proceedings of the CVPR 2009—IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–21 June 2009; pp. 96–103. [Google Scholar]
  42. Min, Z.; Yisha, L.; Yan, Z.; Huosheng, H. Visual Campus Road Detection for an UGV using Fast Scene Segmentation and Rapid Vanishing Point Estimation. IFAC Proc. Vol. 2014, 47, 11898–11903. [Google Scholar]
  43. Deng, T.; Yan, H.; Li, Y. Learning to Boost Bottom-Up Fixation Prediction in Driving Environments via Random Forest. IEEE Trans. Intell. Transp. Syst. 2018, 19, 3059–3067. [Google Scholar] [CrossRef]
  44. Moghadam, P.; Starzyk, J.A.; Wijesoma, W.S. Fast Vanishing-Point Detection in Unstructured Environments. IEEE Trans. Image Process. 2012, 21, 425–430. [Google Scholar] [CrossRef]
  45. Shi, J.; Wang, J.; Fu, F. Fast and Robust Vanishing Point Detection for Unstructured Road Following. IEEE Trans. Intell. Transp. Syst. 2016, 17, 970–979. [Google Scholar] [CrossRef]
  46. Fan, X.; Shin, H. Road vanishing point detection using weber adaptive local filter and salient-block-wise weighted soft voting. IET Comput. Vis. 2016, 10, 503–512. [Google Scholar] [CrossRef]
  47. Yong, L.; Guofeng, T.; Anan, S.; Weili, D. Road extraction algorithm based on intrinsic image and vanishing point for unstructured road image. Robot. Auton. Syst. 2018, 109, 86–96. [Google Scholar]
  48. Bui, T.H.; Saitoh, T.; Nobuyama, E. Road area detection based on texture orientations estimation and vanishing point detection. In Proceedings of the The SICE Annual Conference, Nagoya, Japan, 14–17 September 2013; pp. 1138–1143. [Google Scholar]
  49. Neethu, J.; Anusha, B.; Krishnan, K. A Reliable Method for Detecting Road Regions from a Single Image Based on Color Distribution and Vanishing Point Location. Procedia Comput. Sci. 2015, 58, 2–9. [Google Scholar]
  50. Zhaozi, Z.; Yingtuan, H.; Dixiao, C.; Jianru, X. Real-time road detection with image texture analysis-based vanishing point estimation. In Proceedings of the IEEE International Conference on Progress in Informatics and Computing (PIC), Nanjing, China, 18–20 December 2015; pp. 454–457. [Google Scholar]
  51. Le, M.C.; Phung, S.L.; Bouzerdoum, A. Pedestrian Lane Detection in Unstructured Environments for Assistive Navigation. In Proceedings of the International Conference on Digital Image Computing: Techniques and Applications (DICTA), Wollongong, Australia, 25–27 November 2014; pp. 1–8. [Google Scholar]
  52. Nguyen, L.; Phung, S.L.; Bouzerdoum, A. Efficient Vanishing Point Estimation for Unstructured Road Scenes. In Proceedings of the International Conference on Digital Image Computing: Techniques and Applications (DICTA), Gold Coast, Australia, 30 November–2 December 2016; pp. 1–6. [Google Scholar]
  53. Negahdaripour, S.; Horn, B.K.P. A direct method for locating the focus of expansion. Comput. Vis. Graph. Image Process. J. 1989, 46, 303–326. [Google Scholar] [CrossRef][Green Version]
  54. Shi, J.; Tomasi, C. Good features to track. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 21–23 June 1994; pp. 593–600. [Google Scholar]
  55. Lucas, B.D.; Kanade, T. An Iterative Image Registration Technique with an Application to Stereo Vision. In Proceedings of the 7th international joint conference on Artificial intelligence (IJCAI’81), Vancouver, BC, Canada, 24–28 August 1981; pp. 674–679. [Google Scholar]
  56. Jiqing Expressway Dataset. Available online: https://github.com/vonsj0210/Multi-Lane-Detection-Dataset-with-Ground-Truth (accessed on 8 February 2021).
  57. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  58. Bay, H.; Tuytelaars, T.; Van Gool, L. SURF: Speeded-Up Robust Features. Comput. Vis. Image Underst. 2008, 404–417. [Google Scholar] [CrossRef]
  59. David, A.F.; Jean, P. Chapter 17: Segmentation and fitting using probabilistic methods. In Computer Vision: A Modern Approach; Prentice Hall Professional Technical Reference; Prentice Hall: Upper Saddle River, NJ, USA, 2002; p. 483. ISBN 978-0-13-085198-7. [Google Scholar]
  60. Zhou, K.; Wang, K.; Yang, K. A Robust Monocular Depth Estimation Framework Based on Light-Weight ERF-PSPNet for Day-Night Driving Scenes. In Proceedings of the International Conference on Machine Vision and Information Technology (CMVIT), Sanya, China, 20–22 February 2020. [Google Scholar]
  61. ZJU Day and Night Driving Dataset. Available online: https://github.com/elnino9ykl/ZJU-Dataset (accessed on 1 March 2021).
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.