You are currently viewing a new version of our website. To view the old version click .
Electronics
  • Article
  • Open Access

15 November 2023

LiDAR Localization by Removing Moveable Objects

,
and
1
Graduate School of Automotive Engineering, Kookmin University, Seoul 02707, Republic of Korea
2
Department of Automotive and IT Convergence, Kookmin University, Seoul 02707, Republic of Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Advancements in Connected and Autonomous Vehicles

Abstract

In this study, we propose reliable Light Detection and Ranging (LiDAR) mapping and localization via the removal of moveable objects, which can cause noise for autonomous driving vehicles based on the Normal Distributions Transform (NDT). LiDAR measures the distances to objects such as parked and moving cars and objects on the road, calculating the time of flight required for the sensor’s beam to reflect off an object and return to the system. The proposed localization system uses LiDAR to implement mapping and matching for the surroundings of an autonomous vehicle. This localization is applied to an autonomous vehicle, a mid-size Sports Utility Vehicle (SUV) that has a 64-channel Velodyne sensor, detecting moveable objects via modified DeepLabV3 and semantic segmentation. LiDAR and vision sensors are popular perception sensors, but vision sensors have a weakness that does not allow for an object to be detected accurately under special circumstances, such as at night or when there is a backlight in daylight. Even if LiDAR is more expensive than other detecting sensors, LiDAR can more reliably and accurately sense an object with the right depth because a LiDAR sensor estimates an object’s distance using the time of flight required for the LiDAR sensor’s beam to detect the object and return to the system. The cost of a LiDAR product will decrease dramatically in the case of skyrocketing demand for LiDAR in the industrial areas of autonomous vehicles, humanoid robots, service robots, and unmanned drones. As a result, this study develops a precise application of LiDAR localization for a mid-size SUV, which gives the best performance with respect to acquiring an object’s information and contributing to the appropriate, timely control of the mid-size SUV. We suggest mapping and localization using only LiDAR, without support from any other sensors, such as a Global Positioning System (GPS) or an Inertial Measurement Unit (IMU) sensor; using only a LiDAR sensor will be beneficial for cost competitiveness and reliability. With the powerful modified DeepLabV3, which is faster and more accurate, we identify and remove a moveable object through semantic segmentation. The improvement rate of the mapping and matching performance of our proposed NDT, by removing the moveable objects, was approximately 12% in terms of the Root-Mean-Square Error (RMSE) for the first fifth of the test course, where there were fewer parked cars and more moving cars.

1. Introduction

A GPS sensor error is shown in Figure 1; errors occur because GPS signals from satellites can be blocked by many tall buildings and refracted by the buildings’ glass. The solid red line shows the GPS sensor’s latitude and longitude on a satellite map; the line is disconnected in spite of a closed-loop course. The experimental vehicle was an autonomous mid-size SUV that was driven by a human driver, and the yellow-highlighted cells indicate both the real starting point and the real end point. In the blue circle, a GPS error was detected; the GPS signal diverged, and the red line was not closed. My personal experience was that when using Google Maps on my mobile phone, the GPS location froze and oscillated in a new city, which has the tallest buildings in the world. In these areas, in the centers of large cities that have taller buildings, GPS sensor errors can occur, and we need to correct the GPS sensor or develop another reliable system. Figure 1 shows the rationale for why we use LiDAR NDT mapping and location in the case of a city with taller buildings and underground parking lots.
Figure 1. Global Positioning System (GPS) error at Seoul City Hall, Seoul, Republic of Korea.
A GPS error analysis is important for understanding how a GPS works and the magnitude of the prediction error. The satellite positioning system corrects incoming clock errors and other effects, but there are some remaining uncorrected errors. A GPS receiver’s position is calculated based on data received from a satellite. A GPS receiver requires the current time, the satellite’s position, and the measurement delay of the received signal. Location accuracy depends primarily on the satellite’s position and signal delay.
In the case of a general road environment with more moveable objects, there are impacts on the localization performance caused by large differences between mapping data and LiDAR sensing data. In order to solve the performance decline problem, we removed the moveable objects via the application of a deep-learning-based semantic segmentation method. As a result, an improvement in the test result is confirmed for mapping and map matching after removing moveable objects in the original LiDAR sensor data. In order to remove the moveable objects, we needed to identify each object’s classification through semantic segmentation, using modified DeepLabV3 (which was developed based on a Python deep learning process) and Velodyne LiDAR learning data, and extract the frozen parameters first. The moveable objects identified using semantic segmentation and frozen parameters were removed using the Robot Operating System (ROS), and the performance was improved compared to the previous NDT logic by removing noise, such as moveable objects, once the NDT program was running.

3. Materials

We analyzed NDT mapping and localization through an experiment to acquire LiDAR cloud data and GPS data using the remodeled autonomous mid-size SUV in Figure 2.
Figure 2. The platform of the remodeled mid-size SUV.
Figure 3 shows a diagram of the LiDAR detection range, and the LiDAR sensor detection range was about 50 m, which is highlighted in the blue circle.
Figure 3. Diagram of LiDAR detection range.
The detailed specifications of the experimental hardware of the sensors and equipment are described in Table 1. There were two sensors attached to the test vehicle. One sensor was a 64-channel Velodyne LiDAR on the top of the vehicle, and the other sensor was a Leica GS15 GPS sensor on the top of the vehicle. The test’s desktop PC used the Melodic Robot Operation System (ROS) in Ubuntu 18.04 LTS. Real-time data were acquired from the LiDAR and GPS by the ROS via a communication method in the master management module. The LiDAR and GPS topics were subscribed from the master PC, which was installed in the autonomous vehicle in an ROS environment. The ROS contains a powerful software library and helps users to communicate sensor data well when building robot applications. The ROS is open-source from the drivers to the accurate and reliable algorithms for robot developers and autonomous vehicle engineers.
Table 1. Experimental hardware specifications.

4. Method

This study aimed to develop exact and on-time mapping and localization with 64-channel Velodyne LiDAR, removing moveable objects, such as vehicles and bicycles. Raw LiDAR data were detected at a frequency of 10 Hz. Figure 4 shows the concept of LiDAR mapping and localization by removing moveable objects. The proposed mapping and localization consists of two categories. One is moveable object elimination, and the other is NDT processing. Moveable object elimination is a key factor in our paper. Moveable object elimination means that we identify moveable objects using the semantic segmentation method and DeepLabV3 and remove the moveable objects in the ROS.
Figure 4. The concept of LiDAR mapping and localization by removing moveable objects.
We failed NDT mapping first due to a delayed process, which required heavier optimization and transformation calculation. Therefore, we applied a voxel filter in the proposed system. NDT mapping was successful with the voxel filter (filter size: 1 m), as shown in the right picture of Figure 5.
Figure 5. NDT mapping comparison with and without voxel filter.
Assuming that there are no rotations, the transformation matrix represents only the translation movement to point P and its translation matrix. This equation is only described for joint-position linear movement, not rotated movement. Based on a Cartesian coordinate system, the translation matrix describes the linear slide cylinder and is used for both the forward and inverse kinematic equations for the position of each robot joint. The P vector length is a key factor when the scale factor is one, and the matrix is described as a 3 × 4 matrix in the case of no scale factors, which is generally not the case [28].
T c a r t =     0   0   0   P x     0   1   0   P y     0   0   1   P z   0   0   0   1  
Three rotations normally exist in an autonomous vehicle based on the gravity center reference frame. The first rotation is a   , which we generally call roll, and the reference axis is the vehicle heading axis. The second rotation is o   , which we generally call pitch, and the reference axis is the lateral axis. The third rotation is n   , which we generally call yaw, and the reference axis is the vertical axis from the ground to the sky. The RPY orientation consists of three kinds of roll, pitch, and yaw, totally integrating and describing the three rotations in sequence. The symbol ‘S’ is expressed for sine, the symbol ‘C’ is expressed for cosine, and the total combined matrix integrates three rotations, which has been described [28].
R P Y a ,   o ,   n   = R o t a , a R o t o , o R o t n , n =     C ø a C ø o   C ø a S ø o S ø n     S ø a C ø n   C ø a S ø o C ø n + S ø a C ø n   0   S ø a C ø o   S ø a S ø o S ø n + C ø a C ø n   S ø a S ø o C ø n     C ø a S ø n   0   S ø o   C ø o S ø n   C ø o C ø n   0     0   0   0   1    
Once we express the coordinates of the i t h points in the grid as ‘ x i ’, the distribution mean ( μ ) and distribution covariance ( Σ ) are expressed in the following equation.
μ = 1 n i = 1 n x i
Σ = 1 n i = 1 n x i μ x i μ T
The normal distribution probability ( p x ) is expressed in the following equation.
p x ~ e x μ T Σ 1 x μ 2
In terms of two dimensions, the spatial mapping (T) for the coordinate transformation is expressed in the following equation.
T :   x y = cos ϕ sin ϕ sin ϕ cos ϕ x y + t x t y = x cos ϕ y sin ϕ + t x x sin ϕ + y cos ϕ + t y
where   t x and t y express the translation movement, and ϕ describes the rotation movement of two scan data in a sequence in terms of the time perspective. The mapping implementation is completed between the first scan datapoint and the next scan datapoint, making a calculation of the normal distribution and completing the optimization using the score equation (Equation (7)). x i describes a sample of the scanned LiDAR points, and x i indicates that x i is sorted by the parameters in P . The equation is x i = T x i , P . The P score describes the matching optimization value in the following equation.
Score P = i = 1 n e x i μ i T Σ 1 x i μ i 2
For mapping optimization, we need to determine the optimal parameters (P) to minimize the function ( f ) using the Newton algorithm. The parameters (P) minimize the function using the gradient ( g ) of the function ( f ) and the Hessian matrix (H). s is described as follows.
H Δ p = g
g i = f p i
H i j = f p i p j
μ = x i μ i ,       s = e μ T Σ 1 μ 2
In terms of P i , the partial derivative of μ can be described by x i , and the score function gradient is completed using Equation (11) as a summand. Therefore, the partial derivative of μ is estimated as the Jacobian matrix ( J T ) of T. The Hessian matrix is estimated using the score function partial derivatives in terms of every parameter.
g ˜ i = s p i = s μ μ p i = μ T Σ 1 μ p i e μ T Σ 1 μ 2
J T = 1 0 x sin ϕ y cos ϕ 1 1 x cos ϕ y sin ϕ
The Hessian matrix (H) is described in Equation (14), and the second partial derivative of μ is described using Equation (15).
H ˜ i j = s p i p j = e μ T Σ 1 μ 2 μ T Σ 1 μ p i μ T Σ 1 μ p j + μ T Σ 1 2 μ p i p j + μ T p j Σ 1 μ p i
2 μ p i p j = x cos ϕ + y sin ϕ x sin ϕ y cos ϕ ,     i = j = 3                 0 0             ,         o t h e r w i s e
The Hessian matrix is used for NDT function optimization and position estimation. Using the Hessian matrix, we calculate the derivatives of the NDT function and make a correction for position estimation. The Hessian matrix gives curvature information and the NDT function’s second-order derivative. It optimizes the function efficiently and refines the forecasted position [22].
The spherical coordinate system in Figure 6 can be used to convert to the image coordinate system, as shown in Equation (16).
u v = 1 / 2 1 arctan y , x π 1 W 1 arcsin z r 1 + f u p f 1 H
where   H is the height of an image, and W is the width of an image, while x, y, and z represent the LiDAR data. f is the Field of View (FOV) of the sensor, as shown in Equation (17).
f = f u p + f d o w n
where r is the distance from the sensor to the point, as shown in Equation (18).
r = x 2 + y 2 + z 2
Figure 6. Spherical coordinates.
The LiDAR data and r are projected onto a transformed coordinate system to generate an image [14]. The converted data are shown in Figure 7. The moveable objects, such as vehicles, were removed from the lower picture, and the vehicles are categorized in blue. The raw LiDAR data were changed for the lower picture, removing the moveable objects such as vehicles, categorized in blue through semantic segmentation.
Figure 7. Point cloud and LiDAR data projected by removing objects.

5. Experiments and Results

We drove a mid-size SUV in a closed loop near the National Assembly. The red circle indicates that we started and finished the journey in the same position. Both absolute coordinate latitude and longitude GPS data are described in Figure 8. The experiment’s resulting data are the extracted averages from four closed-loop journeys in the blue triangle line around the National Assembly in the Republic of Korea.
Figure 8. Latitude and longitude GPS data near National Assembly, Seoul, Republic of Korea.
We mapped the closed-loop course using the NDT program, and the calculated trajectory is shown in Figure 9. Both Figure 10 and Figure 11 show similar trajectories for the test line, and they have triangular shapes.
Figure 9. NDT map near National Assembly, Seoul, Republic of Korea.
Figure 10. Comparison between NDT matching with moveable objects and GT (GPS).
Figure 11. Comparison between NDT matching by removing moveable objects and GPS.
The ground truth is the red solid line (GPS data), which is based on the global absolute coordinates for latitude and longitude, and the blue solid line is NDT matching with moveable vehicles. However, NDT-matching coordinates are relative, and the origin is almost zero. In order to match the GPS absolute coordinates with the NDT coordinates, we moved the first GPS point to the origin (x = 0, y = 0) using the Equation (1) translation transformation matrix through the parallel movement in the left picture in Figure 10. However, the two trajectories were not yet comparable to determine the Root-Mean-Square Error (RMSE), and the trajectory needed to be rotated. First, we determined each unit vector between the start point and the next point to calculate the rotation angle between the GPS trajectory and the NDT-matching trajectory in the right picture of Figure 9 using Equation (2) (RPY orientation).
Figure 11 shows a performance comparison between GPS data and NDT matching by removing moveable objects in converted raw LiDAR data using the modified DeepLabV3.
As a result, we compared the RMSE between NDT matching with moveable objects and NDT matching removing moveable objects, as shown in Table 2. The RMSE of all trajectories of NDT matching with moveable objects is 0.9870, and the RMSE of all trajectories of NDT matching without moveable objects is 1.1623. The performance of our proposed NDT was inferior to normal NDT matching because one lane was full of parked cars, which are a good feature for NDT mapping and matching, after the first fifth of the course in the test area.
Table 2. Comparison of RMSE between proposed and previous NDTs.
We operated the SUV four times by configuring the closed loop with the same starting and stopping positions. The data described in Table 3 are the RMSE of each iteration for the proposed NDT, which was run four times. Figure 12 shows the same position for every iteration with different surroundings, such as vehicles.
Table 3. RMSE for every iteration of proposed NDT.
Figure 12. LiDAR scan for each iteration with different surroundings.
Comparing the RMSE in the first fifth of the course, where there were fewer parked cars and more moving cars, the RMSE of NDT matching with moveable objects was 0.1970, and the RMSE of NDT matching without moveable objects was 0.1752. The improvement rate of the mapping and matching performance of our proposed NDT was approximately 12%. The improvement rate is expected to increase in circumstances where there are more moving vehicles and fewer parked vehicles.

6. Conclusions

We cannot perform GPS-based localization accurately in certain areas, such as underground, in tunnels, and in cities with tall buildings. Therefore, we propose LiDAR-based NDT mapping and localization, which have reliable performance in areas where GPS errors are possible. The proposed method of NDT localization by removing moveable objects had 12% better performance for a specific test area with more moving vehicles.
In the future, we plan to implement the updated NDT algorithm, removing only moving objects and not deleting whole moveable objects, especially parked cars, which are a good feature for NDT mapping. We expect that the new updated algorithm, combining the JPDA-IMM-UKF algorithm, could bring out the best performance.

Author Contributions

Conceptualization, S.J. and M.K.; methodology, S.J. and M.K.; software, S.J.; validation, S.J.; formal analysis, S.J.; investigation, S.J.; resources, S.J.; data curation, S.J.; writing—original draft preparation, S.J.; writing—review and editing, J.K.; visualization, S.J.; supervision, J.K.; project administration, J.K.; funding acquisition, J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the BK21 Four Program (5199990814084) of the National Research Foundation of Korea (NRF), funded by the Ministry of Education, Korea.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xia, X.; Meng, Z.; Han, X.; Li, H.; Tsukiji, T.; Xu, R.; Ma, J. An automated driving systems data acquisition and analytics platform. Transp. Res. Part C Emerg. Technol. 2023, 151, 104120. [Google Scholar] [CrossRef]
  2. Xia, X.; Bhatt, N.; Khajepour, A.; Hashemi, E. Integrated inertial-LiDAR-based map matching localization for varying environments. IEEE Trans. Intell. Veh. 2023. [Google Scholar] [CrossRef]
  3. Jo, K.; Chu, K.; Sunwoo, M. GPS-Bias Correction for Precise Localization of Autonomous Vehicles. In Proceedings of the 2013 IEEE Intelligent Vehicles Symposium (IV), Gold Coast, Australia, 23–26 June 2013; pp. 636–641. [Google Scholar]
  4. Peter, B.; Wolfgang, S. The Normal Distributions Transform: A New Approach to Laser Scan Matching. In Proceedings of the 2003 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2003) (Cat. No.03CH37453), Las Vegas, NV, USA, 27–31 October 2003; Volume 3, pp. 2743–2748. [Google Scholar]
  5. Takubo, T.; Kaminade, T.; Mae, Y.; Ohara, K.; Arai, T. NDT scan matching method for high resolution grid map. In Proceedings of the 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems, St. Louis, MO, USA, 10–15 October 2009; pp. 1517–1522. [Google Scholar]
  6. Chen, S.; Ma, H.; Jiang, C. NDT-LOAM: A Real-Time LiDAR Odometry and Mapping with Weighted NDT and LFA. IEEE Sens. J. 2022, 22, 3660–3671. [Google Scholar] [CrossRef]
  7. Kang, D.; Wong, A.; Lee, B.; Kim, J. Real-Time Semantic Segmentation of 3D Point Cloud for Autonomous Driving. Electronics 2021, 10, 1960. [Google Scholar] [CrossRef]
  8. Kirillov, A.; He, K.; Girshick, R.; Rother, C.; Dollár, P. Panoptic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9404–9413. [Google Scholar]
  9. Baek, S.; Kim, M.; Suddamalla, U.; Wong, A.; Lee, B.; Kim, J. Real-Time Lane Detection Based on Deep Learning. J. Electr. Eng. Technol. 2022, 17, 655–664. [Google Scholar] [CrossRef]
  10. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  11. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 11–18 December 2015; pp. 1440–1448. [Google Scholar]
  12. Chen, L.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [PubMed]
  13. Xiong, Y.; Liao, R.; Zhao, H.; Hu, R.; Bai, M.; Yumer, E.; Urtasun, R. Upsnet: A Unified Panoptic Segmentation Network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8818–8826. [Google Scholar]
  14. Kang, D. Object Detection by Matching Data Representation of LiDAR and Camera. Ph.D. Thesis, Graduate School of Automotive Engineering, Kookmin University, Seoul, Republic of Korea, 2023. [Google Scholar]
  15. Milioto, A.; Vizzo, I.; Behley, J.; Stachniss, C. RangeNet++: Fast and Accurate LiDAR Semantic Segmentation. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; pp. 4213–4220. [Google Scholar]
  16. Pendleton, S.D.; Andersen, H.; Du, X.; Shen, X.; Meghjani, M.; Eng, Y.H.; Rus, D.; Ang, M.H. Perception, Planning, Control, and Coordination for Autonomous Vehicles. Machines 2017, 5, 6. [Google Scholar] [CrossRef]
  17. Laconte, J.; Kasmi, A.; Aufrère, R.; Vaidis, M.; Chapuis, R. A Survey of Localization Methods for Autonomous Vehicles in Highway Scenarios. Sensors 2022, 22, 247. [Google Scholar] [CrossRef] [PubMed]
  18. Yihuan, Z.; Liang, W.; Jun, W. Real-time localization method for autonomous vehicle using 3DLIDAR. Dyn. Veh. Roads Tracks 2017, 1, 271–276. [Google Scholar]
  19. Jens, B.; Cyrill, S. Efficient Surfel-Based SLAM using 3D Laser Range Data in Urban Environments. In Proceeding of Robotics: Science and System XIV, Pittsburgh, PA, USA, 26–30 June 2018; pp. 59–69. [Google Scholar]
  20. Andres, M.; Ignacio, V.; Jens, B.; Cyrill, S. SuMa++: Efficient LiDAR-based Semantic SLAM. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; pp. 4530–4537. [Google Scholar]
  21. Kenji, K.; Jun, M.; Emanuele, M. A Portable 3D LiDAR-based System for Long-term and Wide-area People Behavior Measurement. Int. J. Adv. Robot. Syst. 2019, 16, 1729881419841532. [Google Scholar]
  22. Kim, M.; Kwon, O.; Kim, J. Vehicle to Infrastructure-Based LiDAR Localization Method for Autonomous Vehicles. Electronics 2023, 12, 2684. [Google Scholar] [CrossRef]
  23. Pagad, S.; Agarwal, D.; Narayanan, S.; Rangan, K.; Kim, H.; Yalla, G. Robust method for removing dynamic objects from point clouds. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 10765–10771. [Google Scholar]
  24. Suleymanov, T.; Gadd, M.; Kunze, L.; Newman, P. LiDAR lateral localisation despite challenging occlusion from traffic. In Proceedings of the 2020 IEEE/ION Position, Location and Navigation Symposium (PLANS), Portland, OR, USA, 20–23 April 2020; pp. 334–341. [Google Scholar]
  25. Ding, P.; Wang, Z. 3D LiDAR point cloud loop detection based on dynamic object removal. In Proceedings of the 2021 IEEE International Conference on Real-time Computing and Robotics (RCAR), Xining, China, 15–19 July 2021; pp. 980–985. [Google Scholar]
  26. Arora, M.; Wiesmann, L.; Chen, X.; Stachniss, C. Mapping the static parts of dynamic scenes from 3D LiDAR point clouds exploiting ground segmentation. In Proceedings of the 2021 European Conference on Mobile Robots (ECMR), Bonn, Germany, 31 August–3 September 2021; pp. 1–6. [Google Scholar]
  27. Wissem, S.; Yacine, M.; Mohand, S.D. Multiple Sensors and JPDA-IMM-UKF Algorithm for Tracking Multiple Maneuvering Targets. Int. J. Electr. Comput. Eng. 2007, 1, 1494–1499. [Google Scholar]
  28. Niku, S.N. Introduction to Robotics Analysis, Systems, Applications; Pearson Education: San Luis Obispo, CA, USA, 2001; pp. 55–61. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.