Next Article in Journal
Early Findings on the Effectiveness of Novel Awakening Stimuli for Patients with Congenital Central Hypoventilation Syndrome
Previous Article in Journal
YOLO-WAD for Small-Defect Detection Boost in Photovoltaic Modules
Previous Article in Special Issue
Efficient Large-Scale Point Cloud Geometry Compression
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Human Tracking Using a 3D LiDAR and Point Cloud Projection for Human-Following Robots

1
Graduate School of Robotics and Design, Osaka Institute of Technology, Osaka 530-8568, Japan
2
Faculty of Robotics and Design, Osaka Institute of Technology, Osaka 530-8568, Japan
3
Faculty of Information Sciences and Technology, Osaka Institute of Technology, Hirakata 573-0196, Japan
4
Graduate School of Engineering, Tohoku University, Sendai 980-8579, Japan
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(6), 1754; https://doi.org/10.3390/s25061754
Submission received: 15 February 2025 / Revised: 6 March 2025 / Accepted: 11 March 2025 / Published: 12 March 2025

Abstract

:
Human tracking is a fundamental technology for mobile robots that work with humans. Various devices are used to observe humans, such as cameras, RGB-D sensors, millimeter-wave radars, and laser range finders (LRF). Typical LRF measurements observe only the surroundings on a particular horizontal plane. Human recognition using an LRF has a low computational load and is suitable for mobile robots. However, it is vulnerable to variations in human height, potentially leading to detection failures for individuals taller or shorter than the standard height. This work aims to develop a method that is robust to height differences among humans using a 3D LiDAR. We observed the environment using a 3D LiDAR and projected the point cloud onto a single horizontal plane to apply a human-tracking method for 2D LRFs. We investigated the optimal height range of the point clouds for projection and found that using 30% of the point clouds from the top of the measured person provided the most stable tracking. The results of the path-following experiments revealed that the proposed method reduced the proportion of outlier points compared to projecting all the points (from 3.63% to 1.75%). As a result, the proposed method was effective in achieving robust human following.

1. Introduction

Human tracking is a fundamental technology for mobile robots that work with humans. Various devices are used for human tracking, such as cameras with computer vision technology [1,2,3], RGB-D sensors [4,5,6], millimeter-wave radar [7,8,9], and laser range finders [10,11]. Observations from multiple sensors are often combined [12,13,14].
Laser range finder (LRF)-based human tracking has several advantages over methods that use other sensors. Measurements using LRFs are more robust to lighting conditions, more accurate than camera-based methods, and require less computation than camera-based measurements. A drawback of LRF-based measurement is that an ordinary LRF observes only a specific horizontal plane, which can cause misdetection of the human body because the measurement plane may miss the target position (such as the torso) when the height of the human varies.
A 3D LiDAR (Light Detection and Ranging) [15] is a laser-based sensor that measures distances to objects across multiple planes. Several studies have used a 3D LiDAR as a human detection and tracking sensor because it can capture the surface of surrounding objects as point clouds [16,17,18]. However, using the observed point cloud directly for human tracking requires high computational costs.
Here, we develop an energy-efficient human-tracking method for a small mobile robot that follows a person. We have already developed a simple, robust, and computationally low-cost human-tracking method based on 2D LRFs [19,20]. Therefore, this paper proposes a technique that projects the point cloud onto a 2D plane to apply the 2D-LRF-based human-tracking method and improve robustness to human height variations.
This paper is organized as follows. We investigate related works in Section 2 and explain the proposed method in Section 3. In Section 4, we describe two experiments: objective and subjective evaluations. In Section 5, we conclude this paper.

2. Related Works

2.1. Human Tracking Using 2D LRFs

A number of human-tracking methods using LRFs have been proposed so far. The method by Fod et al. was one of the earliest works [21]. Their work first detects objects in the observation plane as blobs and then merges the blobs of the same person as a post-processing step. Other methods aim to either detect and track humans with lightweight processing for low-computational-resource devices [22,23,24] or use machine learning/deep neural networks for accurate detection and tracking [25,26,27].
When tracking humans, an LRF can only observe one horizontal plane. Therefore, it is important to determine the appropriate measurement height of the LRF. Small robots often observe human legs [22,23]; however, the detection of human legs is often unstable, especially in environments where there are multiple people.
Tracking the torso is another common method in LRF-based human tracking [24,25]. There are two problems with torso detection: first, it is affected by arm movements, and second, its performance depends on the person’s height. In particular, detecting shorter people, such as children, is challenging, as shown in Figure 1.
Our previous work belongs to the latter, where the human torso is detected and tracked using an LRF [24]. This method is stable, robust, lightweight, and suitable for mobile robots. A similar method can detect and track multiple people moving quickly around the robot [19,20].
More recent studies have also applied LRFs for human tracking. Hasan et al. proposed a person identification technique named “PerFication” that utilizes LRF data to evaluate gait. The system uses ankle-level LRF data for individual tracking and identification, especially in environments where video monitoring is ineffective [28]. Aguirre et al. developed a method that combines observations from an LRF and a vision-based deep learning technique [27,29]. Kutyrev et al. applied an LRF to trajectory control for autonomous vehicles used in a horticultural plant field [30].
When observing humans using an LRF, the height of the LRF should be adjusted to match the height of the humans being tracked. For example, we previously demonstrated a robot that tracked humans at events such as our institute’s open house [31], where many children visited and played with robots. In those events, we needed to tune the height of the LRF for children because the robot’s tracking system was developed for adults.
Therefore, robots with human-tracking functions intended for general public use should be robust to various user traits, including height.

2.2. Human Tracking Using a 3D LiDAR

A 3D LiDAR observes the environment in multiple planes. Since a 3D LiDAR obtains more information than an LRF, it is expected to enable more robust and accurate human tracking.
Human detection methods using a 3D LiDAR [16,17,18,32] first obtain a point cloud of the environment, segment the point cloud into individual objects, and classify the objects to detect humans. For example, the method by Goméz et al. [18] first detects the moving part of the point cloud and then creates voxels from the cloud. Then, the voxels are segmented and classified as humans or other objects. Although Goméz et al.’s method requires less computation than similar methods, it still involves voxelization and voxel segmentation, which are costly. The problem with these methods is that human detection is computationally expensive, even when relatively inexpensive machine learning techniques are employed.

3. Method

3.1. Overview

The problem with the LRF method is that an LRF observes only one plane. A two-dimensional scan is enough for human tracking if the observation plane is appropriate for capturing the human body. Here, “appropriate” means that the plane includes the body part the tracking method aims to capture, such as the torso. However, it is impossible to find an appropriate plane without knowing the shape of the human body. As described in the previous section, the body part captured by the plane changes with different heights. If a human is taller than the expected height, the plane captures the arms, which makes tracking unstable. Conversely, if a human is shorter, the plane captures the head, which causes detection failure. Therefore, the proposed method aims to generate an appropriate 2D scan from the 3D point cloud. After obtaining the 2D scan, we apply the existing human-tracking method for 2D LRFs [24]. Converting a 3D point cloud to a 2D map is not a novel idea. Yoon et al. proposed a method to generate a 2D floor map from a 3D point cloud [33]. However, no previous studies have used 2D-converted point clouds for human tracking.
Figure 2 shows the use of a 3D LiDAR. As shown in Figure 2a, a 3D LiDAR casts beams into the environment at several planes (shown in different colors in the figure). As a result of the measurement, it obtains the distance r ( θ , ϕ ) for a specific azimuth θ Θ and elevation ϕ Φ , where Θ and Φ are sets of discrete angles. This observed point can be converted to a Cartesian coordinate ( x , y , z ) as follows:
x = r cos θ cos ϕ
y = r sin θ cos ϕ
z = h s + r sin ϕ
where h s is the height of the sensor.
When we project the observed points onto the X-Y plane, we obtain the points on the plane, as shown in Figure 2b. Here, we have multiple points for a specific azimuth θ . Then, we choose the nearest point in that azimuth.
r ^ ( θ ) = min ϕ V r ( θ , ϕ ) cos ϕ
Here, V Φ is the set of elevation angles for projection, which is a subset of all elevation angles.
The problem is how to determine the vertical range of the projection V. If we project all points (i.e., V = Φ ), human tracking may become unstable because of arm movements. Conversely, when observing only a single horizontal plane, even if it corresponds to the most stable body region such as the chest, changes in height due to leg bending can cause the system to detect the neck or head, resulting in unstable tracking. By slightly expanding the observation volume downward and projecting the resulting point cloud onto a plane, the system becomes more robust to vertical body movements. Therefore, we determine the observation range from the top of a person’s head using a fixed ratio of their height, and the point cloud within that range is projected in two dimensions. When the person’s height is h and the ratio is 0 < r o b s < 1 , then
V = { ϕ | ϕ Φ   and   h · r o b s h s + D sin ϕ h }
where D is the distance between the 3D LiDAR and the point on the object.
To determine V, we need to measure the person’s height. Since this method is designed for a human-following robot, we assume that the person is first registered, after which the robot starts following them. When registering the person, the robot measures the height of that person.

3.2. Measurement of Human Height

Figure 3 shows the measurement of a person’s height. It is assumed that the robot tracks a single pre-registered target. The target person stands 2.5 in front of the robot, and the robot measures the person’s height using a 3D LiDAR. Since the approximate distance between the robot and the person is known, the person’s height can be determined by identifying the highest elevation angle that detects an object around 2.5 m away. When registering the person, it is possible that the robot cannot maintain the required distance from the person. We do not address this case in this paper. Instead, we assume that the person is observed at a pre-determined height. Once the robot can measure the height of the person, the proposed method is applied to improve measurement accuracy.
Since the 3D LiDAR is installed at a height of 1.2 m ( h s = 1.2 m) and has an elevation angle range of ±15 degrees, the measurable height range is given by 1.2 ± 2.5 tan 15 π / 180 = [ 0.53 , 1.87 ] m.

3.3. Human Following

After obtaining the point cloud, the points are projected onto a plane using Formula (4). After that, the 2D point cloud data are treated in the same way as measurements by a 2D LRF.
We used the human-following method by Hiroi et al. [19,24]. We briefly describe the method.
Figure 4 shows the detection of the human body. When we obtain the observation r ^ ( θ i ) ,   i = 1 , , N , we calculate the distance difference, Δ D i = | r ^ ( θ i ) r ^ ( θ i + 1 ) | , and detect the range, [ j , k ] , where Δ D i < D t h for j i k , Δ D j 1 > D t h , and Δ D k + 1 > D t h . Here, θ j and θ k are the rightmost and leftmost angles of the object. D t h is the object detection threshold, and we use D t h = 0.15 m in a later experiment. Then, we calculate the width of the object as
W = r ^ ( θ j ) 2 + r ^ ( θ k ) 2 2 r ^ ( θ j ) r ^ ( θ k ) cos ( θ k θ j ) .
We regard the object as a candidate for a person when 100 W 800 mm, and the center point of a candidate is the midpoint of the line segment between the leftmost and rightmost points [24].
After detecting the coordinates of the candidates, the robot determines the target person to follow. Let ( x ( i , t ) , y ( i , t ) ) be the coordinates of the i-th candidate in the robot’s coordinate system at time t, and let p ( t ) be the index of the target person among the candidates at time t. The origin of the robot’s coordinate system is the position of the LRF. Then, we determine p ( t ) as follows:
Δ x ( i , t ) = x ( i , t ) x ( p ( t 1 ) , t 1 )
Δ y ( i , t ) = y ( i , t ) y ( p ( t 1 ) , t 1 )
inarea ( Δ x , Δ y ) = True Δ x 2 + Δ y 2 < R m i n False otherwise
p ( t ) = arg min i x ( i , t ) 2 + y ( i , t ) 2 s.t. inarea ( Δ x ( i , t ) , Δ y ( i , t ) )
These formulas indicate that we choose the nearest candidate to the LRF who is not too far from the target person’s location at the previous time step. We used R m i n = 0.6 m.
After determining the target person, we control the rotation of the robot’s wheels to keep a constant distance from the target [19,24].

4. Experiment

4.1. Height Measurement

We experimented with measuring human height using a 3D LiDAR. We used a Velodyne VLP-16-LITE (https://www.mapix.com/lidar-scanner-sensors/velodyne/velodyne-vlp16-lite/, accessed on 1 February 2025) as the 3D LiDAR. Table 1 shows the specifications of the VLP-16-LITE.
Figure 3 shows the participants standing 2.5 m in front of the 3D LiDAR. We invited eleven participants, whose heights ranged from 1.58 to 1.73 m. Figure 5 shows the measurement results. Since the elevation angles are discrete, the measured height values are also discrete. The average RMS error was 3.83 cm.

4.2. Optimal Projection Range

Next, we investigated the optimal projection range. As described in Formula (5), the range is determined by the ratio r o b s . Therefore, we carried out human-following experiments with different values of r o b s and observed the stability of human size. Human size and position estimation are stable if the projection range is appropriate.
Figure 6 shows the robot. The dimensions of the robot are 0.50 × 0.57 × 1.2 (WDH m), and its weight is 25 kg. The robot is equipped with four omni wheels to enable omnidirectional movement. The tread is 0.45 m, and the maximum translation speed is 1.4 m/s. The 3D LiDAR is installed at a height of 1.2 m. The 3D LiDAR observes objects at 10 Hz.
The specifications of the PC controlling the robot were as follows: an HP Pavilion Gaming Laptop 15-dk0000 (HP Japan Inc., Tokyo, Japan), with an Intel Core i7-9750H CPU 2.60 GHz CPU (Intel Corp., Santa Clara, CA, USA), and 16 GB of memory. The OS was Ubuntu 20.04.6 LTS with ROS Noetic.
In this experiment, three participants walked along the paths shown in Figure 7. We chose participants with various heights. The participants’ heights were 1.58, 1.70, and 1.83 m. We prepared two paths: a straight path and a circular path. We marked every 0.6 m on the path, and the participants stepped on the marks synchronously with a metronome sound. The metronome’s speed was 120 BPM; thus, a participant’s walking speed was 1.2 m/s. The robot moved behind the walking participants and measured them. This experiment used four projection ranges: r o b s { 0.2 , 0.3 , 0.4 , 0.5 } . Thus, we conducted 24 experiments (three participants × two paths × four ranges).
After the experiment, we investigated the results from two points of view: the stability of the human figure and the stability of human tracking. As described in the previous section, an object’s width is the key to detecting a human. Therefore, human tracking becomes difficult if human width estimation is unstable.
Figure 8 shows the human width estimation results for Participant 2. When r o b s = 0.2 , the estimated width was narrower than that with other values of r o b s . When a participant walked along a circular path, the following robot observed the participant from an angle, making the observed width narrower than when observed from the front.
Figure 9 shows the distribution of the estimated human widths. The figure indicates that the estimated human width with r o b s = 0.2 was narrower than under the other conditions for all participants, suggesting that the head width was measured when r o b s = 0.2 , while the body width was measured under the condition r o b s 0.3 . We conducted an ANOVA, considering the ratio, participants, and paths as factors. The results showed that all factors were statistically significant at the 1% level. Then, we further conducted a Tukey honest significant difference test to compare pairwise significance and found 5% significant differences between the ratios 0.4 and 0.5 and 1% significant differences between all other conditions.
Table 2 shows the mean and standard deviation of the human width measurements. The standard deviation values are related to the stability of human measurement, which is important for human tracking. The standard deviation decreased when r o b s was small. One possible reason for the increased standard deviation at large r o b s was the influence of arm movements. According to the average human body size [34], the average ratio of head-to-axilla length to height is 0.264 for males and 0.260 for females. Therefore, if r o b s exceeds this value, the arms are included in the observed body, causing the measured body width to fluctuate with arm movements.
We conducted a statistical test based on these results. First, we applied Bartlett’s test to examine whether the variances in these categories were different. The results confirmed statistically significant differences among these ratios ( p = 1.93 × 10 10 ). As a post hoc test, we conducted an F-test with Bonferroni correction, as shown in Table 3. The results showed that the conditions r o b s { 0.2 , 0.3 } and r o b s { 0.4 , 0.5 } formed distinct groups, and their variances were significantly different.
From these observations, we decided that the optimal value of r o b s was 0.3.

4.3. Human-Tracking Experiment with Various Paths

Next, we conducted an experiment comparing the optimal r o b s value ( r o b s = 0.3 ) with the condition in which the entire point cloud was used (i.e., r o b s = 1 ) across ten different paths. Figure 10 shows the paths used in the experiment. The experimental paradigm and participants were the same as those in the previous experiment.
We counted outlier points in the measured human positions as an evaluation metric. Here, we measured the distance between contiguous human positions. Since the walking speed of a participant was 1.2 m/s and the observation frequency was 10 Hz, the distance between the contiguous points should be 0.12 m. However, the distance fluctuated because of the measurement error. Figure 11 shows the distribution of the distances. In this figure, two histograms (pink for the “30%” condition ( r o b s = 0.3 ) and blue for the “all” ( r o b s = 1 ) condition) are superimposed. We used all data (three participants and ten paths) to calculate the histogram. We can see that the distribution for the “30%” condition is more concentrated at 0.12 m and has fewer outliers than the “all” condition. Here, we regard points with distances over 0.2 m as outliers. Figure 12 shows an example of the observed points with outlier points. The outlier points happened because of the effect of arm motion. When one or both arms were regarded as part of the body, the center position of the body fluctuated with the arm motion.
Figure 13 shows the distribution of the outlier point ratio among all observed points. This figure shows that the ratio of outliers using a 30% point cloud from the top of the head (1.65%) was smaller than that using the entire point cloud (3.63%). We tested the difference with a single-sided unpaired t-test. As a result, we found a statistically significant difference between the conditions ( p = 0.0055 ).
Figure 14 shows the outlier ratio of each path. The figure shows that the proposed condition (30%) resulted in fewer outlier points than using the entire point cloud, except for the L-shaped path. The outlier ratio varied across paths, with more complex paths, such as zig-zag or S-shaped ones, tending to produce more outliers.

5. Discussion

5.1. Robustness of Human Tracking

In the previous sections, we conducted two human-tracking experiments. In the first experiment, we examined two paths and determined the optimal r o b s value based on the stability of the observed human width. In the second experiment, we used ten paths to compare two conditions, with r o b s = 0.3 and r o b s = 1 . Although we used different measures for comparison, both measures are closely related to the stability of human tracking.
Figure 15 shows an example of point clouds of walking people observed by both methods. We can confirm that the arms were detected as part of the point cloud in the “all” method, and those clouds were misdetected as the body. In this situation, the body width suddenly became smaller, and the distance from the previous human position fluctuated. These misdetections did not appear in the “30%” condition.

5.2. Computational Efficiency

One advantage of the proposed method is its computational efficiency. Methods that use 3D point clouds directly to detect and track humans often employ neural networks, which makes them computationally expensive. For example, the method by Yin et al. tracks humans at 11 to 16 frame/s using a computer with an Intel Core i7 CPU (Intel Corp., Santa Clara, CA, USA) and a Titan RTX GPU (NVIDIA Corp., Santa Clara, CA, USA) [32], which means their method requires 60 to 90 ms to process one frame. On the other hand, we tested our method on an Intel Core i7-9750H CPU @ 2.60 GHz (Intel Corp., Santa Clara, CA, USA) without a GPU and processed the point cloud in 2.175 ms (1.49 ms for point cloud projection, 0.574 ms for human detection, and 0.111 ms for human tracking).

6. Conclusions

This paper introduced a new method for tracking people using a 3D LiDAR. Traditional methods using 2D scanners struggle with people of different heights. Our approach addressed this by projecting the point cloud obtained by a 3D LiDAR onto a 2D plane and then using a standard 2D LRF-based human-tracking method. A key innovation is that we used the top 30% of the 3D point cloud representing the person, making tracking much more reliable regardless of a person’s height.
Through experiments, we found that using the top 30% of the point cloud was the most effective for consistent tracking. Our experiments showed that this method significantly reduced errors and performed well in various scenarios. Specifically, the number of tracking errors was reduced by half compared to using all the 3D data.
The major limitation of the proposed method is the assumption that there is only one target person and that the person is registered before tracking begins. We also assume that registration is performed while the person stands at a certain distance from the robot. We need to develop methods that can relax these assumptions. One idea, as stated above, is for the robot to observe the person at a predetermined plane and then begin using the projected point cloud after observing the entire body of the target person.
We also plan to test this system in more complex environments and with groups of people. Furthermore, we aim to develop a more sophisticated tracking method by considering how people move and their posture. Finally, we will work on making the system faster and more energy-efficient, and we will further explore combining our approach with other sensor data to improve accuracy and robustness. This research offers a robust and efficient way to track people of varying heights, offering more advanced human tracking.

Author Contributions

Conceptualization, S.K.; validation, K.M.; software, A.I.; methodology, formal analysis, investigation, Y.H.; resources, S.K.; experiment, Y.H.; data curation, S.K. and Y.H.; writing—original draft preparation, A.I.; writing—review and editing, A.I.; visualization, A.I.; supervision, A.I.; project administration, Y.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This study was approved by the Ethics Committee of the Osaka Institute of Technology (approval number 2015-52-9).

Informed Consent Statement

Informed consent was obtained from all subjects involved in this study.

Data Availability Statement

The data are available upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Brunetti, A.; Buongiorno, D.; Trotta, G.F.; Bevilacqua, V. Computer vision and deep learning techniques for pedestrian detection and tracking: A survey. Neurocomputing 2018, 300, 17–33. [Google Scholar] [CrossRef]
  2. Dutta, A.; Mondal, A.; Dey, N.; Sen, S.; Moraru, L.; Hassanien, A.E. Vision tracking: A survey of the state-of-the-art. SN Comput. Sci. 2020, 1, 57. [Google Scholar] [CrossRef]
  3. Xin, S.; Zhang, Z.; Wang, M.; Hou, X.; Guo, Y.; Kang, X.; Liu, L.; Liu, Y. Multi-modal 3d human tracking for robots in complex environment with siamese point-video transformer. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 337–344. [Google Scholar]
  4. Munaro, M.; Menegatti, E. Fast RGB-D people tracking for service robots. Auton. Robot. 2014, 37, 227–242. [Google Scholar] [CrossRef]
  5. Rasoulidanesh, M.; Yadav, S.; Herath, S.; Vaghei, Y.; Payandeh, S. Deep attention models for human tracking using RGBD. Sensors 2019, 19, 750. [Google Scholar] [CrossRef]
  6. Liu, H.; Luo, J.; Wu, P.; Xie, S.; Li, H. People detection and tracking using RGB-D cameras for mobile robots. Int. J. Adv. Robot. Syst. 2016, 13, 1729881416657746. [Google Scholar] [CrossRef]
  7. Cui, H.; Dahnoun, N. High precision human detection and tracking using millimeter-wave radars. IEEE Aerosp. Electron. Syst. Mag. 2021, 36, 22–32. [Google Scholar] [CrossRef]
  8. Shen, Z.; Nunez-Yanez, J.; Dahnoun, N. Advanced Millimeter-Wave Radar System for Real-Time Multiple-Human Tracking and Fall Detection. Sensors 2024, 24, 3660. [Google Scholar] [CrossRef]
  9. Zhao, P.; Lu, C.X.; Wang, J.; Chen, C.; Wang, W.; Trigoni, N.; Markham, A. Human tracking and identification through a millimeter wave radar. Ad Hoc Netw. 2021, 116, 102475. [Google Scholar] [CrossRef]
  10. Ishihara, Y.; Uchitane, T.; Ito, N.; Iwata, K. Validation of Multiple Visitor Tracking with a Laser Rangefinder Using SMC Implementation of PHD Filter. In Proceedings of the 2022 Joint 12th International Conference on Soft Computing and Intelligent Systems and 23rd International Symposium on Advanced Intelligent Systems (SCIS&ISIS), Ise-Shima, Japan, 29 November–2 December 2022; pp. 1–6. [Google Scholar]
  11. Hu, W.; Fang, S.; Wang, Y.; Luo, D. A 2D-Laser-based Pedestrian Detector for Mobile Robots via Self-Supervised Transfer Learning. In Proceedings of the 2023 IEEE International Conference on Mechatronics and Automation (ICMA), Harbin, China, 6–9 August 2023; pp. 291–295. [Google Scholar]
  12. Chebotareva, E.; Safin, R.; Hsia, K.H.; Carballo, A.; Magid, E. Person-following algorithm based on laser range finder and monocular camera data fusion for a wheeled autonomous mobile robot. In Proceedings of the International Conference on Interactive Collaborative Robotics, St Petersburg, Russia, 7–9 October 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 21–33. [Google Scholar]
  13. Bozorgi, H.; Truong, X.T.; La, H.M.; Ngo, T.D. 2D laser and 3D camera data integration and filtering for human trajectory tracking. In Proceedings of the 2021 IEEE/SICE International Symposium on System Integration (SII), Iwaki, Japan, 11–14 January 2021; pp. 634–639. [Google Scholar]
  14. Funato, K.; Tasaki, R.; Sakurai, H.; Terashima, K. Development and experimental verification of a person tracking system of mobile robots using sensor fusion of inertial measurement unit and laser range finder for occlusion avoidance. J. Robot. Mechatron. 2021, 33, 33–43. [Google Scholar] [CrossRef]
  15. Raj, T.; Hanim Hashim, F.; Baseri Huddin, A.; Ibrahim, M.F.; Hussain, A. A survey on LiDAR scanning mechanisms. Electronics 2020, 9, 741. [Google Scholar] [CrossRef]
  16. Wang, H.; Wang, B.; Liu, B.; Meng, X.; Yang, G. Pedestrian recognition and tracking using 3D LiDAR for autonomous vehicle. Robot. Auton. Syst. 2017, 88, 71–78. [Google Scholar] [CrossRef]
  17. Yan, Z.; Duckett, T.; Bellotto, N. Online learning for 3D LiDAR-based human detection: Experimental analysis of point cloud clustering and classification methods. Auton. Robot. 2020, 44, 147–164. [Google Scholar] [CrossRef]
  18. Gómez, J.; Aycard, O.; Baber, J. Efficient detection and tracking of human using 3D LiDAR sensor. Sensors 2023, 23, 4720. [Google Scholar] [CrossRef] [PubMed]
  19. Nakamori, Y.; Hiroi, Y.; Ito, A. Multiple player detection and tracking method using a laser range finder for a robot that plays with human. ROBOMECH J. 2018, 5, 25. [Google Scholar] [CrossRef]
  20. Kasai, Y.; Hiroi, Y.; Miyawaki, K.; Ito, A. Development of a mobile robot that plays tag with touch-and-away behavior using a laser range finder. Appl. Sci. 2021, 11, 7522. [Google Scholar] [CrossRef]
  21. Fod, A.; Howard, A.; Mataric, M. A laser-based people tracker. In Proceedings of the 2002 IEEE International Conference on Robotics and Automation (Cat. No. 02CH37292), Washington, DC, USA, 11–15 May 2002; Volume 3, pp. 3024–3029. [Google Scholar]
  22. Horiuchi, T.; Thompson, S.; Kagami, S.; Ehara, Y. Pedestrian tracking from a mobile robot using a laser range finder. In Proceedings of the 2007 IEEE International Conference on Systems, Man and Cybernetics, Montreal, QC, Canada, 7–10 October 2007; pp. 931–936. [Google Scholar]
  23. Chung, W.; Kim, H.; Yoo, Y.; Moon, C.B.; Park, J. The detection and following of human legs through inductive approaches for a mobile robot with a single laser range finder. IEEE Trans. Ind. Electron. 2011, 59, 3156–3166. [Google Scholar] [CrossRef]
  24. Hiroi, Y.; Matsunaka, S.; Ito, A. A mobile robot system with semi-autonomous navigation using simple and robust person following behavior. J. Man Mach. Technol. 2012, 1, 44–62. [Google Scholar]
  25. Zainudin, Z.; Kodagoda, S.; Dissanayake, G. Torso detection and tracking using a 2D laser range finder. In Proceedings of the 2010 Australasian Conference on Robotics and Automation, ACRA 2010, Brisbane, Australia, 1–3 December 2010. [Google Scholar]
  26. Kohara, Y.; Nakazawa, M. Human Tracking of Single Laser Range Finder Using Features Extracted by Deep Learning. In Proceedings of the 2019 Twelfth International Conference on Mobile Computing and Ubiquitous Network (ICMU), Kathmandu, Nepal, 4–6 November 2019; pp. 1–5. [Google Scholar] [CrossRef]
  27. Abrego-González, J.; Aguirre, E.; García-Silvente, M. People detection on 2D laser range finder data using deep learning and machine learning. In Proceedings of the XXIV Workshop of Physical Agents, Alicante, Spain, 5–6 September 2024; pp. 235–249. [Google Scholar]
  28. Hasan, M.; Uddin, M.K.; Suzuki, R.; Kuno, Y.; Kobayashi, Y. PerFication: A Person Identifying Technique by Evaluating Gait with 2D LiDAR Data. Electronics 2024, 13, 3137. [Google Scholar] [CrossRef]
  29. Aguirre, E.; García-Silvente, M. Detecting and tracking using 2D laser range finders and deep learning. Neural Comput. Appl. 2023, 35, 415–428. [Google Scholar] [CrossRef]
  30. Kutyrev, A.I.; Kiktev, N.A.; Smirnov, I.G. Laser Rangefinder Methods: Autonomous-Vehicle Trajectory Control in Horticultural Plantings. Sensors 2024, 24, 982. [Google Scholar] [CrossRef]
  31. Hiroi, Y.; Ito, A. Realization of a robot system that plays “Darumasan-Ga-Koronda” Game with Humans. Robotics 2019, 8, 55. [Google Scholar] [CrossRef]
  32. Yin, T.; Zhou, X.; Krahenbuhl, P. Center-based 3D object detection and tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11784–11793. [Google Scholar]
  33. Yoon, S.; Choi, S.; An, J. Effective Denoising Algorithms for Converting Indoor Blueprints Using a 3D Laser Scanner. Electronics 2024, 13, 2275. [Google Scholar] [CrossRef]
  34. Kouchi, M.; Mochimaru, M. AIST Anthropometric Database. H16PRO 287, 2005. Available online: https://www.airc.aist.go.jp/dhrt/91-92/index.html (accessed on 1 February 2025).
Figure 1. Observation of humans using a 2D LRF. Since the LRF observes the environment at a specific height, it captures different parts of the body depending on a person’s height.
Figure 1. Observation of humans using a 2D LRF. Since the LRF observes the environment at a specific height, it captures different parts of the body depending on a person’s height.
Sensors 25 01754 g001
Figure 2. Observation by a 3D LiDAR. The 3D LiDAR scans beams in multiple planes, producing sets of 2D scans. (a) Observation by a 3D LiDAR. (b) Example of point clouds projected onto a horizontal plane. Different color indicates the measurement by beams with different elevation angles shown in (a).
Figure 2. Observation by a 3D LiDAR. The 3D LiDAR scans beams in multiple planes, producing sets of 2D scans. (a) Observation by a 3D LiDAR. (b) Example of point clouds projected onto a horizontal plane. Different color indicates the measurement by beams with different elevation angles shown in (a).
Sensors 25 01754 g002
Figure 3. Measurement of human height. The target person first stands 2.5 m in front of the robot. Then, the robot measures the person’s height using a 3D LiDAR.
Figure 3. Measurement of human height. The target person first stands 2.5 m in front of the robot. Then, the robot measures the person’s height using a 3D LiDAR.
Sensors 25 01754 g003
Figure 4. Overview of the human detection method. We segment the observation into objects using an LRF based on the measured distances and identify human bodies based on the widths of the objects.
Figure 4. Overview of the human detection method. We segment the observation into objects using an LRF based on the measured distances and identify human bodies based on the widths of the objects.
Sensors 25 01754 g004
Figure 5. Results of height measurement.
Figure 5. Results of height measurement.
Sensors 25 01754 g005
Figure 6. The robot used in the experiment.
Figure 6. The robot used in the experiment.
Sensors 25 01754 g006
Figure 7. The walking paths used in the experiment. We used a straight path (upper figure) and a circular path (lower figure).
Figure 7. The walking paths used in the experiment. We used a straight path (upper figure) and a circular path (lower figure).
Sensors 25 01754 g007
Figure 8. Estimation results of human width (Participant 2). (a) Straight path. (b) Circular path.
Figure 8. Estimation results of human width (Participant 2). (a) Straight path. (b) Circular path.
Sensors 25 01754 g008
Figure 9. Distribution of human widths. (a) Straight path. (b) Circular path.
Figure 9. Distribution of human widths. (a) Straight path. (b) Circular path.
Sensors 25 01754 g009
Figure 10. The ten paths examined in the experiment.
Figure 10. The ten paths examined in the experiment.
Sensors 25 01754 g010
Figure 11. The distribution of distances between contiguous observation points. Two histograms with different colors are superimposed.
Figure 11. The distribution of distances between contiguous observation points. Two histograms with different colors are superimposed.
Sensors 25 01754 g011
Figure 12. An example of observed points: the path was “zig-zag”, the participant was “Person 3”, and the entire point cloud was used ( r o b s = 1 ) .
Figure 12. An example of observed points: the path was “zig-zag”, the participant was “Person 3”, and the entire point cloud was used ( r o b s = 1 ) .
Sensors 25 01754 g012
Figure 13. Ratio of outlier points to all the observed points. The error bars show the standard error.
Figure 13. Ratio of outlier points to all the observed points. The error bars show the standard error.
Sensors 25 01754 g013
Figure 14. Ratio of outlier points for each path.
Figure 14. Ratio of outlier points for each path.
Sensors 25 01754 g014
Figure 15. Examples of point clouds of walking people. We can see that the arms are included in the point cloud of the human and misdetected as the body when the entire point cloud is projected. These misdetections can be avoided by using 30% of the point cloud.
Figure 15. Examples of point clouds of walking people. We can see that the arms are included in the point cloud of the human and misdetected as the body when the entire point cloud is projected. These misdetections can be avoided by using 30% of the point cloud.
Sensors 25 01754 g015
Table 1. Specifications of the VLP-16-LITE.
Table 1. Specifications of the VLP-16-LITE.
Sensors16 laser emitters and receivers
Field of ViewHorizontal 360 degrees, vertical ± 15 degrees
Range0.1 to 100 m
Sampling frequency5 to 20 Hz
Sampling speed≈300,000 point/s
Precision ± 3 cm ( 1 σ @ 25 m)
Angle resolutionHorizontal 0.1 to 0.4 degrees, vertical 2.0 degrees
Table 2. Mean and standard deviation of human width measurements.
Table 2. Mean and standard deviation of human width measurements.
RatioMean [m]Standard Dev. [m]
0.20.3920.087
0.30.5320.092
0.40.5640.110
0.50.5790.108
Table 3. p-values from the multiple comparison test of variances between different ratios, with Bonferroni correction.
Table 3. p-values from the multiple comparison test of variances between different ratios, with Bonferroni correction.
Ratio0.20.30.4
0.31.0--
0.4 1.69 × 10 7 2.08 × 10 5 -
0.5 2.27 × 10 6 1.87 × 10 4 1.0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kitamoto, S.; Hiroi, Y.; Miyawaki, K.; Ito, A. Robust Human Tracking Using a 3D LiDAR and Point Cloud Projection for Human-Following Robots. Sensors 2025, 25, 1754. https://doi.org/10.3390/s25061754

AMA Style

Kitamoto S, Hiroi Y, Miyawaki K, Ito A. Robust Human Tracking Using a 3D LiDAR and Point Cloud Projection for Human-Following Robots. Sensors. 2025; 25(6):1754. https://doi.org/10.3390/s25061754

Chicago/Turabian Style

Kitamoto, Sora, Yutaka Hiroi, Kenzaburo Miyawaki, and Akinori Ito. 2025. "Robust Human Tracking Using a 3D LiDAR and Point Cloud Projection for Human-Following Robots" Sensors 25, no. 6: 1754. https://doi.org/10.3390/s25061754

APA Style

Kitamoto, S., Hiroi, Y., Miyawaki, K., & Ito, A. (2025). Robust Human Tracking Using a 3D LiDAR and Point Cloud Projection for Human-Following Robots. Sensors, 25(6), 1754. https://doi.org/10.3390/s25061754

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop