You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

17 October 2025

High-Precision Extrinsic Calibration for Multi-LiDAR Systems with Narrow FoV via Synergistic Planar and Circular Features

,
,
and
1
School of Mechanical Engineering, Hebei University of Technology, Tianjin 300401, China
2
College of Electrical Engineering, Hebei University of Science and Technology, Shijiazhuang 050000, China
*
Author to whom correspondence should be addressed.

Abstract

Precise extrinsic calibration is a fundamental prerequisite for data fusion in multi-LiDAR systems. However, conventional methods are often encumbered by dependencies on initial estimates, auxiliary sensors, or manual feature selection, which renders them complex, time-consuming, and limited in adaptability across diverse environments. To address these limitations, this paper proposes a novel, high-precision extrinsic calibration method for multi-LiDAR systems with a narrow Field of View (FoV), achieved through the synergistic use of circular and planar features. Our approach commences with the automatic segmentation of the calibration target’s point cloud using an improved VoxelNet. Subsequently, a denoising step, combining RANSAC and a Gaussian Mean Intensity Filter (GMIF), is applied to ensure high-quality feature extraction. From the refined point cloud, planar and circular features are robustly extracted via Principal Component Analysis (PCA) and least-squares fitting, respectively. Finally, the extrinsic parameters are optimized by minimizing a nonlinear objective function formulated with joint constraints from both geometric features. Simulation results validate the high precision of our method, with rotational and translational errors contained within 0.08° and 0.8 cm. Furthermore, real-world experiments confirm its effectiveness and superiority, outperforming conventional point-cloud registration techniques.

1. Introduction

In recent years, rapid advancements in laser technology have facilitated the emergence of solid-state LiDAR. Due to its inherent advantages—such as high reliability, lightweight design, and low cost—it is gradually replacing conventional mechanical LiDAR as a key component in environmental perception systems for autonomous vehicles [1]. However, a single solid-state LiDAR is typically limited by a narrow field of view (FoV), which hinders the wide-area coverage required by mobile platforms such as Automated Guided Vehicles (AGVs) and Unmanned Aerial Vehicles (UAVs). Furthermore, in geometrically sparse environments—such as long corridors, tunnels, and open fields—single-LiDAR systems are prone to feature degradation [2,3], thereby undermining the robustness of perception algorithms. Consequently, deploying a multi-LiDAR perception system has become a mainstream solution for achieving a wider perceptual field and richer environmental features [4]. However, the efficacy of this approach critically depends on accurate and robust extrinsic calibration among individual sensors, which remains the main technical challenge.

1.1. Motivation

Accurate extrinsic calibration is a fundamental prerequisite for the effective fusion of data from multiple solid-state LiDAR sensors with narrow FoV [5]. Although multi-LiDAR calibration techniques have seen rapid progress in recent years, significant challenges remain in achieving high levels of automation, ensuring robustness under diverse environmental conditions, and improving operational efficiency. Addressing these challenges is the core objective of this study. Specifically, lack of automation poses a major obstacle, some existing methods [6] require manual selection or segmentation of geometric primitives (e.g., planes, corners, or edges) from raw point clouds. This manual process is not only labor-intensive and time-consuming but also limits the scalability and deployment speed of calibration systems in large-scale applications such as autonomous driving. Furthermore, dependence on priors and motion constraints presents another critical challenge, since other approaches [7,8] rely on auxiliary sensors such as IMUs, wheel odometry, or GNSS to provide initial pose estimates, or require the LiDAR platform to follow specific motion trajectories that sufficiently excite the sensor system. Consequently, these methods tightly couple calibration performance with external conditions and the quality of prior data. As a result, large initial pose errors, insufficient movement, or operation in static environments can easily cause such methods to converge to local optima or fail entirely, thereby compromising their robustness and applicability in real-world scenarios.
In light of these limitations, there is a pressing need for a high-precision, fully automated extrinsic calibration method for solid-state LiDAR systems with narrow FoV—one that does not rely on specific motion patterns or auxiliary sensors and remains robust under complex environmental conditions. This study addresses this gap by proposing a novel calibration framework tailored to such requirements.

1.2. Contribution

To address the aforementioned challenges, this study proposes an extrinsic calibration method for multiple solid-state LiDARs with narrow fields of view, leveraging a customized calibration board. The primary innovations and contributions of this work can be summarized as follows:
  • We propose an automatic method for calibration board detection and segmentation using an improved VoxelNet, which ensures efficient and robust extraction of the board’s point cloud even in complex environments.
  • We develop a planar point cloud filtering technique using the GMIF to effectively suppress noise, thereby significantly enhancing the quality of subsequent feature extraction.
  • We design a nonlinear optimization framework that jointly constrains planar and circular features. This framework incorporates an innovative adaptive weighting model to balance the contributions of different geometric primitives, leading to substantially improved calibration accuracy.

3. Methodology

The overall framework of the proposed calibration system is illustrated in Figure 1. The core pipeline comprises three key modules: automatic calibration target recognition, accurate feature extraction, and coupled nonlinear optimization.
Figure 1. Overview of the proposed multi-LiDAR calibration system. The system comprises three main components: (1) automatic identification and precise segmentation of calibration boards using an improved VoxelNet; (2) extraction of planar and circular geometric features; and (3) joint nonlinear optimization based on multi-feature constraints.
The process begins with the automatic recognition and instance segmentation of the calibration target, which is performed using an enhanced VoxelNet algorithm. To ensure the accuracy of the final calibration result, outliers and noise are subsequently removed using the RANSAC algorithm and a GMIF, respectively. The denoised high-quality point cloud is then used to accurately extract planar surfaces and circular center features. In the nonlinear optimization stage, we devise a two-step strategy. First, a reliable initial estimate of the calibration parameters is computed by exploiting correspondences between planar normal vectors. Then, a joint cost function is constructed to refine this initial estimate globally. This function simultaneously incorporates both planar and circular geometric features. This leads to a high-precision estimation of the extrinsic calibration parameters.

3.1. Notation

Before presenting the proposed method, we define the notations and conventions used in this paper. The system consists of two solid-state LiDAR sensors. Their coordinate frames are denoted as the primary LiDAR frame L M and the secondary LiDAR frame L S . The primary frame L M is used as the reference for calibration. The rigid-body transformation from the secondary frame L S to the primary frame L M is denoted as T S M SE ( 3 ) . It consists of a rotation matrix R S M SO ( 3 ) and a translation vector t S M R 3 :
T S M = R S M t S M 0 1 SE ( 3 )
where SE ( 3 ) denotes the special Euclidean group representing 3D rigid-body transformations, SO ( 3 ) denotes the special orthogonal group of 3D rotations, and R 3 represents the three-dimensional Euclidean space.
Since the 3D rotation group SO ( 3 ) is locally diffeomorphic to its tangent space at the identity, it allows for a minimal local parameterization. Therefore, we adopt a rotation vector φ R 3 as a minimal representation of rotation, which facilitates subsequent nonlinear optimization. The mapping between the rotation vector φ and the corresponding rotation matrix R is established via the exponential and logarithmic maps, defined as follows:
R = exp ( ( φ ) ) SO ( 3 ) φ = ( log ( R ) ) R 3
here, ( φ ) denotes the skew-symmetric matrix of the vector φ , and the exponential map ( φ ) can be efficiently computed using Rodrigues’ rotation formula.
For ease of subsequent derivations and explanations, we introduce the following notations:
Exp ( φ ) = R Log ( R ) = φ

3.2. Target Detection and Extraction

VoxelNet [28] is a prominent end-to-end deep learning architecture designed for 3D point cloud processing. The architecture consists of three primary components: a feature learning network, convolutional middle layers, and a region proposal network (RPN). In this work, we modify the convolutional middle layers to enhance the network’s ability to identify and extract geometric features. This improvement is particularly beneficial when dealing with calibration target point clouds under complex environmental conditions. The modified VoxelNet architecture is illustrated in Figure 2.
Figure 2. Architecture of the VoxelNet framework. The feature learning network first partitions raw point clouds into voxels, aggregating the points within each voxel into representative feature vectors, resulting in a sparse 4D tensor. This tensor is then passed through intermediate 3D convolutional layers to extract spatial contextual features. Finally, the Region Proposal Network (RPN) generates 3D object proposals for downstream detection tasks.
The first step in our processing pipeline involves voxelizing the raw point clouds acquired from the LiDARs. This process involves partitioning the 3D space into a grid of voxels and grouping the points accordingly. Since the point clouds from the two LiDARs often contain millions of points, a downsampling strategy is employed. This step improves both processing efficiency and target detection accuracy. If the number of points within a voxel exceeds a predefined threshold S, S points are randomly sampled from it. For a voxel containing s < S points, the point set is represented as follows:
V = P q x q , y q , z q , I q R 4 q = 1 s
each point consists of four attributes ( x , y , z , I ) , where x, y, and z denote spatial coordinates and I denotes intensity.
Next, we compute the offset of each point in the voxel relative to the point cloud centroid. The resulting offset-augmented point set is denoted as V i n , and is expressed as follows:
V i n = P q x q , y q , z q , I q , x q x c , y q y c , z q z c R 7 q = 1 s
The input P q is processed by a fully convolutional network (FCN) [29], which transforms primitive inputs (e.g., coordinates and intensity) into a learnable high-dimensional feature vector e q R m , which facilitates subsequent spatial aggregation. Max pooling is applied to e q to obtain the local aggregated feature e R m corresponding to V. Point-wise concatenation of the features of e q and e yields the output multi-dimensional feature vector f i o u t = [ e i , e ] R 2 m . The same encoding is applied to all non-empty features to obtain the output feature set:
V o u t = f q o u t q = 1 s
Multiple Voxel Feature Encoding (VFE) layers are used to transform the input feature dimension c i n into the output feature dimension c o u t . The weight matrix of the linear layer has a shape of c i n × ( c o u t / 2 ) , and the concatenation of intermediate features produces an output with feature dimension c o u t . The m-dimensional convolution operator is represented as C o n v M D ( c i n , c o u t , k , s , p ) . In this formulation, k denotes the kernel size, s denotes the stride, and p denotes the m-dimensional vector specifying per-dimension kernel sizes.
Each intermediate block consists of a 3D convolution, followed by a Rectified Linear Unit (ReLU) and a Batch Normalization (BN) layer. In this work, we adjust the convolutional parameters to better fit the geometric structure of the calibration board in complex scenes. The output feature maps from these layers are passed to a Region Proposal Network (RPN) [30]. The outputs from each module are then upsampled to a common spatial resolution and concatenated into a single high-resolution feature map. This fused representation is mapped to a classification score map and a regression map. The region with the highest confidence score is finally selected as the predicted location of the calibration board.

3.3. Plane and Circle Feature Extraction

To obtain a rich and diverse set of co-visible planar and circular features across the multi-LiDAR system, a dynamic data acquisition strategy is used. During calibration, the target is placed at n different poses, each visible to all sensors at the same time. This setup provides enough geometric constraints for the following optimization process.

3.3.1. Plane Feature Extraction

Due to the low single-frame point cloud density of non-repetitive scanning solid-state LiDARs, we synchronously collect data from LiDAR sensors L M and L S for τ seconds at each calibration pose to obtain a complete and reliable point cloud of the calibration board.
Let the aggregated point cloud from L M be represented as Q L M = { C 0 L M , C 1 L M , , C n l L M } , and the corresponding dataset from L S as Q L S = { C 0 L S , C 1 L S , , C n l L S } . This temporal aggregation improves point cloud density. However, it also introduces cumulative noise. This effect is shown in Figure 3a,b. These subfigures show single-frame and multi-frame aggregated point clouds from frontal and side perspectives, respectively. The accumulated points do not lie on a perfect plane; instead, the side view clearly reveals that they form a point cloud with a noticeable thickness and significant out-of-plane deviations.
Figure 3. Point cloud processing: (a) single-frame point cloud; (b) multi-frame accumulated point cloud; (c) filtered point cloud after refined segmentation; (d) extracted planar and circular features.
To accurately estimate the planar parameters of the calibration board from the point clouds Q L M and Q L S , we apply a three-step procedure. This procedure is designed to mitigate the effects of accumulated noise. First, we use the RANSAC algorithm [31] to perform an initial filtering of the point cloud. However, RANSAC may fit a non-target plane due to the board’s imperfect planarity. Such an error would degrade the performance of the subsequent optimization. To address this, we use RANSAC to identify inliers rather than to determine the final plane model. Specifically, we retain only the inlier subset associated with the board’s dominant plane and discard non-target points, such as outliers or p. Second, we perform precise clustering and segmentation on the filtered point cloud using prior knowledge of the board’s geometry. The segmentation result is illustrated in Figure 3c. Finally, the GMIF is applied to further refine the segmented plane points. GMIF computes a geometric mean intensity I M and a threshold σ based on the reflectivity of the points. These metrics are then used to filter out points exhibiting abnormal reflectivity, which often correspond to sensor artifacts or edge points. This process effectively reduces noise and enhances planarity. The resulting point cloud exhibits high fidelity and is well-suited for subsequent accurate plane fitting. The GMIF algorithm is presented in Algorithm 1.  
Algorithm 1: Mean Intensity filtering based on Gaussian Newton
Input: segmented point cloud: C = { C 1 , C 2 , , C N } , p j C k , I j p j
Output: C GMIF = { C 1 , C 2 , , C N }
1:
for  k = 1  to  N do
2:
M k getNumber ( C k )
3:
s u m 0.0
4:
for  j = 1  to  M k  do
5:
   s u m s u m + I j k
6:
end for
7:
I mean k s u m / M k
8:
σ 0.0
9:
for  j = 1  to  M k  do
10:
   σ σ + I j k I mean k 2
11:
end for
12:
σ k σ / M k
13:
lower I mean k σ k
14:
upper I mean k + σ k
15:
for  j = 1  to  M k  do
16:
  if  I j k  lower and  I j k upper  then
17:
    C k p j
18:
  end if
19:
end for
20:
C GMIF C k
21:
end for
22:
return C GMIF
We further apply Principal Component Analysis (PCA) [32] to the planar point cloud set obtained after GMIF refinement in order to estimate the best-fitting plane. The centroid of the point cloud is denoted by m i R 3 , and the unit normal vector of the fitted plane is denoted by n i R 3 . Let X = ( x , y , z ) T be an arbitrary inlier point on the plane. The plane equation is then given by
n i T ( X m i ) = 0

3.3.2. Circular Feature Extraction

To accurately extract the 3D circular features of the calibration board, the filtered point cloud is orthogonally projected onto a designated reference plane. Throughout this process, the original 3D coordinates of all points are retained, allowing their projected counterparts to be accurately restored in subsequent steps.
As shown in Figure 4, the reference plane is defined to be perpendicular to the LiDAR’s line of sight, which is aligned with the positive X-axis. To extract the 2D boundary of the projected point set, we apply the Alpha Shapes algorithm [33], which effectively handles both complex and concave geometries. This ensures robust and accurate contour delineation. After boundary extraction, the original 3D coordinates are utilized to recover the X-values of the projected boundary points. This operation lifts the 2D boundary into 3D space, reconstructing a contour point cloud denoted as C b d p . A minimal subset of points is randomly sampled from C b d p to construct a candidate 3D circle, characterized by the following equations:
n · ( p p 0 ) = 0 p c 2 2 = R 2
where n is the normal vector of the plane containing the 3D circle, p is any arbitrary point on the 3D circle, p 0 is an arbitrary point from the minimal subset, c is the coordinate of the circle’s center, and R denotes the radius.
Figure 4. Point cloud geometric projection. The point cloud is orthogonally projected along the positive X-axis onto a plane located at the maximum X-coordinate of all points.
The distances between all points in C b d p and the candidate 3D circles are computed. Points whose distances fall below a predefined threshold are marked. This process is repeated until the maximum number of iterations is reached. The 3D point cloud with the highest number of inlier markings is then selected. The least squares method is then used to refit the 3D point cloud, yielding the optimized 3D circle contour and its corresponding center. The fitted contour points and the circle’s center are shown in Figure 3d.

3.4. Nonlinear Optimization

To achieve high-precision calibration, the entire nonlinear optimization process is divided into two stages.

3.4.1. Parameter Initialization

The initial estimation of extrinsic parameters is critical to ensure the convergence and final accuracy of subsequent nonlinear optimization in multi-LiDAR calibration. This section focuses exclusively on the initial estimation of the rotational component of the extrinsic parameters. This is primarily because rotational parameters lie on the Special Orthogonal group SO ( 3 ) , which forms a nonlinear manifold. This structural property makes the optimization of rotation highly sensitive to its initial value. In contrast, the translation parameters lie in the Euclidean space R 3 , where the optimization is less dependent on the initial guess and generally exhibits better convergence properties. Therefore, this section is dedicated to estimating the initial rotation to establish a solid foundation for the subsequent fine-grained optimization of the full six-degree-of-freedom (6-DoF) extrinsic parameters.
We construct the constraints by matching the normal vectors of the same calibration board plane captured from different viewpoints by the LiDARs. Let n i L M and n i L S denote the normal vectors of the calibration board plane extracted at the i-th pose from LiDARs L M and L S , respectively. Accordingly, the residual is defined as the angle between the corresponding plane normal vectors from the two LiDARs. The cost function is formulated as follows:
L ( R L S L M ) = i = 1 N e i R 2 = i = 1 N ( R L S L M n i L S ) × n i L M 2
During the optimization process, the rotational parameters are iteratively updated to minimize the angular residual between the corresponding plane normal vectors. As a result, the transformed normal vector gradually aligns with the target normal vector. The entire initialization process can be represented as follows:
φ ^ L S L M = arg min R L S L M SO ( 3 ) L R L S L M

3.4.2. Refined Calibration

Building on the initial rotation parameters obtained in Section 3.4.1, this section aims to refine the full extrinsic parameters between multiple LiDARs. To achieve this, a multi-constrained coupled optimization model is constructed by jointly calibrating both planar and circular features of the calibration board. This approach enhances the accuracy and robustness of the resulting calibration. The feature constraint relationships are shown in Figure 5.
Figure 5. Joint feature constraints in the calibration process. Planar and circular geometric constraints are jointly formulated as part of the optimization objective function. These constraints guide the iterative minimization of residuals for accurate estimation of inter-LiDAR extrinsic parameters.
(1)
Centroid-to-Plane Constraint
Unlike Section 3.4.1, which formulates constraints based solely on minimizing the angular discrepancy between corresponding plane normals, this section introduces the Euclidean distance from the calibration board’s centroid to the matched plane as a primary constraint term. This formulation ensures compatibility with the subsequent circle constraint, which is likewise defined in terms of Euclidean distance.
The residual term for the Euclidean distance between the centroid m i L S and the plane Π i L M is formulated as follows:
m ^ i L S = R L S L M m i L S + t L S L M e i p ( R L S L M , t L S L M ) = n i L M T m ^ i L S m i L M 2
(2)
Center-to-Center Constraint
The constraint on circular features is defined as the Euclidean distance residual between corresponding circle centers. To mitigate the influence of potential circular fitting errors and to balance the relative contributions of the planar and circular constraints within the objective function, we introduce a weighting factor for the circular constraint:
ω j = 1 if | ξ j | δ δ | ξ j | if | ξ j | > δ
where ξ j = r j true r j est , r j true and r j est represent the true and estimated radii of the circular feature, respectively, and δ is a predefined threshold.
The residual term for the distance between c j L M and c j L S is defined as follows:
e j c ( R L S L M , t L S L M ) = ω j c ^ j L S c j L M 2
herein, c ^ j L S = R L S L M c j L S + t L S L M .
(3)
Iterative Optimization
During the refined optimization stage, the cost function is defined as follows:
L ( R L S L M , t L S L M ) = i = 1 m e i p 2 + j = 1 n e j c 2
The refined calibration is formulated as the following optimization problem:
( φ L S L M , t L S L M ) = arg min R L S L M SO ( 3 ) , t L S L M R 3 L ( R L S L M , t L S L M )

4. Experiments

To comprehensively evaluate the accuracy and robustness of the proposed algorithm, we conducted extensive experiments using both simulated and real-world datasets.

4.1. Implementation Details

To evaluate the proposed algorithm, simulations and real-world experiments were conducted using two Livox Mid-40 solid-state LiDAR sensors. These sensors were manufactured by DJI Technology Co., Ltd., located in Shenzhen, China. Each sensor has a circular field of view (FoV) of 38.4°, an angular accuracy of 0.05°, and a sampling frequency of 10 Hz. A customized calibration board with geometric features was used as the target. The board consisted of four square regions, each with a side length of 0.6 m. Circular holes with a diameter of 0.3 m were cut into the top-left and bottom-right regions. These two regions were primarily used during calibration, as they contained both the planar and circular features required by our method. The training platform was a desktop computer running Ubuntu 20.04, equipped with an NVIDIA Quadro A6000 GPU (32 GB VRAM). The deployment platform was an NVIDIA Jetson Orin Nano, also running Ubuntu 20.04. A custom-developed mobile robot served as the experimental platform. All programs were implemented in C++. The Point Cloud Library (PCL) was used for preprocessing, while Eigen and Sophus were employed for matrix operations. Nonlinear optimization was performed using the Ceres Solver.

4.2. Simulated Experiment

Gazebo was used to generate simulated data for our experiments. As an advanced open-source simulator, Gazebo is capable of simulating various types of sensor data, including that from the Livox Mid-40 solid-state LiDAR used in this study. To evaluate the accuracy and robustness of the proposed algorithm, a complex simulation environment featuring people, vehicles, a calibration board, and buildings was set up, as illustrated in Figure 6. Gaussian noise was added to the simulated point cloud data to better mimic LiDAR behavior in real-world scenarios. Specifically, zero-mean Gaussian noise with a standard deviation of 0.01 m was applied to each point to simulate the LiDAR’s scanning behavior under realistic conditions.
Figure 6. Validation in the simulated environment.
Given that the Livox Mid-40 has a limited field of view (FoV) of only 38.4°, the roll and pitch angles between the two LiDAR sensors were restricted to ±10°, while the yaw angle was constrained to ±15. This constraint ensures sufficient overlap between the point clouds of the two sensors, enabling reliable acquisition of calibration data. The distance between the calibration board and the mobile robot varied throughout the process. Experimental observations indicated that maintaining a distance between 5 and 10 m led to more uniform and sufficient point cloud coverage on the calibration board surface.
Properly selecting the number of calibration board point cloud samples is crucial for accurate estimation of the calibration parameters. An excessively large sample size significantly increases data collection time. In contrast, too few samples may reduce calibration accuracy. To explore this trade-off, we quantitatively evaluate the relationship between sample size and calibration accuracy using simulation experiments. The specific procedure is as follows:
  • Define the ground-truth calibration parameters.
  • Collect a dataset of 100 calibration board point cloud instances.
  • Generate subsets of varying sizes via random sampling without replacement.
The relative pose between the two LiDARs is known in the simulation environment. This allows calibration accuracy to be evaluated in a rigorous and quantitative manner. To this end, we define the ground-truth transformation as T true = { R true , t true } . The corresponding estimated transformation is denoted as T est = { R est , t est } . Based on this, we define the following error metrics:
ε R = Log R true R est 2 ε t = t true t est 2
here, ε R and ε t denote the rotational and translational errors, respectively. Figure 7 quantitatively illustrates the variation in calibration error with respect to the number of calibration board samples across five independent experiments. Figure 7a,b illustrate the trends of rotation and translation calibration errors as a function of the number of randomly selected calibration patterns, respectively. In both plots, the horizontal axis represents the number of samples, and the vertical axis indicates the error between the estimated values and the ground truth. The experimental results demonstrate that the system achieves convergence when the sample size, n, is greater than 5 (n > 5).
Figure 7. Impact of the number of calibration boards on the accuracy of extrinsic calibration: (a) variation in rotational calibration error with respect to the number of calibration boards; (b) variation in translational calibration error with respect to the number of calibration boards.
To evaluate the robustness and generalization capability of our calibration algorithm, we conducted simulation experiments using five distinct ground-truth configurations. These configurations were further used to investigate the influence of varying ground-truth parameters on algorithm performance. Each parameter set comprises five calibration board point cloud samples. For each set, we conducted 10 repeated trials under identical initial conditions to reduce the influence of random factors. This enhances the stability and reproducibility of the experimental results. The five sets of ground-truth parameters, along with their corresponding Euler angle representations, are summarized in Table 2. Table 3 presents the number of independent experiments conducted under different pose conditions. It also lists the rotation and translation errors of the estimated calibration results with respect to the ground truth. For a more intuitive visualization of these simulation results, the error distributions for rotation and translation are plotted in Figure 8. Specifically, Figure 8a,b illustrate the distribution of rotation and translation errors across different ground truth poses, respectively. In these plots, the horizontal axis represents the various ground truth poses, while the vertical axis indicates the error between the estimated results and their corresponding ground truth.
Table 2. Extrinsic parameter configurations.
Table 3. Performance evaluation under varying ground truth configurations.
Figure 8. Distribution of errors between the estimated calibration results and the ground truth: (a) distribution of rotation errors; (b) distribution of translation errors.
Figure 9 illustrates the registration results demonstrating the alignment between object detection outputs and LiDAR point clouds in a simulated environment. Specifically, Figure 9a,b show the calibration board detected by the primary LiDAR and the target LiDAR, respectively. Figure 9c presents the final point cloud registration result obtained from both LiDARs.
Figure 9. LiDAR point cloud registration using target detection in a simulated environment: (a) detection of the calibration pattern in the primary LiDAR’s point cloud; (b) detection of the calibration pattern in the target LiDAR’s point cloud; (c) registration result of the point clouds from the primary and target LiDARs.
The simulation results show that across all five experimental sets, the median rotational error remained below 0.08°, and the median translational error stayed below 0.8 cm. This indicates that the overall calibration error was effectively controlled. These results confirm that the proposed algorithm achieves consistently high calibration accuracy across varying parameter settings.

4.3. Real-World Experiments

Figure 10 illustrates the experimental setup and real-world environment. As shown in Figure 10a, the calibration board used in the real-world experiments has the same dimensions and shape as the one used in simulation. It is suspended in the air by a metal support frame. Figure 10b shows two Livox Mid-40 LiDAR sensors mounted on the chassis of a mobile robot. The sensors are secured using custom 3D-printed connectors. Figure 10c presents the actual outdoor test environment. During data acquisition, the procedure was similar to that in the simulation. The calibration board was placed sequentially at distances ranging from 5 to 10 m from the mobile robot. For each position, both LiDAR sensors simultaneously collected data for 10 s. Figure 11 shows the results of target detection and point cloud registration in the real-world setting. Figure 11a,b display the calibration board detected by the primary and target LiDARs, respectively. Figure 11c shows the final registration result of the point clouds from both sensors.
Figure 10. Setup and results of the real-world experiments: (a) calibration board; (b) installed LiDAR sensors; (c) configured multi-LiDAR experimental environment.
Figure 11. Performance of object detection and point cloud registration in realistic: (a) detection of the calibration pattern in the primary LiDAR’s point cloud; (b) detection of the calibration pattern in the target LiDAR’s point cloud; (c) registration result of the point clouds from the primary and target LiDARs.
To evaluate the performance of the proposed calibration method, we benchmarked it against four recent algorithms: NDT [34], GICP [35], FGR [36], and TEASER++ [37]. To ensure a fair and rigorous comparison, the raw point cloud data for all methods underwent preprocessing using a voxel grid filter with a leaf size of 0.05 m. Furthermore, each method was subjected to 10 independent trials to ensure statistical robustness. Table 4 provides a quantitative comparison of the evaluated methods. It includes the average execution time, as well as the median and standard deviation of the estimated extrinsic parameters across all six degrees of freedom (6-DoF). For a more intuitive visualization of the results, Figure 12 shows the distribution of the rotational (roll, pitch, yaw) and translational (x, y, z) components for each method.
Table 4. Comparison of different registration methods.
Figure 12. Field evaluation using Livox Mid-40 LiDARs. The figure compares the proposed method with NDT, GICP, FGR, and TEASER++. Each box summarizes the results over 10 independent runs.
According to Table 4 and Figure 12, NDT is the fastest method. However, its large standard deviation makes it unsuitable for high-precision tasks. GICP is also computationally efficient and more accurate than NDT. Nevertheless, compared to the state-of-the-art methods FGR and TEASER++, GICP shows less consistency, indicated by a wider box plot. In contrast, both FGR and TEASER++ deliver highly consistent results. Their small standard deviations and narrow box plots reflect excellent precision. However, both methods suffer from occasional failures, which are visible as prominent outliers in their error distributions. Notably, our proposed method achieves a computational time comparable to FGR’s. More importantly, it outperforms all four benchmark algorithms by achieving the lowest median error and superior consistency. Furthermore, our method is highly robust, showing no significant outliers across any of the experimental trials.
Overall, the box plot results demonstrate that the proposed method outperforms the baseline methods in both accuracy and robustness. It provides more reliable LiDAR-to-LiDAR extrinsic calibration.

5. Conclusions and Future Work

This paper proposes an automatic and high-precision calibration method for multiple narrow FoV LiDARs. The method automatically extracts and segments the calibration board point cloud using an improved VoxelNet-based neural network. RANSAC and GMIF filters are applied to remove noise, thereby improving plane extraction accuracy. During optimization, a weighting strategy is introduced to reduce the influence of noise in circular features. Multiple simulation and real-world experiments were conducted to validate the effectiveness of the proposed calibration method. In simulation experiments, the estimated results were compared with ground-truth data, demonstrating a rotation error below 0 . 08 and a translation error under 0.8 cm. In real-world experiments, the proposed method was compared with NDT, GICP, FGR, and TEASER++ and consistently outperformed these baselines in both accuracy and robustness. These findings demonstrate that the proposed method achieves high calibration accuracy and strong robustness in both simulated and real-world scenarios.
Despite the promising performance of our method in the current experiments, we also acknowledge its limitations. The real-world experiments in this paper were primarily conducted in relatively static and structured environments. This setup provided a clear baseline for validating the core performance of our proposed framework. However, real-world applications such as autonomous driving and robotic navigation often involve dynamic and highly cluttered scenes. Interference and severe occlusion from dynamic objects, such as pedestrians and moving vehicles, undoubtedly pose new challenges to the stable extraction and matching of features. To address these challenges and further enhance the practical utility of our method, we plan to pursue the following research in our future work. First, we will conduct supplementary experiments in dynamic and cluttered scenarios to comprehensively evaluate and optimize the generalization ability and robustness of our method. Second, we will incorporate advanced 2D or 3D segmentation networks to achieve precise extraction of target features from complex backgrounds. Finally, another important direction for future research is to extend our calibration framework to the joint calibration of heterogeneous sensors such as LiDAR and cameras.

Author Contributions

Conceptualization, J.L.; Methodology, X.S.; Software, X.S. and Z.Z.; Validation, X.S. and Z.Z.; Formal analysis, X.S.; Investigation, X.S.; Resources, J.L.; Data curation, X.S., Z.Z. and S.X.; Writing—original draft, X.S.; Writing—review and editing, all authors; Visualization, X.S. and Z.Z.; Supervision, J.L. and S.X.; Project administration, J.L.; Funding acquisition, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Key R&D Program of China (Grant No. 2024YFC3811103), the National Natural Science Foundation of China (Grant No. U24A6005), Hebei Province Major Scientific and Technological Achievements Transformation Project (Grant No. 23281902Z).

Data Availability Statement

The data are not publicly available.

Conflicts of Interest

The authors declare no conflicts of interes.

References

  1. Li, Y.; Ibanez-Guzman, J. Lidar for autonomous driving: The principles, challenges, and trends for automotive lidar and perception systems. IEEE Signal Process. Mag. 2020, 37, 50–61. [Google Scholar] [CrossRef]
  2. Zhang, J.; Wang, S.; Tan, X. ODLC_SAM: A novel LiDAR SLAM system towards open-air environments with loop closure. Ind. Robot 2023, 50, 1011–1023. [Google Scholar] [CrossRef]
  3. Xu, W.; Zhang, F. Fast-LIO: A fast, robust lidar-inertial odometry package by tightly-coupled iterated Kalman filter. IEEE Robot. Autom. Lett. 2021, 6, 3317–3324. [Google Scholar] [CrossRef]
  4. Xu, J.; Huang, S.; Qiu, S.; Zhao, L.; Yu, W.; Fang, M.; Li, R. Lidar-Link: Observability-aware probabilistic plane-based extrinsic calibration for non-overlapping solid-state lidars. IEEE Robot. Autom. Lett. 2024, 9, 2590–2597. [Google Scholar] [CrossRef]
  5. Pusztai, Z.; Eichhardt, I.; Hajder, L. Accurate calibration of multi-lidar-multi-camera systems. Sensors 2018, 18, 2139. [Google Scholar]
  6. Zhang, J.; Lyu, Q.; Peng, G.; Wu, Z.; Yan, Q.; Wang, D. LB-L2L-Calib: Accurate and robust extrinsic calibration for multiple 3D LiDARs with long baseline and large viewpoint difference. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 926–932. [Google Scholar]
  7. Liu, W.; Li, Z.; Malekian, R.; Sotelo, M.A.; Ma, Z.; Li, W. A novel multifeature based on-site calibration method for LiDAR-IMU system. IEEE Trans. Ind. Electron. 2019, 67, 9851–9861. [Google Scholar]
  8. Pentek, Q.; Kennel, P.; Allouis, T.; Fiorio, C.; Strauss, O. A flexible targetless LiDAR–GNSS/INS–camera calibration method for UAV platforms. ISPRS J. Photogramm. Remote Sens. 2020, 166, 294–307. [Google Scholar] [CrossRef]
  9. Tsai, R.Y.; Lenz, R.K. A new technique for fully autonomous and efficient 3D robotics hand/eye calibration. IEEE Trans. Robot. Autom. 1989, 5, 345–358. [Google Scholar] [CrossRef]
  10. Huang, K.; Stachniss, C. Extrinsic multi-sensor calibration for mobile robots using the Gauss-Helmert model. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 1490–1496. [Google Scholar]
  11. Schneider, S.; Luettel, T.; Wuensche, H.J. Odometry based online extrinsic sensor calibration. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Tokyo, Japan, 3–7 November 2013; pp. 1287–1292. [Google Scholar]
  12. Wan, E.A.; Van Der Merwe, R. The unscented Kalman filter. The unscented Kalman filter. In Kalman Filtering and Neural Networks; Haykin, S., Ed.; John Wiley & Sons: New York, NY, USA, 2001; pp. 221–280. [Google Scholar]
  13. Taylor, Z.; Nieto, J. Motion-based calibration of multimodal sensor extrinsics and timing offset estimation. IEEE Trans. Robot. 2016, 32, 1215–1229. [Google Scholar] [CrossRef]
  14. Das, S.; Mahabadi, N.; Djikic, A.; Nassir, C.; Chatterjee, S.; Fallon, M. Extrinsic calibration and verification of multiple non-overlapping field of view lidar sensors. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 919–925. [Google Scholar]
  15. Das, S.; af Klinteberg, L.; Fallon, M.; Chatterjee, S. Observability-aware online multi-lidar extrinsic calibration. IEEE Robot. Autom. Lett. 2023, 8, 2860–2867. [Google Scholar]
  16. Kim, J.; Kim, C.; Han, Y.; Kim, H.J. Automated extrinsic calibration for 3D LiDARs with range offset correction using an arbitrary planar board. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 5082–5088. [Google Scholar]
  17. Lai, Z.; Jia, Z.; Guo, S.; Li, J.; Han, S. Extrinsic calibration for multi-LiDAR systems involving heterogeneous laser scanning models. Opt. Express 2023, 31, 44754–44771. [Google Scholar]
  18. Lee, H.; Chung, W. Extrinsic calibration of multiple 3D LiDAR sensors by the use of planar objects. Sensors 2022, 22, 7234. [Google Scholar] [CrossRef]
  19. Nie, M.; Shi, W.; Fan, W.; Xiang, H. Automatic extrinsic calibration of dual LiDARs with adaptive surface normal estimation. IEEE Trans. Instrum. Meas. 2022, 72, 1000711. [Google Scholar] [CrossRef]
  20. Shi, B.; Yu, P.; Yang, M.; Wang, C.; Bai, Y.; Yang, F. Extrinsic calibration of dual LiDARs based on plane features and uncertainty analysis. IEEE Sens. J. 2021, 21, 11117–11130. [Google Scholar]
  21. Liu, X.; Zhang, F. Extrinsic calibration of multiple lidars of small FOV in targetless environments. IEEE Robot. Autom. Lett. 2021, 6, 2036–2043. [Google Scholar]
  22. Lin, J.; Liu, X.; Zhang, F. A decentralized framework for simultaneous calibration, localization and mapping with multiple LiDARs. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 25–29 October 2020; pp. 4870–4877. [Google Scholar]
  23. Jiao, J.; Ye, H.; Zhu, Y.; Liu, M. Robust odometry and map for multi-LiDAR systems with online extrinsic calibration. IEEE Trans. Robot. 2021, 38, 351–371. [Google Scholar] [CrossRef]
  24. Zhang, F.; Zhang, Z.; Yang, L. A new PHD-SLAM method based on memory attenuation filter. Meas. Sci. Technol. 2021, 32, 095104. [Google Scholar] [CrossRef]
  25. Wang, F.; Zhao, X.; Gu, H.; Wang, L.; Wang, S.; Han, Y. Multi-Lidar system localization and mapping with online calibration. Appl. Sci. 2023, 13, 10193. [Google Scholar] [CrossRef]
  26. Cao, W.; Song, H. Targetless extrinsic calibration of multiple LiDARs based on pose graph optimization. In Proceedings of the 2024 5th International Conference on Machine Learning and Computing Applications (ICMLCA), Shenzhen, China, 23–25 February 2024; pp. 311–315. [Google Scholar]
  27. Chang, D.; Zhang, R.; Huang, S.; Hu, M.; Ding, R.; Qin, X. Versatile multi-lidar accurate self-calibration system based on pose graph optimization. IEEE Robot. Autom. Lett. 2023, 8, 4839–4846. [Google Scholar] [CrossRef]
  28. Zhou, Y.; Tuzel, O. VoxelNet: End-to-end learning for point cloud based 3D object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 4490–4499. [Google Scholar]
  29. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  30. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef]
  31. Chum, O.; Matas, J. Optimal randomized RANSAC. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 1472–1482. [Google Scholar] [CrossRef]
  32. Hotelling, H. Analysis of a complex of statistical variables into principal components. J. Educ. Psychol. 1933, 24, 417–458. [Google Scholar] [CrossRef]
  33. Edelsbrunner, H.; Mücke, E.P. Three-dimensional alpha shapes. ACM Trans. Graph. 1994, 13, 43–72. [Google Scholar] [CrossRef]
  34. Biber, P.; Straßer, W. The normal distributions transform: A new approach to laser scan matching. In Proceedings of the 2003 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 27–31 October 2003; pp. 2743–2748. [Google Scholar]
  35. Segal, A.; Haehnel, D.; Thrun, S. Generalized-ICP. In Proceedings of the Robotics: Science and Systems (RSS), Seattle, WA, USA, 28 June–1 July 2009; pp. 435–442. [Google Scholar]
  36. Mints, M.O.; Abayev, R.; Theisen, N.; Paulus, D.; von Gladiss, A. Online calibration of extrinsic parameters for solid-state LiDAR systems. Sensors 2024, 24, 2155. [Google Scholar] [CrossRef]
  37. Yang, H.; Shi, J.; Carlone, L. TEASER: Fast and certifiable point cloud registration. IEEE Trans. Robot. 2021, 37, 314–333. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.