Next Article in Journal
An Improved Map Information Collection Tool Using 360° Panoramic Images for Indoor Navigation Systems
Next Article in Special Issue
Research on a Prediction Method for Maintenance Decision of Expressway Asphalt Pavement Based on Random Forest
Previous Article in Journal
Experimental and Analytical Study of an Anode-Supported Solid Oxide Fuel Cell
Previous Article in Special Issue
Hybrid Rule-Based and Reinforcement Learning for Urban Signal Control in Developing Cities: A Systematic Literature Review and Practice Recommendations for Indonesia
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Multi-Sensor Calibration for Autonomous Vehicles: A Rapid Approach to LiDAR and Camera Data Fusion

by
Stefano Arrigoni
*,
Francesca D’Amato
and
Hafeez Husain Cholakkal
Department of Mechanical Engineering, Politecnico di Milano, 20156 Milan, Italy
*
Author to whom correspondence should be addressed.
Appl. Sci. 2026, 16(3), 1498; https://doi.org/10.3390/app16031498
Submission received: 11 December 2025 / Revised: 31 December 2025 / Accepted: 5 January 2026 / Published: 2 February 2026

Abstract

Precise sensor integration is crucial for autonomous vehicle (AV) navigation, yet traditional extrinsic calibration remains costly and labor-intensive. This study proposes an automated calibration approach that uses metaheuristic algorithms (Simulated Annealing (SA), Genetic Algorithms (GA), and Particle Swarm Optimization (PSO)) to independently optimize rotational and translational parameters, reducing cross-compensation errors. Bayesian optimization is used offline to define the search bounds (and tune hyperparameters), accelerating convergence, while computer vision techniques enhance automation by detecting geometric features using a checkerboard reference and a Huber estimator for noise handling. Experimental results demonstrate high accuracy with a single-pose acquisition, supporting multi-sensor configurations and reducing manual intervention, making the method practical for real-world AV applications.

1. Introduction

The rapid advancement of autonomous vehicle (AV) technology has made precise sensor integration crucial for safe and efficient navigation. Sensor fusion relies on accurate extrinsic calibration to interpret complex environments and support real-time decisions. As AVs become more reliant on multiple sensors, such as cameras and LiDAR systems, ensuring that these devices are properly aligned is vital to achieving reliable, real-time data interpretation. In autonomous driving, precise extrinsic calibration is a prerequisite for reliable multi-sensor fusion; errors in the relative pose between LiDAR, cameras, and radar directly degrade detection, tracking, and localization performance as highlighted by several recent surveys on sensor fusion for autonomous vehicles [1,2].
Extrinsic calibration, which determines the relative position and orientation between different sensors, is key to effective sensor fusion and, by extension, to the safety and performance of autonomous systems. The operational importance of accurate calibration is also reflected in major autonomous driving benchmarks, which provide fully calibrated sensor suites (e.g., nuScenes, Waymo Open Dataset) and document detailed mounting parameters to ensure the validity of perception and fusion evaluations [3,4].
Manual and automatic target-based calibration methods differ primarily in the level of human intervention required during the calibration process. Manual approaches require engineers to explicitly establish correspondences between LiDAR point clouds and camera images using calibration targets, such as checkerboards [5,6,7], fiducial markers (e.g., ArUco tags) [8], or custom planar targets [9]. Zhang’s method [5], which remains a cornerstone in camera calibration, estimates the extrinsic transformation by observing a planar target from multiple viewpoints. Manual methods, though still a benchmark for calibration, have paved the way for increasingly automated approaches. The literature [10,11] highlights a clear trend toward developing calibration toolboxes that reduce reliance on calibration objects and human intervention. With the rise in autonomous vehicles, there is a growing need for calibration methods that are fully automatic, swift, and adaptable to various configurations, enabling faster, robust, and accurate sensor fusion. Automatic target-based calibration methods aim to reduce human intervention by automatically estimating correspondences between sensor data and calibration targets. In industrial automotive settings, offline calibration procedures are still the de facto standard, as they provide repeatable and certifiable results required by safety regulations. However, such procedures typically involve highly controlled environments, precise target placement, multiple acquisitions, and expert supervision, resulting in calibration times ranging from several tens of minutes to several hours per vehicle, depending on sensor configuration and required accuracy [12]. These constraints translate into significant operational costs, mainly driven by dedicated infrastructure and skilled personnel rather than by the calibration algorithm itself. Industry reports indicate that even minor sensor misalignments following low-speed collisions can lead to recalibration expenses of up to $5000 (USD), largely due to the need for specialized facilities and manual setup [13]. Consequently, calibration methods that relax environmental constraints while maintaining accuracy are highly desirable. The proposed approach addresses this need by increasing flexibility in target placement and acquisition geometry. Unlike conventional procedures that require close-range positioning and repeated fine adjustments, the method allows for reliable calibration with targets placed at distances of up to seven meters while robustly filtering noise and outliers. This flexibility reduces setup time and lowers dependency on rigid infrastructure, thereby enabling cost-efficient deployment in existing industrial workflows without sacrificing accuracy. Recent research has explored online and learning-based calibration techniques that estimate extrinsic parameters during vehicle operation by exploiting natural scene structure and motion cues [2,10,14]. While promising, these approaches typically require extended data collection, sufficiently rich environments, and vehicle motion. Moreover, their convergence properties and safety assurances remain challenging to certify in safety-critical workflows. For these reasons, online and learning-based calibration is commonly treated as a complementary refinement rather than a replacement for offline calibration, which remains a necessary initial step in industrial settings. More broadly, generative-AI and hybrid AI frameworks are increasingly investigated to support autonomy-related tasks such as simulation, behavior modeling, and system-level development, providing a complementary context for AV research [15,16]. Offline automatic target-based methods remain widely used in industrial and pre-deployment contexts due to their accuracy and repeatability. Representative approaches, such as the method proposed by Geiger et al. [6], leverage automatic detection of multiple checkerboards to reduce manual effort. Nevertheless, these pipelines still rely on controlled calibration environments, precise target placement, and multiple observations to achieve robust results. In practice, offline target-based calibration procedures often require dedicated infrastructure, including leveled floors, fixed target rigs, and controlled lighting conditions, and are associated with non-negligible operational costs and calibration times ranging from several minutes to over an hour when setup and validation are included [14]. Within this offline target-based category, Zhou et al. [17] introduces a single-pose calibration method that estimates the extrinsic transformation between a LiDAR sensor and a camera using geometric constraints derived from a checkerboard plane and its edges. This approach relies on an iterative optimization process that minimizes three error terms, as outlined below:
  • Checkerboard normal-vector alignment: Ensures that the normal vector of the checkerboard in the LiDAR frame, denoted as n L , aligns with the normal in the camera frame, n C .
  • Alignment of the Checkerboard’s Edges: The direction of the edges in the LiDAR frame, represented by the vector d L , must match the direction of the edges in the camera frame, d C , ensuring consistency in their representation.
  • Accurate Projection of Edge Centroids: The edge centroids must project consistently in the image plane, providing a positional constraint that ensures the checkerboard’s edges are correctly positioned in the camera frame relative to the LiDAR frame.
Together, these error terms interact to refine the overall calibration process, with each term influencing the others. A discrepancy in one term, such as misalignment of the checkerboard’s normal vector, can propagate and affect the accuracy of the edge alignment and centroid projection, potentially leading to cumulative errors. Therefore, minimizing all three terms simultaneously is essential for achieving high-precision calibration. The objective of this process is to determine the optimal transformation—rotation R C L and translation t C L —that minimizes the sum of misalignments in the checkerboard’s normal, edges, and centroids:
min R C L , t C L i = 1 N ( 1 N i m = 1 N i n C · ( R C L p L + t C L ) + d C 2 + j = 1 K i k = 1 K i j A i j ( R C L q L + t C L p C ) 2 )
Specifically, p L and q L denote points in the LiDAR frame, while p C represents corresponding points in the camera frame. The term A i j is a projection matrix that removes components parallel to the edge direction d C . The variable N i indicates the number of points on the checkerboard plane in the LiDAR frame, K i represents the number of edges on the checkerboard in the camera frame, and K i j denotes the number of points on the j-th edge in the camera frame. Although Zouh’s method offers clear advantages in terms of efficiency, its reliance on a single observation of a planar calibration target introduces both strengths and weaknesses. On the one hand, this design significantly reduces data acquisition time and eliminates the need for repeated target placement or complex motion sequences. Consequently, the optimization problem involves only a small set of variables and constraints, ensuring fast convergence and low computational overhead—features that make the approach appealing in controlled calibration scenarios. On the other hand, the single-pose formulation inherently limits geometric redundancy, reducing the ability to compensate for errors. As a result, the accuracy of extrinsic parameter estimation becomes highly sensitive to plane fitting, edge extraction, and target placement. Measurement noise, partial occlusions, or uneven sampling of the checkerboard surface can propagate directly into the final solution, without the benefit of averaging across multiple observations. Verma et al. [18] explicitly analyze the impact of edge-based constraints on calibration accuracy. Their study shows that edge estimation accumulates uncertainty from both plane fitting and line fitting, whereas constraints based on the board normal depend solely on plane estimation. In addition, LiDAR beam divergence introduces systematic deviations in edge measurements, often on the order of several centimeters. To mitigate these effects, Verma et al. propose a cost function optimized via a Genetic Algorithm. While this approach improves robustness, it significantly increases computational cost and requires multiple target poses to ensure stable convergence, thereby increasing acquisition time and reducing operational flexibility. More recent offline target-based approaches, such as Jeong et al. [19], focus on improving numerical stability through refined non-linear optimization. By introducing weighted residuals and iterative refinement, these methods reduce sensitivity to noise and outliers while preserving the same set of unknowns and geometric constraints as Zhou’s formulation. However, the acquisition requirements remain unchanged, and in practice the reported accuracy gains are achieved through increased computational effort and the use of multiple observations. Gentilini et al. [20] address a different calibration scenario by jointly estimating extrinsic parameters across multiple sensors and multiple poses. Although this global optimization framework achieves high accuracy through redundancy, it substantially increases the dimensionality of the optimization problem and requires extended data acquisition, making it unsuitable for single-pose or rapid calibration scenarios. From an operational perspective, three quantitative factors are critical in practical calibration procedures: the number of required poses, which directly affects acquisition time; the dimensionality and complexity of the optimization problem, which influence convergence behavior and computational cost; and flexibility with respect to target placement and environmental conditions. Existing offline target-based methods primarily improve robustness by increasing redundancy, either through additional poses or higher computational effort. In contrast, the method proposed in this work operates under the same single-pose acquisition assumption as [17], while reducing sensitivity to noise and target placement constraints without increasing the number of poses or the dimensionality of the optimization problem. For this reason, Zhou’s method represents the most appropriate baseline for quantitative comparison, as it enables a direct evaluation of robustness improvements under identical acquisition and operational conditions.
In the literature, offline target-based LiDAR–camera calibration is typically addressed using conventional numerical optimization methods. [17] formulate the single-pose calibration as a non-linear least squares problem solved iteratively with the Levenberg–Marquardt algorithm. [19] extend this approach through weighted residuals and iterative refinement, while [20] employ a Gauss–Newton scheme for multi-pose, multi-sensor calibration, leveraging redundancy across observations to improve stability. These gradient-based strategies perform well for smooth, differentiable cost functions but become less effective when the optimization landscape exhibits strong nonlinearity, multimodality, or non-differentiable components. The calibration problem considered in this work involves a custom cost function with precisely these characteristics, making conventional numerical methods inadequate. To overcome this challenge, metaheuristic optimization algorithms are investigated. Metaheuristics have proven successful in autonomous driving tasks such as nonlinear MPC-based motion planning and obstacle avoidance, where Firefly algorithms achieved efficient convergence under non-convex constraints [21]. Although the focus here is extrinsic calibration, the underlying optimization landscape shares similar properties—strong nonlinearity, multimodality, and lack of analytical gradients—justifying the adoption of metaheuristic techniques. Despite extensive research on LiDAR–camera calibration, robust and rapid procedures remain essential for in-vehicle deployment, where frequent recalibration and minimal human intervention are required; recent frameworks emphasize this need in real autonomous platforms [22]. To address these requirements, three widely used metaheuristic algorithms—Simulated Annealing (SA), Genetic Algorithms (GA), and Particle Swarm Optimization (PSO)—are evaluated and compared against traditional numerical methods. Unlike gradient-based approaches, which are prone to local minima in highly non-linear and multimodal spaces, metaheuristics employ global search strategies that balance exploration and exploitation, enabling efficient approximation of optimal solutions [23]. The comparison of different optimization algorithms such as SA, GA, and PSO was motivated by the “No Free Lunch” theorem [24], which states that no single algorithm universally outperforms all others across every problem. Consequently, the most effective algorithm for this calibration task was identified through experimental evaluation. According to literature guidelines, GA is typically well suited for problems requiring extensive global exploration, PSO is effective for rapid convergence in smoother search spaces, and SA performs well for stochastic local search. However, these tendencies are general observations and do not guarantee superior performance for every specific problem. To exploit the complementary strengths of these algorithms, reduce cross-compensation effects, and simplify the optimization problem, the procedure was divided into two independent 3D phases—one for rotation and one for translation. Each phase was solved using the algorithm shown to perform best in practice, with the rotation matrix represented via Euler angles to both reduce the number of optimization variables and maintain the orthogonality of the matrix during the optimization process. To maximize the performance of the selected metaheuristic algorithms, Bayesian optimization was employed to define the search space and tune algorithm parameters offline. This approach automatically selects ideal ranges for relative yaw, pitch, roll, and translation between sensors, enhancing both optimization efficiency and adaptability to various sensor configurations.Furthermore, computer vision tools were employed to automate the selection of geometric features in both camera and LiDAR frames using checkerboard dimensions as the sole reference, while an innovative strategy incorporating a Huber estimator was devised to filter out errors and outliers. This approach ensured that the optimization process focused on reliable data, minimized noise impact, and improved overall precision, even when calibration relies on a single pose.
In Section 2, we present the methodology of our framework, detailing the implementation of the cost functions. Section 3 details the architecture of the proposed method, emphasizing Bayesian optimization for process automation. Section 4 outlines the experimental setup, followed by Section 5, which presents the results. Finally, Section 6 discusses conclusions and future directions.

2. Fundamentals of Metaheuristic Algorithms

The term metaheuristic stems from the Greek words “meta” (beyond) and “heuristic” (to find) [25]. According to Laporte and Osman [23], metaheuristics comprise iterative methods that guide subordinate heuristics by combining diverse strategies to explore and exploit the search space, using learning mechanisms to approach near-optimal solutions efficiently. While no universal definition exists, metaheuristics generally refer to adaptable algorithms that overcome the limitations of traditional optimization methods—particularly for nonlinear, multimodal, or high-dimensional problems where gradient-based or direct search approaches often fail. They rely on diversification (global search) and intensification (local exploitation) [26]. Widely applied across domains like NP-hard problem solving, robotics, and image processing [27], metaheuristics are typically classified by their source of inspiration: evolutionary algorithms, swarm intelligence, physical-law-based methods, and miscellaneous strategies [28,29]. Building on this foundation, the metaheuristic algorithms employed in this study—each representing a different category—are described in detail below:
  • Genetic algorithm: Genetic Algorithms (GAs), developed by John Holland in the 1960s, emulate the process of natural adaptation through computational means [30]. Inspired by biological evolution, GAs operate on a population of candidate solutions called “chromosomes,” each composed of “genes” representing variables or traits that can take on different values, known as “alleles.” The fitness of a chromosome, evaluated by a fitness function to be maximized, determines its probability of being selected for reproduction.
    The algorithm follows a cycle of steps: it begins with the initialization of a population of n p o p individuals encoded as bit strings, balancing randomness, compositionality, and generality to ensure a diverse search space. Next, each chromosome’s fitness is evaluated [31]. Then, offspring are generated by applying genetic operators—selection, crossover, and mutation—to produce m new individuals. The parent and offspring populations are merged, evaluated, and sorted, and the best individuals are selected to form the next generation. This process iterates for a predefined number of generations, typically ranging from 50 to 500. The overall workflow of the Genetic Algorithm is illustrated in Figure 1.
    Because GAs are stochastic, repeated runs with the same parameters can yield different results. Thus, results are usually averaged over multiple runs to provide robust performance metrics. In the automotive domain, Genetic Algorithms have also been successfully applied to nonlinear Model Predictive Control (MPC) trajectory planning for autonomous driving, demonstrating high convergence reliability under complex vehicle-dynamics constraints [32]. Such applications confirm the suitability of GAs for non-convex optimization problems similar to those addressed in this work. The three key genetic operators balance exploration and exploitation within the search:
    -
    Selection favors fitter individuals to propagate their traits. A common method is roulette wheel selection, where the probability of selecting chromosome i is proportional to its fitness:
    P ( i ) = f i j = 1 n p o p f j
    Here, f i is the fitness of chromosome i and n p o p is the population size. Although simple, this method may cause premature convergence by quickly favoring highly fit individuals. Alternatives such as tournament selection—choosing the best from random subsets—and ranking selection—assigning selection probabilities based on relative fitness ranks—help maintain diversity and prevent dominance [33].
    -
    Crossover mimics biological recombination by exchanging gene segments between two parents. Techniques like one-point, two-point, or uniform crossover combine traits to create potentially superior offspring, enhancing the search for optimal solutions [34].
    -
    Mutation introduces random changes in gene alleles, injecting novel genetic material to prevent premature convergence and maintain population diversity. Mutation promotes exploration, complementing the exploitative nature of crossover.
    Convergence is typically defined when 95% of the population shares the same allele at each gene locus, indicating a high degree of genetic uniformity. This balance of operators allows GAs to effectively navigate complex search spaces, evolving populations toward fitter solutions over successive generations.
  • Particle Swarm Optimization: Particle Swarm Optimization (PSO), introduced by Kennedy and Eberhart in 1995, is inspired by the social behavior of bird flocks and is designed to optimize continuous nonlinear functions [35]. The algorithm employs a swarm of particles, where each particle represents a candidate solution characterized by its position and velocity within the search space. Over successive iterations, particles adjust their trajectories by leveraging both their individual experience and the collective knowledge of the swarm, ultimately converging toward optimal solutions.
    Starting from randomly initialized positions and velocities, the velocity and position of each particle k at iteration i + 1 are updated as follows:
    V k ( i + 1 ) = w · V k ( i ) + c 1 · r 1 · ( p b e s t k X k ( i ) ) + c 2 · r 2 · ( g b e s t X k ( i ) )
    X k ( i + 1 ) = X k ( i ) + V k ( i + 1 )
    The velocity update equation consists of three components: the momentum term w · V k ( i ) , which helps maintain the particle’s current direction and balances exploration and exploitation; the cognitive term c 1 · r 1 · ( p b e s t k X k ( i ) ) , which draws the particle toward its personal best position to encourage individual learning; and the social term c 2 · r 2 · ( g b e s t X k ( i ) ) , which attracts the particle toward the global best position discovered by the swarm, promoting collective learning.
    Here, the parameters are defined as follows: w is the inertia weight that controls the influence of the particle’s previous velocity; c 1 and c 2 are acceleration coefficients that weigh the relative contributions of personal and social experiences; r 1 and r 2 are random scalars uniformly sampled from the interval [ 0 , 1 ] at each iteration, introducing stochastic variability to avoid premature convergence; V k ( i ) and X k ( i ) represent the current velocity and position of particle k, respectively; pbest k is the best position found so far by particle k; and gbest is the best position found by the entire swarm.
    The effectiveness of PSO is governed by key parameters that balance exploration and exploitation. The inertia weight w influences how strongly a particle relies on its previous velocity, with larger values encouraging broader exploration and smaller values focusing the search around promising areas. The acceleration coefficients c 1 and c 2 control the relative importance of individual learning versus social learning within the swarm. Additionally, although not explicitly included in the standard velocity update equations, a maximum velocity parameter V max is often introduced in some implementations to cap the particle velocity, preventing excessive jumps in the search space and facilitating smoother convergence.
    By iteratively updating particles’ velocities and positions in accordance with these principles, PSO efficiently balances the trade-off between global exploration and local exploitation, thereby converging toward near-optimal solutions. Moreover, PSO and its accelerated variants have been integrated into nonlinear MPC frameworks for real-time motion planning by the authors, achieving both computational efficiency and solution robustness [36]. This evidence further supports the adoption of PSO in the present study as a reliable global-search optimizer for highly nonlinear calibration problems.
  • Simulated Annealing: Simulated Annealing (SA), introduced by Kirkpatrick et al. [37], is an optimization algorithm inspired by the metallurgical process of annealing, where a crystalline solid is heated to a high-energy state and then slowly cooled to reach a stable, low-energy configuration with minimal defects. Analogously, SA starts from a random solution and gradually lowers a control parameter called temperature T, exploring neighboring solutions at each step. Unlike greedy methods, SA probabilistically accepts worse solutions early on, enabling escape from local minima and a more thorough search of the solution space. As T decreases, the algorithm becomes more selective, converging to an optimal or near-optimal solution.
    The analogy between physical annealing and simulated annealing is summarized as follows:
    Physical AnnealingSimulated Annealing
    System StatesSolutions
    EnergyCost
    Perturbed StateNeighboring Solutions
    TemperatureControl Parameter
    Final StateHeuristic Solution
    The SA algorithm proceeds through these steps:
    1.
    Initialization: Start with a random initial solution S 0 and initial temperature T 0 .
    2.
    Main Loop: While the stopping condition is unmet:
    -
    Generate a neighboring solution S by perturbing the current solution S and compute its cost E ( S ) .
    -
    Calculate the energy difference Δ E = E ( S ) E ( S ) .
    *
    If Δ E 0 , accept S as the new solution.
    *
    If Δ E > 0 , accept S with probability e Δ E / T , allowing occasional uphill moves to avoid local minima.
    -
    Update the temperature according to the cooling schedule, typically exponentially:
    T i = α T i 1
    where α ( 0 , 1 ) controls the cooling rate [38].
    The overall workflow of the SA Algorithm is illustrated in Figure 2.
    This gradual cooling and probabilistic acceptance strategy enables SA to balance exploration and exploitation effectively, making it a powerful metaheuristic for global optimization.

3. Methodology

In this section, the methodology and the main steps of the extrinsic calibration process are presented. The sensor architecture consists of three main devices: an OS1 LiDAR (Ouster, Inc., San Francisco, CA, USA), a Ladybug5+ omnidirectional camera (Teledyne FLIR Integrated Imaging Solutions, Inc., Richmond, BC, Canada), and a ZED X stereo camera (Stereolabs, Inc., San Francisco, CA, USA), as summarized in Table 1. These sensors include both active (LiDAR) and passive (cameras) devices, whose combination enables sensor fusion for a more complete and reliable perception of the environment [39].
The procedure begins with an initial recording of approximately one minute to analyze and minimize measurement noise. The data from all sensors are recorded and synchronized. A checkerboard calibration target with a 9 × 7 grid of 0.108 m squares is employed to establish correspondences across the modalities.
Specifically, rather than relying on traditional board edges, which can introduce fitting errors, the focus is on the checkerboard’s center and normal, thereby reducing inaccuracies. RANSAC and DBSCAN clustering are used to isolate the checkerboard plane from the LiDAR data to accurately detect the chosen features. An approach using the Huber estimator is implemented to refine the results by prioritizing points that closely match the expected dimensions of the checkerboard. This ensures more reliable coordinates even in the presence of noise.
The procedure calculates the width and height of the checkerboard for each acquisition i by measuring the distances between corner points. The total deviation is then computed as:
total _ deviation i = | width i W | + | height i H | W + H
with H and W being the known dimensions of the board. This deviation is used to assign weights to the acquisitions, which are then combined in the weighted estimator:
X ^ = w i X i w i
where X i represents the data from each acquisition, and w i is the weight assigned to that acquisition.
For the camera data, the coordinates of the four corners are derived from the checkerboard’s dimensions and expressed in the standard reference frame at the upper-left corner in the pixel frame. These points, along with the center (calculated as the average of the detected corners), are then transformed into the world coordinate system. The normal vector is obtained by transforming a point along the checkerboard’s z-axis into the camera frame. Despite some intrinsic noise, high resolution minimizes errors, and the Huber estimator refines the coordinates, ensuring reliable calibration.
At the end of this process, the features are represented as 3xNposes matrices, where the rows represent the x, y, z coordinates (Figure 3) while the columns represent the number of poses:
Lidar frame : C L . j , j = 1 4 , C L , N l Camera frame : C C , j , j = 1 4 , C C , N c
with C j denoting the corners, C the center of the board, and N the normals. For reference, the corners and edges of the board expressed in the LiDAR reference frame are shown in Figure 3.
At this point, the extrinsic calibration problem is redefined as a streamlined optimization task that directly minimizes alignment errors between sensor frames, diverging from traditional methods that rely on point cloud-to-pixel correspondences and varied feature sets. By structuring the problem in this way, straightforward optimization algorithms can target solutions directly, bypassing the need for preliminary approximations or suboptimal results requiring later refinement with non-linear methods, thus achieving both efficiency and precision in calibration. Further optimization is achieved by representing rotations using Euler angles rather than a full 3 × 3 matrix. This reduces the number of optimization variables from nine to three and inherently preserves the orthogonality of the rotation matrix throughout the optimization process. Directly optimizing the individual matrix elements would require enforcing orthogonality at each iteration, as independent updates by the optimization algorithms could otherwise produce invalid transformations. By using Euler angles, the optimization is simplified while ensuring that all resulting transformations remain valid homogeneous transformations. The overall formulation is guided by geometric observability and numerical stability considerations rather than empirical tuning. In particular, the decomposition of the optimization into sequential stages and the selection of complementary geometric error terms are designed to reduce parameter coupling and improve convergence robustness. Detailed definitions of each cost component and their role in the optimization are provided in the following subsections. This approach leads to a multi-objective cost function that integrates both rotational and translational discrepancies, with each component balanced by experimentally determined weights to ensure accurate alignment. To avoid cross-compensation between translation and rotation, calibration is divided into two phases and implemented in MATLAB as follows.
Phase 1: The first phase aims to align the checkerboards within the camera’s reference frame by minimizing three angular errors, producing an initial guess for the rotation that maps the LiDAR points first to the ZED X frame and then to the Ladybug5+ frame.
  • Normal Error Alignment: The first error is the alignment between normal vectors. Given the normal matrices N l and N c ( 3 × N extracted _ point ), and the Euler angles, a rotation matrix R is derived. For each point i, the normal vectors in Lidar frame n l i and camera frame n c i are computed as follows:
    n l i = N l [ : , i ] N l [ : , i ] , n c i = N c [ : , i ] N c [ : , i ]
    Applying the rotation to the normal vector in the Lidar frame:
    n l i = R · n l i
    The alignment error is computed using the dot product, clamped between 1 and 1 to prevent numerical errors:
    dot _ product = n c i · n l i = cos ( θ )
    The angular error θ (Figure 4) is computed as:
    θ = arccos ( n c i · ( R · n l i ) )
    Afterwards, Huber loss is applied to handle variations in reflections, distances, and other uncertainties affecting checkerboard poses.
    This choice prioritizes reliable data while mitigating measurement errors, ensuring a robust cost function that improves algorithm convergence:
    huber _ err = 1 2 · angle _ err 2 if | angle _ err | δ δ · | angle _ err | 0.5 · δ otherwise
    where δ = 0.05 rad based on the observed range of angular deviations in the calibration data. Finally, the Huber error for each normal vector pair is accumulated across all points. The final output, e n , is the average Huber error across all N points:
    e n = i = 1 N p o i n t s normal _ huber _ error i N p o i n t s
  • Directional Alignment Error: The second error reflects the degree to which the diagonals in each frame are misaligned (Figure 5), following the same logic as the previous one, with the only difference being that the calculations are now performed on the vectors pointing from the center to each corner:
    D i , j , L = C i , j , L C i , L , D i , j , C = C i , j , C C i , C , j = 1 , , 4 , i = 1 , , N p o i n t s
    Applying the rotation to the normalized vectors:
    D i , j , l = R · D i , j , l , j = 1 , , 4 , i = 1 , , N p o i n t s ,
    At this point, the function performs the same calculations shown in Equations (10) and (12), applying a threshold of 0.05 rad for consistency, based on the same rationale outlined for the normal vectors. This yields the final directional alignment error:
    e D , j = i = 1 N p o i n t s direction _ huber _ error i N p o i n t s ( j = 1 , , 4 )
    Summing over all corners:
    e D , tot = j = 1 4 e D , j
  • Perpendicularity Error: The third error ensures perpendicularity between the transformed normal vector and the corner directional vectors (Figure 6), computed via:
    e P , tot = j = 1 4 e P , j
    To ensure this perpendicularity, the function calculates the arcsine of the cross product rather than using the arccosine, allowing it to capture small deviations effectively. Finally, the first objective function is formulated with weights:
    w n = 0.5 , w d = 0.25 , w p = 0.25
    Minimizing:
    arg min Euler angles J = w n · e n + w d · e D , t o t + w p · e P , t o t
    The weights were selected based on geometric robustness considerations, prioritizing normal vector alignment due to its higher stability compared to corner-based features. These parameters were fixed once and applied consistently across all experiments, and the optimization was observed to be stable under moderate variations of their values.
Phase 2: The second phase aims to determine the rigid transformation that precisely aligns the LiDAR point clouds with those in the camera frame, ensuring accurate calibration. To enhance the initial rotation estimate, both directional e d and normal errors e n are reintroduced to add redundancy, considering the variability of real-world data.
  • Centers Alignment Error: A new function is implemented to refine the rigid transformation by ensuring precise alignment of centers and corners (Figure 7a,b).
    This function retrieves the center points in both frames across multiple poses and transforms the LiDAR centers into the Camera frame using the estimated rotation and translation:
    C i , L = R · C i , L + t i = 1 , , N p o i n t s
    The alignment error is then quantified through the Euclidean distance:
    error _ value i = C i , C C i , L , i = 1 , , N p o i n t s
    To mitigate the influence of outliers, the Huber loss function is applied with a selected threshold δ = 0.0074 m following Equation (12).
    The final error metric is computed as the average Huber loss across all poses:
    e f = i = 1 N p o i n t s centers _ huber _ error i N p o i n t s
This process is systematically repeated for all corner points, producing e corners as the alignment error for the corners, and resulting in an objective function that combines multiple weighted error components:
arg min eul , t J = w n · e n + w d · e d + w f c · e f + w f · e corners
where the weights are empirically determined:
w n = 0.4 , w f c = 0.3 , w d = 0.05 , w f = 0.25
The weights are experimentally selected, guided not only by the sources of error discussed previously but also by the amplification of the angular errors over distance. For example, when a rotational misalignment occurs, the perceived positional error of an object increases as the distance grows. This understanding also leads to the decision to divide the calibration problem into two stages, with a stronger emphasis on rotation in the initial phase to ensure precise angular alignment across multiple poses.

4. Bayesian Optimization Approach

To obtain an initial estimate of the rotation matrix, the normal vector was used due to its robustness, being influenced primarily by the plane-fitting error from RANSAC in the LiDAR frame:
R = ( N L N L ) 1 · N L · N
where N L and N represent the normal vectors in the LiDAR and camera frames. This matrix was then converted into Euler angles, providing a structured starting point for the optimization process. The next step consisted of defining the search space within which the optimal rotation matrix could be located. This required establishing a sufficiently good lower and upper bound around the initial estimate, taking into account both the relative orientation of the sensors and potential mounting errors, which are inherently unknown.
Manually exploring this search space through trial and error would have been time-consuming, prone to inconsistencies and contrary to the goal of an automated calibration process. To address this challenge, Bayesian optimization was employed to systematically identify optimal bounds for each algorithm (GA, SA, PSO), effectively defining the search space required to minimize the cost function across all approaches.
Bayesian optimization has proven to be a powerful technique for complex, expensive optimization problems, particularly when the objective function is nonlinear, non-convex, and difficult to model analytically. By using a Gaussian Process (GP) to probabilistically model the objective function and iteratively refining estimates through an acquisition function such as Expected Improvement (EI), the method balances exploration and exploitation, enabling an efficient search without requiring explicit derivatives. This approach has been widely adopted in machine learning, especially for hyperparameter tuning in high-dimensional spaces, where traditional methods like grid search or random search are inefficient.
The potential of Bayesian optimization extends beyond machine learning to metaheuristic algorithms, which also rely on careful parameter tuning for optimal performance. Similar to neural networks, the relationship between parameters and outcomes in metaheuristics is often complex and problem-dependent. Bayesian optimization overcomes the need for an explicit model by using sample evaluations of the objective function, making it particularly valuable for sensor calibration, where the search space is initially undefined and optimal solutions are not readily apparent.
Although only a few studies, such as [40], have explored the combination of Bayesian optimization with metaheuristics, the results demonstrate its effectiveness in improving algorithm performance through optimal parameter selection. This approach was successfully applied to the Hybrid Kernel EDA algorithm, and future work could extend its application to other metaheuristics.
By integrating Bayesian optimization, the search boundaries are tuned offline in a one-time configuration stage, prior to the actual calibration. Since the extrinsic calibration is static, this procedure is not executed at each calibration run but only when the sensor setup changes, and the resulting bounds are reused thereafter, significantly reducing the runtime of the subsequent optimization process. For this study, a broad initial range of [ [ 200 , 0 ] , [ 0 , 200 ] ] was chosen, enabling the optimizer to iteratively refine the bounds and efficiently guide the algorithms toward the optimal solution.
For each candidate configuration, GA, SA, and PSO were executed 20 times, and the resulting cost values and execution times were averaged and normalized, yielding AvgCost and AvgTime. The optimization objective was defined as:
obj = 0.5 · AvgCost Cost min Cost max Cost min + 0.5 · AvgTime Time min Time max Time min
ensuring a trade-off between accuracy and efficiency.
The comparative analysis showed that all three algorithms achieved similar accuracy within the defined bounds, but their computational profiles differed significantly. SA offered stable performance but required substantially longer execution times. PSO performed competitively in terms of accuracy, but its sensitivity to parameter choices made it less consistent. GA, instead, consistently combined low average cost with the shortest execution times, confirming its suitability for the first calibration phase. For this reason, only GA results are shown in the following figures, as it was ultimately selected for the subsequent parameter-tuning stage (see Figure 8, Figure 9 and Figure 10). Table 2 summarizes the performance of all three algorithms after Bayesian optimization. From the table, it is evident that GA (roulette selection) achieved the lowest average cost and shortest average execution time, confirming the choice of GA for the subsequent parameter tuning stage. The table also highlights that although SA was the most robust to parameter changes, its execution time was a limiting factor, and PSO required careful bound selection to achieve similar performance.
Detailed analysis of GA parameters showed that higher crossover percentage (Pc) reduced cost but increased execution time, while mutation percentage (Pm) introduced cost fluctuations and slightly increased execution time. The crossover range factor (gamma) maintained stable times until 0.5, then spiked, and the mutation rate (mu) minimized cost at 0.2 while increasing execution time at 0.15.
After completing the robustness analysis in phase one, the optimized rotation matrix R was exported as the starting point for phase two. This matrix provided a reliable initial estimate, which was then used both to refine R and to define the translation vector:
t l c = mean ( C C R · C L )
where C C and C L represent the corresponding calibration points in the camera and LiDAR frames.
The search space for phase two was established around these initial estimates, with lower and upper bounds adjusted to account for sensor orientation and potential mounting errors. Bayesian optimization was again employed to refine the bounds and guide parameter selection. All three algorithms—GA, SA, and PSO—were evaluated to assess cost and execution time. However, only PSO results are presented in the figures, as it provided the most favorable compromise between accuracy and computational efficiency.
Cost analysis revealed that SA generally maintained lower values (around 4.4 × 10 4 ) across the search space, while PSO achieved comparable results, with slightly higher peaks at the bounds. GA, in contrast, reached costs up to 10 3 . Execution time considerations highlighted that SA required up to 30 s, whereas PSO and GA completed the computations up to 100 times faster (See Table 3).
Following parameter tuning, PSO consistently produced solutions close to the minimum within reasonable time frames. Table 3 reports the numerical results, while Figure 11, Figure 12 and Figure 13 illustrate fitness and execution time trends for PSO within the refined search space.
As a result of the experimental evaluation guided by the No Free Lunch theorem, the optimization process adopts a hybrid approach, using a Genetic Algorithm (GA) for Phase 1 and Particle Swarm Optimization (PSO) for Phase 2. During Phase 1 (GA) of the hybrid optimization process, an adaptive mutation function is employed, as summarized in Algorithm 1. Instead of using a standard fixed mutation, the mutation intensity σ is dynamically adjusted based on the parent’s fitness:
δ = 1 parent . Fitness min ( pop . Fitness ) , σ = δ · ( VarMax VarMin )
Selected genes of the parent are mutated using σ , enabling adaptive exploration of the search space while keeping values within the predefined bounds, VarMin and VarMax. This strategy improves convergence, reduces the number of tunable parameters, and prevents premature convergence during the GA phase. Additionally, 50% of the first-phase GA population is carried over to initialize Phase 2 (PSO), providing a heuristic that leverages promising solutions. Overall, this approach balances exploration and exploitation, enhances search efficiency, and ensures the discovery of high-quality solutions, as illustrated in the pseudocode below.
Algorithm 1: Hybrid GA-PSO Optimization
Def.   C o s t F u n c t i o n G A , C o s t F u n c t i o n P S O
Init. GA ( n p o p , m a x _ g e n , P c , P m , b o u n d s )
GA Phase1: Init. Pop ∈ bounds
for   g e n = 1 m a x _ g e n   do
       Eval fitness
       Selection → Crossover → AdaptiveMutation
       Update Pop
end for
Output:   e u l f i r s t _ g u e s s , P o p G A
PSO Phase2: Def. bounds for e u l , t
 Init. Swarm: 50% PopGA, 50% rand
for   i t e r = 1 m a x _ i t e r   do
       for each particle do
           Eval fitness
           Update velocity, position ∈ bounds
           Update local, global best
       end for
end for
Return:  e u l b e s t , t b e s t

5. Experimental Setup

The design of the new platform benefits from the experience gained in our previous prototypical autonomous-vehicle testbed [41], where the overall system architecture, interfacing strategies and validation methodology were established on a full-scale electric vehicle. In the present work, a different hardware and computing architecture is adopted, tailored to the specific multi-sensor configuration and calibration objectives described in Section 3.
The sensor system used in this study integrates multiple sensing modalities within a Nissan Townstar van to achieve precise and reliable environmental perception (see Figure 14). The external sensor setup is composed of an OS1 LiDAR (mounted 270 mm above the base with a custom cylindrical flange), a Ladybug 5+ camera (elevated to 578 mm via a tripod adapter), and a ZED X stereo camera (positioned at 175 mm with an adjustable “L” bracket for pitch alignment). Due to the elevated mounting height of the external sensors, the resulting geometric configuration differs from that of standard passenger vehicles, affecting viewing angles and the spatial distribution of LiDAR points on calibration targets. To account for this aspect and to avoid overfitting the calibration procedure to a single fixed geometry, experiments were conducted by varying the distance between the sensor suite and the calibration target across multiple acquisition sessions. Inside the vehicle, additional modules are installed, including the IMU, the GPS board, the computing and storage units, and the power electronics required to supply the entire system. In particular, power is provided by two 24 V, 70 Ah battery groups, regulated through inverters that convert 24 V DC into 230 V AC, ensuring stable distribution while continuously monitoring voltage and current.
Data processing is distributed across two computing units: a NUC, which manages the inputs from the Ladybug camera, the IMU, and the DURO GPS, and a Jetson, which processes the data from the OS1 LiDAR and the ZED X stereo camera.
All components are interconnected via a network switch, which implements a modular software architecture that logically separates sensing, processing, and storage functions. This architecture enables efficient data flow between subsystems, ensures scalability for future expansion, and simplifies the integration and maintenance of heterogeneous modules.
The acquisition system is based on the open-source framework ROS. Sensor data streams are synchronized and collected in real time, then stored into dedicated log files. To optimize storage, data are automatically compressed and transferred to a NAS for long-term retention.
A modem ensures network synchronization across all devices, while a manual interface in the passenger compartment allows for real-time control of power distribution and system settings. To maintain signal integrity and minimize interference, all cables are routed through protective housings beneath the structure, resulting in a clean and efficient integration of power and data management for high-fidelity perception.

6. Results

To demonstrate the effectiveness of the proposed calibration method, we evaluated its performance when reducing the number of target poses from nine to a single pose. The validation relied on a task-driven evaluation criterion directly relevant to autonomous driving: the lateral error perceived by the sensor system for an obstacle placed at a known distance from the sensor setup. This metric is critical because it directly impacts vehicle maneuvering and obstacle localization accuracy. In the experimental setup, a red spherical target (Figure 15) was positioned at a known distance from the sensor setup. Its center was detected in the image plane and reprojected into the LiDAR reference frame. From this reprojection, the azimuth deviation Δ θ and elevation deviation Δ ψ were computed. The azimuth deviation was then converted into a lateral error L error at the obstacle distance d obstacle (measured from the sensor setup), according to:
L error = d obstacle · tan ( Δ θ )
When reducing the calibration from nine poses to a single pose, the increase in lateral error was negligible:
L error = 0.15 mm for d obstacle = 10 m from the sensor setup .
Detailed results for both LiDAR-to-Ladybug and LiDAR-to-ZED X transformations are reported in Table 4. This minimal increase confirms that the single-pose configuration satisfies the accuracy requirements of autonomous driving applications while significantly reducing calibration complexity.
Once the number of poses was reduced to one, we compared the proposed method with the approach in [7]. Zhou’s method (Table 5), while achieving a significantly lower elapsed time—processing a single frame in just 3.1 s compared to the 10.986 s required by the proposed method—relies exclusively on automatic detection of checkerboard dimensions. This reliance, heavily influenced by external environmental conditions, leads to azimuth and elevation errors that are, on average, 217 times higher than those obtained using the proposed method, highlighting the necessity of managing uncertainties and leveraging stable features for precise and reliable extrinsic calibration.
For completeness, the hybrid metaheuristic optimization approach was compared with classic gradient descent, included as a representative local optimization baseline. As shown in Figure 16 and Figure 17 and Table 6, gradient descent frequently becomes trapped in local minima, resulting in suboptimal calibration. This comparison highlights the ability of metaheuristic strategies to explore the non-linear and multimodal cost landscape more effectively, achieving higher accuracy and robustness even when the initial estimate is poor. In contrast, the proposed approach successfully completes the optimization in just 0.284 s.
Building on the same task-driven validation strategy, the influence of target distance on calibration accuracy was systematically analyzed. This analysis serves a dual purpose: first, to identify a practical distance range that ensures calibration reliability while relaxing setup constraints and increasing flexibility in target placement; and second, to evaluate the robustness of the proposed method under varying effective geometric configurations induced by sensor mounting height and viewing angle. As illustrated in Figure 18 and summarized in Table 7, calibration error exhibits only a marginal increase when the target distance from the sensor setup varies between 4.53 m and 6.71 m. At approximately 7 m, the resulting angular deviations translate into lateral errors L error that remain within the centimeter range. Such deviations are well below the thresholds that affect obstacle localization, lane-level perception, and trajectory planning in urban and highway autonomous driving scenarios. Beyond this distance, error growth becomes non-linear. Angular deviations Δ θ of 1–2 degrees produce lateral displacements ranging from a few centimeters for near-field objects—such as those encountered during assisted parking—to approximately 30 cm at d obstacle = 10 m . This larger distance is considered representative of scenarios where vehicles ahead are detected at extended ranges, for example, during highway driving. The comparison between these two cases illustrates the impact of distance on calibration accuracy: small lateral errors at short ranges may be acceptable for low-speed maneuvers, whereas the substantially larger deviations observed at longer ranges compromise the requirements of safety-critical autonomous driving tasks that demand precise obstacle positioning. Since sensor mounting height primarily affects calibration through changes in viewing geometry and angular sensitivity, expressing calibration validity in terms of target distance provides a vehicle-agnostic criterion that captures these effects in a practical and measurable form. Based on this analysis, 7 m is identified as the maximum reliable target distance for static offline calibration, providing a balanced compromise between operational flexibility and the accuracy constraints of autonomous driving applications.

7. Conclusions and Future Work

This paper presented a metaheuristic-based framework for extrinsic sensor calibration tailored to autonomous-vehicle platforms. The proposed method satisfies the demanding requirements of AV perception pipelines by combining speed, adaptability, and precision. These properties align with the needs observed in large-scale benchmarks, where accurate and up-to-date calibration underpins reliable multi-sensor fusion [1,3].
For pairwise calibration tasks (e.g., Ladybug+LiDAR or ZED+LiDAR), the process completes in approximately 11 s at runtime, with the search space and metaheuristic parameters already defined during the offline configuration stage. Optimization parameters selected via Bayesian optimization yield robust and more accurate results than Zhou’s state-of-the-art method. Although the computational time is slightly higher, this increase is justified by improved noise handling, enabling superior precision even up to a 7-m range without requiring a strictly controlled environment. The experiments also confirmed consistent performance across different sensor pairs and scene configurations, with repeatability errors below 0.5°, indicating readiness for integration into autonomous-vehicle calibration pipelines.
Future extensions of this work will address four key aspects:
  • porting the implementation to C++ for real-time and embedded deployment;
  • performing a systematic sensitivity analysis of weighting coefficients to better quantify their influence on calibration accuracy;
  • extending the calibration range through distance-weighted costs and improved target reflectivity;
  • validating robustness under diverse environmental conditions;
  • unifying intrinsic and extrinsic calibration into a single optimization stage.
Collectively, these enhancements will strengthen the robustness and flexibility of the proposed calibration framework, extending its applicability to real-world autonomous vehicles. The achieved accuracy, repeatability, and execution time demonstrate the maturity of the approach as a reliable calibration tool for current AV research and development while paving the way toward fully automated, high-precision calibration pipelines for next-generation intelligent transportation systems.

Author Contributions

Conceptualization, F.D. and H.H.C.; methodology, F.D. and S.A.; software, F.D.; validation, S.A., F.D. and H.H.C.; formal analysis, H.H.C.; investigation, F.D.; data curation, F.D.; writing—original draft preparation, F.D.; writing—review and editing, S.A.; supervision, S.A.; project administration, S.A.; funding acquisition, S.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available on request from the corresponding author due to (specify the reason for the restriction).

Acknowledgments

The authors would like to thank Milano Serravalle—Milano Tangenziali S.p.A. for their support in this work, in particular for providing access to the test vehicle and the measurement equipment used in the experimental campaign. During the preparation of this manuscript/study, the author(s) used ChatGPT 5.0 for the purposes of textual revision. The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yeong, S.M.; Velasco-Hernandez, J.; Barry, C.; Walsh, J.J. Sensor and Sensor Fusion Technology in Autonomous Vehicles: A Review. Sensors 2021, 21, 2140. [Google Scholar] [CrossRef] [PubMed]
  2. Wang, H.; Liu, J.; Dong, H.; Shao, Z. A Survey of the Multi-Sensor Fusion Object Detection Task in Autonomous Driving. Sensors 2025, 25, 2794. [Google Scholar] [CrossRef]
  3. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. nuScenes: A Multimodal Dataset for Autonomous Driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11618–11628. [Google Scholar] [CrossRef]
  4. Sun, P.; Kretzschmar, H.; Dotiwalla, X.; Chouard, A.; Patnaik, V.; Tsui, P.; Guo, J.; Zhou, Y.; Chai, Y.; Caine, B.; et al. Scalability in Perception for Autonomous Driving: Waymo Open Dataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 2446–2454. [Google Scholar] [CrossRef]
  5. Zhang, R.P.Q. Extrinsic Calibration of a Camera and Laser Range Finder (Improves Camera Calibration). In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Sendai, Japan, 28 September–2 October 2004; Volume 3, pp. 2301–2306. [Google Scholar]
  6. Geiger, A.; Moosmann, F.; Car, Ö.; Schuster, B. Automatic Camera and Range Sensor Calibration Using a Single Shot. In Proceedings of the 2012 IEEE International Conference on Robotics and Automation (ICRA), St. Paul, MN, USA, 14–18 May 2012; pp. 3936–3943. [Google Scholar]
  7. Zhou, L.; Deng, Z. A New Algorithm for Computing the Projection Matrix Between a LiDAR and a Camera Based on Line Correspondences. In Proceedings of the 2012 IV International Congress on Ultra Modern Telecommunications and Control Systems, St. Petersburg, Russia, 3–5 October 2012; pp. 436–441. [Google Scholar]
  8. Dhall, A.; Chelani, K.; Radhakrishnan, V.; Krishna, K.M. LiDAR-Camera Calibration Using 3D-3D Point Correspondences. arXiv 2017, arXiv:1705.09785. [Google Scholar]
  9. Guindel, C.; Beltrán, J.; Martín, D.; García, F. Automatic Extrinsic Calibration for LiDAR-Stereo Vehicle Sensor Setups. In Proceedings of the 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC), Yokohama, Japan, 16–19 October 2017; pp. 1–6. [Google Scholar]
  10. An, P.; Ding, J.; Quan, S.; Yang, J.; Yang, Y.; Liu, Q.; Ma, J. Survey of Extrinsic Calibration on LiDAR-Camera System for Intelligent Vehicle: Challenges, Approaches, and Trends. IEEE Trans. Intell. Transp. Syst. 2024, 25, 15342–15366. [Google Scholar] [CrossRef]
  11. Beltrán, J.; Guindel, C.; de la Escalera, A.; García, F. Automatic Extrinsic Calibration Method for LiDAR and Camera Sensor Setups. IEEE Trans. Intell. Transp. Syst. 2022, 23, 17677–17689. [Google Scholar] [CrossRef]
  12. TEXA S.p.A. ADAS Radar & Camera Calibration Kit. Available online: https://www.texa.com/products/adas-radar-camera-calibration-kit/#ADAS_RCCS3 (accessed on 4 January 2026).
  13. American Automobile Association. Cost of Advanced Driver Assistance Systems (ADAS) Repairs. Technical Report, American Automobile Association, Inc. 2023. Available online: https://newsroom.aaa.com/wp-content/uploads/2023/11/Report_Cost-of-ADAS-Repairs-FINAL-23.pdf (accessed on 4 January 2026).
  14. Levinson, J.; Thrun, S. Automatic Online Calibration of Cameras and Lasers. In Robotics: Science and Systems (RSS); KIT-MRT: Karlsruhe, Germany, 2013; Available online: https://www.roboticsproceedings.org/rss09/p29.pdf (accessed on 4 January 2026).
  15. Wang, Y.; Xing, S.; Can, C.; Li, R.; Hua, H.; Tian, K.; Mo, Z.; Gao, X.; Wu, K.; Zhou, S.; et al. Generative AI for Autonomous Driving: Frontiers and Opportunities. arXiv 2025, arXiv:2505.08854. [Google Scholar] [CrossRef]
  16. Bilal, H.; Rehman, A.; Aslam, M.S.; Ullah, I.; Chang, W.J.; Kumar, N.; Almuhaideb, A.M. Hybrid TrafficAI: A Generative AI Framework for Real-Time Traffic Simulation and Adaptive Behavior Modeling. IEEE Trans. Intell. Transp. Syst. 2025, 1–17. [Google Scholar] [CrossRef]
  17. Zhou, L.; Li, Z.; Kaess, M. Automatic Extrinsic Calibration of a Camera and a 3D LiDAR Using Line and Plane Correspondences. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 5562–5569. [Google Scholar] [CrossRef]
  18. Verma, S.; Berrio, J.S.; Worrall, S.; Nebot, E. Automatic Extrinsic Calibration Between a Camera and a 3D LiDAR Using 3D Point and Plane Correspondences. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; pp. 3906–3912. [Google Scholar]
  19. Jeong, S.; Lee, J.; Kim, J.; Cho, S. LiDAR-camera calibration based on the characteristics of LiDAR sensor. Sensors 2024, 24. [Google Scholar] [CrossRef]
  20. Gentilini, L.; Serio, P.; Donzella, V.; Pollini, L. A target-based multi-LiDAR multi-camera extrinsic calibration system. arXiv 2025. [Google Scholar] [CrossRef]
  21. Inghilterra, G.; Arrigoni, S.; Braghin, F.; Cheli, F. Firefly Algorithm-Based Nonlinear MPC Trajectory Planner for Autonomous Driving. In Proceedings of the 2018 International Conference of Electrical and Electronic Technologies for Automotive (EETA), Milan, Italy, 9–11 July 2018; pp. 1–6. [Google Scholar] [CrossRef]
  22. Walters, R.; Tan, L.; Ferrari, F. A Robust Framework for Online Extrinsic Calibration of Multi-Sensor Vehicle Platforms. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Kyoto, Japan, 23–27 October 2022; pp. 8432–8439. [Google Scholar] [CrossRef]
  23. Laporte, G.; Osman, I.H. Routing Problems: A Bibliography. Ann. Oper. Res. 1995, 61, 227–262. [Google Scholar] [CrossRef]
  24. Igel, C. No Free Lunch Theorems: Limitations and Perspectives of Metaheuristics. In Theory and Principled Methods for the Design of Metaheuristics; Springer: Berlin/Heidelberg, Germany, 2014; pp. 1–23. [Google Scholar]
  25. Lazar, A.; Reynolds, R.G. Heuristic Knowledge Discovery for Archaeological Data Using Cultural Algorithms and Rough Sets. In Heuristics and Optimization for Knowledge Discovery; Abbass, H.A., Newton, C.S., Sarker, R.A., Eds.; Idea Group Publishing: Hershey, PA, USA, 2002; pp. 263–278. [Google Scholar]
  26. Yang, X.S. Nature-Inspired Optimization Algorithms; Academic Press: Boca Raton, FL, USA, 2020. [Google Scholar]
  27. Panwar, K.; Deep, K. Discrete Grey Wolf Optimizer for Symmetric Travelling Salesman Problem. Appl. Soft Comput. 2021, 105, 298. [Google Scholar] [CrossRef]
  28. Molina, D.; Poyatos, J.; Ser, J.D.; García, S.; Hussain, A.; Herrera, F. Comprehensive Taxonomies of Nature- and Bio-Inspired Optimization: Inspiration Versus Algorithmic Behavior, Critical Analysis Recommendations. Cogn. Comput. 2020, 12, 897–939. [Google Scholar] [CrossRef]
  29. Rajwar, K.; Deep, K.; Das, S. An Exhaustive Review of the Metaheuristic Algorithms for Search and Optimization: Taxonomy, Applications, and Open Challenges. Artif. Intell. Rev. 2023, 56, 13187–13257. [Google Scholar] [CrossRef] [PubMed]
  30. Holland, J.H. Adaptation in Natural and Artificial Systems: An Introductory Analysis with Application to Biology, Control, and Artificial Intelligence; University of Michigan Press: Ann Arbor, MI, USA, 1975; pp. 439–444. [Google Scholar]
  31. Goldberg, D.E.; Deb, K. A Comparative Analysis of Selection Schemes Used in Genetic Algorithms. In Foundations of Genetic Algorithms; Rawlins, G.J.E., Ed.; Morgan Kaufmann: Los Altos, CA, USA, 1991; pp. 69–93. [Google Scholar]
  32. Arrigoni, S.; Braghin, F.; Cheli, F. MPC Trajectory Planner for Autonomous Driving Solved by Genetic Algorithm Technique. Veh. Syst. Dyn. 2022, 60, 4118–4143. [Google Scholar] [CrossRef]
  33. Vanneschi, L.; Silva, S. Genetic Algorithms. In Lectures on Intelligent Systems; Natural Computing Series; Springer: Cham, Germany, 2023. [Google Scholar] [CrossRef]
  34. De Jong, K. The Analysis and Behaviour of a Class of Genetic Adaptive Systems. Ph.D. Thesis, University of Michigan, Ann Arbor, MI, USA, 1975. [Google Scholar]
  35. Kennedy, J.; Eberhart, R. Particle Swarm Optimization. In Proceedings of the IEEE International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; pp. 1942–1948. [Google Scholar]
  36. Arrigoni, S.; Trabalzini, E.; Bersani, M.; Braghin, F.; Cheli, F. Nonlinear MPC Motion Planner for Autonomous Vehicles Based on Accelerated Particle Swarm Optimization Algorithm. In Proceedings of the 2019 AEIT International Conference of Electrical and Electronic Technologies for Automotive (AEIT AUTOMOTIVE), Turin, Italy, 2–4 July 2019; pp. 1–6. [Google Scholar]
  37. Kirkpatrick, S.; Gelatt, C.D.; Vecchi, M.P. Optimization by Simulated Annealing. Science 1983, 220, 671–680. [Google Scholar] [CrossRef] [PubMed]
  38. Donnett, J.G. Simulated Annealing and Code Partitioning for Distributed Multimicroprocessors; Technical report; Department of Computer and Information Science, Queen’s University: Kingston, ON, Canada, 1987. [Google Scholar]
  39. Woo, A.; Fidan, B.; Melek, W.W. Localization for Autonomous Driving. In Handbook of Position Location: Theory, Practice, and Advances, 2nd ed.; Zekavat, S., Buehrer, R.M., Eds.; Wiley-IEEE Press: Hoboken, NJ, USA, 2019; pp. 1051–1087. [Google Scholar]
  40. Roman, I.; Ceberio, J.; Mendiburu, A.; Lozano, J.A. Bayesian Optimization for Parameter Tuning in Evolutionary Algorithms. In Proceedings of the 2016 IEEE Congress on Evolutionary Computation (CEC), Vancouver, BC, Canada, 24–29 July 2016; pp. 4839–4845. [Google Scholar]
  41. Arrigoni, S.; Mentasti, S.; Cheli, F.; Matteucci, M.; Braghin, F. Design of a Prototypical Platform for Autonomous and Connected Vehicles. In Proceedings of the 2021 AEIT International Conference on Electrical and Electronic Technologies for Automotive (AEIT AUTOMOTIVE), Torino, Italy, 17–19 November 2021; pp. 1–6. [Google Scholar] [CrossRef]
Figure 1. Genetic Algorithm workflow.
Figure 1. Genetic Algorithm workflow.
Applsci 16 01498 g001
Figure 2. SA Algorithm workflow.
Figure 2. SA Algorithm workflow.
Applsci 16 01498 g002
Figure 3. Geometrical features extracted from the LiDAR reference frame.
Figure 3. Geometrical features extracted from the LiDAR reference frame.
Applsci 16 01498 g003
Figure 4. Illustration of the angular error between the checkerboard normal vectors. The black and white checkerboard represents the plane observed in the camera frame, while the orange checkerboard corresponds to the same plane represented in the LiDAR frame. The red arrows denote the respective normal vectors in each coordinate frame, and the angular discrepancy between them defines the alignment (normal-to-normal) error.
Figure 4. Illustration of the angular error between the checkerboard normal vectors. The black and white checkerboard represents the plane observed in the camera frame, while the orange checkerboard corresponds to the same plane represented in the LiDAR frame. The red arrows denote the respective normal vectors in each coordinate frame, and the angular discrepancy between them defines the alignment (normal-to-normal) error.
Applsci 16 01498 g004
Figure 5. Illustration of the diagonal-direction angular error. The black and white checkerboard represents the plane in the camera frame, and the orange checkerboard represents the same plane in the LiDAR frame. The red arrows indicate the center-to-corner directional vectors defined for each checkerboard. The angular deviation between corresponding vectors quantifies the diagonal-alignment error.
Figure 5. Illustration of the diagonal-direction angular error. The black and white checkerboard represents the plane in the camera frame, and the orange checkerboard represents the same plane in the LiDAR frame. The red arrows indicate the center-to-corner directional vectors defined for each checkerboard. The angular deviation between corresponding vectors quantifies the diagonal-alignment error.
Applsci 16 01498 g005
Figure 6. Illustration of the perpendicularity error. The orange checkerboard represents the plane in the LiDAR frame, with its associated normal vector shown in red. The black and white checkerboard corresponds to the plane in the camera frame, where the red arrow denotes a center-to-corner directional vector. The perpendicularity error is defined as the angular deviation from the expected orthogonality between the transformed normal vector and the corresponding directional vector.
Figure 6. Illustration of the perpendicularity error. The orange checkerboard represents the plane in the LiDAR frame, with its associated normal vector shown in red. The black and white checkerboard corresponds to the plane in the camera frame, where the red arrow denotes a center-to-corner directional vector. The perpendicularity error is defined as the angular deviation from the expected orthogonality between the transformed normal vector and the corresponding directional vector.
Applsci 16 01498 g006
Figure 7. Visualization of the alignment errors used to refine the rigid transformation between LiDAR and Camera frames. The black and white checkerboard is represented in the Camera frame, the orange checkerboard in the LiDAR frame, and green points indicate the detected features. (a) shows the Euclidean distances between corresponding centers, defining the center alignment error. (b) shows the Euclidean distances between corresponding corners, defining the corner alignment error.
Figure 7. Visualization of the alignment errors used to refine the rigid transformation between LiDAR and Camera frames. The black and white checkerboard is represented in the Camera frame, the orange checkerboard in the LiDAR frame, and green points indicate the detected features. (a) shows the Euclidean distances between corresponding centers, defining the center alignment error. (b) shows the Euclidean distances between corresponding corners, defining the corner alignment error.
Applsci 16 01498 g007
Figure 8. GA fitness and execution time vs. lower and upper bounds expressed in degrees.
Figure 8. GA fitness and execution time vs. lower and upper bounds expressed in degrees.
Applsci 16 01498 g008
Figure 9. GAfitness trend with respect to parameters during phase one.
Figure 9. GAfitness trend with respect to parameters during phase one.
Applsci 16 01498 g009
Figure 10. GA execution time with respect to parameters during phase one.
Figure 10. GA execution time with respect to parameters during phase one.
Applsci 16 01498 g010
Figure 11. Trend of the average cost with respect to the bounds during phase two: ub 1 , lb 1 in degrees, and ub 2 , lb 2 in meters.
Figure 11. Trend of the average cost with respect to the bounds during phase two: ub 1 , lb 1 in degrees, and ub 2 , lb 2 in meters.
Applsci 16 01498 g011
Figure 12. Cost trend of PSO with respect to parameters for phase two.
Figure 12. Cost trend of PSO with respect to parameters for phase two.
Applsci 16 01498 g012
Figure 13. Average execution time trend of PSO with respect to parameters for phase two.
Figure 13. Average execution time trend of PSO with respect to parameters for phase two.
Applsci 16 01498 g013
Figure 14. Setup system mounted on the experimental vehicle.
Figure 14. Setup system mounted on the experimental vehicle.
Applsci 16 01498 g014
Figure 15. Validation setup: the sensor system is shown on the left, while the red spherical target is shown on the right, with its center selected for computing the azimuth and elevation error deltas.
Figure 15. Validation setup: the sensor system is shown on the left, while the red spherical target is shown on the right, with its center selected for computing the azimuth and elevation error deltas.
Applsci 16 01498 g015
Figure 16. Comparison of optimization performance for single-pose objective functions: (a) Phase 1 GA-based optimization; (b) Phase 2 PSO-based optimization; (c) Optimization through Gradient Descent Method.
Figure 16. Comparison of optimization performance for single-pose objective functions: (a) Phase 1 GA-based optimization; (b) Phase 2 PSO-based optimization; (c) Optimization through Gradient Descent Method.
Applsci 16 01498 g016
Figure 17. Visualization of the point cloud after transformation into the camera coordinate frame using the pose estimates obtained through (a) Gradient Descent optimization; (b) the proposed Hybrid GA–PSO optimization method.
Figure 17. Visualization of the point cloud after transformation into the camera coordinate frame using the pose estimates obtained through (a) Gradient Descent optimization; (b) the proposed Hybrid GA–PSO optimization method.
Applsci 16 01498 g017
Figure 18. Percentage increase in calibration errors as the target distance grows.
Figure 18. Percentage increase in calibration errors as the target distance grows.
Applsci 16 01498 g018
Table 1. Sensor setup used for the extrinsic calibration.
Table 1. Sensor setup used for the extrinsic calibration.
Sensor TypeSensor Name
LiDARApplsci 16 01498 i001
OS1 LiDAR
Omnidirectional cameraApplsci 16 01498 i002
Ladybug5+
Stereo cameraApplsci 16 01498 i003
ZED X
Table 2. Performance of GA, PSO, and SA after Bayesian optimization for phase one (one pose). GA was selected as it offered the best trade-off between Avg. Cost and Avg. Time. Only GA parameter trends are reported in the following figures.
Table 2. Performance of GA, PSO, and SA after Bayesian optimization for phase one (one pose). GA was selected as it offered the best trade-off between Avg. Cost and Avg. Time. Only GA parameter trends are reported in the following figures.
AlgorithmAvg. Time (s)Avg. CostBest Cost
GA (Roulette Selection)0.0640.00330.0024
GA (Tournament Selection)0.0800.00350.0024
PSO0.0830.00350.0024
SA1.10.00440.0024
Table 3. Values obtained after Bayesian optimization for phase two, with only one pose.
Table 3. Values obtained after Bayesian optimization for phase two, with only one pose.
AlgorithmAvg. Time (s)Avg. CostBest Cost
GA (Roulette Wheel Selection)0.08350.000650.00044
GA (Tournament Selection)0.10.000650.00044
PSO0.220.000510.00044
SA9.030.000470.00044
Table 4. Azimuth ( Δ θ ) and elevation ( Δ ψ ) differences for LiDAR-LADYBUG and LiDAR-ZED X transformations. The target was placed 4.5–5.5 m from the board center.
Table 4. Azimuth ( Δ θ ) and elevation ( Δ ψ ) differences for LiDAR-LADYBUG and LiDAR-ZED X transformations. The target was placed 4.5–5.5 m from the board center.
TransformationPoses Δ θ (°) Δ ψ (°)
LiDAR-to-Ladybug10.01520.11
LiDAR-to-ZED X10.00270.013
Table 5. Performance evaluation of Zhou’s method.
Table 5. Performance evaluation of Zhou’s method.
Method Δ θ (°) Δ ψ (°)Time (s)
Zhou0.99850.91383.1
Table 6. Results with the Gradient Descent Method.
Table 6. Results with the Gradient Descent Method.
Delta Azimuth (°)Delta Elevation (°)
0.691013.38
Table 7. Normalized percentage increment for azimuth and elevation changes across distances.
Table 7. Normalized percentage increment for azimuth and elevation changes across distances.
Range (m)ΔAzim (%)ΔElev (%)
4.53–6.718.2012.50
6.71–9.4988.739.77
9.49–11.783.0677.73
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Arrigoni, S.; D’Amato, F.; Cholakkal, H.H. Automatic Multi-Sensor Calibration for Autonomous Vehicles: A Rapid Approach to LiDAR and Camera Data Fusion. Appl. Sci. 2026, 16, 1498. https://doi.org/10.3390/app16031498

AMA Style

Arrigoni S, D’Amato F, Cholakkal HH. Automatic Multi-Sensor Calibration for Autonomous Vehicles: A Rapid Approach to LiDAR and Camera Data Fusion. Applied Sciences. 2026; 16(3):1498. https://doi.org/10.3390/app16031498

Chicago/Turabian Style

Arrigoni, Stefano, Francesca D’Amato, and Hafeez Husain Cholakkal. 2026. "Automatic Multi-Sensor Calibration for Autonomous Vehicles: A Rapid Approach to LiDAR and Camera Data Fusion" Applied Sciences 16, no. 3: 1498. https://doi.org/10.3390/app16031498

APA Style

Arrigoni, S., D’Amato, F., & Cholakkal, H. H. (2026). Automatic Multi-Sensor Calibration for Autonomous Vehicles: A Rapid Approach to LiDAR and Camera Data Fusion. Applied Sciences, 16(3), 1498. https://doi.org/10.3390/app16031498

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop