You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

2 December 2025

Multi-Calib: A Scalable LiDAR–Camera Calibration Network for Variable Sensor Configurations

,
,
,
and
1
School of Mechanical Engineering, Beijing Institute of Technology, Beijing 100081, China
2
China North Vehicle Research Institute, Beijing 100072, China
*
Author to whom correspondence should be addressed.
Sensors2025, 25(23), 7321;https://doi.org/10.3390/s25237321 
(registering DOI)
This article belongs to the Section Vehicular Sensing

Abstract

Traditional calibration methods rely on precise targets and frequent manual intervention, making them time-consuming and unsuitable for large-scale deployment. Existing learning-based approaches, while automating the process, are typically limited to single LiDAR–camera pairs, resulting in poor scalability and high computational overhead. To address these limitations, we propose a lightweight calibration network with flexibility in the number of sensor pairs, making it capable of jointly calibrating multiple cameras and LiDARs in a single forward pass. Our method employs a frozen pre-trained Swin Transformer as a shared backbone to extract unified features from both RGB images and corresponding depth maps. Additionally, we introduce a cross-modal channel-wise attention module to enhance key feature alignment and suppress irrelevant noise. Moreover, to handle variations in viewpoint, we design a modular calibration head that independently estimates the extrinsics for each LiDAR–camera pair. Through large-scale experiments on the nuScenes dataset, we show that our model, requiring merely 78.79 M parameters, attains a mean translation error of 2.651 cm and a rotation error of 0 . 246 , achieving comparable performance to existing methods while significantly reducing the computational cost.

1. Introduction

LiDAR and cameras constitute the primary sensing modalities in intelligent perception, as their capabilities complement one another. LiDAR point clouds provide accurate spatial structure and precise 3D position, while camera images offer dense appearance information and rich semantic contexts. Effectively combining these two modalities significantly boosts performance for downstream applications, including object detection [1,2,3] and semantic segmentation [4,5,6]. However, the benefits of multi-modal fusion hinge critically on the quality of extrinsic calibration. A precise calibration matrix is essential for accurate spatial alignment between the 3D point clouds and the 2D images, enabling reliable fusion at both data and feature levels.
The objective of calibration is to estimate the transformation matrix that aligns the LiDAR and camera coordinate systems. Existing approaches are generally classified into two groups, offline calibration [7,8,9,10,11,12] and online calibration [13,14,15,16,17,18], as illustrated in Figure 1. Offline strategies typically rely on dedicated calibration targets and manual feature selection, which is labor-intensive and unable to automatically compensate for perturbations during operation. By contrast, online methods often substitute the handcrafted feature extraction step with learning-based techniques that identify salient structures directly from multi-modal sensor data. Consequently, online methods are proposed to conduct calibration in an automated manner. With the rapid progress of deep learning techniques, such as CNNs, numerous studies [13,14,18,19,20,21] have proposed supervised data-driven approaches to handle LiDAR–camera calibration task. Specifically, recent approaches treat calibration as a deep regression problem, aiming to estimate the 6-DoF (Degree of Freedom) extrinsic parameters between LiDAR and the camera using learning-based methods. Most of these methods adopt the general network structure illustrated in Figure 1a. The input typically consists of the original image paired with a misaligned depth map, and the pipeline is organized into two separate streams—one for image features and another for depth features—followed by feature extraction, correspondence computation, and parameter regression. Some of them also take the architecture shown in Figure 1b to design the network model, which projects the raw point cloud into related images for input generation; then they are extracted by a uniform CNN backbone to obtain uniform feature maps. Due to CNN’s strong fitting capacity, their performance has been facilitated significantly.The majority of methods perform calibration in a single-shot manner [22,23,24], whereas others employ multiple networks [13,14,19,20,21] for iterative refinement.
Figure 1. Comparison of technical pipelines: existing approaches vs. ours. (a) Traditional techniques often rely on a dual-branch framework that leverages independent backbones to separately encode features from images and depth inputs. (b) Lightweight approaches fuse image and depth features via separate convolutions, and then transform them into a unified representation for a shared CNN. (c) Our method merges the strengths of both approaches by concatenating image and depth data in batches, extracting features via a unified Transformer backbone, and enhancing them with a cross-modal channel-wise attention module.The blue pipeline corresponds to the image processing module, whereas the yellow pipeline corresponds to the depth module.
Figure 2 illustrates the sensor configurations of two commonly used open-source datasets, KITTI [25] and nuScenes [26].
Figure 2. Comparison of multiple sensor configurations in mainstream autonomous driving datasets. (a) Sensor setups in the KITTI Dataset: single LiDAR sensor and four cameras. (b) Sensor configuration in the nuScenes dataset, featuring a single LiDAR sensor and six surround-view cameras.
Although state-of-the-art deep learning-based methods have achieved remarkable progress, we argue that they still suffer from two fundamental limitations: (1) One-to-one calibration: In practical deployment scenarios—such as those illustrated in Figure 2a,b—autonomous platforms typically employ multi-camera and multi-LiDAR configurations. However, most existing calibration methods are designed for single LiDAR–camera pairs. Extending these approaches to multi-view settings (e.g., front, rear, left, and right) is non-trivial due to significant viewpoint variations, leading to poor generalization, increased calibration errors, and higher memory consumption. (2) Inefficiencies in the calibration architecture: Most traditional methods separate calibration into dual-stream networks, enabling modality-specific feature extraction but incurring high computational and memory costs, especially during iterative inference. In contrast, recent lightweight approaches unify inputs by projecting 3D point clouds as 2D depth channels, reducing overhead but struggling to fully exploit transformer-based backbones. Since transformers require large-scale data and are hard to train on limited samples, these methods often yield poor convergence and degraded calibration performance.
To overcome these two challenges, we introduce the following three-fold enhancement strategy:
  • First, we adopt a large-scale pre-trained Swin Transformer as a unified backbone for joint feature extraction from both RGB (Red Green Blue) images and depth maps. Leveraging pre-trained weights significantly reduces the training cost and improves memory efficiency during iterative inference.
  • Second, we design a cross-modal channel-wise attention mechanism that facilitates multi-scale feature integration between camera and LiDAR modalities. This mutually guided attention module enhances salient and correspondingly aligned features via context-aware reinforcement.
  • Finally, we implement a configurable calibration head architecture that accommodates varying numbers of cameras. Each head independently predicts the extrinsics for its corresponding monocular image–depth map pair. This modular design avoids cross-view feature entanglement, effectively mitigating the artifacts caused by shared-head architectures and preserving viewpoint-specific calibration fidelity.
Figure 3 provides an overview of the full calibration workflow.
Figure 3. Our method estimates LiDAR–camera extrinsics as follows. Multi-view images from calibrated cameras and projected depth maps from miscalibrated LiDAR are concatenated and processed by a unified Swin Transformer backbone for feature extraction (The asterisk (*) denotes element-wise multiplication with the same dimension.). Multi-scale features are obtained through two trainable FPNs (Feature Pyramid NetworkS) and fused with a cross-modal channel-wise attention module. The fused features are passed to calibration heads that regress 6-DoF transformations T p r e d , which represent the deviation between the initial extrinsics T i n i t and the ground truth T g t .
Results obtained on the nuScenes dataset [26] indicate that the proposed method performs slightly worse than state-of-the-art approaches in calibration accuracy while achieving significant model size reduction and supporting flexible calibration configurations beyond the conventional one-to-one setting. The main contributions of this work are summarized as follows:
  • To address multi-sensor calibration, we present Multi-Calib, an online deep learning model with explicit depth supervision capable of aligning multiple LiDAR sensors and cameras.
  • We introduce a cross-modal channel-wise attention module for mutually guided fusion between camera and LiDAR features, achieving highly efficient integration.
  • A novel loss function is developed based on the intrinsic correlation between translation and rotation errors, effectively improving LiDAR–camera registration accuracy.
The structure and content of this article are summarized as follows: Section 2 primarily elaborates on relevant research works in the field. Section 3 systematically presents the network architecture and its underlying principles. Section 4 comprehensively documents experimental results, implementation details, and ablation studies. Section 5 summarizes the main findings and outlines directions for future research.

3. Method

This work focuses on a mobile robotic platform equipped with multiple cameras and LiDAR sensors under the following assumptions: (i) Each LiDAR–camera pair has an overlapping field of view (FOV). (ii) Coarse initial extrinsics T i n i t and camera intrinsics K are available.

3.1. Problem Formulation

Using multi-sensor inputs, the objective of the online calibration task is to refine the coarse initial estimate T i n i t i and recover the accurate extrinsics T g t i . Following previous works [14,19,20], to generate sufficient training data, a random transform matrix, T r a n d , is applied to T i n i t i to obtain the ground truth T g t i . The miscalibrated depth map of the point cloud is then generated as
Z c i u i v i 1 = K i ( T i n i t i P ) = K i ( R i n i t i P + t i n i t i )
T g t i = T i n i t i × T r a n d = R g t i t g t i 0 1
Here, i 0 , 5 indexes the LiDAR–camera pairs, P = ( X , Y , Z , 1 ) represents a 3D LiDAR point, and x = ( u i , v i ) denotes its corresponding 2D projection. K i stands for the camera intrinsic matrix, and Z c i is an arbitrary depth-related scale factor. The initial transformation matrix T i n i t i comprises a rotation component, R i n i t i , and a translation vector, t i n i t i . The network receives a randomly miscalibrated depth map together with a resized RGB image as input, and the goal of online calibration is to estimate T r a n d to correct the extrinsic misalignment.

3.2. Network Architecture

The technical workflow of the proposed Multi-Calib network for LiDAR–camera calibration is illustrated in Figure 3, and a more detailed network subfigure is provided in Figure 4.
Figure 4. Multiple LiDAR–camera pairs are batched and fed into a frozen pretrained Swin Transformer backbone (shown in green). The resulting multi-level features (Feats1, Feats2, and Feats3) are forwarded to two trainable FPN necks—one for image features (blue) and one for depth features (yellow). The output features (Feats4 and Feats5) are then refined using a cross-modal channel-wise Attention block(shown in purple) to generate importance-weighted features, which are subsequently concatenated into Feats6. Finally, Feats6 is passed to the calibration heads to produce the predicted translation and rotation parameters.
The system takes multiple images from different onboard cameras (top) and the corresponding depth maps projected from LiDAR point clouds (bottom) as input. To enhance robustness, the images are randomly scaled, while the point clouds undergo random affine transformations and shuffling. The resulting depth maps are further refined using a local window filter. These inputs are batch-concatenated and fed into a unified Transformer backbone for feature extraction using both modalities—unlike traditional dual-stream architectures that treat each modality separately. The extracted features are then processed by FPNs to obtain multi-scale representations, which are subsequently fused via a cross-modal channel-wise attention module. Finally, a configurable calibration head processes the fused features to output the 6-DoF extrinsic parameters for each LiDAR–camera pair in one forward pass.

3.3. Loss Functions

For a set of N RGB images and their corresponding initial depth maps, two loss terms are applied during training: a regression loss, L T , and a point cloud distance loss, L P , as formulated in [20].
L = ( 1 λ P ) L T + λ P L P
where λ P denotes the point cloud distance loss weight.

Regression Loss

For the predicted translation vector t pred , we apply the Huber loss function. As for the rotation loss L q , we adopt the angular distance to quantify the discrepancy between them, as defined below:
L R = D i s t a n c e a ( q g t , q p r e d )
where q g t denotes the ground-truth quaternion, q p r e d is the predicted quaternion, and D i s t a n c e a represents the angular distance between them.
However, it is important to emphasize that rotation and translation are not independent components; they are intrinsically coupled within the rigid-body transformation matrix. Ignoring this interdependence and optimizing their losses separately may lead to suboptimal calibration performance. To better capture this coupling and enhance the overall calibration accuracy, we introduce a matrix-based loss function, L M , which directly penalizes the discrepancy between the predicted and ground-truth transform matrices, as defined below:
L M = Σ i = 1 5 H u b e r L o s s ( T p r e d i × i n v ( T r a n d i ) , E )
where T p r e d i denotes the predicted transform matrix, which should ideally be the identity matrix E when multiplied by the inverse of the random transformation. The total regression loss L T is defined as the combination of the translation, rotation, and matrix losses:
L T = λ t L t + λ q L R + λ m L M
where λ t , λ q , and λ m denote the respective weights for the translation, rotation, and matrix loss terms.

4. Experiments

The proposed calibration approach is evaluated on the nuScenes dataset. This section describes dataset preparation and implementation details and presents an analysis of the experimental results.

4.1. Dataset Preparation

Dataset Generation: We validate our proposed network using the full nuScenes dataset [26], which contains 1000 multi-modal driving sequences captured with six cameras. We use the officially designated splits of the dataset: 700 sequences for training, 150 for validation, and 150 for testing. Each scene provides six RGB images from the front, front-left, front-right, rear, rear-left, and rear-right cameras, together with LiDAR point clouds obtained from the roof-mounted LiDAR sensor.
To simulate real-world calibration errors, we follow the protocol in [20] and apply a random affine transformation to the LiDAR point clouds. The transformation parameters are sampled from uniform distributions over translation ranges [ x , x ] with x 1.5 m, 1.0 m, 0.5 m, 0.25 m, 0.12 m and rotation ranges [ y , y ] with y 20 , 15 , 10 , 5 , 2 . The whole process is depicted in Figure 5. The range is selected based on our test results.
Figure 5. Dataset preparation process.
Evaluation Metrics: The experimental results are assessed in terms of both rotation and translation of the calibration parameters. The translation vector is quantified using the Euclidean distance, and its absolute error is computed as follows:
E t = | | t p r e d t g t | | 2
Here, | | · | | 2 represents the L2-norm of a vector. In addition, the absolute errors of the translation vector are evaluated separately along the X, Y, and Z axes.

4.2. Implementation Details

Training Details: The original image resolution is 1600 × 900 , which is resized to 704 × 256 as the input to our model. The LiDAR sensor used in the nuScenes dataset is the Velodyne HDL-32E sensor (Velodyne LiDAR, San Jose, CA, USA), which features 32 laser channels and a 360 horizontal field of view. Each LiDAR frame contains up to 55,000 points. Depth maps are generated using the camera intrinsics and the original LiDAR-to-camera extrinsics, and they are subsequently resized to 704 × 256 .
We train the proposed network for 60 epochs using four NVIDIA RTX 4090 GPUs (NVIDIA Corporation, Santa Clara, CA, USA) with a batch size of 10. The AdamW optimizer is employed with an initial learning rate of 1 × 10 3 . During the first 300 iterations, the learning rate is linearly warmed up using a small multiplicative factor. Then, it is halved every 10 epochs for the subsequent 50 epochs. Finally, a cosine annealing schedule is applied during the last 10 epochs.
To ensure sufficient point cloud density, we limit the point cloud range from [ 54.0 , 54.0 , 5.0 ] to [ 54.0 , 54.0 , 5.0 ] . To maintain balanced optimization across the sub-tasks and ensure equal contribution from each objective during training, all loss function weights are set to 1.0 , except for the point cloud distance loss, which is weighted at 0.5 to achieve equilibrium between the regression and point cloud distance components. For a fair comparison with LCCNet, the iterative step size is kept at five.

4.3. Comparison to State-of-the-Art Models

Baselines: Two different methods were selected as baselines, and their experimental settings were maintained for a fair comparison.
Comparison of basic parameters: Table 1 presents the model sizes of CMRNet [39], LCCNet [20], SCNet [40], and our proposed Multi-Calib model. To the best of our knowledge, no existing published and open-source methods are capable of calibrating multiple cameras simultaneously within a unified framework. Therefore, we compare our approach with representative single-camera calibration models, which remain the strongest competitors in the current literature. It can be seen that
Table 1. Comparison of model size, iterations, and camera num.
  • CMRNet uses the fewest iterations, but its three separate models do not share parameters, resulting in a total model size of 106.20 M—which is significantly larger than ours (78.79 M). In addition, our model is scalable to any number of LiDAR–camera pairs, which exceeds the capacity of all of them.
  • Unlike LCCNet, which requires training five separate whole models for iterative refinement, our method adopts a unified frozen Swin Transformer as a shared backbone across all iterations, significantly reducing model size and the training cost. As shown in Table 1, the total size of our model is only 78.79 M, with just 5.99 M parameters requiring training. In contrast, LCCNet trains five independent models (66.75 M each), resulting in a total size exceeding 333.75 M.
Results: Table 2 presents the comparison with the baseline methods. Multi-Calib achieves performance comparable to all baselines across the tested environments, indicating that the combination of frozen shared backbones and the proposed iterative training strategy is effective, without relying on additional sensor data or multi-frame inputs.
Table 2. Comparison of calibration methods under a miscalibration range of [−1.5 m, 1.5 m]/[−20°, 20°] in nuScenes.
Figure 6 illustrates the predicted depth maps and the results of the iterative calibration, providing an intuitive demonstration of Multi-Calib’s effectiveness.
Figure 6. Visualization of the progressive refinement process from the initial input to the final ground truth. (a) Iteration Results: Each subfigure shows intermediate results. Arrows indicate forward refinement stages. (b) Visualization of predicted extrinsics after 4 iterations of network inference. (c) Ground Truth: The red boxes highlight the most noticeable differences.

4.4. Ablation Studies

4.4.1. Backbone Selection

To validate our backbone network strategy, we first selected two representative multi-sensor fusion perception tasks: 3D object detection and BEV (Bird’s Eye View) map segmentation. For both tasks, we adopted BEVFusion [2], a multi-task perception model that integrates surround-view cameras and LiDAR. To simulate extrinsic errors caused by long-term platform operation and vibrations, we injected average noise into the extrinsic matrices, following the design methodology of prior work. The corresponding perception results are summarized in Table 3, where the NDS (nuScenes Detection Score) and mAP (mean Average Precision) are reported for 3D object detection, and the mIoU (Mean Intersection over Union) is used as the evaluation metric for BEV map segmentation.
Table 3. Performance comparison of 3D object detection and BEV map segmentation under different noise levels.
In the setting without noise injection (w/o Noise), which corresponds to using ground-truth extrinsics, BEVFusion achieves an NDS of 71.15 and an mAP of 68.41 on the 3D object detection task. For BEV map segmentation, the best performance is observed in the Drivable Area, with an mIoU as high as 83.04.
In the first experiment, we injected average noise values of ±0.12 m in translation and ±1° in rotation. Under this setting, both the NDS and mAP exhibited a slight decrease of 2.08 and 2.68, respectively. For the BEV map segmentation task, degradation was relatively minor, with the most significant drop observed in the Carpark Area (↓ 7.05).
In the second experiment, when the noise level was increased to ±0.5 m/ ± 10 , BEVFusion suffered a sharp performance decline. On the 3D object detection task, the NDS dropped to 49.69 (one-third lower), while the mAP decreased by nearly half (↓ 31.40). The BEV map segmentation task also experienced substantial degradation, with the Carpark Area decreasing by 26.63 and Divider by 22.91, resulting in an average drop of 17.03.
In the third experiment, with noise increased to ±1.5 m/ ± 20 , the 3D object detection task almost completely lost detection capability, with the mAP reduced to 5.64 and the NDS to 26.44. Meanwhile, BEV map segmentation showed an average drop of 32.85, where the Carpark Area decreased the most (↓ 60.13), while the Drivable Area decreased the least (↓ 19.53).
These results clearly demonstrate that extrinsic errors have a critical impact on multi-sensor fusion perception. As the error magnitude increases, perception accuracy consistently degrades, exhibiting a gradual decline at lower noise levels and a drastic collapse at higher noise levels. Therefore, we argue that the primary goal of online calibration is to constrain extrinsic errors within ±0.12 m/ ± 1 while maintaining model efficiency.
To further investigate the impact of backbone networks on LiDAR–camera calibration, we designed a second experiment comparing representative backbone architectures commonly used in calibration models. The evaluation was divided into two groups: (1) fully frozen backbones, where all backbone parameters are fixed (this setting is advantageous for multi-round iterative calibration since it significantly reduces memory consumption and allows for larger batch sizes, thereby improving training efficiency), and (2) non-frozen backbones, including fully trainable networks and a partially frozen Swin Transformer (with only the final output layer unfrozen).
The experimental setup was as follows: translation noise: ±1.5 m, rotation noise: ± 20 , and training conducted on four GTX 4090 GPUs(NVIDIA Corporation, Santa Clara, CA, USA). According to the available GPU memory, the batch size was set to 24 for frozen backbones, while for non-frozen backbones, it was 4 for ResNet50, 3 for ResNet101, and 2 for the Swin Transformer. Training was performed for 60 epochs, with an initial learning rate of 0.001, which was halved every 5 epochs. The final calibration results are reported in Table 4.
Table 4. Comparison of backbone freezing strategies under a miscalibration range of [−1.5 m, 1.5 m]/[−20°, 20°] in nuScenes.
From the first-round calibration results, the partially frozen Swin Transformer achieves the best performance, with an average translation error of 19.41 cm and an average rotation error of 2 . 18 . The second-best results are obtained by the trainable ResNet50 model and the trainable Swin Transformer, while the fully frozen ResNet101 and ResNet50 models yield the worst performance.
From the perspective of iterative calibration, the frozen Swin Transformer demonstrates the highest efficiency, requiring only 78.79 M parameters over five iterations, thereby minimizing memory usage and training time. Although the partially frozen Swin Transformer achieves the best accuracy, it requires 130 M parameters, which represents a 64% increase compared to the fully frozen variant. The trainable ResNet50 model ranks second in terms of accuracy but demands 148 M parameters, an 88% increase, while providing only marginal improvements of 1.01 cm in translation error and 0 . 08 in rotation error.
We attribute these results to the inherent advantages of the Swin Transformer’s global attention mechanism for calibration tasks. Since most of its parameters are concentrated in the latter layers, fully training the backbone is suboptimal given the limited size of the nuScenes dataset. In contrast, partial freezing balances stability and adaptability, thereby achieving the best accuracy. However, considering the significant growth in parameter size with only limited accuracy gains, we conclude that adopting a fully frozen Swin Transformer backbone represents the most balanced and efficient choice.

4.4.2. Module Ablation

To evaluate the effectiveness of each proposed module, we conduct an ablation study at the first calibration iteration under a challenging miscalibration range of [ 1.5 m , 1.5 m ] / [ 20 , 20 ] , as shown in Table 5. Starting from a baseline model without any enhancement modules (FPN, CCFusion, or Matrix Loss), we progressively introduce each component to assess its individual and joint contributions.
Table 5. Ablation study on module effectiveness at the first iteration under a miscalibration range of [−1.5 m, 1.5 m]/[−20°, 20°] in nuScenes.
Adding the FPN module alone slightly improves rotation accuracy (e.g., R o t X decreases by 0 . 23 ) and reduces translation errors along the Y-axis by 5.22 cm , though some metrics (e.g., T X ) increase slightly, indicating limited benefits when used in isolation.
Introducing the CCFusion module significantly enhances performance across all metrics. Translation errors along all three axes are notably reduced (e.g., T X by 9.59 cm , T Z by 7.89 cm , and T Z by 7.11 cm ), and rotation accuracy improves as well (e.g., R o t Z decreases by 1 . 31 ). These improvements validate the effectiveness of the cross-modal fusion model in aligning image and depth features.
Further incorporating the Matrix Loss leads to the best overall performance across both translation and rotation metrics. Compared to the baseline, the translation error along the X-axis is reduced from 38.98 cm to 24.06 cm , and the rotation error on the X-axis drops significantly from 3 . 64 to 1 . 43 , yielding a 2 . 21 improvement. This substantial reduction highlights the effectiveness of enforcing geometric consistency between translation and rotation predictions. By regularizing the network through matrix-level constraints, the model achieves more stable and accurate extrinsic estimation, particularly under large initial miscalibrations.
In summary, all three modules exhibit strong complementary effects. Their combined application achieves the highest calibration accuracy and validates the design of our full model.

4.4.3. Calibration Pairs Analysis

To highlight the capability of our Multi-Calib network in simultaneously calibrating multiple LiDAR–camera pairs, we conducted experiments on the nuScenes dataset using different numbers of sensor pairs. The injected perturbations were kept consistent with the previous settings, with miscalibration ranges of [−1.5 m, 1.5 m] and [−20°, 20°]. The detailed calibration performance under each configuration is summarized in the Table 6.
Table 6. Translation and rotation errors across six calibration pairs at the first iteration under a miscalibration range of [−1.5 m, 1.5 m]/[−20°, 20°] in nuScenes.
The experimental results show that the calibration accuracy remains highly consistent across different numbers of LiDAR–camera pairs. Both translation and rotation errors exhibit only small variations, and no performance degradation trend is observed as the number of calibrated pairs increases.
This behavior can be explained by the network design: the Multi-Calib framework allocates an independent calibration head to each LiDAR–camera pair, and all pairs share a common feature extraction backbone. As a result, increasing the number of pairs mainly adds parameters to the pair-specific calibration heads but does not alter the core feature representation or the learning dynamics of the backbone.
Consequently, the increase in calibration pairs has minimal influence on overall accuracy. The results indicate that the system maintains stable and reliable calibration performance even at the largest tested configuration, demonstrating strong scalability and robustness for multi-sensor calibration.

5. Conclusions

This paper proposes a lightweight, learning-based method for LiDAR–camera extrinsic calibration. Unlike existing approaches that are limited to fixed sensor configurations, our model supports flexible calibration of multiple LiDAR–camera pairs, making it adaptable to various autonomous driving platforms. We adopt a unified Transformer backbone to extract features from both RGB images and LiDAR depth maps, enabling efficient end-to-end learning with reduced computational costs. A cross-modal channel-wise attention module is introduced to enhance feature fusion between modalities. To handle varying sensor setups, we design configurable calibration heads that adjust to the number of sensor pairs and support simultaneous prediction in a single forward pass. Additionally, a novel loss function jointly optimizes translation and rotation by modeling their correlation, improving calibration stability and accuracy. Extensive experiments on the nuScenes dataset validate the effectiveness and scalability of our approach.

Author Contributions

Conceptualization, methodology, and writing, L.H.; review and editing, C.W.; validation, M.W.; project administration Z.W. and Y.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of Beijing Municipality, Grant 4242047.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data used to support the finding of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Song, Z.; Yang, L.; Xu, S.; Liu, L.; Xu, D.; Jia, C.; Jia, F.; Wang, L. GraphBEV: Towards robust BEV feature alignment for multi-modal 3D object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Milan, Italy, 29 September–4 October 2024; Springer Nature: Cham, Switzerland, 2024; pp. 347–366. [Google Scholar]
  2. Liu, Z.; Tang, H.; Amini, A.; Yang, X.; Mao, H.; Rus, D.L.; Han, S. BEVFusion: Multi-task multi-sensor fusion with unified bird’s-eye view representation. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 2774–2781. [Google Scholar]
  3. Li, Y.; Yu, A.W.; Meng, T.; Caine, B.; Ngiam, J.; Peng, D.; Shen, J.; Lu, Y.; Zhou, D.; Le, Q.V.; et al. Deepfusion: Lidar-camera deep fusion for multi-modal 3D object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17182–17191. [Google Scholar]
  4. Zhao, L.; Zhou, H.; Zhu, X.; Song, X.; Li, H.; Tao, W. LIF-Seg: LiDAR and camera image fusion for 3D LiDAR semantic segmentation. IEEE Trans. Multimed. 2023, 26, 1158–1168. [Google Scholar] [CrossRef]
  5. Gu, S.; Yang, J.; Kong, H. A cascaded LiDAR-camera fusion network for road detection. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 13308–13314. [Google Scholar]
  6. Fan, R.; Wang, H.; Cai, P.; Liu, M. SNE-RoadSeg: Incorporating Surface Normal Information into Semantic Segmentation for Accurate Freespace Detection. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 340–356. [Google Scholar]
  7. Zhang, J.; Liu, Y.; Wen, M.; Yue, Y.; Zhang, H.; Wang, D. L2V2T2-Calib: Automatic and unified extrinsic calibration toolbox for different 3D LiDAR, visual camera and thermal camera. In Proceedings of the IEEE Intelligent Vehicles Symposium (IV), Anchorage, AK, USA, 4–7 June 2023; pp. 1–7. [Google Scholar]
  8. Yan, G.; He, F.; Shi, C.; Wei, P.; Cai, X.; Li, Y. Joint camera intrinsic and LiDAR-camera extrinsic calibration. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 11446–11452. [Google Scholar]
  9. Jiao, J.; Chen, F.; Wei, H.; Wu, J.; Liu, M. LCE-Calib: Automatic LiDAR-frame/event camera extrinsic calibration with a globally optimal solution. IEEE/ASME Trans. Mechatron. 2023, 28, 2988–2999. [Google Scholar] [CrossRef]
  10. Lai, Z.; Wang, Y.; Guo, S.; Meng, X.; Li, J.; Li, W.; Han, S. Laser reflectance feature assisted accurate extrinsic calibration for non-repetitive scanning LiDAR and camera systems. Opt. Express 2022, 30, 16242–16263. [Google Scholar] [CrossRef] [PubMed]
  11. Beltrán, J.; Guindel, C.; De La Escalera, A.; García, F. Automatic extrinsic calibration method for LiDAR and camera sensor setups. IEEE Trans. Intell. Transp. Syst. 2022, 23, 17677–17689. [Google Scholar] [CrossRef]
  12. Domhof, J.; Kooij, J.F.; Gavrila, D.M. A joint extrinsic calibration tool for radar, camera and LiDAR. IEEE Trans. Intell. Veh. 2021, 6, 571–582. [Google Scholar] [CrossRef]
  13. Xiao, Y.; Li, Y.; Meng, C.; Li, X.; Ji, J.; Zhang, Y. CalibFormer: A transformer-based automatic LiDAR-camera calibration network. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 16714–16720. [Google Scholar]
  14. Zhu, J.; Xue, J.; Zhang, P. CalibDepth: Unifying depth map representation for iterative LiDAR-camera online calibration. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 726–733. [Google Scholar]
  15. Sun, Y.; Li, J.; Wang, Y.; Xu, X.; Yang, X.; Sun, Z. ATOP: An attention-to-optimization approach for automatic LiDAR-camera calibration via cross-modal object matching. IEEE Trans. Intell. Veh. 2022, 8, 696–708. [Google Scholar] [CrossRef]
  16. Ren, S.; Zeng, Y.; Hou, J.; Chen, X. CorrI2P: Deep image-to-point cloud registration via dense correspondence. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 1198–1208. [Google Scholar] [CrossRef]
  17. Liao, Y.; Li, J.; Kang, S.; Li, Q.; Zhu, G.; Yuan, S.; Dong, Z.; Yang, B. SE-Calib: Semantic edge-based LiDAR–camera boresight online calibration in urban scenes. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1000513. [Google Scholar] [CrossRef]
  18. Jeon, Y.; Seo, S.W. EFGHNet: A versatile image-to-point cloud registration network for extreme outdoor environment. IEEE Robot. Autom. Lett. 2022, 7, 7511–7517. [Google Scholar] [CrossRef]
  19. Yuan, K.; Guo, Z.; Wang, Z.J. RGGNet: Tolerance aware LiDAR-camera online calibration with geometric deep learning and generative model. IEEE Robot. Autom. Lett. 2020, 5, 6956–6963. [Google Scholar] [CrossRef]
  20. Lv, X.; Wang, B.; Dou, Z.; Ye, D.; Wang, S. LCCNet: LiDAR and camera self-calibration using cost volume network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; pp. 2894–2901. [Google Scholar]
  21. Zhu, A.; Xiao, Y.; Liu, C.; Tan, M.; Cao, Z. Lightweight LiDAR-camera alignment with homogeneous local-global aware representation. IEEE Trans. Intell. Transp. Syst. 2024, 25, 15922–15933. [Google Scholar] [CrossRef]
  22. Shi, J.; Zhu, Z.; Zhang, J.; Liu, R.; Wang, Z.; Chen, S.; Liu, H. CalibRCNN: Calibrating camera and LiDAR by recurrent convolutional neural network and geometric constraints. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October 2020–24 January 2021; pp. 10197–10202. [Google Scholar]
  23. Ye, C.; Pan, H.; Gao, H. Keypoint-based LiDAR-camera online calibration with robust geometric network. IEEE Trans. Instrum. Meas. 2021, 71, 2503011. [Google Scholar] [CrossRef]
  24. Lv, X.; Wang, S.; Ye, D. CFNet: LiDAR-camera registration using calibration flow network. Sensors 2021, 21, 8112. [Google Scholar] [CrossRef] [PubMed]
  25. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? The KITTI vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar]
  26. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. nuScenes: A multimodal dataset for autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Jeju, Republic of Korea, 13–19 June 2020; pp. 11621–11631. [Google Scholar]
  27. Zhou, L.; Li, Z.; Kaess, M. Automatic extrinsic calibration of a camera and a 3D lidar using line and plane correspondences. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 5562–5569. [Google Scholar]
  28. Pusztai, Z.; Hajder, L. Accurate Calibration of LiDAR-Camera Systems Using Ordinary Boxes. In Proceedings of the IEEE International Conference on Computer Vision Workshops (ICCVW), Venice, Italy, 22–29 October 2017; pp. 394–402. [Google Scholar]
  29. Tsai, D.; Worrall, S.; Shan, M.; Lohr, A.; Nebot, E. Optimising the Selection of Samples for Robust LiDAR-Camera Calibration. In Proceedings of the 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), Indianapolis, IN, USA, 19–22 September 2021; pp. 2631–2638. [Google Scholar]
  30. Kim, E.s.; Park, S.Y. Extrinsic calibration between camera and LiDAR sensors by matching multiple 3D planes. Sensors 2019, 20, 52. [Google Scholar] [CrossRef] [PubMed]
  31. Yuan, C.; Liu, X.; Hong, X.; Zhang, F. Pixel-level extrinsic self calibration of high resolution LiDAR and camera in targetless environments. IEEE Robot. Autom. Lett. 2021, 6, 7517–7524. [Google Scholar] [CrossRef]
  32. Ou, N.; Cai, H.; Yang, J.; Wang, J. Targetless extrinsic calibration of camera and low-resolution 3-D LiDAR. IEEE Sens. J. 2023, 23, 10889–10899. [Google Scholar] [CrossRef]
  33. Tu, D.; Wang, B.; Cui, H.; Liu, Y.; Shen, S. Multi-camera-LiDAR auto-calibration by joint structure-from-motion. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Kyoto, Japan, 23–27 October 2022; pp. 2242–2249. [Google Scholar]
  34. Peršić, J.; Petrović, L.; Marković, I.; Petrović, I. Spatiotemporal multisensor calibration via Gaussian processes moving target tracking. IEEE Trans. Robot. 2021, 37, 1401–1415. [Google Scholar] [CrossRef]
  35. Pandey, G.; McBride, J.R.; Savarese, S.; Eustice, R.M. Automatic extrinsic calibration of vision and lidar by maximizing mutual information. J. Field Robot. 2015, 32, 696–722. [Google Scholar] [CrossRef]
  36. Pandey, G.; McBride, J.; Savarese, S.; Eustice, R. Automatic targetless extrinsic calibration of a 3D lidar and camera by maximizing mutual information. In Proceedings of the Twenty-Sixth AAAI Conference on Artificial Intelligence, Toronto, ON, Canada, 22–26 July 2012; pp. 2053–2059. [Google Scholar]
  37. Jiang, P.; Osteen, P.; Saripalli, S. SemCal: Semantic LiDAR–Camera Calibration Using Neural Mutual Information Estimator. In Proceedings of the 2021 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI), Karlsruhe, Germany, 23–25 September 2021; pp. 1–7. [Google Scholar]
  38. Park, C.; Moghadam, P.; Kim, S.; Sridharan, S.; Fookes, C. Spatiotemporal camera-LiDAR calibration: A targetless and structureless approach. IEEE Robot. Autom. Lett. 2020, 5, 1556–1563. [Google Scholar] [CrossRef]
  39. Cattaneo, D.; Vaghi, M.; Ballardini, A.L.; Fontana, S.; Sorrenti, D.G.; Burgard, W. CMRNet: Camera to LiDAR-map registration. In Proceedings of the IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; pp. 1283–1289. [Google Scholar]
  40. Duan, Z.; Hu, X.; Ding, J.; An, P.; Huang, X.; Ma, J. A robust LiDAR-camera self-calibration via rotation-based alignment and multi-level cost volume. IEEE Robot. Autom. Lett. 2023, 9, 627–634. [Google Scholar] [CrossRef]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 17–30 June 2016; pp. 770–778. [Google Scholar]
  42. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Article metric data becomes available approximately 24 hours after publication online.