Next Article in Journal
Development of a Transient Wellbore Heat Transfer Model Validated with Distributed Temperature Sensing Data
Previous Article in Journal
ACGAN-Based Multi-Target Elevation Estimation with Vector Sensor Arrays in Low-SNR Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

LiDAR Point Cloud Colourisation Using Multi-Camera Fusion and Low-Light Image Enhancement

1
School of Minerals and Energy Resources Engineering, University of New South Wales, Sydney, NSW 2052, Australia
2
School of Surveying and Built Environment, University of Southern Queensland, Toowoomba, QLD 4350, Australia
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(21), 6582; https://doi.org/10.3390/s25216582 (registering DOI)
Submission received: 29 September 2025 / Accepted: 21 October 2025 / Published: 25 October 2025
(This article belongs to the Special Issue Advances in Point Clouds for Sensing Applications)

Abstract

In recent years, the fusion of camera data with LiDAR measurements has emerged as a powerful approach to enhance spatial understanding. This study introduces a novel, hardware-agnostic methodology that generates colourised point clouds from mechanical LiDAR using multiple camera inputs, providing complete 360-degree coverage. The primary innovation lies in its robustness under low-light conditions, achieved through the integration of a low-light image enhancement module within the fusion pipeline. The system requires initial calibration to determine intrinsic camera parameters, followed by automatic computation of the geometric transformation between the LiDAR and cameras—removing the need for specialised calibration targets and streamlining the setup. The data processing framework uses colour correction to ensure uniformity across camera feeds before fusion. The algorithm was tested using a Velodyne Puck Hi-Res LiDAR and a four-camera configuration. The optimised software achieved real-time performance and reliable colourisation even under very low illumination, successfully recovering scene details that would otherwise remain undetectable.
Keywords: point cloud colourisation; low-light image enhancement; 360° coverage; multi-camera integration; data fusion; single-shot calibration; object-free calibration point cloud colourisation; low-light image enhancement; 360° coverage; multi-camera integration; data fusion; single-shot calibration; object-free calibration

Share and Cite

MDPI and ACS Style

Ranasinghe, P.; Patra, D.; Banerjee, B.; Raval, S. LiDAR Point Cloud Colourisation Using Multi-Camera Fusion and Low-Light Image Enhancement. Sensors 2025, 25, 6582. https://doi.org/10.3390/s25216582

AMA Style

Ranasinghe P, Patra D, Banerjee B, Raval S. LiDAR Point Cloud Colourisation Using Multi-Camera Fusion and Low-Light Image Enhancement. Sensors. 2025; 25(21):6582. https://doi.org/10.3390/s25216582

Chicago/Turabian Style

Ranasinghe, Pasindu, Dibyayan Patra, Bikram Banerjee, and Simit Raval. 2025. "LiDAR Point Cloud Colourisation Using Multi-Camera Fusion and Low-Light Image Enhancement" Sensors 25, no. 21: 6582. https://doi.org/10.3390/s25216582

APA Style

Ranasinghe, P., Patra, D., Banerjee, B., & Raval, S. (2025). LiDAR Point Cloud Colourisation Using Multi-Camera Fusion and Low-Light Image Enhancement. Sensors, 25(21), 6582. https://doi.org/10.3390/s25216582

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop