Next Article in Journal
Refining WLC Benchmark Strategies for Italy’s Building Sector: The Role of Seismic Zoning and Dataset Specificity
Previous Article in Journal
LaserCAD—A Novel Parametric, Python-Based Optical Design Software
Previous Article in Special Issue
Target Tracking with Adaptive Morphological Correlation and Neural Predictive Modeling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Self-Supervised Depth and Ego-Motion Learning from Multi-Frame Thermal Images with Motion Enhancement

1
School of Automation, Nanjing University of Science and Technology, Nanjing 210094, China
2
School of Energy and Power Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(22), 11890; https://doi.org/10.3390/app152211890 (registering DOI)
Submission received: 8 October 2025 / Revised: 31 October 2025 / Accepted: 6 November 2025 / Published: 8 November 2025
(This article belongs to the Special Issue Application of Artificial Intelligence in Image Processing)

Abstract

Thermal cameras are known for their ability to overcome lighting constraints and provide reliable thermal radiation images. This capability facilitates methods for depth and ego-motion estimation, enabling efficient learning of poses and scene structures under all-day conditions. However, the existing studies on depth prediction for thermal images are limited. In practical applications, thermal cameras capture sequential frames. Unfortunately, the potential of this multi-frame aspect is underutilized by the previous methods, resulting in limitations on the depth prediction accuracy of thermal videos. To leverage the multi-frame advantages of thermal videos and to improve the accuracy of monocular depth estimation from thermal images, we propose a framework for self-supervised depth and ego-motion learning from multi-frame thermal images. We construct a multi-view stereo (MVS) cost volume from temporally adjacent thermal frames. The construction process is adjusted based on the estimated pose, which serves as a motion hint. To stabilize the motion hint and improve pose estimation accuracy, we design a motion enhancement module that utilizes self-generated poses for additional supervisory signals. Additionally, we introduce RGB images in the training phase to form a multi-spectral loss, thereby augmenting the performance of the thermal model. The experimental results, conducted on a public dataset, demonstrate the proposed method’s accurate estimation of depth and ego-motion across varying light conditions, surpassing the performance of the self-supervised baseline.

1. Introduction

Understanding depth within images and the ego-motion of cameras is crucial for 3D vision perception. Its significance extends to practical applications like autonomous driving [1], object detection [2], and mixed reality [3]. Although sensors like LiDAR can directly measure object distances, their high expense and sparse depth motivate the investigation of visual self-supervised methods for their versatility and applicability. Self-supervised learning of depth and pose, extensively researched [4,5,6], demands neither expensive hardware nor intricate labels. It adeptly generates dense depth solely from visual data. Nevertheless, most studies concentrate on estimating depth in RGB images. The sensitivity of RGB image quality to illumination conditions results in information loss and heightened noise [7], especially in dim environments like nighttime or tunnels. These challenges, inherent in the real world, compromise the performance and generalization capabilities of RGB-based methods in low-light conditions.
Self-supervised learning from thermal images emerges as a potential remedy to close the performance gap. In contrast to RGB cameras, thermal cameras function independently of extra light sources and can seamlessly operate in complete darkness. Thermal cameras detect the Long-Wave Infrared (LWIR) radiation emitted from objects directly and then convert the raw radiation data into a temperature distribution represented in a visible image for human perception [8]. LWIR radiation’s ability to penetrate certain transparent mediums empowers thermal cameras to operate effectively in adverse weather, like rain and fog. In real-world scenarios, the thermal camera can provide more than one frame during testing, such as vehicles carrying thermal cameras for inspection tasks. However, the existing methods for self-supervised depth and ego-motion learning [9,10,11] fall short in fully utilizing the potential of multiple frames provided by thermal cameras.
To utilize the benefits of the multi-frame property and optimize the network predictions from thermal videos, we propose a self-supervised depth and ego-motion estimation framework based on multi-view stereo (MVS) for thermal videos. A cost volume is constructed by considering temporally adjacent frames from thermal videos with various hypothesized depth values. Then, the cost volume is decoded to generate an MVS depth. However, in large-scale outdoor scenes with slow camera movement, MVS methods face challenges. Owing to minimal disparities and insufficient triangulation [12] in such situations, the network acquires limited new information, potentially resulting in geometry ambiguity in the predicted depth results.
In response to this problem, we adopt a strategy that adaptively adjusts the range of depth candidate sampling guided by motion hints. Depending on the poses generated by the network, the MVS depth will be shrunk to the single-frame depth when the camera is static. To bolster motion hint stability and refine pose estimation accuracy, we design a motion enhancement module based on epipolar geometry constraints. This module utilizes self-generated supervisory signals to aid in network training. Moreover, to address the lower resolution and unclear structures and edges in thermal images, we integrate a depth warping operation into the proposed framework. It combines the thermal camera with a visible camera. This integration facilitates the combination of thermal loss and photometric loss into a multi-spectral loss, providing rich self-supervisory signals. The experimental results on the public dataset demonstrate that our method is capable of predicting depth and ego-motion under low-light conditions, exhibiting performance superior to state-of-the-art baselines, as shown in Figure 1.
In summary, we make the following contributions:
  • We propose a self-supervised multi-frame depth and ego-motion learning framework for monocular thermal videos. It successfully leverages the temporal advantages of multiple thermal frames by constructing the multi-view stereo. The accuracy of prediction results is further improved through the integration of motion hints and multi-spectral properties.
  • We design a motion enhancement module that utilizes self-generated motion constraints to complement self-supervised signals. It enhances the stability of motion hints within the framework, consequently improving the accuracy of pose estimation.
  • We employ an efficient objective function that combines photometric loss, thermal loss, and motion loss to address the challenges in real-world scenarios, including unavoidable low-light conditions.

2. Related Work

2.1. Self-Supervised Depth and Ego-Motion Learning

The goal of depth and ego-motion learning [13,14,15] is to simultaneously estimate the depth of each pixel in an image and the camera poses between image sequences. Self-supervised methods avoid the use of challenging-to-obtain ground-truth by using the view synthesis mechanism. SfM-Learner, proposed by Zhou et al. [4], stands as a pioneering work in this field. It utilizes the generated single-frame depth and poses to project source frame images onto the target frame, synthesizing an image. Subsequently, it calculates the photometric loss between the original and synthesized images for training. Some self-supervised methods [5,16,17,18] build upon this approach to improve depth prediction performance. Extending the work of SfM-Learner, Bian et al. [17] introduce a scale-consistency loss to address scale-ambiguity in each frame. Bian et al. [18] integrate estimation networks with the conventional visual odometry ORB-SLAM2 [19], achieving heightened accuracy. EDS-Depth [20] enhances self-supervised depth estimation in dynamic scenes using frame interpolation and pseudo-labels derived from optical flow. To improve depth estimation with dynamic objects or a static camera, Feng et al. [21] separate static from temporal features via a dual-path encoder and synthesize full-resolution depth using a learnable offset field.
Recent works [22,23,24,25] introduce multi-view stereo (MVS) into depth learning. These methods can generate dense geometry structures of scenes using multiple frames. Assuming static scenes, Watson et al. [24] introduce MVS into self-supervised depth learning. The method requires neither ground-truth nor any pretrained networks. IAFMVS [26] enhances iterative MVS by leveraging deformable convolutions for adaptive feature extraction and attention for feature matching, specifically targeting texture-less regions. Cheng et al. [22] improve the robustness of depth estimation systems under inaccurate camera poses by adaptively fusing high-confidence single-view and multi-view depth information. Emphasizing its MVS foundation, Wang et al. [27] use a cross-attention mechanism to fuse geometric constraints from multi-frame optical flow, enabling recurrent refinement of scene detail. Compared to single-frame depth estimation methods, MVS-based methods can effectively utilize information from various perspectives, exhibiting more robust geometry structures and reduced sensitivity to occlusion. MVS methods hold significant research potential in self-supervised depth prediction, offering broad applicability across diverse scenarios. The potential value of MVS methods in self-supervised depth prediction suggests promising applications in various scenarios.

2.2. Depth and Ego-Motion Learning with Thermal Images

In contrast to the abundant methods for depth and pose estimation from RGB images, the research on estimation from thermal images is limited. RGB cameras are susceptible to illumination conditions, whereas thermal cameras offer image capture under broader conditions, including nighttime and adverse weather environments. This provides a distinct advantage to thermal images in 3D vision perception. Some works [28,29] employ additional motion and ranging sensors, such as IMU and radar, to assist thermal cameras in ego-motion estimation. However, these methods primarily concentrate on localization rather than environmental perception.
Recently, only a limited number of works have focused on depth learning from thermal images. Kim et al. [10] design a system composed of two RGB cameras and one thermal camera. The system utilizes the spatial relationship between the stereo cameras to predict the depth of the current frame. However, this method requires a complex geometry configuration between the cameras and does not estimate the camera poses. Lu et al. [9] propose a similar method to Kim et al. [10], incorporating two RGB cameras and two thermal cameras. They initialize the disparity map with RGB images and design a specialized network module to translate a thermal image from RGB images. In their final step, triangulation using the translated thermal images and original thermal images provides depth ground-truth. Nevertheless, their system still does not estimate ego-motion. Shin et al. [11] propose the first self-supervised method for simultaneously estimating ego-motion and depth from thermal images based on the SfM-Learner framework. They train pose and depth networks by reconstructing thermal images. However, networks from these methods utilize the single-frame image to predict depth, insufficiently exploiting the temporal and multi-frame advantages of thermal videos, thereby limiting improvements in estimation accuracy.

3. Method

3.1. Method Overview

Our framework is designed to jointly train neural networks for estimating depth and camera pose from monocular thermal videos. For this goal, we utilize a self-generated multi-spectral supervision signal and a motion supervision signal from a motion enhancement module to optimize the network.
To refine estimation accuracy, we construct a cost volume between temporally adjacent thermal frames, enabling the direct utilization of geometry constraints from MVS. This approach exploits the strengths of multi-frame thermal images. Additionally, the motion enhancement module, as designed, aims to improve pose estimation accuracy and maintain the stability of cost volume construction. Since the proposed framework employs self-supervised learning, no ground-truth is required at training time.
The main architecture of the proposed framework is illustrated in Figure 2. For a given thermal video input, (a) the PoseNet is employed initially to estimate the camera poses between thermal frames. This combines the poses from the motion enhancement module to compute the motion loss. (b) Subsequently, we utilize the DepthNet to predict the single-frame depth of the current thermal image. (c) We construct a cost volume from the multiple thermal frames. During this process, the predicted single-frame depth serves as a geometry prior, and the predicted poses act as a motion hint to assist the construction of the cost volume. Upon decoding the cost volume, the MVS depth and confidence map are generated. A more accurate depth is achieved through fusing the single-frame depth with the MVS depth.

3.2. Single-Frame Self-Supervised Depth Learning

As illustrated in Figure 2a,b, single-frame self-supervised depth learning relies on the view synthesis mechanism [4] and reconstruction loss. The current thermal frame I t T can be reconstructed from the previous frame I t 1 T . This mechanism is based on the warping operation:
p t 1 = K T P t 1 t T D M o n o , t T ( p t ) K T 1 p t ,
in which D M o n o , t T ( p t ) is the generated depth of I t T from DepthNet. P t 1 t T is the estimated pose between I t T and I t 1 T . K T is the intrinsic matrix of the thermal camera. The reconstructed frame I ^ t T is then synthesized by a bi-linear sampling operation. Utilizing I ^ t T and I t T , we can compute the multi-spectral reconstruction loss, which is described in detail in the subsequent section.

3.3. Multi-Frame Depth Learning from Thermal Images with Motion Hint

Our multi-frame pipeline, detailed in Figure 3, employs a shared MVS encoder to process adjacent thermal frames I t T , I t 1 T . The encoder begins with a convolutional stem, followed by downsampling blocks using 5 × 5 and 3 × 3 convolutions to create a feature pyramid [30]. These features are then warped into a cost volume, with the depth sampling range adaptively guided by a motion hint derived from prior depth and pose estimates. A U-Net-like MVS decoder [31], built upon 3 × 3 × 3 3D convolutional layers, processes this volume. The decoder leverages skip connections to fuse multi-scale information and uses a final softmax layer to compute a depth probability map. Finally, a lightweight confidence network processes this probability map to generate a weighting map, which fuses the geometrically robust MVS depth with the single-frame depth to produce the refined output.
Following the previous works [23,24,25], our MVS depth estimation part constructs a cost volume by warping the encoded source frame to the target frame. The cost volume is derived from the measurement of geometric similarity between two thermal frames. It signifies the likelihood of the correct depth for each pixel among a set of different depth values. The decoded cost volume then generates the highest-activated depth value for each pixel. The MVS-based method mitigates geometry ambiguity in single-frame depth and enhances prediction accuracy.

3.3.1. Cost Volume Construction

The MVS encoder (Figure 2c) extracts the feature of the current thermal frame I t T and the previous frame I t 1 T to produce feature maps denoted as F t + i R H / 4 × W / 4 × C , i { 1 , 0 } . Then, the previous feature map is warped to the current feature map:
p t 1 , j = K T P t 1 t T D j ( p t ) K T 1 p t ,
with p t being the pixels on the current feature map, D j ( p t ) stands for the depth value of the j-th hypothesized depth value, and p t 1 , j corresponds to the pixel on the previous frame’s feature map. Through the warping, we construct the feature volume Z t 1 R H / 4 × W / 4 × C × N D , in which N D represents the number of depth values.
The cost volume between the two feature maps is constructed by computing visual similarity:
q i = z i , f i ,
here, · , · is the inner product operation. z i R C × N D is the feature of the i-th pixel in feature volume Z t 1 , and f i R C × 1 denotes the feature of the i-th pixel in the feature map F t . Then, all pixel features form the cost volume Q t R H / 4 × W / 4 × G × N D .

3.3.2. Depth Sampling with Motion Hint

Within the construction of the cost volume, the candidate depth set is formed by various depth values. However, the observed outdoor scenes are not static, and specifying a fixed range of depth values with predetermined maximum and minimum values would result in redundant depth candidates. To mitigate this, we leverage single-frame depth as prior information. Subsequently, we employ motion hints to adjust the adaptive range of the depth set, guiding the depth sampling process.
As the thermal camera moves, the velocity of the camera significantly influences the thermal frame disparity [12]. Higher velocities lead to more abundant triangulation for multi-view geometry, resulting in larger disparities. Conversely, lower velocities reduce triangulation, yielding smaller disparities and hindering the MVS network to acquire new information. Leveraging this knowledge, we utilize the estimated ego-motion as the motion hint to adaptively adjust the depth range:
d min = 1 γ t r a n s P t 1 t T 2 D M o n o , t T d max = 1 + γ t r a n s P t 1 t T 2 D M o n o , t T ,
where t r a n s P t 1 t T 2 denotes the camera’s velocity, and t r a n s · indicates the translational part of the camera’s pose. γ is a hyperparameter determined by the camera’s frame rate. To maintain training stability, γ t r a n s P t 1 t T 2 is clamped to the range ( 0 , 1 ) . We set γ = 0.15 . For depth sampling, an inverse sampling strategy [23] is employed to ensure that sampling results are closer to D M o n o , t T when the camera is static or moving at a low velocity:
d j = d max 1 + j N D 1 d min 1 d max 1 1 .
Following the sampling process, the cost volume is fed into the MVS decoder to produce the depth probability map M p r o b R H / 4 × W / 4 × N D . The MVS depth D M V S , t T for the current thermal image is then computed from M p r o b R H / 4 × W / 4 × N D by the local-max function [32]:
D M V S ( p t ) = j = J ϵ J + ϵ D j ( p t ) M p r o b ( j , p t ) Σ i = J ϵ J + ϵ M p r o b ( i , p t ) .
Here, J represents the index of the maximum value within the per-pixel probability vector M p r o b ( · , p t ) . The parameter ϵ is the index radius along the depth dimension, which we set to 1. The resulting low-resolution depth map D M V S ( p t ) is then upsampled to its original resolution.

3.3.3. Depth Fusing

To further enhance the accuracy of MVS depth and leverage the strengths of single-frame depth, we rectify regions with lower confidence [33] in the MVS depth using the results from the single-view depth. The confidence map M c o n f is generated by computing the information entropy of the decoded depth probability map M p r o b :
M c o n f = θ c o n f j = 0 D 1 M p r o b , j log M p r o b , j ,
here, * denotes element-wise multiplication, and θ c o n f represents the confidence network. The final fused depth is calculated using the following formula:
D F u s e d , t T = 1 M c o n f D M o n o , t T + M c o n f D M V S , t T .

3.4. Motion Enhancement Using Epipolar Geometry Constraint

Visual localization systems [19,34,35] often rely on matching sparse stable features like keypoints and lines to achieve robust localization. These systems then employ well-established geometry tools to compute relative camera poses. Namely, the relative ego-motion between two frames can be calculated using the epipolar geometry constraint, and iterative optimization diminishes the estimation error, providing effective guidance for training networks. Therefore, to guarantee stability for the motion hint-driven method and enhance pose estimation accuracy, we design a motion enhancement module, as illustrated in Figure 4. Through offline processing, these geometry elements serve as prior information and self-supervised signals integrated into the neural network training.

3.4.1. Epipolar Geometry Constraint

As the scenes observed by the thermal frames I t 1 T , I t T are similar, we extract keypoints from the thermal frames and obtain a set of feature matches for keypoint pairs by associating their SIFT feature descriptors [36]. According to the epipolar geometry constraint, the relationship between keypoint pairs and the relative camera pose is as follows:
p t K T E K T 1 p t 1 = 0 , p t = H p t 1 ,
with p t 1 and p t denoting the homogeneous coordinates of a pair of keypoints between two frames. H is the homography matrix, and E is the essential matrix. The pose R t 1 t | t t 1 t between the two frames can be determined using the formula E = t t 1 t × R t 1 t , where · × represents the skew-symmetric matrix of a three-dimensional vector.

3.4.2. Matrix Scoring and Selection

We calculate the homography matrix and the essential matrix separately and select the one that best fits the current frame situation. With an RANSAC scheme, E and H are computed using the eight-point algorithm [37] and the DLT algorithm [38], respectively. At each iteration, the score for the solution is calculated using the following formula:
s = i e t 1 t 2 p t , i , p t 1 , i + e t t 1 2 p t , i , p t 1 , i ,
where e 2 represents the symmetric transfer error [38], calculated through mutual projection between the two frames. p t , i and p t 1 , i are the i-th keypoint pairs. Finally, we preserve E and H associated with the highest score.
In situations where the scene is planar or exhibits low parallax, explaining it using the essential matrix may lead to wrong results in ego-motion recovery. We should choose the homography matrix for pose retrieval. Conversely, in scenes with sufficient parallax, the homography matrix can only solve for pose using keypoint pairs on the planes, whereas the essential matrix yields more accurate results. Thus, matrix selection is determined by the following:
s r a t i o = s H s H + s E ,
with s r a t i o being the ratio of scores from the homography matrix. If s r a t i o > 0.45 , it indicates situations involving either a low parallax or the camera detecting the extensive planar. In such cases, the scene is processed using the homography matrix. Conversely, the essential matrix is selected.

3.4.3. Montion Recovery

After the matrix selection, to guarantee the validity of the pose solving, we examine the motion hypothesis. In the case of the homography matrices, we perform the hypothesis verification [39] on the eight solutions. Solutions where keypoints go behind the camera are discarded. From the remaining solutions, those with significant parallax and minimal reprojection errors are selected. For essential matrices, the four solutions are examined using the Singular Value Decomposition (SVD), and erroneous solutions are removed using the similar approach as with homography matrices. Then, the pose between the two frames is determined.

3.4.4. Motion Loss

The poses generated by the motion enhancement module are converted from the transformation matrix to the 6DoF representation. We incorporate the pose, computed based on the epipolar geometry constraint, into the motion loss function for optimization:
L m o t i o n = w r r ^ t 1 t r t 1 t 2 + w t t ^ t 1 t t t 1 t 2 ,
where r ^ t 1 t , t ^ t 1 t and r t 1 t , t t 1 t are the 6-DoF relative poses from the motion enhancement module and PoseNet, respectively. w r and w t denote the weights corresponding to rotation and translation. Notably, normalization of the translation is applied for scale consistency. We empirically set w r = 10 and w t = 1 .

3.5. Multi-Spectral Loss

In the training phase, conducting view synthesis on thermal and RGB images allows us to calculate thermal loss and photometric loss, respectively. This provides the network with sufficient multi-spectral self-supervision signals. In the testing phase, our framework solely requires thermal frames to estimate depth. To calculate loss terms from visible images, a depth warping operation is employed to establish the relationship between the two cameras, allowing the depth warping from the thermal camera coordinate system to the RGB camera coordinate system.

3.5.1. Thermal Reconstruction Loss

After the networks generate depths at three levels (single-frame depth D M o n o , t T , MVS depth D M V S , t T , and fused depth D F u s e d , t T ), the thermal image I ^ t T is synthesized using the estimated pose P t 1 t T and the previous thermal frame I t 1 T , as described in Section 3.2. Utilizing both the original thermal image I t T and the synthesized image I ^ t T , we compute the thermal reconstruction loss:
L r e c T = min t e I t T , I ^ t T ,
in which t e is a thermal error and is defined as a combination of an SSIM loss and an L1 loss:
t e I t T , I ^ t T = w T 2 1 S S I M I t T , I ^ t T + 1 w T I t T I ^ t T 1 ,
here, S S I M · , · represents the Structural Similarity Index (SSIM) loss [40], and w T is the weight of the SSIM loss. The SSIM loss remains effective for training due to the diverse structures and shapes of thermal sources in the scenes. Empirically, we set w T = 0.15 .

3.5.2. Depth Warping

The photometric loss derived from RGB images effectively supplements self-supervision signals, mitigating inherent deficiencies in thermal images, such as weak textures and blurred boundaries. However, the photometric loss typically demands an additional RGB network to process visible images. To address this challenge, a depth warping operation is employed to enable the depth warping from the thermal camera to the RGB camera, eliminating the requirement for an extra neural network to infer visible images. Initially, exploiting the rigid geometry association of the two cameras and the generated depth maps, we calculate the thermal–visible rigid flow between the thermal and visible frames:
S T V p t = K V P T V D t T p t ( K V ) 1 p t p t ,
where D t T is the depth of the current thermal frame, and p t represents the pixels in the current thermal frame. K V is the intrinsic matrix of the visible camera, and P T V is the extrinsic matrix between the two cameras.
Subsequently, with the flow reversal operation and the differentiable sampling, the depth of the current visible frame is obtained as follows:
D t V = f s a m p l e D t T , f r e v e r s e S T V ,
where f r e v e r s e · is the flow reversal layer [41], which is designed to reverse the computed optical flow of pixels. f s a m p l e · , · denotes the spatially sampling function [42], utilized for calculating the visible frame depth, which is transferred with the pixel offset of the optical flow. As the two operations are differentiable, they are allowed for backward propagation during training.

3.5.3. Photometric Reconstruction Loss

The depth D t V , computed through the depth warping module, can be utilized for synthesizing visible images I ^ t V . Following [5], the photometric reconstruction loss for visible images is expressed as
L r e c V = min p e I t V , I ^ t V ,
p e I t V , I ^ t V = w V 2 1 S S I M I t V , I ^ t V + 1 w V I t V I ^ t V 1 ,
where p e is the photometric error. w V is the SSIM loss weight for visible images and is set as w V = 0.85 .
Then, the multi-spectral loss L r e c M S is formed by the thermal loss and the photometric loss:
L r e c M S = λ T L r e c T + λ V L r e c V ,

3.5.4. Total Loss Function

Our framework is trained in a self-supervised manner, and we have the total loss, which is formulated as
L t o t a l = L r e c M S + λ M L m o t i o n + λ S L S T + L S V ,
where L S represents the edge-aware smoothness loss [43], which is commonly incorporated in existing works to deal with the discontinuity in depth map:
L S = i N D t p i · e | I t ( p i ) | ,
where N is the number of pixels in an image. ∇ represents the 2D differential operator. L S is applicable to both RGB part (V) and thermal part (T). L r e c M S signifies the multi-spectral reconstruction loss, composed of the thermal loss and the photometric loss. L m o t i o n denotes the motion loss term from the motion enhancement module. λ S and λ M are weights assigned to the smoothness loss and the motion loss.

4. Experiments

4.1. Datasets

To validate the effectiveness of our framework, we conduct training and testing using the public and available VIVID dataset [44]. This dataset is highly suitable for our task as it comprises multimodal data streams with temporal properties, containing raw thermal images, RGB images, sensor extrinsic calibration results, camera poses (serving as ego-motion ground-truth), and LiDAR point clouds (serving as depth ground-truth). It encompasses challenging scenes from both daytime and nighttime, various lighting conditions, and scenarios with different motion conditions. For our experiments, we utilize 6958 samples for training and 1058 samples for testing, and all images are processed at a 320 × 256 resolution. As our method is self-supervised, the ground-truth data (e.g., LiDAR depth and camera poses) is not required during training. The training phase solely utilizes the thermal images, the corresponding RGB images, and the sensor calibration parameters to compute the loss functions. During inference, our model is fully independent and requires only the monocular thermal frames as input. The ground-truth data is used exclusively for the quantitative evaluation of our method’s performance during the testing phase. Hence, the VIVID dataset provides satisfactory and comprehensive resources for training and evaluating the proposed framework.

4.2. Thermal Image Preprocessing

To meet the training requirements, we perform a multi-step preprocessing pipeline. Typically, the 8-bit thermal images produced by the thermal camera undergo an automatic frame-wise min–max rescaling. The 8-bit thermal images lack temporal consistency, which is unsuitable for the self-supervised training of consecutive image sequences. On the other hand, the 14-bit raw thermal images, due to the expansive temperature measurement range, are impractical for depth and ego-motion estimation, leading to the compression of a substantial amount of valuable information.
Therefore, to ensure temporal consistency while enhancing structural details, we employ a comprehensive pipeline. The resized 14-bit raw data ( I T , r a w ) is first clamped to the predefined temperature hyperparameters { τ min , τ max } and normalized to the [0, 1] range. This normalized image, which often suffers from low contrast, then undergoes a mild contrast adjustment. We apply Contrast Limited Adaptive Histogram Equalization (CLAHE), which boosts local features while avoiding the introduction of excessive noise. Finally, this single-channel image is remapped using a jet colormap to 3-channels and standardized. This full pipeline can be formulated as a single operation:
I T = f s t d f m a p f a d j f c l a m p ( I T , r a w , τ max , τ min ) τ min τ max τ min
where f c l a m p ( · ) is the clamping function, f a d j ( · ) is the contrast adjustment function, f m a p ( · ) is the colormap function, and f s t d ( · ) is the standardization. Figure 5 illustrates the visualization of the results of this preprocessing. The subsequent ablation study demonstrates the effectiveness of this processing in improving the performance of the networks.

4.3. Implementation Details

Training Configuration

For the depth prediction network of single-frame part (DepthNet) and PoseNet, we utilize the architectures from Monodepth2 [16] and SfMLearner [4] with the ResNet-18 [45] encoder, respectively. The confidence network is constructed with a three-layer 2D CNN and a Sigmoid layer. The ResNet blocks are pretrained with ImageNet, and other network layers are trained from scratch. Our method was implemented using the PyTorch (version 2.2.2) deep learning library [46]. All networks are trained on a single NVIDIA RTX 3080Ti GPU using the ADAM optimizer [47].
For data augmentation, we employed an offline strategy. Multiple augmented sequences were generated to expand the original dataset by applying varied parameters for horizontal flipping, scaling, and cropping. We shuffle and randomly sample from this entire original and augmented pool. During training, the single-frame depth prediction network and pose estimation network receive three consecutive thermal frames I t 1 T , I t T , I t + 1 T , and the MVS depth network takes two frames I t T , I t 1 T . To balance the loss, we assign weights λ T , λ V , λ S , λ M as 1.0 , 0.1 , 0.001 , 1.0 , respectively. The temperature parameters τ min , τ max in the thermal image preprocessing are set as 0.0 , 30.0 (in degrees Celsius). The batch size is set to 4, and the learning rate is 0.0001. A total of 300 epochs are conducted for training.

4.4. Depth Prediction Results

In evaluating the performance of the proposed method, we compare it with state-of-the-art self-supervised depth estimation methods. The work of Shin et al. [11] is set as the baseline. Additionally, the methods employing monocular RGB supervision are included for comparison. The evaluation uses the metrics proposed by Eigen et al. [14], which are widely employed in modern depth estimation methods for RGB images, to measure the performance. The metrics include the error metrics (smaller is better) and the accuracy rate metrics (larger is better):
Abs Rel : 1 N i = 1 N D ( p i ) D gt ( p i ) D gt ( p i )
Sq Rel : 1 N i = 1 N D ( p i ) D gt ( p i ) 2 D gt ( p i )
RMSE : 1 N i = 1 N D ( p i ) D gt ( p i ) 2
RMSE log : 1 N i = 1 N log D ( p i ) log D gt ( p i ) 2
Accuracy : % of D ( p i ) s . t . max D ( p i ) D gt ( p i ) , D gt ( p i ) D ( p i ) = δ < thr
where D ( p i ) and D gt ( p i ) are the estimated and ground-truth depth values, respectively. thr is a threshold, which is set to 1.25 , 1.25 2 , and 1.25 3 .
Our framework’s performance in monocular depth estimation was rigorously evaluated on the VIVID dataset. The experiments, summarized in Table 1 and Figure 6 and Figure 7, conclusively demonstrate the superiority of our approach, which arises from the synergistic integration of multi-frame geometric constraints and multi-spectral supervision.
As shown in Table 1, our method consistently outperforms baseline self-supervised techniques across all scenarios and metrics. The key to this performance lies in our departure from traditional single-frame estimation. Unlike the baseline [11], which predicts depth from a single image and thus fails to exploit temporal data, our framework is built upon an MVS cost volume. By constructing this volume from temporally adjacent thermal frames, our model effectively leverages multi-view geometry to resolve ambiguities inherent in a single perspective, leading to more robust and accurate geometric structures. This architectural advantage is the primary driver behind the significant 19.8% reduction in absolute relative error (Abs. Rel.) compared to the baseline.
The results particularly highlight our method’s ability to maintain performance across extreme variations in illumination, especially when compared to RGB-based approaches. As predicted, RGB-based methods suffer a catastrophic performance degradation at night, underscoring their unreliability for continuous operation. While thermal imaging is inherently advantageous in these conditions, our method proves more stable than even the thermal-based baseline. From day to night, our model’s Abs. Rel. error increases by a mere 2.55%, whereas the baseline degrades by 8.72%. This enhanced stability can be attributed to the multi-spectral loss employed during training. By using a depth warping operation to incorporate a photometric reconstruction loss from registered RGB images, our thermal network learns richer, more generalized features that better capture object boundaries and fine-grained textures—details often blurred or subtle in thermal data alone. This robust feature representation, learned via multi-spectral supervision, translates directly to more consistent and accurate predictions when presented with only thermal images during inference.
To visually corroborate the quantitative results from Table 1, Figure 6 provides a qualitative comparison. The visualization highlights a key limitation of relying on standard cameras: in the challenging nighttime scene (bottom row), the RGB-based method (e) fails completely, producing incoherent geometry. In stark contrast, both our method (c) and the thermal baseline (d) robustly generate plausible depth maps that align with the ground-truth structure (f), underscoring the advantage of using thermal imagery for all-day operation.
A more granular analysis in Figure 7 reveals the improvements our method offers over the thermal baseline. This figure provides a detailed side-by-side comparison with the baseline, including error maps. The baseline frequently produces noticeable geometric distortions and artifacts. These are common failure modes for single-frame methods. In contrast, our results are geometrically coherent and artifact-free, a direct benefit of the MVS cost volume that aggregates spatial information over time to form a more complete and consistent 3D understanding of the scene.
Furthermore, the error maps (e, f) provide a granular view of our method’s precision. Our error map (e) is predominantly blue (low error), whereas the baseline’s (f) shows significant concentrations of red and orange (high error), especially around object edges. This visual evidence confirms that our framework, which combines the strengths of MVS with a final depth fusing step, achieves a superior level of per-pixel accuracy and produces more reliable depth maps for the downstream tasks.

4.5. Pose Estimation Results

In addition to depth, accurate ego-motion estimation is critical for autonomous navigation. The results presented in Table 2 demonstrate that our framework surpasses the baseline method in self-supervised pose estimation from thermal video.
As detailed in Table 2, our method achieves the lowest Absolute Trajectory Error (ATE) [48] across all the daytime and nighttime sequences. The most significant contributor to this success is the proposed motion enhancement module. This module provides a geometrically grounded supervisory signal by leveraging epipolar geometry constraints to compute a highly accurate pose offline. This extra poses supervision signal is then used to formulate a motion loss that guides the training of our PoseNet. This additional supervision stabilizes the learning process and forces the network to adhere to geometric principles, resulting in a 24.94% improvement in ATE over the baseline. This module also ensures the stability of the motion hint used for MVS depth sampling, creating a virtuous cycle where better poses lead to better depth and vice versa.
Moreover, similar to our depth results, our pose estimation network maintains exceptional performance from day to night, with its ATE degrading by only 1.51%. Compared to RGB-supervised methods, our method exhibits robustness to changes in illumination. In well-lit daytime scenes, RGB-based self-supervised methods perform well, but, as the scenes transition to poorly lit nighttime conditions, the errors in these methods increase. For instance, the method proposed by Bian et al. [18] shows that performance decreases by 74.43%.

4.6. Ablation Study

In the evaluation experiments, our method reveals superior performance on ego-motion and depth estimation compared to the other approaches. However, the individual contributions and improvements made by each component of the proposed framework remain unclear. Therefore, we conduct ablation experiments on the VIVID dataset to analyze and evaluate the contributions of each component and loss term within the framework. The results, presented in Table 3 and Table 4, systematically deconstruct our framework and quantify the importance of each element.

4.6.1. Effect of Each Component

This study, summarized in Table 3, assesses the impact of each core architectural module by selectively toggling them on or off within the proposed framework. By toggling various components on or off within the proposed framework, we show their respective importance. Table 3 presents the performance test results for depth prediction.
“Ours(full)” represents the results using all the components, and the other rows correspond to the results with the removal of each respective component. The results indicate that “Ours(full)” achieves the best performance. Removing different components leads to varying degrees of performance decline, with increased prediction errors and decreased accuracy. For “Ours (w/o depth warping),” we deactivate the depth transfer module, thus not receiving any supervisory signals from visible images. In the case of “Ours (w/o motion enhancement),” the motion enhancement module is turned off, omitting the utilization of the epipolar geometry constraint for pose generation. In the “w/o motion hint” experiment, the depth candidate range is half of the single-frame depth value. As for “Ours (w/o remapping),” training is conducted using the original 14-bit thermal images. For “Ours (w/o depth fusing),” the model solely outputs MVS depth without performing depth map fusion.
The underlying rationale for the motion enhancement module is to provide a supervisory signal with a different failure mode than the primary reconstruction-based supervision. The dense reconstruction signal can be ambiguous in photometrically poor regions, where an incorrect pose might still yield a low error. Our module provides an additional signal that relies on sparse structural features rather than dense texture content. To validate the quality of this mechanism, Figure 8 visually demonstrates its core feature-matching process. The module successfully extracts a dense and geometrically consistent set of feature correspondences, confirming that it provides a reliable geometric constraint. The benefit of adding this geometrically rigid constraint is then shown in Table 3. The “Ours (w/o motion enhancement)” experiment, which relies solely on the potentially ambiguous dense signal, shows a corresponding drop in pose estimation accuracy.

4.6.2. Effect of Each Loss Term

The ablation study is conducted on different loss terms by setting them to zero. The depth prediction evaluation results for the experiments are presented in Table 4. “Ours(full)” employs all the loss terms and achieves the best results. The thermal reconstruction loss is found to be crucial and foundational to our framework as removing it results in non-convergent training. Although removing the smoothness loss enhances the visibility of the structural details in the depth map, it results in an overall increase in errors. Similarly, removing the photometric reconstruction loss or the motion loss leads to increased errors and decreased accuracy in depth prediction.

5. Discussion

While our framework demonstrates robust performance, we acknowledge several avenues for future work. Our method does not explicitly model dynamic objects, a common challenge for self-supervised methods relying on view synthesis. Thus, incorporating dynamic-aware modules remains a valuable direction. We also acknowledge that the multi-spectral loss requires RGB data during the training phase. Although this is a training-only optimization (our model is fully independent at inference and our ablation study shows that it remains effective without this signal), this reliance does reduce data collection flexibility. A promising future direction is to enhance this flexibility by applying domain adaptation techniques to reduce this RGB dependence. In terms of deployment, while the MVS cost volume adds computational load, our full model achieves 43 FPS (23 ms per frame), comfortably meeting real-time requirements. We also note that the motion enhancement module is a training-only component and does not impact runtime. For more resource-constrained hardware, model compression and knowledge distillation remain promising avenues for optimization. Finally, while validated on a comprehensive benchmark, extending the framework’s applicability presents a clear direction for future work, such as evaluating its generalization to broader domains, including diverse scenarios and fusion with sensors like LiDAR and IMU.

6. Conclusions

In this paper, we introduce a self-supervised learning framework for depth and pose estimation from multi-frame thermal images. Firstly, the proposed framework constructs an MVS cost volume from multiple thermal frames. The poses generated by the network adaptively guide and adjust the construction process, directing MVS depth towards single-frame depth in scenarios with slow or static camera movement. Secondly, to bolster the stability of this mechanism, a motion enhancement module based on epipolar geometry constraints provides self-generated supervision signals to optimize the training process. Finally, a depth warping operation links visible and thermal cameras, establishing a loss term based on the multi-spectral property to aid in training the thermal network.
The experimental evaluations on the public dataset showcase that the proposed method demonstrates improved accuracy in depth and pose estimation compared to state-of-the-art methods. It exhibits effectiveness and accuracy even under challenging light conditions. We expect further improvements in the results by incorporating additional geometry features and employing multi-sensor fusion methods to boost the accuracy and robustness of depth and pose predictions.

Author Contributions

Conceptualization, R.Y., G.M., and J.G.; methodology, R.Y. and L.X.; investigation, R.Y. and L.X.; resources, G.M. and J.G.; supervision, G.M. and J.G.; writing—original draft, R.Y.; writing—review and editing, G.M. and J.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Agricultural Science and Technology Independent Innovation Fund of Jiangsu Province (Grant No. CX(24)1023) and the Key Research and Development Program of Jiangsu Province (Industrial Foresight and Key Core Technologies) (Grant No. BE2021016).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://visibilitydataset.github.io/, accessed on 8 October 2025.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, C.; Lee, D. EASD: Exposure Aware Single-Step Diffusion Framework for Monocular Depth Estimation in Autonomous Vehicles. Appl. Sci. 2025, 15, 9130. [Google Scholar] [CrossRef]
  2. Shi, P.; Dong, X.; Ge, R.; Liu, Z.; Yang, A. Dp-M3D: Monocular 3D object detection algorithm with depth perception capability. Knowl.-Based Syst. 2025, 318, 113539. [Google Scholar] [CrossRef]
  3. Yang, W.J.; Tsai, H.; Chan, D.Y. High-Precision Depth Estimation Networks Using Low-Resolution Depth and RGB Image Sensors for Low-Cost Mixed Reality Glasses. Appl. Sci. 2025, 15, 6169. [Google Scholar] [CrossRef]
  4. Zhou, T.; Brown, M.; Snavely, N.; Lowe, D.G. Unsupervised learning of depth and ego-motion from video. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1851–1858. [Google Scholar] [CrossRef]
  5. Godard, C.; Mac Aodha, O.; Firman, M.; Brostow, G.J. Digging into self-supervised monocular depth estimation. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3828–3838. [Google Scholar] [CrossRef]
  6. Lin, H.; Peng, S.; Chen, J.; Peng, S.; Sun, J.; Liu, M.; Bao, H.; Feng, J.; Zhou, X.; Kang, B. Prompting depth anything for 4k resolution accurate metric depth estimation. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 11–15 June 2025; pp. 17070–17080. [Google Scholar]
  7. Aditya, N.; Dhruval, P. Thermal voyager: A comparative study of rgb and thermal cameras for night-time autonomous navigation. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 14116–14122. [Google Scholar] [CrossRef]
  8. Nawaz, M.; Khan, S.; Daud, M.; Asim, M.; Anwar, G.A.; Shahid, A.R.; Ho, H.P.A.; Chan, T.; Pak Kong, D.; Yuan, W. Improving Autonomous Vehicle Cognitive Robustness in Extreme Weather With Deep Learning and Thermal Camera Fusion. IEEE Open J. Vehic. Tech. 2025, 6, 426–441. [Google Scholar] [CrossRef]
  9. Lu, Y.; Lu, G. An alternative of lidar in nighttime: Unsupervised depth estimation based on single thermal image. In Proceedings of the Winter Conference on Applications of Computer Vision (WACV), Virtual, 5–9 January 2021; pp. 3833–3843. [Google Scholar] [CrossRef]
  10. Kim, N.; Choi, Y.; Hwang, S.; Kweon, I.S. Multispectral transfer network: Unsupervised depth estimation for all-day vision. In Proceedings of the Conference AAAI, New Orleans, LA, USA, 2–7 February 2018; pp. 6983–6991. [Google Scholar] [CrossRef]
  11. Shin, U.; Lee, K.; Lee, S.; Kweon, I.S. Self-supervised depth and ego-motion estimation for monocular thermal video using multi-spectral consistency loss. IEEE Robot. Autom. Lett. 2022, 7, 1103–1110. [Google Scholar] [CrossRef]
  12. Schönberger, J.L.; Zheng, E.; Frahm, J.M.; Pollefeys, M. Pixelwise view selection for unstructured multi-view stereo. In Proceedings of the European Conference, Amsterdam, The Netherlands, 11–14 October 2016; pp. 501–518. [Google Scholar] [CrossRef]
  13. Chen, W.; Fu, Z.; Yang, D.; Deng, J. Single-image depth perception in the wild. Adv. Neural Inf. Process. Syst. 2016, 29, 730–738. [Google Scholar]
  14. Eigen, D.; Puhrsch, C.; Fergus, R. Depth map prediction from a single image using a multi-scale deep network. Adv. Neural Inf. Process. Syst. 2014, 27, 2366–2374. [Google Scholar]
  15. Dong, Q.; Zhou, Z.; Qiu, X.; Zhang, L. A Survey on Self-Supervised Monocular Depth Estimation Based on Deep Neural Networks. IEEE Trans. Neural Netw. Learn. Syst. 2025, 36, 15622–15642. [Google Scholar] [CrossRef]
  16. Guizilini, V.; Ambrus, R.; Pillai, S.; Raventos, A.; Gaidon, A. 3D packing for self-supervised monocular depth estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 2485–2494. [Google Scholar] [CrossRef]
  17. Bian, J.; Li, Z.; Wang, N.; Zhan, H.; Shen, C.; Cheng, M.M.; Reid, I. Unsupervised scale-consistent depth and ego-motion learning from monocular video. Adv. Neural Inf. Process. Syst. 2019, 33, 35–45. [Google Scholar]
  18. Bian, J.; Zhan, H.; Wang, N.; Li, Z.; Zhang, L.; Shen, C.; Cheng, M.M.; Reid, I. Unsupervised scale-consistent depth learning from video. Int. J. Comput. Vis. 2021, 129, 2548–2564. [Google Scholar] [CrossRef]
  19. Mur-Artal, R.; Tardós, J.D. ORB-SLAM2: An open-source SLAM system for monocular, stereo, and RGB-D cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef]
  20. Yu, S.; Wu, M.; Lam, S.K.; Wang, C.; Wang, R. EDS-Depth: Enhancing Self-Supervised Monocular Depth Estimation in Dynamic Scenes. IEEE Trans. Intell. Transp. Syst. 2025, 26, 5585–5597. [Google Scholar] [CrossRef]
  21. Feng, C.; Zhang, C.; Chen, Z.; Hu, W.; Lu, K.; Ge, L. Self-Supervised Monocular Depth Estimation With Dual-Path Encoders and Offset Field Interpolation. IEEE Trans. Image Process. 2025, 34, 939–954. [Google Scholar] [CrossRef] [PubMed]
  22. Cheng, J.; Yin, W.; Wang, K.; Chen, X.; Wang, S.; Yang, X. Adaptive Fusion of Single-View and Multi-View Depth for Autonomous Driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024; pp. 10138–10147. [Google Scholar] [CrossRef]
  23. Wang, X.; Zhu, Z.; Huang, G.; Chi, X.; Ye, Y.; Chen, Z.; Wang, X. Crafting monocular cues and velocity guidance for self-supervised multi-frame depth learning. In Proceedings of the Conference AAAI, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 2689–2697. [Google Scholar] [CrossRef]
  24. Watson, J.; Mac Aodha, O.; Prisacariu, V.; Brostow, G.; Firman, M. The temporal opportunist: Self-supervised multi-frame monocular depth. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021; pp. 1164–1174. [Google Scholar] [CrossRef]
  25. Bae, G.; Budvytis, I.; Cipolla, R. Multi-view depth estimation by fusing single-view depth probability with multi-view geometry. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 2842–2851. [Google Scholar] [CrossRef]
  26. Zhao, G.; Wei, H.; He, H. IAFMVS: Iterative Depth Estimation with Adaptive Features for Multi-View Stereo. Neurocomputing 2025, 629, 129682. [Google Scholar] [CrossRef]
  27. Wang, L.; Liang, Q.; Che, Y.; Wang, L.; Wang, G. IFDepth: Iterative fusion network for multi-frame self-supervised monocular depth estimation. Knowl.-Based Syst. 2025, 318, 113467. [Google Scholar] [CrossRef]
  28. Wang, Y.; Chen, H.; Liu, Y.; Zhang, S. Edge-Based Monocular Thermal-Inertial Odometry in Visually Degraded Environments. IEEE Robot. Autom. Lett. 2023, 8, 2078–2085. [Google Scholar] [CrossRef]
  29. Doer, C.; Trommer, G.F. Radar visual inertial odometry and radar thermal inertial odometry: Robust navigation even in challenging visual conditions. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; pp. 331–338. [Google Scholar] [CrossRef]
  30. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar] [CrossRef]
  31. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Munich, Germany, 5–9 October 2015; Volume 9351, pp. 234–241. [Google Scholar]
  32. Wang, F.; Galliani, S.; Vogel, C.; Pollefeys, M. IterMVS: Iterative probability estimation for efficient multi-view stereo. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 8606–8615. [Google Scholar] [CrossRef]
  33. Zhang, J.; Li, S.; Luo, Z.; Fang, T.; Yao, Y. Vis-mvsnet: Visibility-aware multi-view stereo network. Int. J. Comput. Vision 2023, 131, 199–214. [Google Scholar] [CrossRef]
  34. Mur-Artal, R.; Montiel, J.M.M.; Tardos, J.D. ORB-SLAM: A versatile and accurate monocular SLAM system. IEEE Trans. Robot. 2015, 31, 1147–1163. [Google Scholar] [CrossRef]
  35. He, M.; Zhu, C.; Huang, Q.; Ren, B.; Liu, J. A review of monocular visual odometry. Vis. Comput. 2020, 36, 1053–1065. [Google Scholar] [CrossRef]
  36. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  37. Zhang, Z. Eight-Point Algorithm. In Computer Vision: A Reference Guide; Ikeuchi, K., Ed.; Springer: Cham, Switzerland, 2021; pp. 370–371. [Google Scholar] [CrossRef]
  38. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar] [CrossRef]
  39. Faugeras, O.D.; Lustman, F. Motion and structure from motion in a piecewise planar environment. Int. J. Pattern Recognit. Artif. Intell. 1988, 2, 485–508. [Google Scholar] [CrossRef]
  40. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  41. Xu, X.; Siyao, L.; Sun, W.; Yin, Q.; Yang, M.H. Quadratic video interpolation. Adv. Neural Inf. Process. Syst. 2019, 32, 1647–1656. [Google Scholar]
  42. Jaderberg, M.; Simonyan, K.; Zisserman, A.; Kavukcuoglu, K. Spatial transformer networks. Adv. Neural Inf. Process. Syst. 2015, 29, 2017–2025. [Google Scholar]
  43. Godard, C.; Mac Aodha, O.; Brostow, G.J. Unsupervised monocular depth estimation with left-right consistency. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 270–279. [Google Scholar] [CrossRef]
  44. Lee, A.J.; Cho, Y.; Shin, Y.S.; Kim, A.; Myung, H. ViViD++: Vision for visibility dataset. IEEE Robot. Autom. Lett. 2022, 7, 6282–6289. [Google Scholar] [CrossRef]
  45. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  46. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Curran Associates, Inc.: Nice, France, 2019; Volume 32. [Google Scholar]
  47. Kingma, D.P.; Ba, J.L. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  48. Zhang, Z.; Scaramuzza, D. A tutorial on quantitative trajectory evaluation for visual (-inertial) odometry. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 7244–7251. [Google Scholar] [CrossRef]
Figure 1. (a) The overview of our framework. (b) The RGB–Thermal image pair at the test time. (c) The depth and error map from the baseline model [11], with smaller errors represented in blue and larger errors in red. (d) The depth and error map from our method, showing superior accuracy.
Figure 1. (a) The overview of our framework. (b) The RGB–Thermal image pair at the test time. (c) The depth and error map from the baseline model [11], with smaller errors represented in blue and larger errors in red. (d) The depth and error map from our method, showing superior accuracy.
Applsci 15 11890 g001
Figure 2. The main architecture of the proposed framework. (a) Utilizing PoseNet to estimate ego-motion between thermal frames, and integrating the pose generated from the motion enhancement module to compute the motion loss. (b) Employing DepthNet to predict the depth of the single thermal frame and after depth warping. (c) The encoded thermal feature maps are used to construct a cost volume with the assistance of motion hints. The cost volume is then decoded into the MVS depth and confidence map. Each generated depth map can be used to calculate multi-spectral losses. The confidence map is instrumental in fusing the single-frame depth and MVS depth, resulting in a more refined fused depth.
Figure 2. The main architecture of the proposed framework. (a) Utilizing PoseNet to estimate ego-motion between thermal frames, and integrating the pose generated from the motion enhancement module to compute the motion loss. (b) Employing DepthNet to predict the depth of the single thermal frame and after depth warping. (c) The encoded thermal feature maps are used to construct a cost volume with the assistance of motion hints. The cost volume is then decoded into the MVS depth and confidence map. Each generated depth map can be used to calculate multi-spectral losses. The confidence map is instrumental in fusing the single-frame depth and MVS depth, resulting in a more refined fused depth.
Applsci 15 11890 g002
Figure 3. Detailed architecture of the multi-frame depth estimation part. It consists of three main components: (1) a shared MVS encoder for feature extraction; (2) a module that constructs a cost volume using motion hint-based sampling and feature warping; and (3) a 3D Convolutional Neural Network (CNN)-based MVS decoder that infers depth and a confidence map for the final fusion step.
Figure 3. Detailed architecture of the multi-frame depth estimation part. It consists of three main components: (1) a shared MVS encoder for feature extraction; (2) a module that constructs a cost volume using motion hint-based sampling and feature warping; and (3) a 3D Convolutional Neural Network (CNN)-based MVS decoder that infers depth and a confidence map for the final fusion step.
Applsci 15 11890 g003
Figure 4. In the upper branch, the motion enhancement module constructs epipolar geometry constraints between matched keypoints in thermal frames and computes the essential matrix ( E ) and homography matrix ( H ) using the two algorithms. The suitable matrix is selected according to the scene structure, and the camera pose is then computed. The lower branch utilizes PoseNet to generate poses between the thermal frames. Ultimately, the two poses are used to calculate motion loss.
Figure 4. In the upper branch, the motion enhancement module constructs epipolar geometry constraints between matched keypoints in thermal frames and computes the essential matrix ( E ) and homography matrix ( H ) using the two algorithms. The suitable matrix is selected according to the scene structure, and the camera pose is then computed. The lower branch utilizes PoseNet to generate poses between the thermal frames. Ultimately, the two poses are used to calculate motion loss.
Applsci 15 11890 g004
Figure 5. Preprocessing of thermal images. The details are compressed in the raw thermal image. Clamping, adjustment, and remapping operations are performed based on the temperature range to make efficient use of the thermal image information.
Figure 5. Preprocessing of thermal images. The details are compressed in the raw thermal image. Clamping, adjustment, and remapping operations are performed based on the temperature range to make efficient use of the thermal image information.
Applsci 15 11890 g005
Figure 6. Qualitative comparison of depth prediction under varying illumination. The top row shows a daytime scene, while the bottom row depicts a challenging nighttime scene. Note the severe degradation of the RGB-based depth (e) at night, whereas our thermal-based approach (c) remains robust.
Figure 6. Qualitative comparison of depth prediction under varying illumination. The top row shows a daytime scene, while the bottom row depicts a challenging nighttime scene. Note the severe degradation of the RGB-based depth (e) at night, whereas our thermal-based approach (c) remains robust.
Applsci 15 11890 g006
Figure 7. Qualitative comparison results of depth prediction on the VIVID dataset. We present (a) RGB scene images, (b) thermal images, (c) depth maps generated by our method, (d) depth maps generated by the baseline, (e) error maps from our method, and (f) error maps from the baseline. The visualized error maps are derived from the absolute relative errors between the predicted depth and the ground-truth provided by LiDAR, where cool colors (blue) indicate smaller errors and warm colors (red) represent larger errors [11].
Figure 7. Qualitative comparison results of depth prediction on the VIVID dataset. We present (a) RGB scene images, (b) thermal images, (c) depth maps generated by our method, (d) depth maps generated by the baseline, (e) error maps from our method, and (f) error maps from the baseline. The visualized error maps are derived from the absolute relative errors between the predicted depth and the ground-truth provided by LiDAR, where cool colors (blue) indicate smaller errors and warm colors (red) represent larger errors [11].
Applsci 15 11890 g007
Figure 8. Qualitative visualization of the feature-matching process in our motion enhancement module. Robust feature correspondences (inliers shown as connecting lines) are successfully extracted from two consecutive thermal frames. These matches provide the geometric constraints for the self-generated supervisory signal.
Figure 8. Qualitative visualization of the feature-matching process in our motion enhancement module. Robust feature correspondences (inliers shown as connecting lines) are successfully extracted from two consecutive thermal frames. These matches provide the geometric constraints for the self-generated supervisory signal.
Applsci 15 11890 g008
Table 1. Quantitative comparison between our method and other self-supervised methods for depth prediction on the VIVID dataset. The evaluation encompasses scenes from daytime and nighttime, testing each method’s performance under varying light conditions. The “Input” column specifies the types of images fed into the network. “V” and “T” represent visible RGB images and thermal images, respectively. The table highlights results with the minimum errors and the highest accuracy in bold.
Table 1. Quantitative comparison between our method and other self-supervised methods for depth prediction on the VIVID dataset. The evaluation encompasses scenes from daytime and nighttime, testing each method’s performance under varying light conditions. The “Input” column specifies the types of images fed into the network. “V” and “T” represent visible RGB images and thermal images, respectively. The table highlights results with the minimum errors and the highest accuracy in bold.
SceneMethodsInputError (Lower Is Better)Accuracy (Higher Is Better)
Abs. Rel.Sq. Rel.RMSERMSElog δ < 1.25 δ < 1.25 2 δ < 1.25 3
Outdoor
day1
SfMLearner [4]V0.2001.5496.3940.2480.6840.9220.975
Monodepth2 [5]V0.4366.3189.9670.4540.4910.7190.840
Bian et al. [17]V0.1361.4645.6380.1790.8440.9770.991
Watson et al. [24]V0.1661.4726.3580.2200.7540.9450.986
Shin et al. [11] (T)T0.1430.7664.5470.1800.7900.9790.996
Shin et al. [11] (MS)T0.1420.7924.7300.1810.7880.9800.996
OursT0.1150.6834.3100.1590.8690.9800.996
Outdoor
day2
SfMLearner [4]V0.1921.6526.5110.2350.7020.9410.978
Monodepth2 [5]V0.4356.9159.5760.4380.5910.7520.843
Bian et al. [17]V0.1311.3765.8990.1850.8220.9760.991
Watson et al. [24]V0.1972.2817.7070.2410.6950.9220.982
Shin et al. [11] (T)T0.1480.9334.7360.1880.8020.9700.991
Shin et al. [11] (MS)T0.1450.9174.7840.1870.8040.9710.991
OursT0.1220.9574.7870.1700.8550.9680.991
Outdoor
night1
SfMLearner [4]V0.4294.5848.6240.4450.4680.6980.834
Monodepth2 [5]V0.70411.7512.530.6360.3620.5590.701
Bian et al. [17]V0.5206.41310.400.5160.3810.5960.755
Watson et al. [24]V0.4695.69010.960.4660.3320.6120.839
Shin et al. [11] (T)T0.1580.8444.6340.1920.7540.9770.996
Shin et al. [11] (MS)T0.1560.8564.8130.1920.7520.9760.996
OursT0.1190.7474.1890.1620.8500.9790.996
Outdoor
night2
SfMLearner [4]V0.3734.2158.2940.3960.5480.7730.879
Monodepth2 [5]V0.60210.8411.720.5620.4770.6500.759
Bian et al. [17]V0.4646.3769.8870.4720.5110.6850.807
Watson et al. [24]V0.4164.86010.160.4280.3830.6920.873
Shin et al. [11] (T)T0.1591.0845.1150.2040.7720.9570.989
Shin et al. [11] (MS)T0.1561.0495.1660.2020.7750.9570.989
OursT0.1240.8994.6940.1790.8290.9630.990
Table 2. Quantitative comparison of our method to other self-supervised methods for pose estimation. The comparison methods include the trained RGB-based self-supervised methods, and the scenes encompass both daytime and nighttime situations. Smaller errors in the table indicate better performance. Results with the minimum errors are highlighted in bold.
Table 2. Quantitative comparison of our method to other self-supervised methods for pose estimation. The comparison methods include the trained RGB-based self-supervised methods, and the scenes encompass both daytime and nighttime situations. Smaller errors in the table indicate better performance. Results with the minimum errors are highlighted in bold.
MethodsInputOutdoor Day1Outdoor Day2Outdoor Night1Outdoor Night2
MeanStd.MeanStd.MeanStd.MeanStd.
SfMLearner [17]V0.07740.04070.08670.04140.06570.03420.05790.0279
Monodepth2 [5]V0.05250.03050.05440.02660.05520.03020.05450.0277
Bian et al. [17]V0.05030.02550.05140.02850.08860.04390.08880.0390
Shin et al. [11] (T)T0.07510.03710.07840.03910.07440.04170.07930.0402
Shin et al. [11] (MS)T0.05410.03070.06430.03650.05900.03150.06040.0315
OursT0.04290.02510.04500.02480.04640.02880.04420.0252
Table 3. Ablation study on various components. The best results are marked in bold.
Table 3. Ablation study on various components. The best results are marked in bold.
MethodsAbs. Rel.RMSE δ < 1.25
Ours (full)0.1274.3050.831
Ours (w/o depth warping)0.1444.7290.815
Ours (w/o motion enhancement)0.1384.7630.824
Ours (w/o motion hint)0.1404.7980.816
Ours (w/o remapping)0.1404.8320.813
Ours (w/o depth fusing)0.1364.4780.816
Table 4. Ablation study on various loss terms. The experiment labeled “w/o RGB loss” indicates setting the reconstruction loss based on photometric consistency to zero. The best results are marked in bold.
Table 4. Ablation study on various loss terms. The experiment labeled “w/o RGB loss” indicates setting the reconstruction loss based on photometric consistency to zero. The best results are marked in bold.
MethodsAbs. Rel.RMSE δ < 1.25
Ours (full)0.1274.3050.831
Ours (w/o thermal loss)---
Ours (w/o motion loss)0.1384.7630.824
Ours (w/o smoothness loss)0.1404.7420.799
Ours (w/o RGB loss)0.1444.7290.815
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, R.; Ma, G.; Guo, J.; Xu, L. Self-Supervised Depth and Ego-Motion Learning from Multi-Frame Thermal Images with Motion Enhancement. Appl. Sci. 2025, 15, 11890. https://doi.org/10.3390/app152211890

AMA Style

Yu R, Ma G, Guo J, Xu L. Self-Supervised Depth and Ego-Motion Learning from Multi-Frame Thermal Images with Motion Enhancement. Applied Sciences. 2025; 15(22):11890. https://doi.org/10.3390/app152211890

Chicago/Turabian Style

Yu, Rui, Guoliang Ma, Jian Guo, and Lisong Xu. 2025. "Self-Supervised Depth and Ego-Motion Learning from Multi-Frame Thermal Images with Motion Enhancement" Applied Sciences 15, no. 22: 11890. https://doi.org/10.3390/app152211890

APA Style

Yu, R., Ma, G., Guo, J., & Xu, L. (2025). Self-Supervised Depth and Ego-Motion Learning from Multi-Frame Thermal Images with Motion Enhancement. Applied Sciences, 15(22), 11890. https://doi.org/10.3390/app152211890

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop