Next Article in Journal
ILViT: An Inception-Linear Attention-Based Lightweight Vision Transformer for Microscopic Cell Classification
Previous Article in Journal
Imaging Evaluation of Periarticular Soft Tissue Masses in the Appendicular Skeleton: A Pictorial Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Parallel Multi-Scale Semantic-Depth Interactive Fusion Network for Depth Estimation

1
Department of Computer Science and Engineering, Southeast University, Nanjing 210000, China
2
North Information Control Research Academy Group Co., Ltd., Nanjing 210000, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Imaging 2025, 11(7), 218; https://doi.org/10.3390/jimaging11070218
Submission received: 11 May 2025 / Revised: 8 June 2025 / Accepted: 24 June 2025 / Published: 1 July 2025

Abstract

Self-supervised depth estimation from monocular image sequences provides depth information without costly sensors like LiDAR, offering significant value for autonomous driving. Although self-supervised algorithms can reduce the dependence on labeled data, the performance is still affected by scene occlusions, lighting differences, and sparse textures. Existing methods do not consider the enhancement and interaction fusion of features. In this paper, we propose a novel parallel multi-scale semantic-depth interactive fusion network. First, we adopt a multi-stage feature attention network for feature extraction, and a parallel semantic-depth interactive fusion module is introduced to refine edges. Furthermore, we also employ a metric loss based on semantic edges to take full advantage of semantic geometric information. Our network is trained and evaluated on KITTI datasets. The experimental results show that the methods achieve satisfactory performance compared to other existing methods.

1. Introduction

With the technological breakthroughs in the field of computer vision, research fields such as autonomous driving and 3D reconstruction have also developed rapidly. These fields inevitably apply visual technology to perceive three-dimensional space to assist decision making. The application of 3D vision adds depth information to 2D, paying more attention to information recognition and substance understanding in a 3D space. In addition to directly using sensors to obtain the information on depth, attempts have been made to obtain 3D data directly from 2D data. Depth estimation is widely applied in various fields, playing a key role in robot navigation and autonomous driving scene reconstruction.
Derived from the multi-view stereo geometry theory, the principle of depth estimation is closely related to the natural biological binocular vision system. Early depth estimation methods mainly relied on matching pixels from images obtained from multiple and calculating the depth of pixels through triangular matching of feature points. With the rapid development of deep learning, people have proposed methods based on deep learning. Through continuous training on datasets, deep neural networks can learn the relationship between images and depth. Depth estimation methods based on deep learning are divided into two categories—supervised learning and self-supervised learning. Supervised learning methods [1,2,3] require a large number of manually annotated depth image pairs as training data, whose generalization ability is particularly limited by the size and diversity of the dataset. According to the theory of structure from motion, the idea of self-supervised learning to estimate the depth and the pose from the video sequence has received attention in scientific research [4,5]. Image re-projection is performed using the estimated depth and pose. Then the photometric loss between the re-projected image and the target image provides a supervisory signal for the training of the neural network.
In recent years, many new methods have emerged in the field of self-supervised monocular depth estimation, and the results of models have been improved on Zhou’s baseline [4]. These methods [5,6] mainly rely on pixel loss and smoothness loss, resulting in the loss of weakly textured regions. What is more, tiny objects in the scene cannot be correctly estimated. Godard et al. [6] introduced a minimum re-projection loss and auto uncertainty masking technique to improve training accuracy. Although many studies on self-supervised monocular depth estimation come out one after the other, the accuracy of the results is still affected by dynamic objects, scene occlusions, lighting differences, and sparse textures.
The combination of semantic information and depth information can be used as supplementary knowledge of 3D space, thereby improving the performance of depth estimation. Considering the relevance of semantic segmentation and depth estimation tasks [7,8], we propose a novel multi-task self-supervised monocular video depth estimation pipeline. Taking unlabeled video sequences and the semantic labels generated by a pre-trained model as input, we design a parallel multi-scale semantic-depth interactive fusion network. The features of two tasks at different scales can achieve continuous transmission and parallel interaction, which improves the accuracy of depth estimation.
Although the complementary information in the feature space of different tasks enables the network to autonomously emphasize feature information, the implicit feature interaction fusion method may introduce some feature noise. Therefore, we actively mine the information of edges in semantic images and add a metric loss based on semantic edges to the photometric loss to take full advantage of semantic geometric information. Metric samples are obtained according to semantic boundaries, and the strategy of boundary sample classification based on the reconstruction of semantic images of adjacent frames is designed to increase the robustness of the metric loss.
Models trained on our pipeline outperform most other recent works. The main innovations of the paper are summarized as follows.
  • A scheme that combines semantic segmentation to estimate depth is proposed to implement an end-to-end pipeline. We adopt a multi-stage feature attention network (MSFAN) for the feature extraction of RGB images instead of the traditional U-net structure encoder to improve the accuracy of the depth estimation task.
  • We introduce a semantic segmentation task by sharing feature extraction network MSFAN, and add a parallel semantic depth interactive fusion module (PSDIFM) to achieve the bidirectional complementarity of feature information between different tasks.
  • The total multi-task loss function is designed to adapt to the new pipeline, and the metric loss based on semantic edges is added to refine the depth of the edges, promoting the further improvement of depth estimation results.
  • Our network pipeline is trained on KITTI dataset and evaluated. The results of KITTI and Make3D datasets show that the network pipeline designed by us achieves satisfactory performance compared to other existing methods.

2. Related Work

2.1. Depth Estimation

Before the invention of neural networks, stereo algorithms were used for depth estimation of multiple images. After that, Eigen et al. [1] applied neural networks to introduce depth estimation from a single image by training the network on sparse labels provided by LiDAR scans. This work was refined in the following studies by improving training techniques and frameworks. With the application of fully convolutional neural networks, Laina et al. [9] proposed a fully convolutional network based on residual structure and used a pre-trained encoder for feature extraction, which improved the resolution of results. Mancini et al. [10] used fully convolutional networks and optical flow labels to detect the obstacle. However, existing approaches still suffer from limitations in sensor data quality, insufficient cross-scene generalization, and challenges in modeling complex dynamic environments.

2.2. Self-Supervised Depth Estimation

The bottlenecks of supervised methods are poor generalization performance and difficulty in obtaining ground truth of depth. On the other hand, stereo depth estimation relies on camera extrinsic parameters and complex networks, thereby, self-supervised monocular depth estimation is beginning to be studied. Zhou et al. [4] first proposed a monocular-trained framework combined by a depth network and a pose network that trained on sequences of video frames. Aiming to solve the problems of moving objects and occlusion, Godard et al. [6] designed an auto masking and brought out Monodepth2 framework, which has been one of the most frequently used baselines. Yang et al. [11] introduced a normal vector of surface and consistency constraints of the normal vector of depth to improve the performance. Mahjourian et al. [5] extracted the information on adjacent frames and proposed to use the ICP loss to make the depth results in adjacent frames have stronger consistency. For the uncertainty of depth estimation, Poggi et al. [12] designed a novel method. The uncertainty is estimated by image flipping and integrated, and the statistical mean and variance in the image are estimated according to different models. Many thoughts have been carried out on previous works according to the model architecture [13] and loss functions [14,15] and have achieved significant improvement. Despite these advances, challenges persist in handling intricate motion patterns, robustness to abrupt illumination changes, and the inability to recover absolute metric scale.

2.3. Semantic-Guided Depth Estimation

Semantic segmentation [16,17] is to assign a predefined semantic label to each pixel in the image. Fully convolutional and encoder–decoder structures are used for feature extraction and refinement segmentation, resulting in significantly improved segmentation performance. Due to the similarity of depth estimation and semantic segmentation tasks, semantic information has been introduced to improve depth estimation performance. A scheme to handle moving dynamic objects to avoid miscalculation of photometric losses is proposed in [8]. Chen et al. [18] designed a multi-task framework with shared encoders to perform semantic segmentation and depth estimation with a consistent structure. Choi et al. [19] designed different fusion modules between decoder network layers. Zhu et al. [20] proposed a measure of boundary consistency between segmentation and depth, and Jung et al. [21] designed a semantics-guided metric loss. They explicitly guide the training of depth edges with semantic edges to provide a new signal for optimizing depth results. However, these methods do not fully utilize the feature information in encoder and decoder networks, resulting in the loss of information during network transmission. To make up for that, the pipeline we designed takes into account the parallel interaction of different scale features in different tasks. Inspired by Jung et al. [21], the metric loss based on the strategy of boundary sample classification is added to the photometric loss, improving the accuracy of depth estimation.

3. Proposed Approach

3.1. Problem Statement

Self-supervised monocular depth estimation computes depth values for camera images on a pixel-by-pixel basis without using any ground truth labels. Given three consecutive frames of a video sequence, they are numbered in time order as I t 1 , I t and I t + 1 . We also refer to I t as the target image, I t 1 and I t + 1 as the source images I s . According to Monodepth2 [6], the pipeline is trained with both the depth and six degree-of-freedom relative poses of the source and target images T t s . Based on the world static assumption that the view change in driving scenes is only caused by a moving camera, we can synthesize the corresponding frames of the target image using the source image and the relative poses. With known camera intrinsics K, we can use the following equation to calculate the projected pixel coordinates:
p s K T t s D ^ t ( p t ) K 1 p t ,
where p t is the homogeneous coordinates of the pixel in the target image and p s is the re-projected coordinates of p t by T t s . Through bilinear sampling, the surrounding pixels of p s in I s can be obtained, and is calculated as the pixel value of the corresponding position of the synthetic target image. Based on the theory of structure from motion, the synthesized image I ^ t should be consistent with the original image I t . We calculate the minimized photometric loss using the structural similarity index metric in combination with the L 1 pixel loss.

3.2. Network Architecture

3.2.1. Pipeline

The pipeline of the parallel multi-scale semantic-depth interactive fusion network we designed is shown in Figure 1. The input of the depth estimation task and semantic segmentation task is both single RGB images. Two tasks share a multi-stage feature attention network (MSFAN) and output multiple scales of features. More feature details can be preserved at each resolution, and the parameter redundancy caused by repeated calculations can be effectively avoided. In the next feature fusion process, we designed a parallel semantic-depth interactive fusion module (PSDIFM). The depth and semantic prediction tasks each uses separate network branches to fuse features at multiple scales level by level, and features of different scales are interacted and fused between tasks. At the same time, two adjacent frames of images I s and I t are input to the PoseNet to obtain a 6DoF pose for the calculation of photometric loss.
The boundary set B t of the image is extracted from the semantic pseudo-label, and then the metric loss L m e t r i c of the depth feature at the corresponding position is calculated according to the feature comparison idea in the metric learning method, which realizes the explicit guidance of the depth by the semantic boundary. The semantic segmentation task uses only the semantic pseudo-labels generated by the pre-trained model. The labels are only used to distinguish objects of different semantics and do not need to represent specific categories. This can reduce the reliance on real labels for semantic tasks and make the training data more uniform without introducing additional datasets.

3.2.2. Multi-Stage Feature Attention Network

Many existing depth estimation approaches that obtain high-resolution feature maps use the method of first reducing and then increasing the resolution. The networks are usually built on ResNet to convert images into low-resolution features, similar to U-Net, DeconvNet, etc. Although low-resolution features have rich semantics, images lose spatial structure information in the process of feature extraction. In previous work, we can infer that the performance of the model on the task of depth estimation depends to some extent on the resolution of the input image, i.e., the results of high-quality images are usually better than those of low-quality ones. Inspired by dense prediction tasks such as semantic segmentation, multi-level parallel feature extraction networks can better preserve features at different resolutions. HR-Depth [22] has applied multi-level feature extraction to the network and confirmed the effectiveness of the method, but lacks the interactive fusion of different stages.
HRNet [16] parallelizes feature maps of different resolutions, and adds interactions to feature maps of different resolutions in forward convolution stages. Inspired by the framework, we adopt a multi-stage feature attention network (MSFAN) instead of the classic U-Net structure including a multi-stage details enhancement module. As shown in Figure 2, the network has four data streams and four stages, each of which fuses information between different scales, while extracting the feature maps, we observed a significant improvement in other methods using ResNet as the backbone.
Feature maps are transmitted in parallel at different levels, and as the network is extended, the feature information from the previous stages cannot be effectively preserved and utilized. To enrich the feature information in the final output, the network first stitches multi-stage features together and then performs detail enhancement using the idea of an attention mechanism. Compared to traditional resolution-reducing encoders, this multi-scale parallel design enables better preservation of spatial detail, particularly near depth boundaries. However, maintaining multiple resolution streams throughout the network introduces moderate computational and memory overhead. This is a conscious design trade-off in favor of accuracy over efficiency. In Section 4.3, we provide empirical comparisons demonstrating that our method maintains competitive performance under acceptable resource usage. The traditional spatial attention and channel attention lack the consideration of the relationship between features and coordinate positions in two-dimensional images. Global pooling is used in the channel attention mechanism to encode the spatial information as channel descriptors, making it difficult to preserve location information. Unlike traditional channel attention, we adopt a coordinate attention mechanism [23] to emphasize the interdependence between location information. This is particularly useful in-depth estimation, where depth discontinuities often align with semantic or geometric boundaries. Let f l i denote the feature map from the i-th stage at resolution level l, and ε ( . ) denote channel-wise concatenation. The output feature f l at level l is defined as follows:
f l = C ( f l 0 ) , l = 0 C ( ε ( f l l , f l l + 1 , , f l 4 ) ) , l { 1 , 2 , 3 , 4 }
where C ( . ) represents the coordinate attention module applied to the concatenated features. This allows the network to enhance important features at each resolution stage by leveraging both spatial and contextual cues.

3.2.3. Parallel Semantic Depth Interactive Fusion

In multi-task training, two tasks with high similarity can be considered as the primary and secondary task, and the secondary task can provide more feature information to the primary task [24]. Features at different scales of the auxiliary task have different effects on the main task due to the different receptive field sizes of features. For example, local semantic maps can provide useful information for depth prediction and improve the depth estimation of object edges. On the other hand, local depth maps provide less information about the semantics of the scene, but when the receptive field is expanded, depth maps reveal the shapes of the object, implying semantic information about the scenes.
We designed a parallel semantic-depth interactive fusion module, which allows the features of a task to interact between multiple scales. The interacted features can be used for feature combinations at other scales. In general, this approach can provide more adequate feature information for the training of each task and improve the network performance. In Figure 3, the parallel fusion module of depth estimation task is shown in blue and the parallel fusion module of semantic segmentation is shown in orange. They interact with features at four scales by a designed feature interactive fusion mechanism. The depth estimation task outputs four fused features F 0 , F 1 , F 2 , and F 3 , and the final predicted depth map of the network can be obtained after the convolution layer and activation function. The semantic segmentation task outputs a semantic graph S with the same size as the input image.
As shown in Figure 3, in the depth estimation task, the smaller-sized feature F F l + 1 is first upsampled, and is spliced with f l , and then pass through convolutional and relu layers to obtain the fused feature F e a t l on a single task. Similarly, in the semantic segmentation task, F S l + 1 is upsampled and spliced with s l , and the S e g l feature is output after passing through convolutional and activation layers. The F e a t l and S e g l can be calculated as follows:
F e a t l = F ( ε ( U ( F F l + 1 ) , f l ) ) , l { 0 , 1 , 2 , 3 } f l , l = 4
S e g l = F ( ε ( U ( F S l + 1 ) , f l ) ) , l { 0 , 1 , 2 , 3 } s l , l = 4
where F ( . ) is a block consisting of convolution operation followed by an activation, U ( . ) is an upsampling block.
The parallel interactive fusion mechanism is designed with reference to the idea of spatial self-attention [25]. The Spatial Attention Module (SAM) can be used as a gate function to control the information flow and enable the network to refine useful information autonomously. F l is added with S l distilled by SAM in the semantic segmentation task and will be used for feature fusion in the upper layer. Meanwhile, S l is also distilled using SAM and added with F l in the semantic network.
F F l = F e a t l S A M ( S e g l ) , l { 0 , 1 , 2 , 3 } F e a t l , l = 4
F S l = S e g l S A M ( F e a t l ) , l { 0 , 1 , 2 , 3 } S e g l , l = 4
The features F l that meet the task requirements after fusion can be calculated from F F l , and the final output depth map can be obtained after the convolution layer and activation function.
F l = F ( U ( F F l ) ) , l = 0 F F l 1 , l { 1 , 2 , 3 }

3.3. Boundary Alignment Loss

One of the optimization goals of the depth estimation task is to make the edges of objects in the depth map more clear. After adding the semantic segmentation task, the correspondence between the depth map and the object edges in the semantic map can be established, so that the boundaries of depth images are aligned with semantic images. Since the pixels on the semantic boundary may have large depth differences, metric loss can be calculated inspired by recent work [21], using depth feature maps within the neighborhood locations of the semantic boundary. Based on this idea, our work designs a boundary alignment loss to further enhance the use of semantic boundaries.
When extracting boundary pixel samples, a neighborhood of size Z × Z is constructed with each pixel i on the semantic map as the center. In each neighborhood, the pixels belonging to the same semantic category as the central pixel are marked as a set of positive pixels P i + , and the pixels with different semantic categories are marked as a set of negative pixels P i . The set of boundary pixels B t needs to satisfy the formula as follows:
| c o u n t ( P i + ) c o u n t ( P i ) | T
As shown in Figure 4, both the neighborhood length Z and the threshold T significantly affect the extracted boundary maps. Based on visual quality and consistency with object contours, we empirically set T = 10 in all experiments. This value strikes a balance between suppressing weak/noisy gradients and preserving meaningful edge structures. Further evaluation of sensitivity to T will be considered in future work.
For three consecutive images I t 1 , I t and I t + 1 , we have corresponding semantic images S t 1 , S G T and S t + 1 . By the method of view synthesis, we obtain the synthesized semantic maps corresponding to I t , which are denoted as S ^ t ( t 1 ) and S ^ t ( t + 1 ) . As shown in Figure 5, semantic inconsistency is mainly distributed at the boundary of the image, which is caused by the depth estimation task and the pose prediction task. It can be used as a self-supervised signal to optimize the training of the whole network.
In Algorithm 1, B c refers to the predicted depth boundary pixels, extracted by computing gradient edges from the estimated depth map. B c t denotes the predicted semantic boundary pixels, computed from the semantic segmentation output. S G T is the semantic ground truth label map, used to identify label transitions in the neighborhood of depth edges. On the boundary sample set B t , a classification strategy based on semantic reconstruction between frames is adopted. The set of boundary samples B t are divided into basic boundary set B g t and semantically inconsistent boundary set B c t . A higher weighted metric loss is computed for the semantically inconsistent boundary pixels to enhance the effectiveness of the metric loss. Algorithm 1 describes the extraction process of B c t . B g t is calculated by Equation (9).
B g t = B t B c t
The feature map can be normalized as F ^ l = F l F l .
Algorithm 1: Semantic Inconsistent Boundary Pixel Extraction
Input: Synthetic semantic maps S ^ t ( t + 1 ) and S ^ t ( t 1 ) , Pseudo label S G T , Semantic boundary pixel set B t
Output: Semantically inconsistent boundary pixel set B c t
1:
Initialize the set of semantically inconsistent pixels B c
2:
for all each pixel coordinates p t in I t do
3:
    Obtain pixel categories: S ^ t ( t + 1 ) ( p t ) , S ^ t ( t 1 ) ( p t )
4:
    Obtain pixel class S G T ( p t ) on the pseudo-label map
5:
    if S ^ t ( t + 1 ) ( p t ) S G T ( p t ) then
6:
        Add pt to Bc
7:
    end if
8:
    if p t B c and S ^ t ( t 1 ) ( p t ) S G T ( p t ) then
9:
        Add pt to Bc
10:
    end if
11:
end for
12:
B c t B c B t
13:
return Bct
We grouped the features in each patch of the depth feature map into three classes (i.e., anchor, positive, and negative) following the corresponding pixel locations in the semantic image patch. We defined positive distance d + ( i ) and negative distance d ( i ) as the mean of the cosine distance.
d + ( i ) = 1 | P + | j P i + 2 2 F ^ l ( i ) F ^ l ( j ) d ( i ) = 1 | P | j P i 2 2 F ^ l ( i ) F ^ l ( j )
The triplet metric loss [26] is calculated separately on the two sample point sets B g t and B c t . For the points in the base boundary sample point set B g t , the metric loss on the depth features of the l-th layer is as follows:
L m e t r i c g t ( l ) = i B g t max ( d + ( i ) d ( i ) + m , 0 )
Similarly, for the points in the semantically inconsistent boundary sample point set B c t , the metric loss on the depth features of the l-th layer is as follows:
L m e t r i c c t ( l ) = i B c t max ( d + ( i ) d ( i ) + m , 0 )
The total boundary alignment loss can be calculated as:
L m e t r i c ( l ) = γ 1 L m e t r i c g t ( l ) + γ 2 L m e t r i c c t ( l ) | B g t | + | B c t |

3.4. Multi-Task Loss

The network needs to calculate the loss on the depth estimation results at four different scales. From the source image I s , the pose T t s and the depth value D ^ t calculated by the depth network, the synthesized target image I ^ t can be calculated. If the depth and pose are accurate, the images I ^ t and I t should be consistent at the pixel level, so the pixel difference between them can be used to train the network. Denoting the pixel error between images as L p i x e l we have
L p i x e l = min I s p e ( I t , I ^ t )
In Equation (14), p e ( . ) represents the pixel photometric error consisting of L 1 loss and Structured Similarity (SSIM) loss. The L 1 loss calculates the difference in each valid pixel point. The structured similarity function measures the similarity between two images. p e ( . ) can be calculated as follows:
p e = α 2 ( 1 S S I M ( I t , I ^ t ) ) + ( 1 α ) I t I ^ t 1
In real-world scenes, not all scenes conform to the static assumption that the presence of moving objects in the scene can interfere with pixel errors. We apply a stationary pixel mask [6] to filter the stationary pixels that maintain the same appearance between the front and back frames. The pixel mask M p i x e l is calculated in Equation(16).
M p i x e l = min I s p e ( I t , I ^ t ) < min I s p e ( I t , I s )
For image I t and the corresponding predicted depth D ^ t , the edge smoothing loss [6] is calculated as in Equation (17), with x and y denoting the horizontal and vertical gradients. L s m o o t h can make the gradient of depth and RGB images consistent.
L s m o o t h ( I t , D ^ t ) = | x D ^ t | e | x I t | + | y D ^ t | e | y I t |
The losses on a single scale can be calculated as follows:
L s i n g l e = μ M p i x e l L p i x e l + λ L s m o o t h ( I t , D ^ t )
In order to reduce the influence of noise on the loss calculation, it is necessary to assign different weights to different scales of losses. The total loss is calculated by Equation (19).
L b a s e = 1 | S c a l e | l S c a l e ( 1 2 l L s i n g l e l ( I t , I ^ t ) )
In this study, a pre-trained model is used to construct corresponding semantic labels for RGB images in video sequences, and the difference between the semantic results predicted by the semantic network branches and the pseudo-labels is calculated as the cross-entropy loss L C E = i = 1 n y i log ( y ^ i ) to train the network. To achieve end-to-end training, the loss of the multitask pipeline is calculated as follows:
L m u l t i = μ 1 L b a s e + μ 2 L C E + μ 3 L m e t r i c

4. Experiments

4.1. Datasets and Evaluation Metrics

KITTI. The KITTI dataset [19] has been implemented in training and evaluation, which contains RGB images captured by cameras and real depth data captured by radar scanning equipment. Since the dataset lacks semantic labels, existing semantic model has been adopted generating labels for each image. We use the KITTI Eigen split [1] and remove all static frames that are the same with related works [4,6]. This gives rise to 39,910 samples for training and 4424 samples for validation. The test dataset has a total of 697 samples.
Make3D. The Make3D dataset [27] contains 400 single training RGB with depth map pairs and 134 test samples for depth estimation tasks. The RGB images have high resolution, while the depth maps are provided at low resolution by laser camera.
For semantic supervision, we use pseudo-labels generated by a DeepLabv3+ model with a ResNet-101 backbone, pre-trained on the Cityscapes dataset. Although this model is not fine-tuned on KITTI or Make3D, it produces reasonably accurate segmentation results in urban scenes. Some domain mismatches may still occur, particularly in non-driving environments like those in Make3D, leading to noisy semantic labels in certain regions. The evaluation metrics in this paper are consistent with previous works [4,6]. For mean absolute relative error (AbsRel), square relative error (SqRel), root mean square error (RMSE), and root mean square logarithmic error (RMSE_log), lower is better. For the accuracy under threshold ( δ < 1.25 i , i = 1 , 2 , 3 ), higher is better. All depth images are calculated within 80 m during validation. Because the self-supervised models trained on monocular video sequences cannot recover the true depth, we use the same scale factor as Monodepth2 [6] to determine the true scale information on depth values.

4.2. Network and Training Details

The code implementation is based on PyTorch 2.7.1 library trained with NVIDIA 3090 GPU and Intel(R) Core(TM) i7-9700F CPU @ 3.00 GHz. The size of each input image is 192 × 640 pixels and we do random color jitter augmentation while loading the dataset. We initiate the parameters of some high-resolution representation layers with the HRNet [16] model pre-trained on Cityscapes, and the depth and segmentation layers are randomly initialized.
The network is trained for 20 epoches by the Adam optimizer and the learning rate is set to 10 4 . During training, we set the batch size to 12. We set α to 0.85 in Equation (15), and the weights ( μ 1 , μ 2 , μ 3 ) in the total loss are set to 1.0, 0.2, and 0.02 in Equation (20), respectively. These values were chosen based on empirical tuning to balance the magnitude and impact of each loss term. A larger μ 2 or μ 3 was found to overemphasize auxiliary tasks (semantic segmentation and edge consistency), leading to reduced depth accuracy. This choice provides a stable convergence behavior and the best validation performance.

4.3. Quantitative and Qualitative Results

Evaluation on KITTI Dataset. We evaluated the main performance of the KITTI Eigen split and the quantitative results are shown in Table 1. We compare them to the methods that use monocular image sequences, trained with or without the segmentation task: Y means with segmentation and N means without the task. The evaluation results represent our best practice for the depth estimation task, which outperforms other self-supervised approaches. Compared with works of single task, our proposed approach shows a great improvement, which reflects the effectiveness of the multi-stage feature attention network and the guidance of segmentation task. The shared feature extraction network strengthens information through the transmission of the parallel features and attention mechanism. What is more, our work also has advantages over methods of semantic guidance [8,18,19]. Although they add semantic segmentation tasks to improve depth estimation results, they ignore the loss of information in the transmission of the network. By introducing the parallel multi-scale semantic-depth interactive fusion network, our method improves the a1 index by 0.3% over FeatDepth [28], and by 1.8% over SAFENet [29]. Through qualitative results shown in Figure 6, we compare our method with HR-Depth [22], Monodepth2 [6], and SFMLearner [4]. It can be seen that depth estimation can identify more details of distant scenes and better represent the outlines of objects, which also confirms the effectiveness of our method.
Evaluation on Make3D Dataset. To verify the generalization ability of the proposed method in training scenarios, which have never been seen before, we compare the performance of several self-supervised models on Make3D dataset. As shown in Table 2, under the same evaluation protocol as [27], our model outperforms other works and exhibits good generalization performance. The qualitative results are shown in Figure 7, which optimizes the depth estimation results in terms of details in the scenes.
Although our method is primarily designed to improve depth estimation accuracy, particularly around object boundaries, real-time inference is also important for practical applications such as autonomous driving and robot navigation, while we did not perform a full runtime benchmark, our model is built with efficiency in mind. Specifically, we adopt a lightweight HRNet backbone and streamlined parallel fusion modules to balance accuracy and computational cost. We acknowledge the importance of inference speed and will explore this aspect more thoroughly in future versions of the work.

4.4. Ablation Study

To determine the effectiveness of each contribution proposed in our method, we conducted an ablation study. (The ablation results are based on single-run experiments due to time and resource constraints, while consistent trends were observed across training epochs, future work will include repeated trials and statistical analysis (e.g., standard deviation or t-tests) to better quantify the significance of the observed improvements.) The result is shown in Table 3, starting from the baseline model and individually adding our contributions up to the full method. First, the addition of multi-stage feature attention network (MSFAN) shows an improvement. After that, introducing the semantic segmentation network without interactive fusion of features between tasks further improves the baseline. Then, applying the parallel semantic-depth interactive fusion module (PSDIFM), we achieved a greater performance than before. In the end, the metric loss is added to training. Ablation studies show that all techniques are designed to refine the depth representation. When they are combined, they can offer significant improvements to the depth estimation task.

5. Conclusions

In this paper, we propose a self-supervised monocular depth estimation pipeline incorporating semantic tasks. Through the designed parallel multi-scale semantic-depth interactive fusion network, our self-supervised depth estimation network could learn semantic-aware features to improve the performance of depth estimation. Furthermore, the total multi-task loss function is designed to adapt to the new pipeline, and the metric loss based on semantic edges is added to refine the depth of the edges. To prove the effectiveness of different modules in our method, various experiments were conducted. The experimental results show that our method is more effective than other state-of-the-art methods. While our approach shows strong performance, it also has limitations. In particular, the quality of semantic features may impact depth prediction, as we rely on pseudo-semantic labels generated by a pre-trained segmentation model. Although the depth decoder learns in an end-to-end manner and can tolerate some noise, robustness under imperfect or noisy semantic labels remains a challenge. In future work, we plan to investigate the effect of semantic label noise by introducing synthetic corruption or dropout during training, and evaluate model performance under these settings. Furthermore, we aim to improve the real-time performance of our model through optimization techniques such as pruning, quantization, or knowledge distillation. These approaches will help adapt our method for deployment in resource-constrained environments, such as embedded or mobile platforms used in robotics or autonomous driving.

Author Contributions

Conceptualization, C.F. and N.W.; methodology, C.F. and N.W.; software, C.F.; validation, C.F. and S.S.; formal analysis, S.S.; investigation, C.F. and V.C.; resources, V.C. and N.W.; data curation, N.W.; writing—original draft preparation, S.S. and N.W.; writing—review and editing, S.S.; visualization, X.X.; supervision, W.W.; project administration, X.X.; funding acquisition, C.F. and W.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under grant No. 62232004, 62272099, the Natural Science Foundation of Jiangsu Province under Grant No. BK20231543 and BK20230024, and the Key Laboratory of Computer Network and Information Integration (Southeast University), Ministry of Education.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The experimental data for this paper comes from two publicly available datasets, KITTI and Make3D. They are available at the following links respectively: 10.1109/CVPR.2012.6248074 and 10.1109/TPAMI.2008.132.

Conflicts of Interest

The author Xueyong Xu was employed by the company North Information Control Research Academy Group Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflicts of interest.

References

  1. Eigen, D.; Puhrsch, C.; Fergus, R. Depth map prediction from a single image using a multi-scale deep network. Adv. Neural Inf. Process. Syst. 2014, 27, 2366–2374. [Google Scholar]
  2. Li, B.; Shen, C.; Dai, Y.; Van Den Hengel, A.; He, M. Depth and surface normal estimation from monocular images using regression on deep features and hierarchical crfs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1119–1127. [Google Scholar]
  3. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef] [PubMed]
  4. Zhou, T.; Brown, M.; Snavely, N.; Lowe, D.G. Unsupervised learning of depth and ego-motion from video. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1851–1858. [Google Scholar]
  5. Mahjourian, R.; Wicke, M.; Angelova, A. Unsupervised learning of depth and ego-motion from monocular video using 3d geometric constraints. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–26 June 2018; pp. 5667–5675. [Google Scholar]
  6. Godard, C.; Mac Aodha, O.; Firman, M.; Brostow, G.J. Digging into self-supervised monocular depth estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3828–3838. [Google Scholar]
  7. Meng, Y.; Lu, Y.; Raj, A.; Sunarjo, S.; Guo, R.; Javidi, T.; Bansal, G.; Bharadia, D. Signet: Semantic instance aided unsupervised 3d geometry perception. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9810–9820. [Google Scholar]
  8. Klingner, M.; Termöhlen, J.A.; Mikolajczyk, J.; Fingscheidt, T. Self-supervised monocular depth estimation: Solving the dynamic object problem by semantic guidance. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XX 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 582–600. [Google Scholar]
  9. Laina, I.; Rupprecht, C.; Belagiannis, V.; Tombari, F.; Navab, N. Deeper depth prediction with fully convolutional residual networks. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; IEEE: New York, NY, USA, 2016; pp. 239–248. [Google Scholar]
  10. Mancini, M.; Costante, G.; Valigi, P.; Ciarfuglia, T.A. Fast robust monocular depth estimation for obstacle detection with fully convolutional networks. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Republic of Korea, 9–14 October 2016; IEEE: New York, NY, USA, 2016; pp. 4296–4303. [Google Scholar]
  11. Yang, Z.; Wang, P.; Xu, W.; Zhao, L.; Nevatia, R. Unsupervised learning of geometry from videos with edge-aware depth-normal consistency. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  12. Poggi, M.; Aleotti, F.; Tosi, F.; Mattoccia, S. On the uncertainty of self-supervised monocular depth estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3227–3237. [Google Scholar]
  13. Guizilini, V.; Ambrus, R.; Pillai, S.; Raventos, A.; Gaidon, A. 3d packing for self-supervised monocular depth estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2485–2494. [Google Scholar]
  14. Yin, Z.; Shi, J. Geonet: Unsupervised learning of dense depth, optical flow and camera pose. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1983–1992. [Google Scholar]
  15. Bian, J.W.; Zhan, H.; Wang, N.; Li, Z.; Zhang, L.; Shen, C.; Cheng, M.M.; Reid, I. Unsupervised scale-consistent depth learning from video. Int. J. Comput. Vis. 2021, 129, 2548–2564. [Google Scholar] [CrossRef]
  16. Sun, K.; Zhao, Y.; Jiang, B.; Cheng, T.; Xiao, B.; Liu, D.; Mu, Y.; Wang, X.; Liu, W.; Wang, J. High-resolution representations for labeling pixels and regions. arXiv 2019, arXiv:1904.04514. [Google Scholar]
  17. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  18. Chen, P.Y.; Liu, A.H.; Liu, Y.C.; Wang, Y.C.F. Towards scene understanding: Unsupervised monocular depth estimation with semantic-aware representation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2624–2632. [Google Scholar]
  19. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? the kitti vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; IEEE: New York, NY, USA, 2012; pp. 3354–3361. [Google Scholar]
  20. Zhu, S.; Brazil, G.; Liu, X. The edge of depth: Explicit constraints between segmentation and depth. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 13116–13125. [Google Scholar]
  21. Jung, H.; Park, E.; Yoo, S. Fine-grained semantics-aware representation enhancement for self-supervised monocular depth estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 12642–12652. [Google Scholar]
  22. Lyu, X.; Liu, L.; Wang, M.; Kong, X.; Liu, L.; Liu, Y.; Chen, X.; Yuan, Y. Hr-depth: High resolution self-supervised monocular depth estimation. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 2–9 February 2021; Volume 35, pp. 2294–2301. [Google Scholar]
  23. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13713–13722. [Google Scholar]
  24. Xu, D.; Ouyang, W.; Wang, X.; Sebe, N. Pad-net: Multi-tasks guided prediction-and-distillation network for simultaneous depth estimation and scene parsing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 675–684. [Google Scholar]
  25. Mnih, V.; Heess, N.; Graves, A.; Kavukcuoglu, K. Recurrent models of visual attention. Adv. Neural Inf. Process. Syst. 2014, 27. Available online: https://proceedings.neurips.cc/paper_files/paper/2014/file/3e456b31302cf8210edd4029292a40ad-Paper.pdf (accessed on 10 May 2025).
  26. Dong, X.; Shen, J. Triplet loss in siamese network for object tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 459–474. [Google Scholar]
  27. Saxena, A.; Sun, M.; Ng, A.Y. Make3d: Learning 3d scene structure from a single still image. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 31, 824–840. [Google Scholar] [CrossRef] [PubMed]
  28. Shu, C.; Yu, K.; Duan, Z.; Yang, K. Feature-metric loss for self-supervised learning of depth and egomotion. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 572–588. [Google Scholar]
  29. Choi, J.; Jung, D.; Lee, D.; Kim, C. Safenet: Self-supervised monocular depth estimation with semantic-aware feature extraction. arXiv 2020, arXiv:2010.02893. [Google Scholar]
  30. Casser, V.; Pirk, S.; Mahjourian, R.; Angelova, A. Depth prediction without the sensors: Leveraging structure for unsupervised learning from monocular videos. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 8001–8008. [Google Scholar]
  31. Ranjan, A.; Jampani, V.; Balles, L.; Kim, K.; Sun, D.; Wulff, J.; Black, M.J. Competitive collaboration: Joint unsupervised learning of depth, camera motion, optical flow and motion segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12240–12249. [Google Scholar]
  32. Pnvr, K.; Zhou, H.; Jacobs, D. Sharingan: Combining synthetic and real data for unsupervised geometry estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 13974–13983. [Google Scholar]
  33. Chanduri, S.S.; Suri, Z.K.; Vozniak, I.; Müller, C. Camlessmonodepth: Monocular depth estimation with unknown camera parameters. arXiv 2021, arXiv:2110.14347. [Google Scholar]
  34. Yan, J.; Zhao, H.; Bu, P.; Jin, Y. Channel-wise attention-based network for self-supervised monocular depth estimation. In Proceedings of the 2021 International Conference on 3D Vision (3DV), London, UK, 1–3 December 2021; IEEE: New York, NY, USA, 2021; pp. 464–473. [Google Scholar]
  35. Masoumian, A.; Rashwan, H.A.; Abdulwahab, S.; Cristiano, J.; Asif, M.S.; Puig, D. GCNDepth: Self-supervised monocular depth estimation based on graph convolutional network. Neurocomputing 2023, 517, 81–92. [Google Scholar] [CrossRef]
  36. Liu, M.; Salzmann, M.; He, X. Discrete-continuous depth estimation from a single image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 716–723. [Google Scholar]
  37. Godard, C.; Mac Aodha, O.; Brostow, G.J. Unsupervised monocular depth estimation with left-right consistency. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 270–279. [Google Scholar]
  38. Wang, C.; Buenaposada, J.M.; Zhu, R.; Lucey, S. Learning depth from monocular videos using direct methods. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2022–2030. [Google Scholar]
Figure 1. Overview of the proposed Parallel Multi-Scale Semantic-Depth Interactive Fusion Network. I t and I s represent the target and source images. MSFAN extracts multi-scale semantic and depth features, which are fused by PSDIFM into the semantic prediction S t and depth prediction D ^ t . The network includes a pose estimation module (PoseNet) to compute the transformation T t s for view synthesis. S G T and B t denote the ground-truth semantic map and semantic edge boundary, respectively. Loss functions L C E , L m e t r i c , and L m u l t i supervise the respective tasks.
Figure 1. Overview of the proposed Parallel Multi-Scale Semantic-Depth Interactive Fusion Network. I t and I s represent the target and source images. MSFAN extracts multi-scale semantic and depth features, which are fused by PSDIFM into the semantic prediction S t and depth prediction D ^ t . The network includes a pose estimation module (PoseNet) to compute the transformation T t s for view synthesis. S G T and B t denote the ground-truth semantic map and semantic edge boundary, respectively. Loss functions L C E , L m e t r i c , and L m u l t i supervise the respective tasks.
Jimaging 11 00218 g001
Figure 2. Multi-Stage Feature Attention Network (MSFAN).
Figure 2. Multi-Stage Feature Attention Network (MSFAN).
Jimaging 11 00218 g002
Figure 3. Architecture of the Parallel Semantic-Depth Interactive Fusion Module (PSDIFM). The left and right branches process multi-scale depth features f i and semantic features g i , respectively. At each level, the Spatial Attention Module (SAM) enables bidirectional interaction by enhancing one branch with attention maps generated from the other. F F i and F S i denote the fused outputs at scale i. “Up” indicates bilinear up-sampling. Dashed lines represent cross-branch fusion.
Figure 3. Architecture of the Parallel Semantic-Depth Interactive Fusion Module (PSDIFM). The left and right branches process multi-scale depth features f i and semantic features g i , respectively. At each level, the Spatial Attention Module (SAM) enables bidirectional interaction by enhancing one branch with attention maps generated from the other. F F i and F S i denote the fused outputs at scale i. “Up” indicates bilinear up-sampling. Dashed lines represent cross-branch fusion.
Jimaging 11 00218 g003
Figure 4. Boundary results B t under different threshold T .
Figure 4. Boundary results B t under different threshold T .
Jimaging 11 00218 g004
Figure 5. Images of semantic inconsistency.
Figure 5. Images of semantic inconsistency.
Jimaging 11 00218 g005
Figure 6. Qualitative results of depth estimation on KITTI dataset.
Figure 6. Qualitative results of depth estimation on KITTI dataset.
Jimaging 11 00218 g006
Figure 7. Qualitative results of depth estimation on Make3D dataset.
Figure 7. Qualitative results of depth estimation on Make3D dataset.
Jimaging 11 00218 g007
Table 1. Quantitative results of depth estimation on KITTI dataset for distance within 80 m.
Table 1. Quantitative results of depth estimation on KITTI dataset for distance within 80 m.
MethodSupAbsRel ↓SqRel ↓RMSE ↓RMSE_log ↓a1a2a3
SFMLearner [4]N0.1831.5956.7090.2700.7340.9020.959
Vid2Depth [5]N0.1631.2406.2200.2500.7620.9160.968
GeoNet [14]N0.1531.3285.7370.2320.8020.9340.972
Casser [30]N0.1411.0265.2910.2150.8160.9450.979
CC [31]F0.1401.0705.3260.2170.8260.9410.975
SharinGAN [32]N0.1160.9395.0680.2030.8500.9480.978
SCSFM [15]N0.1140.8134.7060.1910.8730.9600.982
Monodepth2 [6]N0.1120.8514.7540.1900.8810.9600.981
SGDepth [8]Seg0.1120.8334.6880.1900.8840.9610.983
SAFENet [29]Seg0.1120.7884.5820.1870.8780.9630.983
PackNet-sfm [13]N0.1110.7854.6010.1890.8780.9600.982
HR-Depth [22]N0.1090.7924.6320.1850.8840.9620.983
Chanduri [33]N0.1060.7504.4820.1820.8900.9640.983
CADepth [34]N0.1050.7694.5350.1810.8920.9640.983
FeatDepth [28]HR0.1040.7294.4810.1790.8930.9650.984
GCNDepth [35]HR0.1040.7204.4940.1810.8880.9650.984
FSRE [21]Seg0.1020.6754.3930.1780.8930.9640.984
OursSeg0.1010.7184.3760.1760.8960.9660.984
↓ (smaller values preferred) and ↑ (larger values preferred).
Table 2. Quantitative results of depth estimation on Make3D dataset for distance within 80 m.
Table 2. Quantitative results of depth estimation on Make3D dataset for distance within 80 m.
MethodSupAbsRel ↓SqRel ↓RMSE ↓RMSE_log ↓
Liu et al. [36]Y0.4626.6259.9720.161
Laina et al. [9]Y0.2041.8405.6830.084
Godard et al. [37]Y0.4437.1128.8600.142
Zhou et al. [4]N0.3924.4738.3070.194
DDVO [38]N0.3874.7208.0900.204
Monodepth2 [6]N0.3444.0657.9200.197
OursN0.3373.8427.7330.190
↓ (smaller values preferred) and ↑ (larger values preferred).
Table 3. Ablation study on KITTI dataset.
Table 3. Ablation study on KITTI dataset.
MethodAbsRel ↓SqRel ↓RMSE ↓RMSE_log ↓a1a2a3
w/o MSFAN, w/o Seg, w/o PSDIFM0.1090.7924.6320.1850.8840.9620.983
+MSFAN0.1070.7764.6200.1850.8860.9620.983
+MSFAN+Seg0.1070.7664.5110.1830.8920.9650.983
+MAFAN+Seg+ PSDIFM0.1030.7474.4050.1770.8950.9660.984
+MAFAN+Seg+ PSDIFM+Metric Loss0.1010.7184.3760.1760.8960.9660.984
↓ (smaller values preferred) and ↑ (larger values preferred).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fu, C.; Sun, S.; Wei, N.; Chau, V.; Xu, X.; Wu, W. Parallel Multi-Scale Semantic-Depth Interactive Fusion Network for Depth Estimation. J. Imaging 2025, 11, 218. https://doi.org/10.3390/jimaging11070218

AMA Style

Fu C, Sun S, Wei N, Chau V, Xu X, Wu W. Parallel Multi-Scale Semantic-Depth Interactive Fusion Network for Depth Estimation. Journal of Imaging. 2025; 11(7):218. https://doi.org/10.3390/jimaging11070218

Chicago/Turabian Style

Fu, Chenchen, Sujunjie Sun, Ning Wei, Vincent Chau, Xueyong Xu, and Weiwei Wu. 2025. "Parallel Multi-Scale Semantic-Depth Interactive Fusion Network for Depth Estimation" Journal of Imaging 11, no. 7: 218. https://doi.org/10.3390/jimaging11070218

APA Style

Fu, C., Sun, S., Wei, N., Chau, V., Xu, X., & Wu, W. (2025). Parallel Multi-Scale Semantic-Depth Interactive Fusion Network for Depth Estimation. Journal of Imaging, 11(7), 218. https://doi.org/10.3390/jimaging11070218

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop