Next Article in Journal
Regional Time-Series Coding Network and Multi-View Image Generation Network for Short-Time Gait Recognition
Next Article in Special Issue
Backdoor Attack against Face Sketch Synthesis
Previous Article in Journal
Clustering Analysis of Seismicity in the Anatolian Region with Implications for Seismic Hazard
Previous Article in Special Issue
ShrewdAttack: Low Cost High Accuracy Model Extraction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multiscale Attention Fusion for Depth Map Super-Resolution Generative Adversarial Networks

1
School of Computer Science and Technology, Harbin Institute of Technology, Harbin 150001, China
2
Pengcheng Laboratory, Shenzhen 518052, China
3
School of Electronic Engineering and Computer Science, Peking University, Beijing 100871, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(6), 836; https://doi.org/10.3390/e25060836
Submission received: 23 April 2023 / Revised: 17 May 2023 / Accepted: 17 May 2023 / Published: 23 May 2023
(This article belongs to the Special Issue Trustworthy AI: Information Theoretic Perspectives)

Abstract

:
Color images have long been used as an important supplementary information to guide the super-resolution of depth maps. However, how to quantitatively measure the guiding effect of color images on depth maps has always been a neglected issue. To solve this problem, inspired by the recent excellent results achieved in color image super-resolution by generative adversarial networks, we propose a depth map super-resolution framework with generative adversarial networks using multiscale attention fusion. Fusion of the color features and depth features at the same scale under the hierarchical fusion attention module effectively measure the guiding effect of the color image on the depth map. The fusion of joint color–depth features at different scales balances the impact of different scale features on the super-resolution of the depth map. The loss function of a generator composed of content loss, adversarial loss, and edge loss helps restore clearer edges of the depth map. Experimental results on different types of benchmark depth map datasets show that the proposed multiscale attention fusion based depth map super-resolution framework has significant subjective and objective improvements over the latest algorithms, verifying the validity and generalization ability of the model.

1. Introduction

With the increasing emphasis on security, trustworthy artificial intelligence is on the rise. In trustworthy AI, various 3D applications play a crucial role in scene construction, understanding the relationship between entities and the scene, and reasoning about invisible factors outside the scene. In the research of stereo image technology, the quality of depth maps is very significant because the depth value reflects the spatial position of the object in the scene. However, the resolution of depth maps has been very low due to the limited capture capability of depth sensors. Therefore, depth map super-resolution (SR) has become an urgent problem to be solved.
Due to the limited information contained in a single depth map, the corresponding high-resolution (HR) color image is generally used to guide the super-resolution of depth maps. Conventional methods use filters or Markov Random Fields (MRF) to implement depth map super-resolution with the guidance of the color image. Leveraging the HR color image and the given low-resolution (LR) depth map, Kopf et al. [1] proposed a joint bilateral filter (JBU) which combines a range filter and a spatial filter to produce very good full resolution results. Diebel and Thrun [2] first formulated the depth map SR as a multi-labeling optimization problem based on the MRF model by connecting the color image and the depth map as the balance factor of the smooth term.
In recent years, due to the rapid development of convolutional neural networks, color-guided depth map super-resolution methods based on convolutional neural networks have achieved more remarkable results. Hui et al. [3] proposed a multiscale guided convolutional network (MSG-Net) for depth map super-resolution which complements low-resolution depth features with HR intensity features using a multiscale fusion strategy. Ye et al. [4] constructed a convolutional neural network architecture to learn a binary map of depth edge location from a low-resolution depth map and the corresponding color image, and then proposed a fast edge-guided depth filling strategy to interpolate the missing depth.
However, most color-guided depth map super-resolution methods use color images directly. How to quantitatively measure the guiding effect of color images on depth map super-resolution lacks the attention of researchers. In this paper, we propose a depth map super-resolution framework that uses hierarchical attention fusion modules to measure the guidance of color features on depth features. Inspired from the recent emergence of excellent color image super-resolution generative adversarial networks such as SRGAN [5] and ESRGAN [6], our framework uses relativistic standard generative adversarial networks as the backbone. In particular, a loss model generator that includes content loss, adversarial loss, and edge loss helps the proposed generative adversarial networks produce clearer edges of the depth map.
Our main contributions are as follows: (1) We propose a depth map super-resolution framework with multiscale attention fusion based generative adversarial networks to quantitatively measure the effectiveness of color images as a guide to depth map super-resolution. (2) The hierarchical color–depth attention fusion module measures the guidance of the color image on the depth map super-resolution and generates fused features of various scales. (3) The multiscale fused feature balance module evaluates the correlation between scales and fused features, and integrates fused color–depth features of various scales proportionally. (4) A loss function consisting of content loss, adversarial loss, and edge loss helps our method produce clearer edges of the depth map.
We organize the remainder of this paper as follows. After a brief review of related literature in Section 2, we present the framework and introduce the details of our method in Section 3. In Section 4, we conduct an ablation study and comparison experiments on the benchmark depth map datasets, and discuss the performance of our method compared to other methods. Finally, we conclude the whole paper in Section 6.

2. Related Works

In this section, we introduce color-guided depth map super-resolution and color image super-resolution generative adversarial networks methods.

2.1. Conventional Color-Guided Depth Map Super-Resolution

Conventional color-guided depth map super-resolution methods can be divided into three categories: filter based methods, MRF based methods, and sparse representation based methods.
Filter-based methods [1,7,8,9,10,11,12,13] aim to construct upsampling filters to enhance the depth map resolution with the guidance of the registered color image. Leveraging the HR color image and the given low-resolution depth map, Kopf et al. [1] proposed a joint bilateral filter (JBU) which combines a range filter and a spatial filter to produce very good full-resolution results. In [8], Kim et al. proposed a modified JBU called JABDU that computes each depth value as the average of neighboring pixels weighted by color and depth intensity filters, which are formulated as an adaptive smoothing parameter and a control parameter, respectively. Inspirited from the geodesic distance, Liu et al. [9] advanced the resolution of a depth map using geodesic paths to the pixels whose depths are known from the low-resolution ones. A weighted mode filter (WMF) is proposed in [10] by seeking a global mode on the histogram which uses the weight considering color similarity between the reference and neighboring pixels on the color image to upsample the depth map. Furthermore, Fu et al. [11] incorporated a noise-aware filter (NAF) into a WMF. In order to reduce the artifacts such as texture copy and edge discontinuities, Lo et al. [12] constructed a joint trilateral filtering (JTF) algorithm for depth image SR considering spatial distance, color difference, and local depth gradient simultaneously to better preserve the contour information. Filter-based depth map SR methods can remove the external and internal noise of the depth map, and simultaneously preserve contour features of it. However, with the color image guiding them, these methods can produce texture copy artifacts in smooth regions of the depth map which correspond to richly textured regions of the color image.
Optimization-based single depth map SR methods can be generally divided into two classes: Markov Random Fields (MRF) [2,14,15,16,17,18,19] based algorithms and Sparse Representation based algorithms. Diebel and Thrun [2] first formulate the depth map SR as a multi-labeling optimization problem based on the MRF model. The method in [15] extends the MRF model by presenting a novel data term allowing for adaptive pixel-wise determination of an appropriate depth reference value. In [14], Zuo et al. proposed a method to quantitatively measure the inconsistency between the depth edge map and the color edge map and explicitly embedded the measurement into the smoothness term of the MRF model. Utilizing the edges of the low-resolution depth image through a Markov Random Fields optimization in a patch synthesis based manner, Xie et al. [17] constructed a high-resolution edge map to guide the upscaling of the depth map. By solving an MRF labeling optimization problem, Lo et al. presented a learning-based depth map super-resolution framework in [12] which exhibits the capability of preserving the edges of range data while suppressing the artifacts of texture copy due to color discontinuities. Compared with filter-based methods, optimization-based methods are more robust to noise. For the condition that the edges in a depth map correspond to the smooth region of the color image, blurred edge artifacts can be generated in the SR process due to the inconsistency between the edges of the depth map and color image at the same location.
Many sparse representation-based depth map SR methods [20,21,22,23,24,25] have been proposed in the last few years. They usually cut HR color images and LR depth maps into patches and bind them in pairs to train a dictionary. The depth map SR solutions can be represented as a linear combination of elements in the learned dictionary. Ferstl et al. [21] presented a variational sparse representation approach by using a dictionary of edge priors which learned from an external database of high- and low-resolution examples. In [22], Xie et al. reconstructed the corresponding HR depth map through a robust coupled dictionary learning method with locality coordinate constraints. Simultaneously, an adaptively regularized shock filter is introduced to reduce sharpening of the contours and the jagged noise. Zhang et al. proposed a dual sparsity model based single depth map SR method by combining the analysis model and synthesis model in [24]. As this category of methods utilizes amounts of depth map patches in the training stage, the performance of them heavily relies on the selection of external datasets. In addition, few representation-based depth map SR methods suffer from blurring edge artifacts on the depth edges or the overlapping regions of adjacent patches similar to the optimization-based depth map SR methods.
Single depth map SR methods can achieve a promising performance in preserving depth contour while alleviating the noise of the depth map. However, they can produce texture copy artifacts and blurring edge artifacts derived from the depth discontinuities that are not consistent with color discontinuities at the corresponding position.

2.2. Neural-Networks-Based Depth Map Super-Resolution

Depth map super-resolution methods based on neural networks have achieved promising success [3,4,26,27]. The authors of [3] proposed a multiscale guided convolutional network (MSG-Net) for depth map super-resolution which complements low-resolution depth features with HR intensity features using a multiscale fusion strategy. Ye et al. [4] constructed a convolutional neural network architecture to learn a binary map of depth edge location from a low-resolution depth map and corresponding color image. They then proposed a fast edge-guided depth filling strategy to interpolate the missing depth constrained by the acquired edges to prevent predicting across the depth boundaries. Wang et al. [26] proposed a novel depth upsampling framework based on deep edge-aware learning which firstly learns edge information of depth boundaries from the known LR depth map and its corresponding high-resolution (HR) color image as reconstruction cues. Then, two depth restoration modules, i.e., a fast depth filling strategy and a cascaded restoration network, are proposed to recover an HR depth map by leveraging the predicted edge map and the HR color image. In [28], Zuo et al. proposed a novel DCNN to progressively reconstruct the high-resolution depth map guided by the intensity image. Specifically, the multiscale intensity features are extracted to provide guidance for the refinement of depth features as their resolutions are gradually enhanced. In [27] by Zuo et al., a novel depth-guided affine transformation is used to filter out the unrelated intensity features, which is further used to refine the depth features. Since the quality of initial depth features is low, the depth-guided intensity features filtering and the intensity-guided depth features refinement are iteratively performed, which progressively promotes the effects of such tasks.
Images at different scales contain different feature information [3]. However, the guidance of color image features at different scales on depth map super-resolution is not equal. It is not appropriate to cascade or link them directly. As far as we know, quantitative evaluation of the correlation between the scales of features and depth map super-resolution is a topic that has not been discussed. In this article, we use a multiscale fused feature balance module to measure the correlations between different scale features and depth map super-resolution, and further fuse the color–depth features at different scales proportionally.

2.3. Generative Adversarial Network Based Color Image Super-Resolution

Super-resolution methods for color images based on generative adversarial networks [5,6,29,30,31] generate realistic high-resolution color images by means of successive iterations of mutual adversaries between generators and discriminators.
Denton et al. [29] introduced a generative parametric model capable of producing high-quality samples of natural images. It uses a cascade of convolutional networks within a Laplacian pyramid framework to generate images in a coarse-to-fine fashion. At each level of the pyramid, a separate generative convnet model is trained using the generative adversarial networks (GAN) approach (Goodfellow et al.). Samples drawn from their model are of significantly higher quality than alternate approaches. The key idea of [30] is to grow both the generator and discriminator progressively: starting from a low resolution and adding new layers that model increasingly fine details as training progresses. This both speeds the training up and greatly stabilizes it to produce images of unprecedented quality. Ledig et al. [5] presented SRGAN, a generative adversarial network (GAN) for image super-resolution (SR). To our knowledge, it is the first framework capable of inferring photo-realistic natural images for 4× upscaling factors. To achieve this, they propose a perceptual loss function which consists of an adversarial loss and a content loss. The adversarial loss pushes their solution to the natural image manifold using a discriminator network that is trained to differentiate between the super-resolved images and original photo-realistic images. The super-resolution generative adversarial network (SRGAN) is a seminal work that is capable of generating realistic textures during single image super-resolution. However, the hallucinated details are often accompanied with unpleasant artifacts. To further enhance the visual quality, Wang et al. [6] thoroughly studied three key components of SRGAN, network architecture, adversarial loss, and perceptual loss, and improve each of them to derive an enhanced SRGAN (ESRGAN).
Some excellent methods for color image super-resolution generative adversarial networks have emerged. However, they produce more artifacts and textures. Obviously, these networks are not suitable for depth map super-resolution. Therefore, considering the sharp edges and smooth interior of the depth map, we propose a multiscale attention fusion based super-resolution generative adversarial network for depth maps. In particular, building a generator loss function that includes content loss, adversarial loss, and edge loss facilitates the generation of sharper edges.

3. Multiscale Attention Fusion for Depth Map Super-Resolution Generative Adversarial Networks

In this section, we propose a multiscale attention fusion for depth map super-resolution generative adversarial networks.

3.1. Framework

The framework of our proposed method is demonstrated in Figure 1. In Figure 1, our goal is to generate a precise high-resolution depth estimation D H R of the ground truth D G . The generator consists of four parts: a multiscale color and depth feature extraction module, a hierarchical feature attention fusion module, a multiscale fused feature balance module, and a super-resolution module. The multiscale color and depth feature extraction module extracts different scale features using a low-resolution depth map and the corresponding color image as inputs. It consists of two convolutional layers and n residual dense blocks (RDBs), where n is the scale of feature extraction. The settings of RDBs are consistent with those in [32]. The depth feature and color feature passing through the ith RDB are represented as F D i and F I i , respectively. Previous methods have used these to directly concatenate depth features and color features together. However, the guidance of color features on depth features should not be just a simple link. How to quantitatively measure the guidance of color features on depth features has become a key issue. In this article, we propose using an attention module to measure the guiding effect of color features on depth features. F D i and F I i form a color–depth fused feature F f i at the ith scale through the attention module. In this way, we obtain color–depth fused features F f 1 , F f 2 , , F f n at n scales. Images at different scales contain different geometric structures. The contribution of fused features at different scales to depth map super-resolution is not equal. We input F f 1 - F f n into the multiscale fused feature balance module to evaluate the correlations between the scales and fused features, and obtain a final fused feature F f . We choose UPNet as [32] as the super-resolution module of the generator. The high-resolution depth map D H R is generated by F f through UPNet.

3.2. Hierarchical Color and Depth Attention Fusion Module

The details of the proposed hierarchical color and depth attention fusion module are shown in Figure 2. Before inputting them into the module, we first concatenate the color feature F I i and the depth feature F D i at the ith scale to form the merged feature F C i . Then, F C i is fed into global average pooling and global sum pooling, respectively. Global average pooling and global sum pooling are followed by two convolutional layers and one ReLU, respectively. By processing the convolutional features through the sigmoid function, two coefficient matrices are obtained. By adding and splitting the coefficient matrices in place we can obtain the weight coefficient vector C i of F I i and F D i as in Equation (1).
C i = f a t t F D i , F I i ,
where f a t t denotes the color and depth attention fusion module. Multiplying F I i and F D i element-wise by the corresponding coefficient vector C i , we obtain the fused color–depth fused feature F f i at the ith scale as in Equation (2).
F f i = F D i , F I i C i

3.3. Multiscale Fused Feature Balance Module

After obtaining the color–depth fused features F f i using the n attention modules, we concatenate these features and denote them as F f C .
F f C = F f 1 , F f 2 , , F f i , , F f n
Then F f C is fed to the multiscale fused feature balance module.
W f = f b a l F f C
where W f is a vector of balance factors and f b a l is the multiscale fused feature balance module. The multiscale fused feature balance module is used to evaluate the correlations between the scales and the fused features as shown in Figure 3. It consists of two branches. These two branches start with a global average pooling and a global sum pooling, separately, followed by two convolutional layers, a ReLU layer, and a sigmoid function. F f C generates two weight coefficient matrices through these two branches. The two weight coefficient matrices are summed and separated to obtain W f . The balanced multiscale color–depth feature F f is generated by multiplying the concatenated sequence F f C with the corresponding balance factor vector as in Equation (5).
F f = F f C W f

3.4. Relativistic Standard Generative Adversarial Networks

In the standard GAN, the discriminator outputs the probability that the input image is real to determine whether the input is real or fake. The type of GAN can be defined with the discriminator. In general, the discriminator loss of a standard GAN with the assumption of cross-entropy loss can be expressed as follows:
L D = E x r P f 1 D x r + E x f Q f 2 D x f
where x r and x f indicate the real depth map and the fake one, respectively. The adversarial loss of the generator is expressed as
L G = E x r P g 1 C x r + E x f Q g 2 C x f ,
where
f 1 D x = g 2 D x = log D x f 2 D x = g 1 D x = log 1 D x
D x is the activation function of the non-tranformed layer C x as in Equation (9).
D x = sigmoid C x
In the discriminator, D G and D H R are input as x r and x f . Because the gradient of g 1 is 0, only half of the generator is involved during the training process.
In this paper, we adopt the relativistic standard GAN (RGAN) [33] structure to achieve the full participation of the generator. The discriminator of RGAN estimates the probability that the given real data is more realistic than a randomly sampled fake data. It is denoted as Equation (10).
D x r , x f = sigmoid C x r C x f
Correspondingly, the loss of the discriminator is expressed as follows:
L D = E x r , x f P , Q log sigmoid C x r C x f E x r , x f P , Q log 1 sigmoid C x f C x r
The adversarial loss of the generator is expressed as Equation (14).
L G = E x r , x f P , Q log 1 sigmoid C x r C x f E x r , x f P , Q log sigmoid C x f C x r
The discriminator extracts features using the RDBs, and then performs a discriminant classification using the sigmoid function to determine whether the input depth map is fake or real. Compared to the standard GAN, the relativistic GAN can generate high-resolution depth maps from relatively small samples. Furthermore, the training time to achieve optimal performance is significantly reduced.
Due to the fact that depth maps are a kind of piece-wise smooth images, they are characterized by sharp edges and smooth interiors. Conventional GAN-based color image super-resolution methods that only use mean squared error (MSE) as content loss are not suitable for depth map super-resolution. In order to improve the edge sharpness of the generated high-resolution depth maps, we propose a loss function consisting of content loss, adversarial loss, and edge loss, which is expressed by Equation (13).
L ˜ G = L C + μ L G + γ L e
where
L C = 1 N i = 1 N D H R D G 2 L e = 1 N i = 1 N e H R e G 2
μ and γ are the scale factors which balance the adversarial loss and the edge loss.

4. Experimental Results

4.1. Parameter Setting

We train our network with 80 color and depth pairs. In the training dataset, 52 color–depth pairs are from the Middlebury dataset and others are from the MPI Sintel depth dataset. The color images are downsampled to the images of corresponding scale factor by interval interpolation. The patch size is 128 × 128 and the batch size is 16. To enrich data diversity, we flip the patches horizontally and vertically, and rotate them by 90 . The kernel size of all convolution layers is 3 and the channels of the feature map number 64. We take ReLU function as the activation function after all convolution layers. Adam is set as the optimizer with β 1 = 0.9 and β 2 = 0.999 . Our proposed method is implemented on two Nvidia RTX 2080ti GPUs. We train our network for 1000 epochs, and the initial learning rate is 10 4 and halved every 200 epochs.

4.2. Datasets

  • Training Datasets
In the training phase, we use two datasets, the Middlebury datasets and the MPI Sintel depth dataset. The Middlebury datasets [34] are a stereo dataset widely used in applications related to stereo matching, 3D reconstruction, and stereo quality evaluation. It consists of five versions constructed in different years, namely the 2001 dataset, 2003 dataset, 2005 dataset, 2006 dataset, 2014 dataset, and 2021 moving dataset. We randomly selected 52 datasets from the Middlebury datasets of different view, size, illuminations, and exposures.
The MPI Sintel depth dataset [35] is a synthetic stereo dataset which provides naturalistic video sequences. The depth values in the MPI Sintel depth dataset are returned from Blender with an additional Z-buffer pass, similar to the optical flow.
  • Testing Datasets
Among the Middlebury Stereo Datasets [34], we use six color–depth pairs as the testing samples. They are Art, Books, Moebius, Reindeer, Laundry, and Dolls. To better demonstrate the effectiveness of our method, we also conduct experiments on the Multi-view depth (MVD) test sequences [36] and ToFMark dataset [37].
The Multi-view depth (MVD) test sequences consist of multi-view video sequences and corresponding pixel-by-pixel depth information to support flexible synthesis of virtual views during rendering. It is widely used in studies on 3D applications such as free-view video, binocular stereoscopic video, and naked eye 3D stereo video, making it the most promising form of 3D video data representation today.
ToFMark dataset containes three real-scene depth maps captured by ToF sensors. The low-resolution depth maps in it were acquired using the PMD Nano ToF camera with a resolution of 120 × 160, and the high-resolution color images were acquired using the CMOS camera with a resolution of 810 × 610.

4.3. Evaluation Metrics

For reconstructed and enhanced images, many studies have proposed many objective evaluation criteria [38,39]. In this paper, we take three metrics to evaluate the performance of our proposed method in depth map super-resolution. They are RMSE, MAD, and PSNR.
RMSE stands for root mean squared error, as in Equation (15).
RMSE = 1 N i = 1 N D H R D G 2
MAD represents Mean Absolute Difference, described by Equation (16).
MAD = 1 N i = 1 N D H R D G
Peak Signal to Noise Ratio (PSNR) is also a commonly used objective criterion for evaluating image quality,
PSNR = 10 × log 10 2 n 1 2 MSE
where MSE is mean squared error in Equation (18), which is the square of RMSE.
MSE = 1 N i = 1 N D H R D G 2

4.4. Comparison of Different Numbers of RDBs

In this subsection, we explore the correlation between the quantity of scales in the multiscale fusion attention module on the performance of the depth map super-resolution. We tested the experimental results on the Middlebury datasets for four quantities of RDBs: 10, 16, 20, and 22. The selection type of GAN is RGAN and the loss of the generator is set to content loss + edge loss. The experimental results are shown in Table 1. We can see that as the quantity of RDBs increases, the RMSE of the generated depth map decreases. However, after the number of scales exceeds 16, the effect of depth map super-resolution is not significantly improved. Considering the increased storage and computing consumption, we believe that 16 is the most reasonable number of RDBs.

4.5. Comparison of GAN Types

In this subsection, we compare the depth map SR results with different kinds of GANs. Table 2 demonstrates the experimental results of our proposed method with GAN and RGAN. We choose 16 as the number of scales in the multiscale fusion attention module and MSE + edge loss as the loss of generator. It can be seen that the RMSE of our method with RGAN is better than that with GAN. This shows that our method uses RGAN to generate high-resolution depth maps that are closer to the real depth maps.

4.6. Comparison of Generator Losses

In this subsection, we compare the experimental results with generator losses of MSE loss and MSE loss + edge loss, besides adversarial loss, to verify the necessity of edge loss. The general color image super-resolution generative adversarial networks are reconstructed based on MSE loss, which can obtain closer objective experimental results to the ground truth. However, the visual performance of the image generated in this way is not perceptually the closest to ground truth. Therefore, we propose a generator loss function that includes an edge loss for the characteristics of the clear edges and the smooth interior of the depth maps. As shown in Table 3, the RMSE of the depth maps generated by RGAN with edge loss is very close to the RMSE of those generated by a network containing only MSE. However, Figure 4 shows the comparison of two sets of super-resolution results on Art. We can see that the network containing edge loss generates high-resolution depth maps with clearer edges compared to the GAN containing only MSE, thereby verifying the effectiveness of edge loss in generating perceptually high-quality depth maps.

4.7. Experimental Results on Middlebury Datasets

Our baseline state-of-the-art methods are joint bilateral upsampling (JBU) [1], noise-aware filter (NAF) [40], anisotropic diffusion [41], Markov random field (MRF) [2], guided image filtering (GIF) [42], SRF from [43], edge weighted NLM regularization (Edge) [44], joint geodesic filtering (JGF) [9], total generalized variation (TGV) from [37], four deep learning method SRCNNs from [45], deep joint image filter (DJIF) from [46], deep edge-aware network (DSR) from [26] and cross-guided network for depth map enhancement (CGN) from [27], two GAN-based color image super-resolution methods for a super-resolution generative adversarial network (SRGAN) from [5], enhanced SRGAN (ESRGAN) from [6], and dictionary learning method JESR from [20] that are used in comparison to evaluate the performance of our method. We set the number of RDBs to 16 and the type of GAN to RGAN in our method. The depth map upscaling factors are set to 2, 4, 8, and 16.
In Table 4 and Table 5, we can see that both DSR and CGN obtain top-ranked experimental results. Compared with the other two color image super-resolution GAN methods, our proposed method gains the lowest RMSE and MAD. This is because SRGAN and ESRGAN are designed for color images with a structure that produces more texture. However, they are not suitable for the internal smoothing properties of depth maps.
Figure 5 shows the visual comparison of the state-of-the-art baselines with our method. It can be seen that our method produces clearer and sharper edges, and avoids artifacts of blurred edges and texture transfer.

4.8. Experimental Results on Real Datasets

Since depth maps are acquired by depth sensors in real scenes, we not only compare experimental results on the Middlebury datasets, but also conduct experiments and comparisons on real scene depth map datasets. In this article, we selected the ToFMark dataset captured by the ToF sensor and the multi-view depth (MVD) test sequences [36] as the test sets. Our comparison methods are bicubic, joint geodesic filtering (JGF) [9], total generalized variation (TGV) from [37], SRGAN from [5], enhanced SRGAN (ESRGAN) from [6], deep joint image filter (DJIF) from [46], deep edge-aware network (DSR) from [26], and cross-guided network for depth map enhancement (CGN) from [27]. The depth map upscaling factors are set to 2, 4, 8, and 16.
Table 6 and Table 7 demonstrate the quantitative depth upsampling results on ToFMark dataset and MVD dataset, respectively. Our proposed method shows the best objective performance over other the state-of-the-art methods.

5. Discussion

In this section, we briefly discuss our proposed method and the directions we can focus on in the future. In the edge region of the depth map, the introduced color image corresponds to a smooth region, resulting in the generated high-resolution depth map with occasional edge blurring. In the future, we will focus on introducing color image edges aligned with the edges of the depth map into the framework to achieve more accurate depth map super-resolution as well as to generate sharper edges.

6. Conclusions

In this paper, we propose a multiscale attention fusion based depth super-resolution generative adversarial networks for 3D reconstruction in trustworthy AI. Specifically, a hierarchical color–depth attention fusion module measures the guidance of the color image on the depth map super-resolution and generates fused features of various scales. The multiscale fused feature balance module evaluates the correlation between scales and fused features, and integrates fused color–depth features at different scales in a proportional manner. By constructing a loss function model consisting of content loss, adversarial loss, and edge loss, our proposed generative adversarial networks produce high-resolution depth maps with sharper edges. The robustness and generalization of the model is demonstrated by extensive experiments that show satisfactory subjective and objective results of our proposed method on several types of depth map datasets.

Author Contributions

Conceptualization, methodology, software, validation, writing—original draft preparation, D.X.; writing—review and editing, supervision, funding acquisition, X.F.; project administration, W.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National High Technology Research and Development Program of China grant number [2021YFF0900500], and the National Natural Science Foundation of China (NSFC) under grants [61972115 and 61872116].

Data Availability Statement

Data will be made available on request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kopf, J.; Cohen, M.; Lischinski, D.; Uyttendaele, M. Joint Bilateral Upsampling. ACM Trans. Graph. 2007, 26, 96-es. [Google Scholar] [CrossRef]
  2. Diebel, J.; Thrun, S. An Application of Markov Random Fields to Range Sensing. In Proceedings of the 18th International Conference on Neural Information Processing Systems, Cambridge, MA, USA, 5–8 December 2005; pp. 291–298. [Google Scholar]
  3. Hui, T.W.; Loy, C.C.; Tang, X. Depth Map Super-Resolution by Deep Multi-Scale Guidance. In Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Volume 9907, pp. 353–369. [Google Scholar] [CrossRef]
  4. Ye, X.; Duan, X.; Li, H. Depth Super-Resolution with Deep Edge-Inference Network and Edge-Guided Depth Filling. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 1398–1402. [Google Scholar] [CrossRef]
  5. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 105–114. [Google Scholar] [CrossRef]
  6. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Qiao, Y.; Loy, C.C. ESRGAN: Enhanced Super-Resolution Generative Adversarial Networks. In Proceedings of the Computer Vision—ECCV 2018 Workshops, Munich, Germany, 8–14 September 2018; pp. 63–79. [Google Scholar]
  7. Park, J.; Kim, H.; Tai, Y.-W.; Brown, M.S.; Kweon, I. High quality depth map upsampling for 3D-TOF cameras. In Proceedings of the International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 1623–1630. [Google Scholar] [CrossRef]
  8. Jeong, J.; Kim, J.; Jeon, G. Joint-adaptive bilateral depth map upsampling. Signal Process. Image Commun. 2014, 29, 506–513. [Google Scholar] [CrossRef]
  9. Liu, M.; Tuzel, O.; Taguchi, Y. Joint Geodesic Upsampling of Depth Images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 169–176. [Google Scholar] [CrossRef]
  10. Min, D.; Lu, J.; Do, M.N. Depth Video Enhancement Based on Weighted Mode Filtering. IEEE Trans. Image Process. 2012, 21, 1176–1190. [Google Scholar] [CrossRef] [PubMed]
  11. Fu, M.; Zhou, W. Depth map super-resolution via extended weighted mode filtering. In Proceedings of the Visual Communications and Image Processing, Chengdu, China, 27–30 November 2016; pp. 1–4. [Google Scholar]
  12. Lo, K.; Wang, Y.F.; Hua, K. Edge-Preserving Depth Map Upsampling by Joint Trilateral Filter. IEEE Trans. Cybern. 2018, 48, 371–384. [Google Scholar] [CrossRef] [PubMed]
  13. Song, Y.; Gong, L. Analysis and improvement of joint bilateral upsampling for depth image super-resolution. In Proceedings of the 2016 8th International Conference on Wireless Communications & Signal Processing (WCSP), Yangzhou, China, 13–15 October 2016; pp. 1–5. [Google Scholar]
  14. Zuo, Y.; Wu, Q.; Zhang, J.; An, P. Explicit Edge Inconsistency Evaluation Model for Color-Guided Depth Map Enhancement. IEEE Trans. Circuits Syst. Video Technol. 2018, 28, 439–453. [Google Scholar] [CrossRef]
  15. Lu, J.; Min, D.; Pahwa, R.S.; Do, M.N. A revisit to MRF-based depth map super-resolution and enhancement. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; pp. 985–988. [Google Scholar] [CrossRef]
  16. Mac Aodha, O.; Campbell, N.D.F.; Nair, A.; Brostow, G.J. Patch Based Synthesis for Single Depth Image Super-Resolution. In Computer Vision—ECCV 2012; Springer: Berlin/Heidelberg, Germany, 2012; pp. 71–84. [Google Scholar]
  17. Xie, J.; Feris, R.S.; Sun, M. Edge-Guided Single Depth Image Super Resolution. IEEE Trans. Image Process. 2016, 25, 428–438. [Google Scholar] [CrossRef] [PubMed]
  18. Lo, K.; Hua, K.; Wang, Y.F. Depth map super-resolution via Markov Random Fields without texture-copying artifacts. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 1414–1418. [Google Scholar]
  19. Li, J.; Lu, Z.; Zeng, G.; Gan, R.; Zha, H. Similarity-Aware Patchwork Assembly for Depth Image Super-resolution. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3374–3381. [Google Scholar]
  20. Li, Y.; Xue, T.; Sun, L.; Liu, J. Joint Example-Based Depth Map Super-Resolution. In Proceedings of the 2012 IEEE International Conference on Multimedia and Expo, Melbourne, VIC, Australia, 9–13 July 2012; pp. 152–157. [Google Scholar] [CrossRef]
  21. Ferstl, D.; Rüther, M.; Bischof, H. Variational Depth Superresolution Using Example-Based Edge Representations. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 513–521. [Google Scholar]
  22. Xie, J.; Feris, R.S.; Yu, S.; Sun, M. Joint Super Resolution and Denoising From a Single Depth Image. IEEE Trans. Multimed. 2015, 17, 1525–1537. [Google Scholar] [CrossRef]
  23. Kiechle, M.; Hawe, S.; Kleinsteuber, M. A Joint Intensity and Depth Co-sparse Analysis Model for Depth Map Super-resolution. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, NSW, Australia, 1–8 December 2013; pp. 1545–1552. [Google Scholar]
  24. Zhang, Y.; Zhang, Y.; Dai, Q. Single depth image super resolution via a dual sparsity model. In Proceedings of the 2015 IEEE International Conference on Multimedia Expo Workshops (ICMEW), Turin, Italy, 29 June–3 July 2015; pp. 1–6. [Google Scholar]
  25. Zheng, H.; Bouzerdoum, A.; Phung, S.L. Depth image super-resolution using internal and external information. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), South Brisbane, QLD, Australia, 19–24 April 2015; pp. 1206–1210. [Google Scholar]
  26. Wang, Z.; Ye, X.; Sun, B.; Yang, J.; Xu, R.; Li, H. Depth upsampling based on deep edge-aware learning. Pattern Recognit. 2020, 103, 107274. [Google Scholar] [CrossRef]
  27. Zuo, Y.; Fang, Y.; An, P.; Shang, X.; Yang, J. Frequency-Dependent Depth Map Enhancement via Iterative Depth-Guided Affine Transformation and Intensity-Guided Refinement. IEEE Trans. Multimed. 2021, 23, 772–783. [Google Scholar] [CrossRef]
  28. Zuo, Y.; Fang, Y.; Yang, Y.; Shang, X.; Wang, B. Residual dense network for intensity-guided depth map enhancement. Inf. Sci. 2019, 495, 52–64. [Google Scholar] [CrossRef]
  29. Denton, E.L.; Chintala, S.; Szlam, A.; Fergus, R. Deep Generative Image Models using a Laplacian Pyramid of Adversarial Networks. In Advances in Neural Information Processing Systems; Cortes, C., Lawrence, N., Lee, D., Sugiyama, M., Garnett, R., Eds.; 2015; Volume 8, Available online: https://proceedings.neurips.cc/paper/2015/hash/aa169b49b583a2b5af89203c2b78c67c-Abstract.html (accessed on 16 May 2023).
  30. Karras, T.; Aila, T.; Laine, S.; Lehtinen, J. Progressive Growing of GANs for Improved Quality, Stability, and Variation. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  31. Huang, X.; Li, Y.; Poursaeed, O.; Hopcroft, J.; Belongie, S. Stacked Generative Adversarial Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1866–1875. [Google Scholar] [CrossRef]
  32. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual Dense Network for Image Super-Resolution. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2472–2481. [Google Scholar] [CrossRef]
  33. Jolicoeur-Martineau, A. The relativistic discriminator: A key element missing from standard GAN. arXiv 2018, arXiv:1807.00734. [Google Scholar]
  34. Hirschmuller, H.; Scharstein, D. Evaluation of Cost Functions for Stereo Matching. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar] [CrossRef]
  35. Butler, D.J.; Wulff, J.; Stanley, G.B.; Black, M.J. A naturalistic open source movie for optical flow evaluation. In Proceedings of the European Conference on Computer Vision (ECCV), Florence, Italy, 7–13 October 2012; Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C., Eds.; Springer: Berlin/Heidelberg, Germany, 2012. Part IV, LNCS 7577. pp. 611–625. [Google Scholar]
  36. MVD. Multi-View Video Plus Depth MVD Format for Advanced 3D Video Systems. In Document Joint Video Team (JVT) of ISO/IEC MPEG ITU-T VCEG, and JVT-W100; MVD: Online, 2007. [Google Scholar]
  37. Ferstl, D.; Reinbacher, C.; Ranftl, R.; Ruether, M.; Bischof, H. Image Guided Depth Upsampling Using Anisotropic Total Generalized Variation. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, NSW, Australia, 1–8 December 2013; pp. 993–1000. [Google Scholar] [CrossRef]
  38. Liu, Y.; Gu, K.; Wang, S.; Zhao, D.; Gao, W. Blind Quality Assessment of Camera Images Based on Low-Level and High-Level Statistical Features. IEEE Trans. Multimed. 2019, 21, 135–146. [Google Scholar] [CrossRef]
  39. Hu, R.; Liu, Y.; Gu, K.; Min, X.; Zhai, G. Toward a No-Reference Quality Metric for Camera-Captured Images. IEEE Trans. Cybern. 2023, 53, 3651–3664. [Google Scholar] [CrossRef] [PubMed]
  40. Chan, D.; Buisman, H.; Theobalt, C.; Thrun, S. A Noise-Aware Filter for Real-Time Depth Upsampling. In Proceedings of the Workshop on Multi-Camera and Multi-Modal Sensor Fusion Algorithms and Applications—M2SFA2 2008, Marseille, France, 18 October 2008; pp. 1–12. [Google Scholar]
  41. Liu, J.; Gong, X. Guided Depth Enhancement via Anisotropic Diffusion. In Proceedings of the Pacific-Rim Conference on Multimedia, Nanjing, China, 13–16 December 2013; Springer: Berlin/Heidelberg, Germnay, 2013; pp. 408–417. [Google Scholar] [CrossRef]
  42. He, K.; Sun, J.; Tang, X. Guided Image Filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1397–1409. [Google Scholar] [CrossRef] [PubMed]
  43. Huang, J.; Singh, A.; Ahuja, N. Single image super-resolution from transformed self-exemplars. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 5197–5206. [Google Scholar] [CrossRef]
  44. Liu, W.; Chen, X.; Yang, J.; Wu, Q. Robust Color Guided Depth Map Restoration. IEEE Trans. Image Process. 2017, 26, 315–327. [Google Scholar] [CrossRef] [PubMed]
  45. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image Super-Resolution Using Deep Convolutional Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 295–307. [Google Scholar] [CrossRef] [PubMed]
  46. Li, Y.; Huang, J.; Ahuja, N.; Yang, M. Deep Joint Image Filtering. In Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; pp. 154–169. [Google Scholar]
Figure 1. Framework of the multiscale attention fusion for depth map super-resolution generative adversarial networks. D L R and I L R are the low-resolution depth map and the corresponding downsampled color image. D H R is the high-resolution depth map generated by the generator of our proposed GAN. Furthermore, D G is the ground truth depth map.
Figure 1. Framework of the multiscale attention fusion for depth map super-resolution generative adversarial networks. D L R and I L R are the low-resolution depth map and the corresponding downsampled color image. D H R is the high-resolution depth map generated by the generator of our proposed GAN. Furthermore, D G is the ground truth depth map.
Entropy 25 00836 g001
Figure 2. Hierarchical color and depth attention fusion module.
Figure 2. Hierarchical color and depth attention fusion module.
Entropy 25 00836 g002
Figure 3. Details of multiscale fused feature balance module.
Figure 3. Details of multiscale fused feature balance module.
Entropy 25 00836 g003
Figure 4. Visual comparison of Art with cropped zoomed regions (scaling factor = 4). They are (a) depth map, (b) ground truth, (c) our method with MSE loss, (d) our method with MSE + edge loss.
Figure 4. Visual comparison of Art with cropped zoomed regions (scaling factor = 4). They are (a) depth map, (b) ground truth, (c) our method with MSE loss, (d) our method with MSE + edge loss.
Entropy 25 00836 g004
Figure 5. Visual comparison of Moebius with cropped zoomed regions (scaling factor = 4). They are (a) depth map, (b) ground truth, (c) JBU [1], (d) MRF [2], (e) TGV [37], (f) ESRGAN [6], (g) DJIF [46], (h) DSR [26], (i) CGN [27], and (j) ours.
Figure 5. Visual comparison of Moebius with cropped zoomed regions (scaling factor = 4). They are (a) depth map, (b) ground truth, (c) JBU [1], (d) MRF [2], (e) TGV [37], (f) ESRGAN [6], (g) DJIF [46], (h) DSR [26], (i) CGN [27], and (j) ours.
Entropy 25 00836 g005
Table 1. Quantitative comparison of depth upsampling results (RMSE) on the Middlebury datasets regarding the quantity of RDBs.
Table 1. Quantitative comparison of depth upsampling results (RMSE) on the Middlebury datasets regarding the quantity of RDBs.
Algorithm Art Book Moebius Reindeer Laundry Dolls
16× 16× 16× 16× 16× 16×
RDB_101.192.403.133.840.791.261.582.310.520.811.271.691.451.712.483.351.321.752.163.090.871.131.371.69
RDB_160.812.152.813.470.490.941.301.830.290.580.951.521.281.492.183.031.111.471.962.800.610.841.031.46
RDB_200.792.132.783.430.460.911.271.800.260.550.901.481.241.452.162.981.061.451.922.780.590.820.981.43
RDB_220.78 2.102.763.390.440.891.251.790.230.520.871.461.211.422.152.971.031.441.902.750.560.810.961.38
Table 2. Quantitative comparison of depth upsampling results (RMSE) on the Middlebury datasets regarding GAN types.
Table 2. Quantitative comparison of depth upsampling results (RMSE) on the Middlebury datasets regarding GAN types.
Algorithm Art Book Moebius Reindeer Laundry Dolls
16× 16× 16× 16× 16× 16×
GAN0.962.373.023.680.611.251.582.190.460.731.211.741.501.812.493.271.451.752.323.330.921.141.491.87
RGAN0.81 2.152.813.470.490.941.301.830.290.580.951.521.281.492.183.031.111.471.962.800.610.841.031.46
Table 3. Quantitative comparison of depth upsampling results (RMSE) on the Middlebury datasets regarding generator loss.
Table 3. Quantitative comparison of depth upsampling results (RMSE) on the Middlebury datasets regarding generator loss.
Algorithm Art Book Moebius Reindeer Laundry Dolls
16× 16× 16× 16× 16× 16×
MSE0.792.132.753.420.480.911.261.800.250.540.931.481.211.462.133.001.061.451.922.770.580.811.011.44
MSE + Edge0.812.152.813.470.490.941.301.830.290.580.951.521.281.492.183.031.111.471.962.800.610.841.031.46
Table 4. Quantitative depth upsampling results (RMSE) on Middlebury datasets.
Table 4. Quantitative depth upsampling results (RMSE) on Middlebury datasets.
Algorithm Art Book Moebius Reindeer Laundry Dolls
16× 16× 16× 16× 16× 16×
JBU [1]3.495.086.269.741.782.502.975.441.502.142.994.292.463.294.085.862.423.084.125.841.381.772.433.30
NAF [40]3.525.106.3910.451.852.443.035.761.512.273.014.382.483.364.496.342.493.134.446.201.451.962.833.51
AD [41]4.164.886.659.711.662.232.955.281.442.153.114.402.593.354.516.372.513.174.255.961.311.802.693.59
MRF [2]3.744.756.489.921.732.353.175.341.402.113.174.482.583.294.416.262.543.174.195.921.321.792.663.55
GIF [42]3.154.115.738.531.412.032.583.671.151.652.584.122.192.984.446.581.882.604.025.891.181.672.103.24
SRF [43]2.653.895.518.241.061.622.383.410.901.372.062.991.952.844.105.971.612.403.505.241.141.391.982.79
Edge [44]2.583.244.306.031.211.521.932.600.861.271.992.681.962.893.583.991.622.393.224.291.121.321.512.20
JESR [20]2.633.665.137.051.051.591.832.910.871.211.592.241.952.693.554.881.612.342.844.441.131.321.672.25
JGF [9]3.083.945.257.131.321.822.383.491.141.592.343.472.172.783.504.461.872.593.685.241.131.501.982.71
TGV [37]2.603.344.106.431.201.471.822.630.821.221.642.411.802.713.154.601.612.392.644.171.011.311.612.22
SRCNN [45]2.633.535.347.681.201.471.842.840.861.201.872.672.072.783.544.861.672.182.784.491.151.331.662.64
SRGAN [5]2.023.574.255.901.081.431.852.790.781.231.602.381.862.683.434.371.542.062.753.961.161.371.642.21
ESRGAN [6]1.763.293.865.490.751.371.692.580.651.011.422.121.732.513.194.081.451.992.523.720.971.251.482.02
DJIF [46]1.833.464.074.700.771.501.782.610.561.041.472.092.152.593.244.122.042.232.863.880.911.161.451.94
DSR [26]1.413.033.594.020.631.361.622.380.480.851.291.941.521.982.823.891.641.972.413.560.861.041.271.77
CGN [27]1.272.913.463.880.681.251.552.160.330.791.131.711.491.702.653.621.481.722.353.190.751.021.251.73
Ours0.812.152.813.470.490.941.301.830.290.580.951.521.281.492.183.031.111.471.962.800.610.841.031.46
Table 5. Quantitative depth upsampling results (MAD) on Middlebury datasets.
Table 5. Quantitative depth upsampling results (MAD) on Middlebury datasets.
Algorithm Art Book Moebius Reindeer Laundry Dolls
16× 16× 16× 16× 16× 16×
JBU [1]0.721.131.953.470.300.410.691.210.310.410.691.240.530.650.942.060.430.641.122.030.330.440.631.18
NAF [40]0.721.241.983.680.300.400.671.240.310.410.611.260.540.650.982.040.450.691.132.010.310.450.661.27
AD [41]0.751.222.064.290.330.450.851.540.350.450.741.560.500.641.092.170.480.701.052.750.380.470.811.40
MRF [2]0.801.292.154.250.360.470.861.580.370.430.701.500.520.621.041.960.510.791.102.290.300.460.881.35
GIF [42]0.631.011.703.460.220.350.581.140.230.370.591.160.420.530.881.800.380.520.951.900.280.350.561.13
SRF [43]0.460.971.833.440.150.320.591.120.140.320.511.100.300.551.041.850.230.541.061.990.200.350.561.13
Edge [44]0.410.651.032.110.170.300.561.030.180.290.511.100.200.370.631.280.170.320.541.140.160.310.561.05
JESR [20]0.450.761.512.980.150.270.480.900.160.300.441.010.310.470.691.420.230.500.961.470.200.320.510.92
JGF [9]0.290.470.781.540.150.240.430.810.150.250.460.800.230.380.641.090.210.360.641.200.190.330.591.06
TGV [37]0.450.651.172.300.180.270.420.820.180.290.490.900.320.491.033.050.310.551.223.370.210.330.702.20
SRCNN [45]0.220.530.772.130.090.220.400.790.100.220.420.890.320.470.681.770.240.500.961.540.230.330.571.09
SRGAN [5]0.190.480.702.050.160.280.400.740.150.260.470.810.270.460.651.190.280.440.601.150.200.310.530.88
ESRGAN [6]0.150.360.621.690.140.210.370.650.140.230.420.760.240.430.601.080.230.390.521.080.150.270.480.75
DJIF [46]0.160.380.681.830.180.250.390.680.190.220.400.730.230.390.521.040.200.300.531.120.180.280.450.79
DSR [26]0.130.310.571.460.150.220.340.610.120.200.370.660.210.370.500.960.150.280.461.070.150.240.420.69
CGN [27]0.110.250.481.390.120.180.300.570.130.180.350.620.180.350.460.830.120.260.431.050.140.230.400.65
Ours0.090.230.421.250.090.140.250.480.100.150.260.470.120.240.390.670.110.210.380.940.080.180.330.61
Table 6. Quantitative depth upsampling results (in MAD) on ToFMark dataset.
Table 6. Quantitative depth upsampling results (in MAD) on ToFMark dataset.
BooksDevilShark
Bicubic16.2317.7816.66
JGF [9]17.3919.0218.17
TGV [37]12.8014.9715.53
SRGAN [5]11.7612.8013.92
ESRGAN [6]10.4412.1613.03
DJIF [46]10.8511.6313.50
DSR [26]10.3210.4112.59
CGN [27]10.0110.2311.87
Ours9.699.1411.44
Table 7. Quantitative depth upsampling results (average PSNR) on the MVD test sequences Doorflowers, PoznanStreet, and PoznanCarpark.
Table 7. Quantitative depth upsampling results (average PSNR) on the MVD test sequences Doorflowers, PoznanStreet, and PoznanCarpark.
DoorflowersPoznanStreetPoznanCarpark
Bicubic38.1245.0935.15
JGF [9]38.3645.2835.13
TGV [37]38.4245.5035.18
SRGAN [5]39.9545.6736.60
ESRGAN [6]40.8147.7338.09
DJIF [46]40.6747.6938.26
DSR [26]41.3848.3339.21
CGN [27]41.8048.7239.66
Ours41.9549.1139.84
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, D.; Fan, X.; Gao, W. Multiscale Attention Fusion for Depth Map Super-Resolution Generative Adversarial Networks. Entropy 2023, 25, 836. https://doi.org/10.3390/e25060836

AMA Style

Xu D, Fan X, Gao W. Multiscale Attention Fusion for Depth Map Super-Resolution Generative Adversarial Networks. Entropy. 2023; 25(6):836. https://doi.org/10.3390/e25060836

Chicago/Turabian Style

Xu, Dan, Xiaopeng Fan, and Wen Gao. 2023. "Multiscale Attention Fusion for Depth Map Super-Resolution Generative Adversarial Networks" Entropy 25, no. 6: 836. https://doi.org/10.3390/e25060836

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop