You are currently viewing a new version of our website. To view the old version click .
Electronics
  • Article
  • Open Access

11 May 2023

Graph Convolution Point Cloud Super-Resolution Network Based on Mixed Attention Mechanism

,
,
and
1
The 54th Research Institute of China Electronics Technology Group Corporation, Shijiazhuang 050081, China
2
Institute of Information Science, Beijing Jiaotong University, Beijing 100044, China
*
Author to whom correspondence should be addressed.

Abstract

In recent years, point cloud super-resolution technology has emerged as a solution to generate a denser set of points from sparse and low-quality point clouds. Traditional point cloud super-resolution methods are often optimized based on constraints such as local surface smoothing; thus, these methods are difficult to be used for complex structures. To address this problem, we proposed a novel graph convolutional point cloud super-resolution network based on a mixed attention mechanism (GCN-MA). This network consisted of two main parts, i.e., feature extraction and point upsampling. For feature extraction, we designed an improved dense connection module that integrated an attention mechanism and graph convolution, enabling the network to make good use of both global and local features of the point cloud for the super-resolution task. For point upsampling, we adopted channel attention to suppress low-frequency information that had little impact on the up-sampling results. The experimental results demonstrated that the proposed method significantly improved the point cloud super-resolution performance of the network compared to other corresponding methods.

1. Introduction

The point cloud, a collection of sampling points with spatial coordinates obtained by 3D scanning, finds extensive use in various fields, including autonomous driving and robotics. However, its unstructured and irregular characteristics also pose significant challenges. For instance, the point cloud data collected using acquisition equipment is often locally sparse and non-uniform due to terrain conditions and weather factors. The original point cloud’s characteristics make subsequent processing and applications extremely challenging. As a solution to this problem, the technology of point cloud super-resolution (SR) has drawn broad attention from industry and academia. This technology generates dense, complete, and uniform outputs, thereby resolving the limitations of the original point cloud data set.
Most traditional point cloud SR methods [1,2,3,4] are optimized based on constraints such as local surface smoothness and other shape priors. While these methods are effective for simple models, they fall short when it comes to complex and specialized structures. They require significant computation and perform poorly in practical applications. Hence, there is a pressing need for further improvement in both computation and performance to make them more suitable for real-world scenarios.
Deep learning has achieved significant success in the field of image SR, resulting in the emergence of various excellent image SR networks such as SRCNN [5], MDSR [6], VDSR [7], and ESPCN [8], which have further promoted the development of point cloud SR technology. Currently, researchers use either multi-branch MLPs (Multilayer Perceptron) [9] or methods based on duplicated upsampling operations [10,11]. Multi-branch MLPs are widely used as a general feature extraction method to extract features from point cloud data in deep learning. However, they may overlook critical information from neighboring nodes in 3D point cloud data while performing feature extraction. Moreover, methods based on duplicated upsampling operations not only lead to the loss of local information in the generated high-resolution point clouds, but they also increase the model’s computational parameters, reducing its training efficiency and consuming significant computational resources.
In this paper, we propose a graph convolution point cloud super-resolution network based on a mixed attention mechanism (GCN-MA) to address the challenges faced by traditional point cloud SR methods. GCN-MA consists of two primary parts. The first part is designed for feature extraction, where an improved dense connected network (DCN) comprising multiple feature extraction modules (FEM) is employed to generate high-quality feature maps. Each FEM consists of a dilated graph convolution block (DGCB) and mixed attention mechanism. The mixed attention mechanism involves channel attention within a DGCB to filter information on the channel dimension and multi-head attention between a DGCB to make the network focus on the most valuable information from the convolution layer of different depths. Therefore, the mixed attention mechanism can extract useful high-frequency information for point cloud detail recovery. Finally, the different feature information of each FEM is assigned adaptive weights to enhance the network’s learning ability for better results. The second part of the GCN-MA is the point upsampling module followed by a standard coordinate reconstructor. The output features from the DCN are first extended through the point upsampling module. Then, the channel attention is adopted to suppress low-frequency information that has little impact on the upsampling results. Finally, the feature vectors are rearranged to generate high-quality point cloud data using a standard coordinate reconstructor.
In summary, the main contributions of this paper are summarized as follows:
(1)
We propose the Graph Convolution Point Cloud Super-Resolution Network based on a Mixed Attention Mechanism (GCN-MA), which is designed to fully exploit the global and local features of point cloud through the attention mechanism and improve the network’s learning ability. The GCN-MA pipeline comprises two major parts: feature extraction and point upsampling.
(2)
For feature extraction, we employ an improved dense connected network (DCN) consisting of multiple feature extraction modules (FEM) to generate high-quality feature maps. Each FEM includes a dilated graph convolution block (DGCB) and a mixed attention mechanism. Moreover, adaptive weight allocation between FEMs enhances the learning ability for better point cloud detail recovery.
(3)
The mixed attention mechanism is designed to integrate channel attention within DGCB and multi-head attention between DGCB to make the network focus on the most valuable information from the convolution layer of different depths, making good use of global and local features of point cloud.
(4)
In the point upsampling module, we also adopt channel attention to suppress low-frequency information with little impact on the upsampling results. Following the point upsampling module, feature vectors are rearranged to generate high-quality point cloud data through a standard coordinate reconstructor.
The remainder of the paper is structured as follows. Section 2 presents a review of related work, while Section 3 outlines the details of our proposed point cloud super-resolution method. In Section 4, we conduct an ablation study and provide experimental comparisons with other relevant point cloud super-resolution methods. Finally, in Section 5, we offer our conclusions and discuss potential future research directions.

3. Proposed Method

In this section, we will present the design methodology of our proposed GCN-MA for point cloud SR. We will begin by introducing the overall architecture of our network, followed by a detailed explanation of each individual module.

3.1. Overview

The proposed GCN-MA architecture consists of two main parts: an improved dense connected network (DCN) and a point upsampling module, which is depicted in Figure 1. Initially, the GCN-MA network takes the point cloud’s 3D coordinates as input and encodes the coordinate information through KNN and single-layer graph convolution. Then, DCN extracts deep features using multiple feature extraction modules with a mixed attention mechanism. Each feature extraction module adaptively assigns different weights to enhance the learning ability, producing the final output features. In the point upsampling module, the output features are first channel expanded through a bottleneck layer comprising single-layer graph convolution. To suppress low-frequency information that has little impact on upsampling results, a channel attention mechanism is employed on the expanded feature map. Finally, the feature maps after the output of the channel attention mechanism are coordinate reconstructed to output a super-resolution reconstructed point cloud of shape N × 3 .
Figure 1. The whole architecture of the proposed GCN-MA. The GCN-MA pipeline comprises two major parts: feature extraction and point upsampling.

3.2. Improved Dense Connected Network (DCN)

Figure 2 shows the detailed structure of the improved DCN used in the proposed GCN-MA, which is designed to generate high-quality feature maps. The improved DCN consists of multiple feature extraction modules (FEMs), each composed of a dilated graph convolution block (DGCB) with channel and multi-head attention mechanisms. The mixed attention mechanism in FEM combines channel attention and multi-head attention to make optimal use of the global and local features of the point cloud. The channel attention within DGCB filters information on the channel dimension, while multi-head attention between DGCB enables the network to focus on the most valuable information from the convolution layer of different depths. Furthermore, adaptive weights can be assigned to different extracted features from each FEM, allowing the high-frequency feature information that is useful for the upsampling performance to be utilized, resulting in higher-precision point cloud super-resolution results. More details of the improved DCN will be presented in the following sub-sections.
Figure 2. The structure of improved DCN in the proposed GCN-MA. DCN consists of multiple feature extraction modules (FEM) to generate high-quality feature maps. Each FEM includes a dilated graph convolution block (DGCB) and mixed attention mechanism.

3.2.1. DGCB with Channel Attention

Multi-branch feature fusion is an effective approach to learning features from different branches, where each branch’s feature extraction module can achieve distinct feature extraction results by adaptively learning different parameters. Therefore, in the improved DCN, we employ a multi-branch structure to extract richer features. Additionally, on each branch, dilated graph convolution is utilized to increase the receptive field of the network.
Figure 3 shows the structure of the dilated graph convolution block (DGCB) in the feature extraction module. The DGCB consists of two dilated graph convolution layers, a global maximum pooling layer, and a channel attention mechanism module. To reduce computation, the input features first pass through a bottleneck layer with a 1 × 1 convolution kernel, which reduces the dimension to 64. The feature information computed in the bottleneck layer is then fed into the two dilated graph convolution layers and the global maximum pooling layer. The two dilated graph convolution layers have different dilation rates to capture different receptive fields. The feature information from the dilated graph convolution layer and the global pooling layer are fused at the channel level using the concatenate operation. Due to the disordered nature of point clouds, only the channel attention mechanism is used to filter information on the channel dimension. The final output features are obtained by summing the jump connections from the initial input features. By introducing the channel attention mechanism to the DGCB, the model can refine high-frequency information for point cloud detail recovery and improve the model’s training capability.
Figure 3. DGCB with Channel Attention. The DGCB consists of two dilated graph convolution layers, a global maximum pooling layer, and a channel attention mechanism module.

3.2.2. Multi-Head Attention Mechanism

To enable the network to learn the influence of different point features on point cloud SR, the improved DCN integrates Multi-head Attention with channel attention as a mixed attention mechanism. The channel attention within DGCB filters information on the channel dimension, while the multi-head attention between DGCB helps the network focus on more valuable information from the convolution layer of different depths. By using the multi-head attention mechanism, the network can screen out important information in the feature and ignore most of the unimportant information. The attention is reflected in the weight coefficients, where larger weight coefficients signify more important information.
Figure 4 shows the multi-head attention mechanism used in this paper. In traditional multi-head attention, each single head calculates global information, which often includes a significant amount of irrelevant data that can adversely affect point cloud super-resolution results. To address this issue, we introduce a mask matrix M that suppresses nodes beyond the adjacent points, preventing them from contributing to the weight calculation. The mask matrix is a symmetric matrix centered on the main diagonal, where the number of sub-diagonals on either side represents the number of selected adjacent points. Specifically, the values of the selected primary and secondary diagonals are set to 1, while the others are set to 0. When combined with the weight matrix through a Hadamard product, the positions set to 0 effectively exclude information from points far from the center point. As a result, the mask matrix can significantly reduce computation while improving the point cloud recovery performance.
Figure 4. Multi-head Attention Mechanism. Compared with the traditional multi-head attention mechanism, we introduce a mask matrix M that suppresses nodes beyond the adjacent points, preventing them from contributing to the weight calculation.
Inspired by [28], the attention function of each head can be represented as follows
h e a d i = A t t e n t i o n Q i , K i , V i , M i = softmax Q i K i T d k · M i × V i
where Q , K , V , and d k have the same meaning with the traditional attention function in [28]. Furthermore, in Figure 4, Q i = Q W i Q ; K i = K W i K ; and V i = V W i V . M i denotes the proposed mask matrix of h e a d i . As a result, the final multihead attention function can be indicated as
M u l t i H e a d Q , K , V , M = C o n c a t h e a d 1 , , h e a d h W O
where W O means the projection parameter matrices for the output features.

3.2.3. Adaptive Weights Allocation between FEMs

Residual connection and dense connection are two commonly used network structures for feature extraction in deep learning tasks. While the traditional dense connection method allows each layer in the network to fully utilize the features of its previous layer, it can lead to increased problems as the network deepens. With more layers, the dense connection requires more parameters, leading to a significant increase in computational complexity, training time, and even overfitting.
To overcome these challenges, we propose an approach to adaptively allocate weights between FEMs, which reduces the redundancy of feature information across layers. By preserving the network’s ability to learn effectively, we further reduce computational complexity through the use of fewer parameters. Specifically, we introduce a multilayer perceptron to control the final output features through adaptive learning weight parameters. This enables us to assign the proportion of output features from each layer to achieve more efficient feature learning.
Figure 5 illustrates the process of encoding point cloud features and extracting feature information from each FEM, which is then fused using a dense connection. Unlike traditional dense connections, which usually use channel-level concatenation for feature information, this paper proposes adaptive weight allocation to achieve feature fusion. In this method, feature information generated by each FEM is multiplied by a weight coefficient of adaptive learning, and the final output feature is obtained by summing the multiplication results. The weight coefficient is calculated by MLPs and SoftMax normalization, which are learned using the back-propagation mechanism. Compared to traditional dense connections, the improved dense connection not only learns the weight coefficient adaptively but also focuses on the feature information that is more valuable to the model generation results, which further enhances the model’s learning ability.
Figure 5. Adaptive Weights Allocation. The weights between FEMs can be adaptively allocated to reduce the redundancy of feature information across layers.

3.3. Point Cloud Upsampling Module

In the field of image super-resolution, rearranging image feature information is a common method for image reconstruction. Drawing inspiration from PixelShuffle [8], which is used for image super-resolution, NodeShuffle [27] has been proposed to effectively upsample point clouds. In this paper, we have combined NodeShuffle with channel attention to design the point cloud upsampling module of GCN-MA, as shown in Figure 6. The node feature F l with shape N × C , after feature extraction, is input to the upsampling module, which first goes through a graph convolution layer to expand the number of channels C to rC , where r is the upsampling rate, to obtain the feature information with shape N × rC . The expanded node features are then rearranged to obtain node features of shape rN × C . Next, the channel attention module is applied to extract high-frequency node information that is valid for upsampling in the global and neighborhood through the attention mechanism. Finally, the feature information is input into the coordinate reconstructor to obtain the new point cloud super-resolution image with shape rN × 3 . This approach enables us to obtain better quality point cloud reconstructions.
Figure 6. Point Cloud Upsampling Module. In the point upsampling module, we also adopt channel attention to suppress low-frequency information with little impact on the upsampling results.

3.4. Loss Function

The loss function is a crucial component in point cloud SR, as it greatly influences the overall performance of the network. By accurately representing the difference between the high-resolution original point cloud and the SR reconstructed point cloud estimated by the network, a well-designed loss function can significantly improve the quality of the super-resolution results. Ultimately, a smaller difference value between the two point clouds indicates a more robust network.
In point cloud SR reconstruction models, the loss function typically uses the minimized Chamfer distance to calculate the absolute distance between the SR-reconstructed point cloud image and the original high-resolution point cloud image. The Chamfer distance loss function can be calculated as follows.
L o s s C D P , Q = 1 P p P min q Q p q 2 2 + 1 Q q Q min p P p q 2 2
where P and Q represent the SR point cloud image generated by model prediction and the original high-resolution point cloud image, respectively, while p and q represent any points in P and Q , respectively. · 2 2 denotes the squared Euclidean norm. The aim of this loss function is to calculate the point-to-point absolute distance between two point cloud images and minimize this loss value. By training the network model through the back-propagation algorithm, the generated SR point cloud image becomes more approximate to the original high-resolution point cloud image.
In the field of point cloud SR reconstruction, the most frequently used objective evaluation metric for the quality of the SR reconstructed point cloud image is the Chamfer distance. As mentioned above, the Chamfer distance as the evaluation metric is also calculated as the absolute distance between two point cloud images. The smaller the value of the Chamfer distance metric, the closer the generated SR point cloud image is to the original high-resolution point cloud image, representing better model training performance.

4. Experiments

This section presents the datasets used to train and test our models, followed by the implementation details of our proposed framework. Next, we analyze the different modules in our proposed network and compare our results with several corresponding methods, in terms of both quantitative evaluations and visual quality.

4.1. Datasets

In this paper, we utilized the training set from the PU1K [27] and PUGAN [1] datasets to train our proposed network model and evaluate its performance on the corresponding test sets also provided in the datasets. The PU1K dataset comprised 1147 3D models, with 1020 samples in the training set and 127 samples in the test set. The dataset included models with both simple and complex shapes, allowing us to assess the model’s generalization ability. The PUGAN dataset comprised 147 3D models, with 120 samples in the training set and 27 samples in the test set.
In this paper, we used Chamfer distance (CD) and Hausdorff distance (HD) as evaluation metrics to compare our point cloud SR method with others in a quantitative analysis. CD has been previously introduced in Section 3.4, while HD can be calculated as follows:
D i s t a n c e H D S 1 , S 2 = max h S 1 , S 2 , h S 2 , S 1
h S 1 , S 2 = max x S 1   min y S 2 x y h S 2 , S 1 = max y S 2   min x S 1 y x
where D i s t a n c e H D S 1 , S 2 is the bi-directional Hausdorff distance, and h S 1 , S 2 is the one directional Hausdorff distance. Similar to the Chamfer distance metric, a smaller HD value indicates better reconstruction quality, while a larger HD value indicates more severe distortion in the reconstructed point cloud.

4.2. Implementation Details

In this paper, we have implemented our proposed model using the deep learning framework of TensorFlow. The difference in model parameters could greatly impact experimental results, thus we adjusted the parameters through several experiments combined with experience to achieve an acceptable effect. The network parameters were set as follows: the learning rate was 0.0001, and the batch size was 64. We optimized the models using the Adam optimizer and trained for 200 epochs to obtain our experimental results. Specifically, batchsize denoted the size of the batch, and epoch denoted the number of trainings.

4.3. Ablation Study

In this subsection, we began by examining the fundamental network parameters of the proposed GCN-MA, specifically, the number of FEMs. We then explored the mask matrix settings in the multi-head attention and analyzed the positions of the channel attention. Finally, we conducted a detailed analysis of the proposed components.

4.3.1. Number of FEMs

Each FEM in our proposed GCN-MA network consisted of a dilated graph convolution block (DGCB) with channel and multi-head attention mechanisms. To investigate the impact of the number of FEMs on point cloud super-resolution, we conducted comparison experiments with the same hyperparameters and experimental settings, except for varying the number of FEMs. Table 1 presents the results of the experiments, indicating that the number of FEMs had a significant effect on the performance of the network. We found that the performance improvement of the point cloud super-resolution was the greatest when the number of FEMs was three. Our results demonstrated that increasing the number of FEMs led to a deeper network and helped improve network performance.
Table 1. Performance test of different FEMs on point cloud super-resolution.

4.3.2. Mask Matrix Settings in the Multi-Head Attention

In this paper, we employed the multi-head attention mechanism with a mask matrix, and different mask matrix settings could have varying effects on experimental results. The mask matrix could suppress nodes beyond neighboring points, making full use of local structural features for SR reconstruction tasks.
In this subsection, we further investigated the influence of different numbers of nearest neighbor points on SR performance of the network model. The comparison experiments used the same hyperparameters and experimental settings, except for the number of nearest neighbor points selected by the mask matrix. Table 2 presents the SR performance test with different mask matrix settings, where the last row corresponded to the method without a mask matrix. From Table 2, we can see that the selection of the number of nearest neighbor points had an impact on the SR performance of the model. The SR performance of the compared methods was similar when a small number of nearest neighbor points were selected by the mask matrix, as shown in the first five rows of the table. However, the network SR performance decreased when a large number of points were selected, as in the last two rows of the table.
Table 2. SR performance comparisons with different mask matrix settings.

4.3.3. Positions of the Channel Attention

The proposed GCN-MA incorporated a channel attention mechanism in both the FEM and point cloud upsampling module, which improved the model’s ability to learn feature information in the channel dimension. Table 3 presents the impact of the channel attention mechanism at different positions in the network. The first row of Table 3 shows the performance of the channel attention mechanism only in the FEM, the second row shows its performance only in the upsampling module, and the third row shows its performance in both FEM and the upsampling module. The last row corresponds to the model’s performance without the channel attention mechanism. The experimental results demonstrate that the channel attention mechanism, used at different locations in the network, had different effects on the SR performance and incorporating the channel attention mechanism in both FEM and the upsampling module achieved the best performance compared to the other methods.
Table 3. SR performance comparisons with different positions of the channel attention mechanism.

4.3.4. Study of the Proposed Components

In this subsection, we verified the performance of the proposed components, namely, the attention mechanism and mask matrix. Note that the following experiments were based on the parameter settings described above. Table 4 shows the ablation experiments involving different components. The first row represents adding only the channel attention mechanism, the second row represents adding only the multi-head attention mechanism, the third row represents using the mask matrix in adding the multi-head attention mechanism, and the last row represents adding both the channel attention mechanism module and the multi-head attention mechanism module with the mask matrix. The experiments demonstrated that the different components had varying effects on the network performance, and the proposed method in this paper outperformed the other methods.
Table 4. Ablation experiments involving different components.

4.4. Comparisons with Corresponding Methods

In order to evaluate the effectiveness of the proposed GCN-MA network, we compared the experimental results with several corresponding point cloud SR algorithms, including PU-Net, 3PU, PU-GACNet, AR-GCN, PU-GAN, and PU-GCN.
Firstly, Table 5 presents a comparison of objective evaluation metrics on the PU1K dataset, with Chamfer distance (CD) and Hausdorff distance (HD) being used as evaluation standards. As shown in the table, the proposed model in this paper, which utilized dilated graph convolution and mixed attention mechanisms, achieved lower Chamfer and Hausdorff distances. Table 6 presents a comparison of objective evaluation metrics on PUGAN dataset with CD and HD metrics, which shows that the proposed GCN-MA had better reconstruction performance.
Table 5. Objective quality evaluation results of different algorithms on PU1K dataset.
Table 6. Objective quality evaluation results of different algorithms on PUGAN dataset.
Moreover, we conducted a subjective evaluation of the reconstruction quality by comparing it with other advanced methods. As depicted in Figure 7, while the point cloud SR images generated by the other compared algorithms exhibited some SR effect when compared to the point cloud low-resolution images, the subjective quality could still be further improved. In contrast, the proposed GCN-MA produced smoother edge features and enhanced the overall SR effect, as shown in Figure 7. For instance, in the reconstruction of the seat backrest corners, our method provided clearer details and less noise. Similarly, in the reconstruction of the camel’s ears and feet, as well as the bird’s beak and feet, our proposed method was more effective than the other algorithms.
Figure 7. Subjective reconstruction quality comparison of various methods.

5. Conclusions

In this paper, we proposed a point cloud SR network that combined graph convolution with attention mechanisms. Graph convolution effectively extracted neighborhood information, and a well-designed attention mechanism filtered meaningful information to improve the feature extraction ability of the network. Specifically, we incorporated two different attention mechanisms: the multi-head attention mechanism with mask matrix, which utilized global and neighborhood information and prevented overfitting, and the channel attention mechanism, which improved the learning efficiency and perceptual ability of the model from the channel dimension. Compared with other methods, our proposed GCN-MA achieved better SR reconstruction results in both subjective quality and objective evaluation metrics.

Author Contributions

Conceptualization, T.C. and H.B.; methodology, T.C.; software, T.C. and Z.Q.; validation, Z.Q. and C.Z.; formal analysis, Z.Q. and C.Z.; investigation, Z.Q. and C.Z.; resources, T.C. and Z.Q.; data curation, T.C. and Z.Q.; writing—original draft preparation, T.C. and H.B.; writing—review and editing, T.C. and C.Z.; visualization, T.C. and H.B.; supervision, C.Z. and H.B.; project administration, C.Z. and H.B.; funding acquisition, C.Z. and H.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Fundamental Research Funds for the Central Universities (2019JBZ102).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Alexa, M.; Behr, J.; Cohen-Or, D.; Fleishman, S.; Levin, D.; Silva, C.T. Computing and Rendering Point Set Surfaces. IEEE Trans. Vis. Comput. Graph. 2003, 9, 3–15. [Google Scholar] [CrossRef]
  2. Lipman, Y.; Cohen-Or, D.; Levin, D.; Tal-Ezer, H. Parameterization-free Projection for Geometry Reconstruction. ACM Trans. Graph. 2007, 26, 22–30. [Google Scholar] [CrossRef]
  3. Huang, H.; Li, D.; Zhang, H.; Ascher, U.; Cohen-Or, D. Consolidation of Unorganized Point Clouds for Surface Reconstruction. ACM Trans. Graph. 2009, 28, 1–7. [Google Scholar] [CrossRef]
  4. Huang, H.; Wu, S.; Gong, M.; Cohen-Or, D.; Ascher, U.; Zhang, H. Edge-Aware Point Set Resampling. ACM Trans. Graph. 2013, 32, 9. [Google Scholar] [CrossRef]
  5. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image Super-Resolution Using Deep Convolutional Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 295–307. [Google Scholar] [CrossRef] [PubMed]
  6. Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced Deep Residual Networks for Single Image Super-resolution. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 1132–1140. [Google Scholar]
  7. Kim, J.; Lee, J.K.; Lee, K.M. Accurate Image Super-resolution Using Very Deep Convolutional Networks. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2015; pp. 1646–1654. [Google Scholar]
  8. Shi, W.; Caballero, J.; Huszar, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time Single Image and Video Super-resolution Using an Efficient Sub-pixel Convolutional Neural Network. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2016; pp. 1874–1883. [Google Scholar]
  9. Yu, L.; Li, X.; Fu, C.; Cohen-Or, D.; Heng, P.A. PU-Net: Point Cloud Upsampling Network. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2018; pp. 2790–2799. [Google Scholar]
  10. Wang, Y.; Wu, S.; Huang, H.; Cohen-Or, D.; Olga, S. Patch-based Progressive 3d Point Set Upsampling. In Proceedings of the 2019 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 5958–5967. [Google Scholar]
  11. Yu, L.; Li, X.; Fu, C.; Cohen-Or, D.; Heng, P.A. EC-Net: An Edge-Aware Point Set Consolidation Network. In Proceedings of the 2018 European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 386–401. [Google Scholar]
  12. Heimann, V.; Spruck, A.; Kaup, A. Frequency-selective Geometry Upsampling of Point Clouds. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 1511–1515. [Google Scholar]
  13. Zhang, Y.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image Super-resolution Using Very Deep Residual Channel Attention networks. In Proceedings of the 2018 European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  14. Dai, T.; Cai, J.; Zhang, Y.; Xia, S.T.; Zhang, L. Second-order Attention Network for Single Image Super-resolution. In Proceedings of the 2019 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 11057–11066. [Google Scholar]
  15. Yan, Y.; Liu, C.; Chen, C.; Sun, X.; Jin, L.; Peng, X.; Zhou, X. Fine-grained Attention and Feature-sharing Generative Adversarial Networks for Single Image Super-resolution. IEEE Trans. Multimed. 2022, 24, 1473–1487. [Google Scholar] [CrossRef]
  16. Liang, G.; KinTak, U.; Yin, H.; Liu, J.; Luo, H. Multi-scale Hybrid Attention Graph Convolution Neural Network for Remote Sensing Images Super-resolution. Signal Process. 2023, 207, 108954. [Google Scholar] [CrossRef]
  17. Zhang, Q.; Feng, L.; Liang, H.; Yang, Y. Hybrid Domain Attention Network for Efficient Super-Resolution. Symmetry 2022, 14, 697. [Google Scholar] [CrossRef]
  18. Qian, Y.; Hou, J.; Kwong, S.; He, Y. PUGeo-Net: A Geometry-centric Network for 3d Point Cloud Upsampling. In Proceedings of the 2020 European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 752–769. [Google Scholar]
  19. Bai, Y.; Wang, X.; Ang, M., Jr.; Rus, D. BIMS-PU: Bi-Directional and Multi-Scale Point Cloud Upsampling. IEEE Robot. Au-Tomation Lett. 2022, 7, 7447–7454. [Google Scholar] [CrossRef]
  20. Akhtar, A.; Li, Z.; Auwera, G.V.; Li, L.; Chen, J. PU-Dense: Sparse Tensor-Based Point Cloud Geometry Upsampling. IEEE Trans. Image Process. 2022, 31, 4133–4148. [Google Scholar] [CrossRef] [PubMed]
  21. Li, R.; Li, X.; Fu, C.; Cohen-Or, D.; Heng, P.A. PU-GAN: A Point Cloud Upsampling Adversarial Network. In Proceedings of the 2019 IEEE International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7203–7212. [Google Scholar]
  22. Liu, H.; Yuan, H.; Hou, J.; Hamzaoui, R.; Gao, W. PUFA-GAN: A Frequency-Aware Generative Adversarial Network for 3D Point Cloud Upsampling. IEEE Trans. Image Process. 2022, 31, 7389–7402. [Google Scholar] [CrossRef] [PubMed]
  23. Liu, H.; Yuan, H.; Hou, J.; Hamzaoui, R.; Gao, W.; Li, S. PU-Refiner: A Geometry Refiner with Adversarial Learning for Point Cloud Upsampling. In Proceedings of the 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022; pp. 2270–2274. [Google Scholar]
  24. Liu, C.; Jiang, A.; Kwan, H. Sparse to Dense: LiDAR Point Cloud Upsampling by Mult-modal GAN. In Proceedings of the 2022 IEEE Region 10 Conference (TENCON), Hongkong, China, 1–4 November 2022. [Google Scholar] [CrossRef]
  25. Wu, H.; Zhang, J.; Huang, K. Point Cloud Super Resolution with Adversarial Residual Graph Networks. arXiv 2019, arXiv:1908.02111. [Google Scholar]
  26. Han, B.; Zhang, X.; Ren, S. PU-GACNet: Graph Attention Convolution Network for Point Cloud Upsampling. Image Vis. Comput. 2022, 118, 104371. [Google Scholar] [CrossRef]
  27. Qian, G.; Abualshour, A.; Li, G.; Thabet, A.; Ghanem, B. PU-GCN: Point Cloud Upsampling Using Graph Convolutional Networks. In Proceedings of the 2021 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021; pp. 11683–11692. [Google Scholar]
  28. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, N.A.; Kaiser, L.; Polosukhin, I. Attention is All Your Need. In Proceedings of the 2017 International Conference on Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.