Next Article in Journal
Ubiquitination Occurs in the Mitochondrial Matrix by Eclipsed Targeted Components of the Ubiquitination Machinery
Next Article in Special Issue
Optimization of 3D Immunofluorescence Analysis and Visualization Using IMARIS and MeshLab
Previous Article in Journal
Cathepsins Trigger Cell Death and Regulate Radioresistance in Glioblastoma
Previous Article in Special Issue
Inhibition of EZH2 Ameliorates Sepsis Acute Lung Injury (SALI) and Non-Small-Cell Lung Cancer (NSCLC) Proliferation through the PD-L1 Pathway
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scope Feature Extraction for Intracranial Aneurysm 3D Point Cloud Completion

1
School of Advanced Technology, Xi’an Jiaotong-Liverpool University, Suzhou 215123, China
2
Data Science Research Center, Duke Kunshan University, Kunshan 215316, China
3
Department of Computer Science, University of Liverpool, Liverpool L69 3BX, UK
*
Author to whom correspondence should be addressed.
Cells 2022, 11(24), 4107; https://doi.org/10.3390/cells11244107
Submission received: 14 November 2022 / Revised: 3 December 2022 / Accepted: 13 December 2022 / Published: 17 December 2022
(This article belongs to the Special Issue Bioinformatics and Cells)

Abstract

:
3D point clouds are gradually becoming more widely used in the medical field, however, they are rarely used for 3D representation of intracranial vessels and aneurysms due to the time-consuming data reconstruction. In this paper, we simulate the incomplete intracranial vessels (including aneurysms) in the actual collection from different angles, then propose Multi-Scope Feature Extraction Network (MSENet) for Intracranial Aneurysm 3D Point Cloud Completion. MSENet adopts a multi-scope feature extraction encoder to extract the global features from the incomplete point cloud. This encoder utilizes different scopes to fuse the neighborhood information for each point fully. Then a folding-based decoder is applied to obtain the complete 3D shape. To enable the decoder to intuitively match the original geometric structure, we engage the original points coordinates input to perform residual linking. Finally, we merge and sample the complete but coarse point cloud from the decoder to obtain the final refined complete 3D point cloud shape. We conduct extensive experiments on both 3D intracranial aneurysm datasets and general 3D vision PCN datasets. The results demonstrate the effectiveness of the proposed method on three evaluation metrics compared to baseline: our model increases the F-score to 0.379 ( + 21.1 % )/ 0.320 ( + 7.7 % ), reduces Chamfer Distance score to 0.998 ( 33.8 % )/ 0.974 ( 6.4 % ), and reduces the Earth Mover’s Distance to 2.750 ( 17.8 % )/ 2.858 ( 0.8 % ).

1. Introduction

The intracranial aneurysm can be a life-threatening disease requiring a complicated and costly diagnosis and treatment process. The design of a surgical plan to prevent an aneurysm from rupturing and endangering life is of paramount importance. This kind of surgery requires physicians to operate with pinpoint accuracy since they need clip the neck of aneurysms to prevent rupture. Compared with 2D magnetic resonance angiography (MRA) images, 3D models can provide physicians with more detailed and intuitive information for simulation, diagnosis, and treatment planning. Due to the sensitivity of medical data, there are currently few datasets on intracranial vessels. To our knowledge, the IntrA dataset [1] is the only existing public dataset for intracranial aneurysms neurosurgery simulation, published in 2020, it contains 3D surface models for segmentation and classification in both 3D mesh and point cloud formats. However, the manual preprocessing of raw 3D data is a challenge for experts; the 3D models should be reconstructed to be smooth on their surface (Figure 1), which is a very time-consuming task. For example, the experts spent 50 days to complete this IntrA dataset.
Three-dimensional point clouds have been widely used to represent 3D objects in non-medical fields, because they require less memory than mesh and voxel representations and contain a more comprehensive perspective than multiview images. However, owing to limitations of acquisition devices or occlusions by other objects in the area being imaged [2], 3D point clouds scanned from the real world also suffer from incomplete and uneven shapes, affecting the implementation of downstream tasks, such as segmentation and classification. This problem also occurs in medical fields since CT and MRI are limited by inherent physical limitations [3], noise and ambient light will appear in scanned 2D images, resulting in the rough of 3D models constructed from 2D images. In order to provide more efficient 3D point cloud models for other tasks, point cloud completion has recently attracted significant attention. There are two main challenges in point cloud completion, i.e., feature extraction from the partial shape and completion based on the extracted feature. Based on scholars’ exploration of 3D point clouds, two pioneering works, PointNet [4] and PointNet++ [5], are usually taken as point cloud feature extraction techniques. Specifically, PointNet proposes to extract global features for the three properties of point cloud data format, including unordered points, interaction among points, and invariance under transformations, while PointNet++ partitions the position of each point and aggregates the information of its neighbors to obtain local geometric information. Starting from PointNet, PF-Net [6,7] were proposed based on multi-resolution convolution, which offers richer information for the decoder by aggregating multi-resolution point cloud features. Following feature extraction, the the challenge arises on how to use the decoder to restore the geometric features of the point cloud. Recent attempts [6,8,9,10] have made significant progress in 3D point cloud completion. For example, FoldingNet [11] regards a 3D object as a folding deformation of a 2D grid, which assigns the features extracted by the encoder to each point on the 2D grid, and transforms the 2D grid into a 3D shape through a multilayer perceptron (MLP). Another work AtlasNet [12] splits a 3D object into more 2D planes and obtains the geometry of the 3D object by deforming each plane. However, these works focus mainly on objects with obvious geometric or symmetric structural information, such as airplanes, boats, and lamps (see Figure 2 for illustration).
Despite its importance, there have been few completion tasks in the medical field. In the previous medical-related completion tasks, the objects processed merely miss a small part in their complete shape. For instance, missing teeth completion [13] and skull completion [14]. These instances differ from what is required by our motivation, i.e., generating ideal and smooth 3D models from raw datasets. Different from other common objects, the intracranial vessel and aneurysm fragments are typically asymmetric, with varying geometry and topology. Additionally, the number of data points is limited because of their size. Therefore, the completion model mentioned for ordinary objects may not be suitable for completing intracranial blood vessels and aneurysms.
In this paper, we propose an encoder–decoder-framed deep learning model called the Multi-Scope Feature Extraction Network, which takes one partial intracranial vessel or aneurysm point cloud as its input. we then predicts its complete 3D structure. Concretely, we develop a Multi-Scope feature Extraction encoder to obtain different scope neighborhoods for one point, then extract and fuse the multi-scope features together. For the decoder, we adopt style-based folding using SpareNet [15], but add original partial points coordinates to guide it to recover more structural information from the input. Furthermore, we take the advantage of the pioneer 3D point cloud completion work PCN [2] by utilizing a coarse-to-fine framework to calculate both coarse point cloud and refined point cloud loss between the ground truth with shape completion during the training process.

2. Materials and Methods

2.1. Data Sources

In this paper, we adopt the IntrA dataset [1] for the experiments to evaluate the proposed method, which was published in 2020 as a point-based and a mesh-based 3D aneurysm model collection. It contains three data types, including 103 complete models, 1909 generated segments, and 116 annotated segments. The 103 complete brain vessels were reconstructed by scanned 2D MRA images from patients using life sciences software, Amira 2019 (Thermo Fisher Scientific, Waltham, MA, USA), and can take 50 workdays to process [1]. The generated segments are split from the complete models after manually cleaning the data and re-meshing, resulting in 1694 healthy vessels and 215 aneurysms. Each segment has approximately 500 to 1700 points at Geodesic Distance 30. According to IntrA, their dataset is provided to support classification, segmentation, reconstruction. During the IntrA dataset generating process, we found that the authors face the challenge that the restored 3D data model from 2D scanned images were not complete, and could only be restored manually by experts. After carefully studying 103 complete models, we were able to identify incomplete regions as shown in Figure 3 and Figure 4. We can see that even after a long time of reconstruction processing period, the reconstructed 3D models still have noise and their surfaces are not smooth [1].

2.2. Experimental Dataset

Some previous works [7,9] set the partial point cloud number equal to the complete point number which transfers the point completion task to a point rearrangement task. This setting is, however, limited by the choice of datasets. The model needs duplicate points or add zero points when the input points are insufficient. The number of output points is also limited to a fixed number. The point cloud completion task can also be seen as a missing part prediction from the existing part [6,16,17] where the final shape isformed by merging the existing partial part and the predicted missing part. These methods need filling the gap between the two parts, where the missing part should shape the same distribution as the input. Since the input shape is incomplete already, its distribution is not guaranteed to be consistent with the complete shape of the ground truth.
Another typical setting that follows PCN [2] is to assume that the input incomplete point cloud and the output complete point cloud are independently distributed. They both represent the same object but are separately sampled from the ground truth. The purpose of this setting is that the incomplete point cloud only provides the model with a clue to predict the complete point cloud, which has structural characteristics similar to the predicted part of the point cloud, but it is not part of the complete point cloud. This setting is more flexible for the number of input and output points set by the network. Since the output point cloud is a complete shape, its distribution is consistent.
Our data setting follows PCN to make a different distribution that is more accurate for intracranial vessel situations. Since we do not have a paired complete model and incomplete model of 103 brain vessels, we choose the IntrA dataset’s 1909 generated segments subset as our dataset. We follow PCN and deal with pure point cloud dataset completion tasks, to generate an extended point cloud completion dataset, called IntrACompletion dataset.
To generate the IntrACompletion dataset from the original IntrA dataset, we take all 1909 segments models, including two categories: 1694 vessels and 215 aneurysms. The complete point clouds are sampled from these 3D meshes uniformly. Similar to PCN, we use back-projected depth images to generate partial point clouds. This operation can separate the distribution of the partial point cloud and its complete shape. More specifically, the partial point clouds are not subsets from their complete point cloud. In addition, we use eight random angles to generate eight different viewpoints of partial point clouds to enrich the dataset (Figure 5). To divide training, validation, and test groups, we use 8:1:1 ratio. That means we have 1356:169:169 for vessels and 173:21:21 for aneurysms (Figure 6) in the following experiments.

2.3. Proposed Methods

In this paper, we proposed a Multi-Scope Feature Extraction Network for intracranial aneurysm vessel point cloud completion on the generated IntrACompletion dataset. This model exploits an encoder–decoder framework and adopts a coarse-to-fine pipeline from PCN [2]. It has a multi-scope aggregate-based encoder to extract global feature G from partial input, and a style-based folding decoder predicts the corresponding coarse complete shape. Last, a refinement network is considered to refine the coarse to fine output. The overall architecture is shown in Figure 7a.

2.3.1. Multi-Scope Aggregate Encoder

The characteristic of 3D point clouds is that they are represented by discrete coordinates of points. This representation makes it lighter to store. However, point clouds do not have vertices and surface structures like 3D meshes, therefore, we cannot directly obtain one point’s local features from it. The popular feature extraction method for point clouds is to find a point and its k neighbor points with the closest Euclidean Distance, or neighbors within a spherical distance. Aggregating the coordinates and feature information of surrounding points to ensure a point has not only its own features but also local features provided by its neighbors [5,18]. These works usually use a single k only in information extraction, resulting in weak performance in real-world datasets with an uneven point distribution. To mitigate this issue, and inspired by VRCNet [19], we propose to apply multi-scope feature extraction. Different from VRCNet, which uses the point self-attention kernel module to select points, we follow [15,20], using EdgeConv [18] to extract features and engage Squeeze-and-Extraction [21] to enhance the features.
Our proposed Multi-Scope Aggregate Module (MSA) is shown in Figure 7b. It is combined with several Single-Scope Modules (SSM) (Figure 8). Each SSM module takes an N × C i n unordered partial point cloud as its input, denoted as P i n . For each point p i P i n , we find its k-nearest neighbors (k-NN) p i j P i n and their corresponding features { e i j R C | j = 1 , 2 , 3 , , k } . Then we use EdgeConv [18] to make a directed graph between p i and p i j . It combines the local neighborhood information ( e i j e i ) with global information e i on each edge as ( e i , e i j e i ) . The merged features are sent to a shared MLP to learn each edge’s feature. Accordingly, the learned new features e i j of the point p i are assigned as follows:
e i j = max j : ( i , j ) ε MLP ( e i , e i j e i ) ,
where e i j R C , i { 1 , 2 , , N } , j { 1 , 2 , , k } . All new assigned features e i as graph information are squeezed into a channel descriptor through N × k dimensions followed by a squeeze operation in the SE block [21]. We obtain a gating vector s with a Sigmoid activation called excitation as follows:
s = σ ( MLP ( 1 N × k i = 1 N j = 1 k e i j ) ) .
After obtaining new features e i j from EdgeConv [18] and the gating vector s from the SE block, the MSA layer multiplies the new features by the gating vector. It uses max-pooling to reduce the edges of each graph to find the most representative edge and its features as the final features N × C o u t for P i n points.
Use of a fixed k for the neighborhood may cause unbalanced information. Even if we can rebalance by using multi-resolution processes, such as PFNet [6] and Multiresolution tree networks [7], using one k value in different downsampled point sets, the uneven density distribution problem cannot be avoided. To this end, we borrow the R-PSK module provided by VRCNet [19]. The R-PSK module is designed to fuse different scope neighborhood relationships for each point in the point cloud. Our MSA module assigns multiple k values for SSM layers. After getting a C i n dimensional feature, the MSA module will take one MLP to enhance the feature’s expression ability to C o u t and send the new higher dimensional feature to multiple SSM layers. Then average-pooling can be applied to merge the features from different scopes. These merged features are used for weights through Softmax and MLP. The weights will guild the feature before SSM modules to realize self-attention.
Our encoder contains four MSA modules which aim to extract different features from low to high-dimension feature space. In the end, the features from the last MSA module will be expanded by both max-pooling and average-pooling to filter important information as global features G.

2.3.2. Style-Based Folding Decoder

Folding-based point cloud deformation [2,11,12,22,23,24,25] usually concatenates the global shape, 2D grid, or point cloud coordinates together and learns a mapping between 2D grid and 3D form. This mapping is learned from multiple MLP layers. Since the global shape remains at the first layer, its effect on the overall shape generation is diminished in later layers. To overcome this shortcoming, SpareNet [15] proposes the style-based folding decoder inspired by StyleGAN [26]. It injects the style into each folding internal layer, which is from the global features generated by the encoder. Style-based folding improves the quality of the generated point cloud. Our folding-based decoder borrows SpareNet’s style-based folding decoder that injects style information from global features into MLP layers. Furthermore, we directly add coordinate information of the input point cloud to assist the MLP in capturing more primitive input details.
In summary, Figure 7c shows that our style-based folding decoder takes three different representations for the object: the global features G generated from the MSA encoder, initial 2D grids P n , and partial input point cloud P i n as inputs. In the style-based module (Figure 9), we learn two modulation parameters γ g and β g from the global features G through MLP, and then utilize an Adaptive Instance Normalization (AdaIN) proposed by StyleGAN [26] to transfer feature x i to new features:
x i = A d a I N x i ,
where
A d a I N x i = γ g x i μ x i σ x i + β g .
In the above, μ and σ denote the mean and variance, respectively. After AdaIN, the feature will be transferred through the SE layer and ReLU to a new feature. As with AtlasNet [12], we assume that each 3D object is combined with K patches to K decoders ( K = 32 in our experiments). Each patch is deformed from an initial 2D grid, and each grid is evenly divided into small grids. The vertex coordinates of each small grid are taken as inputs for each decoder. After the style-based folding decoder, each 2D grid learns a mapping to generate a 3D surface. All surfaces are combined directly as a complete coarse shape.

2.3.3. Refinement Module

The 32 surfaces are generated separately. To avoid overlapping between each surface, our model follows MSN [24] by using an expansion penalty (5) and minimum density sampling (MDS) (6) to merge and refine the coarse points from the 32 surfaces as follows:
L e x p a n s i o n = 1 K N 1 i K ( u , v ) T i 𝟙 d i s ( u , v ) λ l i d i s ( u , v ) ,
where 𝟙 denotes the indicator function sharing the filtering distance that is shorter than λ l i . Denoting P i = p j | 1 j i as the set of first i sampled points, p i is the ith sampled point. MSD will return a point that has minimum density as follows:
P i = arg min x P i 1 p j P i 1 exp x p j 2 / 2 σ 2 .
In the expansion penalty, every point in each patch will be treated as a vertex set. Moreover, a minimum spanning tree T i [27] is built based on Euclidean distances. Distance between vertex u and vertex v denotes d i s ( u , v ) . Each minimum spanning tree can express the distribution of a patch, and longer distances are penalized until they converge.

2.3.4. Joint Loss Function

Earth Mover’s Distance (EMD) is widely used in the evaluation of point cloud [28], which measures the similarity between the generated point cloud and ground truth ( g t ) as defined by:
L E M D ( P 1 , P 2 ) = min ϕ : P 1 P 2 1 P 1 x S 1 x ϕ x 2 ,
where ϕ is a bijection. In our experiments, we use EMD loss (7) to train the model. To better guide the model optimization, we set two g t points for the coarse point cloud and refined point cloud, respectively. Both g t points are randomly sampled from the dataset. Finally, we adopt a joint loss method to combine the loss between each point cloud and supervise the network’s learning ability in segments, which is defined by
L j o i n t = L E M D ( P c o a r s e , P g t _ c o a r s e ) + L E M D ( P r e f i n e , P g t _ r e f i n e ) + α L e x p a n s i o n ,
where we set α = 0.1 in our experiments.

2.4. Experimental Setting

Our MSENet is implemented using Pytorch [29] and CUDA. The optimizer is important and it can affect the training result [30], we use the Adam [31] optimizer with β 1 = 0 and β 2 = 0.9 . The partial input number is 512 and randomly sampled from one of eight angles. The corresponding complete points number is 2048 and randomly sampled from the datasets. We set a batch size of 32, learning rate as 0.0001, and epoch as 150. We train all the models using the joint loss (8) on a single Quadro RTX 8000. To evaluate our proposed model, we use three metrics in the following experiments, including EMD (7), CD (9), and F-score @1%. The CD measure function [28] is calculated by
L C D ( P 1 , P 2 ) = 1 2 ( 1 P 1 x P 1 min y P 2 x y + 1 P 2 y P 2 min x P 1 x y ) .
Comparing the above-mentioned two similarity measure metrics EMD and CD, CD is more computationally efficient while EMD can better capture the shape similarity between two point clouds [32].

2.5. Evaluations

We compare our MSENet with other typical methods in this research. AtlasNet [12] generates a point cloud from several patches, FolidingNet [11] reconstructs a point cloud from a 2D grid, and PCN [2] adopts it to generate a complete point cloud. MSN [24] is the baseline of point cloud completion, GRNet [33] proposed a 3DConv gridding method, SpareNet [15] proposes a style-based folding decoder. All the methods are evaluated on two datasets, including our generated IntrACompletion dataset in the medical field and the PCN dataset provided by PCN [2] in the general 3D vision field, and the experimental results are shown in the following section.

3. Results

3.1. Evaluation on IntrACompletion Dataset

The results are compared in Table 1, Table 2 and Table 3 for the evaluation metrics F-score @1%, CD, and EMD, respectively. We can see that our proposed model offers the best performance on all the metrics for both aneurysm and vessel data point cloud completion, which demonstrates the effectiveness of the proposed method. Compared to point cloud baseline work MSN [24], both our model and MSN exploit 32 surfaces to reconstruct the complete point cloud based on the partial point cloud, but our MSA module takes the multi-scope nearest neighbor range to learn richer features for the decoder. Our MESNet improves the F-score from 0.313 to 0.379, reduces the EMD from 3.345 to 2.750, and 34 % lower in CD. Style-based folding then helps each layer of folding obtain global information. Compared to SpareNet [15], which also uses a style-based folding decoder, our model reduces the CD value by 15 % and the EMD value by 6 % . Our decoder makes this improvement through partial point cooperation. We will discuss the effect of the MSA module and partial point cooperation in Section 3.3 Ablation Study. Comparing the performance over the two types of data for aneurysm and vessel, we can observe that all models have better performance on vessel data because we have more vessel data resulting in model bias to the vessel data.
Note that GRNet [33] was proposed to complete the dense point cloud. It originally uses 64 3 grids to generate and sample a 2048 point coarse cloud from 2048 partial input points, and, finally, generate a 16,384 point dense point cloud. For fair comparison, we use the same ratio to set its network: 32 3 grids to generate and sample a 512 point coarse cloud from 512 partial input points, and finally generate a 2048 point point cloud.
As the purpose of completion is to offer a better 3D model for diagnoses and practices, we make a visualization comparison here. We take six examples from blood vessels and aneurysms for visual comparison. In Figure 10, the first row is the incomplete point cloud input, the following rows are the performance of different models, and the last row is the complete ground truth for comparison. Compared to the folding-based completion models PCN [2], MSN [24], SpareNet [15], we can see that in Aneurysm (a), Vessel (a), and Vessel (c), SpareNet and our model can capture the details of vessel bifurcations, but the others can only recover rough outlines. This is because we apply the style-based folding to inject the global feature in each folding layer. Vessel (b) shows that not every model can reconstruct the tube-like shape. Aneurysm (b) and Aneurysm (c) give us more complex structures.

3.2. Evaluation on PCN Dataset

To demonstrate the effectiveness of the proposed method in the general point cloud completion in 3D vision, we also conduct the experiments on the PCN dataset [2], which contains airplanes, cabinets, cars, chairs, lamps, sofas, tables, and vessels (boats). For a fair comparison, we use the same train/val/test splits with SpareNet [15] and GRNet [33]. For the overall evaluation, our model is the best in F-score and EMD, as shown in Table 4 and Table 5. For each category, our performance is better in sofa, table, and vessels in Table 6 evaluated by CD, and outperforms 7 out of 8 categories in F-score. The visualization comparison is shown in Figure 11. We can see AtlasNet [12], PCN [2], MSN [24], SpareNet, and our MSENet completing the point cloud using multiple surfaces are able to better represent object surfaces. However, FoldingNet [11] uses a single surface to complete the point cloud, making it hard to separate different parts of objects, such as the wheel section in the third column. In the first column we displayed, due to the missing middle region of the plane the head part of the input point cloud, the full plane head part recovered by MSN and SpareNet is also missing, but our planes head part is complete. In the fifth column, only SpareNet and our model using style-based folding can recover the details of the pole part of the lamp. Compared to SpareNet, our completed point cloud has fewer holes, such as the plane in the first column and the table in the seventh column.

3.3. Ablation Study

The proposed method contains two key components, which are MSA module and partial point cooperation. We add them step by step to examine their effectiveness. The ablation studies on IntrACompletion are presented in Table 7. The experimental setting and evaluations are the same as those used in Section 2.4 and Section 2.5. We first set our baseline as using one SSM module assigned a k value equal to 5 in our MSA encoder, and remove the partial input for the style-based folding decoder. The EMD score is 3.064 × 10 3 , CD score is 1.267 × 10 3 , and F-score @ 1 % is 0.354 . Compared to the baseline, adding multiple SSM modules in our encoder can reduce 0.091 × 10 3 EMD score, and 0.086 × 10 3 CD score. Partial input can reduce 0.133 × 10 3 EMD score, and 0.149 × 10 3 CD score. The partial input can also increase 0.004 F-score @ 1 % compared to the baseline. With all parts added, the EMD score was reduced 0.314 × 10 3 , and 0.296 × 10 3 was reduced in the CD score. Additionally, the two parts increase 0.025 for F-score @ 1 % .
We also performed the ablation study on the PCN dataset as shown in Table 8. Compared to the baseline in the first row, our multiple SSM modules and partial input can reduce the EMD score from 2.890 × 10 3 to 2.858 × 10 3 , and enhance the F-score @ 1 % from 0.313 to 0.320 . We can note that any addition of MSA module or partial point cooperation can improve the model performance, and MSENet performs better when the two modules are combined.

4. Discussion

The development of deep learning in the medical field has been applied to many areas, such as anonymous detection, biological data mining, disease detection, education, etc. [34]. However, some of the open-access data sources listed in the article for disease diagnosis and segmentation only contain hundreds of patient samples. Dataset amount can directly affect the training efficiency, and more samples are helpful. Therefore, how to deal with the original data to make it usable is also worth studying.
Three-dimensional representations are more expressive than two-dimensional data. Furthermore, in medical diagnosis, 3D representations can provide physicians with a clearer and more complete model. Three-dimensional data can help improve the accuracy and efficiency of diagnosis. The full 3D model can also provide clearer objects to practice on in experiments. Since intracranial aneurysm’s are life-threatening, 3D data can provide doctors with better imaging results during diagnosis and aid in developing more accurate surgical plans. The current 3D data are mainly composed of multiple 2D MRA images, and the quality of the synthesis is usually incomplete and noisy, as shown in Figure 3. Such rough data also need be manually refined by experts, which is very time-consuming. We propose a 3D point cloud completion model called Multi-Scope Feature Extraction Network (MSENet) that can complete point clouds of incomplete vessels and aneurysms.
To our knowledge, our model is the first to address the problem of missing data in 3D medical data acquisition. Different from other functional completion problems in the medical field, such as tooth loss [13] or skull completion [14], our model mainly promotes the integrity of the data from the source of data collection and provides more convenient and fast data preprocessing for medical diagnosis and simulation exercises. Lack of data problems are common in the medical field, e.g., in the IntrACompletion dataset which has eight times more vessels than aneurysms as shown in Figure 6. Table 1, Table 2 and Table 3 show that all of the methods show better performance in vessel completion. At present, there is little research considering the differences between different categories when completing point clouds, which we will explore for future research.

5. Conclusions

This paper aims to mitigate the incomplete data issue regarding the 3D image reconstruction process for intracranial vessels and aneurysms. We propose a Multi-Scope Feature Extraction Network to complete the partial 3D point cloud. This network engages a multi-scope aggregate module to explore and merge different scopes of neighbors to enhance feature extraction. Then we design a style-based folding decoder to concatenate the original partial points coordinates directly, where the coordinates can offer more existing structure to guide the decoder’s ability. We take the 3D intracranial vessels and aneurysms dataset from IntrA [1] to evaluate our proposed model. We also use a general dataset PCN [2] to confirm this model’s generalizability for other 3D objects. Extensive experimental results demonstrate the effectiveness of the proposed method. Compared to the baseline work MSN [24] on the two datasets, our model increases the F-score to 0.379 ( + 21.1 % )/ 0.320 ( + 7.7 % ), reduces Chamfer Distance score to 0.998 ( 33.8 % )/ 0.974 ( 6.4 % ), and reduces the Earth Mover’s Distance to 2.750 ( 17.8 % )/ 2.858 ( 0.8 % ).

Author Contributions

The experiment was mainly completed by W.M.; W.M., X.Y., and Q.W. participated in the method discussion; X.Y. gave guidance in the medical field; Q.W., K.H., and X.H. are the supervision team; all the authors participated in the writing and revision of this paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China under no. 61876154 and no. 61876155; Jiangsu Science and Technology Programme (Natural Science Foundation of Jiangsu Province) under no. BE2020006-4; and Natural Science Foundation of the Jiangsu Higher Education Institutions of China under no. 22KJB520039.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All datasets used in this paper are public for research usage. IntrA dataset [1] can be found here: [https://github.com/intra3d2019/IntrA, accessed on 6 September 2022], PCN [2] dataset can be found here: [https://github.com/wentaoyuan/pcn, accessed on 9 October 2021].

Acknowledgments

The authors are very grateful to Mark Leach [[email protected]] for his help with the English editing of the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yang, X.; Xia, D.; Kin, T.; Igarashi, T. INTRA: 3D intracranial aneurysm dataset for deep learning. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2653–2663. [Google Scholar] [CrossRef]
  2. Yuan, W.; Khot, T.; Held, D.; Mertz, C.; Hebert, M. PCN: Point completion network. In Proceedings of the 2018 International Conference on 3D Vision, Verona, Italy, 5–8 September 2018; 2018; pp. 728–737. [Google Scholar] [CrossRef] [Green Version]
  3. Goyal, B.; Dogra, A.; Khoond, R.; Al-Turjman, F. An Efficient Medical Assistive Diagnostic Algorithm for Visualisation of Structural and Tissue Details in CT and MRI Fusion. Cogn. Comput. 2021, 13, 1471–1483. [Google Scholar] [CrossRef]
  4. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep learning on point sets for 3D classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 77–85. [Google Scholar] [CrossRef] [Green Version]
  5. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep hierarchical feature learning on point sets in a metric space. Adv. Neural Inf. Process. Syst. 2017, 5100–5109. [Google Scholar]
  6. Huang, Z.; Yu, Y.; Xu, J.; Ni, F.; Le, X. PF-Net: Point fractal network for 3D point cloud completion. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; Volume 4, pp. 7659–7667. [Google Scholar] [CrossRef]
  7. Gadelha, M.; Wang, R.; Maji, S. Multiresolution tree networks for 3D point cloud processing. In Computer Vision—ECCV 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2018; Volume 11211, pp. 105–122. [Google Scholar] [CrossRef] [Green Version]
  8. Wang, X.; Ang, M.H.; Lee, G.H. Cascaded Refinement Network for Point Cloud Completion. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 787–796. [Google Scholar] [CrossRef]
  9. Sarmad, M.; Lee, H.J.; Kim, Y.M. RL-GAN-net: A reinforcement learning agent controlled gan network for real-time point cloud shape completion. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5891–5900. [Google Scholar] [CrossRef] [Green Version]
  10. Hu, T.; Han, Z.; Shrivastava, A.; Zwicker, M. Render4Completion: Synthesizing multi-view depth maps for 3D shape completion. In Proceedings of the International Conference on Computer Vision Workshop, Long Beach, CA, USA, 15–20 June 2019; pp. 4114–4122. [Google Scholar] [CrossRef] [Green Version]
  11. Yang, Y.; Feng, C.; Shen, Y.; Tian, D. FoldingNet: Point Cloud Auto-Encoder via Deep Grid Deformation. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 206–215. [Google Scholar] [CrossRef]
  12. Groueix, T.; Fisher, M.; Kim, V.G.; Russell, B.C.; Aubry, M. A Papier-Mâché Approach to Learning 3D Surface Generation. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 216–224. [Google Scholar] [CrossRef]
  13. Ping, Y.; Wei, G.; Yang, L.; Cui, Z.; Wang, W. Self-attention implicit function networks for 3D dental data completion. Comput. Aided Geom. Des. 2021, 90, 102026. [Google Scholar] [CrossRef]
  14. Kodym, O.; Li, J.; Pepe, A.; Gsaxner, C.; Chilamkurthy, S.; Egger, J.; Španěl, M. SkullBreak/SkullFix—Dataset for automatic cranial implant design and a benchmark for volumetric shape learning tasks. Data Brief 2021, 35, 106902. [Google Scholar] [CrossRef] [PubMed]
  15. Xie, C.; Wang, C.; Zhang, B.; Yang, H.; Chen, D.; Wen, F. Style-based Point Generator with Adversarial Rendering for Point Cloud Completion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021. [Google Scholar] [CrossRef]
  16. Zhang, W.; Yan, Q.; Xiao, C. Detail Preserved Point Cloud Completion via Separated Feature Aggregation. In Computer Vision—ECCV 2020; Lecture Notes in Computer Science; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Springer: Cham, Switzerland, 2020; Volume 12370, pp. 512–528. [Google Scholar] [CrossRef]
  17. Mendoza, A.; Apaza, A.; Sipiran, I.; Lopez, C. Refinement of Predicted Missing Parts Enhance Point Cloud Completion. arXiv 2020, arXiv:2010.04278. [Google Scholar]
  18. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic graph Cnn for learning on point clouds. ACM Trans. Graph. 2019, 38, 1–12. [Google Scholar] [CrossRef] [Green Version]
  19. Pan, L.; Chen, X.; Cai, Z.; Zhang, J.; Zhao, H.; Yi, S.; Liu, Z. Variational Relational Point Completion Network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  20. Hong, J.; Kim, K.; Lee, H. Faster dynamic graph CNN: Faster deep learning on 3d point cloud data. IEEE Access 2020, 8, 190529–190538. [Google Scholar] [CrossRef]
  21. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  22. Wen, X.; Li, T.; Han, Z.; Liu, Y.S. Point Cloud Completion by Skip-Attention Network with Hierarchical Folding. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1936–1945. [Google Scholar] [CrossRef]
  23. Li, R.; Li, X.; Fu, C.W.; Cohen-Or, D.; Heng, P.A. PU-GAN: A point cloud upsampling adversarial network. In Proceedings of the IEEE International Conference on Computer Vision, Long Beach, CA, USA, 15–20 June 2019; pp. 7202–7211. [Google Scholar] [CrossRef] [Green Version]
  24. Liu, M.; Sheng, L.; Yang, S.; Shao, J.; Hu, S.M. Morphing and sampling network for dense point cloud completion. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 11596–11603. [Google Scholar] [CrossRef]
  25. Wang, H.; Liu, Q.; Yue, X.; Lasenby, J.; Kusner, M.J. Unsupervised Point Cloud Pre-training via Occlusion Completion. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021. [Google Scholar] [CrossRef]
  26. Karras, T.; Laine, S.; Aila, T. A Style-Based Generator Architecture for Generative Adversarial Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 4217–4228. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Prim, R.C. Shortest connection networks and some generalizations. Bell Syst. Tech. J. 1957, 36, 1389–1401. [Google Scholar] [CrossRef]
  28. Fan, H.; Su, H.; Guibas, L. A point set generation network for 3D object reconstruction from a single image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2463–2471. [Google Scholar] [CrossRef] [Green Version]
  29. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32. [Google Scholar]
  30. Huang, K.; Hussain, A.; Wang, Q.F.; Zhang, R. Deep Learning: Fundamentals, Theory and Applications; Springer: Berlin/Heidelberg, Germany, 2019. [Google Scholar] [CrossRef]
  31. Kingma, D.P.; Ba, J.L. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015; pp. 1–15. [Google Scholar]
  32. Yu, L.; Li, X.; Fu, C.-W.; Cohen-Or, D.; Heng, P.-A. PU-Net: Point Cloud Upsampling Network Lequan. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2790–2799. [Google Scholar]
  33. Xie, H.; Yao, H.; Zhou, S.; Mao, J.; Zhang, S.; Sun, W. GRNet: Gridding Residual Network for Dense Point Cloud Completion. In Computer Vision—ECCV 2020; Lecture Notes in Computer Science; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Springer: Cham, Switzerland, 2020; Volume 12354, pp. 365–381. [Google Scholar] [CrossRef]
  34. Mahmud, M.; Kaiser, M.S.; McGinnity, T.M.; Hussain, A. Deep Learning in Mining Biological Data. Cogn. Comput. 2021, 13, 1–33. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Examples of reconstructing ideal mesh from raw data in IntrA dataset. (a) Original raw data built from 2D images. (b) Smooth and complete part of aneurysms. (c) Smooth and complete part of the healthy vessel.
Figure 1. Examples of reconstructing ideal mesh from raw data in IntrA dataset. (a) Original raw data built from 2D images. (b) Smooth and complete part of aneurysms. (c) Smooth and complete part of the healthy vessel.
Cells 11 04107 g001
Figure 2. Airplanes, cars, and lamps examples in PCN dataset.
Figure 2. Airplanes, cars, and lamps examples in PCN dataset.
Cells 11 04107 g002
Figure 3. Visualization of incomplete IntrA raw dataset. Incomplete parts are circled.
Figure 3. Visualization of incomplete IntrA raw dataset. Incomplete parts are circled.
Cells 11 04107 g003
Figure 4. Comparison of IntrA data before and after manually refined: (a) The coarse surface before manually refined, (b) The smooth surface after manually refined.
Figure 4. Comparison of IntrA data before and after manually refined: (a) The coarse surface before manually refined, (b) The smooth surface after manually refined.
Cells 11 04107 g004
Figure 5. One example of the IntrACompletion dataset. The complete point cloud is shown in the middle, and its 8 different angles partial point clouds are shown to surround it.
Figure 5. One example of the IntrACompletion dataset. The complete point cloud is shown in the middle, and its 8 different angles partial point clouds are shown to surround it.
Cells 11 04107 g005
Figure 6. Vessels and aneurysms distribution in the train, validation, and test.
Figure 6. Vessels and aneurysms distribution in the train, validation, and test.
Cells 11 04107 g006
Figure 7. Overall architecture of the proposed Multi-Scope Feature Extraction Network (MSENet). (a) Our MSENet contains three modules: a feature extractor built of several MSA modules to get the global features G for partial point cloud input, a style-based folding decoder to generate a coarse but complete point cloud, and a merge and sample module to refine the coarse point cloud; (b) the detailed structure of the MSA module; and (c) the detailed structure of the style-based folding.
Figure 7. Overall architecture of the proposed Multi-Scope Feature Extraction Network (MSENet). (a) Our MSENet contains three modules: a feature extractor built of several MSA modules to get the global features G for partial point cloud input, a style-based folding decoder to generate a coarse but complete point cloud, and a merge and sample module to refine the coarse point cloud; (b) the detailed structure of the MSA module; and (c) the detailed structure of the style-based folding.
Cells 11 04107 g007
Figure 8. This is the Singe-Scope Module (SSM) which is the main module in our MSA module. Each SSM module is assigned a different k value k i in one MSA module.
Figure 8. This is the Singe-Scope Module (SSM) which is the main module in our MSA module. Each SSM module is assigned a different k value k i in one MSA module.
Cells 11 04107 g008
Figure 9. Style-based module with AdaIN (Adaptive Instance Normalization [26]).
Figure 9. Style-based module with AdaIN (Adaptive Instance Normalization [26]).
Cells 11 04107 g009
Figure 10. Visual comparison of point cloud completion on the IntrACompletion dataset.
Figure 10. Visual comparison of point cloud completion on the IntrACompletion dataset.
Cells 11 04107 g010
Figure 11. Visual comparison of point cloud completion on the PCN dataset.
Figure 11. Visual comparison of point cloud completion on the PCN dataset.
Cells 11 04107 g011
Table 1. Point completion results on IntrACompletion in F-score @ 1 % (higher is better).
Table 1. Point completion results on IntrACompletion in F-score @ 1 % (higher is better).
MethodsAneurysmVesselOverall
AtlasNet [12]0.1680.1960.193
FoldingNet [11]0.1320.1610.158
PCN [2]0.1400.1540.153
MSN [24]0.2780.3180.313
GRNet [33]0.2240.2470.245
SpareNet [15]0.3220.3760.370
Ours0.3430.3840.379
Table 2. Point completion results on IntrACompletion in CD × 10 3 (lower is better).
Table 2. Point completion results on IntrACompletion in CD × 10 3 (lower is better).
MethodsAneurysmVesselOverall
AtlasNet [12]2.1562.0492.061
FoldingNet [11]5.1243.7813.929
PCN [2]3.9723.9713.971
MSN [24]1.6231.4931.508
GRNet [33]2.0682.0062.013
SpareNet [15]1.4531.1401.174
Ours1.1730.9760.998
Table 3. Point completion results on IntrACompletion in EMD × 10 3 (lower is better).
Table 3. Point completion results on IntrACompletion in EMD × 10 3 (lower is better).
MethodsAneurysmVesselOverall
AtlasNet [12]4.1583.9834.002
FoldingNet [11]5.4014.7244.799
PCN [2]5.4665.3815.391
MSN [24]3.5243.3233.345
GRNet [33]4.1644.0484.061
SpareNet [15]3.3792.8852.940
Ours3.0002.7192.750
Table 4. Point completion results on PCN in F-score @ 1 % (higher is better).
Table 4. Point completion results on PCN in F-score @ 1 % (higher is better).
MethodsAirplaneCabinetCarChairLampSofaTableVesselOverall
AtlasNet [12]0.5900.0960.1600.1490.2110.1060.1990.2630.222
FoldingNet [11]0.5090.1340.1590.1610.2470.1170.2290.2790.229
PCN [2]0.6110.1480.2000.1850.2930.1280.2400.3120.265
MSN [24]0.6700.1170.2020.2290.3800.1480.2760.3570.297
GRNet [33]0.5540.1330.1760.1970.3650.1320.2330.3150.263
SpareNet [15]0.6640.1620.2040.2400.4360.1530.2750.3780.314
Ours0.6610.1640.2060.2490.4390.1670.2970.3800.320
Table 5. Point completion results on PCN in EMD × 10 3 (lower is better).
Table 5. Point completion results on PCN in EMD × 10 3 (lower is better).
MethodsAirplaneCabinetCarChairLampSofaTableVesselOverall
AtlasNet [12]1.9263.8093.0903.5813.9204.0013.2442.8793.307
FoldingNet [11]2.0903.4583.2293.7113.6093.9403.1882.9133.267
PCN [2]1.8673.2612.7203.2753.2683.5522.9352.6642.943
MSN [24]1.7043.6682.8443.0492.9683.6002.7272.4862.881
GRNet [33]2.0133.4742.9543.1882.8363.7052.9772.5832.966
SpareNet [15]1.7623.4422.9133.2312.6913.6592.9372.5292.895
Ours1.7783.3992.8893.1762.6973.6002.8542.4722.858
Table 6. Point completion results on PCN in CD × 10 3 (lower is better).
Table 6. Point completion results on PCN in CD × 10 3 (lower is better).
MethodsAirplaneCabinetCarChairLampSofaTableVesselOverall
AtlasNet [12]0.4311.8360.9771.4701.9522.0141.4050.9481.379
FoldingNet [11]0.5821.7461.2081.9462.2032.1511.6891.1821.589
PCN [2]0.4081.4200.7751.2911.2341.5271.2230.8161.087
MSN [24]0.3151.6910.8281.1531.0941.6070.9590.6841.041
GRNet [33]0.4411.3600.8741.1890.9941.8261.1010.6911.059
SpareNet [15]0.3201.3740.8801.1920.8911.5010.9730.6490.972
Ours0.3311.3740.8741.1980.9451.4850.9510.6330.974
Table 7. Ablation study on partial input in decoder and multi-scope in the encoder using IntrACompletion dataset.
Table 7. Ablation study on partial input in decoder and multi-scope in the encoder using IntrACompletion dataset.
Multi-ScopePartial InputEMD × 10 3 (Lower Is Better)CD × 10 3 (Lower Is Better)F-Score @ 1 % (Higher Is Better)
3.0641.2670.354
2.9731.1810.351
2.9311.1180.358
2.7500.9980.379
Table 8. Ablation study on partial input in decoder and multi-scope in the encoder using PCN dataset.
Table 8. Ablation study on partial input in decoder and multi-scope in the encoder using PCN dataset.
Multi-ScopePartial InputEMD × 10 3 (Lower Is Better)CD × 10 3 (Lower Is Better)F-Score @ 1 % (Higher Is Better)
2.8900.9550.313
2.8780.9770.312
2.8621.0110.314
2.8580.9740.320
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ma, W.; Yang, X.; Wang, Q.; Huang, K.; Huang, X. Multi-Scope Feature Extraction for Intracranial Aneurysm 3D Point Cloud Completion. Cells 2022, 11, 4107. https://doi.org/10.3390/cells11244107

AMA Style

Ma W, Yang X, Wang Q, Huang K, Huang X. Multi-Scope Feature Extraction for Intracranial Aneurysm 3D Point Cloud Completion. Cells. 2022; 11(24):4107. https://doi.org/10.3390/cells11244107

Chicago/Turabian Style

Ma, Wuwei, Xi Yang, Qiufeng Wang, Kaizhu Huang, and Xiaowei Huang. 2022. "Multi-Scope Feature Extraction for Intracranial Aneurysm 3D Point Cloud Completion" Cells 11, no. 24: 4107. https://doi.org/10.3390/cells11244107

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop