Next Article in Journal
Locking Multi-Laser Frequencies to a Precision Wavelength Meter: Application to Cold Atoms
Previous Article in Journal
Performability Evaluation of Load Balancing and Fail-over Strategies for Medical Information Systems with Edge/Fog Computing Using Stochastic Reward Nets
Previous Article in Special Issue
Effect of Spectral Signal-to-Noise Ratio on Resolution Enhancement at Surface Plasmon Resonance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Unsupervised Learning-Based Multi-Organ Registration Method for 3D Abdominal CT Images

1
School of Automation, Central South University, Changsha 410083, China
2
Hunan Xiangjiang Artificial Intelligence Academy, Changsha 410083, China
3
Hunan Engineering Research Center of High Strength Fastener Intelligent Manufacturing, Changde 415701, China
4
School of Computer Science and Engineering, Hunan University of Science and Technology, Xiangtan 411201, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(18), 6254; https://doi.org/10.3390/s21186254
Submission received: 24 July 2021 / Revised: 22 August 2021 / Accepted: 26 August 2021 / Published: 18 September 2021
(This article belongs to the Special Issue Advanced Optoelectronic Sensors and Biomedical Application)

Abstract

:
Medical image registration is an essential technique to achieve spatial consistency geometric positions of different medical images obtained from single- or multi-sensor, such as computed tomography (CT), magnetic resonance (MR), and ultrasound (US) images. In this paper, an improved unsupervised learning-based framework is proposed for multi-organ registration on 3D abdominal CT images. First, the explored coarse-to-fine recursive cascaded network (RCN) modules are embedded into a basic U-net framework to achieve more accurate multi-organ registration results from 3D abdominal CT images. Then, a topology-preserving loss is added in the total loss function to avoid a distortion of the predicted transformation field. Four public databases are selected to validate the registration performances of the proposed method. The experimental results show that the proposed method is superior to some existing traditional and deep learning-based methods and is promising to meet the real-time and high-precision clinical registration requirements of 3D abdominal CT images.

1. Introduction

Sensors and sensing systems play important roles in various medical applications, including disease diagnosis, monitoring, preoperative planning, surgical navigation, and so on [1,2,3,4]. Registration is one of the fundamental technologies that enable the sensing systems to be used in the above-mentioned applications [5,6,7,8]. Abdominal images taken from inter individuals are different in shape and texture, due to the complex intensity distribution of multi-organ, susceptibility of respiratory movement, and so on. Most existing methods cannot simultaneously meet the clinical requirements of high-accuracy and real-time performance for full abdominal image registration. To solve the above problems, many researchers pay attention to the segmented-based abdominal image registration methods. For example, Li et al. [9] proposed a liver MR image registration method based on the respiratory sliding motion segmentation that achieves more accurate registration results. Xie et al. [10] proposed a lung and liver 4D-CT image registration method based on tissue features and ROI segmentation, which can be implemented on clinical data.
In clinical practice, experts usually delineate the regions of interest (ROIs) of the target organ and organs at risk for treatment planning, and thereby the multi-organ ROIs of the abdominal medical images can be naturally obtained. For the registration stage, the traditional registration methods usually obtain the optimized deformation field by iteratively minimizing the custom energy function including data and regularization terms [11,12,13,14], and the deformation field of each pair of fixed and moving images is calculated independently by the traditional methods. The registration time of the traditional methods is always substantial, especially when the pair-wise images have a large anatomical difference. For example, the registration time of the methods including Demons [15], Elastix [16], and free-form deformation with b-splines [17] on medical images are ranged from minutes to hours.
To address this issue, many researchers have begun to pay more attention to the learning-based registration methods implemented by convolutional neural networks (CNNs), which can be divided into supervised and unsupervised [18,19,20,21]. These learning-based methods use the CNN model to obtain good initialization parameters for medical image registration. Most supervised methods require ground truth of deformation fields or anatomical segmentation masks, which are obtained from the traditional registration tools or manual delineation. These approaches entail much effort on data labeling, and the registration performance is influenced by the quality of the labels.
The unsupervised methods can directly use the unlabeled data to train CNNs model, avoiding the expensive and time-consuming labeling work. For example, Lei et al. [22] presented a multi-scale unsupervised registration framework for abdominal 4D-CT images. It has three loss functions to train the global and local subnetworks, including the similarity, adversarial, and regularization losses. Heinrich et al. [23] used a discrete displacement layer to improve the accuracy of the unsupervised learning-based 3D abdominal CT image registration framework. Balakrishnan et al. [24] proposed a U-net unsupervised registration framework, namely Voxelmorph, for 3D brain MR images. The regular similarity and regularization loss functions are used to train the framework, and then an auxiliary data loss function is added in the testing stage. Zhao et al. [25] proposed a U-net unsupervised registration framework for liver CT and brain MR images, namely VTN. Specifically, the affine transformation is integrated into the framework as a subnetwork to reduce the pre-processing time. Subsequently, the recursive cascaded networks (RCNs) by Zhao et al. [26] can be embedded into any base network as general architecture. Both Kuang et al. [27] and Mok et al. [28] emphasized that the distortion of the transformation field is non-negligible. They integrated the topology-preserving loss into the total loss functions for preventing the distortion.
Aiming at avoiding the time-consuming pre-processing and maintaining the topology-preserving property of the transformation, we developed an unsupervised learning-based registration framework for the segmented multi-organ from 3D abdominal CT images. First, the recursive cascaded network (RCN) modules are embedded into a basic U-net framework for promoting the unsupervised end-to-end learning. Secondly, the affine transformation subnetwork is cascaded with the subsequent fine registration subnetworks to implement the integration of the transformation field prediction from coarse to fine. Then, a topology-preserving loss is added in the total loss function for the training the registration framework, and then the transformation field is obtained for abdominal multi-organ registration.
The main contributions of this paper are as follows. First, an unsupervised learning-based registration framework is proposed, which can automatically learn from unlabeled data avoiding the time-consuming expert labeling work. Secondly, the coarse-to-fine RCN modules are embedded into the framework, which can efficiently deal with the large-scale deformation and improve the accuracy of the transformation field prediction. Moreover, an additional loss is integrated into the total loss function, which can ensure the registration with the property of the topology preservation. Finally, the proposed method proved to be more precise and faster than some existing registration methods on multi-organ from 3D abdominal CT images.

2. Methods

The essential of image registration is to find the mapping relationship between the fixed image I f and the moving image I m , which enables the I m align to the I f with a reasonable transformation T . Generally, the energy function of optimizing T is as follows [29]:
T ^ = arg min T L S ( I f , I m T ) + λ L R ( T )
where L S denotes the similarity term, L R represents the regularization term to constraint the L S , and λ is the empirical constant. The uniform image domain of I m , I f , and T is Ω in D-dimension, where the value of D is 3.

2.1. Optimization Problem Formulation

In this paper, the optimal parameters for transformation T are estimated by an improved unsupervised learning-based model. The task of the model is to use a flow prediction function F ( I m , I f ) to obtain the transformation field T ( I m , I f ) during a recursive procedure. As shown in Figure 1, there are n modules cascaded in the model. Each module contains a predicted transformation field T in a subnetwork. Therefore, the final warped I m can be written as [25]:
W a r p e d m ( n ) = I m F ( I m , I f )
and the output of the model is composited by:
F ( I m , I f ) = T 1 T n
where T 1 , …, T n represent the transformation field of the modules from 1st to nth, respectively. For instance, the k th predicted transformation field T k can be represented as T k = f k ( W a r p e d m ( k 1 ) , I f ) , where f k is the k th prediction function of F . Therefore, the moving image is gradually warped by the modules.

2.2. Architecture of the Unsupervised Learning-Based Networks

First, we embed the coarse-to-fine modules of RCNs into the basic U-net framework, intending to improve its registration performance on multi-organ from 3D abdominal CT images. The RCN modules can be used to successively predict their corresponding flow field f , as shown in Figure 1. Then, we integrate a topology-preserving loss into the total loss function to avoid the distortion of the predicted transformation T of the registration framework.

2.2.1. Coarse Registration

Affine transformation is widely applied to coarsely register the pair-wise medical images as pre-processing because it can reduce the registration errors caused by the difficulty of predicting the large deformation between the two images. Researchers commonly use the conventional software to perform the above process in a traditional way. However, it is time-consuming and requires researchers’ manual operations. To solve the above problems, the framework of the proposed method assigns the first subnetwork (namely coarse-subnetwork), to predict the affine transformation field with a small computational burden. The coarse-subnetwork contains a series of downsampling operations followed by a fully connected layer. The architecture of the network is the same as that of [25]. First, the input images are sequentially resampled to 643, 323, 163, 83, and 43 by the convolution layers with the uniform kernel size as 33 and stride as 2. Then, the flow transformation field is predicted according to the output parameters of the fully connected layer. The flow transformation field is represented as [25]:
T ( x ) = A x + b
where x denotes the voxel in the image domain Ω , A is a transform matrix, and b is a displacement vector. The moving image I m is first warped by the predicted affine transformation parameters, and then becomes the initial input for the fine registration subnetworks (namely fine-subnetworks).

2.2.2. Fine Registration

For fine registration, each subnetwork has the same encode–decoder architecture as U-net [30]. Both of them contain five resolutions to obtain multiple receptive fields, including 643, 323, 163, 83, and 43. The skip connection is used to concatenate the features of the same resolution from the encoding to the decoding stages, enabling the subnetwork to obtain more accurate predictions for the transformation field. For a pair of images, the transformation prediction is implemented by first extracting features in the encoding stage, and then restoring the image resolution in the decoding stage. The uniform kernel size and stride of the subnetworks are 33 and 2, respectively. All the subnetworks are used to progressively predict the flow transformation field for fine registration except the coarse-subnetwork.

2.3. Loss Functions

The unsupervised learning-based networks are trained by minimizing the following loss function:
L Total = L Coarse + L Fine
L Coarse = L S + λ 1 L R 1 + λ 2 L R 2
L Fine = L S + λ 3 L R 3 + λ 4 L R 4
where L Total denotes the total loss, L Coarse and L Fine represent the subtotal losses for the coarse- and the fine-subnetworks, respectively. L S is the similarity loss of both L Coarse and L Fine . L R 1 to L R 4 are the regularization terms of their corresponding subtotal loss functions. λ 1 to λ 4 are the empirical constants.

2.3.1. Similarity Loss

For mono-modal medical images, the correlation coefficient (CC) is suitable to be the similarity loss L S both for L Coarse and L Fine , which is defined as:
L S = 1 C C ( I f , I m T )
C C [ I f , I m T ] = σ [ I f , I m T ] σ [ I f , I f ] σ [ I m T , I m T ]
where σ [ ] denotes the covariance. The values of C C ranges from −1 to 1, indicating that the degree of correlation between two images changes from completely anti-correlated to correlated.

2.3.2. Regularization Terms for the Coarse-Subnetwork

Dealing with the large-scale deformation, we first employ the coarse-subnetwork to implement the affine transformation of the moving image I m . To avoid over non-rigid transformation field prediction, the orthogonality loss L R 1 and determinant loss L R 2 are used to regularize the similarity loss L S in the coarse-subnetwork, as shown in Equation (6). The L R 1 and L R 2 are respectively defined as [25,26]:
L R 1 = 6 + i = 1 3 μ i 2 + μ i 2
L R 2 = ( 1 + det ( A + I ) ) 2
where μ i is the singular values obtained from A + I , i = 1 , 2 , 3 . I is the identity matrix, and A is the affine transform matrix.

2.3.3. Regularization Terms for the Fine-Subnetworks

Generally, researchers only use a regular regularization term to penalize the similarity measure, such as L1 norm, L2 norm, and total variation. The smoothness property of the transformation field can be maintained by the above-mentioned term. First, we combine the similarity loss L S and the smooth loss L R 3 as a conventional group for transformation field prediction. However, one of the desirable properties of the transformation field is always ignored, namely topology-preserving. The relevant regularization term imposing on the similarity measure can prevent the distortion of the transformation field. Therefore, we integrate the topology-preserving loss L R 4 into the conventional group, aiming to obtain a more physically possible and accurate transformation field, as shown in Equation (7). The L2 variation loss L R 3 is defined as [31]:
L R 3 = 1 3 N x i = 1 3 ( T ( x + ρ i ) T ( x ) ) 2
where N is the total number of voxel x , ρ i forms the basis of , i = 1 , 2 , 3 .
Since the Jacobian determinant or its variations are commonly used to evaluate the topology preservation of the dense vector field, they can be used as a regularization loss for avoiding the distortion of the transformation field. Therefore, we use the negative Jacobian determinants as the topology-preserving loss L R 4 to further constrain the similarity loss of the fine-subnetworks. The L R 4 is defined as [27,28]:
L R 4 = 1 N x Ω σ ( J T ( x ) J T ( x ) )
where N is the total number of voxel x , J T ( x ) denotes the Jacobian determinant; σ ( ) is defined as m a x ( 0 , ) that can linearly activate the positive values and change the others from negative to zero. Therefore, σ ( ) decides whether L R 4 will penalize x . If the orientation of x is inconsistent with those of the neighbors, L R 4 can be activated (and vice versa).

3. Results and Discussion

3.1. Database

We selected 30, 22, 19, and 5 abdominal CT volumes from the publicly available training datasets from BTCV [32], LiTS [33], Sliver07 [34], and 3Dircadb [35], respectively, to validate the proposed method. The original size of the volumes is 512 × 512 × d e p t h .
BTCV is a dataset for the performance comparisons of 3D abdominal CT image segmentation methods. Its training dataset contains 30 objects with the liver, left kidney, right kidney, and spleen segmentation masks. All of them were selected for this study.
LiTS is a challenging dataset for liver tumor segmentation. A total of 130 training objects are provided with the liver and tumor segmentation masks. In this study, our experts randomly selected 22 objects and manually supplemented the corresponding left kidney, right kidney, and spleen masks.
Sliver07 is a challenging dataset for liver segmentation. Its training dataset includes 20 objects with liver segmentation masks. First, one of the objects was excluded considering the unclear structure of the spleen. Then, the remaining 19 objects were selected for this study, and our experts manually supplemented their left kidney, right kidney, and spleen masks.
3Dircadb is a 3D abdominal CT images dataset. It contains 20 training objects with multiple structures’ segmentation masks, including liver, left kidney, liver tumor, and so on. In this study, six objects were selected that simultaneously include liver, left kidney, right kidney, and spleen segmentation masks.
Therefore, 76 abdominal multi-organ CT volumes were formed by combining the above selected original volumes with their masks. One of them was used as the atlas (i.e., the fixed image), which is randomly selected from LiTS. A total of 15 volumes were evenly selected from LiTS, BTCV, and Sliver07, and paired with the atlas as the testing groups. The remsining 60 volumes were paired with the atlas as the training groups. The training and testing data were unified to 1283 due to the limited memory usage of the GPU (NVIDIA RTX 2080 Ti, 11G).

3.2. Evaluation Indexes

For registration analysis, the global intensity differences between two images are first evaluated by the root mean squared error (RMSE) and the peak signal-to-noise ratio (PSNR) [36]:
R M S E ( I f , I m T ) = M S E
P S N R = 10 × log 10 ( M A X I 2 M S E )
where N is the total number of voxel x . M A X I is the maximum possible voxel value of the image, M S E = 1 N x Ω ( I f ( x ) I m T ( x ) ) 2 .
Secondly, the local intensity differences between two images are evaluated by the structural similarity index (SSIM) [36]:
S S I M ( I f , I m T ) = ( 2 μ I f μ I m T + c 1 ) ( 2 σ I f I m T + c 2 ) ( μ I f 2 + μ I m T 2 + c 1 ) ( σ I f 2 + σ I m T 2 + c 2 )
where μ I f and μ I m T are the mean values, σ I f 2 and σ I m T 2 are the variances, and σ I f I m T is the covariance for I f and I m T ; c i = ( k i L ) 2 denotes the constant, which i = 1 , 2 , k 1 = 0.01 , k 2 = 0.03 , and L is the range of voxel value.
Moreover, the dice coefficient (DICE), Haustorff distance (HD), contour mean distance (CMD), intersection over union (IOU), sensitivity (SS), and specificity (SC) are also included [37]:
D I C E = X Y X + Y
H D = max h ( X , Y ) , h ( Y , X )
C M D = max h ( X b , Y b ) , h ( Y b , X b )
I O U = T P T P + F N + F P
S S = T P T P + F N
S C = T P T N + F P
where X and Y are the multi-organ segmentation masks of the fixed and warped images, respectively; X b and Y b are the boundaries for X and Y , respectively T P , T N , F N , and F P is the true positive, true negative, false negative, and false positive voxels in the multi-organ segmentation masks of the fixed and warped images, respectively. h ( X , Y ) = max x X min y Y x y is the distance from X to Y , and h ( X b , Y b ) = max x X b min y Y b x y the distance from X b to Y b .

3.3. Experimental Analysis

3.3.1. Internal Comparisons

We use the basic U-net (namely Base-Net) integrated with our total loss function to explore the optimal number of the RCN modules. First, the coarse-subnetwork-related module is integrated into the Base-Net without counting. Then, the Base-Net is cascaded with 1, 2, 3, 5, and 7 RCN modules, namely Base-Net-1, Base-Net-3, Base-Net-5, and Base-Net-7, respectively, for experimental analysis. The maximum number of the RCN modules is determined by the memory usage of the GPU. Subsequently, we chose the best performing model as the proposed method for external comparisons, i.e., Base-Net-7. The values of λ 1 and λ 2 are the same as those in [25], and the value of λ 4 is the same as that in [27], i.e., λ 1 = 10 1 , λ 2 = 10 1 , and λ 4 = 10 5 . The optimal value of λ 3 is selected according to the experimental results. The uniform batch size, epoch, and learning rate of these comparison models are set as 2, 5, and 104, respectively.
Figure 2 shows the total loss trends of different models on the training dataset. The curves of the Baseline and Baseline-RCNs-Topology are obtained from the Base-Net and the Base-Net-7, respectively. The curve of the Baseline-RCNs is acquired from the same model as the Base-Net-7 without the topology-preserving loss. Compared with the Baseline, the Baseline-RCNs has lower total loss values, indicating that the RCN modules are effective in decreasing the registration errors; meanwhile, the Baseline-RCNs-Topology proves that the topology-preserving loss function modules can further improve the multi-organ registration performance on 3D abdominal CT images. Furthermore, Figure 3 displays the total loss trends of the Base-Net-7 with different values of λ 3 . It can be seen that the Base-Net-7 with λ 3 = 10 5 achieves the optimal performance on the training dataset. Therefore, the value of λ 3 is set as 10 5 in the proposed method.
Table 1 presents the internal comparisons of embedding different number of the RCNs modules into our Base-Net. As observed, the average values of the RMSE, HD, and CMD obtained by the Base-Net-1, 3, and 7 are smaller than those of the Base-Net. Meanwhile, the values of the PSNR, SSIM, DICE, IOU, SS, and SC obtained by the Base-Net-1, 3, 5, and 7 are higher than those of the Base-Net. It may be because the architecture of the Base-Net is changed to a recursive one by embedding the RCN modules, and the improved Base-Net can progressively predict a more accurate transformation field. Besides, the RMSE, HD, and CMD from Base-Net-7 are 54.01%, 31.45%, and 82.01% lower than those of the Base-net, respectively. The PSNR, SSIM, DICE, IOU, SS, and SC from Base-Net-7 are 38.16%, 7.92%, 33.81%, 65.66%, 17.52%, and 48.29% higher than those of the Base-Net, respectively. Moreover, the Base-Net-7 outperforms all of the other models on the multi-organ registration from 3D abdominal CT images. It can be therefore concluded that the optimal number is seven for embedding the RCNs modules into the Base-Net.
Figure 4 displays the registration progress of the proposed method on the 41st, 57th, 65th, 73rd, and 81st slices of the randomly selected testing group, where T 1 is the affine transformation field, and T 2 to T 8 are the transformation field obtained by 1st to 7th RCN modules, respectively. It can be observed that there exist big intensity differences between the fixed image I f and the moving image I m . However, the RCN modules can help the model progressively obtain the final warped images, which are exactly similar to the fixed images. Therefore, the proposed method has good performance on multi-organ registrations from 3D abdominal CT images.

3.3.2. External Comparisons

We compared our method against three traditional methods, Demons [38], Hybrid [39], MSI [40], and two state-of-art unsupervised learning-based methods, VTN [25] and Voxelmorph [24]. We ran the experimental methods on an Intel i5-GTX1060 CPU and an NVIDIA RTX 2080 Ti GPU, respectively.
Table 2 shows the external comparisons of the abdominal multi-organ registration results on different methods. As observed, the proposed method performs comparably to MSI in terms of RMSE, and is superior to Demons, Hybrid, VTN, and Voxelmorph. Moreover, the average values of PSNR, SSIM, DICE, IOU, SS, SC, HD, and CMD from the proposed method are 29.5584, 0.9732, 0.9775, 0.9562, 0.9982, 0.9578, 12.2646, and 3.9296, respectively, which are better than those of the traditional and unsupervised learning-based methods. The registration time in Table 2 illustrates that the unsupervised learning-based methods are obviously faster than the traditional ones. Although VTN achieves the shortest registration time, its performances on the other evaluation indicators are barely satisfactory. Generally, the proposed method gives a good compromise to meet the real-time and high-accuracy clinical requirements.
Figure 5 and Figure 6 directly display the histograms and boxplots of the evaluation metrics of 15 pair-wise testing groups’ registration results with different methods, respectively. It can be found that the distributions of the RMSE, PSNR, DICE, CMD, and IOU values in Figure 5 and Figure 6 are consistent with those in Table 2. Hence, the proposed method has stable registration performance, and is superior to the other competing methods in the above evaluation indicators.
Figure 7 presents the intensity differences between the fixed and moving images from four randomly selected testing groups. The intensity values of the image range from 0 to 255 and its corresponding color varies from blue to red. As shown in Figure 7, all pair-wise groups have significant differences before registration. Then, the Demons, VTN, and Voxelmorph methods produce considerable differences, indicating that their performances may be influenced by the noise and artifacts in the abdominal CT images. Moreover, the proposed method produces the smallest differences and outperforms the other competing methods. It can be concluded that our method can avoid the impact of the noise and artifacts and perform more accurately and robustly on multi-organ registration for 3D abdominal CT images.

4. Conclusions

In this paper, we present an improved unsupervised learning-based framework for multi-organ registration from 3D abdominal CT images. The coarse-to-fine RCNs modules are embedded into a basic U-net model, which can hence inherit the advantages of the recursive model and achieve better performance on multi-organ registration from 3D abdominal CT images. In addition, a topology-preserving loss is added in the total loss function, which can penalize the similarity loss to avoid the distortion of the predicted transformation field. The experimental results show that the proposed method has the optimal PSNR, SSIM, DICE, IOU, SS, SC, HD, and CMD average values and is competitive with the traditional and unsupervised learning-based methods, and therefore has the potential to be used in clinical practice.

Author Contributions

Writing-original draft preparation, conceptualization, investigation, methodology, and software, S.Y.; methodology, supervision, writing-reviewing and editing, project administration, resources, and funding acquisition, Y.Z.; validation and supervision, M.L.; visualization and validation, F.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the National Natural Science Foundation of China (Grant nos. 62076256, 61772555, and 52005520), and 111 Project (Grant No. B18059).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bielecki, Z.; Stacewicz, T.; Wojtas, J.; Mikołajczyk, J.; Szabra, D.; Prokopiuk, A. Selected optoelectronic sensors in medical applications. Opto-Electron. Rev. 2018, 26, 122–133. [Google Scholar] [CrossRef]
  2. David, D.D.S.; Parthiban, R.; Jayakumar, D.; Usharani, S.; RaghuRaman, D.; Saravanan, D.; Palani, U. Medical Wireless Sensor Network Coverage and Clinical Application of Mri Liver Disease Diagnosis. Eur. J. Mol. Clin. Med. 2021, 7, 2559–2571. [Google Scholar]
  3. Gao, L.; Zhang, G.; Yu, B.; Qiao, Z.; Wang, J. Wearable human motion posture capture and medical health monitoring based on wireless sensor networks. Measurement 2020, 166, 108252. [Google Scholar] [CrossRef]
  4. Luo, X.; He, X.; Shi, C.; Zeng, H.-Q.; Ewurum, H.C.; Wan, Y.; Guo, Y.; Pagnha, S.; Zhang, X.-B.; Du, Y.-P. Evolutionarily Optimized Electromagnetic Sensor Measurements for Robust Surgical Navigation. IEEE Sens. J. 2019, 19, 10859–10868. [Google Scholar] [CrossRef]
  5. Kok, E.N.D.; Eppenga, R.; Kuhlmann, K.F.D.; Groen, H.C.; Van Veen, R.; Van Dieren, J.M.; De Wijkerslooth, T.R.; Van Leerdam, M.; Lambregts, D.M.J.; Heerink, W.J.; et al. Accurate surgical navigation with real-time tumor tracking in cancer surgery. NPJ Precis. Oncol. 2020, 4, 1–7. [Google Scholar] [CrossRef] [PubMed]
  6. Ahn, S.J.; Lee, J.M.; Lee, D.H.; Lee, S.M.; Yoon, J.-H.; Kim, Y.J.; Yu, S.J.; Han, J.K. Real-time US-CT/MR fusion imaging for percutaneous radiofrequency ablation of hepatocellular carcinoma. J. Hepatol. 2017, 66, 347–354. [Google Scholar] [CrossRef]
  7. Li, K.; Su, Z.; Xu, E.; Huang, Q.; Zeng, Q.; Zheng, R. Evaluation of the ablation margin of hepatocellular carcinoma using CEUS-CT/MR image fusion in a phantom model and in patients. BMC Cancer 2017, 17, 1–10. [Google Scholar] [CrossRef] [Green Version]
  8. Radu, C.; Fisher, P.; Mitrea, D.; Birlescu, I.; Marita, T.; Vancea, F.; Florian, V.; Tefas, C.; Badea, R.; Ștefănescu, H.; et al. Integration of Real-Time Image Fusion in the Robotic-Assisted Treatment of Hepatocellular Carcinoma. Biology 2020, 9, 397. [Google Scholar] [CrossRef]
  9. Li, D.; Zhong, W.; Deh, K.M.; Nguyen, T.D.; Prince, M.R.; Wang, Y.; Spincemaille, P. Discontinuity Preserving Liver MR Registration with Three-Dimensional Active Contour Motion Segmentation. IEEE Trans. Biomed. Eng. 2018, 66, 1884–1897. [Google Scholar] [CrossRef]
  10. Xie, Y.; Chao, M.; Xing, L. Tissue Feature-Based and Segmented Deformable Image Registration for Improved Modeling of Shear Movement of Lungs. Int. J. Radiat. Oncol. Biol. Phys. 2009, 74, 1256–1265. [Google Scholar] [CrossRef] [Green Version]
  11. Fu, Y.; Lei, Y.; Wang, T.; Curran, W.J.; Liu, T.; Yang, X. Deep learning in medical image registration: A review. Phys. Med. Biol. 2020, 65, 20TR01. [Google Scholar] [CrossRef] [Green Version]
  12. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.; van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [Green Version]
  13. Nazib, A.; Fookes, C.; Perrin, D. A comparative analysis of registration tools: Traditional vs deep learning approach on high resolution tissue cleared data. arXiv 2018, arXiv:1810.08315. preprint. [Google Scholar]
  14. Villena-Martinez, V.; Oprea, S.; Saval-Calvo, M.; Azorin-Lopez, J.; Fuster-Guillo, A.; Fisher, R.B. When Deep Learning Meets Data Alignment: A Review on Deep Registration Networks (DRNs). Appl. Sci. 2020, 10, 7524. [Google Scholar] [CrossRef]
  15. Thirion, J.P. Image matching as diffusion process: An analogy with Maxwell’s demons. Med. Image Anal. 1998, 2, 243–260. [Google Scholar] [CrossRef] [Green Version]
  16. Klein, S.; Staring, M.; Murphy, K.; Viergever, M.A.; Pluim, J.P.W. elastix: A Toolbox for Intensity-Based Medical Image Registration. IEEE Trans. Med. Imaging 2009, 29, 196–205. [Google Scholar] [CrossRef] [PubMed]
  17. Modat, M.; Ridgway, G.; Taylor, Z.; Lehmann, M.; Barnes, J.; Hawkes, D.J.; Fox, N.; Ourselin, S. Fast free-form deformation using graphics processing units. Comput. Methods Programs Biomed. 2010, 98, 278–284. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Cao, X.; Yang, J.; Zhang, J.; Nie, D.; Kim, M.-J.; Wang, Q.; Shen, D. Deformable image registration based on similarity-steered CNN regression. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Quebec City, QC, Canada, 10–14 September 2017; Springer: Cham, Switzerland, 2017; pp. 300–308. [Google Scholar]
  19. Ferrante, E.; Oktay, O.; Glocker, B.; Milone, D.H. On the adaptability of unsupervised CNN-based deformable image registration to unseen image domains. In Proceedings of the International Workshop on Machine Learning in Medical Imaging, Granada, Spain, 16 September 2018; Springer: Cham, Switzerland, 2018; pp. 294–302. [Google Scholar]
  20. Blendowski, M.; Hansen, L.; Heinrich, M.P. Weakly-supervised learning of multi-modal features for regularised iterative descent in 3D image registration. Med. Image Anal. 2021, 67, 101822. [Google Scholar] [CrossRef]
  21. Xu, Z.; Niethammer, M. DeepAtlas: Joint semi-supervised learning of image registration and segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019. [Google Scholar]
  22. Lei, Y.; Fu, Y.; Wang, T.; Liu, Y.; Patel, P.; Curran, W.J.; Liu, T.; Yang, X. 4D-CT deformable image registration using multiscale unsupervised deep learning. Phys. Med. Biol. 2020, 65, 085003. [Google Scholar] [CrossRef]
  23. Heinrich, M.P.; Hansen, L. Highly Accurate and Memory Efficient Unsupervised Learning-Based Discrete CT Registration Using 2.5D Displacement Search. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru, 4–8 October 2020; Springer: Cham, Switzerland, 2020; pp. 190–200. [Google Scholar]
  24. Balakrishnan, G.; Zhao, A.; Sabuncu, M.R.; Guttag, J.; Dalca, A.V. VoxelMorph: A Learning Framework for Deformable Medical Image Registration. IEEE Trans. Med. Imaging 2019, 38, 1788–1800. [Google Scholar] [CrossRef] [Green Version]
  25. Zhao, S.; Lau, T.; Luo, J.; Chang, E.I.-C.; Xu, Y. Unsupervised 3D End-to-End Medical Image Registration with Volume Tweening Network. IEEE J. Biomed. Health Inform. 2020, 24, 1394–1404. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Zhao, S.; Dong, Y.; Chang, E.; Xu, Y. Recursive cascaded networks for unsupervised medical image registration. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 10600–10610. [Google Scholar]
  27. Kuang, D.; Schmah, T. Faim—A convnet method for unsupervised 3d medical image registration. In Proceedings of the International Workshop on Machine Learning in Medical Imaging, Shenzhen, China, 13 October 2019; Springer: Cham, Switzerland, 2019; pp. 646–654. [Google Scholar]
  28. Mok, T.C.W.; Chung, A. Fast symmetric diffeomorphic image registration with convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4644–4653. [Google Scholar]
  29. Ferrante, E.; Paragios, N. Slice-to-volume medical image registration: A survey. Med. Image Anal. 2017, 39, 101–123. [Google Scholar] [CrossRef] [Green Version]
  30. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  31. Rudin, L.I.; Osher, S.; Fatemi, E. Nonlinear total variation based noise removal algorithms. Phys. D Nonlinear Phenom. 1992, 60, 259–268. [Google Scholar] [CrossRef]
  32. Xu, Z.; Lee, C.P.; Heinrich, M.P.; Modat, M.; Rueckert, D.; Ourselin, S.; Abramson, R.G.; Landman, B.A. Evaluation of Six Registration Methods for the Human Abdomen on Clinically Acquired CT. IEEE Trans. Biomed. Eng. 2016, 63, 1563–1572. Available online: https://competitions.codalab.org/competitions/17094 (accessed on 1 November 2018). [CrossRef] [PubMed] [Green Version]
  33. Bilic, P.; Christ, P.F.; Vorontsov, E.; Chlebus, G.; Chen, H.; Dou, Q.; Fu, C.-W.; Han, X.; Heng, P.-A.; Hesser, J.; et al. The Liver Tumor Segmentation Benchmark (LiTS). arXiv 2019, arXiv:1901.04056. preprint. [Google Scholar]
  34. Heimann, T.; Ginneken, B.V.; Styner, M.A. Segmentation of the Liver 2007(SLIVER07). Available online: http://sliver07.isi.uu.nl/ (accessed on 12 December 2018).
  35. Soler, L.; Hosttettle, A.; Charnoz, A.; Fasquel, J.; Moreau, J. 3D Image Reconstruction for Comparison of Algorithm Database: A Patient Specific Anatomical and Medical Image Database. Available online: https://www.ircad.fr/research/3dircadb/ (accessed on 16 April 2018).
  36. Sara, U.; Akter, M.; Uddin, M.S. Image quality assessment through FSIM, SSIM, MSE and PSNR—A comparative study. J. Comput. Commun. 2019, 7, 8–18. [Google Scholar] [CrossRef] [Green Version]
  37. Pei, H.-Y.; Yang, D.; Liu, G.-R.; Lu, T. MPS-Net: Multi-Point Supervised Network for CT Image Segmentation of COVID-19. IEEE Access 2021, 9, 47144–47153. [Google Scholar] [CrossRef]
  38. Lombaert, H.; Grady, L.; Pennec, X.; Ayache, N.; Cheriet, F. Spectral Log-Demons: Diffeomorphic Image Registration with Very Large Deformations. Int. J. Comput. Vis. 2014, 107, 254–271. [Google Scholar] [CrossRef] [Green Version]
  39. Chan, C.L.; Anitescu, C.; Zhang, Y.; Rabczuk, T. Two and Three Dimensional Image Registration Based on B-Spline Composition and Level Sets. Commun. Comput. Phys. 2017, 21, 600–622. [Google Scholar] [CrossRef]
  40. Aganj, I.; Iglesias, J.E.; Reuter, M.; Sabuncu, M.R.; Fischl, B. Mid-space-independent deformable image registration. NeuroImage 2017, 152, 158–170. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Framework of the proposed method.
Figure 1. Framework of the proposed method.
Sensors 21 06254 g001
Figure 2. Total loss curves of different models.
Figure 2. Total loss curves of different models.
Sensors 21 06254 g002
Figure 3. Total loss curves of the Base-Net-7 with different values of λ 3 .
Figure 3. Total loss curves of the Base-Net-7 with different values of λ 3 .
Sensors 21 06254 g003
Figure 4. Visualization of a randomly selected testing group on the proposed method.
Figure 4. Visualization of a randomly selected testing group on the proposed method.
Sensors 21 06254 g004
Figure 5. Evaluation metrics of 15 pair-wise testing groups’ registration results with different methods. (a) RMSE; (b) PSNR; (c) DICE; (d) CMD; (e) IOU.
Figure 5. Evaluation metrics of 15 pair-wise testing groups’ registration results with different methods. (a) RMSE; (b) PSNR; (c) DICE; (d) CMD; (e) IOU.
Sensors 21 06254 g005
Figure 6. Boxplots for evaluation metrics of 15 pair-wise testing groups’ registration results with different methods. (a) RMSE; (b) PSNR; (c) DICE; (d) CMD; (e) IOU.
Figure 6. Boxplots for evaluation metrics of 15 pair-wise testing groups’ registration results with different methods. (a) RMSE; (b) PSNR; (c) DICE; (d) CMD; (e) IOU.
Sensors 21 06254 g006
Figure 7. The intensity differences between the fixed and moving images from the paired testing groups with different methods.
Figure 7. The intensity differences between the fixed and moving images from the paired testing groups with different methods.
Sensors 21 06254 g007
Table 1. The average metrics of abdominal multi-organ registration results with the different number of RCN modules.
Table 1. The average metrics of abdominal multi-organ registration results with the different number of RCN modules.
MetricBase-NetBase-Net-1Base-Net-3Base-Net-5Base-Net-7
RMSE0.08980.05160.04300.04190.0413
PSNR21.393826.863928.957429.325929.5584
SSIM0.90180.95720.97010.97200.9732
DICE0.73050.91550.96210.97240.9775
IOU0.57720.84510.92720.94450.9562
SS0.84940.97570.99440.99740.9982
SC0.64590.86310.93200.94680.9578
HD (mm)17.892415.009013.123112.624512.2646
CMD (mm)21.83837.07334.79724.33413.9296
Table 2. The average metrics of abdominal multi-organ registration results with different methods.
Table 2. The average metrics of abdominal multi-organ registration results with different methods.
MetricsDemonsHybridMSIVTNVoxelmorphProposed
RMSE0.09870.04740.03550.09280.09060.0413
PSNR20.696926.912129.341820.745621.321229.5584
SSIM0.89250.94120.95920.87690.90090.9732
DICE0.66780.58910.87150.70780.72530.9775
IOU0.51130.58330.83860.55690.57090.9562
SS0.80770.91890.92680.82560.84500.9982
SC0.58360.61370.89970.62680.64070.9578
HD (mm)27.784330.737825.712318.153617.731912.2646
CMD (mm)55.645680.764110.702220.879021.84323.9296
CPU (s)381.0294510.1189-7.56128.62479.8922
GPU (s)--536.43781.39981.52121.6371
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, S.; Zhao, Y.; Liao, M.; Zhang, F. An Unsupervised Learning-Based Multi-Organ Registration Method for 3D Abdominal CT Images. Sensors 2021, 21, 6254. https://doi.org/10.3390/s21186254

AMA Style

Yang S, Zhao Y, Liao M, Zhang F. An Unsupervised Learning-Based Multi-Organ Registration Method for 3D Abdominal CT Images. Sensors. 2021; 21(18):6254. https://doi.org/10.3390/s21186254

Chicago/Turabian Style

Yang, Shaodi, Yuqian Zhao, Miao Liao, and Fan Zhang. 2021. "An Unsupervised Learning-Based Multi-Organ Registration Method for 3D Abdominal CT Images" Sensors 21, no. 18: 6254. https://doi.org/10.3390/s21186254

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop