Next Article in Journal
Electrodynamics Sensor for the Image Reconstruction Process in an Electrical Charge Tomography System
Next Article in Special Issue
A 3D Sensor Based on a Profilometrical Approach
Previous Article in Journal
A Multi-Agent System Architecture for Sensor Networks
Previous Article in Special Issue
Sensors for 3D Imaging: Metric Evaluation and Calibration of a CCD/CMOS Time-of-Flight Camera
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Nonrigid Registration of Brain Tumor Resection MR Images Based on Joint Saliency Map and Keypoint Clustering

Department of Biomedical Engineering, School of Life Sciences & Biotechnology, Shanghai Jiao Tong University, Shanghai 200240, China
*
Author to whom correspondence should be addressed.
Sensors 2009, 9(12), 10270-10290; https://doi.org/10.3390/s91210270
Submission received: 23 October 2009 / Revised: 1 December 2009 / Accepted: 9 December 2009 / Published: 17 December 2009
(This article belongs to the Special Issue Image Sensors 2009)

Abstract

:
This paper proposes a novel global-to-local nonrigid brain MR image registration to compensate for the brain shift and the unmatchable outliers caused by the tumor resection. The mutual information between the corresponding salient structures, which are enhanced by the joint saliency map (JSM), is maximized to achieve a global rigid registration of the two images. Being detected and clustered at the paired contiguous matching areas in the globally registered images, the paired pools of DoG keypoints in combination with the JSM provide a useful cluster-to-cluster correspondence to guide the local control-point correspondence detection and the outlier keypoint rejection. Lastly, a quasi-inverse consistent deformation is smoothly approximated to locally register brain images through the mapping the clustered control points by compact support radial basis functions. The 2D implementation of the method can model the brain shift in brain tumor resection MR images, though the theory holds for the 3D case.

1. Introduction

Image registration is an important enabling technology for neuronavigation [1, 2] due to its mapping pre-operative images to the patient's anatomy in physical space and augmenting the intra-operative images with the pre-operative image. Image registration can be classified into intensity- and landmark/feature-based methods [3]. It can be considered as an optimization problem, posed as finding the optimal transformation T between the reference image IR and the floating image IF to maximize a defined similarity measure such as mutual information (MI) [4, 5]. The transformation space includes rigid and nonrigid that compensates for deformations. During the last decade, nonrigid registration of MR brain images has attracted much attention at the brain shift estimation [1, 2] in image-guided neurosurgery.
The key challenge for the nonrigid registration of pre- and intra-operative MR images is to compensate for the local large tissue distortion caused by the tumor resection. The local large tissue deformations with irregular shapes violate the usual assumption of smoothness of the deformation fields. An additional challenge exists when the unmatchable outlier features (i.e., a tumor in the preoperative image may not even exist in the intra-operative image) are introduced. Moreover, the local large deformation, sharp geometric difference between the pre- and intra-operative MR images and the confounding effects of edema and tumor infiltration, render the outlier problem more intractable.
To reject the outliers, many intensity-based registration approaches are proposed including M-estimator [6] or mixture-based similarity measure [7], graph-based multifeature MI [8], least-trimmed square based outlier rejection [1], consistency test [9, 10] and intensity modification [11]. However, the intensity similarity does not necessarily mean anatomical similarity and easily suffers from local and biased maxima [1215] when outliers are presented in images. Additionally, by forcibly matching non-corresponding structure features, the extra flexibility of the complex deformation in intensity-based methods may make the results unpredictable and less reliable. To somewhat alleviate these problematic aspects, modifications have been added in intensity-based methods to include higher level feature information such as landmark [16, 17]. Despite these modifications, the presence of local large deformation and the outliers still remains an unresolved problem for most intensity-based methods.
Recently, landmark-based registration methods using local invariant features, such as salient region features [1820], multi-scale Laplacian blob [21, 22], SIFT keypoint [2328], attribute vector [2931] and multiscale wavelet-based region [32], have been proven effective than intensity-based methods in compensating for the local deformations around small anatomical structures. The SIFT detector has been successfully applied to various applications such as face recognition [33] and medical image registration [2427]. In image registration, the SIFT keypoints [24, 27] localized at specific anatomical structures could be automatically selected for the adaptive setup of the irregular control point grids for the local deformation of specific anatomical structures. Without tedious manual selection of control points, the adaptive setup of irregular control point grids could alleviate the computational cost and the registration inaccuracy that are related to the regular grids [34] of control points arranged for the local large deformation at the tumor resection region. In general, it is often difficult to correctly match local keypoints [26, 28] by using only the similarity between SIFT descriptors when complex local deformation and outliers exist in brain images. The deformation invariant local feature descriptor was presented in [29], however this topic is beyond the scope of this paper.
Although the results of these above methods clearly demonstrate the power of local invariant feature-based nonrigid deformations, the desired landmark-based registration algorithm should establish robust control point correspondence to accurately model the complex local deformation around the tumor resection region. To find robust point correspondence, some approaches are proposed including soft correspondence detections [20, 35], joint clustering-matching strategy [36] and modeling point sets by kernel density function [37]. Compared with the classical template matching, the iterative closest point [38] and the correspondence by sensitivity to movement [39], the self-organizing map [40] algorithm was considered in [41] to be the most effective method in 2D feature point correspondence detection. However, these methods do not consider the complexity of correspondence detection in the context of local large structure distortion combined with the outliers. In this work, we first compute the global correspondence between the contiguous matching areas of normal tissues and the tumor regions in the two images by using MI-based rigid registration method. The global correspondence is then used to introduce the cluster correspondence between the paired pools of DoG keypoints [23] that are detected and clustered at those contiguous matching areas in the two images. An important contribution is that we have proposed the cluster-to-cluster correspondence can be introduced as a useful constraint for the local point correspondence detection within the paired pools of keypoints.
To make the rigid registration more accurate and robust for matching brain tumor resection images, we use joint saliency map (JSM), first proposed in our previous work [15], to highlight the corresponding salient structures in the two images, and emphatically group those salient structures into the smoothed compact clusters in the weighted joint histogram for MI based registration. Another important contribution is that we have demonstrated the JSM can be combined with the keypoint clustering for nonrigid landmark-based registration scheme, ensuring that the structural mismatches at the tumor regions could be detected and the underlying outlier DoG keypoint could be rejected automatically. Specifically, after implementing keypoint clustering, we identify the pair of tumor resection clusters in both images owing to its average JSM value being below a threshold value. We then remove the outlier keypoints inside the paired tumor areas, use the cluster-to-cluster correspondence to guide the point-to-point correspondence detection of the significant keypoints within the paired matching keypoint groups. Once the significant keypoint sets are selected and classified into a number of local adaptive setup of irregular control point grids, the nonrigid local registration is carried out within deterministic annealing framework with the implementation of quasi-inverse consistency [16, 42] radial basis functions (RBF) [16, 40, 43] as a warping method.
The proposed algorithm has been applied to the nonrigid registration of 2D brain tumor resection MR images. Experimental results show that, compared to other classic nonrigid registration methods, our proposed method can provide better robustness and higher accuracy for the registration of brain tumor resection images. The rest of this paper is organized as follows. Section 2 briefly summarizes the definition of JSM and its application in global rigid registration, which are the same as in [15]. Section 3 integrates the JSM into the keypoint clustering for robust correspondence detection with outlier rejection, and implements the irregular control point setup for RBF-based local nonrigid registration within a deterministic annealing framework. In Section 4, using clinical brain tumor resection images, we report preliminary experiment results to identify the registration performance on accuracy and robustness. The conclusion and future work are discussed in Section 5.

2. Global Registration Based on Joint Saliency Map

Since 1995 [4, 5], mutual information (MI) has proved effective in intensity-based image registration. For a reference image IR and a floating image IF with intensity bins r and f, the MI between IR and IF is defined as:
MI = H ( I R ) + H ( I F ) H ( I R , I F )
where H(I) = −Σip(i) log p(i) and H (IR, IF) = −Σr,fp (r, f) log p (r, f) are the entropy of the intensities of image I and the joint entropy of the two images, p(i) is the intensity probabilities with p (r) = Σfp (r, f) and p (f) = Σrp (r, f), p (r, f) is the joint intensity probabilities estimated by a joint histogram h (r, f).
MI-based registration methods take advantage of the fact that properly registered images usually correspond to compactly-clustered joint histograms [44]. They measure the joint histogram dispersion by computing the entropy of the joint intensity probabilities. When the images become misregistered, the compact clusters become disperse sets of points in the joint histogram. The JSM not only emphatically groups the corresponding salient pixel pairs into the compact histogram clusters for robust global rigid registration, but also determines the structural mismatches for subsequent local nonrigid registration. The idea of JSM is demonstrated schematically in Figure 1h, where the high fractional joint saliency values (between 0 and 1) are assigned to the corresponding salient pixel pairs rather than the structural mismatches (tumor areas) and the homogeneous pixel pairs.

2.1. Joint Saliency Map

Noting that many techniques have defined the saliency of image to enhance the image pixels we are interested in, i.e., using edge gradient, corner, salient regions [18] and keypoint [23]. Nevertheless, gradient is a local feature and sensitive to noise, corner and keypoint saliency are not to be defined for each pixel, and salient regions suffers from high complexity. In this study, we first use multi-scale saliency map [45] Sl(v) = ΣuNv (Il (v) − Il (u))2, where Nv is the 1-pixel radius circular neighborhood of the pixel position v = (x, y) at scale l, Sl(v) is the local saliency computed from the intensity Il(v) of the pixel position v in the Gaussian image pyramid [46] at scale l, Il(u) is the intensity of the pixel in the Il(v)'s neighborhood. The final saliency map S(x, y) is reconstructed by summing up all the saliency maps at the coarser scales.
In the second step, a principal axis analysis of regional saliency distribution assigns the regional saliency vector to each pixel based on the inertia matrix:
M = [ φ 20 φ 11 φ 11 φ 02 ]
where φmn = Σ(xgx)m(ygy)nS(x, y), (gx, gy) = (ω1000, ω0100), ωmn = ΣxmynS(x, y) are the central (m, n)- moment, the centroid and the (m, n)- moment of the regional saliency distribution S(x, y) in a 5.5-pixel radius circular neighborhood around the center pixel. This regional saliency distribution describes a 2D regional salient structure, within which the two eigenvectors of the M represent the orthogonal coordinate system. The first eigenvector referred as the RSV is enough to represent the regional salient structure around a pixel due to its storing the regional information about the orientation of the salient structure.
Given the two RSVs sr and sf at an overlapping location v, the JSM is constructed to indicate the corresponding salient structures in the two images by computing the inner product of two RSVs. The essential idea of JSM is always valid in practice in image registration: for two precisely aligned images, the majority of the corresponding pixel locations are very likely to produce RSVs with similar orientations. This is because the two images under registration fundamentally depict the same image structures. As a result, the RSVs of the corresponding pixel locations from two images could present relatively coincident orientations in general. Therefore, the angle γ between the two vectors is simply calculated, making cos γ the scalar measure of the JSM value w (v) = cos γ (sr, sf) = 〈sr, sf 〉/‖sr‖ · ‖sf‖.
A JSM value near one suggests that the underlying pixel pair comes from the corresponding spatial structures. Contrarily, a JSM value near zero indicates that the pixel pair originates from either the outliers or a homogeneous region. To speed up the rigid registration procedure without reducing accuracy, the registration only uses the salient pixel pairs with large saliency values. The pixel with a small saliency value below a fixed threshold value (10 percent of the maximum saliency value) is assigned zero JSM value directly. Generally, the JSM would primarily respond to the high-gradient common edges in both images if a high threshold is chosen to exclude more pixels from estimating the JSM. However, as in the Figure 1c–e representing the image gradient and the JSM profiles of the same line at the two registered images (see Figure 1a,b), the JSM (see Figure 1e) consistently captures the corresponding regional salient structures in the two images while the image gradient features (see Figure 1c,d) are very noisy and do not agree with each other at each overlapping location.

2.2. JSM-Weighted Mutual Information

The accurate rigid registration can be iteratively achieved by maximizing the MI between the corresponding salient structures that are highlighted by the JSM. The JSM is updated with the transformation T changing the overlapping area of the two images during the rigid registration. Specifically, the JSM weights each overlapping pixel pairs for the joint histogram so that the contribution of the interpolated floating intensity f (vf) to the histogram is weighted by the w (v) of the JSM. For example, if using nearest neighbor or linear interpolation, the value w(v) should be added to the histogram entry h(r, f) for the overlapping intensity pair (r, f). In the JSM-weighted joint histogram, the histogram entries related to the outliers and homogeneous regions have little impact on the histogram dispersion. Furthermore, each histogram entry for the corresponding salient structures is the sum of smoothly varying fractions of one. As a result, the smoothed compact clusters (see Figure 1g) in the neighboring bins for the corresponding salient structures is introduced to solve the outlier-induced dispersion and the undesired clotted clusters caused by the homogeneous regions (see Figure 1f) in the histogram.

3. Local Nonrigid Registration Based on Keypoint Clustering

Given the globally registered images and the affiliated JSM (see Figure 1a,b,h), there is a global correspondence between the paired contiguous matching areas from the corresponding anatomical tissues and tumor regions in the two images. Moreover, the deformations of the corresponding anatomical tissues and tumor regions occur locally in the paired matching areas. For example, the tumor-induced local large deformation occurs in the paired corresponding tumor regions (see Figure 1a,b). To facilitate characterization of those local deformations, keypoints belong to those contiguous matching areas can be further grouped into paired keypoint clusters, since they presumably represent the local deformations of the paired contiguous matching areas in the brain tumor resection images.

3.1. EM Algorithm-Based Keypoint Clustering

Ideas related to clustering based control point setup was first suggested by Chui et al. [36]. The cluster centers of point sets is provided for a concise representation of the original point data and is used as control points for deformation. Recently, clustering-based registration of brain white matter fibers has been developed in [47]. In this work, we automatically detect 2D stable DoG keypoints by adopting an open implementation [48] of the SIFT detector [23]. Figure 1i shows that the many DoG keypoints are extracted from the intra-operative MR images. Subsequently, keypoint clustering is introduced so that the paired keypoint groups can be provided for a point cloud representation of the paired contiguous matching areas in the two images. This could introduce the intermediate cluster-to-cluster keypoint correspondence which represents a useful constraint on the local point-to-point keypoint correspondence detection for the local nonrigid registration.
In order to divide the keypoints into groups, we make use of the EM algorithm [49] to determine the maximum likelihood parameters of a mixture of K Guassian components in the keypoint space. The EM algorithm is used for finding maximum likelihood parameter estimates when there is missing or incomplete data. In our case, the missing data is the Gaussian cluster to which the points in the keypoint space belong. Assume that we are using K Gaussians in the mixture model. The probability density function of the mixture model can be stated as follows:
p ( y | θ ) = i = 0 K α i p i ( y | θ i )
where y = (x, y) is a coordinate vector of a 2D keypoint, i is the cluster index, the αi's represent the mixing weights ( ( i = 0 K α i = 1 ), θ represents the collection of parameters (α1,…, αK, θ1,…, θK), and pi is a two-dimensional Gaussian density parameterized by θi = (μi, Σi) (i.e., the means and covariance):
p i ( y | θ i ) = 1 2 π | i | 1 / 2 e 1 2 ( y μ i ) T i 1 ( y μ i )
The first step in applying the EM algorithm to the Gaussian mixture problem is to initialize K mean vectors μ1,…, μK and K covariance matrices Σ1,…,ΣK to represent each of the K groups. We initialize these parameters using improved K-means clustering (We will return to the details of choosing K and the initial parameters shortly.). On subsequent restarts of the EM iteration, we have found that using improved K-means clustering initialization yields good results with fast convergence. The update equations take on the following form:
α i new = 1 N j = 1 N p ( i | y j , θ old )
μ i new = j = 1 N y j p ( i | y j , θ old ) j = 1 N p ( i | y j , θ old )
i new = j = 1 N p ( i | y j , θ old ) ( y j μ i new ) ( y j μ i new ) T j = 1 N p ( i | y j , θ old )
where N is the total number of keypoints, and p (i| yj, θold) is the probability that Gaussian i fits the keypoint yj, given the data θold:
p ( i | y j , θ old ) = α i p i ( y j | θ i old ) k = 1 N α k p k ( y j | θ i old )
The above update equations are iteratively computed until the log likelihood log j = 1 N p ( y j | θ ) increases by less than 1 percent from one iteration to the next.
We have thus far not discussed how to choose K and the initial parameters (α1,…, αK, θ1,…, θK) of the EM algorithm. Choosing the number of clusters in a Gaussian mixture model [50] is a famous unsolved problem. Ideally, we would like to choose the number of clusters and those initial parameters that best suit the natural characteristic patterns of keypoint groups around the salient anatomical structures. Note that those keypoint groups may include a number of spatially disjoint regions in the brain images, and one readily available pattern representing the distribution for the random variable (x, y) of the keypoints is the joint histogram. It is generally believed that the number of peaks marked on the histogram may correspond to the number of Gaussians while the valleys specify the means and variances of Gaussian mixture models. Based on this knowledge, we can automatically detect the peaks and valleys in a smoothed histogram [51] as follows: (1) exponentially smooth the keypoint histogram to remove noisy peaks and valleys. The simplest form of exponential smoothing for each direction (x, y) of the histogram is given by the formula:
E 1 = H 1 E j = ρ H j + ( 1 ρ ) E j 1 = E j 1 + ρ ( H j E j 1 )
where ρ is the smoothing factor, and 0 < ρ < 1, Hj is jth the histogram value. We choose the desired ρ for minimizing the mean square error (MSE) according to the criterion:
ρ = arg min ρ M S E = arg min ρ ( 1 N i , j = 1 max i , max j ( H i j E ( i 1 ) ( j 1 ) ) 2 )
(2) detect peaks and valleys. A point Hi in histogram is a peak or a valley if and only if Hi is the largest or the smallest among its 8 neighbors; (3) remove pseudo peaks (or valleys) if the distance between neighboring peaks (or valleys) is less than a distance threshold. A reasonable distance threshold is (dmaxdmin)/2 < Dthreshold < dmax; (4) reject each keypoint which is far from any peaks obtained from step 3 under the distance threshold (1-2 times Dthreshold); (5) compute the number of peaks and the initial parameters of Gaussian mixture model. If the set of peaks and valleys are P = {t0, t1,…, tl}, V = {v0, v1,…, vl}, the initial parameters are given by the formula:
Ω j 0 = i = v j 1 v j h i
μ j 0 1 Ω j 0 i = v j 1 v j i h i
j 0 = 1 Ω j 0 i = v j 1 v j ( i μ j 0 ) h i
where hi is the probability that a pixel has a gray level intensity i.
Given a set of keypoints (y1, y2,…, yN) and the initial peak locations which are used as initial K means μ 1 1 , , μ K 1, K-means clustering is started to partition the N observations into K sets (K < N) C = {C1,C2,…,CK} so as to minimize the within-cluster sum of squares:
arg min C i = 1 K y j C i y j μ i 2
where μi is the mean of Ci. The algorithm proceeds by alternating between two steps. For Step 1, every point is assigned to the cluster with the closest mean. In Step 2, the centroid is computed for each set. The algorithm is stopped when the assignments of the keypoints no longer change. The output of the K-means clustering is the ( μ 1 1 , , μ K 1 ) and the C = {C1,C2,…,CK}, on which we obtain the mixing weights by computing α i 1 = card ( C i ) / N. The covariance matrices 1 1 , , K 1 could be obtained by computing the within-cluster statistics of keypoints. These values provide the ideal initial parameters of the EM algorithm.
The results of the EM clustering method are probabilistic. This means that every keypoint belongs to all clusters, but each assignment of a keypoint to a cluster has a different probability. Given the resultant parameters (α1,…, αK, θ1,…, θK) for a mixture of Gaussians, we search for the cluster index zi that meets zi = arg max αkpk (yi | θk) to estimate the Gaussian cluster to which the each keypoint yi belongs. Figure 1j shows that the circles representing floating keypoints and the crosses depicting the reference keypoints are classified into nine corresponding clusters.

3.2. Outlier Rejection and Correspondence Detection

Note that if one cluster pair of tumor regions suffers from the large structural mismatches with local large deformation and outliers, this cluster pair will bring together many keypoints and small average JSM value ζk (computed as ζk = Σw (yi)/card (Ck) (yiCk)). Therefore, the keypoint cluster of tumor region in each image can be localized if its ζk is below a threshold value (ζthreshold = 0.4). In Figure 1j and Figure 2a,b, the keypoint clusters in both images are sorted according to the average JSM values {0.266, 0.404, 0.494, 0.533, 0.560, 0.565, 0.590, 0.639, 0.734}. The tumor resection clusters with ζ1 = 0.266 could be automatically detected from the nine clusters. Figure 2e,f shows that the two images have ten clusters with the tumor resection cluster having 0.234 average JSM value. Furthermore, the keypoint clusters of tumor regions in both images could be regarded as an ellipse model based on the mean and variation of the cluster, and then the outlier keypoints inside the ellipse are removed while the boundary significant keypoints (Figure 2c,d,g,h) are saved for subsequent correspondence detection, i.e., delete x = (x, y) if [ (xμx)2/(κσx)2 + (yμy)2/(κσy)2] < 1 with κ = 0.1. After outlier keypoint rejection, we approach the control point setup and related correspondence detection as follows.
In general, correspondences are found by choosing the two points with the optimum similarity measure (such as mutual information and cross correlation) between the points' surrounding regions, but this template matching method can give unsatisfactory correspondence due to the large number of keypoints detected in each image and the region around each keypoint being too small to include sufficient information. To enhance the confidence of template matching-based correspondence detection, we use two steps to find the robust correspondence:
  • Choose the significant keypoints for irregular control point setup. The SIFT keypoint detector has assigned a location and a scale to each stable DoG keypoint. The scale defines the saliency measure of each keypoint such that the keypoint with a large scale could be identified at the same location in the noisy pre- and intra-operative MR brain images. Based on the above consideration, a keypoint with the largest scale measure within a neighborhood could be saved as the significant keypoint in a cluster. In Figure 2a,b,e,f, the significant keypoints selected from each cluster well represent the irregular control point setup in the salient structures. Figure 2c,d,g,h shows the boundary significant keypoints at the tumor resection clusters.
  • Use the cluster-to-cluster correspondence and inverse consistent correspondence calculation to enhance the confidence of MI-based template matching. In detail, we use a local MI (see the details in the following section) to match a reference point ri in cluster k to the floating points within a certain search radius δk in the corresponding cluster at floating image. The floating points, together with their similarity values to the reference point, define a forward search map of the inverse consistent correspondence detection. The floating point with the maximum local MI in the forward search map can be appointed as a candidate corresponding point fi. For inverse consistency, to mach a floating point fi in cluster k, a backward search map defines a candidate corresponding point ri. When the constraint of inverse consistent correspondence detection is satisfied, i.e., ri = ri and fi = fi, the point pair (ri, fi) is accepted as a corresponding control point pair.
In recent years, increased attention has been paid to the local similarity measure of small image regions for local nonrigid registration [5254]. Among various local similarity measures, the information-theoretic local similarity measures such as local MI [52] have proved effective in nonrigid registration. Noting that a local MI between two corresponding areas can be low if the two local salient areas are mismatched, or if either of the corresponding areas is featureless and therefore has low entropy (This is because MI is bounded by the entropies of the individual areas, i.e., MI (a, b) ≤ min {H(a), H(b)}), we solved this ambiguity by adopting the idea of local mismatch measure [52] to evaluate the local MI (LMI) as LMI = MI (a, b)/min (H(a),H(b)), where MI (a, b), H (a) and H (b) are local measures which are computed over a finite subblock in the floating and reference images. Note that the subblock size cannot be too small, we choose subblock size of 21 × 21 for local MI.

3.3. Quasi-Inverse Consistent Deformation Modeling

Although we have initialized correspondence detection in the context of local large deformation and outliers, it may not yet lead to 100% accurate matching of DoG control points, a few of the corresponding pairs are likely to be incorrect. We adopt an approximating [32, 43, 55] rather than interpolating RBF to remove the effects of those correspondence outliers on deformation modeling. Furthermore, a desirable property of inverse consistency for nonrigid registration should be considered in our work. That is, when the roles of reference image and floating images are swapped, the forward transformation T1 : rifi, is the inverse of the backward transformation T2 : firi, i.e., T 1 = T 2 1. Let {ri} {fi}, i = {1,…,N} denote a set of N reference and floating control points, the inverse consistent registration [16, 42] cost function is:
D = 1 N i = 1 N r i T 1 ( f i ) + 1 N i = 1 N T 2 1 ( f i ) T 1 ( f i ) + J T
where T1 is the RBF forward transformation, T 2 1 is the inverse of the backward transformation. The first term measures the distance between the two control point sets and can be thought of as a registration accuracy constraint, and the second [16] measures inverse consistency at the point location, the third term represents the smoothness of the transformation and is formalized within the context of RBF transformation.
In this work we have used the Wendland's function as RBFs [43] to parameterize the transformation:
R ( d / ψ ) = { ( 1 d / ψ ) 4 ( 4 d / ψ + 1 ) , ( 1 d / ψ ) > 0 0 else
where the d is a Euclidean distance and the ψ is a radius of support which defines the locality of the RBF deformation. The minimum ψ depends on the control point displacement Δ = max (Δ1, Δ2)(Δ1 and Δ2 are the displacement from the floating control point to the reference control point in x and y coordinate direction, respectively) with the constraint ψ > 2.98Δ.
Generally, an interpolation transformation function T : R2 → R2 for 2D images based on control points must fulfill the constraints T (fi) = ri, i = 1,…,N. Often, each coordinate of the transformation function is calculated separately, i.e., the interpolation problem Tl : R2 → R is solved for each coordinate l = 1, 2 with the corresponding constraints Tl (fi) = rl,i. With N reference and floating control points in the globally registered images, the nonrigid registration could estimate the displacements for a floating points as follows [43]:
T l ( f ) = i = 1 N β i R ( f f i )
where the ‖ffi‖ is the Euclidean distance from f to fi. R(d) = R(‖d‖) is the Wendland's radial basis function depending only on the Euclidean distance d. The RBFs R(‖ffi‖) are centered around the N control points. The RBFs coefficient βl = (βl,1,…βl,N)T can be calculated to maximize smoothness whilst interpolating through the control points by solving the following linear system:
β l = ( K + χ W 1 ) 1 r l
where rl = (rl,1,…, rl,N)T is an N vector of the lth coordinate of the reference points and K an N × N positive definite matrix (Kij = R(‖fifj‖)). The matrix W = diag { 1 / τ 1 2 , , 1 / τ N 2 } and τi represent point localization errors. χ is an approximating factor to control the tradeoff between the smoothness of deformation and the accuracy of control point matching, which is very important to maximize smoothness whilst preventing the folding or the tearing of the deformation field.
Nonrigid registration is started by the correspondence detection and the local deformation computation for the tumor resection area, subsequently modified by the surrounding clusters in the ascending order of ζk. Nevertheless, the entire deformation field (especially the transition between each cluster's deformation field) would be non-smooth and even tearing, if the nonrigid registration is implemented once when the local correspondence detection and the local RBF-based deformation for each cluster are finished independently. Therefore, our method is processed within a deterministic annealing iteration framework (the maximum number of iterations is 5), both in terms of the inverse consistent correspondence detection as well as the approximating local transformation model.
Specifically, in the initial registration iteration, the search radius δk of inverse consistent correspondence detection and the approximating factor χ in local RBF modeling are inversely proportional to the average JSM value ζk of each cluster. With the process of registration, the contiguous area of the current cluster in floating image is locally changed by the previous iteration's and the neighboring cluster's local deformations, the template-matching based inverse consistent point correspondence should be adjusted around the current cluster before implementing the local RBF-based deformation. The correspondence search radius and the approximating factor values are gradually reduced, making the RBF-based local deformation field gradually moving from an approximation formulation to an interpolation formulation, thus effectively increasing the specificity of final correspondence detection and the smoothness of deformation fields.

4. Experimental Results

We preliminarily evaluated our algorithm on 5 pairs of pre- and intra-operative (or post-operative) MR T1 images that are corresponding slices of rigidly transformed 3D MR datasets. We first remove the skull near to the tumor resection area to prevent from introducing the deformation of rigid skull. We use 18 × 18 bins for 2D DoG keypoint joint histogram, which is exponentially smoothed with the exponential smoothing factor ρ being selected as 0.5 ∼ 0.6. A pseudo peak in the histogram is removed if a 30-pixel radius circular neighborhood has more than two neighboring peaks. We reject all the DoG keypoints that are more than 80 pixels away from any peaks in the histogram. In irregular setup of DoG control point grids and the correspondence detection, the maximum number of keypoints chosen as candidate control points in each cluster is equal to 30. With the process of registration in deterministic annealing framework, the search radius is iteratively reduced from 10 pixels to 4 pixels for the local MI-based correspondence detection. The optimal support radius ψ of RBF in this work is a fixed value of 60 while the approximating factor χ is iteratively reduced from 0.5 to 0.01. The registration is terminated when either the value of the cost function is below a predefined threshold (1 ∼ 1.5) or the maximum number of iterations is reached.
In contrast to four state-of-the-art intensity-based registration methods, including B-Spline with correlation ratio [56] (BCR), B-Spline with normalized MI (BMI) [34, 57], Demons (DEM) [58] and diffeomorphic Demons (DIF) [59], which are failed or unsatisfactory in tumor resection with outliers and local large deformation, our proposed method based on JSM & keypoint clustering (JKC) successfully model the tumor resection-induced brain shift (see Figures 34 and Table 1 cases 1–3). Table 1 summarizes the registration quality in terms of correlation ratio (CR) and normalized MI (NMI) [4, 60]. Figure 3a–d and Figure 4a–d show that the regions around the tumor areas have smoothly deformed to shift towards the tumor resection areas in the JKC registration results. The smoothly constructed deformation field is displayed by means of displacement vector field with variations of the vector color (from blue to red according to the displacement amplitude in pixels). The quality of these registration results also can be validated by the respective small CRs listed in the Table 1 cases 1–3. We also define the average error distance between the manually defined reference landmarks and the floating landmarks in the registered images (Figure 5 shows that most landmarks are around the tumor resection areas and the neighboring normal tissues), our method's average accuracy for the three cases of local large deformation (cases 1–3) achieves an average error distance of less than 1.2 pixel while the average accuracy for the two cases of small deformation achieves an average error distance of less than 1 pixel.
The BCR and BMI registration are implemented at two pass with the different transformation options (B-Spline degree for all axes: 1, 2; B-Spline control points for all axes: 8, 16; gradient descent minimize step size: 1.0, 0.5; gradient descent minimize maximum search steps: 10, 10) and the different iteration options (the convergence limit of minimum change rate for one iteration: 0.1, 0.01; maximum number of iterations: 10, 10) [57]. The dark areas on the deformation image (Figure 3f,h and Figure 4f,h) are related to the areas on the two images which did not perform a deformation during the registration or the deformation was relatively small. The light areas are related to the areas on the two images which perform a bigger deformation. In contrast to the BMI (Figure 3e,f and Figure 4e,f) that fails to compensate the local large deformation around the tumor areas, the BCR (Figure 3g and Figure 4g) can guide the anatomical structure deformation around the tumor resection regions (This also can be validated from the cases 1–3 at Table 1, at which the BCR have smaller CRs than the BMI, DEM and DIF). Nevertheless, the BCR has blurred and excessively shrunk the pre-operative anatomical structures around the tumor areas. From this experiment, we also confirm that the CR is more appropriate to be a similarity measure for the nonrigid monomodal registration of small structures than the NMI.
The DEM and DIF registration are conducted with a maximum step length of 2 pixels, 1.0 standard deviations of the Gaussian smoothing, a maximum number of 200 iterations and 0.001 intensity difference threshold. Treating each image as a set of iso-intensity contours and assuming the same anatomical point having the same intensity level in both images, the DEM and DIF easily distort the data to some extent, which may introduce strange artifacts similar to pieces of small mosaic patterns in the deformed pre-operative images (Figure 3i,k and Figure 4i,k). Additionally, the ”demons algorithm” with its large number of degrees-of-freedom allows to run into problems with the physical fidelity of the deformation field (Figure 4k and the DIF at case 2 in Table 1).
In the two cases of local small structure distortion, all five methods get good results (see Figure 6 and Table 1 cases 4–5). To facilitate the visual assessment of registration accuracy, we also use a mosaic (or checkerboard) pattern to fuse the two images. Figure 6a–b display the multi-temporal brain changes with a small tumor presented in the preoperative MRI image but not in the post-operative MRI image. From clustering analysis, the brain images have seven clusters (Figure 6c,d) with no cluster's average JSM (Figure 6e) value less than the average JSM threshold, so that there is no need for any cluster to reject outliers before registration procedure. Although there exist a few outlier keypoints in the small tumor, the clustering-based correspondence detection and the approximating RBF modeling could suppress the effect of outlier keypoints on registration result (Figure 6f). We also applied the method to multimodal brain image registration. Figure 7a–b show the T1 MRI (reference image) and proton density MRI (floating image) distorted by known deformations, which have six clusters with average JSM (Figure 7c–e) values {0.400, 0.556, 0.631, 0.652, 0.687, 0.741}. There is no need for any cluster to reject outliers in locally distorted structures.

5. Conclusions

We have presented a new hybrid nonrigid registration of brain MRI images to model tumor resection-induced brain shift. While SIFT keypoint detector was designed under the assumption of linear changes in intensity, the DoG keypoint detected by the SIFT detector can be effective in robustly matching intra- and pre-operative MR image pairs taken under substantially different illumination condition due to the spatially-varying intensity inhomogeneity and large intra-operative noise. Nevertheless, the keypoint detection (and the keypoint description) could not sufficient [28] for the unambiguous correspondence detection in the nonrigid image registration with the complex local large deformation and outliers, and should be incorporated with the structural mismatch detection mechanism and other useful correspondence constraints. By testing on a few typical clinical images and comparing four state-of-the-art nonrigid registration methods, this work introduces the JSM and keypoint clustering towards the robust keypoint correspondence detection for nonrigid registration of brain tumor resection MR images.
The main drawback of our approach is that a number of parameters should be set. By applying more experiments on the brain tumor resection images, the future study must be carried out determining the optimal parameters and comparing the JSM reconstruction methodologies with discussion their robustness. An interesting future development is to design a robust deformation invariant keypoint detector [29] which should meet both needs for reconstructing JSM to detect the structural mismatches and finding good landmark with high repeatability for nonrigid registration.

Acknowledgments

The authors would like to thank Simon K.Warfield for allowing the use of the images, all reviewers for their useful comments and Andrea Vedaldi for the open-source implementation of the SIFT algorithm. The authors thank the open source implementation of diffeomorphic demons and the open source ITK project. The authors also give thanks to Wang Yikang, Zhuang Tiange, Zhu Yisheng and Xu Yongjiang for their help to our work. This work was supported in part by the NSFC (Grant 60872102), the National Basic Research Program of China (Grant 2010CB834303), the Science Foundation of Shanghai Municipal Science & Technology Commission (Grant 04JC14060), Shanghai Municipal Health Bureau (Grant 2008115) and the small animal imaging project (Grant 06-545).

References

  1. Clatz, O.; Delingette, H.; Talos, I.F.; Golby, A.J.; Kikinis, R.; Jolesz, F.A.; Ayache, N.; Warfield, S.K. Robust nonrigid registration to capture brain shift from intraoperative MRI. IEEE Trans. Med. Imag. 2005, 24, 1417–1427. [Google Scholar]
  2. Ji, S.; Wu, Z.; Hartov, A.; Roberts, D.W.; Paulsen, K.D. Mutual-information-based patient registration using intraoperative ultrasound in image-guided neurosurgery. Med. Phys. 2008, 35, 4612–4624. [Google Scholar]
  3. Zitova, B.; Flusser, J. Image registration methods: a survey. Image Vision Comput. 2003, 21, 977–1000. [Google Scholar]
  4. Maes, F.; Collignon, A.; Vandermeulen, D.; Marchal, G.; Suetens, P. Multimodality image registration by maximization of mutual information. IEEE Trans. Med. Imag. 1997, 16, 187–198. [Google Scholar]
  5. Viola, P.; Wells, W.M. Alignment by maximization of mutual information. Int. J. Comput. Vis. 1997, 24, 137–154. [Google Scholar]
  6. Hellier, P.; Barillot, C.; Memin, E.; Perez, P. Hierarchical estimation of a dense deformation field for 3-D robust registration. IEEE Trans. Med. Imag. 2001, 20, 388–402. [Google Scholar]
  7. Hachama, M.; Desolneux, A.; Richard, F. Combining registration and abnormality detection in mammography. Proceedings of International Workshop on Biomedical Image Registration, Utrecht, The Netherland, July 9–11, 2006; pp. 178–185.
  8. Staring, M.; van der Heide, U.A.; Klein, S.; Viergever, M.A.; Pluim, J.P.W. Registration of cervical MRI using multifeature mutual information. IEEE Trans. Med. Imag. 2009, 28, 1412–1421. [Google Scholar]
  9. Auer, M.; Regitnig, P.; Holzapfel, G.A. An automatic nonrigid registration for stained histological sections. IEEE Trans. Image Process. 2005, 14, 475–486. [Google Scholar]
  10. Cooper, L.; Sertel, O.; Kong, J.; Lozanski, G.; Huang, K.; Gurcan, M. Feature-based registration of histopathology images with different stains: an application for computerized follicular lymphoma prognosis. Comput. Methods Programs Biomed. 2009, 96, 182–192. [Google Scholar]
  11. Gao, S.; Zhang, L.; Wang, H.; Crevoisier, R.; Kuban, D.D.; Mohan, R.; Dong, L. A deformable image registration method to handle distended rectums in prostate cancer radiotherapy. Med. Phys. 2006, 33, 3304–3312. [Google Scholar]
  12. Salvado, O.; Wilson, D.L. Removal of local and biased global maxima in intensity-based registration. Med. Image Anal. 2007, 11, 183–196. [Google Scholar]
  13. Tsao, J. Interpolation artifacts in multimodality image registration based on maximization of mutual information. IEEE Trans. Med. Imag. 2003, 22, 854–864. [Google Scholar]
  14. Lu, X.; Zhang, S.; Su, H.; Chen, Y. Mutual information-based multimodal image registration using a novel joint histogram estimation. Comput. Med. Imaging Graph. 2008, 32, 202–209. [Google Scholar]
  15. Qin, B.; Gu, Z.; Sun, X.; Lv, Y. Registration of images with outliers using joint saliency map. IEEE Signal Process. Lett. 2010, 17, 91–94. [Google Scholar]
  16. Rogers, M.; Graham, J. Robust and accurate registration of 2-D electrophoresis gels using point-matching. IEEE Trans. Imag. Proc. 2007, 16, 624–635. [Google Scholar]
  17. Fei, B.; Duerk, J.L.; Sodee, D.B.; Wilson, D.L. Semiautomatic nonrigid registration for the prostate and pelvic MR volumes. Acad. Radiol. 2005, 12, 815–824. [Google Scholar]
  18. Kadir, T.; Brady, M. Saliency, scale and image description. Int. J. Comput. Vis. 2001, 45, 83–105. [Google Scholar]
  19. Huang, X.; Sun, Y.; Metaxas, D.; Sauer, F.; Xu, C. Hybrid image registration based on configural matching of scale-invariant salient region features. Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshop, Washington, DC, USA, June 27–July 02, 2004; pp. 167–179.
  20. Yang, Y.; Williams, J.P.; Sun, Y.; Blum, R.S.; Xu, C. Non-rigid image registration using geometric features and local salient region features. Proceedings of 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, New York, NY, USA, June 17–22, 2006; pp. 825–832.
  21. Lindeberg, T. Feature detection with automatic scale selection. Int. J. Comput. Vis. 1998, 30, 77–116. [Google Scholar]
  22. Zhan, Y.; Ou, Y.; Feldman, M.; Tomaszeweski, J.; Davatzikos, C.; Shen, D. Registering histological and MR images of prostate for image-based cancer detection. Acad. Radiol. 2007, 14, 1367–1381. [Google Scholar]
  23. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar]
  24. Franz, A.; Carlsen, I.; Renisch, S. An adaptive irregular grid approach using SIFT features for elastic medical image registration. In Bildverarbeitung fr die Medizin; Springer: Berlin, Germany, 2006; pp. 201–205. [Google Scholar]
  25. Moradi, M.; Abolmaesoumi, P.; Mousavi, P. Deformable registration using scale space keypoints. Proc. SPIE Med. Imaging: Image Process 2006, 6144, 791–798. [Google Scholar]
  26. Ni, D.; Chui, Y.P.; Qu, Y.; Yang, X.; Qin, J.; Wong, T.T.; Ho, S.S.H.; Heng, P.A. Reconstruction of volumetric ultrasound panorama based on improved 3D SIFT. Comput. Med. Imaging Graph. 2009, 33, 559–566. [Google Scholar]
  27. Kashani, R.; Hub, M.; Balter, J.M.; Kessler, M.L.; Dong, L.; Zhang, L.; Xing, L.; Xie, Y.; Hawkes, D.; Schnabel, J.A.; McClelland, J.; Joshi, S.; Chen, Q.; Lu, W. Objective assessment of deformable image registration in radiotherapy: A multi-institution study. Med. Phys. 2008, 35, 5944–5953. [Google Scholar]
  28. Kelman, A.; Sofka, M.; Stewart, C.V. Keypoint descriptor for matching across multiple image modalities and non-linear intensity variations. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, June 18–23, 2007; pp. 1–7.
  29. Li, G.; Guo, L.; Liu, T. Deformation invariant attribute vector for deformable registration of longitudinal brain MR images. Comput. Med. Imaging Graph. 2009, 33, 384–398. [Google Scholar]
  30. Liao, S.; Chung, A.C.S. Feature based non-rigid brain MR image registration with symmetric alpha stable filters. IEEE Trans. Med. Imag. 2009. [Google Scholar] [CrossRef]
  31. Ou, Y.; Davatzikos, C. DRAMMS: Deformable registration via attribute matching and mutual-saliency weighting. Proceedings of Information Processing in Medical Imaging, Williamsburg, WV, USA, July 5–10, 2009; pp. 50–62.
  32. Marias, K.; Behrenbruch, C.; Parbhoo, S.; Seifalian, A.; Brady, M. A registration framework for the comparison of mammogram sequences. IEEE Trans. Med. Imag. 2005, 24, 782–790. [Google Scholar]
  33. Mian, A. S.; Bennamoun, M.; Owens, R. Keypoint detection and local feature matching for textured 3D face recognition. Int. J. Comput. Vis. 2008, 79, 1–12. [Google Scholar]
  34. Rueckert, D.; Sonoda, L.I.; Hayes, C.; Hill, D.L.G.; Leach, M.O.; Hawkes, D.J. Nonrigid registration using free-form deformations: Application to breast MR images. IEEE Trans. Med. Imag. 1999, 18, 712–721. [Google Scholar]
  35. Chui, H.; Rangarajan, A. A new point matching algorithm for non-rigid registration. Comput. Vision Image Understanding 2003, 89, 114–141. [Google Scholar]
  36. Chui, H.; Win, L.; Schultz, R.; Dncan, J.; Rangarajan, A. A unified non-rigid feature registration method for brain mapping. Med. Image Anal. 2003, 7, 113–130. [Google Scholar]
  37. Tustison, N.J.; Awate, S.P.; Song, G.; Cook, T.S.; Gee, J.C. A new information-theoretic measure to control the robustness-sensitivity trade-off for DMFFD point-set registration. Proceedings of Information Processing in Medical Imaging, Williamsburg, WV, USA; 2009; pp. 215–226. [Google Scholar]
  38. Besl, P.J.; McKay, N.D. A method for registration of 3-D shapes. IEEE Trans. Patt. Anal. Mach. Intell. 1992, 14, 239–256. [Google Scholar]
  39. Guest, E.; Berry, E.; Baldock, R.A.; Fidrich, M.; Smith, M.A. Robust point correspondence applied to two- and three-dimensional image registration. IEEE Trans. Patt. Anal. Mach. Intell. 2001, 23, 165–179. [Google Scholar]
  40. Matsopoulos, G.K.; Mouravliansky, N.A.; Asvestas, P.A.; Delibasis, K.K.; Kouloulias, V. Thoracic non-rigid registration combining self-organizing maps and radial basis functions. Med Image Anal. 2005, 9, 237–254. [Google Scholar]
  41. Economopoulos, T.L.; Asvestas, P.A.; Matsopoulos, G.K. Automatic correspondence on medical images: A comparative study of four methods for allocating corresponding points. J. Digital Imaging 2009. [Google Scholar] [CrossRef]
  42. Johnson, H.; Christensen, G. Consistent landmark and intensity-based image registration. IEEE Trans. Med. Imag. 2002, 21, 450–461. [Google Scholar]
  43. Fornefett, M.; Rohr, K.; Stiehl, H.S. Radial basis functions with compact support for elastic registration of medical images. Image Vis. Comput. 2001, 19, 87–96. [Google Scholar]
  44. Orchard, J. Multimodal image registration using floating regressors in the joint intensity scatter plot. Med. Image Anal. 2008, 12, 385–96. [Google Scholar]
  45. Liu, F.; Gleicher, M. Region enhanced scale-invariant saliency detection. Proceedings of the IEEE International Conference on Multimedia and Expo, Toronto, Ontario, Canada, July 9–12, 2006; pp. 1477–1480.
  46. Jenkinson, M.; Smith, S. A global optimization method for robust affine registration of brain images. Med. Image Anal. 2001, 5, 143–156. [Google Scholar]
  47. Xu, Q.; Anderson, A.W.; Gore, J.C.; Ding, Z. Unified bundling and registration of brain white matter fibers. IEEE Trans. Med. Imag. 2009, 28, 1399–1411. [Google Scholar]
  48. Vedaldi, A. An Open Implementation of the SIFT Detector and Descriptor. UCLA CSD Technical Report 070012, 2006; Available online: http://www.vlfeat.org/~vedaldi/code/siftpp.html (Accessed on 25 November 2009).
  49. Dempster, A.; Laird, N.; Rubin, D. Maximum likelihood from incomplete data via the EM algorithm. J. Royal Statistical Soc. 1977, 39, 1–38. [Google Scholar]
  50. Constantinopoulos, C.; Titsias, M.K.; Likas, A. Bayesian feature and model selection for gaussian mixture models. IEEE Trans. Patt. Anal. Mach. Intell. 2006, 28, 1013–1018. [Google Scholar]
  51. Ren, B.; Wang, B.; Luo, B. Method of threshold and peak detection based on histogram exponent smoothing. China J. Image Graphics 1997, 2, 230–233. [Google Scholar]
  52. Park, H.; Bland, P.H.; Brock, K.K.; Meyer, C.R. Adaptive registration using local information measures. Med. Image Anal. 2004, 8, 465–473. [Google Scholar]
  53. Rogelj, P.; Kovacic, S.; Gee, J.C. Point similarity measures for non-rigid registration of multi-modal data. Comput. Vis. Image Understanding 2003, 92, 112–140. [Google Scholar]
  54. Karacali, B. Information theoretic deformable registration using local image information. Int. J. Comput. Vis. 2007, 72, 219–237. [Google Scholar]
  55. Zagorchev, Z.; Goshtasby, A. A comparative study of transformation functions for nonrigid image registration. IEEE Trans. Imag. Proc. 2006, 15, 529–538. [Google Scholar]
  56. Roche, A.; Malandain, G.; Pennec, X.; Ayache, N. The correlation ratio as a new similarity measure for multimodal image registration. Proceedings of 1st International Conference on Medical Image Computing and Computer-Assisted Intervention, Cambridge, MA, USA, October 11–13; 1998; pp. 1115–1124. [Google Scholar]
  57. MIPAV (Medical image processing, analysis, and visualization), Center for Information Technology, U.S. National Institutes of Health. Available online: http://mipav.cit.nih.gov/ (Accessed on 19 September 2009).
  58. Thirion, J.P. Image matching as a diffusion process: an analogy with maxwell's demons. Med. Image Anal. 1998, 2, 243–260. [Google Scholar]
  59. Vercauteren, T.; Pennec, X.; Perchant, A.; Ayache, N. Diffeomorphic demons: Efficient non-parametric image registration. NeuroImage 2009, 45, S61–S72. [Google Scholar]
  60. Studholme, C.; Hill, D.L.G.; Hawkes, D.J. An overlap invariant entropy measure of 3D medical image alignment. Patt. Recogn. 1999, 32, 71–86. [Google Scholar]
  • Classification: PACS 87.57.nj, 87.85.Pq, 87.61.-c, 87.57.nm, 87.57.-N, 87.57.-s.
Figure 1. (a)–(b) Intra- and pre-operative MR images with tumor resection. (c)–(d) Gradient value profiles of the lines in (a)–(b), which are marked as dashed lines. (e) JSM value profiles of the lines in (a)–(b). (f) Joint histogram dispersion with two clotted clusters (dark red in pseudo color). (g) The JSM-weighted joint histogram with smoothed compact clusters for (a)–(b). (h) JSM for the two images in (a)–(b) with low JSM values at the tumor resection area. (i) The intra-operative MR image and the circle marked DoG keypoints. (j) The pre-operative image and resultant keypoint clustering with circle marked floating keypoints and cross marked reference keypoints. Different colors mean different clusters.
Figure 1. (a)–(b) Intra- and pre-operative MR images with tumor resection. (c)–(d) Gradient value profiles of the lines in (a)–(b), which are marked as dashed lines. (e) JSM value profiles of the lines in (a)–(b). (f) Joint histogram dispersion with two clotted clusters (dark red in pseudo color). (g) The JSM-weighted joint histogram with smoothed compact clusters for (a)–(b). (h) JSM for the two images in (a)–(b) with low JSM values at the tumor resection area. (i) The intra-operative MR image and the circle marked DoG keypoints. (j) The pre-operative image and resultant keypoint clustering with circle marked floating keypoints and cross marked reference keypoints. Different colors mean different clusters.
Sensors 09 10270f1
Figure 2. (a)–(b) and (e)–(f) Clustered significant keypoints in the two images. (c)–(d) and (g)–(h) Boundary significant keypoints around tumor resection regions with circles defining the scale measures of the points.
Figure 2. (a)–(b) and (e)–(f) Clustered significant keypoints in the two images. (c)–(d) and (g)–(h) Boundary significant keypoints around tumor resection regions with circles defining the scale measures of the points.
Sensors 09 10270f2
Figure 3. (a)–(b) Intra- and pre-operative MR images. (c) JKC. (d) Displacement vector field with the vector orientation and the variations of the vector color (the color scale encodes the norm of the displacement vector, in pixels). (e) BMI. (f) BMI deformation image. (g) BCR. (h) BCR deformation image. (i) DEM. (j) DEM deformation image (the color scale encodes the norm of the displacement vector, in pixels). (k) DIF. (l) DIF deformation image.
Figure 3. (a)–(b) Intra- and pre-operative MR images. (c) JKC. (d) Displacement vector field with the vector orientation and the variations of the vector color (the color scale encodes the norm of the displacement vector, in pixels). (e) BMI. (f) BMI deformation image. (g) BCR. (h) BCR deformation image. (i) DEM. (j) DEM deformation image (the color scale encodes the norm of the displacement vector, in pixels). (k) DIF. (l) DIF deformation image.
Sensors 09 10270f3
Figure 4. (a)–(b) Intra- and pre-operative MR images. (c) JKC. (d) Displacement vector field. (e) BMI. (f) BMI deformation image. (g) BCR. (h) BCR deformation image. (i) DEM. (j) DEM deformation image. (k) DIF. (l) DIF deformation image.
Figure 4. (a)–(b) Intra- and pre-operative MR images. (c) JKC. (d) Displacement vector field. (e) BMI. (f) BMI deformation image. (g) BCR. (h) BCR deformation image. (i) DEM. (j) DEM deformation image. (k) DIF. (l) DIF deformation image.
Sensors 09 10270f4
Figure 5. (a)–(b) Corresponding landmarks for intra- and pre-operative MR images at Figure 3(a)–(b). (c)–(d) Corresponding landmarks for intra- and pre-operative MR images at Figure 4(a)–(b).
Figure 5. (a)–(b) Corresponding landmarks for intra- and pre-operative MR images at Figure 3(a)–(b). (c)–(d) Corresponding landmarks for intra- and pre-operative MR images at Figure 4(a)–(b).
Sensors 09 10270f5
Figure 6. (a)–(b) The pre- and post-operative brain tumor images. (c)–(d) Significant keypoints selected from the different clusters. (e) JSM between the two images. (f) Fused image using check pattern after registration. (g) Warped mesh after registration. (h) Displacement vector field.
Figure 6. (a)–(b) The pre- and post-operative brain tumor images. (c)–(d) Significant keypoints selected from the different clusters. (e) JSM between the two images. (f) Fused image using check pattern after registration. (g) Warped mesh after registration. (h) Displacement vector field.
Sensors 09 10270f6
Figure 7. (a)–(b) The MR-T1 and MR proton density weighted images. (c)–(d) Significant keypoints selected from the different clusters. (e) JSM between the two images. (f) Fused image using check pattern after registration. (g) Warped mesh after registration. (h) Displacement vector field.
Figure 7. (a)–(b) The MR-T1 and MR proton density weighted images. (c)–(d) Significant keypoints selected from the different clusters. (e) JSM between the two images. (f) Fused image using check pattern after registration. (g) Warped mesh after registration. (h) Displacement vector field.
Sensors 09 10270f7
Table 1. CR and NMI values for different registration methods (The smaller values usually mean the better registration results).
Table 1. CR and NMI values for different registration methods (The smaller values usually mean the better registration results).
casesCRNMI

JKCBCRBMIDEMDIFJKCBCRBMIDEMDIF
10.06130.05530.10681.00000.99980.86190.85950.87130.98450.9846
20.08890.07430.21670.09740.03490.86070.85900.87750.84950.8195
30.06950.07040.26730.93540.89170.84210.86110.86970.97830.9624
40.06370.09740.41810.05900.06490.87450.82390.85800.78800.7925
50.07490.08580.31770.05820.06380.86570.84710.86930.79190.8142

Share and Cite

MDPI and ACS Style

Gu, Z.; Qin, B. Nonrigid Registration of Brain Tumor Resection MR Images Based on Joint Saliency Map and Keypoint Clustering. Sensors 2009, 9, 10270-10290. https://doi.org/10.3390/s91210270

AMA Style

Gu Z, Qin B. Nonrigid Registration of Brain Tumor Resection MR Images Based on Joint Saliency Map and Keypoint Clustering. Sensors. 2009; 9(12):10270-10290. https://doi.org/10.3390/s91210270

Chicago/Turabian Style

Gu, Zhijun, and Binjie Qin. 2009. "Nonrigid Registration of Brain Tumor Resection MR Images Based on Joint Saliency Map and Keypoint Clustering" Sensors 9, no. 12: 10270-10290. https://doi.org/10.3390/s91210270

Article Metrics

Back to TopTop