Next Article in Journal
Quantifying Streamflow Variations in Ungauged Lake Basins by Integrating Remote Sensing and Water Balance Modelling: A Case Study of the Erdos Larus relictus National Nature Reserve, China
Previous Article in Journal
In-Situ Measurement of Soil Permittivity at Various Depths for the Calibration and Validation of Low-Frequency SAR Soil Moisture Models by Using GPR
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Remote Sensing Image Registration Using Multiple Image Features

1
School of Information Science and Technology, Yunnan Normal University, Kunming 650092, China
2
The Engineering Research Center of GIS Technology in Western China of Ministry of Education of China, Yunnan Normal University, Kunming 650092, China
3
Laboratory of Pattern Recognition and Artificial Intelligence, Yunnan Normal University, Kunming 650092, China
4
Department of Electrical and Computer Engineering, National University of Singapore, Singapore 117576, Singapore
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2017, 9(6), 581; https://doi.org/10.3390/rs9060581
Submission received: 10 March 2017 / Revised: 27 May 2017 / Accepted: 4 June 2017 / Published: 12 June 2017

Abstract

:
Remote sensing image registration plays an important role in military and civilian fields, such as natural disaster damage assessment, military damage assessment and ground targets identification, etc. However, due to the ground relief variations and imaging viewpoint changes, non-rigid geometric distortion occurs between remote sensing images with different viewpoint, which further increases the difficulty of remote sensing image registration. To address the problem, we propose a multi-viewpoint remote sensing image registration method which contains the following contributions. (i) A multiple features based finite mixture model is constructed for dealing with different types of image features. (ii) Three features are combined and substituted into the mixture model to form a feature complementation, i.e., the Euclidean distance and shape context are used to measure the similarity of geometric structure, and the SIFT (scale-invariant feature transform) distance which is endowed with the intensity information is used to measure the scale space extrema. (iii) To prevent the ill-posed problem, a geometric constraint term is introduced into the L2E-based energy function for better behaving the non-rigid transformation. We evaluated the performances of the proposed method by three series of remote sensing images obtained from the unmanned aerial vehicle (UAV) and Google Earth, and compared with five state-of-the-art methods where our method shows the best alignments in most cases.

Graphical Abstract

1. Introduction

Remote sensing image registration refers to the fundamental task in image processing to align two or more images of the same scene (i.e., the sensed images and the reference images), which can be multiview (obtained from different viewpoint), multitemporal (taken at different times) and multisource (derived from different sensors). In this paper, we mainly focus on registering the remote sensing images taken from different viewpoint. As a basic technology in the field of remote sensing image processing, it has been widely used in the field of military and civilian such as natural disaster damage assessment, resource census, assignment of climate changes, military damage assessment, environment monitoring and ground targets identification, etc.
Existing remote sensing image registration methods can be approximately classified into two categories: area-based methods and feature-based methods. Various reviews on image registration methods can be found in [1,2,3,4,5]. Area-based methods use the pixel intensities of identical image windows to estimate similarity while feature-based methods extract image features (e.g., points, lines, and regions) and match them using similarity criteria [6]. When there are noises, complex distortion and significant radiometric differences between the images to be aligned, the computation complexity or the search space of the area-based methods will increase nonlinearly with the transformation complexity and feature-based methods are more robust and preferable. In this paper, the captured remote sensing images exist the local non-rigid geometric distortions caused by ground relief variations and imaging viewpoint changes, thus we mainly focus on feature-based methods for registration. Such methods generally consists of three steps [7]: (i) feature descriptors extraction; (ii) feature point sets registration; (iii) image transformation and resampling.
Generally, a good feature descriptor should be distinctive and at the same time robust to changes in viewing conditions as well as to errors of the point detector [8]. For desirable local image descriptors, to improve distinctiveness while maintaining robustness is the main concern [9]. Among the popular local invariant features which have been proposed in remote sensing image registration (e.g., Harris [10], scale-invariant feature transform (SIFT) [11,12] and speeded-up robust features (SURF) [13]), Mikolajczyk et al. [8] demonstrated that the performance of SIFT [11] under affine transformation, scale change, rotation, image blur, jpeg compression, and illumination change outperforms other local invariant descriptors in most of the tests. There are various researches for remote sensing image registration based on the variants of SIFT algorithm or the combination of SIFT and some other algorithms. Li et al. [14] proposed a new criterion named scale-orientation joint restriction criteria in order to overcome the intensity difference between remote sensing image pairs. Sedaghat et al. [6] introduced an automatic registration algorithm by extracting high-quality SIFT features in the uniform distribution of both the scale and image spaces. In addition, other variants of SIFT such as PCA-SIFT [15], SAR-SIFT [16], AB-SIFT [17] and SIFT-DRS [18] are also proposed in literatures. Furthermore, Goncalves et al. [19] developed a new AIR (Automatic image registration) method based on the combination of image segmentation and SIFT, and the method is complemented by a robust procedure of outlier removal. In [20], a novel coarse-to-fine scheme for automatic image registration based on SIFT and MI is proposed, and their method achieved the outlier removal and also can generally reject most incorrect matches.
Feature-based methods are typically formulated as a point set registration problem, since point representations are general and easy to extract. In order to achieve a robust point registration, it is crucial to construct putative correspondences based on local invariant feature similarity at first and then estimate the spatial transformation based on geometric structure feature constraint [21,22,23]. Here, we briefly review some point set registration methods since our method is based on feature point. Fischler et al. proposed the classical random sample consensus (RANSAC) [24] algorithm. Myronenko and Song [25] introduced a probabilistic method, called the coherent point drift (CPD) algorithm, for both rigid and non-rigid point set registration. Moreover, Liu and Yan [26] investigated how to discover common visual pattern discovery via spatially coherent correspondences and recover the correct correspondences. Zhang et al. [27] defined the spatially consistent topic model by making full use of the correlation between image classification and annotation. Recently, Ma et al. proposed a robust L 2 -minimizing estimate ( L 2 E ) [28] for non-rigid point set registration, they later proposed a flexible and general algorithm called locally linear transforming (LLT) [29] for both rigid and non-rigid registration on remote sensing images. More recently, Yang et al. [30] proposed a new method named GLMDTPS, which considers global and local structural differences as a linear assignment problem.
Although the aforementioned methods have been proposed for different applications, there still exists the following problems for remote sensing image registration. (i) These methods use either the local invariant features or geometric features, thus the distinctiveness capability of the feature points is lost in the complex remote sensing registration patterns. For example, methods for image registration which use only SIFT [11,12] or its variants [6,14,15,16,17,18,19,20] suffer from inaccurate registration of SIFT. In addition, the performances of the methods [23,24,25,26,27,28,29,30] which consider only the geometric structure features is limited by the assumption that the corresponding points have similar structures. (ii) Using a rigid or affine transformation model is infeasible for registering remote sensing images with different viewpoints since the images usually contain local non-rigid distortion. (iii) It is ill-posed for mapping one feature point set to another by a non-rigid transformation model since the solution is not unique.
In this paper, we present a new method for remote sensing image registration with different viewpoints. Compared with the current methods, the major differences and advantages of this paper are as follows. (i) Probabilistic model construction: the mixture-feature Gaussian mixture model (MGMM) is constructed for simultaneously dealing with different types of image features. (ii) Feature complementation: Making the respective advantages of Geometric structure and intensity information to complement each other, i.e., the Euclidean distance and shape context to measure the similarity of global and local geometric structure discrepancies, and the SIFT distance to measure the scale space extrema, both of which are combined and substituted into the mixture model by which the reliable correspondence is obtained. (iii) Geometric constraint: a regularization term to prevent the ill-posed problem based on coherent velocity field is introduced into the L2E-based energy function for better behaving the non-rigid transformation.
The rest of this paper is organized as follows. Section 2 introduces our method in detail. Section 3 demonstrates the registration performance of our method on various types of remote sensing images with different viewpoints against other methods, followed by some concluding remarks in Section 4.

2. Methodology

We first introduce the three major contributions of the proposed method: (i) modeling of the MGMM; (ii) feature specification and combination; and (iii) geometric constrained energy function. Whereafter the main process is demonstrated, followed by the method analysis.

2.1. Mixture-Feature Gaussian Mixture Model (MGMM)

The motivation behind the development of MGMM is the need for estimating reliable correspondence using multiple image features whereas the Gaussian mixture model (GMM) can only work on single feature. In this subsection we detail the derivation of the MGMM.
Based on the reasonable assumption that points from one set are normally distributed around points belonging to the other set, we therefore consider the registration of two point sets as a GMM probability density estimation problem. Let y j be the centroid of the j t h Gaussian component, x i the i t h data. The probability density function (PDF) is obtained as:
p ( x i ) = ( 1 ζ ) j = 1 m C i j p ( x i | y j ) + ζ 1 n ,
where p ( x i | y j ) = 1 ( 2 π σ 2 ) d e 2 exp x i y j 2 2 σ 2 with the equal isotropic covariances σ 2 shared, C i j = 1 m are nonnegative equal quantity with j = 1 m C i j = 1 , which are called the priors of the GMMs. 1 n is an additional uniform distribution with a weighting parameter ζ , 0 ζ 1 for outlier dealing. However, p ( x i | y j ) considers only the Euclidean distance and might lead to insufficient robustness in some registration scenarios, e.g., although the local structure around x a and x b are different, p ( x a | y j ) = p ( x b | y j ) if x a y j 2 = x b y j 2 .
Consequently, the MGMM which is capable of dealing with multiple image features is developed. The PDF of the MGMM is formulated as:
f ( x i ) = ( 1 ζ ) j = 1 m M i j f ( x i | y j ) + ζ 1 n ,
where f ( x i | y j ) = 1 ( 2 π σ 2 ) d e 2 exp G i j + α L ij and α is a constant. G m × n and L m × n exploit the global and local geometric structure discrepancies, priors M m × n are specified by measuring the discrepancy of intensity information, which is analogous to confidence.
Once we have the PDF of the MGMM, the correspondence between the two point sets can be easily inferred through the posterior probability of the MGMM which is written as:
p i j = M i j f ( x i | y j ) k = 1 m M i k f ( x i | y k ) + ζ 1 n ,
where P m × n is the posterior probability matrix by which the newly estimated coordinates of x i can be obtained. It is worth nothing that two strategies for constraining P are herein under consideration: (i) a single normalization enforcing j = 1 m p i j = 1 and (ii) a double normalization which additionally requires i = 1 n p i j = 1 . Intuitively, the former acts like an one-to-many correspondence estimation, whereas the latter provides an one-to-one counterpart. In our method, the number of the feature points are strictly equal because the applied SIFT feature extractor. In addition the outlier to data ratios, however, are unknown and varied due to the different overlap ratios of each image pair. In this case it is inappropriate for requiring an one-to-one correspondence since the presence of outliers might trap the estimation, whereby the first strategy is chosen in this paper.

2.2. Combination and Complementation of Multiple Image Features

The distinctiveness of feature points, which affects the accuracy of the feature matching, are mainly determined by the robustness of the involved measures. However, different types of measures have their own advantages and limitations, where our idea is to make their respective advantages complementary to each other. In order to make the paper more self-contained, we succinctly discuss the concept of the three features we used, i.e., the Euclidean distance, the SC and the SIFT distance. The combination of features and the complete form of posterior probability function Equation (3) are then discussed.

2.2.1. Euclidean Distance

The squared Euclidean distance of x i and y j is simply written as:
g i j = x i y j 2 .
It acts a straightforward global estimation with insufficient robustness in many scenarios. To alleviate this problem, g i j is commonly treated as an argument of a Gaussian function, which therefore plays a flexible global estimation by the help of the changeable covariances.

2.2.2. Shape Context (SC)

The work in [31,32] proposed the SC. A detailed illustration of the SC is shown in Figure 1. The SC constructs a polar coordinate with B R bins in the radial direction and B T bins in the tangential direction, by centering the polar coordinate at x i and y j , it counts the number of points within each bin and obtains two 1 × B sets { h i ( b ) } b = 1 B and { h j ( b ) } b = 1 B , respectively, where B = B R × B T . The SC discrepancy of x i and y j is measured using Chi-square distribution as:
l i j = 1 2 b = 1 B [ h i ( b ) h j ( b ) ] 2 h i ( b ) + h j ( b ) .
The SC is robust especially in scenarios like contour point set registration and hand-written characters matching, etc., because the geometric structure of these shapes are quite distinctive. However, it can also be deteriorated if x i and y j have similar neighborhood structure.

2.2.3. SIFT Distance

The SIFT algorithm introduced by Lowe is a classic algorithm, it first repeatedly convolves the images with Gaussians to produce the set of scale space images and obtains the difference-of-Gaussian images. The feature points are detected by comparing a pixel with its neighbors at the current and adjacent scales. Histograms of local oriented gradients around each feature points are then computed, obtaining an 128-dimensional vector as the SIFT descriptor, as shown in Figure 2. Finally, a process for matching the feature points are carried out. A matching pair is detected, if its distance is closer than τ times the distance of the second nearest neighbor. Thus, the feature points extracted from the sensed and reference images are of the same number, i.e., | X | = | Y | , where | · | denotes the cardinality of a set, and each pair with the same suffix are matched already from the view of intensity information.
The SIFT distance is defined as:
s i j = u i v j 2 ,
where S is of m × n dimension. u i and v j are the SIFT descriptors of x i and y j , respectively. We have introduced three relevant features of our method, i.e., G , L and S . The next concern is to substitute them into our MGMM.

2.2.4. Multiple Image Feature Based Correspondence Estimation

The global geometric feature discrepancy are defined by G i j = g i j 2 σ 2 , where σ 2 are the covariances of the MGMM. It actually forms the Gaussian function of the pairwise Euclidean distance in the context of f ( x i | y j ) . The local geometric feature discrepancy L i j are defined as the Chi-square distribution of the SC histograms, i.e., L i j = l i j . The Priors of the MGMMs are formulated as M i j = 1 s ˜ i j , where matrix S ˜ is obtained by row-by-row rescaling which makes each entry s ˜ i j [ 0 , 1 ] , and are therefore taken as the confidence. Substituting G i j , L i j and M i j into Equation (3), we can therefore rewrite it as:
p i j * = ( 1 s ˜ i j ) exp g i j 2 σ 2 + α l i j k = 1 m ( 1 s ˜ i k ) exp g i k 2 σ 2 + α l i k + ( 2 π σ 2 ) d e 2 ζ n .
The underlying assumption of density function Equation (7) is the decomposition of the process for human to recognize and categorize objects. Supposing such process is based on the linear combination among features such as Euclidean distance and density, etc. The priority of certain features may change during the process. For instance, one can easily categorize different letters according to the feature of shape at the very beginning, whereafter the accuracy can be further optimized by involving other features in. Following this idea, an deterministic annealing scheme is applied to control the fuzziness, which progressively decreases the covariances by σ 2 ϵ σ 2 from a large value, where ϵ is a constant. In the early stage of iterations, the two point sets have the biggest difference, estimation based on Euclidean distance is less accurate, yet the local geometric feature discrepancy and the intensity information are relative strong and stable. The unreliable global correspondence is filtered due to the property of negative nature exponential function. At the final stage of iterations, x i and y j are very similar, a direct estimation using the global geometric feature is desirable. In addition, the statuses of these features interchange as σ 2 is small, leading to binary-like correspondences.
Furthermore, the correspondences estimation of the MGMM is a fuzzy-to-binary process, by which the correspondences are able to be improved gradually and continuously during the optimization without jumping around in the space of binary permutation matrices.

2.3. Geometric Constraint for L2E Based Energy Optimization

Based on the reliable correspondence estimated by the MGMM, the target coordinate is determined by x i * = j = 1 n p i j x j . Our next concern is to formulate a criterion, by which a reasonable position T ( y i ) of y i is determined. This position in turn improves the accuracy of the correspondence estimation in the next iteration interlockingly. In this subsection, we formulate the L2E based energy function with geometrical constraint. Followed by sketching out the related concepts and techniques, i.e., the L 2 -minimizing estimate ( L 2 E ) and the motion coherent based geometric constraint.
The unknown coordinates of the source point set to be transformed is estimated by using the L2E based energy function with geometric constraint which is written as:
Q ( T , ρ 2 ) = L 2 E ( T , ρ 2 ) + λ 2 R ( T ) ,
where T is the non-rigid transformation, ρ 2 is the parameter of the density models we used to represent the deviation between the source and target point sets, L 2 E ( · ) is the L2E estimator and R ( · ) is the geometric constraint on T based on the Tikhonov regularization theory [33], the constant λ controls the strength of the constraint.

2.3.1. L 2 E

L 2 E [34] is a robust estimator which minimizes the L 2 distance between densities, and is particularly appropriate for analyzing massive datasets where outlier rejecting is impractical.
Suppose we have a density model p ( z | θ ) , our goal is to minimize the estimate of L 2 distance with respect to θ as θ * = arg min θ p ( z | θ ) p ( z | θ 0 ) 2 d z , where the true parameter θ 0 is unknown. After omitting the constant p ( z | θ 0 ) 2 dz, parameter θ is estimated by minimizing the L 2 E criterion as:
θ L 2 E * = arg min θ p ( z | θ ) 2 d z 2 m i = 1 m p ( z | θ ) .
The robustness of the L2E estimator can be shown by comparing against the maximum log-likelihood estimator (MLE). Consider a sample of size 500 from a normal distribution N ( 0 , 1 ) which stands for the inliers, five normal distributed samples N ( 5 , 1 ) in a tendency of increasing size (e.g., 25, 50, 100, 150, 250) act as outliers. The estimated means are shown in Figure 3. L 2 E has a stable global minimum at approximately 0, as well as a local minimum at approximately 5 which becomes deeper as the number of outliers increases. This is reasonable since the outliers are from N ( 5 , 1 ) . By contrast, MLE is not resistant to outliers, since the global minimum deviates more under heavier contamination.

2.3.2. Motion Coherent Based Geometric Constraint

The key constraint of a rigid transformation is that all distances are preserved. However, once non-rigidity is allowed, there are an infinite number of ways to map one point set onto another. An appropriate constraint is necessary to solve this ill-posed problem. To this end, the motion coherent based geometric constraint is introduced into our method.
In motion perception, there are a number of important phenomena involving coherence. Velocity coherence is a particular way of imposing smoothness on the underlying transformation. The concept of motion coherence is proposed in the motion coherence theory [35] which is intuitively interpreted as that points close to one another tend to move coherently. Examples of velocity fields with two different levels of motion coherence for two different point correspondences are illustrated in Figure 4. Since our focus is on its application in remote sensing image registration, we will not drill-down further into the theoretical model but directly write the geometric constraint R ( T ) in the form as:
R ( T ) = T 2
It implies that we are imposing motion coherence among the SIFT feature points, and discouraging the undesired transformation, e.g., as shown in Figure 4c, over the whole warping plane. With a large λ , the constraint produces globally smooth transformation, while it produces more arbitrary transformation with small values.

2.4. Main Process

Given a sensed image I s of size N w × N h and a reference image I r of size N w × N h , our goal is to obtain a transformed image I t of size N w × N h by recovering the underlying geometric transformation from I s to I r . The main process of our method consists of three steps as shown in Figure 5: (i) extraction of the feature point sets, (ii) registration of the extracted feature point sets, and (iii) image Transformation and resampling, where the second step comprises an alternating two-step process: correspondence estimation and transformation updating of the feature point sets.

2.4.1. Extraction of the Feature Point Sets

SIFT algorithm is employed for feature point sets extraction. By { Y m × d e , V m × d s } and { X n × d e , U n × d s } we denote the feature point sets and SIFT descriptor sets extracted from the sensed and reference images, respectively, where Y = { y 1 , y 2 , . . . , y m } T and X = { x 1 , x 2 , . . . , x n } T are the d e dimensional geometrical coordinates of the source and target point sets , and V = { v 1 , v 2 , . . . , v m } T and U = { u 1 , u 2 , . . . , u n } T are the intensity information, namely the d s dimensional SIFT descriptors for X and Y , respectively. Note that T ( Y ) (initial T ( Y ) = Y ) denotes the transformed set Y in each iteration, and a general derivation is realized by using different suffixes n and m where m = n actually.

2.4.2. Feature Point Set Registration

Correspondence estimation: The pairwise global and local geometric structure discrepancies G and L , and the SIFT distance S are obtained by Equations (4)–(6), respectively. The posterior probability matrix, which is also known as the correspondence matrix is then written as:
p i j * = ( 1 s ˜ i j ) exp x i y j 2 2 σ 2 + α 2 b = 1 B [ h i ( b ) h j ( b ) ] 2 h i ( b ) + h j ( b ) k = 1 m ( 1 s ˜ i k ) exp x i y k 2 2 σ 2 + α 2 b = 1 B [ h i ( b ) h k ( b ) ] 2 h i ( b ) + h k ( b ) + ( 2 π σ 2 ) d e 2 ζ n ,
where P * is of size m × n . s ˜ i j , s ˜ i j [ 0 , 1 ] is the rescaled SIFT distance, σ 2 are the equal covariances of the MGMM, constant ζ is the outlier weighting, and B bins are used to construct the SC log-polar coordinate. Hence, the estimated corresponding set is obtained by
X * = P * X .
After computing X * , the non-rigid transformation is modeled by the current correspondence X * and the source point set Y .
Transformation updating: The Resiz representation theorem states that if ψ is a bounded linear function on a Hilbert space H , then there is a unique vector ν in H , making ψ t = t , ν for all ψ H . Thus the cumbersome mapping problem reduces to find the unique vector ν . This motivates us to model the non-rigid transformation T by requiring it to lie within the reproducing kernel Hilbert space (RKHS).
We first define a RKHS H by choosing a positive definite kernel, here we adopt the Gaussian kernel, which is in the form Θ ( y i , y j ) = exp ( 1 2 β 2 y i y j 2 ) , where β is a constant to control the spatial smoothness and Θ is of size m × m . According to the representation theorem, the optimal transformation function T takes the form
T ( Y ) = Y + Θ W ,
where W m × d e = { w 1 , w 2 , . . . , w m } T is the coefficient matrix. Hence, the minimization over energy function Equation (8) boils down to finding a finite coefficients matrix W .
The non-rigid transformation T ( Y ) is equivalent to the initial position plus a displacement function i.e., T ( Y ) = Y + V ( Y ) . It is worth nothing that R ( T ) and R ( V ) are exactly the same, i.e., R ( T ) = R ( Y + V ( Y ) ) = R ( V ) , since R ( T ) is invariant under affine transformations. In this paper, we solve for V instead of T . The complete energy function Equation (8) takes the form as:
Q ( W , ρ 2 ) = 1 2 d e ( π ρ ) d e 2 2 m i = 1 n 1 ( 2 π ρ 2 ) d e 2 exp x i * j = 1 m Θ ( y i , y j ) w j 2 2 ρ 2 + λ t r ( W Θ W ) ,
where constant λ controls the strength of the constraint, t r ( · ) denotes the trace of a matrix. Taking derivative of Equation (14) with respect to coefficient matrix W , we obtain:
Q W = 2 Θ [ U ( E 1 ) ] m ρ 2 ( 2 π ρ 2 ) d e 2 + 2 λ W ,
where U m × d e = Θ W X * , E m × 1 = exp { d i a g ( U U T ) / 2 ρ 2 } , 1 1 × d e is a row vector with all ones, d i a g ( · ) is the diagonal of a matrix, symbols ⊙ and ⊗ denote the Hadamard product and the Kronecker product.
A gradient-based numerical optimization technique is employed to solve this optimization problem based on Equation (15). Moreover, non-rigid point set registration has high degrees of flexibility, trapping the optimization process by local extrema. Therefore, we improve the convergence by adopting another deterministic annealing on the covariances ρ 2 . It initializes with a large value for ρ 2 which progressively decreases by ρ 2 = κ ρ 2 , where κ is a constant. And the energy function Equation (14) will converge to an optimal solution eventually.
After updating the coordinates of the source point set by Y T ( Y ) , we anneal the covariances of the MGMM by σ 2 ϵ σ 2 , then return to correspondence estimation and continue the registration process until the maximum iteration number is reached. The pseudo-code of the feature point sets registration of our method is outlined in Algorithm 1. In addition,
Algorithm 1: Feature matching using multiple features for remote sensing registration
Remotesensing 09 00581 i001

2.5. Image Transformation and Resampling

Once we have the transformed sensed feature point set Y ^ , a mapping function can be constructed based on the corresponding set, i.e., I = { y i , y ^ i } i = 1 m and thus to register the images. There are two choices. (i) Forward approach: directly transforming the sensed image I s using the mapping function. (ii) Backward approach: determining the transformed image I t from I s using the grid of the reference image I r and the inverse of the mapping. Since (i) is complicated to implement, as it can produce holes and/or overlaps in the output image due to the discretization and rounding, we use the backward approach for image transformation, and the flowchart is shown in Figure 6.
Let Y ^ be the source, Y the target, the TPS kernel is defined as K ( y ^ i , y ^ j ) = y ^ i y ^ j 2 log y ^ i y ^ j , thus, the TPS transformation model is obtained by
P = K Q Q T O 1 Y 0 0 0
where the TPS model P is of size ( m + 3 ) × 3 , O is a 3 × 3 matrix of zeros and Q is the m × 3 matrix with the i t h row denotes ( 1 , y ^ i a , y ^ i b ) , where y ^ i a and y ^ i b indicate the coordinates of y ^ i .
A regular grid Γ Z × 2 r = { γ 1 r , γ 2 r , . . . γ Z r } T is obtained by a pixel-by-pixel indexing process on the reference image I r , where Z = N w × N h . Let grid Γ r be the source point set, P the TPS transformation model, the transformed grid is obtained by first computing
Γ ^ Z × 3 r = K Q P ,
then restoring the dimension of the grid to 2 by Γ ^ r Γ ^ ( · , 1 ) r Γ ^ ( · , 2 ) r , where the Z × m kernel K = γ i r y ^ j 2 log γ i r y ^ j , Q is the Z × 3 matrix with the i t h row denotes ( 1 , ( γ i r ) a , ( γ i r ) b ) and Γ ^ ( · , i ) r denotes the i t h column of Γ ^ r . Let Γ s be the grid obtained on I s , we have
Γ ^ * = Γ ^ r Γ s .
Finally, the transformed image I t is obtained by getting intensities from the sensed image I s based on Γ ^ * , and setting the rest of pixels to black. Note that the bicubic interpolation is used to improve the smoothness of I t , to be more precisely, the intensities of each pixel in I t is determined by summing the weighted neighbor pixel intensities within a 4 × 4 window.

2.6. Method Analysis

Figure 7 demonstrates the advantage of the MGMM by comparing against the single feature based method, e.g., CPD. At the first iteration, we can see that the initial estimated coordinates by CPD are the regional center of masses, while they are close to the real target coordinates using our method. This helps us recover a good initial pose for aligning the two point sets. Iterating CPD even more, the estimated coordinates are still regional center of masses, and a much better source coordinates obtained by our method. Finally, our method outperforms with less iterations. The obvious difference within red rectangles results from that the estimating ranges i.e., covariances σ 2 of the GMMs in CPD wane to too small a magnitude whereas the distance between points are relatively large. The meaningless probabilities P * provide a bad X * , which can not lead T ( Y ) anymore. Moreover, we see that many green circles are missing from the red rectangle in CPD, and collapsed to ( 0 , 0 ) , in the contrary, all the green circles correctly distribute to where they should be, yet only the outliers have no circles in our method.
To examine by which the reliability of our method is mainly contributed, we also additionally compare the accuracy of the feature point sets registration using our method against GMMREG [36], which considers the registration problem as one of the aligning two Gaussian mixture models, and estimates the transformation by minimizing the L 2 distance between the two models. As shown in Figure 8, though the L 2 E estimator is employed, there exists obvious deviations in the result of GMMREG, while our method shows feasible performance. This implies that the accuracy of correspondence and transformation is mainly improved by the strategy based on the complementation and combination of multiple image features using the MGMM.

2.6.1. Computational Complexity

The local geometric feature discrepancy is measured by using the SC, it takes O ( m ) time to build the SC for one point, which therefore takes O ( m 2 ) time to build for m points. Note that the Hungarian algorithm which is used to solve the problem of bipartite graph matching with worst-cost time O ( m 3 ) is not included in our method. The Matlab Optimization toolbox, e.g., the Matlab function fminunc, which implicitly uses the BFGS Quasi-Newton method with a mixed quadratic and cubic line search procedure,. is used for solving the numerical optimization method, and the total complexity is approximately O ( m 3 + 2 m 2 ) . Overall, the time complexity of our method is O ( m 3 ) . For storing kernel Θ , the space complexity of our method is O ( m 2 ) .

2.6.2. Parametric Setting

The default threshold 1.5 is used to extract the putative corresponding set { x i , y i } i = 1 n as well as the SIFT descriptors { u i , v i } i = 1 n . The bins of the SC is set to default, i.e., B = 12 × 5 . The constant α controls the trade-off between the global and local geometric feature discrepancy. The constant ζ weights the uniform distribution 1 n for outlier dealing. The two annealing schemes, i.e., gradually update σ 2 ϵ σ 2 and ρ 2 κ ρ 2 , which deal with the non-convexity play important role in our method. A slow annealing, e.g., close to 1, is always preferred without the consideration of the efficiency issue. The constant t σ and t ρ determine the maximum iteration numbers of our method and the function fminunc in Matlab. β determines the neighborhood size of the source point set. λ controls the influence of the geometric constraint on the transformation T . We set α = 10 , ζ = 0.3 , ϵ = 0.9 , κ = 0.75 , β = 2 and λ = 3 , and initializing σ 2 and ρ 2 to 1 and 0.05 , respectively. Due to the iterations need termination conditions, we use the function fminunc in Matlab with the options: { M a x I t e r = 50 } , namely t ρ = 50 , and set t σ = 100 .

3. Experiments and Results

SIFT [11], SURF [13], CPD [25], GLMDTPS [30], RSOC [21], totally five state-of-the-art methods are compared against our method in the experiments. These methods can mainly categorize into three types based on the methods of feature extraction. Type 1: using open source V L F e a t toolbox with default threshold 1.5 , i.e., SIFT, CPD, GLMDTPS and ours. Type 2: using open source VLFeat toolbox with specific built-in setting, i.e., RSOC. Type 3: using Matlab open source O p e n S U R F function with default setting, i.e., SURF. Since both the feature matching and image registration are considered in our method, we fairly design three series of experiments. (I) Due to the employment of the same feature point sets, quantitative comparison on feature matching is carried out on methods of Type 1 using the precision ratio (PR). However, since no inlier set is outputted from GLMDTPS, the compared methods are SIFT, CPD and ours. (II) Quantitative comparison and Qualitative demonstration on image registration are carried out on all the methods using the root of mean square error (RMSE), mean absolute error (MAE) and standard deviation (SD). (III) By using the different datasets, quantitative and qualitative demonstration on feature matching and image registration are carried out to examine the availability and robustness of our method using the PR, RMSE, MAE and SD. Three datasets are used which are: (i) 100 pairs of UAV images with horizontal viewpoint changes; (ii) 150 pairs of UAV images pairs with vertical viewpoint changes; and (iii) 50 pairs of satellite images registration with horizontal and vertical viewpoint changes. The experimental design is shown in Table 1. All experiments are tested on a PC with 2.20 GHz Intel CPU and 8 GB memory.
Generally, the downsampled image size is selected based on the performance and accuracy requirements of the underlying system. High sampling rates might lead to more locally accurate at the cost of the far more time-consuming, whereas the globality is our main concern. Therefore, a procedure of downsample, which is based on the bicubic interpolation using Matlab i m r e s i z e function, is carried out on all image pairs. The downsampled images have a resolution in the range from 640 × 450 to 1100 × 850 pixels, thus the more compact feature points can therefore be extracted. The number of the extracted feature points for Type I to III methods are shown in Table 2.

3.1. Evaluation Criterion

P r e c i s i o n = T P T P + F P ,
where T P denotes true positive, F P denotes false positive, 0 P r e c i s i o n 1 . To obtain the estimated inlier set I , different approaches are used based on the characteristic of each method. For SIFT, it is just an array from 1 to n; for CPD and our method, it is obtained by:
I = { i : ( P T 1 ) i > ϱ } ,
where the threshold is set to ϱ = 0.75 empirically.
The RMSE, MAE and SD are used to quantify the registration accuracy. We manually determine at least 10 pairs of landmarks between the sensed image and the reference image as ground-truth, and all the landmarks are well-distributed and selected on the easily identified places around the interest areas. The related formulations and the definitions in statistics are as follows:
R M S E = 1 m i = 1 m ( y i t x i t ) ,
The root mean square error is a frequently used measure of the distance between selected landmark and its corresponding point actually located, where m is the total number of the selected landmark, and y i t is the landmark that corresponds to x i t ;
M A E = i = 1 m | y i t x i t | m ;
the mean absolute error is a quantity used to measure how close the landmark are to its corresponding point;
S D = 1 m i = 1 m d ( x i t , y i t ) R M S E 2 ,
the standard deviation informally measures how far the distance of a landmark pair are spread out from its RMSE, where d ( · , · ) denotes the distance.

3.2. Results on Feature Matching

The quantitative results are shown in Table 3. Feature matching results on eight typical image pairs are demonstrated in Figure 9, as well as the quantitative results of the eight selected image pairs.
The extraction of the SIFT feature point sets is based on the intensity information, or, to be more precisely, the scale-space extrema. However, for all the data which contain viewpoint changes, the objects captured in one image may be missing or distorted non-rigidly from another image, which further lead to false matching. Based upon the SIFT putative corresponding set, CPD estimates the correspondence using Euclidean distance, a single global geometric structure discrepancy for correspondence estimation. The motion coherent based geometric constraint is employed to regularize the displacement field between the point sets. However, as we mentioned before, the regular Gaussian mixture model is incapable of dealing with mixture features, and for Euclidean distance, p ( x a | y j ) = p ( x b | y j ) if x a y j 2 = x b y j 2 even when the neighborhood structure of x a and x b are totally different. In our method, both the intensity information and the global and local geometric structure discrepancies are considered, the complementation and combination of these features provide a reliable correspondence estimation. We summarize the key components employed in these three method as: (I) intensity information discrepancy; (II/III) global/local geometric structure discrepancy; (IV) geometric constraint. Thus, SIFT use only (I), CPD employs (II) and (IV), and our method adopts both. Therefore, the result shows a increasing tendency on PR from left to right.

3.3. Results on Image Registration

Quantitative comparison using the mean RMSE, MAE and SD are shown in Table 4. The transformed images and 5 × 5 checkboards on eight typical image registration examples from dataset (i) and (ii) are shown (four examples per dataset) in Figure 10 and Figure 11, respectively.
A low computation complexity is one of the advantages for registering image based on the extracted feature points. However, its registration accuracy might be poor which in general cannot obtain accurate dense correspondences. Therefore, the number of the extracted feature points plays a crucial role to alleviate this problem. In this experiment, SURF and RSOC, which extract relative small number of feature points, are sensitive to false matches. Since RSOC employs robust graph matching technique to remove dubious matches, its performance remains relative high. By contrast, the result shows that directly yielding the transformed images by the putative corresponding set of SIFT with a default threshold is not a good idea when viewpoint changes exist. Fewer and more reliable feature points can generate by setting a high threshold, which in turn limits the registration accuracy based on more sparse correspondences. GLMDTPS employs mixture features as well. The global feature is the pairwise Euclidean distance, and the local feature first respectively generates two index matrices of K nearest neighbors for two point sets. After translating the two currently computing points together, the local distance is obtained by summing the squared Euclidean distance between the i t h neighbors. The defect of the employed features is that it is sensitive to outliers and similar neighborhood structure. Unlike the contour point set, the feature point set extracted in remote sensing images distributes irregularly, the local distance employed by GLMDTPS is therefore ambiguous, resulting in more dubious estimation in local area.

3.4. Reliability and Availability Examination of Our Method

In this series of experiment, we examine the reliability and availability of our method using Satellite Image with Horizontal and Vertical Viewpoint Changes. The PR, RMSE, MAE and SD are shown in Table 5, seven typical images are selected to show the image registration results in Figure 12. We see that our method maintains accurate alignments in all experiments. Which means that our proposed method can successfully handle the satellite image registration problem in most time.

4. Conclusions

In this paper, we proposed an accurate method for remote sensing image registration with different viewpoint. The main contributions of this paper are considered as follows. (i) The MGMM is constructed for simultaneously dealing with different types of image features. (ii) Two types of features are using to form a feature complementation, i.e., the Euclidean distance and shape context to measure the global and local geometric structure discrepancies, and the SIFT distance which is endowed with the intensity information are combined and substituted into the MGMM by which the reliable correspondence is obtained. (iii) To prevent the ill-posed problem, a geometric constraint term based on coherent velocity field is introduced into the L2E-based energy function and achieves accurate transformation updating. Experiments on three series of remote sensing images obtained from the unmanned aerial vehicle (UAV) and Google Earth with different viewpoint demonstrate that our method shows best registration performances against five state-of-the-art methods.

Acknowledgments

The authors wish to thank David G. Lowe, Herbert Bay, Andriy Myronenko, Zhaoxia Liu and Yang Yang for providing their implementation source codes and test data sets. This greatly facilitated the comparison experiments. This work was supported by (i) National Nature Science Foundation of China [41661080]; (ii) Scientific Research Foundation of Yunnan Provincial Department of Education [2017TYS045]; (iii) Doctoral Scientific Research Foundation of Yunnan Normal University [01000205020503065]; (iv) National Undergraduate Trainking Program for Innovation and Entrepreneurship [201610681002].

Author Contributions

Yang Yang, Anning Pan and Su Zhang developed the method; Kun Yang, Anning Pan and Su Zhang conceived and designed the experiments; Kun Yang, Su Zhang and Anning Pan performed the experiments and analyzed the data; Sim Heng Ong and Haolin Tang helped technology implementation of the method; Su Zhang and Yang Yang wrote the paper. All the authors reviewed and provided valuable comments for the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zitov, B.; Flusser, J. Image registration methods: A survey. Image Vis. Comput. 2003, 21, 977–1000. [Google Scholar] [CrossRef]
  2. Brown, L. A survey of image registration techniques. ACM Comput. Surv. 1992, 41, 325–376. [Google Scholar] [CrossRef]
  3. Maintz, J.; Viergever, M. A survey of medical image registration. Med. Image Anal. 1998, 2, 1–36. [Google Scholar] [CrossRef]
  4. Wang, X.; Li, Y.; Wei, H.; Liu, F. An asift-based local registration method for satellite imagery. Remote Sens. 2015, 7, 7044–7061. [Google Scholar] [CrossRef]
  5. Liu, S.; Tong, X.; Chen, J.; Liu, X.; Sun, W.; Xie, H.; Chen, P.; Jin, Y.; Ye, Z. A Linear Feature-Based Approach for the Registration of Unmanned Aerial Vehicle Remotely-Sensed Images and Airborne LiDAR Data. Remote Sens. 2016, 8, 82. [Google Scholar] [CrossRef]
  6. Sedaghat, A.; Mokhtarzade, M.; Ebadi, H. Uniform Robust Scale-Invariant Feature Matching for Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2011, 49, 4516–4527. [Google Scholar] [CrossRef]
  7. Jensen, J. Introductory Digital Image Processing; Prentice Hall: Upper Saddle River, NJ, USA, 2004. [Google Scholar]
  8. Mikolajczyk, K.; Chmid, C. A performance evaluation of local descriptors. In Proceedings of the 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Madison, WI, USA, 18–20 June 2003; Volume 2. [Google Scholar]
  9. Fan, B.; Wu, F.; Hu, Z. Rotationally Invariant Descriptors Using Intensity Order Pooling. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 2031–2045. [Google Scholar] [CrossRef] [PubMed]
  10. Harris, C. A combined corner and edge detector. In Proceedings of the 4th Alvey Vision Conference, Manchester, UK, 31 August–2 September 1988; Volume 1988, pp. 147–151. [Google Scholar]
  11. Lowe, D. Object recognition from local scale-invariant features. In Proceedings of the IEEE Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; Volume 2, p. 1150. [Google Scholar]
  12. Lowe, D. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 21, 91–110. [Google Scholar] [CrossRef]
  13. Bay, H.; Ess, A.; Tuytelaars, T.; Van Gool, L. Speeded-Up Robust Features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  14. Li, Q.; Wang, G.; Liu, J.; Chen, S. Robust Scale-Invariant Feature Matching for Remote Sensing Image Registration. IEEE Geosci. Remote Sens. Lett. 2009, 6, 287–291. [Google Scholar]
  15. Ke, Y.; Sukthankar, R. PCA-SIFT: A more distinctive representation for local image descriptors. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004, Washington, DC, USA, 27 June–2 July 2004; Volume 2, pp. 506–513. [Google Scholar]
  16. Dellinger, F.; Delon, J.; Gousseau, Y.; Michel, J. SAR-SIFT: A SIFT-Like Algorithm for SAR Images. IEEE Trans. Geosci. Remote Sens. 2015, 53, 453–466. [Google Scholar] [CrossRef]
  17. Sedaghat, A.; Ebadi, H. Remote Sensing Image Matching Based on Adaptive Binning SIFT Descriptor. IEEE Trans. Geosci. Remote Sens. 2015, 53, 5283–5293. [Google Scholar] [CrossRef]
  18. Liu, F.; Bi, F.; Chen, L.; Shi, H. Feature-Area Optimization: A Novel SAR Image Registration Method. IEEE Geosci. Remote Sens. Lett. 2016, 13, 242–246. [Google Scholar] [CrossRef]
  19. Goncalves, H.; Corte-Real, L.; Goncalves, J. Automatic Image Registration Through Image Segmentation and SIFT. IEEE Trans. Geosci. Remote Sens. 2011, 49, 2589–2600. [Google Scholar] [CrossRef]
  20. Gong, M.; Zhao, S.; Jiao, L.; Tian, D. A Novel Coarse-to-Fine Scheme for Automatic Image Registration Based on SIFT and Mutual Information. IEEE Trans. Geosci. Remote Sens. 2014, 52, 4328–4338. [Google Scholar] [CrossRef]
  21. Liu, Z.; An, J.; Jing, Y. A simple and robust feature point matching algorithm based on restricted spatial or derconstraints for aerial image registration. IEEE Trans. Geosci. Remote Sens. 2012, 50, 514–527. [Google Scholar] [CrossRef]
  22. Ma, J.; Zhao, J.; Tian, J.; Bai, X.; Tu, Z. Regularized vector field learning with sparse approximation for mismatch removal. Pattern Recognit. 2013, 46, 3519–3532. [Google Scholar] [CrossRef]
  23. Ma, J.; Zhao, J.; Tian, J.; Yuille, A.; Tu, Z. Robust point matching via vector field consensus. IEEE Trans. Image Proc. 2014, 23, 1706–1721. [Google Scholar]
  24. Fischler, M.; Bolles, R. Random Sample Consensus: A Paradigm for Model Fitting with Applications to Image Analysis and Automated Cartography. Commun. ACM 1980, 24, 381–395. [Google Scholar] [CrossRef]
  25. Myronenko, A.; Song, X. Point set registration: coherent point drift. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 2262–2275. [Google Scholar] [CrossRef] [PubMed]
  26. Liu, H.; Yan, S. Common visual pattern discovery via spatially coherent correspondences. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognit, San Francisco, CA, USA, 13–18 June 2010; pp. 1609–1616. [Google Scholar]
  27. Zhang, Z.; Yang, M.; Zhou, M.; Zeng, X. Simultaneous remote sensing image classification and annotation based on the spatial coherent topic model. In Proceedings of the IGARSS 2014—2014 IEEE International Geoscience and Remote Sensing Symposium, Quebec, QC, Canada, 13–18 July 2014; pp. 1698–1701. [Google Scholar]
  28. Ma, J.; Qiu, W.; Zhao, J.; Ma, Y. Robust L2E Estimation of Transformation for Non-Rigid Registration. IEEE Trans. Signal Process. 2015, 63, 1115–1129. [Google Scholar] [CrossRef]
  29. Ma, J.; Zhou, H.; Zhao, J.; Gao, Y.; Jiang, J.; Tian, J. Robust Feature Matching for Remote Sensing Image Registration via Locally Linear Transforming. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6469–6481. [Google Scholar] [CrossRef]
  30. Yang, Y.; Ong, S.; Foong, K. A robust global and local mixture distance based non-rigid point set registration. Pattern Recognit. 2015, 48, 156–173. [Google Scholar] [CrossRef]
  31. Belongie, S.; Malik, J.; Puzicha, J. Shape Matching and Object Recognition Using Shape Contexts. IEEE Trans. Pattern Anal. 2002, 24, 509–522. [Google Scholar] [CrossRef]
  32. Kortgen, M.; Park, G.; Novotni, M.; Klein, R. 3D shape matching with shape context. In Proceedings of the 7th Central European Seminar on Computer Graphics, Budmerice, Slovakia, 22–24 April 2003; pp. 22–24. [Google Scholar]
  33. Tikhonov, A.N.; Arsenin, V.Y. Solutions of Ill-Posed Problems; Winston & Sons Inc.: Washington, DC, USA, 1977. [Google Scholar]
  34. Scott, D. Parametric statistical modeling by minimum integrated sqaure error. Technometrics 2001, 43, 274–285. [Google Scholar] [CrossRef]
  35. Yuille, A.; Grzywacz, N. A Mathematical Analysis of the Motion Coherence Theory. Int. J. Comput. Vis. 1989, 3, 155–175. [Google Scholar] [CrossRef]
  36. Jian, B.; Vemuri, B. Robust Point Set Registration Using Gaussian Mixture Models. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 1633–1645. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Illustration of the SC. (a) and (b): diagrams of log-polar histogram bins centered at x i and y j used in computing the shape contexts. (c) and (d): each shape context e.g., h i or h j is a log-polar histogram of the coordinates of the rest of the point set measured using the centered point as the origin, where darker denotes larger value.
Figure 1. Illustration of the SC. (a) and (b): diagrams of log-polar histogram bins centered at x i and y j used in computing the shape contexts. (c) and (d): each shape context e.g., h i or h j is a log-polar histogram of the coordinates of the rest of the point set measured using the centered point as the origin, where darker denotes larger value.
Remotesensing 09 00581 g001
Figure 2. Illustration of how the SIFT feature descriptor is obtained. (a): Repeatedly convolving initial image with Gaussians to produce the set of scale space images. (b): Subtracting adjacent Gaussian images to produce the difference-of-Gaussian images. (c): Extrema detection by comparing a pixel (marked with red circle) to its 26 neighbors in 3 × 3 regions at the current and adjacent scales (marked with green circles). (d): Feature descriptor creation by computing the gradient magnitude and orientation at each image sample point. a Gaussian window indicated by the overlaid circle is used as weighting. (e) Orientation histograms accumulation by summarizing the contents over 4 × 4 subregions (a 2 × 2 subregions is shown for convenience), with the length of each arrow corresponding to the sum of the gradient magnitudes near that direction within the region, generating 4 × 4 × 8 descriptors, where 8 is the number of directions.
Figure 2. Illustration of how the SIFT feature descriptor is obtained. (a): Repeatedly convolving initial image with Gaussians to produce the set of scale space images. (b): Subtracting adjacent Gaussian images to produce the difference-of-Gaussian images. (c): Extrema detection by comparing a pixel (marked with red circle) to its 26 neighbors in 3 × 3 regions at the current and adjacent scales (marked with green circles). (d): Feature descriptor creation by computing the gradient magnitude and orientation at each image sample point. a Gaussian window indicated by the overlaid circle is used as weighting. (e) Orientation histograms accumulation by summarizing the contents over 4 × 4 subregions (a 2 × 2 subregions is shown for convenience), with the length of each arrow corresponding to the sum of the gradient magnitudes near that direction within the region, generating 4 × 4 × 8 descriptors, where 8 is the number of directions.
Remotesensing 09 00581 g002
Figure 3. The robustness comparison between L 2 E and M L E . We estimate the mean of a normally distributed sample N ( 0 , 1 ) with contaminations of extra samples N ( 5 , 1 ) . The outlier to the inlier ratios are 5 % , 10 % , 20 % , 30 % and 50 % . The vertical dash lines indicate the extrema. L 2 E has a global minimum at approximately 0 and a local minimum at approximately 5, both of which conform to the inlier and outlier distributions, respectively. However, the deviation of MLE increases as the ratio grows.
Figure 3. The robustness comparison between L 2 E and M L E . We estimate the mean of a normally distributed sample N ( 0 , 1 ) with contaminations of extra samples N ( 5 , 1 ) . The outlier to the inlier ratios are 5 % , 10 % , 20 % , 30 % and 50 % . The vertical dash lines indicate the extrema. L 2 E has a global minimum at approximately 0 and a local minimum at approximately 5, both of which conform to the inlier and outlier distributions, respectively. However, the deviation of MLE increases as the ratio grows.
Remotesensing 09 00581 g003
Figure 4. Illustration of the velocity field. (a): two given point pairs. (b): a coherent velocity field. (c): a velocity field that is less coherent.
Figure 4. Illustration of the velocity field. (a): two given point pairs. (b): a coherent velocity field. (c): a velocity field that is less coherent.
Remotesensing 09 00581 g004
Figure 5. The summary of our method. (a): Extracting putative corresponding set { x i , y i } i = 1 n and SIFT descriptor set { u i , v i } i = 1 n using SIFT algorithm, and obtaining two types of discrepancies, i.e., the global and local geometric structure discrepancies G and L , and intensity information discrepancy S . (b): substituting the combined features into the optimization framework and obtaining the transformed point set Y ^ , note that a loop is included. (c): Image registration based on the backward approach.
Figure 5. The summary of our method. (a): Extracting putative corresponding set { x i , y i } i = 1 n and SIFT descriptor set { u i , v i } i = 1 n using SIFT algorithm, and obtaining two types of discrepancies, i.e., the global and local geometric structure discrepancies G and L , and intensity information discrepancy S . (b): substituting the combined features into the optimization framework and obtaining the transformed point set Y ^ , note that a loop is included. (c): Image registration based on the backward approach.
Remotesensing 09 00581 g005
Figure 6. Illustration of the image transformation and resampling. (a) Computing a TPS transformation model P which maps Y ^ back onto Y , meanwhile, constructing a grid Γ r which is of the same size as I r ; (b) Mapping Γ r using P , obtaining Γ ^ r ; (c) Limiting indexes to boundaries of Γ ^ r Γ s ; (d) Getting intensities from I s within boundaries to generate I t , each pixel in I t is determined by the bicubic interpolation.
Figure 6. Illustration of the image transformation and resampling. (a) Computing a TPS transformation model P which maps Y ^ back onto Y , meanwhile, constructing a grid Γ r which is of the same size as I r ; (b) Mapping Γ r using P , obtaining Γ ^ r ; (c) Limiting indexes to boundaries of Γ ^ r Γ s ; (d) Getting intensities from I s within boundaries to generate I t , each pixel in I t is determined by the bicubic interpolation.
Remotesensing 09 00581 g006
Figure 7. Demonstration of the advantage of the MGMM. Red *: the target feature point set X . Green ∘: the estimated corresponding point set X * . Blue +: the source feature point set T ( Y ) . Upper row and lower row: registration process of CPD and our method. X and Y are extracted from a remote sensing image pair.
Figure 7. Demonstration of the advantage of the MGMM. Red *: the target feature point set X . Green ∘: the estimated corresponding point set X * . Blue +: the source feature point set T ( Y ) . Upper row and lower row: registration process of CPD and our method. X and Y are extracted from a remote sensing image pair.
Remotesensing 09 00581 g007
Figure 8. Examination of the availability of the MGMM. Red *: the target feature point set X . Blue ∘: the source feature point set T ( Y ) . (a,b) registration result of GMMREG and our method. Obvious deviation exists in the result of GMMREG.
Figure 8. Examination of the availability of the MGMM. Red *: the target feature point set X . Blue ∘: the source feature point set T ( Y ) . (a,b) registration result of GMMREG and our method. Obvious deviation exists in the result of GMMREG.
Remotesensing 09 00581 g008
Figure 9. Feature matching demonstrations on eight typical image pairs of dataset (i) and (ii). For each image pair, rows from the first to the fourth are: the image pair, the results on feature matching of ours, SIFT and CPD, respectively. Blue lines indicate true positive and true negative, red lines indicate false positive and false negative. The PRs of three methods on each image pairs are listed from (a) to (h) as follows. Ours: 99.00, 99.29, 97.31, 97.40, 99.32, 99.62, 98.31, 98.89; SIFT: 85.68, 85.18, 45.12, 60.26, 82.43, 80.34, 82.43, 83.55; CPD: 86.86, 88.14, 85.02, 88.24, 87.22, 85.55, 89.76, 85.23.
Figure 9. Feature matching demonstrations on eight typical image pairs of dataset (i) and (ii). For each image pair, rows from the first to the fourth are: the image pair, the results on feature matching of ours, SIFT and CPD, respectively. Blue lines indicate true positive and true negative, red lines indicate false positive and false negative. The PRs of three methods on each image pairs are listed from (a) to (h) as follows. Ours: 99.00, 99.29, 97.31, 97.40, 99.32, 99.62, 98.31, 98.89; SIFT: 85.68, 85.18, 45.12, 60.26, 82.43, 80.34, 82.43, 83.55; CPD: 86.86, 88.14, 85.02, 88.24, 87.22, 85.55, 89.76, 85.23.
Remotesensing 09 00581 g009
Figure 10. Registration examples on four typical image pairs from dataset (i). The first two rows: the sensed and reference image, with the yellow crosses denoting the landmarks. Rows from the third to the end are the registration results of ours, SIFT, SURF, CPD, RSOC and GLMDTPS. For each method, the registration results are shown using two rows, with the upper row showing the transformed image, and the lower row showing the 5 × 5 checkboard. The registration errors are highlighted using the red rectangles.
Figure 10. Registration examples on four typical image pairs from dataset (i). The first two rows: the sensed and reference image, with the yellow crosses denoting the landmarks. Rows from the third to the end are the registration results of ours, SIFT, SURF, CPD, RSOC and GLMDTPS. For each method, the registration results are shown using two rows, with the upper row showing the transformed image, and the lower row showing the 5 × 5 checkboard. The registration errors are highlighted using the red rectangles.
Remotesensing 09 00581 g010
Figure 11. Registration examples on four typical image pairs from dataset (ii). The first two rows: the sensed and reference image, with the yellow crosses denoting the landmarks. Rows from the third to the end are the registration results of ours, SIFT, SURF, CPD, RSOC and GLMDTPS. For each method, the registration results are shown using two rows, with the upper row showing the transformed image, and the lower row showing the 5 × 5 checkboard. The registration errors are highlighted using the red rectangles.
Figure 11. Registration examples on four typical image pairs from dataset (ii). The first two rows: the sensed and reference image, with the yellow crosses denoting the landmarks. Rows from the third to the end are the registration results of ours, SIFT, SURF, CPD, RSOC and GLMDTPS. For each method, the registration results are shown using two rows, with the upper row showing the transformed image, and the lower row showing the 5 × 5 checkboard. The registration errors are highlighted using the red rectangles.
Remotesensing 09 00581 g011
Figure 12. Registration examples on the seven typical satellite image pairs. From (a) to (g) are: Berlin, Las Vegas, Mekong River, Paris, WuHan, Washington and Hawaii. Columns from the left to the right are: the sensed, reference and transformed images, the 5 × 5 checkboards, and the feature matching results. Blue lines indicate true positive and true negative, red lines indicate false positive and false negative.
Figure 12. Registration examples on the seven typical satellite image pairs. From (a) to (g) are: Berlin, Las Vegas, Mekong River, Paris, WuHan, Washington and Hawaii. Columns from the left to the right are: the sensed, reference and transformed images, the 5 × 5 checkboards, and the feature matching results. Blue lines indicate true positive and true negative, red lines indicate false positive and false negative.
Remotesensing 09 00581 g012
Table 1. The experimental design.
Table 1. The experimental design.
SeriesCriteriaCompared MethodsDatasets Used
IPRSIFT, CPD, Ours(i), (ii)
IIAllAll(i), (ii)
IIIAllOurs(iii)
Table 2. Numbers of the extracted feature points.
Table 2. Numbers of the extracted feature points.
Type 1Type 2Type 3
87 to 5053035 to 165
Table 3. Experimental results of series (I). Quantitative comparisons on the mean PR of Type I method are carried out. Bold fonts indicate the best results. All Units are in percentage.
Table 3. Experimental results of series (I). Quantitative comparisons on the mean PR of Type I method are carried out. Bold fonts indicate the best results. All Units are in percentage.
DatasetSIFTCPDOurs
(i)78.3090.8198.25
(ii)72.7890.1797.15
Table 4. Experimental results on series (II). Quantitative comparisons on image registration measured using the mean RMSE, MAE and SD are carried out. Bold fonts indicate the best results. All units are in pixel.
Table 4. Experimental results on series (II). Quantitative comparisons on image registration measured using the mean RMSE, MAE and SD are carried out. Bold fonts indicate the best results. All units are in pixel.
DatasetSIFTSURFCPDGLMDTPSRSOCOURS
RMSE(i)13.52877.98373.23863.01522.27371.0171
(ii)11.44667.06277.26455.59914.17431.4331
MAE(i)16.008012.28037.54047.14596.44484.0271
(ii)14.29899.841110.37789.11027.35853.9188
SD(i)14.182610.22415.85945.43534.80133.1957
(ii)11.96138.25234.36197.65315.78443.0021
Table 5. Experimental results on series (III). Quantitative tests on feature matching and image registration are carried out to examine the availability and robustness of our method using the PR, RMSE, MAE and SD. (a) to (g): results on the corresponding pairs of Figure 12. Mean: the mean of results on all image pairs of dataset (iii). For the PR, the units are in percentage, for the RMSE, MAE and SD, the units are in pixel.
Table 5. Experimental results on series (III). Quantitative tests on feature matching and image registration are carried out to examine the availability and robustness of our method using the PR, RMSE, MAE and SD. (a) to (g): results on the corresponding pairs of Figure 12. Mean: the mean of results on all image pairs of dataset (iii). For the PR, the units are in percentage, for the RMSE, MAE and SD, the units are in pixel.
(a)(b)(c)(d)(e)(f)(g)Mean
P R 96.8896.7795.9795.7795.0998.1995.5796.54
R M S E 0.83581.39631.95561.42720.77431.43420.31711.1628
M A E 2.14584.02082.90971.93172.57782.35422.29582.6051
S D 1.85622.54982.51711.80612.14292.00982.07172.1362

Share and Cite

MDPI and ACS Style

Yang, K.; Pan, A.; Yang, Y.; Zhang, S.; Ong, S.H.; Tang, H. Remote Sensing Image Registration Using Multiple Image Features. Remote Sens. 2017, 9, 581. https://doi.org/10.3390/rs9060581

AMA Style

Yang K, Pan A, Yang Y, Zhang S, Ong SH, Tang H. Remote Sensing Image Registration Using Multiple Image Features. Remote Sensing. 2017; 9(6):581. https://doi.org/10.3390/rs9060581

Chicago/Turabian Style

Yang, Kun, Anning Pan, Yang Yang, Su Zhang, Sim Heng Ong, and Haolin Tang. 2017. "Remote Sensing Image Registration Using Multiple Image Features" Remote Sensing 9, no. 6: 581. https://doi.org/10.3390/rs9060581

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop