Next Article in Journal
Mapping Cropland Burned Area in Northeastern China by Integrating Landsat Time Series and Multi-Harmonic Model
Previous Article in Journal
Outbreak of Negative Narrow Bipolar Events in Two Mid-Latitude Thunderstorms Featuring Overshooting Tops
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Multimodal Remote Sensing Image Registration Methods and Advancements: A Survey

1
School of Mathematics, Northwest University, Xi’an 710127, China
2
School of Computer Science, Shaanxi Normal University, Xi’an 710119, China
3
Department of Computing Science, University of Alberta, Edmonton, AB T6G 2E8, Canada
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(24), 5128; https://doi.org/10.3390/rs13245128
Submission received: 17 October 2021 / Revised: 10 December 2021 / Accepted: 11 December 2021 / Published: 17 December 2021

Abstract

:
With rapid advancements in remote sensing image registration algorithms, comprehensive imaging applications are no longer limited to single-modal remote sensing images. Instead, multi-modal remote sensing (MMRS) image registration has become a research focus in recent years. However, considering multi-source, multi-temporal, and multi-spectrum input introduces significant nonlinear radiation differences in MMRS images for which researchers need to develop novel solutions. At present, comprehensive reviews and analyses of MMRS image registration methods are inadequate in related fields. Thus, this paper introduces three theoretical frameworks: namely, area-based, feature-based and deep learning-based methods. We present a brief review of traditional methods and focus on more advanced methods for MMRS image registration proposed in recent years. Our review or comprehensive analysis is intended to provide researchers in related fields with advanced understanding to achieve further breakthroughs and innovations.

Graphical Abstract

1. Introduction

Image registration is the process of geometrical alignment and matching of two or more images of the same scene, acquired from different sensors, with different views and at different times [1]. In the field of remote sensing, the accuracy of image registration plays a crucial role in subsequent applications, such as image fusion, map correction and change analysis [2,3]. It is possible to obtain multi-source remote sensing image data with the rapid development of aerospace technology and remote sensing. Comprehensive utilization of multi-source remote sensing data has been widely used to realize the uniqueness and complementarity of different remote sensing images, in order to acquire images containing more information.
The registration of MMRS images still has difficulties in applications due to significant geometric distortion and nonlinear intensity differences between these images. However, it is necessary to integrate these images for earth observation applications [4]. For instance, different sensors capture optical and synthetic aperture radar (SAR) images, and different imaging mechanisms produce distinct characteristics of an area. Therefore, fusing two types of images is more conducive to representing a given area. Simultaneously, when encountering emergencies, such as weather disasters, only SAR images are useful since they can work during both day and night and see-through cloud and fog to capture images. In this case, there is an inevitable problem in combining traditional optical images with a currently acquired SAR image to analyze the imaged area [5].
Although significant advances have been made in automatic image registration technologies in the past few decades, due to the high-performance requirements of MMRS image registration, scholars strive to propose technologies advancing the state-of-the-art to address the problems of geometric differences and time efficiency.
The field of remote sensing can be roughly divided into remote sensing image acquisition technology and remote sensing information processing technology. Photogrammetry technology that mainly relies on aerospace and ground imaging platforms is different from the remote sensing technology that relies on satellite platforms. There are also differences in imaging bands and imaging methods. Photogrammetry is mainly to obtain accurate geographic information from remote sensing, and its application also has the above-mentioned problems. Therefore, remote sensing image acquisition technology and close-range application of photogrammetry cannot benefit from this survey.
Existing surveys on remote sensing image registration mostly include single-modal general registration frameworks and matching methods, while MMRS image registration, as a very important branch, only occupies a small part of the article [1,6,7,8]. A large part of the multi-modal image processing work is focused on medical image registration [9,10,11], and few papers specifically review the part of multi-modal images in remote sensing image processing. Common MMRS images include cross-temporal, cross-season, optical to SAR, optical to infrared, optical to Light Detection and Ranging (LIDAR), map to visible, etc. In the existing literature, research on optical-SAR [12,13,14] and optical-infrared [15,16,17] is most common. We discuss registration methods for MMRS images, rather than the types of modal image pairs. Readers who want to know about more registration methods can refer to [18]. In this regard, we review the general methods of MMRS image registration, especially classification according to the registration method category, and introduce recently popular learning-based methods, so that readers can learn about cutting-edge methods in the field at a glance. The integrated structural framework of this review is shown in Figure 1.

2. General Framework of MMRS Image Registration Methods

Area-based methods (also called intensity-based methods) generally use a robust similarity measure to search for the optimal geometric transformation through a predefined template window. Feature-based methods primarily include three steps: feature detection, feature description and feature matching [12].

2.1. General Framework of Area-Based Methods

The framework of area-based method is illustrated in Figure 2. This framework consists of three major components: namely, similarity measures, geometric transformation models and optimizations.

2.1.1. Similarity Measures

Similarity measurement as the core factor in the iterative process of MMRS image registration is used to design various standards based on the intensity difference between two images [10]. Common similarity measures include the sum of square differences (SSD), sum of absolute differences (SAD), cross-correlation and normalized cross-correlation (NCC). Based on the assumption that two images have similar intensities, SSD directly calculates the distance between the template window and the corresponding pixels of the original image. Due to the simplicity of SSD, it has high computational efficiency. SSD can be expressed by the following formula:
D ( x , y ) = i = 1 m j = 1 n I ( x + i 1 , y + j 1 ) T ( i , j ) 2
where T indicates the template window of size m × n , I represents the original image, in which we need to find an area matching the template. Taking x , y in the upper left corner of the original image I as the starting point, the subgraph of size m × n traverses the whole image and calculates its similarity with the template and finds the subgraph with the maximum similarity as the final matching result. NCC [19,20] judges the linear correlation of two images by calculating the correlation coefficient matrix. NCC can be expressed as:
ρ ( x , y ) = σ ( I x , y , T ) D x , y D = 1 m n i = 1 m j = 1 n ( I x , y ( i , j ) I ¯ x , y ) ( T ( i , j ) T ¯ ) 1 m n i = 1 m j = 1 n ( I x , y ( i , j ) I ¯ x , y ) 2 1 m n i = 1 m j = 1 n ( T ( i , j ) T ¯ ) 2
where I x , y indicates that the sub-block with the same size in the original image which takes x , y as the upper left corner point. σ · is the covariance. D x , y and D are the variances of I x , y and T respectively. I ¯ x , y and T ¯ are expressed as the mean gray value.
Note that SSD and NCC are sensitive to nonlinear radiation differences, they are not suitable for MMRS image registration. Mutual information (MI) is more robust to radiation changes [21,22], whose objective, analogous to NCC, is to maximize the MI between the two images. The MI of random variables A and B can be expressed as [23]:
I ( A , B ) = E ( A ) + E ( B ) E ( A , B )
where E · are the entropies, and E A , B is the joint entropy. The MI metric specifies that when I A , B reaches its maximum value, two images are registered. The entropy and joint entropy can be calculated as:
E ( A ) = a P A ( a ) log P A ( a )
E ( B ) = b P B ( b ) log P B ( b )
E ( A , B ) = a , b P A , B ( a , b ) log P A , B ( a , b )
where P A ( a ) = b P A , B ( a , b ) and P B ( b ) = a P A , B ( a , b ) are the marginal probability mass functions, and P A , B ( a , b ) = h ( a , b ) / a , b h ( a , b ) is the joint probability mass function, h is the joint histogram of the two images, which can be represented by the following matrix:
h = h ( 0 , 0 ) h ( 0 , 1 ) h ( 0 , N 1 ) h ( 1 , 0 ) h ( 1 , 1 ) h ( 1 , N 1 ) h ( M 1 , 0 ) h ( M 1 , 1 ) h ( M 1 , N 1 )
where h a , b is the number of point pairs for which the intensity in one image is a and for the other image is b , M and N are the ranges of the intensity values. We determine the MI value between two images by calculating the joint histogram of each window to be matched. Although MI is more suitable for MMRS image registration due to its robustness to nonlinear radiation differences, researchers are committed to improving this method to enhance the registration performance of MMRS by reducing the computational cost. For instance, Yang et al. [24] and Lehureau et al. [25] combined MI with feature-based methods to improve the registration accuracy of optical-to-SAR images.

2.1.2. Geometric Transformation Models and Optimizations

The other key steps of area-based MMRS image registration are the selection of geometric transformation models and optimizations. The transformation model, also called the mapping function, requires estimation of model parameters for image warping and resampling, and finally alignment and registration. Existing transformation models can be simply divided into linear models (for example, rigid models, affine transformation models, and projection models) and non-rigid models, such as physical models and interpolation models [10].
After selecting the metric and transformation model, it is necessary to find the optimal transformation in the iterative process of MMRS image registration to achieve optimal matching between two images. However, it is easy to fall into a local optimum during the search process, which leads to a decrease in registration performance [26]. It can be seen that the choice of the transformation model and optimization method largely determines the accuracy of registration. Readers can refer to [10,27] for more methods and details.

2.2. General Framework of Feature-Based Methods

Feature-based MMRS image registration includes main three steps. First, feature detection, which selects the prominent features between two images, such as point features [28], edge features [13] and region features [29]. Second, feature description, which refers to describing the extracted features for the next step of matching; and third, feature matching, which designs a specific similarity measure for the descriptor to establish a geometric transformation model, to realize the alignment and registration between the two images. Figure 3 illustrates the feature-based framework.

2.2.1. Feature Detection

Feature detection and extraction can be divided into corner and blob detection. Corners usually refer to the position in the image where the gray value changes substantially, such as an edge intersection point. Moravec first proposed a corner detector, which mainly extracts key points from the intensity difference in a local image area [30]. After this, based on the Moravec detector, the Harris corner detector was developed by the image gradient and the response function was proposed [31]. Subsequently, the Gaussian Laplace (LOG) detector motivated Mikolajczyk [32] to combine the LOG and Harris detectors to develop a new version that locates key points in scale space, i.e., the scale-invariant Harris detection (Harris-Laplace).
Blobs generally refer to areas that are different in color and gray scale from the surrounding area (also called singular areas). The scale-invariant feature transform (SIFT) [33,34] is a classical algorithm for blob detection. Owing to its invariance to scale changes, translation and rotation, and robustness to geometric distortion, it is widely used in feature detection. The SIFT feature detector generates the key points of position and the scale by constructing a Gaussian pyramid and locating the extreme points in the difference of Gaussian (DOG) space. Due to the scale invariance and positioning accuracy of SIFT, SIFT-like methods have been widely used in various applications. Many academics have successfully improved SIFT to eliminate multi-modal difference, thereby achieving MMRS image registration, including multispectral [35,36], optical-to-SAR images [5,14] and visible-to-infrared images [37].

2.2.2. Feature Description

Feature description refers to generating specific descriptors for the extracted feature points to prepare for the follow-up feature matching. This step directly determines the registration performance, which requires generating stable descriptors for the matched features and being robust to geometric transformation and image resolution. SIFT mentioned above and SURF (speeded up robust features) [38] are the two most common feature descriptors. The major idea is to calculate the gradient amplitude and direction around the interest point to generate a histogram descriptor. At the same time, Gaussian derivative, moment invariant, and shape context are also frequently used descriptors.

2.2.3. Feature Matching

Assuming that two sets of interest points have been extracted from the reference image and the sensed image, feature matching finds the correspondence between them based on a specific algorithm to obtain more matching point pairs. A direct method is to correspond the two interest points by directly using spatial geometrical relations like graph matching and point set registration. An indirect method is based on the similarity of local feature descriptors with false matches being rejected using local and/or global geometrical constraints. The most typical methods for mismatch removal are resampling-based methods, which are also known as fast sample consensus (FSC) and random sample consensus (RANSAC) [39].

2.3. General Framework of Deep Learning-Based Methods

Traditional image registration technology is constrained by image resolution, intensity differences and low computational efficiency. With the rapid development of deep learning, learning-based methods can promote an iterative area-based method, directly estimate the geometric transformation parameters, and reduce the computational complexity. At the same time, in the process of MMRS image registration, deep learning shows great potential in detecting feature points of images with appearance differences, further improving the registration performance. Common methods for solving MMRS image registration problems based on deep learning are described in detail in Section 3.4.

2.3.1. Convolutional Neural Network

The model construction of deep learning is based on a neural network and consists of many layers [40]. These layers transform input data into output by learning features. The “hidden layers” are usually between the input and output layers. Deep learning refers to a large number of hidden layers [40]. Convolutional neural network (CNN) is one of the most common models in the field of deep learning. Because of the structure characteristic of continuous layers, it can capture more complex image features and learn features for registration tasks. These continuous layers are specifically divided into convolutional layer, pooling layer and fully connected layer. The convolution operation is mainly to extract image features. With the increase of convolutional layers, multi-layer networks can extract richer image features. After convolution, there are still many dimensional features of the image. The pooling layer divides the feature matrix into several individual blocks and takes the maximum or average value; that is, maximum pooling and average pooling, which play a role in dimensionality reduction. Finally, the fully connected layer non-linearly combines all the local features and the feature matrix of each channel to obtain the output.

2.3.2. Generative Adversarial Networks

At present, Generative adversarial networks (GANs) [41] is a very popular technology for deep learning. It is composed of the generator and the discriminator. In the learning process, the generator makes the output image as real as possible, while the discriminator has to work hard to identify the true and false images. This process is similar to a “two-person game”. Both networks try to optimize two completely opposite loss functions [42]. Unlike CNN’s powerful ability to analyze data and extract features, GANs focus on generating data, enhancing data by the adversarial network, or generating fake images to eliminate modal differences.

2.3.3. Other Deep Learning Models

Widely used deep learning models include recurrent neural networks (RNN) [43], autoencoders (AEs) [44], stacked autoencoders and Restricted Boltzmann machines [45]. They are mostly utilized in the fields of sequence analysis [46], feature representation [47,48], data compression and dimensionality reduction [49], and image classification [50,51,52,53], and do not involve remote sensing image registration.

3. Review of MMRS Image Registration Methods

Remote sensing images are obtained from remote sensing satellites, which record the magnitude of electromagnetic waves of various ground objects. Depending on the imaging system, it mainly includes optical, SAR, Light Detection and Ranging (LIDAR) and Radio Detection and Ranging (RADAR) data, which could be combined for remote sensing applications. For MMRS image registration, there are several types of multi-source, multi-spectrum, and multi-temporal algorithms. Common research mainly includes the following types: optical-to-SAR, optical-to-LIDAR, infrared-to-visual, and visual-to-map images. As shown in Figure 4, in the optical and SAR images of the same area, due to the differences among imaging systems and the presence of speckle noise, it is very difficult to make use of traditional methods to register two images.
For this type of MMRS image registration problem, with significant non-linear radiation differences, the area-based method is susceptible to problems of image overlap, geometric distortion, and high-resolution images with high computational cost. However, the typical feature-based method (SIFT) is not suitable for multi-modal images. It is difficult to extract a large number of matching point pairs and find a suitable descriptor to match them. In order to improve the accuracy and efficiency of MMRS image registration, researchers have proposed many improvements. This section will review and discuss these methods in detail. The following mainly include area-based, feature-based and deep learning-based improvements.

3.1. Review of Area-Based Methods

As mentioned in Section 2, an area-based method finds an appropriate similarity measure in two images to accurately estimate the optimal parameters of the geometric transformation in the iterative process. However, this method needs a large amount of computation and is easily trapped into a local optima. Thus, relevant researchers have made improvements in the following aspects:
  • Improving the similarity measure MI that is the most suitable for MMRS image registration;
  • Establishing a similarity measure to eliminate modal differences based on the shape and structure of an image;
  • Improving the optimization method for high-resolution and noise-affected images to enhance registration performance.
Table 1 shows a summary of the improvements on area-based algorithms in MMRS image registration.

3.1.1. Methods to Improve Similarity Measures

Multi-resolution remote sensing image registration using stochastic gradient to optimize the MI similarity metric was proposed in [21]. First, it improved the joint histogram of MI into 64 bins to produce a smoother surface, which can better integrate the optimization method. In addition, the author proposed using the simultaneous perturbation stochastic approximation (SPSA) algorithm, which relies on an efficient gradient approximation instead of an accurate target gradient value. Then, the global maximum value was found through a local extrema of the objective function, which greatly accelerates the entire registration process. Chen et al. [23] put forward a new joint histogram estimation algorithm called generalized partial volume estimation (GPVE, with a second-order B-spline function) for computing MI to register multi-temporal remote sensing images, to overcome the problem of interpolation-induced artifacts.
Regardless of whether mutual information (MI) or normalized mutual information (NMI) overcomes the overlapping area problem, the estimation of the similarity measure requires calculating its joint histogram. Shadaydeh et al. [54] proposed a weight-based joint histogram estimation method (WJH-MI). That is, each bin in the joint intensity histogram is calculated as the sum of the weights corresponding to the pixel intensity value of the bin. The weight of each pixel is defined as the exponential function of the distance image and the normalized gradient image. This method reduces the peaks in the joint histogram caused by the background or homogeneous regions by assigning more weight to the high gradient pixels containing more registration related information. Experiments show that the proposed method produces a better similarity measurement surface, with more obvious peaks and fewer registration errors. In order to solve the problem of MI (also known as Kullback–Leiber divergence) facing insufficient image overlapping area, Xu et al. [55] proposed using the symmetric form of Kullback–Leiber divergence, namely Jeffrey’s divergence as similarity measure. Jeffrey’s divergence quantifies the distinction of image pairs by bidirectionally measuring the “distance” between the joint histogram and the product edge histogram of two images. Derived from the definition, the registration model based on Jeffrey’s divergence can provide a larger feasible search space. Experiments showed that the sensitivity of Jeffrey’s divergence to scene overlap is lower than other measures, and it is more suitable for registering small size images.
Structure and shape features have been used as similarity measures in multi-modal medical image registration, achieving better performance than traditional similarity measures. The features extracted by the phase correlation or phase congruency (PC) algorithm proposed by De Castro [65] and Reddy [66] are widely used in robust MMRS image registration due to their insensitivity to changes in illumination and contrast. Xie et al. [56] proposed the MLPC method, based on the multi-scale Log-Gabor [67] filter combined with traditional phase congruency methods [65,66], effectively resolving the non-linear intensity difference in MMRS image registration. The authors obtained the amplitude of the reference and sensed images in the frequency domain after Fourier transform, and then used log-Gabor filters with different central frequencies to filter the amplitude to obtain a series of filtered images. After filtering, a multi-scale map space was constructed and each image pair in the map was phase-correlated. The maximum response peak was considered as the optimal solution, and the coordinates of the maximum response peak were used to obtain the rotation angle and scale factor. Finally, the cross-power spectrum between the superimposed structure spectrum of the reference image and the corrected sensed image was calculated to eliminate the radiation difference. The inverse Fourier transform was used to obtain the maximum peak value, and the coordinates of the maximum peak value determined the transformation. Subsequently, based on this article, the authors proposed an improved LGEPC [57], which is an extended phase correlation algorithm based on log-Gabor. Similarly, the filtered amplitude image obtained by log-Gabor filters with different central frequencies was used to construct a multi-scale atlas space and obtain rotation and scale factors. In addition, the filter structure spectrum obtained by log-Gabor filters of different central frequencies was superimposed together to enhance the overall structure information, which helps eliminate the influence of radiation differences as much as possible on the step of solving translation in the extended phase correlation.

3.1.2. Methods to Improve Optimization

A new similarity measure, called cross-cumulative residual entropy (CCRE), was successfully applied to multi-modal medical image registration to adapt to images with different brightness and contrast, while being more robust to noise. Hasan et al. [58] used CCRE to align SAR and Google satellite images, and at the same time extended the Parson-window optimization method proposed by Thevenaz [68]. The authors applied partial volume interpolation to calculate the gradient of the similarity measure instead of the joint histogram, which allowed the authors to directly implement the optimization process based on partial volume interpolation. Results showed that the use of partial volume interpolation in the optimization process significantly improved the registration success rate and accuracy of CCRE-based and MI-based algorithms. The authors defined a new inverse combinatorial optimization method in [59] to handle the quasi-concave shape of MI, where the required derivative can be pre-calculated allowing the Hessian matrix to be estimated after convergence. Thus, this method can reduce the calculation time and estimate an accurate parameter. Note that the definition of MI has been adapted to the differential image alignment problem, so that the alignment function is as smooth and concave as possible and retains robustness to multi-modal image intensity changes. In addition, utilizing a new method based on reference pixel selection greatly reduces time consumption, which results in an accurate, fast and robust registration process. In [60,61], the authors used the ant colony optimization (ACO) algorithm to optimize the similarity measure to maximize MI. However, the similarity curve of MI has been proved to have many local optimal values. Thus, Yan et al. [26] was inspired by transfer learning and proposed a method of using transfer optimization (TO) to maximize MI, optimize conversion parameters and transfer better results to another optimizer in the iterative process. This helps enhance the global search capability of TO and avoid a local optima.

3.1.3. Hybrid Methods

Liu et al. [62] proposed an image registration method that combines local self-similarity (LSS) and MI, which combined local internal features and global intensity information to achieve a better registration effect. A likelihood map was calculated for each CP in the image using the Gaussian pyramid weighted Bayesian probabilistic model. The position of the peak in the likelihood map represented the corresponding CPs in the image. At the same time, the authors took the maximum of MI as the optimization goal and introduced the particle swarm optimization (PSO) algorithm to search for the best parameter of the geometric mapping function. Considering effectiveness and performance, experiments show that the proposed method can register multi-sensor images with different resolutions and handle the geometric differences between grayscale and corresponding pixels and regions. Ye et al. [63] proposed a fast and robust template matching framework for MMRS images. The first step extracts a local descriptor at each pixel of the image, which can be HOG, LSS, or SURF to generate a pixelized feature representation. Here, the author introduced a novel pixelized feature representation called channel features of orientated gradients (CFOG), which is an extension of the pixelated HOG descriptor. The pixelated feature representation can be constructed by using HOG descriptors with a single unit block. The trilinear interpolation in a single unit block can be regarded as a convolution operation with a triangular kernel. Thus, the HOG descriptor was reconstructed by convolution in the image gradient of a specific direction, and the convolution was performed by the Gaussian kernel instead of the triangular kernel. The reconstructed CFOG descriptor suppressed noise more effectively and reduced the contribution of the gradient away from the center of a region, improving the computational efficiency. Subsequently, a new template matching similarity measure based on pixel feature representation was introduced, and fast Fourier transform was used for faster computation. Yan et al. [2] proposed a new similarity measure, the histogram of oriented gradient distance (HOGD). In order to avoid falling into local optimality, the grey wolf optimizer (GWO) was introduced since it exhibited global search capabilities for complex optimization problems. GWO searched for optimal transformation parameters by minimizing HOGD. To reduce computation time, GWO was combined with a data-driven strategy, namely DDGWO. In DDGWO, a support vector machine (SVM) model was trained to predict HOGD instead of calculating it directly, which can lead to a significant reduction in computation time. Liang et al. [64] proposed a fast-matching method based on dominant orientation of gradient (DOG), which constructed a feature map by extracting the DOG feature of each pixel in an image. The author defined a new similarity measure called the sum of cosine difference, which can be accelerated by the Fast Fourier Transform (FFT). In order to improve the matching performance, a new variable template matching (VTM) method was proposed to determine the correspondence between images. Experimental results showed that the proposed matching method had good robustness to nonlinear light intensity changes, and effectively improved the matching accuracy.

3.2. Review of Feature-Based Methods

Researchers have made many improvements and contributions to the area-based methods. However, there are still cumbersome and time-consuming calculations in the iterative optimization process caused by high resolution and strong noise. Most methods rely on geometric correction based on the geographic information of an image. However, when geographic information is not available, an image is difficult to optimize considering large geometric changes. Feature-based methods are robust to rotation, translation and geometric distortion; but, due to speckle noise and intensity differences, it is very difficult to extract feature points between two images and it is easy to remove outliers. Hence, how to detect reasonable features between reference and sensed images and design suitable descriptors for them has become a challenge for scholars. Table 2 shows a summary of the improvements on feature-based algorithms in MMRS image registration.

3.2.1. Methods to Improve SIFT or Harris Operator

SIFT and Harris are relatively mature and complete image registration methods. Therefore, researchers have improved on both methods directly, considering MMRS image registration. In [28], the authors combined the advantages of SIFT and Harris to propose a coarse to fine MMRS image registration. First, SIFT was used to automatically find the CPs between the images to roughly match the reference and sensed images. Then, the affine transformation model was utilized to detect the transformation. Next, the algorithm applied Harris detector to perform fine registration on the coarsely matched image while processing the local deformation in the image. Then, a set of dense feature points from the input image were found. After that, the wavelet pyramid was used to search for the corresponding connection points of these feature points. Simultaneously, the error in matching was eliminated by the global consistency check method. Finally, the input image was corrected by piecewise linear transformations and the triangular irregular network (TIN). In [35], the authors proposed a registration method, called Uniform Robust SIFT (UR-SIFT), suitable for a variety of optical multi-source remote sensing images with illumination, rotation and up to five times the scale difference. A symmetric SIFT [82] algorithm adapted to multimodal invariance was widely used in MMRS image registration. However, in the process of descriptor merging, this algorithm led to many other combinations to be merged into the same final descriptor, which greatly reduced the registration accuracy. In order to address the above problems, Hossain et al. [69] proposed an improved symmetric SIFT method. The overall direction difference between the given images was estimated by analysing the initial matching set, and the estimated value was used to rotate and normalize the descriptor in the second step. This way of normalization has not caused area reversal. Hence, the descriptor merging step in symmetric SIFT can be skipped to obtain excellent matching accuracy.
Fan et al. [5] proposed a novel spatial consensus matching (SCM) algorithm, which used improved SIFT and K-nearest neighbors (KNN) to obtain an initial set of matching features and utilized spatial consistency constraints (the low distortion constraint was used) for refinement. The algorithm obtained matching features by gradually adding features that are spatially consistent with the currently obtained matching features. Finally, RANSAC was used to estimate the transformation parameters from spatially consistent matching features. In [70], a new algorithm for improving feature extraction and feature matching was introduced. The Harris operator was utilized to extract the CPs of the reference image, and the Canny operator was used to extract the edge features of the reference and sensed images, while processed by the dilation algorithm to retain useful edge information. Finally, the improved shape context descriptor was applied to match the image by comparing the edge feature distribution in the defined circular template.
Xiang et al. [12] proposed a SIFT-like algorithm (OS-SIFT) in order to achieve high-resolution optical-to-SAR image registration. Experimental results on simulated images and multiple high-resolution satellite images showed that the algorithm was excellent in registration performance and accuracy. Two different operators were introduced in this algorithm, multi-scale ratio of exponentially weighted averages (ROEWA) [83] and multi-scale Sobel, to calculate the consistent gradient of SAR and optical images to demonstrate robustness to noise. Then, two Harris scale spaces were constructed for SAR and optical images, and a location refinement method based on key point spatial information was proposed. That is, key points of different scales belonging to the same corner point should have similar structural properties, which means they were in a straight line. Finally, orientation restrictions and multiple image blocks were used to construct GLOH-like descriptors, which contained more structural information and produced more stable registration results. In successive research, Xiang et al. proposed a dense registration algorithm that combines robust features and optical flow, namely OS-flow [84]. First, this algorithm extracted two dense feature descriptors for each pixel in the optical and SAR images, i.e., the optical GLOH descriptor and the SAR GLOH descriptor. Using dense descriptors instead of brightness values can satisfy the assumption of brightness constancy. Then, the global and local methods were constructed. The global method estimated the flow graph by optimizing the objective function, and the local method iteratively estimated the flow vector in the local neighborhood. Finally, a coarse-to-fine matching strategy was adopted to handle large displacements and improve efficiency. In [85], the authors proposed a subpixel registration method that combined robust features and 3-D PC (phase correlation). Robust features captured the inherent properties of two images and retained their structural information. 3-D PC used image cubes as a substitute for the two original images, which estimated the 2-D translation by locating the peak in the spatial domain or directly in the Fourier domain. In addition, in order to improve the accuracy of 3-D PC, a constrained energy minimization algorithm was introduced, which was used to find the Dirac delta function after the inverse Fourier transform, as well as fast sample consistency fitting to estimate the phase difference after high-order singular value decomposition of the PC matrix.

3.2.2. Methods to Improve Feature Descriptor

Designing a robust and effective descriptor for MMRS images has become the most popular improvement method nowadays. In order to avoid the lack of correlation of gradients caused by nonlinear intensity differences, Aguilera et al. [71] proposed the edge-oriented histogram (EOH) descriptor to merge the spatial information of the contour from each key point without using the gradient information. The descriptor contained the information of the contour near each feature point to describe the shape and contour of the image. Finally, a large number of multispectral image pairs were used to evaluate the proposed method.
Local self-similarity (LSS) [86] is a local feature descriptor that captures the internal geometric structure of the image based on a grid of log-polar coordinates, which is insensitive to color and intensity changes. Therefore, in [72], the authors used SR-SIFT with scale limitation to first detect key points, then they combined Harris and LSS descriptors to establish a piecewise linear transformation to achieve reliable registration of multi-spectral remote sensing images. However, the motivation for this method is to minimize the impact of the low distinguishability of the LSS descriptor. In general, a good descriptor should be robust and distinctive. In order to satisfy the above properties, Sedaghat et al. [73] proposed an advanced version of the self-similarity descriptor, which has high distinguishability, called distinctive order based self-similarity (DOBSS) descriptor. First, the UR-SIFT proposed by the author’s previous work was used to extract a set of uniform and dense features. Second, the authors adopted a correlation value-based direction assignment method to construct descriptors, and at the same time utilized correlation values to group pixels in a local area to improve the separability of the descriptors. Finally, all the log-polar coordinate calculation descriptors were connected to form the final descriptor. Experiments showed that the DOBSS descriptor has better recall, precision and positioning accuracy. Based on the internal self-similarity of images, Ye et al. [74] introduced a shape descriptor of dense local self-similarity (DLSS). First, this method selected a template window on the image and divided the window into spatial regions, called “cells” containing pixels. Then, it extracts and normalizes the LSS descriptor of each cell and passes through a dense overlapping network that covers the template window. The LSS descriptors collected by all the cells in the grid were combined to construct a DLSS. Finally, the normalized cross-correlation of DLSS descriptors was used to define a similarity measure (called DLSC), and a template matching strategy was utilized to detect the correspondence between images. Based on DLSS and inspired by the Spearman’s rank correlation coefficient in statistics, Xiong et al. [75] proposed rank-based local self-similarity (RLSS) to describe the local shape of an image in a distinguishable manner. The rank value was used as a substitute for the correlation value to indicate the relative relationship of the correlation value. In addition, similar to the scheme of Ye et al., this descriptor was integrated into a dense sampling grid to obtain a dense RLSS descriptor (DRLSS), which further improved distinguishability of feature descriptors.
Another popular approach is to use PC to describe the detected key points in the frequency domain and design efficient descriptors. Cui et al. [76] proposed a multi-scale phase-congruency descriptor (MS-PC), which captured the shape and structural characteristics of an image to compensate for the sensitivity of traditional descriptors to radiation differences. First, the extreme values were detected in the DOG space, the log-Gabor wavelet was used to construct the phase congruency map, and the MS-PC feature vector was established. Finally, the Euclidean distance was introduced as the matching metric to obtain the final reliable matching point pair. Ye et al. [4] were inspired by medical image processing and combined the PC describing structural features with the histogram of gradient directions (HOG) as a feature descriptor, called HOPC. The author first detected the CPs through the block-based Harris operator, at the same time sorted the Harris values in each block from large to small and selected the first k points as interest points. Then, they defined a similarity measure named according to the NCC of the HOPC descriptor and used the fast template matching scheme to match the control points. Mismatches were eliminated by the global constraints in the projection transformation model, and ultimately estimated by a piecewise linear model to achieve precise registration. The authors verified the robustness and accuracy of the proposed method on multiple sets of MMRS images. In [15], Liu et al. proposed a method of registering infrared and visible images, namely maximum stable phase congruency (MSPC), which combined affine invariant region extraction and image structural features. They used moment ranking analysis to detect feature points and extracted structural features by merging phase congruency images in multiple orientations. After obtaining the neighborhood centered on the feature point through the log-Gabor filter response, the maximum stable extreme value region (MSER) was used to determine the affine invariant region of the feature point. Finally, the orientated phase congruency was constructed from these regions to describe the structural features, and registration was achieved according to the correspondence of the descriptors.

3.2.3. Hybrid Methods

Traditional methods are often unsatisfactory in extracting reliable matching point pairs, so they cannot generate robust descriptors to estimate geometric transformations accurately. Based on the feature-based registration method, researchers have improved both feature point extraction and feature description to achieve MMRS image registration that is more adaptable to nonlinear intensity differences.
For solving the problem of local invariant features being sensitive to the radiation difference between multi-sensor images, Ye et al. [77] proposed a local invariant feature that was robust to geometric distortion and radiation changes. It consisted of a feature detector named MMPC-lap and a feature descriptor named local histogram of orientated phase congruency (LHOPC). Inspired by Harris-Laplace, MMPC-lap combined the minimum moment of phase congruency with LOG, used the MMPC to detect key points, and then utilized LOG for scale location. Once a set of key points was detected, the phase congruency magnitudes and orientations were applied to construct the descriptor according to the spatial arrangement of DAISY [87]. In addition to adapting to geometric distortions such as scale and rotation, this method also solved the radiation differences caused by spectral and time changes between MMRS images. Compared to Gaussian smoothing (GS), nonlinear diffusion [88,89] can better preserve edge features and details while suppressing noise. At the same time, previous experiments also show that uniformly distributed Harris corners are more reliable in the matching process. Thus, Fan et al. [78] proposed an UND-Harris detector that introduced nonlinear diffusion, feature ratio and block strategy. Secondly, the phase congruency structure descriptor (PCSD) was inspired by the LSS descriptor and was built on the PC structure image of each extracted point in the form of PC sequential grouping. Experimental results on different SAR and optical image pairs showed the effectiveness of this method, which can obtain better registration results and improve registration accuracy.
The greatest challenge for the registration of infrared and visible images is the grayscale difference between the two images, but they both retain obvious contour edges. Therefore, morphological methods can be used to retain the grayscale edges of the image and improve the similarity. Consequently, Zeng et al. [37] proposed an infrared-to-visible image registration method based on morphological gradient and C-SIFT. First, the rough edges of the image were extracted by grayscale morphology. Using structural elements as a template, the authors retrieved the maximum (minimum) gray value of the template size in the image to obtain the morphologically dilated (eroded) image; thereby obtaining the gray morphological gradient image, which contained the rough edges in the image. In C-SIFT, DOG was used to extract feature points to ensure scale invariance, and the centroid method was used to obtain the main directions of feature points. In addition, BRIEF feature descriptors with fast calculation speed and strong gray invariance were introduced, and Hamming distance was utilized as a similarity measure to match feature points. Through in-depth research on phase congruency, Li et al. [79] proposed a radiation insensitive image registration method based on phase congruency (PC) and a maximum index image (MIM), which was called radiation variation insensitive feature transform (RIFT). First, the precise edge map, i.e., the PC map, was used to obtain the minimum moment map corner features and the maximum moment map edge features, to obtain a large number of repeatable features. Then, the algorithm employed the MIM based on the log-Gabor convolution sequence to describe the feature and constructed multiple MIMs to realize the rotation invariance by analyzing the internal influence of the rotation on the MIM value. The authors verified the reliability and accuracy of RIFT in the registration process through multiple qualitative and quantitative comparisons.
LSD [90] is a line segment detection algorithm, which can obtain high-precision line segment detection results in a short time. The LSD line detection algorithm first calculates the gradient size and direction of all points in the image, and then takes the adjacent points with small gradient direction changes as a connected domain. Subsequently, according to the rectangular degree of each domain, we need to decide whether to disconnect it to form a plurality of domains with a larger rectangular degree. Finally, all the generated domains are improved and filtered, and the domains that meet the conditions are retained as the final straight line detection result. The advantage of this algorithm is that the detection speed is fast, there is no need for parameter adjustment, and the error control method is used to improve the accuracy of straight-line detection. Sui et al. [13] proposed an iterative process combining line segment extraction and line intersection matching based online segment extraction, and integrated Voronoi polygons into spectral point matching (SPM) to obtain the correspondences between line intersections. At the same time, an iterative strategy of “re-extraction” and “re-matching” mechanisms was adopted to enhance feature extraction and matching performance. Zhao et al. [80] used Kovesi corner point extraction and line segment detection methods based on phase congruency and local direction. Compared to other edge extraction methods, this method extracted more equivalent line segments. Then, a new multi-modal robust line segment descriptor (MRLSD) was proposed, which was calculated by using line segment information located in a circular feature region. Based on this, the MRLSD matching method was presented. Xu et al. [81] proposed a combination of feature descriptors called “Angle Histogram and Maximum Edge Orientation Distribution” (HAED). First, they used the angle and edge direction distribution to extract the image information to generate contour segment features and capture the local and global textures respectively. Second, similarity was calculated using the Fréchet distance metric between the curves, which is the weight parameter of the histogram of each contour segment. Finally, the precise bilateral matching rules were used to match the corresponding contour segments. This method can effectively reduce the influence of radiation distortion and is better than current popular multi-source image matching methods.

3.3. Review of Combined Area-Based and Feature-Based Methods

Template matching and feature-based methods have their own advantages and disadvantages in MMRS image registration. By extracting reliable matches between two images and finding the corresponding relationship between them, they can handle rotation, translation, scale difference and geometric distortion. The robust similarity measure can effectively eliminate the nonlinear radiation difference caused by multiple sensors. Based on the above reasons, the researchers combined these two types of methods to achieve more accurate matching. Table 3 shows a summary of the improvements on integrated algorithms in MMRS image registration.
Gong et al. [3] proposed a novel coarse-to-fine scheme for automatic image registration. In coarse registration (pre-registration), the scale histogram was used to remove the outliers detected by SIFT. The dense clusters in the histogram were the true scale differences between the images. Key point pairs that contribute to the cluster were considered to be correct matches, and those far from the cluster were considered incorrect matches and removed. Due to the robustness of MI to noise and its adaptability to different image intensity values, the authors adopted it to refine the pre-registration results in the multiresolution framework. Finally, the improved Marquardt–Levenberg search strategy was developed to find the global optimum. In [72], the authors put forward a local descriptor-based registration method for multispectral remote sensing images. In the first stage, pre-registration used SR-SIFT and projection transformation to eliminate the obvious rotation and scale differences between the reference and sensed images. In the second stage, first a set of evenly distributed interest points were extracted from the pre-registered image based on the block-Harris method. Then, through the bidirectional matching technique, LSS was used as a new similarity measure for connection point detection (named LSCC). Finally, the pre-registered image was corrected using a piecewise linear model. Experimental results showed that three pairs of multispectral remote sensing image pairs with significant nonlinear intensity differences and geometric distortions from different sensors could achieve reliable registration accuracy.
As mentioned earlier, the edge information is relatively stable and consistent in optical and SAR images, and the line intersection point as a matching primitive is a great choice. Thus, Xiong et al. [91] proposed coarse registration based on intersections of straight lines and fine registration based on MI from separated patches. Zhang et al. [92] used optical and SAR images as the reference image and the sensed image respectively. They first used the feature-based SAR-SIFT algorithm to complete the coarse registration, and then utilized the area-based ROEWA-HOG method to complete the fine registration to achieve high-precision automatic registration of the hybrid model.

3.4. Review of Deep Learning-Based Methods

Existing methods of using deep learning to solve the MMRS image registration problem can be roughly divided into three categories. First, deep learning algorithms (such as CNN and GAN) are integrated into the existing MMRS image registration method to generate high-level and deep features. Second, generating an end-to-end network framework based on a Siamese network for MMRS image registration, and third, using deep learning to eliminate the difference between modalities, and then combining traditional methods for image registration. Table 4 summarizes improvements on deep learning-based algorithms for MMRS image registration.

3.4.1. Methods to Improve CNN and GAN Algorithms

Ye et al. [93] used CNN to extract the middle and high-level features of an image and combined them with the low-level features extracted by SIFT. Since the PSO-SIFT algorithm achieved advanced performance in remote sensing image registration, features of CNN and SIFT were incorporated into the PSO-SIFT algorithm for registration. Visual Geometry Group (VGG) is a commonly used high-level image feature extraction network. Ma et al. [94] used VGG-16 to approximate spatial relationships and proposed a new point matching strategy based on spatial relationships and combined the local feature-based methods. Due to the powerful feature extraction capability of CNN, the traditional matching method is stable and effective. Considering the spatial relationship, the matching result is more robust and accurate. In [95], the authors adopted a pre-trained VGG network to generate multi-scale descriptors through high-level convolution information features and utilized the gradually expanding dynamic inlier selection to optimize the registration details by increasing the number of feature points.
Quan et al. [96] applied GAN for MMRS images data augmentation, which could immensely enhance the accuracy and robust of registration process. The author used a dual-channel deep network as a matching network, which helped to extract different features of optical and SAR images. CNN was introduced as a feature extractor to save the spatial information of the image. Through multiple constraints, such as associative constraints and geometric constraints to delete incorrect matching points. Experimental results showed that this method was superior to traditional methods and had a great registration performance for optical and SAR images. Merkle et al. [97] proposed a GAN-based method for dealing with optical and SAR image registration. First, a pseudo-SAR image was generated from the optical image by training a GAN-based network, so that the two images had similar intensity information. Subsequently, the area-based method and the feature-based method were introduced to match the pseudo-SAR with the real SAR image. The experimental results validated the necessity of this method, especially the generator in GAN. It accurately and reliably retained the geometric structure of the optical image, opening up new possibilities for MMRS.

3.4.2. Methods to Improve Siamese Network

The methods of using deep learning to solve MMRS image registration problems are mostly based on Siamese networks. These methods are usually trained to learn the features and differences between multi-modal image pairs. It is then used to measure the similarity between these image pairs.
Hughes et al. [98] produced a specific pseudo-Siamese network to register the optical and SAR images. The two convolutional streams of this network are identical and independent, and there is no parameter sharing to process the different intensity information of the two images. Finally, the fully connected layer is used to fuse the different features learned by the two branches. While Zhang et al. [99] shared parameters between the two branches to solve the problem of lack of data sets. The author proposed a fully convolutional Siamese network, which used an end-to-end training process. It built a general framework for MMRS image registration based on depth features, which used the image block selected by the Harris detector to input the network, and finally obtained the similarity between the sub-image and the remote sensing image. Merkle et al. [100] trained the Siamese network to learn the spatial transformation between optical and SAR images. First, the features of the two images were extracted through CNN, and then the dot product layer [105] was produced to measure the similarity. The point with the highest final response value was regarded as the matching point. The effectiveness of this approach for the generation of reliable and robust matching points between optical and SAR images had been demonstrated by experiment. He et al. [101] proposed a novel multi-scale remote sensing image registration deep network, which includes the following three steps: corner detection based on s-Harris, a search strategy based on Gaussian pyramid coupled quadtree to narrow the search space and realize multi-scale comparison of conjugate patches. Finally, global to local quadratic polynomial constraints and RANSAC were utilized to remove mismatches. The visualization results showed that this method can make the matching evenly distributed and obtain satisfacting matching accuracy. The most important thing was to avoid the influence of complex background changes on the registration results.

3.4.3. Others

In [102], the authors extended the image visual attribute transfer method to pre-process an image in order to eliminate the intensity differences between the multi-modal image pairs, and make the color, texture and other characteristics consistent in a similar structural area. Then, it used traditional local feature-based algorithms for image registration. Wang et al. [103] proposed a remote sensing image registration framework based on deep learning, which directly learns the mapping function between image patch pairs and their matching labels through closed-loop information. A self-learning method was introduced to avoid a few image data and can learn the mapping function from itself without relying on other data. At the same time, the application of transfer learning further improves the registration accuracy and reduces training costs. Zampieri et al. [104] completed the alignment between remote sensing images and maps by designing a neural network with a specific scale to learn image features.

3.5. Multi-Modal Image Registration Methods in Other Fields

In other related fields, image processing methods such as photogrammetry, medical imaging, and computer vision have their characteristics and innovations. Some cutting-edge image processing methods will help researchers achieve new breakthroughs. Table 5 summarizes image registration algorithms in other fields.
Maximal Self-dissimilarity (MSD) [106] is taking the lead for photogrammetric based multimodal registration. The features detected through the Non-Maxima Suppression (NMS) stage are local maxima of the Contextual Self-dissimilarity (CSD) operator, herein the detector will be referred to as Maximal Self-Dissimilarity interest point detector (MSD).
In the field of medical image processing, multi-modal image registration is more mature and advanced. Its main methods can also be divided into similarity measurement problems, which are intensity-based, and control point pair extraction based on features. Using deep learning strategies to solve the above problems has become a new trend. In terms of learning similarity measures, Sedghi et al. [107] built a five-layer neural network to learn a similarity metric that can measure the level of registration to estimate the model parameters in the matching of 3D-US and MR abdominal scans. Then, Powell’s method is used to optimize the learned metric in an iterative manner. Wang et al. [108] constructed a content-adaptive weakly supervised deep learning framework for multi-modal retinal image registration, which is composed of three neural networks for blood vessel segmentation, feature detection and description, and outlier elimination. This method is superior to other learning-based methods, achieved the highest success rate and Dice coefficient, and has significant robustness for poor quality images. Huang et al. [109] proposed a novel unsupervised learning-based framework. This is an end-to-end network architecture composed of affine and deformable transformations. In addition, double consistency constraints and a new loss function based on prior knowledge have been developed to achieve accurate and efficient multi-contrast MR image registration.
For cross-spectral image matching in computer vision, Jhan et al. [110] proposed a new N-SURF matching method for multi-spectral camera (MSCs) image registration. This method has good accuracy and can obtain more correct matches (CMs) that are evenly distributed. Based on the N-SURF and EPT models, the authors developed a general tool for image registration of various MSCs, which can quickly and accurately co-register a large number of MS images, has the advantages of accuracy and efficiency, and does not require prior knowledge on sensor calibration. For cross-temporal image matching, Zhou et al. [111] were inspired by the GAN method and proposed a potential generation model for cross-weather image alignment. Based on intensity constancy and image manifold characteristics, this method describes the image registration task as a constrained deformation flow estimation problem with a potential encoding process.

4. Experiments

In this section, we select 8 representative methods and 2 assessment metrics to conduct experiments, which can provide an objective performance reference for different MMRS image registration methods and hence support relative engineering with credible evidence. The experiments are conducted on a computer with 2.5 GHz Intel Core CPU, 16 GB memory, and MATLAB codes. The codes of the 8 representative image registration methods are all publicly available.
We test the 8 representative methods on the 6 MMRS image pairs from CoFSM (CoFSM: https://skyearth.org/publication/project/CoFSM/ (accessed on 10 December 2021)), which consist of modality pairs, namely, depth-optical, optical-optical (cross-temporal), infrared-optical, map-optical, day-night and SAR-optical. The six pairs of original images selected in the database are shown in Figure 5.
Figure 6 shows the performances of 8 methods on six pairs of MMRS images. It is easy to find that the traditional single-mode image registration algorithm is inferior to some recently popular improved algorithms in terms of registration performance. SIFT [33,34], SAR-SIFT [112] and PSO-SIFT (SIFT, SAR-SIFT and PSO-SOFT available at https://github.com/ZeLianWen/Image-Registration (accessed on 10 December 2021)) [113] are all based on the image gradient to extract features and construct feature descriptors, which are very sensitive to the strong radiation difference of MMRS images. The core idea of SURF (Available at https://www.mathworks.com/matlabcentral/fileexchange/28300-opensurf-including-image-warp/?ivk_sa=1024320u (accessed on 10 December 2021)) [38] is similar to SIFT, which is an accelerated version of the SIFT algorithm, and the registration performance is also not good enough. In contrast, HOPC [4], CFOG (Available at https://github.com/yeyuanxin110/CFOG (accessed on 10 December 2021)) [63], RIFT [79] (Available at http://www.escience.cn/people/lijiayuan/index.html (accessed on 10 December 2021)) and RIFT-LAF [114] (Available at https://github.com/StaRainJ/LAF (accessed on 10 December 2021)) methods, which lead the development of MMRS image registration, show better registration performance.
In addition, we select two frequently used assessment metrics, namely the correct match ratio (CMR) and root mean square error (RMSE), to evaluate the performances of the different MMRS image registration methods. Table 6, Table 7 and Table 8 report the results of the two metrics and run times using the 8 representative methods on the 6 image pairs. In each table, we use ‘/’ to indicate that this method cannot be applicable to multimodal image pairs, so there is no data information.
From the results, we can see that the performance of classic descriptors (SIFT, SURF, etc.) is not perfect for most multimodal image pairs, and the number of CM of these methods is significantly less than the improved algorithm based on multimodality. Especially for SAR-optical image pairs, due to the significant non-linear intensity difference between the two modalities and the presence of SAR image speckle noise, both have a great impact on the registration results of the classic method. In our experiments, both RIFT and RIFT-based LAF matching algorithms have shown satisfactory performance. However, it should be noted that the RIFT algorithm used in the experiment does not consider rotation invariance. First, it needs to be unified with the geographic registration technology used by other methods. Second, when the rotation invariance is considered in the model, the computation time is very long. The HOPC algorithm based on structural similarity achieves good performance on multi-modal image pairs, but it performs poorly on cross-temporal (optical-optical) image pairs. As a multi-modal image template matching framework, CFOG has excellent performance in matching performance and computational efficiency.

5. Challenges and Future Directions

Through the review of the most cutting-edge MMRS image registration method in the previous section, current research has largely solved the nonlinear radiation distortion caused by modal differences and the geometric distortions caused by various factors. However, there are still several issues that need further consideration:
  • There is insufficient data for different modal conditions in the field of remote sensing, and there is no complete and comprehensive database containing all types of MMRS image pairs. Therefore, for deep learning methods that are developing rapidly and showing great progress, most of the work uses this technology for feature detection and feature description. The lack of training and test data greatly limits its application in MMRS image registration.
  • With the development of remote sensing technology, images from different sensors will have higher resolution. The details within high-resolution images are complex, the amount of data is large, and the local geometric deformation caused by the undulation of the terrain cannot be ignored. This results in a significant challenge for image registration.
  • By comparing existing methods on MMRS image registration, we found that there are many registration strategies for optical-to-SAR and infrared-to-visible. However, learning-based methods can only learn the characteristics and differences between specific multi-modal image pairs. In the future, how to fuse and learn the data between multi-modalities should become the focus. For example, rather than eliminating the modal difference between two images, the registration method should be suitable for remote sensing images with multiple modalities.
  • Area-based methods still face the impact of overlapping block regions and low computational efficiency on the registration performance, whereas feature-based methods face the challenge of nonlinear intensity differences.
In future research, equipping the MMRS image database and improving the registration performance through deep learning will become more mainstream. Of course, the improvement and fusion of traditional methods can also further improve the registration accuracy, and better integrate image registration into other imaging applications. Furthermore, future research trends will focus on such as solving the problem of multi-modal image pair registration at the same time through data fusion and so on. As for multi-modal high-resolution remote sensing images, how to improve the registration accuracy under significant geometric deformation and global radiation difference will be a tough challenge.

6. Conclusions

In recent years, MMRS image registration has attracted widespread attention given its important imaging applications, such as image fusion and target recognition [115]. Thus, this article comprehensively investigates existing MMRS image registration methods, which can be divided into three categories: area-based, feature-based and deep learning-based. Through a brief introduction to the major methods and theoretical frameworks, this paper provides a basis for further research to improve the performance of MMRS image registration. The latest research results are classified according to image registration methods, providing future research ideas for people in related fields. In addition, we use 8 representative methods to perform experimental evaluation on 6 common multi-modal remote sensing image pairs, and visually demonstrate the performance of different methods for MMRS image registration.

Author Contributions

Conceptualization, X.Z. and Y.H.; methodology, X.Z. and C.L.; formal analysis, X.Z., C.L., A.B. and Z.P.; investigation, X.Z. and Y.H.; resources, X.Z. and C.L.; data curation, C.L., Z.P.; writing—original draft preparation, X.Z. and Y.H.; writing—review and editing, C.L., A.B. and I.C.; visualization, X.Z. and Y.H.; supervision, C.L., I.C. and Z.P.; funding acquisition, C.L., A.B., I.C.; All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China under Grant Nos. 61702251, 61971273, the Key Research and Development Program in Shaanxi Province under Grant no. 2021GY-032, in part by the Natural Sciences and Engineering Research Council of Canada, and Youth Academic Talent Support Program of Northwest University under Grant No. 360051900151.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://skyearth.org/publication/project/CoFSM/ (accessed on 10 December 2021).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zitová, B.; Flusser, J. Image registration methods: A survey. Image Vis. Comput. 2003, 21, 977–1000. [Google Scholar] [CrossRef] [Green Version]
  2. Yan, X.H.; Zhang, D.J.; Hou, N. Multimodal image registration using histogram of oriented gradient distance and data-driven grey wolf optimizer. Neurocomputing 2020, 392, 108–120. [Google Scholar] [CrossRef]
  3. Gong, M.G.; Zhao, S.M.; Jiao, L.C.; Tian, D.Y.; Wang, S. A novel coarse-to-fine scheme for automatic image registration based on SIFT and mutual information. IEEE Trans. Geosci. Remote Sens. 2014, 52, 4328–4338. [Google Scholar] [CrossRef]
  4. Ye, Y.X.; Shan, J.; Bruzzone, L.; Shen, L. Robust registration of multimodal remote sensing images based on structural similarity. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2941–2958. [Google Scholar] [CrossRef]
  5. Fan, B.; Huo, C.; Pan, C.; Kong, Q. Registration of optical and SAR satellite images by exploring the spatial relationship of the improved SIFT. IEEE Geosci. Remote Sens. Lett. 2013, 10, 657–661. [Google Scholar] [CrossRef] [Green Version]
  6. Ma, J.; Jiang, X.; Fan, A.; Jiang, J.; Yan, J. Image matching from handcrafted to deep features: A survey. Int. J. Comput. Vis. 2021, 129, 23–79. [Google Scholar] [CrossRef]
  7. Dawn, S.; Saxena, V.; Sharma, B. Remote sensing image registration techniques: A survey. In Proceedings of the International Conference on Image and Signal Processing, Québec, QC, Canada, 30 June–2 July 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 103–112. [Google Scholar]
  8. Wu, Y.; Liu, J.W.; Zhu, C.Z.; Bai, Z.F.; Miao, Q.G.; Ma, W.P.; Gong, M.G. Computational Intelligence in Remote Sensing Image Registration: A survey. Int. J. Autom. Comput. 2020, 18, 1–17. [Google Scholar] [CrossRef]
  9. Haskins, G.; Kruger, U.; Yan, P. Deep learning in medical image registration: A survey. Mach. Vis. Appl. 2020, 31, 8. [Google Scholar] [CrossRef] [Green Version]
  10. Sotiras, A.; Davatzikos, C.; Paragios, N. Deformable medical image registration: A survey. IEEE Trans. Med. Imaging 2013, 32, 1153–1190. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Fu, Y.; Lei, Y.; Wang, T.; Curran, W.J.; Liu, T.; Yang, X. Deep learning in medical image registration: A review. Phys. Med. Biol. 2020, 65, 20TR01. [Google Scholar] [CrossRef] [Green Version]
  12. Xiang, Y.M.; Wang, F.; You, H.J. OS-SIFT: A robust SIFT-like algorithm for high-resolution optical-to-SAR image registration in suburban areas. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3078–3093. [Google Scholar] [CrossRef]
  13. Sui, H.G.; Xu, C.; Liu, J.Y.; Hua, F. Automatic optical-to-SAR image registration by iterative line extraction and Voronoi integrated spectral point matching. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6058–6072. [Google Scholar] [CrossRef]
  14. Xu, C.; Sui, H.; Li, H.; Liu, J. An automatic optical and SAR image registration method with iterative level set segmentation and SIFT. Int. J. Remote Sens. 2015, 36, 3997–4017. [Google Scholar] [CrossRef]
  15. Liu, X.; Ai, Y.; Zhang, J.; Wang, Z. A novel affine and contrast invariant descriptor for infrared and visible image registration. Remote Sens. 2018, 10, 658. [Google Scholar] [CrossRef] [Green Version]
  16. Du, Q.; Fan, A.; Ma, Y.; Fan, F.; Huang, J.; Mei, X. Infrared and visible image registration based on scale-invariant piifd feature and locality preserving matching. IEEE Access 2018, 6, 64107–64121. [Google Scholar] [CrossRef]
  17. Min, C.; Gu, Y.; Li, Y.; Yang, F. Non-rigid infrared and visible image registration by enhanced affine transformation. Pattern Recognit. 2020, 106, 107377. [Google Scholar] [CrossRef]
  18. Jiang, X.; Ma, J.; Xiao, G.; Shao, Z.; Guo, X. A review of multimodal image matching: Methods and applications. Inf. Fusion 2021, 73, 22–71. [Google Scholar] [CrossRef]
  19. Zhao, F.; Huang, Q.M.; Gao, W. Image matching by normalized cross-correlation. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Toulouse, France, 14–19 May 2006; Volume 2. [Google Scholar]
  20. Luo, J.W.; Konofagou, E.E. A fast normalized cross-correlation calculation method for motion estimation. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2010, 57, 1347–1357. [Google Scholar]
  21. Cole-Rhodes, A.A.; Johnson, K.L.; LeMoigne, J.; Zavorin, I. Multiresolution registration of remote sensing imagery by optimization of mutual information using a stochastic gradient. IEEE Trans. Image Process. 2003, 12, 1495–1511. [Google Scholar] [CrossRef]
  22. Chen, H.M.; Arora, M.K.; Varshney, P.K. Mutual information-based image registration for remote sensing data. Int. J. Remote Sens. 2003, 24, 3701–3706. [Google Scholar] [CrossRef]
  23. Chen, H.M.; Varshney, P.K.; Arora, M.K. Performance of mutual information similarity measure for registration of multitemporal remote sensing images. IEEE Trans. Geosci. Remote Sens. 2003, 41, 2445–2454. [Google Scholar] [CrossRef]
  24. Yang, W.; Han, C.Z.; Sun, H.; Cao, Y.F. Registration of high resolution SAR and optical images based on multiple features. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Seoul, Korea, 29 July 2005; Volume 5, pp. 3542–3544. [Google Scholar]
  25. Lehureau, G.; Tupin, F.; Tison, C.; Guillaume, O.; Petit, D. Registration of metric resolution SAR and optical images in urban areas. In Proceedings of the European Conference on Synthetic Aperture Radar, Friedrichshafen, Germany, 2–5 June 2008; pp. 1–4. [Google Scholar]
  26. Yan, X.H.; Zhang, Y.J.; Zhang, D.J.; Hou, N.; Zhang, B. Registration of multimodal remote sensing images using transfer optimization. IEEE Geosci. Remote Sens. Lett. 2020, 17, 2060–2064. [Google Scholar] [CrossRef]
  27. Tondewad, P.S.; Dale, M.P. Remote sensing image registration methodology: Review and discussion. Procedia Comput. Sci. 2020, 171, 2390–2399. [Google Scholar] [CrossRef]
  28. Yu, L.; Zhang, D.R.; Holden, E.J. A fast and fully automatic registration approach based on point features for multi-source remote sensing images. Comput. Geosci. 2008, 34, 838–848. [Google Scholar] [CrossRef]
  29. Gonçalves, H.; Gonçalves, J.; Corte-Real, L. HAIRIS: A method for automatic image registration through histogram-based image segmentation. IEEE Trans. Image Process. 2011, 20, 776–789. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Moravec, H. Obstacle Avoidance and Navigation in the Real World by a Seeing Robot Rover. Ph.D. Thesis, Standford University, Stanford, CA, USA, 1980. [Google Scholar]
  31. Harris, C.; Stephens, M. A combined corner and edge detector. In Proceedings of the Alvey Vision Conference, Manchester, UK, 31 August–2 September 1988; pp. 147–151. [Google Scholar]
  32. Mikolajczyk, K.; Schmid, C. Scale & affine invariant interest point detectors. Int. J. Comput. Vis. 2004, 60, 63–86. [Google Scholar]
  33. Lowe, D.G. Object recognition from local scale-invariant features. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; pp. 1150–1157. [Google Scholar]
  34. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  35. Sedaghat, A.; Mokhtarzade, M.; Ebadi, H. Uniform robust scale-invariant feature matching for optical remote sensing images. IEEE Trans. Geosci. Remote Sens. 2011, 49, 4516–4527. [Google Scholar] [CrossRef]
  36. Sun, H.; Lei, L.; Zou, H.; Wang, C. Multimodal remote sensing image registration using multiscale self-similarities. In Proceedings of the IEEE International Conference on Computer Vision in Remote Sensing, Xiamen, China, 16–18 December 2012; pp. 199–202. [Google Scholar]
  37. Zeng, Q.; Adu, J.; Liu, J.; Yang, J.; Xu, Y.; Gong, M. Real-time adaptive visible and infrared image registration based on morphological gradient and C_SIFT. J. Real-Time Image Process. 2020, 17, 1103–1115. [Google Scholar] [CrossRef]
  38. Bay, H.; Tuytelaars, T.; Van Gool, L. SURF: Speeded up robust features. In Proceedings of the European Conference on Computer Vision, Graz, Austria, 7–13 May 2006; pp. 404–417. [Google Scholar]
  39. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  40. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; Sanchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Goodfellow, I.; Abadie, J.; Mirza, M.; Xu, B.; Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. arXiv 2014, arXiv:1406.2661v1. [Google Scholar]
  42. Oliehoek, F.A.; Savani, R.; Gallego-Posada, J.; Van der Pol, E.; De Jong, E.D.; Groß, R. GANGs: Generative adversarial network games. arXiv 2017, arXiv:1712.00679v2. [Google Scholar]
  43. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  44. Zhu, X.X.; Tuia, D.; Mou, L.; Xia, G.; Zhang, L.; Xu, F.; Fraundorfer, F. Deep learning in remote sensing: A comprehensive review and list of resources. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–36. [Google Scholar] [CrossRef] [Green Version]
  45. Hinton, G.E. A Practical Guide to Training Restricted Boltzmann Machines. In Neural Networks: Tricks of the Trade. Lecture Notes in Computer Science; Montavon, G., Orr, G.B., Müller, K.R., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; p. 7700. [Google Scholar]
  46. Sutskever, I.; Martens, J.; Hinton, G.E. Generating text with recurrent neural networks. In Proceedings of the 28th International Conference on Machine Learning, Washington, DC, USA, 28 June–2 July 2011; pp. 1017–1024. [Google Scholar]
  47. Abdi, G.; Samadzadegan, F.; Reinartz, P. Spectral-spatial feature learning for hyperspectral imagery classification using deep stacked sparse autoencoder. J. Appl. Remote Sens. 2017, 11, 042604. [Google Scholar]
  48. Gong, M.; Yang, H.; Zhang, P. Feature learning and change feature classification based on deep learning for ternary change detection in SAR images. ISPRS J. Photogramm. Remote Sens. 2017, 129, 212–225. [Google Scholar] [CrossRef]
  49. Zabalza, J.; Ren, J.; Zheng, J.; Zhao, H.; Qing, C.; Yang, Z.; Du, P.; Marshall, S. Novel segmented stacked autoencoder for effective dimensionality reduction and feature extraction in hyperspectral imaging. Neurocomputing 2016, 185, 1–10. [Google Scholar] [CrossRef] [Green Version]
  50. Mou, L.; Ghamisi, P.; Zhu, X.X. Deep recurrent neural networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef] [Green Version]
  51. Wang, C.; Zhang, L.; Wei, W.; Zhang, Y. When low rank representation based hyperspectral imagery classification meets segmented stacked denoising auto-encoder based spatial-spectral feature. Remote Sens. 2018, 10, 284. [Google Scholar] [CrossRef] [Green Version]
  52. Lin, D.; Fu, K.; Wang, Y.; Xu, G.; Sun, X. MARTA GANs: Unsupervised representation learning for remote sensing image classification. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2092–2096. [Google Scholar] [CrossRef] [Green Version]
  53. Zhan, Y.; Hu, D.; Wang, Y.; Yu, X. Semisupervised hyperspectral image classification based on Generative Adversarial Networks. IEEE Geosci. Remote Sens. Lett. 2018, 15, 212–216. [Google Scholar] [CrossRef]
  54. Shadaydeh, M.; Sziranyi, T. An improved mutual information similarity measure for registration of multi-modal remote sensing images. In Image and Signal Processing for Remote Sensing XXI; Bruzzone, L., Ed.; International Society for Optics and Photonics: Washington, DC, USA, 2015; Volume 9643, p. 96430F. [Google Scholar]
  55. Xu, X.; Li, X.; Liu, X.; Shen, H.; Shi, Q. Multimodal registration of remotely sensed images based on jeffrey’s divergence. ISPRS J. Photogramm. Remote Sens. 2016, 122, 97–115. [Google Scholar] [CrossRef]
  56. Xie, X.; Zhang, Y.; Ling, X.; Wang, X. A new registration algorithm for multimodal remote sensing images. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 7011–7014. [Google Scholar]
  57. Xie, X.; Zhang, Y.; Ling, X.; Wang, X. A novel extended phase correlation algorithm based on log-gabor filtering for multimodal remote sensing image registration. Int. J. Remote Sens. 2019, 40, 5429–5453. [Google Scholar] [CrossRef]
  58. Hasan, M.; Pickering, M.R.; Jia, X. Robust automatic registration of multimodal satellite images using ccre with partial volume interpolation. IEEE Trans. Geosci. Remote Sens. 2012, 50, 4050–4061. [Google Scholar] [CrossRef]
  59. Dame, A.; Marchand, E. Second-order optimization of mutual information for real-time image registration. IEEE Trans. Image Process. 2012, 21, 4190–4203. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  60. Liang, J.; Liu, X.; Huang, K.; Li, X.; Wang, D.; Wang, X. Automatic registration of multisensor images using an integrated spatial and mutual information (SMI) metric. IEEE Trans. Geosci. Remote Sens. 2014, 52, 603–615. [Google Scholar] [CrossRef]
  61. Wu, Y.; Ma, W.; Miao, Q.; Wang, S. Multimodal continuous ant colony optimization for multisensor remote sensing image registration with local search. Swarm Evol. Comput. 2019, 47, 89–95. [Google Scholar] [CrossRef]
  62. Liu, X.; Chen, S.; Zhuo, L.; Li, J.; Huang, K. Multi-sensor image registration by combining local self-similarity matching and mutual information. Front. Earth Sci. 2018, 12, 779–790. [Google Scholar] [CrossRef]
  63. Ye, Y.; Bruzzone, L.; Shan, J.; Bovolo, F.; Zhu, Q. Fast and robust matching for multimodal remote sensing image registration. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9059–9070. [Google Scholar] [CrossRef] [Green Version]
  64. Liang, D.; Ding, J.; Zhang, Y. Efficient Multisource Remote Sensing Image Matching Using Dominant Orientation of Gradient. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 2194–2205. [Google Scholar] [CrossRef]
  65. De Castro, E.; Morandi, C. Registration of translated and rotated images using finite fourier transforms. IEEE Trans. Pattern Anal. Mach. Intell. 1987, 9, 700–703. [Google Scholar] [CrossRef]
  66. Reddy, B.S.; Chatterji, B.N. An FFT-based technique for translation, rotation, and scale-invariant image registration. IEEE Trans. Image Process. 1996, 5, 1266–1271. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  67. Foroosh, H.; Zerubia, J.B.; Berthod, M. Extension of phase correlation to subpixel registration. IEEE Trans. Image Process. 2002, 11, 188–200. [Google Scholar] [CrossRef] [Green Version]
  68. Thevenaz, P.; Unser, M. Optimization of mutual information for multiresolution image registration. IEEE Trans. Image Process. 2020, 9, 2083–2099. [Google Scholar]
  69. Hossain, M.T.; Lv, G.; Teng, S.; Lu, G.; Lackmann, M. Improved symmetric-SIFT for multi-modal image registration. In Proceedings of the International Conference on Digital Image Computing: Techniques and Applications, Noosa, Australia, 6–8 December 2011; pp. 197–202. [Google Scholar]
  70. Huang, L.; Li, Z.; Zhang, R. SAR and optical images registration using shape context. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Honolulu, HI, USA, 25–30 July 2010; pp. 1007–1010. [Google Scholar]
  71. Aguilera, C.; Barrera, F.; Lumbreras, F.; Sappa, A.D.; Toledo, R. Multispectral image feature points. Sensors 2012, 12, 12661–12672. [Google Scholar] [CrossRef] [Green Version]
  72. Ye, Y.; Shan, J. A local descriptor based registration method for multispectral remote sensing images with non-linear intensity differences. ISPRS J. Photogramm. Remote Sens. 2014, 90, 83–95. [Google Scholar] [CrossRef]
  73. Sedaghat, A.; Ebadi, H. Distinctive order based self-similarity descriptor for multi-sensor remote sensing image matching. ISPRS J. Photogramm. Remote Sens. 2015, 108, 62–71. [Google Scholar] [CrossRef]
  74. Ye, Y.; Li, S.; Hao, M.; Wang, J.; Xu, Z. Robust optical-to-SAR image matching based on shape properties. IEEE Geosci. Remote Sens. 2017, 14, 564–568. [Google Scholar] [CrossRef]
  75. Xiong, X.; Xu, Q.; Jin, G.; Zhang, H.; Gao, X. Rank-based local self-similarity descriptor for optical-to-SAR image matching. IEEE Geosci. Remote Sens. 2020, 17, 1742–1746. [Google Scholar] [CrossRef]
  76. Cui, S.; Zhong, Y. Multi-modal remote sensing image registration based on multi-scale phase congruency. In Proceedings of the 10th IAPR Workshop on Pattern Recognition in Remote Sensing (PRRS), Beijing, China, 19–20 August 2018. [Google Scholar]
  77. Ye, Y.; Shan, J.; Hao, S.; Bruzzone, L.; Qin, Y. A local phase based invariant feature for remote sensing image matching. ISPRS J. Photogramm. Remote Sens. 2018, 142, 205–221. [Google Scholar] [CrossRef]
  78. Fan, J.; Wu, Y.; Li, M.; Liang, W.; Cao, Y. SAR and optical image registration using nonlinear diffusion and phase congruency structural descriptor. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5368–5379. [Google Scholar] [CrossRef]
  79. Li, J.; Hu, Q.; Ai, M. RIFT: Multi-modal image matching based on radiation-variation insensitive feature transform. IEEE Trans. Image Process. 2020, 29, 3296–3310. [Google Scholar] [CrossRef]
  80. Zhao, C.; Zhao, H.; Lv, J.; Sun, S.; Li, B. Multimodal image matching based on multimodality robust line segment descriptor. Neurocomputing 2016, 177, 290–303. [Google Scholar] [CrossRef]
  81. Xu, G.; Wu, Q.; Cheng, Y.; Yan, F.; Li, Z.; Yu, Q. A robust deformed image matching method for multi-source image matching. Infrared Phys. Technol. 2021, 115, 103691. [Google Scholar] [CrossRef]
  82. Chen, J.; Tian, J. Real-time multi-modal rigid registration based on a novel symmetric-SIFT descriptor. Prog. Nat. Sci. 2009, 19, 643–651. [Google Scholar] [CrossRef]
  83. Fjortoft, R.; Lopes, A.; Marthon, P.; Cubero-Castan, E. An optimal multiedge detector for SAR image segmentation. IEEE Trans. Geosci. Remote Sens. 1998, 36, 793–802. [Google Scholar] [CrossRef] [Green Version]
  84. Xiang, Y.; Wang, F.; Wan, L.; Jiao, N.; You, H. Os-flow: A robust algorithm for dense optical and SAR image registration. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6335–6354. [Google Scholar] [CrossRef]
  85. Xiang, Y.; Tao, R.; Wan, L.; Wang, F.; You, H. Os-pc: Combining feature representation and 3-D phase correlation for subpixel optical and SAR image registration. IEEE Trans. Geosci. Remote Sens. 2020, 58, 6451–6466. [Google Scholar] [CrossRef]
  86. Shechtman, E.; Irani, M. Matching local self-similarities across images and videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; Volume 1, pp. 1–8. [Google Scholar]
  87. Tola, E.; Lepetit, V.; Fua, P. DAISY: An efficient dense descriptor applied to wide-baseline stereo. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 815–830. [Google Scholar] [CrossRef] [Green Version]
  88. Perona, P.; Malik, J. Scale-space and edge detection using anisotropic diffusion. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 629–639. [Google Scholar] [CrossRef] [Green Version]
  89. Alcantarilla, P.F.; Bartoli, A.; Davison, A.J. KAZE features. In Proceedings of the European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; pp. 214–227. [Google Scholar]
  90. von Gioi, R.; Jakubowicz, J.; Morel, J.M.; Randall, G. LSD: A fast line segment detector with a false detection control. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 722–732. [Google Scholar] [CrossRef] [PubMed]
  91. Xiong, B.; Li, W.; Zhao, L.; Lu, J.; Zhang, X.; Kuang, G. Registration for SAR and optical images based on straight line features and mutual information. In Proceedings of the IGARSS, IEEE International Geoscience and Remote Sensing Symposium, Beijing, China, 10–15 July 2016; pp. 2582–2585. [Google Scholar]
  92. Zhang, M.X.; Wang, Z.G.; Bai, R.Y. A coarse-to-fine optical and SAR remote sensing image registration algorithm. J. Geo-Inf. Sci. 2020, 22, 2238–2246. [Google Scholar]
  93. Ye, F.; Su, Y.; Xiao, H.; Zhao, X.; Min, W. Remote sensing image registration using convolutional neural network features. IEEE Geosci. Remote Sens. Lett. 2018, 15, 232–236. [Google Scholar] [CrossRef]
  94. Ma, W.; Zhang, J.; Wu, Y.; Jiao, L.; Zhu, H.; Zhao, W. A novel two-step registration method for remote sensing images based on deep and local features. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4834–4843. [Google Scholar] [CrossRef]
  95. Yang, Z.; Dan, T.; Yang, Y. Multi-temporal remote sensing image registration using deep convolutional features. IEEE Access. 2018, 6, 38544–38555. [Google Scholar] [CrossRef]
  96. Quan, D.; Wang, S.; Liang, X.F.; Wang, R.J.; Fang, S.; Hou, B.; Jiao, L.C. Deep generative matching network for optical and SAR image registration. Proceedings of The IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 6215–6218. [Google Scholar]
  97. Merkle, N.; Auer, S.; Müller, R.; Reinartz, P. Exploring the potential of conditional adversarial networks for optical and SAR image matching. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 1811–1820. [Google Scholar] [CrossRef]
  98. Hughes, L.H.; Schmitt, M.; Mou, L.C.; Wang, Y.Y.; Zhu, X.X. Identifying corresponding patches in SAR and optical images with a pseudo-Siamese CNN. IEEE Geosci. Remote Sens. Lett. 2018, 15, 784–788. [Google Scholar] [CrossRef] [Green Version]
  99. Zhang, H.; Ni, W.; Yan, W.; Xiang, D.; Wu, J.; Yang, X.; Bian, H. Registration of multimodal remote sensing image based on deep fully convolutional neural network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 3028–3042. [Google Scholar] [CrossRef]
  100. Merkle, N.; Luo, W.J.; Auer, S.; Muller, R.; Urtasun, R. Exploiting deep matching and sar data for the geo-localization accuracy improvement of optical satellite images. Remote Sens. 2017, 9, 586. [Google Scholar] [CrossRef] [Green Version]
  101. He, H.; Chen, M.; Chen, T.; Li, D. Matching of remote sensing images with complex background variations via siamese convolutional neural network. Remote Sens. 2018, 10, 355. [Google Scholar] [CrossRef] [Green Version]
  102. Zhang, J.; Ma, W.; Wu, Y.; Jiao, L. Multimodal Remote Sensing Image Registration Based on Image Transfer and Local Features. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1210–1214. [Google Scholar] [CrossRef]
  103. Wang, S.; Quan, D.; Liang, X.; Ning, M.; Guo, Y.; Jiao, L. A deep learning framework for remote sensing image registration. ISPRS J. Photogramm. Remote Sens. 2018, 145, 148–164. [Google Scholar] [CrossRef]
  104. Zampieri, A.; Charpiat, G.; Girard, N.; Tarabalka, Y. Multimodal Image Alignment through a Multiscale Chain of Neural Networks with Application to Remote Sensing. 2018. Available online: http://openaccess.thecvf.com/content_ECCV_2018/papers/Armand_Zampieri_Multimodal_image_alignment_ECCV_2018_paper.pdf (accessed on 14 December 2021).
  105. Luo, W.J.; Schwing, A.G.; Urtasun, R. Efficient deep learning for stereo matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 5695–5703. [Google Scholar]
  106. Tombari, F.; Stefano, L.D. The Maximal Self-dissimilarity Interest Point Detector. IPSJ Trans. Comput. Vis. Appl. 2015, 7, 175–188. [Google Scholar] [CrossRef] [Green Version]
  107. Sedghi, A.; Luo, J.; Mehrtash, A.; Pieper, S.; Tempany, C.M.; Kapur, T.; Mousavi, P.; Wells, W.M., III. Semi-supervised deep metrics for image registration. arXiv 2018, arXiv:1804.01565. [Google Scholar]
  108. Wang, Y.; Zhang, J.K.; Cavichini, M.; Bartsch, D.G.; Freeman, W.R.; Nguyen, T.Q.; An, C. Robust Content-Adaptive Global Registration for Multimodal Retinal Images Using Weakly Supervised Deep-Learning Framework. IEEE Trans. Image Process. 2021, 30, 3167–3178. [Google Scholar] [CrossRef]
  109. Huang, W.; Yang, H.; Liu, X.; Li, C.; Zhang, I.; Wang, R.; Zheng, H.; Wang, S. A Coarse-to-Fine Deformable Transformation Framework for Unsupervised Multi-Contrast MR Image Registration with Dual Consistency Constraint. IEEE Trans. Med. Imaging 2021, 40, 2589–2599. [Google Scholar] [CrossRef]
  110. Jhan, J.P.; Rau, J. A Generalized Tool for Accurate and Efficient Image Registration of UAV Multi-lens Multispectral Cameras by N-SURF Matching. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 6353–6362. [Google Scholar] [CrossRef]
  111. Zhou, H.; Ma, J.; Tan, C.C.; Zhang, Y.; Ling, H. Cross-weather image alignment via latent generative model with intensity consistency. IEEE Trans. Image Process. 2020, 29, 5216–5228. [Google Scholar] [CrossRef]
  112. Dellinger, F.; Delon, J.; Gousseau, Y.; Michel, J.; Tupin, F. Sar-Sift: A Sift-Like Algorithm for Sar Images. IEEE Trans. Geosci. Remote Sens. 2015, 53, 453–466. [Google Scholar] [CrossRef] [Green Version]
  113. Ma, W.; Wen, Z.; Wu, Y.; Jiao, L.; Gong, M.; Zheng, Y.; Liu, L. Remote Sensing Image Registration with Modified SIFT and Enhanced Feature Matching. IEEE Geosci. Remote Sens. Lett. 2017, 14, 3–7. [Google Scholar] [CrossRef]
  114. Ma, J.; Zhou, H.; Zhao, J.; Gao, Y.; Jiang, J.; Tian, J. Robust Feature Matching for Remote Sensing Image Registration via Locally Linear Transforming. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6469–6481. [Google Scholar] [CrossRef]
  115. Li, Y.; Ma, J.; Zhang, Y. Image retrieval from remote sensing big data: A survey. Inf. Fusion 2021, 67, 94–115. [Google Scholar] [CrossRef]
Figure 1. Structural framework of this review.
Figure 1. Structural framework of this review.
Remotesensing 13 05128 g001
Figure 2. Framework of area-based image registration.
Figure 2. Framework of area-based image registration.
Remotesensing 13 05128 g002
Figure 3. Framework of feature-based image registration.
Figure 3. Framework of feature-based image registration.
Remotesensing 13 05128 g003
Figure 4. An example of radiation difference between MMRS images: (a) Optical image. (b) SAR image. As shown by the enlarged patch, the radiation differences and noise effects are particularly noticeable.
Figure 4. An example of radiation difference between MMRS images: (a) Optical image. (b) SAR image. As shown by the enlarged patch, the radiation differences and noise effects are particularly noticeable.
Remotesensing 13 05128 g004
Figure 5. Selected original image pairs from the database, which covers six multi-modal pairs in the field of remote sensing.
Figure 5. Selected original image pairs from the database, which covers six multi-modal pairs in the field of remote sensing.
Remotesensing 13 05128 g005
Figure 6. The matching results of 8 methods on typical multi-modal image pairs in the remote sensing research field. (No matching result indicates that the method is not applicable to this multi-modal image pair).
Figure 6. The matching results of 8 methods on typical multi-modal image pairs in the remote sensing research field. (No matching result indicates that the method is not applicable to this multi-modal image pair).
Remotesensing 13 05128 g006
Table 1. Analysis of area-based method improvements.
Table 1. Analysis of area-based method improvements.
AuthorMethod CategoryMethod ImprovementConclusion
Cole-Rhodes et al. [21]Methods to improve similarity measuresImprove the joint histogram of MI and propose the simultaneous perturbation stochastic approximation (SPSA) algorithm.The algorithm relies on gradient approximation instead of the target gradient value and finds the global maximum value through the local extrema of the objective function, which greatly speeds up the entire registration process.
Chen et al. [23]Methods to improve similarity measuresA new joint histogram estimation algorithm is proposed, called Generalized Partial Volume Estimation (GPVE, with a second-order B-spline function).Overcome the problem of interpolation-induced artifacts.
Shadaydeh et al. [54]Methods to improve similarity measuresPropose a weight-based joint histogram estimation method (WJH-MI).Reduces the peaks in the joint histogram caused by the background or homogeneous regions by assigning more weight to the high gradient pixels containing more registration related information.
Xu et al. [55]Methods to improve similarity measuresPropose using the symmetric form of Kullback-Leiber divergence, namely Jeffrey’s divergence as similarity measure.The registration model based on Jeffrey’s divergence can provide a larger feasible search space and solve the problem of MI facing insufficient image overlapping area.
Xie et al. [56]Methods to improve similarity measuresMLPC method combining multi-scale Log-Gabor filter and phase consistencyEffectively resolving the non-linear intensity difference in MMRS image registration.
Xie et al. [57]Methods to improve similarity measuresBased on the extended phase correlation of log-Gabor, an improved LGEPC method is proposedUse the overall structure information to eliminate the influence of radiation differences as much as possible.
Hasan et al. [58]Methods to improve optimizationUsing CCRE to align SAR and Google satellite images and applying partial volume interpolation to calculate the gradient of the similarity measure.Directly implement the optimization process based on partial volume interpolation.
Dame and Marchand [59]Methods to improve optimizationDefine a new inverse combinatorial optimization method to handle the quasi-concave shape of MI, where the required derivative can be pre-calculated allowing the Hessian matrix to be estimated after convergence.Reduce the calculation time and estimate an accurate parameter.
Liang et al. [60] and Wu et al. [61]Methods to improve optimizationThe authors used the ant colony optimization (ACO) algorithm to optimize the similarity measure to maximize MI.The similarity curve of MI has been proved to have many local optimal values.
Yan et al. [26]Methods to improve optimizationPropose a method of using transfer optimization (TO) to maximize MI, optimize conversion parameters and transfer better results to another optimizer in the iterative process.Enhance the global search capability of TO and avoid a local optima.
Liu et al. [62] Hybrid methodsPropose an image registration method that combines local self-similarity (LSS) and MI, which combined local internal features and global intensity information.The proposed method can register multi-sensor images with different resolutions and handle the geometric differences between grayscale and corresponding pixels and regions.
Ye et al. [63] Hybrid methodsA fast and robust template matching framework for MMRS images called channel features of orientated gradients (CFOG).A new general scheme of template matching based on pixel feature representation, suitable for all kinds of multi-modal image registration.
Yan et al. [2] Hybrid methodsIntroduced the histogram of oriented gradient distance (HOGD) and the grey wolf optimizer (GWO).It avoids falling into the local optimum and reduces the calculation time.
Liang et al. [64] Hybrid methodsA variable template matching method based on DOG features and the sum of cosine difference.The proposed matching method has good robustness to nonlinear light intensity changes, and effectively improves the matching accuracy.
Table 2. Analysis of feature-based method improvements.
Table 2. Analysis of feature-based method improvements.
AuthorMethod CategoryMethod ImprovementConclusion
Yu et al. [28]Methods to improve SIFT or Harris operatorCombine the advantages of SIFT and Harris to propose a coarse to fine MMRS image registration.The proposed method is more suitable for MMRS image registration.
Sedaghat et al. [35]Methods to improve SIFT or Harris operatorPropose a registration method, called Uniform Robust SIFT (UR-SIFT), suitable for a variety of optical multi-source remote sensing images with illumination, rotation and up to five times the scale difference.In the process of descriptor merging, this algorithm led to many other combinations to be merged into the same final descriptor, which greatly reduced the registration accuracy.
Hossain et al. [69]Methods to improve SIFT or Harris operatorAn improved symmetric SIFT method.The descriptor merging step in symmetric SIFT can be skipped to obtain excellent matching accuracy.
Fan et al. [5]Methods to improve SIFT or Harris operatorA novel spatial consensus matching (SCM) algorithm.Using improved SIFT and K-nearest neighbors (KNN) to obtain an initial set of matching features and utilized spatial consistency constraints for refinement.
Huang et al. [70]Methods to improve SIFT or Harris operatorA new algorithm for improving feature extraction and feature matching was introduced.Using Harris operator, Canny operator and shape context descriptor to match multimodal images.
Xiang et al. [12]Methods to improve SIFT or Harris operatorPropose a SIFT-like algorithm (OS-SIFT), which introduce multi-scale ratio of exponentially weighted averages (ROEWA) and multi-scale Sobel operators.By calculating the consistent gradients of SAR and optical images, it is proved that the algorithm is robust to noise and has excellent registration performance and accuracy.
Aguilera et al. [71]Methods to improve feature descriptorThe edge-oriented histogram (EOH) descriptorThe descriptor contained the information of the contour near each feature point without using the gradient information to describe the shape and contour of the image.
Ye et al. [72]Methods to improve feature descriptorCombine Harris and LSS descriptors to establish a piecewise linear transformationThe impact of the low discriminability of the LSS descriptor is reduced, and the reliable registration of multi-spectral remote sensing images is realized.
Sedaghat et al. [73]Methods to improve feature descriptorAn advanced version of the self-similarity descriptor, which has high distinguishability, called distinctive order based self similarity (DOBSS) descriptor.The DOBSS descriptor has better recall, precision and positioning accuracy.
Ye et al. [74]Methods to improve feature descriptorBased on the internal self-similarity of images, the author introduced a shape descriptor of dense local self-similarity (DLSS).Using the internal structure information of the image to construct the descriptor is more suitable for multi-mode image registration.
Xiong et al. [75]Methods to improve feature descriptorA rank-based local self-similarity (RLSS) to describe the local shape of an image in a distinguishable manner.The rank value was used as a substitute for the correlation value to indicate the relative relationship of the correlation value, which further improved distinguishability of feature descriptors.
Cui et al. [76]Methods to improve feature descriptorA multi-scale phase-congruency descriptor (MS-PC), which captured the shape and structural characteristics of an image.The descriptor compensates for the sensitivity of traditional descriptors to radiation differences.
Ye et al. [4]Methods to improve feature descriptorCombine the PC describing structural features with the histogram of gradient directions (HOG) as a feature descriptor, called HOPC.Using the structure information of the image to construct the descriptor can eliminate modal differences and is more suitable for multi-mode image registration.
Liu et al. [15]Methods to improve feature descriptorA maximum stable phase congruency (MSPC), which combined affine invariant region extraction and image structural features.The algorithm extracted structural features by merging phase congruency images in multiple orientations. Registration was achieved according to the correspondence of the descriptors.
Ye et al. [77]Hybrid methodsA local invariant feature that was robust to geometric distortion and radiation changes, which consisted of a feature detector named MMPC-lap and a feature descriptor named local histogram of orientated phase congruency (LHOPC).This method also solved the radiation differences caused by spectral and time changes between MMRS images.
Fan et al. [78]Hybrid methodsAn UND-Harris detector that introduced nonlinear diffusion, feature ratio and block strategy.Experimental results on different SAR and optical image pairs showed the effectiveness of this method, which can obtain better registration results and improve registration accuracy.
Zeng et al. [37]Hybrid methodsAn infrared-to-visible image registration method based on morphological gradient and C-SIFT.The algorithm uses morphological methods to preserve the gray-scale edges of the image and improve the similarity of infrared and visible images.
Li et al. [79]Hybrid methodsA radiation insensitive image registration method based on phase congruency (PC) and a maximum index image (MIM), which was called radiation variation insensitive feature transform (RIFT).This method realizes the insensitivity and rotation invariance to multi-modal image radiation changes.
Sui et al. [13]Hybrid methodsAn iterative process combining line segment extraction and line intersection matching based online segment extraction, and integrated Voronoi polygons into spectral point matching (SPM) to obtain the correspondences between line intersections.An iterative strategy of “re-extraction” and “re-matching” mechanisms was adopted to enhance feature extraction and matching performance.
Zhao et al. [80]Hybrid methodsUsing Kovesi corner point extraction and line segment detection methods based on phase congruency and local direction.Compared to other edge extraction methods, this method extracted more equivalent line segments.
Xu et al. [81]Hybrid methodsA new contour segment representation method based on local histogram of maximal edge orientation and defined angles is proposed, and the Fréchet distance is defined as the weighting parameter of combined histograms to enhance the descriptive ability.The proposed method can effectively reduce the impacts of radiation distortion and is superior to some current popular multi-source image matching methods.
Table 3. Analysis of integrated method improvements.
Table 3. Analysis of integrated method improvements.
AuthorMethod CategoryMethod ImprovementConclusion
Gong et al. [3]Combined area-based and feature-based methodsA novel coarse-to-fine scheme for automatic image registration.In coarse registration (pre-registration), the scale histogram was used to remove the outliers detected by SIFT. Due to the robustness of MI to noise and its adaptability to different image intensity values, adopted it to refine the pre-registration results in the multiresolution framework.
Ye et al. [72]Combined area-based and feature-based methodsA local descriptor-based registration method for multispectral remote sensing images.Pre-registration used SR-SIFT and projection transformation to eliminate the obvious rotation and scale differences between the reference and sensed images. Then the second stage uses LSS as a new similarity measure.
Xiong et al. [91]Combined area-based and feature-based methodsCoarse registration based on intersections of straight lines and fine registration based on MI from separated patches.Using the stable and consistent edge information of the optical and SAR images, the registration accuracy is improved.
Zhang et al. [92]Combined area-based and feature-based methodsUsing the feature-based SAR-SIFT algorithm to complete the coarse registration, and then utilize the area-based ROEWA-HOG method to complete the fine registration.Achieve high-precision automatic registration of the hybrid model.
Table 4. Analysis of deep learning-based method improvements.
Table 4. Analysis of deep learning-based method improvements.
AuthorMethod CategoryMethod ImprovementConclusion
Ye et al. [93]Methods to improve CNN and GAN algorithmsUsing CNN to extract the middle and high-level features of an image and combined them with the low-level features extracted by SIFT.Features of CNN and SIFT were incorporated into the PSO-SIFT algorithm for registration.
Ma et al. [94]Methods to improve CNN and GAN algorithmsUsing VGG-16 to approximate spatial relationships and proposed a new point matching strategy based on spatial relationships and combined the local feature-based methods.Due to the powerful feature extraction capabilities of CNN and the consideration of spatial relationships, the matching results are robust and accurate.
Yang et al. [95]Methods to improve CNN and GAN algorithmsAdopting a pre-trained VGG network to generate multi-scale descriptors through high-level convolution information features.Optimize the registration details by increasing the number of feature points.
Quan et al. [96]Methods to improve CNN and GAN algorithmsApplying GAN for MMRS images data augmentation, which could immensely enhance the accuracy and robust of registration process.A dual-channel deep network and CNN can save the spatial information of the image.
Merkle et al. [97]Methods to improve CNN and GAN algorithmsA GAN-based method for dealing with optical and SAR image registration.The generator in GAN accurately and reliably retained the geometric structure of the optical image, opening up new possibilities for MMRS.
Hughes et al. [98]Methods to improve Siamese networkA specific pseudo-Siamese network to register the optical and SAR images.The two convolutional streams of this network are identical and independent, and there is no parameter sharing to process the different intensity information of the two images.
Zhang et al. [99]Methods to improve Siamese networkA fully convolutional Siamese network, which used an end-to-end training process. It built a general framework for MMRS image registration based on depth features.Sharing parameters between the two branches to solve the problem of lack of data sets.
Merkle et al. [100]Methods to improve Siamese networkBy training the Siamese network to learn the spatial transformation between optical and SAR images.The effectiveness of this approach for the generation of reliable and robust matching points between optical and SAR images had been demonstrated by experiment.
He et al. [101]Methods to improve Siamese networkA novel multi-scale remote sensing image registration deep network, which includes the following three steps: corner detection based on s-Harris, a search strategy based on Gaussian pyramid coupled quadtree. Finally, global to local quadratic polynomial constraints and RANSAC were utilized to remove mismatches.This algorithm can realize the scale comparison of multi-scale conjugate patches, make the matching evenly distributed, and obtain satisfactory matching accuracy. The most important thing is to avoid the influence of complex background changes on the registration results.
Zhang et al. [102]OthersExtending the image visual attribute transfer method to pre-process an image.Eliminate the intensity differences between the multi-modal image pairs, and make the color, texture and other characteristics consistent in a similar structural area.
Wang et al. [103]OthersA remote sensing image registration framework based on deep learning, which directly learns the mapping function between image patch pairs and their matching labels through closed-loop information.A self-learning method was introduced to avoid a few image data and can learn the mapping function from itself without relying on other data. At the same time, the application of transfer learning further improves the registration accuracy and reduces training costs.
Zampieri et al. [104]OthersDesigning a neural network with a specific scale to learn image features.Complete the alignment between remote sensing images and maps
Table 5. Analysis of image registration algorithms in other fields.
Table 5. Analysis of image registration algorithms in other fields.
AuthorMethod CategoryMethod ImprovementConclusion
Tombari et al. [106]Multimodal registration of photogrammetric imagesThe local maximum feature of the Contextual Self-dissimilarity (CSD) operator is detected through the non-maximum suppression (NMS) stage, which is called the Maximal Self-Dissimilarity interest point detector (MSD).MSD is taking the lead for photogrammetric based multimodal registration.
Sedghi et al. [107]Multimodal registration of medical imagesBuilding a five-layer neural network to learn a similarity metric that can measure the level of registration and using Powell’s method to optimize the learned metric in an iterative manner.The results demonstrate the feasibility of learning a useful deep metric from substantially misaligned training data, and results are significantly better than from Mutual Information.
Wang et al. [108]Multimodal registration of medical imagesA content-adaptive weakly-supervised deep learning framework is constructed for multi-modal retinal image registration. The framework is composed of three neural networks for blood vessel segmentation, feature detection and description, and outlier elimination.This method is superior to other learning-based methods, achieved the highest success rate and Dice coefficient, and has significant robustness in poor quality images.
Huang et al. [109]Multimodal registration of medical imagesAn end-to-end network architecture composed of affine transformation and deformable transformation is proposed.Double consistency constraints and a new loss function based on prior knowledge have been developed to achieve accurate and efficient multi-contrast MR image registration.
Jhan et al. [110]Multimodal registration of computer vision imagesA new N-SURF matching method for multi-spectral camera (MSCs) image registration and a general tool for image registration of various MSCs.This method has good accuracy and can obtain more the number of correct matches (CMs) that are evenly distributed, which has the advantages of accuracy and efficiency.
Zhou et al. [111]Multimodal registration of computer vision imagesA potential generation model for cross-weather image alignment based on intensity constancy and image manifold characteristics.Experimental results demonstrate that this approach can significantly outperform the state-of-the-art methods.
Table 6. CMR at different MMRS image pairs.
Table 6. CMR at different MMRS image pairs.
Image PairsDepth-Opti.Opti-Opti.IR-Opti.Map-Opti.Night-DaySAR-Opti.
Method
SIFT7/12215/836/10335/1466/37/
SURF42/83081/102523/94365/375/20/1297
SAR-SIFT4/4325/127//4/20/
PSO-SIFT4/13712/1137/6416/1245/49/
RIFT258/1152285/956319/800554/1351238/958596/1520
RIFT-LAF607/1152496/956283/800410/1351387/958897/1520
HOPC89/2198/186128/230174/498153/202286/328
CFOG98/258213/956135/319193/554168/238303/596
Table 7. RMSE or Precision at different MMRS image pairs.
Table 7. RMSE or Precision at different MMRS image pairs.
Image PairsDepth-Opti.Opti-Opti.IR-Opti.Map-Opti.Night-DaySAR-Opti.
Method
SIFT0.440.520.490.580.22/
SURF0.340.140.240.20/0.21
SAR-SIFT0.120.62//0.06/
PSO-SIFT0.350.640.550.560.63/
RIFT0.640.630.640.700.680.66
RIFT-LAF0.300.360.530.590.380.40
HOPC//////
CFOG//////
Table 8. Operation times at different MMRS image pairs.
Table 8. Operation times at different MMRS image pairs.
Image PairsDepth-Opti.Opti-Opti.IR-Opti.Map-Opti.Night-DaySAR-Opti.
Method
SIFT13.6914.048.3912.9011.44/
SURF2.692.242.151.69/2.55
SAR-SIFT15.0212.59//11.66/
PSO-SIFT12.9112.869.1710.289.34/
RIFT10.939,318.3213.049.6311.03
RIFT-LAF11.049.418.6513.259.8711.31
HOPC61.6561.5040.2455.6062.6962.99
CFOG10.934.291.303.952.732.47
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, X.; Leng, C.; Hong, Y.; Pei, Z.; Cheng, I.; Basu, A. Multimodal Remote Sensing Image Registration Methods and Advancements: A Survey. Remote Sens. 2021, 13, 5128. https://doi.org/10.3390/rs13245128

AMA Style

Zhang X, Leng C, Hong Y, Pei Z, Cheng I, Basu A. Multimodal Remote Sensing Image Registration Methods and Advancements: A Survey. Remote Sensing. 2021; 13(24):5128. https://doi.org/10.3390/rs13245128

Chicago/Turabian Style

Zhang, Xinyue, Chengcai Leng, Yameng Hong, Zhao Pei, Irene Cheng, and Anup Basu. 2021. "Multimodal Remote Sensing Image Registration Methods and Advancements: A Survey" Remote Sensing 13, no. 24: 5128. https://doi.org/10.3390/rs13245128

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop