Next Article in Journal
Dark Spot Detection from SAR Images Based on Superpixel Deeper Graph Convolutional Network
Next Article in Special Issue
Self-Supervised Depth Completion Based on Multi-Modal Spatio-Temporal Consistency
Previous Article in Journal
Assessing Changes in Boreal Vegetation of Kola Peninsula via Large-Scale Land Cover Classification between 1985 and 2021
Previous Article in Special Issue
Dithered Depth Imaging for Single-Photon Lidar at Kilometer Distances
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient and Robust Feature Matching for High-Resolution Satellite Stereos

1
State Key Laboratory of Geo-Information Engineering, Xi’an 710054, China
2
Xi’an Institute of Surveying and Mapping, Xi’an 710054, China
3
School of Geospatial Engineering and Science, Sun Yat-Sen University, Zhuhai 519082, China
4
School of Remote Sensing and Information Engineering, Wuhan University, Wuhan 430079, China
5
College of Geodesy and Geomatics, Shandong University of Science and Technology, Qingdao 266590, China
6
Research Center of Engineering and Technology for “3S”-GPS, RS and GIS, Shandong University of Science and Technology, Qingdao 266590, China
7
Star-Rising Technologies Co., Ltd., Qingdao 266590, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2022, 14(21), 5617; https://doi.org/10.3390/rs14215617
Submission received: 18 September 2022 / Revised: 1 November 2022 / Accepted: 4 November 2022 / Published: 7 November 2022

Abstract

:
Feature matching between high-resolution satellite stereos plays an important role in satellite image orientation. However, images of changed regions, weak-textured regions and occluded regions may generate low-quality matches or even mismatches. Furthermore, matching throughout the entire satellite images often has extremely high time cost. To compute good matching results at low time cost, this paper proposes an image block selection method for high-resolution satellite stereos, which processes feature matching in several optimal blocks instead of the entire images. The core of the method is to formulate the block selection into the optimization of an energy function, and a greedy strategy is designed to compute an approximate solution. The experimental comparisons on various satellite stereos show that the proposed method could achieve similar matching accuracy and much lower time cost when compared with some state-of-the-art satellite image matching methods. Thus, the proposed method is a good compromise between matching accuracy and matching time, which has great potential in large-scale satellite applications.

1. Introduction

Feature matching is a preliminary step in satellite image orientation that aims to find correspondences between satellite images [1]. The feature matching accuracy will significantly influence the orientation results. Therefore, it is still one of the research hotspots in both photogrammetry and computer vision communities [2].
Compared with feature matching in aerial or terrestrial scenarios, the feature matching of satellite stereos meets several unique challenges, such as large image sizes [3], large weak-textured regions [4], geometric distortions [5], cloud occlusions and seasonal changes [6]. The above challenges will greatly increase matching time complexity and influence the final orientation results since the correspondences in these challenging regions are often of low accuracy [7]. In order to achieve good matching results, current work either finds correspondences throughout whole image pairs to guarantee enough good matches for orientation or utilizes some constraints (e.g., pyramid strategy and epipolar strategy) for the efficiency purpose [3]. Therefore, the feature matching of satellite stereos can be categorized into: (1) robustness-first algorithms and (2) efficiency-first algorithms.
The robustness-first algorithms focus on improving the robustness and accuracy of feature matching between satellite stereos, especially in geometric distortion, radiometric distortion and weak-textured scenarios. Traditional methods utilized some famous descriptors (e.g., SIFT [8], SURF [9], ORB [10], BRISK [11] and AKAZE [12]) in satellite image matching, which is able to obtain robust matching results against Euclidean geometric distortions and local linear radiometric distortions [5,13,14,15,16]. However, these traditional ones may meet challenges in the scenarios of more complex geometric and radiometric distortions, which are common in off-track or multi-source satellite stereos. To further improve the matching robustness of satellite stereos, some work either improves the descriptors themselves by merging variant traditional features [17] and integrating affine invariant features [18,19,20] and intensity invariant features [21,22] or imposes spatial regularization constraints to reduce the matching ambiguity [7,23,24]. In recent years, a convolutional neural network (CNN) has been utilized in feature matching, which is able to extract more accurate and more robust deep features, especially in several challenging scenarios [6,25,26,27,28,29,30]. However, to guarantee finding enough good matches, such methods usually involve the entire images in feature matching, which greatly increases the time complexity of feature matching.
The efficiency-first algorithms focus on improving the matching efficiency of satellite stereos. Some work utilized epipolar lines and image pyramids to predict the potential positions of correspondences, which greatly reduce the search space of feature matching [3,4,31,32]. On the other hand, Ling et al. obtained matches from evenly distributed image blocks instead of the whole images, which is an alternative way to reduce the matching time cost [33]. However, the above methods did not distinguish the ready-to-match regions, and the regions in the challenging scenarios may bring obvious mismatches in the next orientation process. Moreover, the image block-based matching algorithm cannot be applied in the scenarios of large positioning errors (e.g., hundreds of pixels) of satellite stereos, since large positioning errors may bring wrong image block pairs in stereos.
For the purpose of efficient and robust feature matching, this paper proposes an image-matching block selection algorithm with the basic assumption that the matches of only a few image blocks could achieve competitive orientation results with the ones of the entire image stereos. The core of the proposed method is to formulate the image block selection into the three-step optimization of a discrete energy function. Since the optimization is an NP-hard problem, a greedy solution is utilized for an approximate optimal result. The proposed method is able to avoid difficult matching regions (e.g., water, clouds) and select optimal image blocks, thus resulting in high-quality matches for orientation. Furthermore, the proposed method imposes block enlargement constraints in the optimization, which can achieve robust matching results against the positioning errors of hundreds of meters. The main contribution of the proposed method is to greatly reduce the matching time cost of satellite stereos while maintaining robust matching results.

2. Methodology

2.1. Workflow

This paper aims to reduce matching time costs while keeping the matching accuracy of satellite stereos at a sub-pixel level by selecting a few optimal image blocks for matching. The basic idea is that evenly distributed, textured and photo-consistency blocks tend to provide good matches for high-accuracy orientation. Therefore, this paper proposes an optimal image block selection method with three-step optimizations, which selects optimal image blocks step by step with the constraints of block textures, block distributions and photo consistency between block pairs. In general, the input of the proposed method is high-resolution (HR) satellite stereos with the corresponding geographic configuration (i.e., rational polynomial coefficient (RPC) files); the workflow is shown in Figure 1. Before block selection, both satellite stereos were rectified on a common height plane so that their GSDs were unified and the rotation distortions could be eliminated, which could simplify the next block selection process.
This paper takes any one of satellite stereos as the basic image I 1 and the other is denoted as I 2 . Before optimization, both images were automatically zoomed out, and the optimization process was conducted on the zoomed image for the computation efficiency purpose. In this paper, the zoomed-out ratio is set as 4 throughout the paper for two reasons: (1) high-resolution satellite images already have sub-meter resolutions (e.g., GSD: 0.5 m) so that the zoomed images with 4× ratio (e.g., GSD: 2.0 m) still have enough texture information for the block selection; (2) the original image block size is set as 2000 × 2000 pixels in this paper, and the larger ratios (>4) will lead to much smaller zoomed sizes (<500 × 500 pixels), which is too small to select optimal blocks. In the first step of the optimization, the overlap regions between the zoomed basic images I 1 and I 2 are divided into a series of nonoverlap fixed blocks, and several candidate optimal blocks S 1 = b 1 1 , b 2 1 , are selected according to their textures and distributions (e.g., red rectangles in Figure 1a), where b I 1 is the ith block in I 1 . Due to seasonal changes or cloud occlusions, it is possible that the potential optimal blocks in I 1 are not fit for I 2 . Therefore, this paper further considers the texture information in I 2 , where the blocks in I 2 are obtained by forward–backward projections on a common height plane, as shown in Figure 1b. The common height plane is acquired by averaging elevations of SRTM within the geographic scope of satellite stereos. Only the blocks with rich textures in both images are selected, e.g., the blue rectangles in Figure 1b. These selected blocks in the second optimization step are termed S 2 = b 1 12 , b 1 12 , with b 1 12 being the ith block pair in both I 1 and I 2 . Due to the underlying positioning errors of satellite images, the projected blocks in I 2 may be inconsistent with the ones in I 1 . This paper therefore enlarges the projected blocks in I 2 to guarantee containing the consistent information with the blocks in I 1 , as shown in Figure 1b. In the third optimization step in Figure 1c, this paper imposes photo-consistency constraints between block pairs in S 2 through feature matching techniques, and selects optimal block pairs with good matches, e.g., the green rectangle in Figure 1c. Finally, the sizes of the projected blocks in I 2 are refined to guarantee that the block sizes in both I 1 and I 2 are similar. These optimal blocks are then resampled in original GSD (ground sampling distance) for feature matching and orientation.
The time complexity of the three-step optimizations is low, which only takes a few seconds. Though feature matching is still used in the optimization to guarantee the photo-consistency between block pairs, the number of these ready-to-match blocks is small after the first and the second optimization. Moreover, the image pyramid strategy also greatly reduces the time complexity. Therefore, the proposed method could achieve good matching results at low time cost.

2.2. Formulation

2.2.1. Global Energy Function

Given a satellite stereo I 1 ,   I 2 and the corresponding zoomed-out stereo I 1 z ,   I 2 z , this paper formulates the image block selection problem as the minimization of an energy function in Equation (1), which consists of cost term and regularization term with the former representing the probability of being selected for each block and the latter imposing spatial distance constraints among blocks.
min E L = b S N C b ,   l b P · b S N d S b N 1 D b ,   d
Here, E is the designed energy function, whose optimal solution is the selected blocks for feature matching; L is the set of the selected blocks; N is the pre-defined block number for selection; S is the set of candidate blocks, whose block number is normally larger than N . The essence of Equation (1) is to select N blocks that minimize the energy function. In general, N and S have different meanings in different steps of the optimization. In order to distinguish different N and S in different steps, the subscripts are added as N i and S i , i = 1 , 2 , 3 . In the first-step optimization, S 1 represents the set of all blocks in I 1 z , and N 1 represents the number of blocks for the optimal solution of the first step. In the second-step optimization, its input block set S 2 is exactly the solution of the first-step optimization; therefore, the element number of S 2 is N 1 . After the optimization, the optimal block number N 2 is reduced to N 1 / s with s being the number scaling factor. Similarly, the input block S 3 is exactly the solution of the former optimization, and the optimal block number N 3 is reduced to N 2 / s . The scaling factor s is related to the solution number during three-step optimization. The larger scaling factor s will increase the time cost of block selection, while the smaller one may ignore some good block pairs. Thus, the scaling factor should be appropriately determined. The candidate block sets from the solution of the former-step optimization may contain more inappropriate blocks than appropriate ones, especially in some difficult matching scenarios (e.g., cloudy imagery). Thus, the solution number should be smaller than half the size of the candidate block set, i.e., s > 2. On the other hand, the time efficiency of the proposed method requires the smallest s . Thus, this paper set s = 3 by considering the above two factors.
The first term of Equation (1) serves as the cost term, which quantitatively evaluates the texture information and photo consistency information of image blocks and sums these evaluation results limited by the predefined block number. The lower cost means the higher probability of being selected. In the first term, b is any one block in S and C b ,   l b is the cost of b at label l b . In the optimization of Equation (1), each block only has binary labels. For example, label 1 means “selected blocks” and label 0 means “not selected blocks”. Since each step of the optimizations considers different cost constraints (e.g., textures in I 1 z for the first-step optimization, textures in both I 1 z and I 2 z for the second-step optimization and the photo-consistency information for the third-step optimization), the cost terms in each step have different mathematical formulations, which will be introduced in Section 2.2.2.
The second term of Equation (1) serves as the regularization term, which tries to make the selected blocks stay away from each other for the purpose of evenly distributed matches. In the second term, P is a penalty coefficient, which is used to balance the impact of the regularization term on the block selection results. The optimal value of P will be analyzed in the experimental part. b , d are any two blocks in S and D b ,   d measures the center distances between b and d . Detailed descriptions about the regularization term will be introduced in Section 2.2.3.

2.2.2. Cost Term

To achieve a good compromise between matching accuracy and time cost, the proposed method optimizes the block selection results step by step. Each step has different formulations of the cost term. In the first-step optimization, the cost term only considers texture information in I 1 z so that several textured blocks could be efficiently selected from hundreds of candidate blocks. In this paper, intensity gradients are used to evaluate textures due to its low time cost. The cost term of the first-step optimization is therefore represented as:
C 1 b ,   l b = p b G p i f   l b = 1 0 i f   l b = 0
where C 1 is the cost term in the first-step optimization and p is a pixel which is located in the block b . G p is the intensity gradient of p , which is computed by the Sobel operator in this paper. When the block is selected (i.e., l b = 1 ), the cost term is the summation of G p ; otherwise, it is zero.
Due to seasonal changes and cloud occlusions, the textured block in I 1 z may not be suitable for I 2 z . Thus, the cost term of the second-step optimization considers the texture information not only in I 1 z , but also in I 2 z . In the second-step optimization, only the selected blocks from the former step are considered in the texture evaluation, which is helpful in reducing the time cost. The formulation of the cost term in the second-step optimization is as follows:
C 2 b ,   l b = p b 1 G 1 p p b 2 G 2 p i f   l b = 1 0 i f   l b = 0
where C 2 is the cost term in the second-step optimization; b = b 1 ,   b 2 is a block pair with b 1 being a block in I 1 z and b 2 being a block in I 2 z ; G 1 is the intensity gradient in I 1 z ; G 2 represents the intensity gradient in I 2 z . Considering the positioning errors of high-resolution satellite images, this paper sets the sizes of b 2 being three times larger than the ones of b 1 , which is able to improve the robustness of the texture evaluations in I 2 z .
To ensure that the block pairs are fit for matching, this paper imposes photo-consistency constraints in the third-step optimization through feature matching techniques. Though some feature matching algorithms are complex, the limited block number from the former step will greatly reduce the matching time cost. In the third-step optimization, this paper formulates the photo-consistency constraints as the number of matches, as follows:
C 3 b ,   l b = N u m b 1 , b 2 i f   l b = 1 0 i f   l b = 0
where C 3 is the cost term in the third-step optimization and N u m is a function of counting the matches between the block pair b 1 ,   b 2 . In the experimental part, SIFT descriptors are used in the feature matching. C 3 means that the blocks with more matches have higher probabilities of being selected.
To balance the amount of each cost term, this paper additionally normalizes the cost term by dividing it by the absolution of the minimum cost term, as follows. After the normalization, the penalty coefficient P in Equation (1) can be fixed throughout the three-step optimizations.
C k ¯ b ,   l b = C k b ,   l b / min C k b ,   l b
Here, C k ¯ is the normalized cost term in the kth optimization and min C k b ,   l b means the minimum cost among the set of C k b ,   l b .

2.2.3. Regularization Term

The evenly distributed matches tend to generate high-accuracy orientation results. Therefore, this paper defines the regularization term to separate the image blocks so that the matches from these selected blocks are evenly distributed. The regularization term is formulated as the center distances between image blocks (as shown in Equation (6)), which encourages the selected blocks to stay away from each other.
D b ,   d = x b x d 2 + y b y d 2
Here, x b , y b and x d , y d are the center coordinates of the block b and d .
To balance the amount of the cost terms and the regularization terms, this paper also normalizes the regularization term by dividing it by the diagonal length of the image, which is able to reduce the regularization term to the range [0, 1].

2.3. Solution

The optimization of the global energy function in Equation (1) is an NP-hard problem. For the purpose of efficient matching, this paper proposes a greedy solution, which is able to obtain an approximate optimal result at low time cost. The greedy solution is divided into stepwise iterations. In each iteration, the greedy solution always selects one local optimal result; thus, the iteration times depends on the predefined number of optimal blocks.
Given the current block set S c u r r e n t = b i and the optimal block set S s e l e c t = , the task of the greedy solution is to select the optimal blocks from S c u r r e n t and add these optimal blocks into S s e l e c t . Each step of the optimization has a different block set, which is termed as S c u r r e n t t and S s e l e c t t with t being the step ID of the optimization. In the first-step optimization, S c u r r e n t 1 is the set containing all image blocks in I 1 z , and in the next-step optimization, S c u r r e n t t is defined as the optimal results S s e l e c t t 1 of the former-step optimization. In this paper, the element number of S s e l e c t t is three times larger than that of S s e l e c t t + 1 . In the experimental part, the optimal element number of S s e l e c t 3 will be analyzed and discussed.
All steps of the optimization share the same greedy strategies which iteratively select local optimal results. At the beginning of the iteration, the regularization term is set as zero due to the empty S s e l e c t t . Thus, the local optimal solution at the beginning is obtained by selecting the block with the minimum cost, e.g., the green rectangle in Figure 2a. The optimal block of the beginning iteration b 1 is added into S s e l e c t t with S s e l e c t t = b 1 . In the next iteration, the distances between the remaining blocks in S c u r r e n t t and the ones in S s e l e c t t are formulated as regularization terms, as shown in Figure 2b, where the red rectangles represent blocks in S c u r r e n t t , and the green one is the block in S s e l e c t t . The current local optimal result is obtained by selecting the block with the minimum summation of the cost term and the regularization term. The greedy solution is iterated until the number of the selected blocks reaches the pre-defined value.
In general, the proposed greedy solution can be formulated as follows:
b i = a r g m i n b S c u r r e n t   C b ,   l b       i f   i = 1   a r g m i n b S c u r r e n t   C b ,   l b P · d S S e l e c t D b ,   d       i f   i > 1
where b i is the local optimal solution in the ith iteration; b is an image block in the set S c u r r e n t ; d is a block in the set S s e l e c t . The greedy solution in Equation (7) iteratively selects optimal blocks until the element number of S s e l e c t reaches the predefined value.

2.4. Post-Processing

After the three-step optimization, several optimal blocks in both stereo images are selected. However, considering positioning errors, the block sizes in I 2 z are much larger than the ones in I 1 z , thus only a small region in the blocks of I 2 z is available for matching. In order to further improve the matching efficiency, this paper searches for such available regions through post processing and defines them as the optimal blocks in I 2 z , as shown in Figure 3. The blue circles represent the matches between block pairs after the three-step optimization. To accurately locate the available regions in I 2 z , a block (red rectangle) with the same size of the one in I 1 z moves at a fixed pace in both row and column directions. During this movement, the number of matches within the moving block is counted, and the block with the maximum matches is selected as the optimal block in I 2 z . Finally, both optimal blocks in I 1 z and I 2 z are zoomed to the original sizes for feature matching. In all experimental comparisons, the moving pace is set as 10 pixels.

3. Study Areas and Data

The proposed method was tested on high-resolution satellite stereos including a WorldView-3 satellite dataset with a GSD of 0.3 m and one pair of Gaofen-7 satellites with a GSD of 0.7 m. The WorldView-3 satellite data were off-track stereos, as shown in Figure 4a–c. The first one was near Jacksonville with the imaging time being November 2015 and May 2015, respectively; the second one was near Omaha with the imaging time being September 2015 and October 2015, respectively; the third one was near Buenos Aires with the imaging time being March 2015, June 2015, September 2015 and October 2015, respectively. Both WorldView-3 stereos were provided by the intelligence advanced research projects activity (IARPA) and DigitalGlobe [34,35,36]. On the other hand, the Gaofen-7 stereo was the in-track one near Tibet with the imaging time being March 2021. These satellite stereos would meet challenges in matching with the first one in Figure 4a having large areas of seawater, the second one in Figure 4b having significantly changed farmlands, the third one in Figure 4c having serious radiometric distortions and the last one in Figure 4d having large areas of snow.
To evaluate the feature matching accuracy of the proposed method, this paper manually selected 12 matching points as checking points for the satellite data in Figure 4a,b,d and 9 matching points for the ones in (c), which are illustrated as yellow circles in Figure 4. In experimental comparisons, four metrics were used to comprehensively evaluate the performances of the proposed method, including: (1) matching time A t i m e ; (2) inlier number A n u m ; (3) inlier percentage A p e r ; (4) orientation accuracy A o r i e n t a t i o n . The matching time metric A t i m e measures the time cost of the whole matching processing, including the block selection process and the feature matching process, which can evaluate the efficiency of different matching algorithms. The inlier number metric A n u m finds inliers from the feature matching results by computing their distances to the corresponding epipolar lines. Matches with the epipolar distance shorter than three pixels are selected as inliers. To ensure the geometric accuracy of the epipolar lines, this paper first used the checking points in orientation, and then computed epipolar lines through the true orientation results. The inlier percentage metric A p e r computes the percentage of the inliers in the whole matching results, which indicates the robustness of the feature descriptors. The orientation accuracy metric A o r i e n t a t i o n is the average distances of all checking points to the epipolar lines which are derived from the orientation results of all matches. A o r i e n t a t i o n indicates the actual performances of all matches in the later orientation process. The formulation of the four metrics is shown in Equation (8). Moreover, considering mismatches, the orientation process in this paper utilized the iteration method with variable weights [33], which is able to increase the weights of inliers and decrease the weights of mismatches in the adjustment.
A t i m e = T b l o c k + T m a t c h A n u m = NUM ( m | m M     E p i c h e c k m < 3 ) A p e r = A n u m / N U M ( m | m M ) A o r i e n t a t i o n = AVG ( E p i m a t c h c p   |   c p C P )
Here, T b l o c k is the time cost of the block selection optimization; T m a t c h is the time cost of feature matching in the original GSD; M is the set of all matches; m is a match in M ; E p i c h e c k m is the distance of m to the epipolar line which is computed from the orientation results of checking points; NUM is a function to count the number of points that satisfy a certain condition; C P is the set of checking points; c p is any one checking point in C P ; E p i m a t c h c p is the distance of c p to the epipolar line which is computed from the orientation results of matches in M ; AVG is a function to average the elements that satisfy a certain condition.
In all experiments, the distance ratio criteria were used before obtaining the final matching results, which firstly found distances between feature points and their candidate matching points in the feature space, and then measured the ratio between the minimum distance and the second minimum distance. The matches with a distance ratio less than 0.4 were eliminated, and the remaining high-confidence matches were used in the experimental accuracy evaluations.

4. Results

This paper utilized the proposed method to select optimal blocks, and then applied SIFT to find matches in these blocks. Since the performances of the proposed method are related to the optimal block number and the penalty coefficient in Equation (1), this paper first analyzed the above two parameters of the proposed method on the Jacksonville data in Section 4.1. The best values of the two parameters were then fixed in all the following experiments. Section 4.2 shows the ability of the positioning error correction of the proposed method, which was tested on Omaha data by manually adding various systematic errors into the corresponding RPC parameters. Section 4.3 compared the proposed method with other state-of-the-art methods in the metrics of A t i m e , A n u m , A p e r and A o r i e n t a t i o n . The compared methods include: (1) SIFT matching with fixed image blocks (SFB) [33]; (2) SIFT matching with all image blocks (SAB) [5]; (3) heterologous images matching (HIM) [20] which considers the anisotropic weighted moment and absolute phase orientation in feature descriptors. Finally, the proposed method was further tested on more satellite stereos of Buenos Aires data, which aimed at finding the limitation of the proposed method. In all experiments, the SIFT matching in SFB, SAB and the proposed method were conducted on the same compute configuration with a single core of Intel i9-12900HX, while the proposed block selection optimization used full cores of Intel i9-12900HX. On the other hand, HIM is applied with a single core of AMD 5900HX. The reference clock frequencies of i9-12900HX and 5900HX are similar with the frequency of i9-12900HX being 3.60 GHz and 5900HX being 3.30 GHz. Thus, the small compute configuration differences did not impact the conclusion of the time efficiency comparisons. In the experiments of Section 4.1, Section 4.2 and Section 4.3, the SIFT descriptor was applied after the block selection of the proposed method, and the original block sizes were set as 2000 × 2000 pixels. However, in order to verify that the block selection results were also fit for other feature descriptors, this paper applied SuperPoint descriptors in Section 4.3 and set the original block sizes as 1500 × 1500 pixels, which is recommended in [25,37].

4.1. Analysis on Optimal Parameters of the Proposed Method

The performances of the proposed method are related to two parameters: optimal block number N o p t and penalty coefficient of the regularization term P . The former influences the matching time as well as the orientation accuracy. Higher N o p t will guarantee the orientation accuracy but increase matching time; otherwise, lower N o p t will reduce the matching time while potentially decreasing the orientation accuracy. On the other hand, the penalty coefficient P influences the balances between the block texture information and the block spatial distributions. Lower P will increase the texture richness of the blocks but cannot ensure their even distribution. Higher P will enhance the even distributions of blocks while potentially decreasing the texture information in the blocks. Therefore, both N o p t and P should be analyzed for the best performances of the proposed method.
Jacksonville data with image sizes as 43,032 × 32,805 pixels in Figure 4a were used in the analysis about the optimal value of N o p t and P . This paper firstly set the block number as six and computed the optimal value of P by gradually increasing the values of P from 0 to 2.5 and then evaluating the performances of the proposed method with the various P in the metrics of A n u m and A o r i e n t a t i o n . The evaluation result is shown in Figure 5. The horizontal axis represents the various values of P from 0 to 2.5. The vertical axes represent the evaluation results in A n u m and A o r i e n t a t i o n , respectively. The figures in the bottom of Figure 5 show the spatial distributions of the optimal blocks of the proposed method with a given value of P .
The trend of A n u m in Figure 5 shows that the number of inliers decreases with the increasing P . This is because the increasing P will reduce the weight of the cost term in the optimization and generate blocks with less texture information, thus resulting in fewer matching points. However, the orientation accuracy A o r i e n t a t i o n has a trend of being higher with the increasing P , though the inlier number decreases. This is because the larger P brings more evenly distributed matching points, as shown in the bottom figures of Figure 5. When P = 0 , most blocks concentrate on the left tops of Jacksonville data, which are the most textured regions in Jacksonville. With the increasing P , the image blocks are more evenly distributed. In all the cases of P , the corresponding image blocks always avoid seawater regions, thus providing more robust matches.
In Figure 5, the trend of inlier number A n u m becomes convergent when P = 1.5 , and the highest orientation accuracy is found at P = 1.5 . Therefore, this paper sets P = 1.5 in all the following experiments.
In the analysis of N o p t , this paper varied the value of N o p t from 1 to 10 and evaluated the matching results in the metrics of A t i m e , A n u m and A o r i e n t a t i o n , as shown in Figure 6. The horizontal axis represents the various values of N o p t from 1 to 10. The vertical axes represent the evaluation results in A t i m e , A n u m and A o r i e n t a t i o n , respectively. The figures in the bottom of Figure 6 show the spatial distributions of the optimal blocks of the proposed method with a given value of N o p t .
Figure 6 shows the trends of A t i m e , A n u m and A o r i e n t a t i o n with various N o p t . In the sub-figure of A t i m e , the total matching time is increasing with the increase of the block number N o p t , since more blocks means more feature matching computations. The black digits mean the entire time cost (block optimization time + feature matching time), and the red digits mean the block optimization time of the proposed method. The difference between the minimum and the maximum optimization time is only 4.2 s, which shows that the optimization time of the proposed method is rarely affected by the block number. Furthermore, for the matching in the stereo with the size as 43,032 × 32,805 pixels, the maximum matching time is only 90.3 s at N o p t = 10 , which shows the potential ability of the proposed method in the efficient matching.
In Figure 6, the sub-figure of A n u m shows that the inlier number is increasing with the increasing block number. However, the corresponding orientation accuracy A o r i e n t a t i o n did not have the same trend with A n u m , which shows that more matches cannot guarantee more accurate orientation results. The trend A o r i e n t a t i o n tends to be convergent when N o p t 5 , and it has the top third orientation accuracy at N o p t = 6, 7 and 10. Since the time cost of N o p t = 6 is less than other two cases, this paper sets N o p t = 6 in all the following experiments.
In general, this paper sets P = 1.5 and N o p t = 6 in all the following experiments, which is considered a good compromise between time cost and orientation accuracy. It should be noted that good satellite stereos with rich texture information could provide more optimal blocks, which will result in more matches. Nevertheless, the core idea of the proposed method is that a few high-quality, evenly distributed matches are enough to support high-accuracy orientation results at low time cost. Thus, this paper only selects several optimal blocks instead of all good blocks in matching. The block number N o p t is fixed in all the following experiments, no matter how rich the texture information is that the stereos have.

4.2. Analysis on Positioning Error Correction of the Proposed Method

The block selection results are dependent on the positioning accuracy of the satellite stereos since the blocks in I 2 z are obtained by forward–backward projections of the blocks in I 1 z . Large positioning errors may lead to wrong blocks in I 2 z , which are not consistent with the ones in I 1 z . To reduce the influence of the positioning errors, this paper enlarges the block sizes in I 2 z during the optimization so that enough consistent image information could be contained. To test the validation of the proposed method against the positioning errors, this paper tested the matching results of the proposed method on Omaha data with image sizes as 43,210 × 50,471 pixels. Various positioning errors from 0 m to 1000 m were added into the RPC of Omaha data, and the corresponding matching results were evaluated in the metrics of A n u m and A p e r , as shown in Figure 7.
The sub-figure of A n u m shows that the inlier number is similar when the positioning errors are smaller than 600 m, which verifies the ability of the proposed method against the positioning errors. The proposed method set image block size as 2000 × 2000 pixels. The block sizes in I 2 are three times larger than the ones in I 1 during the three-step optimization. Thus, the proposed method can theoretically correct positioning errors of 2000 pixels. Since the GSD of Omaha data is 0.3 m, the lossless positioning error correction of the proposed method on Omaha data is at most 600 m. However, when the positioning errors are larger than 600 m, the inlier number is significantly decreasing, since the block pairs in the stereo only have small overlaps. These extremely large positioning errors (>600 m) can be further corrected by further expanding the block sizes in I 2 . However, the expanding operation will increase the time cost, and it is rare to find stereos with positioning errors larger than 600 m. Therefore, this paper still set the block enlargement ratio to three times during the three-step optimization. In the other sub-figure, A p e r has no obvious differences in all cases of positioning errors with the maximum percentage being 98.95% and the minimum one being 97.14%, due to the robustness of SIFT descriptors.
To comprehensively verify the ability of the positioning error correction of the proposed method, this paper selected any one of the optimal blocks in I 1 and visually illustrated the corresponding blocks in I 2 with/without positioning error corrections separately. The visual comparisons are shown in Figure 8. The first row in Figure 8 is the image block in I 1 , which is the reference block for the visual comparisons. The third to sixth rows represent the corresponding image blocks in I 2 with various positioning errors. When the positioning error is zero, both scenarios with and without corrections could achieve consistent blocks with the reference block. However, with the increase of positioning errors, the blocks without corrections only had smaller overlaps with the reference block. When the positioning errors were larger than 600 m, the blocks without corrections had no overlap with the reference one, thus generating no matches. On the other hand, the proposed method could keep the consistent blocks with the reference one when the positioning errors are smaller than 600 m. Even in the worst case (error = 1000 m), the proposed method still finds partial overlap between the blocks of the stereo, which shows that the proposed method has a strong ability against the positioning errors.

4.3. Comparison of Experimental Results

To verify the validation of the proposed method, this paper further compared the proposed method with other developed ones, including: (1) SFB [33]; (2) SAB [5]; (3) HIM [20]. This paper tested these methods on Omaha data with image sizes set to 43,210 × 50,471 pixels and Tibet data with image sizes set to 42,676 × 41,249 pixels, and evaluated their matching results in the metrics of A t i m e , A n u m , A p e r and A o r i e n t a t i o n , as shown in Figure 9a,b. Both SAB and HIM involved the entire stereos in matching with a large number of matches. The comparisons with them aim to check whether a smaller number of evenly distributed matches from the optimal blocks could achieve competitive orientation results.
The sub-figure of A t i m e in Figure 9 shows that the matching time of SFB and the proposed method is much less than that of SAB and HIM, since both SFB and the proposed method selected only a few blocks for matching. SAB and HIM methods found matches throughout the entire images, thus resulting in high time cost. The time cost of the proposed method is slightly higher than SFB due to the optimization process of the proposed method. In general, the proposed method can achieve low time cost around 75 s in large satellite stereos, which is at least 22 times and 58 times faster than SAB and HIM methods, respectively. Contrary to A t i m e , both SAB and HIM could obtain many more matches than SFB and the proposed method, since both methods found matches throughout the entire images instead of a few blocks. The HIM method found the most matches due to its robustness descriptors against geometric and radiometric distortions. Though the block number of SFB and the proposed method is the same, the proposed method found many more matches than SFB due to its block selection optimizations. The third row of Figure 9 shows the inlier percentages of each method, where SFB, SAB and the proposed method have similar inlier percentages, since all these methods adopted SIFT descriptors in matching. HIM methods had the lowest inlier percentages, while its large number of matches can still guarantee the robustness of the next orientation process.
In the sub-figure of A o r i e n t a t i o n , SFB achieved the lowest orientation accuracy since the matches in the fixed blocks may be of low quality, as shown in Figure 10a. The first row in Figure 10a shows the matching blocks of SFB, which simply selected blocks at fixed intervals. However, these blocks may fall in changed regions (e.g., the block in the top left of Omaha data in (a) and weak-textured regions (e.g., the blocks in the top right of Omaha data in (a)), thus providing low-quality matches or even mismatches. Therefore, SFB achieved the lowest orientation accuracy. The proposed method obtained the average accuracy of 0.545 pixels on Omaha data and Tibet data, which is slightly higher than SAB with an average accuracy of 0.605 pixels. Considering the inlier number of SAB is much higher than the proposed method, the comparison results show that the orientation accuracy is not totally dependent on the inlier number. The proposed method selected optimal image blocks with even spatial distributions and rich texture information (as shown in Figure 10b), where good-quality inliers can be matched for accurate orientation at low time cost. Finally, HIM achieved the highest orientation accuracy with an average of 0.50 pixels, while its time cost is much higher than the proposed method.
Though the inlier number of the proposed method is much less than SAB and HIM, it could achieve an average of 62 times faster time efficiency while keeping similar matching accuracy, which shows that a few evenly distributed, good matches are enough to support high-accuracy orientation results. In general, the proposed method is a good, compromised solution between matching accuracy and matching time, which can compute accurate matching results at low time cost.
In order to find the limitations of the proposed method, this paper further tested the proposed method on the Buenos Aires dataset, which consists of four overlapping images and therefore generates six pairs of satellite stereos. This paper still utilized SIFT to select photo-consistency blocks during the optimization of block selection. However, after the block selection, SuperPoint descriptors were utilized within the optimal blocks for feature matching, which aims at testing the transferability of the proposed method. In general, this paper compared the proposed method with SFB in the metrics of A n u m and A o r i e n t a t i o n , and recorded the time cost of block selection for each pair, as shown in Figure 11a,b. The horizontal axis shows the stereo pair ID in the format of i - j with i , j being the image ID in the Buenos Aires dataset.
The digits above the orange line in Figure 11a show the time cost of the block selection of the proposed method. Since the original block sizes and the block numbers are the same for SFB and the proposed method, the time cost differences between the two methods mainly focused on the block selection process. The proposed method needs to take extra time in the optimal block selection, which causes higher time cost than SFB. Figure 11a shows that the time cost of the optimal block selection did not change much for different stereo pairs. The average time cost is only 30.373 s, which is acceptable when compared with the time cost of feature matching.
Figure 11a shows that the inlier number of the proposed method is always higher than the ones of SFB. Both methods utilized SuperPoint to find matches within image blocks. During the optimization of the block selection, the proposed method considers texture information constraints and prefers selecting textured blocks. Thus, the proposed method could obtain 2.13 times more inliers than SFB. On the other hand, Figure 11b shows that the orientation accuracy of the proposed method is higher than SFB in most cases with an average accuracy improvement of 28.27%, which shows that the block selection results help to provide more high-quality matches. However, the orientation accuracy of the proposed method is lower than SFB in Pair 2–4 due to the inability of the photo-consistency block selection of the proposed method. This paper adopted the SIFT descriptor in the photo-consistency block selection, though the SIFT matching results may not be reliable, especially in some challenging scenarios. For example, there are serious seasonal changes and radiometric distortions between Pair 2–4, as shown in Figure 12; the SIFT descriptors only found 16 matches from the stereo pair, in which case the photo-consistency constraint is unreliable and therefore disturbs the final block selection results. Adopting more reliable feature descriptors is helpful in improving the block selection results. In general, the proposed method performed well in both time efficiency and orientation accuracy.

5. Conclusions

This paper proposed a novel matching method for high-resolution satellite stereos by selecting optimal image blocks and formulating the block selection as a three-step optimization of an energy function. A greedy strategy is designed to obtain an approximate solution at low time cost. The main contributions of the proposed method include: (1) the proposed method only selects optimal image blocks for matching, which could greatly reduce the matching time cost while keep competitive matching accuracy with the entire-image-based matching algorithms; (2) the proposed method could generate robust matching results against large positioning errors of satellite stereos. Since the proposed method only involves block selections, it can combine various matching descriptors in addition to SIFT. Experiments on high-resolution satellite datasets (including two pairs of WorldView-3 satellites and one pair of Gaofen-7 satellites) show that the proposed method has strong matching abilities for satellite stereos against the positioning errors at most 1 km. When compared with other developed matching methods, the proposed method could achieve the best compromise between the matching time cost and the orientation accuracy. In both the WorldView-3 dataset and the Gaofen-7 dataset, the proposed method could achieve the average time cost of 75.20 s, which is almost 62 times faster than those entire-image-based matching methods (e.g., SAB and HIM) while retaining similar matching accuracy. Though the time cost of the proposed method is about 30–40 s higher than SFB, its matching accuracy is higher than SFB by 28.27%. Thus, the proposed method is able to greatly reduce the matching time cost while retaining high matching accuracy. However, the proposed method is only fit for stereo feature matching instead of the multi-view cases. In future work, the block selection method in multi-view cases will be further developed.

Author Contributions

Conceptualization, D.G.; methodology, D.G. and X.H.; software, D.G. and X.H.; validation, X.H. and Y.Y.; formal analysis, X.H.; investigation, X.H. and J.Z.; resources, X.H. and D.G.; data curation, X.H. and D.G.; writing—original draft preparation, X.H. and J.Z.; writing—review and editing, D.G. and Y.H.; visualization, X.H.; supervision, X.H.; project administration, X.H.; funding acquisition, D.G. and X.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (grant number 41701540), Basic Startup Funding of Sun Yat-sen University (grant number 76230-18841205), and 3D reconstruction technology based on high-resolution aerospace line array images (grant number 76230-71020009).

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the intelligence advanced research projects activity (IARPA) and DigitalGlobe for providing the WorldView-3 satellite imagery in the CORE3D public dataset.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Prakash, P. Introduction to Satellite Photogrammetry, 1st ed.; OrangeBooks Publication: Bhilai, India, 2021. [Google Scholar]
  2. Kocanaogullari, A.; Ataer-Cansizoglu, E. Active Descriptor Learning for Feature Matching. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  3. Ling, X.; Zhang, Y.; Xiong, J.; Huang, X.; Chen, Z. An image matching algorithm integrating global SRTM and image segmentation for multi-source satellite imagery. Remote Sens. 2016, 8, 672. [Google Scholar] [CrossRef] [Green Version]
  4. Yuan, X.; Chen, S.; Yuan, W.; Cai, Y. Poor textural image tie point matching via graph theory. ISPRS J. Photogramm. 2017, 129, 21–31. [Google Scholar] [CrossRef]
  5. Tahoun, M.; Shabayek, A.E.R.; Nassar, H.; Giovenco, M.M.; Reulke, R.; Emary, E.; Hassanien, A.E. Satellite Image Matching and Registration: A Comparative Study Using Invariant Local Features. In Image Feature Detectors and Descriptors: Foundations and Applications; Awad, A.I., Hassaballah, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 135–171. [Google Scholar]
  6. Luo, Q.-Y.; Zhang, J.-D.; Zhu, L.-C.; Huang, X. Research on Feature Matching of High-Resolution Optical Satellite Stereo Imagery under Difficult Conditions. In Proceedings of the 8th China Hight Resolution Earth Observation Conference (CHREOC); Springer: Singapore, 2022. [Google Scholar]
  7. Huang, X.; Wan, X.; Peng, D. Robust Feature Matching with Spatial Smoothness Constraints. Remote Sens. 2020, 12, 3158. [Google Scholar] [CrossRef]
  8. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  9. Bay, H.; Ess, A.; Tuytelaars, T.; Van Gool, L. Speeded-up robust features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  10. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2564–2571. [Google Scholar]
  11. Leutenegger, S.; Chli, M.; Siegwart, R.Y. BRISK: Binary Robust Invariant Scalable Keypoints. In Proceedings of the IEEE International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2548–2555. [Google Scholar]
  12. Alcantarilla, P.F.; Solutions, T. Fast explicit diffusion for accelerated features in nonlinear scale spaces. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 1281–1298. [Google Scholar]
  13. Kuriakose, E.; Viswan, A. Remote sensing image matching using sift and affine transformation. Int. J. Comput. Appl. 2013, 80, 22–27. [Google Scholar] [CrossRef]
  14. Tahoun, M.; Shabayayek, A.E.R.; Hassanien, A.E. Matching and co-registration of satellite images using local features. In Proceedings of the International Conference on Space Optical Systems and Applications, Kobe, Japan, 7–9 May 2014. [Google Scholar]
  15. Zheng, M.; Wu, C.; Chen, D.; Meng, Z. Rotation and affine-invariant SIFT descriptor for matching UAV images with satellite images. In Proceedings of the 2014 IEEE Chinese Guidance, Navigation and Control Conference, Yantai, China, 8–10 August 2014; pp. 2624–2628. [Google Scholar]
  16. Li, X.-X.; Luo, X.; Wu, Y.-X.; Li, Z.-T.; Xu, W.-B. Research on Stereo Matching for Satellite Generalized Image Pair Based on Improved SURF and RFM. In Proceedings of the IGARSS 2020—2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 2739–2742. [Google Scholar]
  17. Karim, S.; Zhang, Y.; Brohi, A.A.; Asif, M.R. Feature Matching Improvement through Merging Features for Remote Sensing Imagery. 3D Res. 2018, 9, 52. [Google Scholar] [CrossRef]
  18. Cheng, L.; Li, M.; Liu, Y.; Cai, W.; Chen, Y.; Yang, K. Remote sensing image matching by integrating affine invariant feature extraction and RANSAC. Comput. Electr. Eng. 2012, 38, 1023–1032. [Google Scholar] [CrossRef]
  19. Sedaghat, A.; Ebadi, H. Remote Sensing Image Matching Based on Adaptive Binning SIFT Descriptor. IEEE Trans. Geosci. Remote Sens. 2015, 53, 5283–5293. [Google Scholar] [CrossRef]
  20. Yao, Y.; Zhang, Y.; Wan, Y.; Liu, X.; Guo, H. Heterologous Images Matching Considering Anisotropic Weighted Moment and Absolute Phase Orientation. Geomat. Inf. Sci. Wuhan Univ. 2021, 46, 1727–1736. [Google Scholar]
  21. Xiang, Y.; Wang, F.; You, H. OS-SIFT: A robust SIFT-like algorithm for high-resolution optical-to-SAR image registration in suburban areas. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3078–3090. [Google Scholar] [CrossRef]
  22. Chen, J.; Yang, M.; Peng, C.; Luo, L.; Gong, W. Robust Feature Matching via Local Consensus. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  23. Li, J.; Hu, Q.; Ai, M. 4FP-Structure: A Robust Local Region Feature Descriptor. Photogramm. Eng. Remote Sens. 2017, 83, 813–826. [Google Scholar] [CrossRef]
  24. Liu, Y.-X.; Mo, F.; Tao, P.-J. Matching Multi-Source Optical Satellite Imagery Exploiting a Multi-Stage Approach. Remote Sens. 2017, 9, 1249. [Google Scholar] [CrossRef] [Green Version]
  25. De Tone, D.; Malisiewicz, T.; Rabinovich, A. Superpoint: Self-supervised interest point detection and description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 224–236. [Google Scholar]
  26. He, H.; Chen, M.; Chen, T.; Li, D. Matching of remote Sensing Images with Complex Background Variations via Siamese Convolutional Neural Network. Remote Sens. 2018, 10, 355. [Google Scholar] [CrossRef] [Green Version]
  27. Yang, Z.; Dan, T.; Yang, Y. Multi-Temporal Remote Sensing Image Registration Using Deep Convolutional Features. IEEE Access 2018, 6, 38544–38555. [Google Scholar] [CrossRef]
  28. Fan, D.; Dong, Y.; Zhang, Y. Satellite image matching method based on deep convolution neural network. Acta Geod. Cartogr. Sin. 2018, 47, 844–853. [Google Scholar]
  29. Dong, Y.; Jiao, W.; Long, T.; Liu, L.; He, G.; Gong, C.; Guo, Y. Local Deep Descriptor for Remote Sensing Image Feature Matching. Remote Sens. 2019, 11, 430. [Google Scholar] [CrossRef] [Green Version]
  30. Xu, C.; Liu, C.; Li, H.; Ye, Z.; Sui, H.; Yang, W. Multiview Image Matching of Optical Satellite and UAV Based on a Joint Description Neural Network. Remote Sens. 2022, 14, 838. [Google Scholar] [CrossRef]
  31. Xiong, J.-X.; Zhang, Y.-J.; Zheng, M.-T.; Ye, Y.-X. An SRTM assisted image matching algorithm for long-strip satellite imagery. J. Remote. Sens. 2013, 17, 1103–1117. [Google Scholar]
  32. Du, W.-L.; Li, X.-Y.; Ye, B.; Tian, X.-L. A fast dense feature-matching model for cross-track pushbroom satellite imagery. Sensors 2018, 18, 4182. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Ling, X.; Huang, X.; Zhang, Y.; Zhou, G. Matching Confidence Constrained Bundle Adjustment for Multi-View High-Resolution Satellite Images. Remote Sens. 2020, 12, 20. [Google Scholar] [CrossRef]
  34. Brown, M.; Goldberg, H.; Foster, K.; Leichtman, A.; Wang, S.; Hagstrom, S.; Bosch, M.; Almes, S. Large-Scale Public Lidar and Satellite Image Data Set for Urban Semantic Labeling. In Proceedings of the SPIE Defense + Security, Orlando, FL, USA, 15–19 April 2018. [Google Scholar]
  35. SpaceNet on Amazon Web Services (AWS). Available online: https://spacenetchallenge.github.io/datasets/datasetHomePage.html (accessed on 15 October 2018).
  36. Dataset Was Created for the IARPA CORE3D Program. Available online: https://www.iarpa.gov/index.php/research-programs/core3d (accessed on 24 April 2022).
  37. Sarlin, P.E.; De Tone, D.; Malisiewicz, T.; Rabinovich, A. SuperGlue: Learning Feature Matching with Graph Neural Networks. In Proceedings of the 2020 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 4937–4946. [Google Scholar]
Figure 1. Workflow of the proposed method.
Figure 1. Workflow of the proposed method.
Remotesensing 14 05617 g001
Figure 2. Greedy solution process. (a) The beginning iteration. (b) The next iteration.
Figure 2. Greedy solution process. (a) The beginning iteration. (b) The next iteration.
Remotesensing 14 05617 g002
Figure 3. Post processing.
Figure 3. Post processing.
Remotesensing 14 05617 g003
Figure 4. Experimental data. (a) WorldView-3 pair near Jacksonville. (b) WorldView-3 pair near Omaha. (c) WorldView-3 pairs near Buenos Aires. (d) Gaofen-7 pair near Tibet.
Figure 4. Experimental data. (a) WorldView-3 pair near Jacksonville. (b) WorldView-3 pair near Omaha. (c) WorldView-3 pairs near Buenos Aires. (d) Gaofen-7 pair near Tibet.
Remotesensing 14 05617 g004
Figure 5. Analysis of P .
Figure 5. Analysis of P .
Remotesensing 14 05617 g005
Figure 6. Analysis of N o p t .
Figure 6. Analysis of N o p t .
Remotesensing 14 05617 g006
Figure 7. The performances of the proposed method against various positioning errors.
Figure 7. The performances of the proposed method against various positioning errors.
Remotesensing 14 05617 g007
Figure 8. Visual comparisons with/without positioning error corrections.
Figure 8. Visual comparisons with/without positioning error corrections.
Remotesensing 14 05617 g008
Figure 9. Evaluations of different methods. (a) Omaha data. (b) Tibet data.
Figure 9. Evaluations of different methods. (a) Omaha data. (b) Tibet data.
Remotesensing 14 05617 g009
Figure 10. Image block distributions of SFB and the proposed method. (a) SFB. (b) The proposed method.
Figure 10. Image block distributions of SFB and the proposed method. (a) SFB. (b) The proposed method.
Remotesensing 14 05617 g010
Figure 11. Comparisons with Buenos Aires dataset. (a) Comparisons in A n u m . (b) Comparisons in A o r i e n t a t i o n .
Figure 11. Comparisons with Buenos Aires dataset. (a) Comparisons in A n u m . (b) Comparisons in A o r i e n t a t i o n .
Remotesensing 14 05617 g011
Figure 12. An example image tile in Pair 2–4.
Figure 12. An example image tile in Pair 2–4.
Remotesensing 14 05617 g012
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gong, D.; Huang, X.; Zhang, J.; Yao, Y.; Han, Y. Efficient and Robust Feature Matching for High-Resolution Satellite Stereos. Remote Sens. 2022, 14, 5617. https://doi.org/10.3390/rs14215617

AMA Style

Gong D, Huang X, Zhang J, Yao Y, Han Y. Efficient and Robust Feature Matching for High-Resolution Satellite Stereos. Remote Sensing. 2022; 14(21):5617. https://doi.org/10.3390/rs14215617

Chicago/Turabian Style

Gong, Danchao, Xu Huang, Jidan Zhang, Yongxiang Yao, and Yilong Han. 2022. "Efficient and Robust Feature Matching for High-Resolution Satellite Stereos" Remote Sensing 14, no. 21: 5617. https://doi.org/10.3390/rs14215617

APA Style

Gong, D., Huang, X., Zhang, J., Yao, Y., & Han, Y. (2022). Efficient and Robust Feature Matching for High-Resolution Satellite Stereos. Remote Sensing, 14(21), 5617. https://doi.org/10.3390/rs14215617

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop