Next Article in Journal
Introduction of the Advanced Meteorological Imager of Geo-Kompsat-2a: In-Orbit Tests and Performance Validation
Previous Article in Journal
Testing and Validating the Suitability of Geospatially Informed Proxies on Land Tenure in North Korea for Korean (Re-)Unification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Information Content to Select Keypoints for UAV Image Matching

1
Geomatics Engineering Faculty, K.N. Toosi University of Technology, Valiasr St., 1996715433 Tehran, Iran
2
3D Optical Metrology (3DOM) Unit, Bruno Kessler Foundation (FBK), Via Sommarive, 18, 38123 Trento, Italy
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(7), 1302; https://doi.org/10.3390/rs13071302
Submission received: 3 February 2021 / Revised: 10 March 2021 / Accepted: 20 March 2021 / Published: 29 March 2021

Abstract

:
Image matching is one of the most important tasks in Unmanned Arial Vehicles (UAV) photogrammetry applications. The number and distribution of extracted keypoints play an essential role in the reliability and accuracy of image matching and orientation results. Conventional detectors generally produce too many redundant keypoints. In this paper, we study the effect of applying various information content criteria to keypoint selection tasks. For this reason, the quality measures of entropy, spatial saliency and texture coefficient are used to select keypoints extracted using SIFT, SURF, MSER and BRISK operators. Experiments are conducted using several synthetic and real UAV image pairs. Results show that the keypoint selection methods perform differently based on the applied detector and scene type, but in most cases, the precision of the matching results is improved by an average of 15%. In general, it can be said that applying proper keypoint selection techniques can improve the accuracy and efficiency of UAV image matching and orientation results. In addition to the evaluation, a new hybrid keypoint selection is proposed that combines all of the information content criteria discussed in this paper. This new screening method was also compared with those of SIFT, which showed 22% to 40% improvement for the bundle adjustment of UAV images.

1. Introduction

Recent developments in Unmanned Arial Vehicles (UAVs) have changed them to a low-cost, flexible and feasible acquisition system for many photogrammetry and remote sensing applications [1,2,3]. UAV images can be used to produce various outputs, such as high-density point clouds, object 3D models, mosaic images, cadastral maps and orthophotos of high resolution. The quality of products for UAV photogrammetry depends primarily on the accuracy and completeness of keypoints extracted from the images.
By far, numerous algorithms [4,5,6,7,8,9] dealing with keypoint extraction have been developed. Due to a large number of keypoint detectors and descriptors, many surveys [10,11] discussing their advantages and drawbacks have been reported. Furthermore, many works have evaluated and compared the performance of keypoint extraction and description techniques for various image matching tasks [12,13,14,15]. Clearly, the number and distribution of the extracted keypoints play a vital role in the reliability and accuracy of the image matching results, which is a very critical factor in high-resolution UAV photogrammetry bundle adjustment [16].
Many keypoints are extracted in current feature-based matching algorithms, the number of which depends directly on the content of the image. In addition to suffering from scale, rotation, viewing angle, image blur and illumination changes [17,18], UAV images are usually taken with inaccurate non-metric cameras and therefore the possibility of getting multiple incorrect matches is usually high. Besides, the huge number of keypoints detected make the matching process time consuming and error prone [19]. The number of extracted keypoints is reduced either during the matching process using some stricter or adaptive thresholds [20,21,22,23] or subsequently through matchability prediction [24,25] or the keypoint screening process [26,27].
In the case of stricter keypoint detection, the number of keypoints is controlled by setting stricter thresholds, which extract fewer but more repeatable and stable keypoints [25]. Adaptive threshold setting methods are also intended to dynamically update these thresholds instead of using a fixed threshold [21,28]. On the other hand, matchability prediction strategies attempt to explicitly predict the matchability of a particular keypoint during descriptor computing and keypoint matching processes [24,26]. A classifier is usually trained on a large set of descriptors that are successfully matched to predict which keypoints would likely survive the matching test.
In addition to eliminating keypoints that lead to incorrect matches or do not produce any match at all, keypoint selection/screening strategies are also capable of providing well-distributed matched keypoints. Furthermore, screening approaches provide the opportunity to use various relevant keypoint selection criteria based on different objectives and applications. These advantages caused keypoint selection strategies to receive considerable attention in recent studies [16,27,29]. However, keypoint selection must be performed very carefully as removing too many points blindly can easily lead to a gap in the reconstructed model.
Up to now, numerous selection criteria/scores have been proposed for keypoint selection purposes. These criteria either result from the keypoint detecting and description processes or are computed using the image’s information content. The first group works by using the keypoint detection and description outputs such as keypoint locations in image space [29], keypoint scale in scale pyramid [16,30] and their equivalent descriptor elements [31]. The second group includes criteria such as spatial saliency [32], entropy [30] and texture coefficient [19], while the latter includes scores such as robustness and scale [16], confusion risk [31] and keypoint distance [28]. In essence, they are measures to distinguish gray-level patterns at a keypoint location and show the variation of the underlying grey values. The suitability of keypoints could be assessed using one or more of these criteria.
Among the above two groups, the information content-based criteria are the most frequently used methods [16,23,30,32,33,34,35], which is why their effect on keypoint selection is studied in this paper. For example, in studies [16,35,36], a selection strategy is proposed that extracts high-quality SIFT keypoints. The entropy of the selected keypoints in the corresponding Gaussian image is considered as a keypoint distinctiveness measure. As another example, in ACT-SIFT [21], robust keypoints against illumination changes are extracted by checking the keypoint entropy through an iterative process. In Mukherjee et al. [32], the selection criterion relies on the detectability, distinctiveness and repeatability of the keypoints. These scores are combined to give a keypoint a saliency score. The keypoints are ranked according to their saliency values to filter out weak or irrelevant keypoints based on a given threshold. Furthermore, in a so-called UC algorithm [14], the quality of the keypoints is controlled using a competency parameter that consists of three quality measures which are robustness, spatial saliency and scale through a gridding schema.

Paper Aims and Contributions

To the best of authors knowledge, although learning-based keypoint detectors and descriptors start to be available [37,38,39,40], most researchers have focused mainly on developing handcrafted new methods/criteria for keypoint detection and/or description. They usually evaluate the overall accuracy of points produced by different detectors. Unfortunately, no work is found that compares the quality of keypoint thresholding methods. However, such information can significantly help researchers develop hybrid detectors (see Section 7) that can produce higher quality keypoints and improve (UAV) image bundle adjustment quality. To this end, in this paper, the quality of keypoint selection criteria is deeply studied. Also, as mentioned, there exist many criteria which can be used for keypoint selection. This paper focuses on evaluating the performance of information content based keypoint selection criteria. This is because they are among the most frequently used methods for keypoint selection. Also, all studies that have evaluated the accuracy of detectors have carried it out only in 2D (i.e., only for matching performance evaluation). However, in this paper, in addition to performing assessments in 2D, we evaluate them in 3D. The goal of 3D evaluations is to study the effect of the selected keypoints on bundle adjustment of UAV images. The reported experiments are extensive and include several 2D and 3D quality measures. In 2D, they include repeatability, recall, precision, positional accuracy and spatial distribution. In 3D, they are the RMSE of re-projection errors, visibility of 3D points in more than three images, the average number of rays and intersection angles per 3D point. Together with the quality measures employed in our experiments, the methods could be used as a new and robust framework to effectively and rigorously evaluate any detection or selection algorithm/criterion accuracy.
The keypoint selection criteria hereafter presented (Section 3) include the essential and popular ones, which are entropy, spatial saliency and texture coefficient. Detectors used for the production of initial keypoints are those which are well known for the extraction of scale and affine invariant features and include scale-invariant feature transform (SIFT) [6], speeded up robust features (SURF) [4], binary robust invariant scalable keypoints (BRISK) [5] and maximally stable extremal regions (MSER) [41]. These methods are used and discussed in different previous studies; therefore, we can compare our results with those reported in previous studies.
The remainder of this paper is organized as follows. A detailed explanation of the criteria for selecting keypoints is provided in Section 2. The proposed evaluation methodology is discussed in Section 3. Used datasets are presented in Section 4 while results and discussion on both simulation and real data are reported in Section 5 and Section 6, respectively. A proposal for a hybrid keypoint selection algorithm is presented in Section 7, before concluding the paper with final remarks and suggestion for future studies.

2. Keypoint Selection Criteria

We hereafter present the criteria used for selecting the best keypoints extracted from an image and used for image orientation. The purpose of the keypoint selection is to obtain an adequate number of high-quality interest points that are likely to be matched correctly. The quality of the extracted keypoints can be assessed using different quality measures. Some of the most popular ones studied in this paper are entropy, spatial saliency and texture coefficient.
  • Entropy: This refers to the uniqueness of the keypoints and their descriptors. The entropy of a keypoint in the relevant Gaussian image and within a local region around the feature point shows its distinctiveness. It can be computed using the following equation.
    E = P i l o g 2 P i .
    In Equation (1), Pi is the histogram-driven probability of the ith pixel within the local region around the detected keypoint. The scale of the extracted feature (σ) is used to select the relevant smoothed image. The local region is considered a surrounding circle with radius r = 3σ as suggested by [30].
2.
Spatial saliency: Another property dealing with image information content level is spatial saliency. This criterion deals with detecting visually distinct and prominent areas inside the image which are more critical and contain valuable information. The visual attention approach is the process that allows machine vision systems to detect the most essential and informative regions in the image. Among the various algorithms computing spatial saliency in images, the Weighted Maximum Phase Alignment (WMAP) model [42] offered appropriate results on various datasets [43]. This model uses the local phase (energy) information of the input data to achieve the saliency map by the integration of the WMAP measures for each pixel of each color channel (c) as follows:
S a l i e n c y ( x . y ) = c = 1 3 W M A P ( x . y ) .
To achieve the WMAP measures, the level of local phase alignment of the Fourier Harmonics is weighted by the strength of the visual structure in each scale to measure its importance. Therefore, the WMAP measure can be computed as follows:
W M A P ( x . y ) = w f d n · m a x i = 1 s { f M . i ( x . y ) · cos θ i } ,
where f M . i ( x . y ) is the monogenic signal for the ith scale and θ i is the angle between vectors f M . i ( x . y ) and E l o c a l . This angle measures the deviation of the local phase in the monogenic signal at the ith scale with respect to the local energy vector in pixel (x, y). A simple implementation of the monogenic signal in the frequency domain and the method to achieve the w f d n factor can be found in [44]. Once the saliency map is computed in the relevant Difference of Gaussians (DoG) image, the average saliency of the pixels inside the local region is computed for each keypoint. The radius of the local region is r = 3σ.
3.
Texture coefficient: The grey values underlying the detected features should have enough variation to generate a unique descriptor for each keypoint. To define this variation, a coefficient is used that can define the local texture of the image. This coefficient is the standard deviation of N × N neighboring pixels surrounding each keypoint. It is computed using the gray values of the smoothed image produced during the keypoint extraction process using the following equation:
t x t c o e f = l a y e r 1 l a y e r + 1 1 N 2 i = 1 n j = 1 m ( x i j x ¯ ) 2 3 .
In Equation (4), m and n are coordinates of the keypoint in the image, N is the number of neighbouring pixels around the target pixel, layer is the relevant DoG scale space image, x i j is the grey level value of the pixels in the relevant scale space and x ¯ is the mean grey level value in the template.

3. Evaluation Methodology

The proposed framework evaluates the performance of selection algorithms in two separate steps using both synthetic and real datasets. In this section, the methodology of evaluating the effect of information content criteria on matching results is described. The evaluations use either synthetic or real images. As for the synthetic data, the evaluation process is used at the end of the descriptor matching phase. This is a 2D process and is done to assess the efficiency of information content criteria in selecting high quality keypoints. In contrast, the assessments carried out using the real images are 3D in nature and their goal is to study the effect of point selection strategies on the quality of UAV image orientation.
The methodology adopted in each case (i.e., using synthetic or real images) is described first. Then, the quality measures used to evaluate the results are explained. Finally, details of applying the suggested criteria for keypoint selection are given in the Implementation sub-section.

3.1. Evaluations Using Synthetic Data

A synthetic data set was prepared to provide the required data for the first phase. In this dataset, one image was considered as a reference and a known geometric transformation was applied to generate the second one. To generate the synthetic dataset, three geometric transformations including rotation, scale and 2D projective transformation were applied. The details are discussed below.
Suppose G ( j . k )   1 j J .   1 k K is a discrete output image which is created by applying a geometric transformation on input image of F ( p . q )   1 p P .   1 q Q . Let u q and v p be the coordinate of a pixel in the input image. For each transformation, the x k and y j are the coordinates of the output pixel.
  • Scaling: Spatial size scaling of an image can be obtained by modifying the coordinates of the input image according to the following equation:
    [ x k y j ] = [ s x s y ] · [ u q v p ] ,
    where s x and s y are positive-valued scaling factors, but not necessarily integer-valued. If s x and s y are each greater than unity, the applied equation will lead to magnification. In our dataset, five scaled images were generated using different scale factors ranging from 1.2 to 2.6.
2.
Rotation: Rotation of an input image about its origin can be accomplished by the following equation:
[ x k y j ] = [ cos θ sin θ sin θ cos θ ] · [ u q v p ] ,
where θ is the counter-clockwise rotation angle with respect to the horizontal axis of the input image. In our prepared synthetic dataset, six images were created with rotation angles varying from 5° to 155° by a difference of 30° at each step.
3.
2D Projective transform: The projective transformation shows how the perceived objects change when the viewpoint of the observer changes. This transformation allows creating perspective distortion and is computed using the following equation:
[ x k y j 1 ] = [ a 1 a 2 b 1 a 3 a 4 b 2 c 1 c 2 1 ] H · [ u q v p 1 ] ,
where c = [ c 1 c 2 ] is the projection vector, K = [ a 1 a 2 a 3 a 4 ] is a rotation matrix and b = [ b 1 b 2 ] is the translation vector. In our prepared synthetic dataset, five images were created with different values for matrix H simulating images captured from different viewing angles. The projection vector (c) values were changed to generate each simulated image. The greater values for projection vector (c) simulated images captured from a larger viewpoint angle. A constant rotation value equal to 1° and translation vector equal to zero was considered for all images.
In this paper, different geometric transformations, such as 2D affine or projective transformations, are considered in order to evaluate the performance of keypoint selection algorithms under different imaging conditions. The synthetic dataset was designed to eliminate the effect of image content and texture quality on the matching results of selection algorithms. Furthermore, the stability of keypoint selection algorithms could be investigated under different imaging conditions. As shown in Figure 1, a keypoint detector was used to extract keypoints and compute descriptors for each of them. After that, the keypoint selection stage was performed based on Algorithm 1 (pseudo-code details are shown in Figure 2) to select high-quality keypoints. Finally, to determine the performance of keypoint selection methods, nominated keypoints were matched.
The pseudo-code of Algorithm1 is shown in Figure 2. In summary it includes the following steps:
  • The initial keypoint locations and scales were extracted in both original and transformed images.
  • The quality values for each keypoint were computed separately. The quality measures of entropy, spatial saliency and texture coefficient were computed using Equation (1), WMAP method and Equation (4), respectively. Each value of entropy, spatial saliency and texture coefficient was considered as the competency value of the initial keypoints.
  • For each criterion, the mean competency value of all initial keypoints was considered as a threshold. Most researchers [30,34,35] define this threshold empirically. However, in this paper, we use the mean competency value which, first of all, is calculated robustly over the corresponding image. Secondly, as the values computed for each criterion are all positive, their average has a meaningful value that could well serve as the threshold for eliminating the weaker points. This way, the selected points all have a competency value that is above the average magnitude of all points.
  • The initial keypoints are ordered based on their competencies. Here, only the keypoints having a competency value higher than the computed threshold were selected. Other keypoints were discarded.

3.2. Evaluations Using Real Images

In the second phase of the evaluation of keypoint selection criteria, the performance was checked using real image blocks and the Structure from Motion (SfM) pipeline. These tests verify the influence of the keypoint selection methods on the orientation of the image. Keypoint selection algorithms certainly affect the number of matches and their distribution. This phase of the evaluation explores how the keypoint selection methods influence the results of the image orientation. In order to have a standard evaluation procedure, the keypoints were extracted and described in the first step. After that, each of the keypoint selection algorithms was applied to the dataset separately. After the matching procedure and the tie point triangulation, the bundle adjustment was initiated. The results achieved are compared to the original SfM pipeline. Figure 3 summarizes the whole process.

3.3. The Quality Measures Used for Comparisons

In the synthetic dataset, five criteria including repeatability, recall, precision, positional accuracy and spatial distribution quality, were used to evaluate the capability of the keypoint selection methods. Repeatability is defined as the percentage of the keypoints detected on the scene part visible in both images. The recall and precision criteria were computed using the following equations.
P r e c i s i o n = C M C M + F M
R e c a l l = C M C M + F N
In Equations (8) and (9), Correct Match (CM) and False Match (FM) are numbers of correctly and falsely matched pairs, respectively. False Negative (FN) is the number of existing correctly matched pairs that are incorrectly rejected. The positional accuracy of the keypoint detector algorithm is calculated using the Root-Mean-Square Error (RMSE). The RMSE is computed using the pairs of keypoints that are correctly matched and the ones determined by the computed transformations. In order to evaluate the quality of the distribution, the global coverage index, based on the Voronoi diagrams, is calculated as follows:
α = i = 1 n A i A T o t a l
where A i is the area of the ith Voronoi cell, n is the number of Voronoi cells and A T o t a l is the area of the whole image. The larger the α value, the better the spatial distribution of the matched pairs.
In a real image block dataset, four criteria, including the following outcomes of bundle adjustment, were analyzed for comparison.
  • RMSE of the bundle adjustment: This criterion expresses the re-projection error of all computed 3D points.
  • Average number of rays per 3D point: It shows the redundancy of the computed 3D object coordinates.
  • Visibility of 3D points in more than three images: It indicates the number of the triangulated points which are visible in at least 3 images in the block.
  • Average intersection angles per 3D points: This criterion shows the intersection angle of 3D points, which are determined by the triangulation. A higher intersection angle of homologues rays provides more accurate 3D information.

3.4. Implementation

The original detectors of SIFT, SURF, MSER and BRISK were used to extract a sufficient number of initial keypoints for each synthetic image. For each keypoint detected, based on the gradient orientation histogram described in [6], one orientation was assigned to achieve rotation invariance. The descriptors were finally computed using the SIFT descriptor for all keypoint detectors in all experiments using the VLFeat toolbox [45]. All detectors were implemented in MATLAB (R2017b). Table 1 shows the parameters set for each detector in the synthetic image dataset.
Each of the entropy, spatial saliency and texture coefficient criterion was then calculated separately for each keypoint. Entropy and spatial saliency were calculated based on the local region with radius r = 3σ. A 7 × 7 template around the target pixel was also considered for texture coefficient criterion computation.
In order to produce the required synthetic dataset, three simple image transformations were applied. The scaled images were generated using a scale transformation with different scale factors from 1.2 to 2.6. A rotation transformation was also applied with rotation angles varying from 5° to 155° by a difference of 30° at each step. To generate an image from a different view, a projective transformation was used to simulate angle differences between 20° and 60°.
To evaluate the matching results in the synthetic dataset, due to the known geometrical relationship between each pair of images, the values CM, FM and FN could be calculated automatically. A spatial threshold of 1.5 pixels was used to separate correct matches from false matches. The precision and recall criteria could be computed using CM, FM and FN values. The RMSE value was also calculated using the location of correctly matched keypoints and their computed location determined by the known transformations. The global coverage index was also calculated using the Voronoi diagrams computed by locations of correctly matched keypoints and Equation (10).
For the real data analysis, the SfM pipeline was implemented in MATLAB with the exception of the bundle adjustment written in C++ using the Levenberg–Marquardt algorithm [46]. Some of the codes provided in [47] were used to implement the proposed algorithm. All experiments were performed on a PC with a 3.40 GHz Intel Core i3-4130 processor and 8 GB of RAM.

4. Description of Datasets

Performance evaluation was carried out on the basis of two different datasets, i.e., synthetic image dataset and real image dataset. In the synthetic dataset, one image was considered as a reference, and a known geometric transformation was used to generate the second one. In this paper, two reference images of different textures were used. The first one includes mostly man-made objects (i.e., buildings, roads and trees) constructed on a flat area, whereas the second one covers a soil-type hilly area. For the simulation of different image conditions, three different types of geometric transformations including rotation, scaling and viewing changes were applied. In the transformed image, no radiometric differences are considered. Figure 4 is an example of the synthetic images generated for each type of transformation used in the experiments.
Three UAV image blocks were used in the real image network dataset. The selected images were captured by low-cost aircraft and cover both urban and rural areas with different patterns of texture. Datasets are characterized by different image resolution, the number of images, the overlap of images, changes in illumination and surfaces with different textures and sizes. Figure 5 shows the datasets used with their sample images and camera networks, and Table 2 summarizes their specifications.

5. Results

This section presents the evaluation results of keypoint selection methods using synthetic and real image datasets (Section 4). In the first part, the effect of selection methods is thoroughly evaluated using the SIFT detector on the synthetic dataset (Section 5.1). In the following, the results obtained using the detectors MSER, SURF and BRISK are shown in a way similar to that of SIFT. In the second part, the results of the evaluations carried out using the real image datasets are presented (Section 5.2). It should be noted that in Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10 the figures shown on the left are obtained using S1 images and those on the right using the S2 images.

5.1. Matching Results Obtained Using the Synthetic Dataset

This section explains the comparative results of the experiments, including repeatability, recall, precision and spatial distribution quality for the three keypoint selection strategies.
Before presenting the results and discussions in the following sections, a sample evaluation is described to illustrate how the results in the following sections are obtained. Figure 6 shows the selected keypoint using the entropy selection criterion, the spatial saliency and the texture coefficient of one of the image pairs in the dataset. A synthetic rotation angle equal to 5° was applied to the reference image to generate geometric differences between the image pairs. Initial keypoints were extracted using SIFT for each image. The keypoints are shown in red. As can be seen, the selected keypoints for the different selection strategies are shown in green. The final matched keypoints are also shown in yellow. Generally, each selection strategy chooses different keypoint subsets with different properties, with a relatively small overlap between the results. The number of keypoints initially extracted in the reference image and the transformed image is 688 and 820, respectively. The keypoint method of selection of entropy, spatial saliency and texture coefficient reduces these numbers to 360,270 and 430 for the reference image and 440,350 and 530 for the transformed one. These numbers show that between 39% and 62% of the initial keypoints are discarded during the selection process, but matching results are not degraded at all.

5.1.1. Repeatability

The comparative results for repeatability of the SIFT detector using different keypoint reduction methods are shown in Figure 7. As shown, the repeatability of the selected keypoints is not significantly degraded by screening strategies compared to the original SIFT. Generally speaking, SIFT-Saliency with average repeatability of 65% has the best performance. SIFT-texture coefficient and SIFT-entropy are ranked second and third with 62% and 56% repeatability, respectively. SIFT-entropy has the lowest repeatability in the experiments conducted.
The results obtained for repeatability tests on images with different scales (Figure 7b) show that the SIFT texture coefficient with an average repeatability increase of 10% has superior results. SIFT-saliency results indicate that as the scale difference between images increases, the number of salient keypoints decreases significantly. This problem affects the repeatability of the selected keypoints. The aggregation of salient objects in a small part of the image is the main reason for this.
As Figure 7c shows, the repeatability of the SIFT detector degrades significantly with an increase in viewpoint angle. Repeatability of screening methods under different viewing angles is reduced by 5% for the image (S1), while in the image (S2), the keypoint selection strategies are superior in comparison to the original SIFT. SIFT-saliency obtains an average repeatability of 35% in the image (S2) and SIFT-entropy obtains 37% of the repeatable keypoints. The weakest results in this regard belong to SIFT-texture coefficient.
From the repeatability tests performed in the synthetic dataset it can be concluded that in spite of keypoint elimination, the repeatability of the selection strategies is reduced by an average of 10% to 15%. As a result, it seems losing the keypoints repeatability is not big enough to be a serious challenge in keypoint reduction strategies. Overall, the SIFT-saliency has the best performance in terms of repeatability results except for scale transformation in low-quality texture images. The SIFT-texture coefficient performs well in rotation and scale transformation but is not successful under viewpoint transformations. SIFT-entropy has the weakest results regarding repeatability quality measures in all experiments.

5.1.2. Precision

The precision of the matching results for all transformations is shown in Figure 8. The results of Figure 8a indicate that the original SIFT detector and screening methods are almost invariant with rotational changes. Selecting keypoints on the basis of their saliency increases the precision performance of SIFT in S1 dataset. SIFT-saliency with an average precision of 96.4% has superior results in the (S1) dataset. However, the SIFT-texture coefficient and the SIFT-entropy perform differently based on the scene type. Using entropy as a feature selection criterion degrades the precision of matching results in low dynamic range images (i.e., image (S2)) under rotation variations. This is mainly due to a large number of mismatches in the matching process.
As shown in Figure 8b, the matching capability of the SIFT detector and the keypoint selection methods deteriorates as the scale difference between the images increases. For the image (S1), the selection of keypoints based on the value of the texture coefficient increases the precision of the matching results. The average value of the SIFT-texture coefficient precision criteria is 85% in the image (S1). Nevertheless, none of the keypoint selection methods perform well in the image (S2). This is mainly due to the low number of extracted keypoints. However, the SIFT-entropy shows different results when the image is changed. The average precision results for SIFT-entropy are 61% for images (S1) and 76% for images (S2).
The results of Figure 8c show that SIFT-texture coefficient and SIFT-saliency improve the efficiency of the keypoint detector in terms of precision results dealing with different viewpoint images. Selected keypoints using SIFT-saliency have achieved an average precision of 78% and 83% in images (S1) and (S2), respectively. However, SIFT-texture coefficient does not perform well in (S2) dataset.
The precision results in the synthetic dataset indicate that the performance of keypoint selection methods varies depending on the image type and the applied keypoint selection criteria. Overall it can be said that the SIFT-saliency and SIFT-texture-coefficient perform better in images that contain surfaces of different textures and sizes. For the areas with poor texture quality and under viewpoint transformation, the SIFT-texture coefficient does not perform well. In all our experiments, SIFT entropy has the weakest results in terms of precision results.

5.1.3. Recall

The results of the calculation of the recall criteria for the different transformations are shown in Figure 9. As can be seen in Figure 9a, the applied selection methods outperform the original SIFT detector in all experiments. SIFT performance improves between 5% to 10% in terms of recall results using different keypoint selection methods. The SIFT-texture coefficient performs the best in images (S1) and (S2) with an average of 95%.
In most experiments for images with different scale factors and viewing angles, as shown by Figure 9b,c, the recall criterion of screening methods is relatively higher than the original SIFT. The top-ranked keypoint selection method in all transformations in the image (S1) and (S2) is SIFT-saliency with an average recall criterion of 88%, 87% and 93%, respectively.
The higher recall results in the synthetic dataset indicate that almost all of the keypoints were selected during the matching process. This means the number of redundant keypoints that are not useful in the matching process are significantly reduced by the applied selection methods.

5.1.4. Global Coverage

The global coverage parameter for matching results can be seen in Figure 10. In experiments, the global coverage criterion for selected keypoints is relatively lower. As Figure 10a shows, SIFT-saliency has the lowest global coverage, with an average of 74% in the image (S1) and 53% in the image (S2). SIFT-texture coefficient has the best performance comparing keypoint discarding methods. The main reason is that the SIFT-texture coefficient method selects a higher number of keypoints during the selection process. Also, as can be seen, the screening-based version of the detectors select a subset of high-quality keypoints from the interest points that were initially extracted. However, the distribution of the selected keypoints is only reduced 5.1% to 14.41% compared to the original method.
As the scale difference between images increases, the area of Voronoi cells computed by the corresponding keypoints decreases. Consequently, the global coverage criterion is gradually reduced. This is mainly because the number of extracted keypoints decreases as the scale differences increase. As can be seen in Figure 10b, the performance of screening approaches is similar with different scale changes.
As Figure 10c shows, among the selection methods, SIFT-saliency has the weakest performance in terms of global coverage criteria, especially for the image (S2). The average global coverage criteria for SIFT-saliency for image (S1) is 63%, while the average value of this criterion is around 23.6% for image (S2). From this point of view, SIFT-entropy and SIFT-texture coefficient methods achieve similar results in the same image.
The results of global coverage in synthetic datasets indicate that, despite the keypoint selection approaches having better performance in terms of precision and recall values, the distribution of the selected keypoints cannot be controlled during the selection process and must be considered.

5.1.5. Matching Results in Detectors Other Than SIFT

Three other keypoint detectors, including MSER, SURF and BRISK, were analyzed to fully investigate the performance of keypoint selection methods. The matching results of these detectors are shown in Table 3, Table 4 and Table 5 and the best results are shown in Bold font. The mean and standard deviations under different geometric transformations were calculated for each assessment criterion. The higher standard deviation results of the image detectors show their different performance. Further discussion is provided in this section.

MSER

The matching results with the MSER detector are shown in Table 3, and the best results are shown in Bold font. By comparing the achieved results under different geometric transformations, it can be seen that MSER extracts the most repeatable keypoints but does not perform well in the matching process. In all experiments, the precision and stability of the obtained results in original MSER detector was lower than the keypoint selection methods. Generally speaking, keypoint selection methods achieved better results compared to the original MSER in all experiments. MSER-entropy with an average precision of 73.41% and an average recall of 24.18% outperforms other keypoint selection methods in images with varying rotation changes. For scaled images, MSER-saliency achieves superior results with average precision and recall criteria of 70.74% and 95.5%, respectively. In viewpoint change transformation, MSER-entropy gets the best results in terms of precision but considering recall parameter, MSER-saliency is the best.

SURF

The matching results using the SURF detector are summarized in Table 4. The original SURF has the best repeatability criteria results. SURF-entropy with an average precision of 81.5% and an average recall of 86% improves the matching capabilities of the detectors. SURF-texture coefficient with an average precision of 77.43% outperforms other keypoint selection algorithms dealing with images with different scale factors. SURF-Saliency with precision of 68.85% and a recall of 99.9% obtains the best results during the matching process in synthetic images with different viewing angles. In all tests, the original SURF has the best results in terms of the global coverage criterion.

BRISK

The matching results using the BRISK detector are shown in Table 5. BRISK gives greater robustness to large-scale changes. It improves matching performance over large-scale changes with an average precision of 71.74% and an average recall rate of 94.73%. Additionally, selecting keypoints based on their saliency measure increases the precision criteria around 9% and the recall criteria around 15%. Concerning the RMSE point of view, BRISK and BRISK-entropy have the best performance. In all tests, the original BRISK has the best results in terms of the global coverage criterion.

5.2. Results Obtained Using Real Images

In this section, keypoint selecting strategies were applied to blocks of real images to evaluate the effect of keypoint selection algorithms. To do so, each algorithm was performed separately on the actual image datasets. Following the pairwise matching process, the results achieved after the bundle solution were compared to the original SfM pipeline. The SIFT detector was selected for the extraction and description of the keypoints.
As the experiments on synthetic image pairs have shown, the number and distribution of extracted keypoints are decreased using keypoint selection methods. This can lead to either orientation failure due to the inadequate number of tie points or incorrect orientation results. Experiments in this section evaluate the impact of keypoint selection algorithms based on the results of the image orientation. The four criteria, as described in Section 3.3, are evaluated in the comparisons. Detailed findings will be discussed below.

5.2.1. RMSE of the Bundle Adjustment

The re-projection error of all computed 3D points is shown in Figure 11a for all datasets. As shown, the SIFT-texture coefficient has the best performance comparing keypoint screening methods in each dataset. SIFT-saliency with 1-pixel re-projection error has the weakest test results. The main reason for this is that the R3 dataset contains uniform and low-contrast images with a lower number of salient objects. For this reason, keypoints with larger scales are selected in a saliency-based strategy, leading to lower spatial resolution. The same is true for SIFT-entropy screening in the R1 dataset. The selected keypoints in SIFT-entropy algorithm are from a larger scale, leading to larger re-projection errors in the bundle adjustment.

5.2.2. Average Angles of Intersection

Since 3D points are calculated by triangulation, a higher angle of intersection of similar rays provides more accurate 3D details. As shown in Figure 11b, higher intersection angles are achieved in R1 and R2 datasets consisting of higher overlapping images (almost more than 70%). For R1 datasets with fewer overlapping images, the intersection angles are relatively smaller. In this respect, however, SIFT saliency performs better.

5.2.3. Average Rays Per 3D Point

As the overlap between images in a dataset increases, more redundancy was expected for 3D object coordinates. As shown in Figure 11c, higher average multiplicity for the tie-points is achieved in datasets R1 and R2 with higher overlapping images. The larger multiplicity belongs to the original SIFT and SIFT-texture coefficient keypoints with an average of 2.83 and 2.72 rays, respectively. However, for R1 datasets containing fewer overlapping images, SIFT-entropy and SIFT-texture coefficients with an average ray of 2.3 outperform other methods.

5.2.4. Visibility of 3D Points in More Than Three Frames

Figure 11d shows the visibility of the 3D points that are normalized with respect to all extracted points for each dataset. The visibility results indicate that 30% of the triangulate points using the SIFT-entropy strategy are visible in three images in the R1 dataset. The original SIFT and SIFT-texture coefficient strategies achieved the best results for R2 and R3 datasets with an average image overlap of 60% and 80%.

5.2.5. The Number of 3D Points

Figure 11e shows the number of the 3D points used for the bundle adjustment process. The results indicate that nearly 50% of the 3D points are eliminated using the SIFT-entropy and SIFT-texture coefficient strategies. The SIFT-saliency method reduces nearly 75% of 3D points compared to the original SIFT. However, the quality of the image orientation is not reduced except for the R3 dataset. This is because it contains a relatively lower number of salient objects.

5.2.6. Processing Time

The time required for the orientation of the image for each dataset is shown in Figure 11f. As shown, using the original SIFT in blocks with a large number of images is more time-consuming. Whereas selection-based approaches are more efficient, SIFT-saliency, SIFT-entropy and SIFT-texture coefficient keypoint selection approaches are the most time-efficient strategies for R1, R2 and R3 datasets.

6. Discussions

The presented evaluation framework (Section 5) compares results in two separate phases. In the first part, the evaluation process is performed at the end of the descriptor matching phase using synthetic datasets. The tests then compare the results of a real image block orientation to understand the performance of keypoint selection methods for 3D reconstruction purposes. The achieved results are respectively discussed in this section.
The achieved results on the synthetic dataset show that the repeatability of the selected keypoints is 10% to 15% degraded by keypoint selection strategies when SIFT detector is applied to extract keypoints. Similarly, repeatability of the selected keypoints is 3% to 20% lower when keypoints are extracted using MSER, SURF and BRISK detectors.
From the precision point of view, the performance of the applied screening methods was different depending on the image type and the keypoint selection criteria. Generally, in most experimental cases, precision was improved 10% to 15%. Furthermore, SIFT performance improves between 5% to 10% in terms of recall results using different keypoint selection methods. For the MSER detector, in all experiments the precision and recall of the obtained results in the original MSER detector was lower than the screening methods. Generally speaking, keypoint selection methods achieved 25% to 40% better precision results compared to the original MSER in the conducted experiments. Keypoint selection methods in the SURF detector had similar performance, enhancing the precision and recall results from 10% to 25% in synthetic dataset. Selecting keypoints based on their saliency measure increases the matching capability of the BRISK detector. Precision and recall criteria for both rotation and viewpoint change tests are improved in the BRISK detector. However, the selection of keypoints based on selection criteria measures does not affect the performance of the BRISK detector in scaled images.
The result of global coverage criteria for all detectors shows that the global coverage criterion for selected keypoints is relatively lower and should be considered.
Experiments on synthetic image pairs showed that screening methods decrease the number and distribution of selected keypoints. Further experiments were performed to evaluate the impact of these shortcomings on the results of the 3D reconstruction. Similar to synthetic dataset, the achieved results showed that screening methods perform well in real images. The re-projection error of computed 3D points is on average 15% to 30% decreased using selected keypoints. This means that the adequate number of tie points with an appropriate distribution are selected by screening methods. However, the keypoint selection strategies are not capable of controlling the scale of the selected keypoints. As a result, keypoints with larger scales may be selected, which leads to less accurate results in bundle adjustment. As far as computing time is concerned, keypoint selection approaches outperform conventional methods and improve time efficiency, especially for blocks with large number images.

7. Development of a New Hybrid Keypoint Selection Algorithm

As shown by the presented results on real and synthetic images (Section 5), keypoint selection methods, thanks to their own specifications and functionalities, have an overall positive impact on the accuracy of image matching and orientation. Each of the keypoint reduction methods discussed perform well when they are separately applied for keypoint selection. However, developing a hybrid keypoint selection algorithm that combines all these criteria simultaneously could be a more efficient solution. This is because it can provide the opportunity to compensate for the shortcomings of each keypoint selection criteria and, and thus, reinforce their strengths. On this basis, in this section, a new hybrid keypoint selection algorithm is proposed, which uses all the keypoint selection criteria at the same time.
In order to combine entropy, saliency and texture coefficient criteria to achieve a single competency value for each keypoint, a simple average ranking method is used. Suppose that n initial keypoints are extracted, and each criterion of entropy, saliency and texture coefficient are computed for them. For each quality measures, the rank vector is computed as follows:
R i = { r i j } j = 1.2.3 i = 1.2. n ,
where r i j represent the rank of ith keypoint in the jth quality measures. The rank of each keypoint in each quality criteria is calculated by sorting the quality measures values in descending order and assigning the rank according to its position.
Afterwards, the average rank for each keypoint ( r ¯ i ) is computed using the following equation:
r ¯ i = ( j n r i j ) n
where n is the number of criteria and r ¯ i is the average rank for the ith keypoint. Once the average rank for each keypoint is computed, the image is gridded and the keypoints with r ¯ i larger than the average rank of each cell is selected as the final keypoints.
To evaluate the proposed hybrid criterion, different experiments are carried out similar to those mentioned in Section 5.2. To this end, the real datasets of R1, R2 and R3 (Section 4) were used to assess the results of the image orientation. The four quality measures mentioned in Section 3.3, including re-projection error of bundle adjustment, average rays per 3D points, visibility of 3D points in more than three images and average intersection angle per each 3D point are used to compare the performance of the proposed algorithm to that of SIFT. The required processing time is also tested to evaluate the time efficiency of the proposed method. Detailed results are discussed below.

7.1. RMSE of the Bundle Adjustment

Figure 12a shows the RMSE of bundle adjustment in the hybrid proposed method and the original SIFT. The re-projection error of bundle adjustment for the proposed hybrid method is lower in all datasets. The re-projection error is on average 22% to 40% improved using the proposed hybrid algorithm.

7.2. Average Angles of Intersection

The average intersection angles in the proposed algorithm, as shown in Figure 12b, for dataset R2 and R3 are relatively smaller. For R1 datasets with fewer overlapping images, the average intersection angle for the proposed method is 3% larger in comparison to the original SIFT.

7.3. Average Rays Per 3D Point

The average multiplicity results are shown in Figure 12c and indicate that the proposed algorithm does not perform well from this aspect. For the R1 dataset, the average multiplicity for the tie-points for both the proposed method and the original SIFT is almost the same. However, for R2 and R3 dataset, the average multiplicity for the proposed method is 15% to 20% lower in comparison with the original SIFT.

7.4. Visibility of 3D Points in More Than Three Frames

The visibility of the 3D points is shown in Figure 12d, which are normalized with respect to all extracted points for each dataset. The visibility results indicate that in the R1 dataset for both the original SIFT and the proposed method, an average of 20% of the triangulate points are visible in more than three images. The visibility results in the proposed method and original SIFT differ 54% and 45% for R2 and R3 datasets, respectively.

7.5. The Number of 3D points

Figure 12e shows the number of the 3D points applied for the bundle adjustment process. A high number of keypoints are eliminated during keypoint selection. Generally, the proposed method eliminates averagely 70% of keypoints which leads to lower multiplicity and average rays per 3D point. However, the quality of the bundle adjustment is not reduced in comparison with the original SIFT.

7.6. Processing Time

The required processing time for the image orientation for each dataset is shown in Figure 12f. Similar to the results of Section 5.2.6, the proposed method is more efficient regarding the required processing time.
The evaluation results of the proposed method show that applying all the discussed criteria simultaneously for keypoint selection could provide better results regarding matching and 3D reconstruction results. The proposed algorithm enhances the performance of bundle adjustment in terms of re-projection error and processing time. However, the problem of setting the optimal threshold for keypoint selection/discarding is still a challenge. As shown by the experiments, a high number of keypoints are eliminated during keypoint selection which leads to lower performance in terms of multiplicity and average rays per each 3D point. Furthermore, despite the obtained results in Section 5.2, the problem of image scene type, which affected the results of the R3 dataset is solved by the combination of selection criteria.

8. Conclusions

Accurate keypoint extraction and matching is a key issue for various UAV photogrammetry applications. Previous studies have not paid due attention to the impact of the information content criteria on the quality of the selected keypoints. Several tests were carried out in this paper using simulated and real data to verify the impact of different information content criteria on keypoint selection tasks. To this end, three strategies have been chosen to select among the detected interest points. In these approaches, different quality measures were applied to the conventional keypoint detectors and extracted keypoints were selected based on their measure values. In summary, the following results have been obtained.
  • Each keypoint detector performs differently depending on the type of image and the type of scene. Enhancing the capabilities of the original detectors, keypoint selection strategies (i.e., entropy, saliency and texture coefficient), are advantageous for accurate image matching and UAV image orientation.
  • Despite the reduction of keypoints, the keypoint selection strategies applied almost preserve the distribution of corresponding points.
  • Information-based selection methods can effectively control the number of keypoints detected. As discussed (Section 5.1 and Section 5.2.5), in our experiments, the remaining keypoints range between 38% to 61% for the synthetic data and between 25% to 50% for the real data. As a result, the large number of detected keypoints that make the matching process time-consuming and error-prone is significantly decreased. The results of the image orientation also confirm this fact.
  • It should be noted that keypoint selection algorithms are not capable of finding matched pairs with lower scale over entire image. This is because there is a lack of control over the scale of the keypoints in the selection process.
It can generally be concluded that keypoint selection methods have a positive impact on the accuracy of image matching. Developing an approach that uses all the keypoint selection criteria simultaneously is, therefore, of interest. To this end, a new hybrid keypoint selection that combines all of the information content criteria discussed in this paper was proposed. The results from this new screening method were also compared with those of SIFT, which showed a 22% to 40% improvement for the bundle adjustment of UAV images. In addition, providing a strategy that considers the scale of keypoints is also suggested for further research. Last but not least, it is possible to use measures other than a criterion’s average as the threshold, below which the keypoints are discarded. This is done to study the effect of new thresholds on the quality of point selection. Therefore, it is suggested that this issue is further explored in future studies.

Author Contributions

Conceptualization, V.M., M.V. and F.R.; methodology, V.M. and M.V.; software, V.M. and M.V.; validation, V.M. and M.V.; formal analysis, V.M. and M.V.; investigation, V.M. and M.V.; resources, V.M. and M.V.; data curation, V.M. and M.V.; writing—original draft preparation, V.M; writing—review and editing, M.V and F.R.; visualization, V.M. and M.V.; supervision M.V. and F.R.; project administration, M.V.; funding acquisition, F.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study [48].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Colomina, I.; Molina, P. Unmanned Aerial Systems for Photogrammetry and Remote Sensing: A Review. ISPRS J. Photogramm. Remote Sens. 2014, 92, 79–97. [Google Scholar] [CrossRef] [Green Version]
  2. Nex, F.; Remondino, F. UAV for 3D Mapping Applications: A Review. Appl. Geomat. 2014, 6, 1–15. [Google Scholar] [CrossRef]
  3. Hassanalian, M.; Abdelkefi, A. Classifications, applications, and design challenges of drones: A review. Prog. Aerosp. Sci. 2017, 91, 99–131. [Google Scholar] [CrossRef]
  4. Bay, H.; Ess, A.; Tuytelaars, T.; van Gool, L. Speeded-up Robust Features (Surf). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  5. Leutenegger, S.; Chli, M.; Siegwart, R. Brisk: Binary Robust Invariant Scalable Keypoints. In IEEE International Conference on Computer Vision (ICCV); IEEE: New York, NY, USA, 2011. [Google Scholar]
  6. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  7. Verdie, Y.; Yi, K.; Fua, P.; Lepetit, V. Tilde: A Temporally Invariant Learned Detector. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  8. Wang, Z.; Fan, B.; Wu, F. Frif: Fast Robust Invariant Feature. In Proceedings of the BMVC, Bristol, UK, 9–13 September 2013. [Google Scholar]
  9. Xing, Y.; Zhang, D.; Zhao, J.; Sun, M.; Jia, W. Robust Fast Corner Detector Based on Filled Circle and Outer Ring Mask. IET Image Process. 2016, 10, 314–324. [Google Scholar] [CrossRef]
  10. Krig, S. Interest Point Detector and Feature Descriptor Survey. In Computer Vision Metrics; Springer: Berlin/Heidelberg, Germany, 2016; pp. 187–246. [Google Scholar]
  11. Lee, H.; Jeon, S.; Yoon, I.; Paik, J. Recent Advances in Feature Detectors and Descriptors: A Survey. IEIE Trans. Smart Process. Comput. 2016, 5, 153–163. [Google Scholar] [CrossRef] [Green Version]
  12. Ahmadabadian, A.H.; Robson, S.; Boehm, J.; Shortis, M.; Wenzel, K.; Fritsch, D. A comparison of dense matching algorithms for scaled surface reconstruction using stereo camera rigs. ISPRS J. Photogramm. Remote Sens. 2013, 78, 157–167. [Google Scholar] [CrossRef]
  13. Heinly, J.; Dunn, E.; Frahm, J.-M. Comparative Evaluation of Binary Features. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2012; pp. 759–773. [Google Scholar]
  14. Miksik, O.; Mikolajczyk, K. Evaluation of Local Detectors and Descriptors for Fast Feature Matching. In Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012), Tsukuba, Japan, 11–15 November 2012. [Google Scholar]
  15. Mousavi, V.; Khosravi, M.; Ahmadi, M.; Noori, N.; Haghshenas, S.; Hosseininaveh, A.; Varshosaz, M. The performance evaluation of multi-image 3D reconstruction software with different sensors. Measurement 2018, 120, 1–10. [Google Scholar] [CrossRef]
  16. Sedaghat, A.; Mohammadi, N. Uniform competency-based local feature extraction for remote sensing images. ISPRS J. Photogramm. Remote Sens. 2018, 135, 142–157. [Google Scholar] [CrossRef]
  17. Karpenko, S.; Konovalenko, I.; Miller, A.; Miller, B.; Nikolaev, D. UAV Control on the Basis of 3D Landmark Bearing-Only Observations. Sensors 2015, 15, 29802–29820. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Konovalenko, I.; Kuznetsova, E.; Miller, A.; Miller, B.; Popov, A.; Shepelev, D.; Stepanyan, K. New Approaches to the Integration of Navigation Systems for Autonomous Unmanned Vehicles (UAV). Sensors 2018, 18, 3010. [Google Scholar] [CrossRef] [Green Version]
  19. Lingua, A.; Marenchino, D.; Nex, F. Performance Analysis of the SIFT Operator for Automatic Feature Extraction and Matching in Photogrammetric Applications. Sensors 2009, 9, 3745–3766. [Google Scholar] [CrossRef] [PubMed]
  20. Lerma, J.L.; Navarro, S.; Cabrelles, M.; Seguí, A.; Hernandez, D. Automatic orientation and 3D modelling from markerless rock art imagery. ISPRS J. Photogramm. Remote Sens. 2013, 76, 64–75. [Google Scholar] [CrossRef]
  21. Safdari, M.; Moallem, P.; Satari, M. SIFT Detector Boosted by Adaptive Contrast Threshold to Improve Matching Robustness of Remote Sensing Panchromatic Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 675–684. [Google Scholar] [CrossRef]
  22. Wu, C. Towards linear-time incremental structure from motion. In Proceedings of the 2013 International Conference on 3D Vision, 3DV 2013, Seattle, WA, USA, 29 June–1 July 2013; pp. 127–134. [Google Scholar] [CrossRef] [Green Version]
  23. Zhao, J.; Zhang, X.; Gao, C.; Qiu, X.; Tian, Y.; Zhu, Y.; Cao, W. Rapid Mosaicking of Unmanned Aerial Vehicle (UAV) Images for Crop Growth Monitoring Using the SIFT Algorithm. Remote Sens. 2019, 11, 1226. [Google Scholar] [CrossRef] [Green Version]
  24. Hartmann, W.; Havlena, M.; Schindler, K. Predicting Matchability. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; Institute of Electrical and Electronics Engineers (IEEE): New York, NY, USA, 2014; pp. 9–16. [Google Scholar]
  25. Hartmann, W.; Havlena, M.; Schindler, K. Recent developments in large-scale tie-point matching. ISPRS J. Photogramm. Remote Sens. 2016, 115, 47–62. [Google Scholar] [CrossRef]
  26. Papadaki, A.I.; Hansch, R. Match or No Match: Keypoint Filtering based on Matching Probability. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; Institute of Electrical and Electronics Engineers (IEEE): New York, NY, USA, 2020; pp. 4371–4378. [Google Scholar]
  27. Royer, E.; Chazalon, J.; Rusinol, M.; Bouchara, F. Benchmarking Keypoint Filtering Approaches for Document Image Matching. In Proceedings of the 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR), Kyoto, Japan, 9–15 November 2017; Institute of Electrical and Electronics Engineers (IEEE): New York, NY, USA, 2017; Volume 1, pp. 343–348. [Google Scholar]
  28. Hossein-Nejad, Z.; Agahi, H.; Mahmoodzadeh, A. Image matching based on the adaptive redundant keypoint elimination method in the SIFT algorithm. Pattern Anal. Appl. 2020, 1–15. [Google Scholar] [CrossRef]
  29. Hossein-Nejad, Z.; Nasri, M. RKEM: Redundant Keypoint Elimination Method in Image Registration. IET Image Process. 2017, 11, 273–284. [Google Scholar] [CrossRef]
  30. Sedaghat, A.; Mokhtarzade, M.; Ebadi, H. Uniform Robust Scale-Invariant Feature Matching for Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2011, 49, 4516–4527. [Google Scholar] [CrossRef]
  31. Royer, E.; Lelore, T.; Bouchara, F. COnfusion REduction (CORE) algorithm for local descriptors, floating-point and binary cases. Comput. Vis. Image Underst. 2017, 158, 115–125. [Google Scholar] [CrossRef] [Green Version]
  32. Mukherjee, P.; Srivastava, S.; Lall, B. Salient keypoint selection for object representation. In Proceedings of the 2016 Twenty Second National Conference on Communication (NCC), Guwahati, India, 4–6 March 2016; Institute of Electrical and Electronics Engineers (IEEE): New York, NY, USA, 2016; pp. 1–6. [Google Scholar]
  33. Apollonio, F.; Ballabeni, A.; Gaiani, M.; Remondino, F. Evaluation of feature-based methods for automated network orientation. Int. Arch. Photogramm. Remote Sens. Spat. Inform. Sci. 2014, 40, 45. [Google Scholar]
  34. Buoncompagni, S.; Maio, D.; Maltoni, D.; Papi, S. Saliency-based keypoint selection for fast object detection and matching. Pattern Recognit. Lett. 2015, 62, 32–40. [Google Scholar] [CrossRef]
  35. Paul, S.; Pati, U.C. Remote Sensing Optical Image Registration Using Modified Uniform Robust SIFT. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1300–1304. [Google Scholar] [CrossRef]
  36. Xing, L.; Dai, W. A local feature extraction mcd on virtual line descriptors. Signal Image Video Process. 2020. [Google Scholar] [CrossRef]
  37. Dusmanu, M.; Ignacio, R.; Tomas, P.; Marc, P.; Josef, S.; Akihiko, T.; Torsten, S. D2-Net: A Trainable Cnn for Joint Detection and Description of Local Features. arXiv 2019, arXiv:1905.03561. [Google Scholar]
  38. Luo, Z.; Lei, Z.; Xuyang, B.; Hongkai, C.; Jiahui, Z.; Yao, Y.; Shiwei, L.; Tian, F.; Long, Q. Aslfeat: Learning Local Features of Accurate Shape and Localization. In Proceedings of the 2020 IEEE/CVF conference on computer vision and pattern recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6588–6597. [Google Scholar]
  39. Revaud, J.; Philippe, W.; César, D.S.; Noe, P.; Gabriela, C.; Yohann, C.; Martin, H. R2d2: Repeatable and Reliable Detector and Descriptor. arXiv 2019, arXiv:1906.06195. [Google Scholar]
  40. Schönberger, J.L.; Hardmeier, H.; Sattler, T.; Pollefeys, M. Comparative evaluation of hand-crafted and learned local features. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6959–6968. [Google Scholar]
  41. Matas, J.; Chum, O.; Urban, M.; Pajdla, T. Robust Wide-Baseline Stereo from Maximally Stable Extremal Regions. Image Vis. Comput. 2004, 22, 761–767. [Google Scholar] [CrossRef]
  42. Lopez-Garcia, F.; Fdez-Vidal, X.R.; Dosil, R. Scene Recognition through Visual Attention and Image Features: A Comparison between Sift and Surf Approaches. In Object Recognition; InTech: Rijeka, Croatia, 2011. [Google Scholar]
  43. Borji, A.; Sihite, D.N.; Itti, L. Quantitative Analysis of Human-Model Agreement in Visual Saliency Modeling: A Comparative Study. IEEE Trans. Image Process. 2013, 22, 55–69. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Kovesi, P.D. Matlab and Octave Functions for Computer Vision and Image Processing. Centre for Exploration Targeting, School of Earth and Environment, The University of Western Australia. Available online: https://www.peterkovesi.com/matlabfns (accessed on 23 March 2021).
  45. Vedaldi, A.; Fulkerson, B. Vlfeat: An Open and Portable Library of Computer Vision Algorithms. In Proceedings of the 18th ACM international conference on Multimedia, Firenze, Italy, 25–29 October 2010. [Google Scholar]
  46. Moré, J. The Levenberg-Marquardt algorithm: Implementation and theory. In Lecture Notes in Mathematicsl; Springer: Berlin/Heidelberg, Germany, 1978; Volume 630. [Google Scholar]
  47. Xiao, J. Sfmedu. Available online: http://vision.princeton.edu/courses/SFMedu/ (accessed on 23 March 2021).
  48. Applied Dataset. Available online: https://drive.google.com/drive/flers/1wAJ1HMv8nck0001wEQkf42OETClA-zvY?usp=sharing. (accessed on 23 March 2021).
Figure 1. The pipeline of image keypoint selection and matching for the synthetic dataset.
Figure 1. The pipeline of image keypoint selection and matching for the synthetic dataset.
Remotesensing 13 01302 g001
Figure 2. Pseudocode of the keypoint selection algorithm (Algorithm 1).
Figure 2. Pseudocode of the keypoint selection algorithm (Algorithm 1).
Remotesensing 13 01302 g002
Figure 3. The keypoint selection module in the Structure from Motion (SfM) pipeline.
Figure 3. The keypoint selection module in the Structure from Motion (SfM) pipeline.
Remotesensing 13 01302 g003
Figure 4. The employed synthetic image dataset with different transformations: (a) original image, (b)rotation, (c) scale, (d)viewpoint transformations.
Figure 4. The employed synthetic image dataset with different transformations: (a) original image, (b)rotation, (c) scale, (d)viewpoint transformations.
Remotesensing 13 01302 g004
Figure 5. The employed real datasets with their images and different camera networks. They feature high-resolution images, variable image overlap, and texture-less surfaces.
Figure 5. The employed real datasets with their images and different camera networks. They feature high-resolution images, variable image overlap, and texture-less surfaces.
Remotesensing 13 01302 g005
Figure 6. An example of the keypoint selection for an image pair for SIFT detector, (a) SIFT-entropy, (b) SIFT-saliency, (c) SIFT-texture-coefficient.
Figure 6. An example of the keypoint selection for an image pair for SIFT detector, (a) SIFT-entropy, (b) SIFT-saliency, (c) SIFT-texture-coefficient.
Remotesensing 13 01302 g006
Figure 7. Repeatability results for the synthetic image pair of S1 (left column) and S2 (right column). (a) Rotation, (b) scale, (c) viewpoint change.
Figure 7. Repeatability results for the synthetic image pair of S1 (left column) and S2 (right column). (a) Rotation, (b) scale, (c) viewpoint change.
Remotesensing 13 01302 g007
Figure 8. Precision results for the synthetic image pair of S1 (left column) and S2 (right column). (a) Rotation, (b) scale, (c) viewpoint change.
Figure 8. Precision results for the synthetic image pair of S1 (left column) and S2 (right column). (a) Rotation, (b) scale, (c) viewpoint change.
Remotesensing 13 01302 g008
Figure 9. Recall results for the synthetic image pair of S1 (left column) and S2 (right column). (a) Rotation, (b)scale, (c) viewpoint change.
Figure 9. Recall results for the synthetic image pair of S1 (left column) and S2 (right column). (a) Rotation, (b)scale, (c) viewpoint change.
Remotesensing 13 01302 g009
Figure 10. Global coverage results for the synthetic image pair of S1 (left column) and S2 (right column). (a) Rotation, (b)scale, (c) viewpoint change.
Figure 10. Global coverage results for the synthetic image pair of S1 (left column) and S2 (right column). (a) Rotation, (b)scale, (c) viewpoint change.
Remotesensing 13 01302 g010
Figure 11. Results for the real image block orientation. (a) Re-projection error of the bundle adjustment for each dataset, (b) average intersection angles, (c) average rays per computed 3D points, (d) the visibility of the derived 3D points in more than 3 images, (e) the number of 3D-points, (f) processing time.
Figure 11. Results for the real image block orientation. (a) Re-projection error of the bundle adjustment for each dataset, (b) average intersection angles, (c) average rays per computed 3D points, (d) the visibility of the derived 3D points in more than 3 images, (e) the number of 3D-points, (f) processing time.
Remotesensing 13 01302 g011
Figure 12. Comparison results of the proposed method. (a) Re-projection error of the bundle adjustment, (b) average intersection angles, (c) average rays per computed 3D points, (d) the visibility of the derived 3D points in more than 3 images, (e) the number of 3D-points, (f) processing time.
Figure 12. Comparison results of the proposed method. (a) Re-projection error of the bundle adjustment, (b) average intersection angles, (c) average rays per computed 3D points, (d) the visibility of the derived 3D points in more than 3 images, (e) the number of 3D-points, (f) processing time.
Remotesensing 13 01302 g012
Table 1. The parameters of detectors.
Table 1. The parameters of detectors.
AlgorithmParameter NameValue
SIFTNumber of octaves4
Number of layers3
Initial Gaussian smoothing1.6
SURFNumber of octaves3
Number of layers3
MSERMinimum region area30 pixels
Maximum area variation0.2%
The step size between the intensity threshold level1.2–5 *%
BRISKNumber of octaves4
Minimum intensity difference between a corner and its surrounding region0.28–0.41 *
Minimum accepted quality of corners0.28–0.41 *
* Minimum and maximum values applied according to the image scene type.
Table 2. Specifications of the real image datasets.
Table 2. Specifications of the real image datasets.
DatasetNo. of ImagesCamera ModelSensor Size (mm)Resolution (pixel)Pixel Size (µm)Nominal Focal Length (mm)
R18DJI FC631013.2 × 8.85472 × 36482.418.8
R218DJI FC631013.2 × 8.85472 × 36482.418.8
R315DJI FC652013 × 17.35280 × 39563.2812
Table 3. The matching results of MSER detector in synthetic images.
Table 3. The matching results of MSER detector in synthetic images.
RotationScaleViewpoint Change
MSERMSER-EntropyMSER-SaliencyMSER -Texture Coef.MSERMSER-EntropyMSER-SaliencyMSER-Texture Coef.MSERMSER-EntropyMSER-SaliencyMSER-
Texture Coef.
Repeatability (%)Mean83.5758.5362.3460.3547.1737.2841.9343.2939.3326.3232.3129.17
Std.6.3211.3914.088.0019.7419.8518.3018.202.032.443.353.68
Precision (%)Mean45.0973.4145.8360.0818.4445.2370.7458.8634.5876.4855.5953.84
Std. 11.0610.4623.3120.2614.4123.1812.6513.887.888.6213.6512.41
Recall (%)Mean19.8424.1817.0419.4099.8367.6495.5485.8419.8224.2917.7020.60
Std. 0.178.9511.3413.760.2117.555.507.470.245.828.026.21
Global coverage (%)Mean57.0729.4910.4720.6514.009.234.3510.9962.0131.6610.6921.23
Std. 5.0112.944.4017.453.432.684871.657.2215.967.439.17
RMSE (pixel)Mean0.720.710.750.721.391.271.4141.4130.700.740.560.64
Std. 0.310.270.430.290.040.280.00060.00070.070.130.220.12
Table 4. The matching results of SURF detector in synthetic images.
Table 4. The matching results of SURF detector in synthetic images.
RotationScaleViewpoint Change
SURFSURF -EntropySURF -SaliencySURF -Texture Coef.SURFSURF -EntropySURF -SaliencySURF -Texture Coef.SURFSURF -EntropySURF -SaliencySURF -
Texture Coef.
Repeatability (%)Mean50.3931.1739.0033.7786.5863.4672.0782.2613.589.1111.6911.74
Std.12.6513.7113.0716.392.695.244.832.022.041.332.352.79
Precision (%)Mean61.2381.5076.9968.4459.1966.0268.8577.4322.0248.7557.8543.96
Std. 16.729.0012.6817.5414.3411.2712.9711.877.479.289.9612.19
Recall (%)Mean45.4586.0878.9079.2599.5599.3399.9099.9525.3755.4261.8854.26
Std. 15.836.6211.399.390.390.800.120.077.6514.3910.027.82
Global coverage (%)Mean81.5680.3934.9562.8925.5621.4711.1021.6735.2744.8331.0139.62
Std. 3.982.011.3914.4315.6214.788.2714.4111.3910.314.3913.47
RMSE (pixel)Mean0.560.620.630.561.391.371.401.4100.640.720.710.71
Std. 0.130.140.130.100.010.0110.0130.0020.030.040.030.06
Table 5. The matching results of BRISK detector in synthetic images.
Table 5. The matching results of BRISK detector in synthetic images.
RotationScaleViewpoint Change
BRISKBRISK -EntropyBRISK -SaliencyBRISK -Texture Coef.BRISKBRISK -EntropyBRISK -SaliencyBRISK -Texture Oef.BRISKBRISK -EntropyBRISK -SaliencyBRISK -
Texture Coef.
Repeatability (%)Mean74.1250.5160.7756.7391.8975.8976.6781.7629.4225.3017.4325.23
Std.5.058.237.105.792.073.464.783.970.931.621.890.93
Precision (%)Mean72.0971.5378.2271.7771.7462.0454.4067.6651.0956.1162.6051.75
Std.11.969.008.0510.3910.5511.2615.7610.8610.748.787.189.40
Recall (%)Mean41.8750.9652.1651.5894.7390.0790.9793.6523.6337.7142.4335.77
Std.16.4912.7913.5114.021.704.364.843.046.239.6610.169.61
Global coverage (%)Mean71.9157.2430.8958.4824.2517.649.7518.0940.3632.7816.7331.53
Std.5.046.132.044.6414.289.906.9310.0811.498.275.709.78
RMSE (pixel)Mean0.400.460.450.441.3971.3801.3821.3920.440.500.500.48
Std.0.030.050.050.050.0020.0050.0070.0040.040.050.050.06
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mousavi, V.; Varshosaz, M.; Remondino, F. Using Information Content to Select Keypoints for UAV Image Matching. Remote Sens. 2021, 13, 1302. https://doi.org/10.3390/rs13071302

AMA Style

Mousavi V, Varshosaz M, Remondino F. Using Information Content to Select Keypoints for UAV Image Matching. Remote Sensing. 2021; 13(7):1302. https://doi.org/10.3390/rs13071302

Chicago/Turabian Style

Mousavi, Vahid, Masood Varshosaz, and Fabio Remondino. 2021. "Using Information Content to Select Keypoints for UAV Image Matching" Remote Sensing 13, no. 7: 1302. https://doi.org/10.3390/rs13071302

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop