Next Article in Journal
Error Resilient Coding Techniques for Video Delivery over Vehicular Networks
Previous Article in Journal
Ni-CNT Chemical Sensor for SF6 Decomposition Components Detection: A Combined Experimental and Theoretical Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved WαSH Feature Matching Based on 2D-DWT for Stereo Remote Sensing Images

1
NASG Key Laboratory of Land and Environment and Disaster Monitoring, China University of Mining and Technology, Xuzhou 221116, China
2
School of Environment Science and Spatial Informatics, China University of Mining and Technology, Xuzhou 221116, China
3
Jiangsu Key Laboratory of Resources and Environmental Information Engineering, China University of Mining and Technology, Xuzhou 221116, China
4
Wits Mining Institute, University of the Witwatersrand, Johannesburg 2050, South Africa
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(10), 3494; https://doi.org/10.3390/s18103494
Submission received: 16 July 2018 / Revised: 10 October 2018 / Accepted: 15 October 2018 / Published: 16 October 2018
(This article belongs to the Section Remote Sensors)

Abstract

:
Image matching is an outstanding issue because of the existing of geometric and radiometric distortion in stereo remote sensing images. Weighted α-shape (WαSH) local invariant features are tolerant to image rotation, scale change, affine deformation, illumination change, and blurring. However, since the number of WαSH features is small, it is difficult to get enough matches to estimate the satisfactory homography matrix or fundamental matrix. In addition, the WαSH detector is extremely sensitive to image noise because it is built on sampled edges. Considering the shortcomings of the WαSH detector, this paper improves the WαSH feature matching method based on the 2D discrete wavelet transform (2D-DWT). The method firstly performs 2D-DWT on the image, and then detects WαSH features on the transformed images. According to the methods of descriptor construction for WαSH features, three matching methods on the basis of wavelet transform WαSH features (WWF), improved wavelet transform WαSH features (IWWF), and layered IWWF (LIWWF) are distinguished with respect to the character of the sub-images. The experimental results on the dataset containing affine distortion, scale distortion, illumination change, and noise images, showed that the proposed methods acquired more matches and better stableness than WαSH. Experimentation on remote sensing images with less affine distortion and slight noise showed that the proposed methods obtained the correct matching rate greater than 90%. For images containing severe distortion, KAZE obtained a 35.71% correct matching rate, which is unacceptable for calculating the homography matrix, while IWWF achieved a 71.42% correct matching rate. IWWF was the only method that achieved the correct matching rate of no less than 50% for all four test stereo remote sensing image pairs and was the most stable compared to MSER, DWT-MSER, WαSH, DWT-WαSH, KAZE, WWF, and LIWWF.

1. Introduction

Image feature matching, which aims to acquire reliable homonymous features—matches between images—is one of the most basic and important processes of remote sensing image processing [1,2]. The accuracy of matching results directly affects the reliability of image registration [3] and fusion, automatic moving objects detection [4], and aerial triangulation and 3D reconstruction [5].
Feature matching includes three steps: feature detecting, feature describing, and similarity matching. At present, the most widely used feature matching algorithm is the Scale Invariant Feature Transform (SIFT) algorithm [6] which firstly uses the difference-of-Gaussian (DOG) operator to identify potential interest features, followed by features localization and orientation assignment, and then describes the features according to local image gradients, finally performing the similarity matching to acquire matches. The SIFT algorithm is adaptable to distortions, such as scale, rotation, translation, etc. Therefore, it is used in most matching processes of aerial or aerospace images. Yu et al. [1] employed the SIFT algorithm to create a set of matches and then removed outliers from the matches and the estimated the transformation relationship between unmanned aerial vehicle (UAV) images by using a maximum likelihood framework. It achieved the feature matching of UAV images under non-rigid transformation in precision agriculture. Chang et al. [7] combined the SIFT algorithm with scale constraints and geometric constraints to achieve highly-accurate registration of high-resolution satellite image. Xiang et al. [8] used a SIFT-liked matching algorithm to extract reliable matches for GF-3 SAR images and then provided a coarse-to-fine registration algorithm, giving a robust, efficient and precise performance of SAR image. Qi and Zhu [9] decomposed remote sensing images by discrete wavelet transform (DWT), and extracted the SIFT features on low-frequency sub-images, then used a coarse-to-fine strategy to improve the precision and efficiency of SIFT.
However, because of the effect of affine distortion as a result of the camera view change at the time of image acquisition, the matching performance of SIFT algorithm decreases sharply result in that reliable potential matches cannot be obtained [10,11]. In this case, affine-invariant features, such as Harris-Affine [12], maximally stable extremal region (MSER) [13], and edge-based region (EBR) [14,15] are required. Mikolajczyk et al. [16] compared six common affine-invariant feature detectors in which the results showed that MSER performs best in most cases and does well on images containing homogeneous regions with distinctive boundaries. Based on the MSER detector, Zhang et al. [17] proposed an automatic coarse-to-fine image registration strategy that achieves high accuracy for images containing affine distortion. Sedaghat and Ebadi [18] extracted the MSER and Harris-Affine features and selected the initial matches using the Euclidean distance between feature descriptors, followed by a consistency check process. Then oriented least square matching is developed to improve the positional accuracy of the affine-invariant features. Their affine-invariant image matching method was successfully applied for various synthetic and real close range and satellite images matching. Yu et al. [19] acquired the initial matches based on the MSER detector and calculated the affine transformations between the initial matches. A small number of high-accuracy matches were detected by combining the SIFT feature matching under the constraint of affine transformation and the neighborhood supporting principle. Then the propagation is performed based on the global affine transformation constraints to find more matches satisfying the accuracy registration of oblique images.
The weighted α-shape (WαSH) [20,21] algorithm has certain stability to the geometric and radiometric distortion of images and can detect invariant local features with high uniqueness. Compared with Hessian-Affine [12], MSER, medial feature detector (MFD) [22], edge foci (EF) [23], and KAZE algorithms [24], Varytimidis et al. [20] found that the WαSH feature detection algorithm performs among the best in all cases of image distortion through a comprehensive analysis of factors, such as repeatability, matching score, and number of matches.
However, there also are some shortcomings of WαSH feature matching: (1) When there are large geometric distortions between images, including affine changes and scale changes, WαSH feature matching detects a few matches that contain a significant proportion of false matches and causes failure in calculating the transformation model between image pairs; (2) Since the WαSH features are extracted based on the image edge information, the repeatability and matching score will be greatly decreased if there is significant noise in images. In order to deal with the problems above, we improve the WαSH feature matching method based on 2D-DWT. This method firstly performs 2D-DWT decomposition on the original image and obtains four sub-images: LL, LH, HL, and HH. We discard the HH layer in which the noise in the original image is concentrated and transform the remaining sub-images to the original image size. WαSH detector is then performed on the transformed images. Depending on the descriptor construction methods for the WαSH features, three matching methods on the basis of wavelet transform WαSH feature (WWF), improved wavelet transform WαSH feature (IWWF), and layered IWWF (LIWWF), respectively, are distinguished. The matches are finally determined by using similarity matching. The procedure of improved WαSH feature matching based on 2D-DWT is illustrated in Figure 1.

2. 2D-DWT

DWT of an image transforms the image data from the spatial domain to frequency domain by using hierarchical decomposing functions [25]. A 2D decomposition is achieved by sequentially applying 1D-DWT along horizontal and vertical directions, respectively [26]. In a 1D decomposition, there are both low-(L) and high-pass (H) filters along each direction, which leads to one “L” and one “H” sub-band, respectively. The “L” sub-band corresponds to low-frequency components in the wavelet domain, while the “H” sub-band to high-frequency ones. After doing the first level 2D decomposition, there are four sub-bands in the wavelet domain, which are labeled as LL, LH, HL, and HH, respectively.
The LL sub-band is generated by convolving the low-pass wavelet filter along horizontal and vertical directions on the image. It is an approximate representation of the original image. The LH sub-band is generated by convolving a low-pass wavelet filter in the horizontal direction and then convolving a high-pass wavelet filter in the vertical direction on the image. It represents a character in the vertical direction of the original image. The HL sub-band is generated by convolving a low-pass wavelet filter in the horizontal direction after a convolution of high-pass wavelet filter in the vertical direction on the image. It represents the horizontal characteristic of the original image. The HH sub-band is a wavelet coefficient generated by convolving high-pass wavelet filters in two directions on the image. It represents the diagonal edge characteristics of the image. Therefore, the LL sub-band contains most features of the original image, while the noise in the original image is concentrated in the HH sub-band.

3. WαSH Feature Detector

The WαSH detector exploits stable and distinctive dominant structural features in an image on the basis of image edge and α shape. It can be divided into four parts, namely edge detecting, edge sampling, WαSH constructing and feature extracting. The processes are as follows:
(1)
Edge detecting: g is the normalized gradient image of the input image in [ 0 , 1 ] . The binary edge image e is acquired by applying the Canny detector on g .
(2)
Edge sampling: With a fixed sampling interval s , e is sampled uniformly along edges to obtain a discrete set of edge points P R 2 . For each point p P , the weight w ( p ) is defined to be multiple of its gradient strength:
w ( p ) = g ( p ) ( s 2 ) 2 ,
where g ( p ) [ 0 , 1 ] is the value of g at p .
(3)
WαSH construction: Regular triangulation of P is calculated. The line segments and triangles of triangulation are added into a collection K and are ordered by descending size, followed by the weighted α-shape constructing. For more details, please see [19].
(4)
Feature extracting: the neighbors of each triangle σ T K are its three edges, while the neighbors of each line segments σ E K are the two adjacent triangles in the triangulation. Since the size of an edge is not larger than that of its two adjacent triangles, and α is decreasing, the intuition is that this edge can keep the two triangles disconnected until it is processed itself. Based on this neighborhood system, each connected weighted α complex is called a component. Considering that each element (line segment or triangle) of K in descending order of size is an independent component, the components are joined with their neighbors that have already been processed. Decreasing the value of α continuously, the strength of a component k U is calculated following Equation (2) when it is joined to another component through its boundary element:
s ( k U ) = a ( k U ) ρ T ,
where a ( k U ) is the total area of k U (the area of the line segment is 0); ρ T is the size of the boundary element. If strength is greater than a fixed threshold τ , the component is determined to be a WαSH feature which is consequently fitted to an ellipse. We assume the image coordinate of the ellipse center is ( x , y ) , and the ellipse parameter is ( a , b , c ) , then the ellipse equation is a ( x u ) ( x u ) + 2 b ( x u ) ( y v ) + c ( y v ) ( y v ) = 1 . The WαSH feature is denoted as x = ( x , y , a , b , c ) T in this paper.

4. Algorithm

4.1. Feature Set Construction

Wavelet transform (the Haar wavelet base is used in this paper) is firstly performed on the reference image and the image to be matched to obtain four sub-bands—LL, LH, HL, and HH. Since the sub-bands are two-dimensional, we consider them as four images of the LL, LH, HL, and HH layer. The LL layer image retains the original image content information, the LH layer, and the HL layer, respectively, maintain the vertical and horizontal information on the original image after wavelet transform. They are favorable for edge information and WαSH local features detecting. The noise in the original image is concentrated in the HH layer. Therefore, we perform WαSH detector on the images of LL layer, LH layer, and HL layer while discarding the HH layer. In order to avoid inconsistent coordinates, we transform the sizes of three-layer images to be equal to the original image. We perform the WαSH detection on the images of the LL layer, LH layer, and HL layer, and obtain the WαSH feature sets S LL = { x i LL | i = 0 , 1 , , n LL } , S LH = { x i LH | i = 0 , 1 , , n LH } , S HL = { x i HL | i = 0 , 1 , , n HL } , respectively, as displayed in Figure 2.
According to the normalization theory in [16], the WαSH feature, an elliptical region, can be normalized into a circular one, on which the description of the feature is more robust to affine distortion.
A variety of feature describing operators were compared in [27], reaching the conclusion that SIFT-based description operators have better robustness than others. Therefore, we use the SIFT algorithm to calculate the descriptors for WαSH features.
Three feature sets—the wavelet transform WαSH feature (WWF) set, the improved wavelet transform WαSH feature (IWWF) set, and the layered IWWF (LIWWF) set are constructed based on the above-mentioned WαSH local invariant features detected on the images of LL layer, LH layer, and HL layer in this paper. As the features consist of coordinates on the image x and the description vectors d , we then denote the features using the representation of f = ( x , d ) . In the following context, we use a superscript to distinguish different feature sets constructing method and a subscript to distinguish the original image for feature sets M containing descriptors.

4.1.1. WWF Set Constructing

Set S LL , S LH , and S HL are combined into set S W = { x i W | i = 0 , 1 , , n W } , where n W = n LL + n LH + n HL is the number of features in WWF. The description vector of feature x i W is calculated by using SIFT algorithm on original image. Then the set of WWF is acquired and denoted as M W = { f i W | i = 0 , 1 , , n W } , where f i W = ( x i W , d i W ) . The process is demonstrated in Figure 3a.

4.1.2. IWWF Set Constructing

For x i LL , the features on the image of LL layer, the description vector is calculated on the image of LL layer, while the description vectors of x i LH and x i HL are calculated on the original image due to the lack of image information on LH and HL layers. We acquired three sets M LL = { f i LL | i = 0 , 1 , , n LL } , M LH = { f i LH | i = 0 , 1 , , n LH } and M HL = { f i HL | i = 0 , 1 , , n HL } , where f i LL = ( x i LL , d i LL ) , f i LH = ( x i LH , d i LH ) and f i HL = ( x i HL , d i HL ) respectively. Merging sets M LL , M LH , and M HL , the set of IWWF is obtained and denoted as M I = { f i I | i = 0 , 1 , , n I } , where n I = n LL + n LH + n HL is the number of features in IWWF. The process is demonstrated in Figure 3b.

4.1.3. LIWWF Set Constructing

Similar to IWWF set constructing method, the description vectors of features in S LL are calculated on the image of LL layer, the descriptor vectors of features in S LH and S HL are calculated on the original image. The difference with IWWF set constructing method is that a variable F is introduced to distinguish the layer to which the features belong in LIWWF. Therefore, we obtain sets M LL = { ( f i LL , F ) | i = 0 , 1 , , n LL , F = 1 } , M LH = { ( f i LH , F ) | i = 0 , 1 , , n LH , F = 2 } , and M HL = { ( f i HL , F ) | i = 0 , 1 , , n HL , F = 3 } , respectively. Merging sets M LL , M LH , and M HL , the set of LIWWF is obtained and denoted as M L = { ( f i L , F ) | i = 0 , 1 , , n L , F = 1 , 2 , 3 } , where n L = n LL + n LH + n HL is the number of features in LIWWF. The process is demonstrated in Figure 3c.

4.2. Similarly Matching

The WWF sets M 1 W and M 2 W are constructed separately for the reference image I 1 and the image to be matched I 2 , and the number of features are n 1 W and n 2 W . For each element in M 1 W , the Euclidean distance between the feature description vector d i and all of the feature description vector d j in M 2 W are calculated. If the nearest distance to nearest neighbor distance ratio (NNDR) between the d i and d j is less than 0.8, the corresponding features with the minimum distance in M 1 W and M 2 W are determined as a match. For sets IWWF of I 1 and I 2 , the matches are determined in the same process with that in WWF. For sets LIWWF of I 1 and I 2 , we first compare the values of F . If the values of F are equal, the Euclidean distance is calculated of the two description vectors in M 1 W and M 2 W . If not, we set the distance to infinity and skip the distance calculation. Finally, the corresponding features are determined under the same criteria as that of WWF.

5. Experiments and Analysis

5.1. Performance Compared to WαSH for Different Distortion

Six representative image sequences in the Krystian Mikolajczyk’s personal homepage [28] with affine distortion, scale distortion, and illumination change, and a group of UAV image pairs with different levels of noise are employed to test the performance of WαSH, WWF, IWWF, and LIWWF. The test images are shown in Figure 4 in which each sequence contains six images.
The number of total matches (Nt), the number of correct matches (Nc) and correct matching rate (Cr) are employed to evaluate the performances of the four methods. The matching results are shown in Figure 5, of which the 1, 2, 3, 4, 5 of the horizontal axis are different image pairs representing image 2 to 1, 3 to 1, 4 to 1, 5 to 1, and 6 to 1, respectively. Cr is calculated as the ratio percentage of Nc to Nt. According to [16], a pair of matches was determined to be the correct one when the overlap error is less than 50%.
According to the above experimental results, the analysis is as follow:
From Figure 5, we can find that the similar orders and changes occurred in the plots of the number of total matches and correct matches. WWF, IWWF, and LIWWF offered higher Nt and Cr than WαSH for most image pairs. That is because the improved methods use features on images of LL layer, LH layer, and HL layer which contain obvious image structure characteristics. Combining the features on all of three sub-images, WWF, IWWF, and LIWWF achieved more matches than the original WαSH method.
Take the 5th image pair of Leuven in Figure 5 for example. The results of WαSH, WWF, IWWF, and LIWWF for this image pair are shown in Figure 6, and the repeatability and correct matching rate are displayed in Table 1. The repeatability measures how large proportion of the detected features is the corresponding scene region. The correct matching rate assesses the distinctiveness of the detected features.
Table 1 shows the WαSH features on LL and LH layer had a relative high repeatability. The reason is that the image of LL and LH layer contains obvious image structure characteristics. In addition, the matching accuracy of features described on LL layer was higher than the matching accuracy of features described on the original image. That means the image context of the LL layer is more distinctive than the image context of the original image in this case. Therefore, the Cr scores of IWWF and LIWWF are higher than the Cr scores of WαSH and WWF.

5.2. Application for Stereo Remote Sensing Images Matching

In order to perform a comprehensive analysis of the proposed methods, we selected four pairs of stereo remote sensing images obtained by UAV to the experimental analysis. Seeing in Figure 7, these UAV images contain different distortion including affine distortion, rotation, blur, and noise. The MSER+SIFT matching algorithm (abbreviated as MSER), DWT-MSER, which applies DWT to smooth images before MSER detecting, the WαSH + SIFT matching algorithm (abbreviated as WαSH), DWT-WαSH, and KAZE [24] are compared in this experiment. KAZE is a multiscale feature detection and description algorithm in nonlinear scale spaces [24]. It is similar to SIFT, but more robust than SIFT. In the MSER/WαSH method, the MSER/WαSH detector is applied to extract local features, which are consequently described by the SIFT-describing algorithm. All of the comparison methods determine matches by using NNDR. The threshold value is set to 0.8, which is equal to that of the proposed method. The matches of the proposed methods are depicted in Figure 7.
The matching results are presented in Table 2. Nt is the number of total matches; Nc is the number of correct matches which have overlap error lower than 50%; Cr is the correct matching rate that is equal to the percentage of Nc and Nt.
Based on the results displayed in Table 1, the analysis is as follow:
Comparing Nt of MSER and DWT-MSER, we can find that the number of total matches decreased when DWT was paired to MSER. Nevertheless, when DWT was paired to WαSH, it increased. The reason is that MSER detected much fewer features on the smoothed images than on the original images, while WαSH detected a similar number due to edge detecting process. The values of Nt and Nc of WWF, IWWF, and LIWWF were improved relative to WαSH, which is consistent with the result in Section 5.1. This is because WWF, IWWF, and LIWWF detect WαSH features on three images, which increase the number of features. Among the WαSH-based methods (i.e., WαSH, DWT-WαSH, WWF, IWWF, and LIWWF), LIWWF achieved the largest number of total matches and correct matches, followed by IWWF. Cr values of IWWF and LIWWF were relatively stable. For images in Figure 7b, which contain simple texture and large rotation distortion, the correct matching rate of MSER, DWT-MSER, WαSH, DWT-WαSH, and WWF were less than 50% with a low Nt that is difficult to estimate the homography matrix between image pairs, while IWWF and LIWWF obtained relatively reliable matches. For images in Figure 7c containing affine distortion and slight noise, Cr values of KAZE and proposed methods were all above 90%, and the correct matching rate of WWF reached to 100%. KAZE obtained the largest value of Nt than other methods. However, KAZE obtained a 35.71% correct matching rate, which is extremely low, in Figure 7a that may cause failure in calculating the homography matrix between the image pairs. IWWF was the only method that achieved the correct matching rate no less than 50% for all of the test images. Therefore, IWWF is the most stable to different kinds of distortion among these comparative methods.

6. Conclusions

The WαSH local invariant feature has advantages in distinctiveness and robustness. However, image matching based on WαSH features performs a small number of matches and noise sensitivity. Focusing on addressing these problems, this paper improved WαSH based on 2D-DWT and proposed three methods of WWF, IWWF, and LIWWF for image matching. Experiments on sequences of affine distortion, scale distortion, illumination change images, and a sequence of simulated noise images showed that the WWF, IWWF, and LIWWF acquired more matches and better performance than WαSH. For remote sensing images with less affine distortion and slight noise, the correct matching rate of KAZE and proposed methods were higher than MSER and WαSH, and the correct matching rate of WWF reached to 100%. Nevertheless, for images containing severe distortion, KAZE obtained extremely low correct matching rate which is unacceptable for the homography matrix calculating, while IWWF achieved 71.42% correct matching rate. IWWF was the only method that achieved the correct matching rate no less than 50% for all of the test images and was the most stable comparing with MSER, DWT-MSER, WαSH, DWT-WαSH, KAZE, WWF, and LIWWF.
The proposed methods are 2–3 times slower than WαSH because they detected WαSH features on three sub-images. In addition, the normalizing and resampling processes make the proposed methods moderate increase in computational cost. The proposed methods improved the number of matches and the stability at the expense of computational efficiency. We are going to focus on speeding up the proposed method in future work.

Author Contributions

Conceptualization, M.Y.; Funding acquisition, K.D.; Methodology, M.Y.; Resources, H.Y.; Supervision, K.D.; Validation, H.Y.; Writing—original draft, M.Y.; Writing—review & editing, C.Q.

Funding

National Key Research and Development Program of China: 2017YFE0107100, Natural Science Foundation of China: 51774270, 41604005, 41371438, Enterprise Commissioned Project: D010102

Acknowledgments

This work was supported by the National Key Research and Development Program of China (No. 2017YFE0107100), the Natural Science Foundation of China (no. 51774270, no. 41604005, and no. 41371438), and the Enterprise Commissioned Project (no. D010102). The authors would like to thank authors who have provided their algorithm WαSH as the free and open-source resource, which was very helpful for the research in this paper.

Conflicts of Interest

The authors declare no conflict of interest

References

  1. Yu, Z.; Zhou, H.; Li, C. Fast non-rigid image feature matching for agricultural UAV via probabilistic inference with regularization techniques. Comput. Electron. Agric. 2017, 143, 79–89. [Google Scholar] [CrossRef]
  2. Sedaghat, A.; Mokhtarzade, M.; Ebadi, H. Uniform Robust Scale-Invariant Feature Matching for Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2011, 49, 4516–4527. [Google Scholar] [CrossRef]
  3. Kahaki, S.M.M.; Arshad, H.; Nordin, M.J.; Ismail, W. Geometric feature descriptor and dissimilarity-based registration of remotely sensed imagery. PLoS ONE 2018, 13, e200676. [Google Scholar] [CrossRef] [PubMed]
  4. Kalantar, B.; Mansor, S.B.; Abdul Halin, A.; Shafri, H.Z.M.; Zand, M. Multiple Moving Object Detection from UAV Videos Using Trajectories of Matched Regional Adjacency Graphs. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5198–5213. [Google Scholar] [CrossRef]
  5. Nex, F.; Remondino, F. UAV for 3D mapping applications: A review. Appl. Geomat. 2014, 6, 1–15. [Google Scholar] [CrossRef]
  6. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef] [Green Version]
  7. Chang, X.; Du, S.; Li, Y.; Fang, S. A Coarse-to-Fine Geometric Scale-Invariant Feature Transform for Large Size High Resolution Satellite Image Registration. Sensors 2018, 18, 1360. [Google Scholar] [CrossRef] [PubMed]
  8. Xiang, Y.; Wang, F.; You, H. An Automatic and Novel SAR Image Registration Algorithm: A Case Study of the Chinese GF-3 Satellite. Sensors 2018, 18, 672. [Google Scholar] [CrossRef] [PubMed]
  9. Youjie, Q.I.; Zhu, E. A New Fast Matching Algorithm by Trans-Scale Search for Remote Sensing Image. Chin. J. Electron. 2015, 24, 654–660. [Google Scholar]
  10. Jiang, S.; Jiang, W. Efficient structure from motion for oblique UAV images based on maximal spanning tree expansion. ISPRS J. Photogramm. 2017, 132, 140–161. [Google Scholar] [CrossRef]
  11. Morel, J.; Yu, G. ASIFT: A New Framework for Fully Affine Invariant Image Comparison. SIAM J. Imaging Sci. 2009, 2, 438–469. [Google Scholar] [CrossRef] [Green Version]
  12. Mikolajczyk, K.; Schmid, C. Scale & Affine Invariant Interest Point Detectors. Int. J. Comput. Vis. 2004, 60, 63–86. [Google Scholar] [CrossRef]
  13. Matas, J.; Chum, O.; Urban, M.; Pajdla, T. Robust Wide-baseline Stereo from Maximally Stable Extremal Regions. Image Vis. Comput. 2004, 22, 761–767. [Google Scholar] [CrossRef]
  14. Tuytelaars, T.; Van Gool, L. Matching Widely Separated Views Based on Affine Invariant Regions. Int. J. Comput. Vis. 2004, 51, 61–85. [Google Scholar] [CrossRef]
  15. Tuytelaars, T.; Van Gool, L. Content-based image retrieval based on local affinely invariant regions. In Proceedings of the International Conference on Visual Information and Information Systems, Amsterdam, The Netherlands, 2–4 June 1999; Volume 1614, pp. 493–500. [Google Scholar]
  16. Mikolajczyk, K.; Tuytelaars, T.; Schmid, C.; Zisserman, A.; Matas, J.; Schaffalitzky, F.; Kadir, T.; Gool, L.V. A Comparison of Affine Region Detectors. Int. J. Comput. Vis. 2005, 65, 43–72. [Google Scholar] [CrossRef] [Green Version]
  17. Zhang, Q.; Wang, Y.; Wang, L. Registration of Images with Affine Geometric Distortion Based on Maximally Stable Extremal Regions and Phase Congruency. Image Vis. Comput. 2015, 36, 23–39. [Google Scholar] [CrossRef]
  18. Sedaghat, A.; Ebadi, H. Accurate Affine Invariant Image Matching Using Oriented Least Square. Photogramm. Eng. Remote Sens. 2015, 81, 733–743. [Google Scholar] [CrossRef]
  19. Yu, M.; Yang, H.; Deng, K.; Yuan, K. Registrating oblique images by integrating affine and scale-invariant features. Int. J. Remote Sens. 2018, 39, 3386. [Google Scholar] [CrossRef]
  20. Varytimidis, C.; Rapantzikos, K.; Avrithis, Y.; Kollias, S. α-shapes for local feature detection. Pattern Recogn. 2016, 50, 56–73. [Google Scholar] [CrossRef] [Green Version]
  21. Varytimidis, C.; Rapantzikos, K.; Avrithis, Y. WαSH: Weighted α-Shapes for Local Feature Detection. In Proceedings of the ECCV’12 12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; pp. 788–801. [Google Scholar]
  22. Avrithis, Y.; Rapantzikos, K. The Medial Feature Detector: Stable Regions from Image Boundaries. In Proceedings of the IEEE International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; Volume 24, pp. 1724–1731. [Google Scholar]
  23. Zitnick, C.L.; Ramnath, K. Edge foci interest points. In Proceedings of the International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 359–366. [Google Scholar]
  24. Alcantarilla, P.F.; Bartoli, A.; Davison, A.J. KAZE features. In Proceedings of the European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; Volume 7577, pp. 214–227. [Google Scholar]
  25. Hussein, A.J.; Hu, F.; He, F. Multisensor of thermal and visual images to detect concealed weapon using harmony search image fusion approach. Pattern Recogn. Lett. 2017, 94, 219–227. [Google Scholar] [CrossRef]
  26. Luo, X.; Bhakta, T. Estimating observation error covariance matrix of seismic data from a perspective of image denoising. Comput. Geosci. 2017, 21, 205–222. [Google Scholar] [CrossRef]
  27. Mikolajczyk, K.; Schmid, C. A Performance Evaluation of Local Descriptors. IEEE Trans. Pattern Anal. 2005, 27, 1615–1630. [Google Scholar] [CrossRef] [PubMed]
  28. Krystian Mikolajczyk Personal Homepage. Available online: http://lear.inrialpes.fr/people/mikolajczyk/ (accessed on 15 January 2018).
Figure 1. Procedure of improved WαSH feature matching based on 2D-DWT. I1 and I2 are the reference image and the image to be matched, respectively.
Figure 1. Procedure of improved WαSH feature matching based on 2D-DWT. I1 and I2 are the reference image and the image to be matched, respectively.
Sensors 18 03494 g001
Figure 2. WαSH features on (a) original image, (b) LL layer, (c) LH layer, and (d) HL layer.
Figure 2. WαSH features on (a) original image, (b) LL layer, (c) LH layer, and (d) HL layer.
Sensors 18 03494 g002
Figure 3. Feature set constructing process of (a) WWF, (b) IWWF, and (c) LIWWF.
Figure 3. Feature set constructing process of (a) WWF, (b) IWWF, and (c) LIWWF.
Sensors 18 03494 g003
Figure 4. Test image sequences with affine distortion, scale distortion, illumination change and noise change in images. (a) graffti: the image affine distortion increasing from left to right; (b) wall: the image affine distortion increasing from left to right; (c) bikes: the image scale distortion increasing from left to right; (d) mosaic: the image illumination increasing from left to right; (e) Leuven: the image illumination decreasing from left to right; (f) library: the image noise increasing from left to right.
Figure 4. Test image sequences with affine distortion, scale distortion, illumination change and noise change in images. (a) graffti: the image affine distortion increasing from left to right; (b) wall: the image affine distortion increasing from left to right; (c) bikes: the image scale distortion increasing from left to right; (d) mosaic: the image illumination increasing from left to right; (e) Leuven: the image illumination decreasing from left to right; (f) library: the image noise increasing from left to right.
Sensors 18 03494 g004
Figure 5. Results of different matching methods for the test data. Nt: number of total matches; Nc: the number of correct matches; and Cr: correct matching rate (%).
Figure 5. Results of different matching methods for the test data. Nt: number of total matches; Nc: the number of correct matches; and Cr: correct matching rate (%).
Sensors 18 03494 g005aSensors 18 03494 g005b
Figure 6. Results of (a) WαSH, (b) WWF, (c) IWWF, and (d) LIWWF for the 5th image pair of Leuven. The red lines indicate correct matches while the blue lines are incorrect matches.
Figure 6. Results of (a) WαSH, (b) WWF, (c) IWWF, and (d) LIWWF for the 5th image pair of Leuven. The red lines indicate correct matches while the blue lines are incorrect matches.
Sensors 18 03494 g006
Figure 7. Results of proposed methods for remote sensing images obtained by UAV. The red lines indicate correct matches while the blue lines are incorrect matches. (a-1), (b-1), (c-1) and (d-1) display result of WWF; (a-2), (b-2), (c-2) and (d-2) display result of IWWF; (a-3), (b-3), (c-3) and (d-3) display result of LIWWF.
Figure 7. Results of proposed methods for remote sensing images obtained by UAV. The red lines indicate correct matches while the blue lines are incorrect matches. (a-1), (b-1), (c-1) and (d-1) display result of WWF; (a-2), (b-2), (c-2) and (d-2) display result of IWWF; (a-3), (b-3), (c-3) and (d-3) display result of LIWWF.
Sensors 18 03494 g007
Table 1. Repeatability and correct matching rate of WαSH features on different image layer of the 5th image pair of Leuven.
Table 1. Repeatability and correct matching rate of WαSH features on different image layer of the 5th image pair of Leuven.
WαSH Features onRepeatabilityCr
Original image53.719%67.68% (described on original image)
LL layer48.7395%79% (described on original image)
88.36% (described on LL layer)
LH layer52.5424%62.5% (described on LH layer)
HL layer38.2022%84.62% (described on HL layer)
Table 2. Results of matching for stereo remote sensing image pairs.
Table 2. Results of matching for stereo remote sensing image pairs.
MSERDWT-MSERWαSHDWT-WαSHKAZEWWFIWWFLIWWF
Figure 7aNt9171028151418
Nc40661071010
Cr44.440.0085.7160.0035.7146.6771.4355.56
Figure 7bNt34111316206212228
Nc736711151114
Cr20.5927.2746.1543.7553.8823.8150.0050.00
Figure 7cNt665515341236183440
Nc463813271114183336
Cr69.7069.0988.6779.4190.13100.0097.0690.00
Figure 7dNt1991430199283642
Nc8471282131920
Cr42.1144.4450.0040.0046.2346.4352.7847.62

Share and Cite

MDPI and ACS Style

Yu, M.; Deng, K.; Yang, H.; Qin, C. Improved WαSH Feature Matching Based on 2D-DWT for Stereo Remote Sensing Images. Sensors 2018, 18, 3494. https://doi.org/10.3390/s18103494

AMA Style

Yu M, Deng K, Yang H, Qin C. Improved WαSH Feature Matching Based on 2D-DWT for Stereo Remote Sensing Images. Sensors. 2018; 18(10):3494. https://doi.org/10.3390/s18103494

Chicago/Turabian Style

Yu, Mei, Kazhong Deng, Huachao Yang, and Changbiao Qin. 2018. "Improved WαSH Feature Matching Based on 2D-DWT for Stereo Remote Sensing Images" Sensors 18, no. 10: 3494. https://doi.org/10.3390/s18103494

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop