Next Article in Journal
Preserving Smart Objects Privacy through Anonymous and Accountable Access Control for a M2M-Enabled Internet of Things
Previous Article in Journal
Spectral and Image Integrated Analysis of Hyperspectral Data for Waxy Corn Seed Variety Classification
 
 
Addendum published on 30 November 2015, see Sensors 2015, 15(12), 29923.
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Assigning Main Orientation to an EOH Descriptor on Multispectral Images

1
Beijing University of Posts and Teles., School of Electronic Engineering, Rd. Xitucheng 10#, Beijing 100876, China
2
Beijing Anzhen Hospital, Capital Medical University and Beijing Institute of Heart Lung and BloodVessel Disease, Beijing 100029, China
*
Author to whom correspondence should be addressed.
Sensors 2015, 15(7), 15595-15610; https://doi.org/10.3390/s150715595
Submission received: 8 May 2015 / Revised: 16 June 2015 / Accepted: 23 June 2015 / Published: 1 July 2015
(This article belongs to the Section Remote Sensors)

Abstract

: This paper proposes an approach to compute an EOH (edge-oriented histogram) descriptor with main orientation. EOH has a better matching ability than SIFT (scale-invariant feature transform) on multispectral images, but does not assign a main orientation to keypoints. Alternatively, it tends to assign the same main orientation to every keypoint, e.g., zero degrees. This limits EOH to matching keypoints between images of translation misalignment only. Observing this limitation, we propose assigning to keypoints the main orientation that is computed with PIIFD (partial intensity invariant feature descriptor). In the proposed method, SIFT keypoints are detected from images as the extrema of difference of Gaussians, and every keypoint is assigned to the main orientation computed with PIIFD. Then, EOH is computed for every keypoint with respect to its main orientation. In addition, an implementation variant is proposed for fast computation of the EOH descriptor. Experimental results show that the proposed approach performs more robustly than the original EOH on image pairs that have a rotation misalignment.

1. Introduction

Keypoint and descriptor techniques have been widely applied in computer vision or pattern recognition. Applications include stereo vision, 3D scene reconstruction, human activity recognition, etc. Keypoints are often matched by computing the distance of their associated descriptors. The matching ability of descriptors is measured with the repeatability and distinctiveness, and in practice, a trade-off is often made between them. On single spectral images, SIFT [1] and its variants with post-processing techniques (e.g., RANSAC) have witnessed many successful applications. On multi-sensor (multispectral) images, SIFT descriptors generate few correct mappings. Recently, the edge-oriented histogram (EOH) [2] was proposed, which utilizes only edge points and five bins for computing descriptors. EOH has a better matching performance on multispectral images than SIFT, but does not assign a main orientation to keypoints, which limits its application to images containing translation misalignment.

1.1. Related Work

Salient points have been widely used in a variety of fields, including object tracking, image fusion, intelligent navigation, etc. [39]. Many keypoint and descriptor detection techniques have been proposed for single spectral images. Lowe [1] proposed SIFT detecting keypoints invariant to scale and rotation. The keypoints are defined to be the extrema of the difference of Gaussians (DOG). The local gradient pattern around a keypoint with respect to an assigned main orientation is computed as its descriptor. Bay et al. [10] proposed SURF (speeded-up robust features). SURF has the same repeatability and distinctiveness as SIFT, but is computed faster than SIFT. Alahi et al. [11] proposed fast retina keypoint (FREAK). FREAK is a cascade of binary strings computed by comparing image intensities over a retinal sampling pattern. Ambai and Yoshida [12] proposed compact and real-time descriptors (CARD). Compared with SIFT and SURF, CARD can be computed rapidly utilizing lookup tables to extract histograms of oriented gradients. Other descriptors include ORB [13] and PCA-SIFT [14].

The above descriptors are devised for single-sensor images and yield a good matching performance on such images. Recently, multispectral systems became an attractive research topic, since they provide a rich representation of scene with images taken by different sensors [15]. Barrera et al. proposed an imaging system for computing depth maps from color and infrared images [16]. Stereo vision can be accomplished by keypoint matches. However, the descriptors, such as SIFT, SURF and ORB, are computed by utilizing the gradient pattern, which may revert on multispectral images [17,18], and hence, their performance deteriorates [19]. Since the computing gradient is a linear operation of original image intensities, the matching ability of descriptors relies on the linear relationship between image intensities.

Three factors contribute to the decrease of matching ability: the repeatability of keypoints, the accuracy of main orientation and the repeatability/distinctiveness of descriptors. From the perspective of descriptors, many techniques have been proposed to adapt descriptors of SIFT/SURF to multispectral images. Chen et al. [18] proposed the partial intensity invariant feature descriptor (PIIFD), which uses gradient orientation instead of direction. The gradient orientation is limited within [0, π), and PIIFD can register poor-quality multi-modal retinal image pairs. Saleem and Sablatnig [20] proposed NG-SIFT, which computes descriptors using a normalized gradient. NG-SIFT outperforms SIFT on image pairs of a visible image and a near-infrared image. Dellinger et al. [21] proposed SAR-SIFT for SAR images. SAR-SIFT uses a new gradient computation method, gradient by ratio (GR), which is robust to speckle noise, so that it can perform better on SAR images than SIFT. Hossain et al. [17] proposed the symmetric-SIFT algorithm for multi-modal image registration. It overcomes the problem that gradient direction could be inverted in different sensors.

Aguilera et al. [2] proposed the edge-oriented histogram (EOH). Unlike SIFT, EOH exploits only edge points in local windows rather than all pixels, since in general, edges are more likely repeatable and, hence, tending to be reliable between multi-sensor images. For an edge pixel, five responses are computed with filters designed in [22]. Edge points are detected by the Canny detector [23]. Note that the first four filters are directional derivatives, and the fifth filter is “no direction”. The problem with EOH is that it does not assign a main orientation to keypoints, which amounts to assuming that the main orientation for all keypoints takes on the same value, e.g., 0°. When the misalignment does not contain a rotation component, EOH works pretty well [2], but this limits the application of EOH to translation only. In real applications, image pairs taken from different views often contain a rotation component in the misalignment, and rotation invariant descriptors are hence desired or necessitated.

1.2. Proposed Method

To adapt EOH to dealing with rotation, we propose assigning a main orientation to keypoints for EOH computation. The main orientation makes EOH invariant to both translation and rotation and, hence, invariant to similarity transformation and partially invariant to affine transformations [1]. Note that the rotation contained in the misalignment is unknown and by no means can one obtain it before building keypoints.

Gauglitz et al. [24] proposed two orientation assignment methods. One method is to utilize the center-of-mass (COM), which is suitable for corners, and the other one is to utilize the histogram of intensities (HOI). Both COM and HOI are more suitable for single-spectral images, since they implicitly use the linear relationship of image intensities. This work utilizes the main orientation provided by PIIFD [18]. When a main orientation is assigned to a keypoint, the computation of its associated EOH descriptor needs interpolation at fractional pixels. Bilinear interpolation is applied to compute the response of the five filters used by EOH.

The rest of the paper is arranged as follows. Section 2 discusses assigning main orientation to keypoints and computing descriptors for keypoints relative to the main orientation, Section 3 gives the matching scheme for keypoints, Section 4 presents experimental results comparing the matching performance, and Section 5 concludes this paper.

2. Assigning Main Orientation to the Keypoint and Compute Descriptor

This section discusses assigning main orientation to keypoints, then the descriptor is computed for every keypoint with respect to the assigned main orientation.

2.1. Why a Main Orientation is Needed for Keypoints

Let Ir(x, y) and It(x, y) denote the reference and the test image to be registered. SIFT [1] calculates the histogram of gradient orientation and finds its peak in a local window to serve as the main orientation. Chen et al. [18] considered the problem of gradient and/or region reversal and square gradient (Gx,Gy) by:

[ G s , x G s , y ] = [ G x 2 G y 2 2 G x G y ]
and then smoothed the squared gradient by convolving it with an average filter hσ,
[ G s , x ¯ G s , y ¯ ] = [ G s , x h σ G s , y h σ ]

The main orientation is calculated as follows,

ϕ = 1 2 { tan 1 ( G s , y ¯ / G s , x ¯ ) + π , G s , x ¯ 0 tan 1 ( G s , y ¯ / G s , x ¯ ) + 2 π , G s , x ¯ < 0 G s , y ¯ 0 tan 1 ( G s , y ¯ / G s , x ¯ ) , G s , x ¯ < 0 G s , y ¯ < 0

A careful derivation shows that the main orientation ϕ and the traditional gradient (Gx, Gy) roughly have the following relationship. Let θ = atan2 (Gy, Gx) be the gradient direction falling in [−π, π]. atan2 is the four-quadrant inverse tangent for a gradient (Gx, Gy), giving the actual gradient direction for (Gx, Gy). In mathematics, it differs from tan−1 in that the range of tan−1 is ( π 2 , π 2 ), while the range of atan2 is (−π, π). θ is mapped to [0, π] by setting θ = mod(θ, π), i.e.,

θ = { θ , θ [ 0 , π ] θ + π , θ [ π , 0 ]

Then:

ϕ = { θ + π / 2 , θ [ 0 , π / 2 ] θ π / 2 , θ [ π / 2 , π ]

Equations (4) and (5) indicate that a gradient direction θ and its reversal direction θ±π will contribute to the same main orientation bin.

EOH is applied to Ir(x, y) and It(x, y) to detect keypoints and descriptors. Let K t i, i = 1, …, Nt, denote the i-th keypoint on the test image It(x, y), and K r j, j = 1, …, Nr, denote the j-th keypoint on the reference image Ir(x, y). Let f t i, i = 1, …, Nt, denote the descriptor of K t i, and f r j, j = 1, …, Nr, denote the descriptor of K r j. Note that both EOH [2] and PIIFD [18] employ the extrema of DOG as keypoints, which was proposed in SIFT. When detecting keypoints, σn = 0.5 is the standard deviation of the Gaussian function used for nominal smoothing of an input image. The threshold on the ratio of principle curvatures is set to 10, the default value in SIFT [1].

This work assigns the main orientation computed with PIIFD to keypoints K t i, i = 1, …, Nt and K r j, j = 1, …, Nr, then computes EOH descriptors. Like SIFT, the process of building keypoint matches includes four steps: (1) detect keypoints to be the extrema of DOG as proposed by SIFT; (2) assign to every keypoint main orientation computed with PIIFD; (3) compute the EOH descriptor for each keypoint with respect to its main orientation. Edge points are detected by the Canny operator, and all parameters of the Canny detectors are set to default values used by the MATLAB implementation, except that the standard deviation of the Gaussian filter σ is set to three, like in the original EOH [2]. The high threshold is defined to be the gradient magnitude ranked as the top 30%, and the low threshold is defined to be 40% of the high threshold. Interpolation is needed here to obtain pixel values at fractional pixels. Finally, (4) match the keypoint with the computed descriptor.

2.2. Compute the Descriptor for Keypoint with the Main Orientation

The EOH computes the gradient orientation at each edge pixel with the following five filters. These filters correspond to the 0°, 45°, 90°, 135° and non-direction, as shown in Figure 1. The filters shown in Figure 1a–d, are called direction filters, while the one shown in Figure 1e is called the non-direction filter. For a pixel, the filter giving the maximum response is defined to be the direction at the pixel. Formally, let fk(x, y),k = 0, 1, 2, 3,4, denote the mathematical representation of the five filters shown in Figure 1, then an edge pixel at (x, y) will contribute one to the bin defined by:

bin EOH ( x , y ) = arg max k | f k ( x , y ) I ( x , y ) |
where ‘•’ is the correlation between image and filter.

The four direction filters shown in Figure 1 are in fact the orientation partition used in PIIFD. SIFT employs eight orientations (bins) for computing descriptors, i.e., for a pixel, its gradient orientation is quantized to eight bins with the center of each bin being 0°, 45°, 90°, 135°, 180°, 225°, 270°, 315°. PIIFD considers the gradient reverse and utilize mod-180° orientation. Specifically, let αx,y be the gradient orientation at (x, y). For SIFT, αx,y contributes to the bin:

bin SIFT ( α x , y ) = round ( α x , y 45 ) % 8
For PIIFD, it contributes to the bin:
bin PIIFD ( α x , y ) = round ( α x , y 22.5 ) % 8
Similar to SIFT, PIIFD uses eight bins in Equation (8). However, Equation (8) maps an orientation αx,y and αx,y + 180° to the same bin, i.e., binPIIFD(αx,y) = binPIIFD(αx,y + 180°). PIIFD partitions [0, 2π] into 16 bins at first, with each bin covering 22.5°, and then “merges” two bins if their center angles differ by 180°.

Equation (7) also says that the centers of the first four orientation bins for SIFT are 0°, 45°, 90°, 135°, exactly the same as EOH bins. Thus,

bin SIFT ( α x , y ) % 4 = b i n EOH ( α x , y )

When main orientation is assigned to a keypoint, we need to compute the maximum response of the five filters. The five filters are rotated by the amount of main orientation, and the rotated pixels for computing the filter response lie in a fractional grid, as shown in Figure 2. To obtain pixel values at the fractional grid, a bilinear interpolation is employed.

After the interpolation step, pixel values for computing the filter response are obtained. The filter giving the maximum response is defined to be the direction at this pixel and contributes to EOH descriptors. As in the original EOH [2], a local window of radius 50 is used for computing an EOH descriptor. Only edge pixels in the window contribute to the descriptor. Alternatively, we can skip the interpolation step and just utilize the gradient orientation at edge pixels, which is discussed in Section 2.3 as a variant of implementing EOH.

2.3. Variant Implementation of EOH

Computing the responses of five filters shown in Figure 1 can be speeded up by fast Fourier transform (FFT). We use It(x, y) as an example and compute the filter responses with FFT. A similar process can be applied to Ir(x, y). Let I t 0 ( x , y ) denote the response of the zeroth filter (0°) applied to It(x, y). Formally,

I t 0 ( x , y ) = I t ( x , y ) f 0 ( x , y ) = I t ( x , y ) * f 0 ' ( x , y )
wherein ‘•’ denotes the correlation, which can be rewritten as the convolution of f 0 ' ( x , y ) and It(x, y). f 0 ' ( x , y ) is the version of f0(x, y) flipped left-right and up-down, i.e., f 0 ' ( x , y ) = f 0 ( x , y ). The convolution I t ( x , y ) * f 0 ' ( x , y ) can be quickly implemented with FFT.

The equivalence of four directional filter responses in EOH to the first four orientation bins in SIFT and the above discussed fast computation by FFT, motivate implementing EOH as follows. Define the directional gradient along the horizontal and vertical axes to be:

G x ( x , y ) = I ( x , y ) f 0 ( x , y ) G y ( x , y ) = I ( x , y ) f 2 ( x , y )
Note, the gradient computation in Equation (11) is nothing but the Sobel operator [25]. Once the directional gradient is obtained, its direction can be simply calculated by:
α ( x , y ) = atan 2 ( G y ( x , y ) , G x ( x , y ) )
Equation (11) together with Equation (9) gives the orientation bin to which every edge pixel contributes. By this means, we implement a variant of EOH, which can be understood from the adaptation of different descriptors.

  • The variant ignores the non-direction bin, which in our experiments proved to have little effect on matching performance. See the analysis in Tables 1 and 2 in Section 4. Furthermore, the orientation bin computed with Equations (12), (11) and (9) may not be identical to that computed with Equation (6). See Section 4 for their matching performance comparison.

  • NG-SIFT [20] utilizes the normalized gradient magnitude Ω ^ ( x , y ) = F ^ h 2 ( x , y ) + F ^ v 2 ( x , y ) to compute the descriptor and does not distribute a pixel to neighboring spatial/orientational bins. See Equations (5)(7) in [20]. EOH is similar to NG-SIFT from this aspect.

  • The proposed variant of EOH utilizes the Sobel operator in Equation (11), and SIFT, PIIFD and NG-SIFT utilize the filters [−1, 0, 1] and [−1, 0, 1]T to calculate the horizontal and vertical directional derivatives.

3. Matching Keypoints with Descriptors

This section discusses matching keypoints with descriptors. The matching ability of descriptors is evaluated with the number of correct keypoint matches. To make a fair comparison for different descriptors, a simple matching approach suggested by SIFT is employed here. A reference keypoint K r j 0 is defined to be matched to a test keypoint K t i 0 if:

D ( f t i 0 , f r j 0 ) < 0.8 D ( f t i 0 , f r j 1 )
where D(·,·) is the Euclidean distance and f r j 1 is the second-closest neighbor to f t i 0. The ‘0.8’ in Equation (13) can be changed to 0.6, which means a tighter matching criterion giving fewer matched keypoints.

Equation (13) is the matching method suggested in the original SIFT [1]. Through Equation (13), a set of keypoint mappings can be established, which will be used to analyze the descriptor performance. See Section 4 for details. Note that post-processing techniques can be applied to keypoints and descriptors for removing outlier keypoint matches. Commonly-used techniques include RANSAC [26,27], its variant fast sample consensus (FSC) [28], etc. However, post-processing is to some extent independent of descriptor matching ability, and the resulting improvement ought to be excluded for comparing descriptors.

4. Experimental Results

This section presents the experimental results. Visual matching results are provided firstly, followed by the quantitative analysis on matching results. The proposed method is compared with the original EOH. Two datasets, EOIRand VS-long-wave infrared (LWIR), are used for investigating the matching performance. EOIR includes 87 image pairs captured by ourselves, 12 Landsat image pairs from NASA. The 87 image pairs include outdoor and indoor scenes with one image taken with visible light and the other taken with middle-wave infrared (MWIR) light. The 12 Landsat image pairs were downloaded from [29] with one taken with the visible band, e.g., Landsat 8 Band 3 Visible (0.53–0.59 µm), and the other taken with middle-wave light or the Thermal Infrared Sensor (TIRS), e.g., Landsat 8 Band 10 TIRS 1 (10.6–11.19 µm). Dataset VS-LWIR is from [2] containing 100 image pairs, one image taken with the visible bandwidth (0.4–0.7 µm) and the other taken with the long-wave infrared bandwidth (LWIR, 8–14 µm).

4.1. Visual Results

This section gives visual matching results. Figure 3 gives the keypoint matchings built with the original EOH without the main orientation and the proposed method. The visible image serves as the reference image, and the infrared image is used as the test image. The test image is rotated by 10°, 20° and 30°. Figure 3a,c,e show the matching result of EOH between the reference and the rotated test image by 10°, 20° and 30°, respectively. Due to the lack of main orientation, the keypoint matches built with the EOH contain very few or no correct matches. As a comparison, the proposed method provides sufficiently many correct matches in Figure 3b,d,f.

Figure 4 shows the keypoint matches on an image pair from dataset EOIR built with EOH, EOH equipped with SIFT main orientation, with COM (center-of-mass) main orientation [24], with HOI (histogram of intensity) main orientation [24] and the proposed method. The infrared image is rotated by 20°. EOH provides five keypoint matches in Figure 4a, and three are visually correct. SIFT main orientation gives seven keypoint matches in Figure 4b, and four matches are visually correct. The COM and HOI main orientations do not give many correct matches, as shown in Figure 4c,d, while the proposed method gives 11 keypoint matches in Figure 4e, and nine matches are visually correct. Visually, the SIFT main orientation and the proposed method give almost the same correct rate of matches, except that the proposed method gives more matches. The reason might be that although this pair of images was taken with a visible camera and an infrared camera, they are very close to single-spectrum images, i.e., brighter (darker) areas in the visible image are also brighter (darker) in the infrared image. However, the relationship between image intensities is not linear, which makes COM and HOI not perform very well.

Figure 5 illustrates the keypoint matches on an image pair from dataset VS-LWIR built with EOH, EOH equipped with SIFT main orientation, COM main orientation, HOI main orientation and the proposed method. The performance of EOH and EOH equipped with SIFT main orientation in Figure 5a,b is inferior to that in Figure 4a,b. The performance of COM and HOI is not good either, as shown in Figure 5c,d. This image pair is taken with a visible camera and an LWIR camera. The multimodality between them causes the inaccuracy of SIFT main orientation, COM and HOI and, hence, the mismatches in Figure 5b–d. The proposed method, for the induction of main orientation to keypoints, performs still well on this image pair.

4.2. Quantitative Results

This section presents quantitative results. The above visual results can only provide a simpler comparison on a few image pairs, and the comparison result is affected by an individual criterion on “correct” matches. To quantitatively assess the performance for different methods, we perform statistics on the number of correct matches. We define a keypoint match to be correct if the distance d between the two keypoints comprising the match is smaller than a threshold d0. In the literature, different values have been used for d0, e.g., d0 = 2, d0 = 4, d0 = 5, etc. [4]. To eliminate the effect of d0 on the performance comparison, d0 = 1, 2, 3, 4, 5, 10, 20, 50, 100 are used in this work. The number of keypoint matches of distance d < d0 is counted and listed in Table 1 for dataset EOIR and Table 2 for dataset VS-LWIR.

In Tables 1 and 2, the test (infrared) image is rotated by 10°, 20°, 30° and 45°. The proposed method outperforms the original EOH for all rotation degrees. For example, when the test image is rotated by 10°, 45.60% of the keypoint matches on dataset EOIR built with the proposed method has a distance less than five, i.e., falling in the range [0, 5], while the EOH has 33.17% falling in [0, 5]. On dataset VS-LWIR, the proposed method has 29.91% matches falling in [0, 5], while the EOH has 19.53%. Additionally, this also indicates that the dataset VS-LWIR is more challenging than EOIR. Both the proposed method and EOH provide superior results on dataset EOIR over VS-LWIR. COM and HOI perform only slightly better than the original EOH without the main orientation on EOIR and VS-LWIR. This to some extent indicates that the main orientations computed with COM and HOI on visible and infrared images are not so accurate as the ones computed with PIIFD, failing to account for the rotation difference between two images.

The performance decreases with the increase of rotation degree for all methods. For example, on dataset VS-LWIR, when the test image is rotated by 10°, the proposed method has 48.13% of matches falling in [0, 10], but this number decreases to 43.08%, 33.13% and 24.02% when the test image is rotated by 20°, 30° and 45°. For EOH, the percent of keypoint matches falling in [0, 10] decreases more than the proposed method, from 41.83% to 5.93%, 1.20% and 0.44%. The performance decrease for EOH is due to the lack of main orientation, while the decrease for the proposed method originates from the inaccuracy of computing main orientation. Figure 6 shows the performance of different methods under rotation. Keypoint matches of distance d ≤ 10 are defined to be correct. From Figure 6, it can be seen that the percent of correct matches for all methods decreases with the increase of rotation degree. On both EOIR and VS-LWIR, the proposed method and the variant implementation of EOH decrease slower than the original EOH without the main orientation, SIFT main orientation, COM main orientation and HOI main orientation.

The variant implementation of EOH with the main orientation performs comparable to the proposed method that utilizes the five filters in Figure 1. On dataset VS-LWIR, the variant gives 18.55% of matches falling in [0, 5] when the rotation is 20°, and the proposed method gives 16.50%. When the rotation gets to 45°, the variant gives 11.79%, and the proposed method gives 12.35%. It can also be observed from Figure 6 that the variant EOH proposed in Section 2.3 performs as well as the proposed method. From Tables 1 and 2, we can conclude that the proposed variant implementation of EOH can yield keypoint matches as reliable as the EOH assigned to the main orientation. This explains and verifies that the non-direction bin does not have a great effect on the matching ability and is not used in descriptors, such as SIFT, SURF and PIIFD.

5. Conclusions

This paper proposed an approach to assigning the main orientation to keypoints. The PIIFD main orientation was calculated for a keypoint, and then, the EOH descriptor is computed with respect to the main orientation. Experimental results show that the assigned main orientation can significantly improve the matching performance of EOH on images of misalignment containing rotation. Additionally, a variant of EOH is proposed that employs the gradient orientation as the filter responses. The variant EOH can be computed with respect to the main orientation more easily and achieve a comparable matching performance to the original EOH, but needs less computational cost.

Acknowledgements

Thanks go to the authors of [2] for kindly providing the source code and test images.

This work was supported by the National Natural Science Foundation of China (Grant Nos. NSFC-61170176, NSFC-61402046, NSFC-61471067), the Fund for the Doctoral Program of Higher Education of China (Grant No. 20120005110002), the Fund for Beijing University of Posts and Telecommunications (Grant Nos. 2013XD-04, 2013XZ10), the Fund for National Great Science Specific Project (Grant No. 2014ZX03002002-004) and the Fund for the Panasonic R&D Center (China).

Author Contributions

Yong Li conceived the idea and writen the paper; Junwei Zou and Xiang Shi executed the simulation and experimental work; Lijun Wei and Fang Chen analyzed the data.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar]
  2. Aguilera, C.; Barrera, F.; Lumbreras, F.; Sappa, A.D.; Toledo, R. Multispectral image feature points. Sensors 2012, 12, 12661–12672. [Google Scholar]
  3. Lindeberg, T. Detecting salient blob-like image structures and their scales with a scale-space primal sketch: A method for focus-of-attention. Int. J. Comput. Vision 1993, 11, 283–318. [Google Scholar]
  4. Mikolajczyk, K.; Schmid, C. A performance evaluation of local descriptors. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1615–1630. [Google Scholar]
  5. Miao, Q.; Wang, G.; Shi, C.; Lin, X.; Ruan, Z. A new framework for on-line object tracking based on surf. Pattern Recognit. Lett. 2011, 32, 1564–1571. [Google Scholar]
  6. Mikolajczyk, K.; Tuytelaars, T.; Schmid, C.; Zisserman, A.; Matas, J.; Schaffalitzky, F.; Kadir, T.; Gool, L.V. A comparison of affine region detectors. Int. J. Comput. Vision 2005, 65, 43–72. [Google Scholar]
  7. Mikolajczyk, K.; Schmid, C. Scale & affine invariant interest point detectors. Int. J. Comput. Vis. 2004, 60, 63–86. [Google Scholar]
  8. Matas, J.; Chum, O.; Urban, M.; Paidla, T. Robust wide baseline stereo from maximally stable extremal regions. Proceedings of the British Machine Vision Conference (BMVC), Cardiff, UK, 2–5 September 2002; pp. 384–393.
  9. Shi, C.; Wang, G.; Lin, X.; Wang, Y.; Liao, C.; Miao, Q. Topology based affine invariant descriptor for msers. Proceedings of the International Conference on Image Processing (ICIP), Hong Kong, China, 26–29 March 2010; pp. 133–136.
  10. Bay, H.; Ess, A.; Tuytelaars, T.; Gool, L.V. Speeded up robust features (surf). Comput. Vision Image Underst. 2008, 110, 346–359. [Google Scholar]
  11. Alahi, A.; Ortiz, R.; Vandergheynst, P. Freak: Fast retina keypoint. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 510–517.
  12. Ambai, M.; Yoshida, Y. Card: Compact and real-time descriptors. Proceedings of the IEEE International Conference on Computer Vision, Barcelona, Spain, 8–11 November 2011; pp. 97–104.
  13. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. Proceedings of the IEEE Computer Vision (ICCV), Barcelona, Spain, 8–11 November 2011; pp. 2564–2571.
  14. Ke, Y.; Sukthankar, R. Pca-sift: A more distinctive representation for local image descriptors. Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Washington, DC , USA, 27 June–2 July 2004; pp. 506–513.
  15. Barrera, F.; Lumbreras, F.; Sappa, A.D. Multispectral piecewise planar stereo using manhattan-world assumption. Pattern Recognit. Lett. 2013, 34, 52–61. [Google Scholar]
  16. Campo, F.B.; Ruiz, F.L.; Sappa, A.D. Multimodal stereo vision system: 3d data extraction and algorithm evaluation. IEEE J. Sel. Top. Signal Process. 2012, 6, 437–446. [Google Scholar]
  17. Hossain, M.T.; Teng, S.W.; Lu, G. Achieving high multi-modal registration performance using simplified hough-transform with improved symmetric-sift. Proceedings of the International Conference on Digital Image Computing: Techniques and Applications (DICTA), Queensland, Australia, 6–8 December 2012; pp. 1–7.
  18. Chen, J.; Tian, J.; Lee, N.; Zheng, J.; Smith, R.T.; Laine, A.F. A partial intensity invariant feature descriptor for multimodal retinal image registration. IEEE Trans. Biomed. Eng. 2010, 57, 1707–1718. [Google Scholar]
  19. Ricaurte, P.; Chilán, C.; Aguilera-Carrasco, C.A.; Vintimilla, B.X.; Sappa, A.D. Feature point descriptors: Infrared and visible spectra. Sensors 2014, 14, 3690–3701. [Google Scholar]
  20. Saleem, S.; Sablatnig, R. A robust sift descriptor for multispectral images. IEEE Signal Process. Lett. 2014, 21, 400–403. [Google Scholar]
  21. Dellinger, F.; Delon, J.; Gousseau, Y.; Michel, J.; Tupin, F. Sar-sift: A sift-like algorithm for sar images. IEEE Trans. Geosci. Remote Sens. 2015, 53, 453–466. [Google Scholar]
  22. Manjunath, B.S.; Ohm, J.-R.; Vasudevan, V.V.; Yamada, A. Color and texture descriptors. IEEE Trans. Circuits Sys. Video Technol. 2001, 11, 703–715. [Google Scholar]
  23. Canny, J. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 8, 679–698. [Google Scholar]
  24. Gauglitz, S.; Turk, M.; Höllerer, T. Improving keypoint orientation assignment. Proceedings of the British Machine Vision Conference, Dundee, UK, 29 August–2 September 2011; pp. 93.1–93.11.
  25. Forsyth, D.A.; Ponce, J. Computer Vision; Prentice Hall: Upper Saddle River, NJ, USA, 2003. [Google Scholar]
  26. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar]
  27. Tan, H.; Zhou, Y.; Zhu, Y. A novel curve lane detection based on Improved River Flow and RANSAC. Proceedings of IEEE 17th International Conference on Intelligent Transportation Systems (ITSC), Qingdao, China, 8-11 October 2014; pp. 133–138.
  28. Wu, Y.; Ma, W.; Gong, M. A novel point-matching algorithm based on fast sample consensus for image registration. IEEE Geosci. Remote Sens. Lett. 2015, 12, 43–47. [Google Scholar]
  29. USGS Landsat. Available online: http://landsat.usgs.gov/ accessed on 13 February 2015.
Figure 1. The five filters used in [2,22]. The filters compute directional derivatives in 0°, 45°, 90°, 135° and the non-direction showed in (a–d) and (e) respectively. (a–d) are the direction filters and (e) is the non-direction filter.
Figure 1. The five filters used in [2,22]. The filters compute directional derivatives in 0°, 45°, 90°, 135° and the non-direction showed in (a–d) and (e) respectively. (a–d) are the direction filters and (e) is the non-direction filter.
Sensors 15 15595f1 1024
Figure 2. When the main orientation is assigned to a keypoint, the filters shown in Figure 1 ought to be rotated with respect to the main orientation. Black dots represent the integer pixel grid, and red dots are the fractional pixel locations, whosewhole values are used by the rotated filters.
Figure 2. When the main orientation is assigned to a keypoint, the filters shown in Figure 1 ought to be rotated with respect to the main orientation. Black dots represent the integer pixel grid, and red dots are the fractional pixel locations, whosewhole values are used by the rotated filters.
Sensors 15 15595f2 1024
Figure 3. The matching performance under rotation. The test image is rotated by 10°, 20°, and 30° from top to bottom line. The left column is the result of EOH without main orientation. The right column is the result of EOH equipped with the partial intensity invariant feature descriptor (PIIFD) main orientation.
Figure 3. The matching performance under rotation. The test image is rotated by 10°, 20°, and 30° from top to bottom line. The left column is the result of EOH without main orientation. The right column is the result of EOH equipped with the partial intensity invariant feature descriptor (PIIFD) main orientation.
Sensors 15 15595f3 1024
Figure 4. The matched keypoints built with descriptors. (a) The original EOH without main orientation; (b) the main orientation computed by SIFT, ranging from [0, 2π]; (c) the main orientation computed by center-of-mass (COM); (d) the main orientation computed by the histogram of intensities (HOI); (e) the proposed method that utilizes the main orientation computed by PIIFD. The test (IR) image is rotated by 20°.
Figure 4. The matched keypoints built with descriptors. (a) The original EOH without main orientation; (b) the main orientation computed by SIFT, ranging from [0, 2π]; (c) the main orientation computed by center-of-mass (COM); (d) the main orientation computed by the histogram of intensities (HOI); (e) the proposed method that utilizes the main orientation computed by PIIFD. The test (IR) image is rotated by 20°.
Sensors 15 15595f4 1024
Figure 5. The matched keypoints built with descriptors. (a) The original EOH without main orientation; (b) the main orientation computed by SIFT, ranging from [0, 2π]; (c) the main orientation computed by COM; (d) the main orientation computed by HOI; (e) the proposed method that utilizes the main orientation computed by PIIFD. The test (IR) image is rotated by 20°.
Figure 5. The matched keypoints built with descriptors. (a) The original EOH without main orientation; (b) the main orientation computed by SIFT, ranging from [0, 2π]; (c) the main orientation computed by COM; (d) the main orientation computed by HOI; (e) the proposed method that utilizes the main orientation computed by PIIFD. The test (IR) image is rotated by 20°.
Sensors 15 15595f5 1024
Figure 6. Comparison of keypoint matches for different methods. The horizontal axis represents the degree of rotation, and the vertical axis represents the percent of correct matches. Correct matches are defined to be of a distance falling in (0 10]. (a) The result on dataset EOIR; (b) the result on dataset VS-long-wave infrared (LWIR). On both EOIR and VS-LWIR, the performance of every method decreases when the rotation degree increases. The proposed method and the variant EOH with main orientation decrease significantly slower than the original EOH, SIFT main orientation, COM and HOI on both datasets, varying the effectiveness of the main orientation.
Figure 6. Comparison of keypoint matches for different methods. The horizontal axis represents the degree of rotation, and the vertical axis represents the percent of correct matches. Correct matches are defined to be of a distance falling in (0 10]. (a) The result on dataset EOIR; (b) the result on dataset VS-long-wave infrared (LWIR). On both EOIR and VS-LWIR, the performance of every method decreases when the rotation degree increases. The proposed method and the variant EOH with main orientation decrease significantly slower than the original EOH, SIFT main orientation, COM and HOI on both datasets, varying the effectiveness of the main orientation.
Sensors 15 15595f6 1024
Table 1. Comparison of keypoint matches for different methods on dataset EOIR. The cumulative percent of correct matches (CPCM) is presented for different d0.
Table 1. Comparison of keypoint matches for different methods on dataset EOIR. The cumulative percent of correct matches (CPCM) is presented for different d0.
RotationError[0 1](1 2](2 3](3 4](4 5](5 10](10 20](20 50](50 100]>100
10°Proposed142134985048153995659196
CPCM (%)13.7226.6736.1440.9745.6060.3969.9575.3681.06100.00

EOH74735537321321345975146
CPCM (%)9.0617.9924.7229.2533.1749.3365.7372.9582.13100.00

COM1715111066365103142437
CPCM (%)1.963.684.956.106.7914.0421.5233.3749.71100.00

HOI322214141610411089123236
CPCM (%)4.217.118.9510.7912.8926.5841.0552.7668.95100.00

VAR-EOH143131996051155696955216
CPCM (%)13.6526.1535.5941.3246.1860.9767.5674.1479.39100.00

20°Proposed97131837054112834972203
CPCM (%)10.1723.9032.6039.9445.6057.3466.0471.1778.72100.00

EOH108112440528469176
CPCM (%)2.193.956.366.807.6816.4527.8546.2761.40100.00

COM202622142277105148182448
CPCM (%)1.884.326.397.719.7717.0126.8840.7957.89100.00

HOI6128121684120179307535
CPCM (%)0.471.412.032.974.2210.7920.1734.1758.17100.00

VAR-EOH101129756349103717475199
CPCM (%)10.7624.4932.4839.1944.4155.3862.9470.8278.81100.00

30°Proposed85121715132115815584234
CPCM (%)9.1522.1729.8235.3138.7551.1359.8565.7774.81100.00

EOH001008164850194
CPCM (%)0.000.000.320.320.322.847.8923.0338.80100.00

COM145444364135214564
CPCM (%)0.100.480.961.351.735.8812.0425.0545.66100.00

HOI011101194752224
CPCM (%)0.000.290.580.870.874.056.6520.2335.26100.00

VAR-EOH90115766845113895177213
CPCM (%)9.6121.8829.9937.2542.0554.1163.6169.0577.27100.00

45°Proposed61874338311005458137463
CPCM (%)5.6913.8117.8221.3624.2533.5838.6244.0356.81100.00

EOH00000022253406
CPCM (%)0.000.000.000.000.000.000.414.9715.94100.00

COM0213333521051831200
CPCM (%)0.000.130.190.380.572.655.9412.5824.15100.00

HOI00000132195469
CPCM (%)0.000.000.000.000.000.170.684.2420.37100.00

VAR-EOH53745646341166857167463
CPCM (%)4.6711.2016.1420.1923.1933.4239.4244.4459.17100.00
Table 2. Comparison of keypoint matches for different methods on dataset VS-LWIR. The cumulative percent of correct matches (CPCM) is presented.
Table 2. Comparison of keypoint matches for different methods on dataset VS-LWIR. The cumulative percent of correct matches (CPCM) is presented.
RotationError[0 1](1 2](2 3](3 4](4 5](5 10](10 20](20 50](50 100]>100
10°Proposed16658150441561098250203
CPCM (%)1.879.4618.9324.7729.9148.1360.8670.4476.29100.00

EOH9274137271611467039165
CPCM (%)1.254.9910.6615.7919.5341.8362.0571.7577.15100.00

COM81913151186118123112450
CPCM (%)0.842.834.195.766.9115.9228.2741.1552.88100.00

HOI8115917791249473280
CPCM (%)1.142.713.434.717.1418.4336.1449.5760.00100.00

VAR-EOH17667370401621067141216
CPCM (%)1.979.6318.1026.2230.8649.6561.9570.1974.94100.00

20°PROPOSED22445232461621366542230
CPCM (%)2.657.9414.2018.0523.5943.0859.4567.2772.32100.00

EOH0042321334639358
CPCM (%)0.000.000.791.191.785.9312.4521.5429.25100.00

COM3545149394128176458
CPCM (%)0.310.821.221.733.1612.6522.2435.3153.27100.00

HOI04812750103163211836
CPCM (%)0.000.290.861.722.225.8113.2024.8940.03100.00

VAR-EOH21474037471591266941227
CPCM (%)2.588.3513.2717.8123.5943.1258.6067.0872.11100.00

30°Proposed7313133341371058837321
CPCM (%)0.854.618.3712.3816.5033.1345.8756.5561.04100.00

EOH00000943262642
CPCM (%)0.000.000.000.000.001.201.746.0114.29100.00

COM02913669102106127771
CPCM (%)0.000.170.911.992.498.2216.6825.4836.02100.00

HOI1000112194853337
CPCM (%)0.210.210.210.210.422.977.0117.2028.45100.00

VAR-EOH8255135291211128555277
CPCM (%)1.004.1410.5314.9118.5533.7147.7458.4065.29100.00

45°Proposed102624232710488652450
CPCM (%)1.124.046.739.3212.3524.0234.0143.6649.49100.00

EOH100004842127967
CPCM (%)0.090.090.090.090.090.441.134.7915.84100.00

COM1037333591091611448
CPCM (%)0.050.050.220.600.772.585.8111.7920.61100.00

HOI100002124976347
CPCM (%)0.210.210.210.210.210.623.0813.1428.75100.00

VAR-EOH92423252794987066480
CPCM (%)0.983.606.118.8411.7922.0532.7540.3947.60100.00

Share and Cite

MDPI and ACS Style

Li, Y.; Shi, X.; Wei, L.; Zou, J.; Chen, F. Assigning Main Orientation to an EOH Descriptor on Multispectral Images. Sensors 2015, 15, 15595-15610. https://doi.org/10.3390/s150715595

AMA Style

Li Y, Shi X, Wei L, Zou J, Chen F. Assigning Main Orientation to an EOH Descriptor on Multispectral Images. Sensors. 2015; 15(7):15595-15610. https://doi.org/10.3390/s150715595

Chicago/Turabian Style

Li, Yong, Xiang Shi, Lijun Wei, Junwei Zou, and Fang Chen. 2015. "Assigning Main Orientation to an EOH Descriptor on Multispectral Images" Sensors 15, no. 7: 15595-15610. https://doi.org/10.3390/s150715595

Article Metrics

Back to TopTop