Next Article in Journal
Laterally Driven Resonant Pressure Sensor with Etched Silicon Dual Diaphragms and Combined Beams
Previous Article in Journal
High-Accuracy, Compact Scanning Method and Circuit for Resistive Sensor Arrays
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spectral Similarity Assessment Based on a Spectrum Reflectance-Absorption Index and Simplified Curve Patterns for Hyperspectral Remote Sensing

1
College of Resource and Environmental Science, Fujian Agriculture and Forestry University, Fuzhou 350002, China
2
Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
3
Department of Geography, Hong Kong Baptist University, Hong Kong, China
4
College of Electrical and Information Engineering, Hunan University, Hunan 410082, China
*
Author to whom correspondence should be addressed.
Sensors 2016, 16(2), 152; https://doi.org/10.3390/s16020152
Submission received: 27 October 2015 / Revised: 20 January 2016 / Accepted: 21 January 2016 / Published: 26 January 2016
(This article belongs to the Section Remote Sensors)

Abstract

:
Hyperspectral images possess properties such as rich spectral information, narrow bandwidth, and large numbers of bands. Finding effective methods to retrieve land features from an image by using similarity assessment indices with specific spectral characteristics is an important research question. This paper reports a novel hyperspectral image similarity assessment index based on spectral curve patterns and a reflection-absorption index. First, some spectral reflection-absorption features are extracted to restrict the subsequent curve simplification. Then, the improved Douglas-Peucker algorithm is employed to simplify all spectral curves without setting the thresholds. Finally, the simplified curves with the feature points are matched, and the similarities among the spectral curves are calculated using the matched points. The Airborne Visible Infrared Imaging Spectrometer (AVIRIS) and Reflective Optics System Imaging Spectrometer (ROSIS) hyperspectral image datasets are then selected to test the effect of the proposed index. The practical experiments indicate that the proposed index can achieve higher precision and fewer points than the traditional spectral information divergence and spectral angle match.

1. Introduction

A hyperspectral remote sensing image has the combined characteristics of an image and spectrum; in addition, undetectable substances can be easily detected in multi-spectral remote sensing images [1]. In recent years, hyperspectral remote sensing technologies have developed rapidly and are being widely applied in many industries. Remote sensing image classification has evolved from rough recognition using multi-spectral images to undertake spectrum-analysis-based fine identification using hyperspectral images [2,3]. However, due to the increased cost of hyperspectral remote sensing image data volume, high correlation among the bands and training samples, the traditional hyperspectral remote sensing image classification and recognition technologies cannot satisfy the requirements of hyperspectral remote sensing applications [4].
To address issues in hyperspectral remote sensing image feature identification and classification, scholars have proposed various methods, which can be roughly divided into two categories: ones based on feature space and those based on spectral space. For the first category, according to the Hughes phenomenon, the number of training samples can sharply increase with the increase of band number. The common solution is to reduce the dimensionality of the hyperspectral image by using the methods of principal component analysis [5,6], independent component analysis [7], information entropy [8], genetic algorithm, projected track [9], wavelet analysis [10], minimum noise fraction transformation [11], neural networks [12], particle swarm optimization, manifold learning algorithm, and so on. The Hughes phenomenon can be avoided once the dimensionality of the hyperspectral remote sensing image is controlled. The hyperspectral remote sensing image classification can be completed using conventional remote sensing image classification methods, such as the maximum likelihood method [13], Bayesian classifier method [14], decision tree, artificial neural networks [15], and so on, Another interesting classification approach is based on kernel space, and includes support vector machines (SVM) and kernel fisher discriminates (KFD), to name a few. Such methods have been widely used because they are not limited by the Hughes phenomenon [16].
The other category is based on spectral space, and includes Euclidean distance (ED), spectral absorption index (SAI), spectral angle mapper (SAM) [17,18], spectral correlation mapper (SCM), mutual information [19], spectral gradient angle [20] spectral information divergence (SID) [21], set methodology, and so on. Based on the fine spectral information of hyperspectral remote sensing images, such methods do not require complex analysis and dimension reduction.
As the essence of remote sensing image classification is to divide objects into several different categories by extracting the characteristics of different objects and using a certain similarity measurement combining some classification criteria, the key aspect of hyperspectral remote sensing image classification is the assessment of spectral similarities among different objects. Some scholars have carried out impact studies on the similarity assessment of the classification results by combining spectral amplitude and shape.
Du et al. [22] proposed a classification method combining the spectral information dispersion and spectral cosine angular, while Kumar et al. [23] proposed a method using the combination of the correlation coefficient (CC) and spectral information dispersion. Kong et al. [24] presented a spectral similarity metric that combines various spectral features using geometric distance, the CC, and relative entropy. Meanwhile, Du et al. [25] extracted the characteristics of spectral absorption/reflection, central moment, fractal dimension and information entropy for image retrieval, and concluded that these are actually inappropriate parameters for image retrieval. Spectral curve feature extraction has also been applied to various fields, such as crop identification and chemistry analysis [26,27]. SAM has been used extensively for distinguishing different objects because it is capable of repressing the influence of shading to enhance the target reflectance [28]. However, this method can only distinguish between negative and positive correlations because only the absolute value is considered [29]. In comparison, SID models the spectrum as a probability distribution and describes the spectral features by using statistical moments; it also considers the spectral variability as a random uncertainty [21]. Based on these advantages, some studies have combined SAM and SID to increase the discriminatory capacity for effective image retrieval [30]. These studies have proven that the classification accuracy of using a combination of two or more similarity measures is higher than that using a single similarity measure. However, the comparisons and evaluations of similarity measures have been carried out using only spectral data or a few samples of land cover types. Furthermore, the comparison methods do not use uniform standards and, to date, there have been few systematic evaluations of similarity measure methods.
All the bands of reference and measured spectra are involved in the computation when using various similarity metrics. In this case, the enormous amount of required calculation can lead to low efficiency in similarity assessment because there may be hundreds of thousands of bands involved. Thus, certain simplified or filtration methods must be used to reduce the amount of computation. The Douglas-Peucker (DP) algorithm is an efficient and extensively used method to simplify a curve. However, the simplification effect depends on the threshold, which cannot be adaptively set. Therefore, in this paper, an improved DP (IDP) algorithm without threshold under spectral reflection-absorption index (SRAI)-restriction is proposed to meet this requirement. In addition, a similarity index is proposed to assess the similarities among different spectrum curves. Section 2 reviews two widely used similarity metrics, i.e., SAM and SID. Section 3 presents the proposed SRAI-restrained IDP method. Section 4 illustrates the assessing experiments using the proposed method and traditional ones. Then, the effects of the parameters are evaluated. The final section draws the conclusions based on the obtained results.

2. Related Similarities of Spectral Vectors

A spectral vector (curve) is the representation of ground feature that is used in hyperspectral remote sensing. Ground feature identification can be achieved by measuring the similarities of the spectral vectors. Two typical similarity indices are shown below.

2.1. SID

SID is based on the information entropy; it can measure the similarities and separability of pixels. First, the information entropy of each point is calculated; then, they are compared to measure the similarities of the spectral vectors using the formula:
S I D ( A , B ) = D ( A | | B ) + D ( B | | A )
where:
D ( A | | B ) = i = 1 N p i log ( p i / q i )
D ( B | | A ) = i = 1 N q i log ( q i / p i )
p i = A i / i = 1 N A i
q i = B i / i = 1 N B i
In the equations above, N refers to the number of bands, and A = ( A 1 , A 2 , , A N ) and B = ( B 1 , B 2 , , B N ) refer to the two spectral vectors, respectively. Here, the lower the SID value, the higher the similarity of both spectral vectors.

2.2. SAM

SAM, also known as vector-included angle cosine method, assess similarity by calculating the angle between two spectral vectors. The spectral angle refers to the angle of two spectral vectors with the same wavelength range, which is given by:
cos α = A B | A | | B | = i = 1 N A i B i / ( i = 1 N A i A i i = 1 N B i B i )
where A and B refer to the two spectral vectors, respectively, and α refers to the spectral angle. The cosine value of α is needed in the computation. Here, the higher the SAM value, the higher the similarity of the two spectral vectors.

3. Proposed Approach

3.1. SRAI

The spectral curves of various ground features have different properties in terms of absorption wave peak and valley, location, width, depth, and symmetry. Therefore, SRAI can be used to illustrate the spectral curve identification feature.
Spectral absorption is characterized by the valley (M) and its corresponding peaks (S1 and S2), or the peak and its corresponding valleys valleys as shown in Figure 1. H represents the spectral absorption depth, and is the distance between the non-absorption baseline formed by valley (M) and its corresponding peaks. Let us suppose that ρ S 1 , ρ S 2 , and ρ M are the reflectances of S1, S2 and M, respectively, and λ S 1 , λ S 2 and λ M are the wavelengths of S1, S2, and M, respectively. Thus, the width of absorption band, asymmetry parameter, and the reflectance rate of peak can be respectively obtained using the expressions:
W = λ S 2 λ S 1
d = λ S 2 λ M W
Δ ρ S = ρ S 2 ρ S 1
Therefore, the formula of the non-absorption baseline is:
W ρ Δ ρ S λ = ρ S 1 Δ ρ S λ S 1
Equation (10) expresses the spectral contribution and spectral behavior of ground features without the spectral absorption characteristics. In this case, SRAI can be defined as the ratio of the spectral value of absorption and the corresponding baseline, as expressed by:
S R A I = d ρ S 1 + ( 1 d ) ρ S 2 ρ M
Figure 1. Illustrations of SRAI (a) absorption and (b) reflection.
Figure 1. Illustrations of SRAI (a) absorption and (b) reflection.
Sensors 16 00152 g001
Figure 2 shows the spectral curves of tree and wheat from an AVIRIS hyperspectral image, from which 20 SRAI feature points (including 10 valley points and 10 peak points) are extracted.
Figure 2. The SRAI features extracted from the spectral curves of a (a) tree and (b) wheat.
Figure 2. The SRAI features extracted from the spectral curves of a (a) tree and (b) wheat.
Sensors 16 00152 g002

3.2. Classical DP Algorithm

Hundreds of bands are involved in calculating similarities among hyperspectral images. One possible way to reduce the computation load is to simplify the spectral vector. The DP algorithm, vertical distance algorithm, and Li-Openshaw algorithm are among the most common methods used, of which the DP algorithm is the most popular. The basic principles of using the DP algorithm are listed below:
(1)
Connect both ends of the curve with a straight line and calculate the distances (d) from all the points on curve to the line.
(2)
Compare the maximum distance (dmax) and threshold D; if dmax < D, then eliminate all points on the curve; otherwise, retain the point with the maximum distance (dmax) and split the curve into two parts;
(3)
Repeat steps 1 and 2 until dmax < D is true for all points on the curve, and all the retained points comprise the final simplified spectral curve.
Figure 3 illustrates the simplified spectral curve using the DP algorithm with different thresholds (D). As can be seen, as the threshold increases, more points are removed and the spectral curve becomes more simplified.
Figure 3. The simplified spectral curve using the DP algorithm with different thresholds. (a) The original spectral curve with 183 points, (b) D = 0.05, 30 retained points, (c) D = 0.1, 23 retained points, and (d) D = 0.2, 13 retained points.
Figure 3. The simplified spectral curve using the DP algorithm with different thresholds. (a) The original spectral curve with 183 points, (b) D = 0.05, 30 retained points, (c) D = 0.1, 23 retained points, and (d) D = 0.2, 13 retained points.
Sensors 16 00152 g003
The DP algorithm is able to preserve the curve pattern feature; however, recursion is needed for processing, which further complicates the process. This process also has three main disadvantages, which are listed below:
(1)
The threshold must be specified as it can significantly affect the performance of curve simplification. On the one hand, if the threshold becomes higher, fewer points are retained, which may destroy the pattern. On the other hand, if the threshold becomes lower, more points are retained; therefore, less simplification is achieved.
(2)
Knowing how many points are left using a threshold before computing is impossible; in other words, setting a proper threshold to maintain the feature of a spectral curve using the necessary points is a difficult and complex task.
(3)
For the effective simplification of multiple curves, particularly for hyperspectral images, different thresholds are needed for various curves; moreover, to retain enough points, every curve should set different thresholds.

3.3. IDP Algorithm

To address the problems mentioned above, this paper proposes an improved version of the DP algorithm, which we call IDP. This algorithm does not require setting the threshold but is able to preserve the curve pattern. The major steps are listed below:
(1)
Connect the starting point S and ending point E, calculate the distance from each point to segment SE, and retain the point M, which has the maximum distance; this is the same as step one in the traditional DP algorithm.
(2)
Connect SM and ME, calculate the distance from each point to segment SM and ME, and retain the point N, which has the maximum distance.
(3)
Divide the curve into three parts using points M and N. Repeat step (2) until it meets the point number retaining requirement.
(4)
If the distances are equal in step (2), calculate the ratio of the distance to the line sector and the length of sector, the point with the lower ratio is then retained.
In other words, in every iteration, only one point with maximum distance is retained, and once the total number of retained points reach the specified number of points, the iteration stops. Hence, the problems mentioned in the above section could be solved without the threshold. Figure 4 shows the simplified results using different numbers of points. As shown in the figure, the proposed IDP algorithm can well maintain the feature pattern of the spectral curve. Comparing Figure 4b and Figure 3d, we find that the retained points of IDP are the same as those of the traditional DP algorithm, thus indicating the robust results of the proposed IDP.
Figure 4. The simplified spectral curves using the IDP algorithm with different numbers. (a) The original spectral curve with 183 points, (b) 13 retained points, (c) 40 retained points, and (d) 70 retained points.
Figure 4. The simplified spectral curves using the IDP algorithm with different numbers. (a) The original spectral curve with 183 points, (b) 13 retained points, (c) 40 retained points, and (d) 70 retained points.
Sensors 16 00152 g004aSensors 16 00152 g004b
By specifying the number of retained points, the algorithm can lead to convergence and ensure effective simplification. At the same time, as the threshold is not required, the algorithm can adapt to different curve patterns.

3.4. The IDP Algorithm under SRAI-Restriction

As described in Section 3.1, the SRAI points are essential for representing the basic features of the spectral curve; hence they must be preserved. The IDP algorithm can retain the specified number of points. Combining these two requirements, we propose the IDP algorithm under SRAI-restraint, which consists of the following steps:
(1)
Specify the number of SRAI points and then calculate the SRAI feature points.
(2)
Specify the number of all the retained points and then set the SRAI points as the initial points of the IDP algorithm. Finally, run the IDP algorithm to determine the remaining points.
Figure 5 presents the 20 SRAI feature points extracted from the spectral curve. Using the improved algorithm, retention of 30, 50 and 70 feature points are achieved, as shown in Figure 5b–d, respectively, so compared with Figure 5a, the improved algorithm can better preserve the overall pattern of the curve.
Figure 5. The results of the IDP algorithm under SRAI-restriction with10 peak and 10 valley points. (a) The original spectral curve, (b) 30 retained points, (c) 50 retained points, and (d) 70 retained points.
Figure 5. The results of the IDP algorithm under SRAI-restriction with10 peak and 10 valley points. (a) The original spectral curve, (b) 30 retained points, (c) 50 retained points, and (d) 70 retained points.
Sensors 16 00152 g005aSensors 16 00152 g005b

3.5. SRAI for the Hyperspectral Curves

For hyperspectral curve similarity assessment, the spectral curves are first simplified by SRAI-restrained IDP algorithm, after which the similarity of the simplified curves are compared with the same number of feature points. The calculation of similarity assessment index consists of two steps: band matching and distance calculation:
(1)
For the simplified spectral curves A and B, for a point Ai on curve A, if a point Bj on curve B has the same band number as Ai, then Bj is matched with Ai. Repeat this procedure until all points on curve A have been processed. Let variable N denote the total number of matched points.
(2)
For every matched point, calculate the ED between the reflectance of every matched point, and obtain the sum of all the distances as the final distance.
The final similarity assessment index is thus defined as:
S i m = ( C N ) 2 × 1 N i N ( R A _ i R B _ i ) 2
where N is the number of matching points and C is the number of all feature points; R A _ i and R B _ i are the reflectance of the matched points Ai and Bi, respectively; and C N is the punishment coefficient. This means that fewer matched points result in higher C N , thus leading to lower similarity. Hence, the lower value of Sim indicates the higher similarity of two curves.

4. Experiment and Analysis

To test the performance of the proposed similarity assessment index, two datasets, i.e., AVIRIS dataset and ROSIS dataset were used, and the typical indices, i.e., correlation coefficient (CC), ED, first-order approximation of Kullback-Leibler divergence (KL), SID and SAM, were compared. The mathematical expression of KL is given by:
K L = i = 1 N [ r 1 ( i ) r 2 ( i ) ] 2 r 1 ( i ) + r 2 ( i )
where r1 and r2 are the two spectrum curves, respectively.

4.1. Airborne Visible Infrared Imaging Spectrometer (AVIRIS) Dataset

The AVIRIS hyperspectral image from the Purdue University Remote Sensing Image Processing Lab was used for our experiments. The original image was captured in June 1992 in an experiment field located in Indiana, USA. The image had 224 bands, the spectral range was from 0.4 to 2.45 μm, the spatial resolution was 20 m, and the image size was 145 × 145 pixels. A total of 183 bands were used for the experiment after removing water absorption and noise bands. Figure 6a presents the RGB composite image of bands 97, 37, and 7.
Figure 6. RGB color composite AVIRIS image. (a) True color image and (b) mean spectral curve of three types of land cover.
Figure 6. RGB color composite AVIRIS image. (a) True color image and (b) mean spectral curve of three types of land cover.
Sensors 16 00152 g006
According to the ground truth, three cover types [i.e., grass (A), wheat (B), and forest (C)] were chosen for the similarity assessment. For simplicity, three pixels of every type were selected for the experiment. Figure 6b shows the similar mean spectral curves of the three types. The evaluation results of our method were compared with those obtained using the CC, ED, SID, KL and SAM methods. The results are listed in Table 1, Table 2, Table 3, Table 4, Table 5 and Table 6, where A1, A2 and A3 correspond to grass; B1, B2 and B3 correspond to wheat; and C1, C2 and C3 correspond to forest. A total of 20 SRAI feature points were exacted and then used to retain 50 points on the spectral curve. The bold back values are the best rank results. In addition, the rank of similarity are listed in the last column. For example, looking at the horizontal view of A3, the similarity between A3 and B2 (0.9977) is bigger than that between A3 and A2 (0.9960), which is obviously wrong (Table 1). Meanwhile, for the rank view of A3, the rank is “A3 > A1 > B2,” where B2 should not be in the third order, thus indicating an error.
Table 1. The performance of CC for the AVIRIS dataset.
Table 1. The performance of CC for the AVIRIS dataset.
CC
A1A2A3B1B2B3C1C2C3Rank
A110.99800.99820.99790.99770.99670.99640.99690.9968A1, A3, A2
A20.998010.99600.99830.99720.99620.99690.99760.9968A2, B1, A1
A30.99820.996010.99670.99770.99730.99630.99680.9968A3, A1, B2
B10.99790.99830.996710.99830.99680.99820.99830.9982B1, B2, C1/C2
B20.99770.99720.99770.998310.99820.99710.99780.9978B2, B1, B3
B30.99670.99620.99730.99680.998210.99620.99690.9965B3, B2, A3
C10.99640.99690.99630.99820.99710.996210.99940.9993C1, C2, C3
C20.99690.99760.99680.99830.99780.99690.999410.9993C2, C1, C3
C30.99680.99680.99680.99820.99780.99650.99930.99931C3, C1, C2
Table 2. The performance of ED for the AVIRIS dataset.
Table 2. The performance of ED for the AVIRIS dataset.
ED
A1A2A3B1B2B3C1C2C3Rank
A100.37390.32800.92410.81881.03861.34151.33671.2500A1,A3,A2
A20.373900.30880.62760.34710.80561.04581.04170.9627A2,A3,B2
A30.32800.308800.69420.57210.79331.08701.08150.9946A3,A2,A1
B10.92410.62760.694200.24990.42510.47840.48230.4068B1,B2,C3
B20.81880.34710.57210.249900.45700.60840.59830.4155B2,B1,A2
B31.03860.80560.79330.42510.457000.44600.51340.4949B3,B1,C1
C11.34151.04581.08700.47840.60840.446000.16180.1800C1,C2,C3
C21.33671.04171.08150.48230.59830.51340.161800.1874C2,C1,C3
C31.25000.96270.99460.40680.41550.49490.18000.18740C3,C1,C2
Table 3. The performance of SID for the AVIRIS dataset.
Table 3. The performance of SID for the AVIRIS dataset.
SID
A1A2A3B1B2B3C1C2C3Rank
A100.00870.00660.01540.01830.01010.01590.01090.0147A1,A3,A2
A20.008700.00820.00850.01140.01410.01050.00640.0117A2,A3,C2
A30.00660.008200.01100.01260.01170.01120.00830.0109A3,A1,A2
B10.01540.00850.011000.00780.02100.00590.00690.0076B1,C1,C2
B20.01830.01140.01260.007800.02140.01000.00850.0075B2,C3,B1
B30.01010.01410.01170.02100.021400.02190.01500.0214B3,A1,A3
C10.01590.01050.01120.00590.01000.021900.00710.0078C1,B1,C2
C20.01090.00640.00830.00690.00850.01500.007100.0076C2,A2,B1
C30.01470.01170.01090.00760.00750.02140.00780.00760C3,B2,B1/C2
Table 4. The performance of KL for the AVIRIS dataset.
Table 4. The performance of KL for the AVIRIS dataset.
KL
A1A2A3B1B2B3C1C2C3Rank
A100.24410.20800.81860.76341.33211.42621.42271.2981A1,A3,A2
A20.244100.21090.45120.23231.12030.97070.96560.9177A2,A3,B2
A30.20800.210900.48350.43730.98660.92970.95180.8453A3,A1,A2
B10.81860.45120.483500.22070.74100.29340.36870.3044B1,B2,C1
B20.76340.23230.43730.220700.84170.48770.50640.3786B2,B1,A2
B31.33211.12030.98660.74100.841700.65290.45340.6356B3,C1,C2
C11.42620.97070.92970.29340.48770.652900.12720.1388C1,C2,C3
C21.42270.96560.95180.36870.50640.45340.127200.1473C2,C1,C3
C31.29810.91770.84530.30440.37860.63560.13880.14730C3,C1,C2
Table 5. The performance of SAM for the AVIRIS dataset.
Table 5. The performance of SAM for the AVIRIS dataset.
SAM
A1A2A3B1B2B3C1C2C3Rank
A110.99880.99920.99810.99800.99860.99720.99790.9976A1,A3,A2
A20.998810.99820.99910.99860.99770.99830.99890.9984A2,B1,C2
A30.99920.998210.99800.99850.99860.99770.99820.9981A3,A1,B3
B10.99810.99910.998010.99910.99750.99920.99930.9992B1,C2,C1/C3
B20.99800.99860.99850.999110.99770.99870.99900.9990B2,B1,C2/C3
B30.99860.99770.99860.99750.997710.99660.99750.9970B3,A1,A3
C10.99720.99830.99770.99920.99870.996610.99900.9987C1,B1,C2
C20.99790.99890.99820.99930.99900.99750.999010.9991C2,B1,C3
C30.99760.99840.99810.99920.99900.99700.99870.99911C3,B1,C2
Table 6. The performance of the proposed index for the AVIRIS dataset.
Table 6. The performance of the proposed index for the AVIRIS dataset.
Proposed
A1A2A3B1B2B3C1C2C3Rank
A100.06370.07480.14950.19600.21510.31670.22710.2295A1,A2,A3
A20.063700.05680.12020.13430.15160.18290.16780.1838A2,A3,A1
A30.07480.056800.17970.14120.24170.18190.21160.2304A3,A2,A1
B10.14950.12020.179700.05180.07260.11010.10010.0934B1,B2,B3
B20.19600.13430.14120.051800.08830.15090.14450.1443B2,B1,B3
B30.21510.15160.24170.07260.088300.11220.12420.0966B3,B1,B2
C10.31670.18290.18190.11010.15090.112200.02930.0412C1,C2,C3
C20.22710.16780.21160.10010.14450.12420.029300.0461C2,C1,C3
C30.22950.18380.23040.09340.14430.09660.04120.04610C3,C1,C2
From Figure 6b, we can see that these three types can be easily confused with one another. For SID results, there are 11 confusing pairs (i.e., A2 and C2, B1 and C1, B1 and C2, B2 and C3, B3 and A1, B3 and A3, C1 and B1, C2 and A2, C2 and B1, C3 and B2, C3 and B1), which can also be seen from the rank column. For the SAM results, there are 13 confusing pairs (i.e., A2 and B1, A2 and C2, A3 and B3, B1 and C2, B1 and C1/C3, B2 and C2/C3, B3 and A1, B3 and A3, C1 and B1, C2 and B1, C3 and B1). For the CC, ED and KL indices, there are five, four and five confusing pairs, respectively. In comparison, for the proposed index, all the types are clearly distinguished, which is also reflected in both the similarity assessment values and ranking. The results indicate that the proposed index is effective in distinguishing different vegetation types with similar spectra.

4.2. Reflective Optics System Imaging Spectrometer (ROSIS) Dataset

The ROSIS dataset was acquired using Reflective Optics System Imaging Spectrometer-DLR (ROSIS_03), Schneider Systemtechnik-HySens Pavia Campaign (2002). The data set acquired from the ROSIS sensor does not fully cover the areas of interest because of the narrower field of view of the DAIS instrument on which the flight lines have been designed. Data were atmospherically corrected but not geometrically corrected. The number of spectral bands was 103 for the “University” dataset, and geometric resolution was 1.3 meters. The images were acquired during a flight campaign over Pavia, in northern Italy (45°11’ N, 9°9’ E), on the 8th of July 2002 from 10:30 a.m. to 12:00 noon. Figure 7a presents the RGB composite image.
Figure 7. RGB color composite ROSIS image. (a) True color image and (b) mean spectral curve of three types of land cover.
Figure 7. RGB color composite ROSIS image. (a) True color image and (b) mean spectral curve of three types of land cover.
Sensors 16 00152 g007
According to the ground truth land cover map, three cover types, i.e., Asphalt (A), Self-Blocking Bricks (B), and Meadows (C), on the image were chosen for the similarity assessment. For simplicity, three pixels of every type were selected for the experiment. Figure 7b shows the mean spectral curves of the three types, from which it can be observed that Asphalt and Self-Blocking Bricks are similar while Meadows is different. The results of SID, SAM, and the proposed index are listed in Table 4, Table 5 and Table 6, respectively, in which A1, A2 and A3 denote Asphalt; B1, B2 and B3 denote Self-Blocking Bricks; and C1, C2 and C3 denote Meadows. A total of 50 points were retained on the spectral curve, and 10 SRAI feature points were selected. The bold back values show the best ranking results. In addition, the similarity rankings are listed in the last column.
Similar to Figure 6b, in Figure 7b, the spectral curves of cover types A and B can be easily confused with each other. In Table 7, Table 8 and Table 9 there are nine confusing pairs in total. Table 10 shows three confusing pairs (A2 and B2, A3 and B1, B2 and A2), while Table 11 shows five confusing pairs (A2 and B3, A3 and B2, B1 and A2, B2 and A3, B3 and A2). These results are also shown in the ranking columns. In comparison, in Table 12, all cover types are clearly distinguished, indicating that the proposed index shows better performance compared with CC, ED, KL, SID and SAM.
Table 7. The performance of CC for the ROSIS dataset.
Table 7. The performance of CC for the ROSIS dataset.
CC
A1A2A3B1B2B3C1C2C3Rank
A11.0000 0.9775 0.9798 0.9024 0.9156 0.9054 0.9441 0.9376 0.9437 A1,A3,A2
A20.9775 1.0000 0.9645 0.9570 0.9673 0.9579 0.9362 0.9327 0.9382 A2,A1,B2
A30.9798 0.9645 1.0000 0.9560 0.9643 0.9593 0.9392 0.9350 0.9406 A3,A1,A2
B10.9024 0.9570 0.9560 1.0000 0.9524 0.9961 0.9128 0.9152 0.9176 B1,B3,A2
B20.9156 0.9673 0.9643 0.9524 1.0000 0.9937 0.9243 0.9264 0.9278 B2,B3,A2
B30.9054 0.9579 0.9593 0.9961 0.9937 1.0000 0.9202 0.9223 0.9245 B3,B1,B2
C10.9441 0.9362 0.9392 0.9128 0.9243 0.9202 1.0000 0.9991 0.9994 C1,C3,C2
C20.9376 0.9327 0.9350 0.9152 0.9264 0.9223 0.9991 1.0000 0.9994 C2,C3,C1
C30.9437 0.9382 0.9406 0.9176 0.9278 0.9245 0.9994 0.9994 1.0000 C3,C1,C2
Table 8. The performance of ED for the ROSIS dataset.
Table 8. The performance of ED for the ROSIS dataset.
ED
A1A2A3B1B2B3C1C2C3Rank
A10.0000 0.19230.48870.34290.41390.45011.57702.00891.6241A1,A2,B1
A20.19230.0000 0.07870.27050.35970.38671.49221.93601.5395A2,A3,A1
A30.48870.07870.0000 0.31120.40910.43141.53481.98201.5831A3,A2,B1
B10.34290.27050.31120.0000 0.34440.14911.37851.79781.4199B1,B3,A2
B20.41390.35970.40910.34440.0000 0.09041.30431.71011.3443B2,B3,B1
B30.45010.38670.43140.14910.09040.0000 1.27871.68471.3175B3,B2,B1
C11.57701.49221.53481.37851.30431.27870.0000 0.48710.0842C1,C3,C2
C22.00891.93601.98201.79781.71011.68470.48710.0000 0.4313C2,C3,C1
C31.62411.53951.58311.41991.34431.31750.08420.43130.0000 C3,C1,C2
Table 9. The performance of SID for the ROSIS dataset.
Table 9. The performance of SID for the ROSIS dataset.
SID
A1A2A3B1B2B3C1C2C3Rank
A10.0000 0.0165 0.0190 0.0468 0.0456 0.0594 0.5079 0.5212 0.5029 A1,A2,A3
A20.0165 0.0000 0.0020 0.0166 0.0138 0.0224 0.3945 0.4035 0.3859 A2,A3,B2
A30.0190 0.0020 0.0000 0.0018 0.0164 0.0234 0.4069 0.4172 0.3992 A3,A2,B1
B10.0468 0.0166 0.0018 0.0000 0.0040 0.0028 0.3804 0.3839 0.3682 B1,B3,B2
B20.0456 0.0138 0.0164 0.0040 0.0000 0.0046 0.3675 0.3703 0.3568 B2,A2,B1
B30.0594 0.0224 0.0234 0.0028 0.0046 0.0000 0.3516 0.3542 0.3395 B3,B2,B1
C10.5079 0.3945 0.4069 0.3804 0.3675 0.3516 0.0000 0.0035 0.0028 C1,C3,C2
C20.5212 0.4035 0.4172 0.3839 0.3703 0.3542 0.0035 0.0000 0.0029 C2,C3,C1
C30.5029 0.3859 0.3992 0.3682 0.3568 0.3395 0.0028 0.0029 0.0000 C3,C1,C2
Table 10. The performance of KL for the ROSIS dataset.
Table 10. The performance of KL for the ROSIS dataset.
KL
A1A2A3B1B2B3C1C2C3Rank
A10.00000.15420.15400.40200.49140.59964.71236.05484.8043A1,A3,A2
A20.15420.00000.02030.36060.08280.42343.92585.31173.9991A2,A3,B1
A30.15400.02030.00000.09550.46060.50814.10375.56364.1915A3,A2,A1
B10.40200.36060.09550.00000.09160.05423.57614.65013.5883B1,B3,A2
B20.49140.08280.46060.09160.00000.04113.41304.32123.4172B2,B3,B1
B30.59960.42340.50810.05420.04110.00003.27014.17493.2621B3,B2,B1
C14.71233.92584.10373.57613.41303.27010.00000.30480.0321C1,C3,C2
C26.05485.31175.56364.65014.32124.17490.30480.00000.2392C2,C3,C1
C34.80433.99914.19153.58833.41723.26210.03210.23920.0000C3,C1,C2
Table 11. The performance of SAM for the ROSIS dataset.
Table 11. The performance of SAM for the ROSIS dataset.
SAM
A1A2A3B1B2B3C1C2C3Rank
A11.00000.99320.99470.9841 0.9845 0.9805 0.8631 0.8601 0.8628 A1,A3,A2
A20.9932 1.00000.99930.9950 0.9954 0.99910.9009 0.8985 0.9011 A2,A3,B3
A30.9947 0.99931.00000.9940 0.99890.9930 0.8967 0.8941 0.8968 A3,A2,B2
B10.9841 0.9950 0.9940 1.00000.99480.9933 0.9014 0.9006 0.9025 B1,A2,B2
B20.9845 0.9954 0.99890.9948 1.00000.99890.9063 0.9054 0.9070 B2,B3,A3
B30.9805 0.99910.9930 0.9933 0.99891.00000.9114 0.9106 0.9124 B3,A2,B2
C10.8631 0.9009 0.8967 0.9014 0.9063 0.9114 1.00000.99960.9997C1,C3,C2
C20.8601 0.8985 0.8941 0.9006 0.9054 0.9106 0.99961.00000.9997C2,C3,C1
C30.8628 0.9011 0.8968 0.9025 0.9070 0.9124 0.99970.99971.0000C3,C1,C2
Table 12. The performance of the proposed index for the ROSIS dataset.
Table 12. The performance of the proposed index for the ROSIS dataset.
Proposed
A1A2A3B1B2B3C1C2C3Rank
A10.00000.08290.06220.1473 0.1192 0.2132 0.5461 0.9955 0.6399 A1,A3,A2
A20.08290.00000.03390.1180 0.1333 0.1645 0.4875 1.0929 0.5742 A2,A3,A1
A30.06220.03390.00000.1149 0.1514 0.1349 0.6716 0.6859 0.4885 A3,A2,A1
B10.1473 0.1180 0.1149 0.00000.05100.07360.6502 0.7601 0.4397 B1,B2,B3
B20.1192 0.1333 0.1514 0.05100.00000.05010.5167 0.5091 0.5821 B2,B3,B1
B30.2132 0.1645 0.1349 0.07360.05010.00000.4990 0.5677 0.3716 B3,B2,B1
C10.5461 0.4875 0.6716 0.6502 0.5167 0.4990 0.00000.18520.0376C1,C3,C2
C20.9955 1.0929 0.6859 0.7601 0.5091 0.5677 0.18520.00000.1232C2,C3,C1
C30.6399 0.5742 0.4885 0.4397 0.5821 0.3716 0.03760.12320.0000C3,C1,C2

4.3.The Impacts of the Parameters

There are two parameters, i.e., the number of retained points and the number of SRAI feature points, in the previous experiments. In this experiment, the impacts of these two parameters were tested. The mean spectral curves of the AVIRIS dataset used in the first experiment (A, B, and C) were used, after which the similarity index using different numbers of retained points and number of SRAI feature points were calculated. When testing the impact of the number of retained points, the number of SRAI feature points was kept at 10, and the number of retained points increased from 30 to 90. In every increasing step, the similarities of A and B, B and C, and C and A were calculated. Meanwhile, when testing the impact of number of SRAI feature points, the number of retained points was kept at 50, whereas the number of SRAI feature points increased from 5 to 20. The above three similarities were calculated. The results are shown in Figure 8 and Figure 9, respectively. As shown in the figures, the x axis indicates the number of retained points or SRAI feature points, and the y axis shows the similarities among A, B, and C.
Figure 8. The impacts of parameters on the AVIRIS dataset. (a) Impact on the number of SRAI feature points and (b) on the number of retained points.
Figure 8. The impacts of parameters on the AVIRIS dataset. (a) Impact on the number of SRAI feature points and (b) on the number of retained points.
Sensors 16 00152 g008
Analogous to the AVIRIS dataset in Figure 8, the ROSIS dataset was also used to demonstrate the impacts of the two parameters of the proposed method. The results are shown in Figure 9.
Figure 9. The impacts of parameters on the ROSIS dataset. (a) Impact on the number of SRAI feature points and (b) on the number of retained points.
Figure 9. The impacts of parameters on the ROSIS dataset. (a) Impact on the number of SRAI feature points and (b) on the number of retained points.
Sensors 16 00152 g009
From Figure 8 and Figure 9, we can see that a higher number of retained points does not necessarily imply a higher number of matched points; hence, the value of the similarity index may be different. No matter how the similarity index values changed, the similarity trends were coincident, as shown in Figure 8 and Figure 9. These results indicated that for practical use, any number of SRAI feature points or retained points could be set. However, as a more proper choice, we can choose a point at which the difference between similarities is the largest.

5. Conclusions

Spectral similarity assessment is a basic issue in hyperspectral image classification and object recognition. As the most intuitive way to express the spectral features, the spectrum curve and its features have always been investigated in the literature. Given that a hyperspectral image usually contains hundreds of bands, reducing the dimensions of a hyperspectral image is an important research topic.
As a solution, the DP algorithm can be used in a single spectrum curve and, therefore, is more suitable for single spectrum curve similarity comparison. However, given that there is a distance threshold in the traditional DP algorithm, it is difficult to determine the most proper threshold, which can retain enough points to complete the similarity assessment.
This paper proposes an improved DP algorithm to solve this problem. After setting the number of retained points, the proposed method could automatically extract the expected points without using the threshold. This algorithm can consistently achieve convergence, resulting in a higher number of extracted points that, in turn, facilitates a more effective similarity assessment. As the reflection and absorption features are important for spectral recognition, SRAI is introduced to restrain the simplification process. Finally, enough points (including the important SRAI features) are obtained for the proposed method.
To test the performance of the proposed method, two widely used hyperspectral image datasets, i.e., AVIRIS and ROSIS, were used. The assessment results were compared with the results of existing indices, i.e., CC, ED, SID, SAM, and KL. The experiment results indicate that the existing indices may lead to obvious incorrect results, especially when the testing curves are similar. In comparison, our proposed method could achieve high accuracy using fewer points, because the most import SRAI features are retained. Finally, the impacts of the number of SRAI points and the number of all retained points have been analyzed. Further research may focus on investigating the practicality of this method.

Acknowledgments

This work is jointly supported by the National Natural Science Foundation program (No. 41301403, 41471340 and 61301255) and the Basic Research Program of Shenzhen (No. JCYJ20150831194441446, JCYJ20150630114942312 and JCYJ20150630114942260); Chongqing Basic and Advanced Research General Project (No. cstc2013jcyjA40010), Hunan Provincial Natural Science Foundation of China (No. 13JJ4039), China Postdoctoral Science Foundation (NO. 2013M531782 and 2014T70768), the Research Grants Council (RGC) of Hong Kong General Research Fund (GRF) (Project No. 203913), and Hong Kong Baptist University Faculty Research Grant (FRG1/12-13/070 and FRG2/11-12/030).

Author Contributions

Jun Liu is the corresponding author who theoretically proposed the whole method; Dan Ma implemented and improved the method and wrote the whole paper; Junyi Huang and Ping Liu helped to implement the experiments and tested method on various hyperspectral remote sensing images, Huali Li, Huijuan Chen and Jing Qian gave much contributions for revising the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Van der Woerd, H.J.; Wernand, M.R. True Colour Classification of Natural Waters with Medium-Spectral Resolution Satellites: SeaWiFS, MODIS, MERIS and OLCI. Sensors 2015, 15, 25663–25680. [Google Scholar] [CrossRef] [PubMed]
  2. Yang, X.L.; Hong, H.M.; You, Z.H.; Cheng, F. Spectral and Image Integrated Analysis of Hyperspectral Data for Waxy Corn Seed Variety Classification. Sensors 2015, 15, 15578–15594. [Google Scholar] [CrossRef] [PubMed]
  3. Zhang, X.L.; Liu, F.; He, Y.; Li, X.L. Application of Hyperspectral Imaging and Chemometric Calibrations for Variety Discrimination of Maize Seeds. Sensors 2012, 12, 17234–17246. [Google Scholar] [CrossRef] [PubMed]
  4. Stratoulias, D.; Balzter, H.; Sykioti, O.; Zlinszky, A.; Tóth, V.R. Evaluating Sentinel-2 for Lakeshore Habitat Mapping Based on Airborne Hyperspectral Data. Sensors 2015, 15, 22956–22969. [Google Scholar] [CrossRef] [PubMed]
  5. Gmur, S.; Vogt, D.; Zabowski, D.; Moskal, L.M. Hyperspectral Analysis of Soil Nitrogen, Carbon, Carbonate, and Organic Matter Using Regression Trees. Sensors 2012, 12, 10639–10658. [Google Scholar] [CrossRef] [PubMed]
  6. Kaarna, A.; Zemcik, P.; Kalviainen, H. Compression of multispectral remote sensing images using clustering and spectral reduction. IEEE Trans. Geosci. Remote Sens. 2000, 38, 1588–1592. [Google Scholar] [CrossRef]
  7. Du, Q.; Fowler, J.E. Low-complexity principal component analysis for hyperspectral image compression. Int. J. High Perform. Comput. Appl. 2008, 22, 438–448. [Google Scholar] [CrossRef]
  8. Wang, J.; Chang, C.I. Independent component analysis-based dimensionality reduction with applications in hyperspectral image analysis. IEEE Trans. Geosci. Remote Sens. 2006, 44, 1586–1600. [Google Scholar] [CrossRef]
  9. Guo, B.; Gunn, S.R.; Damper, R.I. Band selection for hyperspectral image classification using mutual information. IEEE Geosci. Remote Sens. Lett. 2006, 3, 522–526. [Google Scholar] [CrossRef]
  10. Jimenez, L.O.; Landgrebe, D.A. Hyperspectral data analysis and supervised feature reduction via projection pursuit. IEEE Trans. Geosci. Remote Sens. 1999, 37, 2653–2667. [Google Scholar] [CrossRef]
  11. Banskota, A.; Wynne, R.H.; Kayastha, N. Improving within-genus tree species discrimination using the discrete wavelet transform applied to airborne hyperspectral data. Int. J. Remote Sens. 2011, 32, 3551–3563. [Google Scholar] [CrossRef]
  12. Amato, U.; Cavalli, R.M.; Palombo, A. Experimental approach to the selection of the components in the minimum noise fraction. IEEE Trans. Geosci. Remote Sens. 2009, 47, 153–160. [Google Scholar] [CrossRef]
  13. Licciardi, G.A.; Frate, F.D. A comparison of feature extraction methodologies applied on hyperspectral data. In Proceedings of ESA Hyperspectral 2010 Workshop, Frascati, Italy, 17–19 March 2010.
  14. Jia, X.; Riehards, J.A. Segmented prineipal components transformation for efficient hyperspectral remote sensing image display and classification. IEEE Trans. Geosci. Remote Sens. 1999, 37, 538–540. [Google Scholar]
  15. Maghsoudi, Y.; Zoej, M.J.V.; Collins, M. Using class-based feature selection for the classification of hyperspectral data. Int. J. Remote Sens. 2011, 32, 4311–4326. [Google Scholar] [CrossRef]
  16. Goel, P.K.; Prasher, S.O.; Patel, R.M. Classification of hyperspectral data by decision trees and artificial neural networks to identify weed stress and nitrogen status of corn. Comput. Electron. Agric. 2003, 39, 67–93. [Google Scholar] [CrossRef]
  17. Tarabalka, Y.; Fauvel, M.; Chanussot, J. SVM- and MRF-based method for accurate classification of hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2010, 7, 736–740. [Google Scholar] [CrossRef]
  18. Kruse, F.A.; Lefkoff, A.B.; Boardman, J.W. The spectral image processing system (SIPS)–interactive visualization and analysis of imaging spectrometer data. Remote Sens. Environ. 1993, 44, 145–163. [Google Scholar] [CrossRef]
  19. Liu, W.; Chang, Q.R.; Guo, M. Extraction of first derivative spectrum features ofsoil organic matter via wavelet de-noising. Spectrosc. Spectr. Anal. 2011, 31, 100–104. [Google Scholar]
  20. Pluim, J.P.W.; Maintz, J.B.A.; Viergever, M. Mutual information based registration of medical images: A survey. IEEE Trans. Med. Imag. 2003, 22, 986–1004. [Google Scholar] [CrossRef] [PubMed]
  21. Chang, C.I. An information-theoretic approach to spectral variability, similarity, and discrimination for hyperspectral image analysis. IEEE Trans. Inf. Theory 2000, 46, 1927–1932. [Google Scholar] [CrossRef]
  22. Li, Q.L.; Xue, Y.Q.; Wang, J.Y. Automated tongue segmentation algorithm based on hyperspectral image. J. Infrared Millim. Waves 2007, 26, 77–80. [Google Scholar]
  23. Du, Y.Z.; Chang, C.; Ren, H. New hyperspectral discrimination measure for spectral characterization. Opt. Eng. 2004, 43, 1777–1786. [Google Scholar]
  24. Kumar, M.N.; Seshasai, M.V.R.; Prasad, K.S.V. A new hybrid spectral similarity measure for discrimination among Vigna species. Int. J. Remote Sens. 2011, 32, 4041–4053. [Google Scholar] [CrossRef]
  25. Kong, X.B.; Shu, N.; Tao, J.B. A new spectral similarity measure based on multiple features integration. Spectrosc. Spectr. Anal. 2011, 31, 2166–2170. [Google Scholar]
  26. Du, P.J.; Chen, Y.H.; Fang, T. Spectral feature-based hyperspectral RS image retrieval. Spectrosc. Spectr. Anal. 2005, 25, 1171–1175. [Google Scholar]
  27. Sun, H.; Li, M.Z.; Zhang, Y.E. Spectral characteristics of corn under different nitrogen treatments. Spectrosc. Spectr. Anal. 2010, 30, 715–719. [Google Scholar]
  28. Clark, M.L.; Roberts, D.A.; Clark, D.B. Hyperspectral discrimination of tropical rain forest tree species at leaf to crown scales. Remote Sens. Environ. 2005, 96, 375–398. [Google Scholar] [CrossRef]
  29. Bakker, W.H.; Schmidt, K.S. Hyperspectral edge filtering for measuring homo-geneity of surface cover types. ISPRS J. Photogramm. Remote Sens. 2002, 56, 246–256. [Google Scholar] [CrossRef]
  30. Du, Y.; Ives, R.; Etter, D.; Welch, T.; Chang, C.-I. A one-dimensional approach for iris recognition. In Proceedings of the SPIE Biometric Technology for Human Identification, Orlando, FL, USA, 12–16 April 2004; pp. 237–247.

Share and Cite

MDPI and ACS Style

Ma, D.; Liu, J.; Huang, J.; Li, H.; Liu, P.; Chen, H.; Qian, J. Spectral Similarity Assessment Based on a Spectrum Reflectance-Absorption Index and Simplified Curve Patterns for Hyperspectral Remote Sensing. Sensors 2016, 16, 152. https://doi.org/10.3390/s16020152

AMA Style

Ma D, Liu J, Huang J, Li H, Liu P, Chen H, Qian J. Spectral Similarity Assessment Based on a Spectrum Reflectance-Absorption Index and Simplified Curve Patterns for Hyperspectral Remote Sensing. Sensors. 2016; 16(2):152. https://doi.org/10.3390/s16020152

Chicago/Turabian Style

Ma, Dan, Jun Liu, Junyi Huang, Huali Li, Ping Liu, Huijuan Chen, and Jing Qian. 2016. "Spectral Similarity Assessment Based on a Spectrum Reflectance-Absorption Index and Simplified Curve Patterns for Hyperspectral Remote Sensing" Sensors 16, no. 2: 152. https://doi.org/10.3390/s16020152

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop