Next Article in Journal
A Component-Based Model for Novel Modular Connections with Inbuild Component
Previous Article in Journal
At the End of a Slippery Slope: A Pilot Study of Deceleration Mats for Snow Tubing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Autonomous Learning Interactive Features for Hyperspectral Remotely Sensed Data

School of Geomatics Science and Technology, Nanjing Tech University, Nanjing 211816, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(21), 10502; https://doi.org/10.3390/app112110502
Submission received: 5 October 2021 / Revised: 28 October 2021 / Accepted: 4 November 2021 / Published: 8 November 2021

Abstract

:
In the field of remote sensing, most of the feature indexes are obtained based on expert knowledge or domain analysis. With the rapid development of machine learning and artificial intelligence, this method is time-consuming and lacks flexibility, and the indexes obtained cannot be applied to all areas. In order to not rely on expert knowledge and find the effective feature index with regard to a certain material automatically, this paper proposes a data-driven method to learn interactive features for hyperspectral remotely sensed data based on a sparse multiclass logistic regression model. The key point explicitly expresses the interaction relationship between original features as new features by multiplication or division operation in the logistic regression. Through the strong constraint of the L1 norm, the learned features are sparse. The coefficient value of the corresponding features after sparse represents the basis for judging the importance of the features, and the optimal interactive features among the original features. This expression is inspired by the phenomenon that usually the famous indexes we used in remote sensing, like NDVI, NDWI, are the ratio between different spectral bands, and also in statistical regression, the relationship between features is captured by feature value multiplication. Experiments were conducted on three hyperspectral data sets of Pavia Center, Washington DC Mall, and Pavia University. The results for binary classification show that the method can extract the NDVI and NDWI autonomously, and a new type of metal index is proposed in the Pavia University data set. This framework is more flexible and creative than the traditional method based on laboratory research to obtain the key feature and feature interaction index for hyperspectral remotely sensed data.

1. Introduction

The extraction of land cover information has always been a core and hot issue in the field of remote sensing [1]. Land-cover issues using remote sensing imagery enable us to recognize the physical composition and characteristics of different land-cover types. The remotely sensed data for image information identification and extraction is widely used in geospatial information services. Many researchers have contributed to this issue: Priem [2] proposed a synergistic workflow based on high-resolution APEX hyperspectral image to cope with the strong diversity of materials in urban areas, as well as with the presence of shadow; in order to make full use of high-resolution remote sensing imagery and hyperspectral imagery, Zhong [3] proposed a spatial-spectral-emissivity land-cover method based on the fusion of visible and hyperspectral imagery; Park [4] analyzed the discrepancy between the existing cadastral map and the actual land use on hyperspectral unmanned aerial vehicle (UAV) images. Land cover/use and its dynamic information is the basis for characterizing surface conditions, supporting land resource management and optimization, and assessing the impacts of climate change and human activities. Therefore, it has important guiding significance for surveys of land, as well as scientific management analysis [5], and it is a prerequisite for various tasks of land management.
The initial extraction of land use and land cover information is achieved through manual visual interpretation, which places high professional requirements on operators. However, the operation speed is slow, thus effective information cannot be obtained in time. The rapid development of computer technology and the abundance of multiple sources of remotely sensed data have led to the emergence of a large number of automatic identification and classification methods [6]. Among them, the method of extracting remotely sensed information based on the spectral index of interaction has been widely used due to its simplicity and speed [7].
The concept of interaction is widely used in the discussion of statistics [8]. In most cases, the discussion is based on the familiar linear and quadratic statistics in the analysis of variance and linear regression. Rothman gave the definition of statistical interaction. The term “statistical interaction” is intended to indicate the interaction dependence between the influence of two or more factors within the scope of a given risk model [9]. From a statistical point of view, Cox believes that if the individual effects of various factors do not add up, it can be said that an interaction has occurred, that is, the interaction is a special non-additive [10].
The common index of remotely sensed information is almost a kind of interaction between bands [11]. It is proposed based on ratio calculation and normalization processing. The purpose of ratio calculation is to eliminate the influence of terrain differences. The specific principle is to find the strongest and weakest reflection bands and place them in the positions of the numerator and denominator respectively. Through the ratio operation, the gap between the two can be further expanded in a geometric progression, to obtain the maximum brightness enhancement of the features on the image [12]. The Normalized Differential Vegetation Index (NDVI) proposed by Rouse uses the interaction between the red light (R) band and the near-infrared (NIR) band [13] to reflect the coverage of land vegetation; Liu [14] use the soil regulation vegetation index, the normalized building index and the modified normalized water body index for constructing a new type of vegetation index called IVI (Index-based Vegetation Index), which can better suppress the influence of building land and water body information; Mcfeeters is inspired by NDVI [15], using the interaction between the reflected near-infrared and visible green light to propose the Normalized Difference Water Index (NDWI) to describe the characteristics of open water and enhance its presence in remotely sensed digital images; Xu proposed the Modified Normalized Difference Water Index (MNDWI) on the basis of NDWI [16], which takes advantage of the interaction between the green light (G) band and the near-infrared (NIR) band to better extract urban water information; the Multi-Band Water Index (MBWI) based on the characteristics of the water body index and the difference in reflectance between water body and low-reflective surfaces was constructed by Wang [17], the research is based on the average spectral reflectance data of various types in the Qinhuai River Basin; Zha [18] used the interaction between the near-infrared band and the short-wave infrared band to propose the Normalized Difference Barren Index (NDBI), whose effect is to extract residential area information; in the study of the change of land coverage, Deng and Wu [19] considered the lack of direct relationship between soil abundance and its spectral characteristics, it is difficult to identify soil features using remote sensing technology, so they used the interaction between the green light (G) band and the short-wave infrared band. They proposed a Normalized Difference Soil Index (NDSI), which can well extract soil information from vegetation and impervious surfaces; in order to reduce the potential agricultural losses, a new drought index IDI (Integrated Agricultural Drought Index) [20] integrating multiple indicators is proposed by Liu, which can effectively monitor the occurrence of drought events. Therefore, the interaction between wavebands plays a key role in the extraction of feature information from remotely sensed images, but how to automatically learn effective interaction when extracting different types of features is still a question worth exploring.
The traditional method of extracting ground features through laboratory research and analysis is good for specific areas. However, it is difficult to directly apply to other areas because it often relies on expert experience and existing reference data [21]. With the development of machine learning and artificial intelligence, more and more researchers use machine learning methods to extract ground feature information. Yang proposed method of a neural network based on Back Propagation (BP) to extract water automatically [22], this method can better extract water without setting a threshold; Luo used the C5.0 decision tree method based on machine learning to extract wetland vegetation [23], and the effect was significant; Hong integrated water index and regional FCM clustering method to extract urban surface water body [24] so that urban surface water boundaries have good regional integrity and local details. At the same time, this method has a better suppression effect on the complex background noise of urban surface water body; Zhen proposed a feature extracting framework for hyperspectral data, which constructed a structured dictionary to encode spectral information and apply machine learning to map coding coefficients [25]; Pu used three machine-learning algorithms to extract information on rocky desertification in a typical karst region of southwest China, which provide important information on the total area and spatial distribution of different levels of rocky desertification in the study area to support decision making by local governments for sustainable development [26].
Based on the research background above, this paper proposes a data-driven method to learn interactive features for hyperspectral remotely sensed data based on sparse multiclass logistic regression (SMLR) model. SMLR is supported by a complete and rigorous mathematical theory and has strong interpretability. The interaction relationship between original features is expressed as new features by multiplication or division operation, which can be used as a new index for extracting ground feature information. Compared with the index obtained by laboratory research, the method of this article is creative and more flexible in exploring unknown ground feature index.

2. Materials and Methods

The key point of the method in this paper is explicitly to express the interaction relationship between original features as new features by multiplication or division operation in the logistic regression. This method automatically learns effective features based on data dimensionality reduction and normalization of hyperspectral remotely sensed images. Then the method uses these effective features as new indexes to extract ground features, subsequently compare them with the traditional indexes obtained by laboratory research. Meanwhile, a new type of metal index is proposed. Finally, the method of visual interpretation is used to evaluate the results, together with two indicators: OA (Overall Accuracy) and Recall. The overall technique flowchart is shown in Figure 1.

2.1. Hyperspectral Image Preprocessing

Hyperspectral remote sensing imagery has the characteristics of extremely high spectral resolution. At the same time, it also brings about the problem of high data dimension and a large amount of redundant information between adjacent bands [27]. Therefore, in order to avoid redundant information, data dimensionality reduction is required [28]. Considering that adjacent bands have similar spectral characteristics, the median value is calculated every 9 bands and used as the representative band, which greatly reduces the amount of experimental data. The pixel values of all bands of the image are normalized so that all pixel values are limited between 0–1. This is to eliminate the adverse effects of singular sample data and reduce training time. The normalized conversion function is as follows:
X = x min ( x ) max ( x ) min ( x )
Among them, x represents the pixel value in the original image, and X represents the pixel value in the normalized image.

2.2. Representation of Interaction between Features

For the problem of remote sensing feature extraction, considering the interaction between spectral features will enhance the independence between categories and improve the recognition accuracy of different categories. The most typical example is the NDVI, which uses the ratio of reflectivity between the R (red) band and the NIR (near-infrared) band to improve vegetation identification. Similarly, NDWI, NDBI, etc. all use the interaction between bands. These ratio indexes come from the analysis of the spectral characteristics presented in laboratory observations and remotely sensed data. The purpose of this article is to express the interaction relationship between original features as new features by multiplication or division operation and generalize these ratio indexes.
For a logistic regression with K-category, the softmax transformation can be used as the discriminative model to calculate the posterior probability directly for the response variable belonging to class k:
P ( Y ( k ) = k | X , α ) = exp ( j α j ( k ) X j ) i = 1 K exp ( j α j ( i ) X j )
Among them, Χ = X 1 , X 2 , , X P P is a P-dimensional predictor variable and Y K represents the response variable, K is the total number of classes, α is the vector of parameters for each classes regression, α j k is the jth parameter for class k.
The interaction between features is often expressed as a paired product relationship:
P ( Y ( k ) = k | X , α ) = exp ( j α j ( k ) X j + j , k α j k X j X k ) i = 1 K exp ( j α j ( i ) X j + j , k α j k X j X k )
the summation part j α j ( k ) X j represents the main feature item, and the quadratic term j , k α j k X j X k represents the item of the interaction relationship between the features.
Based on the interaction framework of Equation (3), the interaction between features can be expressed in the form of a ratio, which can be further written as:
P ( Y ( k ) = k | X , α ) = exp ( j α j ( k ) X j + j , k α j k ( k ) X j X k ) i = 1 K exp ( j α j ( i ) X j + j , k α j k ( i ) X j X k )
The purpose of changing the form of the product is to capture the phenomenon similar to NDVI or NDWI. The advantage of this formula is that it covers all the pair-wise ratio relationships between features. However, not all the ratio relationships of the feature pairs contribute to the separation of feature categories, so it is necessary to find the ratio of useful feature pairs, sparse optimization is an effective method.

2.3. Autonomous Learning Based on Sparse Multiclass Logistic Regression

The core of sparse optimization is to use the L1 norm as the constraint term of the regression problem. Although both the L1 norm and the L2 norm can solve the problem of regression model overfitting, the biggest difference between the two is that the L1 norm can make the part unimportant weight value of the feature becomes 0, so as to achieve the effect of sparse model coefficients. The estimated value of the sparse model with the L1 norm is expressed as follows:
α ^ = arg min i = 1 n y i j α j x i j 2 + λ j α j
Among them, n is the number of samples, x i j represents the predictor variable, y i is the dependent variable of the predictor variable. The smaller the penalty coefficient λ , the lower the degree of compression, and the more variables that are not 0; on the contrary, the larger λ , the greater the degree of compression, and the fewer variables that are not 0.
With the development of the Least Absolute Shrinkage and Selection Operator (LASSO), logistic regression has more fields for applications. Meier extended the Lasso group (gLasso) to logistic regression for high-dimensional problems and generalized linear model to solve the corresponding convex optimization problem [29]. After Böhning proposed multiclass logistic regression [30], Lasso’s contraction estimation method can be extended to multiclass logistic regression, and the resulting likelihood function of multiclass logistic regression can be written as:
L ( y | α ) = i = 1 n k = 1 K exp ( α ( k ) x i ) l = 1 K exp ( α ( l ) x i ) y i k
where x i is the P-dimensional feature vector and y i is the vector of the K-dimensional category label. When the cross-entropy error function is used for multiclass problems, the goal of maximum likelihood estimation can be achieved by minimizing the negative log-likelihood function:
E ( α ) = ln L y | α
= i = 1 n k = 1 K y i k α k x i + ln l = 1 K exp α l x i
The sparse representation can eliminate unimportant features from high-dimensional data and leave effective features [31]. Therefore, this paper introduces the L1 norm on the basis of the multiclass logistic regression model, and the parameter estimation is in the following form:
α = arg min E α + λ j P max k = 1 , 2 K α k
α k represents the L1 norm as a penalty term of the likelihood function. In this way, the sparse processing of the parameters is performed, and many components that are not important to the category are set to zero. However, this process completes high-dimensional data classification and feature selection, and the parameters will be estimated by cross-validation in the experiment. The proposed method is described as follows (Algorithm 1).
Algorithm 1 The modified SMLR model for autonomous learning
Input: An HSI dataset X and the corresponding label vectors Y , the shape of X is
H × W × C and the shape of Y is H × W × 1
Step1: Normalize the pixel values of all bands of the image between 0 and 1.
Step2: Calculate the median value for every 9 bands, the shape of the dataset is changed to
H × W × C 1 , C 1 = C / 9 + 1 . Then, calculate the interaction between any two bands and
express it as X j X k , which can be seen in equation (4).
Step3: Randomly divide the HSI dataset X into X 1 and X 2 , which represent the training data and testing data. Likewise, Y 1 and Y 2 are the corresponding label vector.
Step4: Input X 1 , X 2 and Y 1 , Y 2 into modified SMLR model for autonomous learning.
Step5: Get the optimal penalty coefficient λ through 5-fold cross-validation on the sample data
Step6: According to (6)–(8), calculate the coefficient of class k corresponding to each band, and through automatic learning, some unimportant band coefficients become 0.
Step7: The feature expression corresponding to the largest coefficient is used as the index of the extraction class k.
Output: Feature extraction results

3. Experimental Data

3.1. Pavia University Data Set

The data of Pavia University is obtained by ROSIS sensors. There are 115 bands in total, ranging from 0.43 to 0.86 μm. After processing, there are 103 bands. The data image size is 610 pixels × 340 pixels and its spatial resolution is 1.3 m. The spectral resolution is 3.74 nm. It contains 9 feature types, namely: asphalt, meadows, gravel, trees, metal, bare soil, bitumen, bricks, and shadows. Figure 2 shows the false-color composite image and ground truth information of the data set, and Table 1 shows the corresponding categories and sample size.

3.2. Pavia Center Data Set

The data of the Pavia Center was acquired by ROSIS sensors over Pavia in northern Italy. There are 115 bands, ranging from 0.43 to 0.86 μm, and there are 102 bands after processing. The size of the data image is 1096 pixels × 715 pixels, and the spatial resolution of each pixel is 1.3 m. The spectral resolution is 3.74 nm. It contains 9 feature categories, namely: water, trees, meadow, block, bare soil, asphalt, bitumen, bricks, and shadows. Figure 3 is a false-color composite image of the data set and ground truth information. Table 2 shows the type and size of samples of the data set.

3.3. Washington DC Mall Dataset

The data is acquired by the Hydice sensor and collected by the Virginia Spectrum Information Technology Application Center. The data includes a total of 210 bands ranging from 0.4 to 2.4 µm, omitting the atmospheric opaque areas in the 0.9 and 1.4 µm regions. There are 191 bands left, and the data size is 1280 pixels × 307 pixels. The spectral resolution is 10 nm. Feature categories include roof, block, gravel, meadow, trees, water, and shadows. Figure 4 shows the false-color composite map and ground truth information of the data set, and Table 3 shows the types of ground features and the number of samples.

4. Experimental Results and Analysis

Normalized Difference Vegetation Index (NDVI), Normalized Difference Water Index (NDWI), Normalized Difference Building Index (NDBI), etc. are all manifestations of the interaction between bands in a ratio [32]. Among them, NDVI is expressed as (NIR − R)/(NIR + R), which is related to the reflection value of the NIR band and the R band. The above indexes are obtained through a large number of experiments in the laboratory. This article aims to use a data-driven method to learn interactive features for hyperspectral remotely sensed data based on a sparse multiclass logistic regression model. These features used as new indexes are compared with the indexes obtained by the traditional method. This experiment only takes vegetation and water body as objects for analysis, and subsequent research can be further extended to more types of ground features.
The interaction relationship between original features is expressed as new features by multiplication or division operation, which can be used as a new index for extracting ground feature information. When considering the interaction between bands in the experiment, the following methods are used to reduce data redundancy and make the amount of calculation controllable: considering that in hyperspectral data, adjacent bands have similar spectral characteristics, it is calculated once every 9 bands The median value is used as a representative. It should be noted that only the median value in the odd number of bands can retain the information characteristics of the original band. If the average value or the median value in the even number of bands is taken, the original band needs to be averaged, which will destroy the characteristics of the original band. The reason why we chose 9 bands has been illustrated in Section 4.4. In addition, continuous strengthening of L1 regularization will make the feature matrix sparse. When automatically learning the interaction relationship, if the feature dimension is greater than the number of training samples, the sparsity of the L1 norm is required to reduce the feature dimension, so the use of the interaction between bands to determine the type of features needs to rely on the sparsity of the L1 norm.

4.1. Extraction of Vegetation and Water Body in the Pavia Center Data Set

For the Pavia Center data set with the interaction between bands, 0.5% of each type of feature is selected as the training sample. If the regularization coefficient is too large, it will easily cause under-fitting, and if it is too small, it will easily cause over-fitting. Therefore, the most suitable regularization coefficient is selected automatically through 5-fold cross-validation and controls the value of the regularization coefficient within the range of 0.01 to 100, and the number is set to 30. When the cross-validation error is the smallest, the value of the optimal regularization parameter is 16.7, Since the water body information in the data set is relatively rich, the feature selection for the water body type is shown in Table 4:
88, 94, 101 represent the original band88, 94, and 101, and the corresponding wavelengths are 865 nm, 895 nm, and 930 nm, which belong to the NIR band; BAND13, BAND32, and BAND55 indicate the bands that are added to the interaction by multiplication or division operation. The naming rules for the interaction bands are: for example, the original bands remaining after dimensionality reduction are a band a, band b, band c, band d, and band e, then BAND1 is expressed as the interaction between band a and band b, BAND2 It is expressed as the interaction between band a and band c until BAND4 is expressed as the interaction between band a and band e, the interaction between band a and other bands except itself indicates the end, and then BAND5 indicates the interaction between band b and band c until the interaction of band b and other bands except itself indicates the end, and so on until BAND10 indicates the interaction of band d and band e, so far all the interaction indicates completion. BAND13 means (band11-band44)/(band11 + band44). The wavelength of band11 is 480 nm and belongs to the B (blue) band. The wavelength of band44 is 645 nm and belongs to the R band. So BAND13 describes the interaction between B and R bands. In the same way, BAND32 means (band34-band88)/(band34 + band88) and describes the relationship between G (green) and NIR bands; BAND55 means (band94-band101)/(band94 + band101) and describes the interaction between the NIR bands. The specific interaction relationship is shown in Table 5:
From the coefficients of multiclass logistic regression, BAND32 has the most prominent regression coefficient among all the characteristics, indicating that BAND32 is the most sensitive to water body information and has the best effect on water body recognition. Therefore, BAND32 is used as an interactive water index, which is expressed as follows:
B A N D 32 = band 34 band 88 band 34 + band 88 = G N I R G + N I R
The same method for the feature selection of the feature type of tree is shown in Table 6:
BAND47 represents the interaction between band66 and band88, the wavelength of band66 is 755 nm, which means the R band, and the wavelength of band88 is 865 nm, which represents the NIR band. In the same way, BAND48 and BAND49 are also expressed as the interactive relationship between the R band and the NIR band. Since the regression coefficients of BAND47 and BAND48 are larger than BAND49, it means that for vegetation information, BAND47 and BAND48 have a better recognition effect. Analogous to the expression form of the normalized index, BAND48 can be expressed as:
B A N D 48 = band 66 band 94 band 66 + band 94 = R N I R R + N I R
Compared with NDVI, it can be proved that the interactive relationship between BAND47 and BAND48 can be used as an interactive vegetation index. In order to display the results more intuitively, BAND32 and BAND48 are used as the water body index and vegetation index to extract the water body and vegetation in the Pavia Center data set, and the results of the water body extracted by BAND32 are compared with the results of NDWI, which was extracted by the laboratory. Results of vegetation extracted by BAND48 are compared with NDVI, as shown in Figure 5 and Figure 6:
BAND32 can effectively extract water body and is similar to the extraction results of NDWI; BAND48 can also effectively extract vegetation and is similar to the extraction results of NDVI. To make the result clearer, the accuracy of the correctly extracted pixels in the binary image is counted. The pixel value 0 represents black (background), and the pixel value 255 represents white (target features). The quantitative results of the water body are shown in Table 7. The quantitative results of vegetation are shown in Table 8. The accuracy of BAND32 and BAND48 for the correct extraction of water body and vegetation is higher than that of NDWI and NDVI.

4.2. Extraction of Vegetation and Water Body in the Washington DC Mall Dataset

For the Washington DC Mall data set, because the number of samples is small, 4% of each type of feature is selected as the training sample, and the optimal regularization parameter obtained by 5-fold cross-validation is 25. At this time, the feature matrix is very sparse, with only one feature of water body and vegetation, as shown in Table 9:
BAND9 and BAND41 can be expressed as:
BAND 9 = band 6 band 95 band 6 + band 95 = p u r p l e N I R p u r p l e + N I R
BAND 41 = band 26 band 85 band 26 + band 85 = G N I R G + N I R
The wavelength of band6 is 418 nm, which corresponds to the purple light band, so it is named “purple”; the wavelength of band95 is 1248 nm, which corresponds to the NIR band. The wavelength of band26 is 498 nm corresponding to the G band, and the wavelength of band85 is 1099 nm corresponding to the NIR band. The above two expressions seem to be inconsistent with those of NDVI and NDWI. To verify the experimental results, BAND9 and BAND41 were used to extract water body and vegetation respectively. Then BAND9 is compared with NDWI in terms of water extraction; BAND41 is compared with NDVI in terms of vegetation extraction, as shown in Figure 7 and Figure 8:
From visual judgment, both the effect of BAND9 in extracting water body and the effect of BAND41 in extracting vegetation are similar to those of NDWI and NDVI. To display the extraction results more clearly, the extraction results of water and vegetation are also quantified. The quantitative results of water body are shown in Table 10; the quantitative results of vegetation are shown in Table 11.
It can be seen from Table 10 that for the extraction of water body, the extraction overall accuracy of BAND9 and NDWI are almost the same as 0.893, but for the water body category (pixel value 255), BAND9′s 0.251 is better than NDWI’s 0.209; from Table 11 we can see that for vegetation, the number of pixels correctly extracted by BAND41 is less than NDVI, but the overall accuracy of BAND41 is better than NDVI.

4.3. Extraction of Vegetation from the Data Set of the University of Pavia

For the Pavia University data set, 0.5% of each type of ground feature is selected as the training sample. Because the vegetation information in the data set is relatively rich and there is no water body information, only the vegetation part is tested. The 5-fold cross-validation shows that the value of the regularization parameter is 10. At this time, the feature matrix becomes very sparse, and the interactive features of the vegetation part can be seen in Table 12.
Compared with BAND18, BAND19 is more prominent. The coefficient of BAND19 is larger, indicating that the feature is more important, which can be expressed as:
BAND 19 = band 11 band 94 band 11 + band 94 = B N I R B + N I R
Among them, the wavelength of band11 is 480 nm belonging to the B band, and the wavelength of band94 is 895 nm, which belongs to the NIR band. The vegetation is extracted by BAND19, and the binarization results are shown in Figure 9:
BAND19 is significantly better than NDVI in the Pavia University data set for vegetation extraction. To make the results more convincing, quantitative results are also given, as shown in Table 13.
For the extraction of vegetation, BAND19 is significantly better than NDVI, which proves the effectiveness of BAND19.

4.4. Supplementary Experiment

This part is used to prove the influence of the number of bands selected when taking the median value on the experimental results. Because the selection of the number of bands is not the main content of this article, in order to make the overall structure of the article clearer, only the quantitative results are shown here. We conducted separate experiments on the cases where the number of bands was 3, 5, 7, 9 and 11, and used the overall accuracy as the verification index, concluding that when the number of bands was 9, the overall extraction result was the best. Although the extraction results of other cases are not as good as the case where the number of bands is 9, they are almost all better than the extraction results of the traditional index. The results are shown in Table 14 and Table 15.

4.5. The New Feature Index in the Data Set of the University of Pavia

In addition to comparing new features learned by the data-driven method with traditional indexes NDVI and NDWI, a new type of ground feature index can also be learned automatically. In the Pavia University data set, there is a category of features called metal. There is no traditional index that can extract this type of feature. Using the method proposed in this article, we can automatically learn the interactive features of metal. When the regularization coefficient is 1, the coefficient results are shown in Table 16.
The wavelength of band22 is 535 nm, which represents the G band, indicating that the metals in the data set are more sensitive to the G band. Among interactive features like BAND22, BAND23, and BAND29, BAND22 performs best, and BAND22 can be expressed as:
BAND 22 = band 22 band 52 band 22 + band 52 = G R G + R
The wavelength of band52 is 685 nm, which represents the R band. The metal is extracted by BAND22, and the binarization result is shown in Figure 10:
BAND22 has an outstanding effect on the extraction of metal, and quantitative results are also given in Table 17.
The interactive feature BAND22 can be used as a new ground feature index for extracting metals.

5. Conclusions

This paper proposes a data-driven method to learn interactive features for hyperspectral remotely sensed data based on a sparse multiclass logistic regression model. The key point is explicitly to express the interaction relationship between original features as new features by multiplication or division operation, and the strong constraint effect of the L1 norm makes feature sparseness. The size of the coefficient value of the corresponding feature after sparse is used as the basis for judging the importance of the feature, then the most important interactive feature is used as the index for extracting ground features. Significantly, small samples are taken on three hyperspectral remotely sensed data sets. These experiments were carried out separately, and the following conclusions were drawn:
(1) The interaction automatically learned through the data-driven method as a ground feature index can effectively extract water and vegetation. The interaction is expressed by multiplication or division operation between original features. It is shown in the Pavia Center data set that the interactive features BAND32 can extract water body information availably, BAND48 can effectively extract vegetation information; in the Washington DC Mall data set, BAND9 can effectively extract water information, which is almost indistinguishable from NDWI. Although BAND41 does not perform very well with NDVI, it can still extract most of the vegetation information; in the Pavia University data set, the effect of vegetation information extraction with BAND19 is better than NDVI. These interactive relationships learned by the data-driven method can be used as a new type of feature index, which is of great significance to the development of feature recognition.
(2) A new type of metal index BAND22 is proposed in the Pavia University data set, and it is verified with ground truth information of this category. The experimental result proves that the index can effectively extract the ground feature of metal.
The ground feature types explored in this study are only water body and vegetation. In the future, more ground feature types will need to be further studied in order to learn more ground feature indexes. If the verification is effective, it will have pivotal significance on ground features recognition and even social production activities.

Author Contributions

Conceptualization, L.D. and G.Z.; methodology, L.D. and G.Z.; software, L.D., J.G. and R.Z.; validation, L.D., G.Z., J.G. and R.Z.; formal analysis, L.D. and G.Z.; investigation, L.D. and G.Z.; writing—original draft preparation, L.D. and G.Z.; writing—review and editing, L.D., G.Z., J.G. and R.Z.; supervision, G.Z., J.G., and R.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data sets used in the article are open source.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cruz-Ramos, C.; Garcia-Salgado, B.P.; Reyes-Reyes, R.; Ponomaryov, V.; Sadovnychiy, S. Gabor Features Extraction and Land-Cover Classification of Urban Hyperspectral Images for Remote Sensing Applications. Remote Sens. 2021, 13, 2914. [Google Scholar] [CrossRef]
  2. Priem, F.; Canters, F. Synergistic use of LiDAR and APEX hyperspectral data for high-resolution urban land cover mapping. Remote Sens. 2016, 8, 787. [Google Scholar] [CrossRef] [Green Version]
  3. Zhong, Y.; Jia, T.; Zhao, J.; Wang, X.; Jin, S. Spatial-Spectral-Emissivity Land-Cover Classification Fusing Visible and Thermal Infrared Hyperspectral Imagery. Remote Sens. 2017, 9, 910. [Google Scholar] [CrossRef] [Green Version]
  4. Park, S.; Song, A. Discrepancy Analysis for Detecting Candidate Parcels Requiring Update of Land Category in Cadastral Map Using Hyperspectral UAV Images: A Case Study in Jeonju, South Korea. Remote Sens. 2020, 12, 354. [Google Scholar] [CrossRef] [Green Version]
  5. Kuras, A.; Brell, M.; Rizzi, J.; Burud, I. Hyperspectral and Lidar Data Applied to the Urban Land Cover Machine Learning and Neural-Network-Based Classification: A Review. Remote Sens. 2021, 13, 3393. [Google Scholar] [CrossRef]
  6. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep learning for hyperspectral image classification: An overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef] [Green Version]
  7. Wang, F.; Yao, X.; Xie, L.; Zheng, J.; Xu, T. Rice Yield Estimation Based on Vegetation Index and Florescence Spectral Information from UAV Hyperspectral Remote Sensing. Remote Sens. 2021, 13, 3390. [Google Scholar] [CrossRef]
  8. Chang, C.I.; Chiang, S.S.; Smith, J.; Ginsberg, I. Linear spectral random mixture analysis for hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2002, 40, 375–392. [Google Scholar] [CrossRef]
  9. Rothman, K.J.; Greenland, S.; Walker, A.M. Concepts of interaction. Am. J. Epidemiol. 1980, 112, 467–470. [Google Scholar] [CrossRef]
  10. Amy, B.; Cox, D.R. Interpretation of interaction: A review. Ann. Appl. Stat. 2007, 1, 371–385. [Google Scholar]
  11. Al-Najjar, H.A.H.; Pradhan, B.; Kalantar, B.; Sameen, M.I.; Santosh, M.; Alamri, A. Landslide Susceptibility Modeling: An Integrated Novel Method Based on Machine Learning Feature Transformation. Remote Sens. 2021, 13, 3281. [Google Scholar] [CrossRef]
  12. Xu, H. Fast information extraction of urban built-up land based on the analysis of spectral signature and normalized difference index. Geogr. Res. 2005, 24, 311–320. [Google Scholar]
  13. Rouse, J.W. Monitoring vegetation system in the Great Plains with ERTS. NASA Spec. Publ. 1974, 351, 309. [Google Scholar]
  14. Liu, W.; Xie, Y.; Wang, Z.; Zheng, H. Analysis on Urban Heat Island Effect of the Different Dynamics of Urban Surface Biophysical Descriptors in Shanghai City. Remote Sens. Technol. Appl. 2012, 27, 797–803. [Google Scholar]
  15. Mcfeeters, S.K. The use of the Normalized Difference Water Index (NDWI) in the delineation of open water features. Int. J. Remote Sens. 1996, 17, 1425–1432. [Google Scholar] [CrossRef]
  16. Xu, H. A Study on Information Extraction of Water Body with the Modified Normalized Difference Water Index. J. Remote Sensing 2005, 9, 79–85. [Google Scholar]
  17. Wang, X.; Xie, S.; Du, J. Water index formulation and its effectiveness research on the complicated surface water surroundings. J. Remote Sens. 2018, 22, 360–372. [Google Scholar]
  18. Zha, Y.; Ni, S.; Yang, S. An Effective Approach to Automatically Extract Urban Land-use from TM Imagery. J. Remote Sens. 2003, 7, 37–40. [Google Scholar]
  19. Deng, Y.; Wu, C.; Li, M.; Chen, R. RNDSI: A ratio normalized difference soil index for remote sensing of urban/suburban environments. Int. J. Appl. Earth Obs. Geoinf. 2015, 39, 40–48. [Google Scholar] [CrossRef]
  20. Liu, X.; Zhu, X.; Zhang, Q.; Yang, T.; Pan, Y.; Sun, P. A remote sensing and artificial neural network-based integrated agricultural drought index: Index development and applications. Catena 2020, 186, 104394. [Google Scholar] [CrossRef]
  21. Frazier, P.S.; Page, K.J. Water body detection and delineation with Landsat TM data. Photogramm. Eng. Remote Sens. 2000, 66, 1461–1468. [Google Scholar]
  22. Yang, W.; Yang, M.; Qi, H. Water body extracting from TM image based on BPNN. Sci. Surv. Mapp. 2012, 37, 148–150. [Google Scholar]
  23. Luo, N.; Ruan, R.; Wang, J. Species Identification of Wetland Vegetation: Based on Machine Learning and Hyperspectral Data. Chin. Agric. Sci. Bull. 2019, 35, 157–164. [Google Scholar]
  24. Hong, L.; Huang, Y.; Yang, K.; Peng, S.; Xu, Q. Study on urban surface water extraction from heterogeneous environments using GF-2 remotely sensed images. J. Remote Sens. 2019, 23, 871–882. [Google Scholar]
  25. Li, Z.; Zhao, B.; Wang, W. An Efficient Spectral Feature Extraction Framework for Hyperspectral Images. Remote Sens. 2020, 12, 3967. [Google Scholar] [CrossRef]
  26. Pu, J.; Zhao, X.; Dong, P.; Wang, Q.; Yue, Q. Extracting Information on Rocky Desertification from Satellite Images: A Comparative Study. Remote Sens. 2021, 13, 2497. [Google Scholar] [CrossRef]
  27. Zhang, B. Advancement of hyperspectral image processing and information extraction. J. Remote Sens. 2016, 20, 1062–1090. [Google Scholar] [CrossRef]
  28. Li, N.; Zhou, D.; Shi, J.; Wu, T.; Gong, M. Spectral-LocationalSpatial Manifold Learning for Hyperspectral Images Dimensionality Reduction. Remote Sens. 2021, 13, 2752. [Google Scholar] [CrossRef]
  29. Lukas, M.; Sara, V.D.G.; Peter, B. The group lasso for logistic regression. J. R. Stat. Soc. Ser. B 2008, 70, 53–71. [Google Scholar]
  30. Dankmar, B. Multiclass logistic regression algorithm. Ann. Inst. Statal Math. 1992, 44, 197–200. [Google Scholar]
  31. Dong, L.; Yuan, Y. Sparse Constrained Low Tensor Rank Representation Framework for Hyperspectral Unmixing. Remote Sens. 2021, 13, 1473. [Google Scholar] [CrossRef]
  32. Cavallo, C.; Papa, M.N.; Gargiulo, M.; Palau-Salvador, G.; Vezza, P.; Ruello, G. Continuous Monitoring of the Flooding Dynamics in the Albufera Wetland (Spain) by Landsat-8 and Sentinel-2 Datasets. Remote Sens. 2021, 13, 3525. [Google Scholar] [CrossRef]
Figure 1. Technique flowchart.
Figure 1. Technique flowchart.
Applsci 11 10502 g001
Figure 2. False-color composite map and ground truth information of the Pavia University data set: (a) False-color composite image; (b) Ground truth image.
Figure 2. False-color composite map and ground truth information of the Pavia University data set: (a) False-color composite image; (b) Ground truth image.
Applsci 11 10502 g002
Figure 3. False-color composite map of Pavia Center and ground truth information: (a) False-color composite image; (b) Ground truth image.
Figure 3. False-color composite map of Pavia Center and ground truth information: (a) False-color composite image; (b) Ground truth image.
Applsci 11 10502 g003
Figure 4. False-color composite map of Washington DC Mall and ground truth information: (a) False-color composite image; (b) Ground truth image.
Figure 4. False-color composite map of Washington DC Mall and ground truth information: (a) False-color composite image; (b) Ground truth image.
Applsci 11 10502 g004
Figure 5. Extraction results of water body and partial enlarged image: (a) Ground truth information; (b) NDWI extraction result; (c) BAND32 extraction result.
Figure 5. Extraction results of water body and partial enlarged image: (a) Ground truth information; (b) NDWI extraction result; (c) BAND32 extraction result.
Applsci 11 10502 g005
Figure 6. Extraction results of vegetation and partial enlarged image: (a) Ground truth information; (b) NDVI extraction result; (c) BAND48 extraction result.
Figure 6. Extraction results of vegetation and partial enlarged image: (a) Ground truth information; (b) NDVI extraction result; (c) BAND48 extraction result.
Applsci 11 10502 g006
Figure 7. Extraction results of water body and partial enlarged image: (a) Ground truth information; (b) NDWI extraction result; (c) BAND9 extraction result.
Figure 7. Extraction results of water body and partial enlarged image: (a) Ground truth information; (b) NDWI extraction result; (c) BAND9 extraction result.
Applsci 11 10502 g007aApplsci 11 10502 g007b
Figure 8. Extraction results of vegetation and partial enlarged image: (a) Ground truth information; (b) NDVI extraction result; (c) BAND41 extraction result.
Figure 8. Extraction results of vegetation and partial enlarged image: (a) Ground truth information; (b) NDVI extraction result; (c) BAND41 extraction result.
Applsci 11 10502 g008aApplsci 11 10502 g008b
Figure 9. Extraction results of vegetation and partial enlarged image: (a) Ground truth information; (b) NDVI extraction result; (c) BAND19 extraction result.
Figure 9. Extraction results of vegetation and partial enlarged image: (a) Ground truth information; (b) NDVI extraction result; (c) BAND19 extraction result.
Applsci 11 10502 g009
Figure 10. Extraction results of metal and partial enlarged image: (a) Ground truth information; (b) BAND22 extraction result.
Figure 10. Extraction results of metal and partial enlarged image: (a) Ground truth information; (b) BAND22 extraction result.
Applsci 11 10502 g010
Table 1. Land category and sample size of Pavia University data set.
Table 1. Land category and sample size of Pavia University data set.
#123456789
CategoryAsphaltMeadowsGravelTreesMetalBare soilBitumenBricksShadows
Size663118649209930641345502913303682947
Table 2. Land category and sample size of Pavia Center data set.
Table 2. Land category and sample size of Pavia Center data set.
#123456789
CategoryWaterTreesMeadowBlockBare soilAsphaltBitumenBricksShadows
Size65972759930912686658592497288428272864
Table 3. Land category and sample size of Washington DC Mall data set.
Table 3. Land category and sample size of Washington DC Mall data set.
#1234567
CategoryRoofBlockGravelMeadowTreesWaterShadows
Size38344161751928405122497
Table 4. Feature selection of water body.
Table 4. Feature selection of water body.
CategoryFeatures
Water88, 94, 101, BAND13, BAND32, BAND55
Table 5. Interaction characteristic coefficients of water body.
Table 5. Interaction characteristic coefficients of water body.
Features8894101BAND13BAND32BAND55
Coefficient−2.189−2.941−1.082−0.7515.7061.574
Table 6. Interactive feature coefficients of vegetation.
Table 6. Interactive feature coefficients of vegetation.
FeaturesBAND47BAND48BAND49
Coefficient0.7690.8000.094
Table 7. Quantitative comparison results of water extraction from the Pavia Center data set.
Table 7. Quantitative comparison results of water extraction from the Pavia Center data set.
CategoryRecall of NDWIRecall of BAND32
00.9980.997
2550.9150.969
Overall Accuracy0.9890.994
Table 8. Quantitative comparison results of vegetation extraction from the Pavia Center data set.
Table 8. Quantitative comparison results of vegetation extraction from the Pavia Center data set.
CategoryRecall of NDVIRecall of BAND48
00.9820.979
2550.7830.833
Overall Accuracy0.9350.945
Table 9. Interaction characteristic coefficients of water body and vegetation.
Table 9. Interaction characteristic coefficients of water body and vegetation.
CategoryInteractive FeaturesCoefficient
WaterBAND90.022
MeadowBAND41−0.09
Table 10. Quantitative comparison results of water extraction from the Washington DC Mall data set.
Table 10. Quantitative comparison results of water extraction from the Washington DC Mall data set.
CategoryRecall of NDWIRecall of BAND9
00.9420.939
2550.2090.251
Overall Accuracy0.8930.893
Table 11. Quantitative comparison results of vegetation extraction from the Washington DC Mall data set.
Table 11. Quantitative comparison results of vegetation extraction from the Washington DC Mall data set.
CategoryRecall of NDVIRecall of BAND41
00.6130.862
2550.4580.223
Overall Accuracy0.5510.605
Table 12. Interaction characteristic coefficients of vegetation.
Table 12. Interaction characteristic coefficients of vegetation.
FeatureBAND18BAND19
Coefficient0.730.87
Table 13. Quantitative comparison results of vegetation extraction from the Pavia university data set.
Table 13. Quantitative comparison results of vegetation extraction from the Pavia university data set.
CategoryRecall of NDVIRecall of BAND19
00.6770.909
2550.8440.882
Overall Accuracy0.7680.895
Table 14. The overall accuracy of vegetation extraction for the different number of bands.
Table 14. The overall accuracy of vegetation extraction for the different number of bands.
NumberPavia CenterWashington DCPavia University
NDVIOursNDVIOursNDVIOurs
30.9350.9390.5510.5820.7680.791
50.9510.590.803
70.9420.6110.882
90.9450.6050.895
110.9420.5960.837
Table 15. The overall accuracy of water body extraction for the different number of bands.
Table 15. The overall accuracy of water body extraction for the different number of bands.
NumberPavia CenterWashington DC
NDWIOursNDWIOurs
30.9890.9790.8930.892
50.9810.893
70.9890.893
90.9940.893
110.9810.889
Table 16. Interaction characteristic coefficients of metal plates.
Table 16. Interaction characteristic coefficients of metal plates.
Feature22BAND22BAND23BAND29
Coefficient4.511.820.841.06
Table 17. Quantitative comparison results of metal sheets extraction from the Pavia university data set.
Table 17. Quantitative comparison results of metal sheets extraction from the Pavia university data set.
CategoryBAND22
00.995
2550.848
Overall Accuracy0.994
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dai, L.; Zhang, G.; Gong, J.; Zhang, R. Autonomous Learning Interactive Features for Hyperspectral Remotely Sensed Data. Appl. Sci. 2021, 11, 10502. https://doi.org/10.3390/app112110502

AMA Style

Dai L, Zhang G, Gong J, Zhang R. Autonomous Learning Interactive Features for Hyperspectral Remotely Sensed Data. Applied Sciences. 2021; 11(21):10502. https://doi.org/10.3390/app112110502

Chicago/Turabian Style

Dai, Ling, Guangyun Zhang, Jinqi Gong, and Rongting Zhang. 2021. "Autonomous Learning Interactive Features for Hyperspectral Remotely Sensed Data" Applied Sciences 11, no. 21: 10502. https://doi.org/10.3390/app112110502

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop