Next Article in Journal
Context-Aware Systems Architecture in Industry 4.0: A Systematic Literature Review
Previous Article in Journal
Prevalence of Palatally Displaced Canines and Their Association with Dental and Skeletal Anomalies: A Retrospective Study
Previous Article in Special Issue
An Analysis of Spatial Variation in Human Impact on Forest Ecological Functions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Explainable Two-Layer Mode Machine Learning Method for Hyperspectral Image Classification

1
College of Information Science and Engineering, Shandong Agricultural University, Tai’an 271018, China
2
Postdoctoral Research Station for Agricultural Resource Utilization, Shandong Agricultural University, Tai’an 271018, China
3
College of Mathematical Sciences, University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(11), 5859; https://doi.org/10.3390/app15115859
Submission received: 16 April 2025 / Revised: 12 May 2025 / Accepted: 20 May 2025 / Published: 23 May 2025
(This article belongs to the Special Issue Application of Machine Learning in Land Use and Land Cover)

Abstract

Explainable machine learning methods with a specific mathematical model provide insights into how the model works. We propose a new mode that contains a two-layer architecture for hyperspectral image (HSI) classification. In the front-end learning layer, superpixel segmentation and mathematical models are combined to achieve the band selection, which obtains the data re-expression in a lower dimension. The mathematical model uses the l 2,1 norm and graph regularized term, which helps induce sparsity, improve robustness to outliers and noise, and enhance the explainability of the data re-expression. We employ the support vector machine or the K-nearest neighbor algorithms in the back-end layer to classify low-dimensional data. Finally, the two-layer mode classification method is applied to the three real HSI dataset classifications. Numerical results show that the overall classification accuracy of our method is improved.

1. Introduction

Hyperspectral images (HSIs) are obtained through an imaging spectrometer with hundreds of continuous spectral bands, and they integrate spectral and image information. They have multi-band and high spectral resolution characteristics [1]. Hyperspectral technology has become an effective tool for remote sensing applications, recognizing and learning multidimensional information from hyperspectral images to achieve corresponding research goals [2]. HSI classification has been used in agriculture for crop identification and health monitoring [3], environmental protection for land cover mapping [4], and water quality assessment [5]. In practical applications, the multi-band characteristics of hyperspectral remote sensing data enable a more comprehensive description of the reflection characteristics of ground objects and have a strong ability to identify ground object attributes [6]. However, the large amount of redundant information in the data can lead to high computational complexity, which is not conducive to subsequent image analysis and weakens the classifier’s generalization ability [7]. The curse of dimensionality and the Hughes phenomenon affect classification performance [8].
In recent years, HSI classification research has made significant progress [9]. Traditional machine learning methods, such as support vector machine (SVM) [10] and random forest [11], effectively handle high-dimensional spectral data by leveraging the unique spectral signatures of materials. However, these methods often struggle with capturing complex spatial–spectral relationships in HSI data, which led to the exploration of more advanced techniques.
With the advent of deep learning, convolutional neural networks (CNNs) emerged as a powerful tool for HSI classification. CNNs can automatically extract features from spectral and spatial dimensions of HSI data, significantly improving classification accuracy [12]. For example, 3D-CNNs have been particularly effective in capturing spatial–spectral information by applying convolutional operations across both dimensions. Also, hybrid models that combine CNNs with other architectures, such as recurrent neural networks (RNNs) or graph neural networks (GNNs) [9], have been developed to enhance performance. Due to their ability to handle non-Euclidean data and capture complex relationships between pixels, GNNs perform well.
The decision-making process of deep learning models remains a black box, making it challenging to understand how and why certain classifications are made [13]. Explainable artificial intelligence (EXI) methods [14,15] address this issue by providing interpretability to complex models.
Developing EXI methods based on integrating deep learning and mathematical models can help analyze the model’s decision-making basis and further optimize its performance. The HSI classification methods are divided into two categories: one is the end-to-end learning approach, for example, using deep learning algorithms [16,17,18,19,20,21] to directly train models from input data to classification results; another type has the characteristics of two-layer mode [22,23,24,25,26,27], which completes the data representation through front-end learning and then uses classifiers to determine the classification model in back-end learning.
A new two-layer mode machine learning method for HSI classification is proposed in this paper, which is shown in Figure 1.The classification method contains two layers: the front-end learning layer and the back-end learning layer. The former layer achieves HSI re-expression, which combines superpixel segmentation technology and a regularized subspace clustering method for band selection. The latter layer uses SVM or K-nearest neighbor (KNN) classification algorithms to classify the re-expressed HSI data. Finally, the evaluation metrics from the back-end learning help adjust the number of bands in the front-end learning layer.
To sum up, our main contributions are as follows. First, a two-layer mode machine learning method for HSI classification is proposed, combining the data re-expression and classification layers. Second, entropy-rate superpixel segmentation is used to reduce the impact of homogeneity and heterogeneity caused by different geographical locations. Third, an explainable mathematical model employs the l 2,1 norm, the graph regulation term, which maintains the robustness and inherent geometric local structure of HSI data. Fourth, numerical comparison on three HSI data sets, i.e., TeaFarm, Salinas, and Indian Pines datasets, with the existing HSI classification methods indicate that the proposed method works well.

2. Band Selection with Sparse Space Clustering

This section summarizes the classical Sparse Space Clustering (SSC) method [28,29] for hyperspectral data analysis given the HSI dataset H R m × n × b , where m , n , and b denote the width, height, and the number of bands, respectively. We reshape the original HSI cube H into matrix form X , where X = ( x 1 , x 2 , , x b ) R m n × b . The SSC method assumes that all bands (i.e., band vectors) of the X lie in a union of low-dimensional subspace, and a linear or affine combination of several bands can represent every band vectors. Accordingly, each band vector x i R m n ,   i = 1,2 ,   . . . , b can be represented as
x i = z 1 i x 1 + z 2 i x 2 + + z b i x b ,
where z i j in (1) is the i j th element of matrix Z R b × b . It should be noted that x i cannot be represented by itself, which indicates that z i i = 0 ,   i = 1,2 , . . . b .
We combine b band vectors into a matrix format so that (1) can be organized as
X = X Z ,   d i a g ( Z ) = 0 ,
where Z is the coefficient matrix of all band vectors and d i a g ( Z ) denotes the diagonal vector of matrix Z . By minimizing the following optimization Problem (3), the solution of (2) can be found [29].
Z ^ = a r g m i n Z q 2   s . t .   X   =   XZ ,   diag ( Z )   =   0 .
where Z q represents the l q norm of Z , which is defined as Z q = ( i = 1 b j = 1 b z i j q ) 1 q . For q = 0 , the l 0 norm counts the number of nonzero elements of Z , and it helps to maintain sparse property of Z . The l 1 norm is usually considered as a relaxation l 0 norm.
Then, the sparse matrix Z ^ is used to construct the similarity matrix W , that is, W = ( Z ^ + Z ^ T ) / 2 . With W , the spectral clustering method [30] clusters all band vectors into their underlying subspaces.
We convert the band selection problem into the general form
min X X Z q 2 + λ J ( Z )
where J ( · ) represents the regularized term and λ > 0 is a regularized parameter. The regularized term J ( Z ) can depend on certain properties of the data, such as geometrical information and physical meaning [22].

3. HSI Classification Method in Explainable Two-Layer Mode

This section introduces a new paradigm for HSI classification: explainable two-layer mode. We divide the classification algorithm into two layers: front-end learning and back-end learning [31]. In the front-end learning layer, the band selection is realized by combining superpixel segmentation and SSC, resulting in a low-dimensional HSI data re-expression. In the back-end learning layer, support vector machine (SVM) or K-nearest neighbor (KNN) is employed to classify low-dimensional data. Band selection and classification tasks are combined through the front-end and back-end integration strategy.

3.1. Front-End Learning Layer for Data Re-Expression

The front-end learning layer completes the band selection task, which gives HSI data a low-dimension re-expression. Our sparse SSC method combines the superpixel segmentation strategy, the l 2,1 norm, and graph regularization to explore HSI information fully.

3.1.1. Segment HSI with Superpixel Segmentation

When interpreting HSI, the same land cover type is affected by geographical location, growth conditions, and other factors, presenting different spectral curves in regions, resulting in the same object with a different spectrum. The sensor resolution may yield the same spectral curve containing spectral information of multiple land cover types simultaneously.
Entropy-rate superpixel segmentation is used to reduce the impact of the above issues. It constructs a graph for the HSI image and defines the objective function for superpixel segmentation with two components: the entropy rate of a random walk on the graph and a balance term. The entropy rate favors compact and homogeneous clusters, while the balancing term encourages clusters of similar size. The superpixel segmentation provides spatial support for region-based features [32,33].
Given the HSI cube H R m × n × b , where m and n represent the spatial width and height, respectively, and b denotes the number of bands, superpixel segmentation cuts an image into N parts, where each piece is called a superpixel. We divide H into N non-overlapping parts and obtain N pixel set S i , where S i is the index set of the i th superpixel and i = 1,2 , , N . Each superpixel is a group of adjacent pixels that share similar characteristics.

3.1.2. Explainable Minimization Problem for SSC

We compute the average spectral values at each band of N superpixels. The original matrix X R m n × b transfers into F R N × b . Then, the minimization problem for SSC can be formulated as follows:
m i n F F Z F 2 + λ Z 1
where Z 1 = m a x 1 j b i = 1 b Z i j and J ( Z ) in (4) is set to Z 1 , which helps to maintain a sparse coefficient matrix Z .
As described in (1), each band vector is written as a linear combination of other band vectors. It is reasonable in the re-expression procedure to preserve the inherent geometric local structure among the data in order to design a model that can reflect the relationship between the data f 1 , f 2 , , f N , where F = f 1 T , f 2 T , , f N T T and the re-expressed data f ~ 1 , f ~ 2 , , f ~ N , where F Z = f ~ 1 T , f ~ 2 T , , f ~ N T T . The assumption is that if f i and f j are in close relationship, their corresponding re-expressed data f ~ i , f ~ j should also be close. According to this discussion, we suppose that S = s i j R N × N represents a matrix where s i j quantifies the similarity between the data point f i and f j . The problem of learning local structure can be formulated as
         m i n 1 2 i , j = 1 N f ~ i f ~ j 2 2 s i j = 1 2 i , j = 1 N f ~ i f ~ j f ~ i f ~ j T s i j          = 1 2 i , j = 1 N f ~ i f ~ i T + f ~ j f ~ j T f ~ j f ~ i T f ~ i f ~ j T s i j = i , j = 1 N f ~ i f ~ i T f ~ i f ~ j T s i j = i = 1 N f ~ i d i i f ~ i T i = 1 N f ~ i s i j f ~ j T = t r Z T F T D F Z t r Z T F T S F Z = t r Z T F T L F Z
where   L = D S is the Laplacian matrix related to the data, S = s i j N × N , and D is a diagonal matrix with diagonal elements d i i = j = 1 N s i j for i = 1 , 2 , , N .
s i j = e x p ( f ~ i f ~ j 2 / 2 σ 2 ) , i = 1,2 , . . . N ,
The explainable minimization problem for SSC is defined as follows:
min F F Z 2,1 2 + λ 1 t r Z T F T L F Z + λ 2 Z 2,1 2 ,  
where · 2,1 denotes the l 2,1 norm, Z 2,1 = j = 1 b i = 1 b z i j 2 1 2 .
The gradient of the objective function in (6) is
f Z = F T P 1 F Z F T P 1 F + λ 1 F T L F Z + λ 2 P 2 Z  
We let d f Z / d Z = 0 ; therefore, we have
Z = F T P 1 F + λ 1 F T L F + λ 2 P 2 1 F T P 1 F
where P 1 and P 2 are diagonal matrices. The diagonal elements are
P 1 i i = 1 / j = 1 b F F Z i j 2 ,   P 2 i i = 1 / j = 1 b Z i j 2
We set W = ( Z ^ + Z ^ T ) / 2 where Z ^ is the estimated solution of (5); then, we cluster all band vectors into k underlying subspaces with the spectral clustering method. Finally, we obtain the selected band set B k ; then, after feature selection, the input data to the second layer are F B k R N × k , obtained with band selection.

3.2. Back-End Learning Layer for HSI Classification

Front-end learning obtains HSI data re-expression in a lower dimension, while back-end learning starts up the classification procedure.
KNN or SVM is used to complete the classification task for back-end learning. Given a new sample point to be classified, KNN searches for the nearest training samples in the training set and predicts the new sample point’s category based on the neighbors’ categories. In our numerical test, the distance between samples is calculated with Euclidean distance. The basic idea of SVM is to find hyperplanes that maximize the interval between different categories. In the numerical test, we call Matlab’s function KKN-Classifier or SVM-Classifier to complete the classification work.

3.3. Two-Layer Mode HSI Classification Method

The two-layer mode HSI classification method proposed in this paper fuses the front-end and back-end through the optimal indicators. We set the number of selected bands K to 5, 10, and 15. The indicator is calculated using the classification results. Finally, the optimal indicator determines the number of selected bands. Figure 2 shows the two-layer mode of our method.
Now, we are ready to give Algorithm 1, which contains our band selection method for HSI data.
Algorithm 1: Classification Algorithm under two-Layer Mode (CALM)
Input: HSI data H R m × n × b ,   the   number   of   the   selected   band   K ,   the   number   of   segmented   region   N ,   parameters   λ 1   and   λ 2 ,   Z 0 = r a n d n b , k 0 ,   ε .  
Output: The index set of the selected band ZK.
Step 1: Segment H into N regions with the index set s1, s2, ..., sN via ERS.
Step 2: Compute F with the index set S1, S2, ..., SN.
Step 3: Compute f Z   with   ( 7 ) ,   if   f Z < ε then goto Step 6, else, go to Step 4.
Step 4: Update Z with (5).
Step 5: Update S with (8), go to step 3.
Step 6: Let Z = (Z + ZT)/Z, and generate ZK with spectral clustering algorithm.
Step 7: Using ZK as input, train a classification model using SVM or KNN, andcalculate the evaluation metrics obtained by selecting K bands.
Step 8: If   K < 50 ,   then   K K + 5 ;   go   to   Step   6 .   Otherwise ,   O A = max O A K , K arg K m a x O A K .

4. Experiments

Having presented the explainable two-layer mode HSI classification method, we are ready to test its performance on three HSI datasets.

4.1. Datasets

The three HSI datasets in our experiments are TeaFarm, Salinas, and Indian Pines, whose pseudo-color and ground-truth images are shown in Figure 3, Figure 4 and Figure 5.
TeaFarm: Located in the tea tree planting base of Fanglu village, Changzhou, Jiangsu Province, the data contain 348 × 512 pixels with a spatial resolution of 2.25 m. The dataset consists of 80 spectral bands covering a spectral range of 417–855 nm.
It includes 10 types of land cover classes, which is available at https://www.geodoi.ac.cn/edoi.aspx?DOI=10.3974/geodb.2017.03.04.V1 (access on 21 May 2025).
Salinas: The data have a size of 512 × 217 with a spatial resolution of 3.7 m. The dataset contains 224 contiguous spectral bands. After removing 20 water absorption bands (108–112, 154–167, 224), 204 bands are actually used for training. The dataset includes 16 types of land cover classes.
Indian Pines: Imaged by the AVIRIS spectrometer, with a wavelength range of 0.4–2.5 μm, the data size is 145 × 145, and there are 224 spectral bands, of which 200 are used for experiments. The dataset contains 16 types of land cover classes.

4.2. Comparison Algorithms

The algorithm proposed in this article is compared numerically with the following four band selection algorithms. After implementing band selection for front-end learning, we use the same classifier for classification and compare the performance of five band selection algorithms. The information for all of these methods is summarized as follows.
ASPS_MN [34]: The core idea is to divide the HSI cube into multiple sub-cubes via an adaptive subspace partition strategy and then select the band with minimum noise in each sub-cube as the representative band to reduce the data dimension while retaining more useful information.
SEASP [35]: By calculating the correlation coefficient between adjacent bands to identify the orderliness between bands, a subset of characteristic bands is selected from the partitioned subspace using information entropy as the metric.
S4P [36]: The spatial structure of HSI is captured through superpixel segmentation, and the spectral correlation between bands is learned using a self-representation model. An adaptive and weighted multi-graph fusion term is designed to generate a unified similarity graph between different superpixels, and the importance of the bands is measured by imposing an l 2,1 norm on the self-representation coefficient matrix.
RLFFC [37]: Spatial information is obtained through superpixel segmentation, and the separability among bands is enhanced by a latent feature fusion strategy, thereby improving the performance of band selection.
GRSC [22]: The spatial information of HSI is preserved through superpixel segmentation, and representative bands are selected by combining spectral correlation and spatial structure information using a self-representation subspace clustering model, thereby reducing data redundancy and noisy bands.

4.3. Experimental Setup

In the experiments, three metrics, OA, AA, and Kappa, are used as the evaluation criteria, and two typical traditional classifiers, KNN and SVM, are also used. The number of selected bands is systematically varied from 5 to 50, with increments of 5, to observe the impact of band selection on classification performance. The parameter k of KNN is set to 5 in our experiments. As for the SVM classifier, the cross-validation process determines that the Radial Basis Function (RBF) with the kernel parameter γ = 0.5 performs best for our classification tasks. In order to reduce the sampling random bias, each classification algorithm is run 10 times, and the average classification accuracy is reported.
There are two critical parameters, λ 1 and λ 2 , in our model. We meticulously tune these parameters within a wide range of values, specifically from 10 4   to 10 4 . The optimal settings of these parameters that yield the best classification results are determined, that is, λ 1 = 0.5 and λ 2 = 0.1.

4.4. Experimental Results

Table 1 provides a detailed list of OA, AA, and Kappa values obtained using different band selection algorithms on different datasets.
During the classification stage, the stratified sampling approach helps construct training and testing sets. To form the training set, 70% of the samples from each class for each frequency band are randomly selected. This ensures that the training data are representative of the entire dataset. The remaining 30% of the samples are allocated to the testing set, which is used to evaluate the model’s performance on unseen data.
The results show that CALM showed a certain degree of improvement after introducing the regularization term compared to S4P and GRSC in all three evaluation metrics. Compared with ASPS_MN, introducing superpixel segmentation technology makes the band selection results more effective. Under all classifiers and conditions, the OA value of CALM is generally high, with values generally approaching or exceeding 0.9. As shown in Table 1, the new method, CALM, consistently outperforms other methods in terms of different metrics, i.e., OA, AA, Kappa. Taking TeaFarm datasets as an example, the OA of CALM with the SVM classifier increases by 6.88%, 1.86%, 1.71%, 5.39%, and 2.09%; the value of AA increases by 11.53%, 11.41%, 11.52%, 17.81%, and 12.35%; the value of Kappa increases by 25.08%, 2.68%, 2.48%, 7.83%, and 3.02%.
As Figure 6, Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11 demonstrates the visualization result graph under different classification methods. Taking the TeaFarm dataset as an example, from Figure 6 and Figure 7, the areas classified as bamboo forest (red area) and grassland (orange area) are less affected by other categories and exhibit higher color category purity in the classification map. In the tea tree area, the distribution range of tea trees can be more accurately identified, reducing the occurrence of misclassification and missed classification. Compared with other methods, the CALM algorithm has higher classification accuracy in complex regions and significantly reduces speckle noise. The classification results indicate the effectiveness of the CALM method, which is more suitable for subsequent hyperspectral data classification.
As is well known, raw hyperspectral images often contain noise and a large amount of redundant information, which can significantly negatively impact classification accuracy [38]. Therefore, it is not easy to consistently achieve optimal performance for HSI classification using a complete set of bands. On the contrary, selecting the appropriate number of bands can improve classification performance to a certain extent.
The optimal number of bands for 3 HSI datasets is obtained with the OA curves, which is shown in Figure 12, Figure 13 and Figure 14. The OA curves describe the OA values with different numbers of selected bands.
Figure 12, Figure 13 and Figure 14 show that the OA values increase obviously when the number of selected bands increases from 5 to 10 for all six band selection methods. The CALM method stays stable, and its OA values are better than those of the other methods.

5. Conclusions

This paper proposes a two-layer mode HSI classification method divided into front-end and back-end learning stages. To reduce the dimensionality of HSI data, the representation learning in the front end combines superpixel segmentation techniques and the minimization model defined with norm and graph regularization, which yields the re-expression of HSI data for the back-end classification. After the back-end procedure implements classification, the number of bands selected by the front-end procedure is determined by classification accuracy.
Introducing a two-layer mode machine learning method helps further develop the interpretability of end-to-end deep learning models, and so does the use of certain mathematical models to achieve data re-expression, providing explainable features for analyzing numerical results of related applications.

Author Contributions

Conceptualization, W.C., J.C. and L.S.; data curation, W.C. and S.Y.; investigation W.C., J.C. and L.S.; validation S.Y. and L.S.; formal analysis, S.Y. and L.S.; software, W.C. and S.Y.; supervision, J.C.; visualization, W.C.; writing—original draft, W.C.; funding acquisition, L.S.; methodology, W.C. and L.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded by the Strategic Priority Research Program of Chinese Academy of Sciences, Grant No. XDB 0900201.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article.

Acknowledgments

We would like to thank the anonymous referees for their helpful comments.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Cheng, M.-F.; Mukundan, A.; Karmakar, R.; Valappil, M.A.E.; Jouhar, J.; Wang, H.C. Modern Trends and Recent Applications of Hyperspectral Imaging: A Review. Technologies 2025, 13, 170. [Google Scholar] [CrossRef]
  2. Zhang, Y.; Wu, W.; Zhou, X.; Cheng, J.H. Non-Destructive Detection of Soybean Storage Quality Using Hyperspectral Imaging Technology. Molecules 2025, 30, 1357. [Google Scholar] [CrossRef]
  3. Zhang, Z.; Huang, L.; Wang, Q.; Jiang, L.; Qi, Y.; Wang, S.; Gu, Y. UAV Hyperspectral Remote Sensing Image Classification: A Systematic Review. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 3099–3124. [Google Scholar] [CrossRef]
  4. Pushpalatha, V.; Mallikarjuna, P.B.; Mahendra, H.N.; Subramoniam, S.R.; Mallikarjunaswamy, S. Land use and land cover classification for change detection studies using convolutional neural network. Appl. Comput. Geosci. 2025, 25, 100227. [Google Scholar] [CrossRef]
  5. Valme, D.; Rassõlkin, A.; Liyanage, D.C. From ADAS to Material-Informed Inspection: Review of Hyperspectral Imaging Applications on Mobile Ground Robots. Sensor 2025, 25, 2346. [Google Scholar] [CrossRef]
  6. Ma, K.; Yao, C.; Liu, B.; Hu, Q.; Li, S.; He, P.; Han, J. Segment Anything Model-Based Hyperspectral Image Classification for Small Samples. Remote Sens. 2025, 17, 1349. [Google Scholar] [CrossRef]
  7. Hennessy, A.; Clarke, K.; Lewis, M. Hyperspectral Classification of Plants: A Review of Waveband Selection Generalisability. Remote Sens. 2020, 12, 113. [Google Scholar] [CrossRef]
  8. Mianji, F.A.; Zhang, Y. Robust hyperspectral classification using relevance vector machine. IEEE Trans. Geosci. Remote Sens. 2011, 49, 2100–2112. [Google Scholar] [CrossRef]
  9. Zhao, X.; Ma, J.; Wang, L.; Zhang, Z.; Ding, Y.; Xiao, X. A review of hyperspectral image classification based on graph neural networks. Artif. Intell. Rev. 2025, 58, 172. [Google Scholar] [CrossRef]
  10. He, Z.; Xia, K.; Zhang, J.; Wang, S.; Yin, Z. An enhanced semi-supervised support Vector Machine Algorithm for spectral-spatial hyperspectral image classification. Pattern Recognit. Image Anal. 2024, 34, 199–211. [Google Scholar]
  11. Chen, C.; Yuan, X.; Gan, S.; Kang, X.; Luo, W.; Li, R.; Gao, S. A new strategy based on multi-source remote sensing data for improving the accuracy of land use/cover change classification. Sci. Rep. 2024, 14, 26855. [Google Scholar] [CrossRef] [PubMed]
  12. Banerjee, A.; Swain, S.; Rout, M.; Bandyopadhyay, M. Composite spectral spatial pixel CNN for land-use hyperspectral image classification with hybrid activation function. Multimed. Tools Appl. 2024, 84, 10527–10550. [Google Scholar] [CrossRef]
  13. Liu, J.; Lan, J.; Zeng, Y.; Luo, W.; Zhuang, Z.; Zou, J. Explainability Feature Bands Adaptive Selection for Hyperspectral Image Classification. Remote Sens. 2025, 17, 1620. [Google Scholar] [CrossRef]
  14. Hassija, V.; Chamola, V.; Mahapatra, A.; Singal, A.; Goel, D.; Huang, K.; Hussain, A. Interpreting black-box models: A review on explainable artificial intelligence. Cogn. Comput. 2024, 16, 45–74. [Google Scholar] [CrossRef]
  15. Contreras, J.; Bocklitz, T. Explainable artificial intelligence for spectroscopy data: A review. Pflügers Arch. Eur. J. Physiol. 2024, 477, 603–615. [Google Scholar] [CrossRef]
  16. Li, R.; Zheng, S.; Duan, C.; Yang, Y.; Wang, X. Classification of hyperspectral image based on double-branch dual-attention mechanism network. Remote Sens. 2020, 12, 582. [Google Scholar] [CrossRef]
  17. Yu, H.; Zhang, H.; Liu, Y.; Zheng, K.; Xu, Z.; Xiao, C. Dual-channel convolution network with image-based global learning framework for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  18. Li, T.; Zhang, X.; Zhang, S.; Wang, L. Self-supervised learning with a dual-branch ResNet for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  19. Wang, C.; Zhan, C.; Lu, B.; Yang, W.; Zhang, Y.; Wang, G.; Zhao, Z. SSFAN: A Compact and Efficient Spectral-Spatial Feature Extraction and Attention-Based Neural Network for Hyperspectral Image Classification. Remote Sens. 2024, 16, 4202. [Google Scholar] [CrossRef]
  20. Tang, X.; Zhang, K.; Zhou, X.; Zeng, L.; Huang, S. Enhancing Binary Convolutional Neural Networks for Hyperspectral Image Classification. Remote Sens. 2024, 16, 4398. [Google Scholar] [CrossRef]
  21. Zhao, H.; Lu, Z.; Sun, S.; Wang, P.; Jia, T.; Xie, Y.; Xu, F. Classification of Large Scale Hyperspectral Remote Sensing Images Based on LS3EU-Net++. Remote Sens. 2025, 17, 872. [Google Scholar] [CrossRef]
  22. Wang, J.; Tang, C.; Zheng, X.; Liu, X.; Zhang, W.; Zhu, E. Graph regularized spatial–spectral subspace clustering for hyperspectral band selection. Neural Netw. 2022, 153, 292–302. [Google Scholar] [CrossRef]
  23. Wang, Y.; Ma, H.; Yang, Y.; Zhao, E.; Song, M.; Yu, C. Self-supervised deep multi-level representation learning fusion-based maximum entropy subspace clustering for hyperspectral band selection. Remote Sens. 2024, 16, 224. [Google Scholar] [CrossRef]
  24. Wang, Q.; Li, Q.; Li, X. A fast neighborhood grouping method for hyperspectral band selection. IEEE Trans. Geosci. Remote Sens. 2020, 59, 5028–5039. [Google Scholar] [CrossRef]
  25. Liu, K.; Chen, Y.; Chen, T. A band subset selection approach based on sparse self-representation and band grouping for hyperspectral image classification. Remote Sens. 2022, 14, 5686. [Google Scholar] [CrossRef]
  26. Zhao, H.; Bruzzone, L.; Guan, R.; Zhou, F.; Yang, C. Spectral-spatial genetic algorithm-based unsupervised band selection for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9616–9632. [Google Scholar] [CrossRef]
  27. Habermann, M.; Fremont, V.; Shiguemori, E.H. Supervised band selection in hyperspectral images using single-layer neural networks. Int. J. Remote Sens. 2019, 40, 3900–3926. [Google Scholar] [CrossRef]
  28. Elhamifar, E.; Vidal, R. Sparse subspace clustering: Algorithm, theory, and applications. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2765–2781. [Google Scholar] [CrossRef]
  29. Sun, W.; Zhang, L.; Du, B.; Li, W.; Lai, Y.M. Band selection using improved sparse subspace clustering for hyperspectral imagery classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2784–2797. [Google Scholar] [CrossRef]
  30. Ng, A.; Jordan, M.; Weiss, Y. On spectral clustering: Analysis and an algorithm. Adv. Neural Inf. Process. Syst. 2001, 14, 849–856. [Google Scholar]
  31. Guo, T.; Han, C.; Li, M. Fusion of front-end and back-end learning based on layer-by-layer data re-representation. Sci. Sin. Inform. 2019, 49, 739–759. (In Chinese) [Google Scholar] [CrossRef]
  32. Liu, M.; Tuzel, O.; Ramalingam, S.; Chellappa, R. Entropy rate superpixel segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Colorado Springs, CO, USA, 20–25 June 2011; pp. 2097–2104. [Google Scholar]
  33. Li, X.; Chen, J.; Zhao, L.; Li, H.; Wang, J.; Sun, L.; Guo, S.; Chen, P.; Zhao, X. Superpixel segmentation based on anisotropic diffusion model for object-oriented remote sensing image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 17, 7621–7639. [Google Scholar] [CrossRef]
  34. Wang, Q.; Li, Q.; Li, X. Hyperspectral band selection via adaptive subspace partition strategy. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 4940–4950. [Google Scholar] [CrossRef]
  35. Tang, C.; Wang, J. A hyperspectral band selection method via adjacent subspace partition. Tientsin Univ. J. 2022, 55, 255–262. [Google Scholar]
  36. Tang, C.; Wang, J.; Zheng, X.; Liu, X.; Xie, W.; Li, X.; Zhu, X. Spatial and spectral structure preserved self-representation for unsupervised hyperspectral band selection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–13. [Google Scholar] [CrossRef]
  37. Wang, J.; Tang, C.; Li, Z.; Liu, X.; Zhang, W.; Zhu, E.; Wang, L. Hyperspectral band selection via region-aware latent features fusion based clustering. Inf. Fusion 2022, 79, 162–173. [Google Scholar] [CrossRef]
  38. Wang, H.; Yu, G.; Cheng, J.; Zhang, Z.; Wang, X.; Xu, Y. Fast hyperspectral image classification with strong noise robustness based on minimum noise fraction. Remote Sens. 2024, 16, 3782. [Google Scholar] [CrossRef]
Figure 1. Structure diagram of two-layer mode machine learning method.
Figure 1. Structure diagram of two-layer mode machine learning method.
Applsci 15 05859 g001
Figure 2. Band selection model based on superpixel segmentation.
Figure 2. Band selection model based on superpixel segmentation.
Applsci 15 05859 g002
Figure 3. TeaFarm: (a) pseudo-color image; (b) ground-truth image; (c) class names.
Figure 3. TeaFarm: (a) pseudo-color image; (b) ground-truth image; (c) class names.
Applsci 15 05859 g003
Figure 4. Salinas: (a) pseudo-color image; (b) ground-truth image; (c) class names.
Figure 4. Salinas: (a) pseudo-color image; (b) ground-truth image; (c) class names.
Applsci 15 05859 g004
Figure 5. Indian Pines: (a) pseudo-color image; (b) ground-truth image; (c) class names.
Figure 5. Indian Pines: (a) pseudo-color image; (b) ground-truth image; (c) class names.
Applsci 15 05859 g005
Figure 6. TeaFarm dataset classification results using KNN classifier: (a) ASPS_MN; (b) S4P; (c) GRSC; (d) RLFFC; (e) SEASP; (f) CALM; (g) Color labels.
Figure 6. TeaFarm dataset classification results using KNN classifier: (a) ASPS_MN; (b) S4P; (c) GRSC; (d) RLFFC; (e) SEASP; (f) CALM; (g) Color labels.
Applsci 15 05859 g006
Figure 7. TeaFarm dataset classification results using SVM classifier: (a) ASPS_MN; (b) S4P; (c) GRSC; (d) RLFFC; (e) SEASP; (f) CALM; (g) Color labels.
Figure 7. TeaFarm dataset classification results using SVM classifier: (a) ASPS_MN; (b) S4P; (c) GRSC; (d) RLFFC; (e) SEASP; (f) CALM; (g) Color labels.
Applsci 15 05859 g007
Figure 8. Salinas dataset classification results using KNN classifier: (a) ASPS_MN; (b) S4P; (c) GRSC; (d) RLFFC; (e) SEASP; (f) CALM; (g) Color labels.
Figure 8. Salinas dataset classification results using KNN classifier: (a) ASPS_MN; (b) S4P; (c) GRSC; (d) RLFFC; (e) SEASP; (f) CALM; (g) Color labels.
Applsci 15 05859 g008
Figure 9. Salinas dataset classification results using SVM classifier: (a) ASPS_MN; (b) S4P; (c) GRSC; (d) RLFFC; (e) SEASP; (f) CALM; (g) Color labels.
Figure 9. Salinas dataset classification results using SVM classifier: (a) ASPS_MN; (b) S4P; (c) GRSC; (d) RLFFC; (e) SEASP; (f) CALM; (g) Color labels.
Applsci 15 05859 g009
Figure 10. Indian Pines dataset classification results using KNN classifier: (a) ASPS_MN; (b) S4P; (c) GRSC; (d) RLFFC; (e) SEASP; (f) CALM; (g) Color labels.
Figure 10. Indian Pines dataset classification results using KNN classifier: (a) ASPS_MN; (b) S4P; (c) GRSC; (d) RLFFC; (e) SEASP; (f) CALM; (g) Color labels.
Applsci 15 05859 g010
Figure 11. Indian Pines datasets classification results using SVM classifier: (a) ASPS_MN; (b) S4P; (c) GRSC; (d) RLFFC; (e) SEASP; (f) CALM; (g) Color labels.
Figure 11. Indian Pines datasets classification results using SVM classifier: (a) ASPS_MN; (b) S4P; (c) GRSC; (d) RLFFC; (e) SEASP; (f) CALM; (g) Color labels.
Applsci 15 05859 g011
Figure 12. OA of two different classifiers with varied numbers of selected bands on the TeaFarm dataset. (a) OA obtained by KNN; (b) OA obtained by SVM.
Figure 12. OA of two different classifiers with varied numbers of selected bands on the TeaFarm dataset. (a) OA obtained by KNN; (b) OA obtained by SVM.
Applsci 15 05859 g012
Figure 13. OA of two different classifiers with varied numbers of selected bands on the Salinas dataset. (a) OA obtained by KNN; (b) OA obtained by SVM.
Figure 13. OA of two different classifiers with varied numbers of selected bands on the Salinas dataset. (a) OA obtained by KNN; (b) OA obtained by SVM.
Applsci 15 05859 g013
Figure 14. OA of two different classifiers with varied numbers of selected bands on the Indian Pines dataset. (a) OA obtained by KNN; (b) OA obtained by SVM.
Figure 14. OA of two different classifiers with varied numbers of selected bands on the Indian Pines dataset. (a) OA obtained by KNN; (b) OA obtained by SVM.
Applsci 15 05859 g014
Table 1. Classification results of different methods with three metrics on different datasets.
Table 1. Classification results of different methods with three metrics on different datasets.
DatasetsClassifierMetricsASPS_MNS4PGRSCRLFFCSEASPCALM
TeaFarmKNNOA0.90120.94170.95080.94370.94450.9522
AA0.77470.85760.86380.86330.86470.8902
Kappa0.89320.91520.91400.91800.91940.9302
SVMOA0.88020.93040.93190.89510.92810.9490
AA0.74790.74910.74800.6 8510.73970.8632
Kappa0.67470.89870.90070.84720.89530.9255
SalinasKNNOA0.82820.91500.92420.90740.92010.9260
AA0.88300.95650.96290.95370.96100.9637
Kappa0.81720.90750.91740.89940.91290.9193
SVMOA0.83180.93180.94580.89080.94780.9469
AA0.88490.96730.97380.93200.97210.9757
Kappa0.82100.92550.92590.88120.94180.9417
Indian PinesKNNOA0.71840.76780.86440.74460.76530.8363
AA0.73240.76780.80000.68240.71500.8285
Kappa0.68850.74790.78070.72340.74490.8285
SVMOA0.75520.86160.89710.80090.87990.9016
AA0.72840.86260.89760.74120.88520.8989
Kappa0.73520.84770.88570.78210.86710.8907
NOTE: The best results on the quality index are labeled in bold.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, W.; Cheng, J.; Yang, S.; Sun, L. Explainable Two-Layer Mode Machine Learning Method for Hyperspectral Image Classification. Appl. Sci. 2025, 15, 5859. https://doi.org/10.3390/app15115859

AMA Style

Chen W, Cheng J, Yang S, Sun L. Explainable Two-Layer Mode Machine Learning Method for Hyperspectral Image Classification. Applied Sciences. 2025; 15(11):5859. https://doi.org/10.3390/app15115859

Chicago/Turabian Style

Chen, Wenjia, Junwei Cheng, Song Yang, and Li Sun. 2025. "Explainable Two-Layer Mode Machine Learning Method for Hyperspectral Image Classification" Applied Sciences 15, no. 11: 5859. https://doi.org/10.3390/app15115859

APA Style

Chen, W., Cheng, J., Yang, S., & Sun, L. (2025). Explainable Two-Layer Mode Machine Learning Method for Hyperspectral Image Classification. Applied Sciences, 15(11), 5859. https://doi.org/10.3390/app15115859

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop