Next Article in Journal
DualNet-PoiD: A Hybrid Neural Network for Highly Accurate Recognition of POIs on Road Networks in Complex Areas with Urban Terrain
Next Article in Special Issue
Refined Landslide Susceptibility Mapping Considering Land Use Changes and InSAR Deformation: A Case Study of Yulin City, Guangxi
Previous Article in Journal
Optical-to-SAR Translation Based on CDA-GAN for High-Quality Training Sample Generation for Ship Detection in SAR Amplitude Images
Previous Article in Special Issue
Multiple Land-Use Simulations and Driving Factor Analysis by Integrating a Deep Cascade Forest Model and Cellular Automata: A Case Study in the Pearl River Delta, China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DCFF-Net: Deep Context Feature Fusion Network for High-Precision Classification of Hyperspectral Image

1
Faculty of Resources and Environmental Science, Hubei University, Wuhan 430062, China
2
Hubei Key Laboratory of Regional Development and Environmental Response, Hubei University, Wuhan 430062, China
3
School of Geography and Environment, Jiangxi Normal University, Nanchang 330022, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2024, 16(16), 3002; https://doi.org/10.3390/rs16163002
Submission received: 28 June 2024 / Revised: 5 August 2024 / Accepted: 12 August 2024 / Published: 15 August 2024
(This article belongs to the Special Issue GeoAI and EO Big Data Driven Advances in Earth Environmental Science)

Abstract

:
Hyperspectral images (HSI) contain abundant spectral information. Efficient extraction and utilization of this information for image classification remain prominent research topics. Previously, hyperspectral classification techniques primarily relied on statistical attributes and mathematical models of spectral data. Deep learning classification techniques have recently been extensively utilized for hyperspectral data classification, yielding promising outcomes. This study proposes a deep learning approach that uses polarization feature maps for classification. Initially, the polar co-ordinate transformation method was employed to convert the spectral information of all pixels in the image into spectral feature maps. Subsequently, the proposed Deep Context Feature Fusion Network (DCFF-NET) was utilized to classify these feature maps. The model was validated using three open-source hyperspectral datasets: Indian Pines, Pavia University, and Salinas. The experimental results indicated that DCFF-NET achieved excellent classification performance. Experimental results on three public HSI datasets demonstrated that the proposed method accurately recognized different objects with an overall accuracy (OA) of 86.68%, 94.73%, and 95.14% based on the pixel method, and 98.15%, 99.86%, and 99.98% based on the pixel-patch method.

1. Introduction

Hyperspectral image (HSI) classification has become an essential research area in remote sensing due to its capability to capture rich spectral information [1,2,3]. Traditional methods for hyperspectral image classification face various limitations and drawbacks. A widely used method includes linear discrimination analysis (LDA) [4,5], K Nearest Neighbors [6,7], Naive Bayes [8,9], polynomial logic regression classification [10,11,12], support vector machine [13,14,15], and others. These methods generally rely on statistical principles and typically represent models within a vector space. A data matrix often describes geographical spatiotemporal datasets, where each sample is considered a vector or point within a finite-dimensional Euclidean space. The inter-relationships among samples are defined by the connections between points [16,17]. Another category of classification methods focuses on spectral graph features. These methods transform spectral data into spectral curves and then perform direct classification and analysis based on the spectral curve graph features [18,19]. Techniques such as the cumulative cross-area of spectral curves, fractal feature method [20], spectral curve feature point extraction method [21], spectral angle classification [22], spectral curve matching algorithm [23], and spectral curve graph index method that describes key statistical features [24] are included. Overall, these approaches mainly focus on the global morphological information of the spectral curve and lack a comprehensive understanding of the spatial–structural characteristics inherent in the spectral information [25,26,27]. Effective extraction of spectral curves or features and optimizing model design are critical research areas for enhancing image classification accuracy.
Deep learning, a machine learning approach based on artificial neural networks, traces its origins to the perceptual machine model of the 1950s [28]. Research has demonstrated the promising potential of deep learning methods in classifying graphics [29]. In recent decades, significant efforts have been dedicated to extensively studying and enhancing neural networks. The increased prevalence of deep learning can be attributed to advancements in computing power, the availability of vast datasets, and algorithms’ continuous evolution [30]. Due to limitations in computing resources and imperfect training algorithms, neural networks have faced challenges in practical applications. In 2006, Geoffrey Hinton introduced the Deep Belief Network (DBN), which spearheaded a new wave in deep learning [31,32]. The DBN employs greedy pre-training to iteratively extract abstract data features through multi-layer unsupervised learning. Recently, researchers have increasingly utilized deeper network structures combined with the backpropagation algorithm. Deep neural networks have significantly advanced image classification, voice recognition, and natural language processing [33,34,35]. In 2012, Krizhevsky introduced the Alexnet convolutional neural network model, which secured victory in the ImageNet image classification competition, effectively indicating the immense potential of deep neural networks in handling large-scale complex data [33]. Subsequently, many deep learning models have emerged, such as the VGG [36,37] network models proposed by Simonyan and Zisserman in 2014, and the GoogLeNet [38] introduced by Google. In 2015, Kaiming presented the ResNet neural network model, which introduced residual connections to address the issues of gradient vanishing and explosion in deep neural networks [39]. Since then, numerous subsequent studies have been conducted, such as ResNetV2 [40], ResNetXT [41], and ResNetST [42], which have further enhanced the performance of the model. The ongoing refinement of deep learning models has led to significant advancements in image classification, resulting in a continuous enhancement of the accuracy of image classification results. These developments have provided a wealth of well-established classification model frameworks for future researchers. The models above primarily rely on 2D-CNN, in addition to numerous classification methods based on sequential data, such as the multi-layer perception machine (MLP), 1D-CNN [43], RNN [44], HybridSN [45], and 3D-CNN [46,47] classification methods that incorporate attention mechanisms [48,49]. The MLP is a fundamental feedforward neural network structure for processing various data types, possessing strong fitting and generalization capabilities. 1D-CNN is commonly used for processing sequential data such as time series or sensor data. It effectively captures local patterns and dependencies within the data, making it suitable for speech recognition and natural language processing [43]. RNN is designed to handle sequential data by maintaining an internal memory that allows it to process input sequences of arbitrary length [44]. This feature makes it well-suited for tasks involving sequential dependencies, such as language modeling, machine translation, and time-series prediction. A hybrid spectral–spatial CNN (HybridSN) employs a 3D-CNN to learn the joint spatial–spectral feature representation, followed by encoding spatial features using a 2D-CNN [45]. Extending the concept of traditional 2D-CNN, 3D-CNN processes spatio-temporal data such as videos, volumetric medical images, and high spectral data for classification [50]. By considering both spatial and temporal features, 3D-CNNs effectively capture complex patterns and movements within the data, making them suitable for action recognition, video analysis, and medical image classification. These diverse classification methods are widely used for hyperspectral image classification. In the field of remote sensing image processing, particularly in hyperspectral image classification and super-resolution reconstruction, there have been notable advancements in recent years. Hong [51] introduced SpectralFormer(SF), a transformative approach to hyperspectral image classification that integrates transformer architecture, significantly enhancing both accuracy and efficiency. This study represents not just a technological innovation but also a crucial breakaway from traditional methods. Roy [52] further advanced this field by proposing a spectral–spatial morphological attention transformer, effectively integrating spectral and spatial information to optimize hyperspectral image classification performance. This method considers not only pixel-level spectral features but also spatial information surrounding pixels, thereby improving classification accuracy. With advancements in hyperspectral technology, considerable attention has also been directed towards image super-resolution techniques. Zhang [53] presented a single-source domain expansion network capable of effectively handling cross-scene hyperspectral image classification tasks, demonstrating its superiority in processing complex scene data. Liu [54] introduced a multi-area target attention technique that enhances the accuracy and robustness of hyperspectral image classification through region-specific attention mechanisms. This method enables finer focus on different regions within an image, thereby enhancing the applicability of classification algorithms in complex environments. In the domain of hyperspectral image super-resolution, Liu [55] proposed an efficient ESSAformer model tailored to reconstructing hyperspectral images with enhanced resolution. ESSAformer not only improves the quality of image reconstruction but also provides finer details that are crucial for remote sensing image analysis and applications. These innovative studies have propelled the development of hyperspectral image processing technologies, offering powerful tools and methods for various remote sensing applications [56,57]. By integrating deep learning with traditional image processing techniques [58], these advancements pave the way for extensive prospects in the future of remote sensing image analysis and applications.
Most high spectral deep learning classification methods are currently based on pixel-patch classification. This approach involves selecting central pixels and their surrounding pixels as the objects for classification. However, two major shortcomings persist: firstly, the model training only incorporates the pixel label value at the center of the image block, resulting in the wastage of a large amount of label information from the neighborhood and the disregard of spatial information within the pixel patch. Secondly, overlapping can lead to severe information leakage [59,60]. When using pixel patches as training features, pixel patches containing label information are directly or indirectly utilized for model training. However, these methods cannot generally be generalized to untagged data. In order to address these issues, this study introduces a new method for pixel-based HSI classification, undertaking the following tasks:
(1)
The eigenvalues of each band of the hyperspectral image are transformed into polarization feature maps utilizing the polar co-ordinate conversion method. This process converts each pixel’s spectral value into a polygon, capturing all original pixel information. These transformed feature maps then serve as a novel input form, facilitating direct training and classification within a classic 2D-CNN deep learning network model, such as VGG or ResNet;
(2)
Based on the feature maps generated in the previous step, a novel deep learning residual network model called DCFF-Net is introduced for training and classifying the converted spectral feature maps. This study includes comprehensive testing and validation across three hyperspectral datasets: Indian Pines, Pavia University, and Salinas. The proposed model consistently exhibits superior classification performance across these datasets through comparative analysis with other advanced pixel-based classification methods;
(3)
The response mechanism of DCFF-Net’s classification accuracy to polar co-ordinate maps under different filling methods is analyzed. The DCFF-Net model, evaluated using the pixel-patch input mode, is compared to other advanced models for classification performance, consistently demonstrating outstanding results.

2. Method

The classification methodology introduced in this study presents a novel approach by modifying the model’s data-input process, transforming pixel spectrum information within the hyperspectral image into graphical representations, and utilizing these as input data for model training and classification. The entire model primarily consists of two core components. The first component standardizes the pixel spectroscopy information into graphical data, with detailed processing methods outlined in Section 2.1. The second component involves training and classifying the transformed graphical data using the newly developed classification model, with comprehensive details regarding the model’s framework structure provided in Section 2.2.

2.1. Converting Hyperspectral Pixels into Feature Maps

Traditional hyperspectral data classification relies on vector space and is described as the data matrix. Each sample is considered a vector or point within a defined Euclidean space, ignoring the spatial relationships among the spectral values of each band in the pixel. The spectrum feature map of the image records not only the spectral values of the bands in the pixel but also the spatial relationships between these spectral values. This study introduces a new method for converting to a feature map, transforming the spectral classification problem into one of image recognition. Initially, the polar value normalization method is employed to standardize all bands of hyperspectral images, as calculated in Equation (1). Then, all the standardized optical spectral values of each pixel are converted into a polar co-ordinate diagram to form a two-dimensional polygon (Figure 1).
p i , j , k = p i , j , k σ k m a x p i , j , 1 , p i , j , 2 , , p i , j , k m i n p i , j , 1 , p i , j , 2 , , p i , j , k
where p i , j , k is the spectral value of the k-th band of pixels in the i-th row and j-th column before normalization; p i , j , k is the spectral values after the normalization; σ k is the average value of the k-th band.
The co-ordinate origin (0, 0) was employed to calculate the polar co-ordinate values of the spectral values for each band. In addition, the polar co-ordinate rotation angle, θ k , is determined based on the number of bands in the hyperspectral data, and n represents the number of bands. The calculation method proceeds as follows:
θ k = 2 π n × k ( k = 0,1 ,   , n 1 )
The following formula calculates the co-ordinates x k , y k :
x k = cos θ k × p i , j , k y k = sin θ k × p i , j , k
The co-ordinate starting point of the feature maps in this study originates from the positive direction of the x–axis and involves counter-clockwise rotation, and neighboring co-ordinate points are connected with straight lines. Ultimately, the first and last co-ordinates are connected to form a closed polar co-ordinate graph. After extensive experimentation, the output graph size is set at 224 × 224 × 3.
In the process of converting spectral data into feature maps, hyperspectral imaging (HSI) data and corresponding label data are read simultaneously. The converted feature maps are standardized and annotated to include label information and pixel IDs. IDs are assigned sequentially starting from the top-left corner of the image matrix, proceeding row-wise from top to bottom and column-wise from left to right, with IDs ranging from 0 to H × W − 1. Here, H and W denote the height and width of the image data, respectively. According to the label category of each feature map, the feature maps of the same category were saved into corresponding folders. For instance, the Indian Pines dataset, which includes 16 categories, results in 16 folders containing labeled feature maps. Categories are labeled consistently with those in the image, ranging from 1 to 16, while unlabeled data are categorized as 0 and stored separately. During image prediction, there is an option to predict unlabeled data, which can significantly reduce prediction time. Alternatively, predictions can be made for all feature maps, albeit at the cost of increased processing time. Following prediction, the results are mapped back to an image format based on the ID information within the feature maps.
The method described above transforms pixels from hyperspectral images into feature maps. Figure 1 displays typical feature figures from the three datasets. Variations in sensors and bands among these hyperspectral datasets result in differing characteristics of the hyperspectral feature maps. The method only shows typical feature maps for each land type. In practice, features of the same terrain type vary, exhibiting distinct differences. Within a specific dataset, hyperspectral feature maps of different land types share similarities in their overall characteristics while exhibiting unique local details. These similarities provide an effective basis for classifying various hyperspectral feature maps. Similar spectral feature patterns are observed across different locations, especially in the Indian Pines dataset. Among the 16 types of characteristic graphic, such as 5-Grass/Pasture, Figure 1 highlights shape features where approximately one-third of the patterns resemble 7-Grass-Pasture-Mowed. This similarity is also noted in the nine types of characteristic graphic computed from the Pavia University datasets, although with larger and more varied overall sizes and shapes. In the Salinas dataset, characteristic graphics predominantly take two forms: butterfly-like and jellyfish-like, with notable similarities observed in the sizes of 8-Grapes_untrained and 15-Vineyard_untrained.

2.2. Network Architectures

Deep Context Feature Fusion Network (DCFF-Net) is proposed for the classification of hyperspectral feature diagrams. This network consists of three essential modules: Special Information Embedding (SIE, Figure 2b), Deep Special Feature Extraction (DSFE, Figure 2c), and Context Information Fusion (CIF, Figure 2a). To streamline the model parameters, the input image undergoes a 7 × 7 convolution followed by batch normalization, ReLU activation, and max-pooling. The resulting computations are then fed into the SIE and DSFE modules. The DSFE module extracts high-level feature representations of the spectral feature diagrams through a sequence of convolutional and batch normalization operations, followed by max-pooling. Then, the CIF module integrates the features from the SIE and DSFE modules. The output from this fusion process undergoes further processing through the SIE module. After passing through four layers of DSFE and CIF modules, max-pooling and global average pooling are applied to enrich the characteristic features. Finally, following the fully connected layer, the classification outcome is obtained through the softmax activation function.
The CIF module serves to seamlessly amalgamate context information, effectively integrating characteristics from various levels and yielding the ultimate feature representation for different pixels. Within the spectral feature extraction module, the features extracted by DSFE are bifurcated, with a portion integrated into the SIE module for contextual feature integration, while the remainder proceeds to the subsequent DSFE module. Figure 1 provides a simplified representation of the network architecture. Following each convolutional operation in the model, batch normalization and ReLU activation are applied. CIF comprises five distinct SIE modules and four DSFE and SIE fusion layers.

2.2.1. Spectral Information Embedding

With the advent of residual networks, the potential for deep learning network layers has substantially increased, leading to enhanced network depth and improved classification outcomes [39]. The Spectral Information Embedding (SIE) is configured as a residual block that includes a Squeeze and Excitation (SE) self-attention mechanism [61]. These residual blocks accelerate network training and mitigate challenges such as vanishing and exploding gradients. In the SIE module, the input x first undergoes a 3 × 3 convolution, batch normalization, and ReLU activation. Then, the SE Attention operation integrates all channels to enhance those contributing significantly to the classification task. Further convolution and batch normalization are then performed. When the stride in the input parameter is 1, the results are added back to the input layer. If the stride is greater than 1, a 1 × 1 convolution is applied to restore the size of the convolutional layer before adding it to the input layer to achieve the final feature encoding, which is then activated using the ReLU function. The specific structure is illustrated in Figure 2b.
The residual block is the basic module in the residual network model structure. The function can be expressed as follows:
F x = H x + x
where H x is the residual function, which can be regarded as the part that performs a non-linear transformation on the input x . The addition operation in the formula means adding the input x and the output of the residual function H x to F x , which obtains the output of the residual block.
Convolutional layer operations form the core of deep learning. To enhance model performance, researchers have proposed several processing methods, such as batch normalization (BN) [62], activation functions [63], maximum pooling, and average pooling. These methods improve the model’s overall performance in different ways. The BN and ReLU activation functions are widely used in deep learning models. Batch normalization reduces the distribution variance across different data batches, mitigates gradient vanishing or explosion issues, and improves the training effect, generalization capability, and robustness of the deep neural network model. The ReLU activation function, known for its simplicity, speed, and effectiveness, significantly enhances model accuracy and convergence speed, demonstrating strong performance in numerous deep learning models.
The formulas for these functions are defined as follows:
X n = x x ¯ s + ε
where x ¯ is mean; s is the total variance; ε is a small positive number to avoid division by zero. The Relu activation function is defined as follows:
f x = m a x 0 , x
where x is the element value in the input tensor. For input values greater than or equal to 0, the original value is output; for input values less than 0, the output value is 0.

2.2.2. Deep Spectral Feature Extract

After each convolutional computation, batch normalization and ReLU activation are applied, and a maximum pooling operation is employed to extract image features. This module employs a two-layer 3 × 3 convolutional calculation, incorporating the SE-Attention module to reconstruct the characteristics of different channels. It connects the feature map after the initial convolutional layer with that of the SE-Attention and subsequently with the feature map after the final convolutional layer to produce the ultimate feature map output of the DSFE module.
The max pooling layer preserves the most significant features and reduces the spatial dimension, diminishing the model’s sensitivity to spatial location, decreasing the number of parameters, and improving computational efficiency. Max pooling methods reduce the feature map size through down-sampling operations, allowing the model to learn overall features and local invariance, enhancing the model’s robustness and generalization capability. For example, if the input two-dimensional feature map size is (H, W), the pooling operation employs a window of size k × k and a step size of s . Hence, the size of the feature map output by the pooling is ( H k s + 1 , W k s + 1 ), and the maximum pooling formula is defined as follows:
Y i , j = max p = 0 k 1 max q = 0 k 1 X i × s + p , j × s + q , i = 0 , , H k s , j = 0 , , W k s
where i and j are the row and column index of the output feature map, respectively, and p and q are the row and column index in the pooling window, respectively. The function represents taking the maximum value in the window.

2.2.3. Cross-Entropy Loss Function and Activation Function

Cross-Entropy Loss is a commonly used metric in classification tasks to evaluate the difference between predicted outcomes and actual labels. It draws on the concept of cross-entropy in information theory, assessing the performance of a model by calculating the difference between the predicted probability distribution and the true label distribution. During the training process, model parameters are refined by minimizing the cross-entropy loss function, enabling more accurate predictions of the real labels. Optimization algorithms such as Gradient Descent are frequently employed to update model parameters based on the gradient of the loss function, thereby reducing the loss function. The mathematical expression for sparse categorical cross-entropy loss function is presented in Equation (8):
L y , y ^ = i = 1 k y i log y i ^
where y is the probability distribution of the real label; y ^ is the predicted probability distribution of the model; k is the number of categories.
The softmax activation function was employed following the fully connected layer in DCFF-NET. This function is typically utilized in the output layer of multi-classification challenges, converting the model’s initial outputs into a probability distribution. Each category’s probability value range is between 0 and 1, with the total probability across all categories summing to 1. The formula for the softmax function is provided below:
σ x i = e x i j = 1 k e x j , i = 1,2 , , k
where x i is the value of the i-th element in the input vector; k is the number of categories; σ x i is the predicted probability of the i-th category.
The output from the softmax activation function delivers a probability value for each category, with the category exhibiting the highest probability designated as the final recognized category. The formula for this determination is specified below:
C l a s s X = a r g   m a x   P i , i = 1,2 , , k

3. Results and Analysis

3.1. Experimental Datasets and Implementation

The Indian Pines (IP) Aerial hyperspectral dataset, collected by the AVIRIS sensor on 12 June 1992, encompasses an area of 2.9 km × 2.9 km in northwest Indiana. The dataset features a pixel size of 145 × 145 (21,025 pixels), a spatial resolution of 20 m, a spectral range from 0.4 to 2.5 µm, and includes 16 land cover types, predominantly in agricultural regions. Originally comprising 220 bands, the dataset was refined to 200 bands after excluding 20 bands that captured atmospheric water absorption and exhibited low SNR. The retained data from these 200 bands were utilized in this experiment. Table 1 and Figure 3a,b illustrate the actual marking information of ground objects in this dataset. The sample distribution across different types is highly variable, with quantities ranging from 20 to 2455.
Aerial hyperspectral remote sensing images of Pavia University (PU) were obtained by the German airborne Reflective Optical Spectral Imager (ROSIS) on 8 July 2002 in the campus area of the University of Pavia, Italy. The spatial resolution of the data is 1.3 m, the image size is 610 × 340 pixels (207,400 pixels), and the spectral range of the image spans from 0.43 to 0.86 µm, encompassing 115 spectral channels. The dataset includes nine urban land cover types and 42,776 labeled pixels. The image on the right displays a false-color composite image of the Pavia University data and the actual overlay type of the surface. Due to noise, 12 noise bands were eliminated, leaving 103 bands to verify the performance of the proposed method. The PU dataset contains nine categories of ground objects, as shown in Figure 3c,d.
The Salinas dataset (SA) consists of hyperspectral remote sensing images of the Salinas Valley region in southern California, United States, acquired by the AVIRIS sensor. The image measures 512 × 217 pixels, with a spatial resolution of 3.7 m, a spectral resolution of between 9.7 and 12 nm, and a spectral range from 400 to 2500 nm. It includes 16 types of ground objects and 54,129 labeled samples. The dataset initially had 224 bands, but 20 bands affected by atmospheric water absorption and low signal-to-noise ratios were removed. The remaining 204 bands, retained after processing, were utilized in this experiment, as indicated in Figure 3e,f.
All experiments were conducted in identical environments using two computers. The configuration for these computers was as follows: the hardware platform included an Intel Core i7-9700K (8 cores/8 threading)/i7-1100F (8 cores/16 threading) processor with 12M/16M L3-cache/, 64 GB/128 G DDR4 memory at 3200 MHz serial speed, NVIDIA GeForce RTX 2070 GPU with 8 GB DDR5/RTX 3060 GPU with 12 GB DDR5 video memory, and a 1 TB HDD with 7200 RPM. The software platform comprised the Windows 10 Professional operating system, Keras 2.5.0 based on TensorFlow-gpu 2.5.0, and Python 3.7.7. The SF-Pixel model used the parameters specified in [51], while other models were trained using a batch size of 12, with IP models employing an Adam optimizer with a learning rate of 2 × 10−3, a decay of 1 × 10−4, and PU and SA models using a learning rate of 1 × 10−4 and a decay of 1 × 10−5.

3.2. Evaluation Criterion

Classification accuracy was estimated using evaluation indicators such as overall accuracy (OA), average accuracy (AA) for each category, user accuracy (UA) for each category, and kappa coefficient (KA) [64,65]. In this text, percentages represent the KA. The calculation formula is as follows:
O A = j 1 N T P j + T N j T P j + T N j + F P j + F N j
A A = 1 N j 1 N T P j T P j + F N j
U A j = T P j T P j + F N j
K A = p 0 p e 1 p e
where j is the class j ; T P j (True Positive) is the number of true examples, i.e., the number of samples that are correctly classified as positive examples; T N j (True Negative) is the number of true counterexamples, i.e., the number of samples that are correctly classified as negative examples; F P j (False Positive) is the number of false positives. The number of examples is the number of samples that were incorrectly classified as positive examples; F N j (False Negative) is the number of false counterexamples, that is, the number of samples that were incorrectly classified as negative examples; N is the total number of categories; p 0 is the accuracy of the classifier; p e is the random accuracy that the classifier can achieve.

3.3. Results Analysis Based on Feature Map

Traditional Naive Bayes (NB), K-Nearest Neighbors (KNN), Random Forest (RF), and Multi-Layer Perceptron (MLP), 1D-CNN, VGG16, ResNet50, and SF based on pixel-wise (SF-Pixel) deep learning methods were selected for comparative analysis with DCFF-NET. NB, KNN, and RF are all open-source machine learning libraries in Python. The data partitioning for random seeds was set to fixed values to achieve data recovery. NB, KNN, and RF use automatic searches (Grid Search) to select the optimal parameters with a 50%-off verification method. MLP, 1D-CNN, VGG16, ResNet50, and DCFF-Net methods were implemented using TensorFlow2.5.0, and a grid search algorithm was also employed to obtain optimal parameters, such as batch size, learning rate, and decay. SF-Pixel utilized publicly available code and made modifications according to the experiments described. The VGG16 and ResNet50 models and pre-trained weights were not used. Instead, training and prediction were conducted using feature maps. When using a grid search method, important parameter selection was performed first, reducing the parameter-space range and gradually narrowing the search range of the parameters based on the calculation results to improve search efficiency. This study uses different proportions of training samples for training and prediction. As the proportion of the training sample changed, the training process and performance of the model were also affected. Therefore, under different training sample proportions, the best parameter combination is sought to ensure the best performance and generalization ability of the model. The best parameters and optimal models were selected to predict and evaluate the three hyperspectral datasets.
The results are depicted in Table 2. When various classification methods use different training sample ratios, the accuracy of image prediction results increases significantly as the proportion of training samples increases. The DCFF-Net classification method performs well on the three datasets. The three different datasets exhibit higher performance than the lower training ratios in the case of 10%, 20%, and 30% training ratios. Except for the maximum value of AA, all other values achieve the maximum value and demonstrate the best classification performance overall. With 30% training samples, the OA, KA, and AA of the IP datasets are 86.68%, 85.05%, and 85.08%; the OA, KA, and AA of the PU dataset are 94.73%, 92.99%, and 92.60%; and the OA, KA, and AA of the SA dataset are 95.14%, 94.59%, and 97.48%.

3.3.1. Results of Indian Pines

Figure 4 and Table 3 show the classification result chart obtained using 30% of the training samples from the Indian Pines dataset. The UAs of each category in this dataset are listed in Table 3. It can be observed from the table that, out of the 16 land types, seven types achieved the best classification effects compared to other classification methods when using the DCFF-Net classification method. The OA, KA, and AA were 86.68%, 85.04%, and 85.08%, respectively, higher than other classification methods. The classification effect of SF-Pixel and VGG16 was less than that of DCFF-Net, with OA, KA, and AA at 85.39%, 83.39%, 82.80, 85.18%, 83.16%, and 83.21%, respectively. Secondly, ResNet50 also performed well. These four deep learning methods outperformed other classification methods. The optimal UA values for the 16 land types were scattered across various classification methods and were not concentrated in a specific classification method. In some categories, such as Corn-no Till, Corn, Oats, and others, DCFF-Net achieved a better classification effect. The NB classification methods showed significant variation in UA across different categories. For Alfalfa, Grass-Pasture-Mowed, and Oats the UA was 0.00% due to too few samples, while the highest, Hay-Windrowed, reached an AA of 99.40%. However, the average accuracy across the 16 categories was only 53.85%, far lower than several other methods.

3.3.2. Results of Pavia University

Figure 5 illustrates the classification result diagram obtained using 30% training samples on the Pavia University dataset. The OA, KA, and AA of the DCFF-Net model were 94.73%, 92.99%, and 92.60%, respectively, and these three indicators achieved the highest classification accuracy, as shown in Table 4. Trees and Bitumen’s UA were the highest compared to other methods. The 1D-CNN classification result’s OA was second only to DCFF-Net, reaching 93.97%. The categories Gravel, Painted Metal Sheets, Bare Soil, and Shadows were higher than other methods. For Gravel and Bitumen, no classification methods exceeded 90.00%. In Figure 1, Gravel, Bitumen, and Self-Blocking Bricks had similar ground features, leading to generally low UA in these categories. Among all comparison methods, the KNN classification method had the lowest accuracy.

3.3.3. Results of Salinas

Figure 6 presents the classification results obtained using 30% training samples on the Salinas dataset. The number of training samples is 16,231. Table 5 shows the user accuracy (UA) of various types in SA. DCFF-Net classification OA, KA, and AA were 95.14%, 94.59%, and 97.48%, respectively, better than other comparison methods. VGG16 and ResNet50 also achieved better classification effects, second only to DCFF-Net. Among the 16 types of land, the DCFF-Net classification results were the highest in the UA of six places. Except for the low UA in Celery, the classification accuracy in other categories was not significantly different from other classification methods. The Celery UA was extremely low. Figure 6 indicates that the characteristic graphic features of Celery and Broccoli_green_weeds_1 and Broccoli_green_weeds_2 were highly similar, making them difficult to distinguish effectively and significantly low accuracy. In the current classification methods, Grapes_untrained and Vineyard_untrained are difficult to distinguish. As shown in Table 5, the three methods—DCFF-Net, VGG16, and ResNet50—were significantly higher than other classification methods. DCFF-Net significantly increased the characteristics of similar objects, obtaining higher classification accuracy. Grapes_untrained and Vineyard_untrained achieved 98.96% and 91.71%, respectively. In other methods, the UAs of Grapes_untrained and Vineyard_untrained were lower than 90.00%, with NB, KNN, RF, MLP, 1D-CNN, and other methods achieving less than 80.00%.
The eight classification methods mentioned in the text are NB, KNN, RF, MLP, 1D-CNN, SF-Pixel, VGG16, ResNet50, and DCFF-Net. The first six methods used one-dimensional serial vector data as the model input data, while the latter three were based on two-dimensional images. The latter three methods extracted image features through 2D-CNN, classifying the data into different categories. In the IP data concentration, including Corn-Min Till, Corn, Oats, and Buildings-Grass-Trees-Drives, and on the SA dataset, including Grapes_untrained and Vineyard_untrained, the accuracy of the latter three methods was significantly higher than the first five methods. This indicates that 2D-CNN can achieve better classification effects in some easily mixed categories. However, for some categories, the first five methods performed better than the latter three methods, such as Soybean-Min Till, Woods, and Stone-Steel-Towers in the IP dataset. More suitable classification methods can be selected based on these observations in practical applications.

3.4. Results Analysis Based on Pixel-Patched

An additional evaluation was conducted using the pixel-patched method to examine the classification effectiveness of DCFF-NET and assess the performance of the proposed model. The methods NB, KNN, RF, MLP, 1D-CNN, and SF-Pixel use one-dimensional vector data and cannot directly use pixels as input data. Therefore, a comparative analysis was conducted using HybridSN, 3D-CNN, A2S2KNet, and SF based on patch-wise (SF-Patch) models. The pixel patch size was uniformly extracted at 24 × 24 × C as the input data, where C is the number of image channels, and 10% of the categories were selected as training samples. The models underwent a fine-tuning process to facilitate the utilization of VGG16 and ResNet50 models for training and testing with the provided input data. After model training was completed, the optimal model was selected for testing, using 90% of the data as test data, performing the test 10 times. The average value and standard deviation were selected as the test results, as shown in Table 6, Table 7 and Table 8. Figure 7 shows the classification map for each classification method. The OA and KA on the IP dataset were 98.15% and 97.89%, respectively, second only to A2S2K but higher than all other methods. The AA value of 97.73% was the highest of all methods. The OA, KA, and AA of 3D-CNN were the lowest, with values of 93.18%, 92.26%, and 94.52%, respectively. The DCFF-Net model achieved the best classification performance regarding OA, KA, and AA on the PU and SA datasets. The results on the two datasets were as follows: 99.86%, 99.82%, 99.79%, 99.98%, 99.98%, and 99.94%, respectively.
In addition, all comparison methods, besides SF-Patch on the IP dataset, also achieved very high classification accuracy. The SF-Patch model showed the lowest accuracy among three datasets, especially noticeable in the IP dataset where its performance lagged behind other classification methods due to the dataset’s small sample size and uneven distribution. A detailed analysis of the SF-Patch model’s performance on the IP dataset was conducted by varying the proportion of training samples—specifically selecting 15%, 20%, and 30%. With an increase in the proportion of training samples, the model’s OA improved significantly, reaching 94.67%, 96.35%, and 98.41%, respectively.
The pixel-patched block contains surrounding pixel information, specifically spectral–spatial information, which provides more discriminative cues for the target pixel. Using pixel-patched block data as the input for the model can effectively improve classification accuracy, as mentioned earlier, but it can also lead to potential label information leakage. In addition, the larger the pixel patch, the more spatial information it contains, however, it will significantly increase computational complexity.

4. Discussion

4.1. Effect of Different Filling Methods

This study discussed the impact of feature map-filling characteristics on the classification accuracy of high spectral images. The spectral feature diagram employs a variety of filling methods. In Figure 8, the initial graphic (NotFill) remains unfilled, displaying a default output of white with a pixel value of 255. The second method entails internal filling in yellow, while the exterior remains unfilled (InnerFill), with the default external color being white. The third method involves internal yellow and external green filling (BothFill). The spectral characteristic graph type is determined in blue for all three methods. The tests on the three datasets were trained on the same classification method and parameters of the three characteristic diagrams for training and verification, and the IP dataset epochs were set at 75, while the other two dataset epochs were set at 100%, 10%, 20%, and 30% of the three different proportional training samples.
The verification accuracy curve is illustrated in Figure 9, and the accuracy results are listed in Table 9. The results indicate that the method involving internal and external filling yields the highest classification accuracy of the feature diagram. During the training process, the convergence rates of the three characteristic diagrams were different, and the curve volatility was inconsistent. The third characteristic diagram exhibits fewer fluctuations in training accuracy, better stability, and the highest final verification accuracy. Therefore, this filling method was selected during the spectral feature diagram conversion process. The first graphic displays general fluctuations in training accuracy and a significantly slower convergence rate, whereas the second method’s performance lies between the first and third methods.

4.2. Effect of the Different Percentages of Training Samples for DCFF-NET

Different proportions of sample tag data affected classification accuracy performance significantly. Different proportions of labeled data from three datasets were selected to analyze the accuracy performance of each classification method, utilizing OA as the evaluation indicator. Proportions of 5%, 7%, 10%, 15%, 20%, 25%, and 30% from the Indian Pines dataset were chosen. In addition, 0.5%, 1%, 3%, 5%, 7%, 10%, 20%, and 30% from the Pavia University and Salinas datasets were examined. Figure 10 and Table 10 show that classification accuracy improves as the training sample proportion increases. For the Indian Pines dataset, the DCFF-Net classification accuracy surpassed other methods when the training sample size exceeded 7%. When the training set size was below 7%, VGG16 achieved superior overall classification accuracy. For the Pavia University and Salinas datasets, DCFF-Net outperformed other methods when the sample ratio exceeded 3%. In contrast, 1D-CNN showed better overall classification accuracy when the training sample was below 3%.

4.3. Ablation Analysis

The DCFF-NET model primarily utilizes the SIE and DSFE modules, with the SIE module as the core component, and integrates the DSFE module to enhance classification performance. The effectiveness of these modules was verified through individual and combined tests. When evaluating the performance of the SIE module, only the SIE module in the Figure 1a CIF module was retained without integrating the DSFE module. In contrast, the DSFE module assessment replaced the SIE module in the Figure 1a CIF module while retaining only the primary model for experimentation. This section replicated all experiments ten times independently with random seeds under consistent conditions. Similar to previous sections, experiments on the IP, PU, and SA datasets used three different sample proportions: 30% (Exp 1), 20% (Exp 2), and 10% (Exp 3). The final values were averaged and are displayed in Table 11.
Under three experimental conditions, the fusion of the SIE and DSFE modules enhanced classification results compared to individual modules. On the IP dataset, the fused classification results for Experiment 1 were as follows: OA 86.06%, KA 84.12%, and AA 84.11%. The fused model showed improvements of 1.14% in OA, 1.30% in KA, and 1.35% in AA over the DSFE module. Compared to the SIE module, improvements were 1.34% in OA, 1.56% in KA, and 3.28% in AA. The fused model also exhibited lower variance, indicating a more stable performance. Significant improvements in classification accuracy were observed under other experimental conditions as well. The improvement in classification results from the fused model on the PU and SA datasets was less pronounced than that on the IP dataset. For Experiment 1, the OA and KA of the PU dataset improved by only 0.14% and 0.04%, respectively, compared to the DSFE, with a slight decrease in AA. Compared to the SIE, the OA, KA, and AA improved by 0.40%, 0.41%, and 0.47%, respectively. For the SA dataset, the improvements were 0.27% in OA, 0.30% in KA, and 0.14% in AA compared to the DSFE, and 0.12%, 0.07%, and 0.19% compared to the SIE.

5. Conclusions

This study employed a novel approach by transforming all pixels in hyperspectral images into standardized spectral feature maps using polar co-ordinates. Unlike methods that rely on cube pixel patches, this approach utilized individual pixel spectral information without considering spatial adjacency. These feature maps served as input data for training and prediction using the proposed DCFF-Net model. The DCFF-Net included core functionality modules: SIE, DSFE, and CIF. The CIF module achieved deep integration of the SIE and DSFE modules, effectively enhancing the model’s performance. This study compared the model with other advanced classification methods using polar co-ordinate feature maps and pixel patches, demonstrating its classification performance. After converting the hyperspectral pixels into polar co-ordinate feature maps, the process enhanced the image features but also increased the data volume, which raised the computational load of the model to a certain extent, necessitating additional computational resources. Future research will focus on enhancing the model’s overall performance by reducing the data’s computational load.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/rs16163002/s1.

Author Contributions

Formal analysis, Z.C., Y.W. and X.W. (Xiaoyan Wang); funding acquisition, X.W. (Xinsheng Wang); investigation, Z.C.; methodology, Z.C., Y.C., Y.W. and X.W. (Xinsheng Wang); project administration, X.W. (Xinsheng Wang); software, Z.C. and Y.C.; validation, Z.C., Y.C., Y.W. and X.W. (Xiaoyan Wang); visualization, Z.C., Y.C. and Z.X.; writing—original draft, Z.C., Y.C., Y.W. and X.W. (Xiaoyan Wang); writing—review and editing, Z.C., Y.C., Y.W. and Z.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the “Special projects for technological innovation in Hubei”(2018ABA078), the “Research and demonstration of precision agricultural monitoring technology based on sky and earth cooperative observation”; open fund of “Hubei Key Laboratory of Regional Development and Environmental Response, Hubei University”(2019(B)001), and the “Research on Classification Method of Hyperspectral Remote Sensing Data Based on Graph-Spatial Features”.

Data Availability Statement

Data can be provided upon request (Supplementary Materials).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Plaza, A.; Benediktsson, J.A.; Boardman, J.W. Recent advances in techniques for hyperspectral image processing. Remote Sens. Environ. 2009, 113, S110–S122. [Google Scholar] [CrossRef]
  2. Ghamisi, P.; Couceiro, M.S.; Benediktsson, J.A. Deep learning-based classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2383–2395. [Google Scholar]
  3. Liu, B.; Zhang, L.; Zhang, L.; Du, Q. Hyperspectral image classification using deep pixel-pair features. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3658–3668. [Google Scholar] [CrossRef]
  4. Belhumeur, P.N.; Hespanha, J.P.; Kriegman, D.J. Eigenfaces vs. fisherfaces: Recognition using class specific linear projection. IEEE Trans. Pattern Anal. Mach. Intell. 1997, 19, 711–720. [Google Scholar] [CrossRef]
  5. Tan, S.; Chen, X.; Zhang, D. Kernel discriminant analysis for face recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2106–2112. [Google Scholar]
  6. Li, Y.; Chen, Y.; Jiang, J. A Novel KNN Classifier Based on Multi-Feature Fusion for Hyperspectral Image Classification. Remote Sens. 2020, 12, 745. [Google Scholar]
  7. Li, X.; Liang, Y. An Improved KNN Algorithm for Intrusion Detection Based on Feature Selection and Data Augmentation. IEEE Access. 2021, 9, 17132–17144. [Google Scholar]
  8. Domingos, P.; Pazzani, M. On the optimality of the simple Bayesian classifier under zero-one loss. Mach. Learn. 1997, 29, 103–130. [Google Scholar] [CrossRef]
  9. Zhang, H.; Wang, X. A survey on naive Bayes classifiers. Neurocomputing 2020, 399, 14–23. [Google Scholar]
  10. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Spectral–spatial hyperspectral image segmentation using subspace multinomial logistic regression and Markov random fields. IEEE Trans. Geosci. Remote Sens. 2012, 50, 809–823. [Google Scholar] [CrossRef]
  11. Mo, Y.; Li, M.; Liu, Y. Multinomial logistic regression with pairwise constraints for multi-label classification. IEEE Access 2020, 8, 74005–74016. [Google Scholar]
  12. Zhang, C.; Li, J. Multinomial logistic regression with manifold regularization for image classification. Pattern Recognit. Lett. 2021, 139, 55–63. [Google Scholar]
  13. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 40, 1778–1790. [Google Scholar] [CrossRef]
  14. Cervantes, J.; Garcia-Lamont, F.; Rodriguez-Mazahua, L.; Lopez, A. A comprehensive survey on support vector machine classification: Applications, challenges and trends. Neurocomputing 2020, 408, 189–215. [Google Scholar] [CrossRef]
  15. Zhong, Y.; Hu, X.; Luo, C.; Wang, X.; Zhao, J.; Zhang, J. WHU-Hi: UAV-borne hyperspectral with high spatial resolution (H2) benchmark datasets classifier for precise crop identification based on deep convolutional neural network with CRF. Remote Sens. Environ. 2020, 250, 112012. [Google Scholar] [CrossRef]
  16. Wang, Y.; Yang, B.; Chen, Y.; Liang, F.; Dong, Z. JoKDNet: A joint keypoint detection and description network for large-scale outdoor TLS point clouds registration. Int. J. Appl. Earth Obs. Geoinf. 2021, 104, 102534. [Google Scholar] [CrossRef]
  17. Jia, K.; Li, Q.Z.; Tian, Y.C.; Wu, B.F. A review of classification methods of remote sensing imagery. Spectrosc. Spectr. Anal. 2011, 31, 2618–2623. [Google Scholar]
  18. Liu, S.; Li, T.; Sun, Z. Discrimination of tea varieties using hyperspectral data based on wavelet transform and partial least squares-discriminant analysis. Food Chem. 2020, 325, 126914. [Google Scholar]
  19. He, L.; Huang, W.; Zhang, X. Hyperspectral image classification with principal component analysis and support vector machine. Neurocomputing 2015, 149, 962–971. [Google Scholar]
  20. Beirami, B.A.; Pirbasti, M.A.; Akbari, V. Fractal-Based Ensemble Classification System for Hyperspectral Images. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar] [CrossRef]
  21. Cheng, H.; Zhang, Y.; Yang, Z. Spectral feature extraction based on key point detection and clustering algorithm. IEEE Access 2019, 7, 43100–43107. [Google Scholar]
  22. Huang, H.; Chen, X.; Guo, L. A novel method for spectral angle classification based on the support vector machine. PLoS ONE 2020, 15, e0237772. [Google Scholar]
  23. Wu, D.; Zhang, L.; Shen, X. A spectral curve matching algorithm based on dynamic programming and frequency-domain filtering. IEEE Trans. Geosci. Remote Sens. 2021, 59, 4293–4307. [Google Scholar]
  24. Wang, Y.; Wang, C. Spectral curve shape index: A new spectral feature for hyperspectral image classification. J. Appl. Remote Sens. 2021, 15, 016524. [Google Scholar]
  25. Huang, J.; Zhang, J.; Li, Y. Discrimination of tea varieties using near infrared spectroscopy and chemometrics. J. Food Eng. 2015, 144, 75–80. [Google Scholar]
  26. Huang, M.; Gong, D.; Zhang, L.; Lin, H.; Chen, Y.; Zhu, D.; Xiao, C.; Altan, O. Spatiotemporal Dynamics and Forecasting of Ecological Security Pattern under the Consideration of Protecting Habitat: A Case Study of the Poyang Lake Ecoregion. Int. J. Digit. Earth 2024, 17, 2376277. [Google Scholar] [CrossRef]
  27. Tao, Y.; Liu, F.; Pan, M. Rapid identification of intact paddy rice varieties using near-infrared spectroscopy and chemometric analysis. J. Cereal Sci. 2015, 62, 59–64. [Google Scholar]
  28. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar]
  29. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  30. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  31. Hinton, G.E. Learning multiple layers of representation. Trends Cogn. Sci. 2007, 11, 428–434. [Google Scholar] [CrossRef] [PubMed]
  32. Hinton, G.E.; Osindero, S.; Teh, Y.W. A fast learning algorithm for deep belief nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef]
  33. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  34. Graves, A.; Mohamed, A.R.; Hinton, G.E. Speech recognition with deep recurrent neural networks. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 6645–6649. [Google Scholar]
  35. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed representations of words and phrases and their compositionality. Adv. Neural Inf. Process. Syst. 2013, 26, 3111–3119. [Google Scholar]
  36. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. Int. J. Comput. Vis. 2015, 113, 136–158. [Google Scholar]
  37. Chatfield, K.; Simonyan, K.; Vedaldi, A.; Zisserman, A. Return of the devil in the details: Delving deep into convolutional nets. In Proceedings of the British Machine Vision Conference (BMVC), Nottingham, UK, 1–5 September 2014; pp. 1–12. [Google Scholar]
  38. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  39. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  40. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; pp. 630–645. [Google Scholar]
  41. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1492–1500. [Google Scholar]
  42. Zhang, H.; Wu, C.; Zhang, J.; Zhu, Y.; Zhang, Z.; Lin, H.; Sun, Y. ResNeSt: Split-Attention Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 163–172. [Google Scholar]
  43. Hu, W.; Huang, Y.; Wei, L.; Zhang, F.; Li, H. Deep convolutional neural networks for hyperspectral image classification. J. Sens. 2015, 2015, 258619. [Google Scholar] [CrossRef]
  44. Mao, J.; Xu, W.; Yang, Y.; Wang, J.; Huang, Z.; Yuille, A. Deep Captioning with Multimodal Recurrent Neural Networks (m-RNN). arXiv 2014, arXiv:1412.6632. [Google Scholar]
  45. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. Hybridsn: Exploring 3-d–2-d cnn feature hierarchy for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 17, 277–281. [Google Scholar] [CrossRef]
  46. Hamida, A.B.; Benoit, A.; Lambert, P.; Amar, C.B. 3-d deep learning approach for remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4420–4434. [Google Scholar] [CrossRef]
  47. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectralspatial residual network for hyperspectral image classification: A 3-d deep learning framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
  48. Qing, Y.; Huang, Q.; Feng, L.; Qi, Y.; Liu, W. Multiscale Feature Fusion Network Incorporating 3D Self-Attention for Hyperspectral Image Classification. Remote Sens. 2022, 14, 742. [Google Scholar] [CrossRef]
  49. Qing, Y.; Liu, W. Hyperspectral image classification based on multi-scale residual network with attention mechanism. Remote Sens. 2021, 13, 335. [Google Scholar] [CrossRef]
  50. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep Feature Extraction and Classification of Hyperspectral Images Based on Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef]
  51. Hong, D.; Han, Z.; Yao, J.; Chen, Y.; Li, W. SpectralFormer: Rethinking hyperspectral image classification with transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5518615. [Google Scholar] [CrossRef]
  52. Roy, S.K.; Deria, A.; Shah, C.; Jiang, L.; Li, W. Spectral–spatial morphological attention transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5503615. [Google Scholar] [CrossRef]
  53. Zhang, Y.; Li, W.; Sun, W.; Wang, Y.; Liu, H. Single-source domain expansion network for cross-scene hyperspectral image classification. IEEE Trans. Image Process. 2023, 32, 1498–1512. [Google Scholar] [CrossRef]
  54. Liu, H.; Li, W.; Xia, X.G.; Chen, Y.; Zhang, Y. Multi-Area Target Attention for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5524916, Advance online publication. [Google Scholar]
  55. Liu, Y.; Li, X.; Zhang, Z.; Ma, Y. ESSAformer: Efficient transformer for hyperspectral image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Paris, France, 1–6 October 2023; pp. 17642–17651. [Google Scholar]
  56. Zhang, C.; Zhang, M.; Li, Y.; Gao, X.; Shi, Q. Difference curvature multidimensional network for hyperspectral image super-resolution. Remote Sens. 2021, 13, 3455. [Google Scholar] [CrossRef]
  57. Zhang, M.; Xu, J.; Zhang, J.; Zhao, H.; Shang, W.; Gao, X. SPH-Net: Hyperspectral Image Super-Resolution via Smoothed Particle Hydrodynamics Modeling. IEEE Trans. Cybern. 2024, 54, 4150–4163. [Google Scholar] [CrossRef] [PubMed]
  58. Xia, Z.; Liu, Y.; Li, X.; Zhu, X.; Ma, Y.; Li, Y.; Hou, Y.; Qiao, Y. SCPNet: Semantic Scene Completion on Point Cloud. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 17642–17651. [Google Scholar]
  59. Wu, J.; Sun, X.; Qu, L.; Tian, X.; Yang, G. Learning Spatial–Spectral-Dimensional-Transformation-Based Features for Hyperspectral Image Classification. Appl. Sci. 2023, 13, 8451. [Google Scholar] [CrossRef]
  60. Wei, Y.; Feng, J.; Liang, X.; Cheng, M.-M.; Zhao, Y.; Yan, S. Object Region Mining with Adversarial Erasing: A Simple Classification to Semantic Segmentation Approach. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  61. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  62. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015. [Google Scholar]
  63. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on ImageNet classification. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  64. Bhattacharyya, C.; Kim, S. Black Ice Classification with Hyperspectral Imaging and Deep Learning. Appl. Sci. 2023, 13, 11977. [Google Scholar] [CrossRef]
  65. Li, J.; Huang, X.; Tu, L. WHU-OHS: A benchmark dataset for large-scale hersepctral image classification. Int. J. Appl. Earth Obs. Geoinf. 2022, 113, 103022. [Google Scholar] [CrossRef]
Figure 1. Feature maps of three datasets. (A representative feature map for each category was selected and displayed from three different datasets.)
Figure 1. Feature maps of three datasets. (A representative feature map for each category was selected and displayed from three different datasets.)
Remotesensing 16 03002 g001
Figure 2. DCFF-NET network architectures. (Identical processing operations were represented using the same colour background, e.g. orange for convolution operations, green for batch normalisation, blue for activation, light green for pooling, etc.)
Figure 2. DCFF-NET network architectures. (Identical processing operations were represented using the same colour background, e.g. orange for convolution operations, green for batch normalisation, blue for activation, light green for pooling, etc.)
Remotesensing 16 03002 g002
Figure 3. Hyperspectral dataset: Indian Pines (a,b), Pavia University (c,d), Salinas (e,f).
Figure 3. Hyperspectral dataset: Indian Pines (a,b), Pavia University (c,d), Salinas (e,f).
Remotesensing 16 03002 g003
Figure 4. Predicted classification map of 30% samples for training. (a) Three bands false color composite. (b) Ground truth data. (c) NB. (d) KNN. (e) RF. (f) MLP. (g) 1DCNN. (h) SF-Pixel. (i) VGG16. (j) Resnet50. (k) DCFF-NET.
Figure 4. Predicted classification map of 30% samples for training. (a) Three bands false color composite. (b) Ground truth data. (c) NB. (d) KNN. (e) RF. (f) MLP. (g) 1DCNN. (h) SF-Pixel. (i) VGG16. (j) Resnet50. (k) DCFF-NET.
Remotesensing 16 03002 g004
Figure 5. Predicted classification map of 30% samples for training. (a) Three bands false color composite. (b) Ground truth data. (c) NB. (d) KNN. (e) RF. (f) MLP. (g) 1DCNN. (h) SF-Pixel. (i) VGG16. (j) Resnet50. (k) DCFF-NET.
Figure 5. Predicted classification map of 30% samples for training. (a) Three bands false color composite. (b) Ground truth data. (c) NB. (d) KNN. (e) RF. (f) MLP. (g) 1DCNN. (h) SF-Pixel. (i) VGG16. (j) Resnet50. (k) DCFF-NET.
Remotesensing 16 03002 g005
Figure 6. Predicted classification map of 30% samples for training. (a) Three bands false color composite. (b) Ground truth data. (c) NB. (d) KNN. (e) RF. (f) MLP. (g) 1DCNN. (h) SF-Pixel. (i) VGG16. (j) Resnet50. (k) DCFF-NET.
Figure 6. Predicted classification map of 30% samples for training. (a) Three bands false color composite. (b) Ground truth data. (c) NB. (d) KNN. (e) RF. (f) MLP. (g) 1DCNN. (h) SF-Pixel. (i) VGG16. (j) Resnet50. (k) DCFF-NET.
Remotesensing 16 03002 g006
Figure 7. Classification map of IP (A) PU (B) and SA (C) based on patched-based input. (a) False color composite. (b) Ground truth. (c) VGG16. (d) Resnet50. (e) 3-DCNN. (f) HybridSN. (g) A2S2K. (h) SF-Patch. (i) DCFF-NET.
Figure 7. Classification map of IP (A) PU (B) and SA (C) based on patched-based input. (a) False color composite. (b) Ground truth. (c) VGG16. (d) Resnet50. (e) 3-DCNN. (f) HybridSN. (g) A2S2K. (h) SF-Patch. (i) DCFF-NET.
Remotesensing 16 03002 g007
Figure 8. Different filling methods. ((a). Feature maps were neither filled inside nor outside; (b). Feature maps were filled internally rather than externally; (c). Feature maps were filled both internally and externally).
Figure 8. Different filling methods. ((a). Feature maps were neither filled inside nor outside; (b). Feature maps were filled internally rather than externally; (c). Feature maps were filled both internally and externally).
Remotesensing 16 03002 g008
Figure 9. Three different filling methods training accuracies curve diagrams.
Figure 9. Three different filling methods training accuracies curve diagrams.
Remotesensing 16 03002 g009
Figure 10. Effect of the different numbers of training samples for different methods.
Figure 10. Effect of the different numbers of training samples for different methods.
Remotesensing 16 03002 g010
Table 1. Detailed information of three datasets.
Table 1. Detailed information of three datasets.
ClassesIndian PinesSalinasPavia University
NamesSamplesNamesSamplesNamesSamples
1Alfalfa46Brocoli_green_weeds_12009Asphalt6631
2Corn-no till1428Brocoli_green_weeds_23726Meadows18,649
3Corn-min till830Fallow1976Gravel2099
4Corn237Fallow_rough_plow1394Trees3064
5Grass-pasture483Fallow_smooth2678Painted metal sheets1345
6Grass-trees730Stubble3959Bare Soil5029
7Grass-pasture-mowed28Celery3579Bitumen1330
8Hay-windrowed478Grapes_untrained11,271Self-Blocking Bricks3682
9Oats20Soil_vinyard_develop6203Shadows947
10Soybean-no till972Corn_senesced_green_weeds3278--
11Soybean-min till2455Lettuce_romaine_4wk1068--
12Soybean-clean593Lettuce_romaine_5wk1927--
13Wheat205Lettuce_romaine_6wk916--
14Woods1265Lettuce_romaine_7wk1070--
15Buildings-Grass-Trees-Drives386Vinyard_untrained7268--
16Stone-Steel-Towers93Vinyard_vertical_trellis1807--
Total Samples10,249 54,129 42,956
Table 2. Accuracies (%) of each method with different training samples on three datasets.
Table 2. Accuracies (%) of each method with different training samples on three datasets.
MethodsTrain
/Test
IPPUSA
OAKAAAOAKAAAOAKAAA
NB30%/70%71.5866.9653.8590.9187.7987.3991.1390.1094.22
KNN79.9877.1279.0990.2886.9488.2992.2491.3696.10
RF83.1780.6673.6791.9189.1789.3993.3992.6396.39
MLP77.6574.5277.1293.8891.8690.3192.3191.4395.67
1DCNN79.9076.9476.6593.9792.0092.4293.3492.5896.91
SF-Pixel85.4983.3982.8093.9592.0292.5692.7191.8796.33
VGG1685.1883.1683.2192.1989.6489.5694.2593.5996.66
Resnet5082.1379.6077.7193.4491.2890.4994.6294.0197.23
DCFF-NET86.6885.0485.0894.7392.9992.6095.1494.5997.48
NB20%/80%66.1460.5248.6389.3685.6185.2490.2189.0993.37
KNN78.1275.0076.3889.3485.6187.2591.3890.4095.54
RF80.7577.8566.6091.0487.9588.0592.6091.7595.90
MLP75.2071.5371.9592.3789.9590.5691.3490.3494.60
1DCNN73.8869.5968.6393.0790.8289.7992.3391.4396.29
SF-Pixel82.9180.3876.8792.7790.5091.4191.6290.6895.80
VGG1683.4481.1283.2592.8390.4889.6492.5591.6895.73
Resnet5075.8472.4473.6493.2391.0090.8993.8093.1096.48
DCFF-NET84.0581.77 79.90 94.1192.1891.8694.2493.5896.81
NB10%/90%58.0550.0840.9385.8580.6779.5288.2286.8691.49
KNN74.7671.2070.9787.5383.1285.0190.0388.9094.42
RF75.8772.2760.4289.5285.8685.7991.3190.3194.93
MLP69.2064.9360.6092.0289.4089.3490.4189.3594.91
1DCNN69.3664.9864.4892.7190.3990.8591.5590.5894.98
SF-Pixel75.2771.4764.4490.7187.7289.5890.0088.8794.89
VGG1677.8874.7372.8091.5988.8488.6192.1291.2395.50
Resnet5070.9066.5464.1491.5188.7389.1892.3591.4794.86
DCFF-NET78.2175.0784.3692.5690.4090.3993.0092.2095.66
Table 3. Accuracies (%) of labeled samples per class for the Indian Pines dataset (30% Train Samples).
Table 3. Accuracies (%) of labeled samples per class for the Indian Pines dataset (30% Train Samples).
Classes NameTrain/TestNBKNNRFMLP1DCNNSF-PixelVGG16Resnet50DCFF-NET
Alfalfa13/330.0069.7066.6782.6115.6278.7990.3292.0092.00
Corn-no till428/100060.6070.0076.9047.6273.3080.2077.2771.4983.48
Corn-min till249/58139.0765.5860.2470.2472.9878.0186.7082.3084.42
Corn71/16613.2559.0456.0270.0464.4663.2580.2973.0386.40
Grass-pasture144/33970.8093.5189.0983.6484.6292.5596.9992.5798.65
Grass-trees219/51197.0696.0996.6789.8695.5096.8893.4395.5995.71
Grass-pasture-mowed8/200.0085.0040.0085.7195.0087.5076.4966.6772.03
Hay-windrowed143/33599.4098.5198.8196.4494.0397.5792.3191.4392.86
Oats6/140.0071.4321.4325.0078.5758.3380.9180.3083.78
Soybean-no till291/68166.8179.1582.0973.9760.1581.6572.4167.0877.41
Soybean-min till736/171987.3282.3290.8782.0084.4787.6266.4661.6375.00
Soybean-clean177/41638.2262.5069.7177.0772.5376.3492.2692.6692.66
Wheat61/14493.75100.0095.8399.5198.6098.5597.6094.4896.84
Woods379/88697.9791.9995.0397.0894.1396.9472.2263.6481.82
Buildings-Grass-Trees-Drives115/27116.9754.2457.5660.6256.3055.2798.5899.70100.00
Stone-Steel-Towers27/6680.3086.3681.8292.4786.1595.3857.1418.7556.25
OA3067/718271.5879.9883.1777.6579.9085.3985.1882.1386.68
KA66.9677.1280.6674.5276.9483.3983.1679.6085.04
AA53.8579.0973.6777.1276.6582.8083.2177.7185.08
Table 4. Accuracies (%) of labeled samples per class for the Pavia University dataset (30% train samples).
Table 4. Accuracies (%) of labeled samples per class for the Pavia University dataset (30% train samples).
Class NameTrain/TestNBKNNRFMLP1DCNNSF-PixelVGG16Resnet50DCFF-NET
Asphalt1989/464290.6989.1992.2092.4594.6496.4193.0894.4693.84
Meadows5594/13,05598.4797.8897.9497.8297.6595.0296.5197.3697.54
Gravel629/147066.8774.5674.1565.6688.1688.2171.3473.1377.70
Trees919/214590.2688.2191.9395.2294.0395.1491.0195.4496.00
Painted metal sheets403/94299.1599.4799.1599.26100.0099.5799.3799.3899.17
Bare Soil1508/352174.0769.3377.0890.6292.5396.1389.7590.5892.02
Bitumen399/93177.2387.7681.9575.6886.4777.8880.5377.6388.05
Self-Blocking Bricks1104/257889.9188.1790.1193.1578.2785.4784.7587.0188.35
Shadows284/66399.85100.00100.0099.25100.00100.0099.7099.3999.70
OA12,829/29,94790.9190.2891.9192.7093.9793.9592.1993.4494.73
KA87.7986.9489.1790.2292.0092.0289.6491.2892.99
AA87.3988.2989.3989.9092.4292.5689.5690.4992.60
Table 5. Accuracies (%) of labeled samples per class for the Salinas dataset (30% train samples).
Table 5. Accuracies (%) of labeled samples per class for the Salinas dataset (30% train samples).
Classes NameTrain/TestNBKNNRFMLP1DCNNSF-PixelVGG16Resnet50DCFF-NET
Brocoli_green_weeds_1602/140797.5199.2999.9399.36100.0099.5899.8699.7999.59
Brocoli_green_weeds_21117/260999.1699.9299.9699.4399.9699.9696.1895.2596.82
Fallow592/138496.4699.9399.4291.0499.4299.3698.4399.0898.69
Fallow_rough_plow418/97699.0899.3999.3999.0899.1899.4798.1299.4199.48
Fallow_smooth803/187596.4898.5198.7798.7299.2099.3794.7199.0699.22
Stubble1187/277299.4299.6499.7599.8999.9699.6798.6798.9698.45
Celery1073/250699.2099.6099.6899.8099.8899.7680.9681.3680.64
Grapes_untrained3381/789087.9385.0289.9179.6286.4390.4397.1798.6598.96
Soil_vinyard_develop1860/434399.0699.5299.3698.8399.9399.9198.6699.7799.92
Corn_senesced_green_weeds983/229591.4294.1294.5192.8198.3095.7797.3397.5998.76
Lettuce_romaine_4wk320/74889.4497.8695.9998.4098.93100.0099.3999.2899.17
Lettuce_romaine_5wk578/134999.5699.8599.4185.1799.48100.0098.9799.0599.31
Lettuce_romaine_6wk274/64297.8298.7598.4497.6699.3899.8499.7899.7499.93
Lettuce_romaine_7wk321/74992.7996.6697.0697.4697.3394.7199.8899.6899.52
Vinyard_untrained2180/508865.3371.2972.2778.8373.9264.3488.6189.5191.71
Vinyard_vertical_trellis542/126596.9298.2698.4298.1099.2199.0499.8299.4599.58
OA16,231/37,89891.1392.2493.3991.1393.3492.7194.2594.6295.14
KA90.1091.3692.6390.1492.5891.8793.5994.0194.59
AA94.2296.1096.3994.6496.9196.3396.6697.2397.48
Table 6. Accuracies (%) of Indian Pines dataset.
Table 6. Accuracies (%) of Indian Pines dataset.
Classes NameTrain/TestVGG16Resnet503-DCNNHybridSNA2S2KSF-PatchDCFF-NET
Alfalfa4/42100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.0093.50 ± 7.0474.36 + 16.0498.60 ± 1.15
Corn-no till142/128696.68 ± 0.1997.72 ± 0.1090.86 ± 0.2094.51 ± 0.2097.45 ± 2.6982.22 + 3.4198.03 ± 0.14
Corn-min till83/74789.95 ± 0.2696.48 ± 0.2078.99 ± 0.2993.47 ± 0.2997.54 ± 1.6387.75 + 4.6399.62 ± 0.08
Corn23/21489.69 ± 0.6384.84 ± 0.6189.65 ± 1.1669.14 ± 1.1698.09 ± 1.3089.95 + 9.6795.54 ± 0.48
Grass-pasture48/43583.01 ± 0.8489.3 ± 0.3992.94 ± 0.2494.3 ± 0.2499.66 ± 0.3188.01 + 2.4787.01 ± 0.40
Grass-trees73/65799.08 ± 0.1097.56 ± 0.18100.00 ± 0.0099.02 ± 0.1099.28 ± 0.9196.73 + 3.27100.00 ± 0.00
Grass-pasture-mowed2/26100.00 ± 0.00100.00 ± 0.0088.54 ± 0.00100.00 ± 0.0090.63 ± 13.2668.00 + 33.5100.00 ± 0.00
Hay-windrowed47/431100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.0099.57 ± 0.6094.68 + 2.2999.77 ± 0.00
Oats2/18100.00 ± 0.00100.00 ± 0.0095.54 ± 0.00100.00 ± 0.0072.22 ± 20.7966.67 + 27.4100.00 ± 0.00
Soybean-no till97/87595.85 ± 0.0898.29 ± 0.12100.00 ± 0.2095.94 ± 0.2097.59 ± 0.9784.25 + 3.0996.90 ± 0.14
Soybean-min till245/221098.63 ± 0.0899.57 ± 0.0588.44 ± 0.0896.5 ± 0.0899.11 ± 0.5192.72 + 1.6699.25 ± 0.07
Soybean-clean59/53491.56 ± 0.3295.24 ± 0.2393.28 ± 0.4089.1 ± 0.4098.52 ± 1.3271.40 + 3.8595.47 ± 0.22
Wheat20/18599.61 ± 0.2598.97 ± 0.17100.00 ± 0.00100.00 ± 0.0097.47 ± 1.4694.51 + 5.45100.00 ± 0.00
Woods126/113996.14 ± 0.2099.82 ± 0.0099.91 ± 0.1298.9 ± 0.1299.55 ± 0.5196.74 + 1.2499.84 ± 0.03
Buildings-Grass-Trees-Drives38/34899.45 ± 0.0999.02 ± 0.18100.00 ± 0.0097.91 ± 0.1795.04 ± 3.2077.84 + 7.96100.00 ± 0.00
Stone-Steel-Towers9/8491.22 ± 0.8083.69 ± 1.1594.16 ± 1.2485.64 ± 1.2494.36 ± 6.1075.00 + 16.1793.67 ± 0.80
OA1018/923195.82 ± 0.07597.6 ± 0.03993.18 ± 0.06995.46 ± 0.06398.29 ± 0.34588.50 + 0.80298.15 ± 0.036
KA95.24 ± 0.08497.26 ± 0.04592.26 ± 0.07994.81 ± 0.07298.05 ± 0.39486.86 + 0.90797.89 ± 0.041
AA95.68 ± 0.09496.28 ± 0.10594.52 ± 0.16894.65 ± 0.12395.60 ± 0.64983.80 + 3.99397.73 ± 0.092
Table 7. Accuracies (%) of Pavia University dataset.
Table 7. Accuracies (%) of Pavia University dataset.
Classes NameTrain/TestVGG16Resnet503-DCNNHybridSNA2S2KSF-PatchDCFF-NET
Asphalt663/596899.49 ± 0.0399.45 ± 0.0399.76 ± 0.0199.24 ± 0.0399.91 ± 0.0898.67 ± 0.5899.89 ± 0.01
Meadows1864/16,78599.97 ± 0.0199.98 ± 0.0099.95 ± 0.0199.98 ± 0.0099.97 ± 0.0299.86 ± 0.0899.92 ± 0.01
Gravel209/189099.77 ± 0.0399.66 ± 0.0399.85 ± 0.0297.49 ± 0.0999.80 ± 0.2895.84 ± 1.23100.00 ± 0.00
Trees306/275898.10 ± 0.0698.43 ± 0.0499.32 ± 0.0599.62 ± 0.0399.96 ± 0.0697.21 ± 0.6799.17 ± 0.04
Painted metal sheets134/121199.93 ± 0.03100.00 ± 0.00100.00 ± 0.00100.00 ± 0.0099.97 ± 0.04100.00 ± 0.00100.00 ± 0.00
Bare Soil502/4527100.00 ± 0.00100.00 ± 0.00100.00 ± 0.0099.83 ± 0.0199.93 ± 0.0699.74 ± 0.06100.00 ± 0.00
Bitumen133/119798.71 ± 0.1198.63 ± 0.1099.35 ± 0.1096.98 ± 0.1499.97 ± 0.0495.41 ± 1.4599.20 ± 0.06
Self-Blocking Bricks368/331499.75 ± 0.0299.96 ± 0.0299.89 ± 0.0199.76 ± 0.0398.44 ± 1.1397.34 ± 0.34100.00 ± 0.00
Shadows94/85399.26 ± 0.0899.14 ± 0.0999.88 ± 0.0099.89 ± 0.0499.74 ± 0.2198.59 ± 0.2699.91 ± 0.05
OA4273/38,50399.68 ± 0.00799.71 ± 0.00799.85 ± 0.00799.58 ± 0.00499.81 ± 0.09398.89 ± 0.16499.86 ± 0.004
KA99.58 ± 0.00999.62 ± 0.00999.80 ± 0.00999.45 ± 0.00599.75 ± 0.12398.07 ± 0.23599.82 ± 0.006
AA99.44 ± 0.01399.47 ± 0.01699.78 ± 0.01499.20 ± 0.01799.74 ± 0.10398.53 ± 0.21399.79 ± 0.011
Table 8. Accuracies (%) of Salinas dataset.
Table 8. Accuracies (%) of Salinas dataset.
Classes NameTrain/TestVGG16Resnet503-DCNNHybridSNA2S2KSF-PatchDCFF-NET
Brocoli_green_weeds_1200/1809100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.0098.78 ± 0.84100.00 ± 0.00
Brocoli_green_weeds_2372/3354100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.0099.91 ± 0.04100.00 ± 0.00
Fallow197/1779100.00 ± 0.0099.89 ± 0.0299.96 ± 0.02100.00 ± 0.00100.00 ± 0.0099.72 ± 0.13100.00 ± 0.00
Fallow_rough_plow139/125599.86 ± 0.03100.00 ± 0.00100.00 ± 0.0099.92 ± 0.0099.82 ± 0.1599.84 ± 0.0999.94 ± 0.03
Fallow_smooth267/241199.88 ± 0.0199.70 ± 0.0399.88 ± 0.0299.81 ± 0.0299.95 ± 0.04100.00 ± 0.0099.96 ± 0.01
Stubble395/356499.96 ± 0.02100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00
Celery357/3222100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.0099.63 ± 0.2699.97 ± 0.00
Grapes_untrained1127/10,144100.00 ± 0.00100.00 ± 0.0099.98 ± 0.0199.96 ± 0.0199.95 ± 0.0399.17 ± 0.5499.99 ± 0.00
Soil_vinyard_develop620/5583100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.0099.97 ± 0.0399.80 ± 0.03100.00 ± 0.00
Corn_senesced_green_weeds327/295199.60 ± 0.03100.00 ± 0.0099.81 ± 0.0299.83 ± 0.0499.94 ± 0.0999.93 ± 0.05100.00 ± 0.00
Lettuce_romaine_4wk106/962100.00 ± 0.00100.00 ± 0.0099.51 ± 0.05100.00 ± 0.00100.00 ± 0.0098.96 ± 0.7599.24 ± 0.08
Lettuce_romaine_5wk192/1735100.00 ± 0.00100.00 ± 0.0099.91 ± 0.03100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00
Lettuce_romaine_6wk91/825100.00 ± 0.0099.88 ± 0.0099.89 ± 0.0499.90 ± 0.0599.96 ± 0.0699.88 ± 0.13100.00 ± 0.00
Lettuce_romaine_7wk107/963100.00 ± 0.00100.00 ± 0.00100.00 ± 0.0099.61 ± 0.0499.61 ± 0.5599.79 ± 0.21100.00 ± 0.00
Vinyard_untrained726/654299.99 ± 0.0099.82 ± 0.02100.00 ± 0.00100.00 ± 0.0099.85 ± 0.1898.67 ± 0.73100.00 ± 0.00
Vinyard_vertical_trellis180/162799.82 ± 0.02100.00 ± 0.0099.71 ± 0.0399.77 ± 0.03100.00 ± 0.0099.38 ± 0.63100.00 ± 0.00
OA5403/48,72699.95 ± 0.00399.96 ± 0.00399.95 ± 0.00299.95 ± 0.00399.95 ± 0.03299.48 ± 0.08599.98 ± 0.002
KA99.95 ± 0.00399.95 ± 0.00399.95 ± 0.00299.95 ± 0.00399.94 ± 0.03699.43 ± 0.09799.98 ± 0.003
AA99.94 ± 0.00399.96 ± 0.00299.92 ± 0.00499.92 ± 0.00499.94 ± 0.03999.59 ± 0.10599.94 ± 0.006
Table 9. Overall accuracies under different training sample proportions and different filling methods.
Table 9. Overall accuracies under different training sample proportions and different filling methods.
DatasetsIPPUSA
Filling Method10%20%30%10%20%30%10%20%30%
NotFill73.7480.7384.6292.3893.6294.5592.3493.7294.07
InnerFill76.1881.8984.3791.9093.8194.4492.0994.1794.73
BothFill78.2184.0586.6892.5694.1194.7393.0094.2495.14
Table 10. Overall accuracies (%) of different percentages of training samples on three datasets.
Table 10. Overall accuracies (%) of different percentages of training samples on three datasets.
DatasetTrain PercentageNBKNNRFMLP1DCNNSF-PixelVGG16Resnet50DCFF-NET
Indian pines550.8168.9669.6364.1963.26 55.21 73.0362.11 69.84
753.7572.0471.7467.3265.37 74.8473.43 65.03 73.51
1058.0574.7675.8769.2068.44 75.27 77.88 70.90 78.21
1561.8376.8678.8273.4769.21 81.80 79.45 75.12 79.88
2066.1478.1280.7574.1874.22 82.91 83.44 75.84 84.05
2569.5579.2382.1175.2575.27 83.94 84.39 81.62 84.41
3071.5879.9883.1776.4775.43 85.49 85.18 82.13 86.68
Pavia University0.572.3177.3378.0379.1282.4675.90 78.09 75.06 78.40
176.1279.5481.3881.2184.4078.16 81.95 80.69 83.67
380.4283.6685.5487.5889.14 81.09 88.53 87.75 89.64
582.3385.3287.2389.1890.36 84.82 89.70 89.08 90.97
784.3186.0488.2091.0890.75 86.24 90.95 90.54 91.53
1085.8587.5389.5292.0291.43 90.71 91.59 91.51 92.56
2089.3689.3491.0492.3793.07 92.77 92.19 93.23 94.11
3090.9190.2891.9192.7093.97 93.95 92.83 93.44 94.73
Salinas0.567.8781.4282.1681.5384.61 82.06 82.87 81.17 85.01
177.4284.1485.5083.7287.25 82.06 87.3384.98 87.13
384.6987.6189.3988.9089.51 87.13 89.85 88.03 90.03
586.1988.9190.1389.2090.17 87.94 90.17 90.88 91.42
787.1888.9790.7489.9190.27 88.40 91.79 91.07 92.19
1088.2290.0391.3190.4191.50 90.00 92.12 92.35 93.00
2090.2191.3892.6091.3492.07 91.62 92.55 93.80 94.24
3091.1392.2493.3992.3192.58 92.71 94.25 94.6295.14
Table 11. Comparison of ablation experiment results by different modules on three datasets. The best performing results are shown in bold.
Table 11. Comparison of ablation experiment results by different modules on three datasets. The best performing results are shown in bold.
DatasetModulesExp 1Exp 2Exp 3
SIEDSFEOAKAAAOAKAAAOAKAAA
Indian Pines--84.72 ± 0.7782.56 ± 0.8981.13 ± 1.5881.94 ± 1.2879.38 ± 1.4777.49 ± 2.6675.00 ± 1.9071.46 ± 2.1966.66 ± 4.21
--84.92 ± 0.6382.82 ± 0.7183.05 ± 1.9881.77 ± 0.8979.22 ± 1.0277.10 ± 2.6474.04 ± 1.9270.34 ± 2.1966.44 ± 2.10
86.06 ± 0.3784.12 ± 0.4184.41 ± 1.0783.46 ± 0.6281.01 ± 0.7179.11 ± 2.8777.56 ± 0.9574.22 ± 1.0969.9 ± 2.87
Pavia University--94.18 ± 0.1292.26 ± 0.1691.88 ± 0.2593.36 ± 0.2891.17 ± 0.3790.89 ± 0.3892.28 ± 0.1389.73 ± 0.1889.53 ± 0.37
--94.44 ± 0.1592.63 ± 0.2092.71 ± 0.2293.71 ± 0.2291.66 ± 0.2991.70 ± 0.2791.61 ± 2.9488.92 ± 3.7389.60 ± 1.63
94.58 ± 0.1092.67 ± 0.1492.35 ± 0.1793.85 ± 0.1891.83 ± 0.2491.51 ± 0.3192.65 ± 0.2990.23 ± 0.3989.91 ± 0.39
Salinas--95.07 ± 0.1394.51 ± 0.1497.32 ± 0.0794.25 ± 0.1093.59 ± 0.1196.77 ± 0.1092.89 ± 0.1892.08 ± 0.2095.76 ± 0.17
--94.92 ± 0.2294.34 ± 0.2597.37 ± 0.1594.06 ± 0.1293.39 ± 0.1396.82 ± 0.1392.71 ± 0.1991.88 ± 0.2195.93 ± 0.28
95.19 ± 0.1594.64 ± 0.1797.51 ± 0.0994.44 ± 0.1793.81 ± 0.1997.02 ± 0.1692.94 ± 0.2992.13 ± 0.3295.98 ± 0.18
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Z.; Chen, Y.; Wang, Y.; Wang, X.; Wang, X.; Xiang, Z. DCFF-Net: Deep Context Feature Fusion Network for High-Precision Classification of Hyperspectral Image. Remote Sens. 2024, 16, 3002. https://doi.org/10.3390/rs16163002

AMA Style

Chen Z, Chen Y, Wang Y, Wang X, Wang X, Xiang Z. DCFF-Net: Deep Context Feature Fusion Network for High-Precision Classification of Hyperspectral Image. Remote Sensing. 2024; 16(16):3002. https://doi.org/10.3390/rs16163002

Chicago/Turabian Style

Chen, Zhijie, Yu Chen, Yuan Wang, Xiaoyan Wang, Xinsheng Wang, and Zhouru Xiang. 2024. "DCFF-Net: Deep Context Feature Fusion Network for High-Precision Classification of Hyperspectral Image" Remote Sensing 16, no. 16: 3002. https://doi.org/10.3390/rs16163002

APA Style

Chen, Z., Chen, Y., Wang, Y., Wang, X., Wang, X., & Xiang, Z. (2024). DCFF-Net: Deep Context Feature Fusion Network for High-Precision Classification of Hyperspectral Image. Remote Sensing, 16(16), 3002. https://doi.org/10.3390/rs16163002

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop