Next Article in Journal
The SSR Brightness Temperature Increment Model Based on a Deep Neural Network
Previous Article in Journal
Imitation Learning through Image Augmentation Using Enhanced Swin Transformer Model in Remote Sensing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature-Decision Level Collaborative Fusion Network for Hyperspectral and LiDAR Classification

1
Faculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo 315211, China
2
Department of Geography and Spatial Information Techniques, Ningbo University, Ningbo 315211, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(17), 4148; https://doi.org/10.3390/rs15174148
Submission received: 30 June 2023 / Revised: 15 August 2023 / Accepted: 15 August 2023 / Published: 24 August 2023

Abstract

:
The fusion-based classification of hyperspectral (HS) and light detection and ranging (LiDAR) images has become a prominent research topic, as their complementary information can effectively improve classification performance. The current methods encompass pixel-, feature- and decision-level fusion. Among them, feature- and decision-level fusion have emerged as the mainstream approaches. Collaborative fusion of these two levels can enhance classification accuracy. Although various methods have been proposed, some shortcomings still exist. On one hand, current methods ignore the shared advanced features between HS and LiDAR images, impeding the integration of multimodal features and thereby limiting the classification performance. On the other hand, the existing methods face difficulties in achieving a balance between feature- and decision-level contributions, or they simply overlook the significance of one level and fail to utilize it effectively. In this paper, we propose a novel feature-decision level collaborative fusion network (FDCFNet) for hyperspectral and LiDAR classification to alleviate these problems. Specifically, a multilevel interactive fusion module is proposed to indirectly connect hyperspectral and LiDAR flows to refine the spectral-elevation information. Moreover, the fusion features of the intermediate branch can further enhance the shared-complementary information of hyperspectral and LiDAR to reduce the modality differences. In addition, a dynamic weight selection strategy is meticulously designed to adaptively assign weight to the output of three branches at the decision level. Experiments on three public benchmark datasets demonstrate the effectiveness of the proposed methods.

1. Introduction

Remote sensing imagery as a means of acquiring ground object information [1] has found extensive applications in geological exploration [2], marine monitoring [3] and disaster management [4]. In recent years, the advancement of sensor technology has facilitated the provision of a diverse range of remote sensing images. Different sensors can provide varying information for land-cover objects within the same geographic area. Hyperspectral data have been widely used in the fine classification of land-cover and land-use due to rich spectral-spatial information. However, it is always difficult to accurately classify ground objects only with hyperspectral data because of the phenomenon of the “same spectrum of foreign objects” [5]. LiDAR data provide high-precision three-dimensional spatial information, which is complementary to hyperspectral data. For example, hyperspectral data fail to distinguish the roads and buildings because they are both concrete structures, whereas LiDAR data can effectively distinguish them due to elevation and height information. In contrast, only LiDAR data face challenges in accurately distinguishing different materials of the same height (e.g., lawn and road). Therefore, the integration of hyperspectral and LiDAR data is widely used for land-cover and land-use classification [6,7,8].
In recent years, how to effectively integrate hyperspectral and LiDAR data has become a hot research topic. The current approaches can be categorized into three types: pixel-level fusion, feature-level fusion and decision-level fusion. Pixel-level fusion directly synthesizes the hyperspectral and LiDAR data at the input layer and then performs classification [9,10,11]. However, pixel-level fusion generally entails a significant amount of computational overhead, and it also exhibits a general low robustness due to some disturbances, such as noise. To tackle the aforementioned challenges and optimize the synergistic information between hyperspectral and LiDAR data, feature-level fusion techniques have been extensively employed. In [12], the classification is achieved by combining hyperspectral and LiDAR features extracted through Extended Attribute Profiles (EAPs). However, the mere concatenation or superposition of their features may result in redundant information and impede classification performance. The principal component analysis (PCA) is an effective method for reducing redundancy by mapping high-dimensional data onto an orthogonal low-dimensional space [13]. Meanwhile, Du et al. [14] proposed a graph fusion approach that leverages the correlation between hyperspectral and lidar data to integrate their features. Furthermore, a manifold alignment approach is proposed in [15] to enhance the acquisition of shared features across diverse modalities and integrate them with modality specific features for classification purposes. Besides feature-level fusion, decision-level fusion is also a widely adopted approach. In [16], it was proposed to use the support vector machine (SVM) for each feature, and then use Naive Bayes to fuse classifiers to obtain classification results. In [17], the maximum likelihood classifier (MLC), SVM and multinomial logistic regression classifiers were used to classify the features, and weighted voting was then used to obtain the final classification results. Although both feature-level and decision-level fusion techniques are capable of integrating information from different modalities, they often heavily rely on human experience in parameter selection and feature designment. Moreover, they also face the challenge of achieving a balance between the algorithmic precision and generalizability.
With the increasing popularity of deep learning in remote sensing [18,19,20,21], a plethora of fusion algorithms have been proposed. Among them, the convolutional neural network (CNN) has proven to be effective in extracting deep features from images and is widely utilized for the joint classification of hyperspectral and LiDAR data. In [22], a two-branch CNN is proposed to extract the spatial and spectral features of hyperspectral data, respectively, and then fuse with the features of LiDAR data for classification. The space and spectral characteristics of hyperspectral data are effectively utilized. Hang et al. [23] tried to use the CNN to extract the features of hyperspectral and LiDAR data in the way of parameter sharing, and after the feature-level fusion, the weighted summing method was adopted at the decision level to obtain the final classification results. To better utilize the complementary information between hyperspectral and LiDAR data, many researchers have proposed some methods. For instance, Zhang et al. [24] proposed a bidirectional autoencoder for hyperspectral and LiDAR data fusion, which utilized spectral and texture metrics to constrain the structure of the fused information, while integrating it using the Gram matrix. The resulting fusion outputs were then fed into a two-branch CNN for classification, reducing the reliance on training samples by leveraging complementary information. In [25], a multi-branch fusion network of self-and cross-guided attention is proposed. Specifically, the LiDAR-derived attention mask guides both the HS and the LiDAR itself. Meanwhile, it is sent to the fusion module together with the spectral supplement module for classification fusion. This approach can effectively interact with complementary information from different modalities. Similarly, Fang et al. [26] proposed a spatial-spectral enhancement module that effectively enhances the interaction between hyperspectral and LiDAR modalities by enhancing the spatial features of hyperspectral data with LiDAR features and enhancing the spectral information of LiDAR with hyperspectral features. In [27], Wang et al. proposed a three-branch CNN backbone network that can simultaneously extract spectral, spatial and elevation information. They utilized hierarchical fusion to achieve a feature interaction and the integration of hyperspectral and LiDAR data, resulting in significant improvements in classification accuracy.
Although the above models have achieved acceptable results, most existing methods extract features of different modalities separately, and then directly integrate the complementary information. Intuitively, the complementary information highlights the distinctiveness of each modality, and the direct integration of this unique information leads to suboptimal fusion performance. Shared features across modalities can demonstrate their affinity and facilitate smoother connections between them, serving as a “bridge” to alleviate huge modality differences. Therefore, the integration of complementary and shared features across modalities is essential for their mutual enhancement. By interactively integrating these features, information from different modalities can be optimally combined. Inspired by this, our insight is the first key research question (RQ), RQ1: How can we effectively utilize the shared-complementary information of hyperspectral and LiDAR images? In addition, feature-level fusion enables the comprehensive processing of information, including edges, shapes, contours and local features, while decision-level fusion exhibits excellent error correction capabilities. Generally, the strategy of feature-level and decision-level fusion play distinct yet equally important roles in improving the performance of hyperspectral and LiDAR classification. However, the current feature-level and decision-level joint classification methods exhibit limited adaptability and flexibility. Therefore, the next key research question is raised: RQ2: How can we adaptively collaboratively integrate the strategy of feature-level and decision-level fusion?
In this paper, to address the above RQ1, we propose a multilevel interactive fusion (MIF) module to integrate the shared-complementary information of HS and LiDAR data while weakening the differences between the modalities. Specifically, the MIF module is cascaded by several three-branch feature interaction (TBFI) modules, which introduces an intermediate state branch in addition to the HS and LiDAR branches, with the aim of reducing modal differences and enabling the full interaction and integration of shared and complementary information through multiple TBFI modules. To tackle the above RQ2, the dynamic weight selection (DWS) module in decision-level fusion is attentively designed, which takes the three output layers of the MIF module as the input and adaptively assigns weights to the features obtained in feature-level fusion. The main contributions of this paper are summarized as follows:
(1) A feature-decision level collaborative fusion network for hyperspectral and LiDAR classification was proposed, which collaboratively integrates the strategy of feature-level and decision-level fusion to a unified framework.
(2) In feature-level fusion, an MIF module was developed that incorporates a novel intermediate state branch to optimize the utilization of shared-complementary features between hyperspectral and lidar data while minimizing mode discrepancies. This branch facilitates the information’s interaction and integration with both the HS and Lidar branches within the TBFI module, utilizing a multi-level enhancement method to extract and transfer shared-complementary features between the two streams.
(3) In decision-level fusion, a DWS module was proposed to adaptively optimize the feature representations of the three branches by dynamically weighting their feature outputs to ensure a balanced and effective feature representation.
The rest of this article is organized as follows. The proposed framework is introduced in detail in Section 2. In Section 3, the experimental configuration and parameter setting, the classification results and analysis and ablation studies are given. Finally, a summary is given in Section 4.

2. Materials and Methods

2.1. Overall Framework of the Proposed Model

In this paper, a novel Feature-Decision level Collaborative Fusion Network (FDCFNet) was proposed for hyperspectral and LiDAR classification; the overall framework is shown in Figure 1. The input mainly consists of three branches: hyperspectral, intermediate state and LiDAR. At the feature level, considering directly integrating the complementary HS and LiDAR features may lead to suboptimal fusion performance, due to the distinct modality difference. The shared features between HS and LiDAR can show the affinity between them, promote a smoother connection and become a “bridge” to alleviate the huge modality differences. Therefore, integrating the complementary and shared characteristics of the HS and LiDAR is essential for their mutual enhancement. Inspired by this, we designed a multilevel interaction fusion module and then three-branch feature interaction (TBFI) modules to effectively learn the complementary and shared features of hyperspectral and LiDAR data while reducing the difference between modalities. At the decision level, we designed the DWS module to optimize the information representation by dynamically assigning weights to three outputs of the MIF module. Most of the existing models often simply combine the feature representations together, or stack them together with some fully connected layers [28,29,30]. However, a mere stacking or fusion of these features may lead to an imbalanced representation of information, thereby diminishing the classification accuracy. Therefore, based on feature-level fusion, we further introduce a decision-level fusion component, i.e., the DWS module, which can dynamically allocate weights to the output from three branches in order to balance the representation of information from different modalities.

2.2. Multilevel Interaction Fusion Module

The multilevel interaction fusion module is shown in Figure 2, given a hyperspectral image I H and its corresponding LiDAR image I L . First of all, in order to eliminate redundant information in the hyperspectral image, PCA was used to reduce the number of bands in hyperspectral images to C , where C is the first C principal components. Then, select a cube P H R M × N × C in I H , and select its corresponding patch P L R M × N × 1 in I L , where M, N and C represent the height, width and band number of the patch, respectively. The P H and P L are fed into a convolution layer to initially learn features, followed by the batch normalization (BN) layer and the rectified linear unit (ReLU) to regularize and learn nonlinearities. The output results are F H and F L , respectively. In order to reduce the impact of the modality difference in the interaction of complementary information, we added an intermediate branch F f , indirectly connecting the HS and LiDAR branches, which can better realize the joint learning of shared and complementary features. It can be formulated by
F f = c o n c a t F H , F L ,
where c o n c a t represents the combinatorial operation. After that, F H , F f and F L are inputs to the TBFI module. In order to further mine and integrate the multi-level features while reducing modality differences, the N TBFI modules are sequentially employed, with the output of each preceding module serving as the input to the subsequent one, and by analogy, the F N H , F N f and F N L are assumed to be the outputs of the N-level TBFI modules, which are formulated as
F N H = g N T B F I , , g i T B F I , , g 1 T B F I F H F N f = g N T B F I , , g i T B F I , , g 1 T B F I F f F N L = g N T B F I , , g i T B F I , , g 1 T B F I F f
where F H , F f and F L represent the three inputs, g T B F I represents the function of the TBFI module and i represents the function of the i th TBFI module layer.
The details of the TBFI module are shown in Figure 2. Take the first TBFI module layer as an example. First, the input feature maps of the three branches are fed to the convolutional layer to obtain the standardized feature maps F H , F f , F L :
F H = N c o n v F H F f = N c o n v F f F L = N c o n v F L
where N c o n v includes the convolution operation, BN layer, ReLU activation function and max-pooling layer. Among them, F H , F f and F L contain rich spectral information, spatial information and elevation information, respectively. Then, in order to further obtain their shared and complementary features, the spectral information extracted from the HS branch and the elevation information extracted from the LiDAR branch are simultaneously injected into the intermediate state branch, and the resulting feature map F f s is obtained:
F f s = F f F H F L
where represents the addition operation of elements. Then, to further supplement the complementary information between the two modalities and enhance the shared information, the three branches carry out the feature interaction between modalities through the intermediate branches after the convolution operation, BN layer and ReLU activation function. The procedure can be formulated as
F 1 H = F H c o n c a t B c o n v F H , B c o n v F f s F 1 L = F L c o n c a t B c o n v F L , B c o n v F f s
and among them, F 1 H and F 1 L represent the feature maps of the HS and LiDAR branches, respectively, which are supplemented with complementary information and enhanced by shared information, that is, both branches fully contain spectral, spatial and elevation information. B c o n v includes the convolution operation, BN layer and ReLU activation function. Finally, to further explore the shared high-level features, the feature maps of both the HS and LiDAR branches are fed into the self-attention module in order to generate attention masks. These masks are then injected into the middle branch to enhance focus on the common information; the specific formula is
F 1 f = B c o n v F f s 1 S A F 1 H S A F 1 L
where represents the multiplication operation of elements, and S A represents the operation in which features pass through the self-attention module to generate an attention mask. The self-attention module is shown in Figure 3. Firstly, the matrix multiplication operation is performed on the input feature W and its transpose W T , and then the result is fed into the softmax layer for normalization to obtain the attention mask.
Finally, the output feature maps F N H , F N f and F N L after the N-level TBFI modules are fed to the fully connected layer after the convolution operation, BN layer and ReLU activation function, and the outputs D H , D f , D L can be formulated as
D H = F C F N H D f = F C F N f D L = F C F N L
where F C represents the operation of feeding features to the fully connected layer for high integration.

2.3. Dynamic Weight-Selecting Module

The proposed dynamic weight-selecting (DWS) module is shown in Figure 4. In order to better integrate the information output by the MIF module, D H , D f and D L are first combined to obtain the sum of the information represented, which can be formulated as
S = D H D f D L .
Then, S is fed to the full connection layer and ReLU activation function successively to learn more about the representation of information. After that, the output is placed separately into three fully connected layers, aiming to learn the representation of the information with different weights. Then, the output is integrated through the concat function, which can be specifically expressed as
S 1 = Re L U F C S
S 2 = c o n c a t F C 1 S 1 + F C 2 S 1 + F C 3 S 1
where F C represents the first fully connected layer; Re L U represents learning nonlinearity through the ReLU activation function; F C 1 , F C 2 and F C 3 represent three fully connected layers with different weights; S R Q × 1 , S 1 R Q × 1 and S 2 R 3 Q × 1 ; and Q represents the number of classes of land-covers to be distinguished. Then, the weight selection operation is carried out on the output information representation S 2 :
w H , w f , w L = S P s o f t max S 2
and among them, s o f t max represents the softmax function and S P represents the split operation, which evenly divides the input into three parts. w H , w f and w L are the weight distribution vectors. Finally, the output of the DWS module can be expressed as
D = w H D H + w f D f + w L D L .

2.4. Classification and Loss Function Design

The output D of the DWS module is the preliminary classification map, and the classification result O can be obtained by the following calculation:
O = arg max s o f t max D
where arg max represents the argmax function, which gives the class with the maximum probability.
We send the training set into the network in the form of X H i , X L i , Y i | i = 1 , 2 , , N , where X H i , X L i and Y i represent the input and corresponding class label value, respectively, i represents the i th sample and K represents the aggregate number of samples. Moreover, we employ the cross-entropy loss function to compute the loss value of the predicted output. Y and the class label value Y show where the predicted value represents the probability that the output result is the class label value. The specific calculation formula is shown as
l o s s = 1 K i = 1 K Y i log Y i .
Simultaneously, for enhanced supervision of the feature learning process across the three branches of FDCFNet, the loss value was calculated for the output of the three branches to obtain l o s s H , l o s s f and l o s s L , which, respectively, represent the loss function of the hyperspectral, intermediate state and LiDAR branch. And the loss function l of the whole frame is designed as
l = μ 1 l o s s H + μ 2 l o s s f + μ 3 l o s s L + l o s s
where among them, μ 1 , μ 2 and μ 3 , respectively, represent the weight parameters of l o s s H , l o s s f and l o s s L . Based on the experience of a large number of experiments, we set them as 0.001, 0.01 and 0.001, respectively.

3. Experiment and Result

Three benchmark remote sensing (RS) datasets were utilized as opening references to assess the performance of FDCFNet. First, the experimental datasets were described. Secondly, the details of the experiments were given. After that, we compared and analyzed the classification results of FDCFNet with different comparison methods. Finally, we conducted ablation experiments to demonstrate the efficacy of various modules.

3.1. Experimental Datasets

In this paper, three common RS datasets: the Houston dataset [31,32], Trento dataset [33] and MUUFL dataset [34,35], are selected to evaluate the performance of FDCFNet. The land-cover object classes and corresponding numbers of training and testing samples for the three datasets are presented in Table 1, Table 2 and Table 3. A detailed description of these datasets follows.
(1) Houston dataset: The dataset comprises a hyperspectral image with 144 spectral bands and a wavelength range of 0.38 to 1.05 μ m, as well as LiDAR data, capturing the University of Houston campus at a spatial resolution of 2.5 m over an area size of 349 × 1905 pixels. The LiDAR data contains one band, while the training and testing labels encompass fifteen classes.
(2) Trento dataset: The second dataset covers a rural region situated to the south of Trento, Italy. The hyperspectral image was acquired by the AISA Eagle sensor, which features 63 spectral bands spanning from 0.42 to 0.99 μ m in wavelength, while the LiDAR data was collected using the Optech ALTM 3100EA sensor and consists of one spectral band. This dataset measures at a spatial resolution of 1 m with dimensions measuring at 166 × 600 pixels. The training and testing labels contain six distinct categories.
(3) MUUFL dataset: The third dataset has a size of 325 × 220 pixels and was taken at the University of Mississippi Gulf Coast campus. It contains a stream of hyperspectral imagery and LiDAR data, where the hyperspectral image has 64 spectral bands with wavelengths ranging from 0.38 to 1.05 μ m; LiDAR data have two spectral bands. The training and testing labels contain 11 different categories.

3.2. Experimental Configuration

To assess the performance of different methods, all experiments were conducted on a PC equipped with an Intel(R) Xeon(R) Gold 5218 CPU, operating at 2.30 GHz, an NVIDIA Quadro P5000 GPU, 32 GB of RAM and Windows 10. Our proposed programs were written in the PyCharm compiler with python3.8, and some deep learning networks were implemented using the PyTorch framework. At the same time, we use the Adam optimization algorithm [36] to adaptively optimize these models. Set the batch size and epoch to 64 and 150, respectively, for training.
In this paper, three metrics were employed—average accuracy (AA), overall accuracy (OA) and the kappa coefficient (kappa)—to assess the performance. The OA represents the proportion of correctly classified pixels to total classified pixels. The AA denotes the mean classification accuracy across all categories, while the kappa measures error reduction between classification and random guessing. Kappa can be mathematically expressed as follows:
K a p p a = O A p e 1 p e × 100 %
where
p e = i = 1 Q a i × b i n × n ,
where Q is the number of classes, a i is the number of actual samples of the i th class, b i is the number of predicted samples of the i th class and n represents the aggregate quantity of samples to be classified. These performance indicators can assess the classification accuracy of the model and a higher value indicates a superior classification performance.

3.3. Classification Results and Analysis

In order to verify the validity of FDCFNet, we compare the proposed model with several state-of-the-art methods, including the support vector machine (SVM) [37], random forest (RF) [38], shared and specific feature learning model (S2FL) [14], common subspace learning (CoSpace-ℓ1) [39], Two-branch convolution neural network (two-branch CNN) [21], Coupled CNN [22], Multi-attentive hierarchical dense fusion net (MAHiDFNet) [26] and Spatial–spectral cross-modal enhancement network(S2ENet) [25]. Meanwhile, in order to ensure comparability across experiments, identical training and testing samples were used.
Table 4, Table 5 and Table 6 show the classification results on the three datasets, i.e., Houston, Trento and MUUFL, respectively. The visual classification maps of the three datasets are presented in Figure 5, Figure 6 and Figure 7 for comparison among different methods.
Results on the Houston dataset: Our proposed method surpasses other recent influential deep learning-based methods, as demonstrated in Table 4, including the two-branch CNN, Coupled CNN, MAHiDFNet and S2ENet on the Houston dataset. The OA was increased by 9.43%, 1.27%, 7.03% and 0.83% respectively, while the kappa was increased by 10.22%, 1.38%, 7.6% and 1.65%. Among them, the accuracy of Commercial is much higher than that of the other classification methods, which is attributed to the multilevel interaction fusion module that can better interactively integrate the spectral information of hyperspectral data and the elevation information of LiDAR data. Specifically, it combines the unique information of the two modalities and enhances the common information of the two modalities. In Figure 5, it is evident that the proposed method yields a smoother visual effect and exhibits no classification errors or misjudgments when categorizing stressed grass, soil and water.
Results on the Trento dataset: Table 5 shows the classification accuracy of the FDCFNet and other comparison methods on the Trento dataset; the OA of the proposed method achieves an outstanding 99.11%, in which the classification accuracy of apple trees, woods and vineyard reaches 100%. In Figure 6, it is evident that SVM and RF exhibit more misclassification patterns than certain deep learning algorithms. It may be because the single-input model will carry out multi-mode information fusion before input, which will cause certain information loss. Additionally, the classification results of the traditional methods S2FL and CoSpace-ℓ1 are also unsatisfactory, possibly due to geometric information loss during the transformation of hyperspectral and LiDAR data into vectors. The FDCFNet exhibits a significant classification effect on large-scale objects, which can be attributed to the effective integration and interaction of complementary and shared information across different modalities facilitated by the MIF module. However, in the road classification, the effect is not satisfactory, which may be due to the phenomenon of misjudgment caused by information overlap when the common features and complementary features are fused and enhanced.
Results on the MUUFL dataset: Table 6 presents the classification accuracy of various methods on the MUUFL dataset. The proposed method exhibits a significantly superior classification performance compared to some classical comparison methods. The OA of FDCFNet surpasses that of several recently proposed deep learning methods, including the two-branch CNN, Coupled CNN, MAHiDFNet and S2ENet by 4.23%, 3.7%, 1.07% and 6.34%, respectively. In addition, we can also see the superiority of the proposed method in Figure 7. It can be seen that FDCFNet classifies almost all of the mixed ground surface and trees in the lower right corner of the figure correctly, and at the same time, the edges of trees and other ground objects are clear and very close to the ground truth which is attributed to the dynamic weight selection strategy, which can better balance the feature representation of hyperspectral and LiDAR data, especially in the edge processing of classification. This fully demonstrates the superiority of the proposed model in land-cover classification.
On the whole, our proposed method outperforms other state-of-the-art methods, demonstrating a superior classification performance. The exceptional results of FDCFNet can be attributed to several key factors: firstly, an intermediate branch is introduced into the model to mitigate modality discrepancies. Secondly, multiple layers of the TBFI modules are employed for deep mining and the integration of shared-complementary features from hyperspectral and LiDAR data. Finally, a DWS module is incorporated to dynamically allocate weights to the output of three branches, ensuring a more equitable representation of information across different modalities. Among the three classification result graphs, FDCFNet exhibits a superior classification performance, with clearer boundaries and closer proximity to the ground truth. This further validates the advantages of our proposed model. In summary, our model demonstrates a high competitiveness in joint hyperspectral and LiDAR data classification.

4. Discussion

4.1. Parameter Tuning

As is widely acknowledged, the performance of the overall framework hinges on its submodules. Therefore, optimizing the overall performance by adjusting hyper-parameters in each submodule becomes a crucial issue to be addressed. Specifically, we fine-tune hyper-parameters such as the PCA dimensionality reduction, input patch size, network learning rate, number of TBFI modules in the MIF module and weight of loss function to maximize the overall performance. We assessed the classification performance of our models by computing OA, AA and kappa scores across various hyper-parameters. To ensure validity and reliability, we averaged the results over 10 experiments.
(1) Number of dimensionality reduction: Since the original hyperspectral data have a large number of spectral bands, it is easy to lead to the problem of dimensional disaster. The PCA can effectively alleviate this problem and minimize the loss of effective information while reducing redundant information. The choice of dimensionality reduction C has a certain influence on model performance. Therefore, other parameters were fixed and the sets of different C 1 , 5 , 10 , 15 , 20 , 25 , 30 were tested on three RS data sets, respectively. Figure 8 shows the OA of different C on the three data sets. It can be seen that in the three data sets, when C is 25, the performance is the best. Therefore, we can set the dimensionality reduction C to 25.
(2) Input patch size: The patch size of hyperspectral and LiDAR data determines the amount of information fed into the model. In order to avoid significant information loss during convolution, pooling and other operations, it is crucial to ensure that the input size is not too small. Conversely, if the input size is too large, there will be an increase in computational complexity without a corresponding increase in abstraction level. In order to determine the optimal network input patch size, while keeping other parameters fixed, patches ranging from 7 × 7 to 19 × 19 were discussed on three different remote sensing datasets. Figure 9 illustrates the overall accuracy (OA) of various patch sizes across these datasets, revealing that the model performs best with a patch size of 13 × 13.
(3) Network learning rate: The learning rate is a crucial hyper-parameter for updating parameters in neural networks, as it determines the convergence of the objective function to a local minimum and the time required for such a convergence. If the learning rate is too small, the network convergence becomes more complex and results in slow changes to the loss function. Conversely, if the learning rate is too high, there may be the direct skipping of local optima by the loss function leading to non-convergence. Therefore, a learning rate ranging from 0.0001 to 0.01 was analyzed for training on the three RS datasets in order to obtain the optimal value. As shown in Figure 10, a learning rate of 0.001 yielded the best results.
(4) The number of TBFI modules: As discussed in Section 2.2, after simply extracting the features of the hyperspectral, intermediate state and LiDAR branch, in order to further excavate and integrate multi-level features and weaken the differences between modes, the N-level TBFI module is used to enhance their shared and complementary features. According to Equation (2), FDCFNet is affected by N. In order to discuss the influence of the number N of TBFI modules on model performance, we change the value of N on three data sets, respectively, to compare the classification results. Figure 11 shows the OA index of classification results of three data sets with different values of N. It is evident that the advanced shared and complementary features of the three branches cannot be effectively extracted when N equals 1, leading to a poor model performance. Conversely, if N is too large, overfitting may occur due to a limited number of training samples. The optimal value for N appears to be 2 as it yields the best model performance.
(5) Weight parameter of the loss function: The loss function of the model is shown in Equation (15), where there are three hyperparameters: μ 1 , μ 2 and μ 3 . In order to find the influence of these three hyperparameters on the model performance, we set different values in three data sets for experiments. Specifically, we first fix μ 2 and μ 3 , change the value of μ 1 from the 0.001 , 0.01 , 0.1 , 1 and set to find the μ 1 that can achieve the best performance of the model. Then, we select the best μ 1 and fix the value of μ 3 and change the value of μ 2 from the 0.001 , 0.01 , 0.1 , 1 set, so as to find the best μ 2 value. Similarly, the optimal values of μ 1 and μ 2 are selected, and the values of μ 3 are changed to find the value of μ 3 that optimizes the model performance. Figure 12 shows the parameter experiments carried out on three data sets. The red line represents μ 1 , the yellow line represents μ 2 and the blue line represents μ 3 . It can be seen from the three figures that when μ 1 , μ 2 and μ 3 are 0.001, 0.01 and 0.001, respectively, the model performance reaches the best.

4.2. Ablation Study

In order to validate the efficacy of each model within the proposed framework for classification, we conducted ablation experiments on the intermediate state branch, TBFI module and DWS module, respectively, in this section. To ensure experimental reliability and fairness, all parameters were set to their optimal values, with OA, AA and kappa used as metrics for evaluating classification performance.
(1) Ablation study for intermediate state branch: In addition to the hyperspectral and LiDAR branches, we propose a novel intermediate state branch that indirectly connects the two modalities, reduces the mode gap, weakens modality differences and facilitates better interactive fusion of information from both hyperspectral and LiDAR branches. Table 7 demonstrates the classification performance of the model on three remote sensing datasets, both before and after incorporating an intermediate state branch. The results indicate a significant improvement in classification accuracy following the addition of this branch. In the Houston data set, OA has increased by 2.59%, and in the Trento data set, OA has increased by 3.08%. On the MUUFL dataset, OA improved by 2.19%. At the same time, AA and kappa also improved correspondingly, which indicates that the intermediate state branch can effectively improve the classification performance of the model.
(2) Ablation study for three-branch feature interaction (TBFI) modules: As a crucial component of the overall framework, the three-branch feature interaction module facilitates the interaction between hyperspectral and LiDAR branches by means of intermediate state branches, thereby enabling the further exploration and integration of their shared-complementary features. Through the experimentation of hyperparameter settings, we have discovered that cascading two TBFI modules together can effectively integrate advanced features from different modalities. Table 8 illustrates the classification performance of models before and after incorporating two TBFI modules into three RS datasets. After the incorporation of two TBFI modules, a significant improvement in classification performance is observed. On the Trento data set, OA is enhanced by up to 8.05%, which serves as compelling evidence that TBFI modules are highly effective in enhancing classification accuracy.
(3) Ablation study for dynamic weight selection module: The dynamic weight selection module is designed to achieve a balanced representation of information across different modalities. It enables the adaptive assignment of weights to the output from three branches, followed by decision fusion, thereby addressing the issue of imbalanced data expression across various sensors. Table 9 shows the classification performance of the models before and after the introduction of the dynamic weight selection strategy on the three RS data sets. It can be seen that the classification performance of the model is improved to some extent after the introduction of the dynamic weight selection strategy, which proves that the DWS module can improve the classification accuracy of the model.

5. Conclusions

In this paper, a novel feature-decision level collaborative fusion network was proposed for remote sensing classification with hyperspectral and LiDAR data. The proposed approach seamlessly integrates feature-level and decision-level fusion strategies within a unified framework, effectively harnessing the combined advantages of both techniques. Specifically, at the feature level, we propose a novel three-branch feature interaction module to alleviate modality discrepancies. This enables us to fully leverage the shared and complementary features of hyperspectral and LiDAR data while achieving interaction and integration through intermediate state branches. A novel dynamic weight selection module is proposed at the decision level, which adaptively assigns weights to the outputs of three branches and integrates information from different modalities in a more balanced manner. Experimental results on three common remote sensing datasets fully demonstrate the effectiveness of our proposed framework.
The proposed three-branch feature interaction module and dynamic weight selection module offer a valuable reference for the fusion and classification of multi-modal data in various fields. The proposed multilevel interactive fusion module fully performs feature interactive fusion of two different modal data through the middle branch, thus improving the classification accuracy. Furthermore, we plan to extend these modules to other domains, such as hyperspectral and multispectral fusion, as well as multispectral and panchromatic image fusion. Meanwhile, we will investigate methods to enhance the model’s generalizability and its application in sensor data fusion across diverse modalities.

Author Contributions

Conceptualization, S.Z., Q.L. and X.M.; methodology, S.Z. and X.M.; software, S.Z.; validation, S.Z.; writing—original draft preparation, S.Z.; writing—review and editing, S.Z., Q.L. and X.M.; supervision, Q.L., X.M., G.Y. and W.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (42171326); National Natural Science Foundation of China (42071323); Zhejiang Provincial Natural Science Foundation of China (LR23D010001, LY22F010014); and Ningbo Natural Science Foundation under Grant 2022J076.

Data Availability Statement

The Houston dataset is available at: http://dase.grss-ieee.org/ (accessed on 5 July 2022). The Trento dataset can be obtained from [34]. The MUUFL dataset can be obtained from [35].

Acknowledgments

The authors wish to acknowledge the team led by Hang Renlong of Nanjing University of Information Science & Technology who provided the code of the Coupled CNN.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Rast, M.; Painter, T.H. Earth observation imaging spectroscopy for terrestrial systems: An overview of its history, techniques, and applications of its missions. Surv. Geophys. 2019, 40, 303–331. [Google Scholar] [CrossRef]
  2. Goetz, A.F.; Rock, B.N.; Rowan, L.C. Remote sensing for exploration; an overview. Econ. Geol. 1983, 78, 573–590. [Google Scholar] [CrossRef]
  3. Danovaro, R.; Carugati, L.; Berzano, M.; Cahill, A.E.; Carvalho, S.; Chenuil, A.; Corinaldesi, C.; Cristina, S.; David, R.; Dell’Anno, A.; et al. Implementing and innovating marine monitoring approaches for assessing marine environmental status. Front. Mar. Sci. 2016, 3, 213. [Google Scholar] [CrossRef]
  4. Bello, O.M.; Aina, Y.A. Satellite remote sensing as a tool in disaster management and sustainable development: Towards a synergistic approach. Procedia-Soc. Behav. Sci. 2014, 120, 365–373. [Google Scholar] [CrossRef]
  5. Lv, W.; Wang, X. Overview of hyperspectral image classification. J. Sens. 2020, 2020, 4817234. [Google Scholar] [CrossRef]
  6. Koetz, B.; Sun, G.; Morsdorf, F.; Ranson, K.J.; Kneubühler, M.; Itten, K.; Allgöwer, B. Fusion of imaging spectrometer and LIDAR data over combined radiative transfer models for forest canopy characterization. Remote Sens. Environ. 2007, 106, 449–459. [Google Scholar] [CrossRef]
  7. Dalponte, M.; Bruzzone, L.; Gianelle, D. Fusion of hyperspectral and LIDAR remote sensing data for classification of complex forest areas. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1416–1427. [Google Scholar] [CrossRef]
  8. Mundt, J.T.; Streutker, D.R.; Glenn, N.F. Mapping sagebrush distribution using fusion of hyperspectral and lidar classifications. Photogramm. Eng. Remote Sens. 2006, 72, 47–54. [Google Scholar] [CrossRef]
  9. Rockinger, O.; Fechner, T. Pixel-level image fusion: The case of image sequences. In Signal Processing, Sensor Fusion, and Target Recognition VII, Proceedings of the Aerospace/Defense Sensing and Controls, Orlando, FL, USA, 13–17 April 1998; SPIE: Bellingham, WA, USA, 1998; Volume 3374, pp. 378–388. [Google Scholar]
  10. Mertens, T.; Kautz, J.; Van Reeth, F. Exposure fusion. In Proceedings of the 15th Pacific Conference on Computer Graphics and Applications, Maui, HI, USA, 29 October–2 November 2017; pp. 382–390. [Google Scholar]
  11. Zhang, Z.; Blum, R.S. A categorization of multiscale-decomposition-based image fusion schemes with a performance study for a digital camera application. Proc. IEEE 1999, 87, 1315–1326. [Google Scholar] [CrossRef]
  12. Sun, W.; Yang, G.; Peng, J.; Du, Q. Lateral-slice sparse tensor robust principal component analysis for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 17, 107–111. [Google Scholar] [CrossRef]
  13. Du, X.; Zheng, X.; Lu, X.; Doudkin, A.A. Multisource remote sensing data classification with graph fusion network. IEEE Trans. Geosci. Remote Sens. 2021, 59, 10062–10072. [Google Scholar] [CrossRef]
  14. Hong, D.; Hu, J.; Yao, J.; Chanussot, J.; Zhu, X.X. Multimodal remote sensing benchmark datasets for land cover classification with a shared and specific feature learning model. ISPRS J. Photogramm. Remote Sens. 2021, 178, 68–80. [Google Scholar] [CrossRef] [PubMed]
  15. Bigdeli, B.; Samadzadegan, F.; Reinartz, P. A decision fusion method based on multiple support vector machine system for fusion of hyperspectral and LIDAR data. Int. J. Image Data Fusion 2014, 5, 196–209. [Google Scholar] [CrossRef]
  16. Zhong, Y.; Cao, Q.; Zhao, J.; Ma, A.; Zhao, B.; Zhang, L. Optimal decision fusion for urban land-use/land-cover classification based on adaptive differential evolution using hyperspectral and LiDAR data. Remote Sens. 2017, 9, 868. [Google Scholar] [CrossRef]
  17. Zhang, L.; Zhang, L.; Du, B. Deep learning for remote sensing data: A technical tutorial on the state of the art. IEEE Geosci. Remote Sens. Mag. 2016, 4, 22–40. [Google Scholar] [CrossRef]
  18. Liu, Q.; Meng, X.; Shao, F.; Li, S. Supervised-unsupervised combined deep convolutional neural networks for high-fidelity pansharpening. Inf. Fusion 2023, 89, 292–304. [Google Scholar] [CrossRef]
  19. Wang, N.; Meng, X.; Meng, X.; Shao, F. Convolution-Embedded Vision Transformer with Elastic Positional Encoding for Pansharpening. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–9. [Google Scholar] [CrossRef]
  20. Meng, X.; Liu, Q.; Shao, F.; Li, S. Spatio–Temporal–Spectral Collaborative Learning for Spatio–Temporal Fusion with Land Cover Changes. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  21. Li, R.; Zheng, S.; Duan, C.; Yang, Y.; Wang, X. Classification of hyperspectral image based on double-branch dual-attention mechanism network. Remote Sens. 2020, 12, 582. [Google Scholar] [CrossRef]
  22. Hang, R.; Li, Z.; Ghamisi, P.; Hong, D.; Xia, G.; Liu, Q. Classification of Hyperspectral and LiDAR Data Using Coupled CNNs. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4939–4950. [Google Scholar] [CrossRef]
  23. Zhang, M.; Li, W.; Tao, R.; Li, H.; Du, Q. Information fusion for classification of hyperspectral and LiDAR data using IP-CNN. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–12. [Google Scholar] [CrossRef]
  24. Dong, W.; Zhang, T.; Qu, J.; Xiao, S.; Zhang, T.; Li, Y. Multibranch Feature Fusion Network with Self- and Cross-Guided Attention for Hyperspectral and LiDAR Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–12. [Google Scholar] [CrossRef]
  25. Fang, S.; Li, K.; Li, Z. S2ENet: Spatial–Spectral Cross-Modal Enhancement Network for Classification of Hyperspectral and LiDAR Data. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  26. Wang, X.; Feng, Y.; Song, R.; Mu, Z.; Song, C. Multi-attentive hierarchical dense fusion net for fusion classification of hyperspectral and LiDAR data. Inf. Fusion 2022, 82, 1–18. [Google Scholar] [CrossRef]
  27. Chen, Y.; Li, C.; Ghamisi, P.; Jia, X.; Gu, Y. Deep fusion of remote sensing data for accurate classification. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1253–1257. [Google Scholar] [CrossRef]
  28. Xu, X.; Li, W.; Ran, Q.; Du, Q.; Gao, L.; Zhang, B. Multisource remote sensing data classification based on convolutional neural network. IEEE Trans. Geosci. Remote Sens. 2017, 56, 937–949. [Google Scholar] [CrossRef]
  29. Zhang, M.; Li, W.; Du, Q.; Gao, L.; Zhang, B. Feature extraction for classification of hyperspectral and LiDAR data using patch-to-patch CNN. IEEE Trans. Cybern. 2018, 50, 100–111. [Google Scholar] [CrossRef]
  30. Liao, W.; Pižurica, A.; Bellens, R.; Gautama, S.; Philips, W. Generalized graph-based fusion of hyperspectral and LiDAR data using morphological features. IEEE Geosci. Remote Sens. Lett. 2014, 12, 552–556. [Google Scholar] [CrossRef]
  31. Debes, C.; Merentitis, A.; Heremans, R.; Hahn, J.; Frangiadakis, N.; van Kasteren, T.; Liao, W.; Bellens, R.; Pižurica, R.; Gautama, S.; et al. Hyperspectral and LiDAR data fusion: Outcome of the 2013 GRSS data fusion contest. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2405–2418. [Google Scholar] [CrossRef]
  32. Ghamisi, P.; Benediktsson, J.A.; Phinn, S. Land-cover classification using both hyperspectral and LiDAR data. Int. J. Image Data Fusion 2015, 6, 189–215. [Google Scholar] [CrossRef]
  33. Gader, P.; Zare, A.; Close, R.; Aitken, J.; Tuell, G. Muufl Gulfport Hyperspectral and Lidar Airborne Data Set; University of Florida: Gainesville, FL, USA, 2013. [Google Scholar]
  34. Du, X.; Zare, A. Scene Label Ground Truth Map for MUUFL Gulfport Data Set; Technical report; University of Florida: Gainesville, FL, USA, 2017; Volume 20170417. [Google Scholar]
  35. Glenn, T.; Zare, A.; Gader, P.; Dranishnikov, D. Bullwinkle: Scoring Code for Sub-Pixel Targets (Version 1.0), MUUFL Gulfport, 2013.
  36. Dozat, T. Incorporating Nesterov Momentum into Adam; OpenReview: Amherst, MA, USA, 2016. [Google Scholar]
  37. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  38. Ham, J.; Chen, Y.; Crawford, M.M.; Ghosh, J. Investigation of the random forest framework for classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 492–501. [Google Scholar] [CrossRef]
  39. Hong, D.; Chanussot, J.; Yokoya, N.; Kang, J.; Zhu, X.X. Learning-Shared Cross-Modality Representation Using Multispectral-LiDAR and Hyperspectral Data. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1470–1474. [Google Scholar] [CrossRef]
Figure 1. The overall framework of FDCFNet.
Figure 1. The overall framework of FDCFNet.
Remotesensing 15 04148 g001
Figure 2. The structure of the three-branch feature interaction module.
Figure 2. The structure of the three-branch feature interaction module.
Remotesensing 15 04148 g002
Figure 3. Illustration of self-attention module; denotes element-wise multiplication.
Figure 3. Illustration of self-attention module; denotes element-wise multiplication.
Remotesensing 15 04148 g003
Figure 4. Illustration of self-attention.
Figure 4. Illustration of self-attention.
Remotesensing 15 04148 g004
Figure 5. Classification maps of different methods on the Houston dataset. The pseudo-color image of HS image adopts the spectral bands of 45th, 25th and 10th as R, G and B.
Figure 5. Classification maps of different methods on the Houston dataset. The pseudo-color image of HS image adopts the spectral bands of 45th, 25th and 10th as R, G and B.
Remotesensing 15 04148 g005
Figure 6. Classification maps of different methods on the Trento dataset. The pseudo-color image of the HS image adopts the spectral bands of 25th, 18th and 10th as R, G and B.
Figure 6. Classification maps of different methods on the Trento dataset. The pseudo-color image of the HS image adopts the spectral bands of 25th, 18th and 10th as R, G and B.
Remotesensing 15 04148 g006
Figure 7. Classification maps of different methods on the MUUFL dataset. The pseudo-color image of HS image uses the 30th, 20th and 10th spectral bands as R, G and B.
Figure 7. Classification maps of different methods on the MUUFL dataset. The pseudo-color image of HS image uses the 30th, 20th and 10th spectral bands as R, G and B.
Remotesensing 15 04148 g007
Figure 8. Effect of PCA dimension reduction on model performance (OA) by FDCFNet.
Figure 8. Effect of PCA dimension reduction on model performance (OA) by FDCFNet.
Remotesensing 15 04148 g008
Figure 9. Effect of input patch size on model performance (OA) by FDCFNet.
Figure 9. Effect of input patch size on model performance (OA) by FDCFNet.
Remotesensing 15 04148 g009
Figure 10. Effect of learning rate on model performance (OA) by FDCFNet.
Figure 10. Effect of learning rate on model performance (OA) by FDCFNet.
Remotesensing 15 04148 g010
Figure 11. The effect of the number of layers of TBFI modules on the model performance.
Figure 11. The effect of the number of layers of TBFI modules on the model performance.
Remotesensing 15 04148 g011
Figure 12. The influence of weight parameters μ 1 , μ 2 and μ 3 on the classification performance (OA) by FDCFNet on the three data sets: (Left to right) Houston data, Trento data and MUUFL data.
Figure 12. The influence of weight parameters μ 1 , μ 2 and μ 3 on the classification performance (OA) by FDCFNet on the three data sets: (Left to right) Houston data, Trento data and MUUFL data.
Remotesensing 15 04148 g012
Table 1. Number of training and testing samples for the Houston dataset.
Table 1. Number of training and testing samples for the Houston dataset.
No.ClassNumber of Train SamplesNumber of Test Samples
1Health grass1981251
2Stressed grass1901254
3Synthetic grass192697
4Trees1881244
5Soil1861242
6Water182325
7Residential1961268
8Commercial1911244
9Road1931252
10Highway1911227
11Railway1811235
12Parking lot 11921233
13Parking lot 2184469
14Tennis court181428
15Running track187660
Total283215,029
Table 2. Number of training and testing samples for the Trento dataset.
Table 2. Number of training and testing samples for the Trento dataset.
No.ClassNumber of Train SamplesNumber of Test Samples
1Apple trees1294034
2Building1252903
3Ground105479
4Woods1549123
5Vineyard18410,501
6Roads1223174
Total81930,214
Table 3. Number of training and testing samples for the MUUFL dataset.
Table 3. Number of training and testing samples for the MUUFL dataset.
No.ClassNumber of Train SamplesNumber of Test Samples
1Trees15023,246
2Mostly grass1504270
3Mixed ground surface1506882
4Dirt and sand1501826
5Road1506687
6Water150466
7Building Shadow1502233
8Building1506240
9Sidewalk1501385
10Yellow curb150183
11Cloth panels150269
Total165053,687
Table 4. Classification accuracy of the Houston data by different methods.
Table 4. Classification accuracy of the Houston data by different methods.
No.ClassSVMRFS2FLCoSpace-ℓ1Two-Branch CNNCoupled CNNMAHiDFNetS2ENetProposed
1Health grass94.0692.7590.5789.4598.8085.6198.5386.8197.45
2Stressed grass95.0795.3997.6997.0584.22100.0092.87100.00100.00
3Synthetic grass10095.12100.00100.0096.2797.8591.11100.0096.84
4Trees98.5199.0198.3998.4795.2299.9298.1099.9299.60
5Soil95.7796.0299.3699.3697.56100.0098.38100.00100.00
6Water79.3782.5499.3899.6998.33100.0095.5899.69100.00
7Residential92.4490.4180.5280.5298.0191.0199.1593.6894.64
8Commercial80.5368.3568.5789.3986.3493.7380.9495.8297.11
9Road79.970.0269.8964.8674.0494.7398.0488.9894.68
10Highway93.7878.4970.0166.7590.0591.9372.8192.6794.38
11Railway88.3182.0588.3488.4292.4496.6072.7198.0698.30
12Parking lot 162.1267.5684.2784.0257.5094.2476.8087.1093.03
13Parking lot 237.1412.7583.5878.68100.0095.7495.8094.0391.82
14Tennis court98.3187.47100.00100.0096.3999.0799.53100.0099.77
15Running track98.7586.4198.9498.1898.7998.33100.00100.0090.37
OA (%)87.6182.5086.8187.5287.1895.3489.5895.0996.61
AA (%)86.2780.2988.6388.9990.9395.9291.3695.7896.53
Kappa × 10086.6081.0585.7586.5186.1294.9688.7494.6996.34
Table 5. Classification accuracy of the Trento data by different methods.
Table 5. Classification accuracy of the Trento data by different methods.
No.ClassSVMRFS2FLCoSpace-ℓ1Two-Branch CNNCoupled CNNMAHiDFNetS2ENetProposed
1Apple trees81.9178.6971.0286.0477.9498.8899.1999.83100.00
2Building91.8285.0182.5795.8389.2997.8388.9298.2198.29
3Ground93.5594.6292.0795.6270.0397.4997.53100.0099.48
4Woods98.0794.8686.6998.73100.0099.8999.9899.96100.00
5Vineyard92.1190.2148.9063.5699.60100.0099.9099.84100.00
6Roads80.5577.7577.5789.1697.7492.8299.7890.9393.23
OA (%)90.3688.3470.2083.4894.1898.8298.5998.7899.11
AA (%)88.0086.8676.4788.1689.1097.8297.5598.1398.50
Kappa × 10087.1384.4061.9178.6692.3298.4298.1298.3898.81
Table 6. Classification accuracy of MUUFL data by different methods and the kappa coefficient.
Table 6. Classification accuracy of MUUFL data by different methods and the kappa coefficient.
No.ClassSVMRFS2FLCoSpace-ℓ1Two-Branch CNNCoupled CNNMAHiDFNetS2ENetProposed
1Trees95.1592.8372.44 78.55 98.34 91.67 98.2987.6196.46
2Mostly grass74.3478.4768.67 74.36 81.54 94.70 86.6385.7684.29
3Mixed ground surface80.8474.5753.66 63.18 73.04 67.63 83.6680.8384.82
4Dirt and sand85.1577.3671.08 81.00 85.05 87.90 83.6580.2895.51
5Road93.0589.0158.47 77.99 84.54 88.26 96.6783.6789.43
6Water81.1973.8994.42 96.57 85.98 99.79 78.8489.4898.93
7Building Shadow65.669.9976.62 83.65 80.29 97.49 74.8595.2596.15
8Building84.2782.1780.77 86.17 98.50 96.44 98.7289.8196.55
9Sidewalk48.2535.8951.05 63.25 78.28 91.77 71.9388.6687.65
10Yellow curb66.295.6295.63 96.72 30.72 96.72 40.4095.0897.27
11Cloth panels77.3986.5997.40 98.88 81.13 99.26 64.4399.6399.63
OA (%)87.0583.9568.93 77.28 88.67 89.20 91.8386.5692.90
AA (%)77.4169.474.56 81.85 79.76 91.97 79.8288.7393.33
Kappa × 10082.8178.8261.35 71.30 85.19 86.03 89.3182.8190.68
Table 7. OA, AA and kappa coefficients of the model before and after adding the intermediate state branch in the three RS data sets.
Table 7. OA, AA and kappa coefficients of the model before and after adding the intermediate state branch in the three RS data sets.
DatasetHoustonTrentoMUUFL
OAAAKappaOAAAKappaOAAAKappa
Without intermediate state branch94.0294.4393.5196.0395.1094.7390.7192.4687.94
With intermediate state branch96.6196.5396.3499.1198.5098.8192.9093.3390.68
Table 8. OA, AA and kappa coefficients of the model before and after adding the TBFI modules in the three RS datasets.
Table 8. OA, AA and kappa coefficients of the model before and after adding the TBFI modules in the three RS datasets.
DatasetHoustonTrentoMUUFL
OAAAKappaOAAAKappaOAAAKappa
Without TBFI modules93.9893.4792.9491.0692.4090.0190.0691.8486.42
With TBFI modules96.6196.5396.3499.1198.5098.8192.9093.3390.68
Table 9. OA, AA and kappa coefficients of models before and after the strategy were selected by dynamic weights on three RS datasets.
Table 9. OA, AA and kappa coefficients of models before and after the strategy were selected by dynamic weights on three RS datasets.
DatasetHoustonTrentoMUUFL
OAAAKappaOAAAKappaOAAAKappa
Without DWS module95.0495.0194.6497.1297.2897.0290.9592.0787.45
With DWS module96.6196.5396.3499.1198.5098.8192.9093.3390.68
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, S.; Meng, X.; Liu, Q.; Yang, G.; Sun, W. Feature-Decision Level Collaborative Fusion Network for Hyperspectral and LiDAR Classification. Remote Sens. 2023, 15, 4148. https://doi.org/10.3390/rs15174148

AMA Style

Zhang S, Meng X, Liu Q, Yang G, Sun W. Feature-Decision Level Collaborative Fusion Network for Hyperspectral and LiDAR Classification. Remote Sensing. 2023; 15(17):4148. https://doi.org/10.3390/rs15174148

Chicago/Turabian Style

Zhang, Shenfu, Xiangchao Meng, Qiang Liu, Gang Yang, and Weiwei Sun. 2023. "Feature-Decision Level Collaborative Fusion Network for Hyperspectral and LiDAR Classification" Remote Sensing 15, no. 17: 4148. https://doi.org/10.3390/rs15174148

APA Style

Zhang, S., Meng, X., Liu, Q., Yang, G., & Sun, W. (2023). Feature-Decision Level Collaborative Fusion Network for Hyperspectral and LiDAR Classification. Remote Sensing, 15(17), 4148. https://doi.org/10.3390/rs15174148

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop