Next Article in Journal
Multi-Objective Optimization of Municipal Solid Waste Collection Based on Adaptive Large Neighborhood Search
Next Article in Special Issue
Feasibility Research on the Auxiliary Variables in Scaling of Soil Moisture Based on the SiB2 Model: A Case Study in Daman
Previous Article in Journal
Heterogeneous Edge Computing for Molecular Property Prediction with Graph Convolutional Networks
Previous Article in Special Issue
Integrated Anti-Aliasing and Fully Shared Convolution for Small-Ship Detection in Synthetic Aperture Radar (SAR) Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Calibration-Enhanced Multi-Awareness Network for Joint Classification of Hyperspectral and LiDAR Data

1
School of Mechanical, Electrical Information Engineering, Shandong University, Weihai 264209, China
2
Shandong Key Laboratory of Intelligent Electronic Packaging Testing and Application, Shandong University, Weihai 264209, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Electronics 2025, 14(1), 102; https://doi.org/10.3390/electronics14010102
Submission received: 10 November 2024 / Revised: 25 December 2024 / Accepted: 28 December 2024 / Published: 30 December 2024
(This article belongs to the Special Issue Advances in AI Technology for Remote Sensing Image Processing)

Abstract

:
Hyperspectral image (HSI) and light detection and ranging (LiDAR) data joint classification has been applied in the field of ground category recognition. However, existing methods still perform poorly in extracting high-dimensional features and elevation information, resulting in insufficient data classification accuracy. To address this challenge, we propose a novel and efficient Calibration-Enhanced Multi-Awareness Network (CEMA-Net), which exploits the joint spectral–spatial–elevation features in depth to realize the accurate identification of land cover categories. Specifically, we propose a novel multi-way feature retention (MFR) module that explores deep spectral–spatial–elevation semantic information in the data through multiple paths. In addition, we propose spectral–spatial-aware enhancement (SAE) and elevation-aware enhancement (EAE) modules, which effectively enhance the awareness of ground objects that are sensitive to spectral and elevation information. Furthermore, to address the significant representation disparities and spatial misalignments between multi-source features, we propose a spectral–spatial–elevation feature calibration fusion (SFCF) module to efficiently integrate complementary characteristics from heterogeneous features. It incorporates two key advantages: (1) efficient learning of discriminative features from multi-source data, and (2) adaptive calibration of spatial differences. Comparative experimental results on the MUUFL, Trento, and Augsburg datasets demonstrate that CEMA-Net outperforms existing state-of-the-art methods, achieving superior classification accuracy with better feature map precision and minimal noise.

1. Introduction

In recent years, advancements in remote sensing and imaging technologies have significantly expanded the range and detail of geomorphological data available for analysis [1,2,3,4,5]. Joint classification of hyperspectral image (HSI) and LiDAR data has become increasingly vital, with numerous studies exploring its potential [6]. Hyperspectral remote sensing provides detailed spectral information across continuous bands, often numbering in the hundreds, enabling precise surface characterization [7]. In parallel, LiDAR technology captures elevation data by using laser pulses to measure the spatial distribution of target objects on the surface [8,9]. With improved imaging and data interpretation techniques, HSI-LiDAR data fusion is playing an essential role in geological exploration [10], environmental monitoring [11], urban planning [12], and other applications.
Historically, remote sensing technology was limited in data type and accessibility, impacting the accuracy and efficiency of surface classification tasks, which depend heavily on the quality of the data source [13]. HSI data, by capturing continuous spectral variations, provides a detailed response to the material and spatial characteristics of objects [14]. However, in single-scene classifications, some objects, such as forests and agricultural areas, may exhibit similar spectral signatures, making it challenging to differentiate them based on spectral data alone, especially with HSI’s limited spatial resolution [15]. Adding LiDAR-derived elevation information to HSI data addresses this challenge by enhancing land cover classification with precise height-based distinctions [16]. Consequently, the effective integration of HSI and LiDAR data is essential for improving classification accuracy and efficiency, fostering better applications in land cover analysis and beyond.
Over recent decades, researchers have adapted various machine learning and pattern recognition models for remote sensing image classification, including support vector machines (SVM) [17], random forests (RF) [18], and extreme learning machines (ELM) [19]. These methods laid foundational advancements in image classification by leveraging spectral–spatial features effectively [20,21]. Researchers have also introduced innovative approaches to improve HSI classification. For instance, Hang et al. [22] proposed a matrix-based spectral–spatial feature representation, enhancing ground information accuracy by embedding HSI pixels into a discriminative feature subspace. Wang et al. [23] developed a multi-kernel learning framework to maximize kernel separability, which improves classification without stringent kernel restrictions. Additionally, Xing et al. [24] advanced spectral–spatial feature quality by employing low-rank learning with regularization for better feature separation. However, these methods often rely on extensive prior knowledge and complex tuning, limiting their applicability for highly non-linear and complex data scenarios [15].
More recently, Deep Neural Networks (DNNs) have significantly transformed remote sensing image analysis and classification [7,25,26]. Among these, convolutional neural networks (CNNs) are particularly notable for their ability to automatically extract deep, relevant features, enabling robust classification across diverse data types. Unlike traditional models that require manual feature extraction, CNNs and other DNNs adaptively handle complex data variations, making them suitable for various remote sensing tasks [27]. For example, Paoletti et al. [28] introduced a convolutional residual pyramid network to address the noise and redundancy issues often present in high-dimensional HSI data, providing a model with fast convergence and high accuracy. Similarly, Roy et al. [29] developed a hybrid architecture combining 3D-CNN and 2D-CNN, significantly enhancing spectral–spatial information fusion for improved classification performance. Recently, Vision Transformer (ViT)-based models have gained attention in HSI classification, pushing forward the capabilities of remote sensing models [30,31,32]. For instance, Mei et al. [33] introduced the Group-Aware Hierarchical Transformer, which enhances local and global spectral–spatial interactions through grouped pixel embeddings.
In LiDAR data classification, research primarily focuses on point cloud processing. For example, Wen et al. [34] proposed the Direction-constrained Fully Convolutional Neural Network (D-FCN), which processes raw 3D coordinates and LiDAR intensity data, offering a more efficient approach to semantic labeling of 3D point clouds. Similarly, Zorzi et al. [35] integrated CNNs and Fully Convolutional Networks (FCNs) for point cloud pixel labeling, leveraging spatial and geometric relationships for improved accuracy. Despite these advancements, single-source data classification methods remain limited in comprehensiveness, confidence, and prediction accuracy due to the unique characteristics of HSI and LiDAR data. Integrating HSI and LiDAR data, therefore, offers an important avenue to enhance classification performance by addressing these limitations.
Leveraging the powerful feature extraction and fusion capabilities of deep learning models, both CNN-based and ViT-based methods have demonstrated substantial advantages in the joint classification of HSI and LiDAR data. For instance, Hang et al. [36] proposed a coupled CNN model that not only extracts spectral–spatial features from HSI but also captures elevation information from LiDAR data. Through a parameter-sharing strategy, this model integrates these heterogeneous features, marking a shift from conventional classification approaches. Zhang et al. [37] introduced the Interleaved Perceptual Convolutional Neural Network (IP-CNN), which effectively incorporates both HSI and LiDAR constraints into the fusion of multi-source structural information, achieving notable success even with limited training data. In another study, Lu et al. [38] presented Coupled Adversarial Learning-based Classification (CALC), an adversarial framework with dual generators and a discriminator that extracts shared semantic information and modality-specific details. Zhao et al. [39] designed a two-branch architecture combining CNN and Transformer encoders in a hierarchical structure, facilitating effective joint classification of heterogeneous data. Furthermore, Sun et al. [40] proposed a multi-scale lightweight fusion network, which avoids attention mechanisms, reducing training parameters while effectively capturing depth and high-order features across scales.
Despite the significant progress of DNN-based methods in joint HSI and LiDAR classification, several challenges remain when addressing complex feature environments [41,42,43]. These challenges stem from the different imaging mechanisms of each sensor, requiring separate processing of each data type to ensure accurate representation of their specific information. Additionally, the diverse application contexts of HSI and LiDAR sensors [44] lead to variations in performance emphasis for each data type, necessitating the development of models that can adaptively address these differences while capturing both local details and global context. Moreover, the inherent variation in the features of these data types demands classification models with strong adaptive capabilities to effectively fuse and complement spectral, spatial, and elevation features, ensuring accurate scene classification.
To address these challenges, we propose a novel joint classification approach for HSI and LiDAR data called the Calibration-Enhanced Multi-Awareness Network (CEMA-Net). This method consists of two primary branches: one for processing HSI features and the other for processing LiDAR features. The key innovation in CEMA-Net is the introduction of the Multi-way Feature Retention (MFR) module, which handles HSI and LiDAR data separately to extract rich spectral–spatial and elevation features. This module is designed to adapt to the unique characteristics of each data type, effectively addressing data discretization issues. To capture semantic information at multiple scales, we also propose the Spectral–spatial Aware Enhancement (SAE) and Elevation Aware Enhancement (EAE) modules. These modules are specifically tailored to the characteristics of each data source, enabling dynamic awareness of spectral, spatial, and elevation features. Finally, to address discrepancies and misalignments between feature representations from the HSI and LiDAR branches, we introduce the Spectral–spatial–elevation Feature Calibration Fusion (SFCF) module. This module learns discriminative features from both data types, effectively bridging the gap and ensuring accurate fusion of the features from the two sources.
To summarize, the main contributions of this work are as follows:
(1)
We propose a novel joint HSI-LiDAR classification method called CEMA-Net, utilizing a hybrid two-branch CNN architecture to effectively extract 3D spectral–spatial features from HSI and 2D elevation features from LiDAR, significantly improving classification accuracy.
(2)
We introduce the Multi-way Feature Retention (MFR) module for adaptive feature extraction from HSI and LiDAR, along with the SAE and EAE modules to enhance spectral, spatial, and elevation awareness, improving feature representation.
(3)
We develop the Spectral–spatial–elevation Feature Calibration Fusion (SFCF) module to recalibrate discrepancies between HSI and LiDAR data, addressing feature differences and spatial misalignments for accurate and consistent fusion.
(4)
Our method outperforms state-of-the-art approaches, with experiments on three datasets consistently demonstrating the superior performance, effectiveness, and robustness of CEMA-Net in joint classification tasks.
The structure of this paper is organized as follows: Section 2 provides a detailed introduction to CEMA-Net, highlighting its key components and operational principles. Section 3 offers a comprehensive description of the experimental datasets, outlines the experimental setup, and presents an in-depth analysis of the classification results. Finally, Section 4 concludes the paper by summarizing the main findings and proposing potential directions for future research.

2. Methodology

Figure 1 provides a detailed visual representation of the proposed CEMA-Net, illustrating its comprehensive workflow. The framework adopts a dual-branch architecture to separately process HSI and LiDAR data. Each branch is specifically designed to address the unique characteristics of the corresponding data type, ensuring that their distinctive features are effectively captured and utilized for accurate classification.
The HSI branch focuses on extracting spectral and spatial features, utilizing modules such as the Spectral–spatial Aware Enhancement (SAE) module to enhance the network’s sensitivity to spectral variations and local spatial structures. This branch leverages convolutional layers and attention mechanisms to preserve both local and global spectral–spatial relationships. The LiDAR branch is dedicated to capturing spatial and elevation information. By incorporating the Elevation Aware Enhancement (EAE) module, this branch highlights the elevation-sensitive features that are critical for distinguishing ground objects. The two branches are integrated through the Spectral–spatial–elevation Feature Calibration Fusion (SFCF) module, which ensures effective alignment and fusion of the heterogeneous features from HSI and LiDAR data. This module not only addresses the representation disparities and spatial misalignments between the two modalities but also enhances the discriminative power of the combined features.
This refined structure allows CEMA-Net to maintain critical spectral, spatial, and elevation information throughout its processing pipeline, ultimately achieving superior classification performance compared to state-of-the-art methods.

2.1. HSI and LiDAR Data Preprocessing

For the analysis, HSI data are represented as X H R m × n × l and LiDAR data as X L R m × n . They include the same surface information and therefore have the same ground truth. Here, m and n correspond to the spatial dimensions, while l denotes the number of spectral bands in the HSI dataset. HSI data offer extensive spectral information, with each pixel associated with a one-hot category vector. However, the high dimensionality of spectral bands introduces significant computational challenges and can lead to redundancy, as adjacent spectral bands often carry overlapping information. To address these issues, we apply Principal Component Analysis (PCA) to reduce the dimensionality of the spectral data. PCA effectively retains the most significant spectral features by projecting the original data onto a smaller set of principal components, which not only mitigates redundancy but also enhances computational efficiency. Specifically, PCA extracts the top k principal components from the HSI data, thereby reducing the spectral band count from l to k while preserving the spatial dimensions. This results in a transformed dataset denoted as X H p c a R m × n × k .
Subsequently, we perform patch extraction using a sliding window of size s × s on both X H p c a and X L . This process generates 3D patches from the HSI data, denoted as X H P R s × s × k , and 2D patches from the LiDAR data, represented as X L P R s × s . Each patch is identified by the label of its central pixel. For edge pixels where the window size cannot be fully accommodated, we apply a padding technique with a width of ( s 1 ) / 2 to ensure consistent patch sizes.
Finally, we discard any pixel blocks with a label of zero and proceed to split the remaining samples into training and testing sets for further evaluation.

2.2. Spectral–Spatial–Elevation Feature Extraction

Taking advantage of Convolutional Neural Networks (CNNs) for their exceptional capability in modeling spatial context and extracting features, CNNs are particularly adept at analyzing the spectral–spatial patterns in HSI data and efficiently retrieving elevation information from LiDAR data. To achieve this, we deploy a 3D CNN that processes high-dimensional 3D patches, capturing complex spectral and spatial features for accurate local representation. Concurrently, a 2D CNN is utilized to specifically focus on the extraction of elevation-related features from the LiDAR data.
As depicted in Figure 1, we begin with the HSI data X H p c a R m × n × k by applying a 3D convolution (Conv3-D) to extract meaningful spectral–spatial features. The resulting feature cube’s spatial dimensions are flattened into a 2D vector. Subsequently, a 2D convolution (Conv2-D) is applied to minimize redundancy in both the spectral and spatial information.
In contrast to the processing of HSI data, the LiDAR data X L R m × n are subject to a different approach. We implement two Conv2-D layers to extract surface elevation features, using convolution kernel sizes of 16 @ 3 × 3 and 64 @ 3 × 3 . To accelerate training and enhance the model’s ability to capture nonlinear relationships, we introduce layer normalization and ReLU activation functions after each convolutional layer.

2.3. Multi-Way Feature Retention Module

In HSI classification tasks, relying solely on single-scale feature extraction can lead to the omission of critical spectral and elevation details. To address this, we introduce the Multi-way Feature Retention (MFR) module, designed to thoroughly investigate the rich spectral–spatial–elevation semantic information through an advanced multi-way feature extraction framework. Additionally, we propose the Spectral–spatial Aware Enhancement (SAE) module and the Elevation Aware Enhancement (EAE) module, which aim to enhance the sensitivity of features to spectral and elevation information, distinguishing between local and global perspectives by adjusting the awareness plate size.
The key strength of the MFR module is its capability for multi-level, fine-grained feature extraction. This allows for the capture of detailed features across various scales, effectively refining the representation of spectral–spatial–elevation information. We integrate the MFR module into both the HSI and LiDAR branches of our network, further enhancing them with the SAE and EAE modules.
Specifically, the MFR module operates through three parallel pathways: local, global, and sequential convolution. For the HSI data retention structure, the features are represented as X H c o n v R h × w × c , where h and w denote the height and width, while c represents the number of spectral bands. Initially, we apply point-wise convolution to obtain F H R h × w × c . This is then processed through both global and local pathways to yield F H g l o b a l R h × w × c and F H l o c a l R h × w × c . The SAE module is utilized here to balance and integrate global and local information. After passing through a sequential convolutional layer, we perform element-wise addition to combine the spectral–spatial features from both pathways, followed by another convolution to produce F H s e r i a l R h × w × c . The final output, F H o u t R h × w × c , is obtained by summing the three results.
For the retention structure of the LiDAR data, the processing flow mirrors that of the HSI data. However, the EAE module is employed to handle the global and local information specific to elevation data.

Spectral–Spatial/Elevation Aware Enhancement Module

Balancing the modeling distance selection between samples is a critical challenge when dealing with large-scale, high-dimensional remote sensing images. Extracting overly localized information can distort the overall trends of spectral curves, while focusing too much on global features may overlook important texture details. To address this, we introduce the Spectral–Spatial Aware Enhancement (SAE) and Elevation Aware Enhancement (EAE) modules, which facilitate a dynamic awareness of global and local information, thereby enhancing the understanding of surface spectral changes, spatial structures, and terrain elevations.
SAE Module: As illustrated in the upper section of Figure 2, we adjust the cube parameter U to various sizes for segmenting spectral–spatial features into multiple non-overlapping blocks. Here, U can be set to 4 × 4 for global processing or 2 × 2 for local processing. We then employ aggregation and displacement of these non-overlapping cubes to effectively extract critical features along the spectral dimension. Additionally, we introduce a trainable parameter α to emphasize task-relevant tokens and perform feature selection based on the similarity matrix of the non-overlapping cubes.
Specifically, we apply an unfolding operation to flatten the spectral–spatial features into a single dimension, partitioning them into spatially adjacent, non-overlapping patches, represented as F H h s i U × U × H U × W U × C . To extract detailed features from these spectral–spatial data, we compress the information of each spectral band by averaging, resulting in F H m U 2 × H W U 2 × 1 . We then use a feedforward network (FFN) for linear calculations to enhance feature representation. To assess the relevance of features for remote sensing classification, we apply a sigmoid activation function to derive confidence scores in the spatial dimension, followed by element-wise multiplication to fine-tune the corresponding features.
By integrating task-specific trainable tokens into the network, we generate HSI-weighted features, defined as ( f i ) i = 1 C , where f i R C . These tokens are added to each output feature, and feature selection is performed via a similarity matrix, expressed as follows:
f i ^ = s i m ( f i · α , f i ) · β
Here, α R C represents the learnable token parameter for HSI, highlighting relevant classification features, while β = H W p 2 serves as the learnable token parameter for LiDAR in the lower branch. The function s i m ( · , · ) computes cosine similarity, yielding values between 0 and 1. Each feature f i is reorganized according to its cosine similarity matrix to facilitate efficient simulated feature selection. We then execute reshaping and interpolation on each feature to produce the outputs from the global and local pathways, effectively enhancing the awareness of spectral–spatial features.
EAE Module: The awareness enhancement for elevation information, shown in the lower part of Figure 2, follows a similar processing approach as that for spectral–spatial features. Given that LiDAR elevation data are localized in the spatial dimension, we set different patch parameters P to divide it into multiple non-overlapping patches, denoted as F L p a t c h P × P × H P × W P × C . Typically, P is set to 4 or 2 to represent global and local processing, respectively. We apply similar aggregation and displacement operations among the patches to extract elevation details. The similarity matrix computation for feature selection also follows this process, with a trainable parameter β introduced as a task-specific token. The similarity matrix calculation for feature selection is given by the following:
f i ^ = s i m ( f i · β , f i ) · α
This completes the awareness enhancement of elevation features.

2.4. Spectral–Spatial–Elevation Feature Calibration Fusion

The integration of multi-source remote sensing data is essential for improving classification performance, yet it faces significant challenges due to differences in feature representations and misalignment in spatial data. To address these issues, we introduce a novel spectral–spatial–elevation feature calibration fusion module, which effectively merges complementary features from diverse data sources, ensuring accurate and coherent integration despite inherent disparities. It incorporates two key advantages: (1) efficient learning of discriminative features from multi-source mixed data and (2) adaptive calibration of spatial differences.
Specifically, we first combine HSI and LiDAR features and dynamically learn discriminative features from the multi-source data by considering the global information importance score. To address the substantial representation differences and spatial misalignments between the multi-source features, we separately feed the features from both data types into the calibration module. There, each sub-feature is refined by applying a learnable offset, enabling more precise feature extraction. Subsequently, to achieve effective fusion of multi-source heterogeneous features, we execute element-wise multiplication between the discriminative features and the calibrated features. This facilitates enhanced information interaction and integration. Details of the spectral–spatial–elevation feature calibration module are shown in Figure 3.
A. 
Discriminative Feature Extraction:
In order to obtain unique discriminative features, we first sum the both source features according to the channel dimensions to initially obtain the joint features F J R C × H × W . Following global average pooling (GAP), each channel is condensed into a feature map with spatial dimensions of 1 × 1 . Next, we introduce Sigmoid functions to normalize weights across 0 and 1. These weights are then utilized to adjust the joint feature to F J R C × H × W , thereby amplifying information-enriched bands while diminishing the impact of irrelevant ones. This feature will be added at the end with the two enhanced features from the calibration operation to achieve the alignment of the multi-source data deviation.
B. 
Adaptive Calibration Difference:
As depicted in Figure 3, we divide the spectral–spatial and elevation features into multiple sub-features according to the dimension of the channel, and learn subtle differences in spectral–spatial and elevation information through multiple learnable offsets.
We achieve feature calibration by employing feature resampling. Specifically, given HSI features F H R C × H × W , we first partition F H into G groups along the channel dimension. Spatial coordinates of pixel positions are defined as [ ( 1 , 1 ) , ( 1 , 2 ) , ( H , W ) ] . For each group, a learnable 2D offset Δ κ R 2 × × H × W is defined to learn spatial offsets. Our calibration function ζ ( · ) can be defined as follows:
M h , w = h H w W F H · max ( 0 , 1 | h + Δ h , w 1 h ) · max ( 0 , 1 | w + Δ h , w 2 w )
where M h , w represents the output obtained at pixel point p = ( h + Δ h , w 1 , w + Δ h , w 2 ) , where h and w are the height and width of the feature, respectively. Δ h , w 1 and Δ h , w 2 denote the displacement generated at the pixel point, w and h are the pixel positions after feature subsampling. For any position p, we enumerate all integral positions and use subsampling to obtain precise feature points at the sampling positions. Specifically, we perform subsampling on the current coordinates and obtain the most favorable features to replace the current pixel value for calibration.
F H typically contains rich spectral–spatial features, while F L contains more elevation features. Calibrating and merging offsets alone does not significantly enhance the performance of remote sensing classification missions. A main reason is that simple calibration and fusion cannot eliminate the feature differences of heterogeneous modes. Therefore, we further propose to selectively emphasize the calibrated complementary features through pixel masks to bridge the representational gap between them:
O = ( β H F H + β L F L ) F J
where β H and β L denote the gate mask.

3. Experimental Setup and Discussion

We provide a detailed description of the experimental setup and results. Initially, we outline the specific datasets, experimental configurations, and evaluation metrics used to ensure a fair comparison. Next, we perform quantitative experiments on three representative multi-modal datasets to showcase the superior performance of our proposed CEMA-Net. Additionally, we conduct comprehensive ablation studies to analyze the contribution of each component within the model. Finally, both quantitative and visual results demonstrate that CEMA-Net outperforms existing state-of-the-art methods in remote sensing classification tasks.

3.1. Description of the Datasets Used

Consistent with prior studies [26,38], we selected three widely recognized datasets for remote sensing classification: MUUFL, Trento, and Augsburg. Each dataset contains both HSI and LiDAR data, as outlined in Table 1, which summarize the sample count and land cover categories of each.
It is worth noting that HSI and LiDAR data in these datasets were not captured simultaneously due to the inherent differences in sensor technologies and acquisition conditions. However, the datasets underwent rigorous pre-registration processes to ensure spatial alignment, allowing pixel-to-pixel correspondence between the modalities. This alignment enables the effective fusion of spectral, spatial, and elevation information from HSI and LiDAR data, which is critical for our proposed method.
(1) MUUFL Dataset: Collected at the University of Southern Mississippi’s Gulf Park campus, the MUUFL dataset combines HSI with LiDAR data, captured using a reflective optical spectrometer [45]. The HSI data span 72 spectral bands from 0.38 to 1.05 μ m, while the LiDAR data are represented by two rasters at a wavelength of 1.06 μ m. With a resolution of 325 × 220 pixels, this dataset covers 11 land cover categories. To reduce noise, the initial and final eight spectral bands, which are most affected by noise, are omitted during training. Figure 4 illustrates the dataset with pseudo-color composite images of the HSI data, grayscale LiDAR images, and land cover category maps.
(2) Trento Dataset:
The dataset covers rural areas around Trento in southern Italy and includes six distinct scenes [46]. It combines LiDAR data, collected using the Optech ALTM 3100EA sensor, with hyperspectral data from the AISA Eagle system [39]. Both sensors achieve a spatial resolution of 1 m, with images sized at 600 × 166 pixels. The hyperspectral data contains 63 spectral bands spanning wavelengths from 402.89 nm to 989.09 nm, with spectral resolution ranging from 0.42 μ m to 0.99 μ m. The LiDAR data are represented as a single raster. Figure 5 displays the hyperspectral and LiDAR data, along with the ground truth land cover categories.
(3) Augsburg Dataset:
This dataset, collected from Augsburg, Germany, combines hyperspectral and LiDAR data for land cover analysis [47]. The hyperspectral data includes 180 spectral bands, covering wavelengths from 0.4 μ m (spanning UV and visible light) to 2.5 μ m (extending into the near-infrared). The LiDAR data, captured at a wavelength of 1.06 μ m, is represented as a single raster layer. Both data types have a spatial resolution of 30 m, with each image measuring 332 × 485 pixels. Figure 6 shows visual representations of the data, which categorize seven types of land cover.

3.2. Experimental Configuration

(1) Evaluation Metrics:
We use four standard metrics to assess the performance of our proposed CEMA-Net: overall accuracy (OA), average accuracy (AA), kappa coefficient (k), and per-class accuracy. These metrics serve as indicators of classification accuracy, and our objective is to achieve the highest possible values in each. Specifically, overall accuracy (OA) is the ratio of correctly classified pixels to the total number of pixels in the dataset. Average accuracy (AA) is the mean value of the classification accuracies for each individual class. Kappa coefficient (k) measures the agreement between the predicted and actual class labels, accounting for the possibility of agreement occurring by chance. Per-class accuracy evaluates the classification performance for each individual class, providing a detailed measure of how well each class is recognized. To ensure a fair and consistent comparison, all experiments are carried out with separate training and testing datasets.
(2) Environment Configuration:
Our CEMA-Net model is implemented in PyTorch 2.2.0, utilizing the Adam optimizer with parameters β 1 = 0.9 and β 2 = 0.999 . The training process spans 100 epochs, with a batch size of 64 and an initial learning rate set to 1 × 10 3 . It is executed on an NVIDIA Geforce RTX 4090 GPU with 24 GB of VRAM. In contrast, traditional methods for comparison are run on the MATLAB platform (https://matlab.zszlxx.cn/index.html?bd_vid=8287305137028280579, accessed on 9 November 2024).

3.3. Hyperparameter Configuration

(1) Patch Size: In the joint classification of HSI and LiDAR data, selecting the optimal patch size is crucial for balancing the ability to capture spatial distribution, spectral variations, and elevation information while keeping computational costs manageable. In the previous experiments, all parameters were fixed except for the patch size. To identify the best patch size, we evaluated the accuracy of several patch sizes from the set 7, 9, 11, 13, 15, 17. As illustrated in Figure 7, excessively large patch sizes result in more complex inputs, particularly impacting the scalar fusion module and diminishing the network’s fitting capacity, thus reducing accuracy. On the other hand, too small a patch size leads to a lack of sufficient contextual information, impairing global feature retention and accuracy. Based on these results, a patch size of 11 was found to be optimal for the classification task.
(2) Reduced Spectral Dimension: HSI offers extensive spectral information with hundreds of continuous bands, which can capture detailed features of ground objects. However, the broad spectral range and high sensitivity often result in a significant amount of redundant data, contributing to the problem of “dimensionality curse”. To address this challenge, we apply Principal Component Analysis (PCA) to extract the most critical spectral components, ensuring an optimal trade-off between spectral richness and computational efficiency. As illustrated in Figure 8, retaining too few spectral bands leads to a substantial loss of important information, while a larger number of bands introduces unnecessary redundancy and increases computational costs, thus reducing classification accuracy. In our experiments, we assessed different numbers of retained spectral bands from the set 5, 10, 15, 20, 25, 30, 35, 40. The results show that selecting 30 bands strikes the best balance, significantly enhancing the integration of HSI and LiDAR data and improving classification performance.
(3) Learning Rate: The learning rate is a crucial hyperparameter that governs the size of weight updates during model training, directly impacting the speed and stability of the learning process. A high learning rate can cause instability, leading to oscillations that prevent the model from converging effectively. Conversely, a low learning rate results in slower convergence, increasing both training time and computational cost. In hyperspectral image classification tasks, choosing an optimal learning rate is essential for maintaining training stability and enhancing final classification performance. We tested various learning rates from the set { 1 × 10 5 , 5 × 10 5 , 1 × 10 4 , 5 × 10 4 , 1 × 10 3 , 5 × 10 3 } to evaluate their impact on accuracy. As shown in Figure 9, the experiments reveal that a learning rate of 5 × 10 4 works best for the MUUFL and Augsburg datasets, while a learning rate of 1 × 10 3 yields the most accurate results for the Trento dataset.

3.4. Ablation Experiments

Based on the results of the ablation experiments in Table 2, we evaluate the effectiveness of each component within CEMA-Net, focusing on the MFR, SAE, EAE, and SFCF modules.
Case 1: When the MFR module is removed, the model’s OA, AA, and kappa scores decrease to 89.67%, 80.87%, and 86.65, respectively. This shows that the MFR module contributes significantly to the model’s feature extraction capabilities since its absence reduces classification accuracy across all metrics.
Case 2 and Case 3: These cases exclude the SAE and EAE modules, respectively. Without the SAE module (Case 2), the OA, AA, and kappa scores are 90.27%, 80.95%, and 87.05. Without the EAE module (Case 3), the scores slightly improve to 90.41%, 81.01%, and 87.17. This suggests that while each module independently enhances classification performance, both modules work together to better capture complex spectral–spatial–elevation information since removing either leads to reduced accuracy.
Case 4: In this configuration, the SFCF module is removed, and spectral–spatial and elevation features are directly added to the classifier. This causes the OA, AA, and kappa values to fall to 90.1%, 80.24%, and 86.45, respectively, indicating that the SFCF module is essential for effective feature fusion, mitigating spatial misalignments between HSI and LiDAR data.
Case 5: With all components included, CEMA-Net achieves its best performance, with OA, AA, and Kappa scores reaching 90.88%, 81.56%, and 87.98. This underscores the importance of each module in contributing to the overall classification accuracy, with the full model demonstrating optimal performance across all metrics.

3.5. Classification Result and Analysis

To highlight the effectiveness of our proposed CEMA-Net over other state-of-the-art methods, we selected several well-regarded classification techniques, organized into two categories. The first includes HSI-based classification methods: RF [20], SVM [21], 2D-CNN [48], HybridSN [29], and GAHT [33]. The second category comprises joint HSI and LiDAR fusion methods: CoupledCNN [36], CALC [38], HCTnet [39], and M2FNet [40]. For a fair comparison, we used each method’s default parameters as specified in the respective references, applied the same training set splits, and kept other parameters consistent with our setup. Each experiment was repeated ten times to ensure robustness, and both the mean and variance of the results were calculated to provide a comprehensive evaluation of the model’s performance and stability.
(1) 
Performance Evaluation and Analysis
Table 3, Table 4 and Table 5 provide a quantitative comparison of CEMA-Net against various state-of-the-art methods across three datasets. Each experiment was repeated 10 times to ensure reliability, with averages and standard deviations calculated for fair comparison. In each case, the best-performing results are emphasized in bold red. Our method consistently achieves top scores in overall accuracy (OA), average accuracy (AA), and the kappa coefficient on all datasets.
Table 3 presents detailed results of each method alongside CEMA-Net on the MUUFL dataset. Across all metrics, our CEMA-Net outperforms other approaches. Unlike traditional models like RF and SVM, deep learning methods capture a wider range of features, significantly improving classification accuracy. While models such as 2D-CNN, HybridSN, and GAHT are effective at feature extraction from HSI data, their OA results are lower than those for our CEMA-Net, likely due to limited receptive fields that miss global features. Furthermore, our CEMA-Net excels in joint HSI and LiDAR data classification, showing competitive performance in metrics like OA, AA, and kappa, and demonstrating strong average accuracy across classes. It surpasses methods like CoupledCNN, CALC, HCTnet, and M2FNet in AA, largely due to the multiple receptive fields in the local-global branch and the efficient fusion strategy for spectral-elevation feature alignment.
The Trento dataset, characterized by a limited number of categories and an uneven sample distribution, results in high overall accuracy (OA) and kappa scores across all methods, yet leaves substantial room for improvement in average accuracy (AA). Table 4 shows that joint classification approaches combining HSI and LiDAR data perform well, with methods like HCTnet and M2FNet achieving solid results. HCTnet’s dual Transformer encoder effectively integrates the unique features of these two remote sensing data types; however, its complex fusion process may lead to underfitting, particularly given Trento’s limited sample size. Our proposed CEMA-Net addresses this challenge effectively. The model’s MFR module is tailored to preserve detailed data, even with fewer samples, enabling our method to surpass HCTnet in both OA and kappa. Similarly, while M2FNet leverages multi-scale feature extraction to capture spectral–spatial–elevation information, our CEMA-Net’s SAE and EAE modules, which strengthen global–local feature awareness, provide a decisive edge, outperforming M2FNet in OA, AA, and kappa metrics.
The Augsburg dataset presents distinct challenges, with its high spatial resolution and complex object information, requiring models to balance local and global information effectively. The integration of LiDAR data further enhances the value of joint classification methods over single-source approaches. As Table 5 indicates, CALC’s use of dual adversarial networks for adversarial training in the object space yields high OA, benefiting from an adversarial strategy that merges spatial and elevation data efficiently. Nonetheless, our CEMA-Net’s SFCF module enhances spatial–elevation coupling by introducing additional offsets, leading to superior data fusion. This refinement allows our CEMA-Net to outperform CALC in key metrics such as OA, AA, and kappa, while the offset correction further improves fusion stability. Across all three datasets, our extensive comparisons emphasize that our CEMA-Net marks a significant improvement in joint classification, consistently achieving top results across metrics.
(2) 
Visual Assessment and Analysis
To further highlight the generalizability of our model under different conditions, we evaluated its performance across the MUUFL, Trento, and Augsburg datasets, which differ in environmental conditions, imaging methods, and spectral characteristics. On all datasets, our model demonstrated superior performance compared to others. For example, on the Trento dataset, which includes complex landscapes, CEMA-Net was able to differentiate between small and large objects with much greater clarity, while traditional methods showed blurred boundaries. On the Augsburg dataset, where the terrain features are more varied and the spectral data are more challenging, our method still maintained high classification accuracy, further proving its robustness under varying conditions.
The comparison across these datasets demonstrates the versatility of CEMA-Net, as it consistently produces high-quality results regardless of environmental or data acquisition challenges. Our model shows excellent performance in handling diverse conditions, showcasing its effectiveness in real-world applications where datasets may vary in scale, spectral information, and terrain complexity.
Figure 10, Figure 11 and Figure 12 illustrate the visualization results for various methods, enabling a qualitative comparison of their classification performance. The differences in classification accuracy between the methods are clearly visible. Notably, our proposed CEMA-Net produces cleaner, more accurate feature maps with minimal noise.
In detail, classification methods relying solely on HSI data tend to produce results with indistinct boundaries and substantial noise. While joint classification methods using multi-source data mitigate these issues to some extent, they still underperform in accurately classifying certain ground regions. In contrast, our approach yields results with sharp boundaries and high classification accuracy. For instance, Figure 10 displays the visualization results on the MUUFL dataset, highlighting that most methods struggle with blurriness and noise, especially in differentiating densely packed small areas across multiple classes. The classification map generated by our CEMA-Net, however, aligns more closely with the actual ground truth.
Figure 11 presents the visualization results of the comparison methods on the Trento dataset. This dataset, with larger ground objects and fewer categories compared to the MUUFL dataset, is relatively easier to classify. However, we can still observe that most of the comparison methods introduce significant noise, while our method consistently achieves high-precision classification. Additionally, Figure 12 shows the classification results on the Augsburg dataset, which has a higher resolution and more complex scenes. Even in these denser areas with more categories, our CEMA-Net continues to deliver superior classification performance, demonstrating its robustness across diverse datasets.

3.6. Exploration of Results for Various Data Modalities

The results of the ablation experiments for various data modalities are shown in Table 6, which evaluates the performance of the proposed method under different data configurations on three datasets: MUUFL, Trento, and Augsburg.
Only HSI: When using only HSI data, the classification performance varied across the datasets. The OA for MUUFL was 86.53%, for Trento it was 94.53%, and for Augsburg it was 92.86%. The AA and kappa coefficient (k) were also reported, with the highest AA of 92.28% for Trento. Although HSI data provide rich spectral information, it may not fully capture the spatial and elevation features needed for accurate classification, leading to lower performance in some cases.
Only LiDAR: When using only LiDAR data, the performance dropped significantly, with the OA of 69.26% for MUUFL, 73.25% for Trento, and 65.82% for Augsburg. LiDAR data alone lack the spectral information provided by HSI, and as a result, their ability to accurately classify land cover types is limited, especially in complex or similar terrain.
HSI + LiDAR: The combination of HSI and LiDAR data yielded the best performance across all datasets. For MUUFL, the OA reached 90.88%, for Trento it was 99.41%, and for Augsburg it was 95.91%. In terms of AA and the kappa coefficient, the joint modality significantly outperformed the individual modalities, achieving the highest AA (98.54% for Trento) and kappa score (99.22 for Trento). This demonstrates the complementary nature of HSI and LiDAR data, where the fusion of spectral, spatial, and elevation features allows for more precise land cover classification.
These results highlight the importance of integrating multi-modal data, especially when leveraging both HSI and LiDAR data to enhance classification accuracy in remote sensing applications. The proposed method shows clear improvements over using individual data types, confirming the effectiveness of joint spectral–spatial–elevation feature fusion.

3.7. Model Complexity Analysis

To validate the effectiveness of the proposed model, we conducted computational complexity experiments on the MUUFL dataset, as shown in Table 7. Compared to other state-of-the-art models, our method demonstrates the smallest parameter count (1.34 M), indicating a lightweight network architecture. Although the computational cost (1.34 G FLOPs) is comparable to some methods, the testing time of our model is only 0.92 s, making it the most efficient. Additionally, our model achieves the highest overall accuracy (OA) of 90.88%, slightly outperforming HTCnet (90.80%) and CALC (90.26%). These results highlight that the proposed method achieves a good balance between computational efficiency and classification performance, making it well-suited for resource-constrained scenarios, such as vehicle-based remote sensing or large-scale, real-time applications.

4. Conclusions

This paper presents CEMA-Net, a novel and highly efficient Calibration-Enhanced Multi-Awareness Network, aimed at achieving the effective fusion of spectral–spatial and elevation features to substantially boost accuracy in remote sensing classification tasks. At the core of our approach is the Multi-way Feature Retention (MFR) module, which leverages a multi-branch design to capture deep semantic information across spectral, spatial, and elevation domains, facilitating the comprehensive integration of global and local characteristics. Additionally, we introduce the Spectral–spatial Aware Enhancement (SAE) and Elevation Aware Enhancement (EAE) modules to emphasize features of ground objects that are especially responsive to spectral and elevation cues. To overcome representation discrepancies and spatial misalignments between multi-source features, we propose the Spectral–spatial–elevation Feature Calibration Fusion (SFCF) module, which effectively learns distinctive features from heterogeneous data and adaptively aligns spatial differences. Comprehensive experiments validate the advancement and resilience of our proposed method. As a future direction, we aim to adapt CEMA-Net for real-time remote sensing applications, such as autonomous vehicle navigation and drone-based land cover monitoring, where the timely and accurate classification of large-scale datasets is essential.

Author Contributions

Conceptualization, Q.Z. and Z.C.; Methodology, Q.Z. and Y.X.; Software, Q.Z. and Z.C.; Investigation, T.W.; Data curation, Z.L.; Writing—original draft, Q.Z.; Writing—review & editing, Y.X.; Visualization, Q.Z. and Z.C.; Supervision, Q.Z. and T.W.; Project administration, Q.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhou, Z.; Zheng, C.; Liu, X.; Tian, Y.; Chen, X.; Chen, X.; Dong, Z. A dynamic effective class balanced approach for remote sensing imagery semantic segmentation of imbalanced data. Remote Sens. 2023, 15, 1768. [Google Scholar] [CrossRef]
  2. Wang, J.; Hu, J.; Liu, Y.; Hua, Z.; Hao, S.; Yao, Y. El-nas: Efficient lightweight attention cross-domain architecture search for hyperspectral image classification. Remote Sens. 2023, 15, 4688. [Google Scholar] [CrossRef]
  3. Su, Z.; Wan, G.; Zhang, W.; Guo, N.; Wu, Y.; Liu, J.; Cong, D.; Jia, Y.; Wei, Z. An Integrated Detection and Multi-Object Tracking Pipeline for Satellite Video Analysis of Maritime and Aerial Objects. Remote Sens. 2024, 16, 724. [Google Scholar] [CrossRef]
  4. Zhang, G.; Fang, W.; Zheng, Y.; Wang, R. SDBAD-Net: A spatial dual-branch attention dehazing network based on meta-former paradigm. IEEE Trans. Circuits Syst. Video Technol. 2023, 34, 60–70. [Google Scholar] [CrossRef]
  5. Fang, W.; Zhang, G.; Zheng, Y.; Chen, Y. Multi-Task Learning for UAV Aerial Object Detection in Foggy Weather Condition. Remote Sens. 2023, 15, 4617. [Google Scholar] [CrossRef]
  6. Kuras, A.; Brell, M.; Rizzi, J.; Burud, I. Hyperspectral and lidar data applied to the urban land cover machine learning and neural-network-based classification: A review. Remote Sens. 2021, 13, 3393. [Google Scholar] [CrossRef]
  7. Audebert, N.; Le Saux, B.; Lefèvre, S. Deep learning for classification of hyperspectral data: A comparative review. IEEE Geosci. Remote Sens. Mag. 2019, 7, 159–173. [Google Scholar] [CrossRef]
  8. Khodadadzadeh, M.; Li, J.; Prasad, S.; Plaza, A. Fusion of hyperspectral and LiDAR remote sensing data using multiple feature learning. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2971–2983. [Google Scholar] [CrossRef]
  9. Ghamisi, P.; Benediktsson, J.A.; Phinn, S. Land-cover classification using both hyperspectral and LiDAR data. Int. J. Image Data Fusion 2015, 6, 189–215. [Google Scholar] [CrossRef]
  10. Murphy, R.J.; Taylor, Z.; Schneider, S.; Nieto, J. Mapping clay minerals in an open-pit mine using hyperspectral and LiDAR data. Eur. J. Remote Sens. 2015, 48, 511–526. [Google Scholar] [CrossRef]
  11. Voss, M.; Sugumaran, R. Seasonal effect on tree species classification in an urban environment using hyperspectral data, LiDAR, and an object-oriented approach. Sensors 2008, 8, 3020–3036. [Google Scholar] [CrossRef]
  12. Liu, L.; Coops, N.C.; Aven, N.W.; Pang, Y. Mapping urban tree species using integrated airborne hyperspectral and LiDAR remote sensing data. Remote Sens. Environ. 2017, 200, 170–182. [Google Scholar] [CrossRef]
  13. Gómez-Chova, L.; Tuia, D.; Moser, G.; Camps-Valls, G. Multimodal classification of remote sensing images: A review and future directions. Proc. IEEE 2015, 103, 1560–1584. [Google Scholar] [CrossRef]
  14. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep learning for hyperspectral image classification: An overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef]
  15. Kumar, B.; Dikshit, O.; Gupta, A.; Singh, M.K. Feature extraction for hyperspectral image classification: A review. Int. J. Remote Sens. 2020, 41, 6248–6287. [Google Scholar] [CrossRef]
  16. Ghamisi, P.; Höfle, B.; Zhu, X.X. Hyperspectral and LiDAR data fusion using extinction profiles and deep convolutional neural network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 10, 3011–3024. [Google Scholar] [CrossRef]
  17. Pal, M.; Mather, P.M. Support vector machines for classification in remote sensing. Int. J. Remote Sens. 2005, 26, 1007–1011. [Google Scholar] [CrossRef]
  18. Belgiu, M.; Drăguţ, L. Random forest in remote sensing: A review of applications and future directions. ISPRS J. Photogramm. Remote Sens. 2016, 114, 24–31. [Google Scholar] [CrossRef]
  19. Huang, F.; Lu, J.; Tao, J.; Li, L.; Tan, X.; Liu, P. Research on optimization methods of ELM classification algorithm for hyperspectral remote sensing images. IEEE Access 2019, 7, 108070–108089. [Google Scholar] [CrossRef]
  20. Ham, J.; Chen, Y.; Crawford, M.M.; Ghosh, J. Investigation of the random forest framework for classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 492–501. [Google Scholar] [CrossRef]
  21. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  22. Hang, R.; Liu, Q.; Song, H.; Sun, Y. Matrix-based discriminant subspace ensemble for hyperspectral image spatial–spectral feature fusion. IEEE Trans. Geosci. Remote Sens. 2015, 54, 783–794. [Google Scholar] [CrossRef]
  23. Wang, Q.; Gu, Y.; Tuia, D. Discriminative multiple kernel learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3912–3927. [Google Scholar] [CrossRef]
  24. Xing, C.; Wang, M.; Wang, Z.; Duan, C.; Liu, Y. Diagonalized Low-Rank Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–12. [Google Scholar] [CrossRef]
  25. Tejasree, G.; Agilandeeswari, L. An extensive review of hyperspectral image classification and prediction: Techniques and challenges. Multimed. Tools Appl. 2024, 83, 80941–81038. [Google Scholar] [CrossRef]
  26. Ahmad, M.; Shabbir, S.; Roy, S.K.; Hong, D.; Wu, X.; Yao, J.; Khan, A.M.; Mazzara, M.; Distefano, S.; Chanussot, J. Hyperspectral image classification—Traditional to deep models: A survey for future prospects. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 15, 968–999. [Google Scholar] [CrossRef]
  27. Paoletti, M.E.; Haut, J.M.; Plaza, J.; Plaza, A. Deep learning classifiers for hyperspectral imaging: A review. ISPRS J. Photogramm. Remote Sens. 2019, 158, 279–317. [Google Scholar] [CrossRef]
  28. Paoletti, M.E.; Haut, J.M.; Fernandez-Beltran, R.; Plaza, J.; Plaza, A.J.; Pla, F. Deep pyramidal residual networks for spectral–spatial hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 57, 740–754. [Google Scholar] [CrossRef]
  29. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN Feature Hierarchy for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 277–281. [Google Scholar] [CrossRef]
  30. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking hyperspectral image classification with transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–15. [Google Scholar] [CrossRef]
  31. Zou, J.; He, W.; Zhang, H. Lessformer: Local-enhanced spectral-spatial transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  32. Yuan, D.; Yu, D.; Qian, Y.; Xu, Y.; Liu, Y. S2Former: Parallel Spectral–Spatial Transformer for Hyperspectral Image Classification. Electronics 2023, 12, 3937. [Google Scholar] [CrossRef]
  33. Mei, S.; Song, C.; Ma, M.; Xu, F. Hyperspectral image classification using group-aware hierarchical transformer. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  34. Wen, C.; Yang, L.; Li, X.; Peng, L.; Chi, T. Directionally constrained fully convolutional neural network for airborne LiDAR point cloud classification. ISPRS J. Photogramm. Remote Sens. 2020, 162, 50–62. [Google Scholar] [CrossRef]
  35. Zorzi, S.; Maset, E.; Fusiello, A.; Crosilla, F. Full-waveform airborne LiDAR data classification using convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8255–8261. [Google Scholar] [CrossRef]
  36. Hang, R.; Li, Z.; Ghamisi, P.; Hong, D.; Xia, G.; Liu, Q. Classification of hyperspectral and LiDAR data using coupled CNNs. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4939–4950. [Google Scholar] [CrossRef]
  37. Zhang, M.; Li, W.; Tao, R.; Li, H.; Du, Q. Information fusion for classification of hyperspectral and LiDAR data using IP-CNN. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–12. [Google Scholar] [CrossRef]
  38. Lu, T.; Ding, K.; Fu, W.; Li, S.; Guo, A. Coupled adversarial learning for fusion classification of hyperspectral and LiDAR data. Inf. Fusion 2023, 93, 118–131. [Google Scholar] [CrossRef]
  39. Zhao, G.; Ye, Q.; Sun, L.; Wu, Z.; Pan, C.; Jeon, B. Joint classification of hyperspectral and LiDAR data using a hierarchical CNN and transformer. IEEE Trans. Geosci. Remote Sens. 2022, 61, 1–16. [Google Scholar] [CrossRef]
  40. Sun, L.; Wang, X.; Zheng, Y.; Wu, Z.; Fu, L. Multiscale 3-D–2-D Mixed CNN and Lightweight Attention-Free Transformer for Hyperspectral and LiDAR Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–16. [Google Scholar] [CrossRef]
  41. Mäyrä, J.; Keski-Saari, S.; Kivinen, S.; Tanhuanpää, T.; Hurskainen, P.; Kullberg, P.; Poikolainen, L.; Viinikka, A.; Tuominen, S.; Kumpula, T.; et al. Tree species classification from airborne hyperspectral and LiDAR data using 3D convolutional neural networks. Remote Sens. Environ. 2021, 256, 112322. [Google Scholar] [CrossRef]
  42. Mohla, S.; Pande, S.; Banerjee, B.; Chaudhuri, S. Fusatnet: Dual attention based spectrospatial multimodal fusion network for hyperspectral and lidar classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 13–19 June 2020; pp. 92–93. [Google Scholar]
  43. Wang, X.; Feng, Y.; Song, R.; Mu, Z.; Song, C. Multi-attentive hierarchical dense fusion net for fusion classification of hyperspectral and LiDAR data. Inf. Fusion 2022, 82, 1–18. [Google Scholar] [CrossRef]
  44. Dong, W.; Zhang, T.; Qu, J.; Xiao, S.; Zhang, T.; Li, Y. Multibranch feature fusion network with self-and cross-guided attention for hyperspectral and LiDAR classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–12. [Google Scholar] [CrossRef]
  45. Du, X.; Zare, A. Scene Label Ground Truth Map for MUUFL Gulfport Data Set. 2017. Available online: https://ufdc.ufl.edu/IR00009711/00001/pdf (accessed on 24 December 2024).
  46. Trento Dataset. Available online: https://github.com/A-Piece-Of-Maple/TrentoDateset?tab=readme-ov-file (accessed on 24 December 2024).
  47. Hu, J.; Liu, R.; Hong, D.; Camero, A.; Yao, J.; Schneider, M.; Kurz, F.; Segl, K.; Zhu, X.X. MDAS: A New Multimodal Benchmark Dataset for Remote Sensing. Earth Syst. Sci. Data Discuss. 2022, 15, 113–131. [Google Scholar] [CrossRef]
  48. Zhao, W.; Du, S. Spectral–spatial feature extraction for hyperspectral image classification: A dimension reduction and deep learning approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
Figure 1. Overview of the CEMA-Net Architecture.
Figure 1. Overview of the CEMA-Net Architecture.
Electronics 14 00102 g001
Figure 2. Illustration of spectral–spatial-aware enhancement and elevation-aware enhancement.
Figure 2. Illustration of spectral–spatial-aware enhancement and elevation-aware enhancement.
Electronics 14 00102 g002
Figure 3. Illustration of Spectral–Spatial–Elevation Calibration Fusion Module.
Figure 3. Illustration of Spectral–Spatial–Elevation Calibration Fusion Module.
Electronics 14 00102 g003
Figure 4. Visual depiction of the MUUFL dataset: (a) False-color composite image of the hyperspectral data. (b) Ground truth map showcasing the various land cover categories. (c) indicates the feature category. Red box indicates the enlarged display.
Figure 4. Visual depiction of the MUUFL dataset: (a) False-color composite image of the hyperspectral data. (b) Ground truth map showcasing the various land cover categories. (c) indicates the feature category. Red box indicates the enlarged display.
Electronics 14 00102 g004
Figure 5. Visual depiction of the Trento dataset. (a) False-color composite image of the hyperspectral data. (b) Ground truth map showcasing the various land cover categories. (c) indicates the feature category. Red box indicates the enlarged display.
Figure 5. Visual depiction of the Trento dataset. (a) False-color composite image of the hyperspectral data. (b) Ground truth map showcasing the various land cover categories. (c) indicates the feature category. Red box indicates the enlarged display.
Electronics 14 00102 g005
Figure 6. Visual depiction of the Augsburg dataset. (a) False-color composite image of the hyperspectral data. (b) Ground truth map showcasing the various land cover categories. (c) indicates the feature category. Red box indicates the enlarged display.
Figure 6. Visual depiction of the Augsburg dataset. (a) False-color composite image of the hyperspectral data. (b) Ground truth map showcasing the various land cover categories. (c) indicates the feature category. Red box indicates the enlarged display.
Electronics 14 00102 g006
Figure 7. Impact of patch size on classification metrics: OA, AA, and kappa coefficient. (a) MUUFL. (b) Trento. (c) Augsburg.
Figure 7. Impact of patch size on classification metrics: OA, AA, and kappa coefficient. (a) MUUFL. (b) Trento. (c) Augsburg.
Electronics 14 00102 g007
Figure 8. Impact of reduced spectral dimensionality on classification metrics: OA, AA, and kappa coefficient. (a) MUUFL. (b) Trento. (c) Augsburg.
Figure 8. Impact of reduced spectral dimensionality on classification metrics: OA, AA, and kappa coefficient. (a) MUUFL. (b) Trento. (c) Augsburg.
Electronics 14 00102 g008
Figure 9. Impact of learning rate on classification metrics: OA, AA, and kappa coefficient. (a) MUUFL. (b) Trento. (c) Augsburg.
Figure 9. Impact of learning rate on classification metrics: OA, AA, and kappa coefficient. (a) MUUFL. (b) Trento. (c) Augsburg.
Electronics 14 00102 g009
Figure 10. Maps depicting the classification of MUUFL using various methods. (a) RF. (b) SVM. (c) 2D-CNN. (d) HybridSN. (e) GAHT. (f) CoupledCNN. (g) CALC. (h) HCTnet. (i) M2FNet. (j) our CEMA-Net. Red box indicates the enlarged display.
Figure 10. Maps depicting the classification of MUUFL using various methods. (a) RF. (b) SVM. (c) 2D-CNN. (d) HybridSN. (e) GAHT. (f) CoupledCNN. (g) CALC. (h) HCTnet. (i) M2FNet. (j) our CEMA-Net. Red box indicates the enlarged display.
Electronics 14 00102 g010
Figure 11. Maps depicting the classification of Trento using various methods. (a) RF. (b) SVM. (c) 2D-CNN. (d) HybridSN. (e) GAHT. (f) CoupledCNN. (g) CALC. (h) HCTnet. (i) M2FNet. (j) our CEMA-Net. Red box indicates the enlarged display.
Figure 11. Maps depicting the classification of Trento using various methods. (a) RF. (b) SVM. (c) 2D-CNN. (d) HybridSN. (e) GAHT. (f) CoupledCNN. (g) CALC. (h) HCTnet. (i) M2FNet. (j) our CEMA-Net. Red box indicates the enlarged display.
Electronics 14 00102 g011aElectronics 14 00102 g011b
Figure 12. Maps depicting the classification of Augsburg using various methods. (a) RF. (b) SVM. (c) 2D-CNN. (d) HybridSN. (e) GAHT. (f) CoupledCNN. (g) CALC. (h) HCTnet. (i) M2FNet. (j) Our CEMA-Net. Red box indicates the enlarged display.
Figure 12. Maps depicting the classification of Augsburg using various methods. (a) RF. (b) SVM. (c) 2D-CNN. (d) HybridSN. (e) GAHT. (f) CoupledCNN. (g) CALC. (h) HCTnet. (i) M2FNet. (j) Our CEMA-Net. Red box indicates the enlarged display.
Electronics 14 00102 g012
Table 1. Training and test samples in MUUFL, Trento, and Augsburg.
Table 1. Training and test samples in MUUFL, Trento, and Augsburg.
ClassMUUFLTrentoAugsburg
Land CoverTrainTestLand CoverTrainTestLand CoverTrainTest
01Trees Mostly46522,781Apple Trees413993Forest27113,236
02Grass864184Buildings302873Residential Area60729,722
03Mixed Ground Surface1386744Ground5474Industrial Area783773
04Dirt and Sand371789Woods929031Low Plants53826,319
05Road1346553Vineyard10610,395Allotment12563
06Water10456Roads323142Commercial Area331612
07Buildings Shadow452188   Water311499
08Buildings1256115     
09Sidewalk281357     
10Yellow Curb4179     
11Cloth Panels6263     
Total107852,609Total30629,908Total157076,724
Table 2. Results of ablation experiments. ✓ indicates selected, × indicates unselected. Bold indicate optimal performance.
Table 2. Results of ablation experiments. ✓ indicates selected, × indicates unselected. Bold indicate optimal performance.
CasesComponentIndicators
MFRSAEEAESFCFOA (%)AA (%) k 100
1×89.6780.8786.65
2×90.2780.9587.05
3×90.4181.0187.17
4×90.180.2486.45
590.8881.5687.98
Table 3. Classification accuracy of MUUFL using different methods. Bold and highlighted in red indicate optimal performance.
Table 3. Classification accuracy of MUUFL using different methods. Bold and highlighted in red indicate optimal performance.
No.Only HSI InputHSI and LiDAR Input
RF [20] SVM [21] 2D-CNN [48] HybridSN [29] GAHT [33] CoupledCNN [36] CALC [38] HCTnet [39] M2FNet [40] Ours
195.86 ± 0.6394.67 ± 1.1396.88 ± 0.5594.42 ± 1.5695.13 ± 1.2195.85 ± 0.3196.87 ± 0.496.78 ± 0.3196.74 ± 0.3396.08 ± 0.6
276.33 ± 4.1282.76 ± 3.6879.11 ± 1.2680.39 ± 1.6373.98 ± 3.7578.73 ± 1.0981.07 ± 2.0682.19 ± 2.181.49 ± 1.080.6 ± 3.88
380.68 ± 7.1185.03 ± 2.0882.05 ± 2.3779.59 ± 2.5681.6 ± 2.2586.57 ± 2.4883.84 ± 0.8885.95 ± 0.5584.34 ± 2.5587.44 ± 1.49
486.73 ± 3.5990.54 ± 3.4185.29 ± 4.6788.64 ± 1.4280.28 ± 4.6988.34 ± 3.4489.49 ± 2.3988.06 ± 2.6589.03 ± 2.2195.4 ± 1.68
586.45 ± 1.488.85 ± 1.1386.41 ± 1.1486.56 ± 0.8985.71 ± 3.7292.5 ± 0.7592.79 ± 0.4392.5 ± 0.9888.71 ± 4.3888.92 ± 0.59
690.99 ± 0.3791.76 ± 3.7784.97 ± 4.6372.57 ± 24.6263.68 ± 3.6586.36 ± 3.0690.79 ± 3.3494.34 ± 2.896.71 ± 2.3298.64 ± 1.42
775.41 ± 1.3279.88 ± 7.874.5 ± 4.8881.62 ± 2.5461.5 ± 9.0878.03 ± 3.2176.32 ± 2.3276.35 ± 1.5180.97 ± 6.0687.92 ± 3.07
895.06 ± 0.8194.04 ± 0.2394.99 ± 0.6693.11 ± 1.1892.6 ± 1.6296.07 ± 0.2897.7 ± 0.5196.58 ± 0.5196.29 ± 1.8296.15 ± 0.35
929.64 ± 10.433.33 ± 6.0639.11 ± 0.9834.71 ± 4.1621.4 ± 7.2933.93 ± 4.0832.07 ± 3.8646.96 ± 2.1946.28 ± 5.9341.88 ± 4.49
101.11 ± 0.454.28 ± 1.1513.22 ± 4.3116.76 ± 6.1710.99 ± 4.114.92 ± 2.772.35 ± 1.4713.52 ± 1.8521.45 ± 5.1330.77 ± 3.17
1140.75 ± 1.1161.74 ± 12.1769.95 ± 7.2884.97 ± 4.9653.41 ± 6.2161.29 ± 3.8673.31 ± 3.6968.97 ± 3.4574.14 ± 12.5493.41 ± 1.13
OA(%)87.59 ± 0.6388.85 ± 0.3888.71 ± 0.2787.50 ± 0.9085.67 ± 0.0489.75 ± 0.5890.26 ± 0.3190.80 ± 0.2490.30 ± 0.8490.88 ± 0.29
AA(%)69.00 ± 1.2373.35 ± 0.3973.32 ± 0.5273.94 ± 1.9465.48 ± 1.2972.96 ± 1.5274.24 ± 0.8176.56 ± 0.6477.83 ± 1.0381.56 ± 0.67
k × 10083.51 ± 0.8085.29 ± 0.4884.97 ± 0.3683.55 ± 1.1080.90 ± 0.0986.43 ± 0.7887.08 ± 0.3987.80 ± 0.3287.13 ± 1.1487.98 ± 0.40
Table 4. Classification accuracy of Trento using different methods. Bold and highlighted in red indicate optimal performance.
Table 4. Classification accuracy of Trento using different methods. Bold and highlighted in red indicate optimal performance.
No.Only HSI InputHSI and LiDAR Input
RF [20] SVM [21] 2D-CNN [48] HybridSN [29] GAHT [33] CoupledCNN [36] CALC [38] HCTnet [39] M2FNet [40] Ours
198.59 ± 0.498.67 ± 0.6599.32 ± 0.4196.09 ± 2.8598.74 ± 0.5999.85 ± 0.0899.67 ± 0.1599.32 ± 0.2199.46 ± 0.5298.97 ± 0.53
294.15 ± 1.5896.62 ± 0.2993.91 ± 2.785.59 ± 0.3488.82 ± 6.3499.22 ± 0.398.97 ± 0.7394.87 ± 0.797.42 ± 1.3298.26 ± 0.25
360.69 ± 7.7560.54 ± 21.2166.46 ± 4.0652.95 ± 19.3934.74 ± 6.7659.32 ± 10.5771.98 ± 11.3990.97 ± 3.3780.89 ± 4.596.12 ± 1.38
4100.0 ± 0.0100.0 ± 0.099.97 ± 0.0499.88 ± 0.1299.64 ± 0.48100.0 ± 0.099.97 ± 0.0299.97 ± 0.03100.0 ± 0.0100.0 ± 0.0
599.97 ± 0.0299.98 ± 0.0398.65 ± 0.9697.87 ± 0.8991.29 ± 12.3299.98 ± 0.0299.97 ± 0.0199.9 ± 0.0799.93 ± 0.04100.0 ± 0.0
690.67 ± 0.4490.53 ± 0.7683.57 ± 2.1886.24 ± 2.1295.59 ± 0.7397.13 ± 0.4394.55 ± 0.8895.98 ± 1.1797.26 ± 0.1597.91 ± 0.36
OA(%)97.64 ± 0.2897.87 ± 0.2196.59 ± 0.6795.13 ± 0.3494.12 ± 5.0798.95 ± 0.1598.82 ± 0.2198.81 ± 0.1999.07 ± 0.1299.41 ± 0.06
AA(%)90.68 ± 1.5191.06 ± 3.3590.31 ± 0.4386.44 ± 3.2984.80 ± 4.1992.58 ± 1.7494.19 ± 1.9096.84 ± 0.7295.83 ± 0.7398.54 ± 0.22
k × 10096.84 ± 0.3897.15 ± 0.2895.45 ± 0.9093.48 ± 0.4692.31 ± 6.5598.60 ± 0.2098.43 ± 0.2898.41 ± 0.2598.75 ± 0.1699.22 ± 0.08
Table 5. Classification accuracy of Augsburg using different methods. Bold and highlighted in red indicate optimal performance.
Table 5. Classification accuracy of Augsburg using different methods. Bold and highlighted in red indicate optimal performance.
No.Only HSI InputHSI and LiDAR Input
RF [20] SVM [21] 2D-CNN [48] HybridSN [29] GAHT [33] CoupledCNN [36] CALC [38] HCTnet [39] M2FNet [40] Ours
198.47 ± 0.1496.73 ± 2.0298.28 ± 0.2398.17 ± 0.3797.57 ± 0.6197.12 ± 0.8298.19 ± 0.1698.96 ± 0.3398.66 ± 0.4599.08 ± 0.18
298.19 ± 0.1998.33 ± 0.0897.7 ± 0.1897.0 ± 0.2797.66 ± 0.897.85 ± 0.7498.55 ± 0.1897.8 ± 0.1196.95 ± 1.5197.49 ± 0.26
376.85 ± 3.5876.54 ± 1.9879.26 ± 2.2371.72 ± 2.8275.78 ± 8.0782.67 ± 3.3585.08 ± 1.9776.24 ± 2.677.97 ± 9.2784.21 ± 0.72
498.65 ± 0.2798.82 ± 0.1598.67 ± 0.0798.76 ± 0.0595.49 ± 1.3398.53 ± 0.2798.7 ± 0.0798.61 ± 0.2198.69 ± 0.398.1 ± 0.25
521.46 ± 5.7820.63 ± 6.4163.0 ± 3.1161.41 ± 2.9748.52 ± 15.6776.13 ± 8.5672.47 ± 3.6360.81 ± 11.0782.74 ± 4.5392.27 ± 1.19
619.42 ± 8.619.7 ± 2.5432.71 ± 5.9539.44 ± 2.7759.74 ± 6.3551.81 ± 3.8737.08 ± 4.9835.15 ± 3.945.38 ± 16.6560.4 ± 4.22
750.98 ± 2.0650.13 ± 9.1560.09 ± 2.0562.86 ± 1.844.47 ± 1.458.81 ± 3.3961.53 ± 2.1255.66 ± 0.8566.14 ± 1.6266.76 ± 3.11
OA(%)94.21 ± 0.1693.98 ± 0.5994.87 ± 0.0894.42 ± 0.1393.63 ± 0.5995.32 ± 0.2695.67 ± 0.0994.80 ± 0.2695.12 ± 0.5195.91 ± 0.09
AA(%)66.29 ± 1.6765.84 ± 2.3375.67 ± 0.5175.62 ± 1.0974.18 ± 3.2080.42 ± 1.3978.80 ± 0.6174.75 ± 2.1380.93 ± 1.6185.47 ± 1.23
k×10091.63 ± 0.2591.30 ± 0.8692.63 ± 0.1192.01 ± 0.2090.86 ± 0.8393.28 ± 0.3993.78 ± 0.1292.52 ± 0.3793.01 ± 0.7094.15 ± 0.13
Table 6. Results of ablation experiments for various data modalities. Bold indicate optimal performance.
Table 6. Results of ablation experiments for various data modalities. Bold indicate optimal performance.
CasesMUUFLTrentoAugsburg
OA (%) AA (%) k × 100 OA (%) AA (%) k × 100 OA (%) AA (%) k × 100
Only HSI86.5371.6581.3394.5392.2893.5692.8677.4188.39
Only LiDAR69.2657.8266.4273.2571.7373.3265.8257.6262.42
HSI + LiDAR90.8881.5687.9899.4198.5499.2295.9185.4794.15
Table 7. Compute complexity results on several models on the MUUFL dataset.
Table 7. Compute complexity results on several models on the MUUFL dataset.
MetricsCoupledCNNCALCHTCnetM2FNetOurs
Parameters (M)1.981.751.851.671.34
Flops (G)0.981.561.371.211.34
Testing Time (s)0.761.151.371.050.92
OA (%)89.7590.2690.8090.3090.88
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Q.; Cui, Z.; Wang, T.; Li, Z.; Xia, Y. Calibration-Enhanced Multi-Awareness Network for Joint Classification of Hyperspectral and LiDAR Data. Electronics 2025, 14, 102. https://doi.org/10.3390/electronics14010102

AMA Style

Zhang Q, Cui Z, Wang T, Li Z, Xia Y. Calibration-Enhanced Multi-Awareness Network for Joint Classification of Hyperspectral and LiDAR Data. Electronics. 2025; 14(1):102. https://doi.org/10.3390/electronics14010102

Chicago/Turabian Style

Zhang, Quan, Zheyuan Cui, Tianhang Wang, Zhaoxin Li, and Yifan Xia. 2025. "Calibration-Enhanced Multi-Awareness Network for Joint Classification of Hyperspectral and LiDAR Data" Electronics 14, no. 1: 102. https://doi.org/10.3390/electronics14010102

APA Style

Zhang, Q., Cui, Z., Wang, T., Li, Z., & Xia, Y. (2025). Calibration-Enhanced Multi-Awareness Network for Joint Classification of Hyperspectral and LiDAR Data. Electronics, 14(1), 102. https://doi.org/10.3390/electronics14010102

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop