4.1. Dataset and Evaluation Metrics
Dataset: The experimental data used in this study are derived from the WHU-OPT-SAR dataset [
15], which is the first publicly available optical–SAR paired dataset for land-use classification. Released by Wuhan University, this dataset covers approximately 50,000 km
2 of Hubei Province (30°–33°N, 108°–117°E) and contains 100 coregistered image pairs, each with a spatial resolution of
pixels. The optical images are acquired from the GF-1 satellite (with a resolution of 2 m), while the SAR images are obtained from the GF-3 satellite (with a resolution of 5 m). To ensure pixelwise alignment, the optical images are resampled to a 5 m resolution using bilinear interpolation and aligned with the SAR images at subpixel accuracy. Pixel-level land use labels are generated based on the 2017 National Land Use Change Survey in China. The dataset contains seven land cover categories: cropland, urban, rural, water, forestland, road, and others.
Data Processing and Experimental Details: During the preprocessing phase, the original remote sensing images were cropped into nonoverlapping patches of size . From these, a total of 8850 image slices were randomly selected to construct the experimental sample set. The dataset was subsequently split into training and testing subsets at a ratio of 8:2 to evaluate the performance of multimodal adversarial attacks and defence strategies.
All the experiments were implemented using the PyTorch 1.13.1 framework and were conducted on a workstation equipped with an NVIDIA RTX A6000 GPU (NVIDIA Corporation, Santa Clara, CA, USA). To simulate realistic adversarial scenarios, three classical attack methods—FGSM, PGD, and BIM—were adapted for multimodal inputs and injected into the dataset. These attacks introduce subtle perturbations to the input to mislead the model, with the perturbation magnitude set to 0.01. For multimodal PGD and MIM attacks, the step size was set to one-eighth of the maximum perturbation, and the number of iterations was fixed at 10. During adversarial training, we employed the Adam optimizer with an initial learning rate of , a weight decay of , a batch size of 64, and a total of 40 training epochs. These settings ensured thorough optimization of the proposed adversarial defence method.
Evaluation Metrics: To assess the classification performance of the model under adversarial attacks quantitatively, we adopt two widely used metrics: overall accuracy (OA) and the kappa coefficient. OA measures the proportion of correctly classified samples and is defined as follows:
where
C denotes the total number of classes,
represents the number of correctly classified samples for class
i in the confusion matrix, and
N is the total number of samples.
The kappa coefficient is used to evaluate the agreement between the classification results and those obtained by random chance and is defined as follows:
where
is the observed agreement and
is the expected agreement by chance. Here,
and
denote the sum of the
i-th row and
i-th column of the confusion matrix, respectively. By computing OA and
under different levels of adversarial perturbations (FGSM, PGD, and MIM), we comprehensively evaluate the effectiveness of the proposed adversarial defence method.
4.2. Adversarial Attack Experimental Results
We first evaluate the robustness of both unimodal and multimodal remote sensing image classification models under three classical white-box adversarial attack methods: FGSM, PGD, and MIM. The experimental results are summarized in
Table 1. With respect to unimodal models, adding slight perturbations to the input images leads to a significant degradation in classification performance. Specifically, the unimodal model using only RGB images exhibits a dramatic decrease in mean overall accuracy (OA), which decreases from 0.8237 on clean examples to 0.4313 under adversarial attacks—an absolute reduction of 0.3924. This result indicates that the RGB-based model is highly vulnerable to adversarial perturbations and lacks robustness. In contrast, the unimodal model using only SAR images experiences a smaller decrease in mean OA, from 0.8456 to 0.5925. The relatively lower sensitivity of SAR-based models to adversarial noise suggests that the unique statistical characteristics of SAR data may limit the effectiveness of adversarial perturbation generation, resulting in less severe performance degradation compared with their RGB counterparts.
In the multimodal attack experiments, the remote sensing classification model that integrates both RGB and SAR imagery achieves an OA of 0.8779 under clean conditions. However, when adversarial perturbations are simultaneously applied to both modalities, the average OA decreases to 0.6517, resulting in a performance degradation of 0.3262. These results indicate that although the multimodal model offers superior classification performance under benign conditions, it still suffers from substantial accuracy loss when subjected to adversarial attacks, reflecting its limited robustness. Furthermore, the degree of performance degradation observed in the multimodal model lies between that of the RGB-only and SAR-only unimodal models. This suggests that although multimodal fusion enhances robustness to some extent, it remains insufficient to effectively resist highly targeted adversarial attacks.
We further evaluated the robustness of the multimodal classification model under unimodal adversarial attacks, with the experimental results presented in
Table 2. As shown in the table, perturbations applied to the optical modality exhibit a significantly stronger attack effect than those targeting the SAR modality do. For example, under the FGSM attack, the model’s OA decreases from 0.8779 to 0.7178 when the optical modality is attacked, whereas attacking only the SAR modality results in a smaller decline to 0.7682. Iterative attacks such as PGD and MIM, which leverage multistep optimization, are generally more destructive than single-step FGSM. Nevertheless, the OA of the SAR modality remains relatively high even under these stronger attacks, demonstrating superior adversarial robustness. This performance gap may stem from the intrinsic differences in sensing mechanisms and feature representations between the two modalities. Optical imagery heavily relies on fine-grained visual features such as colour and texture, making it more sensitive to pixel-level perturbations. In contrast, SAR imagery, owing to its unique imaging process and inherent robustness to noise, preserves semantic consistency more effectively when exposed to adversarial interference, thereby exhibiting stronger resistance to attacks.
4.3. Adversarial Defence Experimental Results
To validate the effectiveness of the proposed CAGMC-Defence method, we conducted a systematic evaluation of its defence performance under three typical adversarial attacks (FGSM, PGD, and MIM) and compared it with seven mainstream defence methods, namely, HFS [
28], Lagrangian-AT [
34], DDC-AT [
35], LBGAT [
36], DKL [
37], GAT [
38], and AT-UR [
39]. All methods were trained under the same configuration to ensure fair comparison, with 40 training epochs, a learning rate of
, and the Adam optimizer.
Table 3,
Table 4 and
Table 5 report the overall accuracy (OA) and kappa coefficient of each method under varying perturbation strengths, providing a quantitative assessment of their adversarial robustness.
Under adversarial attacks, all methods showed decreased classification performance as the perturbation strength increased, but the extent of degradation varied. HFS and DDC-AT maintained moderate accuracy under small perturbations, yet their robustness decreased sharply with stronger attacks. Under PGD and MIM attacks with , the OA of HFS decreased to 0.4324 and 0.4474, respectively, whereas that of DDC-AT decreased to 0.5500 and 0.5441, respectively. These results indicate their limited effectiveness against high-intensity and complex attacks. In contrast, the Lagrangian-AT, LBGAT, and DKL performed well under FGSM attacks. The OA of the Lagrangian-AT was above 0.84 across all levels of perturbation strength and still reached 0.8433 at . Both the LBGAT and DKL maintained OA levels consistently above 0.87, reflecting their good adaptability to single-step attacks. However, all three methods experienced significant performance decreases under PGD and MIM attacks. In particular, the OA of DKL decreased to 0.5630 under the MIM with , which was notably lower than those of the other methods, revealing its limitations in handling complex adversarial scenarios.
Furthermore, AT-UR and GAT demonstrated stronger robustness under multistep attacks. For example, AT-UR remained stable under PGD attacks, achieving an OA of 0.8645 when . However, its robustness decreased under MIM attacks. When , the OA decreased to 0.7843, and the kappa coefficient decreased to 0.3149, indicating a weakness in adapting to higher-order attacks. In contrast, the GAT delivered consistently strong performance across all three types of attacks. In particular, under PGD and MIM attacks with , OA values of 0.8584 and 0.8477, respectively, were achieved. These results were higher than those of all the other methods except CAGMC-Defence, demonstrating GAT’s superior generalization ability and stable defence performance.
Overall, CAGMC-Defence consistently demonstrates more robust defence performance across all three types of adversarial attacks. Not only does it achieve the highest classification accuracy under clean conditions, but it also exhibits significantly less performance degradation under strong perturbations. For example, under PGD attacks with a perturbation strength of 0.05, CAGMC-Defence maintains an OA of 0.8682—outperforming both AT-UR and the LBGAT. Similarly, in the presence of MIM attacks, CAGMC-Defence achieves the highest OA at the maximum perturbation level, surpassing the second-best method (LBGAT) and indicating a more balanced robustness and defence capability. Further statistical analysis reveals that the average decrease in the OA of CAGMC-Defence across all three attacks is only 0.0341, which is notably smaller than those of the other compared methods. Moreover, it results in the smallest reduction in the kappa coefficient, suggesting that CAGMC-Defence maintains high classification consistency even under adversarial conditions.
In summary, the above experimental results comprehensively demonstrate the robustness advantages of CAGMC-Defence in multimodal remote sensing image recognition tasks under various adversarial attacks. CAGMC-Defence not only significantly improves recognition accuracy under clean conditions but also maintains stable defence performance against a range of complex attack strategies. These results highlight its strong generalizability and practical applicability. Visual comparisons of different defence methods under various adversarial attacks are illustrated in
Figure 5,
Figure 6 and
Figure 7.
To further evaluate the defence performance of the CAGMC-Defence model against single-modality adversarial attacks, we applied the three typical attack methods to either the optical or SAR modality under a fixed perturbation strength of 0.01. The overall classification accuracies (OAs) after defence are reported in
Table 6. To highlight the effectiveness of the proposed defence strategy, we also compare these results with the corresponding attack outcomes without defence, as shown in
Table 2.
Specifically, when the optical modality was attacked by the FGSM, PGD, and MIM, the OA of the undefended model decreased to 0.7178, 0.6815, and 0.6637, respectively. With CAGMC-Defence, the OA increased to 0.8841, 0.8813, and 0.8741, representing improvements of 16.63%, 19.98%, and 21.04%, respectively, effectively mitigating performance degradation caused by the perturbations. Similarly, when the SAR modality was attacked, the OA values of the original model were 0.7682, 0.7671, and 0.7658, whereas those of the defended model reached 0.8820, 0.8791, and 0.8649, corresponding to relative gains of 11.37%, 11.20%, and 9.91%, respectively. These results highlight the model’s strong crossmodal robustness. Notably, the proposed method outperformed even the clean-sample baseline in certain unimodal attack scenarios, demonstrating its effective use of redundant modalities and its ability to suppress corrupted inputs. Overall, the results confirm that CAGMC-Defence is not only effective against joint multimodal attacks but also provides stable and significant protection under challenging unimodal attack conditions, greatly reducing the impact of adversarial perturbations on model performance.
Furthermore, we evaluated the practicality of CAGMC-Defence by comparing the inference efficiency of various defence methods under different adversarial attack scenarios, as summarized in
Table 7. All methods used their respective pretrained defence models, and during the testing phase, adversarial examples were regenerated for each model under FGSM, PGD, and MIM attacks. The total processing time, including both adversarial example generation and model inference, was recorded to fairly and comprehensively simulate the defence cost in real-world deployment. The values in the table represent the average processing time per adversarial example (in seconds), with lower values indicating higher efficiency.
The results demonstrate that CAGMC-Defence consistently achieves high inference efficiency across all attack types. Under PGD and MIM attacks, the average processing times are 2.87 s and 2.67 s, respectively, which are significantly lower than those of most multistep adversarial training methods, such as the LBGAT (10.64 s/4.03 s) and GAT (12.28 s/4.77 s), highlighting its computational advantage. In comparison, CAGMC-Defence achieves a favourable trade-off between robustness and computational cost, offering both strong defence performance and practical deployment efficiency.
4.4. Iterations
To evaluate the convergence speed and defence performance of the proposed multimodal adversarial defence method, we tested the CAGMC-Defence method both on clean examples and under various adversarial attack scenarios. As shown in
Figure 8, under clean conditions, the model reached an overall accuracy above 0.87 within just 10 training epochs. It then quickly converged and maintained stable accuracy throughout the remaining training process, demonstrating strong classification performance. This rapid convergence significantly reduced the computational overhead, thereby enhancing the model’s real-time capability and robustness.
In addition, under FGSM attacks, CAGMC-Defence consistently maintained an OA of approximately 0.87, with kappa values ranging narrowly between 0.65 and 0.68, indicating the high robustness of the defence model against this type of attack. In contrast, under PGD attacks, the model experienced a notable performance decrease at epoch 20, with the OA decreasing to 0.7465 and the kappa coefficient decreasing to 0.4011, suggesting that the attack effectively compromised the model’s defence at this stage. However, as training progressed to epoch 40, the OA recovered to 0.8682, and the kappa value increased to 0.6518, demonstrating that the model gradually strengthens its resistance to stronger attacks through adversarial training. In the case of MIM attacks, the OA remained below 0.75 between epochs 10 and 30, indicating that the momentum-based gradient accumulation mechanism can more easily overcome static defence strategies. Nevertheless, by epoch 40, the OA increased to 0.8574, reflecting the robust defence capability of CAGMC-Defence even under high-order adversarial attacks.
4.5. Ablation Studies
To gain a deeper understanding of the mechanisms by which each module enhances model robustness, we first revisit the design rationale of the proposed components prior to conducting the ablation experiments. The MFEF module leverages a multimodal crossattention mechanism and feature fusion strategy to jointly model the representations of optical and SAR modalities. This design not only strengthens the collaborative perception between modalities but also helps mitigate abnormal activations induced by adversarial perturbations. In parallel, the MAT module introduces multimodal adversarial examples during training, guiding the model to adapt to perturbation distributions from an optimization perspective, thereby fundamentally improving adversarial robustness.
Building upon these design principles, we constructed four model variants and performed a series of ablation studies to systematically evaluate the individual contributions of each module to the overall robustness of the model: (1) NO_ACTION, the baseline model without the Multimodal Feature Enhancement and Fusion (MFEF) and Multimodal Adversarial Training (MAT) modules; (2) NO_MFEF, which retains MAT but removes MFEF; (3) NO_MAT, which retains MFEF but excludes MAT; and (4) ALL, the complete model integrating both MFEF and MAT. By comparing the performance of these variants under different adversarial attack scenarios, we assess the individual and joint effects of each module, thereby validating the robustness improvements introduced by our method. The experimental results are shown in
Table 8,
Table 9 and
Table 10.
According to the experimental results presented in
Table 8,
Table 9 and
Table 10, all three types of whitebox attacks significantly degraded the performance of the baseline model, with increasing levels of damage observed as the perturbation strength
increased from 0 to 0.05. Among them, the MIM attack exhibited the highest destructiveness. For instance, at
, the OA of the baseline model without any defence modules (denoted as NO_ACTION) decreased to only 0.2578, which was substantially lower than the OAs under the FGSM (0.4688) and PGD (0.2825) attacks. Under clean conditions (
), the baseline model achieved an OA of 0.8779 and a kappa coefficient of 0.6964. Introducing only the Multimodal Feature Enhancement and Fusion module (NO_MAT) increased the OA to 0.9018 and the kappa coefficient to 0.7776, demonstrating that the MFEF module effectively improves the model’s discriminative capacity. However, in high-strength attack scenarios (
), the robustness of the NO_MAT configuration rapidly diminished. For example, under the FGSM attack at
, its OA decreased to merely 0.4886, indicating that the MFEF module alone is insufficient to reinforce the model’s vulnerable decision boundaries against strong adversarial perturbations.
In contrast, incorporating only the Multimodal Adversarial Training module (NO_MFEF) significantly enhanced robustness. For instance, under FGSM attacks with a perturbation strength of , the model achieved an OA of 0.8437 and a kappa coefficient of 0.6653. Similarly, under PGD attacks with the same level, the OA and kappa coefficient remained at 0.8218 and 0.5109, respectively. Although the kappa value exhibited a more pronounced decline, both metrics still demonstrated substantial improvements over the baseline. These results indicate that the MAT module effectively reshapes the decision boundary through Multimodal Adversarial Training, thereby improving the model’s resistance to both gradient-based and iterative attacks.
The complete model (ALL), which integrates both the MAT and MFEF modules, demonstrated the most robust and stable defence performance. For example, under FGSM, PGD, and MIM attacks with a perturbation strength of , the OA reaches 0.8788, 0.8682, and 0.8574, respectively—surpassing any single-module variant. The kappa coefficient also remained above 0.6284 across all three attacks. Compared with (NO_MFEF), the ALL model achieved a 0.0464 improvement in OA and a 0.1409 increase in kappa under PGD attacks with , highlighting the complementary synergy between the MFEF and MAT modules. Overall, the MAT module serves as the primary driver of robustness, whereas the MFEF module enhances classification consistency by leveraging the complementary characteristics of optical and SAR modalities to smooth decision boundaries. Their synergistic integration significantly strengthens the multimodal model’s resilience against various adversarial perturbations without compromising its baseline performance.
4.6. Transferability Experiments
To assess the transferability of the proposed CAGMC-Defence method under different adversarial training strategies, we adopted three representative attack methods—FGSM, PGD, and MIM—to independently perform adversarial training on the model. The defence performance was then evaluated using all three attack types under a fixed perturbation strength (
). This setup enables a systematic evaluation of the generalization capability and stability of each training strategy (see
Table 11).
The experimental results demonstrate that different adversarial training strategies could significantly increase model robustness against their respective attack types. For instance, under MIM adversarial training, the model achieved optimal defence performance when evaluated against MIM attacks, attaining an OA of 0.8574, which was notably higher than its performance under FGSM (0.8072) and PGD (0.7473) attacks. This observation suggests that the model achieves the highest robustness when the adversarial sample generation method used during training aligns with the type of attack encountered during testing. Further analysis reveals that PGD adversarial training exhibited superior transferability across different attack scenarios, achieving OAs of 0.8669 and 0.8587 under FGSM and MIM attacks, respectively, with the smallest performance degradation. This highlights the adaptability of PGD-trained models under diverse adversarial perturbations. This advantage may be attributed to the multistep iterative optimization process inherent in PGD adversarial sample generation, which produces more diverse and representative perturbations, thereby enhancing the model’s generalizability. In contrast, although MIM training achieved strong defence within its native attack domain, its generalization capability under heterogeneous attacks appears to be limited. Specifically, its performance decreased considerably when facing PGD attacks, with the OA decreasing to 0.7473 and the kappa coefficient decreasing to only 0.1499.
In summary, CAGMC-Defence substantially enhances model robustness across different adversarial training paradigms. Notably, PGD-based adversarial training achieves a well-balanced trade-off between defence against PGD attacks and generalization to other attack types (e.g., FGSM and MIM), highlighting its effectiveness in complex adversarial environments.