Next Article in Journal
A Study on the Identification of Geohazards in Henan Province Based on the Basic Deformation Products of LuTan-1
Previous Article in Journal
Predicting Wildfire Risk in Southwestern Saudi Arabia Using Machine Learning and Geospatial Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Few-Shot Class-Incremental SAR Target Recognition with a Forward-Compatible Prototype Classifier

1
College of Combat Support, Rocket Force Engineering University, Xi’an 710025, China
2
College of Information Science and Technology, Beijing University of Chemical Technology, Beijing 100029, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(21), 3518; https://doi.org/10.3390/rs17213518 (registering DOI)
Submission received: 23 August 2025 / Revised: 9 October 2025 / Accepted: 20 October 2025 / Published: 23 October 2025

Highlights

What are the main findings?
  • We propose a Forward-Compatible Prototype Classifier (FCPC) by emphasizing the model’s forward compatibility to continually learn new concepts from limited samples without forgetting the previously learned ones.
  • A Nearest-Class-Mean (NCM) classifier is proposed for prediction by comparing the semantics of unknown targets with prototypes of all classes based on the cosine criterion.
What are the implications of the main findings?
  • The proposed method can continually learn new concepts from limited samples without forgetting the previously learned ones, which can improve the SAR ATR capability.
  • The proposed method powers the DL-based SAR ATR systems with Few-Shot Class-Incremental Learning (FSCIL) ability to satisfy real-world SAR ATR scenarios.

Abstract

In practical Synthetic Aperture Radar (SAR) applications, new-class objects can appear at any time as the rapid accumulation of large-scale and high-quantity SAR imagery and are usually supported by limited instances in most cooperative scenarios. Hence, powering advanced deep-learning (DL)-based SAR Automatic Target Recognition (SAR ATR) systems with the ability to continuously learn new concepts from few-shot samples without forgetting the old ones is important. In this paper, we tackle the Few-Shot Class-Incremental Learning (FSCIL) problem in the SAR ATR field and propose a Forward-Compatible Prototype Classifier (FCPC) by emphasizing the model’s forward compatibility to incoming targets before and after deployment. Specifically, the classifier’s sensitivity to diversified cues of emerging targets is improved in advance by a Virtual-class Semantic Synthesizer (VSS), considering the class-agnostic scattering parts of targets in SAR imagery and semantic patterns of the DL paradigm. After deploying the classifier in dynamic worlds, since novel target patterns from few-shot samples are highly biased and unstable, the model’s representability to general patterns and its adaptability to class-discriminative ones are balanced by a Decoupled Margin Adaptation (DMA) strategy, in which only the model’s high-level semantic parameters are timely tuned by improving the similarity of few-shot boundary samples to class prototypes and the dissimilarity to interclass ones. For inference, a Nearest-Class-Mean (NCM) classifier is adopted for prediction by comparing the semantics of unknown targets with prototypes of all classes based on the cosine criterion. In experiments, contributions of the proposed modules are verified by ablation studies, and our method achieves considerable performance on three FSCIL of SAR ATR datasets, i.e., SAR-AIRcraft-FSCIL, MSTAR-FSCIL, and FUSAR-FSCIL, compared with numerous benchmarks, demonstrating its superiority and effectiveness in dealing with the FSCIL of SAR ATR.

1. Introduction

Synthetic Aperture Radar (SAR) is a landmark achievement in the field of Earth remote sensing science. Thanks to its capability for observing targets in all day and weather conditions, SAR images have been important data sources for various applications in both civil and military fields, like hazard warning [1], change detection [2], and target surveillance [3,4,5]. As a fundamental task of SAR, SAR Automatic Target Recognition (SAR ATR), which aims to identify potential targets in SAR images, can provide diverse and complementary information for decision-making and has been a longstanding and intensive research topic in the past several decades. When entering the big data era, deep-learning (DL) approaches have shown superior performance to their precursors, igniting a new revolution in various SAR ATR tasks [6,7]. After years of innovation, numerous airborne and spaceborne high-resolution SAR imaging systems have been manufactured and are in service, making the collection of targets of interest from SAR images more convenient and timely, which provides more opportunities for researchers to explore solutions to realistic problems. In particular, given the ever-changing worlds humans and SAR ATR systems face, more efforts should be devoted to pushing them forward to satisfy practical demands.
Traditionally, most current SAR ATR systems are usually designed for ideal recognition tasks, where training data is collected in advance from a closed scene and assumed to follow the nearly i.i.d distributions with testing samples. However, real-world SAR ATR scenarios are non-stationary, where new classes can appear at any time with the acquisition of a large quantity of large-scale SAR imagery. A static SAR ATR system that is not adaptive nor robust to changes can quickly become outdated or unreliable. Therefore, powering a DL-based SAR ATR system with the ability to continually learn new concepts without forgetting the previous ones is important. To address this deficiency, incremental learning (IL), especially Class-IL (CIL), which aims at learning new classes in a stream format, has been borrowed from the computer vision (CV) areas and actively explored in the SAR ATR field in recent works. Certainly, these CIL-powered SAR ATR solutions have shown supervising performance in learning new concepts at non-stationary situations compared with traditional ones. However, standard CIL still assumes that there is sufficiently labeled data for new classes, which inevitably impedes its practical applications since it is expensive and even impossible to collect a large amount of annotated data in most SAR ATR scenarios, like boundary management and battlefield surveillance. Henceforth, powering SAR ATR systems with the ability to continuously learn new classes from limited samples is more practical and worth investigating.
In response, Few-Shot Class-Incremental Learning (FSCIL) has been explored and has attracted much attention from the CV communities. The goal of the FSCIL is to power the classifiers with the ability to continually learn new classes from few-shot samples without losing much performance in judging the old ones. As a specific setting of the CIL, two intrinsic issues, including Catastrophic forgetting [8] and Overfitting, should be addressed carefully. The former refers to the fact that a DL-based classifier learning on the new targets can easily encounter an irreversible performance drop when judging old ones, resulting from the inaccessibility to previous data due to data privacy or storage limitations. The latter, however, suggests that a classifier trained by new-class limited samples tends to capture instance-specific but not general patterns of new classes, due to the data scarcity of novel classes in most non-cooperative scenarios, leading to a substantial performance gap in identifying training and testing samples.
Recently, numerous solutions to the problem have been proposed, and significant performance has been achieved in CV fields. Among them, methods following the forward-compatible paradigm, which aims at improving the model’s adaptability to unknown categories in advance, have exhibited significant performance. Typically, FACT [9] pre-allocated and learned numerous virtual prototypes, which were synthesized from real-class semantics features, to sequence the embedding of known classes and reserve them for the new ones. ALICE [10] solved the problem from the open-set perspective by introducing auxiliary classes via mixing pairs of real-class samples. In terms of the FSCIL in the SAR ATR field, DSSC [11] tentatively solved the problem from the forward-compatible aspect by designing two self-supervised learning (SSL) tasks, which include the Scattering Mixup Module (SMM) and the Rotation-aware Transformation Module (RTM), for virtual-class generation while decoupling the model’s parameters with dynamic worlds for rapid knowledge transferring. Despite the tremendous success, all virtual classes in the DSSC were synthesized based on real-class pixel-level scattering information and can be easily interfered with by background clutter. Also, the decoupling operation implicitly weakens the classifier’s continual adaptability to new concepts, leading to a gradual feature misalignment for representing both old and current targets. Furthermore, due to the unique imaging mechanisms of SAR and the complexity and diversity of target structures and postures, their signatures can exhibit significant intra-class variability and inter-class similarity, leading to severe interference for accurate target recognition. Henceforth, domain knowledge of targets in SAR imagery should be carefully considered for stable inference.
In this paper, a novel Forward-Compatible Prototype Classifier (FCPC) is proposed to improve the model’s forward-compatibility to new targets before and after deployment. In the FCPC, compared with scattering mixup, a Virtual-class Semantic Synthesizer (VSS) is designed for virtual-class generation by using base-class semantics. As a result, the model’s representation ability for diverse features of unknown targets can be further improved in advance. After deployment, a Decoupled Margin Adaptation (DMA) strategy is designed to merely update the model’s high-level semantic parameters by improving the consistency of new-class marginal features with respect to class prototypes while enlarging the difference to inter-class ones. For inference, a Nearest-Class-Mean (NCM) classifier is employed for prediction by comparing the cosine distances between features of testing samples and all-class prototypes. Extensive experiments on three FSCIL of SAR ATR datasets, i.e., MSTAR-FSCIL, SAR-AIRcraft-FSCIL, and FUSAR-FSCIL, demonstrate the superiority of our method compared with numerous task-specific benchmarks. Overall, our contributions are three-fold.
  • We explored the FSCIL problem in the SAR ATR field and designed an FCPC framework to further improve the model’s forward compatibility with unpredictable targets in the dynamic world before and after deployment.
  • We designed a VSS for virtual-class synthesis based on real-class semantics, a DMA for making our method continually evolve, and an NCM classifier for general prediction.
  • We showed the state-of-the-art performance of the FCPC in comparison to several advanced benchmarks on three derived FSCIL of SAR ATR datasets.
The rest of the paper is organized as follows. The related works are presented in Section 2. The problem set-up, the motivations, and our method are given in Section 3. The experimental settings, results, and analysis are detailed in Section 4. Section 5 discusses the potential limitations of the proposed model and its future research directions. Section 6 concludes the work.

2. Related Works

2.1. SAR Target Recognition

The methods of SAR target recognition can be broadly categorized into template-based, model-based, and machine-learning-based paradigms [6]. In template-based solutions [12,13,14], unknown targets are classified by matching them to pre-defined templates based on low-level, hand-crafted features such as gray scale, length, edge, and region moment; it is simple and intuitive yet easily influenced by the environment. In model-based approaches [15,16,17], parameterized models with rigorously physical priors and hypotheses are formulated to estimate target electromagnetic scattering structures for similarity comparison. The models have a strong explainability yet are very complicated, caused by complex SAR imagery mechanisms. Recently, machine-learning-based, especially DL-based, solutions [6,18,19,20,21,22,23] have dominated the field, benefiting from their powerful feature-representing and -discriminating abilities. However, most are data-hungry and merely designed for an ideally closed-world recognition scenario, which is incompatible with reality.
Our method still follows the DL-based paradigm and aims to continuously learn new knowledge from scarce samples without seriously forgetting the old ones.

2.2. Few-Shot Class-Incremental Learning

Few-Shot Class-Incremental Learning (FSCIL) [24] as an extension of the CIL aims at teaching a classifier to continually learn new concepts from few-shot samples without forgetting the old ones and has been an active exploration area in recent years. Typically, TOPIC [24] first defined the problem and preserved feature manifolds of old classes by a neural gas network. ERDIL [25] selected old-class exemplars using an exemplar relation graph (ERG) and distilled relational features for knowledge preservation. IDLVQC [26] represented class knowledge as quantized reference vectors and continually adjusted locations of the old vectors for misalignment reduction. Although their effectiveness has been verified, most of them focus on mitigating the forgetting of the old knowledge, i.e., backward compatibility, but overlook the pre-perceiving ability to the new knowledge, weakening its adaptability to dynamic worlds. In contrast, methods with forward compatibility mainly focus on preparing the base stage training to facilitate the better acceptance of novel classes in the future. Typically, CEC [27] proposed a continually evolving classifier leveraged by a meta-learned graph attention network for new knowledge adaptation. FACT [9] pre-allocated numerous orthogonal prototypes as virtual classes to squeeze the embedding of known targets and reserve for the new. ALICE  [10] solved the problem from an open-set perspective by generating and pre-assigning several virtual classes generated by mixing samples of different classes.
In this paper, our method explores the FSCIL in the SAR ATR field following the forward-compatible scheme for better knowledge incorporation and identification.

2.3. FSCIL of SAR ATR

Recently, some works have been carried out on the FSCIL in the SAR ATR field. Among them, Wang et al. proposed a HEIEN network [28] with an adaptive class-incremental learning (ACIL) module. Due to the tight connection between the cosine criterion and target azimuth-aware structures, the CPL framework [29] designed a couple of losses and a Nearest-Class-Mean (NCM) classifier to learn and identify targets at cosine space. Furthermore, the AASC framework [30] proposed losses at both semantic and manifold facets and designed a subspace classifier on the Grassmannian manifold for prediction. ODF [31] proposed an orthogonal distribution optimization method for the FSCIL of SAR ATR. DILHyFS  [32] proposed a dual-branch architecture that focuses on local feature extraction and leverages the discrete Fourier transform and global filters to capture long-term spatial dependencies. The effectiveness of the method was verified on numerous settings derived from the MSTAR dataset. Unlike the above approaches focusing on the model’s backward compatibility, in which the performance on learned targets is highly considered by stabilizing the drift of feature spaces passively, DSSC  [11] and DSAC  [33] addressed the problem from the forward-compatible angle by pre-allocating and learning several virtual classes in advance. Nevertheless, these classes, generated by the scattering mix-up, can easily interferred with by surroundings, making class supervisions inaccurate and unstable. However, in our method, not only are the semantics of base classes adopted for virtual-class generation but the classifier is also decoupled and evolved after deployment for better knowledge retention and transfer.

3. Materials and Methods

3.1. Problem Statement

Few-Shot Class-Incremental Learning (FSCIL) [24] necessitates a method to learn continually from scarcely labeled samples, which usually contains a base and numerous incremental steps, a.k.a. sessions, to mimic a practical recognition scenario. At the base session, classes with sufficient training samples are provided. At incremental sessions, new-class few-shot samples continually appear to serve as novel knowledge collected in dynamic worlds. An ideal FSCIL model should perform well on all seen categories during evaluation once optimized on the current data. Formally, assuming that we have a sequence of labeled data sets D 1 , D 2 , , D t , where D t = { ( x i t , y i t ) } i = 1 N . x i t and y i t refer to a training sample and its category, respectively. C t is the class set of D t . | C t | is the number of categories of C t . Notably, j , k t , j k , C j C k = . The base session dataset D 1 usually contains large-scale training samples for each class. For each incremental session t , t > 1 , D t contains several new classes with few-shot examples and are provided in an N-way K-shot format. At session t, only D t is accessed for optimization, and the model is expected to classify all seen categories C = i = 1 t C i during evaluation.

3.2. Motivations

As the intrinsic obstacles of the FSCIL and characteristics of targets in SAR imagery, the Forward-Compatible and Stable-Discriminating capabilities should be considered for mitigating the FSCIL in the SAR ATR field.
  • Forward-Compatible ability means that a DL-based SAR ATR system can not only represent targets of known categories but can also incorporate new concepts rapidly once optimization has taken place. According to the Attribute Scattering Center (ASC) theory, target information in SAR imagery can be regarded as the specific composition of various basic scattering parts, e.g., dihedral, trihedral, cylinder. Furthermore, target deep semantics can be seen as collections of various convolutional activations, which implicitly represent high-level class-agnostic structures. Henceforth, the SAR ATR system with a strong forward compatibility should possess the ability to capture diverse cues of incoming targets.
  • Stable-Discriminating ability means that an SAR ATR system can accurately identify targets of different classes captured in diverse scenarios and postures. In contrast to rich and stable target cues in optical images, only target-specific structures and parts can be observed in SAR images, providing rare and inconsistent information, caused by the particular SAR imaging mechanisms. Unlike single target instances, prototypes derived from the average of target features at diverse postures, can provide class-related information in a general and stable way for better identification.

3.3. Overall Framework

The overall framework of our method is given by Figure 1 and can be divided into three stages.
  • At the base learning stage ( t = 1 ), a CNN-based feature extractor parameterized by f ( x ; Φ ) is trained on sufficient data D 1 to learn an embedding space for identifying base classes while extracting general patterns of unknown ones. Like the dynamically real-world SAR ATR scenarios and the prominent partiality of targets in SAR imagery, the model’s forward compatibility with unknown targets is promoted by a Virtual-class Semantic Synthesizer (VSS), in which numerous virtual classes with soft labels are synthesized from pre-encoded real-class embeddings h ( x ; ϕ ) . After being post-encoded by g ( · ; φ ) on both real and virtual classes and optimized by a cross-entropy (CE) loss, a well-spanned f ( x ; Φ ) with base-class prototypes can be obtained.
  • At incremental learning stages ( t > 1 ), where new-class data D t , ( t > 1 ) continually appear with few-shot samples, the model’s forward compatibility is dynamically released by a Decoupled Margin Adaptation (DMA) strategy by merely fine-tuning high-level semantic parameters in a timely manner. Therefore, the similarities of few-shot samples of novel classes to class prototypes and the discrepancy with interclass ones can be improved effectively. After optimization, the prototypes of novel classes are obtained for further inference.
  • For the inference of session t, like the general patterns provided by class prototypes, a Nearest-Class-Mean (NCM) classifier is formulated for identification by comparing semantic features of unknown targets with all-class prototypes.

3.4. Forward Compatibility

In FCPC, a Virtual-class Semantic Synthesizer (VSS) and a Decoupled Margin Adaptation (DMA) strategy are designed to improve the model’s forward compatibility with novel targets at the base and incremental stages, respectively.

3.4.1. Virtual-Class Semantic Synthesizer

At the base stage, a Virtual-class Semantic Synthesizer (VSS) is designed to generate numerous virtual classes with soft labels for enhancing the model’s forward compatibility with incoming targets in advance, inspired by the latent connections between physical-aware scattering parts of targets in SAR imagery and the representing learning paradigm of the DL. The former, supported by the Attribute Scattering Center (ASC) theory [34], formulates targets in SAR images as the composition of class-agnostic physical parts and structures, e.g., dihedral, trihedral, cylinder, and so on. The latter serve as the cornerstone of DL-based learners, representing target semantic cues as the collections of convolutional activations. Therefore, the richer semantic patterns captured by our method, the more diverse scattering parts it can perceive. Algorithm 1 gives the process of the VSS concretely.
  • Virtual Class Generation (VCG): As the rich and diverse target parts provided by D 1 , numerous virtual classes are synthesized by mixing up real-class semantic parts within a batch data B . In Algorithm 1, | B | is the sample number of B . The larger the repeated time M, the more diverse the components of virtual classes that can be generated. S saves the virtual-class instances. In Equation (1), we sample λ from a B e t a ( α , α ) to control the overlaps between real and virtual classes. Notably, instead of synthesizing targets in input space, the intermediate spatial manifolds h B with different categories pre-encoded by h 1 ( x ; ϕ ) are adopted, as are their distinctly component-aware characteristics and limited background interference.
  • Soft Label (SL): The more mixing-up cues provided by virtual labels, the more diversified connections the model can learn from the virtual samples. In response, different from assigning virtual targets with newly one-hot labels y ^ i , a Gaussian soft-labeling function varying with the λ given by Equation (2) is designed to generate soft supervisions. In particular, the labels for virtual targets reach the highest values as the λ reaches 0.5.
Algorithm 1 The procedure of VSS.
  • Input A batch data B = { x i , y i } i | B | sampled from D 1
  • Require Repeated time M, target embeddings h B = { h ( x i ; ϕ ) } i | B | of B , α of B e t a ( α , α )
  • for m = 1 to M do
  •       1.
    get h ˜ B = { h ( x ˜ i ; ϕ ) } by shuffling h B randomly.
          2.
    get one-hot labels { y i } i | B | of h B and shuffled one-hot ones { y ˜ i } i | B | of h ˜ B .
          3.
    get virtual-class embeddings h ^ B = { h ( x ^ i ; ϕ ) } from h B and h ˜ B following Equation (1).
          4.
    get corresponding soft labels { y ^ i } i | B | from { y i } and y ˜ following Equation (2)
          5.
    append h ^ B and { y ^ i } i | B | to S .
  • end for
  • Output S
h ( x ^ i ; ϕ ) = λ · h ( x i ; ϕ ) + ( 1 λ ) h ( x ˜ i ; ϕ )
Γ ( λ ) = 1 σ 2 π exp ( λ 2 2 σ 2 ) y ^ i = 1 Γ ( λ ) Γ ( 1 λ ) ( 1 Γ ( 1 ) )
L V S S = 1 | N r | L C E ( g ( h ( x i ; ϕ ) ) , y i ) + 1 | N v | L B C E ( g ( h ( x ^ i ; ϕ ) ) , y ^ i )

3.4.2. Decoupled Margin Adaptation

Given highly confused and unstable target cues provided by few-shot samples, a Decoupled Margin Adaptation (DMA) strategy, which contains a Decoupled Adaptation (DA) strategy and a Prototype Margin Loss (PML), is designed to make the model continually evolve to adapt new concepts while relieving the forgetting problem after deployment.
  • Decoupled Adaptation: As target cues provided by few-shot samples are extremely rare and unstable, directly tuning whole parameters of our method can easily lead to the overfitting problem. In response, a decoupled adaptation (DA) strategy is introduced to divide the method’s whole parameters into low-level and high-level parameters based on the layers. During incremental learning, only the high-level parts containing abstract information are tuned while freezing the low-level ones with general cues to balance the representation of class-agnostic patterns and the rapid adaptation to class-specific patterns.
  • Prototype Margin Loss: Given the rare and unstable target cues provided by few-shot instances, a Prototype Margin Loss (PML) is designed to update the model’s parameters when necessary, avoiding inappropriate adaptation. Unlike the cross-entropy (CE) loss, the PML ensures robust feature learning under limited data. As defined in Equation (4), f t ( x c ; Φ ) and v c represent the deep embeddings of a novel class c , c C t , and the parameterized class-related prototypes, respectively. Additionally, another class prototype v j , corresponding to the top J nearest distances to the f t ( x c ; Φ ) , is selected for comparison. The cosine distance d is adopted for judgment due to its strong connection to azimuth-aware target structures, as demonstrated in [29]. The parameter m controls the discrepancy margin. By minimizing L P M L , the compact intraclass feature spaces and well-separated interclass feature spaces can be learned properly.
L P M L = ( x c , y c ) D t j J max ( d ( f t ( x c ; Φ ) v c ) d ( f t ( x c ; Φ ) v j ) + m , 0 )

3.5. Nearest-Class-Mean Classifier

At inference, a Nearest-Class-Mean (NCM) classifier is maintained for stable discriminability. Specifically, as shown by the bottom subfigure of Figure 1, semantic features of an unknown target x are first extracted by the current feature extractor f t ( x ; Φ ) . Then, the cosine distances d between f t ( x ; Φ ) and all-class prototypes { p i } i = 1 c t are compared. Here, the class prototype p i is calculated by Equation (5), where N i is the total number of few-shot samples with class i. Finally, the prediction y ^ is calculated by Equation (6) and can be regarded as the class with the closest distance between the prototype and the target features.
p i = 1 N i n = 1 N i 1 ( y n = i ) f t ( x n ; Φ )
y ^ = arg max i c t d p i , f t ( x ; Φ )
Notably, instead of using Euclidean distance for evaluation, the cosine distance is employed to measure the normalized similarity between prototypes and test samples. This approach effectively captures target structural characteristics, thereby enhancing discrimination accuracy and reducing background interference.

4. Results

In this part, the experimental datasets, the implementation details, and the evaluation metrics are first depicted. Afterward, ablation studies for the proposed modules and extensive results are provided.

4.1. Dataset Preparation

In the experiments, two FSCILs of the SAR ATR datasets, which contain the MSTAR-FSCIL and the SAR-AIRcraft-FSCIL, are leveraged for validation.

4.1.1. MSTAR-FSCIL

Following the [30], the MSTAR-FSCIL dataset derived from the Standard Operation Condition (SOC) of the MSTAR [12] dataset is leveraged to verify the effectiveness of our method for training fine-grained ground vehicles captured by airborne platform. According to the types of the vehicles, four types of targets, i.e., BTR70, 2S1, BRDM2, and BMP2, with full-azimuth samples imaged at a 17 ° depression angle are adopted to serve as the base session data D 1 . The remaining six with randomly selected few-shot samples are used for incremental learning. Furthermore, the same categories supported by full-azimuth instances imaged at a 15 ° depression angle are used for validation. Table 1 shows the dataset configurations in detail. Figure 2 shows the targets in SAR and optical images.

4.1.2. SAR-AIRcraft-FSCIL

The SAR-Aircraft-FSCIL dataset was constructed from the public dataset named SAR-Aircraft-1.0 [35], which was developed and released by the Chinese Academy of Sciences (CAS) in 2023. This dataset comprises seven distinct classes of aircraft, including the A220, A320, A330, ARJ21, Boeing737, Boeing787, and others. All samples were acquired from three international airports using data captured by the Chinese Gaofen-3 satellite under C-band single polarization in spotlight mode, achieving a spatial resolution of 1 m. Following the framework established in [30], four classes with the largest sample sizes, namely, Other, A220, Boeing787, and Boeing737 were used for base session learning. The remaining three classes (A320, ARJ21, and A330) were designated as incremental classes D t , t > 1 . For training, each base class was represented by 2000 samples, while the incremental classes received only five samples each. During validation, a random selection of 200 samples was drawn across all classes to ensure balanced testing. A detailed overview of the SAR-Aircraft-FSCIL dataset is provided in Table 2, which includes specific configurations such as imaging parameters and class distributions. Additionally, Figure 3 illustrates representative examples of targets in both SAR and optical imagery, offering a comparative perspective on the dataset’s characteristics.

4.1.3. FUSAR-FSCIL

The FUSAR-Ship [36] dataset is an open SAR-AIS matchup dataset of Gaofen-3 satellite released by Fudan University for ship and marine target detection and recognition. The FUSAR-Ship dataset is constructed from over 100 GF-3 scenes covering a large variety of sea, land, coast, river and island scenarios. It includes over 5000 ship image chips covering about 10 types of marine targets. Figure 1 shows the SAR and optical images of the ten selected types of ships in the FUSAR-Ship dataset. Following the process of the MSTAR-FSCIL and SAR-AIRcraft-FSCIL datasets, a FUSAR-FSCIL dataset is constructed from the FUSAR-Ship dataset for task-related ship recognition evaluation. Here, four types of ships, i.e., Cargo, Other, Fishing, and BulkCarrier, with abundant samples are selected for base-session training. The six remaining types of ships, i.e., Tanker, Unspecificied Container, Dredger, Tug, and GeneralCargo, are used for incremental learning. Table 3 shows the configurations of the constructed FUSAR-FSCIL dataset. Additionally, Figure 4 presents both SAR and optical imagery of selected ships of the FUSAR-FSCIL dataset.

4.2. Implementation Details

In all experiments, we randomly selected five-shot samples for each incremental class and conducted T = 10 trials for statistical evaluation.Following the established protocol in [24], a ResNet-18 network pre-trained on ImageNet serves as the feature extractor f ( x ; Φ ) for all compared methods for balancing the model’s performance and computational efficiency. Its architecture comprises 18 convolutional layers enhanced with skip connections to mitigate the vanishing gradient problem. Regarding hyperparameters, for the VSS method, the number of repetitions M is set to 4, and the value of α in the B e t a ( α , α ) is 0.5. For the DMA, the margin m for the PML is set as 0.5, and features from the first residual block of the ResNet-18 are utilized for the synthesis process. For the base session training, all benchmarks are trained using the Stochastic Gradient Descent (SGD) optimizer for 50 epochs. The initial learning rate is set to 1 × 10 2 and is decayed to 1 × 10 3 at epoch 30 and further to 1 × 10 4 at epoch 40. The mini-batch size is set to 32. For incremental learning, the learning rate is initialized to 1 × 10 4 and decayed by a factor of 0.1 at epoch 30. All hyperparameters for the compared benchmarks are configured as described in their original papers. The input images are normalized to the range [0, 1] and resized to 64 × 64 pixels. For data augmentation, all inputs are randomly rotated within ± 5 . All experiments are conducted on a NVIDIA GTX 3080Ti GPU (NVIDIA Corporation, Santa Clara, CA, USA) with CUDA 11.1.

4.3. Evaluation Metrics and Benchmarks

In experiments, three metrics are used for evaluation. First, the classification accuracy (Acc.) is reported to evaluate the benchmark’s performance on categories seen at each session. Second, the average accuracy (Avg. Acc), calculated by averaging the Acc. of all sessions, reflects the model’s comprehensive performance on all categories seen. Third, the performance drop (PD) rate [27] is provided to measure the absolute performance deterioration by subtracting the accuracy of the last session from that of the first.
Following [30], numerous task-specific benchmarks are adopted for evaluation, as described below. Traditional DL-based methods: Ft-CNN and Oracle are employed. The former fine-tunes a CNN-based classifier on the data of the current session using a cross-entropy loss, while the latter is an offline learner trained on all data from seen classes. IL solutions: Three typical incremental learning (IL) classifiers—iCaRL [37], EEIL [38], and LUCIR [39]—are utilized. These methods follow a replay-based paradigm, where all samples of new classes are stored for learning in subsequent sessions to ensure a fair comparison. FSCIL solutions: Methods including TOPIC [24], ERDIL [25], IDLVQC [26], CEC [27], FACT [9], ALICE [10], SAVC [40], CPL [29] and AASC [30] are adopted for comprehensive evaluation. The key features of these algorithms are discussed in the Related Works (FSCIL) section (Section 2.2). Notably, the last two methods (CPL, AASC) are specifically designed for the FSCIL in SAR ATR.

4.4. Ablation Study

The model’s forward compatibility supported by the two modules, named the VCG and the SL, within the VSS and stable discriminability supported by the DA, the PML, and the NCM classifier are explored in this section. The overall numerical performance is summarized in Table 4. For a fair comparison, a baseline model trained solely on the base session data with the NCM classifier is utilized, with its results presented in the first column of the table.

4.4.1. Effects of VSS

The VSS is employed to enhance the model’s ability to perceive unknown categories in advance. The quantitative and qualitative results of this enhancement are presented in the first four rows of Table 4, as well as in Figure 5 and Figure 6.
  • Quantitative performance: The contributions of the two submodules, namely the VCG and the SL, are progressively verified. Compared to the baseline results presented in the first row of the table, our method with the VCG achieves 79.28%, 71.26%, and 35.51% in Avg Acc, Avg. HA, and PD, respectively, as shown in the second row of the table. Additionally, the corresponding results for the method with the SL are 77.64%, 65.11%, and 36.40%. Finally, as indicated in the fourth row of the table, the scores of our method with the VSS, i.e., with both the VCG and SL, reach 82.49%, 76.59%, and 32.83%. These results are 7.23%, 15.22%, and 4.85% superior to the baseline, demonstrating the effectiveness of the VSS and its submodules.
  • Qualitative performance: The feature diversity and significance are given by Figure 5. The former is quantified by its non-sparsity, measured as the ratio of non-zero feature channels to the total number of channels. The latter is represented by the L2-norm, which corresponds to the overall magnitude of the target features. Larger values correspond to features with more diversity and importance. Here, our method, facilitated by the VCG and the SL, can achieve progressively competitive scores in non-sparsity (Figure 5a) and normalization (Figure 5b). Notably, the much more diverse features can be captured by our method by using both the VCG and the SL. More intuitively, the t-SNE embeddings of class features extracted by our method with the VSS are shown in Figure 6. Here, compared to the baseline, features extracted by our method with the VSS can be distributed more uniformly, implicitly reflecting the diverse and rich target cues acquired by our method.

4.4.2. Effects of DMA

Although the model’s forward compatibility with unpredictable incoming classes is prompted by the VSS, directly applying the model trained on the base session data can inevitably result in a performance drop due to the semantic drift between the current feature space and the few-shot new classes encountered in the dynamic world. For complementarity, the designed DMA, composed of the DA and PML, is leveraged for the model’s rapid adaptation during incremental learning stage.
  • DA: The DA is employed to balance the model’s stability for class-agnostic knowledge and its plasticity for class-specific knowledge by leveraging the hierarchical features from different convolutional layers. As shown in the fifth row of Table 4, our method with the DA achieves more competitive results compared to the version without this module, with Avg Acc, Avg. HA, and PD values of 82.75%, 77.64%, and 32.50%, respectively. Additionally, the effects of varying the locations of the DA are explored in Figure 7a; the method with fewer and deeper trainable parameters achieves higher performance in terms of Avg Acc and Avg HA compared to configurations with shallower and more trainable layers. Furthermore, as demonstrated in Figure 7b, the model with the trainable layer4 achieves the best performance in classifying old classes at each session. This is attributed to its ability to balance static extraction of low-level general features and dynamic adaptation for class-aware semantic features, resulting in lower drifts in old-class features compared to other variants.
  • PML: Our method’s proper adaptability to incoming targets is guaranteed by the PML, for which the contributions are shown in the last row of Table 4. Notably, the Avg Acc, the Avg. HA, and the PD of our method with the PML reach 83.03%, 78.13%, and 32.15%, 0.35%, which is superior to the method without the PML, demonstrating the effectiveness of the module. Furthermore, more investigations into session-wise class separations (R) [40] and interclass distances are presented in Figure 8a. Here, scores of the R among all classes learned by our method with the PML are consistently higher than those of methods without the module or with the CE loss across all sessions. Furthermore, as shown in Figure 8b, benefiting from the clear margin constraint between target samples and selected class weights considered by the PML, interclass distances learned by the loss are significantly superior to those learned by the methods without or with the CE. Henceforth, clear separations can be learned by the PML.

4.5. Benchmark Performance

In this section, comprehensive experiments on three derived benchmarks, namely MSTAR-FSCIL, SAR-AIRcraft-FSCIL, and FUSAR-FSCIL, are conducted to evaluate the effectiveness of our method.

4.5.1. Quantitative Evaluation

The quantitative performance of the compared benchmarks, evaluated on the MSTAR-FSCIL, SAR-AIRcraft-FSCIL, and FUSAR-FSCIL datasets, is presented in Table 5, Table 6, and Table 7 respectively. Several conclusions can be drawn from the results.
  • Firstly, owing to the specially designed techniques aiming to enhance forward compatibility and stable discriminability, our method achieves competitive performance in terms of Avg Acc compared to other benchmarks. On the MSTAR-FSCIL dataset, the Avg Acc of our method reaches 83.03%, surpassing Oracle, SAVC, and CPL by 4.35%, 4.83%, and 4.47%, respectively. Similarly, on the SAR-AIRcraft-FSCIL dataset, our method outperforms Oracle, SAVC, and CPL by 6.64%, 2.49%, and 0.86%, respectively. On the FUSAR-FSCIL dataset, the score of the Avg Acc of our method reaches 60.40%, which is 2.12% higher than the second-ranked solution, demonstrating the effectiveness of our method for coping with the FSCIL problem in SAR ATR field.
  • Secondly, our method effectively addresses the catastrophic forgetting issue, resulting in superior performance on the PD metric. Specifically, our method achieves PD scores of 32.15% and 27.57% on the MSTAR-FSCIL and SAR-AIRcraft-FSCIL datasets, respectively. Unlike most IL and FSCIL methods, which commonly rely on less forgetting losses and replay strategies to mitigate forgetting, our approach with the DMA strategy reaches a balance between static feature representation and dynamic class adaptation, overcoming the limitations of sparse and confused information within limited samples.
  • Thirdly, our method achieves the competitive generalization ability in solving the FSCIL in comparison to public benchmarks. For further validation, a combined dataset with ten sessions (one base + nine incremental) is constructed by combining the FSCIL-MSTAR with the SAR-AIRcraft-FSCIL datasets. Methods are first trained on base classes of the two datasets and then optimized on novel classes of the FSCIL-MSTAR and the SAR-AIRcraft-FSCIL incrementally. The results are given in Table 8. Our method achieves competitive performance on the combined dataset compared to other public benchmarks. Furthermore, our method achieves the average accuracy (Avg Acc) of 77.40%, which is 1.7% higher than the second-ranked solution, verifying its strong generalization ability for continual learning across diverse categories.

4.5.2. Qualitative Evaluation

The qualitative performance of compared benchmarks and the corresponding analysis are given in this section.
  • Session performance curves: The accuracy (Acc) line charts for all compared methods evaluated on of the three datasets are illustrated in Figure 9a–c. Overall, our method with the designed modules consistently achieves the most competitive performance in terms of the Acc across all sessions. Furthermore, the richer and clearer the target-discriminating cues from abundant base classes, the stronger the forward compatibility that could obtained with our method. For instance, the MSTAR-FSCIL dataset, which contains more diverse target components provided by its full-azimuth targets, enables our method’s forward compatibility. Consequently, more competitive performance can be obtained by our method on the MSTAR-FSCIL dataset than that on the SAR-AIRcraft dataset.
  • Confusion matrix: The normalized confusion matrices on the final session of the two datasets are shown in Figure 10, Figure 11 and Figure 12, respectively. Overall, benefiting from the emphasis on both the model’s forward compatibility with incoming classes and the stable discriminability based on limited samples, the colors of diagonal blocks of matrices for both base and incremental classes predicted by our method are more harmonic and bright than those by other benchmarks. For example, most approaches perform well on base classes but fail to judge the new ones, inducing biased confusion matrices. In addition, the more clear and discerning cues provided by limited samples, the more competitive and balanced results the compared benchmarks can reach. For example, the compared benchmarks widely perform better on the MSTAR-FSCIL dataset than on the others, thanks to more distinct cues of targets under ideal conditions than that on the SAR-AIRcraft-FSCIL.
  • Session-wise t-SNE results: Considering the limited testing samples of the FUSAR-FSCIL dataset, the t-SNE are solely conducted on the MSTAR-FSCIL and SAR-AIRcraft-FSCIL datasets, and the results are also shown in Figure 13 and Figure 14, respectively. First, our method, with special consideration on its forward compatibility, can reserve more space for the new, leading to more separated distributions of interclass high-dimensional features than those produced by the baseline. Second, the clearer the target-discerning cues provided by new-class samples, the better distinguishing ability our method can possess. Also, as shown by Figure 13 and Figure 14, the t-SNE results for new classes in the MSTAR-FSCIL are more separated than those in the SAR-AIRcraft-FSCIL. Specifically, interclass features are more separated, while the intraclass ones are more compacted from our method, verifying the importance of unleashing the model’s forward compatibility before deployment.

5. Discussion

Analysis of Experimental Results. The comprehensive experimental results clearly demonstrate that both forward compatibility and stable discriminability are crucial for solving the FSCIL in the SAR ATR field. Due to the specific imaging mechanisms of the SAR and the limited novel samples, new-class target cues in openly dynamic environments are certainly rare and instable. Unlike existing algorithms focusing on identifying current classes or solely being optimized on new instances, our method with forward compatibility supported by the VCG and SL can proactively learn class-agnostic generalized information from sufficient base classes. Meanwhile, the model’s stable discriminability, supported by the DA, PML, and NCM, can learn and classify novel targets rapidly by leveraging few-shot samples. As a result, the model’s stability and plasticity can be balanced properly.
Potential Limitations. The model’s representability and discriminability to diverse incoming targets heavily relly on the high-quality features of both base and novel classes. First, as unknown targets are obtained by the mixing of base-class features and labels, significant feature discrepancies may exist between these two types of categories once the base-class features are highly incomplete or biased. Second, although target general features can be provided by class prototypes of the NCM classifier, which are derived from the average of the class-aware features, the model’s stable discriminability may still deteriorate due to the loss of fine-grained details and the diversity of target azimuth-aware features. Thus, effectively integrating both base-class and new-class features remains crucial for enhancing the model’s perception and discrimination of unknown targets.
Future Research. Since current methods primarily rely on semantic information, future work should prioritize scattering characteristics for target representation and discrimination, given the unique complexities of SAR imaging. Moreover, existing approaches largely depend on task-specific learning from new categories, inherently restricting their adaptability in open-world scenarios. Instead, a task-agnostic learning paradigm, e.g., meta-learning, would be far more suitable for enabling rapid and flexible discrimination.

6. Conclusions

We proposed a Forward Compatible Prototype Classifier (FCPC) to power the DL-based SAR ATR systems with Few-Shot Class-Incremental Learning (FSCIL) ability to satisfy real-world SAR ATR scenarios. The FCSC’s forward compatibility and stable discriminability were emphasized and promoted. For forward compatibility, a VSS was designed to synthesize virtual features with soft labels to unleash the ability before deployment leveraged by the intrinsic links between target partiality and DL’s representing learning paradigm. For stable discriminability, our method was decoupled and evolved from knowledge-oriented new-class fingerprints using an DMA strategy to balance the representation of class-agnostic patterns and the prompt adaptation to class-specific ones. An NCM classifier is maintained for identification without losing generalization. In experiments, the contributions of the designed modules were verified. Extensive experiments on two task-related datasets, i.e., MSTAR-FSCIL and SAR-AIRcraft-FSCIL, showed the effectiveness of our method for the FSCIL in openly dynamic SAR ATR scenarios compared with numerous latest benchmarks.

Author Contributions

Conceptualization, D.G., Y.X. and X.Z.; methodology, D.G. and R.F.; formal analysis, D.G. and R.F.; investigation, B.L. and D.X.; data curation, D.G.; writing—original draft preparation, D.G. and R.F.; writing—review and editing, D.G. and X.Z.; funding acquisition, D.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant 12171481.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Acknowledgments

The authors would like to thank the editors and anonymous reviewers for their valuable comments, which can greatly improve our manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yadav, R.; Nascetti, A.; Azizpour, H.; Ban, Y. Unsupervised flood detection on SAR time series using variational autoencoder. Int. J. Appl. Earth Obs. Geoinf. 2024, 126, 103635. [Google Scholar] [CrossRef]
  2. Hou, X.; Bai, Y.; Xie, Y.; Ge, H.; Li, Y.; Shang, C.; Shen, Q. Deep collaborative learning with class-rebalancing for semi-supervised change detection in SAR images. Knowledge-Based Systems 2023, 264, 110281. [Google Scholar] [CrossRef]
  3. Liu, L.; Fu, L.; Zhang, Y.; Ni, W.; Wu, B.; Li, Y.; Shang, C.; Shen, Q. CLFR-Det: Cross-level feature refinement detector for tiny-ship detection in SAR images. Knowl.-Based Syst. 2024, 284, 111284. [Google Scholar] [CrossRef]
  4. Shang, R.; He, J.; Wang, J.; Xu, K.; Jiao, L.; Stolkin, R. Dense connection and depthwise separable convolution based CNN for polarimetric SAR image classification. Knowl.-Based Syst. 2020, 194, 105542. [Google Scholar] [CrossRef]
  5. Zhao, Y.; Zhao, L.; Liu, Z.; Hu, D.; Kuang, G.; Liu, L. Attentional Feature Refinement and Alignment Network for Aircraft Detection in SAR Imagery. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5220616. [Google Scholar] [CrossRef]
  6. Chen, S.; Wang, H.; Xu, F.; Jin, Y.Q. Target classification using the deep convolutional networks for SAR images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4806–4817. [Google Scholar] [CrossRef]
  7. Zhao, Y.; Zhao, L.; Xiong, B.; Kuang, G. Attention receptive pyramid network for ship detection in SAR images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 2738–2756. [Google Scholar] [CrossRef]
  8. McCloskey, M.; Cohen, N.J. Catastrophic interference in connectionist networks: The sequential learning problem. In Psychology of Learning and Motivation; Elsevier: Amsterdam, The Netherlands, 1989; Volume 24, pp. 109–165. [Google Scholar]
  9. Zhou, D.W.; Wang, F.Y.; Ye, H.J.; Ma, L.; Pu, S.; Zhan, D.C. Forward compatible few-shot class-incremental learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 9046–9056. [Google Scholar]
  10. Peng, C.; Zhao, K.; Wang, T.; Li, M.; Lovell, B.C. Few-shot class-incremental learning from an open-set perspective. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 382–397. [Google Scholar]
  11. Zhao, Y.; Zhao, L.; Zhang, S.; Liu, L.; Ji, K.; Kuang, G. Decoupled Self-Supervised Subspace Classifier for Few-Shot Class-Incremental SAR Target Recognition. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 15845–15861. [Google Scholar] [CrossRef]
  12. Novak, L.M.; Owirka, G.J.; Brower, W.S.; Weaver, A.L. The automatic target-recognition system in SAIP. Linc. Lab. J. 1997, 10, 187–202. [Google Scholar]
  13. Novak, L.M.; Owirka, G.J.; Brower, W.S. Performance of 10-and 20-target MSE classifiers. IEEE Trans. Aerosp. Electron. Syst. 2000, 36, 1279–1289. [Google Scholar]
  14. Novak, L.M.; Halversen, S.D.; Owirka, G.; Hiett, M. Effects of polarization and resolution on SAR ATR. IEEE Trans. Aerosp. Electron. Syst. 1997, 33, 102–116. [Google Scholar] [CrossRef]
  15. Ikeuchi, K.; Wheeler, M.D.; Yamazaki, T.; Shakunaga, T. Model-based SAR ATR system. In Proceedings of the Algorithms for Synthetic Aperture Radar Imagery III, Orlando, FL, USA, 8–12 April 1996; International Society for Optics and Photonics: Bellingham, WA, USA, 1996; Volume 2757, pp. 376–387. [Google Scholar]
  16. Hummel, R. Model-based ATR using synthetic aperture radar. In Proceedings of the IEEE International Radar Conference, Alexandria, VA, USA, 12 May 2000; pp. 856–861. [Google Scholar]
  17. Diemunsch, J.R.; Wissinger, J. Moving and stationary target acquisition and recognition (MSTAR) model-based automatic target recognition: Search technology for a robust ATR. In Proceedings of the Algorithms for Synthetic Aperture Radar Imagery V, Orlando, FL, USA, 13–17 April 1998; International Society for Optics and Photonics: Bellingham, WA, USA, 1998; Volume 3370, pp. 481–492. [Google Scholar]
  18. Li, J.; Yu, Z.; Yu, L.; Cheng, P.; Chen, J.; Chi, C. A comprehensive survey on SAR ATR in deep-learning era. Remote Sens. 2023, 15, 1454. [Google Scholar] [CrossRef]
  19. Li, W.; Yang, W.; Liu, T.; Hou, Y.; Li, Y.; Liu, Z.; Liu, Y.; Liu, L. Predicting gradient is better: Exploring self-supervised learning for SAR ATR with a joint-embedding predictive architecture. ISPRS J. Photogramm. Remote Sens. 2024, 218, 326–338. [Google Scholar] [CrossRef]
  20. Li, W.; Yang, W.; Liu, L.; Zhang, W.; Liu, Y. Discovering and explaining the noncausality of deep learning in SAR ATR. IEEE Geosci. Remote Sens. Lett. 2023, 20, 4004605. [Google Scholar] [CrossRef]
  21. Yu, X.; Yu, H.; Liu, Y.; Ren, H. Enhanced prototypical network with customized region-aware convolution for few-shot SAR ATR. Remote Sens. 2024, 16, 3563. [Google Scholar] [CrossRef]
  22. Peng, B.; Peng, B.; Xia, J.; Liu, T.; Liu, Y.; Liu, L. Towards assessing the synthetic-to-measured adversarial vulnerability of SAR ATR. ISPRS J. Photogramm. Remote Sens. 2024, 214, 119–134. [Google Scholar] [CrossRef]
  23. Wang, C.; Xu, R.; Huang, Y.; Pei, J.; Huang, C.; Zhu, W.; Yang, J. Limited-data SAR ATR causal method via dual-invariance intervention. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5203319. [Google Scholar] [CrossRef]
  24. Tao, X.; Hong, X.; Chang, X.; Dong, S.; Wei, X.; Gong, Y. Few-shot class-incremental learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12183–12192. [Google Scholar]
  25. Dong, S.; Hong, X.; Tao, X.; Chang, X.; Wei, X.; Gong, Y. Few-shot class-incremental learning via relation knowledge distillation. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 2–9 February 2021; Volume 35, pp. 1255–1263. [Google Scholar]
  26. Chen, K.; Lee, C.G. Incremental few-shot learning via vector quantization in deep embedded space. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 4 May 2021. [Google Scholar]
  27. Zhang, C.; Song, N.; Lin, G.; Zheng, Y.; Pan, P.; Xu, Y. Few-shot incremental learning with continually evolved classifiers. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 12455–12464. [Google Scholar]
  28. Wang, L.; Yang, X.; Tan, H.; Bai, X.; Zhou, F. Few-shot class-incremental SAR target recognition based on hierarchical embedding and incremental evolutionary network. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5204111. [Google Scholar] [CrossRef]
  29. Zhao, Y.; Zhao, L.; Ding, D.; Hu, D.; Kuang, G.; Liu, L. Few-Shot Class-Incremental SAR Target Recognition via Cosine Prototype Learning. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5212718. [Google Scholar] [CrossRef]
  30. Zhao, Y.; Zhao, L.; Zhang, S.; Ji, K.; Kuang, G.; Liu, L. Azimuth-aware Subspace Classifier for Few-Shot Class-Incremental SAR ATR. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5203020. [Google Scholar] [CrossRef]
  31. Kong, L.; Gao, F.; He, X.; Wang, J.; Sun, J.; Zhou, H.; Hussain, A. Few-shot class-incremental SAR target recognition via orthogonal distributed features. IEEE Trans. Aerosp. Electron. Syst. 2024, 61, 325–341. [Google Scholar] [CrossRef]
  32. Karantaidis, G.; Pantsios, A.; Kompatsiaris, I.; Papadopoulos, S. Few-Shot Class-Incremental Learning For Efficient SAR Automatic Target Recognition. arXiv 2025, arXiv:2505.19565. [Google Scholar]
  33. Zhao, Y.; Zhao, L.; Zhang, S.; Ji, K.; Kuang, G. Few-shot class-incremental sar target recognition via decoupled scattering augmentation classifier. In Proceedings of the IGARSS 2024-2024 IEEE International Geoscience and Remote Sensing Symposium, Athens, Greece, 7–12 July 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 7584–7587. [Google Scholar]
  34. Potter, L.C.; Moses, R.L. Attributed scattering centers for SAR ATR. IEEE Trans. Image Process. 1997, 6, 79–91. [Google Scholar] [CrossRef] [PubMed]
  35. Wang, Z.; Kang, Y.; Zeng, X.; Wang, Y.; Zhang, D.; Sun, X. SAR-AIRcraft-1.0: High-resolution SAR aircraft detection and recognition dataset. J. Radars 2023, 12, 906. [Google Scholar]
  36. Hou, X.; Ao, W.; Song, Q.; Lai, J.; Wang, H.; Xu, F. FUSAR-Ship: Building a high-resolution SAR-AIS matchup dataset of Gaofen-3 for ship detection and recognition. Sci. China Inf. Sci. 2020, 63, 140303. [Google Scholar] [CrossRef]
  37. Rebuffi, S.A.; Kolesnikov, A.; Sperl, G.; Lampert, C.H. iCaRL: Incremental classifier and representation learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2001–2010. [Google Scholar]
  38. Castro, F.M.; Marín-Jiménez, M.J.; Guil, N.; Schmid, C.; Alahari, K. End-to-End Incremental Learning. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 233–248. [Google Scholar]
  39. Hou, S.; Pan, X.; Loy, C.C.; Wang, Z.; Lin, D. Learning A Unified Classifier Incrementally via Rebalancing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 831–839. [Google Scholar]
  40. Song, Z.; Zhao, Y.; Shi, Y.; Peng, P.; Yuan, L.; Tian, Y. Learning with Fantasy: Semantic-Aware Virtual Contrastive Constraint for Few-Shot Class-Incremental Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 24183–24192. [Google Scholar]
Figure 1. Overall framework of the FCPC.
Figure 1. Overall framework of the FCPC.
Remotesensing 17 03518 g001
Figure 2. Ten-class targets of the MSTAR dataset in SAR and optical images.
Figure 2. Ten-class targets of the MSTAR dataset in SAR and optical images.
Remotesensing 17 03518 g002
Figure 3. Seven-class targets of the SAR-AIRcraft-1.0 dataset in SAR and optical images.
Figure 3. Seven-class targets of the SAR-AIRcraft-1.0 dataset in SAR and optical images.
Remotesensing 17 03518 g003
Figure 4. Ten-class targets of the FUSAR dataset in SAR and optical images.
Figure 4. Ten-class targets of the FUSAR dataset in SAR and optical images.
Remotesensing 17 03518 g004
Figure 5. Feature Non-Sparsity and Normalization of our method with the VSS. (a) Feature Non-Sparsity. (b) Feature Normalization.
Figure 5. Feature Non-Sparsity and Normalization of our method with the VSS. (a) Feature Non-Sparsity. (b) Feature Normalization.
Remotesensing 17 03518 g005
Figure 6. t-SNE embeddings of class features extracted by our method with the VSS. (a) t-SNE results of the baseline. (b) t-SNE results of our method with the VCG. (c) t-SNE results of our method with the VSS.
Figure 6. t-SNE embeddings of class features extracted by our method with the VSS. (a) t-SNE results of the baseline. (b) t-SNE results of our method with the VCG. (c) t-SNE results of our method with the VSS.
Remotesensing 17 03518 g006
Figure 7. Performance of our method with varying locations of the DA (a) The Avg Acc and Avg HA of our method at different configurations. (b) The accuracy of our method evaluated on the old categories at each incremental session.
Figure 7. Performance of our method with varying locations of the DA (a) The Avg Acc and Avg HA of our method at different configurations. (b) The accuracy of our method evaluated on the old categories at each incremental session.
Remotesensing 17 03518 g007
Figure 8. The class separation (R) and interclass distances maintained by our method trained by the PML. (a) The class separation (R) maintained by our method at incremental sessions. (b) The interclass distances maintained by our method at incremental sessions.
Figure 8. The class separation (R) and interclass distances maintained by our method trained by the PML. (a) The class separation (R) maintained by our method at incremental sessions. (b) The interclass distances maintained by our method at incremental sessions.
Remotesensing 17 03518 g008
Figure 9. The benchmark curves and bars on the MSTAR-FSCIL, SAR-AIRcraft-FSCIL, and FUSAR-FSCIL datasets.
Figure 9. The benchmark curves and bars on the MSTAR-FSCIL, SAR-AIRcraft-FSCIL, and FUSAR-FSCIL datasets.
Remotesensing 17 03518 g009
Figure 10. The confusion matrices of different benchmarks at the last session tested on the MSTAR-FSCIL dataset.
Figure 10. The confusion matrices of different benchmarks at the last session tested on the MSTAR-FSCIL dataset.
Remotesensing 17 03518 g010
Figure 11. The confusion matrices of different benchmarks at the last session tested on the SAR-AIRcraft-FSCIL dataset.
Figure 11. The confusion matrices of different benchmarks at the last session tested on the SAR-AIRcraft-FSCIL dataset.
Remotesensing 17 03518 g011
Figure 12. The confusion matrices of different benchmarks at the last session tested on the FUSAR-FSCIL dataset.
Figure 12. The confusion matrices of different benchmarks at the last session tested on the FUSAR-FSCIL dataset.
Remotesensing 17 03518 g012
Figure 13. Session-wise t-SNE results for the MSTAR-FSCIL dataset. Arrows represent incremental processes.
Figure 13. Session-wise t-SNE results for the MSTAR-FSCIL dataset. Arrows represent incremental processes.
Remotesensing 17 03518 g013
Figure 14. Session-wise t-SNE results for the SAR-AIRcraft-FSCIL dataset. Arrows represent incremental processes.
Figure 14. Session-wise t-SNE results for the SAR-AIRcraft-FSCIL dataset. Arrows represent incremental processes.
Remotesensing 17 03518 g014
Table 1. Configurations of the MSTAR-FSCIL dataset.
Table 1. Configurations of the MSTAR-FSCIL dataset.
SessionOrderTypeSerial No.TrainTest
Base1BTR70c71233196
22S1b01299274
3BRDM2E-71298274
4BMP29563233196
Incremental5ZIL131E125274
6T62A515273
7D792v130155274
8BTR60k10yt75325195
9T721325196
10ZSU234d085274
Table 2. Configurations of the SAR-AIRcraft-FSCIL dataset.
Table 2. Configurations of the SAR-AIRcraft-FSCIL dataset.
SessionOrderTypeTrainTest
Base1Other2000200
2A2202000200
3Boeing7872000200
4Boeing7372000200
Incremental5A3205200
6ARJ215200
7A3305200
Table 3. Configurations of the FUSAR-FSCIL dataset.
Table 3. Configurations of the FUSAR-FSCIL dataset.
SessionOrderTypeTrainTest
Base1Cargo24030
2Other24030
3Fishing24030
4BulkCarrier24030
Incremental5Tanker530
6Unspecificied530
7Container530
8Dredger530
9Tug530
10GeneralCargo530
Table 4. Effects of the proposed modules evaluated on the MSTAR-FSCIL dataset.
Table 4. Effects of the proposed modules evaluated on the MSTAR-FSCIL dataset.
Forward-CompatibleStable DiscriminatingMSTAR-FSCIL (%)
VCGSLDAPMLNCMAvg. AccAvg. HAPD ↓
----75.2661.3741.25
---79.2871.2635.51
---77.6465.1136.40
--82.4976.5932.83
-82.7577.6432.50
83.0378.1332.15
Table 5. Comparison of benchmarks on the MSTAR-FSCIL dataset.
Table 5. Comparison of benchmarks on the MSTAR-FSCIL dataset.
MethodsSessionsAvg AccPD ↓
1234567
Ft-CNN98.9476.2162.0152.0247.5543.0838.1259.7060.82
Oracle98.9485.0577.1778.3073.8670.3467.1078.6831.84
iCaRL [37]92.1283.0873.4666.4165.0460.4254.2770.6937.85
EEIL [38]98.9481.7670.7762.7863.3561.5556.6870.8342.26
LUCIR [39]99.8990.0776.9772.6070.3366.6760.8976.7739.00
TOPIC [24]91.1085.2772.7763.2561.2456.0350.0368.5341.07
ERDIL [25]98.9487.9276.0270.1468.6664.1757.9474.8341.00
IDLVQC [26]97.0283.7571.1262.2759.5454.7149.2068.2347.82
CEC [27]90.5480.5272.2772.5366.9761.7657.9771.7932.57
FACT [9]98.8587.3667.1464.2449.5046.6945.0165.5453.84
ALICE [10]97.5586.8372.3666.7363.0559.3654.4471.4743.11
SAVC [40]97.1089.4379.6475.6672.4768.4864.6478.2032.46
CPL [29]99.8990.4178.3274.9373.0469.6663.6678.5636.23
AASC [30]99.8989.7376.3373.6471.4666.9762.3177.1937.58
Ours (FCPC)99.4794.4885.2882.1078.3174.2667.3283.0332.15
Table 6. Comparison of benchmarks on the SAR-AIRcraft-FSCIL dataset.
Table 6. Comparison of benchmarks on the SAR-AIRcraft-FSCIL dataset.
MethodsSessionsAvg AccPD ↓
1234
Ft-CNN99.6282.9367.6659.3277.3840.30
Oracle99.4781.2469.8364.3178.7135.16
iCaRL [37]99.5088.8276.2065.3782.4734.13
EEIL [38]99.6285.0073.1367.2681.2532.36
LUCIR [39]99.7587.6875.7668.2382.8631.52
TOPIC [24]99.6489.0976.9567.0683.1832.58
ERDIL [25]99.6286.5075.5168.8082.6130.82
IDLVQC [26]99.3785.5372.8064.1080.4535.27
CEC [27]78.9765.6457.7651.7563.5327.22
FACT [9]99.4479.5566.3656.8875.5642.56
ALICE [10]95.6374.6262.3055.6672.0539.97
SAVC [40]99.0483.4473.7467.9781.0531.07
CPL [29]99.3788.5577.9872.0684.4927.31
AASC [30]99.5089.1277.1869.2883.7730.22
Ours (FCPC)99.3791.0979.1571.8085.3527.57
Table 7. Comparison of benchmarks on the FUSAR-FSCIL dataset.
Table 7. Comparison of benchmarks on the FUSAR-FSCIL dataset.
MethodsSessionsAvg AccPD ↓
1234567
Ft-CNN75.8358.5348.1142.2935.9632.2230.3046.1845.53
Oracle79.5065.2056.4449.3345.7539.4836.4053.1643.10
iCaRL [37]75.8373.4059.3352.5747.1342.4139.1355.6936.70
EEIL [38]75.8362.5350.9446.7640.3335.4134.7349.5041.10
UCIR [39]80.8371.4760.2853.1946.6741.1137.7355.9043.10
TOPIC [24]75.6767.0057.1150.8644.4640.4137.2053.2438.47
ERDIL [25]77.5060.4043.7236.6226.0422.5921.3041.1756.20
IDLVQC [26]75.0065.6054.5644.9039.0835.5630.8349.3644.17
CEC [27]84.1771.5360.6154.6248.7544.3040.4357.7743.74
FACT [9]77.5075.0063.8356.5250.0045.0440.1058.2837.40
ALICE [10]60.0056.6049.0045.1040.4236.1533.2345.7926.77
SAVC [40]78.8372.6060.7254.2949.4244.5640.7357.3138.10
CPL [29]80.1770.5359.6152.6246.2543.2340.1356.0840.04
AASC [30]82.2370.9260.5353.2147.3543.5539.3356.7342.90
Ours (FCPC)80.1776.3364.9457.8651.4647.8544.2060.4035.97
Table 8. Comparison of benchmarks on the Combined dataset.
Table 8. Comparison of benchmarks on the Combined dataset.
MethodsSessionsAvg AccPD ↓
12345678910
Ft-CNN90.1380.2972.0165.2459.8555.2250.9647.8244.9942.4060.8947.73
Oracle91.7788.4883.3581.0977.5871.1770.9767.5363.7761.2875.7030.49
iCaRL [37]85.3182.2173.1167.4265.6262.0958.1856.4253.5750.7265.4634.59
EEIL [38]90.1380.0373.0868.7463.9760.2057.7256.0853.9251.9865.5938.15
UCIR [39]91.1989.0783.5478.0375.7972.6870.8868.2865.2162.8875.7628.31
TOPIC [24]87.5084.2273.5068.0965.4663.1559.6158.0055.1251.2966.5936.21
ERDIL [25]89.3085.1274.5467.5964.1662.0159.1158.2456.3253.0166.9436.29
IDLVQC [26]84.2282.3973.5965.4362.2660.9358.9358.1355.0253.8465.4730.38
CEC [27]91.1888.9879.8673.5270.7966.6264.3561.9258.7857.1571.3134.03
FACT [9]90.1987.6181.1576.0073.4668.4665.1161.8359.5958.3572.1831.84
ALICE [10]90.1986.5183.3577.6574.5470.5366.1964.3460.8759.3373.3530.86
SAVC [40]88.3785.1181.281.4577.7174.4273.6870.568.6666.2176.7322.16
Ours (FCPC)91.1987.8781.8080.5678.5075.1673.5371.0668.5265.8377.4025.36
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guan, D.; Feng, R.; Xie, Y.; Zheng, X.; Li, B.; Xiang, D. Few-Shot Class-Incremental SAR Target Recognition with a Forward-Compatible Prototype Classifier. Remote Sens. 2025, 17, 3518. https://doi.org/10.3390/rs17213518

AMA Style

Guan D, Feng R, Xie Y, Zheng X, Li B, Xiang D. Few-Shot Class-Incremental SAR Target Recognition with a Forward-Compatible Prototype Classifier. Remote Sensing. 2025; 17(21):3518. https://doi.org/10.3390/rs17213518

Chicago/Turabian Style

Guan, Dongdong, Rui Feng, Yuzhen Xie, Xiaolong Zheng, Bangjie Li, and Deliang Xiang. 2025. "Few-Shot Class-Incremental SAR Target Recognition with a Forward-Compatible Prototype Classifier" Remote Sensing 17, no. 21: 3518. https://doi.org/10.3390/rs17213518

APA Style

Guan, D., Feng, R., Xie, Y., Zheng, X., Li, B., & Xiang, D. (2025). Few-Shot Class-Incremental SAR Target Recognition with a Forward-Compatible Prototype Classifier. Remote Sensing, 17(21), 3518. https://doi.org/10.3390/rs17213518

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop