Next Article in Journal
Intercomparison of Antarctic Sea-Ice Thickness Estimates from Satellite Altimetry and Assessment over the 2019 Data-Rich Year
Next Article in Special Issue
A Target Near-Field Scattering Measurement Technique Utilizing 3D Near-Field Imaging via Cylindrical Scanning
Previous Article in Journal
Assessment of the Potential of Spaceborne GNSS-R Interferometric Altimetry for Monthly Marine Gravity Anomaly
Previous Article in Special Issue
A Multiple-Input Multiple-Output Synthetic Aperture Radar Echo Separation and Range Ambiguity Suppression Processing Framework for High-Resolution Wide-Swath Imaging
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Graph-Based Few-Shot Learning for Synthetic Aperture Radar Automatic Target Recognition with Alternating Direction Method of Multipliers

1
Key Laboratory for Information Science of Electromagnetic Waves (MoE), School of Information Science and Technology, Fudan University, Shanghai 200433, China
2
School of Mathematics and Statistics, Xi’an Jiaotong University, Xi’an 710049, China
3
Yiwu Research Institute, Fudan University, Yiwu 322000, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(7), 1179; https://doi.org/10.3390/rs17071179
Submission received: 22 December 2024 / Revised: 16 March 2025 / Accepted: 24 March 2025 / Published: 26 March 2025
(This article belongs to the Special Issue SAR-Based Signal Processing and Target Recognition (Second Edition))

Abstract

:
Synthetic aperture radar (SAR) automatic target recognition (ATR) underpins various remote sensing tasks, such as defense surveillance, environmental monitoring, and disaster management. However, the scarcity of annotated SAR data significantly limits the performance of conventional data-driven methods. To address this challenge, we propose a novel few-shot learning (FSL) framework: the alternating direction method of multipliers–graph convolutional network (ADMM-GCN) framework. ADMM-GCN integrates a GCN with ADMM to enhance SAR ATR under limited data conditions, effectively capturing both global and local structural information from SAR samples. Additionally, it leverages a mixed regularized loss to mitigate overfitting and employs an ADMM-based optimization strategy to improve training efficiency and model stability. Extensive experiments conducted on the Moving and Stationary Target Acquisition and Recognition (MSTAR) dataset demonstrate the superiority of ADMM-GCN, achieving an impressive accuracy of 92.18% on the challenging three-way 10-shot task and outperforming the benchmarks by 3.25%. Beyond SAR ATR, the proposed approach also advances FSL for real-world applications in remote sensing and geospatial analysis, where learning from scarce data is essential.

1. Introduction

Synthetic aperture radar (SAR) provides high-resolution imaging regardless of lighting or weather conditions [1], making it essential for defense surveillance, environmental monitoring, and disaster management [2,3,4,5]. However, achieving robust performance in these critical applications remains challenging due to the scarcity of annotated SAR data.
Historically, SAR ATR methods have relied on manually designed features, such as statistical, spectral, and structural descriptors, processed through complex signal processing frameworks [3,6,7,8,9]. Techniques such as template matching [10] and model-based approaches [11] have been instrumental in advancing the field. However, these methods depend heavily on handcrafted features, requiring significant human effort and limiting scalability. Additionally, their reliance on scenario-specific designs hinders generalization to diverse detection contexts [12,13].
Deep learning has demonstrated remarkable success in various tasks, including emotion recognition [14], healthcare applications [15], and intelligent surveillance [16]. These advancements highlight its strong capability in feature extraction and classification, offering valuable insights for SAR ATR. Deep learning has significantly improved SAR ATR performance by learning multi-level features from raw data [17,18,19]. However, its reliance on large, well-annotated datasets poses a limitation, as SAR data acquisition is costly and requires specialized expertise [20,21,22].
Few-shot learning (FSL), which allows efficient classification with a small number of labeled samples per class, presents a viable solution to the data scarcity issue [23,24]. This capability makes FSL a possible method for SAR ATR, where the specialized nature of SAR targets and the high cost of data collection often render large-scale datasets infeasible. Leveraging task-specific knowledge, FSL generalizes across new tasks, addressing the critical challenges of SAR ATR. Figure 1 provides an overview of the FSL paradigm, illustrating both training and test tasks. It illustrates a five-way one-shot FSL classification task, where the support set provides one labeled sample per class, while the query set consists of unlabeled samples for prediction. The training tasks (top of Figure 1) involve classes encountered during training. In these tasks, the model learns feature representations from the support set and uses them to classify the samples in the query set. The test tasks (bottom of Figure 1) contain novel classes that were not seen during training. The model generalizes to these unseen classes using the learned feature representations, demonstrating its ability to adapt to new SAR target categories.
To address data scarcity challenges, we propose the alternating direction method of multipliers–graph convolutional network (ADMM-GCN) framework, a novel framework for few-shot SAR ATR that integrates a GCN with the ADMM algorithm. By leveraging the representational capabilities of GCNs, ADMM-GCN effectively captures both global and local characteristics of SAR samples while mitigating overfitting through a mixed regularized loss function. Moreover, the ADMM algorithm is introduced to ensure efficient optimization, achieving consistent convergence even with limited data.
Figure 2 visualizes the ADMM-GCN framework, which processes SAR data through graph construction and iterative updates of node and edge features. In Figure 2, the workflow begins with SAR data represented as a graph, where samples correspond to nodes and their relationships to edges. Through iterative updates, node features are refined to encapsulate key characteristics of SAR targets, while edge features capture nuanced interactions between nodes. This iterative process, highlighted in the circular flow, continuously enhances feature representation and improves classification accuracy, making ADMM-GCN well suited for scenarios where data are scarce.
To summarize, this study’s main contributions to the literature are as follows:
  • We propose an innovative framework termed ADMM-GCN for few-shot SAR ATR, which effectively combines global context with local feature analysis by constructing a relational graph among features, thereby enhancing the overall feature representation under limited-data scenarios.
  • A mixed regularized loss function is designed to mitigate the common challenge of overfitting in FSL, enhancing the model’s stability and generalizability across diverse scenarios without relying on extensive data augmentation.
  • The ADMM algorithm is integrated into few-shot SAR ATR to ensure consistent convergence to the global optimum while avoiding local optima, simplifying optimization by decomposing complex problems into tractable subproblems.
  • Extensive experiments conducted on the Moving and Stationary Target Acquisition and Recognition (MSTAR) dataset verify the superiority of the proposed ADMM-GCN, achieving an impressive accuracy of 92.18% on the challenging three-way 10-shot task, outperforming the benchmarks by 3.25%.
The paper is structured as follows: Section 2 discusses related works. Section 3 explains the details of the ADMM-GCN framework. Section 4 presents the experimental validation, followed by a detailed analysis and discussions in Section 5. The paper concludes with Section 6.

2. Related Works

2.1. Few-Shot SAR Target Recognition

Few-shot SAR target recognition focuses on accurately identifying objects in SAR images with limited labeled samples. This subsection explores the development of this field through two main aspects: data-augmentation-based methods and model-optimization-based methods.

2.1.1. Data-Augmentation-Based Methods

Data augmentation enlarges the available dataset by modifying original SAR data to create synthetic samples resembling the real data, thereby enhancing model learning under data constraints. Common techniques in few-shot SAR target recognition include basic transformations such as rotation and scaling [4,25,26], as well as advanced approaches leveraging deep generative models to produce more complex variations [27,28]. For example, Ding et al. applied data augmentation strategies including target position shifts, speckle noise variations, and pose changes to improve model robustness [4]. Similarly, Song et al. proposed an adversarial autoencoder to generate SAR images with different target views, introducing valuable variability into the training data [27].
While data augmentation expands training sets, it often incurs high computational costs and requires expertise to ensure meaningful transformations. Additionally, deep generative models rely on large datasets, which is an obstacle to their effectiveness in FSL settings.

2.1.2. Model-Optimization-Based Methods

Three primary approaches can be distinguished among the current trends in model-optimization-based techniques for few-shot SAR target recognition: transfer learning [29,30], metric learning [31,32], and meta-learning [33].
Transfer learning enables pre-trained models to adapt to new tasks by transferring knowledge across domains. For instance, Rostami et al. [29] address SAR-EO domain transfer using a sliced Wasserstein distance (SWD)-based invariant embedding space, and Tai et al. [30] propose a few-shot transfer learning strategy incorporating selective feature transfer and a Bayesian CNN for SAR image classification; while transfer learning offers high efficiency and minimal fine-tuning, its effectiveness relies on similarity between the source and target domains. Significant domain shifts can degrade performance, requiring specialized adaptation strategies. Additionally, the availability of sufficiently large and well-curated source datasets remains a practical constraint.
Metric learning maps samples into an optimal metric space, emphasizing similarity-based representations. For instance, Wang et al. [31] introduced an attribute-guided multi-scale prototypical network with subband decomposition to enhance feature extraction from limited SAR data. Similarly, Ren et al. [32] proposed a transductive prototype reasoning approach that refines class prototypes, improving target identity reasoning with few labeled samples. However, metric learning often prioritizes local similarity patterns, potentially neglecting broader structural features that are critical for tasks such as object recognition. To improve SAR target classification, an effective balance between local and global feature representations is needed.
Meta-learning trains models to quickly adapt to new tasks, improving few-shot performance. However, it requires extensive meta-training and risks overfitting to small datasets. Meta-learning enhances model adaptability by leveraging cross-task learning, enabling efficient generalization from limited samples. For instance, Fu et al. [33] proposed MSAR, a meta-learning framework that optimizes initialization and update strategies for better adaptation to new tasks. However, meta-learning involves high computational complexity and is prone to overfitting, particularly in data-scarce scenarios.
In general, although prior research has significantly advanced the field of SAR target recognition under few-shot scenarios, challenges such as model complexity, susceptibility to overfitting, and effective handling of highly limited data remain unresolved, warranting further exploration.

2.2. Alternating Direction Method of Multipliers

Gradient-based optimization techniques, such as Adam [34] and stochastic gradient descent (SGD) [35], are widely used in deep learning. However, these methods face several limitations:
  • They often converge to local optima, making it difficult to reach the global optimum and hindering the overall training process;
  • Their effectiveness is highly sensitive to input data quality, requiring meticulous preprocessing to ensure convergence, which complicates training and affects model performance.
These challenges become even more pronounced in FSL, where the extremely limited number of labeled samples exacerbates issues such as overfitting and unstable convergence. Additionally, gradient-based optimizers rely on well-defined loss landscapes and proper initialization, making them vulnerable to poor convergence in low-data regimes.
To address these challenges, alternative optimization approaches have been explored. One such alternative is the alternating direction method of multipliers (ADMM) [36], which has shown significant promise for constrained and structured optimization problems. ADMM has been widely applied in deep learning due to its advantages [37]:
  • ADMM decomposes the optimization problem into smaller, more manageable subproblems, each of which can be solved optimally with theoretical guarantees of convergence. This decomposition is particularly beneficial in FSL, where limited data necessitates a stable and structured training process.
  • Unlike gradient-based methods, ADMM is inherently robust to parameter initialization, ensuring stable convergence even when training data are scarce.
  • By introducing an auxiliary variable, ADMM enforces constraints during optimization, which not only stabilizes training but also enhances generalization, making it well suited for FSL applications.
In few-shot SAR target recognition, these properties make ADMM particularly advantageous. First, the ability of ADMM to break down complex optimization problems helps mitigate the impact of gradient vanishing, a common issue in low-data regimes. Additionally, its built-in regularization mechanisms help prevent overfitting, which is crucial when training data are extremely limited.
Several prior studies have highlighted the effectiveness of ADMM in handling complex regularization functions and improving generalization performance in deep learning applications [38,39]. These works demonstrate that ADMM achieves superior convergence properties compared to standard gradient-based optimizers such as Adam, particularly in scenarios where optimization constraints and structured learning play a crucial role.
Despite its demonstrated success in various deep learning tasks, ADMM remains underexplored in few-shot SAR target recognition. This presents an opportunity to investigate its potential to improve model interpretability, generalization, and stability. The proposed approach aims to bridge this gap by leveraging the properties of ADMM to enhance few-shot SAR target recognition performance.

3. Methodology

3.1. Framework of ADMM-GCN

Figure 3 presents the overall architecture of the ADMM-GCN framework, which consists of three main components: the embedding module (EM), the graph convolutional module (GCM), and the ADMM optimization process.
The EM extracts initial feature representations from SAR images, providing a foundational understanding of the target characteristics. These features are subsequently processed by the GCM, which encodes and propagates global contextual dependencies using graph-based algorithms. By aggregating and refining features, this module enables the model to effectively capture intricate patterns and interdependencies, which are essential for accurate target classification in FSL scenarios.
Furthermore, the ADMM optimization process is integrated into the framework to iteratively refine the model parameters by decomposing the original optimization problem into a sequence of subproblems. In this process, ADMM follows three key steps: updating the primal variable θ (network parameters) via Equation (18a), updating the auxiliary variable z via Equation (18b), and updating the dual variable α via Equation (18c). Figure 3 visually illustrates these updates, demonstrating the interaction between ADMM, the GCM, and the EM. The detailed ADMM methodology and optimization steps are provided in Section 3.4, while further descriptions of each component and their integration within the framework are discussed in subsequent sections.

3.2. Network Architecture

3.2.1. Embedding Module

The EM in the ADMM-GCN framework acts as the feature extractor, transforming SAR images into latent feature representations, as illustrated in Figure 4. This module is a critical preparatory stage, generating feature representations that serve as input for the subsequent GCM, enabling effective contextual information propagation and classification.
For the ith sample X i , the associated embedding vector Ξ i can be expressed as
Ξ i = f θ ( X i )
where f θ : R H × W R P is the embedding function, θ represents the learnable parameters, H and W indicate the dimensions of the input image, and P specifies the embedding vector’s dimensionality.
Starting with a 3 × 3 convolutional layer, the EM extracts fundamental features from the SAR input. Subsequently, batch normalization ensures stability, while the ReLU activation function introduces non-linearity, enabling effective feature extraction. A maxpooling layer then downsamples the feature map, highlighting dominant features while reducing computational complexity.
As the network progresses, two additional convolutional layers with filter sizes increasing from 32 to 128 are employed to capture intricate features. These layers, paired with batch normalization and ReLU activation, ensure high-quality feature representation. Intermittent maxpooling layers further abstract the feature maps, reducing their size while preserving critical information. In the final stage, a 12 × 12 convolutional kernel operates on a 128-channel feature map, synthesizing high-level features into a 64-channel output. This design balances computational efficiency with analytical depth, enabling the module to generate rich, compact feature representations.
The extracted features are seamlessly relayed to the GCM, where the GCN interprets the data structure and relationships for higher-level contextual analysis. This integrated workflow transforms localized features into a global perspective, enabling comprehensive interpretation and analysis in FSL scenarios.

3.2.2. Graph Convolutional Module

GCNs [40] stand at the forefront of graph-based deep learning, building upon the core concepts of CNNs. Mimicking CNNs, modern GCNs learn the common local and global structural patterns of graphs through designed convolution functions [41]. The embedding of a node is generated by collecting features from its neighbors and then enhancing this information through a series of linear transformations and non-linear activations.
Figure 5 illustrates feature aggregation and the edge information interaction within a GCN. Initially, edges AB and CD, represented as dashed lines, are not directly connected. Through the graph convolution process and successive feature aggregation, nodes progressively update their features, exemplified by the path A B C D , and then edge AB and edge CD have feature fusion as well as interaction through edge BC. Blue nodes signify the initial stages of feature aggregation, while red nodes represent the outcome after several iterations of feature aggregation and edge information interaction.
As the process unfolds across layers, the resulting embeddings reflect not only individual node characteristics but also their interconnections within the graph. Such depth of information, combined with the network’s end-to-end trainability, is essential for effectively tackling few-shot SAR target recognition tasks. The intricate interconnectedness and information exchange between nodes as well as edges underscore the effectiveness of GCNs in capturing the relational inferences crucial for advanced recognition tasks.
Let G = ( V , E ) represent a graph, where V is a collection of N nodes and E defines the edges connecting these nodes. In this undirected graph, a node v i V is linked to another node v j via an edge e i , j = { v i , v j } E , which reflects a similarity relationship between the two. The topology of the graph is encoded in the adjacency matrix A R N × N , where a i , j = e i , j .
Each node v i is assigned a feature vector x i , and all node features are stored in the matrix X R N × M , where M indicates the dimension of the feature space. The similarity e i , j between nodes i and j is determined using a function f d , expressed as
e i , j = f d ( x i , x j ) = M L P ( | x i x j | ) ,
where f d measures the similarity by leveraging a multilayer perceptron (MLP) trained on the absolute difference between the feature vectors of nodes v i and v j .
Figure 6 illustrates a graphical representation of data characterized by both node and edge features. Nodes v i in the graph are each associated with an M-dimensional feature vector, represented as [ x 1 , x 2 , , x M ] . Edges e i , j between nodes v i and v j are represented by their own attribute vectors, capturing the similarities as well as the interactions between the connected nodes.
GCNs extend traditional CNNs to the graph domain. Utilizing the simplification introduced in [40], the convolution operation in a GCN is defined as
g υ s = υ I + D 1 2 A D 1 2 s ,
where s denotes the input signal, g υ the spectral filter, I the identity matrix, ★ the convolution operator, and  υ the Chebyshev coefficient.
The degree matrix D of A is defined as
D = diag ( d 1 , d 2 , , d n ) ,
where each d i = j a i j represents the degree of vertex i.
Using the normalization approach introduced by Kipf and Welling [40], the convolution matrix is transformed as
I + D 1 2 A D 1 2 D ˜ 1 2 A ˜ D ˜ 1 2 ,
where A ˜ = A + I and D ˜ = j A ˜ i j .
Extending the convolution definition to cater to a signal with M input channels, that is, X R N × M where each vertex is affiliated with an M-dimensional feature vector, the propagation rule for this model is given by:
H ( l + 1 ) = σ D ˜ 1 2 A ˜ D ˜ 1 2 H ( l ) Θ ( l ) ,
The activation function σ used in the proposed GCN module is the ReLU function, formulated as ReLU ( · ) = max ( 0 , · ) . Considering the balance between model complexity and computational efficiency, a three-layer architecture is adopted. This configuration ensures efficient data processing while retaining key data features. The output Z of the third layer is derived using the established propagation rule:
Z = H ( 3 ) = A ^ ReLU A ^ ReLU A ^ X Θ ( 0 ) Θ ( 1 ) Θ ( 2 ) ,
where A ^ represents the normalized adjacency matrix, defined as
A ^ = D ˜ 1 2 A ˜ D ˜ 1 2 .
To accentuate intra-class similarity while maintaining clear inter-class divergence, this method computes cluster centers for each category. This approach highlights the nuances of each category, ensuring that similar instances within a category are closely aligned, while distinctly separating different categories. The cluster center c n for the nth category is the mean of the embedding vectors in its support set:
c n = 1 K z i S n z i ,
where K denotes the shot number, S n the support set for the nth category, and  z i the corresponding feature vector.
Subsequently, the feature output of a query sample z i is processed through a S o f t m a x classifier to determine the class probability distribution. The predicted label y i ^ for the query sample is calculated using the S o f t m a x function:
p ( y i ^ = n | z i ) = e S ( c n , z i ) j = 1 N e S ( c n , z i ) ,
where S is a metric function adhering to the cosine similarity, defined as
S ( c n , z i ) = c n · z i c n × z i .

3.3. Construction of Regularized Mixed Loss

To address the prevalent challenge of overfitting in FSL, a specialized regularized mixed loss function has been designed within the ADMM-GCN framework F . This method not only streamlines the model architecture but also alleviates overfitting, thereby improving both stability and generalizability. The neural network’s output in the ADMM-GCN framework is formulated as
y i = F θ i ( x ) ,
where y i represents the ith output, F θ i ( x ) represents the function corresponding to the ith output’s response to input x, and  θ symbolizes the weight parameters.
The cross-entropy loss L ( θ ) is defined as
L ( θ ) = y i ^ log y i = y i ^ log f θ i ( x ) ,
where y i ^ is the actual label and y i denotes the neural network’s output.
In ADMM-GCN, a loss function has been devised that extends the original loss function by incorporating a regularization term. For parameters θ , the expression for the L 2 regularization term is formulated as
R ( θ ) = λ 2 θ 2 = λ 2 θ 1 2 + θ 2 2 + + θ n 2 ,
where λ represents the regularization parameter that balances the strength of the regularization, while  θ 2 denotes the L 2 norm of the weight parameter.
Consequently, the mixed regularized loss is given as
L R = L ( θ ) + R ( θ ) = y i ^ log y i + λ 2 θ 2 = y i ^ log f θ i ( x ) + λ 2 θ 2 .

3.4. ADMM Optimizer

Having established the loss function, we now shift our attention to its optimization. The steps of the ADMM algorithm for regularized loss optimization are summarized in Algorithm 1.
Consider the following minimization problem:
arg min θ y i ^ log f θ i ( x ) + λ 2 θ 2 .
Following this, this method examines the constrained version, defined as follows:
arg min θ , z y i ^ log f θ i ( x ) + λ 2 z 2 s . t . z = θ
According to the typical neural network framework, an iterative procedure solution is typically required to solve (15). The ADMM technique, gaining significant attention for its efficacy in non-convex deep learning frameworks [42,43,44], is therefore used in an attempt to address the minimization problem (16). Therefore, the augmented Lagrangian function for the problem is given as
L = y i ^ log f θ i ( x ) + λ 2 θ 2 + α , θ z + ρ 2 θ z 2 2 ,
where α denotes the Lagrangian multiplier corresponding to the constraint z = θ , while ρ > 0 is the penalty parameter.
Following the ADMM paradigm, the optimal point is sought by maximizing over the dual variable α along with minimizing over the primal variables θ and z, respectively. The ADMM algorithm proceeds by alternately updating { θ , z , α } through the resolution of the following subproblems:
{ (18a) arg min θ y i ^ log f θ i ( x ) + ρ 2 θ z + α ρ 2 2 (18b) arg min z y i ^ log f θ i ( x ) + λ 2 z 2 (18c) α t + 1 α t + ρ ( θ z )
An approximate solution to the first problem (18a) can be derived by utilizing a gradient-based strategy over a specified number of iterations. The Adam iterative approach [34] is adopted for this purpose. Additionally, PyTorch’s automatic differentiation functionality could be leveraged for variable-related numerical gradient computations [45].
The second problem, (18b), can be tackled in a manner akin to that of [46], and the exact solution of z is formulated as
z = S o f t ( θ + α ρ ) ,
where S o f t ( ) denotes the soft threshold operator. The corresponding formulation for S o f t ( )  is
S o f t μ 2 ( y ) = arg min x ( y x ) 2 + μ | x | .
Algorithm 1 ADMM Algorithm for Optimization with L 2 Regularization
1:
Input:  λ ; ρ ; initial values of θ , z, and  α
2:
Output: Optimal θ and z
3:
repeat
4:
   Update θ by solving
θ arg min θ y i ^ log f θ i ( x ) + ρ 2 θ z + α ρ 2 2
5:
   Update z by solving
z arg min z y i ^ log f θ i ( x ) + λ 2 z 2
6:
   Update Lagrangian multiplier α :
α α + ρ ( θ z )
7:
until Convergence criteria are met

4. Experiment

4.1. Dataset

The Moving and Stationary Target Acquisition and Recognition (MSTAR) dataset [47] serves as a widely recognized benchmark in the field of SAR image analysis and target recognition. This dataset features X-band imaging radar operating in HH polarization, producing images with dimensions of 128 × 128 pixels and a spatial resolution of 0.3 × 0.3 m, including ten distinct military vehicle targets. In this study, the training and testing sets are constructed using targets imaged at depression angles of 17° and 15°, respectively, following standard experimental protocols. Table 1 summarizes the number of samples for each target category. Figure 7 illustrates optical and SAR image samples of these military vehicles within the MSTAR dataset.
To create the training task set T t r a i n and the test task set T t e s t , the initial dataset is partitioned into two non-overlapping subsets: the training set D t r and the test set D t e , such that D t r D t e = . To ensure methodological rigor and experimental consistency, the categories in the D t r and D t e remain fixed throughout the experiment.
For a five-way classification task, the test set D t e includes the following categories: BTR60, BRDM2, T72, 2S1, and D7. The training set D t r consists of the remaining five categories. In a three-way classification scenario, the test set D t e comprises BTR60, BRDM2, and T72, while the training set D t r includes the remaining seven categories. Table 2 summarizes the dataset categorization under different experimental configurations.

4.2. N-Way K-Shot Task

In FSL, the concept of N-way K-shot is essential for assessing a model’s ability to generalize and adapt to novel target classes with limited labeled samples. Here, N-way specifies the amount of target classes, while K-shot indicates the amount of labeled samples available for each category.
In this study, we consider two main experimental configurations: three-way K-shot tasks and five-way K-shot tasks. For each configuration, the model’s performance is systematically evaluated under varying K-shot settings, including one-, five-, and ten-shot. The N-way configurations involve either three or five distinct target classes, and the K-shot settings determine the training samples per category. These diverse experimental setups provide a comprehensive evaluation of the proposed method.

4.3. Implementation Details

To achieve the experimental objectives, hyperparameters have been carefully selected. Specifically, a batch size of eight is used, with a learning rate of 0.01 for five-way K-shot tasks and 0.001 for three-way classification tasks. To prevent overfitting, the training process is capped at a maximum of 35,000 iterations. Additionally, an early-stopping mechanism is employed, which halts training if the loss function does not decrease for 20 consecutive iterations, thereby mitigating the risk of overfitting. The experiments are implemented using the PyTorch framework (version 2.4.1) with Python 3.8.
In FSL, the episode-based training method is widely employed, as extensively discussed in prior works [48,49,50]. This approach simulates the testing phase during training by aligning the configurations of training tasks T t r a i n and test tasks T t e s t , both adhering to specific N-way K-shot setups. The detailed procedure for episode training in ADMM-GCN is presented in Algorithm 2.
Algorithm 2 Episode Training for ADMM-GCN
 1:
Input: Training dataset D t r a i n , Number of ways N, Number of shots K, Number of episodes P
 2:
Output: Trained model
 3:
for  p 1 to P do
 4:
   Randomly select N categories from D train
 5:
   Initialize empty support set S and query set Q
 6:
   for each category C in the selected N categories do
 7:
     Randomly select K samples from C and add to S
 8:
     Randomly select 1 sample from C and add to Q
 9:
   end for
10:
   while not converged do
11:
     Train the model using the support set S
12:
     Compute the loss L on S
13:
     Backpropagate gradients and update model parameters Θ
14:
     Evaluate the model on the query set Q
15:
     Check convergence criteria
16:
   end while
17:
end for
18:
Return: Trained model
The fundamental principle of episode training lies in organizing training data into discrete episodes, each consisting of two key components: a support set S and a query set Q . The support set S contains a limited number of classes and serves as the foundation for model learning, while the query set Q includes samples from the same classes and acts as the evaluation set. To construct T t r a i n , N random categories are first selected from the training set D t r . Then, K samples are randomly drawn from each of these categories to form the support set S . Additionally, one sample from the remaining data in each category is randomly chosen to form the query set Q . This process is repeated P times, resulting in P distinct tasks within the training set T t r a i n .

5. Discussion and Analysis

5.1. Comparison Experiments

The experimental results of ADMM-GCN in comparison with various FSL methods on the MSTAR dataset under different N-way K-shot settings are summarized in Table 3. To ensure a comprehensive and fair evaluation, we incorporate widely adopted methods in FSL and SAR ATR, including ProtoNet [50], RelationNet [51], TPN [52], MSAR [33], DeepEMD [53], and BSCapNet [54].
In the three-way classification task, ADMM-GCN achieves the highest accuracy across all K-shot settings. Specifically, in the one-shot scenario, ADMM-GCN attains an accuracy of 84.31% ± 0.39, significantly outperforming ProtoNet (71.24% ± 0.45) and RelationNet (75.32% ± 0.49). Furthermore, ADMM-GCN surpasses advanced methods such as DeepEMD (76.01% ± 0.42) and BSCapNet (73.01% ± 0.47), demonstrating its superior feature extraction and classification capability under extreme sample scarcity. As the number of shots increases, all methods exhibit improved classification accuracy. In the 10-shot setting, ADMM-GCN achieves 92.18% ± 0.38, outperforming all other baselines, including TPN (88.93% ± 0.42) and RelationNet (86.76% ± 0.35).
The five-way classification task poses greater challenges due to an increased number of classes, leading to lower overall accuracy compared to the three-way task. Despite this, ADMM-GCN consistently outperforms existing methods. In the one-shot setting, ADMM-GCN achieves 61.79% ± 0.56, surpassing ProtoNet (50.42% ± 0.89) and RelationNet (53.81% ± 0.91). The performance gap remains noticeable in the 10-shot setting, where ADMM-GCN attains 74.01% ± 0.53, compared to TPN (70.37% ± 0.73) and DeepEMD (69.66% ± 0.60). The results suggest that ADMM-GCN maintains robust classification performance in complex scenarios with a higher number of categories. The model’s superior feature learning capability enables it to extract more discriminative representations, even under conditions with limited training samples. Moreover, the inclusion of confidence intervals in the results ensures statistical reliability, further validating the effectiveness of ADMM-GCN in few-shot SAR ATR.

5.2. Performance Assessment Under Varying Conditions

To evaluate the performance of ADMM-GCN in real-world SAR ATR scenarios, we conduct additional experiments under varying conditions, including noise injection, random cropping, and rotation operations. These perturbations simulate common challenges encountered in practical SAR target recognition, such as sensor-induced noise, partial occlusions due to environmental obstacles, and variations in target orientation.
Noise is a common issue in SAR imaging, often resulting from electronic interference, clutter, and sensor limitations. To replicate these effects, Gaussian noise is injected into SAR images, introducing pixel-level distortions to evaluate the model’s ability to extract discriminative features despite signal degradation. In addition to noise, real-world SAR images frequently suffer from partial occlusion due to obstacles such as buildings and vegetation. To simulate this, random cropping is applied, removing portions of the target object and assessing the model’s ability to generalize with incomplete feature representations. Furthermore, SAR target images are often captured from varying aspect angles due to changes in sensor positions and target orientations. To evaluate the model’s robustness to such variations, random rotations are applied, evaluating whether ADMM-GCN can maintain classification accuracy across different orientations.
The experimental results, summarized in Table 4, indicate that ADMM-GCN maintains stable classification performance under various perturbations, with a modest decline in accuracy observed under more challenging conditions. In the table, checkmarks (✓) indicate that the corresponding perturbation was applied. The introduction of noise injection slightly reduces accuracy, reflecting the increased difficulty in feature extraction due to signal distortions. For instance, in the three-way one-shot setting, accuracy decreases from 84.31% to 81.38%, and in the five-way one-shot setting, it drops from 61.79% to 60.12%. Random cropping also leads to minor performance degradation, suggesting that ADMM-GCN can still effectively classify targets even when partial information is missing. In the three-way one-shot scenario, accuracy drops to 80.32%, and in the five-way one-shot scenario, it remains relatively stable at 61.46%. Rotation perturbations introduce variations in accuracy, highlighting the impact of aspect angle differences on SAR image representation. The accuracy under this setting is 82.32% for three-way one-shot and 58.44% for five-way one-shot, indicating that angular variations can affect feature representation but the model still generalizes well. When all three perturbations are applied simultaneously, the model exhibits a moderate performance drop, yet it still maintains acceptable classification accuracy. The most significant decrease is observed in the five-way one-shot scenario, where accuracy drops to 56.98%, while in the three-way 5-shot setting, it stabilizes at 89.58%. Despite these variations in environmental conditions, ADMM-GCN maintains acceptable classification performance, demonstrating its potential for deployment in operational environments where noisy, partially obscured, or variably oriented targets are encountered.

5.3. Ablation Study

This section conducts a series of experiments to validate the effectiveness of each essential part of the suggested ADMM-GCN. The configuration settings listed in Table 2 are used for all ablation assessments.

5.3.1. Effectiveness Assessment of the EM

First, experiments are conducted to assess the effectiveness of the EM within the proposed ADMM-GCN framework. The experimental results are demonstrated in Table 5, where “w” indicates that the EM is included, while “w/o” indicates that the EM is excluded.
As shown in Table 5, the classification accuracy of ADMM-GCN with the EM is consistently higher than that of the model without the EM. These results demonstrate that the EM effectively extracts embedding features, significantly enhancing the performance of few-shot SAR target classification.

5.3.2. Effectiveness Assessment of the GCM

Next, experiments are carried out to verify the effectiveness of the GCM in enhancing the performance. Table 6 shows the corresponding experimental results, where “w” indicates that the GCM is included, while “w/o” denotes that it is excluded.
As shown in Table 6, the ADMM-GCN, which incorporates the GCM, consistently outperforms the model without the GCM across six different experimental settings. Notably, in the five-way one-shot task, the classification accuracy improves by approximately 5% when the GCM is employed. These experimental results demonstrate that the GCM effectively enhances the model’s classification performance. Specifically, the graph structure and convolutional operations improve the discriminative ability of the extracted features and strengthen the overall feature representation by constructing a relational graph among features. This, in turn, significantly boosts the few-shot SAR target classification performance.

5.3.3. Impact Evaluation of Mixed Regularized Loss

Several experiments were conducted to evaluate the impact of the mixed regularized loss. The results, as shown in Table 7 and Table 8, and Figure 8, highlight the performance of models in three-way and five-way few-shot SAR target classification tasks. These tables and figures provide comprehensive insights by presenting average accuracy and standard deviations, capturing both the precision and stability of the models.
In the three-way K-shot tasks, the mixed regularized term significantly enhances both accuracy and stability. For instance, in the one-shot task, the accuracy increases from 80.30% to 84.31%, while the standard deviation decreases from 0.7684 to 0.5467. This demonstrates the mixed regularized term’s effectiveness in improving model generalization and reducing variability, thus ensuring more consistent performance across diverse settings.
Similarly, for five-way K-shot tasks, the introduction of the regularized term leads to consistent improvements. In the one-shot task, accuracy rises from 39.94% to 61.79%, with a corresponding decrease in standard deviation from 1.1498 to 0.7856. A similar trend is observed in 5-shot and 10-shot tasks, where the term contributes to both accuracy enhancements and variability reductions. By boosting accuracy and reducing fluctuations, the regularized term enables more reliable and stable performance, which is essential for practical applications of few-shot SAR target classification.
To further investigate the effectiveness of addressing overfitting challenges, we have conducted additional experiments comparing the proposed mixed regularized loss with two widely used regularization techniques: L1 regularization [55] and ElasticNet [56]. Table 9 presents the accuracy differences across various N-way K-shot settings.
For the three-way one-shot setting, the mixed regularized loss attains 84.31%, outperforming L1 regularization (67.06%) and ElasticNet (80.84%). Similarly, in the 10-shot setting, the mixed regularized loss achieves 92.18%, surpassing ElasticNet (91.44%) and L1 (85.84%).
For the five-way K-shot tasks, the improvements are also evident. The mixed regularized loss achieves 61.79% accuracy in the one-shot setting, compared to 57.96% and 58.04% for L1 and ElasticNet regularization, respectively. As the number of shots increases, the mixed regularized loss maintains its advantage, reaching 74.01% in the 10-shot setting, while L1 and ElasticNet achieve 67.64% and 66.12%, respectively.
These results suggest that the mixed regularized loss provides superior generalization and increased accuracy compared to L1 regularization and ElasticNet, while L1 regularization encourages sparsity, it can overly penalize model parameters in FSL scenarios, leading to suboptimal feature utilization. ElasticNet, which combines L1 and L2 regularization, improves performance but struggles to maintain model expressiveness. In contrast, the mixed regularized loss balances regularization and model complexity, effectively reducing overfitting while preserving informative features, making it suitable for few-shot SAR ATR tasks.

5.3.4. Performance of ADMM Optimizer

To comprehensively evaluate the effectiveness of the ADMM optimizer, we compare its performance with three widely used optimization methods: SGD [35], RMSprop [57], and Adam [34]. The experimental results, presented in Table 10 and Table 11, as well as Figure 9, demonstrate the accuracy improvements of the ADMM optimizer across different task settings.
Table 10 presents the results for three-way K-shot tasks. For the one-, five-, and ten-shot settings, the ADMM optimizer achieves accuracies of 84.31%, 89.70%, and 92.18%, respectively. In comparison, the Adam optimizer attains 82.98%, 88.46%, and 90.62%. The RMSprop and SGD optimizers exhibit lower performance, with SGD achieving 75.54% in the one-shot setting and RMSprop reaching 87.42% in the 10-shot setting. Among all tested optimizers, ADMM consistently yields the highest accuracy, with relative improvements of 1.33%, 1.24%, and 1.56% over Adam. The performance gap is even more pronounced when compared to RMSprop and SGD.
Similarly, Table 11 presents the results for five-way K-shot tasks. In the one-, five-, and ten-shot settings, the ADMM optimizer achieves accuracies of 61.67%, 68.75%, and 74.01%, respectively, outperforming the Adam optimizer, which attains 60.73%, 67.19%, and 72.67%. The accuracy improvements over Adam are 0.94%, 1.56%, and 1.34%, respectively. Additionally, RMSprop and SGD exhibit lower performance, with SGD achieving 54.92% in the one-shot setting and RMSprop reaching 65.28% in the ten-shot setting.
These trends are further illustrated in Figure 9, which compares the test accuracy achieved by the ADMM optimizer and three other optimizers (Adam, RMSprop, and SGD) across three-way and five-way K-shot tasks. Figure 9a shows the test accuracy for three-way K-shot tasks, while Figure 9b presents the results for five-way K-shot tasks. As shown in Figure 9, the ADMM optimizer (red solid line) consistently outperforms all baseline optimizers across different shot numbers. In the three-way one-shot task, ADMM achieves 84.31%, while Adam attains 82.98%, RMSprop reaches 76.66%, and SGD performs the worst at 75.54%. Similarly, in the five-way 10-shot setting, ADMM achieves the highest accuracy of 74.01%, surpassing Adam (72.67%), RMSprop (65.28%), and SGD (61.40%). These visualizations further validate the effectiveness of the ADMM optimizer in few-shot SAR target recognition, demonstrating its capability to achieve higher accuracy and enhance model generalization across different task configurations.

5.4. Hyperparameter λ Analysis

λ is a key parameter that balances the strength of regularization in the proposed method. To evaluate its impact, a series of experiments were conducted across a wide range of λ values for both three-way and five-way tasks, under one-shot, five-shot, and ten-shot settings.
Figure 10 presents the classification accuracy for different λ values in N-way K-shot tasks. In each subplot, the horizontal axis represents the value of λ , while the vertical axis denotes the average classification accuracy (%). For instance, in the three-way one-shot task (subplot a), the classification accuracy remains stable across a wide range of λ values, from 1 × 10 6 to 1 × 10 5 . However, a slight decline in performance is observed when λ increases to 1 × 10 4 , where accuracy decreases from approximately 85% to 82%. The observed trend in the three-way one-shot task is consistent across all other settings. Such consistent behavior across diverse N-way K-shot tasks indicates the method’s robustness to a wide range of λ settings.

6. Conclusions

In this paper, we propose ADMM-GCN, a novel graph-based framework for few-shot SAR target recognition, designed to address the critical challenge of data scarcity in SAR ATR. By leveraging the learning capabilities of GCNs, ADMM-GCN proficiently extracts both global and local features from few-shot SAR samples. To further enhance performance in FSL scenarios, this study designs a mixed regularized loss function to mitigate the risk of overfitting during training. Additionally, an ADMM-based algorithm is developed to address the optimization model. Extensive experiments conducted on the MSTAR dataset validate the effectiveness of ADMM-GCN, which achieves superior performance across all evaluated settings. Notably, the method achieves an accuracy of 92.18% on the challenging three-way 10-shot task, outperforming benchmarks by 3.25%. While ADMM-GCN offers advantages, its optimization process involves iterative updates, leading to a modest increase in computational cost.
To further improve efficiency, future research will explore several strategies. One promising direction is to develop a hybrid ADMM–Adam optimizer for few-shot SAR target recognition, leveraging ADMM’s constrained optimization capabilities alongside Adam’s fast convergence properties to improve computational efficiency while maintaining robust optimization. Additionally, exploring lightweight variants of ADMM-GCN could facilitate its deployment in real-world SAR processing systems with constrained computational resources.

Author Contributions

Conceptualization, J.J., Z.X. and F.W.; methodology, J.J. and Z.X.; software, J.J., Z.X. and N.Z.; formal analysis, J.J. and Z.X.; writing—original draft preparation, J.J., Z.X. and F.W.; writing—review and editing, J.J. and F.W.; supervision, F.W.; project administration, F.W.; funding acquisition, F.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded in part by the Natural Science Foundation of China under Grant 61991421.

Data Availability Statement

The SAR image dataset used in this paper is the MSTAR dataset, which can be found at https://www.sdms.afrl.af.mil/index.php?collection=mstar, accessed on 1 January 2024.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ADMMAlternating Direction Method Of Multipliers
ATRAutomatic Target Recognition
CNNConvolutional Neural Network
EOElectro-Optical
EMEmbedding Module
FSLFew-Shot Learning
GCMGraph Convolutional Module
GCNGraph Convolutional Network
MLPMultilayer Perceptron
MSTARMoving And Stationary Target Acquisition And Recognition
SARSynthetic Aperture Radar
SGDStochastic Gradient Descent
SWDSliced Wasserstein Distance
TPNTransductive Propagation Network
ZSLZero-Shot Learning

References

  1. Moreira, A.; Prats-Iraola, P.; Younis, M.; Krieger, G.; Hajnsek, I.; Papathanassiou, K.P. A tutorial on synthetic aperture radar. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–43. [Google Scholar]
  2. Chen, S.; Wang, H.; Xu, F.; Jin, Y.Q. Target classification using the deep convolutional networks for SAR images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4806–4817. [Google Scholar]
  3. Zhao, Q.; Principe, J.C. Support vector machines for SAR automatic target recognition. IEEE Trans. Aerosp. Electron. Syst. 2001, 37, 643–654. [Google Scholar]
  4. Ding, J.; Chen, B.; Liu, H.; Huang, M. Convolutional neural network with data augmentation for SAR target recognition. IEEE Geosci. Remote Sens. Lett. 2016, 13, 364–368. [Google Scholar]
  5. Cho, J.H.; Park, C.G. Multiple feature aggregation using convolutional neural networks for SAR image-based automatic target recognition. IEEE Geosci. Remote Sens. Lett. 2018, 15, 1882–1886. [Google Scholar]
  6. Novak, L.M.; Owirka, G.J.; Brower, W.S. Performance of 10-and 20-target MSE classifiers. IEEE Trans. Aerosp. Electron. Syst. 2000, 36, 1279–1289. [Google Scholar]
  7. Sun, Y.; Liu, Z.; Todorovic, S.; Li, J. Adaptive boosting for SAR automatic target recognition. IEEE Trans. Aerosp. Electron. Syst. 2007, 43, 112–125. [Google Scholar]
  8. Dong, G.; Kuang, G.; Wang, N.; Wang, W. Classification via sparse representation of steerable wavelet frames on Grassmann manifold: Application to target recognition in SAR image. IEEE Trans. Image Process. 2017, 26, 2892–2904. [Google Scholar]
  9. Dong, G.; Kuang, G.; Wang, N.; Zhao, L.; Lu, J. SAR target recognition via joint sparse representation of monogenic signal. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 3316–3328. [Google Scholar]
  10. Novak, L.M.; Owirka, G.J.; Brower, W.S.; Weaver, A.L. The automatic target-recognition system in SAIP. Linc. Lab. J. 1997, 10, 187–202. [Google Scholar]
  11. Diemunsch, J.R.; Wissinger, J. Moving and stationary target acquisition and recognition (MSTAR) model-based automatic target recognition: Search technology for a robust ATR. In Algorithms for Synthetic Aperture Radar Imagery V, Proceedings of the Aerospace/Defense Sensing and Controls 1998, Orlando, FL, USA, 13–17 April 1998; SPIE: Bellingham, WA, USA, 1998; Volume 3370, pp. 481–492. [Google Scholar]
  12. Zhu, X.X.; Montazeri, S.; Ali, M.; Hua, Y.; Wang, Y.; Mou, L.; Shi, Y.; Xu, F.; Bamler, R. Deep learning meets SAR: Concepts, models, pitfalls, and perspectives. IEEE Geosci. Remote Sens. Mag. 2021, 9, 143–172. [Google Scholar]
  13. Oveis, A.H.; Giusti, E.; Ghio, S.; Martorella, M. A survey on the applications of convolutional neural networks for synthetic aperture radar: Recent advances. IEEE Aerosp. Electron. Syst. Mag. 2021, 37, 18–42. [Google Scholar]
  14. Khan, M.; Saddik, A.E.; Deriche, M.; Gueaieb, W. STT-Net: Simplified Temporal Transformer for Emotion Recognition. IEEE Access 2024, 12, 86220–86231. [Google Scholar] [CrossRef]
  15. Khan, M.; Khan, U.; Othmani, A. PD-Net: Multi-Stream Hybrid Healthcare System for Parkinson’s Disease Detection using Multi Learning Trick Approach. In Proceedings of the 2023 IEEE 36th International Symposium on Computer-Based Medical Systems (CBMS), L’Aquila, Italy, 22–24 June 2023; pp. 382–385. [Google Scholar] [CrossRef]
  16. Khan, M.; Gueaieb, W.; Saddik, A.E.; De Masi, G.; Karray, F. An Efficient Violence Detection Approach for Smart Cities Surveillance System. In Proceedings of the 2023 IEEE International Smart Cities Conference (ISC2), Bucharest, Romania, 24–27 September 2023; pp. 1–5. [Google Scholar] [CrossRef]
  17. Bai, X.; Xue, R.; Wang, L.; Zhou, F. Sequence SAR image classification based on bidirectional convolution-recurrent network. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9223–9235. [Google Scholar]
  18. Zhou, F.; Wang, L.; Bai, X.; Hui, Y. SAR ATR of ground vehicles based on LM-BN-CNN. IEEE Trans. Geosci. Remote Sens. 2018, 56, 7282–7293. [Google Scholar]
  19. Huang, Z.; Pan, Z.; Lei, B. What, where, and how to transfer in SAR target recognition based on deep CNNs. IEEE Trans. Geosci. Remote Sens. 2019, 58, 2324–2336. [Google Scholar]
  20. Wang, P.; Sun, X.; Diao, W.; Fu, K. FMSSD: Feature-merged single-shot detection for multiscale objects in large-scale remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2019, 58, 3377–3390. [Google Scholar]
  21. Sun, X.; Liu, Y.; Yan, Z.; Wang, P.; Diao, W.; Fu, K. SRAF-Net: Shape robust anchor-free network for garbage dumps in remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2020, 59, 6154–6168. [Google Scholar]
  22. He, Q.; Sun, X.; Yan, Z.; Fu, K. DABNet: Deformable contextual and boundary-weighted network for cloud detection in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5601216. [Google Scholar]
  23. Wang, Y.; Yao, Q.; Kwok, J.T.; Ni, L.M. Generalizing from a few examples: A survey on few-shot learning. ACM Comput. Surv. (CSUR) 2020, 53, 1–34. [Google Scholar]
  24. Sun, X.; Wang, B.; Wang, Z.; Li, H.; Li, H.; Fu, K. Research progress on few-shot learning for remote sensing image interpretation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 2387–2402. [Google Scholar] [CrossRef]
  25. Du, K.; Deng, Y.; Wang, R.; Zhao, T.; Li, N. SAR ATR based on displacement-and rotation-insensitive CNN. Remote Sens. Lett. 2016, 7, 895–904. [Google Scholar] [CrossRef]
  26. Wagner, S.A. SAR ATR by a combination of convolutional neural network and support vector machines. IEEE Trans. Aerosp. Electron. Syst. 2016, 52, 2861–2872. [Google Scholar]
  27. Song, Q.; Xu, F.; Zhu, X.X.; Jin, Y.Q. Learning to generate SAR images with adversarial autoencoder. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5210015. [Google Scholar]
  28. Sun, Y.; Wang, Y.; Liu, H.; Wang, N.; Wang, J. SAR target recognition with limited training data based on angular rotation generative network. IEEE Geosci. Remote Sens. Lett. 2019, 17, 1928–1932. [Google Scholar]
  29. Rostami, M.; Kolouri, S.; Eaton, E.; Kim, K. Deep transfer learning for few-shot SAR image classification. Remote Sens. 2019, 11, 1374. [Google Scholar] [CrossRef]
  30. Tai, Y.; Tan, Y.; Xiong, S.; Sun, Z.; Tian, J. Few-shot transfer learning for sar image classification without extra sar samples. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 2240–2253. [Google Scholar]
  31. Wang, S.; Wang, Y.; Liu, H.; Sun, Y. Attribute-guided multi-scale prototypical network for few-shot SAR target classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 12224–12245. [Google Scholar]
  32. Ren, H.; Liu, S.; Yu, X.; Zou, L.; Zhou, Y.; Wang, X.; Tang, H. Transductive Prototypical Attention Reasoning Network for Few-shot SAR Target Recognition. IEEE Trans. Geosci. Remote. Sens. 2023, 61, 5206813. [Google Scholar] [CrossRef]
  33. Fu, K.; Zhang, T.; Zhang, Y.; Wang, Z.; Sun, X. Few-shot SAR target classification via metalearning. IEEE Trans. Geosci. Remote Sens. 2021, 60, 2000314. [Google Scholar] [CrossRef]
  34. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  35. Robbins, H.; Monro, S. A Stochastic Approximation Method. Ann. Math. Stat. 1951, 22, 400–407. [Google Scholar]
  36. Boyd, S.; Parikh, N.; Chu, E.; Peleato, B.; Eckstein, J. Distributed optimization and statistical learning via the alternating direction method of multipliers. Found. Trends® Mach. Learn. 2011, 3, 1–122. [Google Scholar]
  37. Wang, J.; Yu, F.; Chen, X.; Zhao, L. ADMM for Efficient Deep Learning with Global Convergence. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage AK, USA, 4–8 August 2019; ACM: New York, NY, USA, 2019. [Google Scholar] [CrossRef]
  38. Yang, Y.; Sun, J.; Li, H.; Xu, Z. ADMM-CSNet: A Deep Learning Approach for Image Compressive Sensing. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 521–538. [Google Scholar] [CrossRef] [PubMed]
  39. Lin, C.H.; Lin, Y.C.; Tang, P.W. ADMM-ADAM: A New Inverse Imaging Framework Blending the Advantages of Convex Optimization and Deep Learning. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5514616. [Google Scholar] [CrossRef]
  40. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  41. Zhang, Z.; Cui, P.; Zhu, W. Deep learning on graphs: A survey. IEEE Trans. Knowl. Data Eng. 2020, 34, 249–270. [Google Scholar]
  42. Hong, M.; Luo, Z.Q.; Razaviyayn, M. Convergence analysis of alternating direction method of multipliers for a family of nonconvex problems. SIAM J. Optim. 2016, 26, 337–364. [Google Scholar] [CrossRef]
  43. Wang, Y.; Yin, W.; Zeng, J. Global convergence of ADMM in nonconvex nonsmooth optimization. J. Sci. Comput. 2019, 78, 29–63. [Google Scholar]
  44. Cascarano, P.; Calatroni, L.; Piccolomini, E.L. On the inverse Potts functional for single-image super-resolution problems. arXiv 2020, arXiv:2008.08470. [Google Scholar]
  45. Maclaurin, D.; Duvenaud, D.; Adams, R.P. Autograd: Effortless gradients in numpy. In Proceedings of the ICML 2015 AutoML Workshop, Lille, France, 11 July 2015; Volume 238. [Google Scholar]
  46. Xu, Z.; Luo, Y.; Wu, B.; Meng, D. S2S-WTV: Seismic Data Noise Attenuation Using Weighted Total Variation Regularized Self-Supervised Learning. IEEE Trans. Geosci. Remote. Sens. 2023, 61, 5908315. [Google Scholar]
  47. Defense Advanced Research Project Agency (DARPA); Air Force Research Laboratory (AFRL). The Air Force Moving and Stationary Target Recognition (MSTAR) Database. 2014. Available online: https://www.sdms.afrl.af.mil/index.php?collection=mstar (accessed on 1 January 2024).
  48. Koch, G.; Zemel, R.; Salakhutdinov, R. Siamese neural networks for one-shot image recognition. In Proceedings of the ICML Deep Learning Workshop, Lille, France, 6–11 July 2015; Volume 2. [Google Scholar]
  49. Vinyals, O.; Blundell, C.; Lillicrap, T.; Wierstra, D. Matching networks for one shot learning. In Proceedings of the Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, Barcelona, Spain, 5–10 December 2016; Volume 29. [Google Scholar]
  50. Snell, J.; Swersky, K.; Zemel, R. Prototypical networks for few-shot learning. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  51. Sung, F.; Yang, Y.; Zhang, L.; Xiang, T.; Torr, P.H.; Hospedales, T.M. Learning to compare: Relation network for few-shot learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1199–1208. [Google Scholar]
  52. Liu, Y.; Lee, J.; Park, M.; Kim, S.; Yang, E.; Hwang, S.J.; Yang, Y. Learning to propagate labels: Transductive propagation network for few-shot learning. arXiv 2018, arXiv:1805.10002. [Google Scholar]
  53. Zhang, C.; Cai, Y.; Lin, G.; Shen, C. DeepEMD: Few-shot image classification with differentiable earth mover’s distance and structured classifiers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12203–12213. [Google Scholar]
  54. Liu, S.; Yu, X.; Ren, H.; Zou, L.; Zhou, Y.; Wang, X. Bi-similarity prototypical network with capsule-based embedding for few-shot sar target recognition. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; IEEE: New York, NY, USA, 2022; pp. 1015–1018. [Google Scholar]
  55. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B Stat. Methodol. 1996, 58, 267–288. [Google Scholar]
  56. Zou, H.; Hastie, T. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. B Stat. Methodol. 2005, 67, 301–320. [Google Scholar]
  57. Hinton, G.; Srivastava, N.; Swersky, K. Neural networks for machine learning lecture 6a overview of mini-batch gradient descent. Cited 2012, 14, 2. [Google Scholar]
Figure 1. Illustration of a few-shot SAR target recognition task example. The configuration represents a 5-way 1-shot scenario, where the support set provides one labeled sample per class and the query set includes samples for prediction. Each black box represents an individual task, with the blue-bordered samples in the query set corresponding to the same classes in the support set. The figure includes both training tasks (top) and test tasks (bottom).
Figure 1. Illustration of a few-shot SAR target recognition task example. The configuration represents a 5-way 1-shot scenario, where the support set provides one labeled sample per class and the query set includes samples for prediction. Each black box represents an individual task, with the blue-bordered samples in the query set corresponding to the same classes in the support set. The figure includes both training tasks (top) and test tasks (bottom).
Remotesensing 17 01179 g001
Figure 2. Visualization of the ADMM-GCN framework for few-shot SAR target classification. The framework processes SAR data through graph construction, where node colors indicate different target categories, followed by iterative node and edge feature updates for feature fusion and refinement.
Figure 2. Visualization of the ADMM-GCN framework for few-shot SAR target classification. The framework processes SAR data through graph construction, where node colors indicate different target categories, followed by iterative node and edge feature updates for feature fusion and refinement.
Remotesensing 17 01179 g002
Figure 3. Overview of the proposed ADMM-GCN framework.
Figure 3. Overview of the proposed ADMM-GCN framework.
Remotesensing 17 01179 g003
Figure 4. Structure of the EM.
Figure 4. Structure of the EM.
Remotesensing 17 01179 g004
Figure 5. Illustration of feature aggregation and edge information interaction in a GCN. Dashed lines represent initially unconnected edges, while solid lines represent feature-interacting edges during aggregation. Blue nodes represent the initial stages of feature aggregation, while red nodes indicate the outcomes after iterative feature updates and edge information fusion.
Figure 5. Illustration of feature aggregation and edge information interaction in a GCN. Dashed lines represent initially unconnected edges, while solid lines represent feature-interacting edges during aggregation. Blue nodes represent the initial stages of feature aggregation, while red nodes indicate the outcomes after iterative feature updates and edge information fusion.
Remotesensing 17 01179 g005
Figure 6. Graph representation with node and edge attributes.
Figure 6. Graph representation with node and edge attributes.
Remotesensing 17 01179 g006
Figure 7. Examples of optical and SAR images of military vehicles from the MSTAR dataset. The first row displays optical images for each category, while the second row contains their corresponding SAR images. (a) ZSU234. (b) ZIL131. (c) T72. (d) T62. (e) D7. (f) BTR70. (g) BTR60. (h) BRDM2. (i) BMP2. (j) 2S1.
Figure 7. Examples of optical and SAR images of military vehicles from the MSTAR dataset. The first row displays optical images for each category, while the second row contains their corresponding SAR images. (a) ZSU234. (b) ZIL131. (c) T72. (d) T62. (e) D7. (f) BTR70. (g) BTR60. (h) BRDM2. (i) BMP2. (j) 2S1.
Remotesensing 17 01179 g007
Figure 8. Comparison of test accuracy for models with normal loss and mixed regularized loss on (a) 3-way K-shot tasks and (b) 5-way K-shot tasks. Results are reported for 1-shot, 5-shot, and 10-shot settings.
Figure 8. Comparison of test accuracy for models with normal loss and mixed regularized loss on (a) 3-way K-shot tasks and (b) 5-way K-shot tasks. Results are reported for 1-shot, 5-shot, and 10-shot settings.
Remotesensing 17 01179 g008
Figure 9. Comparison of test accuracy between the ADMM optimizer and other optimizers for (a) 3-way K-shot tasks and (b) 5-way K-shot tasks. Results are reported for 1-shot, 5-shot, and 10-shot settings.
Figure 9. Comparison of test accuracy between the ADMM optimizer and other optimizers for (a) 3-way K-shot tasks and (b) 5-way K-shot tasks. Results are reported for 1-shot, 5-shot, and 10-shot settings.
Remotesensing 17 01179 g009
Figure 10. Classification accuracy of the proposed method with different values of λ for N-way K-shot tasks. Subplots represent the (a) 3-way 1-shot task, (b) 3-way 5-shot task, (c) 3-way 10-shot task, (d) 5-way 1-shot task, (e) 5-way 5-shot task, and (f) 5-way 10-shot task.
Figure 10. Classification accuracy of the proposed method with different values of λ for N-way K-shot tasks. Subplots represent the (a) 3-way 1-shot task, (b) 3-way 5-shot task, (c) 3-way 10-shot task, (d) 5-way 1-shot task, (e) 5-way 5-shot task, and (f) 5-way 10-shot task.
Remotesensing 17 01179 g010
Table 1. Numbers of samples for each target category in the MSTAR dataset.
Table 1. Numbers of samples for each target category in the MSTAR dataset.
Target CategoryBRDM2BMP2BTR60BTR70D7T62T72ZIL131ZSU2342S1
Number572428451429573572428573573573
Table 2. Categorization of the MSTAR dataset into disjoint training ( D t r ) and test ( D t e ) datasets under different N-way K-shot experimental settings.
Table 2. Categorization of the MSTAR dataset into disjoint training ( D t r ) and test ( D t e ) datasets under different N-way K-shot experimental settings.
N-Way K-Shot Task D tr Categories D te Categories
5-way K-shotZIL131, BMP2, T62, BTR70, ZSU234BTR60, BRDM2, T72, 2S1, D7
3-way K-shotD7, T62, 2S1, ZIL131, BMP2, ZSU234, BTR70BTR60, BRDM2, T72
Table 3. Accuracy (%) results of ADMM-GCN compared with other FSL methods on the MSTAR dataset across different N-way K-shot settings, with 95% confidence intervals.
Table 3. Accuracy (%) results of ADMM-GCN compared with other FSL methods on the MSTAR dataset across different N-way K-shot settings, with 95% confidence intervals.
Methods3-Way5-Way
1-Shot5-Shot10-Shot1-Shot5-Shot10-Shot
ProtoNet [50]71.24 ± 0.4580.79 ± 0.3882.37 ± 0.3350.42 ± 0.8963.74 ± 0.7867.95 ± 0.70
RelationNet [51]75.32 ± 0.4984.29 ± 0.4386.76 ± 0.3553.81 ± 0.9166.52 ± 0.8472.20 ± 0.62
TPN [52]80.45 ± 0.4887.32 ± 0.4688.93 ± 0.4257.44 ± 0.9265.70 ± 0.8370.37 ± 0.73
MSAR [33]69.23 ± 0.5184.71 ± 0.5987.96 ± 0.4953.50 ± 1.0060.50 ± 0.9064.72 ± 0.71
DeepEMD [53]76.01 ± 0.4283.23 ± 0.3986.24 ± 0.3455.61 ± 0.8265.17 ± 0.7569.66 ± 0.60
BSCapNet [54]73.01 ± 0.4786.62 ± 0.4884.60 ± 0.4264.81 ± 0.8767.50 ± 0.7973.55 ± 0.56
ADMM-GCN (ours)84.31 ± 0.3989.70 ± 0.3992.18 ± 0.3861.79 ± 0.5668.75 ± 0.5274.01 ± 0.53
Table 4. Experimental results of ADMM-GCN under varying conditions across different N-way K-shot settings (Accuracy, %).
Table 4. Experimental results of ADMM-GCN under varying conditions across different N-way K-shot settings (Accuracy, %).
Perturbation Settings3-Way5-Way
Noise InjectionRandom CroppingRotation1-Shot5-Shot10-Shot1-Shot5-Shot10-Shot
84.3189.7092.1861.7968.7574.01
81.3889.6290.2660.1266.0473.74
80.3288.7090.0261.4664.6273.80
82.3289.0690.5058.4465.8272.16
81.1089.5890.8856.9865.4671.10
Table 5. Ablation study results for assessing the effectiveness of the EM on 3-way and 5-way K-shot tasks (Accuracy, %).
Table 5. Ablation study results for assessing the effectiveness of the EM on 3-way and 5-way K-shot tasks (Accuracy, %).
N-WayEM SettingsK-Shot
1-Shot5-Shot10-Shot
3-wayw/o EM72.6175.4378.25
w EM84.31 (11.70 ↑)89.70 (14.27 ↑)92.18 (13.93 ↑)
5-wayw/o EM58.4868.6972.34
w EM61.79 (3.31 ↑)68.75 (0.06↑)74.01 (1.67 ↑)
Table 6. Ablation study results for assessing the effectiveness of the GCM on 3-way and 5-way K-shot tasks (Accuracy, %).
Table 6. Ablation study results for assessing the effectiveness of the GCM on 3-way and 5-way K-shot tasks (Accuracy, %).
N-WayGCM SettingsK-Shot
1-Shot5-Shot10-Shot
3-wayw/o GCM70.6387.9990.16
w GCM84.31 (13.68 ↑)89.70 (1.71 ↑)92.18 (2.02 ↑)
5-wayw/o GCM56.8565.6872.70
w GCM61.79 (4.94 ↑)68.75 (3.07 ↑)74.01 (1.31 ↑)
Table 7. Experimental results of 3-way K-shot tasks: comparison between classical loss and mixed regularized loss in terms of accuracy (%).
Table 7. Experimental results of 3-way K-shot tasks: comparison between classical loss and mixed regularized loss in terms of accuracy (%).
Configurations Accuracy (%)
Loss Settings K -Shot MinMaxMeanStandard Deviation
Classical Loss1-shot 78.6482.5780.300.7684
5-shot 81.4685.2683.780.7745
10-shot 82.0086.7984.400.7697
Mixed
Regularized Loss
1-shot 82.6786.7984.31 (4.01 ↑)0.5467 (0.2217 ↓)
5-shot 88.5090.8689.70 (5.92 ↑)0.5440 (0.2305 ↓)
10-shot 90.5093.4392.18 (7.78 ↑)0.5374 (0.2323 ↓)
Table 8. Experimental results of 5-way K-shot tasks: comparison between classical loss and mixed regularized loss in terms of accuracy (%).
Table 8. Experimental results of 5-way K-shot tasks: comparison between classical loss and mixed regularized loss in terms of accuracy (%).
Configurations Accuracy (%)
Loss Settings K -Shot MinMaxMeanStandard Deviation
Classical Loss1-shot 37.1443.0739.941.1498
5-shot 62.0766.5064.270.9465
10-shot 64.5665.8265.400.9453
Mixed
Regularized Loss
1-shot 59.1465.2161.79 (21.85 ↑)0.7856 (0.3642 ↓)
5-shot 65.2169.5768.75 (4.48 ↑)0.7198 (0.2267 ↓)
10-shot 71.4376.2974.01 (8.61 ↑)0.7384 (0.2069 ↓)
Table 9. Comparison of N-way K-shot accuracy (%) between the mixed regularized loss and other regularization techniques.
Table 9. Comparison of N-way K-shot accuracy (%) between the mixed regularized loss and other regularization techniques.
Regularization Settings3-Way5-Way
1-Shot5-Shot10-Shot1-Shot5-Shot10-Shot
L1 Regularization [55]67.0685.4885.8457.9663.1867.64
ElasticNet [56]80.8487.2891.4458.0467.4066.12
Mixed Regularized Loss84.3189.7092.1861.7968.7574.01
Table 10. Comparison of three-way K-shot accuracy (%) between ADMM optimizer and other optimizers.
Table 10. Comparison of three-way K-shot accuracy (%) between ADMM optimizer and other optimizers.
Optimizer SettingsK-Shot
1-Shot5-Shot10-Shot
SGD [35]75.5486.4686.76
RMSprop [57]76.6685.3287.42
Adam [34]82.9888.4690.62
ADMM (Ours)84.3189.7092.18
Table 11. Comparison of 5-way K-shot accuracy (%) between ADMM optimizer and other optimizers.
Table 11. Comparison of 5-way K-shot accuracy (%) between ADMM optimizer and other optimizers.
Optimizer SettingsK-Shot
1-Shot5-Shot10-Shot
SGD [35]54.9260.2661.40
RMSprop [57]57.3463.3465.28
Adam [34]60.7367.1972.67
ADMM (Ours)61.6768.7574.01
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jin, J.; Xu, Z.; Zheng, N.; Wang, F. Graph-Based Few-Shot Learning for Synthetic Aperture Radar Automatic Target Recognition with Alternating Direction Method of Multipliers. Remote Sens. 2025, 17, 1179. https://doi.org/10.3390/rs17071179

AMA Style

Jin J, Xu Z, Zheng N, Wang F. Graph-Based Few-Shot Learning for Synthetic Aperture Radar Automatic Target Recognition with Alternating Direction Method of Multipliers. Remote Sensing. 2025; 17(7):1179. https://doi.org/10.3390/rs17071179

Chicago/Turabian Style

Jin, Jing, Zitai Xu, Nairong Zheng, and Feng Wang. 2025. "Graph-Based Few-Shot Learning for Synthetic Aperture Radar Automatic Target Recognition with Alternating Direction Method of Multipliers" Remote Sensing 17, no. 7: 1179. https://doi.org/10.3390/rs17071179

APA Style

Jin, J., Xu, Z., Zheng, N., & Wang, F. (2025). Graph-Based Few-Shot Learning for Synthetic Aperture Radar Automatic Target Recognition with Alternating Direction Method of Multipliers. Remote Sensing, 17(7), 1179. https://doi.org/10.3390/rs17071179

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop