Next Article in Journal
First Analyses of the TIMELINE AVHRR SST Product: Long-Term Trends of Sea Surface Temperature at 1 km Resolution across European Coastal Zones
Previous Article in Journal
Real-Time Three-Dimensional Tracking of Distant Moving Objects Using Non-Imaging Single-Pixel LiDAR
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Zero-Shot SAR Target Recognition Based on a Conditional Generative Network with Category Features from Simulated Images

College of Electronic Science and Technology, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(11), 1930; https://doi.org/10.3390/rs16111930
Submission received: 22 March 2024 / Revised: 17 May 2024 / Accepted: 24 May 2024 / Published: 27 May 2024

Abstract

:
SAR image target recognition relies heavily on a large number of annotated samples, making it difficult to classify the unseen class targets. Due to the lack of effective category auxiliary information, the current zero-shot target recognition methods for SAR images are limited to inferring only one unseen class rather than classifying multiple unseen classes. To address this issue, a conditional generative network with the category features from the simulated images for zero-shot SAR target recognition is proposed in this paper. Firstly, the deep features are extracted from the simulated images and fused into the category features that characterize the entire class. Then, a conditional VAE-GAN network is constructed to generate the feature instances of the unseen classes. The high-level semantic information shared in the category features aids in generalizing the mapping learned from the seen classes to the unseen classes. Finally, the generated features of the unseen classes are used to train a classifier that can classify the real unseen images. The classification accuracies for the targets of the three unseen classes based on the proposed method can reach 99.80 ± 1.22% and 71.57 ± 2.28% with the SAMPLE dataset and the MSTAR dataset, respectively. The advantage and validity of the proposed architecture are indicated with a small number of the seen classes and a small amount of the training data. Furthermore, the proposed method can be extended to generalized zero-shot recognition.

1. Introduction

Synthetic Aperture Radar (SAR) is a remote sensing technology that uses microwave signals to observe the Earth’s surface. Unlike optical remote sensing techniques, SAR has the advantage of being able to operate in all weather conditions and at all times of day [1]. SAR images are widely used in various fields due to their unique advantages. SAR Automatic Target Recognition (SAR-ATR) is a crucial component of SAR image processing techniques and plays a significant role in both civil and military applications. A standard SAR ATR system consists of three main components: detection, discrimination, and classification [2]. The classification of targets in SAR has become a challenging research issue.
SAR target classification methods can be categorized into the traditional and the deep learning-based methods. The traditional method comprises the template matching method [3,4], the model-based method [5], and the machine learning-based method [6]. Among these, the machine learning-based method exhibits a fast processing speed, a high recognition capability and a robust performance. However, it relies on a manually designed feature extractor, which requires specialized knowledge and experience. Compared to the traditional method, the deep learning-based approach has the advantage of automatically extracting features, which provides stronger generalization capabilities and enhances recognition performance. The use of neural networks for image classification has rapidly developed since the first application of the AlexNet network [7]. The convolutional neural network (CNN) models, such as VggNet [8] and ResNet [9], have achieved excellent performance in this field. Additionally, network architectures like recurrent neural networks (RNN) [10] and graphic neural networks (GNN) [11] can also play a unique role in the image recognition task. The deep learning methods have demonstrated an outstanding performance in SAR image recognition [12,13,14,15,16]. However, deep learning methods are data-driven and typically require a large amount of data for effective feature extraction. When the data are scarce, the recognition capability of the network significantly decreases. It is difficult to obtain large quantities of SAR images due to various limitations such as technological and policy constraints. Thus, the issue of the small sample size is a prominent concern in the field of SAR image target recognition. There are various methods to address the issue of the small sample size. Expanding the training dataset is a common approach [17,18,19,20,21]. On the other hand, the transfer learning-based methods [22,23,24] and meta-learning methods [25,26] are also effective strategies. While the methods mentioned above can address the issue of scarce data, they are unable to recognize the unseen class targets that lack any training data.
In real-world scenarios, it is often impossible to acquire SAR images of non-cooperative military facilities and other specific targets in advance for network training. Therefore, there is an urgent need for the ability to recognize the targets appearing for the first time. The issue of identifying the unseen classes with no training data is referred to as zero-shot recognition, which is an extreme case of the few-shot recognition problem. Figure 1 illustrates the learning approaches under the conditions of sufficient samples, limited samples, and zero samples. Four types of vehicle targets from the MSTAR SAR image dataset [27], 2S1, BMP2, BTR70, and T72, are used as examples. In cases where there are sufficient samples, a large amount of the labeled data for the 2S1 and BMP2 target classes can be used to train the network. However, in few-shot learning scenarios, the network can only be trained with a small number of samples. In zero-shot learning scenarios, it is impossible to obtain the data for the BTR70 and T72 target classes in advance for training. The network trained with the 2S1 and BMP2 requires to recognize the unseen classes such as BTR70 and T72. The concept of zero-shot learning was first proposed by Larochelle et al. [28], typically referring to training learning models based on the data from the seen classes and the prior auxiliary information, then the training enables the model to recognize the data from the unseen classes. The methods for solving zero-shot recognition include the direct semantic prediction methods [28,29], embedding model-based methods [30,31,32,33], and visual sample generation methods [34,35,36,37].
Although zero-shot recognition has been widely studied in the natural image domain, the research on zero-shot target recognition in SAR images is quite limited. Some studies [38,39,40] draw inspiration from the zero-shot recognition methods in the optical domain. The semantic information such as label one-hot encoding or attribute information is constructed to assist in zero-shot recognition for SAR images. During the training phase, the seen class data are used to construct an embedding space. During testing, the images from unseen classes are mapped to this embedding space, and their relationships are inferred by measuring the distances between embedding points. Reference [38] proposed a SAR image zero-shot recognition method to explore the relationships between the unseen class target T72 and the seven seen class targets based on a two-dimensional embedding space. Reference [39] proposed an architecture consisting of two autoencoders that utilize the reflection information of SAR images to assist in constructing the embedding space. Reference [40] utilized data mining techniques to annotate binary 10-dimensional attribute information for SAR targets and employed a classifier to assist in constructing the embedding space. The main challenge faced by the aforementioned methods is that the manually designed semantic information is too simple to effectively characterize SAR targets. The names of SAR targets, such as 2S1 and T72, are merely symbols composed of characters without actual semantic meaning. Defining effective class attribute information is also a challenging task. Because of the difficulty in assuring the quality of semantic information and the absence of supervised learning, the embeddings of unseen classes frequently encounter significant domain shift issues and become discrete. The embedding model-based method can only infer one unseen class at a time, rather than completing the recognition of multiple unseen class targets.
SAR electromagnetic simulated images generated from CAD models of targets have been used for SAR target recognition due to their provision of many target details similar to real images (the real images from the real measurement of a radar system) [23,41,42]. In certain studies, the networks are trained on fully simulated images and tested on real images [43,44,45,46]. However, the distributional differences between the simulated and real images limit the transferability of the model between these two types of images. The generative networks such as generative adversarial networks (GANs) [47] have been used to convert simulated data into pseudo-real data (the generated data approximating real images) that more closely resemble the distribution of real images in the case of sufficient or small samples [48,49]. The generated pseudo-real data can serve as more effective training data for training the classification network. However, these generation networks cannot generate data for unseen classes in a zero-shot setting because the real images of unseen classes cannot participate in training [46].
To classify the unseen targets with multiple classes, a conditional generative network with category features from simulated images for zero-shot SAR target recognition is proposed for generating pseudo-real data of unseen classes for supervised learning. The feature extraction network called CANet was first designed to extract deep features from both the simulated and real SAR images. The features from the simulated images are fused into the category features which serve as auxiliary information to represent the characteristics of the entire category. A conditional VAE-GAN network is then constructed. The network utilizes the category features of the seen classes as the conditions to learn the mapping relationship between these features and the deep features of the real SAR images. This mapping relationship can be extended to generate the features for the unseen classes. Ultimately, the generated features of the unseen classes are utilized to train a classifier for the classification of the real SAR images. The main contributions of the article are as follows:
  • The category features constructed from the simulated images are proposed. The feasibility of these category features utilized as the category auxiliary information for SAR zero-shot learning is verified.
  • A framework for zero-shot generation of SAR data based on the conditional VAE-GAN is proposed. The network establishes a connection between the seen and the unseen class data through the category features. By learning the mapping from the category features to the real data using the seen class data, it can generate the unseen class data.
  • Compared to the embedding model-based methods assisted by the semantic information, our architecture can recognize multiple unseen class SAR targets instead of inferring a single one.
The rest of this article is organized as follows. Section 2 provides a detailed overview of the method, including the complete architecture and the entire training and testing processes. Section 3 presents the experimental results of our method and further analyzes the impact of various factors on our approach, besides, a generalized zero-shot recognition extension experiment is conducted. In Section 4, the limitations of the proposed method and directions for future improvements are discussed. Finally, Section 5 summarizes the paper.

2. Methodology

Figure 2 shows the overall architecture of the proposed method, which comprises three parts: the feature extraction module (the red dashed box), the feature generation module (the green dashed box), and the classification module (the yellow dashed box). The feature extraction module is the first stage of the overall architecture. The real features and the category features are extracted using the pre-trained CANet in this stage. Then, the feature generation module is the second stage. The real features and the category features of seen classes extracted in the first stage are used to train the feature extraction network. The network learns the mapping from the category features to the real features. The last stage is the classification recognition module. The category features of unseen classes, extracted in the first stage, are used to generate the pseudo-real features of unseen classes using the learned mapping from the second stage. The generated features are used to train a classifier to recognize the real images. The specific modules of the proposed method and the detailed training and testing procedures are explained in this section. Section 2.1 provides the symbolic representation of zero-shot target recognition. Section 2.2, Section 2.3 and Section 2.4 introduce the feature extraction module, the feature generation model, and the classification module, respectively. Section 2.5 presents the detailed processes of the training and the testing.

2.1. Symbolic Representation of Zero-Shot Target Recognition

Suppose there are a total of k target categories, with m categories as the seen classes, and n categories as the unseen classes. The set of the seen class samples can be represented as S = x s , y s , a s x s X s , y s Y s , a s A s , where x s represents the samples from the seen classes, Y s = y 1 s , y 2 s , , y m s represents the labels for the seen classes, and A s = a 1 s , a 2 s , , a m s represents the auxiliary information. The set of the unseen class samples can be represented as U = y u , a u y u Y u , a u A u , where Y u = y 1 u , y 2 u , , y n u represents the labels for the unseen classes, and A u = a 1 u , a 2 u , , a n u represents the auxiliary information for the unseen classes. m + n = k , Y u Y s = . The zero-shot learning task in this paper involves training a network using S to learn the mapping A s X ˜ s , with X ˜ s being the generated pseudo-real data. The network can then achieve A u X ˜ u , with X ˜ u being the generated pseudo-real data of the unseen classes. Finally, a classifier f z s l : X ˜ u Y u can be trained using X ˜ u to classify the unseen class real images.

2.2. The Feature Extraction Module

Feature extraction is the first stage of the overall architecture, consisting of extracting the real features from the real images and the category features from the simulated images. The extracted features are then used to support the subsequent learning of the feature mappings.

2.2.1. Extraction of the Real Features

The feature extraction module called CANet designed in this paper represents an improvement over A-ConvNet [18] which replaces fully connected layers with sparsely connected convolutional structures. First, the CANet kept the first 4 layers of the A-ConvNet to preserve the feature extraction capability. Second, the additional fifth and sixth layer are added to extract features with a shape of 1 × 1 × 256 to facilitate the training of the subsequent feature generation network. The specific structure of CANet is illustrated in Figure 3. The network is composed of six layers. The first three layers consist of a combination of a convolution module, ReLU activation function and maxpooling module. After the fourth layer, a Dropout operation is applied. The design of the first four layers of the network is based on A-ConvNet. On this basis, the last two separate convolution layers have been added to extract effective features. The input image size is initially 128 × 128, and it is resized to 88 × 88 × 3 before being fed into the network. The feature size after the fourth layer is 3 × 3 × 128. The convolutional kernel size of the fifth layer is designed as 3 × 3 × 256, resulting in a 1 × 1 × 256 feature output of the network. The 256-dimensional feature reduces the complexity of subsequent model training while maintaining the representation capability for the SAR image target. The kernel size of the sixth layer is designed as 1 × 1 × n, enabling it to output the probabilities of n categories for the loss calculation and the gradient backpropagation.
The CANet network is pre-trained to extract the real features from the real images. Only the real images of the seen classes are used to train the network, as the real images of the unseen classes cannot participate in training. The real features of the seen classes are extracted for training in the next stage.

2.2.2. Extraction of Category Features

The existing research on generating real images from simulated images has primarily focused on one-to-one mapping [48,49]. In these works, the generative network maps each simulated image to a unique pseudo-real image corresponding to it, which limits the generation of the pseudo-real data. A conditional generative network [50,51] is used in this paper to achieve a one-to-many generation. Therefore, a category feature is constructed from the simulated images for each class and used as the condition to generate the multiple pseudo-real features.
The simulated images are used to extract the category feature. The extraction process is illustrated in Figure 4. First, CANet is pre-trained using real images of the seen classes. Then, for all simulated images in the same class, the pre-trained CANet is used to extract simulated features a i j . Finally, these simulated features are fused into the category feature a i ¯ . To ensure that a category feature accurately reflects the distinctive features of a given class, we define it as the average of all simulated features of that class. The approach is similar to the prototype network used in the few-shot learning [52]:
a ¯ i = 1 k i j = 1 k i a i j
where k i represents the number of the simulated images for the class i. The category features of the seen classes are extracted for training in the next stage. Meanwhile, the category features of the unseen classes are extracted to generate pseudo-real features for the unseen classes.

2.3. The Feature Generation Module

The learning of the feature generation is the second stage of the overall architecture. The goal is to learn the mapping relationship between the category features and the real features of the seen classes, allowing for better generalization to the generation of the unseen classes. A conditional generation network is created by integrating a VAE network and a GAN network. Additionally, a category feature reconstructor and a feedback module are integrated to enhance the generation capabilities of the network. The generation module learns the mappings at the deep feature level without focusing on the image reconstruction. Therefore, it is constructed only with shallow fully connected layers.
Figure 5 shows the composition of the feature generation module, where x is the real features and a represents the category features. The conditional variational autoencoder (CVAE) is formed by conditioning the encoder E ( x , a ) and generator G ( z , a ) on a, while the conditional generative adversarial network (CGAN) is formed by conditioning G ( z , a ) and the discriminator D ( x , x ˜ , a ) on a. The CVAE and CGAN share a generator, forming a conditional VAE-GAN (CVAE-GAN) [53]. The VAE-GAN combines the advantages of the latent space encoding of the VAE with the high-quality feature generation of the GAN, resulting in more stable training and the ability to generate more smoothly varying pseudo-real features.
For the feature generation part of the CVAE, we use the category feature a as a condition and concatenate the real feature x as the input for E ( x , a ) . Firstly, the encoder E ( x , a ) encodes the real feature into a low-dimensional vector z in the continuous latent space. Assuming z follows an isotropic Gaussian distribution, the output of E is the mean vector and variance vector of the Gaussian distribution, denoted as ( μ x , σ x ) . Then, the generator G decodes z and produces the fake feature x ˜ . Finally, ensuring the cyclic consistency of the generated features with the original features is achieved by minimizing the difference between x ˜ and x. CVAE is optimized through the following loss function:
V A E = K L + G K L = K L ( q ( z | x , a )     p ( z | a ) ) G = E q ( z x , a ) log p ( x ˜ z , a )
where the conditional distribution q z x , a represents the probability distribution modeled by E ( x , a ) , while p ( z | a ) is the prior distribution following N ( 0 , 1 ) . K L refers to the Kullback–Leibler divergence between q z x , a and p ( z a ) . p ( x ˜ z , a ) is equal to G ( z , a ) . G is the generation reconstruction loss of VAE, which is set as the cross-entropy loss between the generated feature and the original feature.
For the feature generation part of CGAN, the generator is identical to the generator G of CVAE. The discriminator D ( x , x ˜ , a ) takes the real feature, the fake feature and condition a as input, and outputs a real number indicating the authenticity of the input feature. CGAN is optimized through the improved WGAN loss [54]:
W G A N = E D ( x , a ) E D ( x ˜ , a ) λ E x ^ D x ^ , a 2 1 2
where x ˜ = G ( x , a ) , x ^ = a x + 1 a x ˜ , here a U ( 0 , 1 ) . The penalty term coefficient is represented by λ . The first two terms in the formula approximate the Wasserstein distance between the distribution of the generated features and the real features, and the last term is the gradient penalty term that forces the gradient at any point to be close to the unit norm. The improved WGAN loss enhances training smoothness and stability, mitigating model collapse issues during traditional GAN network training. The CVAE-GAN optimization objective combines the losses from both the CVAE and CGAN parts:
V A E G A N = min E , G max D V A E + α W G A N
where α is a hyperparameter, for specific details please refer to the literature [36]. The generation module introduces two new modules, the category feature reconstruction module ( C F R ) and the feedback module (F), based on VAE-GAN. These modules enhance the generative capabilities of the network. C F R reconstructs the generated fake features into the category features, learning the inverse mapping from category features to real features to ensure the semantic consistency of the generated features. The network is optimized through cycle consistency loss.
R = E a ˜ a 1
where a ˜ = C F R x ˜ . F provides feedback by feeding the intermediate layer embedding h from C F R into G. This allows G to iteratively improve the feature generation and obtain improved feature representations. In summary, the overall loss function of the zero-shot feature generation module is as follows:
T o t a l = V A E G A N + β R
where β is the weighting coefficient, with specific details please refer to in the literature [37]. After training the feature generation module, the network can generate pseudo-real features from category features of unseen classes. These features are then fed into the classification module to train the classifier.

2.4. The Classification Module

The third stage of the model involves the classification module training. The objective is to train the classifier using generated features of unseen classes, allowing it to classify real unseen class images. Figure 6 illustrates the structure of the classification module.
In the classification module, we retained C F R , rather than directly using the generated features x ˜ from unseen classes to train the classifier. C F R is the inverse mapping of G ( z , a ) . The intermediate layer embedding h and the reconstructed category feature a ˜ encode the complementary category information to the generated feature instances, serving as an auxiliary information source to assist in training the classifier. We concatenate x ˜ , a ˜ , and h as classification features c to train the classifier, which outputs the probability of unseen classes and is optimized through the cross-entropy loss:
C = 1 U ( c , y ) U log p y | c ; θ
here, θ represents the parameters of the classifier, and P ( y | c ; θ ) = exp θ y T c i n exp θ i T c denotes the probability of the output class. Once the classifier training is complete, the entire module is trained.

2.5. Training and Test Process

Figure 2 illustrates the complete training process of the proposed method. In the feature extraction stage, as shown in Section 2.2, the seen class features x s are extracted. Simultaneously, the category features a ¯ s and a ¯ u are extracted for both the seen and the unseen classes. x s , a ¯ s , and a ¯ u are all 256-dimensional features. During the feature generation training stage, x s is concatenated with the corresponding category feature a ¯ s and inputted into the encoder E. The encoder outputs a 256-dimensional mean and variance for the latent code z. Then, z is concatenated with a ¯ s and inputted into G, which produces the reconstructed pseudo-real feature x ˜ s . G is composed of two fully connected layers, with the first layer having 2048 neurons and the second layer having 256 neurons as the dimension of the real feature. Next, x ˜ s and a ¯ s are concatenated to train the discriminator D. The optimization of E, G, and D is conducted preliminarily through the loss V A E G A N . Afterward, x ˜ s is input into the C F R module to reconstruct the category features a ˜ s . The C F R module comprises two fully connected layers, with the first layer having 2048 neurons. The second layer has the same number of neurons as the dimension of the category feature. The first layer of C F R produces a 2048-dimensional vector known as the intermediate layer feature h, which is then sent to the feedback layer F. The feature h, after passing through the feedback layer F, is weighted and fused with the feature after the first fully connected layer of the generator G. F is also a two-layer fully connected layer, with both layers consisting of 2048 neurons, the same dimension as the input intermediate layer feature. C F R and F are optimized through the loss R , and R further constraining the learning parameters of E, G, and D. During the classification training phase, the pseudo-real feature x ˜ u for the unseen class is generated by the feature concatenating random noise with a ¯ u . Next, x ˜ u is input into C F R and concatenated with the output fake category feature a ˜ u and the intermediate feature h to train the final classifier. The parameters of the classifier are optimized through the loss function C . During the testing phase, the pre-trained CANet should be used to extract the real features x u of the unseen class images. These features should then be input into the trained classification module for classification recognition. The training and testing processes in this paper are integrated, with testing being conducted after each training iteration.

3. Experiments

To verify the performance of the proposed method, we first introduce the dataset settings in Section 3.1. In Section 3.2, we present the experimental results. Section 3.3 examines how the intermediate layer dimension size affects the effectiveness and efficiency of the method. In Section 3.4 we investigate the impact of different settings for the category feature and the seen class real features. In Section 3.5, we further discuss the impact of the seen class number on the model. Finally, we present extended experiments on the generalized zero-shot recognition in Section 3.6.
All the experiments are carried out with both the Matlab platform and the Pytorch framework on a NVIDIA GeForce RTX 3090 GPU card manufactured by NVIDIA Corporation, which is based in Santa Clara, California, USA.

3.1. Datasets

The experiments are conducted on the SAMPLE public dataset [55] and the MSTAR datasets. The SAMPLE public dataset is a subset of the SAMPLE dataset, consisting of SAR real images and simulated images of 10 vehicle targets. The image size is 128 × 128 pixels with a resolution of 0.3 m × 0.3 m, and there is a one-to-one correspondence between the simulated and the real images. The data cover a range of depression angles from 14° to 17° and a range of aspect angles from 10° to 80°. The MSTAR dataset was developed with funding from the Defense Advanced Research Projects Agency (DARPA) of the United States Department of Defense. It consists of 10 classes of vehicle targets with a resolution of 0.3 m × 0.3 m. The dataset taken at multiple depression angles (15°, 17°, 30°, 45°) over a 0°–360° range of aspect angles. In this paper, the data from the SAMPLE public dataset with the depression angles of 16° and 17°, as well as the data from the MSTAR dataset with a depression angle of 15°, are utilized.
Two experimental groups are established: SAMPLE-MSTAR and SAMPLE-MSTAR. The category features are extracted from the simulated images of the SAMPLE public dataset, while the real images from both the SAMPLE and MSATAR datasets are used for testing. The SAMPLE-SAMPLE experimental group includes all 10 categories from the SAMPLE public dataset, while the SAMPLE-MSTAR experimental group comprises only five categories that are shared between the two datasets. Figure 7 displays the simulated and the real images for each class in the two experimental groups. The unseen classes in both experimental groups are BMP2, BTR70, and T72, which are used for testing. The seen class targets comprise the remaining seven and two classes in the respective groups, which are used for network training. Table 1 and Table 2 show the specific composition.

3.2. Effectiveness of the Method

The effectiveness of the proposed method has been tested on the two experimental groups. Ten independent training and testing runs were conducted. Table 3 and Table 4 present the minimum, the maximum, and the average recognition accuracies (with standard deviations) across 10 training–testing cycles, along with the recognition rates for each of the three unseen target classes. Figure 8 and Figure 9 depict the confusion matrix of the experiment with the highest recognition rate during the 10 training–testing cycles.

3.2.1. The Analysis of the Experimental Results

The classic networks A-ConvNet, ResNet18, and Vgg16 were trained directly using simulated images and then tested on the real images for comparison. RN18+ [46] is a network trained exclusively on simulated data, with its training and classification strategy specifically designed based on the SAMPLE dataset. In the SAMPLE-SAMPLE experimental group, the A-ConvNet and Vgg16 networks were able to recognize the unseen class real images with an average recognition rate above 97%. RN18+ achieves a recognition rate of 99.19%, which is higher than the classic networks. The proposed method has an average recognition rate of 99.80 ± 0.22%, with a minimum of 99.35%, outperforming all compared methods. For RN18+, the recognition rate for BMP2 is 100%, which is better than our method. However, it exhibits more misclassifications for the T72. Although the maximum recognition rate of RN18+ can reach 100%, the average recognition rate of the proposed method over ten 10 training–testing cycles is slightly higher than that of RN18+. In the SAMPLE-MSTAR experimental group, the classic networks achieved recognition rates distributed around 40–50%. The confusion matrix shows that the comparative methods tend to identify the three classes’ targets as two of them. For example, Vgg16 and RN18+ almost exclusively identify all T72 targets as either BMP2 or BTR70. This indicates that the networks trained on the simulated images are prone to confusion when directly tested on the real images. The proposed method achieved a final recognition rate of 71.57 ± 2.28%, with a maximum recognition rate of 75.68%. This represents an improvement of nearly 20–30% compared to the classic networks. The confusion matrix of the proposed method demonstrates that the proposed method correctly identifies the majority of targets in each class, indicating that it significantly reduces class confusion.
Furthermore, the ablation experiments are conducted on the overall feature generation module and the feedback module (F) within the feature generation module. The baseline represents the results of training CANet using only simulated images through cross-entropy loss and directly testing it on real images without training the subsequent feature generation network. Ours -F represents the results when the feature generation module is involved in training but the feedback module (F) is discarded. Firstly, in both experimental groups, the recognition results of the baseline are higher than those of A-ConvNet, indicating that the modification of CANet from A-ConvNet retains the feature extraction capability when adapting to the subsequent feature generation network training. Both the baseline and the proposed method achieve a maximum recognition rate of 100% in the SAMPLE-SAMPLE experimental group, while the proposed method maintains the smallest standard deviation. Secondly, the average recognition rates of Ours -F and Ours are both higher than that of the baseline, indicating that the classifier trained using the generated pseudo-real data has better generalization to real data. This validates the effectiveness of the feature generation module. Finally, the average recognition rate of Ours is higher than that of Ours -F. This indicates that the feedback module (F) effectively improves the capability of feature generation during training, resulting in enhanced feature representations.
The experimental results demonstrate that the proposed method can recognize three unseen class targets for both experimental groups. The recognition performance surpasses that of classic networks trained directly using the simulated images, demonstrating the effectiveness of the proposed method. The reason is that the feature generation architecture can create pseudo-real features for the unseen class targets, which aids in classifier training. The generated data have a stronger resemblance to the real data than the simulated data, resulting in a network with better generalization ability.

3.2.2. The Significance Test of the Experimental Results

The Wilcoxon signed-rank test is a non-parametric statistical test employed to assess the dissimilarities between two sets of related samples. In this context, the Wilcoxon signed-rank test is utilized to ascertain the statistical significance of the discrepancy between the proposed method and the comparison method in two experimental groups, as it cannot be assumed that the recognition rates of each method satisfy the normal distribution assumption. Table 5 and Table 6 show the p-values of the paired Wilcoxon test at the 95% significance level for SAMPLE-SAMPLE and SAMPLE-MSTAR. The null hypothesis of the Wilcoxon test is that there is no significant difference between the proposed method and the comparison method. When the p-value is less than 0.05, the null hypothesis is rejected, indicating that the proposed method is significantly superior to the comparison method.
The results demonstrate that for the SAMPLE-SAMPLE experimental group, the p-values for A-ConvNet, ResNet50, Vgg16, RN18+ and baseline are all significantly less than 0.05, indicating that the proposed method outperforms them. The p-value for Ours -F is 0.13 which is greater than 0.05, indicating that the absence of the feedback module (F) in the feature generation module does not significantly affect the final recognition results for the SAMPLE-SAMPLE experimental group. For the SAMPLE-MSTAR experimental group, the p-values for all comparison methods are less than 0.05, indicating that the proposed method is significantly superior to the comparison methods. In this case, the recognition rate of Ours is superior to that of Ours -F, demonstrating that when there is a significant difference between the simulated and measured images, the feedback module (F) demonstrates its advantage.

3.2.3. The Analysis of the Differences between the Two Experimental Groups

Upon further analysis of the experimental results, a significant difference in the recognition rate of the unseen classes is observed between the SAMPLE-SAMPLE and the SAMPLE-MSTAR experimental groups. The difference is attributed to the varying disparities between the simulated and the real data in the two experimental groups. The T-SNE visualization plots Figure 10 show that the simulated images of the unseen classes in the SAMPLE are mostly distributed near their corresponding real images, indicating a strong similarity between them. However, the distribution of the real images in the MSTAR shows weaker correlations compared to the corresponding simulated images in the SAMPLE. When there is a significant difference between the simulated and the real images, the transferability of the networks trained on simulated images is greatly reduced. This is particularly noticeable for the RN18+ network, which enhances performance in the SAMPLE-SAMPLE experimental group but has no impact in the SAMPLE-MSTAR experimental group. The experimental results indicate that the proposed method is effective in the SAMPLE-MSTAR experimental group. It indicates that even when the quality of the simulated images is not high, the extracted category features can still assist the proposed framework in achieving zero-shot recognition of the targets of the multiple classes

3.3. Impact of the Intermediate Layer Dimension Size

The intermediate layer dimension size of G and C F R is equal to the dimension of the feedback feature F. It is an important parameter that impacts the training effectiveness and complexity of the model. Table 7 and Table 8 display the average recognition rates with different intermediate layer dimension sizes.
Four different intermediate layer dimension sizes: 512, 1024, 2018, and 4096 are tested. The results show that for the intermediate layer dimension size less than 2048, the average recognition rate increases with the increase in the intermediate layer dimension size. However, when the dimension size further increases to 4096, the average recognition rate decreases. The analysis indicates that expanding the intermediate layer dimension size can enhance the representational and learning capabilities of the model, thereby improving the recognition performance. However, enlarging the intermediate layer dimension size also increases the parameter count. If the capacity of the model is too large, it may overfit irrelevant features, resulting in a decrease in the final recognition rate. Analyzing the training efficiency of the model, it was observed that the training time increases as the dimension of the intermediate layer increases. This is mainly due to the increase in the model complexity. However, the increase in the training time was relatively slow compared to the change in the intermediate layer dimension. For example, the training time for the 2048 dimension only increased by less than 20 s compared to the 1024 dimension. Even when the intermediate dimension size is set to 4096, the training time for 1000 epochs does not exceed 4 min. This due to the fact that the generation module in this paper consists entirely of simple, shallow fully connected layers. In the feature generation stage, there is no need to update the weights of the feature extraction module, which significantly improves the training efficiency. For the experimental dataset in this paper, a network with an intermediate layer dimension size of 2048 achieves the best performance without significantly increasing the training time and is considered an appropriate size setting.

3.4. Impact of the Number of Simulated and Real Images

The impact of the quantity of the data on the method is further explored in this section. It is important to note that the final performance of the model may be influenced by the simulated and the real images used by the network. Firstly, The representational capability of the categories may vary depending on the number of simulated images used to extract the category features. Secondly, the mapping between the real features and the simulated category features is learned using the real features of the seen classes, which means that the amount of real data can affect the model training. We focus on the SAMPLE-SAMPLE experimental group. Different numbers of the simulated images are used to extract the category features, and various numbers of the real images are used for the model training. The number of the images is determined by the range of the aspect angle. The aspect angle range is represented by 1°, 5°, 10°, 20°, 30°, 40°, 50°, 60°, 70°, which correspond to 10°–11°, 10°–15°, 10°–20°, 10°–30°, 10°–40°, 10°–50°, 10°–60°, 10°–70°, and 10°–80°, respectively. The specific relationship between the aspect angle range and the number of images used is shown in Table 9, where the quantity represents the total number of images across all categories. The number of simulated or real images decreases as the range decreases.
Table 10 shows the average recognition rate for unseen classes in different combinations. According to the table, the recognition performance reaches its maximum at 99.80% when all simulated and real images are used. The lowest recognition rate of 97.31% occurs when the aspect angle range of the simulated images is 10° and only one real image. Figure 11 presents a visualization of the changes in recognition rates. The high recognition rates are mostly achieved when the number of simulated and real images is greater than 40. The recognition rate generally decreases as the quantity of both types of images is reduced, but the decreasing trend is not significant, with a difference of 3% between the minimum and maximum recognition rates.
First, the impact of the number of simulated images is analyzed. When there are enough simulated images, the coverage of the aspect angle is broad encompassing more information. As a result, the category features possess stronger representational capabilities. A decrease in the number of simulated images results in a reduction in the introduced aspect angle information, which weakens the representational capacity of the category features. This, in turn, affects the generation of pseudo-real features and further impacts the final classification performance. However, the model maintains its effectiveness with a recognition rate above 97% even when using a small number of simulated images. It may be because the deep features still include scatter characteristics and other category information, which to an extent represent the entire category. Second, concerning the impact of the number of real images. When the training data volume is large, the model can learn rich mappings as the aspect angle range of the seen class can cover 0–80°. The features generated for the unseen class incorporate information from multiple aspect angles, resulting in the effective recognition of true samples with varying aspect angles. As the number of real images decreases, the acquisition of aspect angle information also decreases, leading to a slight decline in the recognition performance of the model. However, the model can still achieve a relatively high recognition rate because the model can focus on learning the relatively simple mappings and other features unrelated to the aspect angle.
The results suggest that the proposed method can maintain a relatively high recognition rate when training on a sparse set of simulated and real images. The requirement for the number of simulated and real images is reduced.

3.5. Impact of the Seen Class Number

The ability to generalize to unseen classes of the model may be affected by the number of seen classes in addition to the number of images used by the network. This is because the number of seen classes is related to the coverage of the feature space. More seen classes may provide additional features shared with unseen classes, which can improve the ability to generate unseen class data. In this section, the number of the seen classes will be reduced to explore their impact on the final recognition rate. The unseen classes remain BMP2, BTR70, and T72. The initial number of the seen classes was seven in the SAMPLE-SAMPLE and two in the SAMPLE-MSTAR. Both experimental groups ultimately ended up with a single commonly seen class, which is 2S1. Table 11 and Table 12 present the experimental results.
The results show a decrease in accuracy when the number of seen classes is less than three for the SAMPLE-SAMPLE experimental group. Specifically, the final accuracy for the two and one seen classes is 92.41% and 88.83%, respectively. Similarly, the recognition rate for one seen class in the SAMPLE-MSTAR experimental group is 2.34% lower than the recognition rate for two seen classes. However, the performance of the model is not strictly proportional to the number of the seen classes. In the experimental group SAMPLE-SAMPLE, the recognition rate of the method can reach 99% or higher when the number of seen classes is three or more. This may be due to the fact that when the number of seen classes is more than three, the features selected from the seen classes can already cover most of the common information with the unseen classes. As a result, the mapping learned from the seen classes can generalize well to the unseen classes.
The recognition performance of the model decreases as the number of seen classes decreases, but there is not necessarily a linear relationship between them. Moreover, the model can maintain a certain level of performance even with fewer seen classes.

3.6. Extended Experiments on the Generalized Zero-Shot Recognition

In real-world scenarios, both the seen class samples and the unseen class samples often coexist. Therefore, it is necessary to be able to recognize both types of samples simultaneously. This recognition problem is known as the generalized zero-shot recognition problem, which is an extension of the zero-shot problem. The training process for generalized zero-shot recognition remains consistent with that of zero-shot recognition, allowing only the seen class samples to be used for training. During the testing phase, the network must be able to recognize both seen and unseen class targets simultaneously. To achieve this, the final classifier training includes both the real features of the seen classes and the generated features of the unseen classes. That is, a softmax classifier f g z s l : X s X ˜ u Y s Y u is trained with X s X ˜ u , and the classifier can recognize both the seen and the unseen class samples. Compared to the zero-shot recognition, the generalized zero-shot recognition is more challenging.
This section presents the results of an extension experiment on the generalized zero-shot SAR target recognition with two experimental groups. In the SAMPLE-SAMPLE experimental group, the separation of the seen and the unseen classes aligns with zero-shot target recognition. In the SAMPLE-MSTAR experimental group, the unseen classes are BMP2 and BMP2, with T72 designated as the seen class. In the comparison experiment, the network is trained with a combination of simulated images of the unseen class and real images of the seen class. Table 13 and Table 14 display the experimental results.
In the SAMPLE-SAMPLE experimental group, the classic networks A-Convnets, ResNet50 and Vgg16 almost misclassify all unseen category targets as seen class targets. While RN18+ and baseline have some recognition ability for the unseen classes, the recognition rate does not exceed 45%. In the SAMPLE-MSTAR experimental group, the comparison methods trained directly with supplemented simulated images of unseen classes have an almost 0% recognition rate for unseen class targets. The reason for the results is that there are statistical distribution differences between the real and the simulated images and the model more easily captures features from real images that dominate the training set. In the SAMPLE-SAMPLE experimental group, the recognition rate of the best comparison method, RN18+, reached 85%. At the same time, our method achieved an overall recognition rate of 91.45%. Although the recognition rate for seen class targets has slightly decreased, the recognition rate for unseen category targets has significantly increased, with an average recognition rate of over 70%. For the SAMPLE-MSTAR experimental group, our method achieved an overall recognition rate of 81.15%, which is approximately 15% higher than that of the comparison methods. Our method generates pseudo-real features for unseen classes that have a feature distribution closer to the real features. When mixing training with real samples of the seen classes, the model preference for the seen classes is reduced and achieves a balance between the seen and unseen classes. The results indicate that our method contributes to the generalized zero-shot recognition.

4. Discussion

The experiments in Section 3 demonstrate that the proposed method can achieve zero-shot recognition for multiple classes of targets, rather than inferring only a single class. It represents a breakthrough compared to existing embedding model-based methods for SAR target zero-shot recognition. Furthermore, the proposed feature generation architecture outperforms classical networks trained directly on simulated images in terms of recognition performance. The effectiveness of the architecture is highlighted. Several factors influencing the method were examined, as discussed in detail in Section 3. Additionally, the proposed method has the following limitations.
  • The deep features utilized for network learning in this paper are extracted solely from a model pre-trained on seen class real images. It may limit their representational capacity thereby influencing the learning of the network.
  • The proposed method relies on the similarity of the features between the seen and the unseen class targets, allowing for the transfer of the mapping learned from the seen classes to the unseen classes. If there is a significant difference between the seen and the unseen classes, the proposed method may have limitations.
Based on the analysis above, the proposed method presents a new approach to address zero-shot recognition in SAR images. However, it still has limitations that require further exploration and improvement. In the future, the following directions will be mainly researched. Firstly, multiple feature extraction methods will be adopted. For example, initially using a large-scale SAR image pre-trained feature extraction network, and then transferring it to specific tasks. Alternatively, we will use self-supervised networks like VAE to learn more effective feature representations. Furthermore, the vector embedding of large models can also be utilized. Secondly, the situations where there is a notable contrast between the seen and the unseen classes will be examined. Domain adaptation methods will be introduced into zero-shot recognition tasks to address this issue.

5. Conclusions

The traditional classification networks cannot classify unseen SAR targets. To generate pseudo-real samples of unseen classes for supervised learning and achieve classification of the targets for the multiple unseen classes, a conditional generative network with category features from simulated images for zero-shot SAR target recognition is proposed. Specifically, the process begins with the extraction of the category features from the simulated images. Next, a conditional VAE-GAN network is trained using samples of seen classes. Pseudo-real samples of the unseen classes are then generated using the category features of unseen classes as conditions. Finally, the classification network is trained using the generated samples to achieve supervised learning. The proposed method can recognize three unseen category targets in both the SAMPLE and MSTAR datasets, achieving recognition rates of 99.80 ± 1.22% and 71.57 ± 2.28%, respectively. The recognition performance of the method decreases slightly when fewer simulated images are used to extract category features and when fewer seen class real images are used for training. However, the decrease is not significant, with a difference of no more than 3% between the lowest and highest recognition rates. The proposed method remains effective even with only a few seen classes. The recognition rates for the three unseen class targets in the SAMPLE and MSTAR datasets exceed 90% and 70%, respectively, when only two seen classes are used. Additionally, the proposed method can be expanded to the tasks of the generalized zero-shot recognition tasks.

Author Contributions

Conceptualization, G.C.; methodology, G.C.; software G.C.; validation, G.C.; writing—original draft preparation, G.C.; supervision, S.Z.; project administration, S.Z.; data processing, Q.H.; visualization, Z.S. and X.Z.; Literature collection, L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Acknowledgments

The authors would like to thank the reviewers and editors who provided valuable comments and suggestions for this article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Khoshnevis, S.A.; Ghorshi, S. A tutorial on tomographic synthetic aperture radar methods. SN Appl. Sci. 2020, 2, 1504. [Google Scholar] [CrossRef]
  2. Dudgeon, D.E.; Lacoss, R.T. An overview of automatic target recognition. Linc. Lab. J. 1993, 6, 3–10. [Google Scholar]
  3. Popova, M.; Shvets, M.; Oliva, J.; Isayev, O. MolecularRNN: Generating realistic molecular graphs with optimized properties. arXiv 2019, arXiv:1905.13372. [Google Scholar]
  4. Shaulskiy, D.V.; Evtikhiev, N.N.; Starikov, R.S.; Starikov, S.N.; Zlokazov, E.Y. MINACE filter: Variants of realization in 4-f correlator. In Proceedings of the Optical Pattern Recognition XXV, Baltimore, MD, USA, 5–9 May 2014; SPIE: Bellingham, WA, USA, 2014; Volume 9094, pp. 135–142. [Google Scholar]
  5. Diemunsch, J.R.; Wissinger, J. Moving and stationary target acquisition and recognition (MSTAR) model-based automatic target recognition: Search technology for a robust ATR. In Proceedings of the Algorithms for Synthetic Aperture Radar Imagery V, Orlando, FL, USA, 13–17 April 1998; SPIE: Bellingham, WA, USA, 1998; Volume 3370, pp. 481–492. [Google Scholar]
  6. Li, J.; Yu, Z.; Yu, L.; Cheng, P.; Chen, J.; Chi, C. A Comprehensive Survey on SAR ATR in Deep-Learning Era. Remote Sens. 2023, 15, 1454. [Google Scholar] [CrossRef]
  7. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1–9. [Google Scholar] [CrossRef]
  8. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  9. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  10. Pearlmutter. Learning state space trajectories in recurrent neural networks. In Proceedings of the International 1989 Joint Conference on Neural Networks, Washington, DC, USA, 18–22 June 1989; IEEE: Piscataway, NJ, USA, 1989; pp. 365–372. [Google Scholar]
  11. Scarselli, F.; Tsoi, A.C.; Gori, M.; Hagenbuchner, M. Graphical-based learning environments for pattern recognition. In Proceedings of the Structural, Syntactic, and Statistical Pattern Recognition: Joint IAPR International Workshops, SSPR 2004 and SPR 2004, Lisbon, Portugal, 18–20 August 2004; Proceedings. Springer: Berlin/Heidelberg, Germany, 2004; pp. 42–56. [Google Scholar]
  12. Morgan, D.A. Deep convolutional neural networks for ATR from SAR imagery. In Proceedings of the Algorithms for Synthetic Aperture Radar Imagery XXII, Baltimore, MD, USA, 20–24 April 2015; SPIE: Bellingham, WA, USA, 2015; Volume 9475, pp. 116–128. [Google Scholar]
  13. Chen, S.; Wang, H.; Xu, F.; Jin, Y.Q. Target classification using the deep convolutional networks for SAR images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4806–4817. [Google Scholar] [CrossRef]
  14. Zhang, F.; Hu, C.; Yin, Q.; Li, W.; Li, H.C.; Hong, W. Multi-aspect-aware bidirectional LSTM networks for synthetic aperture radar target recognition. IEEE Access 2017, 5, 26880–26891. [Google Scholar] [CrossRef]
  15. Zhao, C.; Zhang, S.; Luo, R.; Feng, S.; Kuang, G. Scattering features spatial-structural association network for aircraft recognition in SAR images. IEEE Geosci. Remote. Sens. Lett. 2023, 20, 4006505. [Google Scholar] [CrossRef]
  16. Zhang, X.; Feng, S.; Zhao, C.; Sun, Z.; Zhang, S.; Ji, K. MGSFA-Net: Multi-Scale Global Scattering Feature Association Network for SAR Ship Target Recognition. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 2024, 17, 4611–4625. [Google Scholar] [CrossRef]
  17. Ding, J.; Chen, B.; Liu, H.; Huang, M. Convolutional neural network with data augmentation for SAR target recognition. IEEE Geosci. Remote Sens. Lett. 2016, 13, 364–368. [Google Scholar] [CrossRef]
  18. Ding, B.; Wen, G.; Huang, X.; Ma, C.; Yang, X. Data augmentation by multilevel reconstruction using attributed scattering center for SAR target recognition. IEEE Geosci. Remote Sens. Lett. 2017, 14, 979–983. [Google Scholar] [CrossRef]
  19. Guo, J.; Lei, B.; Ding, C.; Zhang, Y. Synthetic aperture radar image synthesis by using generative adversarial nets. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1111–1115. [Google Scholar] [CrossRef]
  20. Cui, Z.; Zhang, M.; Cao, Z.; Cao, C. Image data augmentation for SAR sensor via generative adversarial nets. IEEE Access 2019, 7, 42255–42268. [Google Scholar] [CrossRef]
  21. Niu, S.; Qiu, X.; Peng, L.; Lei, B. Parameter prediction method of SAR target simulation based on convolutional neural networks. In Proceedings of the EUSAR 2018; 12th European Conference on Synthetic Aperture Radar, Aachen, Germany, 4–7 June 2018; VDE: Frankfurt am Main, Germany, 2018; pp. 1–5. [Google Scholar]
  22. Zhai, Y.; Deng, W.; Xu, Y.; Ke, Q.; Gan, J.; Sun, B.; Zeng, J.; Piuri, V. Robust SAR automatic target recognition based on transferred MS-CNN with L 2-regularization. Comput. Intell. Neurosci. 2019, 2019, 9140167. [Google Scholar] [CrossRef] [PubMed]
  23. Malmgren-Hansen, D.; Kusk, A.; Dall, J.; Nielsen, A.A.; Engholm, R.; Skriver, H. Improving SAR automatic target recognition models with transfer learning from simulated data. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1484–1488. [Google Scholar] [CrossRef]
  24. Zelong, W.; Xianghui, X.; Lei, Z. Study of deep transfer learning for SAR ATR based on simulated SAR images. J. Univ. Chin. Acad. Sci. 2020, 37, 516. [Google Scholar]
  25. Wang, K.; Zhang, G. SAR target recognition via meta-learning and amortized variational inference. Sensors 2020, 20, 5966. [Google Scholar] [CrossRef]
  26. Wang, K.; Zhang, G.; Xu, Y.; Leung, H. SAR target recognition based on probabilistic meta-learning. IEEE Geosci. Remote Sens. Lett. 2020, 18, 682–686. [Google Scholar] [CrossRef]
  27. Keydel, E.R.; Lee, S.W.; Moore, J.T. MSTAR extended operating conditions: A tutorial. Algorithms Synth. Aperture Radar Imag. III 1996, 2757, 228–242. [Google Scholar]
  28. Larochelle, H.; Erhan, D.; Bengio, Y. Zero-data learning of new tasks. In Proceedings of the AAAI, Chicago, IL, USA, 13–17 July 2008; Volume 1, p. 3. [Google Scholar]
  29. Lampert, C.H.; Nickisch, H.; Harmeling, S. Attribute-based classification for zero-shot visual object categorization. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 453–465. [Google Scholar] [CrossRef]
  30. Chen, L.; Zhang, H.; Xiao, J.; Liu, W.; Chang, S.F. Zero-shot visual recognition using semantics-preserving adversarial embedding networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1043–1052. [Google Scholar]
  31. Shigeto, Y.; Suzuki, I.; Hara, K.; Shimbo, M.; Matsumoto, Y. Ridge regression, hubness, and zero-shot learning. In Proceedings of the Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2015, Porto, Portugal, 7–11 September 2015; Proceedings, Part I 15. Springer: Berlin/Heidelberg, Germany, 2015; pp. 135–151. [Google Scholar]
  32. Yang, Y.; Hospedales, T.M. A unified perspective on multi-domain and multi-task learning. arXiv 2014, arXiv:1412.7489. [Google Scholar]
  33. Liu, Y.; Zhou, L.; Bai, X.; Huang, Y.; Gu, L.; Zhou, J.; Harada, T. Goal-oriented gaze estimation for zero-shot learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 3794–3803. [Google Scholar]
  34. Mishra, A.; Krishna Reddy, S.; Mittal, A.; Murthy, H.A. A generative model for zero shot learning using conditional variational autoencoders. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2188–2196. [Google Scholar]
  35. Xian, Y.; Lorenz, T.; Schiele, B.; Akata, Z. Feature generating networks for zero-shot learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5542–5551. [Google Scholar]
  36. Xian, Y.; Sharma, S.; Schiele, B.; Akata, Z. f-vaegan-d2: A feature generating framework for any-shot learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 10275–10284. [Google Scholar]
  37. Narayan, S.; Gupta, A.; Khan, F.S.; Snoek, C.G.; Shao, L. Latent embedding feedback and discriminative features for zero-shot classification. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XXII 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 479–495. [Google Scholar]
  38. Song, Q.; Xu, F. Zero-shot learning of SAR target feature space with deep generative neural networks. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2245–2249. [Google Scholar] [CrossRef]
  39. Wei, Q.R.; He, H.; Zhao, Y.; Li, J.A. Learn to recognize unknown SAR targets from reflection similarity. IEEE Geosci. Remote Sens. Lett. 2020, 19, 4002205. [Google Scholar] [CrossRef]
  40. Wei, Q.R.; Chen, C.Y.; He, M.; He, H.M. Zero-Shot SAR Target Recognition Based on Classification Assistance. IEEE Geosci. Remote Sens. Lett. 2023, 20, 4003705. [Google Scholar] [CrossRef]
  41. Cha, M.; Majumdar, A.; Kung, H.; Barber, J. Improving SAR automatic target recognition using simulated images under deep residual refinements. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 2606–2610. [Google Scholar]
  42. Wang, K.; Zhang, G.; Leung, H. SAR target recognition based on cross-domain and cross-task transfer learning. IEEE Access 2019, 7, 153391–153399. [Google Scholar] [CrossRef]
  43. Liping, H.; Chunzhu, D.; Jinfan, L.; Hongcheng, Y.; Chao, W.; Chao, N. Non-homologous target recognition of ground vehicles based on SAR simulation image. Syst. Eng. Electron. 2021, 43, 3518–3525. [Google Scholar]
  44. Zhang, C.; Wang, Y.; Liu, H.; Sun, Y.; Hu, L. SAR target recognition using only simulated data for training by hierarchically combining CNN and image similarity. IEEE Geosci. Remote Sens. Lett. 2022, 19, 4503505. [Google Scholar] [CrossRef]
  45. Song, Q.; Chen, H.; Xu, F.; Cui, T.J. EM simulation-aided zero-shot learning for SAR automatic target recognition. IEEE Geosci. Remote Sens. Lett. 2019, 17, 1092–1096. [Google Scholar] [CrossRef]
  46. Inkawhich, N.; Inkawhich, M.J.; Davis, E.K.; Majumder, U.K.; Tripp, E.; Capraro, C.; Chen, Y. Bridging a gap in SAR-ATR: Training on fully synthetic and testing on measured data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 2942–2955. [Google Scholar] [CrossRef]
  47. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  48. Liu, L.; Pan, Z.; Qiu, X.; Peng, L. SAR target classification with CycleGAN transferred simulated samples. In Proceedings of the IGARSS 2018–2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 4411–4414. [Google Scholar]
  49. Lewis, B.; Liu, J.; Wong, A. Generative adversarial networks for SAR image realism. In Proceedings of the Algorithms for Synthetic Aperture Radar Imagery XXV, Orlando, FL, USA, 15–19 April 2018; SPIE: Bellingham, WA, USA, 2018; Volume 10647, pp. 37–47. [Google Scholar]
  50. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  51. Sohn, K.; Lee, H.; Yan, X. Learning structured output representation using deep conditional generative models. In Proceedings of the 28th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; Volume 28. [Google Scholar]
  52. Snell, J.; Swersky, K.; Zemel, R. Prototypical networks for few-shot learning. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  53. Larsen, A.B.L.; Sønderby, S.K.; Larochelle, H.; Winther, O. Autoencoding beyond pixels using a learned similarity metric. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 20–22 June 2016; PMLR: Cambridge, MA, USA, 2016; pp. 1558–1566. [Google Scholar]
  54. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A.C. Improved training of wasserstein gans. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  55. Lewis, B.; Scarnati, T.; Sudkamp, E.; Nehrbass, J.; Rosencrantz, S.; Zelnio, E. A SAR dataset for ATR development: The Synthetic and Measured Paired Labeled Experiment (SAMPLE). In Proceedings of the Algorithms for Synthetic Aperture Radar Imagery XXVI, Baltimore, MD, USA, 14–18 April 2019; SPIE: Bellingham, WA, USA, 2019; Volume 10987, pp. 39–54. [Google Scholar]
Figure 1. Different types of network learning methods.
Figure 1. Different types of network learning methods.
Remotesensing 16 01930 g001
Figure 2. The overall architecture of the proposed method.
Figure 2. The overall architecture of the proposed method.
Remotesensing 16 01930 g002
Figure 3. The specific structure of CANet.
Figure 3. The specific structure of CANet.
Remotesensing 16 01930 g003
Figure 4. The process of extracting category features.
Figure 4. The process of extracting category features.
Remotesensing 16 01930 g004
Figure 5. The specific structure of the feature generation module.
Figure 5. The specific structure of the feature generation module.
Remotesensing 16 01930 g005
Figure 6. The specific structure of the classification module.
Figure 6. The specific structure of the classification module.
Remotesensing 16 01930 g006
Figure 7. Examples of image pairs in the two experimental groups.
Figure 7. Examples of image pairs in the two experimental groups.
Remotesensing 16 01930 g007
Figure 8. The confusion matrix of SAMPLE-SAMPLE.
Figure 8. The confusion matrix of SAMPLE-SAMPLE.
Remotesensing 16 01930 g008
Figure 9. The confusion matrix of SAMPLE-MSTAR.
Figure 9. The confusion matrix of SAMPLE-MSTAR.
Remotesensing 16 01930 g009
Figure 10. The T-SNE visualization plots of three unseen targets.
Figure 10. The T-SNE visualization plots of three unseen targets.
Remotesensing 16 01930 g010
Figure 11. The visualization of the changes in recognition rates.
Figure 11. The visualization of the changes in recognition rates.
Remotesensing 16 01930 g011
Table 1. The SAMPLE-SAMPLE experimental group.
Table 1. The SAMPLE-SAMPLE experimental group.
Seen ClassesUnseen Classes
Type2S1M1M2M35M60M584ZSU23-4BMP2BTR70T72
Simulated10810310510511110510810792108
Real10810310510511110510810792108
Table 2. The SAMPLE-MSTAR experimental group.
Table 2. The SAMPLE-MSTAR experimental group.
Seen ClassesUnseen Classes
Type2S1ZSU23-4BMP2BTR70T72
Simulated10810810792108
Real274274195196196
Table 3. Experimental results of SAMPLE-SAMPLE.
Table 3. Experimental results of SAMPLE-SAMPLE.
MethodsBMP2 (%)BTR70 (%)T72 (%)Min (%)Max (%)Avg (%)
A-ConvNet99.4498.5995.7496.7498.7097.88 ± 0.69
ResNet5076.6455.9875.6566.4576.5570.10 ± 3.20
Vgg1699.0798.9196.3994.4699.3598.07 ± 1.29
RN18+10099.7897.8798.0510099.19 ± 0.51
baseline98.6998.8098.1597.0710098.53 ± 0.92
Ours -F99.9110099.2699.3510099.71 ± 0.23
Ours99.8199.8999.7299.3510099.80 ± 0.22
The highest recognition rate achieved for each column are marked in bold.
Table 4. Experimental results of SAMPLE-MSTAR.
Table 4. Experimental results of SAMPLE-MSTAR.
MethodsBMP2 (%)BTR70 (%)T72 (%)Min (%)Max (%)Avg (%)
A-ConvNet60.9842.5019.3935.9547.1940.92 ± 2.87
ResNet5043.9533.0167.5543.1053.6648.18 ± 3.67
Vgg1660.7763.672.1438.3344.9742.67 ± 1.90
RN18+69.1854.340.2638.6742.1641.21 ± 1.9
baseline53.1349.4924.2437.4844.9742.27 ± 2.13
Ours -F47.3965.2090.1564.9171.7267.62 ± 2.27
Ours59.5968.4286.8366.7875.9871.57 ± 2.28
Table 5. The p-value of the paired Wilcoxon test at the 95% significance level for SAMPLE-SAMPLE.
Table 5. The p-value of the paired Wilcoxon test at the 95% significance level for SAMPLE-SAMPLE.
MethodA-ConvNetResNet50Vgg16RN18+BaselineOurs -F
p-value 9.8 × 10 4 9.8 × 10 4 9.8 × 10 4 7.0 × 10 3 3.8 × 10 3 0.13
Table 6. The p-value of the paired Wilcoxon test at the 95% significance level for SAMPLE-MSTAR.
Table 6. The p-value of the paired Wilcoxon test at the 95% significance level for SAMPLE-MSTAR.
MethodA-ConvNetResNet50Vgg16RN18+BaselineOurs -F
p-value 9.8 × 10 4 9.8 × 10 4 9.8 × 10 4 9.8 × 10 4 9.8 × 10 4 2.9 × 10 3
Table 7. Experimental results for different intermediate dimension sizes in SAMPLE-SAMPLE.
Table 7. Experimental results for different intermediate dimension sizes in SAMPLE-SAMPLE.
DimensionsMin (%)Max (%)Avg (%)Time (s)
51299.6710099.77 ± 0.15128.75
102499.3510099.80 ± 0.22134.52
204899.6710099.84  ±  0.16158.19
409699.3510099.71 ± 0.23226.76
Table 8. Experimental results for different intermediate dimension sizes in SAMPLE-MSTAR.
Table 8. Experimental results for different intermediate dimension sizes in SAMPLE-MSTAR.
DimensionsMin (%)Max (%)Avg (%)Time (s)
51263.5473.2567.56 ± 2.56130.54
102464.0575.1369.30 ± 3.24140.40
204866.7875.9871.57 ± 2.28164.37
409666.6175.9871.14 ± 2.72237.83
The highest recognition rate achieved for each column are marked in bold.
Table 9. The correspondence between the aspect angle range and the number of images.
Table 9. The correspondence between the aspect angle range and the number of images.
The Aspect Angle Range 10°20°30°40°50°60°70°
The Simulated Image Number735147268395504603670745
The Real Image Number735147268395504603670745
Table 10. Experimental results for different number configurations (%).
Table 10. Experimental results for different number configurations (%).
Real
10°20°30°40°50°60°70°
Simulated97.9297.7998.1498.7998.8698.7098.9998.6398.79
97.6297.6598.5398.5098.0898.4098.3798.5398.60
10°97.3197.8298.4098.3198.3198.4098.3498.2498.21
20°98.3798.3198.5098.5098.5398.4098.7998.9698.70
30°98.9198.7099.0699.0699.0699.2599.3298.9399.35
40°99.1999.5899.3599.5499.4899.5499.5199.6799.35
50°99.4199.5199.6199.4899.6499.5499.6799.5899.67
60°99.4899.4899.7799.6799.4599.6199.8499.6799.67
70°99.5199.6799.6499.7199.5899.5199.7799.7799.80
Table 11. SAMPLE-SAMPLE experimental results for different seen class number (%).
Table 11. SAMPLE-SAMPLE experimental results for different seen class number (%).
The Seen Class Number7654321
Average Accuracy99.8099.7499.9099.3299.4892.4188.83
Table 12. SAMPLE-MSTAR experimental results for different seen class number (%).
Table 12. SAMPLE-MSTAR experimental results for different seen class number (%).
The Seen Class Number21
Average Accuracy71.5769.23
Table 13. Experimental results of the generalized zero-shot recognition in SAMPLE-SAMPLE (%).
Table 13. Experimental results of the generalized zero-shot recognition in SAMPLE-SAMPLE (%).
Seen ClassesUnseen Classes
2S1M1M2M35M548ZSU23-4M60BMP2BTR70T72Accuracy
A-ConvNet99.4398.4596.091001001001000012.0477.55
ResNet5010010010010010010098.851.8700.9377.25
Vgg16100100100100100100100003.7077.47
RN18+97.7010096.8895.351001001003.7460.8764.8185.80
baseline10010099.2210010098.8610031.7845.6525.9384.68
Ours93.1095.3595.3110010010010095.3351.0984.2691.45
Table 14. Experimental results of the generalized zero-shot recognition in SAMPLE-MSTAR (%).
Table 14. Experimental results of the generalized zero-shot recognition in SAMPLE-MSTAR (%).
Seen ClassesUnseen Classes
T72ZSU23-42S1BMP2BTR70Accuracy
A-ConvNet1001001000065.55
ResNet5099.491001002.05065.81
Vgg161001001001.54065.81
RN18+73.9810079.200056.04
baseline10010010000.5165.64
Ours91.3310086.1351.2867.3581.15
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, G.; Zhang, S.; He, Q.; Sun, Z.; Zhang, X.; Zhao, L. Zero-Shot SAR Target Recognition Based on a Conditional Generative Network with Category Features from Simulated Images. Remote Sens. 2024, 16, 1930. https://doi.org/10.3390/rs16111930

AMA Style

Chen G, Zhang S, He Q, Sun Z, Zhang X, Zhao L. Zero-Shot SAR Target Recognition Based on a Conditional Generative Network with Category Features from Simulated Images. Remote Sensing. 2024; 16(11):1930. https://doi.org/10.3390/rs16111930

Chicago/Turabian Style

Chen, Guo, Siqian Zhang, Qishan He, Zhongzhen Sun, Xianghui Zhang, and Lingjun Zhao. 2024. "Zero-Shot SAR Target Recognition Based on a Conditional Generative Network with Category Features from Simulated Images" Remote Sensing 16, no. 11: 1930. https://doi.org/10.3390/rs16111930

APA Style

Chen, G., Zhang, S., He, Q., Sun, Z., Zhang, X., & Zhao, L. (2024). Zero-Shot SAR Target Recognition Based on a Conditional Generative Network with Category Features from Simulated Images. Remote Sensing, 16(11), 1930. https://doi.org/10.3390/rs16111930

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop