Next Article in Journal
Fiber Optic Sensor of Axial Strain and Dynamic Transverse Force Based on Intensity Demodulation
Previous Article in Journal
CONECT: Novel Weighted Networks Framework Leveraging Angle-Relation Connection (ARC) and Metaheuristic Algorithms for EEG-Based Dementia Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semi-Supervised Radar Work Mode Recognition Based on Contrastive Learning

College of Electronic Countermeasure, National University of Defense Technology, Hefei 230071, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(24), 7440; https://doi.org/10.3390/s25247440 (registering DOI)
Submission received: 30 October 2025 / Revised: 25 November 2025 / Accepted: 2 December 2025 / Published: 7 December 2025
(This article belongs to the Section Radar Sensors)

Highlights

What are the main findings?
  • An end-to-end semi-supervised learning framework based on triple-branch contrastive learning is proposed for fine-grained radar work mode recognition in complex electromagnetic environments.
  • Trained with only 10% labeled data, the framework outperforms state-of-the-art contrastive learning methods by 17% and 34% in recognition accuracy on PDW-MP and PDW-FP datasets, respectively, and maintains robustness in non-ideal scenarios.
What is the implication of the main finding?
  • The high accuracy under limited labeled data significantly alleviates the reliance on costly and time-consuming radar data annotation, lowering the threshold for practical deployment of semi-supervised learning in radar systems.
  • Robust performance in non-ideal scenarios validates the framework’s applicability to real-world complex electromagnetic environments, providing a reliable solution for fine-grained radar mode recognition tasks.

Abstract

Deep learning for fine-grained radar mode recognition faces a major bottleneck—its heavy reliance on expensively labeled data. To address this, we propose a novel semi-supervised framework that effectively leverages unlabeled data. Through an end-to-end, triple-branch framework that integrates a dual contrastive learning mechanism with tailored strategies for pulse distortions, our model achieves high accuracy with minimal labels. Experimental results on two challenging datasets demonstrate that the proposed framework boosts accuracy by 17% to 34% using only 10% of the labeled data, establishing a new state-of-the-art performance.

1. Introduction

Modern Electronic Warfare (EW) primarily consists of three key components: Electronic Support (ES), Electronic Attack (EA), and Electronic Protection (EP) [1,2]. ES is responsible for searching, locating, and analyzing intercepted electronic signals to provide decision-making support for other systems. Modern Multifunction Radar (MFR) can leverage programmable parameters and flexible modulation types to generate fine-grained radar work modes, effectively meeting diverse mission requirements [3,4,5,6]. The recognition of MFR modes is critical for ES missions and plays an important role in the fields of electronic intelligence, cognitive radio, and operational decision-making [7,8,9]. The reliable assessment of behavioral posture and potential threats posed by non-cooperative MFRs is a vital and challenging task for ES systems. The key to effective identification lies in learning and distinguishing the patterns presented by the characteristic pulse parameters and waveforms of MFR.
MFRs can flexibly regulate modulation types and parameters to generate diverse fine-grained work modes, enhancing mission adaptability and anti-detection capability [10]. These modes are characterized by pulse descriptor word (PDW) parameters, including frequency (RF), pulse width (PW), and pulse repetition interval (PRI), with each parameter combination representing a distinct pattern. Identifying these modes through inter-pulse parameter analysis provides critical intelligence for electronic surveillance. Recent deep learning advances have demonstrated superior performance in MFR signal analysis. He et al. [11] embedded functional semantics into radar waveforms using LSTM networks, shifting focus from signal properties to behavioral understanding. To address noisy environments, Zhang et al. [12] developed a Residual Shrinkage Convolutional Network that suppresses irrelevant features through adaptive thresholding. Xiong et al. [13] integrated U-Net denoising with multi-scale attention for robust pattern recognition. For uncertainty modeling, Bayesian approaches have shown promise. Du et al. [14] proposed a Bayesian Attention Belief Network using CNN priors to guide inference, while He et al. [15] employed Sparse Bayesian Learning for joint parameter estimation in multipath environments. These methods enhance reliability under dynamic signal conditions. Supervised radar classification faces practical constraints due to the high cost of manual pulse annotation and the difficulty of obtaining quality labels in complex electromagnetic environments. While semi-supervised learning, leveraging both labeled and unlabeled data, offers a promising alternative, existing methods remain susceptible to pulse distortion under non-ideal conditions with limited labeled samples, requiring improved recognition performance.
Under non-ideal conditions with limited labeled samples, the learned feature distributions of existing semi-supervised recognition methods are susceptible to pulse distortion. Therefore, the recognition performance under non-ideal conditions with a small number of labeled samples still needs to be improved. This paper proposes a Tri-branch Semi-supervised Contrastive Learning (TriSCL) approach for radar work mode recognition based on a semi-supervised framework. This paper makes the following key contributions to advance fine-grained radar work mode recognition under limited labeled data and non-ideal conditions:
  • We design a novel three-branch contrastive learning framework. Distinctively, the third branch receives a fusion of the augmented views from the first two branches, serving as a stable anchor. This allows the model to compute more consistent unsupervised contrastive losses, which are then jointly optimized with supervised objectives to effectively tackle label scarcity.
  • We enhance model robustness against realistic signal degradation by designing two dedicated augmentation strategies that simulate false pulses and pulse loss. Integrated into the framework, they force the model to learn distortion-invariant representations.
  • We introduce an encoder equipped with residual shrinkage modules and dual attention to improve feature extraction. This architecture selectively focuses on critical temporal patterns and feature dependencies while suppressing noise, which is crucial for handling complex radar modulations.
  • We conduct Extensive experiments on the PDW-MP and PDW-FP datasets, demonstrating state-of-the-art performance. Our method achieves an average accuracy improvement of 15% using only 10% labeled data, proving its efficacy and practical value.
The structure of this paper is organized as follows. The subsequent section (Section 2) commences with a review of related work in radar emitter recognition and contrastive learning. Section 3 then recalls the foundational concepts of radar pulse stream structure and diverse inter-pulse modulation types. The core of our work—TriSCL—is thoroughly designed and presented in Section 4. Section 5 is dedicated to an extensive experimental evaluation, where the accuracy and robustness of our method are assessed against several state-of-the-art techniques. The paper concludes with a summary of findings and a perspective on future work in Section 6.

2. Related Work

2.1. Semi-Supervised Radar Emitter Identification Methods

Several semi-supervised learning frameworks have been adapted for the radar emitter identification, which can be categorized as follows:

2.1.1. Consistency Regularization

Consistency regularization enhances model robustness by enforcing prediction invariance under data perturbations or augmentations, improving stability without requiring labeled data. Fu et al. [16] present an SEI method based on dual consistency regularization (DCR) evaluated on a WiFi dataset with 16 categories and an ADS-B dataset with 10 categories. Xu et al. [17] developed an SSL-SEI approach with triple consistency on the 20-class ADS-B dataset.

2.1.2. Generative Modeling

Generative modeling learns latent data distributions rather than class-specific features [18]. Gong et al. [19] integrated auto-encoders with Triple-GAN [20] for signal characterization. Tan et al. [21] developed self-classifying GANs using bispectral features, while Xie [22] and Zhou [23] employed auto-encoder reconstructions for representation learning. However, generative approaches are excluded from our framework due to their computational complexity and tendency to learn features not directly optimized for discrimination. The reconstruction-focused objectives may capture signal characteristics irrelevant to emitter identification, reducing efficiency in practical applications.

2.1.3. Pseudo-Labeling

Pseudo-labeling assigns predicted labels to unlabeled data for model training, bridging semi-supervised and supervised learning. Yang et al. [24] automated labeling with parameter migration to avoid retraining, using a dataset of communication signals from seven radiation sources. Ren et al. [25] employed meta-pseudo-labeling, where a teacher model generates labels for student model training on RF fingerprinting tasks. Pseudo-labeling is avoided for its noise sensitivity, unsuitable for our non-ideal operational recognition with varying signal quality.

2.2. Self-Supervised Learning

Self-supervised learning focuses on embedding similarity between augmented samples rather than prediction consistency, learning noise-resistant features without requiring labeled data. Wu et al. [26] combined contrastive learning with pseudo-label selection for AMR, while Wu et al. [27] employed two-stage pretraining for emitter identification. Qiu [28] enhanced classification with UAV and ADS-B datasets. We selected self-supervised learning for its ability to learn robust representations directly from unlabeled signals, effectively handling noise and variations in non-ideal conditions. The SimCLR framework [29] exemplifies this approach by optimizing embedding distances through contrastive learning without label dependencies.

3. Problem Formulation

This paper addresses fine-grained radar work mode recognition under non-ideal conditions using semi-supervised learning. We first define fine-grained radar work modes and mathematically model pulse parameter modulation under both ideal and non-ideal conditions, then present our contrastive learning framework.

3.1. PDW Multi-Level Hierarchical Model

Radar signals are passively intercepted by ESM receivers, where pulse streams undergo deinterleaving to group pulses from the same emitter [30]. Key parameters, including RF, PW, pulse amplitude, angle-of-arrival, and time-of-arrival, are measured and digitized into PDWs. Statistical parameters derived from PDW sequences, such as PRI, collectively characterize the distinctive features of radar emitters. For MFR, employing diverse signal modes across tasks, we formulate work mode recognition as the pattern recognition of PDW combinations. Using RF, PW, and PRI as core PDW parameters, we define radar work modes and propose a two-level hierarchical representation from an implementation perspective [31].
MFR employs a multi-layer signal model to accomplish tasks, relying on a large variety of pulse sequence combinations to achieve higher-level functionalities. This paper models the pulse sequences of MFR from two perspectives: the functional perspective and the modeling perspective, see Figure 1.
From the functional perspective, a radar pulse sequence sample comprises N orderedpulses P = ( p 1 , p 2 , , p N ) . MFR can be configured to schedule diverse pulse sequences for multiple functions, such as search and tracking. P = P A , , P D , , P Z , where P D R P M × N D is the function D [ A , Z ] within the function sequence, which consists of N D pulses, each pulse characterized by P M parameters. MFR can perform multiple missions by altering the parameter modulation style of the transmitted signal, even in one P D . This makes it possible to fulfill various requirements in non-cooperative scenarios, such as tracking, anti-jamming, and improving radar speed and range performance. The subsequences of pulses, composed of parameters with different modulation types, are defined as fine-grained work modes.
From the modeling perspective, a functional sequence sample P D , which may encompass multiple work modes, is defined as a sequence of N D consecutive pulses:
P D = ( p m f , p m f + 1 , , p m f + N D 1 ) R P M × N D ,
where each pulse p t is a vector of P M features (e.g., amplitude, width, and energy), and m f is the starting index of the sequence. This sequence can be partitioned into L contiguous segments. A segment x i , considered as the minimum identification unit in this paper, consists of N k pulses ( N k [ 1 , N D ] ) and serves a specific function and mode:
x i = ( p s i , p s i + 1 , , p s i + N k 1 ) R P M × N k ,
where s i is the start index of the i-th segment. Consequently, the entire sequence sample is described as a collection of these segments: P = ( x 1 , x 2 , , x L ) . Let Y = ( y 1 , y 2 , , y L ) denote the label sequence, where each label y i corresponds to the working pattern of the pulse segment x i .
In summary, the MFR constructs signals in a hierarchical manner to optimize and coordinate lower-level pulse sequences for accomplishing mission goals. At the functional level, a pulse sequence sample P consists of Z functional sequences. At the operational mode level, each functional sequence is further divided into several mode segments, with the corresponding mode labels stored in the label vector Y .

3.2. Modulation Types and Non-Ideal Conditions

Radar systems employ diverse PDW parameter modulations to support multiple functions, including target search, tracking, and missile guidance. We constructed a PDW dataset using three key parameters—RF, PW, and PRI—for radar work mode identification, with modulation patterns determined by mission requirements. RF represents the carrier frequency, enhanced by frequency agility in phased array radars for improved signal processing.PW reflects fundamental radar characteristics, such as short pulses in pulse compression radars.PRI, the interval between pulse arrivals, critically influences range and velocity measurement performance through its inverse relationship with pulse repetition frequency.
For work mode recognition, the modulation modes of PRI and the agility modes of PW and PRI provide critical discriminative information. According to the parameter variation characteristics, the pulse parameter modulation type can be divided into five fundamental forms: constant, jitter, dwell and switch (D-S), stagger, and periodic [14]. Specifically, constant modulation occurs when the parameter fluctuation is less than 1% of the average value; jitter modulation occurs when the fluctuation is more than 30%; D-S modulation is characterized by the parameter value remaining stable in a fixed time period and then jumping; stagger modulation refers to the parameter value around the average value for periodic changes; periodic modulation follows sinusoidal or cosinusoidal patterns.
Additionally, signals from non-cooperative MFRs are often received under non-ideal conditions, including pulse loss, false pulses, and measurement errors. Figure 2 illustrates the impact of these three degradation factors on radar pulses. Consider the RF parameter within the PDWs as an example: the horizontal axis represents the time sequence, the vertical axis indicates RF, and each vertical line denotes an individual pulse. Figure 2a represents the pulse under ideal conditions. Figure 2b shows the case in which the receiver loses the pulse during the measurement. The receiver fails to receive the corresponding pulse properly due to factors such as long distance and mainlobe misalignment, which results in partial signal loss in the received radar pulse data. Figure 2c shows the distortion of the expected sample in terms of RF and PRI. Figure 2d indicates the insertion of false pulses, where height variations correspond to different RF values and color gradients represent PW variations.
Taking PRI as an example, we mathematically model pulse samples under non-ideal conditions through Equations (1) and (2). Let x denote an original pulse sample with its PRI sequence represented as p r i n . When j consecutive pulses are lost after the ith pulse, then the PRI sequence of sample x becomes:
x p r i = p r i n 0 < n < i k = 0 j p r i k n = i p r i n + j n > i
When the kth pulse in a pulse sample is a false pulse, the total pulse count increases, and the PRI values adjacent to the false pulse are altered.
p r i k 1 = g k T O A k 1 p r i k = T O A k + 1 g k
where g k denotes the arrival time of the false pulse.

4. Method

This paper proposes TriSCL (Triple-branch Semi-supervised Contrastive Learning), an end-to-end framework for fine-grained radar work mode recognition under non-ideal conditions using limited labeled samples. As shown in Figure 3, the TriSCL model accepts a semi-labeled dataset consisting of unlabeled and labeled samples. Leveraging the signal characteristics of missing and false pulses, we designed two data enhancement functions. Using these functions for augmentation on unlabeled data produces three distinct views: a tail mask view, a timestamp mask view, and a strongly augmented view (combining both techniques). These views are mapped to an embedding space by encoder f, where the tail mask view and timestamp mask view try to compute the unsupervised comparison loss by instance with the strongly augmented view, respectively. Meanwhile, the labeled samples go through the encoder to compute the supervised contrast loss in the same embedding space, and then through the classifier to compute the classification loss.

4.1. Data Augmentation

To enhance the robustness and generalization capability of radar work mode recognition models under non-ideal conditions, we design two data augmentation techniques for unlabeled samples to simulate pulse loss and false pulse scenarios.
Pulse loss refers to the phenomenon where portions of radar pulses are missing due to noise, interference, or hardware failures. To emulate this condition, we employ a tail masking strategy where the three parameters of the PDW are masked in the posterior part of a sample by a certain percentage. For the input data X R B × T × C , where B is the batch size, T is the number of pulses in a sample, and C is the feature dimension (i.e., the three parameters of the PDW). We calculate the number of pulses to be masked: L = T r , where r is the masking ratio. The masked data X m a s k e d is generated by zeroing the last L pulses across all PDW parameters:
X masked [ b , t , c ] = 0 if t T L X [ b , t , c ] else
False pulses are a non-authentic radar pulse signal that is generated in radar signal processing due to various interferences, noises, or spoofing techniques. To simulate this phenomenon, we adopt a timestamp masking strategy, randomly masking a portion of pulses in a sample pulse train. Specifically, a Bernoulli-distributed mask matrix M 0 , 1 B × T is generated, where each element takes the value 1 (indicating masking) with probability p, and 0 (indicating retention) with probability 1 p . This mask matrix is then applied to the input data X to obtain the masked data X m a s k e d .
X masked [ b , t , c ] = 0 if M b , t = 1 X [ b , t , c ] else
Figure 4 demonstrates the data before and after applying two data augmentation strategies, using the PRI parameter as an example. Through the two strategies for data augmentation, the model significantly enhances its capability to extract discriminative features, improves adaptability to complex signal conditions, and achieves higher recognition accuracy under pulse distortion scenarios.

4.2. Encoder

The encoder will transform input pulse sequences into low-dimensional embedding vectors. These vectors can capture semantic and characteristic information of PDW sequence samples, facilitating subsequent similarity measurement and contrastive loss computation. Different from image processing, signals are usually one-dimensional data. To adapt to the radar pulse dataset, operations such as convolution and pooling used in the encoder are all one-dimensional. Figure 5 shows the structure of the encoder Spatial–Temporal Attention Residual Shrinkage Network (STARSN). The encoder consists mainly of three modules: the convolution unit, the stacked residual shrinkage units, and the spatial–channel attention module. The convolution unit includes a one-dimensional convolutional layer, batch normalization (BN), and the ReLU activation function, which are used to extract the local temporal features of the signal, accelerate convergence speed, and improve computational efficiency. The following provides a detailed introduction to the residual shrinkage module and the spatial–channel attention module adopted in the encoding process.

4.2.1. Spatial and Channel Attention Module

The convolutional operation in neural networks can extract local signal features, but often loses the original temporal information of the signal. While the convolution process treats all channels indiscriminately, each dimension of the PDW parameters contributes differently to radar work mode recognition, necessitating distinct feature extraction strategies. To adaptively emphasize discriminative features while preserving contextual semantic relationships, we firstly introduce the Channel Attention Module (CAM) to guide the encoder network to focus on more discriminative channels and regions. CAM applies global maximum pooling and global average pooling along the feature length dimension of the input feature map x R C × L , where C denotes the number of feature channels and L represents the feature length. This operation compresses the input into two feature vectors of size 1 × C . These vectors are then passed through a shared two-layer fully connected (FC) network with input and output dimensions of C, where ReLU activation is applied between the layers. The outputs of two branches are combined by element-wise addition, followed by a sigmoid activation function to generate the final channel attention weights M c .
M c = σ ( W 2 W 1 f avg c W 2 W 1 f max c )
where σ ( · ) denotes the sigmoid activation function, W 1 and W 2 represent the weight vectors of the two fully connected layers, f avg c and f max c correspond to the global average pooling and global maximum pooling operations applied over the feature map of size L × C , respectively.
Then, the Spatial Attention Module (SAM) adaptively selects weights of different channels and spatial regions, enabling the network to focus on features with greater influence. The SAM compresses the channel information while maintaining attention to the spatial information of the feature map. First, the global average pooling and global maximum pooling are applied in parallel along the channel dimension to the input feature map x R C × L , yielding two feature maps of L × 1 . These two feature maps are then concatenated along the channel dimension to form a merged feature map of L × 2 . Finally, a 3 × 1 convolution operation is employed to fuse the information extracted from both average pooling and maximum pooling, reducing the feature map back to one dimension. The channel attention feature map is obtained after passing through an activation function:
M s = σ Conv 1 D 3 f a v g s , f max s
where C o n v 1 D 3 denotes a 1D convolution with a kernel size of 3 × 1 , f avg s and f max s represent channel-wise global average pooling and global maximum pooling, respectively.

4.2.2. Residual Shrink Module

The primary purpose of introducing this module into the encoder is to improve the ability to extract informative features from noisy samples, eliminate redundant information, and promote classification accuracy. Additionally, by utilizing identity mappings in residual networks, the module facilitates more efficient backpropagation, reduces training difficulty, and mitigates the risk of gradient explosion.
The Residual Shrinkage Module consists of three key components: residual blocks, attention mechanisms, and soft thresholding operations. The residual block is composed of BN, ReLU, and convolution units. These components appear in pairs and also include an identity shortcut connection. Soft thresholding plays a crucial role in many denoising algorithms. Unlike traditional residual networks, the residual shrinkage module embeds an attention subnetwork (SENet), which can adaptively generate thresholds for different samples [32]. Specifically, for a feature map X of size L × C , its absolute value is first compressed into a one-dimensional vector of size 1 × C through global average pooling. This vector is then fed into an attention network composed of a ReLU activation function and two FC layers to generate a scaling parameter α . Finally, α is mapped to the range [0, 1] by the sigmoid function, as expressed by Equation (7):
α = σ ( W 2 W 1 X a v g )
where X denotes the absolute value of the feature map and avg represents the global average pooling operation. W 1 and W 2 are weight matrices of two FC layers. The threshold τ is obtained by multiplying α with X avg .
The soft thresholding operation can identify features that are irrelevant to the current task. It sets the features x in the feature map X whose absolute values are less than the threshold τ to 0, in order to achieve the goal of denoising. It is implemented using Equation (8):
y = x τ if x > τ 0 if τ x τ x + τ if x < τ
where x is the input for soft thresholding, and y is the output. From Equation (8), it can be seen that the derivative of soft thresholding is either 1 or 0. Therefore, soft thresholding can reduce the risks of gradient vanishing and gradient explosion in deep learning and maintain the stability of gradients during the training process.

4.3. Hybrid Loss

For a batch of unlabeled samples X U with batch size N: X U = [ x u 1 , x u 2 , , x u N ] , two augmented versions are generated for each sample through distinct data augmentation operations:
X A U = [ x A u 1 , x A u 2 , , x A u N ]
and
X B U = [ x B u 1 , x B u 2 , , x B u N ]
These augmented samples X A U and X B U will be fed into the feature extraction network and projection network to obtain their embedded representations:
Z A = g ( h ( X A ) ) = [ ( z A 1 ) T , , ( z A N ) T ] T
and
Z B = g ( h ( X B ) ) = [ ( z B 1 ) T , , ( z B N ) T ] T
where z i R M , M denotes the feature dimension. The batch dimensions of Z A and Z B are connected to form Z R 2 N × M . Z is used to compute the following contrastive loss function:
Z = [ ( z A 1 ) T , , ( z A N ) T , ( z B 1 ) T , , ( z B N ) T ] T
Assuming the positive sample feature index for z i is k, the loss function to calculate the similarity between the embedding feature of positive and negative samples is denoted as:
L = 1 2 N i = 1 2 N l i
l i = log exp ( s i , k / τ ) j = 1 , j i 2 N exp ( s i , j / τ )
s i , j = s ( z i , z j ) = z i z j T z i 2 z j 2
where s i , j is the cosine similarity between the features of the ith and the jth sample, τ is the temperature coefficient, and s i , k is the cosine similarity between the features of the ith sample and the features of its positive sample.
The loss function, formulated as a combination of the negative logarithm and SoftMax, drives the inner term towards 1. This property can effectively improve the similarity measure between positive sample pairs while reducing the similarity among negative pairs. Given the bounded range of pairwise similarity, the optimization objective of minimizing the loss function can only be reached when the feature distance between positive samples is reduced, and the separation between negative samples is increased simultaneously.
Within the contrast learning framework, the temperature coefficient is mainly used to control the degree of influence of positive and negative sample pairs in the model training process. Taking negative sample pairs as an example, in the high-dimensional feature space, there are significant differences in the similarity performance between different negative sample pairs, with some negative sample pairs presenting higher similarity and others lower similarity. In this case, a smaller temperature coefficient will make the loss function pay more attention to the pairs of negative samples with high similarity. On the contrary, a larger temperature coefficient will help to mitigate over-emphasis on these similar samples, which, in turn, effectively reduces overfitting risks and improves the model’s generalization ability and robustness.
This paper adopts an end-to-end semi-supervised learning framework. Compared with the two-stage contrastive learning method, the proposed method allows the contrastive loss in unsupervised training to directly fine-tune the downstream classifier. At the same time, it can also make full use of the classification loss of labeled data to achieve the unification of model training and optimization, and improve the recognition accuracy for specific tasks [33].
For each unlabeled data x i U X U , we first employ specialized data augmentation techniques to generate two augmented views and one strongly augmented view that combines both augmentation strategies. The first augmentation applies tail masking with a 30% truncation ratio, processed through encoder f to obtain embedding z i U , while the second employs timestamp masking via binomial sampling (p = 0.2) to produce embedding z j U . The strongly augmented view, combining both augmentation strategies, generates a reference embedding z U through the same encoder. Within the contrastive learning framework, embeddings derived from different views of the same sample form positive pairs, while those from different samples constitute negative pairs. We optimize the model using normalized temperature-scaled cross-entropy loss [29] to minimize distances between positive samples and increase the distance between negative samples. The unsupervised loss L U 1 ( f , X U ) is computed between embeddings z i U and z U , while L U 2 ( f , X U ) is derived from z j U and z U . This dual loss enforces feature consistency across different distortion patterns.
L U 1 ( f , X U ) = 1 N i = 1 N log exp ( sim ( z i U , z U ) / τ ) k = 1 , k i N exp ( sim ( z i U , z k U ) / τ )
L U 2 ( f , X U ) = 1 N j = 1 N log exp ( sim ( z j U , z U ) / τ ) k = 1 , k j N exp ( sim ( z j U , z k U ) / τ )
where sim ( · ) denotes the similarity function, as defined in Equation (16).
For semi-supervised radar work mode recognition tasks, labeled and unlabeled data may originate from the same dataset with identical distributions. For labeled samples, the shared encoder f generates embeddings z L that reside in the same latent space as z i U , z j U , and z U from unlabeled data. The supervised contrastive loss L s ( f , X L , Y ) is computed by treating embeddings from the same class as positive pairs and those from different classes as negative pairs. This objective simultaneously maximizes intra-class similarity and minimizes inter-class similarity of the embeddings. For a labeled PDW sample x i L with ground-truth class y, its embedding z L is obtained through encoder f, and the supervised contrastive loss is formulated as:
L s ( f , X L , Y ) = log y = y + exp ( sim ( z + L , z L ) / T ) y = y exp ( sim ( z L , z L ) / T )
where z + L denotes the embedding of samples belonging to the same class as z L , i.e., y = y + , while z L represents embeddings from different classes, i.e., y y . The optimization of the supervised contrastive loss drives the model to produce more tightly clustered embeddings for intra-class samples while maximizing separation for inter-class instances, ultimately yielding discriminative feature representations that are particularly effective for classification objectives. Subsequently, the labeled sample’s embedding z L is fed into classifier g to predict the label y ^ = g ( z L ) . The standard softmax cross-entropy loss L C ( f , g , X U , Y ) is then computed between y ^ and the ground-truth label y to optimize the model for radar work mode classification. The classifier comprises a single FC layer with an input dimension of 128 (matching the embedding space) and an output dimension of 6 (corresponding to the number of work mode categories), optimized using cross-entropy loss as the classification objective.
L c ( f , g , X U , Y ) = 1 M i = 1 M c = 1 C I ( y i = c ) log y ^ c
where y ^ c represents the predicted probability for class c given a labeled sample, while I ( y i = c ) denotes the indicator function that equals 1 when the true label of the ith labeled sample is c, and 0 otherwise. In terms of the loss function, the unsupervised contrastive loss learns representations by maximizing the consistency between different augmented views of the same sample. Supervised contrastive learning utilizes the supervised contrastive loss to reduce the distance between samples of the same class in the embedding space. To jointly train this end-to-end framework, the final loss function is a hybrid loss that combines weighted contributions from the unsupervised contrastive loss, supervised contrastive loss, and classification loss.
L = λ 1 L U 1 + λ 2 L U 2 + λ 3 L S + λ 4 L C
where L U 1 and L U 2 denote the unsupervised contrastive losses, L S represents the supervised contrastive loss, and L C is the classification loss. The hyperparameters λ 1 , λ 2 , λ 3 , and λ 4 are introduced to balance the contributions of these four loss terms.
During training, the model updates the encoder f and classifier g by minimizing the hybrid loss, allowing it to learn effective representations from both labeled and unlabeled data while directly optimizing predicted class probabilities to enhance classification performance.

4.4. Materials and Datasets

As shown in Table 1, this section conducts simulation experiments using three different datasets, named PDW-N, PDW-MP, and PDW-FP, respectively, where N stands for “Normal”, MP for “Missing Pulse”, and FP for “False Pulse”. Each dataset comprises six categories of fine-grained modes. The PDW-MP and PDW-FP datasets are constructed based on PDW-N to simulate signal quality degradation in practical electronic warfare environments. PDW-MP emulates the missing pulse phenomenon, where pulses are randomly removed from original sequences to replicate conditions caused by receiver insensitivity, electromagnetic interference, or signal occlusion. This configuration evaluates the model’s reasoning capability under incomplete information. In contrast, PDW-FP simulates deceptive false pulse jamming through the injection of noise or interference pulses into clean sequences, representing adversarial electronic countermeasures. This setup assesses anti-jamming robustness and classification stability in noisy and deceptive scenarios. Each pulse is characterized by three modulation parameters, i.e., RF, PW, and PRI. The details of modulation types and parameter ranges considered in this paper are listed in Table 1. Specifically, constant modulation occurs when fluctuation remains below 1% of the average value. Jitter modulation is characterized by fluctuations exceeding 30% of the average value. D-S modulation is manifested as the parameter value remains stable in a fixed time period and then jumps. Stagger modulation refers to the parameter value around the average value for periodic changes. Periodic modulation follows sinusoidal or cosinusoidal patterns.
Each category of radar work mode in the dataset contains 1000 samples, and each sample is a sequence composed of 4000 pulses. In the PDW-N dataset, the measurement errors of PW and PRI are set to 1%, and the measurement error of RF is ±5 MHz. In the simulation of non-ideal conditions, six scenarios are simulated by using the probability ranges of the occurrence of missed pulses and false pulses, ranging from 0 to 50%, with a step size of 10%. Figure 6 provides some examples of three datasets.

5. Experiments and Analysis

5.1. Experimental Configuration

Before training, the samples need to be preprocessed to eliminate the influence of the dimensions among different features. The mean value and standard deviation are calculated for all the samples in the training set, validation set, and test set, and the data is transformed into a distribution with a mean of 0 and a standard deviation of 1 by using Equation (22).
x scaled = x μ σ
The STARSN is composed of adding a residual shrinkage module and a spatial–channel attention mechanism on the basis of ResNet18. Among them, the number of filters in the convolutional layers is 16, 32, and 128, respectively, the size of the convolution kernel is 3 × 1 , and the stride is 1.
The total number of training epochs is set to 100, the batch size is set to 100, temperature coefficient is set to 0.5. We use the AdamW optimizer with an initial learning rate set to 0.001 and a weight decay coefficient of 10 5 . In addition, we introduce a learning rate exponential decay scheduler with warm-up. The warm-up period is 15 epochs, and the exponential decay coefficient is 0.95. In the following experiments, the ratio of labeled to unlabeled data under the default conditions is 1:10. All experimental methods were trained on the PDW-N dataset, where the ratio of the training set to the validation set is 7:3, while evaluated on the PDW-MP and PDW-FP datasets to assess the model’s generalization capability under non-ideal scenarios. The balancing hyperparameters ( λ 1 , λ 2 , λ 3 , and λ 4 ) in the hybrid loss1 (Equation (21)) were manually tuned to ensure comparable magnitudes across all loss terms, guided by established practices in contrastive learning. The final values ( λ 1 = 0.3, λ 2 = 0.3, λ 3 = 1.0, λ 4 = 1.0), which achieved optimal performance on the validation set, were adopted in our experiments.

5.2. Comparison Experiment

For a comprehensive comparative analysis, the proposed model was evaluated against an extensive set of methods spanning different learning paradigms. The comparison includes representative contrastive learning approaches such as SimCLR [29], BYOL [34], and MOCO [35]. Additionally, classical supervised convolutional networks were considered, including AlexNet [36], VGGNet, ConvNet [37], and ResNet [38]. To ensure a thorough evaluation, the study also incorporates recently developed radar-specific recognition methods, particularly RSCN [12] for noise suppression and BABN [14] for Bayesian uncertainty modeling. To adapt to the radar pulse dataset, all convolutional and pooling operations in these networks were modified to one-dimensional implementations. The supervised methods were trained exclusively using limited labeled samples with cross-entropy loss.
To ensure fair comparison, the training configurations of SimCLR, BYOL, MOCO, RSCN, and BABN were kept consistent with TriSCL. Their data augmentation strategies employed two approaches commonly used for temporal data: scaling augmentation and jitter augmentation. The weak augmentation involved only scaling with a standard deviation factor of 1.1, while the strong augmentation first applied scaling ( β = 1.1 ), followed by additive Gaussian noise ( β = 0.8 ). All algorithms adopted identical data preprocessing methods.
The confusion matrices derived from the training, validation, and challenging test sets (PDW-MP and PDW-FP with 10% distortion) are analyzed to evaluate generalization performance and overfitting, as illustrated in Figure 7. All matrices exhibit a near-diagonal distribution, indicating high classification accuracy. The consistent performance and structural alignment between training and validation sets confirm effective generalization without significant overfitting. Notably, the model achieves 100% accuracy on the test sets containing pulse distortions, demonstrating substantial robustness to interference. This performance is attributed to the integrated framework design: tailored data augmentations emulate realistic signal distortions, while the robust encoder architecture, incorporating residual shrinkage and attention mechanisms, facilitates the extraction of distortion-invariant features. The experimental results collectively validate the generalization capability of the proposed approach and its resilience to targeted signal degradation scenarios.
Table 2 demonstrates the recognition accuracy of these algorithms for multifunctional radar work modes under various non-ideal scenarios. When the proportion of missing pulses is below 20%, the recognition accuracy of our algorithm can reach 100%. As the proportions of missing and false pulses increase, recognition accuracy declines. This occurs because higher distortion proportions compromise discriminative information in the samples and amplify interference and uncertainty during model evaluation. When the proportion of missing pulses is 30%, TriSCL can still maintain a relatively high recognition accuracy for both false pulses and missing pulses, while the accuracies of other algorithms drop significantly. Under high proportions of missing pulses and false pulses, our algorithm still outperforms other supervised and semi-supervised algorithms. Although ResNet shows comparable performance to our method at extremely high interference ratios, our model has a great advantage over ResNet when the proportion is 30% or below. This indicates that the proposed TriSCL can better adapt to the complex electromagnetic environment and enable the model to obtain robust features under non-ideal conditions. At the same time, we found that all models show lower accuracy for missing pulses than for false pulses. This is mainly because missing pulses cause irreversible loss of key signal features, leaving insufficient discriminative patterns for the model to learn; False pulses, while disruptive, still retain authentic signal characteristics that provide learnable features.
The confusion matrix analysis demonstrates that the proposed model exhibits confusion between Mode 5 and Mode 6 under 30% pulse missing conditions in Figure 8. As indicated in the parameter configuration Table 1, this misclassification arises from the inherent similarities in their operational parameters: both modes share D-S and Stagger modulation types. Furthermore, 30% pulse missing renders the periodic modulation (in Mode 5) and jitter modulation (in Mode 6) temporally similar in characteristic variations, thereby exacerbating the classification challenge. To mitigate this issue, future work should incorporate a parameter-aware attention mechanism within the feature space, which would adaptively weight the feature representations of different PDW parameters to enhance discriminative capability.
For comprehensive performance assessment, we evaluate our method using three key metrics: classification accuracy, the number of trainable parameters, and Floating Point Operations (FLOPs). All experiments were conducted on the PDW-MP dataset with a 30% missing ratio. It is evident that, compared with other methods in Table 3, the method proposed in this paper has achieved the best performance in terms of performance metrics. While the triple-branch contrastive learning architecture incurs higher computational costs, it also delivers significantly improved accuracy. TriSCL does not have the highest number of parameters or FLOPs; yet, it achieves the highest accuracy, indicating that its model architecture and training strategy are more effective in balancing computational cost and accuracy. Overall, while our method exhibits moderate complexity, it remains feasible with current computing resources and significantly improves performance metrics.

5.3. Ablation Study

Table 4 shows the influence of the temperature coefficient on the model’s recognition ability. When the temperature coefficient is 0.6, the model can achieve the best recognition performance, indicating that this parameter value can effectively enhance the model’s adaptability to the environment under non-ideal conditions. Based on this, a temperature coefficient of 0.6 is adopted in subsequent experiments to ensure the model’s robustness.
We analyze the impact of different data augmentation methods within contrastive learning frameworks, comparing our two proposed techniques with common time-series augmentations such as scaling and jitter. As illustrated in Table 5, compared to the proposed methods (tail mask + timestamp mask), the conventional augmentation techniques (scaling + jitter) exhibit inferior overall performance when handling non-ideal scenarios in radar PDW datasets, particularly under high proportions of false and missing pulses. The conventional augmentation methods, although capable of improving the robustness to the value range of PDW features, have significant limitations: they only perform independent perturbations on the value range of pulse features, failing to preserve the inherent temporal correlations and variation trends among pulse sequences. Electromagnetic interference and other noise sources commonly distort pulse sequences during radar receiver operation. The data augmentation methods proposed in this paper simulate these typical non-ideal conditions: Tail Mask simulates the phenomenon of pulse loss, and Timestamp Mask simulates the phenomenon of temporal misalignment. This physics-based augmentation strategy enables the model not only to learn the fault tolerance for feature values but, more importantly, to capture the essential laws of pulse feature variation patterns, thereby maintaining stable recognition performance even under the interference of false pulses and missing pulses.
Next, to validate the impact of the residual shrinkage module and the spatial–channel attention module on the model’s recognition ability, we constructed four comparative models by sequentially removing the attention mechanism, residual shrinkage module, or both. As shown in Figure 9, the recognition performance degrades when these key modules are removed. Moreover, using the residual shrinkage or the spatial–channel attention mechanism module alone can improve the classification accuracy, demonstrating that both the residual shrinkage module and attention mechanism individually contribute to improved classification accuracy. The main reason is that the soft-thresholding operation can adaptively suppress noise, and the spatial–channel attention mechanism emphasizes important feature regions and channels while suppressing irrelevant backgrounds. Experimental results under varying false pulse ratios reveal that the model without the residual shrinkage module and the model without the attention mechanism exhibit comparable performance when the false pulse ratio remains below 20%, with their accuracy declining to 83.33% as the ratio increases. In the recognition results for missing pulses with different proportions, the model without the attention mechanism shows marginally inferior performance compared to the model without the residual shrinkage module. This suggests that the attention mechanism can partially compensate for the absence of the residual shrinkage module. The model with both modules removed demonstrates the poorest performance, confirming the crucial synergistic effect between the residual shrinkage and attention mechanisms. Overall, both modules play vital roles in enhancing the model’s robustness against non-ideal conditions, including varying false pulse ratios and missing pulses.
To evaluate the impact of the number of contrastive learning branches on the recognition performance of the model under various non-ideal conditions, we compared experimental results between two-branch and three-branch configurations. To ensure the validity of the experimental results, except for the loss functions, all other aspects of the three models were kept consistent. The experimental results are shown in Table 6. Among them, the first variant of TriSCL uses only the unsupervised contrastive loss L U 1 , while the second variant employs the unsupervised contrastive loss L U 2 solely. Both variants, along with the full TriSCL model, exhibit strong performance for false pulses with varying proportions and missed pulses at relatively low proportions, and still hold certain advantages over other semi-supervised methods. When the proportions of false pulses and missed pulses reach 30% or higher, the recognition results of the two variants are relatively close, but both lag behind the three-branch TriSCL model by a certain margin. When the proportion of missed pulses is 30%, the recognition result of TriSCL drops to 0.8333, which is higher than the 0.5 achieved by both variants. Under severe non-ideal conditions, the three-branch architecture demonstrated enhanced stability and recognition capability. This improvement stems from the third branch, which employs a combination of two data augmentations (compared to one in each of the other branches), acting as an extra constraint that generates more contrastive relationships during loss computation. Specifically, it enables dual contrastive calculations between each single-augmentation branch and the composite-augmentation branch, creating multi-layered constraints involving associations between single pulse anomalies (missing or false pulses) and composite pulse anomalies, as well as indirect links between different pulse anomalies through composite scenarios. When calculating contrastive loss, the three branches generate richer contrastive relationships that cover not only individual pulse anomaly patterns but also their combined effects. This prompts the model to learn sample distributions from multiple feature-space dimensions, enabling more generalizable representations that capture the essential patterns amid complex non-ideal conditions. Consequently, performance improves under complex non-ideal conditions—effectively handling false and missing pulses.
To validate the impact of two-stage contrastive learning versus our proposed end-to-end strategy, we conducted comparative experiments on TriSCL using both training paradigms. The two-stage approach (TriSCL-TS) employs triple-branch contrastive learning with dual unsupervised losses ( L U 1 and L U 2 ) during pretraining, and transitions to supervised contrastive loss combined with classification loss during fine-tuning. Experimental results in Table 7 demonstrate that the proposed end-to-end TriSCL exhibits excellent robustness, outperforming TriSCL-TS in scenarios with low to medium proportions of false pulses and missed pulses. However, under high proportions of non-ideal conditions, the performance of all models declines; yet, TriSCL remains relatively superior. The key reason for this performance gap is the end-to-end strategy’s superior ability to leverage labeled data. By integrating discriminative feature learning and adaptation to non-ideal scenarios into a unified optimization process, it avoids potential inconsistency between pretraining and fine-tuning stages. This enables the model to maintain stable performance across different proportion ranges of false and missing pulses. In contrast, the performance of TriSCL-TS highlights the limitations of the two-stage training method under the current experimental setup: its separate pretraining (relying solely on unsupervised losses) and fine-tuning stages may lead to a mismatch in feature learning objectives, making it less effective in capturing the complex patterns of pulse anomalies, especially when dealing with varying degrees of non-ideal conditions.
As illustrated in Figure 10, recognition performance across all test scenarios significantly improves as the number of labeled samples increases. This occurs because additional labeled data provides the model with richer information about the sample distribution, enabling more discriminative feature extraction. Notably, TriSCL maintains a relatively high recognition accuracy even with extremely limited labeled samples, demonstrating its strong semi-supervised learning capability.

6. Conclusions and Future Work

Recognizing fine-grained multifunctional radar modes with limited labeled data is a practically significant, yet challenging task. To address this, we propose a semi-supervised deep learning framework based on triple-branch contrastive learning. This unified framework effectively combines unsupervised contrastive loss, supervised contrastive loss, and classification loss. To bolster robustness against sample quality degradation—specifically false pulses and lost pulses—we introduce tail masking and timestamp masking as novel data augmentation strategies. Furthermore, our encoder incorporates residual shrinkage modules and self-attention mechanisms to adaptively suppress noise and capture key temporal patterns and inter-feature dependencies. Experimental results demonstrate that our method outperforms state-of-the-art supervised and unsupervised approaches, achieving an average accuracy improvement of 15% across varying levels of sample degradation, even when using only 10% labeled data. These results confirm the effectiveness and robustness of the proposed approach.
Future research in radar mode recognition will confront the challenges posed by complex programmable radars, shifting focus to few-shot and zero-shot learning paradigms. This demands models that can generalize to unseen signals by extracting intrinsic features autonomously. Potential paths include metric learning and data augmentation for few-shot learning, graph networks for zero-shot recognition, and hybrid approaches combining sequence matching or envelope analysis with deep models to improve robustness and interpretability. These efforts, though nascent, are crucial for developing systems that operate under authentic battlefield constraints.

Author Contributions

Conceptualization, P.S.; methodology, P.S.; software, P.S. and M.D.; validation, P.S. and M.D.; formal analysis, J.S. and M.D.; investigation, P.S.; resources, J.S. and Z.L.; data curation, P.S. and M.D.; writing—original draft preparation, P.S.; writing—review and editing, M.D., Z.L. and J.S.; visualization, X.C.; supervision, J.S. and Z.L.; project administration, M.D.; funding acquisition, M.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant 62301578.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
EWElectronic Warfare
ESElectronic Support
EAElectronic Attack
EPElectronic Protection
MFRMultifunction Radar
RFRadio Frequency
PWPulse Width
PRIPulse Repetition Interval
LSTMLong Short-Term Memory
CNNConvolutional Neural Network
TriSCLTri-branch Semi-supervised Contrastive Learning
SEISpecific Emitter Identification
DCRDual Consistency Regularization
SSLSemi-Supervised Learning
ADS-BAutomatic Dependent Surveillance–Broadcast
GANGenerative Adversarial Network
SimCLRA Simple Framework for Contrastive Learning of Visual Representations
AMRAutomatic Modulation Recognition
UAVUnmanned Aerial Vehicle
ESMElectronic Support Measures
D-SDwell and Switch
STARSNSpatial-Temporal Attention Residual Shrinkage Network
BNBatch Normalization
ReLURectified Linear Unit
CAMChannel Attention Module
FCFully Connected
SAMSpatial Attention Module
SENetSqueeze-and-Excitation Network

References

  1. Schleher, D.C. Electronic Warfare in the Information Age; Artech House: Norwood, MA, USA, 1999. [Google Scholar]
  2. Du, M.; Pan, J.; Bi, D. A Contrastive Learner for Automatic Modulation Classification. IEEE Trans. Wirel. Commun. 2025, 24, 3575–3589. [Google Scholar] [CrossRef]
  3. Weber, M.E.; Cho, J.Y.; Thomas, H.G. Command and control for multifunction phased array radar. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5899–5912. [Google Scholar] [CrossRef]
  4. Stailey, J.E.; Hondl, K.D. Multifunction phased array radar for aircraft and weather surveillance. Proc. IEEE 2016, 104, 649–659. [Google Scholar] [CrossRef]
  5. Mir, H.S.; Guitouni, A. Variable dwell time task scheduling for multifunction radar. IEEE Trans. Autom. Sci. Eng. 2013, 11, 463–472. [Google Scholar] [CrossRef]
  6. Boers, Y.; Driessen, H.; Zwaga, J. Adaptive MFR parameter control: Fixed against variable probabilities of detection. IEE Proc.-Radar, Sonar Navig. 2006, 153, 2–6. [Google Scholar] [CrossRef]
  7. Richard, G.; Wiley, E. The Interception and Analysis of Radar Signals; Artech House: Boston, MA, USA, 2006. [Google Scholar]
  8. Revillon, G.; Mohammad-Djafari, A.; Enderli, C. Radar emitters classification and clustering with a scale mixture of normal distributions. IET Radar Sonar Navig. 2019, 13, 128–138. [Google Scholar] [CrossRef]
  9. Zhang, Z.; Long, K.; Wang, J. Self-organization paradigms and optimization approaches for cognitive radio technologies: A survey. IEEE Wirel. Commun. 2013, 20, 36–42. [Google Scholar] [CrossRef]
  10. Zhai, Q.; Li, Y.; Zhang, Z.; Li, Y.; Wang, S. Few-shot recognition of multifunction radar modes via refined prototypical random walk network. IEEE Trans. Aerosp. Electron. Syst. 2022, 59, 2376–2387. [Google Scholar] [CrossRef]
  11. He, C.; Zhang, L.; Wei, S.; Fang, Y. Multifunction Radar Working Mode Recognition with Unsupervised Hierarchical Modeling and Functional Semantics Embedding Based LSTM. IEEE Sens. J. 2024, 24, 22698–22710. [Google Scholar] [CrossRef]
  12. Zhang, Y.; Huo, W.; Zhang, C.; Huang, Y.; Pei, J.; Zhang, Y.; Yang, J. Radar operation mode recognition via multifeature residual-and-shrinkage ConvNet. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 6073–6084. [Google Scholar] [CrossRef]
  13. Xiong, J.; Pan, J.; Du, M. A Cascade Network for Pattern Recognition Based on Radar Signal Characteristics in Noisy Environments. Remote Sens. 2023, 15, 4083. [Google Scholar] [CrossRef]
  14. Du, M.; Zhong, P.; Cai, X.; Bi, D.; Jing, A. Robust Bayesian attention belief network for radar work mode recognition. Digit. Signal Process. 2023, 133, 103874. [Google Scholar] [CrossRef]
  15. He, B.; Wang, F. Specific emitter identification via sparse Bayesian learning versus model-agnostic meta-learning. IEEE Trans. Inf. Forensics Secur. 2023, 18, 3677–3691. [Google Scholar] [CrossRef]
  16. Fu, X.; Shi, S.; Wang, Y.; Lin, Y.; Gui, G.; Dobre, O.A.; Mao, S. Semi-supervised specific emitter identification via dual consistency regularization. IEEE Internet Things J. 2023, 10, 19257–19269. [Google Scholar] [CrossRef]
  17. Xu, Q.; Liu, L.; Zhou, X. Boosting Semi-Supervised Specific Emitter Identification with Curriculum Triple Consistency Regularization. In Proceedings of the 2024 International Joint Conference on Neural Networks (IJCNN), Yokohama, Japan, 30 June–5 July 2024; pp. 1–8. [Google Scholar]
  18. Liu, L.; Sun, S.; Zhi, S.; Shi, F.; Liu, Z.; Heikkilä, J.; Liu, Y. A Causal Adjustment Module for Debiasing Scene Graph Generation. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 4024–4043. [Google Scholar] [CrossRef]
  19. Gong, J.; Xu, X.; Qin, Y.; Dong, W. A generative adversarial network based framework for specific emitter characterization and identification. In Proceedings of the 2019 11th International Conference on Wireless Communications and Signal Processing (WCSP), Xi’an, China, 23–25 October 2019; pp. 1–6. [Google Scholar]
  20. Li, C.; Xu, T.; Zhu, J.; Zhang, B. Triple generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems 30, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  21. Tan, K.; Yan, W.; Zhang, L.; Ling, Q.; Xu, C. Semi-supervised specific emitter identification based on bispectrum feature extraction CGAN in multiple communication scenarios. IEEE Trans. Aerosp. Electron. Syst. 2022, 59, 292–310. [Google Scholar] [CrossRef]
  22. Xie, C.; Zhang, L.; Zhong, Z. Few-shot unsupervised specific emitter identification based on density peak clustering algorithm and meta-learning. IEEE Sens. J. 2022, 22, 18008–18020. [Google Scholar] [CrossRef]
  23. Zhou, Z.; Huang, G.; Chen, H.; Gao, J. Automatic radar waveform recognition based on deep convolutional denoising auto-encoders. Circuits Syst. Signal Process. 2018, 37, 4034–4048. [Google Scholar] [CrossRef]
  24. Yang, H.; Zhang, H.; Wang, H.; Guo, Z. A novel approach for unlabeled samples in radiation source identification. J. Syst. Eng. Electron. 2022, 33, 354–359. [Google Scholar] [CrossRef]
  25. Ren, Z.; Ren, P.; Zhang, T. Deep RF device fingerprinting by semi-supervised learning with meta pseudo time-frequency labels. In Proceedings of the 2022 IEEE Wireless Communications and Networking Conference (WCNC), Austin, TX, USA, 10–13 April 2022; pp. 2369–2374. [Google Scholar]
  26. Wu, D.; Shi, J.; Li, Z.; Du, M.; Liu, F.; Zeng, F. Contrastive Semi-supervised Learning with Pseudo-label for Radar Signal Automatic Modulation Recognition. IEEE Sens. J. 2024, 24, 30399–30411. [Google Scholar] [CrossRef]
  27. Wu, Z.; Wang, F.; He, B. Specific emitter identification via contrastive learning. IEEE Commun. Lett. 2023, 27, 1160–1164. [Google Scholar] [CrossRef]
  28. Qiu, J.; Chen, B.; Song, D.; Wang, W. Semi-Supervised Specific Emitter Identification Based on Contrastive Learning and Data Augmentation. IEEE Trans. Aerosp. Electron. Syst. 2025, 61, 8449–8466. [Google Scholar] [CrossRef]
  29. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, PmLR, Virtual, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
  30. Granger, E.; Rubin, M.A.; Grossberg, S.; Lavoie, P. A what-and-where fusion neural network for recognition and tracking of multiple radar emitters. Neural Netw. 2001, 14, 325–344. [Google Scholar] [CrossRef]
  31. Zhai, Q.; Zhang, X.; Zhang, Z.; Liu, J.; Wang, S. Online few-shot learning for multi-function radars mode recognition based on backtracking contextual prototypical memory. Digit. Signal Process. 2023, 141, 104189. [Google Scholar] [CrossRef]
  32. Zhao, M.; Zhong, S.; Fu, X.; Tang, B.; Pecht, M. Deep residual shrinkage networks for fault diagnosis. IEEE Trans. Ind. Inform. 2019, 16, 4681–4690. [Google Scholar] [CrossRef]
  33. Cai, H.; Zhang, X.; Liu, X. Semi-Supervised End-To-End Contrastive Learning For Time Series Classification. arXiv 2023, arXiv:2310.08848. [Google Scholar]
  34. Grill, J.B.; Strub, F.; Altché, F.; Tallec, C.; Richemond, P.; Buchatskaya, E.; Doersch, C.; Avila Pires, B.; Guo, Z.; Gheshlaghi Azar, M.; et al. Bootstrap your own latent-a new approach to self-supervised learning. Adv. Neural Inf. Process. Syst. 2020, 33, 21271–21284. [Google Scholar]
  35. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 9729–9738. [Google Scholar]
  36. Limin, G.; Xin, C. Low probability of intercept radar signal recognition based on the improved AlexNet model. In Proceedings of the 2nd International Conference on Digital Signal Processing, Tokyo, Japan, 25–27 February 2018; pp. 119–124. [Google Scholar]
  37. Li, X.; Huang, Z.; Wang, F.; Wang, X.; Liu, T. Toward convolutional neural networks on pulse repetition interval modulation recognition. IEEE Commun. Lett. 2018, 22, 2286–2289. [Google Scholar] [CrossRef]
  38. Han, J.W.; Park, C.H. A unified method for deinterleaving and PRI modulation recognition of radar pulses based on deep neural networks. IEEE Access 2021, 9, 89360–89375. [Google Scholar] [CrossRef]
Figure 1. Hierarchical modeling diagram of work modes.
Figure 1. Hierarchical modeling diagram of work modes.
Sensors 25 07440 g001
Figure 2. Example of three factors in a non-ideal condition influencing the expected sample of a specific mode. (a) Ideal condition. (b) Pulse missing. (c) Measurement errors. (d) False pulse.
Figure 2. Example of three factors in a non-ideal condition influencing the expected sample of a specific mode. (a) Ideal condition. (b) Pulse missing. (c) Measurement errors. (d) False pulse.
Sensors 25 07440 g002
Figure 3. Framework of the proposed approach for the MFR work mode recognition.
Figure 3. Framework of the proposed approach for the MFR work mode recognition.
Sensors 25 07440 g003
Figure 4. The PRI parameters before and after two types of enhancement.
Figure 4. The PRI parameters before and after two types of enhancement.
Sensors 25 07440 g004
Figure 5. The architecture of the SCARSN encoder.
Figure 5. The architecture of the SCARSN encoder.
Sensors 25 07440 g005
Figure 6. Several examples of three datasets.
Figure 6. Several examples of three datasets.
Sensors 25 07440 g006
Figure 7. Confusion matrices demonstrating model performance and robustness. (a,b) show consistent performance between training and validation, indicating no overfitting. (c,d) demonstrate perfect robustness when handling both PDW-MP and PDW-FP datasets under 10% distortion conditions.
Figure 7. Confusion matrices demonstrating model performance and robustness. (a,b) show consistent performance between training and validation, indicating no overfitting. (c,d) demonstrate perfect robustness when handling both PDW-MP and PDW-FP datasets under 10% distortion conditions.
Sensors 25 07440 g007
Figure 8. Confusion matrices of TriSCL with the missing ratio of 30% on different datasets.
Figure 8. Confusion matrices of TriSCL with the missing ratio of 30% on different datasets.
Sensors 25 07440 g008
Figure 9. The influence of residual shrinkage and attention mechanism on recognition results.
Figure 9. The influence of residual shrinkage and attention mechanism on recognition results.
Sensors 25 07440 g009
Figure 10. Experimental results under different labeled sample ratios.
Figure 10. Experimental results under different labeled sample ratios.
Sensors 25 07440 g010
Table 1. Parameter settings of simulated work modes.
Table 1. Parameter settings of simulated work modes.
Work ModePRIRFPW
20∼50 μs4000∼5500 MHz1∼8 μs
MODE1D-SD-SConstant
MODE2D-SStaggerStagger
MODE3ConstantD-SD-S
MODE4ConstantStaggerConstant
MODE5PeriodicD-SStagger
MODE6JitterStaggerD-S
Table 2. Recognition results of different DL methods under pulse loss and false pulse conditions.
Table 2. Recognition results of different DL methods under pulse loss and false pulse conditions.
Compared MethodsACC (False Pulse (%))ACC (Missing Pulse (%))
0102030405001020304050
AlexNet10051505049.733.31005050504835.5
ResNet10010010010099.999.510010066.7635049.8
CovNet10010010010010010010083.366.4505048
VGG10071.467.450.3505010049.733.3333030
SimCLR10010066.766.760.247.110083.366.766.733.330
BYOL10033.216.716.116.116.110016.116.116.11616
MOCO10010010083.465.11710066.766502016.7
RSCN10010099.883.383.383.310083.366.766.766.131.5
BABN1008032.223.721.6201005031.23017.917.1
TriSCL10010010010010010010010010083.35050
Note: Bold entries represent the results obtained by the method (TriSCL) proposed in this paper.
Table 3. Complexity comparison with other recognition methods.
Table 3. Complexity comparison with other recognition methods.
MethodsParametersFLOPs ( × 10 6 )Accuracy
AlexNet18,7104.420.50
ResNet10,39042.430.67
LeNet359010.750.50
VGG12,54214.110.30
SimCLR60,11839.280.67
BYOL52,37442.440.17
MOCO35,60685.430.50
RSCD19,25442.440.67
BABN12,39639.210.30
TriSCL58,18952.620.83
Note: Bold values correspond to the results of TriSCL, the method proposed in this study.
Table 4. Recognition results under different temperature coefficients.
Table 4. Recognition results under different temperature coefficients.
Temperature CoefficientFalse Pulse (%)Missing Pulse (%)
0102030405001020304050
0.110010010010099.983.310010096.233.333.333.3
0.210010010010098.188.310010010033.333.333.3
0.410010010010010099.710010010070.30.50.5
0.510010010010010010010010010078.30.50.5
0.610010010010010010010010010083.30.50.5
0.810010010010010010010010099.883.30.50.5
1.010010010010010010010010097.983.30.50.5
Note: The bolded row indicates the temperature coefficient value (0.6) selected in the experiments of this paper.
Table 5. Recognition results of different data augmentation methods.
Table 5. Recognition results of different data augmentation methods.
MethodFalse Pulse (%)Missing Pulse (%)
0102030405001020304050
scale + jitter10010083.383.383.383.310083.366.766.750.050.0
tail mask + timestamp mask10010010010010010010010010083.350.050.0
Note: The bolded row represents the results obtained by the data augmentation function proposed and adopted in this paper.
Table 6. The influence of the number of branches on recognition results.
Table 6. The influence of the number of branches on recognition results.
MethodFalse Pulse (%)Missing Pulse (%)
0102030405001020304050
TriSCL- L U 1 10010010010010089.410010097.45046.133.3
TriSCL- L U 2 10010010010010089.510010097.350.046.233.3
TriSCL10010010010010010010010010083.350.050.0
Note: The bolded row represents the experimental results of the model proposed in this paper (TriSCL).
Table 7. The influence of end-to-end training approaches on experimental results.
Table 7. The influence of end-to-end training approaches on experimental results.
MethodFalse Pulse (%)Missing Pulse (%)
0102030405001020304050
TriSCL-TS16.716.716.716.716.716.716.716.716.711.411.411.4
TriSCL10010010010010010010010010083.350.050.0
Note: The bolded row represents the experimental results of the model proposed in this paper (TriSCL).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, P.; Du, M.; Li, Z.; Chen, X.; Shi, J. Semi-Supervised Radar Work Mode Recognition Based on Contrastive Learning. Sensors 2025, 25, 7440. https://doi.org/10.3390/s25247440

AMA Style

Sun P, Du M, Li Z, Chen X, Shi J. Semi-Supervised Radar Work Mode Recognition Based on Contrastive Learning. Sensors. 2025; 25(24):7440. https://doi.org/10.3390/s25247440

Chicago/Turabian Style

Sun, Peishan, Mingyang Du, Zhihui Li, Xuan Chen, and Junpeng Shi. 2025. "Semi-Supervised Radar Work Mode Recognition Based on Contrastive Learning" Sensors 25, no. 24: 7440. https://doi.org/10.3390/s25247440

APA Style

Sun, P., Du, M., Li, Z., Chen, X., & Shi, J. (2025). Semi-Supervised Radar Work Mode Recognition Based on Contrastive Learning. Sensors, 25(24), 7440. https://doi.org/10.3390/s25247440

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop