Next Article in Journal
Monitoring of Spatio-Temporal Variations of Oil Slicks via the Collocation of Multi-Source Satellite Images
Previous Article in Journal
Earth Observation Multi-Spectral Image Fusion with Transformers for Sentinel-2 and Sentinel-3 Using Synthetic Training Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

AOHDL: Adversarial Optimized Hybrid Deep Learning Design for Preventing Attack in Radar Target Detection

School of Electronics and Information, Northwestern Polytechnical University, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(16), 3109; https://doi.org/10.3390/rs16163109
Submission received: 10 July 2024 / Revised: 15 August 2024 / Accepted: 20 August 2024 / Published: 22 August 2024

Abstract

In autonomous driving, Frequency-Modulated Continuous-Wave (FMCW) radar has gained widespread acceptance for target detection due to its resilience and dependability under diverse weather and illumination circumstances. Although deep learning radar target identification models have seen fast improvement, there is a lack of research on their susceptibility to adversarial attacks. Various spoofing attack techniques have been suggested to target radar sensors by deliberately sending certain signals through specialized devices. In this paper, we proposed a new adversarial deep learning network for spoofing attacks in radar target detection (RTD). Multi-level adversarial attack prevention using deep learning is designed for the coherence pulse deep feature map from DAALnet and Range-Doppler (RD) map from TDDLnet. After the discrimination of the attack, optimization of hybrid deep learning (OHDL) integrated with enhanced PSO is used to predict the range and velocity of the target. Simulations are performed to evaluate the sensitivity of AOHDL for different radar environment configurations. RMSE of AOHDL is almost the same as OHDL without attack conditions and it outperforms the earlier RTD implementations.

1. Introduction

The main processing mechanism of Frequency-Modulated Continuous-Wave (FMCW) radar functions transmitting a sequence of linearly modulated frequency signals over time is referred to as a frame. The received signal is combined with a duplicate of the broadcast signal sequence after the process of reflection from the target. This creates a sinusoidal wave whose frequency specifies the distance to the reflecting target [1]. Through the analysis of the phase shifts of this sinusoidal wave in various chirps, the system estimates the relative velocity of the target. In addition, an FMCW radar can determine the direction of an object by analyzing the phase shifts in the middle of elements and the array of antennas [2].
Radars operating using millimeter-wave (mmWave) signals have greatly enhanced radar accuracy [3], offering advantages over other sensors for implementation in train and test decision-based driver guiding system (Automated-DAS) and autonomous vehicle (AV) fields [4]. Automated DAS systems at present utilize radars for various functions such as versatile critical control, front crash avoidance, and lane tracing systems [5]. In more sophisticated systems, these radars can be utilized for a range of image tests by supplying three-dimensional coordinate data to computer vision systems contained by autonomous vehicles [4]. In addition to the vehicle industry, these radars can offer accurate sensing for built-up air transportation applications [6]. As the use of mmWave FMCW radar in automated systems continues to grow, it is imperative to ensure the security of these sensors against possible intimidations.
On the radar line-of-sight (LOS), the received complex signal wave energy’s addition in the manner of coherence from every single range of target scatters. This wave energy indicates the forecast of the complex and gives back echoes from the scattering center of the target against the radar array called “High-Resolution Range profiles-HRRP”. One of the most dynamic study fields in recent radar technology is HRRP-based RATR [7,8]. This is because HRRP encompasses a large quantity of target structural monograms. When it comes to HRRP-based RATR tasks, feature engineering is the most important component.
In 2014, Szegedy C et al. [9] made the initial discovery of an anomalous occurrence. An adversary can create an intentionally unnoticeable information injection by manipulating a single input sample. These examples of input are rather hostile [10]. Based on the arbitrary incorrect outcomes desired by the attacker, they can create a target recognition system using deep learning. Adversarial samples describe these hostile input samples. Both the unique deep learning methods they target and their highly transferable traits make these samples very dangerous. Much study and interest have been piqued by this finding. The process of creating malicious samples has been enhanced throughout time. Additionally, as demonstrated in [11], end-to-end learning communication systems utilizing deep neural network autoencoders are exceedingly susceptible to attacks since wireless communication is inherently open and allows attackers to more easily introduce perturbations. It also demonstrated that noise assaults are less effective against the classifier than artificial perturbation attacks.
Extensive work is being conducted in the fields of radar beamform refinement [9] and target detection adversarial attacks pose a significant security hazard to the usage of radars in automobile applications. For an autonomous car to operate on the road, it is essential to have precise object-detecting capabilities and regular upgrades [10]. It must exhibit a high level of performance reliability to prevent catastrophic accidents. Deep learning models are trained to classify objects; however, an adversarial approach can significantly reduce their ability to accurately classify objects [10]. Adversarial attacks cause artificial intelligence (AI) to make incorrect predictions by making modest changes to the input signal to deliberately incite this incorrect prediction. This is primarily noteworthy since it poses a security threat that is frequently overlooked by humans, or at the very least not identified as an assault due to its subtle nature and exclusive impact on AI algorithms. However, it is crucial to identify and prevent these threats [12].
In the present work, we are investigating the adversarial attack in radar target detection using deep learning design. The main contributions of the proposed work are as follows:
  • We investigated the impact of hostile cases in compromising the security of vehicle radars by analyzing the vulnerability of radar-based deep learning.
  • From the echo radar cube, two different feature maps are generated using two different deep learning architectures, named DAALnet and TDDLnet, in our proposed work, which are a coherent pulse map deep map and RD feature map, respectively.
  • Adversarial learning networks are involved in these two networks, named Radar Generative Adversarial Network (RGAN). After the RGAN generator and discriminator, the features are fused to predict the target range and velocity using the Optimized Hybrid Deep Learning (OHDL) method. The experimental simulations for the proposed work are performed for the verification of adversarial attack prevention in the adversarial OHDL (AOHDL).
The remainder of the paper is structured as follows. Section 2 discusses the earlier implementations regarding these adversarial attacks in radar target detection, and Section 3 describes FMCW radar signal mathematical modeling and the adversarial learning network in deep learning-based RTD. In Section 4, we provide the performance comparison of simulation output. Last of all, Section 5 concludes the development of the proposed work.

2. Related Works

Machine learning classifiers are vulnerable to adversarial instances, which are deliberately crafted from raw data to deceive the classifier into misclassifying the input. These attacks pose a significant risk to the integrity of systems, particularly for applications that are crucial for safety. Numerous adversarial attacks have been created in recent years to compromise various application domains [13]. Although these works emphasized the susceptibility of ML systems to adversarial noise, only a small number of the attack scenarios they examined are feasible in real-world situations.
In [14], the author introduces a unique adversarial assault on Ultra-Wide Band (UltraWB) radars. This approach involves an opponent deliberately injecting hostile radio noise into the wireless channel, failing to recognize obstacles. Initially, utilizing signals obtained from real-life settings, we demonstrate that traditional attacks are unable to produce durable noise in practical circumstances. The author [13] put forward a technique called Adversarial Radio Noise Attack, abbreviated as Adv-RNA, as a solution to address these problems. More precisely, Adv-RNA produces a disruptive disturbance that is effective even without coordination in the middle of the input signal and the disturbance. Furthermore, the noise generated by Adv-RNA is intentionally designed to be resistant to pre-processing and counteract the threat, for instance, defenses concerning the filtering method. Furthermore, Adv-RNA is not only focused on achieving detectability by restricting the amount of noise, but it is also effective when faced with advanced defenses in the spectral domain by establishing a budget for frequencies.
The design of MadRadar, a new black-box physical layer attack architecture for vehicle radars using mmWave FMCW, has been introduced in [15]. This work is distinct from prior studies, which only concentrated on inserting phony items into prey radar’s multidimensional data points. It is the first study to introduce an indication of wrong results for which a condition does not hold and translation assaults, which effectively eliminate or displace detections of existing objects in the prey radar’s multidimensional data points. In addition, all except one of the preceding spoofing attempts (which merely produced false positives) required prior knowledge of the victim radar’s specifications. In contrast, MadRadar can accurately estimate the victim’s chirp period, chirp slope, and frame duration, allowing for effective attacks to be implemented over 95% of the time. The viability and efficiency of the suggested attacks have been experimentally proven by the author [15] through the development of a real-time MadRadar prototype employing SDR platforms. Ultimately, we have showcased the practical applications of MadRadar through tangible case studies in real-world scenarios.
A goal-oriented approach toward achieving autonomous automobile vehicles is the precise sorting of range-dependent radar target detection using deep learning models [16]. Furthermore, the models have to be susceptible to adversarial attacks and possess the ability to truthfully categorize entities in such situations [17]. The creator of [18] utilized specific requirements and methodologies to gather data using FMCW radars. For categorization purposes, four types of structure designs were employed. These models consist of a self-designed CNN-based RadarNet, as well as three state-of-the-art deep learning models based on transfer learning. The evaluation focused on measuring the complexity and accuracy of these models. The models are subjected to FastGradient-SM adversarial attacks and their accuracy is compared to measure the extent of performance loss. Moreover, adversarial retraining has been implemented to alleviate the impact of FastGradient-SM attacks.

3. Proposed System Methodologies

3.1. Automotive FMCW Radar

Radar can be classified into two main categories: pulse radar and continuous-wave radar. Frequency-modulated continuous-wave radar is a modified version of the previous kind, using frequency modulation to enable simultaneous measurement of range and velocity. The FMCW radar is capable of determining both range and velocity by generating a continuous broadcast with a frequency that changes over time. The frequency can exhibit either a linear increase or decrease, or it can be modulated using a bespoke frequency modulation [19]. Typically, the signal undergoes modulation for a brief duration before being returned to the carrier frequency. A modulated transmission is often known as a chirp and has a duration of a chirp period, Tc. It will be demonstrated that a single chirp is sufficient to calculate distance, but numerous successive chirps are necessary to detect the velocity of the target. Therefore, a series of chirps covers a duration referred to as Tc. A prevalent kind of FMCW radar is Linear Frequency-Modulated radar, which usually operates at a carrier frequency (fc) of around 24 GHz or 77 GHz. The frequency sweep of the transmitted LFM signal encompasses a bandwidth BW, which plays a crucial role in determining the characteristics of the radar. Figure 1 depicts a standard LFM chirp sequence within a certain time window.
The rate of the frequency modulation is the sweep rate Rs and is given by:
R s = B W T c
The broadcast signal propagates unless it attenuates owing to signal losses or reflects off a surface [20]. The full duplex radar range equation that includes noise can be used to define the transmitted signal’s range, and is given by:
T a r g e t R = P t .   H 2 .   λ 2 .   φ .   T f 4 π 3 .   η .   k .   T e .   F .   L 4
where L is the total losses due to attenuation, η is the signal-to-noise ratio, Te is the effective noise temperature of the antenna, k is Boltzman’s constant, F is the noise figure, Pt is the peak transmitted power, λ is the carrier wavelength, φ is the radar cross section (RCS), H is the Tx and Rx antenna gain, and Tf is the time frame [21]. Furthermore, it can be shown that by rearranging Equation (2), the power received at the Rx antennas will be directly related to the reciprocal of the range raised to the fourth power. If the focus is just on the signal that is received at the RX antenna rather than the SNR, then the received power may be determined by [22]:
P r = P t H 2 λ 2 φ 4 π 3 ( T a r g e t R ) 4
As can be seen in Equation (3), the power received at the Rx antenna will also depend on the RCS of the target. The RCS of an object is mainly dependent on aspect angle, size, shape, and frequency. The effective area of the trihedral corner reflector is represented as:
A e f f = a 2 3
The RCS of a plane surface is denoted as:
φ = 4 π A e f f 2 λ 2 = 4 π a 4 3 λ 2
The time delay τ stems from the time-of-flight of the propagating signal from the Tx antenna to the target and then back to the Rx antenna. The delay is related to the target distance d and speed of light c as:
τ = 2 d c
The delay gives rise to a difference in frequency between the two signals, which is directly correlated to the sweep rate Rs:
R τ = R s . 2 d c
By introducing multiple chirps in a time frame, a phase shift between consecutive modulation periods will occur, which allows the radar to determine velocity. The model of FMCW is represented in the below content. The instantaneous phase of the transmitted signal is modeled as:
μ t = 2 π 0 t f t d x + μ 0 = 2 π f c t + R s t 2 2 + θ 0
where θ 0 is the initial phase and f(t) is the frequency at time t and it is given by:
f t = f c + R s . t
For the nth sweep, we have t = nTc + ts, where 0 < ts < Tc. The transmitted signal xtx (t) can then be written by:
x t x t = A cos μ t = A cos 2 π f c n T c + t s + R s t s 2 2 + θ 0
where A is the amplitude of the signal.
If the target is located at a distance TargetR from the radar and is moving with relative radial velocity Targetv, then the delay in Equation (6) is remodeled as:
τ = 2 T a r g e t R + T a r g e t v t c = 2 T a r g e t R + T a r g e t v n T c + t s c
Then, the received signal at the reflected target is represented as:
x r x t = B c o s μ t τ = B cos 2 π f c n T c + t s τ + R s t s τ 2 2 + θ 0
The received signal is then mixed with xtx to produce the IF tone. Thus, by performing the subsequent signal processing on a relatively low-frequency range, the processing circuit realization is simplified [23]. The intermediate frequency is then the subtractive term in simplified form, which is given by:
x I F t = A B 2 cos 2 π f c τ + R τ t s R s τ 2 2
By replacing τ with Equation (11), we obtain:
x I F n , t s = A B 2 cos 2 π 2 R s T a r g e t R c t s + 2 f c T a r g e t v c   n T c + 4 π f c T a r g e t R c
By acquiring the frequency of the IF signal, the range can be calculated since Rs is already known. The beat frequency fb is obtained by performing a Fast Fourier Transform on the sampled input, as shown in Figure 2.
T a r g e t R = f B c 2 R s
One chirp is enough to calculate the distance to a target, but at least two chirps are necessary to accurately measure the speed of the target in the direction towards or away from the observer. If a target is moving slowly with a velocity of just a few meters per second, the chirps will not be able to detect the small changes in beat frequency. This becomes problematic when trying to assess the velocity of the target. The resolution may be found in the phasor representation of a chirp. Each range-FFT will provide peaks that may be located at the same frequency, but their phase may differ if the target is in motion.
Just like the range-FFT, the frequency resulting from the phase shift between chirps may be determined using another FFT. The Doppler-FFT refers to the process of determining the angular rotation of a discretized and sampled signal that corresponds to a phasor spinning at a rate of ω rad/sample using the Fast Fourier Transform (FFT), as shown in Figure 3. The conversion of the phase shift rate ω to frequency is achieved by dividing it by 2πTc. The term “Doppler frequency” pertains to the change in frequency induced by the Doppler shift resulting from the movement of an item. The velocity may be determined by relating the term fd to the second term in Equation (14), resulting in the value of:
f d = 2 f c × T a r g e t v c
T a r g e t v = f d c 2 f c

3.2. Proposed AOHDL Methodologies

The proposed implementation workflow is illustrated in Figure 4. Multi-level deep learning along with an adversarial network is designed in this proposal for radar target detection. In our work, an echo data cube is considered as input. A coherent echo pulse is generated from this echo data cube and it is used to extract the deep features using DAALnet. An adversarial network is generated for these deep coherent pulse map features. From the echo data, FELLnet of the deep learning network produces the RD map image as output. From this RD map, we are extracting the deep features using the TDDLnet architecture. Again, we are also introducing the adversarial attack and prediction network into this TDDLnet. After the GAN design of both DAALnet and TDDLnet, we can obtain the fused features of both. Finally, hybrid optimized deep learning is used to detect the range and velocity of the radar signal using AHODnet. Optimization is performed using Enhanced Particle Swarm Optimization (EPSO). In the following section, we have detailed the adversarial network generation for this application.

3.2.1. Coherent Echo Pulse Map

Coherent processing intervals (CPIs) allow the radar to send out many pulses, and the echoes that come back are either integrated coherently or incoherently, allowing them to be analyzed as separate one-dimensional sequences. Discrete time sequences are processed using a variety of approaches, and these methods in turn determine the many kinds of input that detection networks take. Nonetheless, the original information included in the echo signal, known as a coherent pulse map, can be better utilized when employing complex radar data (real and imaginary) directly as opposed to alternative input formats.

3.2.2. DAALnet-Based Feature Map Using Coherent Echo Pulse

Based on the transfer learning concept, we defined a novel deep learning network of Deep Altered Architecture Learning (DAALnet). Transfer learning is a strategy in deep learning; it involves using a model that has been trained on one task to train on another task that is related to the first task. To do this, the trained model may either be utilized as an assigned feature extractor or the weights of the learned model can be fine-tuned to correspond with the new assignment. Transfer learning can potentially save time and money when compared with learning a model from scratch. Additionally, it has the potential to increase the effectiveness of the new model. Resnets have been more successful when used in conjunction with deep neural networks because they incorporate an identity function, which contributes to the achievement of favorable outcomes. For this application, we make use of a deep neural network that has been pre-trained on Resnet18 and apply it to a use case that has a smaller quantity of training data. To adapt ResNet-18 to the development of our DAALnet feature map, the fully connected layer was modified to include three outputs and a weight/bias learning rate factor that was ten times higher than the preceding layers. By making adjustments to this learning rate, we can train the newly created fully connected layer more quickly. The layered architecture of DAALnet is shown in Figure 5.
Each layer’s name and its corresponding dimension of activation function, weights, bias, offset, and scale are detailed in Table 1. In this table for each layer, the output dimensions will be denoted in the activations, especially in the fully connected layer, and we have 1 × 3 vector-sized probability for the classification of three categories.

3.2.3. FELLnet Design Methodologies

From the RTD input cube of radar echo data, we extract and analyze appropriate two-dimensional samples. The pulse repetition frequency and sampling frequency establish the sample points in the transverse direction (fast time domain), whereas the pulses in a CPI are enumerated in order in the longitudinal direction (slow time domain). In traditional radar signal processing, the range is determined by fast time domain echo data and velocity by slow time domain echo data, the latter of which represents Doppler bins. Utilizing the same logic, we employ the FELLnet architecture to slide convolution kernels of predetermined sizes to retrieve data from the frequency and time domains, utilizing rectangular samples as the RD map.
Utilizing a category-wise segmentation approach, the suggested methodology makes use of deep learning-based encoding and decoding strategies. All of the encoders in the design are equipped with their respective decoders, and they are directed towards the regression layer, which is referred to as the Feature Extraction Deep Learning Architecture (FELLnet). To encode the dense input pictures, a deep fully convolutional network is utilized, and a matching decoder is utilized to turn the input features into a pixel-wise regression layer. To provide a feature map that corresponds to the resolution of the input, the decoder is created with the architecture of the encoder being taken into consideration, as shown in Figure 6.
The network is made up of five convolutional layers, and the encoder portion has a kernel that is three layers by three layers. A max-pooling procedure is utilized, and a ReLU activation function is utilized in conjunction with a kernel of 2 × 2. Additionally, there are five transposed convolutional layers with a 3 × 3 kernel size and a 2 × 2 stride in the decoding region. When it comes to the decoding part, each convolutional unit is followed by a ReLU, just like it is in the encoder section. In conclusion, the regression layer is utilized in order to generate the RD map by utilizing the characteristics that were collected from the FELLnet model. Each layer of the FELLNet is detailed in Table 2, along with the activation size for that layer.

3.2.4. TDDLnet Formation

Target Data Detection Learning Network (TDDLnet) is the network model to extract the deep features from the RD map output from FELLnet. To make the system less complex, the number of layers used in this TDDLnet is much less and it is a series network, not the complex Directed Acyclic Graph (DAG) network type of deep learning. The layer design of TDDLnet is shown in Figure 7.

3.2.5. AHODnet Description

A unique met heuristic optimization technique of EPSO and regression by ensemble tree approach were utilized in our suggested deep learning model, which included optimization of hyperparameters. This was accomplished through the implementation of the deep learning model. As a result, the hybrid model of optimization-based deep learning adversarial network, also known as AHODnet, was developed. We can identify targets and make predictions regarding the range and velocity of all objects in the range-Doppler domain by utilizing our AHODnet. It is the active features from DAALnet and TDDLnet that are used as the input for the AHODnet. These hyperparameters include initial learning rate, learn rate schedule, learn rate drop factor, sequence length, gradient threshold method, type of solver, number of epochs, mini Batch Size, momentum, learn rate drop time, and other similar factors.
Learn Rate, Mini Batch Size, and Momentum are the three decision variables that we have chosen to maximize in the model that we have suggested using the Enhanced PSO (EPSO) algorithm. After optimizing the three parameters of deep learning, the activated features of DAALnet and TDDLnet are fused and reshaped to 32 × 32 × 3 image dimension to forward as input to AHODnet. After the fully connected features of the 1 × 150 vector are extracted from the AHODnet, the regression model is used to predict the range and velocity of the radar data cube given.
The flowchart of EPSO is given in Figure 8 and the AHODnet layers’ activation functions, weights, and bias dimensions are described in Table 3.
The sequence of actions that must be carried out before the beginning of the primary EPSO iterations is referred to as the “Initialization”. All of the particles are formed for the multi-objective parameters at this step, with the lowest and maximum range values having been determined. Particle placements are created in a completely random manner. When it comes to this first particle location, the objective values are evaluated, and the particles are classified as either feasible or infeasible.
If a particle is feasible, the associated Objective Function (OF) vector is computed, and the particle’s present position is added to the pbest set as the sole member of the set. There is also a comparison made between the places in gbest and the OF vector. If there are no places in gbest that are dominant in the OF vector of a particle, the current position of the particle is assigned to gbest. Additionally, the places in gbest that are dominated by the present position are eliminated from consideration. EPSO briefly shifts the initialized particles in one principal direction at a time to be able to catch sensitivities that only exist in particular portions of the search domain. This is carried out in order to avoid performing only one evaluation for each direction in the search space.

3.2.6. Adversarial Attack

Our objective is to create FMCW radar adversarial samples to mislead the FMCW radar-based RTD system. The function A design procedure of the FMCW radar-dependent RTD system takes the input as the FMCW radar signal and produces the foretold activity class with the reference of the probability notch p for all the registered classes of activity. Different types of adversarial attacks are listed in Table 4 [24].
A   is   the   deep   learning   model ,   i   is   the   original   sample ,   ψ

3.2.7. Adversarial Attack Networks

In order to comprehensively evaluate our attack strategy, we must take into account together the efficiency of our attack and the degree of distortion caused by the perturbations. To precisely express this evaluation, we formally define the objective function as minimizing T i , i + ψ . T represents the distance metrics ψ p that assess the magnitude of the disruption produced. Nevertheless, as mentioned in prior research [25], finding a direct solution to this challenging non-linear restricted non-convex problem is arduous. Therefore, we redefine the objective function as an instance of optimization based on gradients.
m i n i m i z e   L i + ψ + λ × T i , i + ψ
Measuring the likelihood of initiating adversarial attacks positively known as adversarial loss denoted by L , the total perturbation size is calculated by using the perturbation loss, denoted as T in the Equation (18).
Adversarial loss is defined as:
L = max   Z i + ψ r max f r Z i + ψ f , k
where Z i + ψ r is a probability of estimating i as the original class, Z i + ψ f is the probability of estimating i as another class, and k is a controlling configuration metric of adversarial attack.
Perturbations refer to the discrepancies in the middle of the real activity sample and the adversarial. The L2 Norm, a widely used metric for evaluating adversarial perturbations, determines the Euclidean distance in the middle of two sets [26]. Perturbation loss can be derived as:
T = ψ p 2 2
Intending to ratify the efficacy of the perturbation and advance the efficiency of perturbation followers, we define the dynamic threshold ξ to ensure the target sample:
ψ p 2 2 < ξ ,   s . t   A i + ψ = m
Conventional Deep Learning Network (DLN) security assessment techniques neglect to assess model security and dependability in favor of concentrating mainly on the DLN model classification accuracy [27]. Numerous protections have been presented in recent research to increase the security and robustness of DLN models in order to overcome this issue [28]. To evaluate DLN security, two main concepts describe DNN resistance to adversarial attacks:
  • The first idea is the robustness of the DLN model. This indicates that under this paradigm, the DLN model knows the least perturbation required to transform picture x into an adversarial attack image x ¯ .
  • Adversarial risk, or the gradient descent loss function of the DLN model, is the second property. By reducing errors concerning the input picture, the model aims to improve its prediction score during the DLN learning process. Therefore, the adversary tries to maximize the loss function in order to produce an adversarial picture. To accomplish this, locate the location within x’s neighborhood bounds that can trick the DLN model [29].
One of the main obstacles in this field of study is DLN security and resilience in autonomous cars [30]. The danger of adversarial assaults is one of the most important security problems with DLNs. Creating an adversarial assault, whether it be a white-box or black-box attack, is a difficult undertaking. Developing a white-box assault becomes challenging if the DLN model architecture is intricate and obscures crucial model characteristics that may be utilized to produce the attack.
Moreover, creating a black-box assault is more challenging. This is because black-box assaults may be created using physical or digital techniques, such as altering the surroundings without being aware of the vehicle’s classification architecture, or by poising an attack. Furthermore, every defense that is offered is intended to stop a particular attack.
The low probability of intercept (LPI) characteristics of the frequency-modulated continuous-wave (FMCW) radar make it challenging for an ineffective receiver to differentiate the transmitted waveform from ambient noise [31]. Once acquired, the intercept module has the capability to store the signal in a DRFM (Digital Radio Frequency Memory) and then broadcast a waveform that mimics an echo with certain delay and phase characteristics, corresponding to a particular velocity and range.
The notion is depicted in Figure 9. The impersonating signal would be indistinguishable from a genuine reflection off a target, thereby obtaining a coherent gain when combined with the reference signal. In addition, the resent spoof signal would only experience attenuation due to one-way atmospheric losses, as described by Equation (6), but the actual echo would be affected by two-way losses. Therefore, a falsified signal would possess a significantly greater amount of energy reaching the receiver of the FMCW radar compared to the actual reflection. However, providing too much energy toward the RX antenna might potentially harm the electronics, thereby undermining any false motives.
Figure 9 illustrates the operational concept of a spoofing assault on radar. Figure 9b depicts an adversarial system that transmits a signal that is in phase and aligned with the victim radar. The opponent uses delay and phase compensation techniques to generate a deceptive target with a specific range and velocity, with the intention of misleading the target’s radar system [32]. Furthermore, the hostile signal has unidirectional losses, resulting in a more potent signal compared to the reflected echo.

3.2.8. Radar Generative Adversarial Networks (GANs)

In our proposed model, we used a Generative Adversarial Network (GAN) to prevent radar target detection from adversarial attacks, named RGAN. Generative Adversarial Networks (GANs) have the ability to produce new data samples that possess certain properties by gaining information around the distribution of training samples. Generator G and discriminator D are two major mechanisms involved in a GAN. To understand how adversarial assaults in a radar environment are modelled probabilistically, Radar Generative Adversarial Networks (RGANs) employ a generator to train a generative model. The discriminator compares the generated signal to the real signal in the dataset [33]. One way to use a model that aims to differentiate between actual and fraudulent signals is through the use of a discriminator.
The generator aims to transform dormant information z, which follows the giving out pz(z), hooked on a fresh sample that is supposed to follow the giving out pg. It is desirable for pg to closely approximate the actual data distribution p data. Afterward, the artificially generated sample and a real sample are together submitted to the discriminator, which aims to distinguish the generated sample from the authentic samples, in addition to ascertaining the corresponding level of confidence [34]. Put simply, the generator produces samples with the intention of perplexing the discriminator, while the discriminator endeavors to properly differentiate between actual and produced samples. The whole framework is trained using an adversarial method, which is characterized by the subsequent cost function:
min G max D L R G A N D , G = E x log D x + E z log 1 D G z
The function LRGAN (D, G) represents the standard cross-entropy loss. The mutable x signifies an unaffected sample extracted out of distribution p data, while z embodies hidden noise obtained from distribution pz (z). During the phase of training, G and D are enhanced alternately via a zero-sum game framework [35]. This technique is iterated indefinitely until the model reaches convergence. If D and G are mutually well constructed and possess suitable constructions, they will theoretically achieve Nash equilibrium at the completion of the training procedure. By the side of this juncture, the counterfeit samples produced by G will be indistinguishable from genuine samples, rendering D unable to discern whether the sample originates from the generator or the actual world.
A generator in RGAN serves as a masking purpose denoted as G, which takes a noisy input and generates a clean mask output according to the noisy input. During the training phase, RGAN is provided with clean signals and generated noise from the generator network [36] with a defined standard deviation, which consists of noisy signals and their matching ground truth. The generator maps the adversarial attack as input to the discriminator. The discriminator D differentiates between the generated signal and the matching ground truth (xclean). The adversarial training procedure enhances the closeness of the produced signal to the ground truth, which refers to the actual uncontaminated signal. The training procedure concludes that the minute of the discriminator is no longer able to distinguish between a genuine signal and one that has been created by the generator. Currently, the RGAN settings are set, and the generator is capable of improving the signal-to-noise ratio (SNR) of adversarial signals [37]. During the testing phase, each signal with noise is generated from G, resulting in the retrieval of the equivalent signal without noise. The block diagram of RGAN is shown in Figure 10.
Adversarial Generator Network
Using the signal-to-map translation, the generator G diminishes the difficulties of converting the noisy signal into an improved concealing signal G(xnoise,z). The goal remains to create a mask that closely resembles the actual signal map, rendering it unrecognizable from the original by the discriminator D. In the RGAN model, it is necessary for the generated signal to preserve the desired information while minimizing the presence of unwanted noise [26]. The convolutional layer operates as:
x j h x , y = f ( i = 1 m u = 0 S 1 v = 0 S 1 k j i h u , v . x i h 1 x + u , y + v + b j h )  
The parameter h represents the layer index, the size of the kernel is denoted as S’, and the previous layer’s ith input feature map is expressed as   x i h i = 1,2 , 3 , . m , the current layer’s jth outcome feature map is x j h i = 1 ,   2 ,   3 , . n , the convolutional kernel of k j i h is used to establish the connection of the ith feature map to the jth feature map, and the jth feature map bias is denoted as b j h . The kernels and pace in each layer are configured as 2 × 1 and 2, respectively, as stated in this study. The “leaky ReLU” activation function, denoted as f(x), is a nonlinear activation function:
LReLU x = x   , x 0 α x , x < 0
where α is the hyperparameter used to regulate the sparsity of the network and has been assigned a value of 0.01 in this research study.
In network G, which serves as a generator, the random vectors of size 100 are transformed into 4-by-4-by-512 arrays using a projection and reform operation. Subsequently, the resulting arrays are increased to 64-by-64-by-3 arrays using a sequence of transposed convolution layers with batch normalization and ReLU layer. In total, four transposed convolutional layers are connected in the network in a series manner. The network utilizes transposed convolution layers with a setup that includes 5-by-5 filters, a diminishing quantity of filters for every single layer, a pace of 2, and in addition, cropping of the outcome takes place in every single edge. Select three 5-by-5 filters for the final transposed convolution layer, corresponding to the three RGB channels of the created pictures, and also select the output size of the preceding layer. Include a hyperbolic tangent (tanh) layer at the end of the network.
Adversarial Discriminator Network
This network aims to categorize observations as either “real” or “attack” by using batches of data that include notes starting together with the training data and the generated data produced by the generator [38]. The discriminator network receives 64-by-64-by-3 pictures as input and generates a single forecast score by the use of a sequence of convolution layers through batch normalization and leaky ReLU layers. Apply the dropout technique to introduce noise into the input pictures. A dropout probability of 0.5 is specified for the dropout layer. The convolution layers are set up with 5-by-5 filters, and the number of filters increases for each layer. The stride and padding of the output are also taken into consideration. The Leaky ReLU activation function is employed with a scale constraint of 0.2. In order to generate probabilities within the range of 0 to 1, a convolutional layer is defined using a single 4-by-4 filter, which is then followed by a sigmoid layer.
Cost Function of RGAN
On the other hand, RGAN may have difficulties in achieving optimal training due to the implementation of weight clipping, which enforces the satisfaction of the Lipschitz constraint. Consequently, a gradient penalty element is incorporated into the cost function of RGAN.
L G P = E x ^ D x ^ 2 1 2
x ^ is the random line mixture of the actual sample x and the spawned sample G z x n o i s y , derived as:
x ^ = ω × x c l e a n + 1 ω × G z   x n o i s y )
where ω, a uniform distribution over the interval [0, 1], is sampled in a random manner. In addition, generator G is restricted to producing samples that are similar to the real samples by incorporating L1 and L2 regularization terms into the cost function:
L L 1 = E x n o i s y , x c l e a n , z x c l e a n G z x n o i s y 1
L L 1 = E x n o i s y , x c l e a n , z x c l e a n G z x n o i s y 2
The final cost function of RSEGAN is the combination of LRGAN, LGP, LL1, and LL2, given by:
L f i n a l = min G max D   L RGAN + λ GP × L GP + λ L 1 × L 1 + λ L 2 × L 2
The coefficient of the gradient penalty is denoted as λGP, the coefficient of L1 regularization is denoted as λ1, and the coefficient of L2 regularization is denoted as λ2 [39]. More precisely, RGAN utilizes an alternative optimization technique throughout the training process. The discriminator cost functions and generator can be stated separately, as seen in Equation (29), as follows:
min D   L D l o s s D , G = E x n o i s y , x c l e a n D x n o i s y x c l e a n + E x n o i s y , z D G z x n o i s y + λ G P . E x ^ D x ^ 2 1 2
min G   L G l o s s D , G = E x n o i s y , z D G z x n o i s y + λ L 1 × E x n o i s y , x c l e a n , z [ x c l e a n G z x n o i s y 1 + λ L 2 × E x n o i s y , x c l e a n , z [ x c l e a n G z x n o i s y 2 ]

4. Results and Discussion

In this section, we perform the simulation of the proposed AOHDL-based radar target detection, and the results are compared with the earlier works of the same. For the simulation of the work, we use a simulation tool of MATLAB 2021a in an Intel® Core™ i7-12700, 2100 Mhz, 12 Core(s) CPU of 16 GB RAM, with a 64-bit operating system.
The sawtooth chirp waveform is generated in FMCW radar signal generation with a chirp duration of 32 microseconds and a chirp cycle of 64. The radar frequency is fixed at 76.5 GHz and the sampling frequency is 5 MHz. The Sweeping Bandwidth is configured as 0.2 GHz. Radar targets of three categories were taken for the simulation, such as cars, bikes, and synthetic objects positioned 0.5 m above the street. Signal characteristics of 10 dBm transmit power, 17 dB transmit gain, and 15 dB receiver gain are assigned. In the channel configuration, the −130 dB noise floor and 10 dB receiver noise figure are fixed. The reception angle for radar is 160° in two receive antennas of 15 dB variance amplitude. For training, we created our dataset of echo radar data cubes with a size of 160 × 64 × 8 in seven different SNR ranges from 0 dB to 30 dB in the count of 21,000 samples. Out of this, 85% of samples were used for training and 5% for validation. The remaining 10% is for testing in the deep learning of RTD. We created a dataset consisting of recognized targets in various locations, moving at variable speeds. The focus is on three types of targets: a car, a bike, and synthetic points, utilizing automotive FMCW radar systems. Figure 11 displays the sample images of the radar echo cube.
To prove the efficiency of adversarial attack prevention, the performance of RMSE for different scenarios is created and analyzed. The OHODnet is verified without any adversarial attack and evaluated for case 1. For the second case, with an adversarial attack, the OHODnet is evaluated. Finally, for the third case, the AOHODnet with adversarial attacks, and the RMSE of range and velocity target predictions are calculated and compared.
The generated images and training of adversarial attacks in DAALnet and TDDLnet are shown in Figure 12 and Figure 13, respectively. In both figures, section (a) depicts the generated image slices of corresponding deep learning model feature maps, and section (b) illustrates the training progress of several iterations versus score convergence in both the generator network and discriminator network. It provides a comparative analysis of how each network’s performance evolves over time. By analyzing, we can gain insight into the stability and efficiency of the training process. Furthermore, observing the convergence patterns helps in understanding the dynamic interplay between the generator and discriminator, important for optimizing the overall model performance.
Figure 14 illustrates the performance comparison of RMSE in the range parameter estimation using the proposed OHDL with and without adversarial network and different earlier works of EDACM [40], DAE [41], ‘IYoLov4-tiny’ [42], ‘SALA-LSTM’ [43], ‘CNN’ [44], ‘RNN’ [45], ‘FFT’ [46], and ‘CFAR’ [47]. In all the curves, as SNR increased, the RMSE was reduced. At 0 dB of SNR, the RMSE value for OHDL with and without adversarial network attains 1.3858 and 1.4230, respectively. For the same SNR, 1.7912, 1.7979, 2.4006, 2.3890, 2.7525, 2.7151, 3.2233, and 3.2134 are attained by EDACM, DAE, ‘IYoLov4-tiny’, ‘SALA-LSTM’, ‘CNN’, ‘RNN’, ‘FFT’, and ‘CFAR’, respectively.
Figure 15 shows the RMSE comparison of radar target velocity value for proposed and existing methods. At 30 dB of SNR, the RMSE values of OHDL with adversarial network, OHDL, EDACM, DAE, ‘IYoLov4-tiny’, ‘SALA-LSTM’, ‘CNN’, ‘RNN’, ‘FFT’, and ‘CFAR’ methods are 0.0000, 0.0139, 0.5982, 0.0599, 0.1157, 0.7809, 1.1825, 1.2869, 1.6136, and 1.8503, respectively.
Figure 16 illustrates the performance of RMSE in range estimation using an adversarial attack network in our proposed model OHDL. OHDL is efficiently used for RTD. When an attack involving OHDL performance degrades, which is overcome by introducing AOHDL (adversarial network), the RMSE is improved. The RMSE performance of the proposed velocity estimation is shown in Figure 17 with and without attack on OHDL and adversarial learning-based OHDL of AOHDL. AOHDL attains almost the same performance as OHDL without an attack on the radar system, which shows the efficiency of our prediction.
Figure 18 depicts the performance of time complexity of the adversarial learning system applied to the radar target detection platform with the deep learning model of OHDL. Compared to simple OHDL, AOHDL has high time consumption due to the design injection of an adversarial RGAN network into the DAALnet and TDDLnet. But at the same time, it is less than that of other earlier methods.
The RMSE values of different SNRs and different methods are shown in Table 5 for range estimation, and Table 6 for velocity estimation. In Table 7, we tabulated the evaluation metrics of proposed AOHDL with attack and compared it with OHDL with and without attack scenarios. OHDL without adversarial attack obtains the best performance. When the attack insisted on the radar, OHDL was degraded, showing the lowest results. This degradation is enhanced by our proposed AOHDL, which attains results almost similar to the absence of an attack environment.
Figure 19 shows the performance prediction accuracy in the radar target detection system of multiple implementation algorithms compared with those with the proposed AOHDL. The proposed accuracy reached 100% at SNR of 15 dB. At the same SNR, 99.9%, 99.51%, 99.18%, 99.1%, 98.9%, 98.55%, 98.4%, 98%, and 97% are attained with OHDL, DAE, EDACM, ‘IYoLov4-tiny’, ‘SALA-LSTM’, ‘CNN’, ‘RNN’, ‘FFT’, and ‘CFAR’, respectively.

4.1. Complexity Analysis of Adversarial Learning

It was argued in [48] that there might be an inherent tension between accuracy and adversarial robustness. We argue that this potential tension disappears for a rich class of learning algorithms if we consider q-bounded adversaries. We show that if a learning algorithm satisfies a particular natural property, then there is a lower bound for the complexity of this algorithm that grows with accuracy. For a rich class of learning algorithms, our security guarantee against bounded adversaries increases with accuracy. Figure 20 depicts the performance of multiple interference scenarios in the prediction of range and velocity estimation. As the interferences are reduced in the environment, the prediction RMSE is reduced as much as possible, as shown in Figure 20.
Figure 21 illustrates the RMSE of the proposed AOHDL method in a dynamic radar environment, where targets of varying speed—humans, bicycles, and cars—are analyzed. As the speed of the moving targets increases from humans to cars, the RMSE correspondingly rises, reaching a peak of 1.66 at 0 dB signal-to-noise ratio (SNR) for cars, whereas the RMSE values for bicycles and humans are 1.04 and 0.59, respectively, under the same conditions.

4.2. Time Complexity of Computation in AOHDL

Findings from studies indicate that learning rate and batch size are two characteristics that significantly impact computational complexity [48]. Increasing the batch size will make the model more computationally difficult. Therefore, to obtain a true measure of the model’s complexity, you must multiply these two parameters. You may figure out the convolution layer’s temporal complexity by calculating:
n = 1 d k n 1 × s n 2 × f n × l n 2 × r 1 × b 1
Here, k n 1 defines the number of input channels in the lth layer, s n is the length of the filter, f n is the number of filters in the nth layer, d is the depth of the convolutional layer, l n is the length of the output feature map, r 1 is the learning rate, b 1 is the batch size.
The time complexity of the fully connected layer can be calculated as:
i = 1 f D W H N
The fully connected layer’s depth is denoted by l, while the input/output channel’s dimension is defined by D, and W, H, and N are the input’s width, height, and the number of outputs, respectively.

5. Conclusions

In this paper, we propose a method for enhancing radar target detection in the FMCW system using a deep learning model in the presence of adversarial attack. Adversarial attack prevention is performed by integrating the Radar Generative Adversarial Network (RGAN) into Adversarial Learned Optimized Hybrid Deep Learning (AOHDL). By utilizing an adversarial learning structure, the low SNR signals can combine the advantages of GAN, L1-norm regularization, and L2-norm regularization to train RGAN. The findings of our experiment show that the combination of RGAN and OHDL-based detectors may achieve an outstanding identification probability and a tiny false alarm probability. For prospective implementation, our main focus can be on domain adaptability within adversarial learning networks.

Author Contributions

Conceptualization, methodology, software, and writing—original draft preparation M.M.A.; validation, L.G.; writing—review and editing, Y.T., W.C. and L.D.; supervision, Y.L.; funding acquisition, W.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Key Research and Development Program of Shaanxi under Grant 2023-YBGY-258 and in part by the National Natural Science Foundation of China under Grant 62301431.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kaselimi, M.; Voulodimos, A.; Daskalopoulos, I.; Doulamis, N.; Doulamis, A. A vision transformer model for convolution-free multilabel classification of satellite imagery in deforestation monitoring. IEEE Trans. Neural Netw. Learn. Syst. 2022, 34, 3299–3307. [Google Scholar] [CrossRef]
  2. Park, K.-E.; Lee, J.-P.; Kim, Y. Deep learning-based indoor distance estimation scheme using FMCW radar. Information 2021, 12, 80. [Google Scholar] [CrossRef]
  3. Sun, S.; Petropulu, A.P.; Poor, H.V. MIMO radar for advanced driver-assistance systems and autonomous driving: Advantages and challenges. IEEE Signal Process. Mag. 2020, 37, 98–117. [Google Scholar] [CrossRef]
  4. Patole, S.M.; Torlak, M.; Wang, D.; Ali, M. Automotive radars: A review of signal processing techniques. IEEE Signal Process. Mag. 2017, 34, 22–35. [Google Scholar] [CrossRef]
  5. Lies, W.A.; Narula, L.; Iannucci, P.A.; Humphreys, T.E. Long range, low swap-C FMCW radar. IEEE J. Sel. Top. Signal Process. 2021, 15, 1030–1040. [Google Scholar] [CrossRef]
  6. Ordean, M.; Garcia, F.D. Millimeter-wave automotive radar spoofing. arXiv 2022, arXiv:2205.06567. [Google Scholar]
  7. Yuan, Y.; Wan, J.; Chen, B. Robust attack on deep learning based radar HRRP target recognition. In Proceedings of the 2019 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Lanzhou, China, 18–21 November 2019; pp. 704–707. [Google Scholar]
  8. Lin, W.; Shi, L.; Zhong, Y.; Huang, Y.; Ding, X. Improving Black-box Adversarial Attacks on HRRP-based Radar Automatic Target Recognition. In Proceedings of the 2021 CIE International Conference on Radar (Radar), Haikou, China, 15–19 December 2021; pp. 3206–3209. [Google Scholar]
  9. Geng, L.; Li, Y.; Dong, L.; Tan, Y.; Cheng, W. Efficiently Refining Beampattern in FDA-MIMO Radar via Alternating Manifold Optimization for Maximizing Signal-to-Interference-Noise Ratio. Remote Sens. 2024, 16, 1364. [Google Scholar] [CrossRef]
  10. Zafar, A.; Khan, A.; Younis, S. Classical Adversarial Attack on mm-Wave FMCW Radar. In Proceedings of the 2021 International Conference on Frontiers of Information Technology (FIT), Islamabad, Pakistan, 13–14 December 2021; pp. 281–286. [Google Scholar]
  11. Zheng, Y.; Su, J.; Zhang, S.; Tao, M.; Wang, L. Dehaze-AGGAN: Unpaired remote sensing image dehazing using enhanced attention-guide generative adversarial networks. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
  12. Valtl, J.; Issakov, V. Universal Adversarial Attacks on the Raw Data from a Frequency Modulated Continuous Wave Radar. IEEE Access 2022, 10, 114092–114102. [Google Scholar] [CrossRef]
  13. Tarchoun, B.; Alouani, I.; Khalifa, A.B.; Mahjoub, M.A. Adversarial attacks in a multi-view setting: An empirical study of the adversarial patches inter-view transferability. In Proceedings of the 2021 International Conference on Cyberworlds (CW), Caen, France, 28–30 September 2021; pp. 299–302. [Google Scholar]
  14. Guesmi, A.; Alouani, I. Adversarial attack on radar-based environment perception systems. arXiv 2022, arXiv:2211.01112. [Google Scholar]
  15. Hunt, D.; Angell, K.; Qi, Z.; Chen, T.; Pajic, M. MadRadar: A Black-Box Physical Layer Attack Framework on mmWave Automotive FMCW Radars. arXiv 2023, arXiv:2311.16024. [Google Scholar]
  16. Xu, Y.; Shi, L.; Lin, C.; Cai, S.; Lin, W.; Huang, Y.; Ding, X. A Contrastive-Based Adversarial Training Algorithm for HRRP Target Recognition. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar] [CrossRef]
  17. Narasimhamurthy, R.; Khalaf, O.I. Deep Learning Network for Classifying Target of Same Shape using RCS Time Series. Air Traffic Manag. Control 2021, 9, 25. [Google Scholar]
  18. Cha, D.; Jeong, S.; Yoo, M.; Oh, J.; Han, D. Multi-input deep learning based FMCW radar signal classification. Electronics 2021, 10, 1144. [Google Scholar] [CrossRef]
  19. Klintberg, J.; McKelvey, T.; Dammert, P. A parametric approach to space-time adaptive processing in bistatic radar systems. IEEE Trans. Aerosp. Electron. Syst. 2021, 58, 1149–1160. [Google Scholar] [CrossRef]
  20. Mark, S.; Lokash, S.; Shashi, S. Trihedral Corner Reflector. Millimeter Wave Product. Available online: https://www.miwv.com/trihedral-reflectors-for-radar-applications/ (accessed on 14 June 2020).
  21. Mahafza, B.R. Radar Systems Analysis and Design Using MATLAB; Chapman and Hall/CRC: Boca Raton, FL, USA, 2005. [Google Scholar]
  22. Chen, H.-R. FMCW Radar Jamming Techniques and Analysis. Available online: https://core.ac.uk/download/pdf/36730041.pdf (accessed on 10 September 2013).
  23. Snihs, L. Evaluation of FMCW Radar Jamming Sensitivity. Available online: https://www.diva-portal.org/smash/get/diva2:1767179/FULLTEXT01.pdf (accessed on 14 June 2023).
  24. Xie, Y.; Jiang, R.; Guo, X.; Wang, Y.; Cheng, J.; Chen, Y. Universal Targeted Adversarial Attacks Against mmWave-based Human Activity Recognition. In Proceedings of the IEEE INFOCOM 2023-IEEE Conference on Computer Communications, New York, NY, USA, 17–20 May 2023; pp. 1–10. [Google Scholar]
  25. Carlini, N.; Wagner, D. Towards evaluating the robustness of neural networks. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–24 May 2017; pp. 39–57. [Google Scholar]
  26. Ozbulak, U.; Vandersmissen, B.; Jalalvand, A.; Couckuyt, I.; van Messem, A.; de Neve, W. Investigating the significance of adversarial attacks and their relation to interpretability for radar-based human activity recognition systems. Comput. Vis. Image Underst. 2021, 202, 103111. [Google Scholar] [CrossRef]
  27. Rahman, M.M.; Gurbuz, S.Z.; Amin, M.G. Physics-aware generative adversarial networks for radar-based human activity recognition. IEEE Trans. Aerosp. Electron. Syst. 2022, 59, 2994–3008. [Google Scholar] [CrossRef]
  28. Yue, Y.; Liu, H.; Meng, X.; Li, Y.; Du, Y. Generation of high-precision ground penetrating radar images using improved least square generative adversarial networks. Remote Sens. 2021, 13, 4590. [Google Scholar] [CrossRef]
  29. Chen, C.; Su, Y.; He, Z.; Liu, T.; Song, X. Clutter mitigation in holographic subsurface radar imaging using generative adversarial network with attentive subspace projection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  30. Chen, S.; Shangguan, W.; Taghia, J.; Kühnau, U.; Martin, R. Automotive radar interference mitigation based on a generative adversarial network. In Proceedings of the 2020 IEEE Asia-Pacific Microwave Conference (APMC), Hong Kong, China, 10–13 November 2020; pp. 728–730. [Google Scholar]
  31. Wang, S.; An, Q.; Li, S.; Zhao, G.; Sun, H. Wiring effects mitigation for through-wall human motion micro-doppler signatures using a generative adversarial network. IEEE Sens. J. 2021, 21, 10007–10016. [Google Scholar] [CrossRef]
  32. Pan, Z.; Wang, S.; Li, Y. Residual attention-aided U-Net GAN and multi-instance multilabel classifier for automatic waveform recognition of overlapping LPI radar signals. IEEE Trans. Aerosp. Electron. Syst. 2022, 58, 4377–4395. [Google Scholar] [CrossRef]
  33. Wang, H.; Li, K.; Lu, X.; Zhang, Q.; Luo, Y.; Kang, L. ISAR Resolution Enhancement Method Exploiting Generative Adversarial Network. Remote Sens. 2022, 14, 1291. [Google Scholar] [CrossRef]
  34. Wang, C.; Wang, P.; Wang, P.; Xue, B.; Wang, D. Using conditional generative adversarial 3-D convolutional neural network for precise radar extrapolation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 5735–5749. [Google Scholar] [CrossRef]
  35. Che, J.; Wang, L.; Wang, C.; Zhou, F. A Novel Adversarial Learning Framework for Passive Bistatic Radar Signal Enhancement. Electronics 2023, 12, 3072. [Google Scholar] [CrossRef]
  36. Zhu, Y.; Miao, C.; Xue, H.; Li, Z.; Yu, Y.; Xu, W.; Su, L.; Qiao, C. TileMask: A Passive-Reflection-based Attack against mmWave Radar Object Detection in Autonomous Driving. In Proceedings of the 2023 ACM SIGSAC Conference on Computer and Communications Security, New York, NY, USA, 26–30 November 2023; pp. 1317–1331. [Google Scholar]
  37. Almutairi, S.; Barnawi, A. Securing DNN for smart vehicles: An overview of adversarial attacks, defenses, and frameworks. J. Eng. Appl. Sci. 2023, 70, 16. [Google Scholar] [CrossRef]
  38. Graff, A.M.; Humphreys, T.E. Signal Identification and Entrainment for Practical FMCW Radar Spoofing Attacks. In Proceedings of the 2023 IEEE 97th Vehicular Technology Conference (VTC2023-Spring), Florence, Italy, 20–23 June 2023; pp. 1–7. [Google Scholar]
  39. Xue, W.; Wang, R.; Liu, L.; Wu, D. Accurate multi-target vital signs detection method for FMCW radar. Measurement 2023, 223, 113715. [Google Scholar] [CrossRef]
  40. Chen, S.; Taghia, J.; Fei, T.; Kühnau, U.; Pohl, N.; Martin, R. A DNN autoencoder for automotive radar interference mitigation. In Proceedings of the ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 4065–4069. [Google Scholar]
  41. Liang, S.; Chen, R.; Duan, G.; Du, J. Deep learning-based lightweight radar target detection method. J. Real-Time Image Process. 2023, 20, 61. [Google Scholar] [CrossRef]
  42. Wang, J.; Li, S. SALA-LSTM: A novel high-precision maritime radar target detection method based on deep learning. Sci. Rep. 2023, 13, 12125. [Google Scholar] [CrossRef]
  43. Ishaq, M.; Kwon, S. A CNN-Assisted deep echo state network using multiple Time-Scale dynamic learning reservoirs for generating Short-Term solar energy forecasting. Sustain. Energy Technol. Assess. 2022, 52, 102275. [Google Scholar]
  44. Mikolov, T.; Karafiát, M.; Burget, L.; Cernocký, J.; Khudanpur, S. Recurrent neural network based language model. In Proceedings of the Interspeech 2010, Makuhari, Japan, 26–30 September 2010; Volume 2, pp. 1045–1048. [Google Scholar]
  45. Wan, R.; Song, Y.; Mu, T.; Wang, Z. Moving target detection using the 2D-FFT algorithm for automotive FMCW radars. In Proceedings of the 2019 International Conference on Communications, Information System and Computer Engineering (CISCE), Haikou, China, 5–7 July 2019; pp. 239–243. [Google Scholar]
  46. Lee, S.; Lee, B.-H.; Lee, J.-E.; Kim, S.-C. Statistical characteristic-based road structure recognition in automotive FMCW radar systems. IEEE Trans. Intell. Transp. Syst. 2018, 20, 2418–2429. [Google Scholar] [CrossRef]
  47. Tsipras, D.; Santurkar, S.; Engstrom, L.; Turner, A.; Madry, A. Robustness may be at odds with accuracy. arXiv 2018, arXiv:1805.12152. [Google Scholar]
  48. Kandel, I.; Castelli, M. The effect of batch size on the generalizability of the convolutional neural networks on a histopathology dataset. ICT Express 2020, 6, 312–315. [Google Scholar] [CrossRef]
Figure 1. FMCW chirp sequence.
Figure 1. FMCW chirp sequence.
Remotesensing 16 03109 g001
Figure 2. Range FFT from sampling.
Figure 2. Range FFT from sampling.
Remotesensing 16 03109 g002
Figure 3. Range Doppler FFT from chirp sequence.
Figure 3. Range Doppler FFT from chirp sequence.
Remotesensing 16 03109 g003
Figure 4. Proposed workflow of the RTD.
Figure 4. Proposed workflow of the RTD.
Remotesensing 16 03109 g004
Figure 5. DAALnet architecture.
Figure 5. DAALnet architecture.
Remotesensing 16 03109 g005
Figure 6. FELLLnet architecture.
Figure 6. FELLLnet architecture.
Remotesensing 16 03109 g006
Figure 7. TDDLnet layout.
Figure 7. TDDLnet layout.
Remotesensing 16 03109 g007
Figure 8. EPSO flowchart.
Figure 8. EPSO flowchart.
Remotesensing 16 03109 g008
Figure 9. Adversarial attack in radar system: (a) normal scenario; (b) adversarial attack scenario.
Figure 9. Adversarial attack in radar system: (a) normal scenario; (b) adversarial attack scenario.
Remotesensing 16 03109 g009
Figure 10. RGAN architecture.
Figure 10. RGAN architecture.
Remotesensing 16 03109 g010
Figure 11. Radar echo data cube database (row 1: bike target, row 2: car target, row 3: synthetic target).
Figure 11. Radar echo data cube database (row 1: bike target, row 2: car target, row 3: synthetic target).
Remotesensing 16 03109 g011
Figure 12. (a) Generated images of DAALnet; (b) DAALnet adversarial network learning progress, epoch: 50, iteration: 150, elapsed: 00:18:06.
Figure 12. (a) Generated images of DAALnet; (b) DAALnet adversarial network learning progress, epoch: 50, iteration: 150, elapsed: 00:18:06.
Remotesensing 16 03109 g012
Figure 13. (a) Generated images of TDDLnet; (b) TDDLnet adversarial network learning progress, epoch: 50, iteration: 150, elapsed: 00:10:54.
Figure 13. (a) Generated images of TDDLnet; (b) TDDLnet adversarial network learning progress, epoch: 50, iteration: 150, elapsed: 00:10:54.
Remotesensing 16 03109 g013
Figure 14. RMSE comparison for range estimation.
Figure 14. RMSE comparison for range estimation.
Remotesensing 16 03109 g014
Figure 15. RMSE comparison for velocity estimation.
Figure 15. RMSE comparison for velocity estimation.
Remotesensing 16 03109 g015
Figure 16. Adversarial attack performance in range prediction.
Figure 16. Adversarial attack performance in range prediction.
Remotesensing 16 03109 g016
Figure 17. Adversarial attack performance in velocity prediction.
Figure 17. Adversarial attack performance in velocity prediction.
Remotesensing 16 03109 g017
Figure 18. The time complexity of the system.
Figure 18. The time complexity of the system.
Remotesensing 16 03109 g018
Figure 19. Prediction accuracy of the system.
Figure 19. Prediction accuracy of the system.
Remotesensing 16 03109 g019
Figure 20. Impact of different interference in detection of radar target.
Figure 20. Impact of different interference in detection of radar target.
Remotesensing 16 03109 g020
Figure 21. Impact of dynamic environment on RMSE evaluation.
Figure 21. Impact of dynamic environment on RMSE evaluation.
Remotesensing 16 03109 g021
Table 1. DAALnet layer description.
Table 1. DAALnet layer description.
Layer TypeActivationsLearnablesLayer TypeActivationsLearnables
Image Input224 × 224 × 3-Addition28 × 28 × 128-
Convolution112 × 112 × 64Weights—7 × 7 × 3 × 64ReLU28 × 28 × 128-
Bias—1 × 1 × 64ID Block 328 × 28 × 128Weights—3 × 3 × 128 × 128
Batch Normalization112 × 112 × 64Offset—1 × 1 × 64Bias—1 × 1 × 128
Scale—1 × 1 × 64DS Block 214 × 14 × 256Weights—1 × 1 × 128 × 256
ReLU112 × 112 × 64-Bias—1 × 1 × 256
Max Pooling56 × 56 × 64-Addition14 × 14 × 256-
ID Block 1Convolution56 × 56 × 64Weights—3 × 3 × 64 × 64ReLU14 × 14 × 256-
Bias—1 × 1 × 64ID Block 414 × 14 × 256Weights—3 × 3 × 256 × 256
Batch Normalization56 × 56 × 64Offset—1 × 1 × 16Bias—1 × 1 × 256
Scale—1 × 1 × 16DS Block 37 × 7 × 512Weights—1 × 1 × 256 × 512
ReLU56 × 56 × 64-Bias—1 × 1 × 512
Convolution56 × 56 × 64Weights—3 × 3 × 64 × 64Addition7 × 7 × 512-
Bias—1 × 1 × 64ReLU7 × 7 × 512-
Batch Normalization56 × 56 × 64Offset—1 × 1 × 64ID Block 57 × 7 × 512Weights—3 × 3 × 512 × 512
Scale—1 × 1 × 64Bias—1 × 1 × 512
Addition56 × 56 × 64-Addition7 × 7 × 512-
ReLU56 × 56 × 64-ReLU7 × 7 × 512-
ID Block 256 × 56 × 64Weights—3 × 3 × 64 × 64—Average Pooling1 × 1 × 512-
Bias—1 × 1 × 64Fully Connected1 × 1 × 3Weights—3 × 512
DS Block 1Convolution28 × 28 × 128Weights—3 × 3 × 64 × 128Bias—3 × 1
Bias—1 × 1 × 128Softmax1 × 1 × 3-
Batch Normalization28 × 28 × 128Offset—1 × 1 × 128Classification--
Scale—1 × 1 × 128
ReLU28 × 28 × 128-
Convolution28 × 28 × 128Weights—3 × 3 × 128 × 128
Bias—1 × 1 × 128
Batch Normalization28 × 28 × 128Offset—1 × 1 × 128
Scale—1 × 1 × 128
Convolution28 × 28 × 128Weights—1 × 1 × 64 × 128
Bias—1 × 1 × 128
Batch Normalization28 × 28 × 128Offset—1 × 1 × 128
Scale—1 × 1 × 128
Table 2. FELLnet layer description.
Table 2. FELLnet layer description.
Layer TypeActivationsLearnables Layer TypeActivationsLearnables
Image Input32 × 32 × 1-Decoding LayersTransposed Convolution2 × 2 × 2Weights—4 × 4 × 2 × 2
Encoding LayersConvolution32 × 32 × 32Weights—3 × 3 × 1 × 32Bias—1 × 1 × 2
Bias—1 × 1 × 32ReLU2 × 2 × 2-
ReLU32 × 32 × 32-Transposed Convolution16 × 16 × 16Weights—3 × 3 × 32 × 16
Max Pooling16 × 16 × 32-Bias—1 × 1 × 16
Convolution16 × 16 × 16Weights—3 × 3 × 32 × 16ReLU16 × 16 × 16-
Bias—1 × 1 × 16Transposed Convolution8 × 8 × 8Weights—3 × 3 × 16 × 8
ReLU16 × 16 × 16-Bias—1 × 1 × 8
Max Pooling8 × 8 × 16-ReLU8 × 8 × 8-
Convolution8 × 8 × 8Weights—3 × 3 × 16 × 8Transposed Convolution4 × 4 × 4Weights—3 × 3 × 8 × 4
Bias—1 × 1 × 8Bias—1 × 1 × 4
ReLU8 × 8 × 8-ReLU4 × 4 × 4-
Max Pooling4 × 4 × 8-Transposed Convolution2 × 2 × 2Weights—3 × 3 × 4 × 2
Convolution4 × 4 × 4Weights—3 × 3 × 8 × 4Bias—1 × 1 × 2
Bias—1 × 1 × 4ReLU2 × 2 × 2-
ReLU4 × 4 × 4- Regression Output32 × 32 × 1-
Max Pooling2 × 2 × 4-
Convolution2 × 2 × 2Weights—3 × 3 × 4 × 2
Bias—1 × 1 × 2
ReLU2 × 2 × 2-
Max Pooling1 × 1 × 2-
Table 3. AHODnet layer description.
Table 3. AHODnet layer description.
Layer TypeActivationsLearnablesLayer TypeActivationsLearnables
Image Input32 × 32 × 3-CCUnit(2, 1)16 × 16 × 32Weights—3 × 3 × 16 × 32
Convolution32 × 32 × 16Weights—3 × 3 × 3 × 16Bias—1 × 1 × 32
Bias—1 × 1 × 16Addition16 × 16 × 32-
Batch Normalization32 × 32 × 16Offset—1 × 1 × 64ReLU16 × 16 × 32-
Scale—1 × 1 × 64CCUnit(2, 2)16 × 16 × 32Weights—3 × 3 × 16 × 32
ReLU32 × 32 × 16-Bias—1 × 1 × 32
Combined
Convolutional Unit(1,1)
Convolution32 × 32 × 16Weights—3 × 3 × 16 × 16
Bias—1 × 1 × 16
Addition16 × 16 × 32-
Batch Normalization32 × 32 × 16Offset –1 × 1 × 16ReLU16 × 16 × 32-
Scale—1 × 1 × 16CCUnit(3, 1)8 × 8 × 64Weights—3 × 3 × 32 × 64
ReLU32 × 32 × 16-Bias—1 × 1 × 64
Convolution32 × 32 × 16Weights—3 × 3 × 16 × 16Addition8 × 8 × 64-
Bias—1 × 1 × 16ReLU8 × 8 × 64-
Batch Normalization32 × 32 × 16Offset –1 × 1 × 16CCUnit(3, 2)8 × 8 × 64Weights—3 × 3 × 64 × 64
Scale—1 × 1 × 16Bias—1 × 1 × 64
Addition32 × 32 × 16-Addition8 × 8 × 64-
ReLU32 × 32 × 16-ReLU8 × 8 × 64-
CCUnit(1, 2)32 × 32 × 16Weights—3 × 3 × 16 × 16Average Pooling5 × 5 × 64-
Bias—1 × 1 × 16Fully Connected1 × 1 × 150Weights—150 × 1600
Addition32 × 32 × 16-Bias—150 × 1
ReLU32 × 32 × 16-Softmax1 × 1 × 150-
Table 4. Types of adversarial attack.
Table 4. Types of adversarial attack.
Adversarial Attack TypesDescriptionFunction
Untargeted AttackSample-specific attack design.
Aim to confuse prediction state from the original duplicate.
A(i + ψ) ≠ o
A is the deep learning model, i is the original sample, ψ is generated adversarial perturbation, and o is the original prediction.
Targeted AttackSample-specific attack design.
Aim to make the system predict the desired category.
A(i + ψ) = m
m is a predefined category
Universal AttackWell-designed general perturbation.
Inserting perturbation into unseen samples.
A(iu + ψ) = m
Iu is a different sample from the same class
Black-Box AttackAttacking an unknown ML model.
Perturbation of one model applied to another model.
Ab (ib) = A(ib)
Ab is another model than A and ib is adversarial samples produced using model A
White-Box AttackAttacking a Known ML model.
Perturbation is added to increase the loss.
A(ib+ψ*sign(Δ(ib)) = A(ib)
Δib is the magnitude of the gradient of the sample
Table 5. RMSE comparison of range estimation.
Table 5. RMSE comparison of range estimation.
MethodProposed AOHDLProposed OHDLEDACMDAEIYoLov4-tinySALA-LSTMCNNRNNFFTCFAR
SNR (dB)
01.38581.42301.79121.79792.40062.38902.75252.71513.22333.2134
50.96411.20640.99261.61541.75511.77492.20752.55083.07922.9172
100.40770.61990.97680.85611.11821.12231.56472.01852.58772.7920
150.19740.38590.93230.78071.10740.96061.52591.88162.53172.6641
200.18000.36660.69570.68790.98500.94901.36871.54872.33252.5249
250.16730.19630.66770.65180.83490.91611.23091.46341.88852.3750
300.09940.18160.51910.60900.44450.89861.18481.38581.87452.1016
Table 6. RMSE comparison of velocity estimation.
Table 6. RMSE comparison of velocity estimation.
MethodProposed AOHDLProposed OHDLEDACMDAEIYoLov4-tinySALA-LSTMCNNRNNFFTCFAR
SNR (dB)
00.73640.77380.89771.37871.87961.52432.23062.36002.77362.6874
50.38370.40850.82381.22441.17001.46051.49032.23952.43852.6203
100.27590.30180.80691.05241.16961.11701.47161.93562.40622.5430
150.05020.20570.78390.79191.02181.09091.43051.84071.97352.4357
200.01040.06950.72170.72550.90820.99461.41341.60861.83962.0496
250.01040.03320.64840.66920.35380.93301.31501.56481.76091.9006
300.00000.01390.59820.05990.11570.78091.18251.28691.61361.8503
Table 7. KPI comparison for different scenarios of adversarial attack in proposed AOHDL at 10 dB of SNR.
Table 7. KPI comparison for different scenarios of adversarial attack in proposed AOHDL at 10 dB of SNR.
KPIImplementation Methods
OHDL without AttackOHDL with AttackAOHDL with Attack
Accuracy (%)99.4097.5299.10
Precision (%)99.3997.4999.11
Recall (%)99.4197.7299.10
FPR0.300.870.45
F1-Score (%)99.3997.4899.10
Mathews Correlation Coefficient (%)99.1096.1398.65
Kappa Coefficient (%)98.6594.2297.98
KPI: key performance indicator.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Akhtar, M.M.; Li, Y.; Cheng, W.; Dong, L.; Tan, Y.; Geng, L. AOHDL: Adversarial Optimized Hybrid Deep Learning Design for Preventing Attack in Radar Target Detection. Remote Sens. 2024, 16, 3109. https://doi.org/10.3390/rs16163109

AMA Style

Akhtar MM, Li Y, Cheng W, Dong L, Tan Y, Geng L. AOHDL: Adversarial Optimized Hybrid Deep Learning Design for Preventing Attack in Radar Target Detection. Remote Sensing. 2024; 16(16):3109. https://doi.org/10.3390/rs16163109

Chicago/Turabian Style

Akhtar, Muhammad Moin, Yong Li, Wei Cheng, Limeng Dong, Yumei Tan, and Langhuan Geng. 2024. "AOHDL: Adversarial Optimized Hybrid Deep Learning Design for Preventing Attack in Radar Target Detection" Remote Sensing 16, no. 16: 3109. https://doi.org/10.3390/rs16163109

APA Style

Akhtar, M. M., Li, Y., Cheng, W., Dong, L., Tan, Y., & Geng, L. (2024). AOHDL: Adversarial Optimized Hybrid Deep Learning Design for Preventing Attack in Radar Target Detection. Remote Sensing, 16(16), 3109. https://doi.org/10.3390/rs16163109

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop