Next Article in Journal
A Novel Multistep Wavelet Convolutional Transfer Diagnostic Framework for Cross-Machine Bearing Fault Diagnosis
Previous Article in Journal
A Semi-Supervised Attention-Temporal Ensembling Method for Ground Penetrating Radar Target Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust DOA Estimation via a Deep Learning Framework with Joint Spatial–Temporal Information Fusion

1
School of Automation and Information Engineering, Xi’an University of Technology, Xi’an 710048, China
2
Shaanxi Key Laboratory of Complex System Control and Intelligent Information Processing, Xi’an University of Technology, Xi’an 710048, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(10), 3142; https://doi.org/10.3390/s25103142
Submission received: 7 April 2025 / Revised: 6 May 2025 / Accepted: 13 May 2025 / Published: 15 May 2025

Abstract

:
In this paper, we propose a robust deep learning (DL)-based method for Direction-of-Arrival (DOA) estimation. Specifically, we develop a novel CRDCNN-LSTM network architecture, which integrates a Cross-Residual Depthwise Convolutional Neural Network (CRDCNN) with a Long Short-Term Memory (LSTM) module for effective capture of both spatial and temporal features. The CRDCNN employs multi-level cross-residual connections and depthwise separable convolutions to enhance feature diversity while mitigating issues such as gradient vanishing and overfitting. Furthermore, a customized FD loss function, combining Focal Loss and Dice Loss, is introduced to emphasize low-confidence samples and promote sparsity in the spatial spectrum, thereby improving the precision and overall effectiveness of DOA estimation. A post-processing strategy based on peak detection and quadratic interpolation is also employed to refine DOA estimations and reduce quantization errors. Simulation results demonstrate that the proposed approach achieves significantly higher estimation accuracy and resolution than conventional methods and current DL models under varying SNR and snapshot conditions. In addition, it offers distinct advantages in terms of generalization and computational efficiency.

1. Introduction

Direction-of-Arrival (DOA) estimation constitutes a central focus in array signal processing and is widely applied in fields including radar detection, wireless communication, electronic countermeasures, acoustic direction finding, and astronomy. The primary objective of DOA estimation is to accurately determine the angles of arrival of incoming signals received by an antenna array, thereby providing critical parameters for further processing tasks such as target tracking and localization [1,2,3]. DOA estimation is a fundamental technology underpinning a wide range of applications, including modern wireless communication systems, intelligent transportation, Unmanned Aerial Vehicle (UAV) coordination, and future 6G networks [4,5,6]. For instance, in multiple-input multiple-output (MIMO) communication systems, accurate DOA estimation can significantly enhance beamforming performance and improve spatial multiplexing capabilities [7,8,9]. Similarly, in radar systems, it supports multi-target detection and tracking, thereby improving the accuracy and robustness of target identification. Accordingly, achieving high-precision and low-complexity DOA estimation under complex conditions has remained a critical research focus in both academia and industry [10,11].
Over the past decades, DOA estimation has been extensively studied, leading to the development of various classical methods. One of the earliest algorithms is the conventional beamforming (CBF) method, whose core principle is to exploit the array antenna’s response to signals from different directions and estimate the angles of arrival by constructing a spatial beam pattern [12,13]. However, the performance of this method is constrained by the Rayleigh resolution limit, making it difficult to distinguish signal sources with small angular separations. Additionally, it is highly susceptible to noise in low signal-to-noise ratio (SNR) environments, resulting in a significant degradation of estimation accuracy [14,15,16]. In response, subspace decomposition-based super-resolution algorithms were introduced in the 1980s to address these shortcomings [17,18]. The Multiple Signal Classification (MUSIC) algorithm exploits the orthogonality between the signal subspace and the noise subspace. It performs eigenvalue decomposition of the covariance matrix to extract the signal subspace and estimates the DOA by searching for peaks in the spatial spectrum [19,20]. Nevertheless, MUSIC suffers from high computational complexity in large-scale arrays and considerable performance degradation under coherent source conditions, limiting its practical applicability [21]. To address these challenges, the Estimation of Signal Parameters via Rotational Invariance Techniques (ESPRIT) algorithm was introduced in 1985. This method exploits the rotational invariance between subarrays and directly estimates the DOA parameters using a parameter matrix derived from eigenvalue decomposition, thus eliminating the need for spectral peak search and improving computational efficiency [22,23]. Other DOA estimation algorithms have also been developed in the statistical signal processing area, including the Minimum Variance Distortionless Response (MVDR) method. Although this method can enhance the accuracy of DOA estimation, it is sensitive to noise and the number of snapshots [24,25,26]. Similarly, the Maximum Likelihood (ML) method provides theoretically optimal estimation performance but suffers from prohibitively high computational complexity, making it unsuitable for real-time applications [27,28,29].
Despite the satisfactory performance of traditional DOA estimation methods under certain conditions, they exhibit notable limitations. These include poor estimation accuracy in low SNR environments, limited snapshots, coherent signal scenarios, and complex propagation channels. Moreover, they suffer from excessive computational complexity and strong sensitivity to parameter selection, which significantly hinders their practical deployment [30,31]. In recent years, with the rapid development of artificial intelligence technologies, deep learning (DL) has achieved remarkable success in fields such as computer vision, natural language processing, and medical image analysis. It has also been gradually introduced into DOA estimation research to enhance model adaptability under complex environments [32,33,34]. Unlike conventional model-based approaches, DL adopts a data-driven, end-to-end framework to directly learn the input–output relationship, offering improved generalization and robustness [9,35]. DL-based DOA estimation methods are typically categorized into classification and regression types. The classification methods discretize the angular space into several intervals and model the estimation as a classification problem, where neural networks are trained to predict the interval corresponding to the incident signal angle [36,37]. For example, DOA estimation algorithms based on convolutional neural networks (CNNs) utilize their spatial feature extraction capabilities and apply fully connected layers for classification, thereby improving resolution [36,38]. In contrast, regression-based methods predict the continuous DOA angles directly through deep neural networks, thereby circumventing the accuracy loss due to discretization. However, this also increases the complexity of signal feature extraction and processing [39,40]. Although DL has shown initial progress in the field of DOA estimation, various challenges still need to be addressed. Firstly, existing DL-based DOA estimation methods show limited feature extraction ability and stability when confronted with complex signal source scenarios, frequently encountering challenges such as gradient vanishing and feature degradation. Second, commonly used loss functions—primarily mean squared error (MSE) and binary cross-entropy (BCE)—fail to account for spatial spectrum sparsity and treat all samples equally, regardless of their confidence levels. This can lead to overfitting on easily learned samples and poor detection of low-confidence or weak signal sources. Additionally, many DL methods exhibit instability under extreme conditions like low SNR, small snapshot numbers, or high adjacent source density, generally lacking generalization ability, robustness, and computational efficiency, thereby constraining their practical application.
In this paper, a DL-based DOA estimation method is proposed. The approach employs a deep convolutional neural network enhanced by a cross-residual structure, coupled with a Long Short-Term Memory (LSTM) network to effectively capture both spatial and temporal features. An end-to-end training framework is employed, integrating a composite FD loss function that combines Focal Loss and Dice Loss to improve the sensitivity of model to low-confidence samples and promote sparsity in spatial spectrum predictions. Furthermore, peak detection and quadratic interpolation-based angle regression are introduced to mitigate quantization errors and further refine estimation accuracy. Simulation results confirm that the proposed method consistently outperforms traditional and existing DL approaches across various SNR levels, snapshot numbers, and resolution, offering a robust and high-precision solution for DOA estimation. The main contributions of this paper are summarized as follows:
  • A CRDCNN-LSTM architecture is proposed, designed for joint spatial–temporal feature fusion. The convolutional module includes multi-level cross-residual connections that mitigate the issues of traditional single-path feature flow, enhancing both multi-scale feature representation and feature diversity. The six-layer stacked CRDCNN configuration allows each layer to retain its features while passing the outputs of previous layers to subsequent ones, ensuring effective deep information propagation and addressing challenges such as gradient vanishing. The LSTM module captures temporal dependencies, which significantly enhances the robustness of DOA estimation under noisy conditions.
  • An FD loss function is designed, integrating Focal Loss and Dice Loss. Focal Loss introduces a modulation factor to down-weight easy samples and emphasize hard samples, thereby reducing overfitting and improving the detection of low-confidence or weak signals. The Dice Loss component optimizes the sparsity and distribution consistency of the spatial spectrum by quantifying the overlap between predicted and true spectra. The weighted combination of these two losses accelerates model convergence and improves generalization, particularly under low SNR or data imbalance conditions, thus enhancing the accuracy and robustness of DOA estimation.
  • To mitigate the impact of angle discretization and associated quantization errors in the estimation results, this paper introduces a post-processing strategy that combines peak detection with quadratic interpolation. The method first detects peaks in the spatial spectrum output by the network to locate prominent responses corresponding to signal directions. Then, quadratic interpolation is used to refine these peak positions to sub-pixel accuracy, enabling high-precision estimation in the continuous angular domain. This approach effectively alleviates quantization-induced localization errors and significantly enhances the resolution and stability of DOA estimation, particularly in scenarios involving closely spaced sources.
The remainder of this paper is organized as follows. Section 2 introduces the signal model based on the uniform linear array (ULA) and discusses the signal information contained in the spatial spectrum. Section 3 gives the architecture of the proposed model in detail. The advantages and disadvantages of the proposed framework are explored and compared with other common methods using simulated experiments in Section 4. Finally, Section 5 concludes the paper.

2. Singal Model

Consider a ULA consisting of N elements with an inter-element spacing of d. In the spatial domain, M uncorrelated far-field narrowband signals impinge upon the array from various directions, as illustrated in Figure 1.
Denote the number of snapshots per element as T. The signal received by the n-th array element at time instant t can be expressed as
x n ( t ) = i = 1 M s i ( t ) e j 2 π λ ( n 1 ) d sin θ i + n n ( t ) , n = 1 , 2 , , N
where λ is the signal wavelength, defined by λ = c / f , where c represents the speed of light and f is the frequency of the signal. s i ( t ) denotes the envelope of the i-th narrowband signal at time t, containing both amplitude and phase information. θ i indicates the direction of arrival of the i-th signal, while n n ( t ) represents additive noise, typically assumed to be zero-mean complex Gaussian white noise with known variance. To facilitate matrix-based derivation, let
x ( t ) = x 1 ( t ) , x 2 ( t ) , , x N ( t ) T N × 1
where ( · ) T denotes the transpose operation. Accordingly, the above expression can be rewritten as
x ( t ) = i = 1 M s i ( t ) a ( θ i ) + n ( t )
where n ( t ) is the noise vector, and a ( θ i ) denotes the array manifold vector of the i-th signal arriving from angle θ i , which is given by
a ( θ i ) = 1 , e j 2 π λ d sin θ i , e j 4 π λ d sin θ i , , e j 2 π λ ( N 1 ) d sin θ i T
by concatenating the array manifold vectors of all M signals, we obtain the following:
A ( θ ) = a ( θ 1 ) , a ( θ 2 ) , , a ( θ M ) N × M
All signal envelopes at time t can be arranged into the following vector:
s ( t ) = s 1 ( t ) , s 2 ( t ) , , s M ( t ) T
thus, Equation (1) can be further expressed as follows:
x ( t ) = A ( θ ) s ( t ) + n ( t )
In the case of offline or batch processing, T snapshots can be collected at times t = 1 , 2 , , T . In practice, the array covariance matrix R = E [ x ( t ) x ( t ) H ] is replaced by the sample covariance matrix R ^ , which is computed from the T snapshots of the array output x ( t ) and is given by
R ^ = 1 T t = 1 T x ( t ) x ( t ) H
where ( · ) H indicates the conjugate transpose operation. The covariance matrix R ^ reflects the statistical characteristics of the signals and noise across the array elements and provides essential information for subsequent spatial spectrum calculation. The steering vector based on R ^ is then defined as
a ( θ ) = exp j 2 π λ d sin ( θ )
where d = 0 , d , 2 d , , ( N 1 ) d T represents the spatial positions of the elements in the ULA. If we define
W ( θ ) = a ( θ ) a H ( θ )
the matrix W ( θ ) and the sample covariance matrix R ^ can be vectorized using vec ( · ) , resulting in the column vectors vec ( W ( θ ) ) and vec ( R ^ ) . Subsequently, the beam output amplitude at scanning angle is calculated as follows:
Z ( θ ) = vec ( W ( θ ) ) H vec ( R ^ )
by iteratively scanning θ across the predefined angular grid Θ = { θ 1 , θ 2 , , θ L } , the full spatial spectrum distribution { Z ( θ 1 ) , Z ( θ 2 ) , , Z ( θ L ) } is derived. The spatial spectrum generally shows significant peaks near the true incident angles of the signals, which can be used as feature inputs for neural network training.
In the context of data-driven DOA estimation, the actual signal directions are labeled during network training or testing by assigning 0-1 label vectors to the samples. Specifically, if the discrete angular grid Θ comprises L discrete angles { θ 1 , θ 2 , , θ L } , a label vector of length L is constructed as follows:
Y = y 1 , y 2 , , y L T
where
y l = 1 , θ l is   the   angle   of   sin gal 0 , else
For multi-source situations, one-hot encoding is applied by assigning 1 to the positions corresponding to each true signal direction in the label vector. The resulting 0–1 labels Y serve to guide the network in distinguishing signal components from noise during training, enhancing its ability to accurately localize source peaks in the spatial spectrum when handling new signals.

3. Proposed Method

This paper presents a DL-based DOA estimation approach, with its core being the construction of the Cross-Residual Deep Convolutional and Long Short-Term Memory Network. The network employs a CRDCNN to extract spatial features from the input signals and integrates an LSTM module for temporal sequence modeling, thereby improving DOA estimation performance under low SNR conditions, limited snapshot availability, and complex signal environments. According to Equation (11), the sample features are described by the spatial spectrum, and the feature matrix input to the network is defined as follows:
X = Z θ 1 , Z θ 2 , , Z θ L B × 1 × L
In this formulation, B denotes the batch size and L represents the number of discrete angular grid points. The proposed method operates as follows: The input features are first fed into the CRDCNN to perform multi-scale feature extraction. The LSTM module then captures the temporal dependencies of the sequential features, enabling the model to learn the evolution of DOA signals across different numbers of snapshots and enhancing temporal stability. Finally, a fully connected layer is used to generate the output, followed by a post-processing step that extracts the peak positions to obtain the final DOA estimates. The overall framework is illustrated in Figure 2, where the blue dashed box denotes the CRDCNN.

3.1. Network Architecture

3.1.1. Cross-Residual Deep Convolutional Network

For DOA estimation, the input features X encompass abundant spatial information, necessitating efficient extraction of the discriminative characteristics between signals and noise. Conventional DCNNs tend to suffer from problems such as information loss and gradient vanishing as network depth increases. Therefore, a cross-residual structure is introduced to enhance feature extraction and improve inter-layer information propagation. The CRDCNN consists of six cross-residual deep convolutional layers, each of which employs depthwise separable convolution to reduce computational complexity while maintaining effective feature representation capability. Let the input feature of the l-th layer be X l , the output as H l , and the convolution process is formulated as
H l = σ L BN Conv depthwise ( X l ) + Conv pointwise ( X l )
where Conv depthwise denotes depthwise convolution, Conv pointwise denotes pointwise convolution, BN represents batch normalization, and σ L is the LeakyReLU activation function. After feature extraction at each layer, CRDCNN fuses information H ˜ l through cross-residual connections, which can be written as
H ˜ l = H l + i = 1 l 1 α i H i
where α i is a learnable parameter that controls the residual information flow between different layers.
The CRDCNN consists of six layers designed for progressive feature extraction, gradually enhancing the network’s spatial perception ability. The specific architecture is depicted in Figure 3. The first layer employs depthwise separable convolutions with an increased number of channels and large receptive fields to capture global information. As the network deepens, both the dilation rate and kernel size are progressively reduced across layers, facilitating a transition from coarse to fine-grained feature extraction. Each layer receives inputs not only from the immediately preceding layer but also integrates features from all preceding layers, ensuring the full utilization of multi-scale representations. In the sixth layer, the number of channels is further expanded, and small-scale convolutions are applied to strengthen the expressiveness of the final features based on the fused multi-layer information.
During the final fusion stage of CRDCNN, the outputs from all layers are concatenated and subjected to a 1 × 1 convolution for dimensionality reduction, resulting in the final feature representation:
F ˜ = σ L BN Conv 1 × 1 [ H 1 , H 2 , , H 6 ]
where [ H 1 , H 2 , , H 6 ] denotes the concatenation of outputs from all layers. The cross-residual structure enhances inter-layer information flow, promoting feature reuse and improving the model’s representational capacity. The use of depthwise separable convolutions reduces computational complexity, making the model more trainable on large-scale datasets. Additionally, the combination of batch normalization (BN) and LeakyReLU effectively mitigates the gradient vanishing problem, improving training stability.

3.1.2. Long Short-Term Memory Network and Output Layer

Due to the temporal correlation of DOA signals across snapshots, temporal sequence modeling is essential to improve estimation stability. To this end, an LSTM network is employed to enhance the features extracted by the CRDCNN. LSTM is an improved variant of the recurrent neural network (RNN), featuring a memory cell (cell state) that retains historical information, along with input, forget, and output gates that regulate the flow and updating of information. The structure is illustrated in Figure 4. In this figure, is used to indicate element-wise addition, while   represents element-wise multiplication.
The output dimension of F ˜ is set to B × 16 × L , and the corresponding input feature matrix for the LSTM is given by
X LSTM = F ˜ T B × 16 × L
where 16 is the number of feature channels. The LSTM computation leverages the temporal evolution patterns of DOA signals. Given that the snapshot number T influences the statistical properties of the signals in DOA estimation, the LSTM utilizes feature inputs across time steps to capture the temporal sequence correlations. Let the LSTM hidden state dimension be h t and the cell state be c t , the operation of LSTM is described in Algorithm 1.
Algorithm 1: LSTM Operations at Each Time Step
Inputs:  X LSTM , t , h t 1 , c t 1 , W f , W i , W c , W o , b f , b i , b c , b o
Output:  h t , c t
  • Forget Gate:
          f t = σ S ( W f [ h t 1 , X LSTM , t ] + b f )
  • Input Gate:
          i t = σ S ( W i [ h t 1 , X LSTM , t ] + b i )
          c ˜ t = tanh ( W c [ h t 1 , X LSTM , t ] + b c )
  • Update Cell State:
          c t = f t c t 1 + i t c ˜ t
  • Output Gate:
          o t = σ S ( W o [ h t 1 , X LSTM , t ] + b o )
  • Compute Hidden State:
          h t = o t tan h ( c t )
Return  h t , c t
Where W f , W i , W c , W o and b f , b i , b c , b o are the learnable parameters of the LSTM, σ S denotes the Sigmoid activation function and tanh represents the hyperbolic tangent function. In this paper, a bidirectional LSTM (BiLSTM) structure is adopted to fully exploit temporal information. The computation of the BiLSTM is formulated as
h t = LSTM forward ( X LSTM , t , h t 1 )
h t = LSTM backward ( X LSTM , t , h t + 1 )
H t = [ h t , h t ]
By capturing global temporal dependencies from both past and future time steps, the bidirectional LSTM improves the robustness of DOA estimation. A fully connected (FC) layer is subsequently applied after the LSTM to yield the final angle estimation:
Y ^ = σ S ( W out H t + b out )
where W out and b out are the learnable parameters of the output layer, and Y ^ denotes the final output pseudo-spatial spectrum.

3.2. Loss Function

The design of the loss function is critical in training DL-based DOA estimation models. Conventional loss functions, such as mean squared error (MSE) or binary cross-entropy (BCE), may not adequately reflect the specific challenges of DOA estimation, particularly in environments with low SNR, few snapshots, or coherent signal sources. Under such conditions, networks are prone to noise interference, leading to reduced estimation accuracy. To mitigate this, the proposed method employs a combination of Focal Loss and Dice Loss, termed FD Loss, to increase attention to low-confidence samples, enforce sparsity in the predicted spatial spectrum, and enhance the robustness of the DOA estimation.
Focal Loss is an advanced loss function based on cross-entropy, developed to handle class imbalance in tasks such as object detection. In DOA estimation, true DOA positions occupy only a small fraction of the spatial spectrum, while the majority of positions correspond to near-zero probabilities. This highly imbalanced distribution causes standard cross-entropy to overemphasize high-probability regions and neglect low-probability areas. Focal Loss mitigates this issue by introducing a modulation factor that decreases the influence of well-classified samples and increases the focus on difficult, low-confidence samples, thereby improving DOA estimation accuracy. The mathematical formulation is
L FL = α ( 1 p t ) γ log ( p t )
where p t denotes the predicted probability of the target class by the network, defined as follows:
p t = y i p + ( 1 y i ) ( 1 p )
where y i Y denotes the ground-truth DOA label, while p represents the predicted probability at angle θ i . For positive samples (locations corresponding to DOA), p t = p ; for negative samples (non-DOA locations), p t = 1 p . The modulation factor ( 1 p t ) γ in Focal Loss controls the loss weighting for samples with varying confidence levels. When p t approaches 1 (indicating accurate prediction), the loss contribution diminishes, reducing the influence of easy samples. Conversely, for low-confidence samples (small p t ), the loss increases, prompting the model to focus more on these difficult cases and enhancing DOA estimation stability. The parameter γ adjusts the modulation strength, larger values accentuate hard samples, while smaller values make Focal Loss converge to cross-entropy loss.
The spatial spectrum in DOA estimation is characterized by sparsity, where true DOA targets occupy only a small fraction, and the majority corresponds to background regions (non-DOA positions), resulting in significant class imbalance. The incorporation of Focal Loss allows the model to dynamically adjust the weighting of different samples, ensuring greater focus on DOA signal positions and improving detection performance, particularly in low SNR environments.
Dice Loss, initially introduced in the context of medical image segmentation, is designed to address extreme class imbalance. The central idea is to calculate the Dice similarity coefficient between the predictions of model and the true DOA labels, thereby quantifying their alignment and optimizing estimation performance. Its mathematical expression is formulated as
L Dice = 1 2 p i y i + ϵ p i + y i + ϵ
where p i refers to the predicted probability at the i-th angular position, and ϵ is a small constant (e.g., 10 6 ) introduced to avoid division by zero. Dice Loss enhances DOA estimation accuracy by maximizing the similarity between predicted outputs and true DOA labels via the Dice coefficient. Given the sparse nature of DOA labels, Dice Loss effectively aligns the predicted spatial spectrum with the ground-truth DOA distribution and mitigates the impact of background noise. Moreover, Dice Loss offers stable gradients, enabling faster convergence during early training phases and facilitating efficient learning of DOA features even when data are limited.
In the tasks of DOA estimation, applying either Focal Loss or Dice Loss independently may not effectively balance learning focus on low-confidence samples and global optimization of the spatial spectrum. Therefore, FD Loss is introduced by weighted fusion of Focal Loss and Dice Loss to enhance estimation precision. The formulation is as follows:
L FD = α L FL + β L Dice
where α and β correspond to the weights assigned to Focal Loss and Dice Loss, respectively, balancing their influence throughout training. According to Equation (22), the network output pseudo-spatial spectrum is denoted by Y ^ , and the true DOA labels obtained from Equation (12) are denoted as Y . Thus, the FD loss function is expressed as
L FD = α L FL ( Y ^ , Y ) + β L Dice ( Y ^ , Y )
The first term indicates the Focal Loss, which applies higher weights to low-confidence samples at DOA positions, enhancing the ability of network to capture difficult cases. The second term refers to the Dice Loss, which optimizes the overall correspondence between the predicted and true DOA spatial spectra, thereby improving prediction accuracy.

3.3. Post-Processing

After the input signals are processed by the CRDCNN-LSTM network, a continuous probability distribution Y ^ is produced. Using Y ^ directly for DOA estimation may lead to significant inaccuracies. To address this, a post-processing step is adopted to extract the final estimated DOA angles. As the network output consists of unnormalized activations, it is necessary to apply a Sigmoid function to convert them into probability values:
Ψ = σ ( Y ^ )
This operation maps the raw prediction values into the [0, 1] interval, giving the network output probabilistic significance and facilitating subsequent DOA angle estimation. In this work, two steps—peak detection and angle refinement—are applied to extract peak values from the pseudo-spatial spectrum Ψ and determine the angles of the DOA signals. First, peak detection is used to identify local maxima within Ψ :
Θ peak = { θ i Ψ i > Ψ i 1   &   Ψ i > Ψ i + 1 , i { 2 , , L 1 } }
Specifically, when the probability value Ψ i at angle θ i is greater than those at its neighboring grid points, θ i is identified as a peak DOA angle. Through peak detection, an initial set of candidate DOA angles can be obtained. However, since the discrete angular grid Θ is finite, the true DOA angles may not exactly coincide with the grid points. To address this, an angle refinement process is applied to the detected peaks to further improve estimation accuracy. The core idea of angle refinement is to perform quadratic interpolation around the detected peak positions to estimate the actual DOA angles. Let the network outputs at a peak position θ i and its two adjacent points be Ψ i 1 , Ψ i , Ψ i + 1 , respectively. The final estimated DOA angle θ ^ at the peak position is obtained using quadratic interpolation as follows:
θ ^ = θ i + Ψ i + 1 Ψ i 1 2 ( Ψ i 1 2 Ψ i + Ψ i + 1 ) θ ¯
where θ ¯ denotes the grid step size, i.e., the interval between two adjacent angular grid points. The derivation of this formula is based on fitting the probability distribution near the peak with a quadratic function and solving for its extremum to obtain a refined DOA estimate. This angle regression technique effectively eliminates the quantization errors caused by grid discretization and leads to a significant improvement in the resolution of the final DOA estimation.

4. Simulation Results

4.1. Data Generation

In this section, signal samples are generated using a uniform linear array, as illustrated in Figure 1. The array configuration includes M = 8 elements with an element spacing of d = λ / 2 , and the number of sources is N = 2. The number of snapshots is T = 256. The angle grid for the simulated array covers the range Θ   [ 75 ° , 75 ° ] , partitioned at intervals of 1°. Twenty distinct angular separations are considered, covering intervals of 2 ° , 4 ° , , 40 ° . Signal samples are generated corresponding to each angular separation Δ θ q , and the angles of the generated samples span
θ q = 75 ° + l q 1 ° θ q = 75 ° + l q 1 ° + Δ θ q
The index l q ranges from   0 , , D q , where D q = 150 Δ θ q + 1 . A total of 2600 angle sample sets are generated for the 20 angular intervals. The signal-to-noise ratios are randomly chosen within the range [−10 dB, 10dB], and the corresponding labels are derived from the 0–1 spatial spectrum obtained via Equation (12). Overall, 26,000 simulation samples are generated, covering various angular separations and noise levels.

4.2. DOA Estimation Performance

Two far-field narrowband independent signals with identical SNRs of 0 dB are considered, with angular intervals of { 5 ° , 15 ° , 25 ° } . For each angular separation, Equation (31) is employed to generate the corresponding angle values, which are used to produce the test samples. Following the reconstruction of the spatial spectrum, the DOA estimates for each test sample are calculated according to Equation (30). The DOA estimation results obtained by the three DL methods are displayed in Figure 5, with the estimated DOAs shown on the left and the corresponding estimation errors on the right.
Figure 5a,b illustrate that the proposed method demonstrates exceptional performance in DOA estimation tasks under typical scenarios. It achieves consistently high estimation accuracy across all angular separations and significantly surpasses competing methods, especially in the presence of closely spaced signal sources. Furthermore, the estimation errors are consistently maintained within ±1° even in proximity to the grid boundaries. This superior performance is primarily due to the FD loss function optimization, which allows the model to leverage boundary samples more effectively during training and improves its generalization across the full range of angles. In comparison, while the DCNN method (Figure 5c,d) exhibits generally balanced estimations, its accuracy is notably lower than that of the proposed method, with errors exceeding 2.5° close to the grid boundaries. Despite the enhanced feature extraction capability of the Res-DCNN method (Figure 5e,f), the error distribution is relatively irregular, particularly at the 5° angular separation.

4.3. Generalization Capability Under Multi-Source Scenarios

In order to assess the generalization performance of various DOA estimation algorithms under multi-source conditions, this experiment generates sample data in scenarios with multiple signal sources. The proposed approach is compared with other DL methods as well as traditional techniques such as L1-SVD, MUSIC, and MVDR. Specifically, the number of sources is set to M = 5, with SNR = 0 dB, T = 100, and source angles positioned at [ 46 ° , 23 ° , 0 ° , 23 ° , 46 ° ] . The DOA estimation results are visualized by plotting the spatial spectrum in a 3D space.
As shown in Figure 6, significant differences are observed in the spatial spectrum estimation results of various DOA estimation algorithms under a multi-source environment. Traditional methods such as L1-SVD produce sharp spectral peaks but suffer from the presence of spurious peaks. Both MUSIC and MVDR generate relatively clear spectral peaks, with MUSIC exhibiting sharper peaks; however, these methods require prior knowledge of the number of signal sources. In contrast, the proposed method produces sharper spectral peaks than MUSIC without needing the number of sources as a prior. The DCNN model yields only a three dominant peak, indicating limited generalization capability. Although the Res-DCNN model enhances feature extraction, it still suffers from missed sources and spurious peaks.
To further verify its generalization, the proposed model was tested on data with unknown source numbers M 1 , 2 , 3 , 4 , 5 , with 200 samples generated for each case and no prior knowledge of M provided. As shown in Figure 7, the confusion matrix illustrates that most predictions align with the ground truth, especially for M = 2 to 4. Even for more complex cases with M = 1 and 5, the model remains accurate despite being trained only on two-source samples, demonstrating strong robustness and generalization beyond its training distribution.
The method incorporates cross-residual convolutional layers and LSTM networks to exploit both spatial and temporal information, leading to enhanced DOA feature learning and effective suppression of false peaks. The integration of the FD loss function further optimizes key angle detection and noise robustness, ensuring the method achieves superior spectral peak definition, high resolution, and strong generalization performance.

4.4. Statistical Performance Analysis

In the DOA estimation experiments, we use the root mean square error (RMSE) as the key evaluation metric to assess the estimation accuracy of different models under different experimental setups [41]. The formula for RMSE is as follows:
RMSE = 1 K M k = 1 K m = 1 M ( θ ^ m , k θ m ) 2
where K specifies the number of Monte Carlo runs, M is the number of signal sources, θ ^ m , k denotes the estimated direction of arrival for the m-th signal in the k-th experiment, and θ m corresponds to the true angle. In order to incorporate a more comprehensive lower bound evaluation of the proposed method, we have additionally included the Cramér–Rao Lower Bound (CRLB) as a baseline alongside the previously discussed comparison methods [42].

4.4.1. Impact of Signal-to-Noise Ratio on Estimation Accuracy

In this experiment, the number of snapshots is set to T = 256 and signal source angles of θ 1 = 11 ° and θ 2 = 23 ° . The SNR is selected as the experimental variable, varying from −10 dB to 10 dB in steps of 2 dB, resulting in 11 SNR scenarios. For each scenario, 600 Monte Carlo experiments are conducted to ensure statistical reliability. Under these conditions, only the noise level is altered to assess its impact on DOA estimation accuracy.
Figure 8 shows that RMSE decreases as SNR increases for most methods. The proposed method maintains the lowest RMSE across the entire SNR range and gradually approaches the CRLB when SNR exceeds 0 dB. Under low SNR conditions (SNR < 0 dB), it achieves significantly lower errors and a more rapid RMSE reduction, indicating strong robustness to noise. This advantage results from the cooperative extraction of spatial and temporal features by the CRDCNN-LSTM framework, along with the effective role of the FD loss function in key angle detection and noise reduction. By contrast, although the RMSE of the Res-DCNN and DCNN methods decreases with increasing SNR, both suffer from large estimation errors under low SNR conditions (notably SNR < −5 dB), reflecting limited robustness. The traditional L1-SVD and MUSIC algorithms similarly perform inadequately at low SNRs, with L1-SVD exhibiting consistently high RMSE and MUSIC being highly susceptible to noise, offering marginal advantages only at high SNRs. The MVDR method shows high RMSE under all SNR conditions, particularly struggling to deliver effective DOA estimation in low SNR environments.

4.4.2. Impact of Snapshot Number on Estimation Accuracy

In this experiment, the SNR is set to 0 dB, and the source angles are configured as θ 1 = 11 ° and θ 2 = 23 ° . The experimental variable is the number of snapshots T, varied from 50 to 500 with a step of 50, resulting in multiple snapshot conditions. A total of 600 Monte Carlo trials are performed for each condition to ensure the statistical reliability of the results. Under these conditions, only the snapshot number is altered to investigate its influence on DOA estimation accuracy.
As shown in Figure 9, the number of snapshots has a significant impact on the DOA estimation accuracy of each method. Overall, most methods exhibit a decreasing trend in RMSE as the number of snapshots increases. However, noticeable differences in performance are observed among the methods under both low and high snapshot conditions. For snapshot numbers below 150, the MUSIC method produces notably higher RMSE values than other methods. While the MVDR method shows better performance than MUSIC, its overall error remains considerable, and the RMSE reduction with increasing snapshots is marginal. The L1-SVD approach exhibits relatively stable RMSE values over the entire snapshot range, consistently higher than those of other methods. In contrast, the DCNN and Res-DCNN methods show a reduction in RMSE as the snapshot number increases, but the decrease is gradual, and after T > 300, the convergence rate slows, and even slightly rebounds. Under T < 150, the RMSE of the proposed method is slightly higher than that of the MVDR method. As the number of snapshots increases, the error decreases more noticeably. When the snapshot count exceeds 200, the RMSE becomes significantly lower than that of other methods and gradually approaches the CRLB, demonstrating the effectiveness of the proposed method in different scenarios.

4.5. Resolution Probability Analysis for Closely Spaced Sources

In order to assess the capability of the model in handling closely spaced signal sources, the resolution probability is defined as follows: In each simulation, let θ ^ 1 and θ ^ 2 be the estimated angles corresponding to the true angles θ 1 and θ 2 . If the sum of the absolute differences θ ^ 1 θ 1 + θ ^ 2 θ 2 2 ° , we consider that simulation a successful resolution; otherwise, it is considered a resolution failure, and the resolution probability is obtained by computing the ratio of successful resolution occurrences. Two groups of experiments are configured to examine the effects of SNR and the number of snapshots, with the signal source angles uniformly set at θ 1 = 1.32 ° and θ 2 = 2.57 ° . The first group of experiments involves varying the SNR from −10 dB to 10 dB in 2 dB steps, with a fixed snapshot number of T = 256. In the second group, the number of snapshots is varied from 50 to 500 in steps of 50, while the SNR is held constant at 5 dB. For each SNR or snapshot condition, 600 Monte Carlo simulations are carried out to ensure result reliability.
The experimental results shown in Figure 10 and Figure 11 indicate that both the SNR and the number of snapshots have a significant impact on the ability of each method to resolve closely spaced signal sources. (It should be noted that the L1-SVD and MVDR methods fail to effectively distinguish adjacent signal sources under the current angular separation settings and are therefore excluded from further analysis.) Figure 10 clearly indicates that the proposed method delivers the best performance throughout the full SNR range. It retains the ability to distinguish closely spaced sources even when SNR < −5 dB, demonstrating excellent noise robustness. At 0 dB SNR, the resolution probability approaches 100%, which is significantly higher than that of other methods, confirming its effectiveness in low-SNR conditions. In low SNR scenarios (SNR < 0 dB), the Res-DCNN and MUSIC methods exhibit an inability to effectively resolve signal sources. Although the DCNN method demonstrates a relatively high resolution probability at SNR > –5 dB, its overall performance remains inferior to the proposed method when considering the results of all preceding experiments.
Figure 11 shows that the proposed method maintains high resolution capability even under low snapshot conditions, with the resolution probability rapidly approaching 100% as the number of snapshots increases. Conversely, Res-DCNN exhibits consistently weak resolution ability, and while MUSIC shows improvement with more snapshots, its convergence rate is slow and does not match the performance of the proposed method.

4.6. Computational Efficiency Evaluation

The computational complexity of each method is evaluated by presenting their respective time metrics in Figure 12 and Table 1. We constructed 10,000 random test samples distributed over different angles and SNR ranges, and all other parameters were set in accordance with Section 4.1. Experiments were conducted under Python 3.9 using PyTorch version 1.11.0 (with CUDA 11.3) on an NVIDIA GeForce RTX 4060 Ti GPU with 32 GB RAM. All randomly generated samples were processed using different methods, and the total prediction time was recorded. Compared to the other two DL-based approaches, the proposed model incurs a longer training time owing to its more complex architecture, which allows for direct input-to-angle estimation mapping. Although conventional methods bypass the need for training, experimental results show that DL models, once trained, can compute DOA estimates efficiently even with moderate computational resources. Due to its structural complexity, the proposed model exhibits marginally longer testing times than other DL models, yet remains significantly faster than traditional techniques, offering real-time capability alongside high measurement accuracy.

5. Conclusions

In this paper, we propose an innovative and robust DL-based framework for DOA estimation, featuring a CRDCNN-LSTM network integrated with an enhanced FD loss function. By combining cross-residual convolutional layers with LSTM units, the CRDCNN-LSTM architecture effectively captures both spatial and temporal signal characteristics, thereby enhancing the capability for signal feature extraction. The designed FD loss function, combining Focal Loss and Dice Loss, enhances model focus on weak and low-confidence signals and enforces sparsity in the output spectrum, leading to improved estimation accuracy and robustness under low SNR and limited snapshot conditions. Additionally, applying peak detection and quadratic interpolation in the post-processing stage further refines DOA predictions by mitigating discretization errors. Simulation results demonstrate that, without requiring prior knowledge of the number of sources, the proposed method achieves substantial RMSE reduction under various SNR and snapshot conditions, while offering higher resolution probability and greater computational efficiency compared to conventional algorithms. Moreover, relative to existing DL approaches, it also achieves lower RMSE, higher resolution probability, and notably stronger generalization capability. These quantitative results confirm the significant advantages of the proposed algorithm. Future work will focus on integrating this method into joint communication and sensing systems for real-world applications.

Author Contributions

Conceptualization, Y.Z. and X.F.; methodology, Y.Z.; software, J.L.; validation, J.L.; formal analysis, J.L.; investigation, J.L.; resources, J.L.; data curation, Y.Z.; writing—original draft preparation, J.L. and Y.Z.; writing—review and editing, Y.Z. and X.F.; visualization, J.L. and Y.Z.; supervision, X.F. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Project of the Ministry of Science and Technology under Grant No. 2022041009L, and in part by the key research and development plan of Shaanxi province under Grant No. 2021GY-072.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article; further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Chowdhury, M.W.T.S.; Mastora, M. Performance Analysis of MUSIC Algorithm for DOA Estimation with Varying ULA Parameters. In Proceedings of the 2020 23rd International Conference on Computer and Information Technology (ICCIT), Dhaka, Bangladesh, 19–21 December 2020; pp. 1–5. [Google Scholar]
  2. Rayar, V.; Naik, U.; Manage, P. A Survey on DoA Measurement using ULA and UCA for Wireless Sensor Network Applications. In Proceedings of the 2020 3rd International Conference on Intelligent Sustainable Systems (ICISS), Thoothukudi, India, 3–5 December 2020; pp. 1145–1149. [Google Scholar]
  3. Shi, W.; Li, Y.; Huang, Z. Sparse Linear Array Designed via ULA Fitting. In Proceedings of the 2022 IEEE 10th Asia-Pacific Conference on Antennas and Propagation (APCAP), Xiamen, China, 4–7 November 2022; pp. 1–2. [Google Scholar]
  4. Kassir, H.A.; Rekanos, I.T.; Lazaridis, P.I.; Yioultsis, T.V.; Kantartzis, N.V.; Antonopoulos, C.S.; Karagiannidis, G.K.; Zaharis, Z.D. DOA Estimation for 6G Communication Systems. In Proceedings of the 2023 12th International Conference on Modern Circuits and Systems Technologies (MOCAST), Athens, Greece, 28–30 June 2023; pp. 1–4. [Google Scholar]
  5. Balamurugan, N.M.; Mohan, S.; Adimoolam, M.; John, A.; Gadekallu, T.R.; Wang, W.Z. DOA tracking for seamless connectivity in beamformed IoT-based drones. Comput. Stand. Interfaces 2022, 79, 103564. [Google Scholar] [CrossRef]
  6. Chen, C.B.; Lo, T.Y.; Chang, J.Y.; Huang, S.P.; Tsai, W.T.; Liou, C.Y.; Mao, S.G. Precision Enhancement of Wireless Localization System Using Passive DOA Multiple Sensor Network for Moving Target. Sensors 2022, 22, 7563. [Google Scholar] [CrossRef] [PubMed]
  7. Liu, B.; Chen, B.; Yang, M.; Xu, H. DOA estimation using sparse Bayesian learning for colocated MIMO radar with dynamic waveforms. In Proceedings of the 2020 IEEE 11th Sensor Array and Multichannel Signal Processing Workshop (SAM), Hangzhou, China, 8–11 June 2020; pp. 1–4. [Google Scholar]
  8. Ma, Y.; Cao, X.; Wang, X. Enhanced DOA Estimation for MIMO radar in the Case of Limited Snapshots. In Proceedings of the 2020 IEEE 11th Sensor Array and Multichannel Signal Processing Workshop (SAM), Hangzhou, China, 8–11 June 2020; pp. 1–5. [Google Scholar]
  9. Tang, H.; Zhang, Y.; Luo, J.; Zhang, Y.; Huang, Y.; Yang, J. Sparse DOA Estimation Based on a Deep Unfolded Network for MIMO Radar. In Proceedings of the IGARSS 2023—2023 IEEE International Geoscience and Remote Sensing Symposium, Pasadena, CA, USA, 16–21 July 2023; pp. 5547–5550. [Google Scholar]
  10. Takahashi, Y.; Yamada, H.; Yamaguchi, Y. Array calibration techniques for DOA estimation with arbitrary array using root-MUSIC algorithm. In Proceedings of the 2011 IEEE MTT-S International Microwave Workshop Series on Innovative Wireless Power Transmission: Technologies, Systems, and Applications, Kyoto, Japan, 12–13 May 2011; pp. 235–238. [Google Scholar]
  11. Guan, H.X.; Ding, S.W.; Dai, W.; Tan, X.B.; Long, Y.H.; Liang, J.E. Low complexity DOA estimation based on weighted noise component subtraction for smart-home application. Appl. Acoust. 2025, 231, 110490. [Google Scholar] [CrossRef]
  12. Schmidt, R. Multiple emitter location and signal parameter estimation. IEEE Trans. Antennas Propag. 1986, 34, 276–280. [Google Scholar] [CrossRef]
  13. Xu, Z.Z.; Li, H.; Yang, K. A Modified Differential Beamforming and Its Application for DOA Estimation of Low Frequency Underwater Signal. IEEE Sens. J. 2020, 20, 8890–8902. [Google Scholar] [CrossRef]
  14. Chai, Y. Advanced Techniques in Adaptive Beamforming for Enhanced DOA Estimation. In Proceedings of the 2024 International Wireless Communications and Mobile Computing (IWCMC), Ayia Napa, Cyprus, 27–31 May 2024; pp. 269–273. [Google Scholar]
  15. Lin, T.; Zhou, X.; Zhu, Y.; Jiang, Y. Hybrid Beamforming Optimization for DOA Estimation Based on the CRB Analysis. IEEE Signal Process. Lett. 2021, 28, 1490–1494. [Google Scholar] [CrossRef]
  16. Singh, M.; Wajid, M. Comparative Analysis of Conventional and Adaptive beamforming for Linear Array. In Proceedings of the 2021 6th International Conference on Signal Processing, Computing and Control (ISPCC), Waknaghat, India, 7–9 October 2021; pp. 576–580. [Google Scholar]
  17. Paulraj, A.; Reddy, V.U.; Shan, T.J.; Kailath, T. Performance Analysis of the Music Algorithm with Spatial Smoothing in the Presence of Coherent Sources. In Proceedings of the MILCOM 1986–IEEE Military Communications Conference: Communications-Computers: Teamed for the 90’s, Monterey, CA, USA, 5–9 October 1986; pp. 41.5.1–41.5.5. [Google Scholar]
  18. Zoltowski, M.; Haber, F. A vector space approach to direction finding in a coherent multipath environment. IEEE Trans. Antennas Propag. 1986, 34, 1069–1079. [Google Scholar] [CrossRef]
  19. Ateşavcı, C.S.; Bahadırlar, Y.; Aldırmaz-Çolak, S. DoA Estimation in the Presence of Mutual Coupling Using Root-MUSIC Algorithm. In Proceedings of the 2021 8th International Conference on Electrical and Electronics Engineering (ICEEE), Antalya, Turkey, 9–11 April 2021; pp. 292–298. [Google Scholar]
  20. Vesa, A.; Simu, C. Performances of Uniform Sensor Array Antenna in case of DoA estimation using the MUSIC Algorithm. In Proceedings of the 2022 International Symposium on Electronics and Telecommunications (ISETC), Timisoara, Romania, 10–11 November 2022; pp. 1–4. [Google Scholar]
  21. Khichar, S.; Santipach, W.; Wuttisittikulkij, L. Covariance Matrix Reconstruction to Improve DoA Estimation Using Subspace Method in Low SNR Regime. IEEE Access 2025, 13, 26695–26706. [Google Scholar] [CrossRef]
  22. Paulraj, A.; Roy, R.; Kailath, T. Estimation of Signal Parameters Via Rotational Invariance Techniques-Esprit. In Proceedings of the Nineteeth Asilomar Conference on Circuits, Systems and Computers, Pacific Grove, CA, USA, 6–8 November 1985; pp. 83–89. [Google Scholar]
  23. Roy, R.; Paulraj, A.; Kailath, T. Estimation of Signal Parameters via Rotational Invariance Techniques–ESPRIT. In Proceedings of the MILCOM 1986—IEEE Military Communications Conference: Communications-Computers: Teamed for the 90’s, Monterey, CA, USA, 5–9 October 1986; pp. 41.46.41–41.46.45. [Google Scholar]
  24. Job, M.; Yadav, R.S. High Resolution DOA Estimation of Narrowband Signal for MUSIC, MVDR and Beamscan Algorithm. In Proceedings of the 2023 11th International Symposium on Electronic Systems Devices and Computing (ESDC), Sri City, India, 4–6 May 2023; pp. 1–5. [Google Scholar]
  25. Propastin, A.; Prokhorenko, V. Determining the DOA of Jamming Signals Using Root-Music and MVDR Algorithms for Planar Elliptical Digital Antenna Array. In Proceedings of the 2023 5th International Youth Conference on Radio Electronics, Electrical and Power Engineering (REEPE), Moscow, Russia, 16–18 March 2023; pp. 1–6. [Google Scholar]
  26. Shen, C.C.; Jhang, W. Joint CFO and DOA Estimation Based on MVDR Criterion in Interleaved OFDMA/SDMA Uplink. IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 2024, E107a, 1066–1070. [Google Scholar] [CrossRef]
  27. Liang, N.; Wang, S. Iterative Maximum Likelihood DOA Estimator Aided by Magnitude Measurements. In Proceedings of the 2022 IEEE 8th International Conference on Computer and Communications (ICCC), Chengdu, China, 9–12 December 2022; pp. 1556–1560. [Google Scholar]
  28. Zhao, Y.; Xin, J.; Wu, S. DOA Estimation Method Based on Maximum Likelihood for Nest Array Via Sparse Representation. In Proceedings of the 2021 CIE International Conference on Radar (Radar), Haikou, China, 15–19 December 2021; pp. 358–361. [Google Scholar]
  29. Zhu, Z.; Qiu, W. Maximum Likelihood DOA Estimation using a Clustering Based Hybrid Optimization Algorithm. In Proceedings of the 2024 7th International Conference on Information Communication and Signal Processing (ICICSP), Zhoushan, China, 21–23 September 2024; pp. 80–84. [Google Scholar]
  30. Fang, W.; Yu, D.; Wang, X.; Xi, Y.; Cao, Z.; Song, C.; Xu, Z. A Deep Learning Based Mutual Coupling Correction and DOA Estimation Algorithm. In Proceedings of the 2021 13th International Conference on Wireless Communications and Signal Processing (WCSP), Changsha, China, 20–22 October 2021; pp. 1–5. [Google Scholar]
  31. Hassan, H.; Maud, A.R.; Amin, M. Deep Learning Based DOA Estimation in Low SNR and Multipath Scenarios. In Proceedings of the 2024 IEEE International Symposium on Phased Array Systems and Technology (ARRAY), Boston, MA, USA, 15–18 October 2024; pp. 1–6. [Google Scholar]
  32. Zheng, R.X.; Sun, S.Q.; Liu, H.S.; Chen, H.L.; Li, J. Interpretable and Efficient Beamforming-Based Deep Learning for Single-Snapshot DOA Estimation. IEEE Sens. J. 2024, 24, 22096–22105. [Google Scholar] [CrossRef]
  33. Tian, Q.; Cai, R.Y.; Qiu, G.R.; Luo, Y. Distributed source DOA estimation based on deep learning networks. Signal Image Video Process 2024, 18, 7395–7403. [Google Scholar] [CrossRef]
  34. Song, J.; Zhao, Z.M.; Yang, K.H.; Cao, L.; Wang, D.F.; Fu, C. Deep learning-enhanced atomic norm minimization for DOA estimation of coherent and incoherent sources using coprime array. Meas. Sci. Technol. 2025, 36, 016163. [Google Scholar] [CrossRef]
  35. Su, Y.; Wang, X.; Li, L. A Novel Off-Grid Deep Learning framework for DOA Estimation. In Proceedings of the 2024 IEEE International Conference on Signal, Information and Data Processing (ICSIDP), Zhuhai, China, 22–24 November 2024; pp. 1–4. [Google Scholar]
  36. Dharma Theja ch, L.; Nagaraju, L.; Puli, K.K. Classification based DOA estimation using ANN and CNN Models. In Proceedings of the 2022 IEEE Microwaves, Antennas, and Propagation Conference (MAPCON), Bangalore, India, 12–16 December 2022; pp. 1470–1473. [Google Scholar]
  37. Zhao, Z.; Li, J. DoA Estimation based on Deep Learning in Low SNR. In Proceedings of the 2023 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB), Beijing, China, 14–16 June 2023; pp. 1–4. [Google Scholar]
  38. Tian, Q.; Cai, R.Y.; Luo, Y.; Qiu, G.R. DOA Estimation: LSTM and CNN Learning Algorithms. Circuits Syst. Signal Process. 2025, 44, 652–669. [Google Scholar] [CrossRef]
  39. Wu, X.; Wang, J.; Yang, X.; Tian, F. A Gridless DOA Estimation Method Based on Residual Attention Network and Transfer Learning. IEEE Trans. Veh. Technol. 2024, 73, 9103–9108. [Google Scholar] [CrossRef]
  40. Xu, S.Y.; Wang, Z.H.; Zhang, W.; He, Z.S. End-to-End Regression Neural Network for Coherent DOA Estimation with Dual-Branch Outputs. IEEE Sens. J. 2024, 24, 4047–4056. [Google Scholar] [CrossRef]
  41. Hannan, M.A.; Crisafulli, O.; Giammello, G.; Sorbello, G. On the Error Metrics Used for Direction of Arrival Estimation. Sensors 2025, 25, 2358. [Google Scholar] [CrossRef] [PubMed]
  42. Liu, Z.-M. Conditional Cramér–Rao Lower Bounds for DOA Estimation and Array Calibration. IEEE Signal Process. Lett. 2014, 21, 361–364. [Google Scholar] [CrossRef]
Figure 1. Uniform linear array structure.
Figure 1. Uniform linear array structure.
Sensors 25 03142 g001
Figure 2. Proposed DOA estimation method.
Figure 2. Proposed DOA estimation method.
Sensors 25 03142 g002
Figure 3. Proposed CRDCNN architecture.
Figure 3. Proposed CRDCNN architecture.
Sensors 25 03142 g003
Figure 4. Computation diagram of LSTM unit.
Figure 4. Computation diagram of LSTM unit.
Sensors 25 03142 g004
Figure 5. DOA estimation performance of different DL methods. (a) Proposed method for DOA estimation. (b) Estimation errors of Proposed. (c) DCNN for DOA estimation. (d) Estimation errors of DCNN. (e) Res-DCNN for DOA estimation. (f) Estimation errors of Res-DCNN.
Figure 5. DOA estimation performance of different DL methods. (a) Proposed method for DOA estimation. (b) Estimation errors of Proposed. (c) DCNN for DOA estimation. (d) Estimation errors of DCNN. (e) Res-DCNN for DOA estimation. (f) Estimation errors of Res-DCNN.
Sensors 25 03142 g005
Figure 6. Power spectrum comparison of different DOA estimation methods.
Figure 6. Power spectrum comparison of different DOA estimation methods.
Sensors 25 03142 g006
Figure 7. Confusion matrix under unknown source number conditions.
Figure 7. Confusion matrix under unknown source number conditions.
Sensors 25 03142 g007
Figure 8. RMSE curves of different SNR for various methods.
Figure 8. RMSE curves of different SNR for various methods.
Sensors 25 03142 g008
Figure 9. RMSE curves of different number of snapshots for various methods.
Figure 9. RMSE curves of different number of snapshots for various methods.
Sensors 25 03142 g009
Figure 10. Resolution probability curves of different SNR for various methods.
Figure 10. Resolution probability curves of different SNR for various methods.
Sensors 25 03142 g010
Figure 11. Resolution probability curves of different number of snapshots for various methods.
Figure 11. Resolution probability curves of different number of snapshots for various methods.
Sensors 25 03142 g011
Figure 12. Execution time comparison for various methods.
Figure 12. Execution time comparison for various methods.
Sensors 25 03142 g012
Table 1. Time consumption comparison across different DOA estimation methods.
Table 1. Time consumption comparison across different DOA estimation methods.
Time(s)ProposedRes-DCNNDCNNL1-SVDMUSICMVDR
Training690.0886667.3496540.3342///
Testing0.58060.04801.03332837.47356.6572466.0343
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, Y.; Fan, X.; Liu, J. Robust DOA Estimation via a Deep Learning Framework with Joint Spatial–Temporal Information Fusion. Sensors 2025, 25, 3142. https://doi.org/10.3390/s25103142

AMA Style

Zhao Y, Fan X, Liu J. Robust DOA Estimation via a Deep Learning Framework with Joint Spatial–Temporal Information Fusion. Sensors. 2025; 25(10):3142. https://doi.org/10.3390/s25103142

Chicago/Turabian Style

Zhao, Yonghong, Xiumei Fan, and Jisong Liu. 2025. "Robust DOA Estimation via a Deep Learning Framework with Joint Spatial–Temporal Information Fusion" Sensors 25, no. 10: 3142. https://doi.org/10.3390/s25103142

APA Style

Zhao, Y., Fan, X., & Liu, J. (2025). Robust DOA Estimation via a Deep Learning Framework with Joint Spatial–Temporal Information Fusion. Sensors, 25(10), 3142. https://doi.org/10.3390/s25103142

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop