Next Article in Journal
Robot Programming from a Single Demonstration for High Precision Industrial Insertion
Previous Article in Journal
High-Sensitivity CMOS-Integrated Floating Gate-Based UVC Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Centroid Optimization of DNN Classification in DOA Estimation for UAV

1
School of Computer Science and Technology, Zhejiang Sci-Tech University, Hangzhou 310018, China
2
School of Information Science and Technology, Zhejiang Sci-Tech University, Hangzhou 310018, China
3
Keyi College of Zhejiang Sci-Tech University, Shaoxing 312369, China
4
Institute of Optical Target Simulation and Test Technology, Harbin Institute of Technology, Harbin 150001, China
*
Authors to whom correspondence should be addressed.
Sensors 2023, 23(5), 2513; https://doi.org/10.3390/s23052513
Submission received: 9 January 2023 / Revised: 16 February 2023 / Accepted: 21 February 2023 / Published: 24 February 2023
(This article belongs to the Section Sensor Networks)

Abstract

:
Classifications based on deep learning have been widely applied in the estimation of the direction of arrival (DOA) of signal. Due to the limited number of classes, the classification of DOA cannot satisfy the required prediction accuracy of signals from random azimuth in real applications. This paper presents a Centroid Optimization of deep neural network classification (CO-DNNC) to improve the estimation accuracy of DOA. CO-DNNC includes signal preprocessing, classification network, and Centroid Optimization. The DNN classification network adopts a convolutional neural network, including convolutional layers and fully connected layers. The Centroid Optimization takes the classified labels as the coordinates and calculates the azimuth of received signal according to the probabilities of the Softmax output. The experimental results show that CO-DNNC is capable of acquiring precise and accurate estimation of DOA, especially in the cases of low SNRs. In addition, CO-DNNC requires lower numbers of classes under the same condition of prediction accuracy and SNR, which reduces the complexity of the DNN network and saves training and processing time.

1. Introduction

In recent years, unmanned aerial vehicles (UAV) have gradually appeared on the frontline of disaster rescue [1]. Although video cameras have been used as embedded tools for UAVs to solve such problems [2], the auxiliary role of audio is particularly important in environments with poor visual conditions [3]. Therefore, the direction of arrival (DOA) estimation of signal sources has become one of the important research topics [4].
Many methods have been proposed for DOA estimation using array signal detection, including the multiple signal classification (MUSIC) [5] by the principle of Eigen decomposition and the estimation of Signal Parameters via Rotational Invariance Techniques (ESPRIT) [6]. Nonlinear algorithms, such as deep learning (DL), are also applied to DOA estimation [7]. Compared with DL methods, conventional DOA estimation algorithms are usually limited by many factors [8]. For example, the received coherent signal [9] causes the permeation of the signal subspace [10] and noise subspace, which lowers the accuracy of conventional DOA estimation algorithms. Ge [8] provided a systematic review of research on DOA estimation by deep neural network methods. In his research, the advantages of DL methods include detection flexibility, accurate detection, increased number of resolvable targets by a limited number of detector elements. Goodman et al. [11] found that deep learning was more robust to significant calibration errors by comparing constrained integer optimization and deep learning. Nevertheless, DL methods convert DOA estimation into pattern recognition [12]. The extracted data features help the distinction between signal and noise and improve the accuracy of DOA estimation.
In 2020, a convolutional recurrent neural network (CRNN) for DOA estimation was proposed by Küçüket et al., and implemented in Android smartphones for real-time hearing aid [13]. Xiang et al. proposed a novel feature-to-feature phase enhancement approach to reduce the phase error of received data and improve the DOA estimation accuracy [14]. In 2021, a reverberation-aware network deep-learning training method was proposed by Liu et al. The proposed method integrates the reverberation information with the input network features to alleviate the dependence of DNN on the match of environment data sample [15]. Li et al. discussed the DOA estimation based on the ResNet model for underwater acoustic arrays to extract covariance matrices in different directions [16]. Huang et al. experimentally demonstrated a machine-learning-enabled DOA estimation that relaxes the heavy reliance on complicated antenna arrays and high-cost computers [17].
Deep Neural Network Classification (DNNC) is one of the popular approaches in Deep Learning classification for DOA estimation. However, the classification approaches typically necessitate bulky-sized angular classification techniques, making them ineffective for accurate detection. This paper proposes Centroid Optimization based on Deep Neural Network Classification (CO-DNNC). The CO-DNNC algorithm consists of signal preprocessing, classification network, and Centroid Optimization. Due to the varied frequency by different people and words, the phase differences are not convenient for detection. The signal intensity is considered the input of the network. Simulation and experiments prove that CO-DNNC has high precision and accuracy. Compared with DNNC, the precision can be improved by one to three times. Especially in the case of low signal-to-noise ratio (SNR), the measurement error of CO-DNNC is only 1/2~1/4 of the measurement error of DNNC. The accuracy of CO-DNNC is two to three times the accuracy of DNNC with the same error tolerance under the condition of the same SNR and number of classes. The Centroid Optimization is flexible in the varied implementation of classification networks and helpful to performance improvement.

2. Theory

2.1. Signal Model

This paper is considering the rescue mission of UAVs, focusing on the detection of human voices in close ranges based on passive probing. The detectors are placed under the rotors of UAVs. According to the installation, the layout of the detectors is designed as a circular array to achieve 360° signal detection [18]. The circular array is centered on the mass center of the UAV. Assume that the number of array elements is N and the array is uniformly distributed [19]. The azimuth of the direction of the incident signal is θ, where θ ∈ [0 2π). The received signal at the nth antenna element can be given as
r n = c n s + n o i s e n
Here, cn is the nth antenna response vector, and n ∈ {1…N}. s is the incident signal. noisen is the noise signal. The noise includes the ambient noise and inherent noise of the detector. The ambient noise received by each detector should be almost the same. However, the inherent noise of the detectors is different and independent of each other. It can be represented by the Gaussian white noise.
In the detection of human voices, the probing is taken passively. The antenna does not transmit known signals, which leads to the troublesome design for matched filters. Meanwhile, different languages, words, and speakers result in varied frequencies of voices, which make it hard to measure the phase differences of detected signals. Thereby, the power of voices is measured for azimuth detection instead of phase differences of signals. The power received by the nth array element can be expressed as
P n = P n + P n o i s e
Pnoise represents the noise power. P′n is the received power and can be further denoted as
P n = P t A r 4 π R n 2
Pt is the average power of the incident signal over a short period of time. Ar is the effective area of the receiving antenna. Rn represents the distance between the signal source and each detector. Because it is a short-distance detection, the distance differences from the signal source to different array elements cannot be ignored.

2.2. DNN Classification (DNNC) and Centroid Optimization of DNN Classification (CO-DNNC)

In Figure 1, DOA estimation based on Centroid Optimization of DNN Classification (CO-DNNC) algorithm consists of three parts: Signal preprocessing, classification network (DNNC), and Centroid Optimization.
In preprocessing, an automatic gain setting algorithm is implemented to prevent the signal intensity from being too low or too high, lowering the difficulty of network feature extraction. However, due to different ranges and intensities of incident voices, the power of the signals should be at the same level before being fed to the network. The received signal power by all detectors is also normalized.
x n = P n i = 1 N P i
xn is the ratio of the received signal power by the nth detector to the total received signal power by all array elements.
The DNNC network consists of one-dimensional convolutional layers (Conv1D) and fully connected layers (FCL). The preprocessed input data gets feature extraction by the Conv1Ds. The extracted features are reshaped from 2D to 1D and fed to FCL. In this paper, Relu is used as the activation function, and specify the binary cross-entropy as the loss function. We can train the whole network by minimizing the loss function iteratively. In the end, the probability of each class is output through Softmax. In conventional classification based on Deep Learning, the output is the probability of the existence of a signal in each direction, which is the flag to determine the predicted direction according to the highest probability. In CO-DNNC, the probabilities are fed to the Centroid Optimization.
Based on the probabilities of the existence of the signal in each direction, the Centroid Optimization of CO-DNNC estimates the azimuth of the received signal. The coordinates rm are the classified angles. The weight is the probability of each class αm [20]. The azimuth of received signal can be calculated as
y = m = 1 M r m α m

3. Experiments

3.1. Datasets

Assume that the detectors are omnidirectional antennas and located under the UAV rotor. The layout of the detectors is a circular array in uniform distribution, with the mass center of the UAV as the center of the detector circle. The distance between each detector and the center is 0.5 m. Assume that the signal source is in short distance from the UAV. The distance between the source and the center of the detectors is 10.5 m. The power of received signals is calculated according to Equation (3). Since the power of the speech signal cannot maintain consistency, the ratio of signal to noise (SNR) is employed as the ratio of the mean of signal to the mean of noise, shown as Equation (6). The detected power of the signal will be the addition of the power of the received signal and noise, shown as Equation (7).
S N R = m e a n ( P ˜ ) m e a n ( P ˜ n o i s e )
P ˜ n = P ˜ + S N R P ˜ n o i s e
where P ˜ is the simulated received signal and P ˜ n o i s e is the simulated received noise.
In the following experiments, it is assumed that the UAV has 4 detectors, shown as blue dots in Figure 2. The classification network divides the detection area 0~360° into 8 classes. The interval of each class includes 45°, shown as the shaded area in Figure 2. Any angle generated in the interval is considered to be the same class. The values of the signal and noise employed in this paper are from the IEEESPCup2019, DREGON dataset [21,22]. The clean-recordings-speech is used as the signal source, and the clean-recordings-white noise is used as the noise, including the ambient noise and inherent noise. Each pair of values corresponds to one randomly selected azimuth of incident signal. The detected powers of signal and noise are calculated by Equations (3) and (7) for the 4 detectors. The SNR is set to 10. Eight classes are considered. The azimuth of the signal source is randomly selected. A total 1600 random azimuths of the dataset are generated in each class. Eight hundred of the dataset are taken as the training set. The remaining 800 azimuths are used as the testing set. Therefore, the sizes of the training and testing sets are both 6400 × 4.

3.2. Performance Metrics

Accuracy (ACC) and root mean square error (RMSE) are utilized for quantifying the performance of the DOA estimation.
The root mean square error (RMSE) [23] is defined as
R M S E = 1 K k = 0 K 1 ( θ k - θ ^ k ) 2
where θ k is the true azimuth and θ ^ k is the predicted azimuth. K represents the total number of signal samples.
ACC is defined as the ratio of the number of correct estimations to the number of total samples [24]. The correct estimation is defined as the estimation with an estimation error less than the tolerance. The default tolerance is half the interval between classes.
A C C = 1 K i = 0 K 1 ε ( λ θ k )
which ε ( θ ) can be expanded to
ε ( θ ) = 0 θ < 0 1 θ 0

3.3. Experiment Validation

Table 1 shows the confusion matrix of the classification network, which is DNNC here. As the proposed classification network is very simple, the accuracy of the classification is not ideal. However, the accuracy can still be above 75%.
On the other hand, the distribution of the errors is almost symmetric at the true azimuth of DOA. Meanwhile, it is less likely to generate larger errors. For example, in the case that the true azimuth of the received signal is 0°, the accuracy of classification reaches 78.5%. The probabilities of generating errors of 45° and −45° are 11.25% and 12.125%. The probabilities of generating errors of 90° and −90° are 0.5% and 0.125%. In order to demonstrate the tendency clearly, the interval of the true azimuth of DOA is set to 10°, where the true azimuths of DOA are 0°, 10°, 20°… 90°. The experiment still employs 4 detectors. The number of classes is still 8. The SNR of the testing data is set to 20 dB. The testing dataset for each true azimuth contains 800 samples. The probabilities of the classification network output are shown in Figure 2.
In Figure 3, the horizontal coordinates are the true azimuths of DOA 0°, 10°, 20°… 90°. The yellow solid yellow line, blue dashed line, and green dotted line represent the probabilities of the classification network outputs on class 0°, 45°, and 90°, respectively.
In the case of the true azimuth of 0°, 87.25% of the outputs of DNNC give 0°, 4.75% of the outputs give 45°, and 0% of the outputs give 90°. The accuracy of classification is 87.25%. It seems acceptable. However, in the case of the true azimuth of 70°, 36.25% of the outputs are classified as 45°, shown in the blue dashed line, and 61.5% of the outputs are classified as 90°, shown in the green dotted line. The accuracy of DNNC is only 61.5%. The probability of wrong classification is significant, which leads to relatively great errors. The reason lies in the determination mechanism of the classification network. The classification network heavily depends on the maximum probability of the outputs without considering the gaps between the probabilities of varied classes. The data of the outputs have not been fully utilized. In Figure 3, all three lines show a symmetric shape. The probability of the output always achieves the highest value if the difference between the true azimuth and predicted azimuth is 0°. As the difference between the true azimuth and predicted azimuth increases, the probability decreases. The blue dashed line of class 45° presents the tendency completely. When the difference between the true azimuth and predicted azimuth is small, the classification network is capable of extracting distinctive features of data. As the difference grows, the data of the true azimuth possess partial features of the neighboring class. The boundary of classification turns indistinct.
Due to the symmetric shape, Centroid Optimization (CO-DNNC) is presented. In Figure 3, the red dots are the outputs of one randomly chosen sample for each true azimuth by CO-DNNC. The horizontal coordinates still represent the true azimuth of DOA. The vertical coordinates based on the right y-axis represent the predicted azimuth of DOA. The distribution of the red points seems approximately a straight line. The predicted azimuth of the true azimuth of 70° equals 71.6° by the Centroid Optimization. Compared with DNNC, CO-DNNC apparently has fewer errors.
Figure 4 shows the error histograms of CO-DNNC and DNNC for all eight classes. Each class has 800 samples with random azimuths in the section. The horizontal coordinate represents the error between the true azimuth and the predicted azimuth. The vertical coordinate represents the counts of errors. The red lines represent the error histogram of CO-DNNC, and the blue curve is the one of DNNC.
The errors caused by DNNC are mostly concentrated between −20° and +20°. The errors caused by CO-DNNC are more concentrated in the range of ±10°. The RMSE of the error distribution of DNNC is 15.16° while the one of CO-DNNC is 10.20°. Therefore, the proposed CO-DNNC reduces the errors and improves the detection accuracy.

4. Discussion

The analysis above has shown that the mechanism of the classification network cannot either extract distinctive features or fully utilize the features, even in a situation of high SNR. When the SNR decreases, more noise is introduced, and the features of the signal are concealed, which makes it more difficult to extract the features by the classification network. Meanwhile, the interval between the classes determines the variation degrees. It will be helpful to decrease the interval and increase the number of classes to extract accurate feathers of data.
This paper investigates the impact of the SNR of signals and the number of classes on the accuracy of DOA estimation. In Table 2 and Table 3, the SNRs of the input signal are set to −20 dB, −10 dB, 0 dB, 10 dB, and 20 dB, respectively. The numbers of classes are set to 4, 8, 12, and 18, respectively. The corresponding intervals are 90°, 45°, 30°, and 20°, respectively. One thousand six hundred samples from random azimuths in each interval are taken as the training and testing datasets. Each of them has 800 samples.

4.1. Impact of Number of Classes and SNR

Table 2 lists the accuracies of DNNC in the cases of varied SNR and the numbers of classes. The RMSEs of the predicted DOA of CO-DNNC and DNNC are compared. In Table 2, two RMSEs are listed for each pair of SNR and the number of classes. The first number is the RMSE of DNNC and the second bold number is the RMSE of CO-DNNC.
In Table 2, the accuracy of the classification network decreases along with the increase in the number of classes for the same SNR. As more noise is introduced, the waveform of the signal is interfered with, which results in inaccurate feature extraction by the network. At the meantime, the RMSEs of CO-DNNC and DNNC increase. Taking four classes as an example where the SNR changes from −20 to 20 dB, the network accuracy decreases from 95.6% to 36.4%. The RMSE of DNNC increases from 26.9° to 95.1°, while the RMSE of CO-DNNC increases from 19.9° to 50.9°.
In Table 2, the accuracy of the classification network decreases along with the decrease of SNR for the same number of classes. As more classes are introduced, the interval between the classes shrinks, which results in the increase of the feature similarity and interference between the neighboring classes. However, it also results in the RMSEs of CO-DNNC and DNNC decrease in high SNR. For example, at 20 dB of SNR, the network accuracy can reach 81.8~95.6%, and the RMSEs of CO-DNNC and DNNC are reduced from 19.9° and 26.9° to 6.6° and 9.0°, respectively. In this case, the interference between the neighboring classes is not severe enough. The classification network can distinctively extract the data features. The shrink of the interval leads to a decrease in the RMSEs. However, when the detection situation gets worse and the SNR turns low, the interference of noise establishes supremacy over the signal. The classification accuracy is greatly reduced to 8.4~36.4%. The RMSEs depend on the intensity of noise and vary little. The RMSE of DNNC does not change much at −20 dB, along with the increase in the number of classes. However, CO-DNNC can still reduce the RMSE from 50.9° to 24.6°. All the RMSEs of CO-DNNC are lower than the RMSEs of DNNC. The differences in RMSEs between the two methods are even larger at low SNR. CO-DNNC can fully utilize the extracted features by taking the calculated probabilities as weights to estimate the DOA and reduce the error. It also shows that CO-DNNC can alleviate the interference of noise on detection. This has demonstrated that CO-DNNC is very suitable for target detection with low SNR.
In another way, in Table 2, the error of CO-DNNC with eight classes at 20 dB (10.2°) is similar to the error of DNNC with 12 classes at 20 dB (10.6°). The classification networks by deep learning algorithms are normally capable of increasing the number of classes to reduce the detection error. However, the increased classes raise the complexity of the network. CO-DNNC is capable of lowering the complexity with the satisfaction of required errors at certain SNR.

4.2. Impact of Tolerance

Since CO-DNNC can overcome the low SNR and reduce error, Table 2 shows the measurement accuracy in the cases of a varied number of classes and SNR. According to Equation (9), the accuracy depends on the tolerance λ. In this paper, three tolerances are set, which are λ = 50%φ, λ = 22.5%φ, and the fixed tolerance λ = 10°. φ is the interval between classes. The tolerances in λ = 50%φ and λ = 22.5%φ vary with the number of classes. λ = 50%φ represents that the tolerance is the entire interval as default tolerance, the same as the setting in the above analysis. λ = 22.5%φ represents the tolerance takes only half the interval. In practical applications, users will wisely set the parameters according to the accuracy requirements. Therefore, a fixed tolerance λ = 10° is discussed. The accuracies of DNNC and CO-DNNC (bold font) are compared in Table 3.
In Table 3, the accuracies of both CO-DNNC and DNNC are improved with the same number of classes when SNR is improved. The accuracies of DNNC and CO-DNNC increase from 18.5% and 51.4% to 88.8% and 96.5% in eight classes, respectively, when SNR increases from −20 to 20 dB and a tolerance is 50%φ. When the tolerance is fixed to 10°, the accuracies of DNNC and CO-DNNC increase from 8.4% and 24.4% to 42.4% and 70.8%, respectively.
In the cases of the same SNR, when the tolerances are set to 50%φ and 22.5%φ, the accuracies for both algorithms decrease with the increase of the number of classes. The reason is that the strategy of tolerance is set to varied tolerance according to the number of classes. The requirements for accuracy are more stringent when the interval between classes shrinks. Taking 0 dB as an example. The accuracies of DNNC and CO-DNNC drop from 76.2% and 88.7% to 24.7% and 57.5%, respectively, with a tolerance of 50%φ. The accuracies of CO-DNNC and DNNC drop from 41.8% and 66.2% to 12.5% and 29.3%, respectively, with a tolerance of 22.5%φ. However, the accuracies of DNNC and CO-DNNC improve with the increase of the number of classes when the tolerance is fixed at 10°. Taking 0 dB as an example, the accuracies of DNNC and CO-DNNC increase from 18.7% and 36.2% to 24.7% and 57.5%, respectively.
Overall, the accuracy of CO-DNNC is higher than that of DNNC. In the case of low SNR, the accuracy of CO-DNNC is significantly higher than that of DNNC. Under the circumstance of −20 dB of SNR and eight classes, the accuracy of DNNC is only 18.5% with a tolerance of 50%φ, while the accuracy of CO-DNNC reaches 51.4%. Therefore, CO-DNNC is more suitable for DOA estimation, especially under the condition of low SNR.

5. Conclusions

In order to improve the accuracy of DOA estimation based on classification by deep learning in random azimuths, this paper proposes a Centroid Optimization based on deep neural network classification (CO-DNNC). The received signal is calibrated in the preprocessing stage and classified by the DNN network. The output probabilities of the network are utilized to estimate the incident azimuth by the Centroid Optimization. The accuracy and precision are evaluated according to ACC and RMSE. The results show that the accuracy and precision of CO-DNNC and DNNC are both improved with the increase of SNR and the number of classes. The predicted azimuths of CO-DNNC have fewer errors than those of DNNC, especially at the condition of low SNR. Overall, the CO-DNNC measurement error is 26% to 74% of the DNNC measurement errors. The measurement accuracy of CO-DNNC is 1~3 times the accuracy of DNNC. When the SNR is low, the CO-DNNC measurement error is only 26~54% of the DNNC measurement errors. When the noise tolerance is limited at 50% of intervals, the CO-DNNC measurement accuracy is 20~33% higher than the DNNC measurement accuracy. In addition, CO-DNNC achieves the same prediction accuracy on the condition of fewer classes, which reduces the complexity of the network and the training time. The method proposed in this paper is suitable for short-distance detection. In the future, we will try to improve the performance of the array element or add phase features to get more accurate results and apply this scheme to three-dimensional space.

Author Contributions

Conceptualization, L.W.; Data curation, L.W.; Formal analysis, L.W., Z.Z., X.Y., L.X. and S.C.; Investigation, L.X. and Y.Z.; Methodology, L.W. and Z.Z.; Project administration, L.W. and L.X.; Resources, Y.Z.; Software, Z.Z. and X.Y.; Supervision, J.Z.; Validation, Z.Z. and X.Y.; Visualization, Jianlong Zhang; Writing—original draft, Z.Z.; Writing—review & editing, L.W. and S.C.; Funding acquisition, L.W., X.Y. and L.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 61801429, the Natural Science Foundation of Zhejiang Province under Grant LY20F010001 and LQ20F050010. This work was also supported by the Fundamental Research Funds of Zhejiang Sci-Tech University under Grant 2021Q030.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are openly available at https://github.com/yu-zhongyang/CO-DNNC.git.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Andrea, R.D. Guest editorial can drones deliver? IEEE Trans. Autom. Sci. Eng. 2014, 11, 647–648. [Google Scholar] [CrossRef]
  2. Lopez-Fuentes, L.; van de Weijer, J.; Gonzalez-Hidalgo, M.; Skinnemoen, H.; Bagdanov, A.D. Review on computer vision techniques in emergency situations. Multimed. Tools Appl. 2018, 77, 17069–17107. [Google Scholar] [CrossRef] [Green Version]
  3. Basiri, M.; Schill, F.; Lima, P.U.; Floreano, D. Robust acoustic source localization of emergency signals from micro air vehicles. In Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, Algarve, Portugal, 7 October 2012. [Google Scholar]
  4. Zhang, R.; Shim, B.; Wu, W. Direction-of-arrival estimation for large antenna arrays with hybrid analog and digital architectures. IEEE Trans. Signal Process. 2021, 70, 72–88. [Google Scholar] [CrossRef]
  5. Schmidt, R. Multiple emitter location and signal parameter estimation. IEEE Trans. Antennas Propag. 1986, 34, 276–280. [Google Scholar] [CrossRef] [Green Version]
  6. Roy, R.; Kailath, T. ESPRIT-estimation of signal parameters via rotational invariance techniques. IEEE Trans. Acoust. Speech Signal Process. 1989, 37, 984–995. [Google Scholar] [CrossRef] [Green Version]
  7. Ping, Z. DOA estimation method based on neural network. In Proceedings of the 2015 10th International Conference on P2P, Parallel, Grid, Cloud and Internet Computing (3PGCIC), Krakow, Poland, 4 November 2015. [Google Scholar]
  8. Ge, S.; Li, K.; Rum, S. Deep learning approach in DOA estimation: A systematic literature review. Mob. Inf. Syst. 2021, 2021, 1–14. [Google Scholar] [CrossRef]
  9. Zhang, X.; Yu, J.; Feng, G.; Xu, D. Blind direction of arrival estimation of coherent sources using multi-invariance property. Prog. Electromagn. Res. 2008, 88, 181–195. [Google Scholar] [CrossRef] [Green Version]
  10. Wang, H.; Kaveh, M. Coherent signal-subspace processing for the detection and estimation of angles of arrival of multiple wide-band sources. IEEE Trans. Acoust. Speech Signal Process. 1985, 33, 823–831. [Google Scholar] [CrossRef] [Green Version]
  11. Goodman, J.; Salmond, D.; Davis, C.; Acosta, C. Ambiguity resolution in direction of arrival estimation using mixed integer optimiztion and deep learning. In Proceedings of the IEEE National Aerospace and Electronics Conference, Dayton, Ohio, 15 July 2019. [Google Scholar]
  12. Yu, L.; Li, Z.; Chu, N.; Mohammad-Djafari, A.; Guo, Q.; Wang, R. Achieving the sparse acoustical holography via the sparse bayesian learning. Appl. Acoust. 2022, 191, 1–14. [Google Scholar] [CrossRef]
  13. Küçük, A.; Panahi, I.M.S. Convolutional recurrent neural network based direction of arrival estimation method using two microphones for hearing studies. In Proceedings of the IEEE 30th International Workshop on Machine Learning for Signal Processing (MLSP), Virtual Conference, 19 April 2020. [Google Scholar]
  14. Xiang, H.; Chen, B.; Yang, T.; Liu, D. Improved de-multipath neural network models with self-paced feature-to-feature learning for DOA estimation in multipath environment. IEEE Trans. Veh. Technol. 2020, 69, 5068–5078. [Google Scholar] [CrossRef]
  15. Liu, Y.; Tong, F.; Zhong, S.; Hong, Q.; Li, L. Reverberation aware deep learning for environment tolerant microphone array DOA estimation. Appl. Acoust. 2021, 184, 1–8. [Google Scholar] [CrossRef]
  16. Li, P.; Tian, Y. DOA estimation of underwater acoustic signals based on deep learning. In Proceedings of the 2021 2nd International Seminar on Artificial Intelligence. Networking and Information Technology (AINIT), Shanghai, China, 17 October 2021. [Google Scholar]
  17. Huang, M.; Zheng, B.; Cai, T.; Li, X.; Liu, J.; Qian, C.; Chen, H. Machine–learning-enabled metasurface for direction of arrival estimation. Nanophotonics 2022, 11, 2001–2010. [Google Scholar] [CrossRef]
  18. Qayyum, A.; Hassan, K.; Anika, A.; Shadiq, F.; Rahman, M.; Islam, T.; Imran, S.A.; Hossain, S.; Haque, M.A. DOANet: A deep dilated convolutional neural network approach for search and rescue with drone-embedded sound source localization. EURASIP J. Audio Speech Music Process. 2020, 2020, 1–18. [Google Scholar] [CrossRef]
  19. Yamano, R.; Kikuma, N.; Sakakibara, K. Effect of redundancy of element placement on DOA estimation with circular array. In Proceedings of the 2016 International Symposium on Antennas and Propagation (ISAP), Guilin, China, 18 October 2016. [Google Scholar]
  20. Ji, X.; Wang, L.; Sun, W.; Wang, H. A k-weighted centroid algorithm based on proximity beacon node optimization. In Proceedings of the 2017 13th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery (ICNC-FSKD), Guilin, China, 29 July 2017. [Google Scholar]
  21. IEEE Signal Processing Cup 2019. Available online: http://dregon.inria.fr/datasets/signal-processing-cup-2019/ (accessed on 22 October 2020).
  22. Strauss, M.; Mordel, P.; Miguet, V.; Deleforge, A. DREGON: Dataset and methods for UAV-embedded sound source localization. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1 October 2018. [Google Scholar]
  23. Xiang, H.; Chen, B.; Yang, M.; Xu, S. Angle separation learning for coherent DOA estimation with deep sparse prior. IEEE Commun. Lett. 2021, 25, 465–469. [Google Scholar] [CrossRef]
  24. Cheng, L.; Sun, X.; Yao, D.; Li, J.; Yan, Y. Estimation reliability function assisted sound source localization with enhanced steering vector phase difference. IEEE/ACM Trans. Audio Speech Lang. Process. 2021, 29, 421–435. [Google Scholar] [CrossRef]
Figure 1. Structure of Centroid Optimization of Deep Neural Network Classification (CO-DNNC).
Figure 1. Structure of Centroid Optimization of Deep Neural Network Classification (CO-DNNC).
Sensors 23 02513 g001
Figure 2. Uniform circular array of UAV and incident wave.
Figure 2. Uniform circular array of UAV and incident wave.
Sensors 23 02513 g002
Figure 3. Probabilities of Softmax in DNNC and predicted azimuth by CO-DNNC.
Figure 3. Probabilities of Softmax in DNNC and predicted azimuth by CO-DNNC.
Sensors 23 02513 g003
Figure 4. Error histogram of CO-DNNC and DNNC.
Figure 4. Error histogram of CO-DNNC and DNNC.
Sensors 23 02513 g004
Table 1. Confusion matrix of DNNC classification.
Table 1. Confusion matrix of DNNC classification.
Predicted Azimuth of DOATrue Azimuth of DOA
45°90°135°180°225°270°315°
78.59.6250.250000.511.125
45°11.2579.3758.6250000.1250.125
90°0.12510.7578.510.5000.1250
135°00.12510.37578.62510.6250.2500
180°000.62511.87577.010.3750.1250
225°0000.2512.2576.12510.8750.5
270°0.50.125000.37511.12575.7512.125
315°12.1250.250000.258.7578.625
Table 2. Accuracies of DNNC and RMSEs of CO-DNNC and DNNC with varied SNRs and classes number.
Table 2. Accuracies of DNNC and RMSEs of CO-DNNC and DNNC with varied SNRs and classes number.
Performance MetricsSNR
−20 dB−10 dB0 dB10 dB20 dB
Accuracy (%)36.454.276.289.295.6
RMSE (°) of DNNCRMSE (°) of CO-DNNC95.150.970.344.642.328.930.119.526.919.9
Accuracy (%)18.533.058.777.888.8
RMSE (°) of DNNCRMSE (°) of CO-DNNC91.030.665.834.836.027.620.115.115.210.2
Accuracy (%)13.023.647.671.487.3
RMSE (°) of DNNCRMSE (°) of CO-DNNC87.825.464.236.234.428.216.713.810.67.2
Accuracy (%)8.415.935.962.481.8
RMSE (°) of DNNCRMSE (°) of CO-DNNC93.924.667.63835.630.216.214.79.06.6
Table 3. Accuracies (%) for CO-DNNC and DNNC with varied SNRs and numbers of classes.
Table 3. Accuracies (%) for CO-DNNC and DNNC with varied SNRs and numbers of classes.
Number of Classes λ SNR
−20 dB−10 dB0 dB10 dB20 dB
4
φ = 90 °
45°36.456.354.271.976.288.789.297.995.699.6
22.5°18.329.029.342.541.866.247.477.048.466.3
10°7.812.312.820.318.736.222.635.520.626.5
8
φ = 45 °
22.5°18.551.431.959.058.868.177.888.088.896.5
11.25°9.427.016.230.831.642.843.766.048.778.1
10°8.424.414.627.228.438.738.962.342.470.8
12
φ = 30 °
15°13.044.021.655.847.653.171.481.387.394.9
7.5°6.323.011.128.625.630.140.061.147.880.8
10°8.530.214.838.333.638.951.669.962.488.3
18
φ = 20 °
10°8.431.715.149.924.757.562.469.081.790.3
4.216.27.726.212.529.334.547.844.474.7
10°8.431.715.149.924.757.562.469.081.790.3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, L.; Zhang, Z.; Yang, X.; Xu, L.; Chen, S.; Zhang, Y.; Zhang, J. Centroid Optimization of DNN Classification in DOA Estimation for UAV. Sensors 2023, 23, 2513. https://doi.org/10.3390/s23052513

AMA Style

Wu L, Zhang Z, Yang X, Xu L, Chen S, Zhang Y, Zhang J. Centroid Optimization of DNN Classification in DOA Estimation for UAV. Sensors. 2023; 23(5):2513. https://doi.org/10.3390/s23052513

Chicago/Turabian Style

Wu, Long, Zidan Zhang, Xu Yang, Lu Xu, Shuyu Chen, Yong Zhang, and Jianlong Zhang. 2023. "Centroid Optimization of DNN Classification in DOA Estimation for UAV" Sensors 23, no. 5: 2513. https://doi.org/10.3390/s23052513

APA Style

Wu, L., Zhang, Z., Yang, X., Xu, L., Chen, S., Zhang, Y., & Zhang, J. (2023). Centroid Optimization of DNN Classification in DOA Estimation for UAV. Sensors, 23(5), 2513. https://doi.org/10.3390/s23052513

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop