This paper is an extended version of our paper published in Ruiz-Espitia, O.; Martinez-Carranza, J.; Rascon, C. AIRA-UAS: An Evaluation Corpus for Audio Processing in Unmanned Aerial System. In Proceedings of the 2018 International Conference on Unmanned Aircraft Systems (ICUAS), Amsterdam, The Netherlands, 10–11 May 2018; pp. 836–845.
Received: 27 July 2019 / Accepted: 6 September 2019 / Published: 10 September 2019
Audio analysis over an Unmanned Aerial Systems (UAS) is of interest it is an essential step for on-board sound source localization and separation. This could be useful for search & rescue operations, as well as for detection of unauthorized drone operations. In this paper, an analysis of the previously introduced Acoustic Interactions for Robot Audition (AIRA)-UAS corpus is presented, which is a set of recordings produced by the ego-noise of a drone performing different aerial maneuvers and by other drones flying nearby. It was found that the recordings have a very low Signal-to-Noise Ratio (SNR), that the noise is dynamic depending of the drone’s movements, and that their noise signatures are highly correlated. Three popular filtering techniques were evaluated in this work in terms of noise reduction and signature extraction, which are: Berouti’s Non-Linear Noise Subtraction, Adaptive Quantile Based Noise Estimation, and Improved Minima Controlled Recursive Averaging. Although there was moderate success in noise reduction, no filter was able to keep intact the signature of the drone flying in parallel. These results are evidence of the challenge in audio processing over drones, implying that this is a field prime for further research.
corpus evaluation; AIRA-UAS; AQBNE; IMCRA; unmanned aerial systems
Current advances in technology have greatly improved different sectors of the industry and society. One such sector has been the area of aerial vehicles piloted by radio control. Very recently, Unmanned Aerial Systems (UAS) were used exclusively for military purposes, first in the shape of airplanes and helicopters, and later in the shape of multi-rotors. Nowadays, UAS have been implemented in several sectors of society, such as agriculture, surveillance, transportation, environment monitoring, topography, and recreation [1,2].
In many of these developments, a plethora of sensor modalities have been used for different application cases, such as vision [3,4], infrared , and even audio [6,7]. In the case of audio, an important application case is that of detecting unauthorized drone operations. For instance, the scenario where UAS should be removed because they are flown in restricted zones, such as exclusive aerial spaces for commercial airplanes and privately-owned areas. Tracking them can be carried out by vision, but it could be greatly improved by the use of sound source localization algorithms, such as the one presented in , since audio can be captured omni-directionally.
In addition, one of the major characteristics of an UAS is the sound its rotors produce while flying, which could be used for its detection and tracking in such restricted zones [8,9]. Another scenario of interest is that of carrying out audio signal processing over a drone could be useful in rescue situations [10,11]. However, the high amount of energy of the ego-noise of the drone while flying could be considered a challenge  but this has not been studied extensively. This is the main objective of the work presented here: to evaluate if current popular techniques are able to properly filter out the ego-noise of the drone’s motors. To do this it is important to first analyze the characteristics of such a noise and choose filtering techniques appropriately. Having chosen these, a proper evaluation needs to be carried out to measure their performance.
To this effect, it is indispensable to use adequate evaluation resources and methodologies. We previously introduced the Acoustic Interactions for Robot Audition (AIRA)-UAS corpus in , and we believe that it is perfectly compatible with these objectives. This corpus consists of audio recordings that were captured over a flying drone. Different recordings settings were used, such as the drone being static at different heights, vertical, and horizontal movements, as well as the presence of other drones. This corpus is named under the umbrella of another corpus made for evaluating audio signal processing algorithms called AIRA . AIRA is aimed for evaluating sound source localization and separation algorithms but in the context of service robotics. The AIRA-UAS corpus is an extension of AIRA in the realm of unmanned aerial systems (UAS), and therefore named as such. The full set of recordings can be obtained from .
As it will be seen, there are recordings in AIRA-UAS where another drone is flying in parallel to the capture drone, and they present a very low Signal-to-Noise Ratio (SNR), which is to be expected. Thus, extracting the audio signature of another flying drone is a challenge. Additionally, it will also be seen that the noise spectral signature changes depending if the drone is static or moving. However, in AIRA-UAS the spectral signature of the capture drone is also provided, which may prove useful for its cancellation.
As for which techniques to evaluate, it is important to mention that there is a vast a mount of filtering and noise cancellation techniques available . We decided upon Berouti’s Non-Linear Noise Subtraction  (NLNS), Adaptive Quantile Based Noise Estimation  (AQBNE), and Improved Minima Controlled Recursive Averaging  (IMCRA). We chose these because of their popularity, ease of implementation and, more importantly, the amount of required computing resources is such that they can run out on the drone’s on-board computer. It is important to mention that the adaptive nature of AQBNE and IMCRA is appropriate for this application, since the ego-noise of the drone is dynamic. And, while the NLNS is not adaptive, it can use the drone’s spectral signature as the noise estimation to cancel out.
This paper is organized as follows: Section 2 describes the relevant parts of the AIRA-UAS corpus for completeness sake; Section 3 provides an analysis of the audio recordings in AIRA-UAS from the time-frequency domain and describe the challenges put forward by carrying out audio processing over a flying drone; Section 4 details the filtering techniques to be evaluated; Section 5 describes the evaluation methodology based on the AIRA-UAS recordings, as well as presents the results obtained from such evaluation; and Section 6 offers the work’s conclusions and future work, as well as some ideas for the UAS community.
2. The AIRA-UAS Corpus
The AIRA-UAS corpus was previously introduced in , and its relevant parts are described here for completeness sake. It was captured with the intent of being used for evaluating audio processing techniques carried over an UAS. Techniques, such as sound source localization and separation, are of interest, but all of them require a pre-processing stage that involves the removal of the drone’s motors ego-noise. To this effect, the AIRA-UAS was designed to provide a varied selection of flying patterns and accompanying flying drones. As mentioned before, it is freely available to download at .
2.1.1. Unmanned Aircraft Vehicles
In this work, the microphones were mounted on an UAS, and another two drones were used to fly around that UAS and generate additional audio data.
DJI Matrice 100
This is the drone where the microphones were mounted. The camera and gimbal were removed to provide space and not impact its weight limitations. It is shown in Figure 1a and has the following specifications:
Diagonal wheelbase: 650 mm
Propeller model: DJI 1345 × 2 (13 in diameter and 4.5 in pitch, 2 blades)
Weight (with TB47D battery model): 2355 g
Max. Takeoff Weight: 3400 g
This drone is shown in Figure 1b, and was one of the two drones that flew around the DJI Matrice 100. It has the following specifications:
Diagonal wheelbase: 460 mm
Propeller model: 1045 × 2 (10 in diameter and 4.5 in pitch, 2 blades)
Parrot Bebop 2
This drone is shown in Figure 1c, and was the second of the two that flew around the DJI Matrice 100. It has the following specifications:
Dimensions: 38 × 33 × 9 cm
Propeller model: 5040 × 3 (5 in diameter and 4 in pitch, 3 blades)
The audio capture hardware is composed by a RaspberryPi 2 running Raspbian OS as the main processing unit, and the 8SoundsUSB audio interface  (shown in Figure 2a) with its eight microphones (show in Figure 2b). The reasoning behind using this hardware is because it is lightweight and requires a low amount of energy. At the same time, however, it is able to capture up to 8 channels simultaneously. This amount of microphones is very popular when carrying out multiple sound source localization . The complete system is shown in Figure 2c.
2.2. Recording Protocols
Three recording protocols were used, each with its own set of objectives. In this section a detailed explanation of each protocol is provided.
2.2.1. 1st Protocol
The intent for this set of recordings is to capture the background noise and the ego-noise of the DJI Matrice 100 doing several aerial maneuvers. Additionally, several recordings only differ between them in terms of the height that the UAS was flying at, with the intent of exploring if the ego-noise of the UAS is dependent of its height.
There were 15 recordings captured and their relevant details are presented in Table 1.
A brief summary of each recording is provided here:
Recording 1. A preliminary test to ascertain that all the systems were running correctly. The drone was not moving and motors were not activated.
Recording 2 and 3. Background noise recordings, so that spectral filtering could potentially be used.
Recording 4. Motors activated, but not taken off.
Recording 5, 8, and 11. Vertical displacements at different heights. These tests were only 15 s, since the displacements were carried out very quickly by the drone.
Recording 6, 9, 12, 14, and 15. Drone hovering at different heights, but not moving.
Recording 7, 10, and 13. Horizontal displacements at different heights.
2.2.2. 2nd Protocol
The intent for this set of recordings is to capture the DJI Matrice 100 and the 3DR Solo flying at the same time, as well as obtain the spectral signature of the 3DR Solo from afar. 3 recordings were captured:
Recording 16. The motors of the DJI Matrice 100 were not activated and the 3DR Solo was flying at a horizontal distance of 2 m and a height of 1 m.
Recording 17. The motors of the DJI Matrice 100 were not activated and the 3DR Solo was flying around it forming a complete circle with a radius of 2 m at a height of 1 m.
Recording 18. Both the DJI Matrice 100 and the 3DR Solo were flying in front of each other at a height of 5 m with a distance of 3 m between them.
2.2.3. 3rd Protocol
The intent for this set of recordings is very similar to the one of the 2nd protocol. The only difference is that the 3DR Solo was replaced by the Parrot Bebop 2. The same three recording settings of the 2nd protocol were carried out (Recordings 19, 20, and 21).
3. Analysis of AIRA-UAS Recordings
As stated before, it is important to first analyze the characteristics of the recordings in AIRA-UAS to adequately choose the filtering techniques that are going to be evaluated. To this effect, four different parts of the corpus are analyzed:
The environment in which the recordings were made.
The noise signatures of the drones captured when flying but not moving.
The difference in noise signatures when the drones are moving.
The recordings when two drones are flying in parallel.
These analyses are carried out mainly by visual inspection of the spectrogram in the time-frequency domain of the first channel of the recordings. For completeness, it is important to note that these spectrograms are calculated with a Hann window size of 1024 samples, with a overlap, and normalized by the window size. In some cases, for ease of visualization, the mean-energy-per-frequency of the spectrogram is also presented.
Additionally, other elements, such as the signal energy and SNR, are also used, as well as subjective hearing in some cases. Finally, it is important to mention that the data here shown are of the sections that are relevant to the analysis; the data shown in the website of the AIRA-UAS is only representative of the movements for each recording and may not exactly coincide with the data shown here.
3.1. Recording Environment
The recording used for this part was the Recording 2 of AIRA-UAS, when the Matrice is not active and only the environment was captured. A mean energy value of -61.657 dBFS was measured, which implies that there was a very low amount of environmental noise.
The spectrogram of this recording is shown in Figure 3a and a zoomed version is presented in Figure 3b. It is important to mention that the same color-map is used in all spectrograms.
As it can be seen, although there is not much energy in most of the frequency throughout the recordings, there is a small amount in the range below 5 kHz.
This implies that the system was not well grounded, as it appears that there is some electrical noise, which was confirmed by subjective hearing of the recording. However, as it will be seen in the following sections, this amount of noise is still quite low compared to the recording of the active drone motors.
3.2. Drone Signatures
The recordings used for this part were Recordings 4 (with active motors but not flying) and 6 (flying in a static position) for the Matrice, and Recordings 16 and 19 for the 3DR and Parrot, respectively.
In Figure 4a,b the spectrograms of Recordings 4 and 6 are shown.
As it can be seen, there is a difference of energy presence in the range between 10 and 15 kHz between these two recordings. Additionally, there is an important increase in energy in the range between 2.5 kHz and 7.5 kHz, which can be visualized in the mean-energy-per-frequency of Recordings 4 and 6 in Figure 5a,b.
The increase in energy from Recording 4 to 6 is expected, since the motors create more noise when flying than just laying on the ground.
It could be argued that Figure 5b shows the Matrice noise signature in the frequency domain. The noise signatures of the 3DR and the Parrot are shown in Figure 6a,b.
As it can be seen, the Matrice, the 3DR and the Parrot bare different noise signatures. This implies that the idea of recognizing a drone by its noise signature may have some merit. However, this will only be possible if the ego-noise of the Matrice is filtered out. Comparing the three signatures, there is a significant difference in energy (almost two orders of magnitude) between the signatures, which implies a severely low SNR. Additionally, the correlation between the three signatures is shown in Table 2.
As it can be seen, the correlation factors between the three drone noise signature are quite close; thus, their extraction from a recording of two drones flying in parallel is very challenging. However, this will be discussed further in Section 3.4.
3.3. Drone Displacement
A recording used for this part was Recording 6 (as already shown in Figure 4b) to show when the Matrice is flying in a static point. Additionally, the Recordings 5 (taking off), 8 (vertical displacement), and 7 (horizontal displacement) are also used.
Because the objective of this part of the analysis is to show how dynamic the ego-noise of a drone is, a zoomed version of the spectrograms of Recordings 5 and 6 are shown in Figure 7a,b.
As it can be seen, taking off (which implies a change in velocity) versus moving in a constant velocity exemplifies how the ego-noise of the drone changes when its velocity changes.
In Figure 8a,b the full spectrogram of Recordings 8 and 7 are shown.
Compared to the spectrogram of Recording 6 (flying in a static position) shown in Figure 4b, small variations of energy can be seen throughout the displacement recordings. To better present these changes, the frequency-wise average of the energy variance per frequency is shown in Table 3.
Although there is an increase of energy variance when the drone is moving compared to when it is static, further statistical analysis is required to prove this was because of the ego-noise and not other factors, such as wind, sensor noise, and the accuracy of automatic control algorithm. However, an interesting preliminary observation is that a horizontal displacement presents a much greater energy variance than a vertical displacement. This could be attributed to the fact that a vertical displacement involves the same amount of velocity change in all the motors and a horizontal displacement involves different velocity changes in different motors.
This part of the analysis shows a tendency between the changes in motor activity and the noise captured, which is to be expected. However, this also raises the issue of the challenge of filtering the ego-noise, because it is non-stationary. This points to the necessity of requiring adaptive filtering techniques to remove the ego-noise of the drone.
3.4. Drones Flying in Parallel
In this part of the analysis, the mean-energy-per-frequency of the spectrogram Recording 6 (as already shown in Figure 5b) was considered as the noise signature of the Matrice when flying alone. Additionally, Recordings 18 (the Matrice and the 3DR flying in parallel) and 21 (the Matrice and the Parrot flying in parallel) were also used.
In Figure 9a,b the mean-energy-per-frequency of the spectrogram of Recordings 18 and 21 are shown.
If we visually compare these figures to Figure 5b which presents the mean-energy-per-frequency of the spectrogram of Recording 6 of the Matrice, there is no clear difference between the three. In fact, it would be a near impossible task to pinpoint visually where the influence of the 3DR noise signature (Figure 6a) and of the Parrot noise signature (Figure 6b) are in Figure 9a,b, respectively. This is because, as mentioned earlier, the energy captured from the noise of the Matrice compared to the noise captured from the 3DR and the Parrot differs in several orders of magnitude. This presents a big challenge in terms of the severely low SNR that should be expected in this application study.
However, when carrying out a subjective hearing of Recording 18, it is somewhat possible to pick up the presence of the 3DR by comparing it to the recording of the Matrice alone. Thus, this points to a possibility that noise cancellation may be viable.
4. Filtering Techniques
As mentioned in Section 3, AIRA-UAS provides the noise signature of the three drones in question; thus, noise subtraction techniques should be considered. Additionally, because the noise is dynamic (depending on the motor activity of the drone), adaptive filtering techniques should also be evaluated. To this effect, the following three popular filtering techniques are to be evaluated and are briefly described in this section:
Improved Minima Controlled Recursive Averaging 
In the following descriptions, the signal model presented in Equation (1) is assumed.
where X is the recorded signal; S is the signal of interest; and N is the noise to be removed or filtered out. All signals are assumed to be in the frequency domain.
4.1. Non-Linear Noise Subtraction (NLNS)
Typically, once an estimation of the noise signal in the frequency domain () is obtained, filtering it in the frequency domain should be just a matter of subtracting from the recorded signal in the frequency domain, such as in Equation (2).
where f is the frequency bin. However, doing this directly may insert artifacts (or as Berouti put it: an “annoying musical noise”) if the resulting subtraction has a magnitude is below zero. To counter this, Berouti  proposed to carry out this subtraction based on the frequency-wise SNR and prevent the subtraction result to go below a spectral floor based on the estimated noise signal. Equation (3) presents the method to be used, referred to here as Non-Linear Noise Subtraction (NLNS).
where is a value that establishes the presence of the spectral floor; and is selected based on the estimated frequency SNR as presented in Equation (4).
where and are the maximum and minimum values of , respectively; and and are the maximum and minimum SNR values, respectively.
Typically, the estimated noise is considered as the mean value of the first windows of the recorded signal. However, in this evaluation, the mean-energy-per-frequency of the spectrogram of Recording 6 (as already shown in Figure 5b) is used for this purpose.
Additionally, in our tests the following values were used and chosen based on recommended values obtained from (2): , , , , and .
4.2. Adaptive Quantile Based Noise Estimation (AQBNE)
It can be assumed that the energy of each frequency band is at the noise level a Q-th amount of time . This amount of time can be represented as quantile of the ordered distribution of the energy values. By specifying beforehand the value of the Q percentile, the energy of the noise signal at each frequency can be calculated as in Equation (5).
where is the pre-specified quantile; and is the energy distribution in frequency f from a pre-specified amount of time .
This can be pushed even further by estimating a Q per frequency () . It does this by calculating the intersection between and a Q-estimation curve K defined in Equation (6).
where is the slope of the curve; and is the minimum value of Q. Both these values need to be pre-specified. Once the noise signal is estimated, NLNS is applied for spectral subtraction; thus, also requires being pre-specified.
To choose the values of , , and (for the NLNS part of the filter), a series of experiments were carried out with clean signals of the environment and artificially inserting noise. In this way, an accurate estimation of the SNR could be calculated. A series of values were tested, and the best SNR was obtained with , , and . The rest of the values required for the NLNS part of this filter were set as described in Section 4.1.
4.3. Improved Minima Controlled Recursive Averaging (IMCRA)
The first version of this filter estimated the noise by averaging past energy values over each frequency band, and then weighted over the probability of the presence of speech. The speech probability is calculated based on the minima values of energy, similar to AQBNE, however in this case these values are extracted from the smooth spectrogram.
This filter was later improved upon . It divided the process into two stages:
Smoothing. Its objective is to provide a rough frequency-wide voice activity detection via the original algorithm.
Minimum Tracking. Its objective is to carry out a secondary smoothing, but avoiding areas where speech energy is present.
This process results in making the minimum tracking robust and less variant. However, this makes the computation of the speech presence probability biased toward big values. The authors, thus, provide a bias compensation factor in their noise estimator.
Unfortunately, a detailed explanation of this filter requires much more space than can be adequately allocated in this work, and it is outside the scope of this paper. However, it is important to mention that there are 9 parameters that are part of the calibration of this filter. The authors themselves provide a set of recommended values for each parameter. However, as it will be seen in Section 5.2, this set of recommended values did not provide good results. To counter the possibility of inappropriate calibration, we carried out a similar set of experiments as in Section 4.2 to choose the parameters of this filter. Thus, the results for this filter will be presented with two different configurations: the ‘default’ configuration presented in  and our ‘alternate’ configuration that was obtained from our preliminary experiments. The values for each configuration are presented in Table 4.
5. Evaluation and Results
There were two evaluations that were carried out, each with a separate objective: (1) to measure the amount of ego-noise that was removed or filtered/canceled out, and (2) toevaluate if the noise signature of the other drone flying in parallel (3DR or Parrot) is collaterally extracted.
5.1. Ego-Noise Reduction
The three filters described in Section 4 were applied to Recording 9 to measure their effects on the Matrice’s ego-noise while flying statically. Recording 6 also presents the Matrice flying statically, however, because the noise estimation for NLNS is extracted from this recording, using Recording 9 presents a fairer comparison to the other two filters. These results are shown in Table 5.
To measure the amount of ego-noise removal when the drone is moving, Recordings 5 (moving vertically) and 7 (moving horizontally) were used. Table 6 and Table 7 show the reduction of the ego-noise from Recordings 5 and 7, respectively.
Additionally, in Figure 10, Figure 11 and Figure 12, the spectrogram of the original recordings, as well as the filtered from all four configurations, are shown from Recordings 9, 5, and 7. Further discussion of these results is provided in Section 5.3.
5.2. Observed Drone Signature Corruption
To evaluate if the signatures of the drones flying in parallel to the Matrice are kept intact in the recordings, the mean-energy-per-frequency of the spectrogram of Recordings 16 and 19 are used for the 3DR and Parrot, respectively. In these recordings, shown in Figure 6a,b, the Matrice is turned off so only the noise signatures of the other drones are recorded. Recordings 18 and 21, were the Matrice and the 3DR or the Parrot are recorded flying in parallel, respectively, were used to evaluate if the filters were able to filter out the ego-noise of the matrice but leave intact the noise signature of the 3DR or the Parrot, respectively. In Figure 13 and Figure 14, the mean-energy-per-frequency of the spectrogram of the resulting filtered spectrogram are shown for all filters, as well as the original noise signature from Recording 16 and 19, respectively.
To evaluate in a more objective manner, the correlation was measured between the noise signatures in Recordings 16 and 19 and the filtered results of Recording 18 and 21, respectively. These correlation results are shown in Table 8 and Table 9.
5.3. Results Discussion
As it can be seen in Table 5, the alternate configuration of IMCRA carries out the biggest noise reduction when flying statically, but AQBNE is not that far behind. However, it can be seen from Table 6 and Table 7 that for the rest of the cases, AQBNE removes more ego-noise. Additionally, a visual inspection of the spectrograms of the results shown in Figure 10, Figure 11 and Figure 12, as well as from subjective hearings of the filtered recordings, it is clear that IMCRA in both configurations inserts a far greater amount of artifacts than AQBNE.
Additionally, it can be seen from Figure 13a and Figure 14, that there is still a clear amount of presence of the Matrice noise signature in the filtered results (which can be seen in Figure 5b). If the high correlation factors seen in Table 8 and Table 9 are also considered, both of these results imply that the tendencies of the noise signatures of the three drones are very close to each other. This is further evidence at what was suggested from the correlation between the noise signatures of the three drones presented in Table 2.
All of this in conjunction implies that, even though AQBNE is the preferable filter from the three chosen in terms of ego-noise removal, it is not able to keep intact the noise signature of the other drone flying in parallel. This may not be a fault of the filter itself, however, since both signatures are very correlated, and their removal may require a more sophisticated method of source separation.
It is also important to mention that the configuration process of these filters may be improved, which may reduce the presence of artifacts.
6. Conclusions and Future Work
In this paper, an in-depth analysis of the AIRA-UAS corpus was carried out. It was found that: The environment in which it was recorded had a very small amount of noise; the current drones’ noise signatures may not be sufficient for discrimination, and they have high correlation between them, making their extraction challenging; the ego-noise is dynamic, having a bigger variance when the drone is moving horizontally than when moving vertically or flying statically; and the recordings have a severely low SNR which, again, will make the noise signature extraction very challenging. This was further evidenced when the three popular filtering techniques that were evaluated did not provide convincing results.
The filtering techniques that were evaluated were Berouti’s Non-Linear Noise Subtraction, Adaptive Quantile Based Noise Estimation, and Improved Minima Controlled Recursive Averaging. The latter required two configurations of its parameters to be tested since its default configuration delivered poor performance against the low SNR.
In terms of noise reduction, AQBNE provided the best results, with the proposed alternate configuration of IMCRA coming in close second when the drone is not moving. IMCRA provided a severe amount of artifacts in both configurations. To be fair, however, other configurations of this filter may counter this, which is left for future work.
These results make it clear that there is a prime field of research in audio processing over drones since current popular algorithms are not up to par for its challenge, opening the door to investigate alternative methods to tackle this task.
To this effect, other type of audio processing algorithms will be evaluated in future works, such as beam-forming techniques and source separation techniques. Geometric Source Separation, for example, is a good candidate for further evaluation.
It is important to consider that the ego-noise of the drone is dynamic but not unpredictable. Since the control signal of the drone’s motor is known, a model could be created the uses this information to predict the noise emitted by the drone and provided a better noise estimation to be canceled. The implementation of this algorithm is also left for future work.
Finally, to overcome the issue of high correlation between the drones’ noise signatures, a drone can be designed with this in mind such that its noise signature does not have such a high correlation with other drones. However, this would require a new rotor design, as well as a new propeller design, and as such is left for future work, as well.
Conceptualization, C.R. and J.M.; methodology, C.R.; software, C.R. and O.R.; validation, C.R.; formal analysis, C.R.; investigation, C.R.; resources, C.R. and J.M.; data curation, C.R. and O.R.; writing–original draft preparation, C.R.; writing–review and editing, C.R., O.R. and J.M.; visualization, C.R.; supervision, C.R. and J.M.; project administration, C.R. and J.M.; funding acquisition, C.R. and J.M.
The authors thank the support of CONACYT through the project 251319. This work has also been partially funded by the Royal Society through the Newton Advanced Fellowship with reference NA140454.
Conflicts of Interest
The authors declare no conflict of interest.
Mohammed, F.; Idries, A.; Mohamed, N.; Al-Jaroodi, J.; Jawhar, I. UAVs for smart cities: Opportunities and challenges. In Proceedings of the 2014 International Conference on Unmanned Aircraft Systems (ICUAS), Orlando, FL, USA, 27–30 May 2014; pp. 267–273. [Google Scholar] [CrossRef]
Goodbody, T.R.; Coops, N.C.; Marshall, P.L.; Tompalski, P.; Crawford, P. Unmanned aerial systems for precision forest inventory purposes: A review and case study. For. Chron.2017, 93, 71–81. [Google Scholar] [CrossRef]
de Lima, R.; Martinez-Carranza, J.; Morales-Reyes, A.; Mayol-Cuevas, W. Toward a smart camera for fast high-level structure extraction. J. Real-Time Image Process.2017, 14, 685–699. [Google Scholar] [CrossRef]
Al-Kaff, A.; Martin, D.; Garcia, F.; de la Escalera, A.; Armingol, J.M. Survey of computer vision algorithms and applications for unmanned aerial vehicles. Expert Syst. Appl.2018, 92, 447–463. [Google Scholar] [CrossRef]
Harvey, M.; Rowland, J.; Luketina, K. Drone with thermal infrared camera provides high resolution georeferenced imagery of the Waikite geothermal area, New Zealand. J. Volcanol. Geotherm. Res.2016, 325, 61–69. [Google Scholar] [CrossRef]
Furukawa, K.; Okutani, K.; Nagira, K.; Otsuka, T.; Itoyama, K.; Nakadai, K.; Okuno, H.G. Noise correlation matrix estimation for improving sound source localization by multirotor UAV. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 3943–3948. [Google Scholar] [CrossRef]
Hoshiba, K.; Washizaki, K.; Wakabayashi, M.; Ishiki, T.; Kumon, M.; Bando, Y.; Gabriel, D.; Nakadai, K.; Okuno, H.G. Design of UAV-Embedded Microphone Array System for Sound Source Localization in Outdoor Environments. Sensors2017, 17, 2535. [Google Scholar] [CrossRef] [PubMed]
Park, S.; Shin, S.; Kim, Y.; Matson, E.T.; Lee, K.; Kolodzy, P.J.; Slater, J.C.; Scherreik, M.; Sam, M.; Gallagher, J.C.; et al. Combination of radar and audio sensors for identification of rotor-type Unmanned Aerial Vehicles (UAVs). In Proceedings of the 2015 IEEE SENSORS, Busan, Korea, 1–4 November 2015; pp. 1–4. [Google Scholar]
Blanchard, T.; Thomas, J.H.; Raoof, K. Acoustic Signature Analysis for Localization Estimation of Unmanned Aerial Vehicles Using Few Number of Microphones. In Proceedings of the MATEC Web of Conferences-The 2nd Franco-Chinese Acoustic Conference (FCAC 2018), Le Mans, France, 29–31 October 2018; Volume 283, p. 04002. [Google Scholar] [CrossRef]
Tomic, T.; Schmid, K.; Lutz, P.; Domel, A.; Kassecker, M.; Mair, E.; Grixa, I.L.; Ruess, F.; Suppa, M.; Burschka, D. Toward a Fully Autonomous UAV: Research Platform for Indoor and Outdoor Urban Search and Rescue. IEEE Rob. Autom Mag.2012, 19, 46–56. [Google Scholar] [CrossRef]
Sibanyoni, S.V.; Ramotsoela, D.T.; Silva, B.J.; Hancke, G.P. A 2-D Acoustic Source Localization System for Drones in Search and Rescue Missions. IEEE Sens. J.2018, 19, 332–341. [Google Scholar] [CrossRef]
Ruiz-Espitia, O.; Martinez-Carranza, J.; Rascon, C. AIRA-UAS: An Evaluation Corpus for Audio Processing in Unmanned Aerial System. In Proceedings of the 2018 International Conference on Unmanned Aircraft Systems (ICUAS), Amsterdam, The Netherlands, 10–11 May 2018; pp. 836–845. [Google Scholar]
Rascon, C.; Meza, I.; Millan-Gonzalez, A.; Velez, I.; Fuentes, G.; Mendoza, D.; Ruiz-Espitia, O. Acoustic interactions for robot audition: A corpus of real auditory scenes. J. Acoust. Soc. Am.2018, 144. [Google Scholar] [CrossRef] [PubMed]
Rascon, C.; Ruiz-Espitia, O. Acoustic Interactions for Robot Audition—Unmanned Aerial Systems (AIRA-UAS). 2018. Available online: http://calebrascon.info/AIRA-UAS/ (accessed on 2 September 2019).
Huang, Y.A.; Benesty, J. Audio Signal Processing for Next-Generation Multimedia Communication Systems; Springer Science & Business Media: New York, NY, USA, 2007. [Google Scholar]
Berouti, M.; Schwartz, R.; Makhoul, J. Enhancement of speech corrupted by acoustic noise. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP ’79), Washington, DC, USA, 2–4 April 1979; Volume 4, pp. 208–211. [Google Scholar]
Bonde, C.S.; Graversen, C.; Gregersen, A.G.; Ngo, K.H.; Nørmark, K.; Purup, M.; Thorsen, T.; Lindberg, B. Noise Robust Automatic Speech Recognition with Adaptive Quantile Based Noise Estimation and Speech Band Emphasizing Filter Bank. In Nonlinear Analyses and Algorithms for Speech Processing; Faundez-Zanuy, M., Janer, L., Esposito, A., Satue-Villar, A., Roure, J., Espinosa-Duro, V., Eds.; Springer: Berlin/Heidelberg, Germany, 2005; pp. 291–302. [Google Scholar]
Cohen, I. Noise spectrum estimation in adverse environments: improved minima controlled recursive averaging. IEEE Trans. Speech Audio Process.2003, 11, 466–475. [Google Scholar] [CrossRef]
Grondin, F.; Létourneau, D.; Ferland, F.; Rousseau, V.; Michaud, F. The ManyEars open framework. Auton. Robots2013, 34, 217–232. [Google Scholar] [CrossRef]
Rascon, C.; Meza, I. Localization of sound sources in robotics: A review. Rob. Autom. Syst.2017, 96, 184–210. [Google Scholar] [CrossRef]
Stahl, V.; Fischer, A.; Bippus, R. Quantile based noise estimation for spectral subtraction and Wiener filtering. In Proceedings of the 2000 IEEE International Conference on Acoustics, Speech, and Signal Processing, Istanbul, Turkey, 5–9 June 2000; Volume 3, pp. 1875–1878. [Google Scholar] [CrossRef]