Next Article in Journal
A Convolutional Neural Network and Graph Convolutional Network Based Framework for AD Classification
Previous Article in Journal
Horizontally Distributed Inference of Deep Neural Networks for AI-Enabled IoT
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Urban Traffic Noise Analysis Using UAV-Based Array of Microphones

by
Marius Minea
and
Cătălin Marian Dumitrescu
*
Department Telematics and Electronics for Transports, University Politehnica of Bucharest, 060042 Bucharest, Romania
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(4), 1912; https://doi.org/10.3390/s23041912
Submission received: 16 January 2023 / Revised: 31 January 2023 / Accepted: 6 February 2023 / Published: 8 February 2023
(This article belongs to the Section Remote Sensors)

Abstract

:
(1) Background: Transition to smart cities involves many actions in different fields of activity, such as economy, environment, energy, government, education, living and health, safety and security, and mobility. Environment and mobility are very important in terms of ensuring a good living in urban areas. Considering such arguments, this paper proposes monitoring and mapping of a 3D traffic-generated urban noise emissions using a simple, UAV-based, and low-cost solution. (2) Methods: The collection of relevant sound recordings is performed via a UAV-borne set of microphones, designed in a specific array configuration. Post-measurement data processing is performed to filter unwanted sound and vibrations produced by the UAV rotors. Collected noise information is location- and altitude-labeled to ensure a relevant 3D profile of data. (3) Results: Field measurements of sound levels in different directions and altitudes are presented in the paperwork. (4) Conclusions: The solution of employing UAV for environmental noise mapping results in being minimally invasive, low-cost, and effective in terms of rapidly producing environmental noise pollution maps for reports and future improvements in road infrastructure.

1. Introduction

Urban traffic is, in present days, a major factor of urban environmental pollution. Therefore, most of the EU countries and countries worldwide are trying to alleviate the negative effects of traffic on the environment, seeking solutions both in infrastructure and on the mobile platforms. Both emissions of traffic-generated chemical composites, particulate matter, and noise produce negative effects on residents, traffic participants, and local natural life. Knowing the way that noise from daily road traffic is dispersed and behaves could be beneficial for competent authorities and municipalities to evaluate and propose countermeasures to reduce the impact on environment and citizens’ lives. In this paper, research that has been carried out is presented to determine the main characteristics of the urban noise produced by traffic, and a solution for a cost-effective and efficient collection of noise data is also presented. The data collection is carried out using a UAV-airborne set of microphones and a data processing unit.
The research in this field is quite rich, and several solutions for collecting information on environmental noise pollution have been proposed by different scholars and industrial researchers.
Starting with modeling, both polluting emissions and noise assessment in urban areas represent a measure designed to find the most appropriate solutions for reducing their impact on citizens’ lives, as shown in [1], where Hui Di et al. determined in specific conditions a traffic noise propagation model to predict instantaneous sound levels, based on noise attenuation. The authors consider that noise is affected by air absorption, obstacle reflex, and ground absorption during its transmission process, which causes noise attenuation, and these elements are considered in building a noise mapping and assessment model. In their conclusion, the authors show that the quality of the acoustic environment seems to be more moderate during the daytime and better at night.
Zefreh, M. M. and Torok, A. go deeper in the analysis of noise and consider different traffic conditions to evaluate their effect on noise intensity and dispersion [2], performing theoretically generated conditions by the Monte Carlo Simulation technique, following the distribution of traffic speed in the urban roads. They show in the conclusions that “the free flow condition and the oversaturated condition would generate the same amount of sound energy (sound exposure level) but the vehicles in free flow condition produce higher range of sound pressure level (58.7 dBA to 64.59 dBA)”.
Kai Cussen et al. [3] evaluated the noise produced by a UAV according to the legally mandated ISO 3744 and investigated the suitability of commercial implementations of ISO 9613 for modeling noise emission from UAVs. Finally, they state that models are adequate for assessing UAV’s noise in terms of directivity. In addition, they showed that most UAVs in the actual industrial production may exceed EU limits by around 1.8 dBA, causing possible urban noise issues.
The study published by Fiedler and Zannin [4] was performed in several Latin America cities to determine noise presence and produce noise maps, also including noise-sensitive areas and locations, such as hospitals. The study also shows that noise maps are extremely helpful tools in the processes of diagnosis and evaluation of noise pollution. The researchers employed different scenarios in hypothetical situations to produce and assess the effects of noise pollution on the environment, simulating increase and decrease of traffic volumes by 50–56% and studying the effects on noise production. At the end of their work, they concluded that reducing heavy traffic, or general traffic, by around 50% can produce a correspondent noise reduction of around 3 dBA (or dB(A)—A-weighted decibel that represents an expression of the relative loudness of sounds as perceived by the human ear; the latest standard is ISO 226:2003).
Doygun H. and Gurun, D.K. [5] tried to quantify noise pollution generated by the urban traffic in a Turkish city, producing over 114 measurements of noise levels in 38 different urban locations, classified as residential or industrial areas. The study was aimed at quantifying the temporal and spatial dynamics of urban traffic noise, to compare levels with national and international limits, and determine mitigative measures against noise pollution. Remarkably, in this study is the conclusion that noise pollution is usually ignored in public policies, regulations, and laws regarding environmental conditions. The authors recommend renewing road surfaces to reduce noise levels by around 5 dBA. Similar research is provided in papers [6,7,8,9]. High capacity and speed roads also came into focus for environmental emissions, PMs, and noise evaluation. In [10] is presented a study to gauge the existing public’s attitude and degree of awareness to contemporary vehicular noise pollution. The study revealed that most people are affected by traffic noise, several people attributing increased headaches and stress to the excessive noise levels. It has been shown by field evaluation that “air pollution and smoke disturb people with the percentage of 54.2% and 62.9%. From Pie chart 3 noise pollution also disturbs people surrounding with 60%”. In conclusion, the study recommends reducing traffic noise on urban highways by a program of shared responsibility. Similar information is presented in works [11,12].
Ciaburro G. et al. [13] expanded noise data collection purposes beyond simple environmental effect, giving it even more importance: early detection of security problems related to citizens’ mobility, crime, risk of terrorism. Their study has led to the conclusion that it is important to analyze the sound from different characteristics and tonal components. The ambient noise in UAV and non-UAV scenarios is very complex, it is not possible to distinguish between the different acoustic sources, at least in the time domain, and it is possible to employ a CNN-based classification system to identify the presence of UAVs with an accuracy of 0.91. Research is conducted in several papers to study effects of road gradients’ influence on chemical pollutant emissions (via modeling) and noise. Studies in similar areas of research are shown in papers [14,15,16,17].
The research conducted by Luo L. et al. [18] is oriented, on the other hand, towards collecting noise information via wireless acoustic sensor networks (WASNs). They propose a new system that employs WASNs to monitor the urban noise and recognize acoustic events with a high performance; the system is composed of sensors with the ability to produce local signal processing, convolutional neural networks for classification, and a platform for noise maps visualization. This direction of research is also employed by other authors, as shown in papers [19,20,21,22,23,24,25], employing also modeling of emissions based on Google maps road profiles [24].
Based on noise measurements, evaluations, and mapping, noise reduction strategies can be imagined for different urban or extra-urban scenarios. Hsiao Mun Lee et al. [26] extended the study on traffic-generated noise to affected population, dividing it into group noise indicators (highly annoyed and sleep-disturbed people) and studying the effect of noise reduction measures on these segments of the population. The results of this study showed that installing noise absorbents and barriers is highly recommended because it significantly reduces the influence of noise on nearby located environment. However, this solution solves neither the problem of chemical pollutants emissions nor the PM, also generated by traveling vehicles [27]. Of course, there are different effects on those segments of the population, according to the measures taken. Similar research is given in works [28,29,30].
General model-based noise mapping in urban areas is the focus in paper [31], where Ye Liu et al. discuss the working principle of general model-based noise mapping using mobile crowdsensing and acoustic sensor networks (ASNs). The authors propose an interesting solution, that is, collecting sound information via smart devices carried by citizens in an urban area (smart watches, smart bracelets, etc.) and using GPS-related information to send relevant data regarding levels and features of noise pollution. However, this solution would probably need serious discussions about ensuring privacy, as personal mobile devices are used to collect audio files. Connected with this research and with a view on future possible applications, based on work [32], noise-relevant data collection and, in a certain amount, chemical pollutants information could also be possibly collected by the means of connected vehicles. However, an overall view of existing technologies for noise data collection specific for smart cities is given in [33].
Within the collection of sound information in cities, it is also possible to detect specific events from this collection of data. Such a proposal is shown in work [34], where Soccoro J.C. et al. are investigating the possibility to eliminate anomalous noise events (ANE) unrelated to road traffic.
The preoccupation towards the employment of mobile crowdsourcing for data collection is also growing in this domain, as the new developments for smart city applications have issued mobile crowdsensing (MCS), a service designed to enable users to collect and share sensor data in different urban areas. As shown in [35], “MCS services can produce detailed sensor readings and provide means to discover new phenomena in urban environments that otherwise cannot be measured by individuals, such as the occurrence of traffic congestion, or environmental noise pollution monitoring”.
As it can be observed from the above literature review, the focus on collecting information regarding urban noise pollution is at high preoccupation of different researchers, and it is also in focus for the development and concept of new road infrastructures, especially in smart cities transformation processes. The purpose of the present research is to investigate a simple, but efficient and effective method for noise data collection employing a UAV that can be used both for building 3D noise maps and to detect ANEs on specific paths, contributing to the correct management of environmental protection and also to public security in large cities. As a future development, the solution of employing UAV-borne sensors could be also used for collecting relevant data on other types of road traffic-generated pollutants, such as CO, CO2, NOx, SO2, and PM, most of the latter deriving from harmful products, such as harmful byproducts including nitrogen dioxide, carbon monoxide, hydrocarbons, benzene, and formaldehyde.
The remainder of this paper is organized as follows: Section 2 is dedicated to Materials and Methods; Section 3, Results; Section 4, Discussion; and, finally, Conclusions.

2. Materials and Methods

2.1. General Information

Sound data collection is important in many applications for smart cities: control of environment, reduction of traffic-generated pollution, or even detection of security-related public events. Mapping the sound levels and characteristics could also be beneficial for smart city administrators, helping them to better control traffic flowing, delicate environmental areas such as hospitals, and people’s mobility, and to take the most appropriate measures to ensure a healthy level of ambient noise in urban areas to provide the best environment for living.
Sound pressure level (SPL) represents the difference in the local ambient atmospheric pressure that is caused by the sound wave. Having as a measurement unit the Pascal (Pa), the SPL can be measured in open air using a microphone. Usually, a microphone’s sensitivity is estimated using a sine wave at 94 dB SPL (equivalent to 1 pascal of air pressure) at the frequency of 1 kHz, where the human acoustic sensitivity is high. The microphone’s sensitivity is measured as its response to the 94 dB SPL signal. Such a microphone considers the maximum value of sound pressure (measured in dB SPL) at which the microphone can work correctly before producing audible levels of distortion. Close-passing road traffic usually produces noise ranging from 60 to 100 dB SPL ( dB   SPL = 20 l o g 10 P 1 P 0 P 1 , where P 1 represents the sound pressure level measured [Pa] and P 0 is the reference sound pressure level of 20 [µPa]). The value of 0 dB SPL corresponds to the lowest threshold of human hearing. In normal conditions, pressure levels might reach quite a large range, spreading from 0.00002 to 200 pascals. However, sensors and devices such as microphones do not have the same dependency on SPL as the human ears. Therefore, A-weighting represents a way to make the sound loudness relative to SPL to be perceived by sensors in a similar way as the human ears. While the theoretical sound frequency band perceived by the human ears is 20–20,000 Hz, in reality, it depends quite a lot on individuals, and especially on their age. Older people usually have a much less broad band perception. However, in the 20–20 kHz frequency range, a normal person perceives lower frequencies with lower sensitivity, and human hearing is generally logarithmic in nature, so decibels, as a measurement unit, match in a large amount with human perception and experience. Usually, normal people perceive sounds with a maximum accuracy in the range of 250–5000 Hz. Therefore, a so-called “A-weighted” curve is used to compensate an acoustic sensor’s sensitivity and make it similar, or at least very close to that of the human ears. The A-weighted curve is depicted in Figure 1a, and the conversion from [dB SPL] to [dBA] employing A-weighting adjustments is presented in Figure 1b.
A-weighting is the standard for establishing hearing damage caused by high sound pressure levels and also used to determine noise pollution. International Electrotechnical Commission standard 61672:2003 presents the curves that are employed in sound meters. All modern sound meters are able to provide A-weighted decibel and C-weighted decibel (dBC) measurements (C-weighted is a flatter curve of sensitivity, used beyond 100 dBs, where the human hearing characteristics also change). In this standardization, there is also the “Z-weighted” curve, a specific characteristic where no compensation is applied, defining the band between 20 Hz and 20 kHz with ±1.5 dB.
SPL, in decibels, is described by Equation (1), as defined by ISO 9613-2:1996 standard.
L A T = 10 lg { [ ( γ T ) 0 T p A 2 ( t ) d t ] / p 0 2 }
where:
L A T —equivalent continuous A-weighted sound pressure level;
p A ( t ) —the instantaneous A-weighted sound pressure, [Pa];
p 0 —the reference sound pressure (= 20 × 10 6 ) Pa;
T —specified time interval [s].
The attenuation produced by geometrical sound divergence may be computed according to Equation (2):
A D i v = [ 20 l g ( d d 0 ) + 11 ]
where:
d—distance from source to receiver [m];
d 0 —reference distance (1 m).
The atmospheric absorption is given by:
A a t m = α d 1000
where α—the atmospheric attenuation coefficient [dB/km].

2.2. Instruments and Initial Setup

In order to initially determine the main characteristics of the traffic-generated noise, some field measurements have been conducted. The instrumentation employed for this purpose was composed of:
-
hardware: sonometer (UNI-T UT 353BT) with a Bluetooth connection to a smartphone for data storage and processing) and a smartphone Android (Samsung Galaxy A50 SM-A505FN/DS);
-
software: UTienv 2.0.8 and Sound Analyzer 2.4 running on Android 11.
The tests were performed in a busy junction in Bucharest, Romania (Figure 2).
The location for initial measurements was chosen considering that the respective area is known as very busy and with high levels of traffic-generated noise. The main boulevard is four lanes per direction of traveling, and there are also public transport lines. The location is situated near a very busy subway station and a large market. Ground level initial measurements of traffic-generated noise had the purpose of determining actual levels of noise pollution, frequency range, and main causes. The microphone was placed 1 m above ground, on a tripod, facing traffic, 5 m from the first lane (on a shoulder of the road). Post-processing of collected information consisted of recording levels on different frequencies and analyzing SPL.
However, due to the current period, there were some specific weather conditions when the tests have been performed, meaning the carriageway was wet, causing increased traffic noise with a slightly different frequency range, compared to noise generated by the normal (dry) roadway surface. Still, we consider that this is not a major downside, as, in this case, we managed to collect information in a really peak noise environment, in a very busy area. Further field measurements will be taken also in different humidity conditions, to compare results. Outside temperature was 3 °C. The area is surrounded in the N–E and S by relatively high buildings, favoring propagation of noise with many reflections. The NW part is occupied by a busy market and commercial areas.

2.3. Initial Findings

Based on the field-collected information, several diagrams have been drawn to determine the main components of traffic-generated noise and their amplitudes, knowing the fact that the main health affections caused by traffic noise are increased hypertension risk, myocardial infarction, cardiovascular, and even stroke mortality risk. High noise sensitivity has also been recently associated with lower ischemic heart disease mortality risk, and road traffic-generated noise was also associated with Phase 4 psychological ill-health, but only among those exposed to 56–60 dBA [36]. In the area where the initial measurements were conducted, there are a lot of buildings with numerous flats, so it also represents an important residential area, hence the importance of finding such noise levels. Initial results are presented in the following diagrams.
It can be observed in Figure 3 that the main traffic noise components (with SPL greater than 50 dBA) increase from 100 Hz and extend up to over 15 kHz, occupying most parts of the human hearing spectrum. The peak levels are reached at around 400–500 Hz. An interesting fact is that some medical reports and scientific papers claim that a 432-Hz tuning has better effects on the human body, but there are no scientific studies that support this hypothesis. The sound frequency of “432 Hz … was associated with a slight decrease of mean (systolic and diastolic) blood pressure values (although not significant), a marked decrease in the mean of heart rate (−4.79 bpm, p = 0.05) and a slight decrease of the mean respiratory rate values (1 r.a., p = 0.06), compared to 440 Hz” [37], but this is the case for music, and not for incoherent sounds. Usually, traffic noise is responsible for stress increase.
To determine more precisely the sound components, separate diagrams have been produced for the main harmonic frequencies, as shown in Figure 4, Figure 5 and Figure 6.
When analyzing the diagrams in Figure 4, Figure 5 and Figure 6, another observation that may be drawn is that the levels are most varied in the lower frequency range than in medium and higher frequency ranges (Figure 4 compared to Figure 5 and Figure 6).
Levels appear to be more constant in midrange frequencies (Figure 5), exceeding 55 dBA on a regular basis. Still, considering that the initial field measurements have been conducted with a wet roadway surface, it is possible that a higher range of frequencies (above 2 kHz) of the levels of sound pressure are to be favorized by the humid surface of roads. Finally, sound components recorded for higher frequencies also show an increased variation compared to medium range (Figure 6 compared to Figure 5).
To determine changes in sound characteristics when picking up data at a different altitude, another set of field measurements has been conducted using a flat in a nearby building at 30-m high by placing the instrumentation in this location. The results are presented in the following diagrams.
In Figure 7, recorded higher frequencies experienced a significant decrease, up to 40 dBA, while the peak changed also in frequency, being situated around 1900 Hz. In addition, while the peak level of sound pressure remained around 60 dBA, this was only for a smaller band, from approx. 1 kHz up to 5 kHz, and the rest of the levels experienced a decrease, more obviously in the lower domain.
One observation that is worth mentioning is that noise levels are more variable when recorded at higher altitude; this is probably because of side activities’ influence (especially in the 50-Hz range, Figure 8).
This remarque is also available for the other frequencies (medium range and higher range, Figure 9 and Figure 10).
From the above presented measurements, it results that it is possible to collect relevant information on traffic-generated noise even from above, meaning that a solution to provide a more detailed imaging on sonic pollution is possible to obtain from an automated UAV flight over selected areas, at specific intervals of time.

2.4. Information Collection Process

Data collection regarding traffic-generated noise is important in the development of a friendly environment for the inhabitants of a large city. However, it is not economic to install sensors in every important point of a street’s network, and there are involved important costs in such as an investment. Still, it is advisable that the next generation for this activity employ sensors in a future connected vehicles network, installed onboard vehicles, using those as “floating cars”, or moving sensors, in the traffic flow. Until then, though, a simpler solution is proposed in this research—to employ periodic measurements in different locations using a UAV. The vehicle is equipped with a specific configuration of microphones, able to create a very informative sound imaging and overall information regarding the sonic pollution that moving traffic is producing in large urban agglomerations.
The proposed solution estimates traffic noise based on the use of a drone on which a directional microphone array is mounted. To reduce the background noise, and especially the noise produced by the propeller rotations, an adaptive noise cancellation (ANC) filter developed on the FPGA is employed. Two microphones are pointed upwards, collecting noise from UAV rotors, in order to reduce its level. Adaptive filtering is performed with an optimized Least-Mean-Square (LMS) algorithm. The general architecture of the system is presented in Figure 11, and, in the following subsections, the mathematical concepts and processing modules are described.

2.5. Sound Processing

The directional microphone array is arranged on a Kintex 7 FPGA development board with a square-shaped ultrasound sensor array containing 24 MEMS capsules, spaced at 11 mm. Figure 12 shows the simulation of the square microphone array together with the beamforming analysis using MUSIC DOA (Multiple Signal Classification & Direction of Arrival) [38]. DOA indicates the direction from which a propagating wave typically arrives at a point where a set of sensors is placed [39]. The image on the left of Figure 12 shows the energy detection of the acoustic signal generated by the drone’s rotors, along with the location position labeling (azimuth and elevation), for the two acoustic frequencies characteristic of drones (white color), represented on the frequency spectrum (bottom right).
The configuration of the acoustic processing system is presented in Figure 13—the acoustic signal picked up from the microphone sensor array is converted from pulse density modulation (PDM) to pulse code modulation (PCM) for each sensor. The conversion module consists of a suite of digital filters, and, finally, a multiplication block is inserted (as shown in Figure 14). PDM can be the ideal choice in challenging application environments, thanks to its inherent resilience to noise.
In the first stage, a low-pass comb filter for PDM to PCM conversion is employed. To achieve lossless conversion, we use a half-band compensation filter because the passband of the comb filters is not flat, and the last digital filter of the finite impulse response (FIR) type has the role of eliminating the noise remaining after conversion. Finally, the multiplier converts the signal to a frequency in the audible spectrum. The sampling frequency for PDM is 4.608 MHz for the MEMS sensors used in the construction of the microphone array.

2.6. Delay and Sum Beamforming

To achieve the Delay & Sum mode, we use the delay and sum algorithm (DAS)–Bartlett beam. The formation of the DAS beamformer consists of applying a delay and a weight to the output of each sensor, and, finally, summing the signals [40]. The delays have the role of obtaining the maximum of the sensitivity matrix for the received acoustic signals. Moreover, by adjusting the delays, the directionality of the microphone area is oriented towards the direction of interest for the reception of the signals, thus achieving the consecutive addition of the individual acoustic signals received from the sensors. This aspect means that, at certain angles of incidence, constructive interference is obtained, and, for other angles of incidence, destructive interference is obtained. Figure 15 shows the way the DAS algorithm works.

2.7. Directional Energy

To perform the audio energy analysis, some processing must be performed to identify the energy of the summed audio signals [39,40,41]. Because the energy of a continuous signal x c ( t ) is calculated with the formula:
E c = + | x c ( t ) | 2 d t
for discrete signals x(n), the energy is calculated with the formula:
E d = + | x [ n ] | 2
The hardware implementation for the directional energy calculation is shown in Figure 16. An FPGA multiplier module is employed to form the square DAS beamforming signal. The output of the accumulator module calculates the signal energy based on the acquired samples.

2.8. Adaptive Filtering

The adaptive filter for identifying an unknown system has the configuration shown in Figure 17. In this configuration, x(n) represents the input signal at the discrete time at moment n, while y(n) is the output of the system to be determined (i.e., h) [41]. By filtering the input signal through both systems, we can write y ^ ( n ) = h ^ T ( n 1 ) x ( n ) and y ( n ) = h T x ( n ) , where x ( n ) is a vector containing the most recent L input samples [42,43].
Finally, the output of the adaptive system is subtracted from the desired signal, so that the error signal e(n) results in e ( n ) = d ( n ) y ^ ( n ) .

2.9. Least-Mean-Square (LMS) Algorithm

The Wiener filter facilitated the development of one of the most popular adaptive algorithms, the LMS algorithm. The popularity of the LMS algorithm is due to its ease of understanding, implementation, and use. At first, the Wiener filter aimed to determine an optimal filter, h ^ w , which minimizes the error signal in each iteration [43,44].
In this context, the optimal filter that minimizes the error signal is identified by minimizing a cost function, the mean squared error (MSE), defined as J [ h ^ ( n ) ] = E [ e 2 ( n ) ] , where E [ . ] represents the statistical averaging operator. After several calculations, the optimal Wiener filter is obtained as h ^ W = R x 1 p x d .
Calculation of the p x d correlation vector and the autocorrelation matrix, R x , involves the knowledge of some statistical estimates, making the steepest slope algorithm unsuitable for practical use [45,46]. If we consider the instantaneous values:
p x d ( n ) = d ( n ) x ( n ) ,   R x ( n ) = x ( n ) x T ( n )
the following equation results:
h ^ ( n ) = h ^ ( n 1 ) + 2 μ x ( n ) e ( n )
which defines the updated equation of the LMS algorithm.
The theoretical configuration of an adaptive filter for noise reduction/adaptive noise cancellation (ANC) is presented in Figure 18.

2.10. Optimized LMS Algorithm

To improve the performance of the LMS algorithm, the fixed step of updating the coefficients is replaced with a control matrix. Y(n) is considered the scaled unitary control matrix, and:
Y ( n ) = μ ( n ) M L
where M L is the LxL identity matrix and μ is fixed step size.
The autocorrelation matrix for the error coefficients is then calculated, replacing the misalignments, and obtaining the following result:
R c ( n ) = [ R c ( n 1 ) + σ h 2 M L ] μ ( n ) [ [ R c ( n 1 ) + σ h 2 M L ] R x + R x [ R c ( n 1 ) + σ h 2 M L ] ] + μ 2 ( n ) { 2 R x [ R c ( n 1 ) + σ h 2 M L ] R x + R x t r [ [ R c ( n 1 ) + σ h 2 M L ] R x ] } + μ 2 ( n ) σ w 2 R x
where:
σ h 2 —system noise variation;
R x —autocorrelation matrix of the input signal.
The ANC simulation result with optimized LMS is presented in Figure 19. It is found that the system noise variation parameter has an important role because if it has a high value, it allows the optimized LMS algorithm to determine the new system parameters in a reasonable calculation time.

2.11. Neural Network for Road Traffic Noise Prediction

For the prediction of road traffic noise, a multilayer feed-forward neural network (MFFN) architecture is employed [46,47,48]. The classic structure of the MFFN is shown in Figure 20.
MFFNs are layered feed-forward networks, trained with statistical inverse propagation. The main advantages are determined by the ease of use and implementation, but also that they can approximate any input/output map [49,50]. The input information of the MFFN network for traffic volume, average speed, and vehicles is obtained from a convolutional neural network (CNN), and the gradient and building density information is determined by Equations (9) and (10) presented in the following methodology subsection.

3. Results

3.1. Methodology

The equalization of the continuous sound level, weighted A, is achieved by the constant sound level for a period of time and contains the same acoustic energy as its variation in real time [50,51,52,53,54]. Mathematical expressions are defined by:
L A e q = 10 L o g ( 1 ( t 2 t 1 ) t 1 t 2 p 2 ( t ) p 0 2 ) d t
where p(t) is the A-weighted instantaneous sound pressure, p 0 is the reference sound pressure equal to 20 × 10−5 N/m2, and t1, t2 are the time periods [50].
The hourly traffic volume is defined by the total number of cars that pass on a road section (e.g., intersection) in a specified interval of time. In addition, the total number of cars can be divided into car categories (vehicles classes). The average speed of the cars is set at 50 km/h.
The classification of vehicles and the calculation of average travel speeds was carried out by analyzing the video streams obtained from the video camera mounted on the drone and using an artificial convolutional neural network (CNN).
The road gradient is determined by the relationship:
G r a d i e n t = a b L
where the values of parameters a, b, and L are determined via topographic measurements.
The building density was calculated using the relationship:
D e n s i t y = i = 1 n θ i θ t
where θ i are the visibility angles for the facades of the buildings in the monitored area and θ t is the maximum visibility angle (opening) of the video camera for the monitored area.
The tests were carried out in five residential areas of Bucharest, Romania, using a UAV on which was mounted an array of microphones, and the analysis was carried out at heights between 80 m and 110 m, for 15–20 min.
Figure 21 shows the result of testing the system using a sound source at a height of 100 m. Terms θ and ϕ represent the reception angles of the directional energy vectors for sound capture in relation to the horizontal and vertical movement of the drone on which the microphone array is mounted.

3.2. Collected Data Using the Artificial Neuronal Network

To carry out the data processing, a neural architecture with two analysis modules was used. The first module is the Supervised Convolutional Machine Learning (CNN), and the second module is an MFFN architecture for determining the traffic noise level. The structure is presented in Figure 22.
Considering the large amount of data associated with video information, a fusion system is proposed that uses deep neural networks to process the prediction output associated with multiple end-to-end systems, to combine all these outputs and create the set of predictions. Given the intrinsic ability of dense layers to detect patterns in the input data and to predict with high accuracy the analyzed samples, we theorize that, by using a set of dense layers, we can exploit the correlation between classifier biases, allowing different combinations of classifiers to strengthen or weaken the discriminative power of the pool of classifiers used in aggregation, based on the patterns learned by the dense network. In addition, because dense layers do not need special assumptions about the domain of the input data, we theorize that their use is also useful for creating a domain-independent aggregation system. The network architecture (Figure 23) is determined by constructing a search system designed to vary a set of dense network parameters as follows: (i) the optimal width by testing the following number of neurons per layer: {25, 50, 500, 1000, 2000, 5000}; (ii) network depth by changing the number of dense layers, testing the following values: {5, 10, 15, 20, 25}; (iii) adding or removing training subset normalization layers between dense layers.
First, there is no intrinsic spatial information associated with such an aggregate input (yi), therefore, some data decoration techniques must be developed to generate spatial correlations that can then be exploited by convolutional layers. As such, we create these correlations and relationships, through a process we call the input decorator. We theorize that by creating the decorated input vector for sample-level convolutional processing and applying convolutional filters to this new input, we can create a system where classifiers with similar discriminative powers are arranged in close spatial proximity, supporting or canceling group decisions based on their spatial relationships. Therefore, two problems need to be solved to introduce convolutional layers in aggregation networks: (i) finding a criterion for detecting similarity between classifiers and (ii) creating a similarity-based spatial organization system. Let p j = [ y 1 j ,   y 2 j , y 2 j ,   ,   y M j ]   be a prediction vector representing the outputs of classifier j for all M samples, where the similarity between two classifiers m and n can be calculated by rmn = M (pm, pn). Finally, by ordering the vector of similarity scores between a classifier m and all other classifiers, we can create a list of the most similar classifiers for each of the N classifiers. The second problem involves using the similarity values calculated in the previous step, and decorated using the given values of r, as follows:
d C i ¯ = | r 4 , 1 c 1 , 1 r 1 , N c 4 , 1 s 1   c 2 , N r 3 , 1 c 3 , 1 r 2 , N |
With the introduction of convolutional layers in the aggregation network, a method was created that can process the similarities between classifiers. However, convolutional networks have been proposed for image processing using the same filters for processing the entire image, and would, therefore, share, in the case of assembly systems, the same weights between different centroids. Although this is a step forward in classifier correlation processing, we theorize that this correlation between classifiers is different for each individual classifier, and, thus, weights should not be shared between centroids. With this assumption in mind, we propose to create a new type of DNN layer, which we call the “Cross-Space-Fusion”, or CSF layer. The implementation of the CSF layer is based on creating a new input decoration method as a new layer in the aggregation architecture.
Therefore, we propose a new input decoration method that creates an additional third dimension, which separately stores similar results of classifiers and similarity scores. Thus, the CSF layer processes these details in the third dimension of the matrix, processing the outputs of the classifiers and the similarity scores of the correspondents together, using the same M function for computing the similarity scores. We also need to quantify that regular convolutional filters may not be optimal for learning correlations, as they may be different from centroid to centroid. Thus, in the CSF layer, we design a larger number of parameters, and, although this leads to an additional depth of the neural network, the number of added parameters is still small, especially compared to the depth and width of the dense architecture. Considering these particularities, the following equation presents the operations performed by the new CSF layer:
[ α 1 , i s i + β 1 , i c 1 , i r 1 , i 2 α 2 , i s i + β 2 , i c 2 , i r 2 , i 2 α 3 , i s i + β 3 , i c 3 , i r 3 , i 2 α 8 , i s i + β 8 , i c 8 , i r 8 , i 2 S i α 4 , i s i + β 4 , i c 4 , i r 4 , i 2 α 7 , i s i + β 7 , i c 7 , i r 7 , i 2 α 6 , i s i + β 6 , i c 6 , i r 6 , i 2 α 5 , i s i + β 5 , i c 5 , i r 5 , i 2 ]
where c i represents the matrix of predicted outputs for the eight most similar classifiers for a classifier i, while R i represents the respective similarity scores, calculated using the function M. These two matrices create the third dimension of the decorated input. After the decoration stage, the input is fed into the CSF layer. For each group ( c i , r i ) of centroids, the network needs to create and learn a set of weights that combine the prediction of the original classifier with the prediction results and clustered similarity scores in the centroids. Thus, the CSF layer contains a set of parameters α and β to be learned, where parameters α are used to control the prediction output of each classifier i and parameters β are used to control the prediction outputs and similarity scores for classifiers similar to i.
The output of the CSF layer is then passed through a subsampling layer by the averaging function, yielding a new set of input data of the same size as the original. Two implementations were tested, in which pairs of contrasts were considered for the most similar four classifiers, and for the most similar eight classifiers.
To establish the performance values, it is necessary to calculate the confusion matrix. The confusion matrix consists of real values in one dimension and predicted labels for the second dimension, and each class consists of a row and a column. The diagonal elements of the matrix represent the correctly classified results.
The values calculated from the confusion matrix represent Precision, Recall, and F1-score, which is the harmonic average of the accuracy and recall and the accuracy of the classification:
Precision: it is defined as the number of samples that contain the existence of the drone.
Recall: it is defined as the ratio between the expected number of samples to contain a drone and the number of samples that contain the drone.
F-measure: it is defined as the harmonic average between accuracy and recall.
F1-scores are calculated for each class, followed by the average of the scaled scores with weights. The weights are generated from the number of samples corresponding to each class.
The network is trained for 200 epochs with a batch size of 16. This experiment resulted in a classification accuracy of 91 percent, Recall is 0.96, and micro average F1-score is 0.91. The confusion matrix is shown in Table 1 and the classification report is shown in Table 2.

3.3. Assessment of Noise Pollution in the Study Area

The evaluation of the noise levels at the measurement points, five residential areas in Bucharest, demonstrates the exceeding of the maximum permissible noise level for commercial-residential areas legislated in Romania, which is 60 dB(A), because the samples obtained showed a sound level that exceeds 75 dB(A), shown in Figure 24. Table 3 describes the statistical variables.
Analyzing the results of the L a e q analysis obtained for five residential areas in Bucharest, Romania, presented in Figure 24, it turns out that the permitted noise level is exceeded compared to the legislated level of 60 dB.
Noise pollution has increased to an alarming degree in all areas of the city (residential, commercial, industrial, and quiet) due to the rapid increase in urbanization, industrialization, and other connectivity of transport systems in all areas of the city. Identifying noise hotspots where immediate remedial action is required has always been a challenge. In addition to providing noise propagation in the X and Z directions, 2D noise mapping is an essential method for identifying regions where noise levels may reach a dangerous level. Thus, the comparative examination of all residential areas within a city provides a clear picture of noise exposure. The measurement of noise on the ground is carried out with the help of the sound level meter. In the horizontal direction, the orientation and distance of buildings from the road or flight path, as well as the shape of buildings are the main influencing factors on the noise effect. Based on the analyses carried out from ground level and from a height of 30 m, by making 2D maps, they showed that the surrounding structures (high buildings, brick walls, borders with grills, trees, etc.) significantly influence noise propagation in all directions. That is why the next step was to make 3D noise maps from a height of 100 m (e.g., road traffic and aircraft noise propagate from different directions and produce different effects on the ground floor and top floor of buildings). In particular, the buildings with the highest noise level are not close to the road or railway, and the apartments with high noise are not located close to the ground or the top of the building. The composition of the noise and the existence of multiple 3D noise sources determine the presence or absence of high noise. The high noise level in the residential area may be due to the heavy traffic volume and the road surrounded by tall buildings. 3D maps can also be used by decision makers in the process of formulating noise control strategies or implementing corrective measures.

4. Discussion

The novelties presented by the proposed solution are:
Carrying out the analysis of noise pollution from traffic with the array of microphones mounted on a UAV.
Use of the best algorithms reported in the field (energy, neural networks) for analyzing the solution.
Use of a combined neural network structure, a CNN module for detecting the number of vehicles in traffic, calculating the average travel speed, and identifying vehicles (cars, trucks, and motorcycles) and the 2nd MFFN (multilayer feed-forward) module neural network for prediction.
Proposal to create a new type of CNN layer, named “Cross-Space-Fusion”, or CSF layer. The implementation of the CSF layer is based on creating a new input decoration method as a new layer in the aggregation architecture.
Proposing a new input decoration method for the CSF network, which creates an additional third dimension that separately stores similar results of classifiers and similarity scores. Thus, the CSF layer processes these details in the third dimension of the matrix, processing the outputs of the classifiers and the similarity scores of the correspondents together.
To determine the level of noise pollution, the specialized literature only mentions measuring the ground level, and the solution proposed in the article can determine the sound level from a height (100 m), but also separate the individualized sound levels of vehicles in traffic.
In the capital city of Romania, decibel limits are exceeded in many locations, every day. Unfortunately, the action plan for reducing noise is slow, or not developing, while the residents of the capital are increasingly affected by noise pollution. The level of noise and vibrations generated by the movement of means of transport belonging to the local transport society also have an impact, among the causes of the phenomenon being worn carriageways or the operation of a carriageway whose constructive solution does not correspond to the urban characteristics of the area (e.g., narrow streets, with fronts of buildings located near the roadway). There is a need for the development of an urban noise monitoring plan and adequate solutions for decreasing noise impact. The greatest impact is felt primarily by children and elderly people.
We believe that a solution to making noise reduction measures more efficient, while optimizing costs and efforts, could be the development of combined strategies to reduce traffic noise and air pollution.

5. Conclusions

Noise monitoring and prediction can raise various complex problems determined by non-linearity, and collected data are very large. The present research has been carried out in two phases: (i) collecting relevant information from traffic with traditional instruments, at the ground level and at an altitude of 30 m above and analyzing the data to determine patterns in noise behavior, and (ii) designing and testing a UAV-borne array of microphones and collecting similar data to determine feasibility of the solution. The solution is based also on previous research experience of the authors [55]. The proposed solution presented here is intended to solve many of these problems due to the use of neural network modules for the analysis of large volumes of data for noise prediction. Moreover, neural networks can be implemented in FPGA and VLSI, being superior compared to statistical and deterministic models. As future improvements and developments, we would like to enhance the filtering methods to suppress non-informative noise coming from the UAV’s own rotors [56,57,58] and lateral side noise sources by modifying the pattern of sensitivity in the horizontal plane, according to altitude of the UAV. Another concern is to choose the correct UAV type to keep environmental impact in urban sensitive areas (hospitals, residential areas, etc.) as low as possible [59]. In addition, a second direction of improvement that we are envisaging is to increase the directivity of the microphone array, based on recent research [60,61,62,63,64,65,66], and introduce the possibility to remotely configure it, according to local geometrical and environmental conditions.

Author Contributions

Conceptualization, M.M. and C.M.D.; methodology, M.M.; software, C.M.D.; validation, M.M. and C.M.D.; formal analysis, M.M.; investigation, M.M.; resources, C.M.D.; data curation, M.M.; writing—original draft preparation, M.M. and C.M.D.; writing—review and editing, M.M.; visualization, C.M.D.; supervision, M.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used were obtained in the laboratory of Intelligent Transport Systems within the Faculty of Transport, Polytechnic University of Bucharest. The data obtained are not public.

Acknowledgments

The authors acknowledge the support of the Laboratory of Telematics and Electronics for Transport from the POLITEHNICA University in Bucharest, Romania, for the materials and measuring instruments made available, and the algorithms presented in this paper were made within the project INNOVATIVE SYSTEM FOR COMBATING CROSS-BORDER TERRORISM, ORGANIZED CRIMINALITY, ILLEGAL TRAFFICKING OF BARS AND PEOPLE—357/390033/27.09.2021, SMIS Code: 121596.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Di, H.; Liu, X.; Zhang, J.; Tong, Z.; Ji, M.; Li, F.; Feng, T.; Ma, Q. Estimation of the quality of an urban acoustic environment based on traffic noise evaluation models. Appl. Acoust. 2018, 141, 115–124. [Google Scholar] [CrossRef]
  2. Zefreh, M.M.; Torok, A. Theoretical Comparison of the Effects of Different Traffic Conditions on Urban Road Traffic Noise. J. Adv. Transp. 2018, 2018, 7949574. [Google Scholar] [CrossRef]
  3. Cussen, K.; Garruccio, S.; Kennedy, J. UAV Noise Emission—A Combined Experimental and Numerical Assessment. Acoustics 2022, 4, 297–312. [Google Scholar] [CrossRef]
  4. Fiedler, P.E.K.; Zanin, P.H.T. Evaluation of Noise Pollution in Urban Traffic Hubs–Noise Maps and Measurements. Environ. Impact Assess. Rev. 2015, 51, 1–9. [Google Scholar] [CrossRef]
  5. Doygun, H.; Gurun, D.K. Analysing and Mapping Spatial and Temporal Dynamics of Urban Traffic Noise Pollution: A Case Study in Kahramanmaraș, Turkey. Environ. Monit Assess 2008, 142, 65–72. [Google Scholar] [CrossRef]
  6. Alberola, J.; Flindell, I.H.; Bullmore, A.J. Variability in road traffic noise levels. Appl. Acoust. 2005, 66, 1180–1195. [Google Scholar]
  7. Sommerhoff, J.; Recuero, M.; Suarez, E. Community noise survey of the city Valdivia, Chile. Appl. Acoust. 2004, 65, 643–656. [Google Scholar]
  8. Stoilova, K.; Stoilov, T. Traffic noise and traffic light control. Transp. Res. Part D Transp. Environ. 1998, 3, 399–417. [Google Scholar] [CrossRef]
  9. Öhrström, E.; Skanberg, A.; Svensson, H.; Gidlöf-Gunnarsson, A. Effects of road traffic noise and the benefit of access to quietness. J. Sound Vib. 2006, 295, 40–59. [Google Scholar]
  10. Yusoff, S.; Ishak, A. Evaluation of Urban Highway Environmental Noise Pollution. Sains Malays. 2005, 34, 81–87. [Google Scholar]
  11. Yusoff, S. Study of Characteristic of Transportation Noise Sources in Klang Valley. Malays. J. East. Asia Soc. Transp. Stud. 1997, 2. [Google Scholar] [CrossRef]
  12. SCOPE 24 (Scientific Committee on Problems of the Environment). Noise Pollution: Effect and Control; Saens, L., Stephens, R.W., Eds.; John Wiley & Sons: Hoboken, NJ, USA, 1986. [Google Scholar]
  13. Ciaburro, G.; Iannace, G. Improving Smart Cities Safety Using Sound Events Detection Based on Deep Neural Network Algorithms. Informatics 2020, 7, 23. [Google Scholar] [CrossRef]
  14. Lim, M.; Lee, D.; Park, H.; Kang, Y.; Oh, J.; Park, J.-S.; Jang, G.-J.; Kim, J.-H. Convolutional Neural Network based Audio Event Classification. KSII Trans. Internet Inf. Syst. 2018, 12, 2748–2760. [Google Scholar]
  15. Kiranyaz, S.; Ince, T.; Abdeljaber, O.; Avci, O.; Gabbouj, M. 1-d Convolutional Neural Networks for Signal Processing Applications. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing, Brighton, UK, 12–17 May 2019; Institute of Electrical and Electronics Engineers: Piscataway, NJ, USA, 2019; pp. 8360–8364. [Google Scholar]
  16. Šarkan, B.; Loman, M.; Synák, F.; Skrúcaný, T.; Hanzl, J. Emissions Production by Exhaust Gases of a Road Vehicle’s Starting Depending on a Road Gradient. Sensors 2022, 22, 9896. [Google Scholar] [CrossRef]
  17. Kong, Q.; Xu, Y.; Sobieraj, I.; Wang, W.; Plumbley, M.D. Sound Event Detection and Time–Frequency Segmentation fromWeakly Labelled Data. IEEE/ACM Trans. Audio Speech Lang. Process. 2019, 27, 777–787. [Google Scholar]
  18. Luo, L.; Qin, H.; Song, X.; Wang, M.; Qiu, H.; Zhou, Z. Wireless Sensor Networks for Noise Measurement and Acoustic Event Recognitions in Urban Environments. Sensors 2020, 20, 2093. [Google Scholar] [CrossRef] [Green Version]
  19. Hakala, I.; Kivela, I.; Ihalainen, J. Design of low-cost noise measurement sensor network: Sensor function design. In Proceedings of the First International Conference on Sensor Device Technologies and Applications (SENSORDEVICES 2010), Venice, Italy, 18–25 July 2010; pp. 172–179. [Google Scholar]
  20. Zamora, W.; Calafate, C.T.; Cano, J.C. Accurate ambient noise assessment using smartphones. Sensors 2017, 17, 917. [Google Scholar] [CrossRef]
  21. Bulkin, V.; Khromulina, D.; Kirillov, N. Analysis of the Acoustically Noise Situation in an Urbanized Area. In the Presence of Vehicular and Industrial Noise. In Proceedings of the IOP Conference Series: Earth and Environmental Science, Surakarta, Indonesia, 6–7 November 2019; pp. 222–237. [Google Scholar]
  22. Sevillano, X.; Socoró, J.C.; Alías, F. DYNAMAP—Development of Low-Cost Sensors Networks for Real Time Noise Mapping. Noise Mapp. 2016, 3, 172–189. [Google Scholar] [CrossRef]
  23. Picaut, J.; Can, A.; Fortin, N.; Ardouin, J. Lagrange, Low-cost sensors for noise monitoring networks: A Review. In Proceedings of the 2020 e-Formum Acusticum, Paris, France, 7–11 December 2020. [Google Scholar] [CrossRef]
  24. Naiudomthum, S.; Winijkul, E.; Sirisubtawee, S. Near Real-Time Spatial and Temporal Distribution of Traffic Emissions in Bangkok Using Google Maps Application Program Interface. Atmosphere 2022, 13, 1803. [Google Scholar] [CrossRef]
  25. Salamon, J.; Bello, J.P. Deep convolutional neural networks and data augmentation for environmental sound classification. IEEE Signal Process. Lett. 2017, 24, 279–283. [Google Scholar]
  26. Lee, H.M.; Luo, W.; Xie, J.; Lee, H.P. Traffic Noise Reduction Strategy in a Large City and an Analysis of Its Effect. Appl. Sci. 2022, 12, 6027. [Google Scholar] [CrossRef]
  27. Mądziel, M.; Campisi, T. Investigation of Vehicular Pollutant Emissions at 4-Arm Intersections for the Improvement of Integrated Actions in the Sustainable Urban Mobility Plans (SUMPs). Sustainability 2023, 15, 1860. [Google Scholar] [CrossRef]
  28. Basner, M.; McGuire, S. WHO Environmental Noise Guidelines for the European Region: A Systematic Review on Environmental Noise and Effects on Sleep. Int. J. Environ. Res. Public Health 2018, 15, 519. [Google Scholar]
  29. Kumar, S.; Lee, H.P. The Present and Future Role of Acoustic Metamaterials for Architectural and Urban Noise Mitigations. Acoustics 2019, 1, 590–607. [Google Scholar] [CrossRef]
  30. Titu, A.M.; Boroiu, A.A.; Mihailescu, S.; Pop, A.B.; Boroiu, A. Assessment of Road Noise Pollution in Urban Residential Areas—A Case Study in Pitesti, Romania. Appl. Sci. 2022, 12, 4053. [Google Scholar]
  31. Liu, Y.; Ma, X.; Shu, L.; Yang, Q.; Zhang, Y.; Huo, Z.; Zhou, Z. Internet of Things for Noise Mapping in Smart Cities: State of the Art and Future Directions. IEEE Netw. 2020, 34, 112–118. [Google Scholar] [CrossRef]
  32. Chen, Q.; Tang, S.; Yang, Q.; Fu, S. Cooper: Cooperative Perception for Connected Autonomous Vehicles Based on 3D Point Clouds. In Proceedings of the 2019 IEEE 39th International Conference on Distributed Computing Systems (ICDCS), Dallas, TX, USA, 7–10 July 2019; pp. 514–524. [Google Scholar]
  33. Alías, F.; Alsina-Pagès, R.M. Review of Wireless Acoustic Sensor Networks for Environmental Noise Monitoring in Smart Cities. J. Sens. 2019, 2019, 7634860. [Google Scholar] [CrossRef]
  34. Socoró, J.C.; Alías, F.; Alsina-Pagès, R.M. An Anomalous Noise Events Detector for Dynamic Road Traffic Noise Mapping in Real-Life Urban and Suburban Environments. Sensors 2017, 17, 2323. [Google Scholar] [CrossRef]
  35. Grubeša, S.; Petošić, A.; Suhanek, M.; Đurek, I. Mobile crowdsensing accuracy for noise mapping in smart cities. Automatika 2018, 59, 286–293. [Google Scholar] [CrossRef]
  36. Stansfeld, S.; Clark, C.; Smuk, M.; Gallacher, J.; Babisch, W. Road traffic noise, noise sensitivity, noise annoyance, psychological and physical health and mortality. Environ. Health 2021, 20, 32. [Google Scholar] [CrossRef]
  37. Calamassi, D.; Pomponi, G.P. Music Tuned to 440 Hz Versus 432 Hz and the Health Effects: A Double-blind Cross-over Pilot Study. Explore 2019, 15, 283–290, Erratum in Explore 2020, 16, 8. [Google Scholar] [CrossRef]
  38. Zhao, X.; Song, S.J.; Kim, H.J.; Gang, T.; Kang, S.C.; Choi, Y.H.; Kang, S.S. Determination of incident angle and position of optimal mode ultrasonic beam for flaw detection in anisotropic and inhomogeneous weldments by ray tracing. J. Korean Soc. Nondestruct. Test. 2007, 27, 231–238. [Google Scholar]
  39. Fyleris, T.; Jasiuniene, E. Analytic Approach for 2D Phased Array Delay Law Calculation in Case of Inspection of Objects with Complex Geometry. Elektron. Elektrotechnika 2019, 25, 28–31. [Google Scholar]
  40. Chazan, S.E.; Hammer, H.; Hazan, G.; Goldberger, J.; Gannot, S. Multi-microphone speaker separation based on deep DOA estimation. In Proceedings of the 27th European Signal Processing Conference (EUSIPCO), A Coruna, Spain, 2–6 September 2019; pp. 1–5. [Google Scholar]
  41. Bohlender, A.; Spriet, A.; Tirry, W.; Madhu, N. Neural networks using full-band and subband spatial features for mask based source separation. In Proceedings of the 29th European Signal Processing Conference (EUSIPCO), Dublin, Ireland, 23–27 August 2021; pp. 346–350. [Google Scholar]
  42. Huang, S.C.; Chiou, C.C.; Chiang, J.T.; Wu, C.F. A novel intelligent option price forecasting and trading system by multiple kernel adaptive filters. J. Comput. Appl. Math. 2020, 369, 112560. [Google Scholar]
  43. Chen, K.; Werner, S.; Kuh, A.; Huang, Y.F. Nonlinear adaptive filtering with kernel set-membership approach. IEEE Trans. Signal Process. 2020, 68, 1515–1528. [Google Scholar]
  44. Li, L.; Zhao, H.; Lv, S. Diffusion recursive total least square algorithm over adaptive networks and performance analysis. Signal Process. 2021, 182, 107954. [Google Scholar] [CrossRef]
  45. Fan, Y.; Lin, L.; Ying, L.; Zepeda-Núnez, L. A multiscale neural network based on hierarchical matrices. Multiscale Model. Simul. 2019, 17, 1189–1213. [Google Scholar] [CrossRef]
  46. Zhang, Z.; Xu, Y.; Yu, M.; Zhang, S.-X.; Chen, L.; Yu, D. ADLMVDR: All deep learning MVDR beamformer for target speech separation. In Proceedings of the ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021. [Google Scholar]
  47. Candès, E.J.; Romberg, J.; Tao, T. Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information. IEEE Trans. Inf. Theory 2006, 52, 489–509. [Google Scholar]
  48. Johnson, D.H.; Dudgeon, D.E. Array Signal Processing: Concepts and Techniques; P T R Prentice Hall: Hoboken, NJ, USA, 1993. [Google Scholar]
  49. Trees, H.L.V. Detection, Estimation, and Modulation Theory, Optimum Array Processing; Part IV ed2; Wiley Library-Interscience: New York, NY, USA, 2002. [Google Scholar]
  50. Benesty, J.; Huang, Y. (Eds.) Adaptive Signal Processing–Applications to Real-World Problems; Springer: Berlin, Germany, 2003. [Google Scholar]
  51. Diniz, P.S.R. Adaptive Filtering: Algorithms and Practical Implementation, 4th ed.; Springer: New York, NY, USA, 2013. [Google Scholar]
  52. Benesty, J.; Cohen, I.; Chen, J. Array Processing–Kronecker Product Beamforming; Springer: Cham, Switzerland, 2019. [Google Scholar]
  53. Zhao, H.; Hu, Y.; Sun, H.; Feng, W. A BDS Interference Suppression Technique Based on Linear Phase Adaptive IIR Notch Filters. Sensors 2018, 18, 1515. [Google Scholar]
  54. Widrow, B.; McCool, J.M.; Larimore, M.G.; Johnson, C.R. Stationary and nonstationary learning characteristics of the LMS adaptive filter. Proc. IEEE 1976, 64, 1151–1162. [Google Scholar]
  55. Haykin, S. Neural Networks: A Comprehensive Foundation; Prentice-Hall: Hoboken, NJ, USA, 1999. [Google Scholar]
  56. Hamad, K.; Khalil, M.; Shanableh, A. Modeling roadway traffic noise in a hot climate using artificial neural networks. Transp. Res. Part D Transp. Environ. 2017, 53, 161–177. [Google Scholar] [CrossRef]
  57. WHO. Environmental Noise Guidelines for the European Region; WHO: Copenhagen, Denmark, 2018.
  58. Banerjee, D.; Chakraborty, S.K.; Bhattacharyya, S.; Gangopadhyay, A. Evaluation and Analysis of Road Traffic Noise in Asansol: An Industrial Town of Eastern India. Int. J. Environ. Res. Public Health 2008, 5, 165–171. [Google Scholar]
  59. Vilas, P.; Prashant, N.P. Measurement and Analysis of Noise at Signalised Intersections. J. Environ. Res. Dev. 2015, 9, 662. [Google Scholar]
  60. Bastian-Monarca, N.; Suarez, E.; Arenas, J. Assessment of methods for simplified traffic noise mapping of small cities: Casework of the city of Valdivia, Chile. Sci. Total Environ. 2016, 550, 439–448. [Google Scholar] [CrossRef]
  61. Fyhri, A.; Klboe, R. Road traffic noise, sensitivity, annoyance and self-reported health—A structural equation model exercise. Environ. Int. 2009, 35, 91–97. [Google Scholar]
  62. Dumitrescu, C.; Minea, M.; Costea, I.M.; Cosmin Chiva, I.; Semenescu, A. Development of an Acoustic System for UAV Detection. Sensors 2020, 20, 4870. [Google Scholar] [CrossRef]
  63. Mesquita, G.P.; Mulero-Pázmány, M.; Wich, S.A.; Rodríguez-Teijeiro, J.D. Terrestrial Megafauna Response to Drone Noise Levels in Ex Situ Areas. Drones 2022, 6, 333. [Google Scholar] [CrossRef]
  64. Duporge, I.; Spiegel, M.P.; Thomson, E.R.; Chapman, T.; Lamberth, C.; Pond, C.; Macdonald, D.W.; Wang, T.; Klinck, H. Determination of optimal flight altitude to minimise acoustic drone disturbance to wildlife using species audiograms. Methods Ecol. Evol. 2021, 12, 2196–2207. [Google Scholar]
  65. Torija, A.J.; Nicholls, R.K. Investigation of Metrics for Assessing Human Response to Drone Noise. Int. J. Environ. Res. Public Health 2022, 19, 3152. [Google Scholar] [CrossRef]
  66. Jekosch, S.; Sarradj, E. An Inverse Microphone Array Method for the Estimation of a Rotating Source Directivity. Acoustics 2021, 3, 462–472. [Google Scholar] [CrossRef]
Figure 1. (a) The A-weighted curve for sound data compensation; (b) conversion from dB SPL to dBA using A-weighting adjustments.
Figure 1. (a) The A-weighted curve for sound data compensation; (b) conversion from dB SPL to dBA using A-weighting adjustments.
Sensors 23 01912 g001aSensors 23 01912 g001b
Figure 2. Placement of ground-level measurement point.
Figure 2. Placement of ground-level measurement point.
Sensors 23 01912 g002
Figure 3. SPL versus frequency recorded in traffic-generated noise measurements, ground level.
Figure 3. SPL versus frequency recorded in traffic-generated noise measurements, ground level.
Sensors 23 01912 g003
Figure 4. SPL variation recorded in lower frequency range (50 Hz, 80 Hz, and 125 Hz traffic noise), ground level: blue—50 Hz, orange—80 Hz, and grey—125 Hz.
Figure 4. SPL variation recorded in lower frequency range (50 Hz, 80 Hz, and 125 Hz traffic noise), ground level: blue—50 Hz, orange—80 Hz, and grey—125 Hz.
Sensors 23 01912 g004
Figure 5. Recorded SPL at medium frequencies (630 Hz, 800 Hz, and 1.6 kHz traffic noise, in blue—630 Hz, orange—800 Hz, grey—1.6 kHz), ground level.
Figure 5. Recorded SPL at medium frequencies (630 Hz, 800 Hz, and 1.6 kHz traffic noise, in blue—630 Hz, orange—800 Hz, grey—1.6 kHz), ground level.
Sensors 23 01912 g005
Figure 6. SPL recorded for higher frequency range, ground level (at 8 kHz, 10 kHz and 12.5 kHz, with blue—8 kHz, orange—10 kHz, and grey—12.5 kHz components).
Figure 6. SPL recorded for higher frequency range, ground level (at 8 kHz, 10 kHz and 12.5 kHz, with blue—8 kHz, orange—10 kHz, and grey—12.5 kHz components).
Sensors 23 01912 g006
Figure 7. SPL versus frequency recorded in traffic-generated noise measurements, 30 m above ground.
Figure 7. SPL versus frequency recorded in traffic-generated noise measurements, 30 m above ground.
Sensors 23 01912 g007
Figure 8. SPL recorded for higher frequency range, ground level (at 50 Hz, 80 Hz, and 125 Hz, with blue—50 Hz, orange—80 Hz, and grey—125 Hz components).
Figure 8. SPL recorded for higher frequency range, ground level (at 50 Hz, 80 Hz, and 125 Hz, with blue—50 Hz, orange—80 Hz, and grey—125 Hz components).
Sensors 23 01912 g008
Figure 9. SPL recorded for medium frequency range, 30 m level (630 Hz, 800 Hz, and 1.6 kHz traffic noise, in blue—630 Hz, orange—800 Hz, grey—1.6 kHz).
Figure 9. SPL recorded for medium frequency range, 30 m level (630 Hz, 800 Hz, and 1.6 kHz traffic noise, in blue—630 Hz, orange—800 Hz, grey—1.6 kHz).
Sensors 23 01912 g009
Figure 10. SPL recorded for medium frequency range, 30 m level (at 8 kHz, 10 kHz and 12.5 kHz, with blue—8 kHz, orange—10 kHz, and grey—12.5 kHz components).
Figure 10. SPL recorded for medium frequency range, 30 m level (at 8 kHz, 10 kHz and 12.5 kHz, with blue—8 kHz, orange—10 kHz, and grey—12.5 kHz components).
Sensors 23 01912 g010
Figure 11. Architecture of the ultrasonic sound localization system using a drone.
Figure 11. Architecture of the ultrasonic sound localization system using a drone.
Sensors 23 01912 g011
Figure 12. Simulation square microphone array and beamforming using acoustic frequency characteristic of UAVs (a) and directivity characteristic (b).
Figure 12. Simulation square microphone array and beamforming using acoustic frequency characteristic of UAVs (a) and directivity characteristic (b).
Sensors 23 01912 g012
Figure 13. The acoustic processing system.
Figure 13. The acoustic processing system.
Sensors 23 01912 g013
Figure 14. Convert from PDM to PCM.
Figure 14. Convert from PDM to PCM.
Sensors 23 01912 g014
Figure 15. DAS beamforming algorithm working principle.
Figure 15. DAS beamforming algorithm working principle.
Sensors 23 01912 g015
Figure 16. The method of realization and implementation for Equation (5).
Figure 16. The method of realization and implementation for Equation (5).
Sensors 23 01912 g016
Figure 17. The system model.
Figure 17. The system model.
Sensors 23 01912 g017
Figure 18. Adaptive noise cancellation theoretical configuration.
Figure 18. Adaptive noise cancellation theoretical configuration.
Sensors 23 01912 g018
Figure 19. ANC simulation result with optimized LMS algorithm.
Figure 19. ANC simulation result with optimized LMS algorithm.
Sensors 23 01912 g019
Figure 20. The proposed MFFC architecture for traffic noise analysis.
Figure 20. The proposed MFFC architecture for traffic noise analysis.
Sensors 23 01912 g020
Figure 21. Result of testing the system for capturing sounds from a height of 100 m.
Figure 21. Result of testing the system for capturing sounds from a height of 100 m.
Sensors 23 01912 g021
Figure 22. The neural architecture proposed.
Figure 22. The neural architecture proposed.
Sensors 23 01912 g022
Figure 23. Cross-Space-Fusion aggregation network diagram.
Figure 23. Cross-Space-Fusion aggregation network diagram.
Sensors 23 01912 g023
Figure 24. Noise pollution levels determined for five residential areas in Bucharest, Romania.
Figure 24. Noise pollution levels determined for five residential areas in Bucharest, Romania.
Sensors 23 01912 g024aSensors 23 01912 g024bSensors 23 01912 g024c
Table 1. Confusion matrix.
Table 1. Confusion matrix.
Predict Label
Background NoiseSingle CarTwo Cars
True LabelBackground Noise1.000.000.00
Single Car0.000.940.06
Two Cars0.000.230.77
Table 2. Classification report.
Table 2. Classification report.
ClassesPrecisionRecallF1-Score
Background Noise111
Single Car0.800.940.87
Two Cars0.930.770.84
Avg/total0.910.900.90
Table 3. Descriptive statistics of variables.
Table 3. Descriptive statistics of variables.
Traffic VolumeAverage SpeedNo. of MotorNo. of TrucksNo. of CarsGDLaeq
Mean4556.7859.3675.41224.14146.56322.9920.7574.669
Standard error274.782.5870.2700.1450.830.4350.0450.426
2106.319.971.9451.0371.6472.0470.2693.952
Min.130224.4370.7590.3671.4480.04074.57
Max.9759.594.9788.8264.5239.56590.9987.38
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Minea, M.; Dumitrescu, C.M. Urban Traffic Noise Analysis Using UAV-Based Array of Microphones. Sensors 2023, 23, 1912. https://doi.org/10.3390/s23041912

AMA Style

Minea M, Dumitrescu CM. Urban Traffic Noise Analysis Using UAV-Based Array of Microphones. Sensors. 2023; 23(4):1912. https://doi.org/10.3390/s23041912

Chicago/Turabian Style

Minea, Marius, and Cătălin Marian Dumitrescu. 2023. "Urban Traffic Noise Analysis Using UAV-Based Array of Microphones" Sensors 23, no. 4: 1912. https://doi.org/10.3390/s23041912

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop