Next Article in Journal
Use of Auditory Cues and Other Strategies as Sources of Spatial Information for People with Visual Impairment When Navigating Unfamiliar Environments
Next Article in Special Issue
Requirements for Drone Operations to Minimise Community Noise Impact
Previous Article in Journal
Cross-Country Adaptation of a Psychological Flexibility Measure: The Comprehensive Assessment of Acceptance and Commitment Therapy Processes
Previous Article in Special Issue
Assessment of Sound Source Tracking Using Multiple Drones Equipped with Multiple Microphone Arrays
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Investigation of Metrics for Assessing Human Response to Drone Noise

by
Antonio J. Torija
* and
Rory K. Nicholls
Acoustics Research Centre, University of Salford, Manchester M5 4WT, UK
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2022, 19(6), 3152; https://doi.org/10.3390/ijerph19063152
Submission received: 5 February 2022 / Revised: 1 March 2022 / Accepted: 3 March 2022 / Published: 8 March 2022

Abstract

:
Novel electric air transportation is emerging as an industry that could help to improve the lives of people living in both metropolitan and rural areas through integration into infrastructure and services. However, as this new resource of accessibility increases in momentum, the need to investigate any potential adverse health impacts on the public becomes paramount. This paper details research investigating the effectiveness of available noise metrics and sound quality metrics (SQMs) for assessing perception of drone noise. A subjective experiment was undertaken to gather data on human response to a comprehensive set of drone sounds and to investigate the relationship between perceived annoyance, perceived loudness and perceived pitch and key psychoacoustic factors. Based on statistical analyses, subjective models were obtained for perceived annoyance, loudness and pitch of drone noise. These models provide understanding on key psychoacoustic features to consider in decision making in order to mitigate the impact of drone noise. For the drone sounds tested in this paper, the main contributors to perceived annoyance are perceived noise level (PNL) and sharpness; for perceived loudness are PNL and fluctuation strength; and for perceived pitch are sharpness, roughness and Aures tonality. Responses for the drone sounds tested were found to be highly sensitive to the distance between drone and receiver, measured in terms of height above ground level (HAGL). All these findings could inform the optimisation of drone operating conditions in order to mitigate community noise.

1. Introduction

A scenario with several drones (whether manned or unmanned) flying over cities and rural areas is now more likely than ever. To start with, there is a myriad of potential uses, from recreational to parcel delivery and even surveillance and law enforcement. There are substantial environmental and societal benefits associated with the wider expansion of the drone sector. For instance, medical deliveries to reduce waiting times [1] or reduction of carbon footprint in cargo transport and parcel delivery [2]. However, there are important concerns that can act as barriers for the wider adoption of these technologies: safety and privacy concerns, airspace management and visual and noise impact [3]. Although the main focus to date has been on the effects on human health, drone noise is also a source of concern for animal welfare [4].
The noise emission of drones and isolated drone propellers have been extensively studied [5,6,7,8,9,10,11,12,13,14]. Hui et al. [3] outlined the main sound generation mechanisms in drones, highlighting the contributions of rotor noise and electric motors. Schäffer et al. [15] conducted a systematic review (based on the PRISMA statement) on current methods for acoustic measurements and noise emission characteristics of drones; they also stated the requirements for a drone emission model and proposed a scheme for data acquisition. Overall, the sound emission of drones was found to be mainly influenced by drone size, configuration (number of rotors), payload, operating conditions and flight manoeuvres.
Schäffer et al. [15] carried out a systematic review on the effects of drone noise on humans, finding that the literature on the topic is still very limited. Christian and Cabell’s work [16] was pioneering in understanding the perception of drone noise and how it compared to other sources of transportation, such as road traffic. Gwak et al. [17] investigated the reported annoyance (in a subjective, laboratory-based experiment) to a series of drones of different size and under hover conditions. Hui et al. [3] investigated the perception (in a subjective, laboratory-based experiment) of a series of drone sounds, accounting for different operating conditions in terms of flight mode and height above the ground. Ivošević et al. [18] carried out a survey where a series of participants reported their in situ subjective assessment on drone noise during a series of drone operations in an open field.
Current evidence suggests that drone noise annoyance highly depends on loudness-related metrics [3,15,16]. However, Hui et al. [3] suggested that further research is needed to understand and quantify the effect of spectral and temporal factors (including tonality and impulsiveness) on drone noise annoyance. Schäffer et al. [15] also point out pure tones and high-frequency broadband noise as important contributors to drone noise annoyance. Tonal and high-frequency drone noise is also likely to increase its noticeability in existing soundscapes [19,20,21] and therefore lead to noise annoyance. None of these sound features are appropriately accounted for in current aircraft noise metrics, neither LAeq-related nor even more sophisticated metrics, such as the effective perceived noise level (EPNL) [22]. Christian and Cabell [16] found that the A-weighted sound exposure level (LAE) was not able to account on its own for the difference in annoyance between drones and road vehicles. In their study, for a same value of LAE, drones were found to be more annoying than road vehicles. In fact, drones were reported to be as annoying as road vehicles with a 5.6 dB higher LAE. Torija et al. [21] also found that the LAeq metric does not account for the particular sound features related to drones, and that these particular features highly influence drone noise reported annoyance. Current regulation of drone noise is solely based on A-weighted sound power level (LwA,max), which, as suggested by Torija and Clark [23] might not provide an accurate picture of the noise impact of drones.
The aim of this paper is to investigate the perception of drone noise under controlled laboratory conditions in order to propose noise metrics for effectively assessing human response to drone noise. This research is framed within the perception-driven engineering approach, where sound quality metrics (SQMs) provide an accurate assessment of how the human auditory system responds to key sound features [23,24]. Gwak et al. [17] found that the noise annoyance of three drones (with maximum take-off mass (MTOM) ranging from 113.5 g to 11 kg) hovering was highly related to the SQMs loudness (perception of amplitude of the sound), sharpness (perception of high frequency) and fluctuation strength (perception of slow amplitude modulation). Torija and Li [25] found the noise annoyance of a series of drone (MTOM of 1.2 kg) flyovers highly related to a tonality metric and the interaction between loudness and sharpness. The research presented in this paper expands the number of drones under investigation to eight types, with payloads ranging from 1.2 to 11.8 kg and differing rotor number (including a contra-rotating configuration). Moreover, different manoeuvres are considered, including take-off, hover, flyover and landing, and varying heights above the ground are analysed. This comprehensive set of drone sounds aims to provide robust metric–noise-perception relationships for a representative sample of drone types. The main contributions of this paper are: (i) a better understanding of the noise perception of a comprehensive range of drone types and operating conditions; (ii) an assessment of the contribution of key acoustic and psychoacoustic features to drone noise perception in terms of perceived annoyance, perceived loudness and perceived pitch; and (iii) a quantification of changes in drone noise perception as a function of distance between drone and receiver. These findings can contribute to the development of noise metrics for assessing human response to drone noise and the definition of operational constrains in terms of distance to the receiver in order to minimise the community noise impact of drones.
The structure of the paper is as follows. Section 2 provides a brief overview of the main metrics for aircraft noise and SQMs. Section 3 describes both the selection of drone sounds used in this research and the methodology behind the subjective experiment investigating the relationship between drone noise and perception. Results are presented in Section 4 and include the relationship between drone noise perception and flyover altitude, effect of loudness on drone noise perception and metrics for drone noise assessment. Finally, results are discussed in Section 5, and conclusions and future work are stated in Section 6.

2. Overview of Noise Metrics

The impact of aircraft noise on communities is mainly assessed with exposure metrics related to A-weighted energy equivalent sound pressure level (LAeq) [22,26]. Examples of these metrics are day–night level (DNL), day–evening–night level (DENL) and LAeq,16h. At a vehicle level, broadband frequency-weighted sound pressure levels are used for aircraft noise assessment. These metrics include the maximum A-weighted sound level (LA,max) and the A-weighted sound exposure level (LAE). LA,max is a common metric to assess sleep disturbance [27]. LAE is numerically equivalent to the total sound energy of an aircraft overflight and therefore is very useful to compare total emission of different types of aircraft under different operating and payload conditions. However, these metrics do not account for other important features for noise annoyance, such as tonality [28,29]. The effective perceived noise level (EPNL) is the main metric for the noise certification of fixed-wing and rotary-wing aircraft [30] and is calculated according to a procedure developed by the Federal Aviation Administration [31]. This metric is based on the calculation of the perceived noise level (PNL) as proposed by Kryter [32]. The PNL is a descriptor for the overall perceived loudness, which is based on the Noy scale derived from a combination of amplitude and frequency [23]. The PNL is then corrected by an exposure duration factor and a tonality factor to obtain the EPNL metric. The tonality factor for the EPNL metric is based solely on the level of the strongest protruding tone, and therefore, as suggested by Torija et al. [22], might not be able to account for the perceptual effects of complex tonality (due to blade-passing frequency and harmonics of diverse rotors) typical in drones. Another limitation of the EPNL metric for assessing drone noise is that it does not consider frequency content above 10 kHz (which might be present in some drone types due to the operational frequency range of their electric motors) [7].
SQMs provide an accurate representation of huma hearing perception [15,33]. The most widely used SQMs are loudness (measured in sone), sharpness (measured in acum), fluctuation strength (measured in vacil), roughness (measured in asper), tonality (measured in tonality units—TU), and impulsiveness (measured in impulsiveness units—IU). Loudness measures the perception of sound intensity. Sharpness assesses the perception of spectral imbalances of a given sound towards the high-frequency region. Fluctuation strength and roughness assess the perception of slow and rapid fluctuations of the sound level. Tonality accounts for the perception of spectral irregularities of pure tones. Impulsiveness assesses the perception of sudden, abrupt increases in the sound level. All these metrics combined reflect the perception of sounds with various acoustic characteristics [15]. Further details on these SQMs can be found in Zwicker and Fastl [24] and Sottek et al. [34].
In addition to drone noise, as described in Section 1, SQMs have been used to develop noise-annoyance models for different aircraft types. For instance, Rizzi et al. [35] developed a model based on loudness, roughness and tonality to estimate noise annoyance for electric fixed-wing aircraft. Moreover, Krishnamurthy et al. [36] and Boucher et al. [37] found the SQMs sharpness, fluctuation strength and tonality as main contributors explaining noise annoyance for rotorcraft. In a study involving a psychoacoustic analysis of contra-rotating propellers, Torija et al. [33] suggested that loudness, tonality and fluctuation strength metrics are able to account for the perception of potential field interaction tones (dominating the sound emission with the rotor closely spaced). They also suggested that roughness and impulsiveness metrics are able to account for the perception of unsteadiness due to propeller–turbulence interaction noise.

3. Materials and Methods

3.1. Selection of Drone Sounds

A database of 44 drone sounds were carefully selected for the subjective experiment in order to assess perceived annoyance, loudness and pitch. The criteria for selection was to include sounds encompassing a wide variety of loudness level and other key psychoacoustic parameters for drone noise, including temporal and frequency characteristics (see [23]). The drone sounds were gathered from different sources, including in-house measurements [38], colleagues at the Volpe Center in the US [39] and colleagues from industry. Although the sounds were from various sources, they were selected to maintain a constant level of audio quality (as assessed by the authors of this paper). Each of the original sounds was edited to extract a sample of 4 sec long to use in the subjective experiment (to balance the gathering of perceptual data with participants’ fatigue [22]).
In total, sounds from eight types of drones were used, with the weight of these drones ranging from 1 to 12 kg. The chosen drone sounds also yielded a large variety of noise and operational characteristics. The drone operations included flyovers, hovering, manoeuvring, take-offs and landing. Furthermore, the drones were recorded performing these operations at differing heights above ground level (HAGL) of 2 to 60 m. The LAeq,4s of the drone sounds ranged between 37 dB and 71 dB. The full list of the 44 drone sounds with associated characteristics can be seen in Table 1. Differences in LAeq,4s between sounds from the same drone type with identical weight and HAGL might be attributed to small variations in operating and meteorological conditions, such as small differences in vehicle speed and different rotor rotational speeds to maintain vehicle stability under different wind profiles. It should also be noted that the DJI Phantom 3 was tested with varying payloads (see Torija et al. [38]).
As mentioned above, the drone sounds described in Table 1 were gathered from three different databases. Sounds S1 to S15 were recorded with a TASCAM DR-05 audio recorder, with sound pressure levels measured with a Norsonic 140 Class 1 sound level meter. These drone sounds were recorded in an open field in Alnmouth (northeast England). There were some other sounds present, including distant waves, birdsong and intermittent railway noise. Sounds S16 to S30 were recorded with a Brüel & Kjær 2250 Class 1 sound level meter with sound-recording capabilities. These drone sounds were recorded in an open field in Southampton. There were some other sounds present, including birdsong and a distant road. For further details see Torija et al. [38]. Sounds S31 to S44 were recorded by colleagues of the John A. Volpe National Transportation Systems Center in the Choctaw Nation of Oklahoma. Drone sounds were recorded using GRAS Model 40AO ½ inch pressure microphones and a Sound Devices 744T digital audio recorder. Sound pressure levels were measured with a Larson–Davis 831 Class 1 sound level meter. Recordings took place in a remote and quiet open field, with ambient sound mainly dominated by wind noise and some occasional aircraft flybys. For further details see Read et al. [39]. The ambient sound levels in all locations were considered sufficiently low so that they would not unduly influence the drone sound recordings. During the selection process, the databases available were carefully explored to discard any extraneous sounds.
Temporal and spectral characteristics of drone noise have been found to be highly influenced by the type of vehicle operation and meteorological conditions [15,23]. As the drone operates, adjustments in rotor operating conditions are made to maintain vehicle stability or to propel the vehicle in a given direction at a given speed (with a specific yaw, pitch and roll). These adjustments have been shown to significantly change the sound character [38], leading potentially to important changes in drone-noise perception. Furthermore, as described above, the operational HAGL was varied, as the drone noise spectrum contains higher frequencies than those of conventional aircraft. This high-frequency content is contributed to by harmonics of the rotor-blade-passing frequencies and the electric motors [5]. This, coupled with the decreased effects of air absorption due to the reduced operational HAGL (compared to conventional aircraft), increases the prominence of high-frequency content in drone noise and therefore should be considered in the subjective experiment.

3.2. Subjective Experiment

3.2.1. Calibration of Test Stimuli

In order to set the sound pressure level of the drone sounds used in the subjective experiment to the LAeq,4s targets, a calibration process was designed. The control of the LAeq,4s of the test sounds is especially important when assessing the effect of operational factors, such as type of operation, vehicle weight and HAGL (and consequently loudness) on perception responses. The calibration setup included a class 1 BSWA308 sound level meter, a Bruel and Kjaer head and torso system (HATS), Norsonic front end type 336 microphone pre-amp, M-Audio M-Track 2X2M audio interface, AKG k 501 headphones, Audio Quest Dragonfly Red USB sound cards (24 bits 96 kHz) and a mainstream laptop for stimuli playback and recording. The calibration setup is presented in Figure 1.
To calibrate the drone-sound stimuli to the target LAeq,4s (shown in Table 1), a 1 kHz sine wave was played from the laptop through the headphones and recorded by the microphones in the HATS system. The level of the sine wave was measured using the sound level meter plugged in to the output of the microphone pre-amp. The difference between the measured level of the microphone pre-amp output and the original sine wave level (94 dB) was calculated to be the calibration factor to apply to the drone-sound stimuli. The calibration factor was applied to the drone-sound stimuli, and the stimuli were recorded through the HATS system so that SQM (and other noise metrics used)-analysis of the test sounds is representative of the stimuli as listened to by the participants. As the experiment was conducted online and not in a lab as originally planned (see Section 3.2.2), it was assumed that the sounds recorded through the calibration system described above were representative of the sounds heard by the participants (see Section 4.1 and Section 5 for further discussion).

3.2.2. Experimental Procedure

Due to the COVID-19 pandemic and the restrictions put in place by the UK Government to mitigate its effects, a laboratory-based experiment was not possible to carry out. Instead, an online experiment was designed and built using the Web Audio Evaluation Toolkit (WAET) [40]. An interface was designed for the participants of the experiment in order to allow them to listen to the 44 drone-sound stimuli and provide their responses of perceived annoyance, loudness and pitch. The online experiment was designed to be completed in about 20 min (in order to maximise responses and completions of the full experiment).
The online experiment was accessible via personalised URL links in order to maintain anonymity and security between participant data. The online experiment was advertised on social media and to the staff and students at the University of Salford. Each person interested in participating was provided with a personalized URL link and a participant ID. Overall, 89 participants completed the online experiment in part, with 49 of them completing the full test (32 males and 17 females). Therefore, the responses of the 49 participants who completed the full test have been used for the analysis of this paper. The participants were instructed to complete the test in a quiet, distraction-free environment and to use high-quality headphones.
Each drone sound was presented individually to the participants. Once each drone sound was presented, the participants could listen to it as many times as required. Responses were then given using a set of sliders in the interface. Once the participants were satisfied with their responses, they could progress to the next stimulus, until the whole set of 44 test sounds were heard and assessed. The order of the stimuli for each participant was randomised.
Prior to the commencement of the experiment, each participant went through a pseudo-calibration stage in order to adjust the level of the UAV stimuli. Since the experiment was online and accessed remotely by the participants, the playback hardware used by each participant was unknown and highly likely to vary. This would lead to a variance in the playback quality and level of the stimuli between participants. To try to counter this, the participants were presented with the loudest and quietest UAV stimuli from the experiment and asked to adjust their playback volume so that the loudest stimulus was at a comfortable level and the quietest stimulus was still audible. Once the participant had appropriately adjusted their system playback level, they were asked to not adjust it for the remainder of the experiment. In addition to this, before starting the experiment, the participants were asked to match the sound levels of a series of tones in order to understand their frequency sensitivity (and also to detect substantial anomalies in the frequency response of the headphones used). A channel checking stage was also included to ensure that stereo playback was used. All 49 participants considered in this research paper appropriately passed these pseudo-calibration stages. In order to analyse the consistency of participants’ responses and the potential effect of using different sound reproduction settings (due to the experiment to be online), a statistical analysis was carried out (see Section 3.2.3 and Section 4.1 for more details).
The response variables considered were perceived annoyance, perceived loudness and perceived pitch. These response variables were chosen to be included in the subjective experiment as they relate to the amplitude of the sound event as well as various spectral and temporal characteristics of drone noise that have been shown to influence perception [15,16,21,23,25,33,38]. Perceived loudness was chosen as it is assumed to be a suitable response metric for explaining the effect of the distance of drone operation on perceived response. Perceived pitch was chosen as it is assumed to be a suitable response metric for explaining the effect of drone noise frequency content on perception. The questionnaire was designed according to the multi-dimensional scaling technique (MDS), which is based on dissimilarity ratings (see Susini et al. [41]). A continuous scale (from 0 to 1) was used for each subjective variable, labelled as follows: ‘Not Annoying’ at the left end and ‘Highly Annoying’ at the right end (perceived annoyance); ‘Not Loud’ at the left end and ‘Highly Loud’ at the right end (perceived loudness); and ‘Low Pitch’ at the left end and ‘High Pitch’ at the right end (perceived pitch).
The experiment reported in this paper is the first of two experiments carried out consecutively. The first experiment reported here, which focused on individual and isolated drone operations, investigated the influence of drone noise spectral and temporal characteristics on perception by analysing and assessing the relationships between participants’ responses and a selection of noise metrics (including SQMs). The second experiment (not reported here) will investigate the effects of drones on the soundscape in which they are operating.

3.2.3. Data Analysis

To quantify the spectral and temporal characteristics of the drone-sound stimuli, the HEAD Acoustics ArtemiS Suite 12.5 software was used to calculate a series of SQMs (as described in Section 2). Loudness was calculated according to DIN45631/A1 [42]. This calculation method is based on Zwicker’s loudness model and includes a modification for time-varying signals. The calculation of sharpness was made according to the Aures method [43] due to the observably large variance in the loudness of the stimuli. Tonality was calculated according to the Aures/Terhardt tonality model [44]. Roughness, fluctuation strength and impulsiveness were calculated following the methods derived by Sottek [34,45]. For the calculation of the SQMs described above, the first 0.5 sec of each sound stimulus was omitted in order to remove any potential transient effect in the sound file that had resulted from editing the stimuli to 4 sec samples. As described by Torija et al. [33], the 5th percentile of the SQMs were used for the statistical analysis to investigate the perception of the drone noise samples tested in this paper. The PNL metric was calculated according to Kryter’s model [32] (see Section 2) with code developed in-house.
The statistical analyses were carried out with the IBM SPSS v.25 statistics software. A correlation analysis (including partial correlation) was implemented to give an initial insight into the relationship between each SQM and PNL and the perceived responses of annoyance, loudness and pitch for each drone sound stimulus. A multiple linear regression (MLR) analysis was also carried out to further determine the main contributors for the perceptual variables assessed in this paper. A forward stepwise-regression method (entry criterion for F-value ≤ 0.01) was implemented.
As the experiment was carried out online, with the participants reproducing the test sounds with different reproduction settings, two statistical tests were implemented to assess consistency between participant responses: (i) a Kendall’s W test to investigate the concordance in perceived annoyance, loudness and pitch for each drone sound between participants (see Section 4.1); and (ii) a multilevel analysis to identify the significance of subject-dependent responses and assess consistency in the perceived annoyance, loudness and pitch for the stimuli tested. The multilevel analysis was carried out according to Boucher et al. [37], with pooling of data between subjects. Pooling by subject creates a partial-pooling methodology and assumes normal distribution across subjects.

4. Results

4.1. Analysis of Consistency between Participant Responses

As discussed by Torija and Flindell [46], in experiments involving participants assessing sound stimuli it is expected that there would be a certain degree of variability in the participants’ responses. For this research, the experiment was completed online, with participants using different sound reproduction systems, and, therefore, a question on consistency of the responses might arise. To investigate inter-participant variability and agreement among the different participants in their responses of perceived annoyance, loudness and pitch for each of the test sounds, a statistical analysis was conducted. In the first step, it was found that the participants’ responses of perceived annoyance, loudness and pitch did not follow a normal distribution (based on both Kolmogorov–Smirnov and Shapiro–Wilk tests). A non-parametric, k-related sample statistic, Kendall’s W, was calculated for each perceptual variable, accounting for the whole set of 44 test sounds. Monte Carlo bootstrapping with 10,000 samples was implemented to ensure a robust calculation of the p-values. As can be seen in Table 2, very good agreement was found between participants in the responses of perceived annoyance and loudness for each of the test sounds (Kendall’s W ≥ 0.6, with W = 0 meaning no agreement among responses). Good agreement was also found for perceived pitch, although with a Kendall’s W value of about 0.4. As shown in Table 2, the p-values are smaller than 0.01, which allows certainty in rejecting the null hypothesis of no agreement between participants’ responses.
The coefficient of variation (CV) was also calculated as the standard deviation divided by the mean in order to check consistency in participants’ responses. As can be seen in Table 3, the CV for the three perceptual variables and for each test sound is consistently about 0.2–0.4 (which could be assumed to be an acceptable level). There are only a few test sounds where the CV increases notably. After further exploration, the CV of perceived annoyance and loudness reach higher values for test sounds S6, S12, S36 and S37, which all correspond to flyovers at high HAGL (see Table 1). For these sounds, with HAGL about 46 and 60 m, the reduction in loudness might make the contribution of other psychoacoustic factors more significant to the subjective responses, leading to less agreement in responses. The cases with the higher value of CV for perceived pitch correspond to sounds S41–S44, all from the same drone type (Gryphon GD28X). This particular drone model is based on a contra-rotating propulsion system. The spectral and temporal characteristics of this drone, with its overlapping propellers, seemed to lead to less agreement in participants’ responses of perceived pitch (although further research will be needed to fully understand the reasons behind this finding).

4.2. Peceived Annoyance, Loudness and Pitch as a Function of Height above Ground Level

The aggregated participant responses for each flyover drone sound were used to investigate the relationship between perceived annoyance, loudness and pitch with HAGL. As it can be seen in Figure 2 (perceived annoyance), Figure 3 (perceived loudness) and Figure 4 (perceived pitch), there is a strong logarithmic correlation between the three perceptual variables and HAGL. Although with some variability above 30 m, there is a clear trend of lower perceived values of annoyance and loudness (for the flyover sounds tested) as the HAGL increases. It is also important to note that the perceived loudness decays more rapidly with HAGL than the perceived annoyance, which might suggest the important contribution of psychoacoustic factors (and probably non-acoustic factors, such as perceived safety) other than loudness. Although there is a clear logarithmic correlation between perceived pitch and HAGL, the residual differences from the trendline are greater (see Figure 4). The perceived pitch might be expected to be associated with the high-frequency content of the test sound (accounted for by the sharpness metric). As the flyover is farther away, the high-frequency content becomes less prominent due to atmospheric absorption. However, a weaker logarithmic correlation with HAGL suggests that other psychoacoustic factors might have an important contribution to the perceived pitch.
In order to further investigate the relationship between different psychoacoustic factors and the perceived responses for the flyover drone sounds tested, a correlation analysis was carried out. Zero-order and partial correlation coefficients controlling for HAGL were calculated between PNL and the SQMs (loudness, sharpness, fluctuation strength, Aures tonality, roughness, impulsiveness) and perceived annoyance (Table 4), perceived loudness (Table 5) and perceived pitch (Table 6). Note that the partial-correlation analysis allowed measurement of the correlation coefficients whilst controlling the effect of HAGL.
As seen in Table 4 and Table 5, PNL, loudness and sharpness have the highest correlation (p ≤ 0.05) with perceived annoyance and loudness, respectively, both zero-order and when controlling for HAGL. This suggests that the different HAGL for the flyover drone sounds did not have any influence in the relationship between PNL, loudness and sharpness and responses of perceived annoyance and perceived loudness. For the case of perceived pitch (Table 6), the highest zero-order correlation is with PNL, loudness and sharpness, but when controlling for HAGL there is a statistically significant correlation with roughness as well. This might help explain the results shown in Figure 4, where it was suggested that other psychoacoustic factors were likely to explain the participants’ responses of perceived pitch as a function of HAGL.

4.3. Loudness and PNL vs. Perceived Responses

After investigating the correlation between PNL and the SQMs considered and the perceived responses for flyover drone sounds at varying HAGL, a bivariate correlation analysis was carried out with the whole set of test sounds. As seen in Table 7, PNL, loudness, sharpness and fluctuation strength have statistically significant correlation (p ≤ 0.05) with both perceived annoyance and perceived loudness. Responses on perceived annoyance and loudness seem to have a significant association to PNL, loudness and sharpness.
Table 7 also shows that PNL, loudness, sharpness, Aures tonality, roughness and impulsiveness have statistically significant correlation (p ≤ 0.05) with perceived pitch. For this particular case, sharpness has the highest correlation with perceived pitch, although contribution from the metrics listed above is observed as well.
An important finding, as shown in Table 7, is that PNL has a greater correlation with perceived annoyance and perceived loudness than the loudness SQM (and similar correlation to perceived pitch as loudness). The PNL metric was developed by Kryter [32] to assess the perception of jet aircraft noise. The frequency vs. sound pressure level (defined in the Noy scale) seems to be able to efficiently capture the perception of the amplitude and spectral characteristics of the drone sounds tested. The Noy scale, and therefore the PNL metric, assumes a significant sensitivity to higher frequency noise, which might explain its strong correlation with perceived annoyance and loudness. For this reason, it was decided to use the PNL and not the loudness metric for the subsequent regression analysis presented in Section 4.4.

4.4. Metrics for Drone Noise Assessment

An MLR analysis was undertaken to define the main contributors to perceived annoyance, perceived loudness and perceived pitch. All assumptions of MLR were checked before implementing the analysis:
  • The linear relationship between predictors and dependent variables was verified in Table 7.
  • The value of the variance inflation factor (VIF), well below 10, allowed the assumption of no multicollinearity in the data.
  • The Durbin–Watson statistic was used to test the assumption of being independent. The value of this statistic was 1.30 (perceived annoyance), 1.52 (perceived loudness) and 1.39 (perceived pitch), allowing the assumption of the residuals being independent.
  • Homoscedasticity was assumed by observing a random distribution of values in scatterplots between the regression-standardised predicted values and the regression-standardised residual values.
  • Exploration of the P–P plots, with data points close to the observed vs. expected cumulative probability diagonal line, allowed the assumption of residuals being normally distributed.
  • Values of Cook’s distance below 1 (max values of 0.12 for perceived annoyance, 0.83 for perceived loudness and 0.23 for perceived pitch) allowed the assumption of no influential cases biasing the MLR models presented.
As shown in Table 8, participants’ responses of perceived annoyance are mainly driven by PNL and sharpness. This suggests the perceived loudness and the high-frequency content of drone noise are key elements for perceived annoyance [17]. Perceived loudness is mainly driven by PNL and fluctuation strength. In this case, the beating effect due to the interaction between rotors seems to play an important role in perceived loudness [33]. The main contributors to perceived pitch are sharpness, roughness and Aures tonality. These results seem to indicate that pitch, as perceived by the participants, is highly influenced by the high-frequency and tonal content of drone noise, including the perceptual effect of the interaction between complex tones [22] typical of drone noise [23].
A multilevel analysis was also carried out to investigate whether the main contributors to perceived annoyance, loudness and pitch (as reported in Table 8) were consistent between participants. Table 9 shows the statistically significant predictors for perceived annoyance, loudness and pitch based on a multilevel analysis with subject-dependent intercepts and regression slopes. As seen in Table 9, there is consistency in the main contributors to the three perceptual variables between participants. The similarity of results in Table 8 and Table 9 suggest certainty in the predictors for the subjective models.
Following a ‘one-off’ approach previously implemented by Boucher et al. [37] and Torija et al. [21], the relative importance of a metric was assessed based on model accuracy (in terms of R2) when removing it from the analysis. Figure 5 confirms that participants’ responses of perceived annoyance and loudness are highly determined by PNL, while the responses of perceived pitch are highly determined by Aures tonality and especially roughness. These results suggest that rapid amplitude modulation, for instance due to unsteadiness of the sound signal, might affect the perception of pitch.

5. Discussion

The results presented in this paper, in terms of correlation between metrics and drone noise perception, are consistent with existing literature. In line with Hui et al. [3], the subjective responses evaluated in this paper correlate strongly with loudness-related metrics. Other SQMs driving responses of perceived annoyance were sharpness and fluctuation strength, similar to Gwak et al. [17]. Interestingly, the two metrics driving responses of perceived loudness were PNL and fluctuation strength. The contribution to fluctuation strength to perceived loudness (and also to perceived annoyance) is assumed to be due to accounting for the perceptual effect of interactions between rotors, however, further research is needed to validate this assumption. It has also been found that the perceived pitch for the drones assessed was mainly related to the high frequency, tonality and rapid amplitude modulation (described as due to interaction between discrete tones by Torija et al. [22]).
In Christian and Cabell [16], the annoyance response did not change significantly with drone altitude ranging from 10 to 100 m above ground level with other parameters held constant. However, this research has found the subjective responses evaluated changing significantly with HAGL. For instance, the perceived annoyance varied from about 0.5 to 0.4 when the HAGL changed from 20 to 60 m. Unlike Christian and Cabell [16], where only a SUI Endurance equipped with two-blade props was considered for analysis, this research investigated subjective responses vs. HAGL with flyovers of seven types of drones (DJI Inspire, Intel Falcon, DJI Matrice 200 and 600, DHI Phantom 3 and Yuneec Typhoon). This might be one of the reasons for the different findings, but further research is needed to investigate in more depth the relationship between drone noise perception and distance from the receiver. A better understanding of this relationship is key to defining operational constraints for drones in order to minimise community noise impact.
Due to the COVID-19 pandemic, only the subjective experiment was carried out, and not as it is usually done under controlled laboratory conditions. As recently reported by the technical committee of the Acoustical Society of America on Psychological and Physiological Acoustics, online experiments can provide access to larger sample sizes and ecologically valid responses, but at the cost of compromising the calibration process and finding inconsistencies in participant experiences [47]. Although the statistical tests reported in Section 4 confirmed consistency in participant responses, the results presented in this paper should be interpreted with caution, and important caveats should be considered:
(i) A careful process to calibrate the drone sounds to the target LAeq,4s (shown in Table 1) was carried out. Moreover, a pseudo-calibration stage was included in the online platform, where the participants adjusted their playback volume of the loudest stimulus to a comfortable level and the quietest stimulus to a just-audible level (see Section 3.2.2 for further details). The assumption was that even though the adjusted playback volume was different for each participant, the relative values of LAeq,4s of the individual stimuli were consistent (as they were calibrated in the laboratory as described in Section 3.1). However, the different playback hardware and quality and frequency response of each participant’s headphones might have altered the actual LAeq,4s of the individual stimuli as heard by the participants.
(ii) The participants were instructed to complete the test with adequate headphones and in a quiet environment. However, it can be assumed that the quality and frequency response of the headphones used and the background sound level where they completed the test varied between participants.
(iii) The online experiment was designed for the participants to reproduce the stimuli via headphones. There is uncertainty as to whether the headphones used by the participants were able to recreate the low frequency noise produced by the drones evaluated.
Therefore, the research findings presented and described in this paper will be validated in a subsequent experiment carried out under controlled laboratory conditions and reproducing the test sounds via a loudspeaker array.

6. Conclusions

This paper presents the results of a subjective experiment to investigate the noise perception of a comprehensive set of drone sounds encompassing different flying operations, size, weight and distance from the receiver. Based on a detailed statistical analysis, the responses of the participants for each drone sound were very consistent (even though the experiment was conducted online).
For the drone sounds tested, the participants’ responses of perceived annoyance were mainly driven by PNL and sharpness, confirming the significance of the high-frequency content present in drone noise. For perceived loudness, participants’ responses were mainly driven by PNL and fluctuation strength. In this case, the beating effect due to rotor interactions might affect the perception of loudness. Perceived pitch was found to be highly influenced by sharpness, Aures tonality and roughness. In this case, the perception of pitch seemed to be driven by the high-frequency and tonal content along with the rapid amplitude modulation due to the unsteadiness of the sound signal.
A robust logarithmic relationship was found between the perceived responses and distance from the receiver (quantified in terms of HAGL). An increase in HAGL of drone flyovers led to consistent reductions in perceived annoyance, perceived loudness and, to a lesser degree, perceived pitch. However, it was found that the perceived loudness declines more rapidly than perceived annoyance when HAGL increases. This means that increasing the distance of drone operations from communities is likely to lead to more substantial reductions in perceived loudness than perceived annoyance. This is likely due to other psychoacoustic factors, or even non-acoustic factors, on the perception of annoyance.
The findings of this research could facilitate better understanding of the key psychoacoustic factors to account for in order to mitigate community noise impact when planning drone operations. The metrics proposed in this paper could also aid the effective assessment of human response to drone noise.
Further research is needed to better understand the effects of drone noise on existing soundscapes, how ambient noise may mask drone noise, and the influence of this masking on the perception of drone operations.

Author Contributions

Conceptualization, A.J.T. and R.K.N.; methodology, A.J.T. and R.K.N.; formal analysis, A.J.T. and R.K.N.; data curation, R.K.N.; writing—original draft preparation, A.J.T.; writing—review and editing, A.J.T. and R.K.N.; visualization, A.J.T. and R.K.N.; supervision, A.J.T.; funding acquisition, A.J.T. All authors have read and agreed to the published version of the manuscript.

Funding

A.J.T. would like to acknowledge the funding received by Innovate UK (ref. 73692) and the UK Engineering and Physical Sciences Research Council (EP/V031848/1).

Institutional Review Board Statement

Prior to conducting the subjective experiment, approval was granted by the Ethics Committee of the University of Salford, UK (application ID 1165).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Restrictions apply to the availability of the drone audio recordings used in this paper. Part of the audio recordings were obtained from John A. Volpe National Transportation Systems Center; and from Nathan Green. Access to this data can be requested to David R. Read, Christopher Cutler and Juliet Page (John A. Volpe National Transportation Systems Center); and Nathan Green (PhD student at the University of Salford’s Acoustics Research Centre). Part of the drone recordings were collected by A.J.T., and would be available upon request.

Acknowledgments

The authors would like to acknowledge the drone audio recordings provided by Nathan Green (PhD student at the University of Salford’s Acoustics Research Centre); by David R. Read, Christopher Cutler and Juliet Page (John A. Volpe National Transportation Systems Center), the Federal Aviation Administration and the Choctaw Nation of Oklahoma.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ackerman, E.; Strickland, E. Medical delivery drones take flight in east africa. IEEE Spectr. 2018, 55, 34–35. [Google Scholar] [CrossRef]
  2. Elsayed, M.; Mohamed, M. The impact of airspace regulations on unmanned aerial vehicles in last-mile operation. Transp. Res. Part D Transp. Environ. 2020, 87, 102480. [Google Scholar] [CrossRef]
  3. Hui, C.T.J.; Kingan, M.J.; Hioka, Y.; Schmid, G.; Dodd, G.; Dirks, K.N.; Edlin, S.; Mascarenhas, S.; Shim, Y.-M. Quantification of the Psychoacoustic Effect of Noise from Small Unmanned Aerial Vehicles. Int. J. Environ. Res. Public Health 2021, 18, 8893. [Google Scholar] [CrossRef] [PubMed]
  4. Duporge, I.; Spiegel, M.P.; Thomson, E.R.; Chapman, T.; Lamberth, C.; Pond, C.; Macdonald, D.W.; Wang, T.; Klinck, H. Determination of optimal flight altitude to minimise acoustic drone disturbance to wildlife using species audiograms. Methods Ecol. Evol. 2021, 12, 2196–2207. [Google Scholar] [CrossRef]
  5. Cabell, R.; Grosveld, F.; McSwain, R. Measured noise from small unmanned aerial vehicles. In Inter-Noise and Noise-Con Congress and Conference Proceedings; Institute of Noise Control Engineering: Reston, VA, USA, 2016; pp. 345–354. [Google Scholar]
  6. Zhou, T.; Jiang, H.; Sun, Y.; Fattah, R.J.; Zhang, X.; Huang, B.; Cheng, L. Acoustic characteristics of a quad-copter under realistic flight conditions. In Proceedings of the 25th AIAA/CEAS Aeroacoustics Conference, Delft, The Netherlands, 20–23 May 2019; p. 2587. [Google Scholar]
  7. Alexander, W.N.; Whelchel, J. Flyover Noise of Multi-Rotor sUAS. In Inter-Noise and Noise-Con Congress and Conference Proceedings; Institute of Noise Control Engineering: Reston, VA, USA, 2019; Volume 259, pp. 2548–2558. [Google Scholar]
  8. Tinney, C.E.; Sirohi, J. Multirotor Drone Noise at Static Thrust. AIAA J. 2018, 56, 2816–2826. [Google Scholar] [CrossRef]
  9. Whelchel, J.; Alexander, W.N.; Intaratep, N. Propeller noise in confined anechoic and open environments. In Proceedings of the AIAA Scitech 2020 Forum, Orlando, FL, USA, 6–10 January 2020; p. 1252. [Google Scholar]
  10. Intaratep, N.; Alexander, W.N.; Devenport, W.J.; Grace, S.M.; Dropkin, A. Experimental study of quadcopter acoustics and performance at static thrust conditions. In Proceedings of the 22nd AIAA/CEAS Aeroacoustics Conference, Lyon, France, 30 May–1 June 2016; p. 2873. [Google Scholar]
  11. McKay, R.S.; Kingan, M.J.; Go, S.T.; Jung, R. Experimental and analytical investigation of contra-rotating multi-rotor UAV propeller noise. Appl. Acoust. 2021, 177, 107850. [Google Scholar] [CrossRef]
  12. Yang, Y.; Liu, Y.; Li, Y.; Arcondoulis, E.; Wang, Y. Aerodynamic and Aeroacoustic Performance of an Isolated Multicopter Rotor During Forward Flight. AIAA J. 2020, 58, 1171–1181. [Google Scholar] [CrossRef]
  13. Zawodny, N.S.; Boyd, D.D., Jr.; Burley, C.L. Acoustic characterization and prediction of representative, small-scale rotary-wing unmanned aircraft system components. In Proceedings of the American Helicopter Society (AHS) Annual Forum, West Palm Beach, FL, USA, 17 May 2016. [Google Scholar]
  14. Miljković, D. Methods for attenuation of unmanned aerial vehicle noise. In Proceedings of the 2018 41st International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO), Opatija, Croatia, 21–25 May 2018; pp. 914–919. [Google Scholar]
  15. Schäffer, B.; Pieren, R.; Heutschi, K.; Wunderli, J.M.; Becker, S. Drone Noise Emission Characteristics and Noise Effects on Humans—A Systematic Review. Int. J. Environ. Res. Public Health 2021, 18, 5940. [Google Scholar] [CrossRef]
  16. Christian, A.W.; Cabell, R. Initial investigation into the psychoacoustic properties of small unmanned aerial system noise. In Proceedings of the 23rd AIAA/CEAS Aeroacoustics Conference, Denver, CO, USA, 5–9 June 2017; p. 4051. [Google Scholar]
  17. Gwak, D.Y.; Han, D.; Lee, S. Sound quality factors influencing annoyance from hovering UAV. J. Sound Vib. 2020, 489, 115651. [Google Scholar] [CrossRef]
  18. Ivošević, J.; Ganić, E.; Petošić, A.; Radišić, T. Comparative UAV Noise-Impact Assessments through Survey and Noise Measurements. Int. J. Environ. Res. Public Health 2021, 18, 6202. [Google Scholar] [CrossRef]
  19. Ciaburro, G.; Iannace, G.; Trematerra, A. Research for the presence of unmanned aerial vehicle inside closed environments with acoustic measurements. Buildings 2020, 10, 96. [Google Scholar] [CrossRef]
  20. Iannace, G.; Ciaburro, G.; Trematerra, A. Acoustical unmanned aerial vehicle detection in indoor scenarios using logistic regression model. Build. Acoust. 2021, 28, 77–96. [Google Scholar] [CrossRef]
  21. Torija, A.J.; Li, Z.; Self, R.H. Effects of a hovering unmanned aerial vehicle on urban soundscapes perception. Transp. Res. Part D Transp. Environ. 2020, 78, 102195. [Google Scholar] [CrossRef]
  22. Torija, A.J.; Roberts, S.; Woodward, R.; Flindell, I.H.; McKenzie, A.R.; Self, R.H. On the assessment of subjective response to tonal content of contemporary aircraft noise. Appl. Acoust. 2019, 146, 190–203. [Google Scholar] [CrossRef] [Green Version]
  23. Torija, A.J.; Clark, C. A Psychoacoustic Approach to Building Knowledge about Human Response to Noise of Unmanned Aerial Vehicles. Int. J. Environ. Res. Public Health 2021, 18, 682. [Google Scholar] [CrossRef] [PubMed]
  24. Zwicker, E.; Fastl, H. Psychoacoustics: Facts and Models; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2013; Volume 22. [Google Scholar]
  25. Torija, A.J.; Li, Z. Metrics for assessing the perception of drone noise. In Proceedings of the e-Forum Acusticum 2020, Lyon, France, 7–11 December 2020. [Google Scholar]
  26. Makarewicz, R.; Galuszka, M.; Kokowski, P. Evaluation of aircraft noise measurements. Noise Control Eng. J. 2014, 62, 83–89. [Google Scholar] [CrossRef]
  27. Basner, M.; McGuire, S. WHO Environmental Noise Guidelines for the European Region: A Systematic Review on Environmental Noise and Effects on Sleep. Int. J. Environ. Res. Public Health 2018, 15, 519. [Google Scholar] [CrossRef] [Green Version]
  28. Angerer, J.R.; McCurdy, D.A.; Erickson, R.A. Development of an Annoyance Model Based upon Elementary Auditory Sensations for Steady-State Aircraft Interior Noise Containing Tonal Components; NASA: Hampton, VA, USA, 1991. Available online: https://ntrs.nasa.gov/citations/19920004540 (accessed on 5 February 2022).
  29. More, S. Aircraft Noise Metrics and Characteristics; PARTNER Project 24 Report COE-2011; Purdue University: West Lafayette, IN, USA, 2011; Volume 4. [Google Scholar]
  30. Read, D.; Roof, C. Research to Support New Entrants to Public Airspace and Aircraft Noise Certification. 2020. Available online: https://rosap.ntl.bts.gov/view/dot/54473 (accessed on 5 February 2022).
  31. FAA. Noise Standards: Aircraft Type and Airworthiness Certification, Calculation of Effective Perceived Noise Level from Measured Data; FAA: Washington, DC, USA, 2002.
  32. Kryter, K.D. The Meaning and Measurement of Perceived Noise Level. Noise Control 1960, 6, 12–27. [Google Scholar] [CrossRef]
  33. Torija, A.J.; Chaitanya, P.; Li, Z. Psychoacoustic analysis of contra-rotating propeller noise for unmanned aerial vehicles. J. Acoust. Soc. Am. 2021, 149, 835–846. [Google Scholar] [CrossRef]
  34. Sottek, R.; Vranken, P.; Busch, G. Ein Modell zrur Berechnung der Impulshaltigkeit. Model Calc. Impulsiveness Proc. DAGA 1995, 95, 13–17. [Google Scholar]
  35. Rizzi, S.A.; Palumbo, D.L.; Rathsam, J.; Christian, A.W.; Rafaelof, M. Annoyance to noise produced by a distributed electric propulsion high-lift system. In Proceedings of the 23rd AIAA/CEAS Aeroacoustics Conference, Denver, CO, USA, 5–9 June 2017; p. 4050. [Google Scholar]
  36. Krishnamurthy, S.; Christian, A.; Rizzi, S. Psychoacoustic Test to Determine Sound Quality Metric Indicators of Rotorcraft Noise Annoyance. In Inter-Noise and Noise-Con Congress and Conference Proceedings; Institute of Noise Control Engineering: Reston, VA, USA, 2018; Volume 258, pp. 317–328. [Google Scholar]
  37. Boucher, M.; Krishnamurthy, S.; Christian, A.; Rizzi, S.A. Sound quality metric indicators of rotorcraft noise annoyance using multilevel regression analysis. Proc. Meet. Acoust. 2019, 36, 040004. [Google Scholar]
  38. Torija, A.J.; Self, R.H.; Lawrence, J.L. Psychoacoustic Characterisation of a Small Fixed-pitch Quadcopter. In Inter-Noise and Noise-Con Congress and Conference Proceedings; Institute of Noise Control Engineering: Reston, VA, USA, 2019; pp. 1884–1894. [Google Scholar]
  39. Read, D.R.; Senzing, D.A.; Cutler, C.; Elmore, E.; He, H. Noise Measurement Report: Unconventional Aircraft—Choctaw Nation of Oklahoma, July 2019; DOT/FAA/AEE/2020-04; U.S. Department of Transportation, John A. Volpe National Transportation Systems Center: Cambridge, MA, USA, 2020.
  40. Jillings, N.; Man, B.D.; Moffat, D.; Reiss, J.D. Web Audio Evaluation Tool: A browser-based listening test environment. In Proceedings of the 12th Sound and Music Computing Conference (SMC2015), Maynooth, Ireland, 26 July–1 August 2015. [Google Scholar]
  41. Susini, P.; Lemaitre, G.; McAdams, S. Psychological measurement for sound description and evaluation. Meas. Pers. Theory Methods Implement. Areas 2012, 227, 241–268. [Google Scholar]
  42. DIN 45631/A1-2010; Calculation of Loudness Level and Loudness from the Sound Spectrum—Zwicker Method—Amendment 1: Calculation of the Loudness of Time-Variant Sound. DIN: Berlin, Germany, 2010.
  43. Aures, W. Berechnungsverfahren für den sensorischen Wohlklang beliebiger Schallsignale. Acta Acust. United Acust. 1985, 59, 130–141. [Google Scholar]
  44. Aures, W. Ein Berechnungsverfahren der Rauhigkeit. Acta Acust. United Acust. 1985, 58, 268–281. [Google Scholar]
  45. Sottek, R. Modelle zur Signalverarbeitung im Menschlichen Gehör; RWTH Aachen University: Aachen, Germany, 1993. [Google Scholar]
  46. Torija, A.J.; Flindell, I.H. The subjective effect of low frequency content in road traffic noise. J. Acoust. Soc. Am. 2015, 137, 189–198. [Google Scholar] [CrossRef]
  47. Peng, Z.E.; Buss, E.; Shen, Y.; Bharadwaj, H.; Stecker, G.C.; Beim, J.A.; Bosen, A.K.; Braza, M.; Diedesch, A.C.; Dorey, C.M. Remote testing for psychological and physiological acoustics: Initial report of the p&p task force on remote testing. In Proceedings of Meetings on Acoustics 179ASA; Acoustical Society of America: Melville, NY, USA, 2020; p. 050009. [Google Scholar]
Figure 1. Calibration setup used for subjective test drone stimuli.
Figure 1. Calibration setup used for subjective test drone stimuli.
Ijerph 19 03152 g001
Figure 2. Perceived annoyance vs. height above ground level of the unmanned aerial vehicles under investigation during flyover operation.
Figure 2. Perceived annoyance vs. height above ground level of the unmanned aerial vehicles under investigation during flyover operation.
Ijerph 19 03152 g002
Figure 3. Perceived loudness vs. height above ground level of the unmanned aerial vehicles under investigation during flyover operation.
Figure 3. Perceived loudness vs. height above ground level of the unmanned aerial vehicles under investigation during flyover operation.
Ijerph 19 03152 g003
Figure 4. Perceived pitch vs. height above ground level of the unmanned aerial vehicles under investigation during flyover operation.
Figure 4. Perceived pitch vs. height above ground level of the unmanned aerial vehicles under investigation during flyover operation.
Ijerph 19 03152 g004
Figure 5. Reduction in R2 per predictor removed from multilevel model (using subject-dependent intercepts and regression slopes) for perceived annoyance, loudness and pitch.
Figure 5. Reduction in R2 per predictor removed from multilevel model (using subject-dependent intercepts and regression slopes) for perceived annoyance, loudness and pitch.
Ijerph 19 03152 g005
Table 1. Drone sounds used in the subjective experiment.
Table 1. Drone sounds used in the subjective experiment.
Sound IDDrone ModelDrone Weight (kg)Operating ProcedureHeight above Ground Level, HAGL (m)Calibrated LAeq,4s
S1DJI Inspire2.85Flyover1552
S2DJI Inspire2.85Flyover7.558
S3DJI Inspire2.85Landing7.564
S4DJI Inspire2.85Takeoff270
S5Intel Falcon1.2Flyover3054
S6Intel Falcon1.2Flyover6047
S7DJI Matrice 6009.1Takeoff371
S8DJI Matrice 6009.1Hover4065
S9DJI Matrice 6009.1Flyover4057
S10DJI Mavic0.743Flyover1551
S11DJI Mavic0.743Flyover3046
S12DJI Mavic0.743Flyover6037
S13DJI Mavic0.743Maneuvering7.551
S14DJI Mavic0.743Maneuvering7.553
S15DJI Mavic0.743Takeoff7.559
S16DJI Phantom 31.216Maneuvering268
S17DJI Phantom 31.216Takeoff264
S18DJI Phantom 31.216Landing262
S19DJI Phantom 31.216Hover269
S20DJI Phantom 31.216Ascending264
S21DJI Phantom 31.216Flyover261
S22DJI Phantom 31.216Flyover263
S23DJI Phantom 31.216Flyover266
S24DJI Phantom 31.216Flyover5.456
S25DJI Phantom 31.216Flyover5.459
S26DJI Phantom 31.216Flyover5.457
S27DJI Phantom 31.216Hover2.262
S28DJI Phantom 31.216Hover5.156
S29DJI Phantom 31.216Hover2.267
S30DJI Phantom 31.216Hover3.667
S31DJI Matrice 2004Flyover4656
S32DJI Matrice 2004Flyover4645
S33DJI Matrice 2004Takeoff3050
S34DJI Matrice 2004Landing3052
S35DJI Matrice 2004Hover1.256
S36Yuneec Typhoon2Flyover4648
S37Yuneec Typhoon2Flyover4644
S38Yuneec Typhoon2Takeoff3046
S39Yuneec Typhoon2Landing3052
S40Yuneec Typhoon2Hover1.257
S41Gryphon GD28X11.8Takeoff3053
S42Gryphon GD28X11.8Landing3054
S43Gryphon GD28X11.8Maneuvering3057
S44Gryphon GD28X11.8Hover1.260
Table 2. Results of the Kendall’s W statistic for the responses on perceived annoyance, loudness and pitch.
Table 2. Results of the Kendall’s W statistic for the responses on perceived annoyance, loudness and pitch.
Perceived AnnoyancePerceived LoudnessPerceived Pitch
Kendall’s W0.600.640.41
p-value 10.000.000.00
1 p-value calculated with Monte Carlo bootstrapping with 10,000 samples.
Table 3. Coefficient of variation for each test sound and for perceived annoyance, loudness and pitch.
Table 3. Coefficient of variation for each test sound and for perceived annoyance, loudness and pitch.
Sound IDPerceived AnnoyancePerceived LoudnessPerceived Pitch
S10.550.490.44
S20.320.310.39
S30.200.190.36
S40.170.180.36
S50.360.470.32
S60.620.670.41
S70.150.190.45
S80.360.350.47
S90.300.320.44
S100.480.430.35
S110.520.600.41
S120.870.830.58
S130.330.460.22
S140.280.350.27
S150.250.310.29
S160.120.170.25
S170.190.220.24
S180.240.250.27
S190.130.160.27
S200.150.200.22
S210.210.230.27
S220.230.230.28
S230.190.200.21
S240.350.380.41
S250.250.320.27
S260.310.350.28
S270.200.230.28
S280.340.390.36
S290.160.170.24
S300.220.180.25
S310.570.600.57
S320.540.600.58
S330.350.390.54
S340.360.410.46
S350.260.320.43
S360.640.740.44
S370.750.810.49
S380.490.550.31
S390.370.500.33
S400.580.720.51
S410.310.410.64
S420.380.340.64
S430.300.290.63
S440.240.290.73
Table 4. Zero-order and partial correlation coefficients (controlling for height above ground level (HAGL)) between PNL and the SQMs (loudness, sharpness, fluctuation strength, Aures tonality, roughness and impulsiveness) and perceived annoyance. p-value shown in brackets.
Table 4. Zero-order and partial correlation coefficients (controlling for height above ground level (HAGL)) between PNL and the SQMs (loudness, sharpness, fluctuation strength, Aures tonality, roughness and impulsiveness) and perceived annoyance. p-value shown in brackets.
PNLLoudnessSharpnessFluctuation StrengthAures TonalityRoughnessImpulsiveness
Zero-order0.96 (p = 0.00)0.91 (p = 0.00)0.87 (p = 0.00)0.24 (p = 0.33)0.23 (p = 0.35)0.08 (p = 0.75)0.01 (p = 0.97)
Controlling for HAGL0.88 (p = 0.00)0.77 (p = 0.00)0.76 (p = 0.00)0.30 (p = 0.24)−0.13 (p = 0.62)−0.03 (p = 0.90)−0.17 (p = 0.51)
Table 5. Zero-order and partial correlation coefficients (controlling for height above ground level (HAGL)) between PNL and the SQMs (loudness, sharpness, fluctuation strength, Aures tonality, roughness and impulsiveness) and perceived loudness. p-value shown in brackets.
Table 5. Zero-order and partial correlation coefficients (controlling for height above ground level (HAGL)) between PNL and the SQMs (loudness, sharpness, fluctuation strength, Aures tonality, roughness and impulsiveness) and perceived loudness. p-value shown in brackets.
PNLLoudnessSharpnessFluctuation StrengthAures TonalityRoughnessImpulsiveness
Zero-order0.98 (p = 0.00)0.95 (p = 0.00)0.85 (p = 0.00)0.26 (p = 0.29)0.32 (p = 0.20)0.17 (p = 0.51)0.01 (p = 0.97)
Controlling for HAGL0.92 (p = 0.00)0.86 (p = 0.00)0.72 (p = 0.00)0.38 (p = 0.14)−0.03 (p = 0.91)0.11 (p = 0.67)−0.22 (p = 0.41)
Table 6. Zero-order and partial correlation coefficients (controlling for height above ground level (HAGL)) between PNL and the SQMs (loudness, sharpness, fluctuation strength, Aures tonality, roughness and impulsiveness) and perceived pitch. p-value shown in brackets.
Table 6. Zero-order and partial correlation coefficients (controlling for height above ground level (HAGL)) between PNL and the SQMs (loudness, sharpness, fluctuation strength, Aures tonality, roughness and impulsiveness) and perceived pitch. p-value shown in brackets.
PNLLoudnessSharpnessFluctuation StrengthAures TonalityRoughnessImpulsiveness
Zero-order0.68 (p = 0.00)0.73 (p = 0.00)0.76 (p = 0.00)0.04 (p = 0.89)0.09 (p = 0.73)−0.35 (p = 0.16)−0.08 (p = 0.76)
Controlling for HAGL0.37 (p = 0.14)0.49 (p = 0.05)0.60 (p = 0.01)−0.01 (p = 0.96)−0.21 (p = 0.43)−0.54 (p = 0.03)−0.21 (p = 0.42)
Table 7. Bivariate correlation coefficients (Pearson’s r coefficient) between PNL and the SQMs (loudness, sharpness, fluctuation strength, Aures tonality, roughness and impulsiveness) and perceived annoyance, loudness and pitch. p-value shown in brackets.
Table 7. Bivariate correlation coefficients (Pearson’s r coefficient) between PNL and the SQMs (loudness, sharpness, fluctuation strength, Aures tonality, roughness and impulsiveness) and perceived annoyance, loudness and pitch. p-value shown in brackets.
PNLLoudnessSharpnessFluctuation StrengthAures TonalityRoughnessImpulsiveness
Perceived Annoyance0.96 (p = 0.00) 0.90 (p = 0.00) 0.90 (p = 0.00) 0.40 (p = 0.01) 0.25 (p = 0.10)0.19 (p = 0.21)−0.16 (p = 0.30)
Perceived Loudness0.98 (p = 0.00) 0.92 (p = 0.00) 0.87 (p = 0.00) 0.47 (p = 0.00) 0.15 (p = 0.33)0.29 (p = 0.06)−0.08 (p = 0.59)
Perceived Pitch0.47 (p = 0.00) 0.50 (p = 0.00) 0.55 (p = 0.00) −0.00 (p = 0.98)0.48 (p = 0.00) −0.34 (p = 0.03) −0.37 (p = 0.01)
Table 8. Summary of multiple linear regression models to estimate perceived annoyance, loudness and pitch.
Table 8. Summary of multiple linear regression models to estimate perceived annoyance, loudness and pitch.
R2Adjusted R2PredictorsStandardised Beta CoefficientVariance Inflation Factor
Perceived Annoyance0.930.93PNL0.724.15
Sharpness0.284.15
Perceived Loudness0.970.97PNL0.951.18
Fluctuation Strength0.091.18
Perceived Pitch0.610.59Sharpness0.561.13
Roughness−0.451.06
Aures Tonality0.321.08
Table 9. Statistical significance (p-value) of predictors for perceived annoyance, loudness and pitch with subject-dependent intercepts and regression slopes.
Table 9. Statistical significance (p-value) of predictors for perceived annoyance, loudness and pitch with subject-dependent intercepts and regression slopes.
PredictorsPerceived AnnoyancePerceived LoudnessPerceived Pitch
PNL0.000.000.17
Sharpness0.000.040.00
Fluctuation Strength0.220.001
Aures Tonality0.290.380.00
Roughness0.210.820.00
Impulsiveness0.1410.50
1 Predictor redundant in multilevel analysis.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Torija, A.J.; Nicholls, R.K. Investigation of Metrics for Assessing Human Response to Drone Noise. Int. J. Environ. Res. Public Health 2022, 19, 3152. https://doi.org/10.3390/ijerph19063152

AMA Style

Torija AJ, Nicholls RK. Investigation of Metrics for Assessing Human Response to Drone Noise. International Journal of Environmental Research and Public Health. 2022; 19(6):3152. https://doi.org/10.3390/ijerph19063152

Chicago/Turabian Style

Torija, Antonio J., and Rory K. Nicholls. 2022. "Investigation of Metrics for Assessing Human Response to Drone Noise" International Journal of Environmental Research and Public Health 19, no. 6: 3152. https://doi.org/10.3390/ijerph19063152

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop