Next Article in Journal
Genetics Reveal Long-Distance Virus Transmission Links in Pacific Salmon
Previous Article in Journal
Effect of Black Soldier Fly (Hermetia illucens L.) Fat on Health and Productivity Performance of Dairy Cows
Previous Article in Special Issue
Male African Elephant (Loxodonta africana) Behavioral Responses to Estrous Call Playbacks May Inform Conservation Management Tools
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sound Visualization Demonstrates Velopharyngeal Coupling and Complex Spectral Variability in Asian Elephants

by
Veronika C. Beeck
1,*,
Gunnar Heilmann
2,
Michael Kerscher
2 and
Angela S. Stoeger
1,*
1
Department of Behavioural and Cognitive Biology, University of Vienna, 1030 Vienna, Austria
2
Mecalc Technologies GmbH, 12487 Berlin, Germany
*
Authors to whom correspondence should be addressed.
Animals 2022, 12(16), 2119; https://doi.org/10.3390/ani12162119
Submission received: 15 June 2022 / Revised: 2 August 2022 / Accepted: 13 August 2022 / Published: 18 August 2022
(This article belongs to the Special Issue Elephant Communication)

Abstract

:

Simple Summary

Whether a vocalization is emitted though the mouth or nose impacts on its acoustic shape through specific vocal tract resonances. In human language, nasalized compared to oral vowels can change the meaning of words. African elephants mainly use low-frequency rumbles to communicate and can utter them either through the mouth or trunk. In this study, we used an acoustic camera to visualize the sound emission of rumbles in Asian elephants, which have been studied considerably less than African elephants. We recorded nine captive female Asian elephants and analyzed the acoustic structure of 203 calls. We found that most rumbles (64%) were emitted through the trunk, 21% through the mouth, and surprisingly, 13% simultaneously through the mouth and trunk. Some of the rumbles were combined with orally emitted roars. The nasal rumbles concentrated most spectral energy in lower frequencies, whereas the oral and mixed rumbles showed a broader spectral energy distribution and concentration in higher frequencies, and were louder. The roars were the loudest, broadest and highest in frequency. This study is the first to demonstrate coupled oral and nasal emission in a non–human animal, thereby setting an important framework to study the functions of this acoustic variability in elephant communication and the evolution of vocal flexibility across species.

Abstract

Sound production mechanisms set the parameter space available for transmitting biologically relevant information in vocal signals. Low–frequency rumbles play a crucial role in coordinating social interactions in elephants’ complex fission–fusion societies. By emitting rumbles through either the oral or the three-times longer nasal vocal tract, African elephants alter their spectral shape significantly. In this study, we used an acoustic camera to visualize the sound emission of rumbles in Asian elephants, which have received far less research attention than African elephants. We recorded nine adult captive females and analyzed the spectral parameters of 203 calls, including vocal tract resonances (formants). We found that the majority of rumbles (64%) were nasally emitted, 21% orally, and 13% simultaneously through the mouth and trunk, demonstrating velopharyngeal coupling. Some of the rumbles were combined with orally emitted roars. The nasal rumbles concentrated most spectral energy in lower frequencies exhibiting two formants, whereas the oral and mixed rumbles contained higher formants, higher spectral energy concentrations and were louder. The roars were the loudest, highest and broadest in frequency. This study is the first to demonstrate velopharyngeal coupling in a non-human animal. Our findings provide a foundation for future research into the adaptive functions of the elephant acoustic variability for information coding, localizability or sound transmission, as well as vocal flexibility across species.

Graphical Abstract

1. Introduction

In mammals, the length and shape of the vocal tract modulates the sound produced by the vibration of the vocal folds, and thus the information it conveys to the listener [1,2]. Elephants have evolved the longest and one of the most sophisticated vocal tract elongations in mammals, the trunk [3]. Hence, this flexible muscular hydrostat plays a crucial role in their acoustic communication, apart from serving as a multipurpose tool for manipulating food and objects [3,4,5], for sensory information and for tactile, chemical, and gestural communication [6,7,8,9]. The three recent elephant species—the African savannah elephant (Loxodonta africana, henceforth African elephant), African forest elephant (Loxodonta cyclotis) and Asian elephant (Elephas maximus)—share 8 to 10 call types that appear to constitute a rather limited vocal repertoire at first sight but actually exhibits an impressive flexibility in within-call–type variation and the combinations of trunk and laryngeal sounds [10,11]. In addition, single captive individuals of African elephants and one Asian elephant proved capable of adding novel sounds to their repertoire through vocal production learning, a capability whose functional value in their interspecific communication remains unknown [12,13]. The trunk is used not only to produce the iconic high-frequency trumpet, where the exact sound source is not yet localized, but also squeaks, idiosyncratic high-frequency sounds, throbbing sounds, snorts and blows (Beeck, unpublished data, [14,15,16]). By vibrating their massive vocal folds, the elephants produce tonal rumbles [17] with fundamental frequencies reaching into the infrasonic range (<20 Hz), but with energy present in higher harmonics up to about 1000 Hz [18,19]. The rumble is the most common call type used across elephant species for interspecific communication in their complex fission–fusion societies [20,21,22]. The trunk’s resonant properties have been shown to affect the spectral structure of the rumbles in African elephants [23], but this has not, thus far, been studied in Asian elephants. A better understanding of the adaptive values of the vocal flexibility for functional information-coding across the elephant species requires a better understanding of the relationship between the elephant’s unique morphology and its resulting acoustic properties.

1.1. The Function of the Vocal Tract in Mammals

Source– and filter–related parameters can be modified independently [1,2]. First, the pulmonary air-flow powers the self–sustaining vibration of the vocal folds, which, when regular, translates to the fundamental frequency (F0) and the harmonic overtones of the call. Then, the vocal tract’s filtering function dampens or enhances the specific frequencies, resulting in acoustic energy peaks across harmonics, i.e., formants. The length and shape of the vocal tract determines the location and amplitude of the formants [24,25,26]. In humans, the low position of the larynx and the flexible movements of the supra-laryngeal articulators (e.g., pharynx, tongue, velum, lips and jaw) dynamically shape the vocal tract and enable us to produce consonants and vowels, the latter being determined by the relative position of the lower three formants [24]. The rapid combination of these phonetic elements allow for the almost unlimited amount of information encoded in human speech. The position of the velum determines whether the vocalization is uttered orally, nasally or through a combination of both of the pathways [27]. The latter, referred to as velopharyngeal coupling, is known in human language as the nasalization of vowels, such as in the French or Hindi language [24,28,29]. In 30% of human languages (WALS dataset), the nasalization of vowels can change the meaning of a word, for example in French beau (bo) means “beautiful” and bon (bõ) means “good” [28]. Non–human mammals are generally thought to have less control over the shape and dimension of their vocal tracts than humans [26,30,31].
Orally and nasally coupled vocalizations have, to our knowledge, not been conclusively demonstrated in any non–human mammal. They have been suggested based on the non-uniform formant spacings in rhesus macaques (Macaca mulatta) [32], Diana monkeys (Cercopithecus diana) [33], elephant seals (Mirounga leonina) [34] and fallow deer (Dama dama) [35]. However, velopharyngeal coupling and vocal tract modifications may have similar effects on the spectral structure of the vocal output [28,29,36]. This means that the velopharyngeal coupling cannot be inferred from the acoustic data alone without evidence of sound emission or velum aperture.
Orally or nasally emitted calls, in contrast, have been demonstrated in numerous species. Here, the formant positions are predicted by the vocal tract length, assuming the vocal tract resembles a uniform tube, where the longer the vocal tract, the lower the formant frequencies and the smaller the spacing between them [1,37,38]. The nasal vocal tract is usually longer than the oral one and the formant positions differ accordingly between the oral and nasal calls. This has been shown in goitered gazelles (Gazella subgutturosa) [39], Saiga antelopes (Saiga tatarica) [40], red deer (Cervus elaphus) [41], sheep (Ovis aries) [42] and several lemur species (Eulemur) [43], where the nasal and oral call emission was observed through the mouth openings. The oral or nasal call production were further demonstrated using x-ray videos in dogs (Canis familiaris), goats (Capra hircus), a piglet (Sus scrofa) and cotton-top tamarins (Saguinus oedipus) [27], and an acoustic camera that visualizes sound emission in African elephants [23].
The formant positions have mainly been shown to code “static” information about the stable long-term attributes of the caller (age, sex, gender) in the context of sexual selection or male–male competition [26,44]. Because of the negative relationship between body size and formant dispersion, some mammals extend their vocal tract to exaggerate their size [45]. The oral vocal tract can be enlarged through lip protrusion (e.g., rhesus macaques [37,46], black and white colobus monkeys (Colobus guereza)) or retraction of the larynx (e.g., red deer [47]). The permanent nasal vocal tract elongations, which are more prominent in males, are seen in the proboscic noses of proboscis monkeys [48] and the Saiga antelope’s and elephant seal’s enlarged nasal vestibulums [34,49]. Beyond the mating context, very few studies have examined the variation in the formant patterns caused by articulator modulations in relation to “dynamic” information (arousal, valence, emotion, motivation) [50] or referential signaling [44], except for the studies on the alarm calls of Diana monkeys [33], meerkats (Suricata suricatta) [51] and African elephants [52]. These findings challenge the view that the reduced articulatory capacities of non–human mammals limit the potential information coding in formant patterns [31], and call for a broader consideration of the relative spectral contribution of oral and nasal resonances.

1.2. Call Emission and Information Coding in Elephant Laryngeal Calls

In elephants, the trunk in both of the sexes extremely elongates the nasal vocal tract (up to 2.5 m in African elephant adult females) in relation to the oral path (0.7 to 1 m) (Figure 1) [19]. The rumbles that were uttered through the trunk in three subadult female African elephants exhibited a three-fold reduction in nasal formant positions and spacing (mean F1 = 40 Hz; mean F2 = 169 Hz) as compared to the orally uttered rumbles (mean F1 = 129 Hz; mean F2 = 415 Hz) [23]. During arousal, rumbles often occur in combination with loud roars with a F0 of approximately 200 Hz (in Asian elephants in this current study) [11]. The roars are suggested to be produced laryngeally and the acoustic camera recordings revealed that the African elephants emit their roars orally [16]. In African elephants, rumbles were found to vary gradually [6,53] (but see [54,55]), but to encode perceptional discrete information [52,56,57]. Here, the formant configurations play an important role. They reflect the pronounced sexual dimorphism in the vocal tract dimensions [58], and age and maturity within males [59]. Furthermore, formants vary with individual identity [56], reproductive state [60], dominance rank [61], alarming triggers [52] and also, in calves, the behavioral context [62]. However, the relationship between the morphology of the vocal tract, the sound space and range, and the information encoded is not well understood.
Compared to this extensive body of research on African elephants, the Asian elephant acoustic communication has been considerably less investigated [63,64,65]. Their low-frequency calls were found to vary among individuals (but the formants were not analyzed) [65], and among behavioral contexts, although here individual differences were not considered [66]. DeSilva, 2010, distinguished between growls, which have no spectral energy over 500 Hz and are emitted through a nearly closed mouth, and rumbles, which have a broader frequency spectrum and are uttered through the open mouth. These observations indicate that the Asian elephants, too, emit laryngeal calls through oral or nasal emission.
In the current study, we investigated the sound emission and structure of the rumbles and roars of captive adult female Asian elephants in detail (adult males are more difficult to handle and thus rarer in captivity). We used an acoustic camera array to visualize the sound emission and to measure the sound intensity. By analyzing multiple time and frequency frames per call, we explored the sequential or simultaneous call emission through the oral or vocal tract. We compared the acoustic structure of the call types, estimated the corresponding vocal tract lengths and tested the call–type classification. We discuss commonalities among the mammal vocal tract resonances and the potential function of our findings within the Asian elephant’s vocal communication.
Figure 1. Schematic figure of the vocal tract in sagittal view: (1) larynx (yellow), vocal folds (red), epiglottis (orange), trachea (green), esophagus (brown); (2) velum (dark blue); (3) tongue (pink); (4) nasal cartilages (violet). In this figure, the epiglottis is slightly open unifying the oral and nasal vocal tract. When the (2) velum (dark blue) is lowered and touches the epiglottis (orange) the oral vocal tract can be closed, and the nasal vocal tract opened for purely nasal sound emission (adapted from [67]).
Figure 1. Schematic figure of the vocal tract in sagittal view: (1) larynx (yellow), vocal folds (red), epiglottis (orange), trachea (green), esophagus (brown); (2) velum (dark blue); (3) tongue (pink); (4) nasal cartilages (violet). In this figure, the epiglottis is slightly open unifying the oral and nasal vocal tract. When the (2) velum (dark blue) is lowered and touches the epiglottis (orange) the oral vocal tract can be closed, and the nasal vocal tract opened for purely nasal sound emission (adapted from [67]).
Animals 12 02119 g001

2. Materials and Methods

2.1. Study Subjects and Housing

The subjects in this study were 9 of the 12 captive adult female Asian elephants located at Tiger Tops Elephant Camp within the buffer zone of the Chitwan National Park, Nepal (coordinates 27.52935, 84.35419). They ranged in age from about 27 to 60 years and shoulder height from 2.40 to 2.55 m (Table 1). The elephants were kept chain-free in corrals of around 0.25 to 1.55 ha in size, in groups of two and one group of three based on social affiliations. The animals were taken on daily trips by their handlers to bath in the nearby Narayani River and to browse in the adjacent grassland or in the rainforest of the buffer zone. All of the elephants were trained and habituated to the presence of humans.

2.2. Acoustic Camera and Audio and Video Recordings

We conducted acoustic camera recordings on four days from 9 until 12 of October 2018 at varying times between 6:00 a.m. and 8:00 p.m. The acoustic camera (Star48 AC Pro, gfai Tech GmbH, Berlin, Germany) consists of an array of 48 microphones arranged on a three–armed star around a central camera, for concomitant video and audio recordings. The array is conically tilted forward in the direction of the sound source, creating a back-field suppression of approximately 15 dB. The array was placed about 6–8 m from the vocalizing elephant and connected to a recorder and laptop with the operating program Noise Image (4.10.0). Given the amount of data to be processed, the acoustic camera recordings cannot be conducted continuously, but a pre-trigger and the total recording time (max 360 s) were set before each recording. For the simultaneous and continuous acoustic recordings during the sessions, we used an omni-directional Neumann KM183 condenser microphone that was modified for recording frequencies below 20 Hz (flat-recording down to 5 Hz frequency response: 5 Hz–20 kHz) with a Rycote windshield, which was connected to a Sound Devices 633 (at 48 kHz sampling rate and 16–bit). We used a Sony Camera FD53 for video recordings. The acoustic recordings were stored as .wav files.

2.3. Recording Context

Since we were primarily interested in the sound production of the calls before their acoustic variation relevant to the behavioral context, we aimed at recording as many vocalizations as possible, rather than balancing the observational settings across the individuals. We established contexts in which the individual elephants can be stimulated to vocalize during a previous study period at the same site in February to March 2018, where we had yet collected acoustic and behavioral data, but no acoustic camera data [15,67]. The recording contexts included four settings: separation-reunion experiments at the corrals; vocalizations on the handler’s verbal cue; the elephants bathing at the river; and opportunistic observational recordings between call solicitations.
In the separation-reunion experiment, the focal individual stayed in the corral while the handler led the other elephant away, that was affiliated and housed with the focal elephant, using food and verbal cues, to a spot out of sight for the focal individual (approximate distance 100–200 m) behind dense vegetation, houses and/or straw bales. The elephants from the neighboring corrals were also moved out of sight of the focal before the start of the experiment. To initially distract and calm the focal elephant, the handler gave her three items of food (elephant “sandwiches” composed of straw and grains). After a few minutes, the focal elephant usually started contact calling. The other elephant was allowed to return by itself at any time. This resulted in a greeting ceremony where, upon reunion, the elephants vocalize in overlapping calls, come into close contact and touch and sniff each other repeatedly on different body parts (e.g., mouth, genitals, temporal glands), along with other behaviors, such as spreading and flapping ears, bouncing the trunk on the floor, raising tails and urinating [14,23,68]. Since the elephants were not equally motivated to participate in the experiments, i.e., they could not be separated or did not contact call or return to greet on their own, not every subject successfully acted as a focal, and the call types are therefore not balanced across contexts and individuals. The separations from their group mates were not part of any routine at the study site and the elephants were not used to or trained to be separated. One individual (Sunder) had an impaired hind leg due to a previous injury and we did not directly conduct the experiment with her. She was part of the group of three and was led out of sight before the actual experiment. We opportunistically recorded her calls when she reunited with the group and rated those as greeting calls. She also stayed behind at the corral when the other two elephants she was housed with were taken out for walks, and we recorded her calls as contact calls.
In the handling setting, the elephant groups were brought together in front of the camera and the handlers gave the verbal cue “bol” (speak in Nepali). The elephants are not trained to utter specific calls upon hearing this cue but start vocalizing and exhibit behaviors that are also typical for arousal situations and greeting, such as the frequent mutual touches with the trunk [14,68,69]. At the riverside, the elephants again vocalize and exhibit the greeting and arousal behaviors described above when the handlers descend after riding them to the river, which we coded as arousal, and when the elephants were reunited at the beach after bathing apart, which we coded as greeting. Along with the behavioral context, we also noted the detailed behavior (e.g., trunk touches), distance from the partner, distance from the microphone, and the mouth opening defined by visual mandible lowering (closed when no mandible lowering visible; slightly open up to an angle around 15°; wide open about 20–40°, or chewing when the mandible was moved up and down repeatedly).

2.4. Acoustic Camera Analysis

The acoustic camera recordings were analyzed in Noise Image (4.12.0). To locate the dominant sound source, the sound pressure level (SPL) was displayed by color coding and the resulting acoustic map projected automatically onto the optical image, where the dominant sound source appears as a color–graded sphere. The effective sound pressure at point x on the image plain was calculated by a delay–and–sum beamforming algorithm. The algorithm takes into account the sum of the relative time delays, or the phase shift when analyzed from the frequency domain. It considers each microphone position and compensates the run time or phase shift of the sound arriving at the microphone array (for details see [23]). The ranges within the time and frequency domain can be manually selected in a modifiable rectangular selection window from the waveform or spectrogram to display the acoustic map in the corresponding 2D acoustic photo. We either selected the sound from its fundamental frequency and its visible harmonics from the spectrogram to exclude background noise, specifically selected the dominant frequency contours when the calls overlapped, or specifically selected the lower or upper harmonics to identify the sound emission from multiple sources.
The algorithm assumes that monopolar uncorrelated sound sources are represented in the selected window. Several of the uncorrelated sound sources can appear as two graphic representations. When two or more correlated sound sources are present, the graphic representation of the sound source appears as one at the correct location, if its sound pressure strongly dominates the other sources. If there is a second strong source present, the round sphere of the graphic representation will be distorted in its direction. If the two sound sources are comparably strong, the graphic representation appears in-between the sources, i.e., between the trunk and mouth. By selecting only partials of the spectrogram in the time and frequency dimensions, the multiple correlated sources that occur simultaneously and or in succession can be segregated. The correct localization of the sound source, and especially the separation of the multiple sound sources, is contingent on the distance between the sources (i.e., mouth and trunk) and the distance from the source to the acoustic camera. The localizations of the sound source can be hindered by interfering background noises, strong ground reflection or multiple sound radiators. In these cases, where the sound source was not clearly localized and could not be segregated into oral or nasal parts, the sound source was termed unsure.
To measure the SPL, we only considered the calls where the whole spectrum was measurable by placing the cursor in the middle of the highest intensity color within the sound source from a 1 s selection window in the middle of the call or the whole call if its duration was less than 1 s. The videos were calculated directly from the audio file (.chl) for presentation purposes (overlap 1, framerate 25 f/s).

2.5. Acoustic Data Analysis

For the annotation and formant analysis of the acoustic data, we used STx (Austrian Academy of Sciences, version 4.4.6 [70]). We calculated the spectrograms (Kaiser–Bessel, bandwidth 5 Hz, window length 52 ms, frequency window: 0–1 kHz for nasal, 0–4 kHz for oral and mixed, 0–12 kHz for roars), marked the single calls and extracted their duration. We manually tracked the fundamental frequency contour from the spectrogram (rumbles: frequency 0–0.3 kHz, frame size 400–500 ms, frame increment 20 ms, frequency resolution 0.2 Hz; roars: frequency 0–0.6 kHz, frame size 200 ms, frame increment 10 ms, frequency resolution 0.8 Hz) using a custom–designed tool in Matlab [71] (The MathWorks, version R2017b), and automatically extracted the mean F0 across the call and the frequency variability index (FVI, for the parameter descriptions, see Table 2). In the roars that were both chaotic, i.e., showed high amounts of deterministic chaos, and partially tonal, we tracked the F0 where possible. We measured the spectral centroid frequency, quartiles and Wiener entropy in Rstudio (Version 1.0.153 [72]), using the analyze function of the package, soundgen [73].
To analyze the formant frequencies, we first down–sampled the sound files to 2 kHz for the nasal rumbles and 8 kHz for the mixed and oral rumbles in PRAAT, we did not down–sample the roars. We then used Stx to calculate the spectra with a Linear Predictive Coding (LPC) function over a 0.5 s time window, either directly in the middle of the call or at another location when the formants were more clearly visible, to avoid the background noise or overlapping calls. We looked at frequencies of 0–1 kHz and a LPC function coefficient of 16–20 for the nasal rumbles, 0–4 kHz and 80–100 for the oral and mixed rumbles and 0–10 kHz and 250–270 for the roars (bandwidth 1 Hz, overlap 75%). The LPC coefficients (filter order) are commonly chosen, based on the number of the expected formants 2 + 2*N (formants) [33]. The formant center frequencies can be estimated using the model of a uniform linear tube open at one end (the mouth) and closed at the other (vocal folds) and applying the equation F n = 2 n 1 c 4 V T L , where n is the number of formants, c is the speed of sound in air (350m/s) and VTL is the vocal tract length [1]. The African elephant’s mean oral and nasal VTL is 0.75 m and 2.5 m, respectively [19,74]. In the Asian elephants, no measurements of the VTL have been published so far. However, we assumed a comparable range of the formant frequencies, due to the physical similarities of the species, and chose LPC coefficients based on our previous experience with analyzing the formants of African elephant rumbles [59]. We adjusted the coefficients when the function did not capture the spectral peak concentrations that resembled the formants from a visual inspection of the spectrum and spectrogram. This adjustment included the less prominent formants [34], and those formants that did not comply with a uniform increment across the calls. We took this conservative approach to measuring the formant positions because this is the first study to describe multiple formants in Asian elephants, thus, there are no reference values, and because we expected a deviation in the measured formant positions from those calculated based on the linear tube model, given our discovery of the mixed oral and nasal rumbles.
We determined the formant spacing ( Δ F ) in each call by calculating the regression slope of the measured center formant frequency positions against the (2i − 1)/2 increments of the formant spacing, as predicted by the vocal tract model [38], using the lm function (package lme4) in Rstudio (Version 1.0.153). The predicted VTL for each call was calculated using the equation V T L = c 2 Δ F and their means (+/−) SD calculated per call type.
Table 2. Acoustic parameters measured and their description.
Table 2. Acoustic parameters measured and their description.
Acoustic ParameterDescription
Duration in sTime from the onset until the end of the vocalization measured from the spectrogram.
Mean Fundamental Frequency (F0) Mean of fundamental frequency values of 60 points spaced evenly across the tracked contour.
Frequency Variability Index (FVI) [75]Calculated variable that represents the magnitude of frequency modulation across a call computed by dividing the variance in frequency by the square of the average frequency of a call and then multiplying the value by 10.
Inflection Factor (IF)Percentage of points along the fundamental frequency’s contour showing a reversal in slope.
Jitter Factor (JF) [75]Calculated variable that represents a weighted measure of the amount of frequency modulation by calculating the sum of the absolute value of the difference between two sequential frequencies divided by the mean frequency. The sum result is then divided by the total number of points measured minus 1 and the final value is obtained by multiplying it by 100.
Dominant Frequency (DF) in Hz Frequency with the highest amplitude peak in the power spectrum.
Wiener EntropyMeasurement of tonality defined as the ratio of a power spectrum’s geometric mean to its arithmetic mean, expressed on a log scale. Lower values relate to higher tonality.
Quartile 25 (Q25)Parameter characterizing the spectral energy distribution, i.e., the frequency value where 25% of the total energy is located below this value.
Quartile 50 (Q50)Parameter characterizing the spectral energy distribution, i.e., the frequency value where 50% of the total energy is located below this value.
Quartile 75 (Q70)Parameter characterizing the spectral energy distribution, i.e., the frequency value where 75% of the total energy is located below this value.
Spectral Centroid Frequency (SCF)Weighted average frequency where the weights are the normalized energy of each frequency component.

2.6. Statistical Analysis

Statistical analyses were conducted in Rstudio. We visually inspected the variable distribution in combined boxplot-violin plots and used the Shapiro–Wilks Test to test for normal distribution. Given that most of the variables were not normally distributed, we report the median along with the mean and standard variation. The same individuals contributed calls unequally to the call type groups, resulting in a non–independent and unbalanced dataset that did not allow for controlling for individuality. Hence, we did not test the single variables for significant differences among the call types. To test for the call type discriminability based on the combined acoustic variables, we first tested for correlations among the variables, and selected the variables that did not corelate strongly with any other variable or one representative variable of a group of strongly correlating (r > 0.86) variables. We ran discriminant function analyses with leave–one–out cross classification on the log and z–transformed variables F0, duration, FVI, formant spacing ( Δ F 1 F 10 ), SPL, Wiener entropy and spectral centroid frequency (SCF). We randomly selected 16 calls per call type while maintaining the maximum sample size of individuals. Since the assumption of homogeneity of variance–covariance matrices between the test groups was not met, we applied quadratic discriminant analysis [76].

3. Results

3.1. Sound Emission

In total, we captured 309 laryngeal calls with the acoustic camera. In 106 calls, the sound source was not clearly defined. Of the 203 laryngeal calls in which the sound source was identified, we recorded 189 rumbles and 14 roars. A total of 120 of 189 (64%) rumbles were emitted nasally, and 42 (22%) orally. We found that five of the orally emitted rumbles began nasally for about one second before transitioning to oral sound emission; thus, we classified them as oral rumbles and analyzed only the oral component acoustically. We classified a total of 27 (14%) calls with simultaneous oral and nasal emission (Figure 2b–e; Video S1, Supplementary Materials); six of them were part of a roar–rumble combination (Figure 3; Video S2, Supplementary Materials). We first discovered mixed oral and nasal emissions in rumbles where the sound source was not clearly detected at the mouth or the trunk tip but in–between, or where the sound sphere appeared not round but distorted, thereby indicating two sources. Here, we selected the upper harmonics only, and consequently found the sound source to be located at the mouth. The sound source appeared to be in–between the trunk and mouth or distorted again when we dragged the selection windows lower, to about 200 to 500 Hz. When selecting the lower harmonics only up to 200 to 500 Hz, the nasal sound emission dominated. We reanalyzed the rumbles that we had initially classified as nasal when we selected the spectrogram with all of the visible harmonics but exhibiting acoustic energy beyond 200 to 500 Hz. This revealed mixed emissions in these rumbles, too. All of the 14 roars were dominantly orally emitted and were part of a roar rumble call combination (Figure 3). However, when selecting the narrower frequency windows in two roar rumble combinations, we also detected evidence for radiation from the front (30 ms, 50–450 Hz, see Figure 3f and Video S3, Supplementary Materials, and 575 ms, 400–970 Hz in the second image see Supplementary Materials Figure S1), where the nostril openings are situated (Figure 1). Here as well as in several rumbles and roars we also detected ground reflections suggesting more complex patterns of sound radiation which, however, go beyond the scope of this study.
Table 3 reports the behavioral context of the call types. Most of the calls (161) were recorded in the context of contact and greeting, with 121 of those resulting from the separation–reunion experiments. The oral, nasal, and mixed calls occurred in both of the contexts. Since we did not balance the data recording across individuals nor contexts, nor corrected them for sample effort, we did not further quantify the contextual call type occurrence. The opening of the mouth defined by the lowering of the jaw was not a reliable indicator of the call emission because the nasal rumbles were also emitted with open mouths and the mixed (oral and nasal) and oral emission of rumbles were recorded with a seemingly closed mouth (Figure 2 and Figure 3; Table 4, Video S1, Supplementary Materials). Only during roar production was the mouth always visible to be slightly or widely open (Figure 3).

3.2. Acoustic Data Analysis

The nasal rumbles differed from the rumbles, with oral emission components in the spectral acoustic variables being concentrated in lower frequency ranges and exhibiting a lower SPL. Nonetheless, the distribution of the F0 and related parameters, tonality and duration overlapped (Table 5). The nasal rumbles had a lower SPL (51.57 ± 5.05 dB) than the mixed oral and nasal (73.5 ± 11.28 dB) and oral rumbles (62.16 ± 11.52 dB). The F0 was slightly lower in the nasal rumbles (15.594 ± 3.634) and the oral rumbles (14.98 ± 3.87 Hz) than in the mixed rumbles (22.39 ± 3.37 Hz). The mean FVI was similar in the nasal (0.07 ± 0.05) and oral (0.07 ± 0.07) and mixed rumbles (0.12 ± 0.07), as was the tonality measured as Wiener entropy (nasal: 0.23 ± 0.13, oral: 0.23 ± 0.09, mixed: 0.25 ± 0.11). The duration was similar in the nasal (5.93 ± 3.42 s) and mixed rumbles (4.12 ± 1.74 s) and lower compared to the oral rumbles (7.75 ± 3.86 s), but with overlapping variance. All of the rumble types differed clearly from the orally emitted roars in all of the measured parameters, with roars exhibiting the highest F0 (170.93 ± 45.41 Hz), SPL (88.7 ± 5.87 dB) and Wiener entropy (0.45 ± 0.13), but the shortest duration (1.42 ± 0.79 s). The frequency modulation parameters (FVI, JF, IF) were not considered in the roars because we could not consistently measure the fundamental frequency contour across the whole call in all of the calls.
The spectral acoustic energy distribution clearly differed. The spectral centroid frequency was lowest in the nasal (121.29 ± 79.97 Hz) compared to the oral (157.93 ± 58.24 Hz) and mixed rumbles (149.83 ± 44.92 Hz) and highest in the roars (417.90 ± 95.06 Hz), as were the Quartiles 25, 50 and 75 (Table 5). The dominant frequency was lowest and less variable in the nasal (16.22 ± 3.55 Hz) compared to the oral (25.64 ± 23.51 Hz) and mixed rumbles (69.99 ± 43.61 Hz) and highest in the roars (505.52 ± 392.27 Hz). The first two formants’ positions were lower and less variable in the nasal (F1 = 21.12 Hz ± 2.71, F2 = 103.34 ± 8.96 Hz) than the mixed rumbles (F1 = 33.51 ± 4.48 Hz, F2 = 110.03 ± 6.82 Hz), but not compared to the oral rumbles (F1 = 28.39 ± 6.58, F2 = 101.26 ± 14.12 Hz). The formant values in the roars were highest (F1 = 188.38 ± 89.26 Hz, F2 = 438.91 ± 99.84 Hz).
The most striking difference was found in the number of the measurable formants. They never reached above two in the nasal but up to ten or more in the rumbles with oral components, and up to twenty or more and a maximum frequency bandwidth of up to 13 kHz in the roars (Figure 2 and Figure 3), with higher formant positions exhibiting considerable variance (Table 5; Figure 4). Accordingly, the formant distribution was measured between the first two formants only in the nasal rumbles (ΔF1–F2 = 81.11 ± 9.42 Hz) and translated into a nasal VTL of 218.76 ± 26.99 cm, which fits into the expected values compared to the female African elephants. Interestingly, these values were comparable to the ones of the oral (ΔF1–F2 70.73 ± 12.91 Hz and VTL of 255.57 ± 47.83 cm) and mixed rumbles (ΔF1–F2 76.66 ± 4.56 Hz and VTL of 229.08 ± 14.26 cm). This confirmed the results of the acoustic camera images showing the dominant emission of the lower frequencies through the nasal vocal tract. The calculated VTL in the roars was around 70 cm, regardless of whether the first two (73.11 ± 14.92 cm), the 3rd to 10th (70.77 ± 15.7 cm) or the 1st to 10th formants (70.88 ± 14.27 cm) were considered, and thus comparable to the VTL measured in the African elephants. When calculating the formant spacing from the 3rd formant to up to the 10th formant or across all of the formants measured, the vocal tract length in the oral (VTL ΔF3–F10 =126.02 ± 26.19 cm and VTL ΔF1–F10 = 151.46 ± 58.11 cm) and the mixed rumbles (VTL ΔF3–F10 = 119.46 ± 36.13 cm and VTL ΔF1–F10 = 126.6 ± 38.13 cm) was in-between the values for the nasal rumbles and roars, reflecting the unequal spacing between the formants.
In the spectra of the mixed and oral rumbles, we frequently observed a drop in energy (dB levels of amplitude peaks) around 200 to 300 Hz after the first two to three formants and an energy plateau around 500 to 700 Hz or 1000 to 1200 Hz. On top of this plateau, the spectral peaks occurred with a spacing similarly even to the spacing between the first two formants (see power spectra, Figure 2). This points to interactive subtractive and additive effects of the oral and nasal resonances.
The discriminate function–analysis testing call–classification within the rumble emission types, based on the parameters F0, FVI, duration, sound pressure level, ΔF1–F10, Wiener entropy and spectral centroid frequency, yielded a cross classification success of 77.08% (chance level 0.33%).

4. Discussion

Deploying an acoustic camera to visualize sound emissions and selecting specific frequency ranges from the spectrogram to segregate the sound sources demonstrated that adult female Asian elephants not only emit nasal rumbles, oral rumbles, and oral roars, but also rumbles that are simultaneously emitted through the mouth and trunk. This is, to our knowledge, the first study to conclusively demonstrate the coupling of oral and nasal cavities in a non-human animal. The spectral features of the nasal rumbles differed strongly from the mixed and oral calls in their acoustic energy concentration in lower harmonics but overlapped in the features of the fundamental frequency, modulation, tonality and call duration. The nasal rumbles exhibited only two formants and overall lower sound pressure levels. The lower two formants in the mixed and oral–only rumbles were in the same frequency range as in the nasal rumbles. This confirms the contribution of nasal resonances to the mixed rumbles and suggests them in the oral–only rumbles, too, although we did not detect nasal emission on the acoustic camera image. The estimated vocal tract length from the formant positions matched expected values when compared to the female African elephants for purely nasal rumbles (2.0 to 2.5 m) and oral roars (about 0.7 m) [19,23,74]. The nasal vocal tract could also be estimated from the first two formants in the mixed rumbles with oral contributions, whereas the dispersion between all or only the upper formants did not yield meaningful vocal tract estimates because of the coupled oral and nasal resonances. While the rumble emission types could be successfully classified by discriminant analysis (77%), the acoustic parameters varied considerably among and within individuals. We demonstrated that, by alternating between the oral and nasal call emission and a coupling of both, the Asian elephants increase the acoustic variability in their most frequently used call type, the rumble. This enlarges the parameter space available for biological information coding. How the Asian elephants perceive and use this flexibility to encode information remains to be investigated.

4.1. Velopharyngeal Coupling and Complex Formant Patterns

Our findings demonstrate that velopharyngeal coupling is not unique to humans and that its acoustic effects of enhancing, dampening and shifting the positions of spectral peaks as compared to purely oral or nasal calls in Asian elephants and humans are similar. The nasal rumbles exhibit only two lower formants, with the first around 20–30 Hz, the purely oral roars up to 20 evenly spaced higher formants, with the first formant around 200 Hz. In contrast, the mixed rumbles have both lower “nasal formants” and higher “oral formants” that are unequally spaced. Similar to human vowel production, where the oral emission is normal and thus represents the reference setting (only nasal consonants are nasally produced), adding nasal resonances leads to an overall annihilation of the spectral energy and to both the additive and subtractive effects in the relative spectral energy composition [24,28,29,36]. These effects result in additional nasal formants (poles) or antiformants (zeros), and thus relative shifts of the formant positions, reducing the formant amplitude and widening formant bandwidths for nasalized versus oral vowels. Both of the vowel types can be further quantified by measuring the relative amplitude of the peak harmonics in the formants and the spectral tilt, that is the slope of decrease in energy in the upper harmonics [28]. Applying these more detailed measurements to the elephant calls in future studies may help to better quantify the observed variability in the oral, nasal and mixed calls. The visual inspection of the “oralized” rumbles (since nasal emission is more common in elephants) with spectral energy valleys and plateaus and non-uniform formant spacing suggests spectral commonalties with the “nasalized” vowels in humans [28]. This formant structure also matches the theoretical models for adding a side branch to a linear tube vocal tract [77]. It further agrees with the potentially nasalized groans modelled for male fallow deer after CT scans of their oral and nasal cavities, where an extended gap between the third and fourth formant was described [78].
These commonalities further support the hypothesis that mammals beyond fallow deer apply velopharyngeal coupling in their vocalizations. This mechanism might be further distributed in the animal kingdom, where non-uniform formant dispersion is frequently described, for example, in ungulates [40]. The lack in data may reflect an oversight due to the complex nature of these mixed vocalizations, and methodological difficulties to identify, analyze and conclusively demonstrate their sound emission and resonances. Velopharyngeal coupling had also been suggested based on relatively lowered formant positions and an amplitude in the calls of Diana monkeys [33], rhesus macaques [32] and elephant seals [34]. In the latter, the amplitude of the first “nasal” formant was also considerably lower than the consecutive “oral” formants, and the nasal contribution in the seals, with their largely elongated and inflated vestibulum during calling, seems convincing. Nonetheless, the studies in humans using physiological and imaging techniques visualizing the velum and/or oral articulators’ position and/or nasal airflow showed that, in general, acoustic data alone cannot conclusively prove oral and nasal coupling. This is because changes in the setting of the vocal tract articulators (e.g., oral or nasal cavity constriction, tongue, lips, mouth aperture) can lead to similar acoustic effects on spectral structure [28,29,36]. Disentangling the influences of the oral or nasal resonances and the oral vocal tract configuration has proven difficult even in these human studies, where, compared to the animal models, the physiological measurements and sample sizes can far more easily be applied and the change of meaning in vocalizations is known. A considerable variation of motor settings achieved the same vocal output across the human subjects, and in the vocal output itself, the acoustic parameters differed within and between subjects. In the elephant study here, we observed a comparably large variation in the dominant frequencies and formant positions that could be attributed to different degrees of velum aperture. However, idiosyncratic morphology, or vocal strategies and articulator configuration, might also cause these effects and are aspects that require further investigation.
Contrary to the orally and nasally mixed emissions, purely oral or nasal emissions were shown in several animals. Their acoustic characteristics are concordant with our results, where the longer nasal tract leads to lower formants’ values and spacing and lower overall sound energy. The lower sound pressure levels and the spectral energy bandwidth of nasal rumbles versus those with oral contribution and the loudest, clearly oral roars, are consistent with the absorption of acoustic energy by the increased surface area and soft mucus-layered tissues within the nasal cavity and sinuses described in nasalized vowels in humans [79] and other animals’ soft nasal (e.g., dog whines or pig grunts) but loud oral calls (e.g., dog bark and pig scream) [27]. The lower formants and concentration of energy in the lower harmonics have further been shown in the nasal calls of sheep [42], goitered gazelles [39], Saiga antelopes [40] and red deer [41], and in the previous study using the acoustic camera on African elephants [23]. Interestingly, trumpets are clearly uttered through the nose but are comparably as loud as the roars and exhibit high energy in upper harmonics (Beeck, unpublished data, [15,16]). This was suggested to be caused by the alternative sound production mechanisms, including the formation of a shockwave [80].

4.2. Rumble Sound Production Mechanisms in African and Asian Elephants

In the African elephants, only oral and nasal emissions with significantly different formant positions but no mixed emissions were detected in the three females, so far. The two juvenile males investigated in the previous study did not utter oral calls, however, the male elephants have been observed to vocalize while the mouth is wide open in bonding contexts, apparently emitting rumbles orally (Stoeger, pers. observation). Only the first two formant positions were considered and the formant values of oral rumbles (F1 = 169.21 ± 25.61 Hz and F2 = 415.20 ± 47.71 Hz) in the African elephants match the values we observed in the oral roars of the Asian elephant females in this present study. The nasal rumbles’ formant values (F2 = 39.79 ± 5.78 Hz and F2 = 128.76 ± 32.57 Hz in the African elephants match the values we observed in the nasal, oral and mixed rumbles in the Asian elephants. As noted above, this indicates that the rumbles we classified as oral based on the acoustic image were also in fact mixed rumbles. The nasal emission may not have shown in the acoustic image because its energy might have been considerably lower than that of the oral part. The other potential explanations are that the elephant might not have been positioned in an optimal angle towards the camera, or that the background noise levels may have affected the resolution of the multiple sources. Similarly, a mixed emission in the African elephants might have been overseen due to the different methodological approaches. As opposed to the current study, where frequency ranges were selected from the spectrogram, the previous study on African elephants in 2012 [23] calculated the acoustic image directly from the waveform. This approach did not allow for the separation of the correlated sound sources, and those calls in which the sound emission was unclear were excluded from the analysis. In addition, the LPC function coefficients assume resonances of a linear tube [33]. In the mixed emission calls, this means that the coefficient selection is more subjective [28]. The different values chosen in the previous study might not have picked up the nasal formants if they were considerably lower in amplitude. In theory, we can assume that the Asian elephants are capable of producing oral rumbles with little or no nasal resonance. In turn, there are no anatomical indications that the African elephants are incapable of emitting mixed rumbles, although the frequency and context of the usage might well differ among the species. Note, however, that only the Asian elephants produce the high-pitched squeaks; we suggested by vibrating their lips [67]. Accordingly, the species differences in sound production cannot be generally excluded.
The acoustic properties of the nasal and “oralized” rumbles described here match the previous distinction of the Asian elephants low-frequency vocalization. De Silva, 2010 [65], described “growls”, as a soft low-frequency vocalization that lacks energy above 500 Hz, and is uttered with depressed cheeks and an almost closed mouth. The rumbles, in turn, have a higher spectral energy, fundamental frequencies, and variation and are uttered with the mouth open in generally more agitated contexts. We showed that mouth opening, as defined by the lowering of the jaw alone, is not a reliable indicator of sound emission. This is because the oral rumbles were also emitted through a seemingly closed (probably only very sightly opened) mouth, while the nasal or mixed rumbles were also emitted with the mouth wide open. This can be explained by the Asian elephant’s ability to tightly close the mouth in the longitudinal direction, since the upper lip is fused into the trunk [67]. This fits with the observation of the depressed cheeks when producing “growls” by de Silva. The more frequent occurrence of “growls” (60%), reported by de Silva, matches our result of 64% nasal rumbles. We expect the nasal rumbles to be produced with the larynx in its resting position—less effortful and allowing the elephant to continue feeding while vocalizing [27]. Our observation that the nasal rumbles are often produced whilst the elephant was visibly chewing (Table 3) supports this notion. In most mammals, the larynx is positioned within the nasopharynx in its resting position during breathing, and is actively retracted during loud calls [27]. While the larynx movement has not yet been studied in living elephants, the unique divided hyoid apparatus of elephants suggests great flexibility and potential for retraction [81]. The energy conservation might be enhanced by the properties of the hydrostatic long and narrow tubes that form the nasal passages in the trunk: an acoustic standing might well be induced during the nasal vocalization [82]. This may result in source filter interactions in which the higher impedance of the nasal compared to the oral pathway is more efficient in feeding power back to the source [83], a hypothesis that requires further research.
In contrast to the spectral similarities between our results and de Silva’s earlier description of the rumbles and growls, she reported differences in the fundamental frequency values and modulation, which in our study overlapped between the nasal, oral or mixed rumbles. This might reflect the fact that she recorded rumbles mainly during agitated situations and visually classified the rumble types. Similarly, the fundamental frequency differed significantly between the oral and nasal rumbles in the African elephant study [23], but both of the studies might have missed the rumbles of mixed emission that bring intermediate values and graded variance to the dataset. From our visual inspection of the rumble types, we can confirm that the more modulated and higher-pitched rumbles occur during agitated greeting. In the future, testing for arousal and behavioral context along with individual variation should help to explain more of the overlapping differences in the acoustic parameters that we observed among the emission types in this study.
Importantly, the discriminant function analysis including the overlapping parameters (fundamental frequency and variability, duration, Wiener entropy) and the differing parameters (formant dispersion, spectral centroid frequencies) still successfully classified the rumbles based on their emission, with a success rate of 77%, high above chance level (0.33%). We were unable, however, to control for individual variation and although different individuals contributed calls to each emission group, this might have led to an over classification. We assume that the variations caused by individually distinct morphology traits of the vocal tract are less prominent than the differences of the oral, nasal or mixed vocal tract resonances across individuals. Nonetheless, the classification results should be considered with caution and replicated with a sample size that allows for the testing of, and control for, individual variation in the rumble emission types.

4.3. Potential Functions of the Various Rumble-Production Types

We suggest that the graded variation in the acoustic parameters from the nasal, oral and mixed laryngeal calls in elephants amplifies the potential for information coding and/or long-distance communication. In the African elephants, the relative position of the first formants encodes relevant information—female estrus rumbles differentiate from non-estrus rumbles by lower first formant frequencies, higher first formant amplitudes and lower fundamental frequencies [60]. The codominant females lowered the second formant, decreasing formant dispersion in the dominance interactions compared to the rumbles produced in the low-affect contexts [84]. The alarm call warning regarding bees is distinguished by an upward shift of the first formant [85], whereas the alarm call warning regarding potential human predators is characterized by an upwards shift of the first and second formant, along with a rise in the fundamental frequencies and greater variability in both of the alarm calls compared to control rumbles [52]. How this spectral variability relates to the active modification of the oral or nasal vocal tract to encode information, and to what extent this variability relates to the physiological effects of heightened arousal remains unclear. Across mammals, the louder calls are in general oral and thus may simply reflect and transmit heightened affect. The “oralized” rumbles described here exhibit significantly more energy (around 1000 Hz and above up to 13 kHz), which matches the maximum hearing sensitivity and the maximal detectable frequency in the currently only published audiogram of a 7-year-old Asian elephant. That individual was more sensitive to low frequencies than previously tested mammals (17 Hz at 60 dB), but less sensitive to frequencies below 100 Hz than to those between 100 Hz and 5 kHz [86]. Adding the higher frequency oral vocal tract resonances to the rumbles may enable the maximal exploitation of the receiver’s attention, at least at close range. In turn, this may help the receiver in identifying or localizing the sender during the greeting ceremonies or the arousal bunching of elephants [87], where individuals join in exuberant choruses with simultaneous and overlapping calls [6], a task commonly referred to as the “cocktail party” problem [88].
Lower frequencies, however, are less prone to reverberation and absorption by vegetation and are better suited for long distance communication [89,90]. The high amplitude of the low formant positions in the rumbles may play a crucial role in the long-distance transmission of relevant information. The playback experiments have shown that the first two formants degenerate less than fundamental frequencies and higher frequencies across distances of 2.5 km in the rumbles of African female [56] and male elephants [91]. In the previous study on call emission in African elephants, the nasal sound emission of rumbles was more prominent during long-distance contact calling and oral emission during close-contact greeting [23]. By combining nasal and oral emission, elephants may be able to convey their signals over long distances while simultaneously expanding the acoustic parameter space for potential information coding over short distances.
We observed additional interesting effects in some of the acoustic images. The first is the radiation of acoustic energy from the front, where the nasal passages enter the skull, in roars and rumbles. The second is ground reflections with sometimes equal or even stronger sound pressure levels than the vocalization itself, which may play a role in sound transmission, including seismic transmission [92]. Future research must determine how the acoustic variability in elephant rumbles, caused by the flexible oral and nasal production mechanism we describe here, translates into biologically significant information. This should be coupled with specifying the role that the vocal tract configuration, including the highly flexible trunk, plays in shaping resonances.

5. Conclusions

Using sound imaging technology, our study is the first to conclusively demonstrate velopharyngeal coupling in a non-human animal, the Asian elephant. This adds to the previously described vocal flexibility of elephants and provides a framework for future research on the information coding and sound transmission in the poorly studied communication system of Asian elephants. We suggest that the vocal tract modifications, including the coupled contribution of oral and nasal resonances, might be more prevalent and complex formant patterns might be more relevant in mammal communication than hitherto investigated. Our findings highlight the significance of analyzing graded call variation, and those calls that do not fit into expected or predefined categories, to better understand the species-specific variability of sound production and information coding. This would be an important step forward in defining universal factors that drive the evolution of communication across species.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/ani12162119/s1, Video S1: Acoustic video of rumble emission types. The video shows the acoustic camera image and the waveform below of rumbles emitted in succession by Pawan (f, 55y) during the separation phase of the separation–reunion experiment. The rumbles emitted are the same depicted in Figure 2, with first a nasal rumble emitted while chewing, second a mixed rumble with the oral emission being dominant and third, a mixed rumble with the nasal emission being dominant. The mixed emission is visible by the colored sound sphere moving in between the mouth and trunk, whereas it remains steady at the trunk tip in the first, purely nasally emitted rumble. Video S2: Setup of acoustic camera and acoustic video of a roar–rumble combination. The video first shows the set-up of the acoustic camera in front of an elephant enclosure. In the following it shows Hira (f, 45y) emitting rumbles during the separation phase of a separation–reunion experiment and the acoustic camera image of the roar–rumble combination depicted in Figure 3, followed by another shot of the camera setup. Video S3: Demonstration of SPL measurement from the acoustic image. This video demonstrates the SPL (here Lp) measurement from the acoustic image including the radiation from the elephant’s front and a strong ground reflection, depicted also in Figure 3f. Figure S1: Second example of a roar–rumble combination. Emitted during the separation phase of a separation–reunion experiment by Champa (f, 41y) showing the acoustic camera images of different time and frequency ranges (a–d) and the corresponding spectrogram: (a) shows the oral only emission of the roar part; (b) the frontal sound radiation during the roar, (c) the partial oral emission of the mixed emission rumble and (d) the nasal emission of the rumble’s lower harmonics.

Author Contributions

Conceptualization, V.C.B. and A.S.S.; methodology, V.C.B.; software, G.H.; validation, G.H.; formal analysis, V.C.B.; investigation, V.C.B., G.H. and M.K.; resources, G.H. and A.S.S.; data curation, V.C.B.; writing—original draft preparation, V.C.B.; writing—review and editing, V.C.B. and A.S.S.; visualization, V.C.B. and G.H.; supervision, A.S.S.; project administration, V.C.B.; funding acquisition, A.S.S. and V.C.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Austrian Science Fund (FWF) project number W1262-B29. Veronika C. Beeck was further sponsored by a VDS CoBeNe final fellowship. Angela S. Stoeger was funded by the Austrian Science Fund (FWF) project number P 31034-B29 and a Marie Jahoda-Scholarship.

Institutional Review Board Statement

The animal study protocol was approved by the Animal Ethics and Experimentation Board of the Faculty of Life Sciences from the University of Vienna (No. 2018-001).

Informed Consent Statement

All animal owners gave written and oral consent to the conducted experiments, the data collection, and publication of the data.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

We thank Tiger Tops Nepal, specifically Kristján Edwards, Jack Edwards, Marie Stissing Jensen and Reshmi Parajuli, for enabling us to do our research, and D.B. Chaudhary, Ashik Gurung, Shambu Mahato, Bhim Thanet and the whole staff for their great support in data collection. We thank gfai Tech GmbH, Berlin, for providing the acoustic camera and analysis program. We thank Sabina Shreshta for acting as an interpreter and Natália Cerqueira for documenting our field work on film. We further thank Michel Stachowitsch and Hoi-Lam Jim for revising the manuscript linguistically and two anonymous reviewers for their helpful comments on the earlier version of the manuscript.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Titze, I.R. Principles of Voice Production (Second Printing); National Center for Voice and Speech: Iowa City, IA, USA, 2000. [Google Scholar]
  2. Fitch, W. Production of vocalizations in mammals. Vis. Commun. 2006, 3, 145. [Google Scholar]
  3. Pretorius, Y.; de Boer, W.F.; Kortekaas, K.; van Wijngaarden, M.; Grant, R.C.; Kohi, E.M.; Mwakiwa, E.; Slotow, R.; Prins, H.H.T. Why elephant have trunks and giraffe long tongues: How plants shape large herbivore mouth morphology. Acta Zool. 2016, 97, 246–254. [Google Scholar] [CrossRef]
  4. Schulz, A.K.; Ning Wu, J.; Ha, S.Y.S.; Kim, G.; Braccini Slade, S.; Rivera, S.; Reidenberg, J.S.; Hu, D.L. Suction feeding by elephants. J. R. Soc. Interface 2021, 18, 20210215. [Google Scholar] [CrossRef] [PubMed]
  5. Dagenais, P.; Hensman, S.; Haechler, V.; Milinkovitch, M.C. Elephants evolved strategies reducing the biomechanical complexity of their trunk. Curr. Biol. 2021, 31, 4727–4737. [Google Scholar] [CrossRef] [PubMed]
  6. Poole, J.H. Behavioral contexts of elephant acoustic communication. In The Amboseli Elephants: A Long-Term Perspective on a Long-Lived Mammal; The University of Chicago: Chicago, IL, USA, 2011; pp. 125–161. [Google Scholar]
  7. Bates, L.A.; Sayialel, K.N.; Njiraini, N.W.; Poole, J.H.; Moss, C.J.; Byrne, R.W. African elephants have expectations about the locations of out-of-sight family members. Biol. Lett. 2008, 4, 34–36. [Google Scholar] [CrossRef]
  8. Purkart, L.; Tuff, J.M.; Shah, M.; Kaufmann, L.V.; Altringer, C.; Maier, E.; Schneeweiß, U.; Tunckol, E.; Eigen, L.; Holtze, S.; et al. Trigeminal ganglion and sensory nerves suggest tactile specialization of elephants. Curr. Biol. 2022, 32, 904–910.e903. [Google Scholar] [CrossRef]
  9. Schulte, B.A.; LaDue, C.A. The Chemical Ecology of Elephants: 21st Century Additions to Our Understanding and Future Outlooks. Animals 2021, 11, 2860. [Google Scholar] [CrossRef]
  10. Stoeger, A.S.; de Silva, S. African and Asian Elephant Vocal Communication: A Cross-Species Comparison. In Biocommunication of Animals; Witzany, G., Ed.; Springer: Dordrecht, The Netherlands, 2014; pp. 21–39. [Google Scholar]
  11. Pardo, M.A.; Poole, J.H.; Stoeger, A.S.; Wrege, P.H.; O’Connell-Rodwell, C.E.; Padmalal, U.K.; de Silva, S. Differences in combinatorial calls among the 3 elephant species cannot be explained by phylogeny. Behav. Ecol. 2019, 30, 809–820. [Google Scholar] [CrossRef]
  12. Poole, J.H.; Tyack, P.L.; Stoeger-Horwath, A.S.; Watwood, S. Elephants are capable of vocal learning. Nature 2005, 434, 455–456. [Google Scholar] [CrossRef]
  13. Stoeger, A.S.; Mietchen, D.; Oh, S.; de Silva, S.; Herbst, C.T.; Kwon, S.; Fitch, W.T. An Asian elephant imitates human speech. Curr. Biol. 2012, 22, 2144–2148. [Google Scholar] [CrossRef]
  14. Poole, J.; Granli, P. The Elephant Ethogram: A library of African elephant behaviour. Pachyderm 2021, 62, 105–111. [Google Scholar]
  15. Fuchs, E.; Beeck, V.C.; Baotic, A.; Stoeger, A.S. Acoustic structure and information content of trumpets in female Asian elephants (Elephas maximus). PLoS ONE 2021, 16, e0260284. [Google Scholar] [CrossRef] [PubMed]
  16. Stoeger, A.S.; Baotic, A.; Heilmann, G. Vocal Creativity in Elephant Sound Production. Biology 2021, 10, 750. [Google Scholar] [CrossRef]
  17. Herbst, C.T.; Stoeger, A.S.; Frey, R.; Lohscheller, J.; Titze, I.R.; Gumpenberger, M.; Fitch, W.T. How low can you go? Physical production mechanism of elephant infrasonic vocalizations. Science 2012, 337, 595–599. [Google Scholar] [CrossRef] [PubMed]
  18. Poole, J.H.; Payne, K.; Langbauer, W.R.; Moss, C.J. The social contexts of some very low frequency calls of African elephants. Behav. Ecol. Sociobiol. 1988, 22, 385–392. [Google Scholar] [CrossRef]
  19. Soltis, J. Vocal communication in African elephants (Loxodonta africana). Zoo Biol. 2010, 29, 192–209. [Google Scholar] [CrossRef]
  20. Archie, E.A.; Moss, C.J.; Alberts, S.C. The ties that bind: Genetic relatedness predicts the fission and fusion of social groups in wild African elephants. Proc. Biol. Sci. 2006, 273, 513–522. [Google Scholar] [CrossRef] [PubMed]
  21. Nandini, S.; Keerthipriya, P.; Vidya, T.N.C. Group size differences may mask underlying similarities in social structure: A comparison of female elephant societies. Behav. Ecol. 2017, 29, 145–159. [Google Scholar] [CrossRef]
  22. Turkalo, A.K.; Wrege, P.H.; Wittemyer, G. Long-Term Monitoring of Dzanga Bai Forest Elephants: Forest Clearing Use Patterns. PLoS ONE 2013, 8, e85154. [Google Scholar] [CrossRef]
  23. Stoeger, A.S.; Heilmann, G.; Zeppelzauer, M.; Ganswindt, A.; Hensman, S.; Charlton, B.D. Visualizing sound emission of elephant vocalizations: Evidence for two rumble production types. PLoS ONE 2012, 7, e48907. [Google Scholar] [CrossRef]
  24. Fant, G. Acoustic Theory of Speech Production; Mouton & Co.: The Hague, The Netherlands, 1960. [Google Scholar]
  25. Fitch, W.T. The evolution of speech: A comparative review. Trends Cogn. Sci. 2000, 4, 258–267. [Google Scholar] [CrossRef]
  26. Taylor, A.M.; Charlton, B.D.; Reby, D. Vocal production by terrestrial mammals: Source, filter, and function. In Vertebrate Sound Production and Acoustic Communication; Suthers, R.A., Fitch, W.T., Fay, R.R., Popper, A.N., Eds.; Springer: Berlin/Heidelberg, Germany, 2016; pp. 229–259. [Google Scholar]
  27. Tecumseh Fitch, W. The Phonetic Potential of Nonhuman Vocal Tracts: Comparative Cineradiographic Observations of Vocalizing Animals. Phonetica 2000, 57, 205–218. [Google Scholar] [CrossRef] [PubMed]
  28. Styler, W. On the acoustical features of vowel nasality in English and French. J. Acoust. Soc. Am. 2017, 142, 2469–2482. [Google Scholar] [CrossRef] [PubMed]
  29. Carignan, C. Using ultrasound and nasalance to separate oral and nasal contributions to formant frequencies of nasalized vowels. J. Acoust. Soc. Am. 2018, 143, 2588–2601. [Google Scholar] [CrossRef] [PubMed]
  30. Fitch, W.T. Comparative vocal production and the evolution of speech: Reinterpreting the descent of the larynx. In The Transition to Language; Wray, A., Ed.; Oxford University Press: Oxford, UK, 2002; pp. 21–45. [Google Scholar]
  31. Boë, L.-J.; Sawallis, T.R.; Fagot, J.; Badin, P.; Barbier, G.; Captier, G.; Ménard, L.; Heim, J.-L.; Schwartz, J.-L. Which way to the dawn of speech?: Reanalyzing half a century of debates and data in light of speech science. Sci. Adv. 2019, 5, eaaw3916. [Google Scholar] [CrossRef] [PubMed]
  32. Hauser, M.D. Articulatory and social factors influence the acoustic structure of rhesus monkey vocalizations: A learned mode of production? J. Acoust. Soc. Am. 1992, 91, 2175–2179. [Google Scholar] [CrossRef]
  33. Riede, T.; Zuberbuhler, K. The relationship between acoustic structure and semantic information in Diana monkey alarm vocalization. J. Acoust. Soc. Am. 2003, 114, 1132–1142. [Google Scholar] [CrossRef]
  34. Sanvito, S.; Galimberti, F.; Miller, E.H. Vocal signalling of male southern elephant seals is honest but imprecise. Anim. Behav. 2007, 73, 287–299. [Google Scholar] [CrossRef]
  35. Charlton, B.D.; Reby, D.; McComb, K. Female perception of size-related formant shifts in red deer, Cervus elaphus. Anim. Behav. 2007, 74, 707–714. [Google Scholar] [CrossRef]
  36. Carignan, C. An acoustic and articulatory examination of the “oral” in “nasal”: The oral articulations of French nasal vowels are not arbitrary. J. Phon. 2014, 46, 23–33. [Google Scholar] [CrossRef]
  37. Fitch, W.T. Vocal tract length and formant frequency dispersion correlate with body size in rhesus macaques. J. Acoust. Soc. Am. 1997, 102, 1213–1222. [Google Scholar] [CrossRef] [PubMed]
  38. Reby, D.; McComb, K. Anatomical constraints generate honesty: Acoustic cues to age and weight in the roars of red deer stags. Anim. Behav. 2003, 65, 519–530. [Google Scholar] [CrossRef]
  39. Volodin, I.A.; Lapshina, E.N.; Volodina, E.V.; Frey, R.; Soldatova, N.V. Nasal and Oral Calls in Juvenile Goitred Gazelles (Gazella subgutturosa) and their Potential to Encode Sex and Identity. Ethology 2011, 117, 294–308. [Google Scholar] [CrossRef]
  40. Volodin, I.A.; Sibiryakova, O.V.; Kokshunova, L.E.; Frey, R.; Volodina, E.V. Nasal and oral calls in mother and young trunk-nosed saiga antelopes, Saiga tatarica. Bioacoustics 2014, 23, 79–98. [Google Scholar] [CrossRef]
  41. Sibiryakova, O.V.; Volodin, I.A.; Matrosova, V.A.; Volodina, E.V.; Garcia, A.J.; Gallego, L.; Landete-Castillejos, T. The power of oral and nasal calls to discriminate individual mothers and offspring in red deer, Cervus elaphus. Front. Zool. 2015, 12, 2. [Google Scholar] [CrossRef] [PubMed]
  42. Sèbe, F.; Duboscq, J.; Aubin, T.; Ligout, S.; Poindron, P. Early vocal recognition of mother by lambs: Contribution of low- and high-frequency vocalizations. Anim. Behav. 2010, 79, 1055–1066. [Google Scholar] [CrossRef]
  43. Gamba, M.; Friard, O.; Giacoma, C. Vocal Tract Morphology Determines Species-Specific Features in Vocal Signals of Lemurs (Eulemur). Int. J. Primatol. 2012, 33, 1453–1466. [Google Scholar] [CrossRef]
  44. Charlton, B.D.; Pisanski, K.; Raine, J.; Reby, D. Coding of Static Information in Terrestrial Mammal Vocal Signals. In Coding Strategies in Vertebrate Acoustic Communication; Aubin, T., Mathevon, N., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 115–136. [Google Scholar]
  45. Charlton, B.D.; Reby, D. The evolution of acoustic size exaggeration in terrestrial mammals. Nat. Commun. 2016, 7, 12739. [Google Scholar] [CrossRef]
  46. Hauser, M.; Ybarra, M.S. The role of lip configuration in monkey vocalizations: Experiments using xylocaine as a nerve block. Brain Lang. 1994, 46, 232–244. [Google Scholar] [CrossRef]
  47. Fitch, W.T.; Reby, D. The descended larynx is not uniquely human. Proc. Biol. Sci. 2001, 268, 1669–1675. [Google Scholar] [CrossRef]
  48. Koda, H.; Murai, T.; Tuuga, A.; Goossens, B.; Nathan, S.K.S.S.; Stark, D.J.; Ramirez, D.A.R.; Sha, J.C.M.; Osman, I.; Sipangkui, R.; et al. Nasalization by Nasalis larvatus: Larger noses audiovisually advertise conspecifics in proboscis monkeys. Sci. Adv. 2018, 4, eaaq0250. [Google Scholar] [CrossRef] [PubMed]
  49. Frey, R.; Volodin, I.; Volodina, E. A nose that roars: Anatomical specializations and behavioural features of rutting male saiga. J. Anat. 2007, 211, 717–736. [Google Scholar] [CrossRef] [PubMed]
  50. Briefer, E.F. Coding for ‘Dynamic’information: Vocal expression of emotional arousal and valence in non-human animals. In Coding Strategies in Vertebrate Acoustic Communication; Aubin, T., Mathevon, N., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 137–162. [Google Scholar]
  51. Townsend, S.W.; Charlton, B.D.; Manser, M.B. Acoustic cues to identity and predator context in meerkat barks. Anim. Behav. 2014, 94, 143–149. [Google Scholar] [CrossRef]
  52. Soltis, J.; King, L.E.; Douglas-Hamilton, I.; Vollrath, F.; Savage, A. African elephant alarm calls distinguish between threats from humans and bees. PLoS ONE 2014, 9, e89403. [Google Scholar] [CrossRef]
  53. Langbauer, W., Jr. Elephant communication. Zoo Biol. 2000, 19, 425–445. [Google Scholar] [CrossRef]
  54. Leong, K.M.; Ortolani, A.; Burks, K.D.; Mellen, J.D.; Savage, A. Quantifying Acoustic and Temporal Characteristics of Vocalizations for a Group of Captive African Elephantsloxodonta Africana. Bioacoustics 2003, 13, 213–231. [Google Scholar] [CrossRef]
  55. Wood, J.D.; McCowan, B.; Langbauer, W.R.; Viljoen, J.J.; Hart, L.A. Classification of African Elephantloxodonta Africanarumbles Using Acoustic Parameters and Cluster Analysis. Bioacoustics 2005, 15, 143–161. [Google Scholar] [CrossRef]
  56. McComb, K.; Reby, D.; Baker, L.; Moss, C.; Sayialel, S. Long-distance communication of acoustic cues to social identity in African elephants. Anim. Behav. 2003, 65, 317–329. [Google Scholar] [CrossRef]
  57. Stoeger, A.S.; Baotic, A. Male African elephants discriminate and prefer vocalizations of unfamiliar females. Sci. Rep. 2017, 7, 46414. [Google Scholar] [CrossRef]
  58. Baotic, A.; Stoeger, A.S. Sexual dimorphism in African elephant social rumbles. PLoS ONE 2017, 12, e0177411. [Google Scholar] [CrossRef]
  59. Stoeger, A.S.; Baotic, A. Information content and acoustic structure of male African elephant social rumbles. Sci. Rep. 2016, 6, 27585. [Google Scholar] [CrossRef] [PubMed]
  60. Soltis, J.; Leong, K.; Savage, A. African elephant vocal communication II: Rumble variation reflects the individual identity and emotional state of callers. Anim. Behav. 2005, 70, 589–599. [Google Scholar] [CrossRef]
  61. Soltis, J.; Blowers, T.E.; Savage, A. Measuring positive and negative affect in the voiced sounds of African elephants (Loxodonta africana). J. Acoust. Soc. Am. 2011, 129, 1059–1066. [Google Scholar] [CrossRef]
  62. Wesolek, C.M.; Soltis, J.; Leighty, K.A.; Savage, A. Infant African Elephant Rumble Vocalizations Vary According to Social Interactions with Adult Females. Bioacoustics 2009, 18, 227–239. [Google Scholar] [CrossRef]
  63. Payne, K.B.; Langbauer, W.R.; Thomas, E.M. Infrasonic calls of the Asian elephant (Elephas maximus). Behav. Ecol. Sociobiol. 1986, 18, 297–301. [Google Scholar] [CrossRef]
  64. Nair, S.; Balakrishnan, R.; Seelamantula, C.S.; Sukumar, R. Vocalizations of wild Asian elephants (Elephas maximus): Structural classification and social context. J. Acoust. Soc. Am. 2009, 126, 2768–2778. [Google Scholar] [CrossRef]
  65. de Silva, S. Acoustic communication in the Asian elephant, Elephas maximus maximus. Behaviour 2010, 147, 825–852. [Google Scholar] [CrossRef]
  66. Sharma, N.; Prakash, S.V.; Kohshima, S.; Sukumar, R. Asian elephants modulate their vocalizations when disturbed. Anim. Behav. 2020, 160, 99–111. [Google Scholar] [CrossRef]
  67. Beeck, V.C.; Heilmann, G.; Kerscher, M.; Stoeger, A.S. A novel theory of Asian elephant high-frequency squeak production. BMC Biol. 2021, 19, 121. [Google Scholar] [CrossRef] [PubMed]
  68. Olson, D.E.A. Ethogram of elephant behaviors. In Elephant Husbandry Resource Guide; Olson, D., Ed.; Allen Press: Lawrence, KS, USA, 2004. [Google Scholar]
  69. Plotnik, J.M.; de Waal, F.B. Asian elephants (Elephas maximus) reassure others in distress. PeerJ 2014, 2, e278. [Google Scholar] [CrossRef] [PubMed]
  70. Zala, S.M.; Reitschmidt, D.; Noll, A.; Balazs, P.; Penn, D.J. Automatic mouse ultrasound detector (A-MUD): A new tool for processing rodent vocalizations. PLoS ONE 2017, 12, e0181200. [Google Scholar] [CrossRef] [PubMed]
  71. Stoeger, A.S.; Zeppelzauer, M.; Baotic, A. Age-group estimation in free-ranging African elephants based on acoustic cues of low-frequency rumbles. Bioacoustics 2014, 23, 231–246. [Google Scholar] [CrossRef] [PubMed]
  72. Team, R.C. R: A language and environment for statistical computing. In R Foundation for Statistics; The R Project for Statistical Computing: Vienna, Austria, 2020. [Google Scholar]
  73. Anikin, A. Soundgen: An open-source tool for synthesizing nonverbal vocalizations. Behav. Res. Methods 2019, 51, 778–792. [Google Scholar] [CrossRef] [PubMed]
  74. Sikes, S.K. Natural History of the African Elephant; Weidenfield & Nicolson: London, UK, 1971. [Google Scholar]
  75. Mitani, J.C.; Brandt, K.L. Social factors influence the acoustic variability in the long-distance calls of male chimpanzees. Ethology 1994, 96, 233–252. [Google Scholar] [CrossRef]
  76. Tabachnick, B.; Fidell, L. Using Multivariate Statistics, 6th ed.; Pearson: Boston, MA, USA, 2013. [Google Scholar]
  77. Jackson, M.T.-T.; Espy-Wilson, C.; Boyce, S.E. Verifying a vocal tract model with a closed side-branch. J. Acoust. Soc. Am. 2001, 109, 2983–2987. [Google Scholar] [CrossRef]
  78. Reby, D.; Wyman, M.T.; Frey, R.; Charlton, B.D.; Dalmont, J.P.; Gilbert, J. Vocal tract modelling in fallow deer: Are male groans nasalized? J. Exp. Biol. 2018, 221, 179416. [Google Scholar] [CrossRef]
  79. Stevens, K.N. Acoustic Phonetics; MIT Press: Cambridge, MA, USA, 2000; Volume 30. [Google Scholar]
  80. Gilbert, J.; Dalmont, J.-P.; Potier, R.; Reby, D. Is Nonlinear Propagation Responsible for the Brassiness of Elephant Trumpet Calls? Acta Acust. united Acust. 2014, 100, 734–738. [Google Scholar] [CrossRef]
  81. Shoshani, J.; Tassy, P. Advances in proboscidean taxonomy & classification, anatomy & physiology, and ecology & behavior. Quat. Int. 2005, 126–128, 5–20. [Google Scholar] [CrossRef]
  82. Fabre, B.; Gilbert, J.; Hirschberg, A.; Pelorson, X. Aeroacoustics of Musical Instruments. Annu. Rev. Fluid Mech. 2012, 44, 1–25. [Google Scholar] [CrossRef]
  83. Titze, I.R.; Palaparthi, A.; Cox, K.; Stark, A.; Maxfield, L.; Manternach, B. Vocalization with semi-occluded airways is favorable for optimizing sound production. PLoS Comput. Biol. 2021, 17, e1008744. [Google Scholar] [CrossRef]
  84. Soltis, J.; Leighty, K.A.; Wesolek, C.M.; Savage, A. The expression of affect in African elephant (Loxodonta africana) rumble vocalizations. J. Comp. Psychol. 2009, 123, 222–225. [Google Scholar] [CrossRef]
  85. King, L.E.; Soltis, J.; Douglas-Hamilton, I.; Savage, A.; Vollrath, F. Bee threat elicits alarm call in African elephants. PLoS ONE 2010, 5, e10346. [Google Scholar] [CrossRef] [PubMed]
  86. Heffner, R.S.; Heffner, H.E. Hearing in the elephant (Elephas maximus): Absolute sensitivity, frequency discrimination, and sound localization. J. Comp. Physiol. Psychol. 1982, 96, 926. [Google Scholar] [CrossRef] [PubMed]
  87. Monson, B.B.; Rock, J.; Schulz, A.; Hoffman, E.; Buss, E. Ecological cocktail party listening reveals the utility of extended high-frequency hearing. Hear. Res. 2019, 381, 107773. [Google Scholar] [CrossRef] [PubMed]
  88. Cherry, E.C. Some Experiments on the Recognition of Speech, with One and with Two Ears. J. Acoust. Soc. Am. 1953, 25, 975–979. [Google Scholar] [CrossRef]
  89. Ey, E.; Fischer, J. The “Acoustic Adaptation Hypothesis”—A Review of the Evidence from Birds, Anurans and Mammals. Bioacoustics 2009, 19, 21–48. [Google Scholar] [CrossRef]
  90. Morton, E.S. Ecological Sources of Selection on Avian Sounds. Am. Nat. 1975, 109, 17–34. [Google Scholar] [CrossRef]
  91. Baotic, A.; Garcia, M.; Boeckle, M.; Stoeger, A. Field Propagation Experiments of Male African Savanna Elephant Rumbles: A Focus on the Transmission of Formant Frequencies. Animals 2018, 8, 167. [Google Scholar] [CrossRef]
  92. O’Connell-Rodwell, C.E.; Wood, J.D.; Kinzley, C.; Rodwell, T.C.; Poole, J.H.; Puria, S. Wild African elephants (Loxodonta africana) discriminate between familiar and unfamiliar conspecific seismic alarm calls. J. Acoust. Soc. Am. 2007, 122, 823–830. [Google Scholar] [CrossRef] [PubMed]
Figure 2. Examples of rumble emission types. Rumbles emitted in succession by Pawan (f, 55y) during the separation phase of the separation–reunion experiment showing the acoustic camera images (left), spectrograms (middle) and power spectra (right). The LPC function lines in red indicate the formant positions (F1–F13). (a) Nasal rumble; (b) Mixed emission, with oral emission but with a distorted sound sphere appearing when the whole call is selected; (c) Nasal part of the mixed rumble when lower harmonics are selected; (d) Oral part of a mixed rumble when only upper harmonics are selected; with (e) a more dominant nasal part when only the lower harmonics are selected. Note that during (a) nasal rumble emission the elephant was chewing and the mouth at times wide open as defined by jaw lowering.
Figure 2. Examples of rumble emission types. Rumbles emitted in succession by Pawan (f, 55y) during the separation phase of the separation–reunion experiment showing the acoustic camera images (left), spectrograms (middle) and power spectra (right). The LPC function lines in red indicate the formant positions (F1–F13). (a) Nasal rumble; (b) Mixed emission, with oral emission but with a distorted sound sphere appearing when the whole call is selected; (c) Nasal part of the mixed rumble when lower harmonics are selected; (d) Oral part of a mixed rumble when only upper harmonics are selected; with (e) a more dominant nasal part when only the lower harmonics are selected. Note that during (a) nasal rumble emission the elephant was chewing and the mouth at times wide open as defined by jaw lowering.
Animals 12 02119 g002
Figure 3. Example of a roar–rumble combination. Emitted during the separation phase of a separation–reunion experiment by Hira (f, 45y). This shows acoustic camera images of different time and frequency ranges (af) and the corresponding spectrogram, with a zoom–in on the rumble part (e,f) and the power spectrum (a,b) taken in the middle of the roar between selection window a and b. The LPC function line in red indicates the formant positions (F1–F9); (a,b,d) show the oral only emission of the roar parts; (c) shows the partial oral emission of the mixed emission rumble part, exhibiting a distorted sphere and the (e) nasal emission of the lower harmonics; (f) shows a pattern of simultaneous sound radiation over the front, the mouth and trunk, as well as a strong ground reflection here at the transition between roar and rumble.
Figure 3. Example of a roar–rumble combination. Emitted during the separation phase of a separation–reunion experiment by Hira (f, 45y). This shows acoustic camera images of different time and frequency ranges (af) and the corresponding spectrogram, with a zoom–in on the rumble part (e,f) and the power spectrum (a,b) taken in the middle of the roar between selection window a and b. The LPC function line in red indicates the formant positions (F1–F9); (a,b,d) show the oral only emission of the roar parts; (c) shows the partial oral emission of the mixed emission rumble part, exhibiting a distorted sphere and the (e) nasal emission of the lower harmonics; (f) shows a pattern of simultaneous sound radiation over the front, the mouth and trunk, as well as a strong ground reflection here at the transition between roar and rumble.
Animals 12 02119 g003
Figure 4. Formant position values in Hz per call and emission type plotted against the formant increment. The regression line gives the values for the formant spacing (ΔF1–F10) that is used to calculate the vocal tract length. Formant position values varied more among calls in higher formants.
Figure 4. Formant position values in Hz per call and emission type plotted against the formant increment. The regression line gives the values for the formant spacing (ΔF1–F10) that is used to calculate the vocal tract length. Formant position values varied more among calls in higher formants.
Animals 12 02119 g004
Table 1. Subjects and call sample sizes. Number of call types combined with emission per subject; the subjects estimated age according to their handlers at the time of recording; and the subjects shoulder height as a mean value of two measurements.
Table 1. Subjects and call sample sizes. Number of call types combined with emission per subject; the subjects estimated age according to their handlers at the time of recording; and the subjects shoulder height as a mean value of two measurements.
Call Types and Emission
Subject Names Age
(Years)
Shoulder Height (m)Rumble NasalRumble Oral and NasalRumble
Oral
Roar
Oral
Sum
Champa 412.40249101053
Chan Chun452.4750005
Dhibya482.5011016027
Dipendra 602.432822032
Hira452.55690419
Pawan 552.413640040
Raj422.492114017
Saraswati272.4010001
Sunder462.4172009
Sum 120274214203
Table 3. Behavioral context. Number of call types combined with call emission per behavioral context.
Table 3. Behavioral context. Number of call types combined with call emission per behavioral context.
Rumble NasalRumble Oral and NasalRumble
Oral
Roar
Oral
Sum
Arousal1800018
Command1725024
Contact26156552
Greeting5910319109
Sum120274214203
Table 4. Mouth opening defined by visual jaw lowering during call emission.
Table 4. Mouth opening defined by visual jaw lowering during call emission.
Rumble NasalRumble Oral and NasalRumble OralRoar Oral
Closed65380
Slightly open191383
Wide open69129
Chewing14050
Unknown16292
Table 5. Results of acoustic parameters measured per call emission type. The table shows sample sizes, median and mean values ± SD.
Table 5. Results of acoustic parameters measured per call emission type. The table shows sample sizes, median and mean values ± SD.
Rumble NasalRumble Oral and NasalRumble OralRoar Oral
ParametersNMedianMean ± SDNMedianMean ± SDNMedianMean ± SDNMedianMean ± SD
Duration in s1205.255.93 ± 3.42273.944.12 ± 1.74427.367.753 ± 3.86141.281.42 ± 0.79
SPL in dB7251.751.57 ± 5.05247473.5 ± 11.282765.262.159 ± 11.5241288.788.7 ± 5.87
Mean F07514.9614.98 ± 3.872121.7822.39 ± 3.373715.7815.594 ± 3.63411172.4170.93 ± 45.41
FVI750.060.074 ± 0.05210.120.12 ± 0.07370.060.074 ± 0.068110.04-
IF750.390.39 ± 0.11210.270.28 ± 0.12370.350.33 ± 0.11110.17-
JF753.053.28 ± 1.08213.163.27 ± 0.92372.72.74 ± 1.09112.07-
DF in Hz6416.616.22 ± 3.552097.8169.99 ± 43.612718.5525.643 ± 23.5114378.29505.52 ± 392.27
Spectral
Wiener Entropy660.170.23 ±0.13210.210.25± 0.11290.200.23± 0.09150.470.45 ± 0.13
Q25 in Hz6613.6820.80± 53.342142.1053.93± 28.462927.3630.93± 12.2115300.00278.36± 111.92
Q50 in Hz6634.7353.52± 82.402192.63102.46± 48.842963.1570.54± 27.8715408.42388.93± 138.15
Q 75 in Hz6684.73142.68± 129.5821134.73185.67± 105.9429140.00173.92± 99.8415510.52529.67± 112.71
SPF in Hz6693.74121.29± 79.9721133.44157.93±58.2429148.06149.83± 44.9215421.55417.90± 95.06
Formant
F1 in Hz6121.4821.12 ± 2.711933.233.51 ± 4.482826.8628.39 ± 6.5813157.47188.38 ± 89.26
F2 in Hz34103.25103.34 ± 8.9618110.84110.03 ± 6.8219105.47101.26 ± 14.1213405.03438.91 ± 99.84
F3 in Hz00017242.19260.28 ± 56.4716240.23259.32 ± 83.6613631.71692.63 ± 161.82
F4 in Hz00017437.5408.72 ± 75.5914464.36440.64 ± 79.1313869.04936.07 ± 200.37
F5 in Hz00016598.14583.26 ± 125.4413545.9577.96 ± 117.67131097.531244.10 ± 335.26
F6 in Hz00016750.49755.17 ± 187.5612704.59694.59 ± 144.73131314.151467.41 ± 389.24
F7 in Hz00015889.65908.2 ± 196.6712814.45823.24 ± 176.55131644.111719.8 ± 410.19
F8 in Hz00014960.941030.41 ± 229.549878.91899.31 ± 160.21131881.591949.07 ± 439.4
F9 in Hz000141145.511162.25 ± 243.2461021.61033.32 ± 155.65132139.032270.22 ± 540.77
F10 in Hz000131244.141293.17 ± 289.6661127.931167.91 ± 199.95132315.922533.59 ± 555.53
ΔF1–F10 in Hz34-81.11 ± 9.4218-148.99 ± 40.0219-127.61 ± 36.3814-258.95 ± 66.65
ΔF1–F2 in Hz34-81.11 ± 9.4218-76.66 ± 4.5619-70.73 ± 12.9114-249.73 ± 56.82
ΔF–F10 in Hz0-017-160.96 ± 53.8514-144.93 ± 31.9914-261.33 ± 71.15
VTL
ΔF1–F10 in cm34-218.76 ± 26.9918-126.6 ± 38.1319-151.46 ± 58.1114-70.88 ± 14.27
ΔF1–F2 in cm34-218.76 ± 26.9918-229.08 ± 14.2619-255.57 ± 47.8314-73.11 ± 14.92
ΔF3–F10 in cm0-017-119.46 ± 36.1314-126.02 ± 26.1914-70.77 ± 15.7
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Beeck, V.C.; Heilmann, G.; Kerscher, M.; Stoeger, A.S. Sound Visualization Demonstrates Velopharyngeal Coupling and Complex Spectral Variability in Asian Elephants. Animals 2022, 12, 2119. https://doi.org/10.3390/ani12162119

AMA Style

Beeck VC, Heilmann G, Kerscher M, Stoeger AS. Sound Visualization Demonstrates Velopharyngeal Coupling and Complex Spectral Variability in Asian Elephants. Animals. 2022; 12(16):2119. https://doi.org/10.3390/ani12162119

Chicago/Turabian Style

Beeck, Veronika C., Gunnar Heilmann, Michael Kerscher, and Angela S. Stoeger. 2022. "Sound Visualization Demonstrates Velopharyngeal Coupling and Complex Spectral Variability in Asian Elephants" Animals 12, no. 16: 2119. https://doi.org/10.3390/ani12162119

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop