1. Introduction
Beyond its long-standing relevance in biological research, sound directionality is increasingly recognized as a key dimension for understanding the structure, functioning, and equity of sonic environments in both natural and human-dominated landscapes. The direction from which an acoustic signal originates represents one of the most fundamental cues in ecoacoustics and in various fields of ecological surveying, such as in the study of bird communities, where it has traditionally been used to delineate breeding territories and describe the spatial behavior of vocal species [
1,
2,
3,
4,
5,
6].
The knowledge of sound direction within a landscape is crucial because sound is not merely a physical phenomenon but a spatial, ecological, and perceptual process that reflects both how environments function and how they are experienced by human and non-human organisms. Sound directionality enables the interpretation of ecological landscape structure by supporting the localization of vocal individuals and populations, the reconstruction of territories, the investigation of social interactions, and the assessment of reproductive dynamics [
7]. In this sense, sound acts as a spatial indicator of biological distribution and ecological relationships. The knowledge of sound source direction further supports the distinction between biological and anthropogenic signals, localized versus diffuse noise sources, and local pressures versus distant disturbances. Directionality thus emerges as a key indicator for assessing environmental integrity, naturalness, and ecological health, providing essential information for conservation planning and landscape management actions [
8]. Sound propagation and perception are strongly influenced by terrain morphology, vegetation structure, and the presence of natural and built elements [
9,
10,
11,
12].
For humans—and for many animal species—sound direction constitutes a primary component of experience, influencing spatial orientation, perceptions of safety, alertness, and well-being. From a human-centered perspective, the analysis of sound directionality integrates a psychoacoustic dimension, moving beyond sound pressure levels to consider how landscapes are perceived, interpreted, and inhabited [
13].
Recent advances in ecoacoustics, particularly through the widespread adoption of autonomous recording units [
14,
15,
16,
17], have expanded the study of soundscapes by enabling the estimation of acoustically detectable individual density [
18], the analysis of spatiotemporal dynamics [
19,
20,
21,
22], and detailed behavioral investigations [
23]. However, while the temporal and spectral dimensions of soundscapes have been extensively explored, their spatial and directional components remain comparatively underdeveloped in relation to environmental monitoring, sustainable soundscape design, and biodiversity conservation.
The integration of sound source localization into ecoacoustic survey protocols therefore represents a methodological advancement of particular relevance for landscape ecology, biodiversity monitoring, and regenerative acoustic design [
24]. Accurate spatial positioning of vocalizations not only improves detection capabilities but also enables the extraction of richer ecological information, including territorial organization, social interactions, reproductive strategies, and responses to environmental and trophic variability [
25]. Such spatially explicit information is essential for linking acoustic patterns to landscape configuration and underlying ecological processes. At the same time, incorporating sound directionality fosters closer integration between biological data and the structural, morphological, and anthropogenic features of landscapes [
26,
27]. This integration supports a more comprehensive understanding of how environmental mosaics influence the distribution of soniferous species and facilitates the development of spatially explicit ecological models for biodiversity prediction and management, particularly in heterogeneous and rapidly changing environments [
28,
29,
30,
31].
To fully exploit the ecological and perceptual information embedded in sound directionality, directional data must be interpreted within spatial frameworks capable of organizing acoustic information at the landscape scale. While sound source localization provides point-based and vectorial information on the origin of acoustic signals, its full ecological significance emerges when such directional cues are integrated to describe how acoustic space is structured, connected, and differentiated [
13]
Within this spatially explicit interpretation of soundscapes, directionality becomes a foundational element for identifying coherent acoustic areas shaped jointly by sound sources and landscape configuration. This conceptual shift—from individual sonic events to spatially organized acoustic units—establishes the theoretical conditions for the operational use of the sonotope concept.
The introduction of the sonotope concept [
32]—defined as an acoustically homogeneous spatial unit within a broader soundscape—has provided a powerful lens for interpreting the spatial organization of environmental sounds. Empirical studies have shown the assemblages of biotic, abiotic, and anthropogenic signals form complex sonic mosaics [
33], whose spatial configuration reflects ecological community structure, habitat connectivity, and landscape heterogeneity [
34,
35,
36]. Sonotopes may vary across spatial scales (e.g., adjacent habitats) and temporal scales (diurnal, nocturnal, and seasonal), offering insights into acoustic diversity, ecosystem health, anthropogenic pressure, and actual or potential ecological connectivity [
32,
37].
In turn, the sonotope is closely related to the notion of the sonic field, defined as the ensemble of acoustic signals emitted by stationary or mobile sources, whose perception depends on both the distance and angle of incidence relative to the receiver. Sonic fields are inherently dynamic and are shaped not only by source composition but also by environmental morphology, atmospheric conditions, and the material properties of landscapes.
Characterizing a sonic field therefore extends beyond simple sound pressure measurements and requires the analysis of spatial, directional, and perceptual components of sound, enabling the reconstruction of the sonic architecture of environments with direct ecological, behavioral, and human-centered implications. Despite the availability of advanced technologies for detecting and analyzing sonic fields, their systematic application in ecoacoustics remains limited. For example, ambisonic and omnidirectional microphone arrays enable researchers to analyze sound directionality, source localization, and spatial dynamics within complex environments. However, these technologies are still primarily used in experimental or artistic contexts and are only marginally integrated into large-scale ecological monitoring and analytical frameworks. Methodological complexity, high spatiotemporal variability of signals, and the need for interdisciplinary integration—spanning acoustics, bioacoustics, landscape ecology, computational modeling, and psychoacoustics—have hindered the development of standardized and scalable protocols. Nevertheless, sonic field analysis represents a promising frontier for advancing ecological theory and practice, providing empirical support for the acoustic habitat hypothesis [
38] and for more inclusive, resilient, and equitable approaches to environmental sound assessment [
39].
Sound source localization is therefore essential for linking landscape configuration with the spatial distribution of soniferous species and for enabling spatially explicit analytical and predictive models that integrate geographic positioning with inter-source spatial relationships [
38]. In this context, sound mapping emerges as a key methodological innovation for ecoacoustics, urban soundscape research, sustainable acoustic planning, and a broader range of applied geosciences.
As previously introduced, two main methodologies are currently used in sonic field investigations: omnidirectional microphone arrays [
17,
18,
19,
40,
41] and ambisonic microphones [
42,
43,
44]. While microphone arrays can achieve high spatial resolution [
45,
46,
47], they require substantial logistical effort, large numbers of synchronized devices, and precise deployment strategies, limiting their applicability in complex natural environments and heterogeneous urban contexts [
48]. Ambisonic microphones enable three-dimensional sound field reconstruction but remain costly, sensitive to environmental conditions, computationally demanding, and constrained by the lack of standardized ecoacoustic protocols [
49,
50].
In this study, we explore an alternative exploratory approach to sonic field investigation based on directional supercardioid microphones, offering a practical and scalable solution for long-term and adaptive sonotope monitoring. Supercardioid microphones provide strong directionality with reduced structural complexity, lower sensitivity to adverse weather conditions, and limited amplification of background turbulence [
51,
52,
53], making them particularly suitable for low-impact and regenerative monitoring strategies.
Although it is generally assumed that animal species in landscapes characterized by varying degrees of spatial heterogeneity tend to concentrate in areas with higher resource availability [
54], the relationship between species-produced acoustic signals and landscape spatial structure remains insufficiently explored. This gap constrains the interpretation of sonic patterns as indicators of the ecological processes organizing landscapes. Addressing this limitation is therefore essential to move from descriptive acoustic surveys toward a landscape ecology-oriented approach capable of supporting spatially explicit management of resources, habitats, and biological communities [
55,
56].
Building on these considerations, we developed and tested the Sonic Explorer, a sonic device composed of two opposed supercardioid microphones mounted on a rotating platform and integrated with a custom software framework for quantitative and qualitative sound analysis. The system is designed to investigate the spatial dimension of a soundscape through directional scanning, with particular emphasis on the detection and characterization of sonotopes and their relationship with landscape configuration (
Figure 1). After frequency- and intensity-based filtering, the spatial arrangement of detected signals enables the reconstruction of original sonic fields and supports integrated acoustic, spatial, and visual analyses. The Sonic Explorer is thus proposed as a versatile experimental tool for ecoacoustics research and future integration with computational, AI-based, and human-centered acoustic design frameworks.
2. Materials and Methods
2.1. Sonic Explorer Prototype
The Sonic Explorer prototype was assembled using commercially available components configured to acquire directional acoustic information through controlled rotation.
Rotating Platform: The system is mounted on an electric rotating plate (YWNYT™, diameter 20 cm, maximum load 8 kg), capable of completing a full 360° clockwise rotation in approximately 14 s, corresponding to a rotational speed of 0.448 rad s−1 (≈0.071 rev s−1). The rotation speed was constant during all measurements.
Microphones: Two supercardioid condenser microphones (Movo™ X3-II, San Francisco, CA, and Edina, MN, USA) were mounted coaxially on the rotating platform, with their acoustic axes oriented in opposite directions (180° apart). The distance between the acoustic foci of the two microphones was 68 cm, corresponding to the minimum separation allowed by the physical dimensions of the microphones while maintaining collinear alignment. Supercardioid microphones were selected for their narrower frontal pickup angle and stronger lateral attenuation compared to conventional cardioid microphones. Structurally, a supercardioid microphone consists of a pressure-sensitive diaphragm coupled with an interference tube featuring lateral apertures (“slots”) (
Figure 2).
These apertures allow sound waves arriving from different angles to reach the diaphragm with frequency-dependent phase delays, resulting in constructive or destructive interference. This mechanism shapes the characteristic supercardioid polar pattern, with strong frontal sensitivity, attenuated lateral response, and a residual rear lobe. Manufacturer specifications for the Movo™ X3-II microphones are reported in
Figure 3 and
Table 1.
Microphone Sensitivity Correction: The relative sensitivity of the two microphones was assessed over 15 h of recording in a low-noise indoor environment, with the microphones mounted on the rotating platform. A Root Mean Square (RMS) sensitivity difference of 5.05 ± 0.30 dB was detected between the two channels. To compensate for this mismatch, a channel-specific gain correction was applied to all recordings using the routine
Microphone_correction, described in detail in
Supplementary Material.
2.2. Recorder
Audio recordings were acquired using a Zoom F3 field recorder (Zoom Corporation®, Chiyoda-ku, Tokyo, Japan, 2022) operating in stereo mode and equipped with 32-bit floating-point recording technology. This format enables the capture of signals over an extremely wide dynamic range without the need for manual gain adjustment, minimizing both clipping and loss of low-amplitude signals.
Zoom F3 employs dual analog-to-digital conversion paths per channel (optimized for low- and high-amplitude signals), which are internally combined into a single 32-bit floating-point file. This architecture ensures stable recordings under highly variable acoustic conditions, such as those encountered during unattended or long-term field measurements.
Channel sensitivity parity was evaluated using sinusoidal test tones at 250, 500, 1000, 2000, 4000, and 8000 Hz. Each tone was played for 10 s from a MacBook Air and fed into the recorder using a balanced 3.5 mm TRS to dual XLR stereo cable (Tisino™, 2 m, Dexinuo, Hangzhou, China). RMS and peak levels were compared across channels, revealing no significant differences.
Remote operation was enabled via a Zoom BTA-1 Bluetooth adapter, allowing the control of the recorder from a smartphone at distances up to approximately 20 m.
2.3. Acoustic Marker (Buzzer)
An electronic buzzer (Colexy DC 3–24 V, nominal level 95 dB SPL) was used as an acoustic marker to indicate the start and end of each full rotation. The buzzer was powered by four rechargeable AA batteries (1.5 V each) and activated via a micro-switch (Daokai 20PCCS, 1 A nominal current) being triggered when microphone #1 reached the 0° position.
The buzzer emitted a short acoustic signal (≈0.344 s) with a dominant frequency band between 3287 and 3327 Hz, which was clearly distinguishable from the environmental sounds and used for temporal segmentation of the recordings. The sound produced by the buzzer was assumed as a negligeable disturbance to the sonifeorus species, limited to a few meters around the device.
2.4. Power Supply and Support Structure
The rotating platform and recorder were powered by two USB-C PD power banks (INIU™ (Shenzhen Simple Life Technology Co., Ltd., Shenzhen, China), 10,000 mAh), while the buzzer was independently powered by rechargeable AA batteries (EBL™, 2300 mAh).
The system was mounted on a wooden tripod originally designed to support a theodolite. The tripod was carefully leveled prior to each deployment to ensure mechanical stability and accurate rotation. Wood was preferentially used for structural elements to minimize vibration transmission.
Before each recording session, the scanner was aligned along the south–north axis, with 0° corresponding to south and 180° to north. The prototype configuration is shown in
Figure 4.
2.5. Principles of Operation of the Sonic Explorer
A single supercardioid microphone provides high directivity but exhibits intrinsic ambiguity due to its residual rear lobe and frequency-dependent side lobes. As a result, signal intensity alone cannot reliably discriminate whether a sound originates from the frontal or rear sensitivity region (
Figure 5).
The Sonic Explorer addresses this limitation by employing two supercardioid microphones oriented 180° apart. When a sound source aligns with the frontal lobe of one microphone, it simultaneously aligns with the rear lobe of the other, producing a marked intensity contrast between the two channels. This complementary configuration reduces directional ambiguity and improves robustness in reverberant or acoustically complex environments.
Additionally, the differential comparison between the two channels enhances the contrast between localized directional signals and diffuses background noise. Because each angular position is sampled simultaneously by two opposing polar patterns, the effective acquisition time for a full angular profile is halved compared to a single rotating microphone.
The distance of 68 cm is necessary to keep the microphones perfectly aligned due to their physical size. Therefore, this was a compromise between having the microphones aligned but spaced apart and having them closer together but not perfectly aligned. With this configuration, the combined polar response covers the entire horizontal plane, producing an “octave-petal-like” sensitivity pattern.
As the platform rotates from 0° to 360°, the relative intensity difference between the two microphones varies as a function of source direction. Let θ be the azimuth of the sound source relative to a fixed reference and ϕ the instantaneous rotation angle of the platform. The relative angles perceived by the microphones are θ − ϕ for microphone #1 and θ − ϕ + 180° for microphone #2. The total sensitivity, S, is expressed as:
where
M1 and
M2 denote the responses of microphone #1 and microphone #2, respectively. Peaks in
S correspond to the alignment between the sound source and the frontal lobe of one microphone.
In summary, the combination of controlled rotation, complementary directional patterns, and differential signal analysis enables the reconstruction of sound source direction within the horizontal plane.
2.6. Dedicated Software
Five distinct routines were specifically developed for data processing. In the future, it is desirable that these routines be integrated into a single package to enhance computational automation. Currently, they are presented separately with minimal graphical design. Developed in an open-source environment, they can be independently verified and implemented by interested users. The routines, written in Python 3.11, are fully described in
Supplementary Material and can be directly copied and executed in a Python environment.
These routines process the data through the following steps (
Figure 6):
Microphone correction—routine name: Microphone_correction (used when the two microphones exhibit different sensitivities to the same signal).
Splitting the stereo recordings into 30 min files—routine name: File_division.
Extracting sound segments to be analyzed (each approximately 12.76 s long)—routine name: Signal_segmentation.
Identifying the direction of the signals—routine name: Sound_direction.
Combining the results from the various segments and presenting them in polar diagrams—routine name: Polar_diagram.
In practice, the WAV files, recorded in stereo format on Zoom F3, with a standard duration of 93 min each, were first subjected to microphone correction (Microphone_correction) and then divided into 30 min segments using the File_division routine to facilitate processing. From each 30 min file, the Signal_segmentation routine extracts 130 segments (here referred to as “segment files”), each approximately 14 s long and delimited by the buzzer signal. The routine can also process files shorter than 30 min, generating a number of segments proportional to the file duration. Subsequently, the Sound_direction routine removes all buzzer signals from both stereo channels, yielding a usable signal length of approximately 12.75 s.
The segmentation of the files was performed using the beep recorded by microphone #1. The buzzer is triggered when microphone #1 passes the 0° (South) position. For each interval of approximately 14 s between two consecutive beeps, the selected segment extends from the end of the first beep to the beginning of the next one, with an additional safety margin applied to ensure precise delimitation. The resulting 130 audio segments (for 30 min files), with the beeps removed, are then exported in WAV format. Segments are discarded if there are too few beeps, if they are excessively short, or if the beeps overlap with environmental sounds, as these conditions may lead to inconsistent segment durations. In particular, the overlap between the beep signal and biological or environmental sounds may prevent the algorithm from correctly identifying the beep pattern. In such (rare) cases, the affected segment is excluded from the analysis. In more detail, the Sound_direction routine processes the signals of each segment through the following steps:
Edge cleaning: a 0.20 s safety margin is applied to remove edge artifacts.
Frame segmentation: stereo audio is divided into temporal frames. Frame durations were tested at 0.1 s and 0.3 s.
Energy analysis: the difference in energy between the two stereo channels is used to determine the direction of the signal.
Frequency estimation: the dominant frequency band is identified using an STFT procedure.
Circular Statistics: Directional data are analyzed to compute mean direction, mean resultant length, circular variance, and circular standard deviation.
The mean direction was calculated by transforming the angles into unit vectors in the complex plane, where
C is the mean cosine component, and
S is the mean sine component:
The mean direction is the angle of the mean resultant vector:
The mean resultant length,
R, represents how strongly the data are concentrated around the mean direction, and it ranges from 0 to 1.
The circular variance,
V, measures the dispersion of the angles around the mean direction. It is defined as:
The circular standard deviation,
σ, is defined as
The results were saved in an Excel file, containing:
The relative acquisition time (within the interval 0–12.75 s, after the portion of the signal occupied by the buzzer has been removed);
The direction of the signal of maximum intensity, estimated by comparing the two microphones;
The directions associated with microphone #1 and microphone #2;
The maximum energy value detected from the comparison of the two microphones;
The energy corresponding to microphone #1 and microphone #2;
The indication of the microphone with the highest energy;
The dominant frequency bands, divided into classes of 1500 Hz each.
For each original 30 min WAV file, the results from the 130 segments are consolidated into a summary and exported as an Excel workbook or as a CSV (Comma-Separated Values) file.
Polar diagrams (energy/direction) for each 12.75 s segment files corresponding to a 360° rotation were generated. The energy, appropriately normalized, is represented with the maximum value near the center of the graph, decreasing progressively toward the periphery. For each segment, the prevailing directions are shown, distinguished by symbols (a dot or a cross) corresponding to each microphone. A summary spatial statistic appears in the second Excel table, and the third table reports the energy of all frequency bands. Finally, a total summary Excel file “Summary-Total” is created, which compiles all parameters of each segment into a single file. This summary file will be used by the Polar_diagram routine, which is explained below.
The Polar_diagram routine processes the data contained in the “Summary-Total” file, which includes all data processed from each segment. This allows the generation of a series of distinct polar diagrams for 30 min intervals according to the selected signal intensity level.
Intensity filtering can thus be applied to distinguish low-intensity signals (originating from more distant sources) from high-intensity signals (attributable to closer sources). It should be emphasized that no standardized procedure exists for applying such filtering; the approach proposed here should therefore be considered an experimental methodology, provided as guidance for data interpretation.
For multiple 30 min samples, all data from a day or multiple days can be normalized by identifying the maximum and minimum values. To exclude signals too weak for reliable interpretation, a minimum energy threshold is defined, and signals below this threshold are disregarded. Five intensity classes were arbitrarily chosen for the case studies presented below, although this number can be adjusted. For example, using 10 intensity classes would allow for a finer observation of variations.
The filtered data are displayed in a summary polar plot, representing the output of the Polar_diagram routine and encompassing all data from all segments. In this plot, angles indicate signal directions, dots represent the intensity of individual signals, and their colors correspond to the dominant frequency band. The diagram also indicates the cardinal points (S, W, N, and E). Unlike the polar diagrams for individual segments, signals in this summary polar plot are not distinguished by microphone.
The Polar_diagram routine calculates circular statistical parameters:
Weighted mean direction, circular variance (how dispersed the angles are), circular standard deviation, mean energy, and its corresponding radial position are displayed alongside the graph, and the mean direction is highlighted with a green arrow. Each graph is automatically saved as a PNG file in a dedicated folder.
During execution of the Polar_diagram routine, the user can choose between two operating modes: (i) batch mode, in which all graphs are generated and saved sequentially, and (ii) manual mode, in which the user manually decides whether to proceed to the next graph.
The collected data may contain numerous outliers, both low and high, which can complicate normalization. To address this, a Winsorizing procedure—a statistical transformation that limits extreme values to reduce the impact of spurious outliers—was applied. This technique, named after an engineer and later biostatistician called Charles P. Winsor (1895–1951), does not remove extreme values entirely; instead, it “resets” them to specific percentiles (or to the lower and upper percentiles) of the data. For example, a 90% Winsorization sets all data below the 5th percentile to the value of the 5th percentile and all data above the 95th percentile to the value of the 95th percentile. Winsorized estimators are generally more robust to outliers than standard estimators, although alternative approaches, such as trimming, also exist.
2.7. Study Area and Field Deployment
The Sonic Explorer was experimentally deployed at the Ortolano Rural Sanctuary [
55], an area characterized by minimal anthropogenic disturbance, which provides suitable habitat for a rich animal community. This site has previously been used for ecoacoustic and behavioral studies, making it an ideal setting for testing the scanner [
56].
2.8. Materials
The functionality of the hardware (rotating platform, buzzer, and F3 Zoom) was tested continuously for 10 days at station n. 3 (
Figure 7) in August 2025 without revealing any functioning problems.
The reliability of the Sonic Explorer’s directionality was evaluated according to the experimental protocol described in detail in
File S1 of the Supplementary Materials. The results show that in 98.33% of cases, the detected direction aligns with the angle between the loudspeaker and the Sonic Explorer within ±1 step (8.37°). This angle is obtained by dividing 360° by the 43 assigned angular intervals. The number of intervals could be increased to achieve a finer resolution, but this would generate a larger amount of data for processing and ultimately depends on the objectives of the research.
In addition, the recording sessions were conducted under the most favorable environmental conditions. Rainy days were carefully excluded, as the prototype is not yet waterproof; the exposure to rain could cause hardware malfunctions and compromise recording quality. To mitigate these risks, the installation of a small, lightweight garden gazebo is recommended to protect the scanner from rain, nighttime dew, and direct sunlight, which—especially during the summer time—could induce thermal stress on the device.
The directionality of signals emitted by nocturnal insects producing persistent sounds was tested using a 90 min recording (captured on 26 August 2025, at 10:06 pm, near a small vineyard;
Figure 6, site no. 2), where numerous tree crickets (
Oecanthus pellucens, Insecta, Orthoptera) were present. From this recording, the second 30 min segment—particularly rich in biological signals—was extracted and designated as night #2.
The variation in sound directionality between nighttime and daytime across different intensity classes was analyzed over a 24 h period—from 06:00 on August 30 to 06:00 on 31 August 2025—using recordings from files part_001 to part_024 collected at site 3 (
Figure 7).
4. Discussion
The proposed system was not designed to provide absolute sound level measurements (dB or dBA) comparable to those obtained with a calibrated sound level meter but rather to detect the directionality and spatial distribution of sound sources. The primary objective is therefore qualitative and spatial rather than quantitative in terms of sound pressure level (SPL). Absolute microphone calibration is desirable but not critical for directional analysis under the sonotope model.
The tests were conducted under variable environmental conditions in order to assess the robustness of the Sonic Explorer in real-world scenarios. The effects of wind and precipitation were mitigated through microphone orientation and physical protection of the devices; however, the methodology does not preclude the possibility of additional measurements under controlled conditions. The aim is to evaluate the practicality of the system in natural and urban environments, where wind and atmospheric phenomena are unavoidable. Sound source typology is determined through spectral analysis and qualitative comparison with reference recordings rather than through absolute sound pressure levels.
The Sonic Explorer is not intended to replace systems such as omnidirectional arrays or ambisonic microphones but to provide a practical, scalable, and low-cost method for the directional estimation and spatial characterization of sonotopes. Validation against professional systems may be considered as a future step, but it is not essential for demonstrating the functionality and applicability of the prototype for qualitative soundscape studies. The initial tests conducted on the system’s ability to detect the direction of a sound source yielded highly positive results. Naturally, the accuracy of directional estimation largely depends on environmental conditions, as well as on the type of sound sources and their distances.
We evaluated the functionality of the hardware, and, although this Sonic Explorer is a prototype, it proved to be very reliable and structurally robust. No interruptions or malfunctions were ever recorded despite the simplicity of the hardware configuration. The Zoom F3 recorder also demonstrated consistent reliability over time.
The simplicity of the system requires no setup other than aligning the microphones with the buzzer switch. Since the system consists of a rotating plate and a separate microphone assembly, it can be mounted on a tripod and aligned using a compass and a bubble level. The orientation of the system is defined both by the direction of the rotating plate—which was always fixed toward the north—and by the presence of a buzzer that emits a 0.1 s beep each time microphone #1 passes by, corresponding to stereo channel 1.
From the first beep to the second, the microphones complete a full 360° rotation, allowing the angular direction of each recorded signal to be determined at any given moment. During the 12.75 s of active rotation, two synchronized microphones operate simultaneously, meaning that every point on the horizon is scanned twice within this interval.
We selected a rotation period of 12.75 s, corresponding to a linear speed of 0.14 m/s, as a compromise between higher speeds—which could introduce physical disturbances due to air friction on the microphone capsules (typically noticeable above 1–3 m/s)—and lower speeds, which would reduce scanning precision by capturing different sound events from nearly identical microphone positions. Further optimization of the scanning speed remains possible when operating in different types of landscapes.
Dividing the 12.75 s recording into 0.3 s frames proved to be a reasonable setting, although shorter frames of 0.1 s were also tested providing more details. A shorter frame increases data resolution but introduces the risk of oversampling if the biological signal lasts longer than 0.1 s and also slows down computation. This limitation must therefore be carefully considered. There is, however, no universally optimal frame size when environmental sounds vary greatly in duration—as in the case of long songs, such as those of the blackcap or the red-billed leiothrix, compared to very short alarm calls like those of the European robin (Erithacus rubecula). Naturally, this system has limitations, as sound signals are not point-like in an idealized sense.
The sensitivity of this prototype for adverse weather conditions requires the use of a protective cover, such as a gazebo, which I have subsequently tested successfully, although it is not presented in this initial work. This solution effectively eliminates problems related to the impact of rain and dew on the microphone capsules and the recorder, but it introduces additional noise during rainfall events. Regardless of the waterproofing attempts applied to the apparatus, the friction of raindrops on the hardware— as occurs with all field recorders currently used in ecoacoustic research—can be mitigated but not completely eliminated.
Currently, sound recordings are stored on micro SD cards (32 or 64 GB). In future developments, recordings could be transmitted via cellular networks directly to a remote server, integrating with existing automated systems—such as the one operating through the GSM protocol at
http://ecosoundscape.it (accessed on 29 November 2025)—which has been successfully running for several years in the same environmental context.
In this initial evaluation of the complete procedure, we focused on assessing the system’s ability to identify the dominant direction of each signal. We found it effective to divide the collected data into intensity thresholds or classes, allowing the analysis to focus on higher-energy signals. These, in turn, provide calibration for distinguishing between “near-field” and “far-field” spatial information regarding the location of sound sources, whether geophonic, biological, or technophonic.
The Sonic Explorer eliminates the need for complex microphone arrays and operates effectively even with incoherent or noisy sounds. Rotation enhances angular resolution as the differential response continuously varies with direction. Unlike two fixed microphones—which may yield ambiguous results for lateral or rear sound sources, where inter-microphone differences do not clearly indicate direction—the rotating system ensures that each source passes through successive angles of maximum sensitivity, thereby reducing directional uncertainty.
Although each microphone has a highly directional supercardioid pattern, rotating the system enables scanning of the entire surrounding environment, eliminating blind spots and ensuring that all directions are detectable. By recording across multiple angular positions, the resulting signals can be averaged to better distinguish sound sources.
The use of separate routines has simplified the testing phase of the Sonic Explorer; however, a single integrated routine would be more functional and user-friendly for regular operation. In the near future, the Microphone_correction, File_division, Signal_segmentation, Sound_direction, and Polar_diagram routines could be merged into a unified software package, enhanced with graphical features to facilitate easier navigation and interaction with the processed data.
However, the procedure implemented by the Sonic Explorer represents a promising advancement in the field of ecoacoustics, introducing dedicated hardware and software to investigate the spatial dimension of environmental sound, an important aspect of this field [
21].
This advancement will enable a deeper understanding of the daily and seasonal dynamics of sonotopes—comprising their three components: geophony, biophony, and anthropophony—and will allow for the assessment of the influence of the surrounding landscape on the occurrence of these sound types. This will make it possible to investigate the various spatial configurations of environmental mosaics and their influence on the distribution of soniferous species. In particular, moving the Sonic Explorer across a landscape will enable a more accurate characterization of sonotopes and the assessment of the influence of landscape spatial arrangement on soniferous species activity.
Furthermore, the routines developed within this framework facilitate efficient navigation of large acoustic databases, enabling researchers to identify specific acoustic events or anomalies that would otherwise be extremely time-consuming to detect manually. The ability to isolate individual intensity classes and visualize their spatial distribution opens a new avenue of inquiry, complementing traditional ecoacoustic approaches that rely on the interpretation of acoustic indices.
The Sonic Explorer can be used to monitor landscape structure and dynamics because the spatial dimension of the sounds it detects can be correlated with landscape characteristics, combining its advanced capabilities for investigating sound directionality with the predictive power of ecoacoustic indices [
57,
58]. The Sonic Explorer represents a tool capable of enhancing our understanding of the ecological role of sounds, as it allows the spatial distribution of sounds to be assessed across a wide range of landscape configurations, from intact natural areas to urban environments. Consequently, the spatial characteristics of landscapes can be effectively evaluated by overlaying spatial sound maps with landscape maps, potentially addressing the challenge of determining the appropriate scale at which sound should be correlated with the landscape [
59].
The characterization of sound sources in the prototype was designed to be multimodal, combining both spectral analysis and direct listening of the signals in their dominant direction.
Spectral analysis, as the primary method, allows the identification and discrimination of sound sources based on frequency content, duration, and temporal modulation, providing an objective and repeatable method for signal classification. This approach is particularly effective in complex environmental contexts, where multiple sources coexist.
Selective listening for validation and further interpretation: direct listening to the signals in their dominant direction serves as a qualitative control, useful for confirming source identity or interpreting complex signals that may be difficult to distinguish using spectrograms alone (e.g., similar vocalizations among species or anthropogenic sounds with harmonic components).
Integration of the two approaches: the combination of quantitative spectral analysis and qualitative listening enables a more complete and reliable characterization of sound sources, balancing objective precision with human interpretative capability without introducing significant bias.
The procedure employed with the Sonic Explorer differs from those typically used in ecoacoustics, where autonomous recording stations collect large amounts of data independently of the operator. Conversely, deploying a single Sonic Explorer across different landscape configurations requires greater operator involvement in the field. This aspect may, on the one hand, represent a limitation in the use of the Sonic Explorer, but on the other hand, it can lead to a more informed and conscious application of this technique, enabling a more efficient coupling between the acoustic and landscape dimensions. Furthermore, the ability to operate within predefined intensity classes will make it possible to observe variations between distant and nearby sonic fields and, for sounds with a known source intensity, to estimate not only their direction but also their distance, thereby increasing the amount of useful information collected in the field.
It is evident that specific spatial indices for describing the patterns detected by the Sonic Explorer have not yet been established. Consequently, our analyses are currently limited to the descriptive phase of spatial sonic patterns.
However, we are confident that, in the near future, it will be possible to process spatial data in the form of more advanced indices capable of interacting with some of the most widely used metrics in ecoacoustics. For instance, the spatial dimension of sound quantified with the Sonic Explorer could be associated with sonic indices such as the Sonic Heterogeneity Index (SHI
tf), the Spectral Variability, or the Effective Number of Frequency Bins (ENFB) [
42].
5. Conclusions
The proposed system is not designed to provide absolute sound level measurements comparable to those obtained with a calibrated sound level meter but rather to detect the directionality and spatial distribution of sound sources. Direction is determined by the difference in intensity between two microphones: for the same signal, the microphone recording the highest intensity indicates the source direction within a 360° range. Sound source typology is assessed through spectral analysis and qualitative comparison with reference recordings, consistent with the ecoacoustic objectives of the Sonic Explorer.
The Sonic Explorer is not intended to replace omnidirectional arrays or ambisonic microphones but to provide a practical, scalable, and low-cost method for directional estimation and spatial characterization of sonotopes. It can be easily built without sophisticated engineering, making it accessible to everyone. Horizontal rotation allows the effective sampling of the dominant component of the sound field, which is often concentrated on the terrestrial plane. Continuous scanning enables rapid detection of the main source directions without covering the entire sphere, which is only necessary for highly specialized applications such as fully immersive acoustic studies or complete ambisonic recordings.
The “two opposing microphones” configuration maximizes the discrimination between front and rear lobes, improving directional accuracy in the horizontal plane. While real supercardioid microphones differ slightly from the theoretical models, the system operates effectively in real-world environments swith complex ecological and urban soundscapes.
This system is suitable for landscapes where tree heights do not exceed 25–30 m, but it is not recommended for tropical or boreal old-growth forests, where taller trees and soniferous species active at considerable heights may limit performance.
Overall, the Sonic Explorer represents a methodological advancement in sonotope characterization, offering a practical, cost-effective, and easily replicable approach for the spatial detection of environmental sounds. It provides valuable directional data for sonotopes, ecological mapping, and soundscape studies that would otherwise be difficult to obtain without substantial investment.