Next Article in Journal
The Methods of Determining the Centre of Gravity of a Tilting Body on the Upper Ankle Joint and Weighing the Feet of a Living Human
Next Article in Special Issue
Audio Features in Education: A Systematic Review of Computational Applications and Research Gaps
Previous Article in Journal
Driver Intention Recognition for Mine Transport Vehicle Based on Cross-Modal Knowledge Distillation
Previous Article in Special Issue
Acoustic Analysis and Perceptual Evaluation of Second Language Cantonese Tones Produced by Advanced Mandarin-Speaking Learners
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Minimum Audible Angle in 3rd-Order Ambisonics in Horizontal Plane for Different Ambisonic Decoders

by
Katarzyna Sochaczewska
1,*,
Karolina Prawda
2,
Paweł Małecki
1,
Magdalena Piotrowska
1 and
Jerzy Wiciak
1
1
Faculty of Mechanical Engineering and Robotics, Department of Mechanics and Vibroacoustics, AGH University of Krakow, 30-059 Krakow, Poland
2
AudioLab, School of Physics, Engineering and Technology, University of York, York YO10 5DD, UK
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(12), 6815; https://doi.org/10.3390/app15126815
Submission received: 28 April 2025 / Revised: 9 June 2025 / Accepted: 13 June 2025 / Published: 17 June 2025
(This article belongs to the Special Issue Musical Acoustics and Sound Perception)

Abstract

:
As immersive audio is gaining popularity, the perceptual aspects of spatial sound reproduction become relevant. The authors investigate a measure related to spatial resolution, the Minimum Audible Angle (MAA), which is understudied in the context of Ambisonics. This study examines MAA thresholds in the horizontal plane in three ambisonic decoders—the Sample Ambisonic Decoder (SAD), Energy-Preserving Ambisonic Decoder (EPAD), and All-Round Ambisonic Decoder (AllRAD). The results demonstrate that the decoder type influences spatial resolution, with the EPAD exhibiting superior performance in MAA thresholds ( 1.24 at 0 azimuth) compared to the SAD and AllRAD. These differences reflect the discrepancies in the decoders’ energy vector distribution and angular error. The MAA values remain consistent between decoders up to 30 azimuth but diverge significantly beyond this range, especially in the 60 135 region corresponding to the cone of confusion. The findings of this study provide valuable insights for spatial audio applications based on ambisonic technology.

1. Introduction

In applications such as gaming and virtual and augmented reality, the goal is to provide the user with an immersive experience, both visual and auditory. In terms of audio, spatial sound reproduction enables such immersion through the physically and perceptually accurate reproduction of a virtual sound field. One primary objective of multichannel audio systems is to create precise directional effects that accurately reproduce the intended spatial image, comparable to the spatial image created by real sound sources. This work provides an insight into the Minimum Audible Angle (MAA), a metric that is part of spatial resolution and is linked to the perception of sound-source localisation [1,2].
Two principal methods for reproducing auditory events can be identified in spatial audio systems. The first involves creating a real sound source by reproducing a signal through a specific loudspeaker, resulting in the perception of sound emanating from that loudspeaker’s position. The second method distributes the signal between multiple loudspeakers to create virtual sound sources, a process known as panning [3]. These virtual sources are perceived between physical loudspeakers, with their perceived position dependent on amplitude relationships. Amplitude panning primarily exploits the Interaural Level Difference (ILD), a fundamental auditory localisation mechanism [2]. The manipulation of the gain between two loudspeakers creates a phantom source between them, with its position determined by the specific gain ratios [3].
Among different spatial sound reproduction techniques, Ambisonics is commonly used due to its “first-person” perspective, as well as the ease of audio-scene translation, such as rotation [4,5,6]. It therefore allows for the use of head-tracking for the purpose of natural navigation in the virtual scene and increased localisation accuracy. Ambisonic technology represents a scene-based approach to spatial audio reproduction, distinct from channel-based and object-based methodologies. It does not rely on a predefined playback system, offering greater flexibility [6]. It enables the encoding of directional information about sound sources in a three-dimensional space and facilitates reproduction through spherical harmonic representation. It is also used as a foundation for an emerging open-source Immersive Audio Model and Format [7] known also as Eclipsa Audio [8], which makes Ambisonics more commercially accessible for a wider audience and therefore the perceptual evaluation is even more justified.
Auditory spatial sound location is a complex phenomenon, highly dependent on the environment, the acoustics of the room, and the familiarity with the sound source, among others [2]. It varies significantly within individual cases and changes with age. Of the three planes depicted in Figure 1, the accuracy of localisation is the highest on the horizontal plane (azimuthal angle), which also exhibits less inter-subject variability.
Several different measures refer to the acuity of sound localisation in a three-dimensional space. Quantitative metrics include localisation, localisation blur, MAA [9], and Minimum Audible Moving Angle (MAMA) [10]. On the other hand, qualitative measures comprise locatedness (ease of localisation) or the source image spread [11]. The focus of this study is the MAA, defined as the smallest angle at which a listener can discriminate between two successively presented stationary sound sources [2]. It is part of the spatial resolution attribute, is relatively easy to measure, and can provide useful insight into the perceptual quality of the system and the decoder.
Several studies have investigated the localisation of sound sources in ambisonic scenes reproduced using virtual or real microphones [12]; at times studies also included synthesised scenes [13]. Ref. [14] explored the MAA in the context of reverberation, establishing that reverberation impairs the MAA threshold. The spatial resolution, however, is understudied. The MAA was determined for the Vector Base Amplitude Panning (VBAP) method [15]. A ninth-order ambisonic reproduction system has been shown to produce spatial blurring that remains below the threshold of human spatial auditory acuity [16]. Ref. [17] compared mode matching and pseudoinverse methods on a horizontal layer with AllRAD rendering, proving that there is no statistical difference in the MAA between horizontal and periphonic rendering methods on that layer. However, to the authors’ best knowledge, no study has looked into the differences in the MAA due to sound field reproduction discrepancies caused by ambisonic decoders while looking at both objective metrics and subjective evaluation.
This work focuses on studying the MAA in the horizontal plane for three commonly used ambisonic decoders: the SAD, EPAD, and AllRAD. The rationale behind selecting these three decoders was a comparison between the simplest approach without perceptual consideration and the two most popular perceptual approaches. The objective evaluation of the decoder is relatively easy to perform; however, further evaluation is needed to fully determine its perceptual performance. The localisation accuracy across the decoders is evaluated in a listening test using third-order Ambisonics, focusing on the MAA and its dependency on a specific decoder design.
The structure of this paper is as follows. Section 2 presents the motivation behind investigating the MAA in ambisonic technology and the context of existing studies on the MAA. Section 3 describes the investigated ambisonic decoders, including their objective metrics. Section 4 elaborates on the methodology and experimental design, followed by a comprehensive description of the results in Section 5. In the discussion in Section 6, some additional factors are raised that are beyond the scope of this study but are worth considering in future research. Section 7 offers concluding remarks.

2. Background

In spatial sound reproduction, an ambisonic decoder is responsible for the conversion of encoded spatial information and its distribution to loudspeakers. Gerzon [4], credited with developing ambisonic technology, noted that designing a decoder that would account for all psychoacoustic variables would require extensive computational resources. However, with current technological advancements, this limitation is no longer considered significant [6].
In the theory of localisation in the ambisonic decoder design, the internal auditory system is treated as a black box, while external mechanisms can be conceptualised as analogous to pressure and gradient-pressure microphones [4], as schematically depicted in Figure 2.
These simplifications allow for choosing reduced parameter sets while enabling the design of perceptually effective decoders. Gerzon introduced two critical models for these mechanisms: the velocity model and the energy model. In practical implementation, both the velocity vector r V , sometimes referred to as Makita localisation, and the energy vector r E must be addressed. Further enhancements of perceptual performance are possible to achieve but are not fundamental to basic functionality. To construct perceptually accurate sound fields, decoders should follow these principles (in order of importance) [4]:
  • High frequencies must converge in the source direction (energy vector);
  • Maintain constant amplitude gains for all directions;
  • Ensure equivalence in the perceived direction for both high and low frequencies;
  • Provide constant energy gain for all directions;
  • Accurately reproduce wavefront direction and velocity at low frequencies (velocity vector).
Gerzon’s vector models characterise sound field properties at the listener’s position. All aforementioned aspects should be integrated into the decoder design to successfully implement the “majority verdict”—an auditory direction resolving mechanism in which the human auditory system localises the sound source based on many available cues, choosing the ones that are in agreement [4]. Decoders fulfilling these requirements are classified as “two-band” or “Vienna decoders” [5]. The complexity of the implementation varies according to the specific loudspeaker configuration. The spherical t-designs are considered optimal [19], as they offer simplifications to numerical integration on a sphere for polynomials of degree 2 N t [20]. The investigation in this study, however, focuses on a system that does not constitute a t-design configuration, in which case certain decoder implementations may exhibit equivalent performance characteristics [6].

3. Minimum Audible Angle

The MAA is defined as the minimum displacement of a stationary sound source that can be detected by a human listener when two sound stimuli are presented consecutively [2,9]. The MAA values have been measured in experiments with physical [9,21] and virtual sources [22], showing that, in general, the MAA is the smallest in the frontal listening area (around 1 2 ) and it increases when the stimuli are presented to the side or to the back of the listener. It is also frequency-dependent, as the localisation mechanisms in the human auditory system are different for high and low frequencies [9]. Despite this, many studies use broadband stimuli in MAA experiments and do not differentiate between low- and high-frequency values in their results [21,22].
Despite sharing the same perceptual cues, the MAA and localisation are two different discrimination tasks. The MAA is a Just Noticeable Difference (JND) task, which is a comparison test, and a localisation task is an absolute test [23].

4. Selected Ambisonic Decoders

Ambisonic decoding plays a crucial role in spatial audio reproduction, influencing sound localisation and perceived immersion [24]. Various decoding approaches exist, each with different trade-offs in terms of energy distribution, spatial accuracy, and computational complexity. This section examines three selected ambisonic decoders: the SAD, EPAD, and AllRAD. Their principles, performance characteristics, and practical implications for playback system configurations are analysed in detail.
The Sampling Ambisonic Decoder (SAD) represents the most fundamental algorithm among those investigated in this study. Its core principle involves sampling virtual panning functions at the directions of L loudspeakers [6]. For spherical reproduction systems, the decoder matrix is simply
D = 4 π L Y N T .
The factor 4 π L reflects that each loudspeaker contributes a fraction of the energy (E measure) distributed across the surrounding directional unit sphere. Y T is a transposed vector of the sampled spherical harmonics at the loudspeaker directions. This method provides an even distribution of objective metrics for uniform loudspeaker layouts. However, in non-uniform configurations, energy mismatches result in attenuation in varying degrees [6].
Figure 3 illustrates the distributions of amplitude, velocity, and energy for the investigated setup. Even distribution of all the measures can be observed on the horizontal plane, with fluctuations occurring above and below it. They reveal a strong dependence between the placement of the speaker and the energy measures. The uniformity of the distributions can be improved with mode-matching design (MMD), the primary objective of which is to align the spherical harmonic modes corresponding to loudspeaker signals with those of the sound field decomposed in Ambisonics. However, both the SAD and MMD share a significant limitation: when the playback system is not a t-design, that is, when there is some non-uniformity in the loudspeaker arrangement [6], the objective criteria of energy and the energy vector are not preserved (see Section 2). This limitation inspired researchers [25] to develop a more robust decoding approach.
The Energy-Preserving Ambisonic Decoder (EPAD) maintains the MMD approach but eliminates energy scaling factors. This design is accurate for frequencies above 200 Hz and enables panning-invariant loudness. This preservation occurs for any loudspeaker configuration, provided that the number of spherical harmonics remains adequate [25].
Figure 4 demonstrates the uniform energy distribution alongside other numerical measures, showing significant improvement compared to the SAD algorithm: equalised energy on the whole surface and reduced r V angular error. It is worth noting that if the loudspeaker arrangement was perfectly uniform, both the SAD and EPAD would be numerically equivalent [6].
The All-Round Ambisonic Decoder (AllRAD) represents a hybrid approach that integrates VBAP with SAD principles [26]. This methodology implements a sequential signal processing chain in which sound sources undergo t-design panning before transmission to loudspeakers on a virtual t-design layout and subsequent routing to physical loudspeakers via VBAP transformation matrices. AllRAD implementation requires computational discretisation through implementing a triangular grid between each of the three loudspeakers [6].
This framework accommodates four distinct reproduction scenarios: (1) a virtual source positioned within a speaker triplet, activating all three loudspeakers; (2) a virtual source positioned between two loudspeakers, activating both; (3) a virtual source coincident with a physical loudspeaker position, activating only that speaker; and (4) a virtual source positioned outside the convex hull, resulting in null reproduction. To address the limitations of Scenario 4, Zotter et al. [24] proposed the implementation of “imaginary loudspeakers” to preserve VBAP functionality in non-uniform configurations where the condition t 2 N + 1 cannot be satisfied. These virtual elements facilitate continuous spatial reproduction across otherwise discontinuous regions of the sound field.
The AllRAD maintains numerical stability across diverse loudspeaker configurations while satisfying Gerzon’s objective measures of energy distribution, energy vector ( r E ), and vector magnitude. The algorithm demonstrates superior performance in non-t-design reproduction systems, providing efficient signal distribution for arbitrary multichannel configurations [27].
Quantitative analysis, illustrated in Figure 5, confirms significantly reduced spatial errors and enhanced uniformity in amplitude and energy distribution compared to alternative decoding methodologies. The angular errors of the velocity and energy vectors of the AllRAD are the smoothest among the three decoders. Fluctuations of Amplitude A and Energy E are also negligible. The velocity vector angular error’s overall value spans from less than 0.5 to 2 for the horizontal plane with the highest value around 90 . Irregular fluctuations can be observed across all measures.

4.1. Experimental Setup

The focus of this study is on the accuracy of the sound field reproduction on a horizontal plane of the 3rd-order ambisonic playback system, consisting of 16 Genelec 6010 loudspeakers on a surface of a sphere, as illustrated in Figure 6. The 3rd order is the highest available resolution in the laboratory used for the evaluation and the most popular order to work with in music production at the moment of writing [28]. In the three-dimensional space, the operational system is a spherical coordinate system with a radius of r = 1.5 m, and azimuth and elevation angles, as specified in Table 1. In the AllRAD case, two additional imaginary loudspeakers were included for correct decoder calculations (yellow dots in Figure 6).
The primary objective of this experimental series was to investigate the MAA values for three ambisonic decoders: the EPAD, AllRAD, and SAD. All listening tests were conducted at the Laboratory of Auralisation, AGH University of Science and Technology (UST), which is an acoustically treated room of dimensions 3.9 m × 6.7 m × 2.8 m (length, width, and height, respectively), with an average reverberation time of 0.15 s.

4.2. Participants

The listening panel comprised 15 subjects (6 female, 9 male) in the age group of 23–37 years old, with varying levels of experience in working with stereophonic and/or spatial audio. The medium experienced participants included students of the Acoustic Engineering course, while the highly experienced participants consisted of Acoustic Engineering teachers and professional audio engineers with at least 10 years of field experience. Four subjects—three highly experienced teachers and researchers and one student with a low level of experience—took more than one listening test to facilitate the analysis of the individual perceptual differences between the decoders. The choice of repeated participants was motivated by their level of experience as well as availability. No audiometric test was performed; however, all participants confirmed that they had normal hearing. Based on prior work conducted at our institute [29], demonstrating that listeners with lower hearing thresholds do not necessarily perform better in challenging listening tasks, the approach adopted in this study—using stimuli at a level of 80 dB(A)—is considered methodologically justified.

4.3. Listening Test Procedure

Prior to the presented research, preliminary listening experiments were conducted. This process helped identify methodological optimisations, particularly regarding session duration based on the adaptive method in the context of the MAA [30]. This study implemented a non-parametric adaptive up–down method, which estimates thresholds without explicitly characterising the psychometric function’s shape beyond monotonicity assumptions. Although threshold estimation typically involves averaging stimulus intensities from final reversals [31], this research used a hybrid-adaptive approach [32] that allows data to be matched to an assumed psychometric function. Various transformed up–down procedures are used in psychoacoustics. The Mean Group Length (MGL) parameter characterises procedural complexity—larger values necessitate more reversals before test termination. The optimal measurement point on the psychometric function (the “sweet point”) for minimising threshold estimate variance is generally considered to fall within the X80–X94 confidence range [33]. The research from Levitt [31] determines the X71 rule or X75 rule as the mid-point. The X71 rule (MGL = 1.71) executes approximately 50% faster than the X75 rule and was therefore chosen for the purposes of the presented research.
Based on the aforementioned studies, the adaptive procedure implemented the following parameters:
  • Decrementing/incrementing step sequence: [ 60 , 30 , 15 , 8 , 4 , 2 , 1 , 0.5 , 0.25 ] .
  • Initial angular displacement: 30 from reference position (100% recognition accuracy baseline).
  • A 2-up/1-down procedure with 7 reversal limits. The first reversal was excluded from the calculations.
  • A 75% recognition threshold determined using the Bayesian psignif algorithm.
  • The frontal plane ( 0 90 ) and rear plane ( 90 180 ) were examined separately to prevent systematic errors.
The experimental framework maintained consistency across all tests regarding the interface design, psychoacoustic method, loudspeaker configuration, and ambisonic order. The testing environment used MATLAB 2020b connected to REAPER (digital audio workstation) through the Open Sound Control protocol. Open-source plugins were used for decoding: IEM AllRAD for the AllRAD and Aalto Sparta Decoder [34] for both the SAD and EPAD. The implementation utilised Schmidt semi-normalisation (SN3D) with max- r E weighting, as recommended in the literature [35,36]. The stimulus used in the experiment was a white Gaussian noise burst of 500 ms, with 20 ms fade-in and fade-out and a frequency range 20 Hz–20 kHz. This range was selected to match both the effective operating frequency range of the ambisonic decoder’s energy vector and the human auditory system’s use of spatial localisation cues. Frequencies above approximately 200 Hz are known to contribute significantly to directional hearing, with low frequencies supporting Itearaural Time Difference (ITD)-based localisation and higher frequencies supporting ILD-based localisation mechanisms. A MATLAB-based interface facilitated the execution of the test.
Participants were placed in the sweet spot and instructed to maintain visual focus on a screen behind the central loudspeaker to minimise head movements. The test sequences began with a reference sample played from predetermined azimuthal positions: 15 ,   30 ,   45 ,   60 ,   75 , and 90 (frontal plane); 105 ,   120 ,   135 ,   150 , and 165 (rear plane). Symmetry was assumed between the left and right hemispheres, while the 0 and 180 positions were specifically tested. A second sample was presented with an angular displacement (randomly left or right) equivalent to the initial step size. The participants indicated the perceived displacement direction by choosing the appropriate button with the mouse click. An example of the course of the experiment is presented in Figure 7.
The duration of the procedure varied with the reference position, with the longest at 0 , as shown in Figure 7 and the shortest within the 60 135 region, which corresponds to the cone of confusion. Based on the results of the staircase procedure, the psychometric function threshold at 75% recognition was calculated for each reference angle. MATLAB script [37] was used to predict and plot the functions.

5. Results

The comparative analysis of the SAD, EPAD, and AllRAD revealed large variations in spatial reproduction parameters. For all the results, the highest step in the adaptive procedure— 40 —is assumed to be the recognition threshold. Above that value, the MAA was not determined and the plots are limited on the Y-axis.
Figure 8 shows the MAA across the subjects for the SAD case. For most of the participants, two significant peaks are revealed: around 105 and at 150 . The high values of the MAA at 105 are preceded by a slow rise starting around 30 45 . For subjects S01 and S15, the first peak is shifted towards 90 , while the second peak is shifted to 135 for S01 and does not occur at all for S15. The MAA values across the subjects show considerable variability, with three of them (for subjects S05, S14, and S16) extended beyond the maximum analysed range of 40 .
Figure 9 shows the outcome of the listening experiment for the EPAD. Similar to the SAD, the variability of the results across the subjects is high, with a few MAA values exceeding the analysis threshold of 40 (subjects S03, S05, and S11). The trends in the MAA are also not as visible as in the case of the SAD. However, the majority of the results show a significant peak at 90 , implying a higher MAA in that area. In general, all the results exhibit higher MAA values between 45 and 165 .
The MAA values obtained for the EPAD are generally lower than those for the SAD, cf. Figure 8. This may be a consequence of low fluctuations in amplitude and energy metrics for the EPAD, as shown in Figure 4.
The MAA values for the AllRAD are depicted in Figure 10. They show that the MAA grows considerably between 60 and 120 . For subjects S06 and S08, the MAA increases symmetrically with respect to the 90 azimuth angle, forming two peaks at 75 and 105 separated by a trough. Other results exhibit only one peak, either at 75 (S09) or 105 (S02 and S10). Conversely, the MAA for S01 and S03 is the highest at 90 .
The mean values of the MAA parameter determined for all three decoders are shown in Figure 11. The values remain coherent below 30 . Overall, the EPAD produces the lowest MAA values in relation to the horizontal plane of all the investigated decoders. Furthermore, for reference 0 , the MAA threshold was the lowest for the EPAD and equalled 1.24 , which is a value corresponding to those of the investigations with the physical sound source [2].
Overall, the MAA values for all the investigated decoders show a significant increase in the angular range between 60 and 120 . Such a tendency is in line with similar results from the literature [21] and corresponds to the region of indeterminate sound localisation, referred to as the cone of confusion [38].
Figure 12 shows isolated results for subject S01, who is highly familiar with the ambisonic system. For such an experienced person, the MAA values are within a similar range for all investigated decoders. There is a visible peak in the cone of confusion area at 90 and then another smaller one at 135 . The second peak might be caused by the fact that the shift from the loudspeaker position always appears as a phantom source in between two loudspeakers, causing small fluctuations in energy vector angular errors. The same phenomenon, although on a smaller scale, is observed in the MAA values for 45 .
The results for subject S05, who is highly experienced in audio engineering, with medium familiarity with the playback system used in the experiment, are shown in Figure 13. For the rear plane in the case of the AllRAD algorithm, the reoccurring systematic errors in the answers made detecting the MAA threshold impossible. As the subject had extensive experience with stereo systems, the low resolution above 30 may suggest difficulties with perception above the stereophonic range due to the contrast in the high familiarity with the stereo range and unfamiliarity beyond it.
Figure 14 shows the results for subject S06, who has the lowest familiarity with the ambisonic system. In this case, only the data from two listening tests was available. For the range of the stereophonic scene, which is up to 30 , there is no significant difference in perception between the two decoders. However, above that reference value, the MAA remains lower for the EPAD.

6. Discussion

In this study, the arithmetic mean was employed as a measure of central tendency for the analysed parameters, with full awareness of the methodological limitations inherent to its application in small-sample research. To address these limitations and ensure transparency, the steps were implemented as follows:
  • Presentation of individual participant data alongside aggregated means (see Figure 8, Figure 9 and Figure 10) to visualise inter-subject variability.
  • Cautious interpretation of results, emphasising exploratory trends rather than definitive conclusions.
This approach aligns with recommendations for small-sample studies in the psychoacoustic literature [33], where descriptive statistics remain valid when accompanied by rigorous transparency about data dispersion and reproducibility constraints.
The ANOVA results indicate that decoder performance varies significantly across spatial positions; however, the main effect of the decoder alone is not significant ( p = 0.697 ). Notably, pairwise comparisons between the AllRAD and SAD show a significant difference ( p < 0.05 ) only at 75 and 90 . The overall effect size is small ( R 2 = 0.42 ).
Additional observations within the subjects are captured, suggesting a possible dependency on familiarity with the playback system and the stereophony experience and offering direction for further research. For subjects highly familiar with the system, the differences between the decoders remain within the same order of magnitude. For listeners with high experience in stereophonic sound, the resolution within the stereo base (up to 30 ) remains very high for the VBAP-based decoders. For the subjects without experience, the EPAD produces better results than any other decoder for the whole range of the horizontal plane. That may suggest the overall importance of stable energy distribution and minimising the energy vector angular error to produce the best perceptual results. The results of the three subjects (S01, S05, S06) with varied experience with stereophonic and spatial audio may suggest that, unless the listener is highly familiar with the playback system, the particular experience in stereophony may cause increased alertness in phase differences. This, in turn, may reduce the perceptual resolution of the MAA.

7. Conclusions

This paper provides insight into the directional auditory location in Ambisonics by investigating the MAA for three different ambisonic decoders: the SAD, EPAD, and AllRAD. It describes all considered decoders and discusses the differences in the sound field reproduction when applied to the non-t-design of the loudspeaker layout.
The perceptual evaluation of the MAA reveals different results for each of the decoders, suggesting that the method of sound field reproduction influences the MAA. In particular, the EPAD, which prioritises equal energy distribution, produced the lowest mean values of the MAA. This may be motivated by the more uniform reproduction of amplitude and energy across the sound field compared to the SAD and AllRAD. However, the localisation perception may also be influenced by the familiarity with ambisonic systems and with audio reproduction systems in general. The effect size is too small in this particular study; to better understand potential trends, particularly regarding individual differences related to experience with spatial or stereophonic playback, which show the potential to contribute to the results, future studies should include a larger number of participants.
The white noise stimulus was band-limited to 20 Hz–20 kHz, which broadly encompasses the frequency ranges relevant to ITD and ILD-based localisation cues. However, the precise relationship between the spectral content of the stimulus, auditory spatial mechanisms, and the frequency response characteristics of the energy vector decoder remains a subject for further investigation and systematic analysis.
In the future, the insights regarding the MAA can be useful in research on the human perception of sound in virtual and extended reality environments, where sound reproduction is most often ambisonic-based. The MAA can also provide guidelines for sound design and music production capable of reproduction that meets the creator’s intention in terms of sound-source localisation.

Author Contributions

Conceptualisation, K.S. and P.M.; methodology, K.S.; software, P.M.; validation, P.M., M.P. and J.W.; formal analysis, K.S.; investigation, K.S.; resources, K.S.; data curation, K.S.; writing—original draft preparation, K.S.; writing—review and editing, K.P.; visualisation, K.S.; supervision, P.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AllRADAll-Round Ambisonic Decoder
EPADEnergy-Preserving Ambisonic Decoder
ILDInteraural Level Difference
ITDInteraural Time Difference
MAAMinimum Audible Angle
MAMAMinimum Audible Moving Angle
MGLMean Group Length
MMDMode Matching Decoder
SADSampling Ambisonic Decoder
SN3DSchmidt semi-normalised 3D
VBAPVector Base Amplitude Panning

References

  1. Blauert, J. Spatial Hearing: The Psychophysics of Human Sound Localization; MIT Press: Cambridge, MA, USA, 1997. [Google Scholar]
  2. Carlini, A.; Bordeau, C.; Ambard, M. Auditory localization: A comprehensive practical review. Front. Psychol. 2024, 15, 1408073. [Google Scholar] [CrossRef] [PubMed]
  3. Pulkki, V. Virtual sound source positioning using vector base amplitude panning. J. Audio Eng. Soc. 1997, 45, 456–466. [Google Scholar]
  4. Gerzon, M.A. General metatheory of auditory localisation. In Proceedings of the Audio Engineering Society Convention, San Francisco, CA, USA, 1–4 October 1992; Audio Engineering Society: New York, NY, USA, 1992; Volume 92. [Google Scholar]
  5. Gerzon, M.A.; Barton, G.J. Ambisonic decoders for HDTV. In Proceedings of the Audio Engineering Society Convention 92, San Francisco, CA, USA, 1–4 October 1992; Audio Engineering Society: New York, NY, USA, 1992. [Google Scholar]
  6. Zotter, F.; Frank, M. Ambisonics: A Practical 3D Audio Theory for Recording, Studio Production, Sound Reinforcement, and Virtual Reality; Springer Nature: Berlin/Heidelberg, Germany, 2019. [Google Scholar]
  7. What Is IAMF? Available online: https://aomedia.org/specifications/iamf/ (accessed on 7 March 2025).
  8. Rudzki, T.; Kearney, G.; Skoglund, J. On the Design of the Binaural Rendering Library for Eclipsa Audio Immersive Audio Container. In Proceedings of the 158th Convention of the Audio Engineering Society, Warszawa, Poland, 22–24 May 2025. [Google Scholar]
  9. Mills, A.W. On the minimum audible angle. J. Acoust. Soc. Am. 1958, 30, 237–246. [Google Scholar] [CrossRef]
  10. Perrott, D.R.; Musicant, A.D. Minimum auditory movement angle: Binaural localization of moving sound sources. J. Acoust. Soc. Am. 1977, 62, 1463–1466. [Google Scholar] [CrossRef] [PubMed]
  11. Lee, H.; Johnson, D. 3D microphone array comparison: Objective measurements. J. Audio Eng. Soc. 2021, 69, 871–887. [Google Scholar] [CrossRef]
  12. Bertet, S.; Daniel, J.; Gros, L.; Parizet, E.; Warusfel, O. Investigation of the perceived spatial resolution of higher order ambisonics sound fields: A subjective evaluation involving virtual and real 3D microphones. In Proceedings of the Audio Engineering Society 30th International Conference, Saariselka, Finland, 15–17 March 2007. [Google Scholar]
  13. Braun, S.; Frank, M. Localization of 3D ambisonic recordings and ambisonic virtual sources. In Proceedings of the 1st International Conference on Spatial Audio, Detmold, Germany, 10–13 November 2011. [Google Scholar]
  14. Dietze, A.; Clapp, S.W.; Seeber, B.U. Static and moving minimum audible angle: Independent contributions of reverberation and position. JASA Express Lett. 2024, 4, 054404. [Google Scholar] [CrossRef] [PubMed]
  15. Meng, R.; Xiang, J.; Sang, J.; Zheng, C.; Li, X.; Bleeck, S.; Cai, J.; Wang, J. Investigation of an MAA test with virtual sound synthesis. Front. Psychol. 2021, 12, 656052. [Google Scholar] [CrossRef] [PubMed]
  16. Zargarnezhad, N.; Mesquita, B.; Macpherson, E.A.; Johnsrude, I. Focality of sound source placement by higher (ninth) order ambisonics and perceptual effects of spectral reproduction errors. J. Acoust. Soc. Am. 2025, 157, 2802–2818. [Google Scholar] [CrossRef] [PubMed]
  17. Gerken, M.; Hohmann, V.; Grimm, G. Comparison of 2D and 3D multichannel audio rendering methods for hearing research applications using technical and perceptual measures. Acta Acust. 2024, 8, 17. [Google Scholar] [CrossRef]
  18. Gerzon, M.A. Surround-sound psychoacoustics. Wirel. World 1974, 80, 483–486. [Google Scholar]
  19. Hardin, R.H.; Sloane, N.J. McLaren’s improved snub cube and other new spherical designs in three dimensions. Discret. Comput. Geom. 1996, 15, 429–441. [Google Scholar] [CrossRef]
  20. Zotter, F.; Pasqual, A.M. Radiation Modes of t-Design and Extremal-Points Compact Spherical Loudspeaker Arrays; Fortschritte der Akustik: Berlin, Germany, 2011. [Google Scholar]
  21. Aggius-Vella, E.; Kolarik, A.J.; Gori, M.; Cirstea, S.; Campus, C.; Moore, B.C.J.; Pardhan, S. Comparison of auditory spatial bisection and minimum audible angle in front, lateral, and back space. Sci. Rep. 2020, 10, 6279–6287. [Google Scholar] [CrossRef] [PubMed]
  22. Rummukainen, O.S.; Schlecht, S.J.; Habets, E.A.P. Self-translation induced minimum audible angle. J. Acoust. Soc. Am. 2018, 144, EL340–EL345. [Google Scholar] [CrossRef] [PubMed]
  23. Goldstein, E.B. Blackwell Handbook of Sensation and Perception; John Wiley & Sons: Hoboken, NJ, USA, 2008. [Google Scholar]
  24. Zotter, F.; Frank, M.; Pomberger, H. Comparison of Energy-Preserving and All-Round Ambisonic Decoders; Fortschritte der Akustik, AIA-DAGA (Meran): Berlin, Germany, 2013. [Google Scholar]
  25. Zotter, F.; Pomberger, H.; Noisternig, M. Energy-preserving ambisonic decoding. Acta Acust. United Acust. 2012, 98, 37–47. [Google Scholar] [CrossRef]
  26. Zotter, F.; Frank, M. All-round ambisonic panning and decoding. J. Audio Eng. Soc. 2012, 60, 807–820. [Google Scholar]
  27. Frank, M.; Zotter, F.; Sontacchi, A. Producing 3D audio in ambisonics. In Proceedings of the Audio Engineering Society Conference: 57th International Conference: The Future of Audio Entertainment Technology–Cinema, Television and the Internet, Los Angeles, CA, USA, 6–8 March 2015; Audio Engineering Society: New York, NY, USA, 2015. [Google Scholar]
  28. SPCMIC: Ambisonics for Human Beings. Available online: https://spcmic.com// (accessed on 23 February 2025).
  29. Kleczkowski, P.; Pluta, M.; Macura, P.; Paczkowska, E. Listeners who have low hearing thresholds do not perform better in difficult listening tasks. In Proceedings of the Audio Engineering Society Convention 132, Budapest, Hungary, 26–29 April 2012; Audio Engineering Society: New York, NY, USA, 2012. [Google Scholar]
  30. Sochaczewska, K.; Małecki, P.; Piotrowska, M. Evaluation of the Minimum Audible Angle on Horizontal Plane in 3rd order Ambisonic Spherical Playback System. In Proceedings of the 2021 Immersive and 3D Audio: From Architecture to Automotive (I3DA), Bologna, Italy, 8–10 September 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–5. [Google Scholar]
  31. Levitt, H. Transformed up-down methods in psychoacoustics. J. Acoust. Soc. Am. 1971, 49, 467–477. [Google Scholar] [CrossRef]
  32. Hall, J.L. Hybrid adaptive procedure for estimation of psychometric functions. J. Acoust. Soc. Am. 1981, 69, 1763–1769. [Google Scholar] [CrossRef] [PubMed]
  33. Brown, L.G. Additional rules for the transformed up-down method in psychophysics. Percept. Psychophys. 1996, 58, 959–962. [Google Scholar] [CrossRef] [PubMed]
  34. McCormack, L.; Politis, A. SPARTA & COMPASS: Real-time implementations of linear and parametric spatial audio reproduction and processing methods. In Proceedings of the AES International Conference on Immersive and Interactive Audio, York, UK, 27–29 March 2019; Audio Engineering Society: New York, NY, USA, 2019. [Google Scholar]
  35. Bertet, S.; Daniel, J.; Parizet, E.; Warusfel, O. Investigation on the restitution system influence over perceived Higher Order Ambisonics sound field: A subjective evaluation involving from first to fourth order systems. J. Acoust. Soc. Am. 2008, 123, 3936. [Google Scholar] [CrossRef]
  36. Frank, M.; Zotter, F.; Sontacchi, A. Localization experiments using different 2D ambisonics decoders (lokalisationsversuche mit verschiedenen 2D ambisonics dekodern). In Proceedings of the 25. Tonmeistertagung, Leipzig, Germany, 19–22 November 2008; pp. 696–704. [Google Scholar]
  37. Schütt, H.H.; Harmeling, S.; Macke, J.H.; Wichmann, F.A. Painfree and accurate Bayesian estimation of psychometric functions for (potentially) overdispersed data. Vis. Res. 2016, 122, 105–123. [Google Scholar] [CrossRef] [PubMed]
  38. Letowski, T.R.; Letowski, S.T. Auditory Spatial Perception: Auditory Localization; Technical Report; Army Research Laboratory Aberdeen Proving Ground MD Human Research and Engineering Directorate: Aberdeen, MD, USA, 2012. [Google Scholar]
Figure 1. Coordinate system convention in spatial sound reproduction.
Figure 1. Coordinate system convention in spatial sound reproduction.
Applsci 15 06815 g001
Figure 2. A human hearing system in an analogy to a pressure (large circle) and a velocity (figure-of-eight) microphone. Adapted from [18].
Figure 2. A human hearing system in an analogy to a pressure (large circle) and a velocity (figure-of-eight) microphone. Adapted from [18].
Applsci 15 06815 g002
Figure 3. Distribution of the objective measures for the combination of the investigated playback system and the SAD. Red dots mark the loudspeaker positions. Please note that the scales for r V angular error and energy are not equal to better visualise the differences in distributions.
Figure 3. Distribution of the objective measures for the combination of the investigated playback system and the SAD. Red dots mark the loudspeaker positions. Please note that the scales for r V angular error and energy are not equal to better visualise the differences in distributions.
Applsci 15 06815 g003
Figure 4. Distribution of the objective measures for the combination of the investigated playback system and the EPAD. Red dots mark the loudspeaker positions. Please note that the scales for r V angular error and energy are not equal to better visualise the differences in distributions.
Figure 4. Distribution of the objective measures for the combination of the investigated playback system and the EPAD. Red dots mark the loudspeaker positions. Please note that the scales for r V angular error and energy are not equal to better visualise the differences in distributions.
Applsci 15 06815 g004
Figure 5. Distribution of the objective measures for the combination of the investigated playback system and the AllRAD. Red dots mark the loudspeaker positions. Please note that the scales for r V angular error and energy are not equal to better visualise the differences in distributions.
Figure 5. Distribution of the objective measures for the combination of the investigated playback system and the AllRAD. Red dots mark the loudspeaker positions. Please note that the scales for r V angular error and energy are not equal to better visualise the differences in distributions.
Applsci 15 06815 g005
Figure 6. (Left) Visualisation of the loudspeaker layout in the AGH UST Laboratory of Auralisation rendered with the IEM AllRAD plugin. Yellow dots represent the imaginary loudspeakers. (Right) Photograph of the loudspeaker layout.
Figure 6. (Left) Visualisation of the loudspeaker layout in the AGH UST Laboratory of Auralisation rendered with the IEM AllRAD plugin. Yellow dots represent the imaginary loudspeakers. (Right) Photograph of the loudspeaker layout.
Applsci 15 06815 g006
Figure 7. Example of the course of the staircase procedure for one of the subjects. The threshold oscilates around 1–2° for a 0 reference.
Figure 7. Example of the course of the staircase procedure for one of the subjects. The threshold oscilates around 1–2° for a 0 reference.
Applsci 15 06815 g007
Figure 8. MAA results for the SAD. The Y-axis is cut at 40 —above that value it is assumed that the threshold is not possible to establish, as the biggest angular step in the experiment was 30 . Sab naming convention signifies subjects’ IDs.
Figure 8. MAA results for the SAD. The Y-axis is cut at 40 —above that value it is assumed that the threshold is not possible to establish, as the biggest angular step in the experiment was 30 . Sab naming convention signifies subjects’ IDs.
Applsci 15 06815 g008
Figure 9. MAA results for the EPAD. Similar to Figure 8, the Y-axis is cut at 40 , as above that value it is assumed that the threshold is not possible to establish. Sab signifies subjects’ IDs.
Figure 9. MAA results for the EPAD. Similar to Figure 8, the Y-axis is cut at 40 , as above that value it is assumed that the threshold is not possible to establish. Sab signifies subjects’ IDs.
Applsci 15 06815 g009
Figure 10. MAA results for the All-Round Ambisonic Decoder. As in Figure 8 and Figure 9, the Y-axis is cut at 40 . Sab signifies subjects’ IDs.
Figure 10. MAA results for the All-Round Ambisonic Decoder. As in Figure 8 and Figure 9, the Y-axis is cut at 40 . Sab signifies subjects’ IDs.
Applsci 15 06815 g010
Figure 11. Comparison of the mean results of the MAA threshold for all of the investigated decoders. Loudspeaker positions are 0 ,   45 ,   90 ,   135 ,   180 .
Figure 11. Comparison of the mean results of the MAA threshold for all of the investigated decoders. Loudspeaker positions are 0 ,   45 ,   90 ,   135 ,   180 .
Applsci 15 06815 g011
Figure 12. MAA values for a subject highly experienced with audio engineering, with high familiarity with the playback system.
Figure 12. MAA values for a subject highly experienced with audio engineering, with high familiarity with the playback system.
Applsci 15 06815 g012
Figure 13. MAA values for a subject highly experienced with audio engineering, with medium familiarity with the playback system.
Figure 13. MAA values for a subject highly experienced with audio engineering, with medium familiarity with the playback system.
Applsci 15 06815 g013
Figure 14. MAA values for a subject moderately experienced with audio engineering (student), with low familiarity with the playback system.
Figure 14. MAA values for a subject moderately experienced with audio engineering (student), with low familiarity with the playback system.
Applsci 15 06815 g014
Table 1. Spherical coordinates of the loudspeaker layout of the experimental stand. The radius was 1.5 m.
Table 1. Spherical coordinates of the loudspeaker layout of the experimental stand. The radius was 1.5 m.
IDAZIMUTHELEVATION
1 0 0
2 45 0
3 90 0
4 135 0
5 180 0
6 135 0
7 90 0
8 45 0
9 67.5 45
10 157.5 45
11 112.5 45
12 22.5 45
13 22.5 45
14 112.5 45
15 157.5 45
16 67.5 45
17 0 90
18 0 90
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sochaczewska, K.; Prawda, K.; Małecki, P.; Piotrowska, M.; Wiciak, J. Minimum Audible Angle in 3rd-Order Ambisonics in Horizontal Plane for Different Ambisonic Decoders. Appl. Sci. 2025, 15, 6815. https://doi.org/10.3390/app15126815

AMA Style

Sochaczewska K, Prawda K, Małecki P, Piotrowska M, Wiciak J. Minimum Audible Angle in 3rd-Order Ambisonics in Horizontal Plane for Different Ambisonic Decoders. Applied Sciences. 2025; 15(12):6815. https://doi.org/10.3390/app15126815

Chicago/Turabian Style

Sochaczewska, Katarzyna, Karolina Prawda, Paweł Małecki, Magdalena Piotrowska, and Jerzy Wiciak. 2025. "Minimum Audible Angle in 3rd-Order Ambisonics in Horizontal Plane for Different Ambisonic Decoders" Applied Sciences 15, no. 12: 6815. https://doi.org/10.3390/app15126815

APA Style

Sochaczewska, K., Prawda, K., Małecki, P., Piotrowska, M., & Wiciak, J. (2025). Minimum Audible Angle in 3rd-Order Ambisonics in Horizontal Plane for Different Ambisonic Decoders. Applied Sciences, 15(12), 6815. https://doi.org/10.3390/app15126815

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop