Next Article in Journal
Crack Identification for Bridge Condition Monitoring Combining Graph Attention Networks and Convolutional Neural Networks
Previous Article in Journal
A Formula-Structure-Embedding-Based GNN for Formula Equivalence Determination
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Voice-to-Text Transcription to Examine Outcomes of AirPods Pro Receivers When Used as Part of Remote Microphone System

1
Callier Center for Communication Disorders, Dallas, TX 75080, USA
2
Department of Speech, Language and Hearing Sciences, The University of Texas at Dallas, Richardson, TX 75080, USA
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(10), 5451; https://doi.org/10.3390/app15105451
Submission received: 17 March 2025 / Revised: 1 May 2025 / Accepted: 8 May 2025 / Published: 13 May 2025

Abstract

:

Featured Application

Smartphone-based remote microphone systems may provide speech recognition benefits in noisy environments.

Abstract

Hearing difficulty in noise can occur in 10–15% of listeners with typical hearing in the general population of the United States. Using one’s smartphone as a remote microphone (RM) system with the AirPods Pro (AP) may be considered an assistive device given its wide availability and potentially lower price. To evaluate this possibility, the accuracy of voice-to-text transcription for sentences presented in noise was compared, when KEMAR wore an AP receiver connected to an iPhone set to function as an RM system, to the accuracy obtained when it wore a sophisticated Phonak Roger RM system. A ten-sentence list was presented for six technology arrangements at three signal-to-noise ratios (SNRs; +5, 0, and −5 dB) in two types of noise (speech-shaped and babble noise). Each sentence was transcribed by Otter AI to obtain an overall percent accuracy for each condition. At the most challenging SNR (−5 dB SNR) across both noise types, the Roger system and smartphone/AP set to noise cancelation mode showed significantly higher accuracy relative to the condition when the smartphone/AP was in transparency mode. However, the major limitation of Bluetooth signal delay when using the AP/smartphone system would require further investigation in real-world settings with human users.

1. Introduction

Noise is common in everyday life, affecting communication not only for individuals with hearing impairments but also for those with typical hearing (TH), i.e., bilateral pure-tone thresholds ≤ 20 dB HL for 0.5–8 kHz in octave bands. Although pure-tone audiometry is widely recognized as the “gold standard” for assessing hearing function, a normal audiogram does not necessarily guarantee satisfactory communication experiences in all settings. Over 10% of adults may experience hearing difficulty in noisy environments despite having normal audiograms. Tremblay et al. [1] reported that 12% of 682 individuals aged 21–67 years with TH reported hearing difficulties. Similarly, 15% of 2015 subjects aged 20–69 years with bilateral pure-tone averages of ≤25 dB HL for 0.5, 1, 2, and 4 kHz experienced similar issues [2]. However, standard clinical guidelines from the American Academy of Audiology [3] and the American Speech-Language-Hearing Association [4] provide limited guidance for addressing these communication challenges for listeners with TH, aside from recommendations to use hearing protectors to prevent occupational noise-induced hearing loss [5]. Mealings et al. [6] found that many clients express frustration when told they have “normal hearing” while still experiencing hearing difficulties. Additionally, clinicians expressed frustration with a lack of adequate training or scientific resources to assist these clients beyond suggesting the use of communication strategies [6].
Possible solutions for these challenges are reviewed. The related work is divided into three sections. Section 1.1 presents one potential technological solution, which is the use of clinically available remote microphone (RM) systems, designed to enhance the signal-to-noise ratio (SNR). Section 1.2 addresses the use of smartphones as RM systems. Smartphones equipped with built-in microphones and Bluetooth capabilities (Bluetooth Special Interest Group, Kirkland, WA, USA) have been suggested as components of RM systems for use in small-group conversations for adults. Section 1.3 includes a review of the available research regarding the acoustic properties and potential benefits of using the AirPods Pro (AP) as a hearing device.

1.1. Clinically Available Remote Microphone Systems

These systems consist of a transmitter/microphone worn by the speaker and a receiver worn by the listener [7]. Research has shown that RM systems significantly improve speech recognition in noisy environments for individuals with hearing impairments and typical hearing. Thibodeau [8] reported that the Phonak Roger RM system enhanced speech recognition performance in persons with moderate-to-severe hearing loss from 45% to 61% compared to using hearing aids (HAs) or cochlear implants alone. In 2024, Thibodeau et al. [9] reported that RM systems benefited adults with TH. The observed improvements were 21.37% and 9.87% when using the Phonak Roger Select and Roger Pen, respectively, compared to no RM system in 10 participants with TH, aged 20 to 63 years [9]. Shiels and colleagues [10] found RM benefits for children aged 6 to 12 years. They reported improvements in the sentence recognition scores by an average of 3.03, 21.40, 21.61, and 49.95% when tested in SNRs of +12, +5, 0, and −10 dB, respectively. They also reported improvements in visual and auditory attention with RM use.
Although listeners with TH can potentially benefit from using RM systems designed for persons with hearing loss [9], the adoption of such devices is limited due to their high cost. For example, Mealings et al. [11] reported that none of the participants (n = 27) with TH expressed a preference for purchasing a pair of mild-gain HAs priced at approximately USD 3250, despite the potential benefits for their hearing difficulties. Additionally, the higher expenses associated with RM systems and the inconvenience of carrying multiple devices may deter individuals from using RM systems in daily communication. Furthermore, the social stigma often associated with using hearing devices can be a concern for some people. da Silva et al. [12] reported that persons with hearing impairment might be perceived as “less intelligent” or labelled as “deaf and dumb”, causing feelings of shame and embarrassment. These negative emotional feelings of self-imposed stigma may lead to social isolation and affect decision-making regarding the initial acceptance and pursuit of potential treatment [12]. Therefore, it is essential to find more affordable and approachable alternatives to increase the acceptance and use of hearing devices for addressing hearing difficulties.

1.2. Smartphone-Based Remote Microphone Systems

Although not designed to be worn at an optimal distance from the talker’s mouth, smartphones do not require external components or additional hardware to be used as remote microphones [13]. Furthermore, smartphones are readily used, with nearly everyone carrying one throughout the day. It has been reported that 98% of Americans own a cellphone or smartphone [14]. Moreover, Bluetooth-compatible headphones for iOS (Apple Inc., Cupertino, CA, USA) and Android (Open Handset Alliance, Mountain View, CA, USA) operating systems have been introduced, making it easier to integrate RM system functionality directly through smartphones [15]. However, there are limitations to using Bluetooth receivers as part of an RM system because of the delay caused by Bluetooth transmission. This delay has been reported to range from 30 to 274 ms [7,16]. Coyle reported that the average AP Bluetooth latency across 19 measurements was 144 ms compared to earlier AirPods versions that had latencies up to 274 ms [16]. Goehring et al. [17] reported a tolerable delay limit of 10 ms for TH listeners (n = 20) and 10–30 ms for listeners with hearing loss (n = 20), with annoyance ratings significantly increasing beyond these values. These findings suggest that Bluetooth transmission delays may affect the use of smartphone-based RM systems.
One smartphone-based RM system involves the iPhone and AP earphones (Apple Inc., Cupertino, CA, USA). It is noteworthy that the software available on compatible versions of the AP was recently approved by the U.S. Food and Drug Administration in 2024 as an over-the-counter HA for adults with mild-to-moderate hearing loss [18]. The AP can connect via Bluetooth with iPhones and Android phones and serve as receivers worn by the listener, with the smartphone functioning as a transmitter/microphone when placed close to the talker. In Android phones (i.e., Samsung, Google Pixel), the RM features that can be used for enhancing hearing perception in noise include “Hearing Enhancement” in the Samsung series, and “Sound Amplifier” on Google Pixel. In iPhone devices, Live Listen (LL) is an additional native feature, available since 2014, that enables microphone transmission to compatible hearing devices such as HA or AP headphones with Bluetooth protocols. When activated, LL uses the iPhone’s microphone to capture sound and transmit it wirelessly from the iPhone to the receiver (i.e., AP), amplifying and clarifying the sounds in their immediate environment for users [19]. The AP can be set at three modes: transparency (TP; lets outside sound in), noise cancellation (NC; cancels the external sounds), and off. This could potentially be a lower-cost alternative to a dedicated RM system designed for persons with TH. One such RM system, a Phonak Roger Touchscreen transmitter used with a Roger Focus II receiver, was shown to enhance speech intelligibility by an average of 53% in 16 children (ages 8–16 years) with unilateral hearing loss, compared to their peers with TH [20].

1.3. Previous Research for Use of AirPods Pro as Hearing Devices

Despite growing research interest driven by the AP’s affordability, widespread adoption, and regulatory approval as an over-the-counter hearing aid [18], evidence supporting its clinical use remains limited. Prior to 2025, only six studies systematically evaluated the AP’s performance as hearing devices, as chronologically summarized in Table 1 [21,22,23,24,25,26]. These investigations, comprising peer-reviewed articles, conference proceedings, trade publications, and graduate theses, collectively suggest potential auditory benefits of using the AP for both TH and adults with hearing loss. For instance, Lin et al. [21] verified that AP receivers have comparable electroacoustic results to HAs, providing adequate amplification for individuals with mild-to-moderate hearing loss. Hammond and Diedesch [22] found that listeners appreciated the custom audiogram-driven features in the AP, finding them easy to use and beneficial regardless of hearing status. Valderrama et al. [24] reported that using the AP and an iPhone mitigated hearing challenges for individuals with TH, resulting in a significant 11.8% increase in speech intelligibility and a +5.5 dB SNR advantage compared to baseline conditions without the AP (unaided). Only one master’s thesis by Foroogozar [26] investigated the use of the AP as part of a smartphone-based remote microphone system in 23 adults aged 60 and above with normal to mild/moderate hearing loss. Use of the LL feature on the smartphone versus the AP alone showed significant improvements in memory retention from 43.8% (no LL) to 59.4% (with LL) and mean sentence recognition scores from 81.8% (no LL) to 94.4% (with LL). These promising findings suggest the need for more rigorous clinical studies to validate their efficacy in audiological practice.
Considering the focus of this study was to evaluate the AP when used with a smartphone set to LL compared to a current clinically available RM system, it is essential to provide objective verification and limit potential confounding human factors (e.g., age, gender, education, personality, attention, linguistic familiarity, cognition). An objective, non-biased speech recognition method was developed following the COVID-19 pandemic, which disrupted traditional methods of collecting experimental data involving human subjects. This new approach was developed utilizing voice-to-text transcription (VTT), the Knowles Electronics Manikin for Acoustic Research (KEMAR) with a standardized artificial ear (Zwislocki coupler), to replace human responses. Advancements in artificial intelligence have led to the creation of various transcription tools designed to convert spoken words into text with great accuracy as a way to facilitate closed captioning [27]. Such VTT applications include Otter.ai (https://otter.ai/ (accessed on 4 January 2024)), Sonix (https://sonix.ai (accessed on 4 January 2024)), Trint (https://trint.com/ (accessed on 4 January 2024)), and Google Cloud Speech-to-Text (https://cloud.google.com/speech-to-text/ (accessed on 4 January 2024)). In a comparative study assessing various VTT tools, Otter.ai was reported to provide the most accurate transcripts with an accuracy rate of 99.7% [28]. Given this high accuracy rate, Otter.ai was utilized for VTT in this study.
The aim of this study was to compare the VTT accuracy of two RM systems (the iPhone set to LL and the Roger RM), specifically designed for users with TH. The investigation involved three SNRs, one HINT list, and two types of noise. The research question was as follows: how does the transcription accuracy for AP in TP and NC modes change when used with an iPhone set to LL compared to the Roger RM system across two noise types and three SNR levels? It was hypothesized that transcription accuracy would be superior in speech-shaped noise compared to babble noise, for the higher SNR conditions, and with the Roger RM system relative to the smartphone-based RM system. The results confirmed these hypotheses. The key findings are as follows: (1) The transcription accuracy was significantly affected by the SNR, noise type, RM condition, and the two-way interaction between SNR and RM condition. (2) In the −5 dB SNR noise condition, the smartphone-based RM with the AP in NC mode using the LL feature yielded comparable VTT accuracy relative to the Roger RM system.

2. Materials and Methods

2.1. Equipment

The accuracy of speech recognition in noise of a smartphone-based RM system (iPhone set to LL) was compared with a clinically available RM system (Phonak Roger On with Roger Focus II), both designed for TH users. The equipment used for this comparison is described in Table 2 and illustrated in Figure 1.

2.2. Stimuli

2.2.1. Sentences

A Hearing in Noise Test (HINT) [29] list (i.e., 20) was selected as the input, containing 10 sentences, as shown in Table 3. The 10 sentences consisted of a total of 54 words. All sentences were presented at 65 dB SPL through a speaker positioned at 0° azimuth from the KEMAR dummy head at three feet. The same list was played continuously four times for each test condition, with the last three used for scoring. The sentences were generated on a Lenovo ThinkPad P70 laptop and then transmitted to channel A of the GSI-61 audiometer.

2.2.2. Noise

To assess the efficacy of VTT accuracy, three SNRs were tested: +5, 0, and −5 dB, which commonly represent everyday listening scenarios. For instance, in acoustically untreated classrooms, SNRs can range from −7 to +6 dB [30]. Two noise types were employed: speech-weighted steady-state and multi-talker babble noise [31], representing the most common energetic and informational masking in daily life. Spille et al. [32] found that the use of babble noise generally resulted in higher speech reception thresholds than speech-shaped noise in humans and automatic speech recognition tasks. This suggests the need to evaluate the potential benefits of RM systems in both types of noise to determine if differences are related to the greater difficulty of informational masking over energetic masking during speech recognition. The noise was continuously presented at 60, 65, and 70 dB SPL (SNRs of +5, 0, and −5 dB, respectively) through a speaker located at 180° azimuth from the KEMAR dummy head at eight feet. The babble noise was generated on the Lenovo ThinkPad P70 laptop and transmitted to channel B of the GSI-61 audiometer, while the speech noise was directly produced by the GSI-61 audiometer. The order of noise was fixed for the SNR conditions but counterbalanced across the technology conditions.

2.2.3. Scoring

To transcribe the sentences played in each noise type at different SNRs, the online program Otter.ai (version 2.3.116) was used on a Dell Latitude 7480 laptop connected to the KEMAR dummy head. Each accurately transcribed word was scored one point. The total percentage of correctly transcribed words was calculated for each list, with a denominator of 54. To account for potential delays in Otter’s transcription during the initial trial, only the last three of the four presentations of the list of sentences were scored.

2.3. Procedures

2.3.1. Setup

The setup of the testing procedure is shown in Figure 1. Stimuli were presented to speakers 1 and 2 inside the sound booth, transmitted by laptop 1 (ThinkPad P70; Lenovo) and the audiometer (Grason Stadler GSI 61). Within the sound booth, speaker 1 continuously played sentence signals (HINT list 20) at 65 dB SPL, while speaker 2 continuously played noise signals at 60, 65, and 70 dB SPL. These signals were received by either an iPhone or Roger On through air conduction and then transmitted to the AP/Roger Focus II worn by KEMAR on the right ear. The power module integrated into KEMAR received the transmitted signal and subsequently transmitted it to laptop 2 (Latitude 7480; Dell). Finally, the sentences were transcribed using the Otter VTT program on laptop 2.

2.3.2. Test Conditions

Six test conditions were categorized into three baseline settings and three Roger RM settings. The three baseline settings included no use of a smartphone: (1) AP NC: AP in noise cancellation mode (NC); (2) AP TP: AP in transparency mode (TP); and (3) KEMAR: without any additional technological devices (i.e., no iPhone, AP, or Roger system). The three RM settings were as follows: (1) AP NC + LL: AP set to NC while wirelessly connected to the iPhone using LL; (2) AP TP + LL: AP set to TP while wirelessly connected to the iPhone using LL; and (3) Roger: Roger On transmitter with a Roger Focus II receiver.

3. Results

The mean transcription accuracy scores for No RM conditions (Baseline) and RM conditions are presented in Table 4 and Table 5 and Figure 2. There were three factors of interest, including technology condition, noise type, and SNR. In general, RM conditions yielded greater accuracy than the No RM conditions. Of initial interest was the comparison of baseline conditions as shown in Table 4. The use of AP alone without the smartphone was compared to KEMAR alone to determine improvements related to the AP alone. Contrary to the results obtained with humans, the accuracy of the VTT was higher for the KEMAR alone condition compared to the two AP conditions, most likely due to the occlusion effect of the AP and the lack of binaural processing. Therefore, given the uniqueness of this arrangement, no statistical analyses were completed for the baseline conditions.
Of greater interest was the comparison of the RM technology conditions shown in Table 5. The statistical analyses included a repeated-measures ANOVA for the RM technology conditions of interest (Roger, AP NC + LL, and AP TP + LL), noise type (speech and babble), and SNR (+5, 0, and −5 dB). Significant main effects were observed for technology [F(2,34) = 47.45, p < 0.001, large partial η22p; effect size) = 0.74, Mean Square Error (MSE) = 20.55], noise type [F(1,34) = 8.65, p = 0.006, large η2p = 0.20, MSE = 20.55], and SNR [F(2,34) = 42.91, p < 0.001, large η2p = 0.72, MSE = 20.55]. The accuracy was significantly higher when tested in speech noise compared to babble noise (84.22% vs. 79.89%) with a small effect size (Cohen’s d = 0.35). A significant two-way interaction was only found between the SNR and technology [F(4,34) = 6.03, p = 0.001, large η2p = 0.42, MSE = 20.55]. All other interactions were non-significant (p > 0.05).
Follow-up analyses were performed using Tukey’s adjustment for each significant effect (p-value was marked as padj). For the main effect of RM technology conditions, the transcription accuracy of Roger (95.06%) was significantly better than AP TP + LL (81.22%) (t34 = 9.15, padj < 0.001; large Cohen’s d = 1.50) but not better than AP NC + LL (92.50%) (t34 = 1.69, padj = 0.22, small Cohen’s d = 0.41). There was also a significant difference between the two AP conditions that AP NC + LL was significantly greater than AP TP + LL (t34 = 7.46, padj < 0.001; large Cohen’s d = 1.17). For the main effect of three SNR conditions, all comparisons (+5 vs. 0, 0 vs. −5, and +5 vs. −5 dB) were significant as expected, with t34 = 3.35, 5.81, and 9.15, respectively, and all padj < 0.002. The relevant effect sizes were all large (Cohen’s d = 0.80, 0.85, 1.52, respectively).
Figure 3 illustrates the two-way interaction between the settings of the SNR and RM technology conditions across both noise types. The follow-up pairwise comparisons are shown in Table 6. The differences among the conditions were most evident at the more challenging SNR conditions (0 and −5 dB). The patterns of significance at these SNRs were the same as the main effects, with Roger and AP NC + LL significantly greater than AP TP + LL, but there were no significant differences between them.

4. Discussion

The aim of the study was to compare the transcription accuracy in noise when using AP as part of an RM system (iPhone set to LL) and when using a sophisticated RM system (Phonak Roger). Three factors were involved, including technology conditions, noise type, and SNR. The results indicated that the transcription accuracy was significantly influenced by all three factors and a two-way interaction of SNR by technology. Overall, accuracy scores were the lowest at −5 compared to 0 and +5 dB SNR, and the highest when using an RM such as Roger On transmitting to a Roger receiver or using a smartphone with LL transmitting via Bluetooth low energy to an AP set to NC mode.
Regarding the baseline conditions (Table 4), it is interesting to note that using the AP alone did not increase the accuracy score. Because of the somewhat artificial nature of the VTT arrangement with a manikin, the baseline results are provided for information purposes. As mentioned earlier, the results may be impacted by the fitting of the AP on an artificial pinna. It should also be noted that in this test arrangement using KEMAR, the result is obtained monaurally, so that the benefits of binaural listening that are available with human listeners were not observed.
However, when using an RM system (Table 5), there were improvements in the transcription accuracy relative to using KEMAR alone. Valderrama et al. [24] reported that using the AP on TH adults with self-disclosed difficulties hearing in noise, set for “maximum ambient noise reduction” and “conversation boost” enabled, provided a 5.4 dB SNR advantage and 11.8% intelligibility increase. It is likely that if they had included the use of the smartphone as an RM system, the improvements would have been greater. In addition, there was an 8% reduction in mental demand and listening effort. In the present study, the most accurate VTT score with the AP was obtained when using LL and with the AP also set to NC mode (92.5%). This agrees with Foroogozar [26], who reported 94.4% accurate sentence recognition on average when testing adults with typical hearing wearing the AP with an iPhone set to LL. The NC mode deactivates the microphones on the AP and allows for signals with the highest quality as the input to the Otter transcription program. Further research is required to confirm such benefits in humans with different degrees of hearing impairment.
When considering the interaction between the SNR and technology, the most demanding listening conditions (0 and −5 dB SNR) revealed significant differences in accuracy among the three RM conditions. The accuracy was significantly lower when using AP TP + LL compared to Roger and AP NC + LL, although there was no significant difference between Roger and AP NC + LL. This suggests that RM features offer potential benefits relative to AP TP + LL mode, which allows the transmission of environmental noise. In the AP TP mode, the environmental noise is mixed with the signal arriving from the RM, thus reducing its potential benefit. Differences between smartphone-based RM and Roger systems may stem from Bluetooth transmission delays (averaging 144 ms in AP [16]), which are not present in Roger’s digital modulation transmission. These delays substantially exceed the reported tolerable thresholds of 10–30 ms for hearing devices [17]. The asynchrony becomes particularly apparent when visual cues from the speaker are present, creating a temporal mismatch between auditory and visual inputs. This incongruity may strain working memory capacity during noisy listening conditions, as listeners must reconcile limited acoustic information with potentially conflicting visual cues. Such cognitive demands can delay speech comprehension and increase processing load during ongoing communication.
These results have potential clinical and research implications. Using Roger and smartphone-based RM is beneficial in reducing the challenges caused by noise relative to not using any device across the SNR and noise types, with benefits of 17.95, 15.39, and 4.11% in Roger, AP NC + LL, and AP TP + LL, respectively. Considering the current device setting is mainly for those without hearing loss, such benefits can potentially help improve hearing difficulties in noise for listeners with TH, occupying 12–15% of the general population [1,2]. The benefits of the Roger RM system were confirmed by Thibodeau et al. [9] in 10 subjects with TH despite using different Roger transmitters (Pen, Select) and receivers (Roger Focus-first generation). Similar to the RM system, the benefits of using smartphone-based RM on speech recognition in noise are expected in humans with TH. It is also of possible benefit to those who have hearing aids that connect to smartphones when using a remote mic app on the phone, if the Bluetooth delay can be tolerated. However, many HA manufacturers do offer proprietary RM devices now with personal hearing technology.
In addition, an iPhone and the AP can work as a portable RM system in daily communication for listeners with TH. This is similar to the findings in Table 1 on the benefit of using the AP as a hearing assistance device [21,22,23,24,25,26]. Given the widespread adoption and convenience of smartphones and earbuds, along with the lower cost compared to Roger devices, smartphone-based RM systems can be considered as alternative RM systems for TH individuals with hearing difficulties in noise, particularly those concerned about cost or the stigma associated with traditional hearing devices. Additionally, with the governmental approval of the software for iPhones to adjust the AP for persons with mild-to-moderate hearing impairments, such systems may function both as RM systems and as hearing assistive devices. However, further research on humans is required before these solutions can be suggested as part of clinical protocols.
Using VTT and KEMAR has the potential to provide objective verification of new technological features, especially in situations where real participants are unavailable due to constraints such as COVID-19 or lack of funding. This approach can yield objective results that are not influenced by human factors such as age, gender, emotional status, personality, cognitive functions, etc. However, to ensure comparability across different studies, the version of the transcription application (e.g., Otter) should be specified, given the rapid pace of feature development in this high-tech era.
Although the results suggested a promising testing method for research and highlighted the potential benefits of using RM for listeners with TH, there are limitations to consider. Firstly, the use of KEMAR may restrict the applicability of the results to humans. Further research involving human participants with various profiles (e.g., age, cognition, and hearing status) is necessary to validate these findings before adopting such a smartphone-based RM system for rehabilitative solutions. Secondly, only one smartphone-based RM system using an iOS device and a single type of headphone was evaluated. This limitation restricts the generalizability of the results to other smartphone platforms (e.g., Android) and different headphones with similar features like LL. Finally, while the cost of an iPhone and the AP may be lower than that of a sophisticated Roger system, these devices may still be expensive for individuals, especially if they are not iPhone users.
This study provides preliminary evidence on the AP’s performance as part of an RM system relative to a Phonak RM system, without considering human participant variability. While the results suggest that the AP may help improve speech recognition in noise under controlled conditions, further research with humans is needed to validate these findings. Suggested future steps include the following: (1) Human subject validation: Evaluating performance across diverse populations (varying hearing levels, ages, genders, and cognitive abilities) to confirm real-world applicability. (2) Behavioral response analysis: Examining differences in human performance (speech recognition, listening effort) and subjective evaluation (sound quality perception, technology acceptance) under the same testing conditions in this study. (3) Signal transmission optimization: Engineering studies to address Bluetooth latency issues, ensuring real-time auditory synchronization meets clinical requirements for assistive listening. These investigations will improve the generalizability of the results and provide deeper insights into the clinical feasibility of AP as an assistive listening device.

5. Conclusions

This study presents the first systematic evaluation of using a smartphone as an RM with AP technologies in a manikin-based VTT paradigm, which enables objective assessment of transmission factors (i.e., SNR) and sound characteristics (i.e., noise type) while controlling for human variability. The results demonstrated better transcription accuracy in speech-shaped noise compared to babble noise at higher SNRs and in RM versus baseline (non-RM) conditions. In addition, the AP (NC mode)/iPhone (LL feature) configuration yielded comparable VTT performance to the clinically available Roger system in challenging noise environments, suggesting its potential to work as a hearing technology when properly configured.
However, two challenges must be addressed before the smartphone/AP RM system can be adopted as a solution for people with hearing difficulties in noisy environments: the current lack of human performance data and the presence of Bluetooth transmission delays. Translating these promising findings into practice will require both behavioral validation with human participants and engineering advancements to reduce signal transmission latency. Given the widespread availability and cost-effectiveness of smartphone platforms, overcoming these limitations could lead to more direct access to hearing assistance.

Author Contributions

Conceptualization, L.T.; methodology, L.T.; formal analysis, S.Q.; investigation, S.Q.; data curation, S.Q.; writing—original draft preparation, S.Q.; writing—review and editing, S.Q. and L.T.; visualization, S.Q.; supervision, L.T.; project administration, L.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw VTT data reported in this study are available in Appendix A: Table A1. Raw transcription accuracy data. Further inquiries can be directed to the corresponding author.

Acknowledgments

The authors want to thank Phonak for providing the Roger equipment for the study. During the preparation of this manuscript/study, the first author used [DeepSeek-V3 and GPT-4o] for the purposes of sentence refinement based on the authors’ original wording. The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
APAirPods Pro
AP 2AirPods Pro 2
BKBBamford–Kowal–Bench
CBConversation Boost
HAHearing aid
HINTHearing in noise test
KEMARKnowles Electronics Manikin for Acoustic Research
LLLive Listen
MHINTMandarin hearing in noise test
MSEMean Square Error
NAL-NL2National Acoustics Laboratories-Non-Linear prescription, 2nd generation
NCNoise cancelation
OTCOver-the-counter
PSAPPersonal sound amplification products
RMRemote microphone
SNRSignal-to-noise ratio
SRTSpeech reception threshold
TPTransparency
THTypical hearing
VTTVoice-to-text

Appendix A

Table A1. Raw transcription accuracy data.
Table A1. Raw transcription accuracy data.
NoiseConditionSNR1st2nd3rdMeanSD
SpeechRoger+598.0098.00100.0098.671.15
SpeechRoger098.0098.00100.0098.671.15
SpeechRoger−598.0091.0094.0094.333.51
SpeechAP TP + LL+591.0098.0094.0094.333.51
SpeechAP TP + LL081.0087.0085.0084.333.06
SpeechAP TP + LL−569.0080.0067.0072.007.00
SpeechAP NC + LL+596.00100.00100.0098.672.31
SpeechAP NC + LL094.0098.0094.0095.332.31
SpeechAP NC + LL−583.0087.0089.0086.333.06
SpeechKEMAR+596.0094.0098.0096.002.00
SpeechKEMAR078.0078.0083.0079.672.89
SpeechKEMAR−543.0056.0048.0049.006.56
SpeechAP TP+590.7483.3383.3385.804.28
SpeechAP TP062.9664.8174.0767.285.95
SpeechAP TP−544.4451.8538.8945.066.50
SpeechAP NC+594.4494.4492.5993.831.07
SpeechAP NC087.0472.2277.7879.017.48
SpeechAP NC−551.8540.7446.3046.305.56
BabbleRoger+589.00100.00100.0096.336.35
BabbleRoger093.0098.0096.0095.672.52
BabbleRoger−578.0091.0091.0086.677.51
BabbleAP TP + LL+591.0093.0091.0091.671.15
BabbleAP TP + LL072.0089.0076.0079.008.89
BabbleAP TP + LL−554.0070.0074.0066.0010.58
BabbleAP NC + LL+591.0098.0098.0095.674.04
BabbleAP NC + LL089.0093.0094.0092.002.65
BabbleAP NC + LL−572.0093.0096.0087.0013.08
BabbleKEMAR+594.0091.0096.0093.672.52
BabbleKEMAR081.0083.0087.0083.673.06
BabbleKEMAR−554.0059.0069.0060.677.64
BabbleAP TP+594.4496.3087.0492.594.90
BabbleAP TP092.5966.6775.9378.4013.14
BabbleAP TP−557.4159.2661.1159.261.85
BabbleAP NC+592.5992.5996.3093.832.14
BabbleAP NC070.3781.4883.3378.407.01
BabbleAP NC−540.7451.8542.5945.065.95
NOTE: AP = AirPods Pro; NC = noise cancellation; TP = transparency; SNR = signal-to-noise ratio; KEMAR means the KEMAR alone condition; Roger means the Roger remote microphone condition.

References

  1. Tremblay, K.L.; Pinto, A.; Fischer, M.E.; Klein, B.E.K.; Klein, R.; Levy, S.; Tweed, T.S.; Cruickshanks, K.J. Self-Reported Hearing Difficulties among Adults with Normal Audiograms: The Beaver Dam Offspring Study. Ear Hear. 2015, 36, e290–e299. [Google Scholar] [CrossRef]
  2. Spankovich, C.; Gonzalez, V.B.; Su, D.; Bishop, C.E. Self Reported Hearing Difficulty, Tinnitus, and Normal Audiometric Thresholds, the National Health and Nutrition Examination Survey 1999–2002. Hear. Res. 2018, 358, 30–36. [Google Scholar] [CrossRef] [PubMed]
  3. Practice Guidelines and Standard; American Academy of Audiology: Reston, VA, USA. Available online: https://www.audiology.org/practice-resources/practice-guidelines-and-standards/ (accessed on 16 March 2025).
  4. ASHA Practice Policy; American Speech-Language-Hearing Association: Rockville, MD, USA. Available online: https://www.asha.org/policy/ (accessed on 16 March 2025).
  5. Position Statement: Preventing Noise-Induced Occupational Hearing Loss; American Academy of Audiology: Reston, VA, USA. Available online: https://www.audiology.org/practice-guideline/position-statement-preventing-noise-induced-occupational-hearing-loss/ (accessed on 16 March 2025).
  6. Mealings, K.; Yeend, I.; Valderrama, J.T.; Gilliver, M.; Pang, J.; Heeris, J.; Jackson, P. Discovering the Unmet Needs of People with Difficulties Understanding Speech in Noise and a Normal or Near-Normal Audiogram. Am. J. Audiol. 2020, 29, 329–355. [Google Scholar] [CrossRef] [PubMed]
  7. Thibodeau, L.M. Between the Listener and the Talker: Connectivity Options. Semin. Hear. 2020, 41, 247–253. [Google Scholar] [CrossRef] [PubMed]
  8. Thibodeau, L.M. Benefits in Speech Recognition in Noise with Remote Wireless Microphones in Group Settings. J. Am. Acad. Audiol. 2020, 31, 404–411. [Google Scholar] [CrossRef]
  9. Thibodeau, L.M.; Land, V.; Sivaswami, A.; Qi, S. Benefits of Speech Recognition in Noise Using Remote Microphones for People with Typical Hearing. J. Commun. Disord. 2024, 112, 106467. [Google Scholar] [CrossRef]
  10. Shiels, L.; Tomlin, D.; Rance, G. The Assistive Benefits of Remote Microphone Technology for Normal Hearing Children with Listening Difficulties. Ear Hear. 2023, 44, 1049. [Google Scholar] [CrossRef]
  11. Mealings, K.; Valderrama, J.T.; Mejia, J.; Yeend, I.; Beach, E.F.; Edwards, B. Hearing Aids Reduce Self-Perceived Difficulties in Noise for Listeners with Normal Audiograms. Ear Hear. 2024, 45, 151–163. [Google Scholar] [CrossRef]
  12. da Silva, J.C.; de Araujo, C.M.; Lüders, D.; Santos, R.S.; Moreira de Lacerda, A.B.; José, M.R.; Guarinello, A.C. The Self-Stigma of Hearing Loss in Adults and Older Adults: A Systematic Review. Ear Hear. 2023, 44, 1301. [Google Scholar] [CrossRef]
  13. Shankar, N.; Bhat, G.S.; Panahi, I.M.S.; Tittle, S.; Thibodeau, L.M. Smartphone-Based Single-Channel Speech Enhancement Application for Hearing Aids. J. Acoust. Soc. Am. 2021, 150, 1663–1673. [Google Scholar] [CrossRef]
  14. Mobile Fact Sheet; Pew Research Center: Internet, Science & Tech: Washington, DC, USA, 2024. Available online: https://www.pewresearch.org/internet/fact-sheet/mobile/ (accessed on 16 March 2025).
  15. Thibodeau, L.M.; Panahi, I. Smartphone Technology: The Hub of a Basic Wireless Accessibility Plan. Hear. J. 2021, 74, 6. [Google Scholar] [CrossRef]
  16. Coyle, S. AirPods Pro Bluetooth Latency-Stephen Coyle. Available online: https://stephencoyle.net/airpods-pro (accessed on 29 April 2025).
  17. Goehring, T.; Chapman, J.L.; Bleeck, S.; Monaghan, J.J. Tolerable Delay for Speech Production and Perception: Effects of Hearing Ability and Experience with Hearing Aids. Int. J. Audiol. 2018, 57, 61–68. [Google Scholar] [CrossRef]
  18. U.S. Food and Drug Administration. FDA Authorizes First Over-the-Counter Hearing Aid Software. Available online: https://www.fda.gov/news-events/press-announcements/fda-authorizes-first-over-counter-hearing-aid-software (accessed on 16 March 2025).
  19. Gilmore, J.N. Design for Everyone: Apple AirPods and the Mediation of Accessibility. Crit. Stud. Media Commun. 2019, 36, 482–494. [Google Scholar] [CrossRef]
  20. Nelson, J.; Dunn, A. Roger Focus II in Children with Unilateral Hearing Loss; Phonak: Stäfa, Switzerland, 2021; Available online: https://www.phonak.com/content/dam/phonak/en/evidence-library/field-studies/PH_FSN_Roger_Focus_II_in_children_with_%20UHL_210x280_EN_V1.00.pdf (accessed on 16 March 2025).
  21. Lin, H.-Y.H.; Lai, H.-S.; Huang, C.-Y.; Chen, C.-H.; Wu, S.-L.; Chu, Y.-C.; Chen, Y.-F.; Lai, Y.-H.; Cheng, Y.-F. Smartphone-Bundled Earphones as Personal Sound Amplification Products in Adults with Sensorineural Hearing Loss. iScience 2022, 25, 105436. [Google Scholar] [CrossRef] [PubMed]
  22. Hammond, E.; Diedesch, A.C. Usability and Perceived Benefit of Hearing Assistive Features on Apple AirPods Pro. Proc. Meet. Acoust. 2023, 51, 050004. [Google Scholar] [CrossRef]
  23. Martinez, S. Evaluating Apple AirPods pro 2 Hearing Protection and Listening; The Hearing Review: Overland Park, KS, USA, 2023; Available online: https://hearingreview.com/inside-hearing/research/evaluating-apple-airpods-pro-2-for-hearing-protection-and-listening (accessed on 16 March 2025).
  24. Valderrama, J.T.; Mejia, J.; Wong, A.; Chong-White, N.; Edwards, B. The Value of Headphone Accommodations in Apple Airpods pro for Managing Speech-in-Noise Hearing Difficulties of Individuals with Normal Audiograms. Int. J. Audiol. 2024, 63, 447–457. [Google Scholar] [CrossRef]
  25. Kim, G.-Y.; Yun, H.J.; Jo, M.; Jo, S.; Cho, Y.S.; Moon, I.J. Apple AirPods pro as a Hearing Assistive Device in Patients with Mild to Moderate Hearing Loss. Yonsei Med. J. 2024, 65, 596. [Google Scholar] [CrossRef] [PubMed]
  26. Foroogozar, M. Assessing the Influence of Apple AirPods with Live Listen Feature on Speech Recognition and Memory Retention in Noise Levels Simulating Noisy Healthcare Settings-Insights from QuickSIN; Arizona State University: Tempe, AZ, USA, 2024; Available online: https://keep.lib.asu.edu/items/193345 (accessed on 16 March 2025).
  27. Goldenthal, E.; Park, J.; Liu, S.X.; Mieczkowski, H.; Hancock, J.T. Not All AI Are Equal: Exploring the Accessibility of AI-Mediated Communication Technology. Comput. Hum. Behav. 2021, 125, 106975. [Google Scholar] [CrossRef]
  28. Millett, P. Accuracy of Speech-to-Text Captioning for Students Who Are Deaf or Hard of Hearing. J. Educ. Pediatr. (Re) Habilit. Audiol. 2021, 25, 1–13. [Google Scholar]
  29. Nilsson, M.; Soli, S.D.; Sullivan, J.A. Development of the Hearing in Noise Test for the Measurement of Speech Reception Thresholds in Quiet and in Noise. J. Acoust. Soc. Am. 1994, 95, 1085–1099. [Google Scholar] [CrossRef]
  30. Zanin, J.; Rance, G. Functional Hearing in the Classroom: Assistive Listening Devices for Students with Hearing Impairment in a Mainstream School Setting. Int. J. Audiol. 2016, 55, 723–729. [Google Scholar] [CrossRef] [PubMed]
  31. Van Engen, K.J. Speech-in-Speech Recognition: A Training Study. Lang. Cogn. Process. 2012, 27, 1089–1107. [Google Scholar] [CrossRef]
  32. Spille, C.; Kollmeier, B.; Meyer, B.T. Comparing Human and Automatic Speech Recognition in Simple and Complex Acoustic Scenes. Comput. Speech Lang. 2018, 52, 123–140. [Google Scholar] [CrossRef]
Figure 1. Equipment setup from the overhead view. AP = AirPods Pro; KEMAR = Knowles Electronics Manikin for Acoustic Research.
Figure 1. Equipment setup from the overhead view. AP = AirPods Pro; KEMAR = Knowles Electronics Manikin for Acoustic Research.
Applsci 15 05451 g001
Figure 2. Mean of transcription accuracy scores under different conditions: (A) under speech-shaped noise; (B) under babble noise. NOTE: RM = remote microphone; AP = AirPods Pro; TP = transparency; NC = noise cancellation; LL = Live Listen; KEMAR = Knowles Electronics Manikin for Acoustic Research; SNR = signal-to-noise ratio. Error bars represent + one standard deviation.
Figure 2. Mean of transcription accuracy scores under different conditions: (A) under speech-shaped noise; (B) under babble noise. NOTE: RM = remote microphone; AP = AirPods Pro; TP = transparency; NC = noise cancellation; LL = Live Listen; KEMAR = Knowles Electronics Manikin for Acoustic Research; SNR = signal-to-noise ratio. Error bars represent + one standard deviation.
Applsci 15 05451 g002
Figure 3. Averaged scores across noise types for the three RM conditions. NOTE: RM = remote microphone, AP = AirPods Pro, NC = noise cancellation, TP = transparency, SNR = signal-to-noise ratio, LL = Live Listen. Error bars represent + one standard deviation.
Figure 3. Averaged scores across noise types for the three RM conditions. NOTE: RM = remote microphone, AP = AirPods Pro, NC = noise cancellation, TP = transparency, SNR = signal-to-noise ratio, LL = Live Listen. Error bars represent + one standard deviation.
Applsci 15 05451 g003
Table 1. A summary of studies using the AirPods (Pro) as a hearing assistive device in chronological order.
Table 1. A summary of studies using the AirPods (Pro) as a hearing assistive device in chronological order.
StudyParticipantsMethodsFindings
N and AgeHearingDevicesTests
Lin et al. (2022) [21]
  • N= 21
  • Mean = 42.9 years
Mild to moderate hearing loss
  • iPhone XS
  • AirPods 2
  • AP
  • A premium HA (OTICON Opn 1)
  • A basic HA (Bern-afon MD 1)
  • Electroacoustic analysis
  • SRT and MHINT in quiet and in noise, aided versus unaided
  • AP has comparable electroacoustic results to HAs and thus can provide proper amplification for listeners with mild-to-moderate hearing loss, while AirPods 2 does not.
  • SRT and MHINT scores varied with different directions of the noise source. AP worked similarly to HAs when tested in easier discriminating conditions (noise from another direction), while significantly poorer when tested in harder conditions (speech and noise came in the same direction).
Hammond & Diedesch (2023) [22]
  • N = 25
  • 24 in 18–25 years.
  • one in 46–60 years
7 with self-reported hearing loss and 18 with self-reported TH
  • Participants’ personal iPhone
  • AP
  • Mimi Hearing Test
  • AzBio sentences in quiet and noise (+5 dB SNR)
  • Sound quality ratings
  • A ten-item questionnaire on usability, benefit, and satisfaction.
  • Usability: Participants found the custom features easy to use. All participants felt confident in using and teaching others to use these features regardless of their hearing status.
  • Listening Experience: There was variability in satisfaction with the custom features. Hearing-impaired participants were more likely to recommend and continue using the features than TH participants. Hearing-impaired participants rated the features in noisy environments more positively than TH participants.
  • Perceived Benefit: Overall, participants found the custom features more beneficial in quiet than in noise. Hearing-impaired participants reported greater benefit from the features in both quiet and noisy conditions compared to TH participants.
Martinez (2023) [23]None
  • Head simulator
  • AP
  • AP 2
  • iPhone
Noise attenuation of AP and AP 2 in different modes (NC, TP, and Off) under various noise conditions at different levels.
  • NC mode: AP 2 reduced ambient noise by an average of 27 dB across all frequencies, showing a 12 dB improvement in low-frequency noise reduction compared to the AP.
  • TP Mode: The new adaptive TP feature in AP 2 lowered loud sounds over 85 dBA, ensuring safe listening while maintaining awareness of the environment.
Valderrama et al. (2024) [24]
  • N = 17
  • Ages 21–59
TH but self-reported hearing difficulties in noise
  • iPhone
  • AP
  • AP set with custom headphone accommodations, including TP, Conversation Boost, and Ambient Noise Reduction.
  • Real-ear measures.
  • Speech-in-noise intelligibility tests using the BKB sentences
  • Subjective listening effort evaluation
  • Real-world hearing experience survey
  • CB and the noise cancellation feature provided +5.4 dB SNR advantage, which is comparable to about 3–6 dB SNR benefit provided by directional HA microphones.
  • AP significantly improved speech intelligibility by 11.8%, from 54.6% unaided to 66.4%. Subjective listening effort evaluation suggested AP significantly reduced the mental demand by 8% and effort by 8%.
  • Main barriers to using AP as an assistive listening device:
    Limited Hearing Benefit: 31.6% reported issues with unnaturally high-frequency sound amplification, self-hearing of noises through bone conduction, and insufficient background noise attenuation.
    Discomfort: 18.4% mentioned AP was uncomfortable or did not fit properly. Comfort varied significantly among users.
    Stigma and Embarrassment: 15.5% reported they felt uncomfortable wearing AP in social settings, often needing to explain their use.
  • Despite the barriers, around 30% indicated they would continue using AP for better hearing in noise, suggesting its potential for further enhancements.
Kim et al. (2024) [25]
  • N = 35
  • Median = 63 years
Mild to moderate hearing loss
  • iPhone
  • AP
  • Bean (a PSAP)
  • Functional gain
  • Monosyllabic word in quiet
  • Sentence in noise at 0 dB SNR
  • Real-ear: aided accuracy relative to NAL-NL2 targets
  • Functional gain: No significant difference between the Bean and AP, except at 8 kHz, where the Bean performed better.
  • Word and sentence recognition: Both Bean (median = 84% and 95%) and AP (median = 84% and 93%) significantly improved scores compared to unaided (median = 60% and 65%) conditions, with no significant difference between Bean and AP.
  • Real-ear: Both devices amplified appropriately relative to the NAL-NL2 fitting formula (<10 dB SPL between target and response frequencies across all ranges), though AP showed a deviation at 6 kHz in the left ear (12.25 dB SPL).
Foroogozar (2024) [26]
  • N = 23
  • Age ≥ 60
Normal to mild/moderate hearing loss
  • iPhone with LL
  • AP2
  • A 9-digit span test to assess working memory
  • QuickSIN to assess participant’s SNR loss
The improvement in memory retention and recognition accuracy is significant.
  • Mean memory retention: 43.8% without LL and 59.4% with LL.
  • Mean Recognition score: 81.8% without LL and 94.4% with LL.
A positive correlation between SNR loss and recognition improvement (slope = 1.095, R2 = 0.284) showed the potential of LL feature to benefit those with higher SNR loss.
NOTE: AP = AirPods Pro; AP 2 = AirPods Pro second generation; SRT = speech reception threshold; MHINT = Mandarin hearing in noise test; HA = hearing aid; PSAP = personal sound amplification products; NC = noise cancellation; TP = transparency; SNR = signal-to-noise ratio; BKB = Bamford–Kowal–Bench; NAL-NL2 = National Acoustics Laboratories-Non-Linear prescription; 2nd generation; LL = Live Listen; CB = Conversation Boost.
Table 2. Devices used in the study.
Table 2. Devices used in the study.
FunctionNameDevice
Sentence and noise stimuli inputLaptopsThinkPad P70 (Lenovo)
Otter VTTLatitude 7480 (Dell)
Stimuli transmissionAudiometerGrason Stadler GSI 61
SpeakersTwo free-field corner speakers (Grason Stadler GSI)
KEMARKEMAR Dummy-Head
TransmitterSmartphone-based RM systemiPhone 12 (IOS 15.0, Apple)
ReceiverRight ear AP (1st Gen., Apple)
TransmitterPhonak Roger RM systemRoger On
ReceiverRoger Focus II
NOTE: VTT = voice-to-text transcription; KEMAR = Knowles Electronics Manikin for Acoustic Research; RM = remote microphone; AP = AirPods Pro.
Table 3. Hearing in Noise Test list 20.
Table 3. Hearing in Noise Test list 20.
Sentence PlayedScoring (Points)
A/The clown has/had a/the funny face6
The bath water is/was warm5
She injured four of her fingers6
He paid his bill in full6
They stared at a/the picture5
A/The driver started a/the car5
A/The truck carries fresh fruit5
A/The bottle is/was on a/the shelf6
The small tomatoes are/were green5
A/The dinner plate is/was hot5
Total54
NOTE: The symbol “/” is used to indicate that either description is considered a correct recognition and is counted as one point towards the overall score.
Table 4. Mean transcription accuracy at each baseline condition.
Table 4. Mean transcription accuracy at each baseline condition.
TechnologySpeech Noise (dB SNR)Babble Noise (dB SNR)Mean
+50−5+50−5
AP NC93.8379.0146.3093.8378.4045.0672.74
AP TP85.8067.2845.0692.5978.4059.2671.40
KEMAR96.0079.6749.0093.6783.6760.6777.11
Mean91.8875.3246.7993.3680.1655.0073.75
NOTE: The mean transcription accuracy of three baseline testing conditions is presented as combinations of noise type and SNR: babble/speech noise at +5, 0, and −5 dB SNR. AP = AirPods Pro; NC = noise cancellation; TP = transparency; SNR = signal-to-noise ratio; KEMAR = Knowles Electronics Manikin for Acoustic Research; KEMAR means the KEMAR alone condition.
Table 5. Mean transcription accuracy at each remote microphone condition.
Table 5. Mean transcription accuracy at each remote microphone condition.
TechnologySpeech Noise (dB SNR)Babble Noise (dB SNR)Mean
+50−5+50−5
AP NC + LL98.6795.3386.3395.6792.0087.0092.50
AP TP + LL94.3384.3372.0091.6779.0066.0081.22
Roger RM system98.6798.6794.3396.3395.6786.6795.06
Mean97.2292.7884.2294.5688.8979.8989.59
NOTE: The mean transcription accuracy of three RM testing conditions is presented as combinations of noise type and SNR: babble/speech noise at +5, 0, and −5 dB SNR. AP = AirPods Pro; LL = Live Listen; NC = noise cancellation; TP = transparency; SNR = signal-to-noise ratio.
Table 6. Post hoc analysis of pair-wise comparisons of remote microphone conditions under different signal-to-noise ratios.
Table 6. Post hoc analysis of pair-wise comparisons of remote microphone conditions under different signal-to-noise ratios.
SNR (dB)+50−5
Roger RM system vs. AP NC+LL0.131.341.46
Roger RM system vs. AP TP+LL1.725.92 ***8.21 ***
AP NC+LL vs. AP TP+LL1.594.58 **6.75 ***
NOTE: t value was indicated. Strength of Tukey’s adjusted p-values: ** padj < 0.01, *** padj < 0.001. Same abbreviations as Table 5.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qi, S.; Thibodeau, L. Using Voice-to-Text Transcription to Examine Outcomes of AirPods Pro Receivers When Used as Part of Remote Microphone System. Appl. Sci. 2025, 15, 5451. https://doi.org/10.3390/app15105451

AMA Style

Qi S, Thibodeau L. Using Voice-to-Text Transcription to Examine Outcomes of AirPods Pro Receivers When Used as Part of Remote Microphone System. Applied Sciences. 2025; 15(10):5451. https://doi.org/10.3390/app15105451

Chicago/Turabian Style

Qi, Shuang, and Linda Thibodeau. 2025. "Using Voice-to-Text Transcription to Examine Outcomes of AirPods Pro Receivers When Used as Part of Remote Microphone System" Applied Sciences 15, no. 10: 5451. https://doi.org/10.3390/app15105451

APA Style

Qi, S., & Thibodeau, L. (2025). Using Voice-to-Text Transcription to Examine Outcomes of AirPods Pro Receivers When Used as Part of Remote Microphone System. Applied Sciences, 15(10), 5451. https://doi.org/10.3390/app15105451

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop