Next Article in Journal
PKD-YOLOv8: A Collaborative Pruning and Knowledge Distillation Framework for Lightweight Rapeseed Pest Detection
Previous Article in Journal
HC-SPA: Hyperbolic Cosine-Based Symplectic Phase Alignment for Fusion Optimization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimizing MFCC Parameters for Breathing Phase Detection

by
Assel K. Zhantleuova
1,
Yerbulat K. Makashev
2 and
Nurzhan T. Duzbayev
1,*
1
Department of Computer Engineering, International Information Technology University, Almaty 050040, Kazakhstan
2
Institute of Genetics and Physiology, Almaty 050060, Kazakhstan
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(16), 5002; https://doi.org/10.3390/s25165002
Submission received: 3 July 2025 / Revised: 28 July 2025 / Accepted: 8 August 2025 / Published: 13 August 2025
(This article belongs to the Section Biomedical Sensors)

Abstract

Breathing phase detection is fundamental for various clinical and digital health applications, yet standard Mel Frequency Cepstral Coefficients (MFCCs) settings often limit classification performance. This study systematically optimized MFCC parameters, specifically the number of coefficients, frame length, and hop length, using a proprietary dataset of respiratory sounds (n = 1500 segments). Classification performance was evaluated using Support Vector Machines (SVMs) and benchmarked against deep learning models (VGGish, YAMNet, MobileNetV2). Optimal parameters (30 MFCC coefficients, 800 ms frame length, 10 ms hop length) substantially enhanced accuracy (87.16%) compared to default settings (80.96%) and performed equivalently or better than deep learning methods. A trade-off analysis indicated that a clinically practical frame length of 200–300 ms balanced accuracy (85.08%) and latency effectively. The study concludes that optimized MFCC parameters significantly improve respiratory phase classification, providing efficient and interpretable solutions suitable for real-time clinical monitoring. Future research should focus on validating these parameters in broader clinical contexts and exploring multimodal and federated learning strategies.

1. Introduction

Breathing is a fundamental physiological process governed by complex neural circuits and homeostatic mechanisms, playing a central role in human health, cognition, and behavior [1,2,3,4]. Recent findings indicate that respiratory patterns, including phase transitions, are individually unique, stable over time, and predictive of cognitive and physiological states [5,6]. Automated detection and characterization of breathing phases—specifically, inhalation, exhalation, and pauses—are essential for objective assessment in respiratory medicine, digital therapeutics, stress management, and biofeedback applications [7,8,9,10,11]. The precise detection of breathing phases holds significant clinical relevance in medical diagnostics and health monitoring, enabling the identification of abnormalities associated with respiratory diseases such as asthma, obstructive sleep apnea (OSA) and chronic obstructive pulmonary disease (COPD) [12,13]. Recent statistics underscore this importance: OSA affects an estimated 38% of adults worldwide, while COPD and asthma exhibit a prevalence of approximately 10–11% [14,15,16]. Beyond diagnostics, monitoring breathing phases supports the control of physical load in sports, where respiratory patterns influence performance, focus, and recovery during exercise [17,18]. Furthermore, in psychophysiology, breathing patterns are intricately linked to emotional states, facilitating transitions in psychophysiological flexibility and contributing to mood regulation through practices such as structured respiration [19,20]. In this context, accurate and automated phase detection forms the basis for a range of clinical and digital health applications.
Traditional techniques, including spirometry, pneumography, plethysmography, capnography, accelerometry, and electromyography, despite their utility, present notable limitations, including invasiveness (e.g., requiring mouthpieces or electrodes), elevated costs for equipment and maintenance, and impracticality for long-term monitoring due to patient discomfort and restricted mobility [13]. To address these drawbacks, recent advancements have shifted toward non-invasive acoustic methods, which analyze sounds captured via microphones to identify breathing phases in real-time, offering greater accessibility and comfort for continuous health surveillance [21,22].
Mel Frequency Cepstral Coefficients (MFCCs) have long been recognized as the gold standard for perceptually relevant spectral analysis in speech and biomedical sound processing [23,24]. However, most studies employ default or empirical MFCC parameter settings, potentially limiting classification performance and model generalizability. Recent work has demonstrated that optimizing the selection of MFCC parameters, including the number of coefficients, frame length, and hop length, has been shown to be critical for improving the performance of respiratory disease classification systems [25].
While prior studies have primarily targeted disease detection using default settings, we focus on optimizing MFCC parameters for accurate real-time classification of breathing phases. Furthermore, we benchmark the optimized MFCC configuration against multiple state-of-the-art deep learning models (VGGish, YAMNet, MobileNetV2) to assess the robustness and generalizability of the proposed settings. By doing so, we demonstrate that an appropriately optimized classical machine learning pipeline can achieve performance comparable to deep learning models while offering greater interpretability and lower computational cost—an important consideration for clinical and mobile health applications.
This paper is organized as follows. Section 2 reviews related work on breathing phase detection and MFCC applications in audio signal processing, providing a foundation for the proposed approach. Section 3 details the materials and methods, including dataset description, MFCC parameter configurations, and the machine learning framework employed. Section 4 presents experimental results, evaluating the performance of optimized MFCC parameters in detecting breathing phases. Section 5 discusses the findings, their implications, limitations, and future research directions. Section 6 concludes the paper by summarizing the key contributions.

2. Related Work

The concept of MFCC was introduced by Davis and Mermelstein in the late 20th century as a compact and robust feature extraction method for automatic speech recognition. MFCC combines the mel scale, designed to approximate human auditory perception, with cepstral coefficients to represent the short-term power spectrum of audio signals [23]. In biomedical applications, particularly in respiratory sound analysis, MFCC gained attention starting from the mid-2000s. They have been widely used to extract features from lung and tracheal sounds for detecting adventitious events such as wheezes, crackles, and rhonchi, as well as for classifying respiratory pathologies [26,27]. Since the 2010s, MFCC has been integrated with machine learning and deep learning frameworks to enhance accuracy in tasks like breathing phase detection and disease diagnosis from audio recordings, addressing challenges such as noise robustness and signal variability [28,29,30].
Various approaches have been proposed to optimize MFCC parameters for enhanced performance in respiratory sound analysis. Bardou et al. [31] employed convolutional neural networks for lung sound classification, integrating MFCC-like features to achieve high accuracy in distinguishing normal and abnormal respiratory events, highlighting synergy with deep learning in medical diagnostics. Dash et al. [32] proposed a bio-inspired cepstral method to detect COVID-19 from speech. It showed adaptability in spotting respiratory impairments via pattern recognition. Ingco et al. [33,34] advanced an enhanced MFCC framework for respiratory sound analysis, optimizing parameters to capture breathing frequency characteristics and evaluating an intelligent classifier that improved pathological sound detection with machine learning in noisy environments.
In the context of breathing phases, audio-based methods employing MFCC have been adapted for classifying pause, inhalation, and exhalation periods but often rely on default reference features [8,35,36,37]. While research by Tirronen et al. [38] explored the role of frame length in identifying voice pathologies, it did not extend to evaluating hop length or the quantity of coefficients. In a related advancement, Yan et al. [25] investigated the optimization of MFCC parameters for the automated identification of respiratory diseases, achieving notable enhancements in classification accuracy through adjustments to frame length, hop length, and coefficient count. However, their work focused on pathology detection rather than explicit breathing phase classification.

3. Materials and Methods

3.1. Dataset Description

A proprietary respiratory sound dataset was collected specifically for this study, comprising 1500 manually annotated audio segments evenly distributed among three classes: inhalation, exhalation, and pause (500 samples per class). Recordings were obtained from five healthy adult participants (three males, two females; age range: 27–35 years) in a controlled indoor environment to minimize external noise interference. Audio was captured using two consumer-grade devices: the built-in microphone of a laptop and a mobile device, positioned on a table opposite the participants at a distance of approximately 20 cm from the mouth to ensure consistent signal capture. Participants were instructed to maintain a seated position in a relaxed posture facing the table, promoting uniformity and reducing movement artifacts during recording sessions. All recordings were saved as uncompressed WAV files (mono, 16-bit PCM, 44.1 kHz sampling rate), trimmed, and manually annotated using Audacity software (version 3.7.4) to ensure consistent quality and facilitate subsequent analysis.
Annotations were performed based on acoustic waveforms and spectrograms, delineating phases by amplitude thresholds. The annotation process entailed two independent researchers classifying breathing phases as “inhalation” and “exhalation” using acoustic waveform and spectrogram analysis with amplitude-based phase boundaries. A third researcher evaluated these labels, resolving discrepancies via consensus to maintain precision across all segments. A “pause” category was added for unclassified intervals, representing airflow cessation critical to the natural breathing cycle and marked by low acoustic activity [39], thereby improving the depiction of the full respiratory process.
Normal breath sounds from the dataset, as illustrated in Figure 1, consist of two primary phases—inhalation and exhalation—separated by short pauses. The time-domain waveform displays distinct amplitude bursts corresponding to these phases, while the frequency-domain analysis (FFT) shows a dominant peak at 300 Hz, with additional components at approximately 50 Hz and 100 Hz. In the spectrogram, both phases are clearly distinguishable, with energy concentrations around 90 Hz, 400 Hz, and 1600 Hz persisting throughout the signal, which reflects typical spectral characteristics of healthy respiratory patterns.

3.2. Pre-Processing

Prior to feature extraction, all audio recordings were resampled to 16 kHz, converted to mono, and normalized to zero mean and unit variance. Segments shorter than the designated frame length were zero-padded to ensure uniform input dimensions. Labels were encoded using one-hot or integer encoding as required for specific classification models. No additional filtering or denoising was applied in this controlled study to preserve the original signal characteristics and evaluate baseline performance. For field deployment, where environmental noise from ambient sources, such as traffic or conversations, may contaminate respiratory sound recordings, implementing denoising strategies is essential to enhance signal quality and model robustness.

3.3. Mel Frequency Cepstral Coefficients Feature Extraction

MFCCs are a standard feature set for audio signal analysis, widely used in speech processing and biomedical applications due to their effectiveness in capturing perceptually relevant spectral characteristics of sound.
The extraction process consists of several stages: segmentation of the raw audio signal into short, overlapping frames; application of a windowing function (Hamming window) to reduce spectral leakage; transformation into the frequency domain using the FFT; mapping of the power spectrum onto the mel scale via a filterbank; logarithmic compression; and finally, decorrelation and dimensionality reduction via the discrete cosine transform [24,40,41].
The MFCCs for each frame are computed as shown in Equation (1):
C m = j = 0 J 1 cos ( m π J ( j + 0.5 ) ) log 10 E j ,           0 m L 1
where E j is the energy of the j-th mel filterbank, J is the total number of mel bands, L is the number of cepstral coefficients, and m is the index of the coefficient [25].
MFCCs were selected as the primary feature representation due to their proven effectiveness in respiratory and biomedical audio analysis [25]. Feature extraction followed a standard pipeline as shown in Figure 2.
A stepwise parameter optimization strategy was employed. First, the number of coefficients varied within the range 13–80 while other parameters were kept at default values. The optimal value was fixed and used as a baseline for the subsequent tuning of frame length (25–800 ms) and finally hop length (5–500 ms). The MFCC feature vectors were obtained by calculating the mean of the frame-wise MFCC coefficients across the entire segment. The extraction of MFCC features from each audio recording was performed using the Librosa library [42].

3.4. SVM Classifier

A Support Vector Machine (SVM) classifier was implemented using the Scikit-learn library to evaluate the discriminative power of the optimized MFCC features [43]. SVM is a widely used and versatile algorithm known for its ability to construct optimal separating hyperplanes in high-dimensional feature spaces [44]. The RBF kernel was employed to allow for nonlinear separation in the feature space. The regularization parameter C = 1.0 was chosen to balance training error minimization with margin maximization, while the gamma parameter was set to “scale”, automatically adapting to the variance of the input features and controlling the influence of individual support vectors. MFCC feature vectors, computed as the mean of frame-level coefficients, were standardized to zero mean and unit variance within each fold.
Model performance was assessed using 10-fold stratified cross-validation, implemented via the StratifiedKFold function from the Scikit-learn library, with a fixed random seed of 42 to ensure reproducibility [43]. This approach divided the dataset into ten folds, with nine folds used for training and one-fold used for testing in each iteration. The process was repeated ten times, ensuring that each fold served as the test set exactly once. Stratification preserved the proportion of classes across all folds, thereby preventing potential class imbalance bias. The results from all folds were aggregated, and mean values with standard deviations were reported for accuracy, AUC, F1-score, precision, and equal error rate (EER).

4. Results

We systematically investigated the influence of three critical MFCC extraction parameters—number of coefficients, frame length, and hop length—on the accuracy of respiratory phase classification using an SVM classifier. In addition to traditional feature-based approaches, we benchmarked the optimized MFCC configuration against a range of deep learning and pretrained audio embedding methods, including VGGish, YAMNet, and MobileNetV2, to comprehensively assess model performance on the same dataset.

4.1. Number of Coefficients

The number of MFCC coefficients determines the amount of spectral information retained from the original signal, with higher values theoretically capturing more detail but potentially introducing redundancy. In this study, we systematically varied the number of coefficients from 13 to 80, where 13 corresponds to the minimal widely adopted MFCC set and 80 represents an extended, information-rich configuration. The frame length and hop length were both fixed at default values (25 ms and 10 ms, respectively) to isolate the effect of this parameter.
As shown in Figure 3, the classification accuracy remained stable for n_mfcc between 13 and 30 (accuracy ≈ 0.8104 ± 0.0396–0.0371), with a slight improvement observed at 40 and higher values (accuracy up to 0.8173 ± 0.0358 for n_mfcc = 50 and above). However, the relative gain in accuracy beyond 30 coefficients was marginal, suggesting that the optimal trade-off between informative content and computational complexity is achieved at or slightly above 30 coefficients. Therefore, n_mfcc = 30 was selected as the optimized setting for subsequent experiments.

4.2. Frame Length

The extraction of MFCC features begins by segmenting the original time-domain signal into short frames, where the duration of each frame—termed the frame length—plays a critical role in capturing the temporal dynamics of the signal. This step is essential because the spectral content of respiratory sounds evolves over time, and a Fourier transform performed over the entire recording would obscure these changes.
In this study, we systematically varied the frame length from 25 ms to 800 ms, while keeping the number of MFCC coefficients and the hop length fixed at their optimized or default values (n_mfcc = 30, hop length = 10 ms). As presented in Figure 4, the classification accuracy increased steadily with longer frame lengths: from 0.8104 ± 0.0371 at 25 ms up to 0.8716 ± 0.0397 at 800 ms. Notably, the largest gains were observed when increasing the frame length above 100 ms, suggesting that longer windows allow for the aggregation of more robust time-frequency information relevant to respiratory phase classification. Based on these results, a frame length of 800 ms was selected as optimal for subsequent experiments.

4.3. Latency-Accuracy Trade-Off

Increasing the frame length improves classification accuracy but also introduces proportional latency due to the longer analysis windows, which is critical for real-time respiratory monitoring (Table 1). In respiratory sound analysis, frame lengths of 200–300 ms are often considered optimal, offering a balanced compromise between responsiveness and analytical robustness [9]. A 300 ms window was therefore selected as a practical configuration, ensuring high accuracy with acceptable delay. Additionally, hop length impacts analysis speed: smaller hop values lead to denser frame overlaps and higher computational load per second of audio.

4.4. Hop Length

The hop length parameter specifies the number of samples shifted between consecutive frames during MFCC extraction. Shorter hop lengths result in greater frame overlap, which reduces spectral leakage and allows for a more continuous and robust representation of the signal. In this study, hop length was varied from 5 ms to 500 ms, with n_mfcc fixed at 30.
The influence of hop length on classification accuracy was assessed for both the optimized frame length (800 ms) and a clinically motivated setting (300 ms). As illustrated in Figure 5, shorter hop lengths (5–20 ms) consistently produced the highest and most stable accuracy, regardless of window duration. Accuracy declined gradually as hop length increased beyond 100 ms, with more pronounced decreases observed for hop lengths above 200 ms. These results confirm that substantial frame overlap is beneficial for robust respiratory phase classification across both optimal and low-latency scenarios. A hop length of 10 ms was therefore selected for all further experiments.

4.5. Optimal Combination of Parameters

A systematic grid search across MFCC extraction parameters was performed using nested loops, evaluating every possible combination. This exhaustive evaluation revealed that the combination of 30 coefficients, an 800 ms frame length, and a 10 ms hop length yields optimal performance for respiratory phase classification using an SVM classifier. This configuration resulted in the highest accuracy (0.8716 ± 0.0397) and consistent improvements across all other metrics, including AUC, F1-score, precision, and EER, as summarized in Table 2. The default configuration employs the standard settings commonly used in audio processing literature: 13 MFCC coefficients (the minimal widely adopted set), a frame length of 25 ms, and a hop length of 10 ms. In contrast, the worst configuration represents the poorest-performing combination encountered in the parameter sweeps, consisting of 13 MFCC coefficients, a frame length of 25 ms, and a hop length of 500 ms.
The advantage of parameter optimization is also depicted in Figure 6, which demonstrates a marked increase in classification accuracy for the optimized configuration compared to both the default and worst-case settings. These findings underscore the importance of data-driven MFCC tuning for automatic analysis of respiratory sounds.

4.6. Comparison with Deep Learning Models

Pretrained VGGish [45] and YAMNet [46] models, recognized as state-of-the-art audio feature extractors, were used to generate fixed-length embeddings (128 dimensions for VGGish and 1024 for YAMNet). VGGish is a 24-layer convolutional neural network built on the VGGNet architecture [47], while YAMNet employs a MobileNet-v1 backbone and is optimized for audio event classification in the AudioSet corpus [48]. Both models were used in frozen mode without fine-tuning, and the averaged embeddings were classified using a Random Forest (RF). MobileNetV2, pretrained on ImageNet, was evaluated as a feature extractor for both MFCC and log-mel spectrogram representations [49]. The frozen backbone of MobileNetV2 produced 1280-dimensional feature vectors that were subsequently classified by an RF.
As summarized in Table 3, the SVM trained on optimized MFCC features achieved the highest overall performance. Among deep learning approaches, VGGish + RF performed best, while YAMNet + RF showed the lowest results. Although deep learning models based on pretrained embeddings are effective for general audio classification, classical models such as SVM remain advantageous due to their computational efficiency and interpretability, which are particularly important in clinical and resource-limited settings.

5. Discussion

This study demonstrates that systematic optimization of MFCC extraction parameters markedly enhances the accuracy of automated respiratory phase classification. The optimal parameter set (30 coefficients, 800 ms frame length, 10 ms hop length) consistently outperformed default and suboptimal settings across all evaluation metrics. When compared to deep learning and embedding-based models (VGGish, YAMNet, MobileNetV2), the optimized SVM model achieved equivalent or superior results with significantly lower computational requirements and greater interpretability. This underscores the ongoing relevance of classical machine learning for biomedical audio tasks, particularly where data is limited and transparency is essential. A key translational insight is the trade-off between accuracy and latency: longer frame lengths maximize accuracy but increase response time, potentially limiting real-time clinical usability. Our results indicate that 200–300 ms frames preserve clinically meaningful accuracy (≥0.8462) while keeping latency well below accepted thresholds in respiratory monitoring applications.
Compared to prior research, our findings emphasize the critical role of MFCC parameter tuning for breathing phase detection. Previous studies summarized in Table 4 primarily focused on other problems and modalities. For example, Tirronen et al. [38] applied default MFCC settings for voice pathology detection using speech data from the SVD dataset, varying only the frame length to assess its impact, while keeping the number of coefficients and hop length fixed. In contrast, Yan et al. [25] and our study systematically explored all three key MFCC parameters across multiple datasets and tasks. Importantly, unlike speech or cough, breath sounds exhibit longer, lower-frequency structures and slower temporal dynamics, particularly in resting or paced breathing.
The current study is subject to several limitations that warrant careful consideration for interpreting the results and guiding future investigations. The proprietary dataset includes recordings of five healthy adult participants in a controlled indoor environment, limiting generalizability. Moreover, the dataset lacks diversity in pathological conditions (e.g., asthma, pneumonia, COPD), demographics (e.g., pediatric or elderly populations), and recording devices. Although both a laptop and a smartphone were used, broader variation across device types was not explored. The absence of environmental noise, such as background sounds typical in real-world settings, may overestimate the robustness of the optimized MFCC parameters in practical applications. Although no denoising strategies were employed in the present study, their implementation is crucial for field deployment, where ambient noise from sources like traffic or conversations could compromise respiratory sound recordings, thereby improving signal integrity and model resilience. Traditional approaches, including spectral subtraction and Wiener filtering, are well-known but often involve a trade-off between noise suppression and the preservation of diagnostically relevant high-frequency components, such as subtle crackles or wheezes [50,51,52,53]. Similarly, Butterworth band-pass filtering, while simple and computationally efficient, imposes strict frequency cutoffs (typically 100–2000 Hz) and risks attenuating clinically meaningful information above this range [54,55]. Wavelet-based denoising, despite its adaptability to non-stationary signals, can lead to partial loss of transient features due to its reliance on fixed basis functions [56,57]. More advanced strategies, such as adaptive filtering, have shown promise in preserving the acoustic signatures while reducing background noise overhead [58]. However, these methods require careful tuning and validation to avoid over-smoothing or artifact introduction. Future studies should explore the integration of hybrid or machine-learning-based denoising pipelines, tailored to the spectral and temporal characteristics of respiratory sounds, to ensure robust performance in real-world noisy environments. These omissions may reduce model performance in noisy environments or heterogeneous patient cohorts, where atypical breathing patterns and signal artifacts pose additional challenges.
To address these limitations, future work should prioritize external validation using multi-center studies and publicly available respiratory sound datasets with broader variability. The ICBHI 2017 [59] database includes recordings from patients with various pulmonary pathologies, Coswara [60] offers noisy and ambulatory samples from individuals with COVID-19, and DeepBreath [61] provides pediatric data collected across multiple sites. These resources enable testing across clinical, demographic, and acoustic conditions. In addition, combining audio with physiological signals and applying denoising techniques (e.g., spectral subtraction or bandpass filtering) may further improve robustness in real-world deployments.

6. Conclusions

This study advances the understanding of respiratory sound analysis by demonstrating that MFCC parameter optimization can significantly improve phase detection while retaining model interpretability. Our findings challenge the common reliance on deep learning by showing that classical models, when combined with optimized features, remain competitive. Beyond technical performance, these insights have practical implications for developing lightweight, real-time respiratory monitoring solutions applicable in mobile health, telemedicine, and digital therapeutics. Future work should extend these findings to larger datasets, noisy real-world environments, and hybrid feature-learning frameworks.

Author Contributions

Conceptualization, N.T.D.; methodology, A.K.Z. and Y.K.M.; formal analysis, A.K.Z.; data curation, A.K.Z. and Y.K.M.; writing—original draft preparation, A.K.Z.; writing—review and editing, N.T.D. and Y.K.M.; visualization, A.K.Z.; supervision, N.T.D.; project administration, N.T.D.; funding acquisition, N.T.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded by the Science Committee of the Ministry of Science and Higher Education of the Republic of Kazakhstan (Grant No. AP19680049).

Institutional Review Board Statement

The study was approved by the Ethics Committee of the Institute of Genetics and Physiology (protocol code 6 and date of approval 03.11.2022).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The datasets for this article are not publicly available due to concerns regarding participant anonymity. Requests to access the dataset should be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AUCArea Under the Curve
COPDChronic Obstructive Pulmonary Disease
COVID-19Coronavirus Disease 2019
EEREqual Error Rate
FFTFast Fourier Transform
LSTMLong Short-Term Memory
MFCCsMel Frequency Cepstral Coefficients
OSAObstructive Sleep Apnea
PCMPulse Code Modulation
RBFRadial Basis Function
RFRandom Forest
ROCReceiver Operating Characteristic
STDStandard Deviation
SVDSaarbrücken Voice Disorders
SVMSupport Vector Machine
WAVWaveform audio file format

References

  1. Del Negro, C.A.; Funk, G.D.; Feldman, J.L. Breathing matters. Nat. Rev. Neurosci. 2018, 19, 351–367. [Google Scholar] [CrossRef] [PubMed]
  2. Dempsey, J.A.; Welch, J.F. Control of breathing. Semin. Respir. Crit. Care Med. 2023, 44, 627–649. [Google Scholar] [CrossRef] [PubMed]
  3. Ashhad, S.; Kam, K.; Del Negro, C.A.; Feldman, J.L. Breathing rhythm and pattern and their influence on emotion. Annu. Rev. Neurosci. 2022, 45, 223–247. [Google Scholar] [CrossRef]
  4. Mitsea, E.; Drigas, A.; Skianis, C. Breathing, attention & consciousness in sync: The role of breathing training, metacognition & virtual reality. Tech. Soc. Sci. J. 2022, 29, 79–97. [Google Scholar] [CrossRef]
  5. Soroka, T.; Ravia, A.; Snitz, K.; Honigstein, D.; Weissbrod, A.; Gorodisky, L.; Weiss, T.; Perl, O.; Sobel, N. Humans have nasal respiratory fingerprints. Curr. Biol. 2025, 35, 3011–3021.e3. [Google Scholar] [CrossRef] [PubMed]
  6. Zaccaro, A.; Piarulli, A.; Laurino, M.; Garbella, E.; Menicucci, D.; Neri, B.; Gemignani, A. How breath-control can change your life: A systematic review on psycho-physiological correlates of slow breathing. Front. Hum. Neurosci. 2018, 12, 409421. [Google Scholar] [CrossRef]
  7. Landry, V.; Matschek, J.; Pang, R.; Munipalle, M.; Tan, K.; Boruff, J.; Li-Jessen, N.Y. Audio-based digital biomarkers in diagnosing and managing respiratory diseases: A systematic review and bibliometric analysis. Eur. Respir. Rev. 2025, 34, 240246. [Google Scholar] [CrossRef]
  8. Jeong, H.; Yoo, J.H.; Goh, M.; Song, H. Deep breathing in your hands: Designing and assessing a DTx mobile app. Front. Digit. Health 2024, 6, 1287340. [Google Scholar] [CrossRef]
  9. Shih, C.H.; Tomita, N.; Lukic, Y.X.; Reguera, Á.H.; Fleisch, E.; Kowatsch, T. Breeze: Smartphone-based acoustic real-time detection of breathing phases for a gamified biofeedback breathing training. Proc. ACM Interact. Mob. 2019, 3, 1–30. [Google Scholar] [CrossRef]
  10. Agrawal, V.; Naik, V.; Duggirala, M.; Athavale, S. Calm a mobile based deep breathing game with biofeedback. In Proceedings of the Extended Abstracts of the 2020 Annual Symposium on Computer-Human Interaction in Play, Virtual Event Canada, 2–4 November 2020. [Google Scholar] [CrossRef]
  11. Latifi, S.A.; Ghassemian, H.; Imani, M. Feature extraction and classification of respiratory sound and lung diseases. In Proceedings of the 2023 6th International Conference on Pattern Recognition and Image Analysis (IPRIA), Qom, Iran, 14–16 February 2023. [Google Scholar] [CrossRef]
  12. Lurie, A.; Roche, N. Obstructive sleep apnea in patients with chronic obstructive pulmonary disease: Facts and perspectives. COPD J. Chronic Obs. Pulm. Dis. 2021, 18, 700–712. [Google Scholar] [CrossRef]
  13. Vitazkova, D.; Foltan, E.; Kosnacova, H.; Micjan, M.; Donoval, M.; Kuzma, A.; Kopani, M.; Vavrinsky, E. Advances in respiratory monitoring: A comprehensive review of wearable and remote technologies. Biosensors 2024, 14, 90. [Google Scholar] [CrossRef]
  14. Iannella, G.; Pace, A.; Bellizzi, M.G.; Magliulo, G.; Greco, A.; De Virgilio, A.; Croce, E.; Gioacchini, F.M.; Re, M.; Costantino, A.; et al. The Global Burden of Obstructive Sleep Apnea. Diagnostics 2025, 15, 1088. [Google Scholar] [CrossRef] [PubMed]
  15. Adeloye, D.; Song, P.; Zhu, Y.; Campbell, H.; Sheikh, A.; Rudan, I. Global, regional, and national prevalence of, and risk factors for, chronic obstructive pulmonary disease (COPD) in 2019: A systematic review and modelling analysis. Lancet Respir. Med. 2022, 10, 447–458. [Google Scholar] [CrossRef]
  16. Song, P.; Adeloye, D.; Salim, H.; Dos Santos, J.P.; Campbell, H.; Sheikh, A.; Rudan, I. Global, regional, and national prevalence of asthma in 2019: A systematic analysis and modelling study. J. Glob. Health 2022, 12, 04052. [Google Scholar] [CrossRef] [PubMed]
  17. Migliaccio, G.M.; Russo, L.; Maric, M.; Padulo, J. Sports performance and breathing rate: What is the connection? A narrative review on breathing strategies. Sports 2023, 11, 103. [Google Scholar] [CrossRef]
  18. Sikora, M.; Mikołajczyk, R.; Łakomy, O.; Karpiński, J.; Żebrowska, A.; Kostorz-Nosal, S.; Jastrzębski, D. Influence of the breathing pattern on the pulmonary function of endurance-trained athletes. Sci. Rep. 2024, 14, 1113. [Google Scholar] [CrossRef]
  19. Balban, M.Y.; Neri, E.; Kogon, M.M.; Weed, L.; Nouriani, B.; Jo, B.; Holl, G.; Zeitzer, J.M.; Spiegel, D.; Huberman, A.D. Brief structured respiration practices enhance mood and reduce physiological arousal. Cell Rep. Med. 2023, 4, 100895. [Google Scholar] [CrossRef] [PubMed]
  20. Toussaint, L.; Nguyen, Q.A.; Roettger, C.; Dixon, K.; Offenbächer, M.; Kohls, N.; Hirsch, J.; Sirois, F. Effectiveness of progressive muscle relaxation, deep breathing, and guided imagery in promoting psychological and physiological states of relaxation. Evid.-Based Complement. Altern. Med. 2021, 2021, 5924040. [Google Scholar] [CrossRef] [PubMed]
  21. Cook, J.; Umar, M.; Khalili, F.; Taebi, A. Body acoustics for the non-invasive diagnosis of medical conditions. Bioengineering 2022, 9, 149. [Google Scholar] [CrossRef] [PubMed]
  22. Lalouani, W.; Younis, M.; Emokpae Jr, R.N.; Emokpae, L.E. Enabling effective breathing sound analysis for automated diagnosis of lung diseases. Smart Health 2022, 26, 100329. [Google Scholar] [CrossRef]
  23. Davis, S.; Mermelstein, P. Comparison of parametric representations for monosyllabic word recognition in continuously spoken sentences. IEEE Trans. Acoust. 1980, 28, 357–366. [Google Scholar] [CrossRef]
  24. Abdul, Z.K.; Al-Talabani, A.K. Mel frequency cepstral coefficient and its applications: A review. IEEE Access 2022, 10, 122136–122158. [Google Scholar] [CrossRef]
  25. Yan, Y.; Simons, S.O.; van Bemmel, L.; Reinders, L.G.; Franssen, F.M.; Urovi, V. Optimizing MFCC parameters for the automatic detection of respiratory diseases. Appl. Acoust. 2025, 228, 110299. [Google Scholar] [CrossRef]
  26. Ruinskiy, D.; Lavner, Y. An effective algorithm for automatic detection and exact demarcation of breath sounds in speech and song signals. IEEE Trans. Audio Speech Lang. Process 2007, 15, 838–850. [Google Scholar] [CrossRef]
  27. Duckitt, W.D.; Tuomi, S.K.; Niesler, T.R. Automatic detection, segmentation and assessment of snoring from ambient acoustic data. Physiol. Meas. 2006, 27, 1047–1056. [Google Scholar] [CrossRef]
  28. Kim, T.; Kim, J.W.; Lee, K. Detection of sleep disordered breathing severity using acoustic biomarker and machine learning techniques. Biomed. Eng. Online 2018, 17, 16. [Google Scholar] [CrossRef]
  29. Srivastava, A.; Jain, S.; Miranda, R.; Patil, S.; Pandya, S.; Kotecha, K. Deep learning based respiratory sound analysis for detection of chronic obstructive pulmonary disease. PeerJ Comput. Sci. 2021, 7, e369. [Google Scholar] [CrossRef] [PubMed]
  30. Pahar, M.; Niesler, T. Machine learning based COVID-19 detection from smartphone recordings: Cough, breath and speech. arXiv 2021, arXiv:2104.02477. [Google Scholar]
  31. Bardou, D.; Zhang, K.; Ahmad, S.M. Lung sounds classification using convolutional neural networks. Artif. Intell. Med. 2018, 88, 58–69. [Google Scholar] [CrossRef]
  32. Dash, T.K.; Mishra, S.; Panda, G.; Satapathy, S.C. Detection of COVID-19 from speech signal using bio-inspired based cepstral features. Pattern Recognit. 2021, 117, 107999. [Google Scholar] [CrossRef]
  33. Ingco, W.E.M.; Reyes, R.S.; Abu, P.A.R. Development of a spectral feature extraction using enhanced MFCC for respiratory sound analysis. In Proceedings of the 2019 International SoC Design Conference (ISOCC), Jeju, Republic of Korea, 6–9 October 2019. [Google Scholar] [CrossRef]
  34. Ingco, W.E.M.; Abu, P.A.R.; Reyes, R.S. Performance evaluation of an intelligent lung sound classifier based on an enhanced MFCC model. In Proceedings of the 2021 7th International Conference on Electrical, Electronics and Information Engineering (ICEEIE), Malang, Indonesia, 2 October 2021. [Google Scholar] [CrossRef]
  35. Fahed, V.S.; Doheny, E.P.; Lowery, M.M. Random forest classification of breathing phases from audio signals recorded using mobile devices. In Proceedings of the Interspeech 2023, Dublin, Ireland, 20–24 August 2023. [Google Scholar] [CrossRef]
  36. Mehrban, M.H.K.; Voix, J.; Bouserhal, R.E. Classification of breathing phase and path with in-ear microphones. Sensors 2024, 24, 6679. [Google Scholar] [CrossRef] [PubMed]
  37. Tran-Anh, D.; Vu, N.H.; Nguyen-Trong, K.; Pham, C. Multi-task learning neural networks for breath sound detection and classification in pervasive healthcare. Pervasive Mob. Comput. 2022, 86, 101685. [Google Scholar] [CrossRef]
  38. Tirronen, S.; Kadiri, S.R.; Alku, P. The effect of the MFCC frame length in automatic voice pathology detection. J. Voice 2024, 38, 975–982. [Google Scholar] [CrossRef] [PubMed]
  39. Laborde, S.; Iskra, M.; Zammit, N.; Borges, U.; You, M.; Sevoz-Couche, C.; Dosseville, F. Slow-paced breathing: Influence of inhalation/exhalation ratio and of respiratory pauses on cardiac vagal activity. Sustainability 2021, 13, 7775. [Google Scholar] [CrossRef]
  40. Gupta, S.; Jaafar, J.; Ahmad, W.W.; Bansal, A. Feature extraction using MFCC. Signal Image Process Int. J. 2013, 4, 101–108. [Google Scholar] [CrossRef]
  41. Siam, A.I.; Elazm, A.A.; El-Bahnasawy, N.A.; El Banby, G.M.; Abd El-Samie, F.E. PPG-based human identification using Mel-frequency cepstral coefficients and neural networks. Multimed. Tools Appl. 2021, 80, 26001–26019. [Google Scholar] [CrossRef]
  42. McFee, B.; Raffel, C.; Liang, D.; Ellis, D.P.W.; McVicar, M.; Battenberg, E.; Nieto, O. librosa: Audio and music signal analysis in python. In Proceedings of the 14th Python in Science Conference, Austin, TX, USA, 11–17 July 2015. [Google Scholar] [CrossRef]
  43. Kramer, O. Scikit-learn. In Machine Learning for Evolution Strategies; Springer: Berlin/Heidelberg, Germany, 2016; pp. 45–53. [Google Scholar] [CrossRef]
  44. Cervantes, J.; Garcia-Lamont, F.; Rodríguez-Mazahua, L.; Lopez, A. A comprehensive survey on support vector machine classification: Applications, challenges and trends. Neurocomputing 2020, 408, 189–215. [Google Scholar] [CrossRef]
  45. Hershey, S.; Chaudhuri, S.; Ellis, D.P.W.; Gemmeke, J.F.; Jansen, A.; Moore, R.C.; Plakal, M.; Channing Moore, R.; Platt, D.; Saurous, R.A.; et al. CNN architectures for large-scale audio classification. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017. [Google Scholar] [CrossRef]
  46. Gemmeke, J.F.; Ellis, D.P.W.; Freedman, D.; Jansen, A.; Lawrence, W.; Moore, R.C.; Plakal, M.; Ritter, M. Audio set: An ontology and human-labeled dataset for audio events. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017. [Google Scholar] [CrossRef]
  47. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  48. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  49. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: The next generation of on-device computer vision networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  50. Haider, N.S. Respiratory sound denoising using empirical mode decomposition, hurst analysis and spectral subtraction. Biomed. Signal Process. Control 2021, 64, 102313. [Google Scholar] [CrossRef]
  51. Emmanouilidou, D.; McCollum, E.D.; Park, D.E.; Elhilali, M. Adaptive noise suppression of pediatric lung auscultations with real applications to noisy clinical settings in developing countries. IEEE Trans. Biomed. Eng. 2015, 62, 2279–2288. [Google Scholar] [CrossRef]
  52. Upadhyay, N.; Karmakar, A. Speech enhancement using spectral subtraction-type algorithms: A comparison and simulation study. Procedia Comput. Sci. 2015, 54, 574–584. [Google Scholar] [CrossRef]
  53. Ramasubramanian, V.; Vijaywargi, D. Speech enhancement based on hypothesized Wiener filtering. In Proceedings of the Interspeech 2008, Brisbane, Australia, 22–26 September 2008. [Google Scholar] [CrossRef]
  54. Podder, P.; Hasan, M.M.; Islam, M.R.; Sayeed, M. Design and implementation of Butterworth, Chebyshev-I and Elliptic Filter for Speech Signal Analysis. Int. J. Comput. Appl. 2014, 98, 12–18. [Google Scholar] [CrossRef]
  55. Haider, N.S.; Behera, A.K. Respiratory sound denoising using sparsity-assisted signal smoothing algorithm. Biocybern. Biomed. Eng. 2022, 42, 481–493. [Google Scholar] [CrossRef]
  56. Ali, M.A.; Shemi, P.M. An improved method of audio denoising based on wavelet transform. In Proceedings of the 2015 IEEE International Conference on Power, Instrumentation, Control and Computing (PICC), Thrissur, India, 9–11 December 2015. [Google Scholar] [CrossRef]
  57. Lee, C.S.; Li, M.; Lou, Y.; Dahiya, R. Restoration of lung sound signals using a hybrid wavelet-based approach. IEEE Sens. J. 2022, 22, 19700–19712. [Google Scholar] [CrossRef]
  58. Dong, G.; Zhang, Z.; Sun, P.; Zhang, M. Adaptive Differential Denoising for Respiratory Sounds Classification. arXiv 2025, arXiv:2506.02505. [Google Scholar]
  59. Rocha, B.M.; Filos, D.; Mendes, L.; Serbes, G.; Ulukaya, S.; Kahya, Y.P.; Jakovljevic, N.; Turukalo, T.L.; Vogiatzis, I.M.; Perantoni, E.; et al. An open access database for the evaluation of respiratory sound classification algorithms. Physiol. Meas. 2019, 40, 035001. [Google Scholar] [CrossRef]
  60. Bhattacharya, D.; Sharma, N.K.; Dutta, D.; Chetupalli, S.R.; Mote, P.; Ganapathy, S.; Chandrakiran, C.; Nori, S.; Suhail, K.K.; Gonuguntla, S.; et al. Coswara: A respiratory sounds and symptoms dataset for remote screening of SARS-CoV-2 infection. Sci. Data 2023, 10, 397. [Google Scholar] [CrossRef]
  61. Heitmann, J.; Glangetas, A.; Doenz, J.; Dervaux, J.; Shama, D.M.; Garcia, D.H.; Benissa, M.R.; Cantais, A.; Perez, A.; Müller, D. DeepBreath—Automated detection of respiratory pathology from lung auscultation in 572 pediatric outpatients across 5 countries. NPJ Digit. Med. 2023, 6, 104. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Representative analysis of normal breath sounds: (upper) time-domain waveform; (middle) frequency spectrum (FFT); (lower) spectrogram.
Figure 1. Representative analysis of normal breath sounds: (upper) time-domain waveform; (middle) frequency spectrum (FFT); (lower) spectrogram.
Sensors 25 05002 g001
Figure 2. Schematic overview of the MFCC extraction pipeline.
Figure 2. Schematic overview of the MFCC extraction pipeline.
Sensors 25 05002 g002
Figure 3. Effect of the number of MFCC coefficients on accuracy.
Figure 3. Effect of the number of MFCC coefficients on accuracy.
Sensors 25 05002 g003
Figure 4. Effect of frame length on accuracy.
Figure 4. Effect of frame length on accuracy.
Sensors 25 05002 g004
Figure 5. Effect of hop length on classification accuracy: (a) Optimized frame length of 800 ms; (b) Real-time motivated frame length of 300 ms.
Figure 5. Effect of hop length on classification accuracy: (a) Optimized frame length of 800 ms; (b) Real-time motivated frame length of 300 ms.
Sensors 25 05002 g005
Figure 6. Accuracy of SVM classifier for MFCC parameter configurations.
Figure 6. Accuracy of SVM classifier for MFCC parameter configurations.
Sensors 25 05002 g006
Table 1. Trade-off between MFCC frame length, system latency, and classification accuracy.
Table 1. Trade-off between MFCC frame length, system latency, and classification accuracy.
Frame Length (ms)Latency (s)Accuracy (±STD)
2000.2000.8462 ± 0.0308
3000.3000.8508 ± 0.0273
4000.4000.8589 ± 0.0337
8000.8000.8716 ± 0.0397
Table 2. Comparison of SVM model performance.
Table 2. Comparison of SVM model performance.
ConfigurationAccuracyAUCF1PrecisionEER
Worst0.73520.88550.74370.73820.1760
Default0.80960.93380.81650.80790.1295
Optimized0.87160.96630.87640.87250.0899
Table 3. Comparison of optimized SVM and deep learning-based approaches.
Table 3. Comparison of optimized SVM and deep learning-based approaches.
Model/FeatureAccuracyAUCF1PrecisionEER
SVM (MFCC, optimized)0.87160.96630.87640.87250.0899
VGGish + RF0.87130.95980.88020.88220.0960
YAMNet + RF0.78710.93720.79840.79520.1295
MobileNetV2 + MFCC0.84160.95940.85480.85380.1088
MobileNetV2 + Spectrogram0.84160.96610.84960.85170.1041
Table 4. Previous work on MFCC parameter optimization for respiratory sound analysis.
Table 4. Previous work on MFCC parameter optimization for respiratory sound analysis.
StudyProblemDatasetModalityClassifiern_mfccFrame Length (ms)Hop Length (ms)Accuracy
Tirronen et al., 2024 [38]Voice pathology detectionSVD 1SpeechSVM13 (default)5005 (default)66.4%
Yan et al., 2025 [25]Automatic detection of respiratory diseasesCambridge COVID-19 Sound databaseSpeech, Cough, and BreathSVM/LSTM3025581.1%/79.2%
CoswaraSpeech, Cough, and BreathSVM/LSTM4025580.6%/79.6%
SVD 1SpeechSVM/LSTM3025571.7%/71.9%
Current studyBreathing phase detectionProprietary datasetBreathSVM303001085.08%
1 Saarbrücken voice disorders.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhantleuova, A.K.; Makashev, Y.K.; Duzbayev, N.T. Optimizing MFCC Parameters for Breathing Phase Detection. Sensors 2025, 25, 5002. https://doi.org/10.3390/s25165002

AMA Style

Zhantleuova AK, Makashev YK, Duzbayev NT. Optimizing MFCC Parameters for Breathing Phase Detection. Sensors. 2025; 25(16):5002. https://doi.org/10.3390/s25165002

Chicago/Turabian Style

Zhantleuova, Assel K., Yerbulat K. Makashev, and Nurzhan T. Duzbayev. 2025. "Optimizing MFCC Parameters for Breathing Phase Detection" Sensors 25, no. 16: 5002. https://doi.org/10.3390/s25165002

APA Style

Zhantleuova, A. K., Makashev, Y. K., & Duzbayev, N. T. (2025). Optimizing MFCC Parameters for Breathing Phase Detection. Sensors, 25(16), 5002. https://doi.org/10.3390/s25165002

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop