Next Article in Journal
Voxel Extraction and Multiclass Classification of Identified Brain Regions across Various Stages of Alzheimer’s Disease Using Machine Learning Approaches
Next Article in Special Issue
Lung Ultrasound Reduces Chest X-rays in Postoperative Care after Thoracic Surgery: Is There a Role for Artificial Intelligence?—Systematic Review
Previous Article in Journal
SkinNet-INIO: Multiclass Skin Lesion Localization and Classification Using Fusion-Assisted Deep Neural Networks and Improved Nature-Inspired Optimization Algorithm
Previous Article in Special Issue
A Comparative Study of Deep Neural Networks for Real-Time Semantic Segmentation during the Transurethral Resection of Bladder Tumors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Harnessing Machine Learning in Vocal Arts Medicine: A Random Forest Application for “Fach” Classification in Opera

1
Institute for Digital Transformation, University of Applied Sciences Ravensburg-Weingarten, Doggenriedstraße, 88250 Weingarten, Germany
2
Occupational College of Music BFSM Krumbach, Mindelheimer Str. 47, 86381 Krumbach, Germany
3
School of Computing, Communication and Business, HTW Berlin University of Applied Sciences, Treskowallee 8, 10318 Berlin, Germany
4
Department of Audiology and Phoniatrics, Charité—Universitätsmedizin Berlin, Campus Charité Mitte, Charitéplatz 1, 10117 Berlin, Germany
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Diagnostics 2023, 13(18), 2870; https://doi.org/10.3390/diagnostics13182870
Submission received: 31 July 2023 / Revised: 28 August 2023 / Accepted: 4 September 2023 / Published: 6 September 2023
(This article belongs to the Special Issue Artificial Intelligence Approaches for Medical Diagnostics in Europe)

Abstract

:
Vocal arts medicine provides care and prevention strategies for professional voice disorders in performing artists. The issue of correct “Fach” determination depending on the presence of a lyric or dramatic voice structure is of crucial importance for opera singers, as chronic overuse often leads to vocal fold damage. To avoid phonomicrosurgery or prevent a premature career end, our aim is to offer singers an improved, objective fach counseling using digital sound analyses and machine learning procedures. For this purpose, a large database of 2004 sound samples from professional opera singers was compiled. Building on this dataset, we employed a classic ensemble learning method, namely the Random Forest algorithm, to construct an efficient fach classifier. This model was trained to learn from features embedded within the sound samples, subsequently enabling voice classification as either lyric or dramatic. As a result, the developed system can decide with an accuracy of about 80% in most examined voice types whether a sound sample has a lyric or dramatic character. To advance diagnostic tools and health in vocal arts medicine and singing voice pedagogy, further machine learning methods will be applied to find the best and most efficient classification method based on artificial intelligence approaches.

1. Introduction

In addition to the artistic aspect, professional singing can also be considered a high-performance sport for the larynx, and just like in high-performance athletes, professional voice care is extremely important [1,2]. Phoniatrics or “voice medicine” is a highly specialized medical field that deals with impairments in the voice in terms of diagnosis, therapy and science [3,4]. Since human vocal function is very complex, the protocol of the European Laryngological Society requires the use of various objective and subjective parameters to describe a voice as accurately and comprehensively as possible [5]. The established diagnostic examination tools are (1) auditory-perceptual voice assessment [6,7,8,9], (2) videolaryngostroboscopy [10,11,12,13], (3) voice range profile measurements for the assessment of vocal performance [14,15,16,17], (4) acoustic-aerodynamic analyses for the evaluation of vocal impairment [18,19,20,21], as well as (5) self-assessment of the voice [22,23,24,25]. Vocal arts medicine is using these diagnostic tools to care for and prevent professional voice disorders in performing artists [26]. In cases of recurrent or persistent voice problems, phoniatric specialists are often confronted with unsolved classification questions: What is the right voice type and “Fach”, lyric or dramatic [27,28]? Since the above-mentioned established examination tools do not include a connection to fach-specific requirements, the precise causes of diagnosed voice problems are currently often difficult to establish adequately.
To date, there are no internationally published studies (apart from [27,28]) within the fields of voice science, pedagogy, or engineering on the objective distinction between lyric and dramatic fach. The scientific literature search revealed only one paper using a ML approach to investigate a timing-based classification method for human voice in opera recordings [29]. However, the aim was merely an automatic identification of famous tenors. In the reality of singing voice pedagogy, the fach question is traditionally answered with an overall view on the respective artist’s personality based on experience and intuition. This traditional approach includes not only the purely vocal requirements, but also all other aspects needed for the realization of an opera role in the corresponding fach. Of course, such a subjective, intuitive approach is also prone to error. However, the issue of correct fach determination is of crucial importance for opera singers to prevent dysodia and vocal fold damage [30,31,32,33]. The German fach system described in detail by Kloiber has occasionally been questioned for some years [34]. In 2008, Ling called it anachronistic and presented a new fach classification for male voices, using “voice structure” [35] (pp. 27–29, 66) and “voice category” [35] (pp. 257–258), among others, as a source of information. He wrote in his proposed solution, “The distinction between lyric and dramatic Fächer … could be retained …” [35] (p. 258) because it primarily describes voice structures to which lyric and dramatic roles can be easily assigned. The present study classifies precisely according to this voice structure, which can be regarded as a necessary prerequisite for mastering a particular fach. In 2016, the editors of the latest edition of Kloiber’s renowned opera guide also decided to reintroduce the critically discussed “Kloiber system” in the detailed chapters on “Fachpartien” [fach-specific roles] and “Besetzungsfragen” [casting concerns], restoring central attention to the distinction between lyric and dramatic Fächer [34] (p. 925). Thus, in addition to the artistically motivated discussions of historical performance practice and contemporary stage direction, the purely physical strain required by a lyric or a dramatic role is given the necessary consideration. In a dramatic role, high volume levels with great carrying power and projection are often required for long periods of time. In a lyric role, the average volume of the accompanying orchestra is lower; it is rather a matter of a special sound aesthetic, often a high agility of the voice.
Misclassification and chronic overuse leading to diagnostically confirmed voice disorders are associated with functional, physical and psycho-emotional impairment, often resulting in a shortened career, occupational disability or phonomicrosurgery [36,37,38,39,40]. Top-level vocalists are therefore not only coached by singing teachers [41], but also supported by phoniatricians or otorhinolaryngologists with voice specialization [2,42]. This interdisciplinary cooperation is not only useful for preventive, diagnostic and therapeutic reasons in individual singers, but is also suitable to answer scientifically still unclear, singing-associated questions. However, this requires examining large amounts of data from a high number of singers, making the inclusion of machine learning (ML) methods extremely helpful. The central and innovative goal of the present work is to offer singers an improved, objective fach, counseling using digital sound analyses and ML procedures, to advance diagnostic tools and health in vocal arts medicine and singing voice pedagogy.

2. Materials and Methods

This section describes our interprofessional methodological approach, starting from the initial musical score studies for the identification of suitable phrases via a selection of sound samples from complete recordings of the respective operas, the construction of the database, definition of appropriate parameters for classification, and up to the final steps of statistical analyses and ML application. To introduce and better understand the process, the workflow chart in Figure 1 gives a general overview of the applied procedures, which will be explained in more detail below.
The extensive database contains sound samples of healthy, well-known opera singers performing in either the lyric or dramatic fach. For the digital fach classification, 281 mezzo-soprano sound samples were added to the 1723 originally collected data. As of 30 June 2023, our database contained a total of 2004 sound samples for the voice types soprano (n = 774), mezzo-soprano (n = 281), tenor (n = 389), baritone (n = 343) and bass (n = 217). In terms of voice structure, our cohort was classified as follows: 422 dramatic vs. 352 lyric sopranos, 147 dramatic vs. 134 lyric mezzo-sopranos, 190 dramatic vs. 199 lyric tenors, 186 dramatic vs. 157 lyric baritones, and 157 dramatic vs. 60 lyric basses.
The sounds have a duration of 6 up to 15 s and are encoded using 16-bit PCM samples at a rate of 22,050 Hz. The criteria for selecting the sounds included the following: (1) a single note in musical notation, (2) no vocal changes and (3) no orchestral accompaniment. Acoustic analysis was limited to this one tone; however, each sample contains a few seconds of musical context before and after the analysis interval. The sound samples were extracted from commercially available complete recordings of the operas. Within each voice type, the sound samples were labeled either as dramatic (1) or as lyric (2) based on the musical role from which the samples were extracted. Kloiber [34] and Ling [35] (pp. 257–286) were used as established references for this attribution. In addition to the lyric or dramatic voice structure, the samples were labeled with further codes containing the following information: voice type, gender, name of singer, musical part, vowel and musical pitch. Labeling examples for selected sound samples in every investigated voice type are shown in Figure 2. A more detailed description of the database can be found in a previous study [27].
For the analysis of the sound files, a development environment was programmed with the software MATLAB version R2017b (The MathWorks, Inc., Natick, MA, USA) which can automatically calculate various features of the sound files. The expanded database makes it possible, for the first time, to present meaningful analysis data of professional opera voices on this scale. Based on the labels assigned in the file names, the system learns which objectively recognizable characteristics a lyric or a dramatic sound sample has.
In order to facilitate the application of ML techniques for resolving the fach classification issue, a collection of 13 distinctive features were meticulously computed and extracted from the original sound files. These features include the following:
  • Start and Stop of Formant (StaFo and StoFo), providing insight into the frequency range that most significantly amplifies vocal tract resonance;
  • Position of Half Energy (PHE) and Frequency of Half Energy (FHE), offering valuable information about the spectral balance of the voice by representing the midpoint position of the energy distribution;
  • Vibrato Rate (VR) and Vibrato Extent (VE), which elucidate details about the speed and extent of pitch oscillation;
  • Perturbation measures Jitter (Jitt) and Shimmer (Shim), which quantify voice stability by calculating cycle-to-cycle variations of the fundamental frequency and amplitude;
  • Strength of Formant Energy (StreFo), signifying the energy intensity within the singer’s formant frequency bands;
  • Further spectral features, i.e., Spectral Centroid (SC), Variance (SV), Skewness (SS), and Kurtosis (SK), delivering essential data about the distribution of energy in the singer’s formant frequency bands.
For a detailed definition of these features, please refer to our previous publications [27,28].
Descriptive statistics were used to determine the basic characteristics of the data. For each voice type and fach, minimum and maximum values, medians, quartiles, means (M), standard deviations (SD) and reference ranges (RR = M ± 1.96 × SD) were computed. To keep the statistical tests consistent throughout all categories, non-parametric tests (which do not rely on normality) were used whenever possible. The Mann–Whitney U test for independent samples was applied because a normal distribution of the data could not be established in each voice type. Histograms were plotted to display the timbre data for fach-specific differences. Pearson correlation coefficients for interval-scaled variables were calculated and visualized in a heat map to identify the relationships between features and to test for collinearity. All statistical analyses were performed using SPSS version 28.0.1 (IBM, Armonk, NY, USA). The level of significance was set at α = 0.05.
The numerical features were fed into a ML classifier in order to distinguish between lyric and dramatic categories based on these specific voice characteristics. The Random Forest (RF) procedure was used for fach classification, an ensemble learning algorithm conceived by Breiman in 2001 [43]. The underlying concept of RF is based on the wisdom of crowds. It includes numerous decision trees and implements random feature and data selection during the training phase. The results from each decision tree are combined, leading to the final decision. Because the RF algorithm is considered one of the most proficient algorithms in the domain of ensemble learning [44], it was the optimal choice for our task. To classify vocal characteristics into lyric or dramatic voice structures, a RF classifier from the Scikit-Learn library (scikit-learn version 1.2.2) [45] was harnessed within a Python environment (Python version 3.9, Python Software Foundation, Wilmington, DE, USA). For the purpose of enhancing the robustness and generalizability of our ML model, we leveraged a stratified 10-fold cross-validation process. This approach ensured an equitable representation of each class across the folds, thereby mitigating potential biases in the training set. This careful methodology should provide reliable classification results in order to objectively delineate between lyric and dramatic voices. The pseudocode-like algorithm used to obtain the classification results is presented in Table 1. The corresponding Python source code is linked in the Supplementary Materials section.
Subsequently, the ML model calculations were scrutinized based on the confusion matrix shown in Table 2.
The efficacy of the model was ascertained using the metrics of accuracy and Balanced Error Rate (BER) [46]. Accuracy describes the performance of a classification model as the number of correct predictions (both true positives and true negatives) divided by the total number of cases examined:
A c c u r a c y = T P + T N T P + T N + F P + F N
With perfect accuracy, the formula equals 1 (i.e., 100% of correct classifications). However, achieving 100% ML model accuracy is typically a sign of some error, such as overfitting. On the other hand, the BER balances both possible error rates by calculating the average between the FP rate and FN rate:
B E R = 0.5 × F P T N + F P + F N T P + F N
The further the result tends towards 0 (or 0%), the better the BER.
Finally, to explore the role of features in the RF classifier, the importance of features during the classification process was determined for each voice type. To provide an explanation of the predictions of the RF classifier, we used SHapley Additive exPlanations (SHAP). This tool for visualizing ML models measures how much each feature affects the prediction, i.e., with SHAP we can understand whether an increase in the value of the feature improves or worsens the prediction.

3. Results

As expected, descriptive statistics provided summarizing information of all vocal parameters and shed light on the distribution of values in our datasets. Table 3 shows the means, SD and 95% reference rages (RR) for the recently introduced timbre parameters in all voice types and for both voice structures (lyric vs. dramatic).
In general, the mean values for PHE, FHE and SC were larger for the higher voice types than for the lower ones. However, the 95% RR revealed a substantial overlap between adjacent voice types within females and males. A difference between both voice structures was determined within each voice type: lyric sopranos, tenors and baritones had significantly higher timbre parameters than dramatic ones (p < 0.001). For the mezzo-soprano and bass voices, the mean differences regarding voice structure were much smaller and statistically not significant. These relations between the abstract p-values and the timbre parameters were also audible: the larger the perceptible discrepancy of the vowel-independent timbre between both voice structures, colloquially referred to as brightness, the more likely this difference is statistically significant.
The histograms of PHE and SC sorted by voice type and fach are shown in Figure 3.
The correlations among all 13 features were successfully computed with Pearson’s correlation coefficient and visualized in a heatmap, as shown in Figure 4.
It is evident from the results that the potential collinearity among features falls within an acceptable range. This observation supported the assumption that every feature imparts unique and additional information to the classification model, thereby enhancing its overall performance. Considering that the characteristics defined by FHE and PHE are identical in practice (namely, the frequency or position of half energy), we chose to retain only PHE (instead of FHE) because its correlations with most other features are lower in the context of this categorization research. Therefore, only 12 features were ultimately used for further fach classification.
Looking at the classification performance for each voice type, Figure 5 presents the values of the confusion matrices, visualizing the prediction results of the RF model for the test dataset.
From this observation, the classification accuracy for dramatic sound samples was generally high in all voice types. However, the model performed differently for the lyric voice structure. While sopranos and tenors showed the best performance for lyric sound samples, the prediction accuracies for mezzo-sopranos and basses were lower.
Using the established evaluation metrics, the performance of our classification model applied to a range of voice types is given in Table 4, which provides a brief summary of the effectiveness of our model. As shown, our model maintained an accuracy of over 70% across all voice types, with the tenor classification achieving the peak at 82.05%. These results illustrated the consistent predictive ability of our model. The BER demonstrated more disparity, ranging from low rates for sopranos, tenors and baritones to significantly higher rates for mezzo-sopranos and basses. For these two voice types, the elevated BER pointed to a less balanced performance of the model in this particular classification.
According to the principle of the RF classifier, the importance of each feature is depicted in Figure 6. Each bar on the chart corresponds to one of the 12 features considered in our model. The height of a bar indicates the degree of importance a feature holds in the classification process for the respective voice type.
To investigate the role each feature plays in the predictions of the model, we displayed SHAP values for the class ‘dramatic’ in each voice type in Figure 7. This provided further insight into the relative importance of all 12 features for the prediction model. In essence, the SHAP values indicated the extent of influence each feature exerts on the model’s prediction. A positive SHAP value suggests that the presence of a feature pushes the prediction higher, implying a stronger association with the predicted class, while a negative SHAP value means the feature tends to push the prediction lower, suggesting a strong association with the other class.

4. Discussion

The developed classification system was able to successfully decide whether a sound sample has a lyric or dramatic character. The results varied according to the voice types. Predictions were excellent for soprano, tenor and baritone voices. Accurate detection rates of about 80% and BER of about 20% can be considered very good, since the discriminatory power of this classification is not very high. This is proven by the fact that there are singers who perform both dramatic and lyric roles, such as the world-famous tenor Luciano Pavarotti [47] (pp. 462–464). Another factor is that there are so-called “Zwischenfächer”, e.g., cavalier baritone [34] (p. 926). This illustrates that the transition between both voice structures can be fluid. On the other hand, close attention must be paid to whether the demands of the respective fach on the vocal folds are manageable for the performers [48,49,50,51]. This applies in particular to the frequently observed development from the lyric to the dramatic fach [33]. In addition, signs of voice changes may be associated with aging [2,52]. That is true, for example, for the parameter VR. On average, dramatic voices have a slower vibrato than lyric ones, but the slowing down of the vibrato is also an age-related phenomenon [27].
For mezzo-sopranos, the classification results were somewhat worse, with an accurate detection rate of about 70% and a BER of about 30%. The spectral timbre parameters SC and FHE confirmed the tendency for lower voice types to have lower values than higher voice types [28]. Traditionally, there is a clear fach distinction in mezzo-sopranos [34] (pp. 931–932), which may well be perceived auditorily in practice. However, our timbre parameters were low for both voice structures; their mean differences were small and statistically not significant. This impaired the fach classification as well, since the ML algorithm could not properly discriminate between lyric and dramatic voice structure for these spectral features. Moreover, the SC and PHE means of lyric mezzo-sopranos were close to those of dramatic sopranos. An overlap of the 95% RR was found for the easily comprehensible and communicable parameter FHE, which can be well translated into the figurative language of vocal pedagogy as well as vocal arts medicine in the context of professional singers’ counseling [53]. Thus, fach changes from lyric mezzo-sopranos to dramatic sopranos like in the famous singer Waltraud Meier [54] (pp. 2157–2160) can be traced with objective parameters. This fach change was also accompanied by a change in voice type. The interaction of the categories voice type and fach is a challenge for objective analysis as well as for auditory perception.
In bass voices, the large difference between BER and accuracy is due to the considerable sample imbalance between the two investigated voice structures. The dark color of all bass voices makes it impossible to classify them with the features used in this study [27]. Therefore, the classification for the bass voice type is not pursued further. Kloiber also does not use the categories lyric/dramatic when describing the fach for bass voices. In contrast, he writes that the division into the categories lyric fach–Zwischenfach–dramatic fach “can be carried out in all voice types” [34] (p. 925). However, the terms “lyric” and “dramatic” are not mentioned in the concrete naming of bass roles, in contrast to all other voice types.
The meticulous methodology of our study facilitated the attainment of reliable and insightful classification outcomes, contributing significantly to our understanding of the delineation between lyric and dramatic voice structure. The RF classifier has demonstrated a generally satisfactory performance in fach classification. One of the key strengths of RF is its robustness to overfitting, thanks to the ensemble learning technique [55]. This model builds multiple decision trees and combines their results, which helps to create a more generalized model that performs well with real-life data [56,57]. Another strength lies in its ability to handle a large number of features without feature scaling or selection. In addition, it provides an estimate of feature importance, which holds instructive significance for vocal arts medicine and singing voice pedagogy. The output of the RF model is also quite interpretable due to its decision tree structure.
However, only basic ensemble learning algorithms were used. Especially in the opera domain, the accuracy of the fach classification can be increased in our classification model. In order to potentially achieve better fach prediction results, more advanced ensemble learning techniques like boosting, stacking, or even integrating deep learning models should be considered in future studies. A further increase in performance is intended and can be achieved by the following means: (1). expanding the amount of tenor, baritone and mezzo-soprano voices in the database to reasonably balance the number of samples regarding voice types and voice structures within a voice type (lyric vs. dramatic), and (2). optimizing the programming of some features. Once these actions have been carried out, the classification results and sound analyses will be supported by an even more solid data basis. This will improve medical advice to opera singers, for example, regarding the following questions:
  • In the phoniatric check-up, how are my current possibilities in solo opera singing assessed with the help of AI-supported evaluation of my objective vocal parameters?
  • Is my voice suitable for a desired change of fach?
  • Are my recent vocal fatigue symptoms due to the selection of my current roles?
  • Are my persistent vocal problems related to my implemented fach change?
To answer these questions, sound samples of new singers are recorded that are comparable in formal structure to the sound samples of our database. Thereafter, the system is trained with all database examples while the new sound recordings are used as test data. The predictions of the system provide the basis for individualized digital voice counseling as an objective aid in the often existential fach classification of opera singers.

5. Conclusions

Our contribution describes a novel ML application in the field of medical voice diagnostics that will soon be adopted into our daily practice as specific fach counseling for professional classical opera singers. It provides a clinical decision support system with risk assessment on whether to sing in the lyric or dramatic fach. This approach advances knowledge management and trustworthy artificial intelligence in objective voice diagnostics to prevent vocal damage, professional disability or premature career end in high-performance singers. In the realm of vocal arts medicine and singing voice pedagogy, the enhancement of diagnostic tools and preservation of vocal health remain essential. In future research, we aim to expand the scope of our dataset and delve deeper into advanced ensemble learning algorithms, striving to identify the optimal and most efficient classification methods underpinned by ML techniques.

Supplementary Materials

The Python source code corresponding to Table 1 can be downloaded from https://github.com/wzehui/Stimmfach-Classification-System.git (accessed on 30 July 2023).

Author Contributions

Conceptualization, M.M. and P.P.C.; methodology, Z.W., M.M. and P.P.C.; software, Z.W. and M.M.; validation, Z.W., M.M., F.C. and P.P.C.; formal analysis, Z.W., M.M. and F.C.; investigation, Z.W. and M.M.; resources, M.M. and P.P.C.; data curation, Z.W. and M.M.; writing—original draft preparation, Z.W., M.M. and P.P.C.; writing—review and editing, Z.W., M.M., F.C. and P.P.C.; visualization, Z.W. and F.C.; supervision, P.P.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used for this study is under a Non-Disclosure Agreement (NDA) and is therefore not available to the public.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sataloff, R.T. Professional singers: The science and art of clinical care. Am. J. Otolaryngol. 1981, 2, 251–266. [Google Scholar] [CrossRef] [PubMed]
  2. Sataloff, R.T. Vocal Health and Pedagogy: Science, Assessment, and Treatment, 3rd ed.; Plural Publishing: San Diego, CA, USA, 2021. [Google Scholar]
  3. Am Zehnhoff-Dinnesen, A.; Wiskirska-Woznica, B.; Neumann, K.; Nawka, T. Phoniatrics I: Fundamentals–Voice Disorders–Disorders of Language and Hearing Development (European Manual of Medicine), 1st ed.; Springer: Berlin, Germany, 2020. [Google Scholar]
  4. Hammarberg, B. Voice research and clinical needs. Folia Phoniatr. Logop. 2000, 52, 93–102. [Google Scholar] [CrossRef] [PubMed]
  5. Dejonckere, P.H.; Bradley, P.; Clemente, P.; Cornut, G.; Crevier-Buchman, L.; Friedrich, G.; Van De Heyning, P.; Remacle, M.; Woisard, V.; Committee on Phoniatrics of the European Laryngological Society (ELS). A basic protocol for functional assessment of voice pathology, especially for investigating the efficacy of (phonosurgical) treatments and evaluating new assessment techniques. Guideline elaborated by the Committee on Phoniatrics of the European Laryngological Society (ELS). Eur. Arch. Otorhinolaryngol. 2001, 258, 77–82. [Google Scholar] [CrossRef] [PubMed]
  6. Dejonckere, P.H.; Remacle, M.; Fresnel-Elbaz, E.; Woisard, V.; Crevier, L.; Millet, B. Reliability and clinical relevance of perceptual evaluation of pathological voices. Rev. Laryngol. Otol. Rhinol. 1998, 119, 247–248. [Google Scholar]
  7. Yamauchi, E.J.; Imaizumi, S.; Maruyama, H.; Haji, T. Perceptual evaluation of pathological voice quality: A comparative analysis between the RASATI and GRBASI scales. Logop. Phoniatr. Vocology 2010, 35, 121–128. [Google Scholar] [CrossRef]
  8. Dos Santos, P.C.M.; Vieira, M.N.; Sansão, J.P.H.; Gama, A.C.C. Effect of Auditory-Perceptual Training with Natural Voice Anchors on Vocal Quality Evaluation. J. Voice 2019, 33, 220–225. [Google Scholar] [CrossRef]
  9. Mahalingam, S.; Venkatraman, Y.; Boominathan, P. Cross-Cultural Adaptation and Validation of Consensus Auditory Perceptual Evaluation of Voice (CAPE-V): A Systematic Review. J. Voice 2021. S0892-1997(21)00359-3. [Google Scholar] [CrossRef]
  10. Nawka, T.; Konerding, U. The Interrater Reliability of Stroboscopy Evaluations. J. Voice 2012, 26, 812.e1–812.e10. [Google Scholar] [CrossRef]
  11. Caffier, P.P.; Schmidt, B.; Gross, M.; Karnetzky, K.; Nawka, T.; Rotter, A.; Seipelt, M.; Sedlmaier, B. A comparison of white light laryngostroboscopy versus autofluorescence endoscopy in the evaluation of vocal fold pathology. Laryngoscope 2013, 123, 1729–1734. [Google Scholar] [CrossRef]
  12. Powell, M.E.; Deliyski, D.D.; Hillman, R.E.; Zeitels, S.M.; Burns, J.A.; Mehta, D.D. Comparison of Videostroboscopy to Stroboscopy Derived From High-Speed Videoendoscopy for Evaluating Patients With Vocal Fold Mass Lesions. Am. J. Speech Lang. Pathol. 2016, 25, 576–589. [Google Scholar] [CrossRef]
  13. Caffier, P.P.; Nawka, T.; Ibrahim-Nasr, A.; Thomas, B.; Müller, H.; Ko, S.R.; Song, W.; Gross, M.; Weikert, S. Development of three-dimensional laryngostroboscopy for office-based laryngeal diagnostics and phonosurgical therapy. Laryngoscope 2018, 128, 2823–2831. [Google Scholar] [CrossRef] [PubMed]
  14. Pabon, P.; Ternström, S.; Lamarche, A. Fourier descriptor analysis and unification of voice range profile contours: Method and applications. J. Speech Lang. Hear. Res. 2011, 54, 755–776. [Google Scholar] [CrossRef] [PubMed]
  15. Ternström, S.; Pabon, P.; Södersten, M. The Voice Range Profile: Its Function, Applications, Pitfalls and Potential. Acta Acust. United Acust. 2016, 102, 268–283. [Google Scholar] [CrossRef]
  16. Caffier, P.P.; Möller, A.; Forbes, E.; Müller, C.; Freymann, M.L.; Nawka, T. The Vocal Extent Measure: Development of a Novel Parameter in Voice Diagnostics and Initial Clinical Experience. BioMed Res. Int. 2018, 2018, 3836714. [Google Scholar] [CrossRef] [PubMed]
  17. Freymann, M.L.; Mathmann, P.; Rummich, J.; Müller, C.; Neumann, K.; Nawka, T.; Caffier, P.P. Gender-specific reference ranges of the vocal extent measure in young and healthy adults. Logop. Phoniatr. Vocology 2020, 45, 73–81. [Google Scholar] [CrossRef]
  18. Wuyts, F.L.; De Bodt, M.S.; Molenberghs, G.; Remacle, M.; Heylen, L.; Millet, B.; Van Lierde, K.; Raes, J.; Van de Heyning, P.H. The dysphonia severity index: An objective measure of vocal quality based on a multiparameter approach. J. Speech Lang. Hear. Res. 2000, 43, 796–809. [Google Scholar] [CrossRef]
  19. Hakkesteegt, M.M.; Wieringa, M.H.; Brocaar, M.P.; Mulder, P.G.H.; Feenstra, L. The Interobserver and Test-Retest Variability of the Dysphonia Severity Index. Folia Phoniatr. Logop. 2008, 60, 86–90. [Google Scholar] [CrossRef]
  20. Patel, R.R.; Awan, S.N.; Barkmeier-Kraemer, J.; Courey, M.; Deliyski, D.D.; Eadie, T.; Paul, D.; Švec, J.; Hillman, R. Recommended Protocols for Instrumental Assessment of Voice: American Speech-Language-Hearing Association Expert Panel to Develop a Protocol for Instrumental Assessment of Vocal Function. Am. J. Speech Lang. Pathol. 2018, 27, 887–905. [Google Scholar] [CrossRef]
  21. Barsties V Latoszek, B.; Mathmann, P.; Neumann, K. The cepstral spectral index of dysphonia, the acoustic voice quality index and the acoustic breathiness index as novel multiparametric indices for acoustic assessment of voice quality. Curr. Opin. Otolaryngol. Head. Neck Surg. 2021, 29, 451–457. [Google Scholar] [CrossRef]
  22. Jacobson, B.H.; Johnson, A.; Grywalski, C.; Silbergleit, A.; Jacobson, G.; Benninger, M.S.; Newman, C.W. The Voice Handicap Index (VHI): Development and Validation. Am. J. Speech Lang. Pathol. 1997, 6, 66–70. [Google Scholar] [CrossRef]
  23. Portone, C.R.; Hapner, E.R.; McGregor, L.; Otto, K.; Johns, M.M., 3rd. Correlation of the Voice Handicap Index (VHI) and the Voice-Related Quality of Life Measure (V-RQOL). J. Voice 2007, 21, 723–727. [Google Scholar] [CrossRef] [PubMed]
  24. Nawka, T.; Verdonck-de Leeuw, I.M.; De Bodt, M.; Guimaraes, I.; Holmberg, E.B.; Rosen, C.A.; Schindler, A.; Woisard, V.; Whurr, R.; Konerding, U. Item Reduction of the Voice Handicap Index Based on the Original Version and on European Translations. Folia Phoniatr. Logop. 2009, 61, 37–48. [Google Scholar] [CrossRef] [PubMed]
  25. Caffier, F.; Nawka, T.; Neumann, K.; Seipelt, M.; Caffier, P.P. Validation and Classification of the 9-Item Voice Handicap Index (VHI-9i). J. Clin. Med. 2021, 10, 3325. [Google Scholar] [CrossRef] [PubMed]
  26. Benninger, M.S.; Jacobson, B.H.; Johnson, A.F. Vocal Arts Medicine: The Care and Prevention of Professional Voice Disorders; Thieme: Stuttgart, Germany, 1994. [Google Scholar]
  27. Müller, M.; Schulz, T.; Ermakova, T.; Caffier, P.P. Lyric or Dramatic-Vibrato Analysis for Voice Type Classification in Professional Opera Singers. IEEE/ACM Trans. Audio Speech Lang. 2021, 29, 943–955. [Google Scholar] [CrossRef]
  28. Müller, M.; Wang, Z.; Caffier, F.; Caffier, P.P. New objective timbre parameters for classification of voice type and fach in professional opera singers. Sci. Rep. 2022, 12, 17921. [Google Scholar] [CrossRef]
  29. Marinescu, M.-C.; Ramirez, R. A Timing-Based Classification Method for Human Voice in Opera Recordings. In Proceedings of the 2009 International Conference on Machine Learning and Applications, ICMLA Proceedings, Miami, FL, USA, 13–15 December 2009; pp. 577–582. [Google Scholar] [CrossRef]
  30. Ropero Rendón, M.D.M.; Ermakova, T.; Freymann, M.L.; Ruschin, A.; Nawka, T.; Caffier, P.P. Efficacy of Phonosurgery, Logopedic Voice Treatment and Vocal Pedagogy in Common Voice Problems of Singers. Adv. Ther. 2018, 35, 1069–1086. [Google Scholar] [CrossRef]
  31. Kwok, M.; Eslick, G.D. The Impact of Vocal and Laryngeal Pathologies Among Professional Singers: A Meta-analysis. J. Voice 2019, 33, 58–65. [Google Scholar] [CrossRef]
  32. Rubin, A.D.; Codino, J. The Art of Caring for the Professional Singer. Otolaryngol. Clin. N. Am. 2019, 52, 769–778. [Google Scholar] [CrossRef]
  33. Caffier, P.P.; Weikert, S.; Nawka, T. Acute Vocal Fold Hemorrhage While Singing. Dtsch. Arztebl. Int. 2023, 120, 114. [Google Scholar] [CrossRef]
  34. Kloiber, R.; Konold, W.; Maschka, R. Handbuch der Oper, 14th ed.; J.B. Metzler: Stuttgart, Germany, 2016. [Google Scholar]
  35. Ling, P.A. Stimme, Stimmfach, Fachvertrag; Wißner: Augsburg, Germany, 2008. [Google Scholar]
  36. Benninger, M.S. The professional voice. J. Laryngol. Otol. 2011, 125, 111–116. [Google Scholar] [CrossRef]
  37. Sataloff, R.T. Professional voice users: The evaluation of voice disorders. Occup. Med. 2001, 16, 633–647,v. [Google Scholar] [PubMed]
  38. McKinney, J. The singing/acting young adult from a singing instruction perspective. J. Voice 1997, 11, 153–155. [Google Scholar] [CrossRef] [PubMed]
  39. Titze, I.R. Mechanical stress in phonation. J. Voice 1994, 8, 99–105. [Google Scholar] [CrossRef] [PubMed]
  40. Zeitels, S.M. The Art and Craft of Phonomicrosurgery in Grammy Award-Winning Elite Performers. Ann. Otol. Rhinol. Laryngol. 2019, 128, 7S–24S. [Google Scholar] [CrossRef] [PubMed]
  41. Chapman, J.L.; Morris, R. Singing and Teaching Singing: A Holistic Approach to Classical Voice, 4th ed.; Plural Publishing: San Diego, CA, USA, 2021. [Google Scholar]
  42. Harris, T.; Howard, D.M. The Voice Clinic Handbook, 2nd ed.; Compton Publishing: Braunton, UK, 2018. [Google Scholar]
  43. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  44. Prasad, A.M.; Iverson, L.R.; Liaw, A. Newer Classification and Regression Tree Techniques: Bagging and Random Forests for Ecological Prediction. Ecosystems 2006, 9, 181–199. [Google Scholar] [CrossRef]
  45. scikit-learn.org. Available online: https://scikit-learn.org/stable/modules/generated/sklearn.ensemble.RandomForestClassifier.html (accessed on 24 July 2023).
  46. Zhou, Z.H. Model Selection and Evaluation. In Machine Learning; Zhou, Z.H., Ed.; Springer: Singapore, 2021; pp. 25–56. [Google Scholar] [CrossRef]
  47. Mecke, A.-C.; Pfleiderer, M.; Richter, B.; Seedorf, T. Lexikon der Gesangsstimme, 2nd ed.; Laaber: Laaber, Germany, 2018. [Google Scholar]
  48. Coleman, R.F. Performance demands and the performer’s vocal capabilities. J. Voice 1987, 1, 209–216. [Google Scholar] [CrossRef]
  49. Scherer, R.C.; Brewer, D.W.; Colton, R.; Rubin, L.S.; Raphael, B.N.; Miller, R.; Howell, E.; Moore, G.P. The Integration of Voice Science, Voice Pathology, Medicine, Public Speaking, Acting, and Singing. J. Voice 1994, 8, 359–374. [Google Scholar] [CrossRef]
  50. Sataloff, R.T. The Professional Voice: Part I. J. Voice 1987, 1, 92–104. [Google Scholar] [CrossRef]
  51. Klingholz, F. Die Akustik der gestörten Stimme; Thieme: Stuttgart, Germany, 1986. [Google Scholar]
  52. Titze, I. Critical periods of vocal change–advanced age. NATS J. 1993, 49, 27. [Google Scholar]
  53. Richter, B.; Echternach, M. Medical treatment of singers’ voices. HNO 2011, 59, 547–555. [Google Scholar] [CrossRef] [PubMed]
  54. Kesting, J. Die Großen Sänger; Hoffmann und Campe: Hamburg, Germany, 2008. [Google Scholar]
  55. Biau, G. Analysis of a random forests model. J. Mach. Learn. Res. 2012, 13, 1063–1095. [Google Scholar] [CrossRef]
  56. Sharaf, A.I. Sleep Apnea Detection Using Wavelet Scattering Transformation and Random Forest Classifier. Entropy 2023, 25, 399. [Google Scholar] [CrossRef] [PubMed]
  57. Zhang, H.; Chi, M.; Su, D.; Xiong, Y.; Wei, H.; You, Y.; Zuo, Y.; Yang, L. A random forest-based metabolic risk model to assess the prognosis and metabolism-related drug targets in ovarian cancer. Comput. Biol. Med. 2023, 153, 106432. [Google Scholar] [CrossRef]
Figure 1. Chart of the overall proposed workflow for fach classification in opera voices.
Figure 1. Chart of the overall proposed workflow for fach classification in opera voices.
Diagnostics 13 02870 g001
Figure 2. Labeling examples for selected sound samples in every investigated voice type (S—soprano, M—mezzo-soprano, T—tenor, B—baritone, L—bass). Voice structure: 1—dramatic, 2—lyric; gender: m—male, w—female; musical part: Isol—Isolde (from Tristan and Isolde by R. Wagner), Dora—Dorabella (from Così fan tutte by W.A. Mozart), Lohe—Lohengrin (from Lohengrin by R. Wagner), Mala—Maltesta (from Don Pasquale by G. Donizetti), Hage—Hagen (from Twilight of the Gods by R. Wagner); vowel sound symbols according to International Phonetic Alphabet: I—/i:/, U—/u:/, O—/o:/, A—/ɑ:/, E—/e:/; musical pitch according to scientific pitch notation (tuning of concert pitch A4 = 442 Hz): G4–394 Hz, G5–788 Hz, G♯4–417 Hz, C4–263 Hz, E♭3–156 Hz; index pitch/vowel: 1… n series number of the analyzed pitch/vowel in case of multiple occurrences.
Figure 2. Labeling examples for selected sound samples in every investigated voice type (S—soprano, M—mezzo-soprano, T—tenor, B—baritone, L—bass). Voice structure: 1—dramatic, 2—lyric; gender: m—male, w—female; musical part: Isol—Isolde (from Tristan and Isolde by R. Wagner), Dora—Dorabella (from Così fan tutte by W.A. Mozart), Lohe—Lohengrin (from Lohengrin by R. Wagner), Mala—Maltesta (from Don Pasquale by G. Donizetti), Hage—Hagen (from Twilight of the Gods by R. Wagner); vowel sound symbols according to International Phonetic Alphabet: I—/i:/, U—/u:/, O—/o:/, A—/ɑ:/, E—/e:/; musical pitch according to scientific pitch notation (tuning of concert pitch A4 = 442 Hz): G4–394 Hz, G5–788 Hz, G♯4–417 Hz, C4–263 Hz, E♭3–156 Hz; index pitch/vowel: 1… n series number of the analyzed pitch/vowel in case of multiple occurrences.
Diagnostics 13 02870 g002
Figure 3. Histograms of timbre parameters PHE (top) and SC (bottom), sorted by voice type (soprano to bass) as well as voice structure (lyric = blue, dramatic = green). Visually distinct plots when comparing lyric versus dramatic voice structure within a voice type can be easily differentiated (e.g., PHE baritone, SC soprano). Many histograms followed a normal distribution, with the most notable exception being lyric basses, who also suffered from the smallest sample size (n = 60).
Figure 3. Histograms of timbre parameters PHE (top) and SC (bottom), sorted by voice type (soprano to bass) as well as voice structure (lyric = blue, dramatic = green). Visually distinct plots when comparing lyric versus dramatic voice structure within a voice type can be easily differentiated (e.g., PHE baritone, SC soprano). Many histograms followed a normal distribution, with the most notable exception being lyric basses, who also suffered from the smallest sample size (n = 60).
Diagnostics 13 02870 g003
Figure 4. Correlation heatmap of selected acoustic features in voice classification, showing the relationships between candidates for the input features in a RF model.
Figure 4. Correlation heatmap of selected acoustic features in voice classification, showing the relationships between candidates for the input features in a RF model.
Diagnostics 13 02870 g004
Figure 5. Confusion matrices of all examined voice types for the test dataset.
Figure 5. Confusion matrices of all examined voice types for the test dataset.
Diagnostics 13 02870 g005
Figure 6. Importance of different features for the classification of voice structure in all investigated voice types, based on impurity.
Figure 6. Importance of different features for the classification of voice structure in all investigated voice types, based on impurity.
Diagnostics 13 02870 g006
Figure 7. SHAP summary plots of input features for sopranos and mezzo-sopranos (upper row), as well as tenors, baritones and basses (lower row). Features were sorted in descending order of the impact on the model output. In all plots, only the results for the dramatic voice structure are shown.
Figure 7. SHAP summary plots of input features for sopranos and mezzo-sopranos (upper row), as well as tenors, baritones and basses (lower row). Features were sorted in descending order of the impact on the model output. In all plots, only the results for the dramatic voice structure are shown.
Diagnostics 13 02870 g007
Table 1. Step-by-step instructions for predicting “Fach” labels using the Random Forest classifier.
Table 1. Step-by-step instructions for predicting “Fach” labels using the Random Forest classifier.
Diagnostics 13 02870 i001
Diagnostics 13 02870 i002
Diagnostics 13 02870 i003
Table 2. 2 × 2 Confusion matrix visualizing all possible ML model prediction outcomes.
Table 2. 2 × 2 Confusion matrix visualizing all possible ML model prediction outcomes.
Predicted DramaticPredicted Lyric
Actual DramaticTrue Positive (TP)False Negative (FN)
Actual LyricFalse Positive (FP)True Negative (TN)
Table 3. Timbre parameters Frequency of Half Energy (FHE), Position of Half Energy (PHE) and Spectral Centroid (SC) depending on voice structure (lyric vs. dramatic) in all voice types.
Table 3. Timbre parameters Frequency of Half Energy (FHE), Position of Half Energy (PHE) and Spectral Centroid (SC) depending on voice structure (lyric vs. dramatic) in all voice types.
Vocal Characteristics of Investigated Opera SingersFrequency of Half Energy
(FHE in Hz)
Position of Half Energy
(PHE in %)
Spectral Centroid
(SC in Hz)
Voice TypeVoice Structure (n Sound Samples)Mean (SD)95% RRp-Value
(MWU-Test)
Mean (SD)95% RRp-Value
(MWU-test)
Mean (SD)95% RRp-Value
(MWU-Test)
sopranoall (n = 774)3092 (284)3072; 3112 35.98 (12.92)35.07; 36.89 3207 (154)3196; 3218
lyric (n = 352)3169 (307)3136; 3201p < 0.00139.46 (13.94)38.00; 40.92p < 0.0013271 (159)3255; 3288p < 0.001
dramatic (n = 422)3028 (247)3004; 305233.08 (11.23)32.00; 34.153153 (128)3141; 3166
mezzo-sopranoall (n = 281)2985 (206)2960; 3009 31.10 (9.36)30.00; 32.20 3097 (121)3083; 3111
lyric (n = 134)2990 (217)2953; 3027p = 0.76631.36 (9.85)29.68; 33.04p = 0.7673110 (132)3088; 3133p = 0.096
dramatic (n = 147)2979 (196)2947; 301130.86 (8.92)29.41; 32.313085 (108)3067; 3103
tenorall (n = 389)2705 (221)2683; 2727 44.05 (13.82)42.67; 45.43 2740 (143)2725; 2754
lyric (n = 199)2760 (205)2732; 2789p < 0.00147.49 (12.77)45.70; 49.28p < 0.0012789 (129)2771; 2807p < 0.001
dramatic (n = 190)2648 (224)2616; 268040.45 (13.99)38.45; 42.452688 (139)2668; 2708
baritoneall (n = 343)2454 (206)2432; 2476 28.34 (12.86)26.98; 29.71 2532 (144)2516; 2547
lyric (n = 157)2576 (206)2543; 2608p < 0.00135.94 (12.86)33.91; 37.97p < 0.0012625 (135)2604; 2646p < 0.001
dramatic (n = 186)2351 (140)2331; 237121.93 (8.72)20.67; 23.202453 (97)2439; 2467
bassall (n = 217)2384 (164)2363; 2406 24.01 (10.24)22.64; 25.38 2480 (92)2468; 2493
lyric (n = 60)2379 (200)2327; 2430p = 0.27023.65 (12.46)20.43; 26.87p = 0.2652491 (98)2466; 2516p = 0.252
dramatic (n = 157)2387 (149)2363; 241024.15 (9.30)22.68; 25.612476 (89)2462; 2490
Legend: MWU = Mann–Whitney U; n = number; RR = reference range (lower bound; upper bound); SD = standard deviation.
Table 4. Classification results in all investigated voice types.
Table 4. Classification results in all investigated voice types.
Voice TypeAccuracyBalanced Error Rate (BER)
Soprano81.94%16.91%
Mezzo-soprano70.18%29.44%
Tenor82.05%17.91%
Baritone79.71%22.83%
Bass79.55%28.39%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Z.; Müller, M.; Caffier, F.; Caffier, P.P. Harnessing Machine Learning in Vocal Arts Medicine: A Random Forest Application for “Fach” Classification in Opera. Diagnostics 2023, 13, 2870. https://doi.org/10.3390/diagnostics13182870

AMA Style

Wang Z, Müller M, Caffier F, Caffier PP. Harnessing Machine Learning in Vocal Arts Medicine: A Random Forest Application for “Fach” Classification in Opera. Diagnostics. 2023; 13(18):2870. https://doi.org/10.3390/diagnostics13182870

Chicago/Turabian Style

Wang, Zehui, Matthias Müller, Felix Caffier, and Philipp P. Caffier. 2023. "Harnessing Machine Learning in Vocal Arts Medicine: A Random Forest Application for “Fach” Classification in Opera" Diagnostics 13, no. 18: 2870. https://doi.org/10.3390/diagnostics13182870

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop