Next Article in Journal
Clinical Predictors of Grade Group Upgrading for Radical Prostatectomy Specimens Compared to Those of Preoperative Needle Biopsy Specimens
Previous Article in Journal
Morphological Diversity of Permanent Maxillary Lateral Incisors and Their Impact on Aesthetics and Function in Orthodontically Treated Patients
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Employing Energy and Statistical Features for Automatic Diagnosis of Voice Disorders

by
Avinash Shrivas
1,*,
Shrinivas Deshpande
1,
Girish Gidaye
2,
Jagannath Nirmal
3,
Kadria Ezzine
4,
Mondher Frikha
4,
Kamalakar Desai
5,
Sachin Shinde
6,
Ankit D. Oza
7,
Dumitru Doru Burduhos-Nergis
8 and
Diana Petronela Burduhos-Nergis
8,*
1
Department of Computer Science & Technology, Degree College of Physical Education, Sant Gadge Baba Amravati University, Amravati 444605, India
2
Department of Electronics and Computer Science, Vidyalankar Institute of Technology, Mumbai University, Mumbai 400037, India
3
Department of Electronics Engineering, Somaiya Vidyavihar University, Mumbai 400077, India
4
ATISP, ENET’COM, Sfax University, Sfax 3000, Tunisia
5
Department of Electronics and Telecommunication Engineering, Bharati Vidyapeeth’s College of Engineering, Shivaji University, Kolhapur 416013, India
6
Department of Mechanical Engineering, Datta Meghe College of Engineering, Mumbai University, Airoli, Navi Mumbai 400708, India
7
Department of Computer Sciences and Engineering, Institute of Advanced Research, The University for Innovation, Gandhianagar 382426, India
8
Faculty of Materials Science and Engineering, Gheorghe Asachi Technical University of Iasi, 700050 Iasi, Romania
*
Authors to whom correspondence should be addressed.
Diagnostics 2022, 12(11), 2758; https://doi.org/10.3390/diagnostics12112758
Submission received: 23 September 2022 / Revised: 30 October 2022 / Accepted: 9 November 2022 / Published: 11 November 2022
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
The presence of laryngeal disease affects vocal fold(s) dynamics and thus causes changes in pitch, loudness, and other characteristics of the human voice. Many frameworks based on the acoustic analysis of speech signals have been created in recent years; however, they are evaluated on just one or two corpora and are not independent to voice illnesses and human bias. In this article, a unified wavelet-based paradigm for evaluating voice diseases is presented. This approach is independent of voice diseases, human bias, or dialect. The vocal folds’ dynamics are impacted by the voice disorder, and this further modifies the sound source. Therefore, inverse filtering is used to capture the modified voice source. Furthermore, the fundamental frequency independent statistical and energy metrics are derived from each spectral sub-band to characterize the retrieved voice source. Speech recordings of the sustained vowel /a/ were collected from four different datasets in German, Spanish, English, and Arabic to run the several intra and inter-dataset experiments. The classifiers’ achieved performance indicators show that energy and statistical features uncover vital information on a variety of clinical voices, and therefore the suggested approach can be used as a complementary means for the automatic medical assessment of voice diseases.

1. Introduction

A key component of human conversation is speech signals. However, the multiple laryngeal illnesses harm the parts of the voice box that are responsible for speech, creating an abnormal voice that is unfit for everyday uses [1]. As a result, it is essential to identify the source of a voice anomaly as soon as possible. Traditional methods for examining voice dysfunction in medical practice are subjective and invasive. The grade, roughness, breathiness, asthenia, and strain (GRBAS) and the Consensus Auditory Perceptual Evaluation–Voice (CAPE-V) scales are used in the traditional perceptual analysis [2]. Such perceptual examination is subjective, imprecise, and necessitates expert professionals. Laryngoscopy, stroboscopy, and endoscopy are methods used in the obstructive clinical process to observe the larynx’s inner structure [3]. These testing methods necessitate a skilled ear, nose, and throat doctor and a cutting-edge laboratory outfitted with instruments. The patient may perform (present) a reflex response in the orinasal cavity while using these methods, resulting in erroneous assessments [4]. For the early and reliable assessment of laryngeal illnesses, an effective alternative approach focused on an acoustical analysis of the clinical voice may be possible. There are two main types of established voice pathology identification frameworks: classical pipeline frameworks and newer end-to-end frameworks [5]. The conventional pipeline scheme is made up of two parts: extraction of features and identification [5,6]. The feature extraction section attempts to obtain useful information from speech records. The attribute sets employed by the research community for the identification of voice abnormalities can be divided into four types: (1) perturbation attributes; (2) spectral and cepstral attributes; (3) complexity attributes; and (4) glottal attributes [5,6,7]. A classification algorithm is used in the detection phase to categorize the source speech as normal or pathological. The numerous acoustic analysis-based approaches focused on machine learning are addressed in [8]. In the end-to-end voice analysis approach, huge speech records are required. However, because of a scarcity of records, end-to-end schemes have still not been extensively used in the research area of voice pathology identification [6,7]. The analysis of pathological voice is challenged by the scarcity of available speech samples, as these samples come from patients whose illness may be so severe that only short recordings are possible. As a result of the scarcity of the large quantity of training samples required to educate end-to-end schemes, the conventional pipeline scheme remains a workable option [7].
Perturbation metrics detect the occurrence of amplitude and frequency instability in the voice signal induced by the abnormal vibration of vocal tract tissues and the inability of glottal closure. Shimmer and jitter are the two most common metrics of this category employed in auditory analysis; however, there are numerous variants of jitter and shimmer in use to recognize voice pathologies. The framework based on these features is presented in [9]. The recordings of sustained vowels as well as continuous speech mined from the Massachusetts Eye and Ear Infirmary (MEEI) dataset were employed to carry out the voice assessment task. Two sets of experiments were carried out: in the first set, an individual metric is chosen, while in the second set, a combination of all nine metrics is considered. The categorization was more correct when the individual metrics were selected from sustained verbal recordings than with those based on continuous speech samples. The accuracy rate increased to 96% by considering a combination of the metrics. The correlation dimension is used along with the two main perturbation metrics to separate laryngeal paralysis subjects from healthy ones [10]. The spectral jitter estimation method is used to obtain the jitter, since it is resilient against inaccuracies in estimating the pitch time. The voice recordings from the MEEI and Principe de Asturias (PdA) datasets are characterized. For the decision making, a separate threshold is decided for sustained vowel and running speech-based jitter. The area under the curve (AUC) obtained 0.94 and 0.84 for the jitter extracted from vowels for MEEI and PdA, respectively. However, the achieved AUC for running speech is 0.96 [11]. Other widely accepted disturbance metrics that assess the existence of aspiration noise include the harmonic-to-noise ratio (HNR), normalized noise entropy (NNE), voice turbulence index, and glottal-to-noise excitation (GNE) ratio. To assess the noise elements embedded in the pathological audio signal, an NNE metric is introduced and its efficacy in the diagnosis of larynx diseases is examined with 250 vocal samples (64 normal people and 186 sufferers with diverse laryngeal illnesses) [12]. To quantify such short-term perturbation metrics, precise fundamental frequency (Fo) mining is required. The precise estimation of the Fo in pathological voice is difficult [13,14], so all of these metrics are rejected from the attributes endorsed by the American Speech–Language–Hearing Association [15].
Spectrum and cepstrum-derived parameters have been widely adopted for voice disorder assessment, as these measures are usually simple to estimate and do not require an F0 calculation [14,16]. These measures have the potential to characterize both continuous speech and steady vowels. The widely accepted attributes in this type are mel-frequency cepstral coefficients (MFCC), linear predictive coding coefficients (LPC), linear predictive cepstral coefficients (LPCC), perceptual linear prediction (PLP), and cepstral peak prominence (CPP). In [17], auditory processed spectrum (APS) and cepstral coefficients based on all-pole model (APCC) measures are used for the detection and labeling of voice pathologies. The running speech recordings mined from the MEEI dataset were used to test the proposed method. The best-achieved voice pathology detection rate is 98.22% and 99.56% and the classification rate obtained is 93.33% and 89.33% with APS and APCC measures, respectively, using the Gaussian mixture model (GMM).
MFCC and their first (Δ) and second (ΔΔ) derivatives, along with the energy, are used as features for voice assessment in [18]. The proposed system is assessed using recordings of the sustained vowel /a/ quarried from the MEEI dataset. For decision making, multilayer perceptron (MLP) and learning vector quantization (LVQ) detectors are used. The best detection rate attained is 96% using LVQ detector and has performed better than MLP. The optimization of the preceding algorithm is addressed by Godino-Llorente et al. [19], which replaced the previous classification technique with a GMM and uses the extended subset speech recording from MEEI. In addition, to analyze the impact of derivatives in the voice pathology detection process, a feature selection approach centered on the F-Ratio and Fisher’s discriminant ratio is used. The authors noted a detection efficiency of 94%, suggesting that the addition of the ΔΔ derivative does not further boost the system’s accuracy. Then again, MFCC attributes as well as the same MEEI speaker audio files are selected in [20]. Numerous methodological problems are addressed with regard to developing systems that identify pathologies automatically. With the suggested technique, which is intended to enable comparisons with other relevant literature, an artificial neural network (ANN) efficiency of 90% is achieved. Ali et al. [21] have conducted various intra- and cross-dataset tests to verify the efficacy of the MFCC-based framework for voice impairment assessment. The system is tested using three diverse datasets: MEEI, SVD, and AVPD. The simulation results show that the MFCC measure-based system’s performance differs from dataset to dataset. The recognition accuracy for the intra-database varies considerably from 72% to 95% and for the inter-database varies from 47% to 82%. The results conclude that existing traditional speech attributes are not strongly linked with speech and are therefore not credible for screening and diagnostic tools. The authors of [14] tested the diagnostic importance of spectral/cepstral measurements, namely, cepstral peak prominence (CPP) and low–high spectral ratio (LHr) to distinguish pathological from non-pathological speech using sustained vowel and running speech records from the MEEI. The CPP (AUC = 0.95) performs better than LHr (AUC = 0.74) in diagnosis when these parameters are extracted from sustained vowels. A similar trend is observed when these parameters are extracted from running speech recordings, i.e., the AUC = 90 achieved with CPP and the AUC = 79 achieved with LHr.
Several researchers have found that certain nonlinear processes occur in speech production and cannot be characterized by traditional metrics such as shimmer, jitter, GNE, HNR, NNE, MFCC, LPCC, and PLP [22]. This is partly related to nonlinearities linked with pressure in the glottis, stress–strain curves of vocal cord muscle tissues, and vocal cord collision [23]. Complex metrics have been developed to characterize the nonlinearity in the audio signals. The correlation dimension (D2) is one of the most-studied complexity attributes of automatic voice quality evaluation systems. For example, the authors in [24] use it to evaluate signals as per their categorization in a dataset of 122 pathological records of the vowel /a/. The results show that D2 uncovers substantial differences between all kinds of signals. Subsequently, the authors of [11] considered D2 with jitter and shimmer to examine the prolonged vowels of persons with normophonic and unilateral laryngeal paralysis in an MEEI dataset subset (57 normal and 67 pathologic). Jitter and shimmer are only measured in nearly periodic sounds, resulting in considerable distinctions in both abnormal and normal voice groups (AUC = 0.7 for jitter and AUC = 0.8 for shimmer). D2 is determined for all voices having clear differences among pathological and normal states (AUC = 0.74). Studies suggest that perturbation and complexity metrics may categorize speech signals owing to their complementarity. A further investigation, including D2, jitter, shimmer, and second order entropies, is reported by Zhang and Jiang [25]. For the study of continuous speech production, the vowel /a/ and the vowels derived from the reading passage of the MEEI corpus are used. The findings revealed the ability of complexity attributes to distinguish among disordered and normal speech cases. On the other hand, perturbation metrics worked well with sustained vowels, but should be applied with precaution in vowels derived from continuous speech. Little et al. [26] presented a couple of new complexity attributes for the diagnosis of voice impairments: recurrence period density entropy (RPDE) and detrended fluctuations analysis (DFA). The recognition rate of 92% was attained with the proposed new attributes by employing a quadratic discriminant analysis for classification and bootstrap resampling for validation in the MEEI dataset. This recognition rate is greater than that attained by integrating other conventional parameters that do not surpass 81%. Two new entropy attributes based on the Markov models introduced in [27] are: Renyi HMM entropy and Shannon HMM entropy. These new attributes have been considered to characterize the MEEI dataset, together with nine other complexity attributes, three perturbation parameters, and MFCC. With all combined parameters, an accuracy of 98% is attained using a GMM.
Despite the fact that spectral and cepstral attributes have really been extensively employed in the assessment of voice impairment, it is important to note that they are quite likely to portray the vocal tract system than variations in the vibratory pattern of the vocal folds directly caused by voice illnesses. As a result, glottal source attributes derived from the glottal wave have a high research value. Forero et al. [28] captured the vocal tract excitation by glottal inverse filtering (GIF). Sixteen time instant-based, quotient-based, and frequency domain glottal metrics are extracted from the glottal wave and 12 MFCC extracted from speech signal to carry out the voice assessment. Three classes of subjects, namely, vocal fold paralysis, nodule, and healthy, are considered for the experiments. The classification accuracy attained with glottal parameters is 95.80%, 82.00%, and 96.20%; however, with the MFCC parameters, these values are 75.20%, 87.00%, and 80.00% using ANN, HMM, and SVM respectively. With the fusion of glottal and MFCC parameters, the achieved classification rate is 96.60%, 92.00%, and 97.20% using ANN, HMM, and SVM, respectively. Similarly, in [29], the authors proposed a voice disorder identification and classification system using an interlaced derivative pattern (IDP). The authors compared the IDP based system with MDVP and MFCC parameters. The suggested scheme attains average accuracies of 99.20%, 93.20%, and 91.50% in disorder identification tests using the SVM classifier with recordings from MEEI, SVD, and AVPD, respectively. In [30], different time-based, frequency-based, and Liljencrants–Fant (LF) model-based attributes were mined from the glottal pulse-form applying the well-known Aalto Aparat voice inverse filtering and parameterization tool. The normal pitch utterance of the sustained vowel /a/ mined from German, English, Arabic, and Spanish voice records is used. The highest rates of voice disorder identification achieved are 99.80%, 99.70%, 99.80%, and 99.80% respectively, over SVD, PdA, AVPD, and MEEI. The best voice illness categorization rates attained are 90.80%, 99.30%, 99.80%, and 90.10% for SVD, PdA, AVPD, and MEEI, respectively. The LF model-based attributes have demonstrated good discrimination power compared to other attributes for the assessment of voice disfunction.
This paper proposes an acoustical assessment-based technique for assessing voice illnesses induced by a malfunctioning vocal fold(s). A multi-resolution transform, such as a stationary wavelet transform (SWT), can detect the voice produced due to disorder since it has a significant amplitude fluctuation of extremely low scale. Sub-band distribution can identify unusual spectrum distribution, speech non-periodicity, and unexpected energy variations in the speech signal [31]. This is how we came up with the idea for using SWT in our research.

2. Materials and Methods

2.1. Voice Disorder Databases

This article employs four distinct language voice disease records. Table 1 shows the number of voice signals in each database, and Table 2 shows the database characteristics. The speech records with sustained vowel/a/are mined from these databases. Since the sampling frequency of each database varies, all speech records were resampled to 22.05 KHz to create a consistent (unique) sampling frequency across all datasets.

2.2. Proposed Speech Pathology Assessment Technique

This paper details a SWT-based framework for assessing voice illnesses. The block diagram in Figure 1 depicts the distinct steps employed in the voice assessment system. The methodology comprises two main phases: (a) learning and (b) evaluation. During the learning stage, a 10-fold cross-validated classification model has been developed using healthful and pathological speech training recordings. In the evaluation phase, unseen speech records are tested on a cross-validated trained model. The normalization of the speech signal amplitude, framing, and windowing are the three main steps in pre-processing. The true voice source is then acquired from the pre-processed voice signal using the GIF method, which isolates the glottal pulse from the spoken signal [35]. In this study, an iterative adaptive inverse filtering (IAIF) technique is applied to obtain the glottal flow and is endorsed due to its robustness in noisy surroundings [36]. In IAIF, the vocal tract response is characterized by the discrete all-pole (DAP) model as it performs better compared to linear predictive coding (LPC) [37]. In an IAIF algorithm, the model of the acoustic tube is estimated in two phases and, hence, is named iterative inverse filtering. In the first step of IAIF, the input speech signal is pre-emphasized. The final refined glottal source is obtained by canceling the influence of lip radiation from the previous stage [38]. Figure 2 gives the illustration of the speech signal and voice source signal acquired using the IAIF. SWT is used to decompose the extracted glottal flow. To represent the voice source in a numerical form, energy and statistical metrics are obtained from each sub-band’s coefficients.
The information gain-centered feature selection method is used to pick the optimal subset of features and omit redundant and irrelevant features. The following section briefly describes the various statistical attributes and SWT. Before feeding the metrics into the classifier, min/max normalization is used to scale all descriptors. This procedure limits the dataset to the range [0, 1]. After normalizing the training data, the min and max values for each attribute of the training data are applied to standardize the test data.
For voice evaluation, two supervised classification algorithms are used: support vector machine (SVM) and stochastic gradient descent (SGD). The typical classifier quality measurement metrics, such as the area under the curve (AUC), classification accuracy (CA), precision (PPV), recall, and F1 score, are calculated. The suggested approach has been validated against four different databases.
Support vector machine (SVM)
An SVM is a linear two-class classifier, and its non-probabilistic facet is a key feature. An SVM partition sample along a decision boundary (plane) is obtained by just a small data subset. The subset of data that supports the decision boundary is adequately termed the support vector. The leftover training samples have no impact on the location of the plane in the feature space. However, the classification model is built in probabilistic classifiers by considering all of the data and thus requiring more computational resources [39].
Furthermore, the binary and linear facets are two constraints of the SVM. The latest developments using “Kernel Trick” handled the linearity constraint of the decision boundary. Additionally, the lack of ability to categorize the data in more than two categories seems to be a topic of current research. The techniques to date support the formation of various SVMs that compare input data to one another in a variety of means, such as one-to-all or all-to-all (the latter is also termed one-to-one) [39,40].
We are provided with training data of n samples in the form of: s 1 , t 1 , s n , t n , where t i is the class label and is either 1 or −1. The label specifies the class of each sample s i   to which it belongs and every s i is a real vector of the x-dimension. While learning, the SVM algorithm determines a decision border inside the input space that splits all data points in the two different classes. In this algorithm, the optimization task is the determination of the linear hyperplane that will have the maximal margin among the two categories. The SVM now employs this hyperplane to evaluate the category of new unseen test samples [40]. The hyperplanes are shown in Figure 3. Thus, the hyperplane splits the set of points s i for which label t i = 1 and t i = 1 . Any hyperplane may be given as a set of data points s i , satisfying the equation:
v . s k = 0 ,
where v is the vector perpendicular (normal vector) to the hyperplane and need not be a unit vector. The factor k v is used to calculate the offset of a hyperplane from the origin along the normal vector v . If the input data used to train the SVM are linearly distinguishable, we can select two parallel hyperplanes that split two data classes. Moreover, one has to ensure that the separation of them is as high as possible. The area constrained by two hyperplanes such as these is termed the “margin”, and the max-margin hyperplane is the hyperplane that exists midway between them. These hyperplanes can be characterized by equations with standardized input data [39].
For v . s k = 1 , any data points above this plane are of one kind of class with the label 1, and for v . s k = 1 , any data points above this plane are of one kind of class with the label −1. The spacing between these two planes is 2 v and to maximize the margin, we need to minimize the v . We also need to avoid data samples’ falling inside the margin, and this is achieved by adding a constraint: for every i, either
v . s i k 1 ,     if   t i = 1 ,   or   v . s i k 1 ,     if   t i = 1 ,
These restrictions assert that every data value will be fall on the right side of the margin. One can rewrite this as: t i v . s i k 1 ,   for all 1 i n .
To have the optimization problem, it can be put together as: “Minimize v subject to t i v . s i k 1 ,   for i = 1 , n ”. The v and k, which solve this optimization problem and will determine classifier, s s g n v . s k . An important consequence of this geometric description is that the max-margin hyperplane is completely determined by those s that lie nearest to it. These s i   are called support vectors [40].
Stochastic gradient descent (SGD)
The SGD is simply an optimization method, but it does not necessarily apply to a particular family of ML models. It is just a means to train a classifier model and is employed in ML for training a wide range of models. However, in this work, it is employed to train the linear SVM, where it plays the role of optimizer [41].
The term “batch” refers to the number of samples employed to compute the gradient in each iteration in gradient descents (GD) from a set of data. In classic GD optimization, such as Batch GD, the batch is drawn to be the entire dataset. Although it is very useful to use the entire dataset to obtain the minima in a more noise-free and far less random way, the trouble starts when our dataset is very big. If one is performing a classification task on two million data samples, then all two million samples must be used for one iteration when performing a GD optimization technique, and this will be done for each iteration till the minima are attained. Therefore, this optimization technique is computationally expensive [41]. This issue is resolved by SGD. In SGD, only one sample, i.e., one batch size, is used to execute each iteration. The training sample is randomly mixed and chosen for the iteration. The path followed by this algorithm is generally noisy when reaching the minima compared to classical GD, since in each iteration, a single sample is drawn randomly. However, this noisy path does not affect the performance as long as the algorithm achieves the minima with a considerably smaller training time. The paths followed in GD and SGD are shown in Figure 4. Further, it normally takes relatively high iterations to achieve the minima, due to the random descent. Although it takes a higher number of iterations to achieve the minima than traditional GD, it is still less of a computational burden. Therefore, in most cases, SGD is preferred over traditional GD to optimize the ML algorithm [41]. Let us look at the problem of objective function minimization in the form:
P v = 1 n m = 1 n P m v
where v is the parameter that minimizes P v .
Every summand function P m corresponds to the mth observation of the set of data considered for the training of the model. If the normal GD is employed to minimize the abovementioned function, then it will execute iterations as [42]:
v : = v η P v = v η n m = 1 n P m v
where the parameter η is the learning rate.
In SGD, the real gradient of P v is approximated by a gradient at the individual observation:
v : = v η P m v
Figure 4. Path followed in GD and SGD to attain minima [43].
Figure 4. Path followed in GD and SGD to attain minima [43].
Diagnostics 12 02758 g004
When the algorithm runs on through the training data, it conducts the update above for every observation of training. You can make multiple runs over all the training dataset before the algorithm converges. When that is done, the training data (samples) can be shuffled to avoid cycles for each run. Classic frameworks can utilize an adaptive learning rate to converge the algorithm. The SGD may be presented in pseudo code as follows [43]:
  • Select an initial vector of parameters v and step size (learning rate) η
  • Repeat it until an approximate minimum is obtained:
    Shuffle the observations in the training dataset randomly.
    For m = 1 , 2 , 3 ,   , n , do:
    v : = v η P m v .

2.2.1. Stationary Wavelet Transform

In the study of transient phenomena, wavelet transform is widely used to obtain temporal and frequency information from the signal [44,45]. A four-level decomposition of the signal employing the SWT is shown in Figure 5. SWT generates two sets of coefficients for a signal s[x] of length L in the first stage: approximation coefficients A1 and detail coefficients D1. Convoluting s with the low-pass filter (LF_D) for the approximation component and the high-pass filter (HF_D) for the detail component yields these coefficients. Using a similar strategy, the approximation coefficients A1 are separated into two pieces in the next level. However, in this stage, up-sampling is performed, and s[x] is substituted by A1, resulting in A2 and D2. In this research, the glottal wave is decomposed up to level four. The Haar wavelet is used, as the energy and statistical metrics mined from the sub-bands have better discrimination ability compared to higher-order Daubechies wavelets.

2.2.2. SWT Sub-Band Features

As stated earlier, the presented scheme employs wavelet-based energy and statistical attributes extraction. Following SWT decomposition, every sub-band yields a total of seven features. The following is a concise description of the different attributes employed:
1.
Energy: the energy of each is calculated using:
E n e r g y = k = 1 L W S B i 2
The sub-band is denoted by WSBi, and the number of coefficients is denoted by L.
2.
Spectral entropy: the sub-band entropy is calculated as follows:
E n t r o p y = k = 1 L W S B i l o g ( W S B i )
3.
Mean (μ): the mean of both coefficients of the sub-band is calculated as:
M e a n = 1 L k = 1 L W S B i
4.
Variance: the variance of both coefficients of the sub-band is calculated as:
V a r i a n c e = 1 L 1 k = 1 L W S B i μ 2
5.
Standard deviation: this is calculated as:
S t d _ d e v = 1 L 1 k = 1 L W S B i μ 2
where μ denotes the mean of the sub-band.
6.
Kurtosis: this is evaluated as:
K u r t o s i s = E ( W S B i μ ) 4 σ 4
where E(i) denotes the expected value.
7.
Skewness: skewness computes the asymmetry of the data in relation to the sample mean and is estimated as:
S k e w n e s s = E ( W S B i μ ) 3 σ 3

2.2.3. Evaluation of Discernment Potential of Attributes

It is extremely important to test the discernment ability of the descriptors before applying to a classifier for a decision to be made. It aids in selecting the best features and removing those that are unwarranted. Subjective and objective methods are used in this study. The probability density plots (PDF) were used to subjectively test the discernment ability of all of the descriptors. The spacing between the peaks in the PDF plots were visually inspected for all four datasets. The greater the distance between the peaks, the greater the discrimination power of the descriptor. For the SVD dataset, it is observed that the spacing between the peaks of the PDF plots of healthy and unhealthy classes are the maximum for the Mean-D3 descriptor, followed by Mean-D4 and Mean-D1. For the PdA dataset, it is observed that the spacing between the peaks of the PDF plots of healthy and unhealthy classes are the maximum for the Mean-D2 descriptor, followed by Mean-D3 and Mean-D1. For the AVPD dataset, it is observed that the spacing between the peaks of the PDF plots of healthy and unhealthy classes are the maximum for the Mean-D3 descriptor, followed by Mean-D4 and Mean-D2. For the MEEI dataset, it is observed that the spacing between the peaks of the PDF plots of healthy and unhealthy classes are the maximum for the Pentropy-A4 descriptor, followed by Mean-D4 and Mean-D2. Figure 6 shows the density plot for these top three descriptors. The density plots of the Mean-D3 feature for SVD dataset is shown in Figure 6A, where Mean-D3 signifies the mean of the detail coefficients of the sub-band after the third decomposition level. One more example, as the PDF plots of the Mean-D2 feature for the PdA dataset, is shown in Figure 6D, where Mean-D2 signifies the mean of the detail coefficients of the sub-band after the second decomposition level. The Mean-D1 and Mean-D4 descriptors represent the mean of the detail coefficients of the sub-band after the first and fourth decomposition levels, respectively. A similar procedure was used to assess the discernment ability of the descriptors for the classification of speech disorders. Figure 7 shows the density plots for the top three features in each dataset for the classification of speech disorders. The density plots for the four classes are plotted for the PdA and AVPD datasets, three of which are pathologic and the fourth is healthful. However, the graphs for three groups are shown in the SVD and MEEI databases, two of which are diseased and the third one is healthful. The quantity of cyst recordings in the SVD and MEEI databases is minimal and of short duration, so only three categories were investigated. Each kind of disorder yields distinct hills with far-flung peaks, implying high discrimination capacity.
The Information Gain (IG) feature-scoring method is used to objectively assess all 56 attributes’ discriminating abilities. Later, the attributes are placed in descending order based on their discernment power. Table 3a,b shows the top 10 attributes with their IG values across each dataset for voice dysfunction identification and classification, respectively. The size of the feature vector depends on the decomposition level. The dimension of the feature vector grows in proportion to the decomposition level. Hence, to reduce the dimension of the feature vector, the feature-selection method is harnessed. It aids in the selection of the best features while eliminating the ones that are redundant. By deleting unnecessary characteristics, feature selection approaches pick a small group of characteristics that boosts the classification rate.

3. Results

Two supervised classifiers, SVM and SGD, are used to evaluate vocal abnormalities. The radial basis function (RBF) kernel is utilized in the SVM classifier, and parameter C is set at 1. The hinge as a classification loss function, squared loss as a regression loss function, and ridge (L2) as a regularization approach were chosen as the various parameters for the SGD, also known as linear SVM. A stratiform k-fold cross-validation resampling procedure is more preferable than general cross-validation in terms of bias and variance, so it is used to develop machine learning models for the assessment of voice abnormalities. On four separate datasets, many inter- and intra (cross)-dataset tests were undertaken to determine the utility of the extracted descriptors for voice disease identification and categorization. The acquired glottal wave is dissected up to level 4, and seven descriptors are mined from each sub-band’s approximation as well as detailed coefficients ((7_approximation coefficients + 7_detailed coefficients) × 4 = 56).

3.1. Intra-Dataset Voice Pathology Detection Experiments

Two experiments are carried out to validate the efficacy of extracted wavelet-based features for voice disfunction recognition, one for each of the top three ranked descriptors and one for all 56 descriptors. Table 4 displays the calculated performance-assessing indicators of the classifiers. The observation that the detection rates for the top three and all 56 features are nearly equal leads to the conclusion that only the top three descriptors are vital to detect voice pathology. The suggested system’s computational complexity decreases because of the fewer descriptors and smaller feature vectors. In addition, these results reveal that the features are not noisy, as increasing the number of descriptors has no negative influence on the detection accuracy. The values of classifier performance metrics look identical in many tables because of the very small number of false positives (FP) and false negatives (FN). The scatter plot for these top three descriptors is shown in Figure 8 and clearly shows the power of discrimination as the clusters of healthy and pathological data points (samples) do not overlap. The outliers are not removed; hence, some data points become mixed into other class data points.

3.2. Voice Disorder Categorization

Using the information gain (IG) feature ranking method, 12 experiments with distinct sets of selected descriptors are conducted in each database for multiclass classification. These experiments are being carried out to see whether the derived sub-band descriptors can distinguish between healthy people and those with a clinical condition of a vocal cord(s) cyst, paralysis, or polyp. The outcomes of these tests are presented in Table 5, which unmistakably demonstrates that the categorization accuracy (CA) rises as the number of descriptors increases. It assures that the features are not noisy and do not impact the categorization accuracy negatively.
The classification rate with all 56 features and the top five features is nearly identical across all datasets. As a result, it is shown that less than 9% of the entire number of descriptors is sufficient to achieve an almost equal recognition rate. The suggested system’s computational complexity decreases because of the fewer descriptors and smaller feature vectors. The highest accuracy obtained with the AVPD database using the top five descriptors was 99.87%. In the case of the SVD, AVPD, and MEEI databases, the top three descriptors performed extraordinarily. The classification accuracy attained with these top three features is nearly identical to that achieved only with the top five and all 56 features. Figure 3 shows the PDF plots of the top three descriptors, which are not overlapping and endorse the high discrimination capacity.

3.3. Dataset-Independent Voice Illness Assessment System

A key objective of this study is to propose a voice pathology identification and classifying framework that is not affected by language, accent, age group, ethnic background, sex, or other factors. There is no single (particular) feature that can provide a high identification and classifying rate across all datasets. Therefore, three common features from the top five features are tactfully picked from all datasets. The chosen three descriptors are Mean-D2, Mean-D3, and Mean-D4. This makes the proposed dataset independent, and as a result, the method is not affected by factors such as sex, age, culture, ethnicity, or language.
To ascertain the discrimination power of these three selected common descriptors, scatter plots and PDF plots are plotted. Figure 9 shows the scatter plots of these descriptors for the SVD, PdA, AVPD, and MEEI datasets for the pathology identification process. In the scatter plot, hardly any of the samples are mixed, indicating that the three descriptors chosen have excellent discrimination ability to separate pathological samples. The quality metrics of the classifiers for speech condition identification with the three most common features are presented in Table 6, and the maximum recognition rate attained is 99.97% for the Arabic dataset. Figure 10 shows the density graphs for the three most common descriptors for speech condition classification. As there are few cyst pathology samples in the SVD and MEEI datasets, three groups—healthy, paralyzed, and polyp—are analyzed. The AVPD and PdA datasets are classified into four categories: normal, cystic, paralyzed, and polyp. The classifiers’ key parameters for categorization are presented in Table 7, and the AVPD dataset attained the best categorization accuracy of 99.87%.

3.4. Voice Disorder Detection (Inter-Database)

Inter-database tests are carried out as an extra task to examine the discriminative power of features for voice pathology identification and to ascertain the proposed voice evaluation system’s independence from language, accent, age, social background, sex, and so on. This set of experiments aided in determining how well a voice pathology identification system trained by one dataset would discriminate speech recordings from another dataset. In the inter (cross)-dataset type, 14 experiments are performed using same three most common features, namely, Mean-D2, Mean-D3, and Mean-D4, considered in Section 3.3. The classifier was trained on one of the datasets for the first four tests, and the learned model was tested on the other three datasets individually. For experiments five through ten, all conceivable pairings of two datasets were employed for learning, and the learned model was tested using the leftover two datasets individually. In the last three experiments, every possible combination of the three datasets was used to train the classifier; however, the leftover dataset was used to assess the educated model. The detection rate of these 14 experiments is shown in Table 8 and reveals that the proposed approach is unbiased in terms of spoken language, age, social background, accent, sex, etc.

3.5. Voice Pathology Independent System

One more goal of this study is to propose an independent voice pathology system. To ascertain this, extra tests are carried out using variety of pathologies. Up to Section 3.4, all of the experiments are performed using healthy, cyst, paralysis, and polyp speech samples. In the SVD, PdA, and MEEI databases, laryngitis and Reinke’s edema are the most common pathologies, along with paralysis, polyp, and cyst. So, these five pathologies were considered with the healthy speech samples to perform voice identification experiments. The descriptor rankings obtained with these five pathologies are given in Table 9. The descriptor rankings for voice disorder identification shown in Table 3 and Table 9 do not change much. In addition, the descriptors Mean-D1, Mean-D2, Mean-D3, and Mean-D4 have retained their position among the top eight descriptors. The classifier performance measures of this experiment are shown in Table 10. To confirm that the system is independent of voice pathologies, the same three most common descriptors Mean-D2, Mean-D3, and Mean-D4 mentioned in Section 3.3 are used. The highest detection rate obtained is more than 99% in all databases by using only the three most common descriptors. The scatter plots are shown in Figure 11 to justify the high detection rate achieved. It is clear from Figure 11 that these three common descriptors are highly discriminative.
The PdA databank covers 15 organic and traumatic etiologies and the AVPD database consists of five organic speech pathology samples. All speech samples available in the PdA and AVPD databases are also considered in the next voice disorder identification experiment. This experiment consists of wide variety of voice pathology samples. The top 10 ranked descriptors are shown in Table 11, and it is found that these are almost the same as those listed in Table 3 and Table 9. In addition, the Mean-D1, Mean-D2, Mean-D3, and Mean-D4 descriptors are at rank 1, rank 2, rank 3, and rank 4, respectively. Thus, although the voice pathologies increased compared to earlier experiments, the most discriminatory descriptors remain the same. The classifier performance measures of this experiment are shown in Table 12. Here, two separate experiments are performed with the three most common (Mean-D2, Mean-D3, and Mean-D4) descriptors and all 56 descriptors. Again, the three most common descriptors perform well, and the highest detection rate obtained is 99.99%. The scatter plots for these three common descriptors are shown in Figure 12 and show that the healthy and unhealthy classes’ data points are non-overlapping. The outcome of this experiment ensures that the proposed system is independent of voice pathologies.
Further, two multiclass experiments are performed with PdA and AVPD dataset speech samples. In the PdA dataset multiclass experiment, healthy and seven pathological classes—cyst, paralysis, polyp, laryngitis, Reinke’s edema, nodules, and sulcus—are used. However, in the AVPD dataset multiclass experiment, healthy and five organic pathological classes—cyst, paralysis, polyp, nodules, and sulcus—are used. The top 10 descriptor rankings obtained are shown in Table 13 and are almost identical to the descriptor rankings shown in Table 3. Thus, though the pathologies increased, the ranking of the descriptors does not change much, and the Mean-D1, Mean-D2, Mean-D3, and Mean-D4 descriptors have retained their position in the top 4. The results of these experiments are shown in Table 14 and it is observed that the accuracy obtained with the top 10 and all 56 descriptors is nearly the same as using the SGD classifier. The performance of the system using the three most common descriptors is not as good for the PdA database, as the number of classes in the multiclass experiments is higher compared to the experiments performed in Section 3.2. However, in the AVPD database, the classification rate achieved with the three common (Mean-D2, Mean-D3, and Mean-D4) features is almost identical. Thus, the proposed framework can be used for the assessment of almost any voice pathology that affects the vibratory pattern of vocal folds. Moreover, to perform voice disorder identification or classification, the three most common (Mean-D2, Mean-D3, and Mean-D4) features are adequate to attain sufficient accuracy.

3.6. Impact of Wavelet Decomposition Level on the Accuracy of Detecting and Categorization

The feature vector’s dimension is determined by the decomposition level. As the decomposition level grows, so does the dimensionality of the feature vector, increasing the system’s computational complexity and response time. As a result, determining the optimal level of decomposition is required to obtain satisfactory detection and categorization accuracy. Table 15 depicts the impact of the decomposition level on the detection rate and reveals that the detection rate is nearly same even when the decomposition level is increased. Table 16 portrays the impact of the decomposition level on the classifying accuracy, revealing that level 4 achieves the highest classification accuracy. The glottal waveform is decomposed to level 4 in this study because the suggested system achieves both the desired identification and classification accuracy. Table 17 compares the accuracies achieved in this study to those stated in the existing state-of-the-art literature.

4. Discussion

This study uses the well-known IAIF method to acquire the glottal wave, which is then dissected using SWT up to level 4. The obtained glottal wave is then numerically quantified using all of the attributes that were mined from each sub-band. The usefulness of extracted features for the identification and categorization of vocal disorders is evaluated using a variety of intra- and inter-database tests. The statistical features have demonstrated outstanding performance for assessing voice diseases. The findings reveal that the descriptors behaved differently for each database, resulting in variations in detection and classification accuracy. The following factors could account for the minor variance in accuracy and other metrics.
  • The voice-acquiring equipment and acoustic settings in each corpus were different.
  • The sampled frequency varies by dataset. This has an impact on the accuracy and reliability of a voice analysis.
  • In the MEEI registry, healthy people were not clinically evaluated. As a result, there is no way of knowing whether these people were truly healthy.
  • Although the recordings are altered to obtain just the stable section of the phonation, the research has revealed that the start and end of the vocalization have more acoustic clues than the steady section.
  • In the German and English datasets, many audios are labeled with multiple diseases.
  • For certain disorders, there are only a few recordings available.
  • The severity of pathology is different in each dataset
  • The distribution of the subject’s age and sex varies by dataset.
  • The SVD dataset does not include relevant information on paralysis, whether unilateral or bilateral.
  • The PdA dataset includes audio files with strong background noise or barely discernible vocalizations.
The wavelet transform has the power to extract spatial–frequency information from a non-stationary signal, and since disordered speech is transient in nature, we used the orthogonal Haar wavelet for multilevel decomposition. We examined Daubechies wavelets over four datasets to validate the use of the Haar wavelet. Table 18 and Table 19 show the calculated detection and classification rates, and Haar outperforms Daubechies wavelets in terms of the classification rate, justifying our choice of Haar wavelets for decomposition. When it comes to detecting voice disorders, both Haar and Daubechies have been proven effective. To confirm that the classifier model developed is independent of the database, it is tested with three more databases other than the database by which it was trained. This set of experiments helped to investigate how efficiently a voice pathology detection system that is educated by one database can distinguish samples from the other database. However, the results were not encouraging. To bring uniformity to the application of the system to various databases, various common descriptors of voice pathologies are identified. The proposed system for identifying and classifying voice disorders is therefore evaluated based on three common features collected from all databases. In terms of the identification and classification accuracy, 99.9% has been attained so far.

5. Conclusions

The variation in speech quality due to voice pathology is directly linked to the real biological source of voiced excitations emerging from the glottis. This source of voiced excitations is obtained by IAIF, which gives better insight into the cause of changes in speech quality. For the modeling of the vocal tract, a discrete all-pole model was used in inverse filtering as it performs better compared to LPC in a high-pitched voice. This study proposes a novel voice dysfunction diagnostic and categorization system based on SWT. Multilevel SWT decomposition has been used to detect variation in speech quality, and energy as well as statistical parameters were derived from each sub-band to qualify the glottal wave. The derived parameters are independent of the fundamental frequency and were found to be very successful in distinguishing healthful and disordered speech. The Mean-D1, Mean-D2, Mean-D3, and Mean-D4 descriptors extracted from the detailed coefficients of decomposition level 1, level 2, level 3, and level 4 have very good discrimination power compared to other descriptors for speech dysfunction assessment. The proposed system achieved an average recognition rate of 99.99% and 99.60% for detection and classifying, respectively. The SGD classification model outperformed the SVM classifier, signifying that the extracted features are linearly separable.
The purpose of this study is to create a voice condition assessment system that is independent of voice pathologies as well as the language, accent, age, cultural background, and gender of the speaker by utilizing three common descriptors from the top ten descriptors.
Additionally, this article presents an automated voice disease diagnostic algorithm that is recommended for a wide variety of voice diseases. Our proposed system is among a few speech impairment diagnostic systems that have identified a broad range of laryngeal pathologies. This framework can be introduced as an android app that can be used to check the voice quality in daily life at any instant. The conceived laryngeal pathology evaluation framework can be used to diagnose voice concerns objectively and without invasion. It can be used as a supplemental aid to track the degree of recovery both during and after the cycle of voice therapy.

Author Contributions

Conceptualization, A.S. and S.D.; methodology, A.S., G.G. and S.D.; software, A.S. and G.G.; validation, S.S., G.G. and K.E.; formal analysis, A.S., G.G. and M.F.; investigation, A.S. and G.G.; resources, A.D.O. and K.D.; data curation, A.S., G.G. and A.D.O.; writing—original draft preparation, A.S. and G.G.; writing—review and editing, S.S., G.G., D.D.B.-N., D.P.B.-N. and S.D.; visualization, A.S.; supervision, S.D., J.N. and M.F.; project administration, S.D., J.N. and M.F.; funding acquisition, D.D.B.-N. and D.P.B.-N. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Gheorghe Asachi Technical University of Iaşi—TUIASI-Romania, Scientific Research Funds, FCSU-2022.

Institutional Review Board Statement

This study did not require ethical approval.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Verdolini, K.; Ramig, L. Review: Occupational risks for voice problems. Logoped. Phoniatr. Vocol. 2001, 26, 37–46. [Google Scholar] [CrossRef] [PubMed]
  2. Nemr, K.; Marcia Simões-Zenari, G.F.; Cordeiro, D.; Tsuji, A.I.; Ogawa, M.; Menezes, M. GRBAS and Cape-V scales: High reliability and consensus when applied at different times. J. Voice 2012, 26, 812-e17. [Google Scholar] [CrossRef] [PubMed]
  3. Arjmandi, M.; Pooyan, M. An optimum algorithm in pathological voice quality assessment using wavelet-packet-based features, linear discriminant analysis and support vector machine. Biomed. Signal Process. Control 2012, 7, 3–19. [Google Scholar] [CrossRef]
  4. Rosa, M.; Pereira, J.; Grellet, M. Adaptive estimation of residue signal for voice pathology diagnosis. IEEE. Trans. Biomed. Eng. 2000, 47, 96–104. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Narendra, N.P.; Alku, P. Glottal source information for pathological voice detection. IEEE Access 2020, 8, 67745–67755. [Google Scholar] [CrossRef]
  6. Kadiri, S.R.; Alku, P. Analysis and detection of pathological voice using glottal source features. IEEE J. Sel. Top. Signal Process. 2020, 14, 367–379. [Google Scholar] [CrossRef] [Green Version]
  7. Reddy, M.K.; Alku, P. A Comparison of Cepstral Features in the Detection of Pathological Voices by Varying the Input and Filterbank of the Cepstrum Computation. IEEE Access 2021, 9, 135953–135963. [Google Scholar] [CrossRef]
  8. Hegde, S.; Shetty, S.; Rai, S.; Dodderi, T. A survey on machine learning approaches for automatic detection of voice. J. Voice 2019, 33, 947-e11. [Google Scholar] [CrossRef]
  9. Parsa, V.; Jamieson, D. Acoustic discrimination of pathological voice. J. Speech Lang. Hear. Res. 2001, 44, 327–339. [Google Scholar] [CrossRef]
  10. Zhang, Y.; Jiang, J.; Biazzo, L.; Jorgensen, M. Perturbation and nonlinear dynamic analyses of voices from patients with unilateral laryngeal paralysis. J. Voice 2005, 19, 519–528. [Google Scholar] [CrossRef]
  11. Michaelis, D.; Gramss, T.; Strube, H. Glottal-to-noise excitation ratio a new measure for describing pathological voices. Acta Acust. United Acust. 1997, 83, 700–706. [Google Scholar]
  12. Manfredi, C.; D’Aniello, M.; Bruscaglioni, P.; Ismaelli, A. A comparative analysis of fundamental frequency estimation methods with application to pathological voices. Med. Eng. Phys. 2000, 22, 135–147. [Google Scholar] [CrossRef]
  13. Watts, C.; Awan, S. Use of spectral/cepstral analyses for differentiating normal from hypofunctional voices in sustained vowel and continuous speech contexts. J. Speech Lang. Hear. Res. 2011, 54, 1525–1537. [Google Scholar] [CrossRef]
  14. Patel, R.R.; Awan, S.N.; Barkmeier-Kraemer, J.; Courey, M.; Deliyski, D.; Eadie, T.; Paul, D.; Švec, J.G.; Hillman, R. Recommended protocols for instrumental assessment of voice: American speech-language hearing association expert panel to develop a protocol for instrumental assessment of vocal function. Am. J. Speech Lang. Pathol. 2018, 27, 887–905. [Google Scholar] [CrossRef] [PubMed]
  15. Benba, A.; Jilbab, A.; Hammouch, A. Discriminating between patients with Parkinson’s and neurological diseases using cepstral analysis. IEEE Trans. Neural Syst. Rehabil. Eng. 2016, 24, 1100–1108. [Google Scholar] [CrossRef]
  16. Ali, Z.; Elamvazuthi, I.; Alsulaiman, M.; Muhammad, M. Automatic voice pathology detection with running speech by using estimation of auditory spectrum and cepstral coefficients based on the all-pole model. J. Voice 2016, 30, 757-e7. [Google Scholar] [CrossRef]
  17. Godino-Llorente, J.; Gomez-Vilda, P. Automatic detection of voice impairments by means of short-term cepstral parameters and neural network based detectors. IEEE. Trans. Biomed. Eng. 2004, 51, 380–384. [Google Scholar] [CrossRef]
  18. Godino-Llorente, J.; Vilda, P.; Blanco-Velasco, M. Dimensionality reduction of a pathological voice quality assessment system based on Gaussian mixture models and short-term cepstral parameters. IEEE. Trans. Biomed. Eng. 2006, 53, 1943–1953. [Google Scholar] [CrossRef]
  19. Sáenz-Lechón, N.; Godino-Llorente, J.; Osma-Ruiz, V.; Gómez-Vilda, P. Methodological issues in the development of automatic systems for voice pathology detection. Biomed. Signal Process. Control 2006, 1, 120–128. [Google Scholar] [CrossRef]
  20. Godino-Llorente, J.; Fraile, R.; Sáenz-Lechón, N.; Osma-Ruiz, V.; Gómez-Vilda, P. Automatic detection of voice impairments from text-dependent running speech. Biomed. Signal Process. Control 2009, 4, 176–182. [Google Scholar] [CrossRef]
  21. Ali, Z.; Alsulaiman, M.; Muhammad, G.; Elamvazuthi, I.; Al-Nasheri, A.; Mesallam, T.M.; Farahat, M.; Malki, K.H. Intra-and inter-database study for Arabic, English, and German databases: Do conventional speech features detect voice pathology? J. Voice 2017, 31, 386-e1. [Google Scholar] [CrossRef] [PubMed]
  22. Arias-Londoño, J.; Godino-Llorente, J. Entropies from Markov models as complexity measures of embedded attractors. Entropy 2015, 17, 3595–3620. [Google Scholar] [CrossRef] [Green Version]
  23. Titze, I. The Myoelastic Aerodynamic Theory of Phonation; National Center for Voice and Speech: Salt Lake City, UT, USA, 2006. [Google Scholar]
  24. Zhang, Y.; Jiang, J. Nonlinear dynamics analysis in signal typing of pathological human voices. Electron. Lett. 2003, 39, 1021–1023. [Google Scholar] [CrossRef]
  25. Zhang, Y.; Jiang, J. Acoustic analyses of sustained and running voices from patients with laryngeal pathologies. J. Voice 2008, 22, 1–9. [Google Scholar] [CrossRef]
  26. Little, M.; McSharry, P.; Roberts, S.; Costello, D.; Moroz, I. Exploiting nonlinear recurrence and fractal scaling properties for voice disorder detection. Biomed. Eng. 2007, 6, 1–23. [Google Scholar]
  27. Arias-Londoño, J.; Godino-Llorente, J.; Sáenz-Lechón, N.; Osma-Ruiz, V.; Castellanos-Domínguez, G. Automatic detection of pathological voices using complexity measures, noise parameters, and melcepstral coefficients. IEEE. Trans. Biomed. Eng. 2011, 58, 370–379. [Google Scholar] [CrossRef] [PubMed]
  28. Forero, L.; Kohler, M.; Vellasco, M.; Cataldo, E. Analysis and classification of voice pathologies using glottal signal parameters. J. Voice. 2016, 30, 549–556. [Google Scholar] [CrossRef]
  29. Muhammad, G.; Alsulaiman, M.; Ali, Z.; Mesallam, T.A.; Farahat, M.; Malki, K.H.; Al-nasheri, A.; Bencherifa, M.A. Voice pathology detection using interlaced derivative pattern on glottal source excitation. Biomed. Signal Process. Control 2017, 31, 156–164. [Google Scholar] [CrossRef]
  30. Gidaye, G.; Nirmal, J.; Ezzine, K.; Shrivas, A.; Frikha, M. Application of glottal flow descriptors for pathological voice diagnosis. Int. J. Speech Technol. 2020, 23, 205–222. [Google Scholar] [CrossRef]
  31. Farouk, M. Clinical Diagnosis and Assessment of Speech Pathology, 1st ed.; Springer International Publishing: Cham, Switzerland, 2018. [Google Scholar]
  32. Saarbrucken Voice Database. Available online: http://www.Stimmdatenbank.coli.uni-saarland.de (accessed on 2 June 2021).
  33. Mesallam, T.A.; Farahat, M.; Malki, K.H.; Alsulaiman, M.; Ali, Z.; Al-Nasheri, A.; Muhammad, G. Development of the Arabic voice pathology database and its evaluation by using speech features and machine learning algorithms. J. Healthc. Eng. 2017, 2017, 1–13. [Google Scholar] [CrossRef] [Green Version]
  34. MEEI: Disordered Voice Database; Version 1.03 (CD-ROM); Voice and Speech Lab, Kay Elemetrics Corp.: Lincoln Park, NJ, USA, 1994.
  35. Drugman, T.; Bozkurt, B.; Dutoit, T. A comparative study of glottal source estimation techniques. Comput. Speech Lang. 2012, 26, 20–34. [Google Scholar] [CrossRef] [Green Version]
  36. El-Jaroudi, A.; Makhoul, J. Discrete all-pole modelling. IEEE Trans. Signal Process. 1991, 39, 411–423. [Google Scholar] [CrossRef]
  37. Alku, P. Glottal wave analysis with pitch synchronous iterative adaptive inverse filtering. Speech Commun. 1992, 11, 109–118. [Google Scholar] [CrossRef]
  38. Airas, M. TKK Aparat: An environment for voice inverse filtering and parameterization. Logoped. Phoniatr. Vocol. 2008, 33, 49–64. [Google Scholar] [CrossRef]
  39. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  40. Ben-Hur, A.; Horn, D.; Siegelmann, H.; Vapnik, V. Support vector clustering. J. Mach. Learn. Res. 2001, 2, 125–137. [Google Scholar] [CrossRef]
  41. Ferguson, T. An inconsistent maximum likelihood estimate. J. Am. Stat Assoc. 1982, 77, 831–834. [Google Scholar] [CrossRef]
  42. Xing, C.; Arpit, D.; Tsirigotis, C.; Bengio, Y. A walk with SGD. arXiv 2018, arXiv:1802.08770. [Google Scholar]
  43. Kohavi, R. A study of cross-validation and bootstrap for accuracy estimation and model selection. In Proceedings of the Fourteenth International Joint Conference on Artificial Intelligence, San Mateo, CA, USA, 20–25 August 1995; pp. 1137–1143. [Google Scholar]
  44. Sreehari, V.; Mary, L. Automatic speaker recognition using stationary wavelet coefficients of lp residual. TENCON 2018. In Proceedings of the IEEE Region 10 Conference, Jeju, Korea, 28–31 October 2018. [Google Scholar]
  45. Nongpiur, R.; Shpak, D. Impulse-noise suppression in speech using the stationary wavelet transform. J. Acoust. Soc. Am. 2013, 133, 866–879. [Google Scholar] [CrossRef] [Green Version]
  46. Al-Nasheri, A.; Muhammad, G.; Alsulaiman, M.; Ali, Z.; Malki, K.; Mesallam, T.; Ibrahim, M. Voice pathology detection and classification using auto-correlation and entropy features in different frequency regions. IEEE Access. 2018, 6, 6961–6974. [Google Scholar] [CrossRef]
  47. Al-Nasheri, A.; Muhammad, G.; Alsulaiman, M.; Ali, Z. Investigation of Voice Pathology Detection and Classification on Different Frequency Regions Using Correlation Functions. J. Voice 2017, 31, 3–15. [Google Scholar] [CrossRef] [PubMed]
  48. Al-Nasheri, A.; Muhammad, G.; Alsulaiman, M.; Ali, Z.; Mesallam, T.A.; Farahat, M.; Malki, K.H. Bencherif, M.A. An investigation of multidimensional voice program parameters in three different databases for voice pathology detection and classification. J. Voice 2016, 31, 113-e9. [Google Scholar] [PubMed]
  49. Akbari, A.; Arjmandi, M. Employing linear prediction residual signal of wavelet sub-bands in automatic detection of laryngeal pathology. Biomed. Signal Process. Control 2015, 18, 293–302. [Google Scholar] [CrossRef]
Figure 1. Voice disorder assessment system.
Figure 1. Voice disorder assessment system.
Diagnostics 12 02758 g001
Figure 2. Illustration of speech signal and voice source signal acquired using the IAIF: (A) recorded healthy speech signal, (B) glottal flow signal of healthy speech, (C) recorded pathological speech signal, and (D) glottal flow signal of pathological speech.
Figure 2. Illustration of speech signal and voice source signal acquired using the IAIF: (A) recorded healthy speech signal, (B) glottal flow signal of healthy speech, (C) recorded pathological speech signal, and (D) glottal flow signal of pathological speech.
Diagnostics 12 02758 g002
Figure 3. Hyperplanes in SVM for binary classification [39].
Figure 3. Hyperplanes in SVM for binary classification [39].
Diagnostics 12 02758 g003
Figure 5. A four-stage decomposition.
Figure 5. A four-stage decomposition.
Diagnostics 12 02758 g005
Figure 6. Density plots for the top three descriptors used in the diagnosis of voice illnesses: (A) Density plot for Mean-D3 parameter of SVD, (B) Density plot for Mean-D4 parameter of SVD, (C) Density plot for Mean-D1 parameter of SVD, (D) Density plot for Mean-D2 parameter of PdA, (E) Density plot for Mean-D3 parameter of PdA, (F) Density plot for Mean-D1 parameter of PdA, (G) Density plot for Mean-D3 parameter of AVPD, (H) Density plot for Mean-D4 parameter of AVPD, (I) Density plot for Mean-D2 parameter of AVPD, (J) Density plot for Pentropy-A4 parameter of MEEI, (K) Density plot for Mean-D4 parameter of MEEI, (L) Density plot for Mean-D2 parameter of MEEI.
Figure 6. Density plots for the top three descriptors used in the diagnosis of voice illnesses: (A) Density plot for Mean-D3 parameter of SVD, (B) Density plot for Mean-D4 parameter of SVD, (C) Density plot for Mean-D1 parameter of SVD, (D) Density plot for Mean-D2 parameter of PdA, (E) Density plot for Mean-D3 parameter of PdA, (F) Density plot for Mean-D1 parameter of PdA, (G) Density plot for Mean-D3 parameter of AVPD, (H) Density plot for Mean-D4 parameter of AVPD, (I) Density plot for Mean-D2 parameter of AVPD, (J) Density plot for Pentropy-A4 parameter of MEEI, (K) Density plot for Mean-D4 parameter of MEEI, (L) Density plot for Mean-D2 parameter of MEEI.
Diagnostics 12 02758 g006
Figure 7. Density plots for top three descriptors for multiclass classification process: (A) Density plot for Mean-D4 parameter of SVD, (B) Density plot for Mean-D3 parameter of SVD, (C) Density plot for Mean-D2 parameter of SVD, (D) Density plot for Mean-D2 parameter of PdA, (E) Density plot for Mean-D3 parameter of PdA, (F) Density plot for Mean-D1 parameter of PdA, (G) Density plot for Mean-D2 parameter of AVPD, (H) Density plot for Mean-D4 parameter of AVPD, (I) Density plot for Mean-D1 parameter of AVPD, (J) Density plot for Mean-D1 parameter of MEEI, (K) Density plot for Mean-D2 parameter of MEEI, (L) Density plot for Mean-D3 parameter of MEEI.
Figure 7. Density plots for top three descriptors for multiclass classification process: (A) Density plot for Mean-D4 parameter of SVD, (B) Density plot for Mean-D3 parameter of SVD, (C) Density plot for Mean-D2 parameter of SVD, (D) Density plot for Mean-D2 parameter of PdA, (E) Density plot for Mean-D3 parameter of PdA, (F) Density plot for Mean-D1 parameter of PdA, (G) Density plot for Mean-D2 parameter of AVPD, (H) Density plot for Mean-D4 parameter of AVPD, (I) Density plot for Mean-D1 parameter of AVPD, (J) Density plot for Mean-D1 parameter of MEEI, (K) Density plot for Mean-D2 parameter of MEEI, (L) Density plot for Mean-D3 parameter of MEEI.
Diagnostics 12 02758 g007aDiagnostics 12 02758 g007b
Figure 8. Three-dimensional scatter plots for each database’s top three descriptors: (A) Three-dimensional scatter plot of top three descriptors of SVD, (B) Three-dimensional scatter plot of top three descriptors of PdA, (C) Three-dimensional scatter plot of top three descriptors of AVPD, (D) Three-dimensional scatter plot of top three descriptors of MEEI.
Figure 8. Three-dimensional scatter plots for each database’s top three descriptors: (A) Three-dimensional scatter plot of top three descriptors of SVD, (B) Three-dimensional scatter plot of top three descriptors of PdA, (C) Three-dimensional scatter plot of top three descriptors of AVPD, (D) Three-dimensional scatter plot of top three descriptors of MEEI.
Diagnostics 12 02758 g008
Figure 9. Scatter plots of three most common descriptors for pathology identification process: (A) Scatter plot of top three descriptors of SVD, (B) Scatter plot of top three descriptors of PDA, (C) Scatter plot of top three descriptors of AVPD, (D) Scatter plot of top three descriptors of MEEI.
Figure 9. Scatter plots of three most common descriptors for pathology identification process: (A) Scatter plot of top three descriptors of SVD, (B) Scatter plot of top three descriptors of PDA, (C) Scatter plot of top three descriptors of AVPD, (D) Scatter plot of top three descriptors of MEEI.
Diagnostics 12 02758 g009aDiagnostics 12 02758 g009b
Figure 10. PDF plots of three most common descriptors for multiclass experiments: (A) Density plot for Mean-D4 parameter of SVD, (B) Density plot for Mean-D3 parameter of SVD, (C) Density plot for Mean-D2 parameter of SVD, (D) Density plot for Mean-D4 parameter of PdA, (E) Density plot for Mean-D3 parameter of PdA, (F) Density plot for Mean-D2 parameter of PdA, (G) Density plot for Mean-D4 parameter of AVPD, (H) Density plot for Mean-D3 parameter of AVPD, (I) Density plot for Mean-D2 parameter of AVPD, (J) Density plot for Mean-D4 parameter of MEEI, (K) Density plot for Mean-D3 parameter of MEEI, (L) Density plot for Mean-D2 parameter of MEEI.
Figure 10. PDF plots of three most common descriptors for multiclass experiments: (A) Density plot for Mean-D4 parameter of SVD, (B) Density plot for Mean-D3 parameter of SVD, (C) Density plot for Mean-D2 parameter of SVD, (D) Density plot for Mean-D4 parameter of PdA, (E) Density plot for Mean-D3 parameter of PdA, (F) Density plot for Mean-D2 parameter of PdA, (G) Density plot for Mean-D4 parameter of AVPD, (H) Density plot for Mean-D3 parameter of AVPD, (I) Density plot for Mean-D2 parameter of AVPD, (J) Density plot for Mean-D4 parameter of MEEI, (K) Density plot for Mean-D3 parameter of MEEI, (L) Density plot for Mean-D2 parameter of MEEI.
Diagnostics 12 02758 g010
Figure 11. Three-dimensional scatter plots for detection process: (A) Three-dimensional scatter plot of SVD for detection process, (B) Three-dimensional scatter plot of PdA for detection process, (C) Three-dimensional scatter plot of MEEI for detection process.
Figure 11. Three-dimensional scatter plots for detection process: (A) Three-dimensional scatter plot of SVD for detection process, (B) Three-dimensional scatter plot of PdA for detection process, (C) Three-dimensional scatter plot of MEEI for detection process.
Diagnostics 12 02758 g011
Figure 12. Three-dimensional scatter plots for detection process with full PdA and AVPD datasets: (A) Three-dimensional scatter plots for detection process with full PdA dataset, (B) Three-dimensional scatter plots for detection process with AVPD dataset.
Figure 12. Three-dimensional scatter plots for detection process with full PdA and AVPD datasets: (A) Three-dimensional scatter plots for detection process with full PdA dataset, (B) Three-dimensional scatter plots for detection process with AVPD dataset.
Diagnostics 12 02758 g012
Table 1. Healthy and pathological speakers’ voice (vowel /a/) samples.
Table 1. Healthy and pathological speakers’ voice (vowel /a/) samples.
DatasetNormalCystParalysisPolypLaryngitisReinke’s EdemaNodulesSulcus
SVD [32]262061944414068----
PdA [27]23820182830292924
AVPD [33]118244534----2041
MEEI [34]531052150853----
In Section 3.1, Section 3.2, Section 3.3 and Section 3.4, speakers’ voice (vowel /a/) samples of healthy, cyst, paralysis, and polyp were used.
In addition to the speech recording indicated in rows 1 to 4, the following pathological PdA [27] database speech samples were used for experiments in Section 3.5:
Acquired iatrogenic trauma on the vocal cords = 2, Upper motor neuron injury = 14, Extrapyramidal alterations = 1, and Lack of closure = 6.
Table 2. Characteristics of the databases.
Table 2. Characteristics of the databases.
Sr. No.CharacteristicsSVD [32]PdA [27]AVPD [33]MEEI [34]
1LanguageGermanSpanishArabicEnglish
2Recording
place
Institute of Phonetics of the Saarland University, GermanyPríncipe de Asturias Hospital, MadridKing Abdul Aziz University Hospital, Saudi ArabiaMassachusetts Eye and Ear Infirmary voice and speech laboratory, USA
3Sampling rate of speech samplesAll speech samples 50 KHzAll speech samples
25 KHz
All speech samples
48 KHz
10 KHz, 16 KHz, 25 KHz, and 50 KHz
4Audio format.WAV and .NSP.WAV.WAV and .NSP.NSP
5Recording
session(s)
• Sustained vowels /a/, /i/, and /u/
• A sentence
• Sustained vowel /a/• Vowels /a/, /i/, and /u/
• Al-Fateha (continuous speech)
• Arabic digits
• Isolated common words
• Sustained vowel /a/
• Rainbow passage (continuous speech)
6Duration of
recorded speech samples
• Vowels: 1–3 s
• Sentence: 2 s
Vowels: 1–3 s• Vowel: 5 s
• Al-Fateha: 18 s
• Digits: 10 s
• Words: 3 s
Healthy
• Vowel: 3 s
• Passage: 12 s
Pathological
• Vowel: 1 s
• Passage: 9 s
7Proportion of healthy and
pathological subjects
Healthy: 33%
Pathological: 67%
Healthy: 54.32% Pathological: 45.68%Healthy: 51%
Pathological: 49%
Healthy: 7%
Pathological: 93%
8Severity of
disordered voice quality
Not availableSeverity is rated
on a GRABS scale
Severity is rated
on a scale of 1 (low) to 3 (high)
Not available
9Voice illness typesFunctional and organicOrganic and traumatic etiologiesOrganicOrganic and functional
10Evaluation of
normal subjects
No such
information is
available
No such
information is
available
Evaluation conducted
using laryngeal
stroboscope
No such
information is
available
Table 3. (a) Top ten parameters for recognizing voice dysfunction. (b) Top ten parameters for categorizing voice dysfunction.
Table 3. (a) Top ten parameters for recognizing voice dysfunction. (b) Top ten parameters for categorizing voice dysfunction.
(a)
RankDetection
SVDPdAAVPDMEEI
ParameterIG ValueParameterIG ValueParameterIG ValueParameterIG Value
1Mean-D40.9904Mean-D30.9844Mean-D30.9857Mean-D40.9866
2Mean-D30.9841Mean-D20.9711Mean-D40.9746Mean-D20.9453
3Mean-D10.8880Mean-D10.8929Mean-D20.8959Pentropy-A40.8862
4Mean-D20.5902Mean-D40.6226Mean-D10.8802Mean-D10.8672
5Pentropy-D20.2034Skewness-D40.5664Kurtosis-A10.2414Mean-D30.8611
6Pentropy-D10.2019Skewness-D30.5128Kurtosis-A20.2412Pentropy-A30.7601
7Pentropy-A20.1758Skewness-D20.2123Kurtosis-A30.2391Pentropy-A10.7408
8Skewness-D40.1436Energy-D10.2015Pentropy-D30.2149Pentropy-A20.7178
9Pentropy-A40.1383Variance-D10.2015Kurtosis-A40.2075Skewness-D20.5615
10Pentropy-D30.1282Std-Dev-D10.1736Pentropy-D40.1985Skewness-D10.3802
(b)
RankClassification
SVDPdAAVPDMEEI
ParameterIG ValueParameterIG ValueParameterIG ValueParameterIG Value
1Mean-D41.1168Mean-D11.5628Mean-D21.9189Mean-D11.0684
2Mean-D31.0786Mean-D21.5520Mean-D41.8214Mean-D21.0468
3Mean-D20.9874Mean-D31.3169Mean-D11.7703Mean-D30.9598
4Mean-D10.9810Mean-D41.1685Mean-D31.7252Mean-D40.6896
5Pentropy-A30.8053Variance-A40.6850Mean-A10.3323Pentropy-A40.5580
6Pentropy-A20.7660Variance-A30.6830Mean-A20.3323Pentropy-A30.4609
7Pentropy-A40.7356Variance-A20.6821Mean-A30.3323Kurtosis-D30.4108
8Pentropy-A10.6911Variance-A10.6806Mean-A40.3323Pentropy-A20.4023
9Pentropy-D40.5450Kurtosis-D40.6631Energy-A40.2966Pentropy-A10.3881
10Pentropy-D30.4738Variance-D40.6384Energy-A30.2963Variance-D20.3307
Table 4. Performance indicators for SVM and SGD classifiers for trials identifying vocal disorders.
Table 4. Performance indicators for SVM and SGD classifiers for trials identifying vocal disorders.
Number
of Features
DatabasesClassifier
SVMSGD
AUCCA (%)F1 ScorePPV (%)Recall (%)AUCCA (%)F1 ScorePPV (%)Recall (%)
All 56SVD0.999999.550.995599.5599.550.993099.300.993099.3199.30
PdA0.999999.060.990699.0799.060.981598.150.981498.1898.15
AVPD0.999999.870.998799.8799.870.999799.970.999799.9799.97
MEEI1.000099.930.999399.9399.930.999399.930.999399.9399.93
Top-3SVD0.999999.970.999799.9799.970.999899.980.999899.9899.98
PdA1.000099.970.999799.9799.970.998899.880.998899.8899.88
AVPD0.999999.970.999799.9799.970.999999.990.999999.9999.99
MEEI1.000099.930.999399.9399.930.999999.990.999999.9999.99
Table 5. Performance indicators for SVM and SGD classifiers for multiclass.
Table 5. Performance indicators for SVM and SGD classifiers for multiclass.
Number
of Features
DatabasesClassifier
SVMSGD
AUCCA (%)F1 ScorePPV (%)Recall (%)AUCCA (%)F1 ScorePPV (%)Recall (%)
All 56SVD0.999998.620.986298.6898.620.990498.720.987298.7698.72
PdA0.999898.930.989398.9498.930.993999.090.990999.1199.09
AVPD0.999999.850.998599.8599.850.998399.750.997599.7599.75
MEEI0.999999.470.994799.4899.471.000099.990.999999.9999.99
Top 5SVD0.999998.860.988698.8998.860.994499.260.992699.2799.26
PdA0.999298.240.982498.2598.240.980597.080.970597.1197.08
AVPD0.999999.870.998799.8799.870.998799.810.998199.8199.81
MEEI1.000099.650.996599.6599.650.998699.820.998299.8299.82
Top 3SVD0.998298.190.981898.2698.190.986198.150.981498.1898.15
PdA0.777973.830.666368.4973.830.962894.420.943994.5694.42
AVPD0.999999.730.997399.7399.730.995999.390.993999.4099.39
MEEI1.000099.650.996599.6599.650.998699.820.998299.8399.82
Table 6. SVM and SGD classifier metrics for three most common descriptors.
Table 6. SVM and SGD classifier metrics for three most common descriptors.
Databases Classifier
SVM SGD
AUCCA (%)F1 ScorePPV (%)Recall (%)AUCCA (%)F1 ScorePPV (%)Recall (%)
SVD0.999999.960.999699.9699.960.999599.950.999599.9599.95
PdA0.999999.770.997799.7799.770.997199.710.997199.7199.71
AVPD0.999999.970.999799.9799.970.999699.960.999699.9699.96
MEEI1.000099.860.998699.8699.860.999399.930.999399.9399.93
Table 7. Performance parameters of classifiers SVM and SGD for three most common descriptors.
Table 7. Performance parameters of classifiers SVM and SGD for three most common descriptors.
DatabasesClassifier
SVMSGD
AUCCA (%)F1 ScorePPV (%)Recall (%)AUCCA (%)F1 ScorePPV (%)Recall (%)
SVD0.998298.190.981898.2698.190.986198.150.981498.1898.15
PdA0.997597.290.972797.2997.290.979896.970.969596.9996.97
AVPD0.999899.870.998799.8799.870.998999.830.998399.8399.83
MEEI0.999999.470.994799.4899.470.998699.820.998299.8299.82
Table 8. Identification rate (%) for cross-dataset experiments with SVM and SGD.
Table 8. Identification rate (%) for cross-dataset experiments with SVM and SGD.
Experiment
No.
Classifier
SVMSGD
Model Tested onSVDPdAAVPDMEEISVDPdAAVPDMEEI
Model Trained on
1SVD--99.9899.9799.99--99.9299.9899.99
2PdA99.98--99.9999.9999.99--99.9999.99
3AVPD99.8899.76--99.8199.9999.93--99.93
4MEEI99.8699.1499.57--99.8098.8999.54--
5SVD + PdA----99.9999.99----99.9999.99
6AVPD + SVD--99.99--99.99--99.99--99.99
7PdA + AVPD99.99----99.9999.99----99.99
8MEEI + SVD--99.9899.97----99.9999.98--
9PdA + MEEI99.98--99.99--99.99--99.99--
10MEEI + AVPD99.9799.97----99.9999.94----
11SVD + PDA + MEEI----99.99------99.98--
12SVD + PDA + AVPD------99.99------99.99
13PDA + MEEI + AVPD99.99------99.99------
14SVD + MEEI + AVPD--98.99---- 99.19----
Table 9. Top 10 ranked descriptors with healthy and five pathologies.
Table 9. Top 10 ranked descriptors with healthy and five pathologies.
RankSVDIG ValuePdAIG ValueMEEIIG Value
1Mean-D40.9915Mean-D40.9956Mean-D40.9870
2Mean-D30.9901Mean-D30.9906Mean-D20.9788
3Pentropy-A40.9853Mean-D20.9618Mean-D30.9691
4Pentropy-A10.9850Mean-D10.8881Pentropy-A40.9059
5Pentropy-A20.9848Energy-D30.6470Mean-D10.8725
6Pentropy-A30.9847Variance-D30.6470Pentropy-A30.7945
7Mean-D20.9622Energy-D40.6448Pentropy-A20.7448
8Mean-D10.8922Variance-D40.6448Pentropy-A10.7307
9Pentropy-D40.2313Energy-D20.6350Variance-A40.5659
10Kurtosis-A40.1651Variance-D20.6350Variance-A10.5628
Table 10. Classifier performance measures for voice pathology identification for SVD, PdA, and MEEI.
Table 10. Classifier performance measures for voice pathology identification for SVD, PdA, and MEEI.
Number of DescriptorsDatabasesClassifier
SVMSGD
AUCCA (%)F1 ScorePPV (%)Recall (%)AUCCA (%)F1 ScorePPV (%)Recall (%)
All 56SVD0.999999.940.999499.9499.940.999599.950.999599.9599.95
PdA0.999999.990.999999.9999.990.999299.920.999299.9299.92
MEEI0.999999.990.999999.9999.990.998899.880.998899.8899.88
Common 3SVD0.999999.990.999999.9999.990.999799.970.999799.9799.97
PdA0.999999.990.999999.9999.990.999999.990.999999.9999.99
MEEI0.999999.990.999999.9999.990.999999.990.999999.9999.99
Table 11. Top 10 ranked descriptors with 15 pathologies from PdA and five from AVPD.
Table 11. Top 10 ranked descriptors with 15 pathologies from PdA and five from AVPD.
RankPdAIG ValueAVPDIG Value
1Mean-D40.9962Mean-D40.9956
2Mean-D30.9883Mean-D20.9660
3Mean-D20.9619Mean-D30.9650
4Mean-D10.8768Mean-D10.8829
5Variance-D30.6070Variance-A40.8494
6Energy-D30.6070Variance-A10.8493
7Variance-D40.6012Variance-A20.8493
8Energy-D40.6012Variance-A30.8490
9Variance-D20.5928Std-Dev-A40.8473
10Energy-D20.5928Std-Dev-A20.8470
Table 12. Classifier performance measures for voice pathology identification for PdA and AVPD.
Table 12. Classifier performance measures for voice pathology identification for PdA and AVPD.
Number of DescriptorsDatabasesClassifier
SVMSGD
AUCCA (%)F1 ScorePPV (%)Recall (%)AUCCA (%)F1 ScorePPV (%)Recall (%)
All 56PdA0.999999.990.999999.9999.990.999499.940.999499.9499.94
AVPD0.999999.920.999299.9299.920.999799.970.999799.9799.97
Common 3PdA0.999999.990.999999.9999.990.999999.990.999999.9999.99
AVPD0.999999.990.999999.9999.990.999799.970.999799.9799.97
Table 13. Top 10 ranked descriptors for multiclass.
Table 13. Top 10 ranked descriptors for multiclass.
RankPdAIG ValueAVPDIG Value
1Mean-D21.6651Mean-D21.4317
2Mean-D11.3641Mean-D41.4194
3Mean-D31.2962Mean-D11.1262
4Mean-D41.0894Mean-D30.9220
5Variance-A20.7819Skewness-D40.6898
6Variance-A10.7819Skewness-D30.4225
7Variance-A30.7817Variance-D40.3733
8Variance-A40.7784Energy-D40.3714
9Variance-D40.7726Variance-D30.3618
10Energy-D40.7726Energy-D30.3601
Table 14. Classifier performance measures of multiclass classification for PdA and AVPD.
Table 14. Classifier performance measures of multiclass classification for PdA and AVPD.
Number of DescriptorsDatabasesClassifier
SVMSGD
AUCCA (%)F1 ScorePPV (%)Recall (%)AUCCA (%)F1 ScorePPV (%)Recall (%)
All 56PdA0.999999.010.990199.0199.010.989998.240.982598.3198.24
AVPD0.999797.610.976297.7797.610.990398.380.983898.4598.38
Top 10PdA0.9135 75.21 0.7063 75.84 75.210.9602 93.04 0.9298 93.14 93.04
AVPD0.9998 97.58 0.9760 97.77 97.580.9868 97.80 0.9779 97.84 97.80
Common 3PdA0.9481 73.69 0.6749 74.8973.690.8786 78.76 0.743870.8878.76
AVPD0.9991 98.11 0.9811 98.12 98.110.9707 95.12 0.9506 95.44 95.12
Table 15. Decomposition level effect on identification rate (%).
Table 15. Decomposition level effect on identification rate (%).
DatabaseClassifier
SVMSGD
Decomposition LevelDecomposition Level
1234512345
SVD99.8999.5099.5399.5599.8899.4299.2499.3299.3099.35
PdA99.8899.6599.6899.0699.9199.8898.7799.8098.1599.85
AVPD99.9799.9699.9299.8799.9699.9399.9999.9899.9799.93
MEEI99.9999.9999.9999.9399.9999.9399.9399.9399.9399.99
Table 16. Decomposition level effect on classifying rate (%).
Table 16. Decomposition level effect on classifying rate (%).
DatabaseClassifier
SVMSGD
Decomposition LevelDecomposition Level
1234512345
SVD96.7598.0598.4998.6298.2596.2499.1998.2298.7298.69
PdA94.1092.1497.9298.9399.0493.2096.4998.9399.0999.94
AVPD95.6699.5598.4599.8599.8597.9199.6198.6299.7599.67
MEEI97.2298.9599.1399.4798.9596.0099.1399.6599.9998.78
Table 17. Comparison with existing approaches.
Table 17. Comparison with existing approaches.
Ref. No.Descriptors UsedConveyed Accuracies (%) for Voice Disorder Identification
Database
SVDPdAAVPDMEEIPrivate
[21]MFCC80.20--83.6594.60--
[29]IDP93.20--91.5099.40--
[30]Glottal flow descriptors99.8099.7099.8099.80--
[46]Auto-correlation
and entropy
92.79--99.7999.69--
[47]Peak and lag90.97--91.1699.80--
[48]MDVP99.68--72.5388.21--
[49]Linear prediction of discrete wavelet
coefficients
------97.01--
Proposed approachSWT-based energy and statistical features99.9899.9899.9999.99--
Conveyed accuracies (%) for voice disorder classification
[28]Short-time and glottal flow descriptors--------96.20
[30]Glottal flow descriptors90.4099.3099.8090.10--
[46]Auto-correlation and entropy features99.53 96.0299.54--
[47]Peak and lag98.94--95.1899.25--
Proposed approachSWT-based energy and statistical features99.8698.9399.8599.65--
Table 18. Voice dysfunction detection accuracy (%) using Haar and Daubechies.
Table 18. Voice dysfunction detection accuracy (%) using Haar and Daubechies.
Database SVM SGD
Haardb2db4db6db8db10db12Haardb2db4db6db8db10db12
SVD99.5562.5073.9961.8064.8561.1077.0599.3099.2896.9199.3998.4999.1298.92
PdA99.0687.5382.5891.0079.7072.8286.6298.1598.9798.0699.0393.4597.5596.92
AVPD99.8795.9692.5783.5185.9490.9482.5099.9798.9099.8699.3999.0098.1699.11
MEEI99.9399.7299.8699.9399.7999.8699.7999.9398.6899.3799.7999.1699.6599.37
Table 19. Voice dysfunction categorization accuracy (%) using Haar and Daubechies.
Table 19. Voice dysfunction categorization accuracy (%) using Haar and Daubechies.
DatabaseSVMSGD
Haardb2db4db6db8db10db12Haardb2db4db6db8db10db12
SVD98.6281.6087.1690.0186.1691.2290.7598.7295.2796.5196.5195.4194.3797.42
PdA98.9383.4386.9479.7290.9281.8493.2599.0995.6491.9892.9491.8290.9295.06
AVPD99.8571.5562.1683.0259.7366.3457.7499.7591.0690.8993.5191.9594.2687.93
MEEI99.4793.7591.3194.4488.7188.3686.9799.9998.4395.6593.5789.9392.7088.71
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shrivas, A.; Deshpande, S.; Gidaye, G.; Nirmal, J.; Ezzine, K.; Frikha, M.; Desai, K.; Shinde, S.; Oza, A.D.; Burduhos-Nergis, D.D.; et al. Employing Energy and Statistical Features for Automatic Diagnosis of Voice Disorders. Diagnostics 2022, 12, 2758. https://doi.org/10.3390/diagnostics12112758

AMA Style

Shrivas A, Deshpande S, Gidaye G, Nirmal J, Ezzine K, Frikha M, Desai K, Shinde S, Oza AD, Burduhos-Nergis DD, et al. Employing Energy and Statistical Features for Automatic Diagnosis of Voice Disorders. Diagnostics. 2022; 12(11):2758. https://doi.org/10.3390/diagnostics12112758

Chicago/Turabian Style

Shrivas, Avinash, Shrinivas Deshpande, Girish Gidaye, Jagannath Nirmal, Kadria Ezzine, Mondher Frikha, Kamalakar Desai, Sachin Shinde, Ankit D. Oza, Dumitru Doru Burduhos-Nergis, and et al. 2022. "Employing Energy and Statistical Features for Automatic Diagnosis of Voice Disorders" Diagnostics 12, no. 11: 2758. https://doi.org/10.3390/diagnostics12112758

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop