Next Article in Journal
Consistency of Suspended Particulate Matter Concentration in Turbid Water Retrieved from Sentinel-2 MSI and Landsat-8 OLI Sensors
Next Article in Special Issue
Characterization of Changes in P-Wave VCG Loops Following Pulmonary-Vein Isolation
Previous Article in Journal
Sensitivities of Rheological Properties of Magnetoactive Foam for Soft Sensor Technology
Previous Article in Special Issue
Wavelet Analysis of Overnight Airflow to Detect Obstructive Sleep Apnea in Children
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Monophonic and Polyphonic Wheezing Classification Based on Constrained Low-Rank Non-Negative Matrix Factorization

by
Juan De La Torre Cruz
1,*,
Francisco Jesús Cañadas Quesada
1,
Nicolás Ruiz Reyes
1,
Sebastián García Galán
1,
Julio José Carabias Orti
1 and
Gerardo Peréz Chica
2
1
Department of Telecommunication Engineering, University of Jaen, Campus Cientifico-Tecnologico de Linares, Avda. de la Universidad, s/n, Linares, 23700 Jaen, Spain
2
Pneumology Clinical Management Unit of the University Hospital of Jaen, Av. del Ejercito Espanol, 10, 23007 Jaen, Spain
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(5), 1661; https://doi.org/10.3390/s21051661
Submission received: 25 January 2021 / Revised: 17 February 2021 / Accepted: 22 February 2021 / Published: 28 February 2021
(This article belongs to the Special Issue Biomedical Signal Processing for Disease Diagnosis)

Abstract

:
The appearance of wheezing sounds is widely considered by physicians as a key indicator to detect early pulmonary disorders or even the severity associated with respiratory diseases, as occurs in the case of asthma and chronic obstructive pulmonary disease. From a physician’s point of view, monophonic and polyphonic wheezing classification is still a challenging topic in biomedical signal processing since both types of wheezes are sinusoidal in nature. Unlike most of the classification algorithms in which interference caused by normal respiratory sounds is not addressed in depth, our first contribution proposes a novel Constrained Low-Rank Non-negative Matrix Factorization (CL-RNMF) approach, never applied to classification of wheezing as far as the authors’ knowledge, which incorporates several constraints (sparseness and smoothness) and a low-rank configuration to extract the wheezing spectral content, minimizing the acoustic interference from normal respiratory sounds. The second contribution automatically analyzes the harmonic structure of the energy distribution associated with the estimated wheezing spectrogram to classify the type of wheezing. Experimental results report that: (i) the proposed method outperforms the most recent and relevant state-of-the-art wheezing classification method by approximately 8% in accuracy; (ii) unlike state-of-the-art methods based on classifiers, the proposed method uses an unsupervised approach that does not require any training.

1. Introduction

Chronic Respiratory Diseases (CRDs) are increasingly a huge and growing public health problem due to their high prevalence, high morbidity and mortality, and socio-economic cost. CRDs can be defined as disorders of the airways and other physiological structures of the respiratory system [1]. Some of the most common and relevant CRDs are asthma and Chronic Obstructive Pulmonary Disease (COPD). According to the World Health Organization (WHO), there were 417,918 deaths due to asthma at the global level in 2016 [2] and approximately three million people die from COPD every year, which is 6% of all deaths worldwide [3]. Although chronic diseases currently have no medical cure, early detection can lead to appropriate treatment when the disease is in its early stages, thus improving people’s quality of life.
The auscultation examination is considered a widely used method of detecting CRDs because it is a non-invasive, inexpensive, easy, comfortable, and fast method regardless of age [4]. However, the auscultation process has several limitations that reduce the reliability of the diagnosis: (i) high subjectivity conditioned by the physician’s training to recognize and interpret the sounds captured by the stethoscope [5,6]; (ii) the discrimination between adventitious sounds with similar characteristics, such as monophonic and polyphonic wheezing sounds, is a harder task to perform by means of auscultation [7]; and (iii) normal respiratory sounds and adventitious sounds (abnormal and indicative of a lung disorder) are simultaneously mixed in the time and frequency domain, complicating the physician’s analysis of the valuable clinical information contained in adventitious sounds [5,8,9]. Considering the above, a misdiagnosis is the main cause of the patient returning to the health center with a worsening of the disease that was not detected in the first medical examination performed by auscultation, so in recent years, it has become crucial to develop novel approaches to help physicians provide reliable diagnoses applied to lung disorders, with the implicit fact of reducing health care costs [10,11].
In general, the sounds generated during breathing can be classified into two main categories: normal respiratory sounds and adventitious sounds. Both sounds are mixed in the time-frequency domain as they are simultaneously generated by the same air flow through the bronchial tree of the lungs and also share part of the spectral bands in which they are active [5,9]. Normal respiratory sounds are represented by a wideband spectrum where most of the energy is concentrated in the frequency band of 60 Hz–1000 Hz [12]. Adventitious sounds can be classified into two categories: discontinuous and continuous sounds. Discontinuous Adventitious Sounds (DASs) are characterized by a short duration of less than 25 ms, such as coarse and fine crackles [13]. Continuous Adventitious Sounds (CASs) are characterized by a long duration of more than 100 ms, such as rhonchi, stridor, and wheezing [14]. In recent years, several works have been published that carried out an exhaustive review of lung acoustic measurements [15] and signal processing methods [16,17] applied to adventitious sounds, most of them focused on detection [16,17,18,19,20,21,22,23] and classification tasks [16,17,24,25,26,27,28,29].
Focusing on wheezing or wheeze sounds, the guidelines established by Computerized Respiratory Sound Analysis (CORSA) [5,30] define them as a pitch located between 100 Hz and 1000 Hz whose duration is greater than 100 ms, displaying trajectories of narrowband spectral peaks over time. The appearance of wheezing is widely considered by doctors as a clue to be able to detect either respiratory diseases or the severity associated with CRDs early, as occurs in the case of asthma and COPD [31,32]. For this reason, many research efforts have been applied in biomedical signal processing in order to develop reliable methods for early wheezing detection. In this sense, many wheezing detection algorithms, based on different approaches, can be found in the state-of-the-art literature: the Autoregressive (AR) model [33], auditory modeling [34], entropy [35], Neural Networks (NN) [36,37], wavelet transform [38,39], tonal index [40,41], Mel-Frequency Cepstral Coefficients (MFCCs) [42,43], Gaussian Mixture Models (GMMs) [44,45], spectral peaks identification [46,47,48], the Hidden Markov Model (HMM) [49], and recently, Non-negative Matrix Factorization (NMF) [9,50,51].
In addition, wheezing can be classified into two main categories according to the spectral behavior [52]: (i) wheezes that occur with a single peak or with the harmonics associated with that single basal peak are called Monophonic (MP) wheezes (as can be seen in Figure 1); and (ii) wheezes that occur with variable peaks that differ in harmonics are called Polyphonic (PP) wheezes (as can be seen in Figure 2). The scientific interest in the field of biomedical sound signal processing in automatically performing this classification lies in the fact that PP wheezes are usually caused by the pathology of small airways and MP wheezes are caused by the pathology of larger airways [53]. In fact, several studies [4,54,55,56] have shown that MP and PP wheezes exhibit distinctive physiological and pathological characteristics: (i) in physiological analysis, MP wheezes are caused by a single bronchial narrowing, while PP wheezes are caused by multiple central bronchial compression; and (ii) in pathological analysis, MP wheezes are an indicator of the presence of asthma, while PP wheezes can be considered as a sound marker of COPD.
Despite advances in the analysis of respiratory sounds, MP/PP wheezing classification is a critical step in the diagnosis of asthma [4,54,55] and COPD diseases [54,55,56], so it is still a challenging topic in biomedical signal processing [7] since both types of wheezes are sinusoidal in nature. Although there are relatively few works [7,18,47,57,58,59,60] in which the analysis of MP/PP wheezing is treated, the only works focused on the task of classifying MP/PP wheezing in depth are [7,57,59,60] to our knowledge. All these MP/PP wheezing classification approaches are based on the feature extraction and classifier configuration. Ulukaya et al. [7] proposed to extract a single feature, the Peak Energy Ratio (PER), from a RAtional Dilation Wavelet Transform (RADWT) to discriminate between MP and PP wheezes. Specifically, PER is obtained from the first and second peak with the highest energy of all sub-bands of the wavelet coefficients (considering that the second peak is not consecutive to the first one). Moreover, the authors applied a robust evaluation methodology in which most of the relevant feature extraction methods [57,59,60] were evaluated using some of the most popular classifiers (SVM, KNN, and ELM) and Leave-One-Out (LOO) cross-validation schemes. The results reported that the proposed method, based on only one feature (PER), obtained the best MP/PP wheezing classification performance showing an accuracy equal to 86%.
However, none of the state-of-the-art methods consider the interference generated by normal respiratory sounds that can affect the MP/PP wheezing classification task. In this work, our proposal is based on the Non-negative Matrix Factorization (NMF) approach in order to classify MP/PP wheezing sounds according to the harmonic structure shown by removing the sound interference caused by normal respiratory sounds. The first contribution of this work proposes a novel Constrained Low-Rank Non-negative Matrix Factorization (CL-RNMF) approach, which allows the spectral patterns associated with wheezing sounds to be extracted with the least possible sound interference from normal breath sounds. Specifically, we propose a low-rank configuration using a reduced number of wheeze bases to compact the frequency components into the fewest possible bases for further analysis without loss of relevant wheeze content. In addition, the proposed CL-RNMF approach incorporates a set of constraints to model the spectro-temporal behavior of wheezing and normal respiratory sounds. These constraints help to acoustically isolate the wheezing spectral patterns from normal respiratory sounds. To classify between MP or PP wheezing sounds, the second contribution analyzes the harmonic structure of the previous reduced number of wheezing bases based on the spectral location of the wheezing components, rather than the energy of their components.
The structure of the paper is as follows. Section 2 briefly reviews the principles of non-negative matrix factorization, focusing on the standard approach and some regularizations used to model the properties of the sounds of interest. The proposed MP/PP wheezing classification method is presented in Section 3. Section 4 details and discusses the experimental evaluation. Finally, we conclude in Section 5 and provide perspectives on further research.

2. Theoretical Background

2.1. Non-Negative Matrix Factorization

Non-negative Matrix Factorization (NMF) or standard NMF [61,62] is a decomposition technique that has attracted special attention in different fields of biomedical signal processing in the last few years [63,64]. Previous works showed the efficiency of the NMF approach at detecting [9,50,51] and improving the audio quality of wheezing [65,66]. In general terms, NMF can be defined as an unsupervised learning tool used for linear representation of non-negative two-Dimensional (2D) data where its main advantage is to reduce the dimensionality of a large amount of data in order to find hidden structures by means of part-based representation with non-negative patterns. From a mixture signal x ( t ) , its magnitude spectrogram X R + F × T is obtained by means of the Short-Time Fourier Transform (STFT) applying a window function (e.g., Hamming or Hann) and inter-window overlap to increase the temporal resolution, F being the number of frequency bins and T the number of time frames. Here, standard NMF decomposes the magnitude spectrogram X into the product of two non-negative matrices: spectral basis matrix (patterns) B R + F × K and temporal activation matrix (weights) A R + K × T , K being the rank or the number of components (spectral bases),
X X ^ = B A
where X ^ R + F × T is the estimated spectrogram. Each column of the basis matrix B defines a spectral pattern that describes the spectral behavior of an active sound event in the input spectrogram X . Each row of the activation matrix A represents a temporal gain for a spectral pattern. In other words, the matrix B provides a dictionary composed of K spectral bases, and the matrix A defines the weight with which the different spectral bases appear along the temporal frames. Due to the non-negativity property, NMF underlies an additive linear interpolation model that results in the so-called part-based representation [61].
The decomposition or factorization of the input magnitude spectrogram X into the product B A is usually sought minimizing a defined scalar-valued divergence,
arg min B , A D ( X | B A ) B , A 0
This divergence function measures the error made in the approximation of the observed spectrogram X and the reconstruction B A . Typically, the divergence is computed entry-wise:
D ( X | X ^ ) = D ( X | BA ) = f = 1 F t = 1 T d ( X f , t | X ^ f , t )
where d ( i , j ) is a function of two scalar variables i , j . It is often called the cost function and is a positive function of i R + given j R + with a single minimum for i = j . Some of the most popular cost functions are the Euclidean distance, the generalized Kullback–Leibler divergence, the Itakura–Saito divergence, and the Cauchy distribution [67,68]. In this paper, we propose to minimize the generalized Kullback–Leibler divergence D K L ( X | X ^ ) (see Equation (4)) because previous works [9,50,51,63,65,66] obtained promising results in biomedical signal processing since D K L ( X | X ^ ) provides a scale-invariant factorization, that is low energy sound components of X bear the same relative importance as high energy ones in the decomposition process.
D K L X | X ^ = f = 1 F t = 1 T X f , t log X f , t X ^ f , t X f , t + X ^ f , t
The most popular minimization method to solve the problem shown in Equation (2) is based on the so-called multiplicative update rules, initially proposed by Lee and Seung [61]. This method obtains the basis and activation matrices, minimizing the Kullback–Leibler divergence function D K L ( X | X ^ ) and ensuring the non-negativity of the estimated matrices. These rules are obtained directly from the negative and positive terms of the partial derivative of the divergence function D K L ( X | X ^ ) with respect to the parameters B and A ,
B B D K L X | X ^ B D K L X | X ^ B + = B X BA A T ( [ 1 ] A T )
A A D K L X | X ^ A D K L X | X ^ A + = A B T X BA ( B T [ 1 ] )
where [ 1 ] R + F × T represents an all-ones matrix, T is the transpose operator, ⊙ is the element-wise multiplication, and ⊘ is the element-wise division. This procedure always maintains the non-negativity of both parameters, since the used terms in the updating are also non-negative.
As previously described, NMF models the magnitude spectrogram of an input mixture signal as a product of a basis matrix and an activation matrix with the only constraint of the element-wise non-negativity of all matrices. Under this constraint, the aim is to minimize the cost function of the reconstruction error. However, the main problem of NMF is the trade-off between signal reconstruction and the physical interpretation of the factorized part-based objects. In other words, this non-negativity of the parameters does not guarantee a meaningful part-based representation when dealing with real-world mixture signals [69,70]. Several properties can be used to improve the uniqueness of the local minima obtained by NMF, incorporating physical meaning into the basis functions and activations. In particular, these properties can be implemented using regularizations, which are added to the global cost function in the factorization model. The main constraints, sparseness and smoothness, used in this paper to model the spectro-temporal behavior of wheezing and normal respiratory sounds are briefly described below.

2.2. Spectral Sparseness

Spectral sparseness ψ ( B ) denotes that, for each source, most of its frequencies are zero or close to zero [71,72]. This constraint enforces that only a few frequency bins predominate in each spectral basis, whilst the other bins are canceled. It is implemented by incorporating a penalty term into the NMF objective function. In practice, the L 1 -norm is often used because it was demonstrated to be less sensitive to changes of the parameter that controls the importance of the constraint in the factorization process. Then, the optimization problem can be expressed as,
arg min B , A D ( X | B A ) + α B 1 B , A 0
where α is the weight parameter that adjusts the influence of the constraint.

2.3. Temporal/Spectral Smoothness

Generally, smoothness ϕ means how continuous or smooth the spectral or temporal changes related to a source are [72]. Smoothness constraints have been defined for both activation A and basis functions B and added to the global cost function as penalty terms as follows,
arg min B , A D ( X | B A ) + λ ϕ ( A ) + β ϕ ( B ) B , A 0
where ϕ ( A ) and ϕ ( B ) are the functions that penalize non-smooth temporal activations or spectral patterns and the parameters λ and β control the effect of the regularization in the decomposition procedure.
Temporal smoothness (also known as smooth activations) ϕ ( A ) , applied to the estimated activation matrix A , reports how slow the amplitude variations over time are. In other words, temporal smoothness accounts for the fact that real-world sounds usually have a temporal structure, and their acoustic characteristics vary slowly as a function of time. In [72], the authors proposed to model the temporal smoothness regularization ϕ ( A ) by applying a high cost to large changes produced between adjacent frames in the activation matrix A as follows,
ϕ ( A ) = k = 1 K 1 σ k 2 t = 2 T ( A k , t A k , t 1 ) 2
where σ k = 1 T t = 1 T A k , t 2 indicates the standard deviation used to normalize the activation functions. This normalization provides that the cost of regularization is independent of the numerical scale of activation [69,72].
Spectral smoothness (also known as smooth basis) ϕ ( B ) , applied to the estimated basis matrix B , measures how fast the amplitude changes along the frequency axis, that is it allows modeling the behavior of those sounds that are represented by a wideband spectrum. In [69,73], the authors proposed to model the spectral smoothness regularization ϕ ( B ) by applying a high cost to large changes produced between adjacent bins in the basis matrix B as follows,
ϕ ( B ) = k = 1 K 1 σ k 2 f = 2 F ( B f , k B f 1 , k ) 2
where σ k = 1 F f = 1 F B f , k 2 represents the standard deviation used to normalize the basis functions. This normalization achieves that the cost of regularization is independent of the numerical scale of the basis [69,73].

3. Proposed Method

The main problem in classifying wheezes from a mixture is that both wheezing sounds and normal respiratory sounds occur simultaneously in the time and frequency domain. Considering the acoustic interference caused by normal respiratory sounds, the proposed signal model is composed of two stages: Modelling and separation of wheezing spectral patterns from normal respiratory sounds based on CL-RNMF (stage I) and Classification between MP/PP wheezing according to its harmonic structure (stage II). In this manner, the goal of the stage I is to model the spectral patterns that characterize wheezing sounds by isolating them from respiratory interference. The aim of stage II is to analyze the location of the frequency components extracted from the previous stage to determine the type, monophonic or polyphonic, of wheezing according how the wheezing energy is locating in the frequency domain. The flowchart of the proposed method is shown in Figure 3, and details are depicted in Section 3.1, Section 3.2 and Section 3.3.

3.1. Time-Frequency Signal Representation

Time-frequency representation by means of spectrograms has been demonstrated to be useful for visualizing the characteristics and behavior of both wheezing and normal respiratory sounds [9,50,51,65,66]. The input mixture signal x ( t ) is composed of wheeze sounds x w ( t ) (MP or PP wheezing) and normal respiratory sounds x r ( t ) overlapping in the time and frequency domain. We assume that the mixture of these sounds is additive and can be expressed as x ( t ) = x r ( t ) + x w ( t ) . The input magnitude spectrogram X R + F × T of the input mixture signal can be represented as X = X R + X W , being X R R + F × T the magnitude spectrogram of only respiratory sounds and X W R + F × T the magnitude spectrogram of only wheeze sounds. Specifically, each magnitude spectrogram is composed of T frames, F frequency bins and a set of time-frequency units X f , t , being f = 1 , , F and t = 1 , , T . Each unit X f , t is defined by the fth frequency bin at the tth frame and is calculated from the magnitude of the Short-Time Fourier Transform (STFT) using a Hamming windows of N samples with 10% overlap. In this work, a normalization process is applied in order to achieve independence regarding the size and scale of the input spectrogram X . Thus, the normalized magnitude spectrogram X ¯ is computed as follows,
X ¯ = X f = 1 F t = 1 T X f , t F T
To avoid complex nomenclature throughout the paper, the variable X is hereinafter referred to the normalized magnitude spectrogram previously computed in Equation (11).

3.2. Stage I: Constrained Low-Rank Non-Negative Matrix Factorization

As mentioned above, it is common that normal respiratory sounds mask the presence of the wheezing sounds. As a result, this sound mask makes the task of wheezing classification difficult since the spectral patterns associated with normal respiratory sounds can be confused with wheezing spectral content. Therefore, the aim of this stage is to provide a reliable modeling of the different frequency components (spectral patterns) that compose a wheeze, removing any sound interference from normal respiratory sounds. For this purpose, we propose a CL-RNMF approach because, as far as the authors’ knowledge, the non-negative matrix factorization approach has never been applied before to MP/PP wheezing classification. In addition, our approach is an unsupervised method because it does not require any training of the sounds to classify. Therefore, the proposed method decomposes a magnitude mixture spectrogram X into two estimated spectrograms: X ^ R (only normal respiratory sounds without wheezing) and X ^ W (only wheeze sounds without normal respiratory sounds). In this manner, each estimated spectrogram can be factorized into the product of its corresponding estimated basis and activation matrices: (i) B R R + F × K r and A R R + K r × T to the factorization of X ^ R , K r being the number of respiratory components; and (ii) B W R + F × K w and A W R + K w × T to the factorization of X ^ W , K w being the number of wheezing components. The proposed separation model can be formulated with the following objective function,
X X ^ R + X ^ W = B R A R + B W A W
where, considering the non-negative property that characterizes the NMF approach, all the matrices that compose the previous model are non-negative.
As previously mentioned, this stage attempts to ensure that B W contains reliable modeling of the wheezing spectral patterns by means of narrowband spectral peaks that typically characterize the wheeze content. The key assumptions behind the proposed CL-RNMF approach to model wheezing spectral patterns are the following:
  • Low-rank: The number of wheezing components should be much less than the number of normal respiratory components, that is K w K r . This assumption allows that the number of frequency components can be reduced in the least number of bases possible for their posterior analysis, while normal respiratory sounds are modeled using a higher range of components. Experimental results showed that the best classification performance was obtained when 2 K w 6 and K r 32 . In particular, when K w = 1 , the proposed CL-RNMF approach tends to converge very quickly at the expense of losing relevant wheezing content. On the other hand, when K w > 6 , the spectral wheezing patterns tend to be split into different components of the matrix B W .
  • Constraints: These characterize wheezing sounds and normal respiratory sounds using opposite restrictions between both sounds. The use of constraints allows isolating the spectral wheezing patterns from the spectral patterns of normal respiratory sounds. Therefore, in order to find a better NMF decomposition that shows spectro-temporal features of the wheezing and normal respiratory sounds as can be observed in the real world, we propose to incorporate sparseness and smoothness into the NMF decomposition process. As shown in Figure 1 and Figure 2, wheezing sounds can be considered sparse in frequency because MP wheezing or PP wheezing is characterized by one or more than one narrowband spectral peak. Moreover, wheezing sounds can be considered smooth or continuous events in time, that is slow variation of the magnitude spectrogram along time. On the other hand, normal respiratory sounds can be considered smooth in frequency, that is they can be modeled assuming wideband spectral patterns. Therefore, B W should contain wheezing spectral patterns composed of one or more than one narrowband spectral peak, depending on the spectral complexity of each wheezing, and B R should be composed of a set of wideband spectral patterns that model the behavior of normal respiratory sounds.
Considering the key assumptions mentioned above, the global objective function D X | X ^ that must be minimized in order to estimate the basis ( B R , B W ) and activation ( A R , A W ) matrices is composed of: (i) the Kullback–Leibler divergence cost function D K L X | X ^ to minimize the reconstruction error between the input spectrogram X and the estimated spectrogram X ^ , (ii) the spectral sparseness ψ B W and temporal smoothness ϕ A W restrictions applied to B W and A W , respectively, to model the wheezing spectral patterns, and (iii) the spectral smoothness ϕ B R restriction applied to B R , to model the spectral patterns of normal respiratory sounds. The global objective function D X | X ^ is detailed as follows,
D X | X ^ = D K L X | X ^ + α ψ B W + λ ϕ A W + β ϕ B R
where the equations of terms D K L X | X ^ , ψ B W , ϕ A W , and ϕ B R can be found in Section 2. The parameters α , λ , and β define the weight to control the effect of the regularization. Experimental results showed that the best classification performance is obtained when all weights are equal α = λ = β , the optimal value being α = λ = β = 0.5 . Analyzing the sound separation performance of the previous decomposition, we observed empirically that the acoustic interference suffered by wheezing sounds from normal respiratory sounds is minimum, and no significant loss of wheezing content occurs when α = λ = β . However, significant losses of wheezing content appear when α = λ > β , or significant sound interference by normal respiratory sounds can be observed when α = λ < β .
From Equation (13), the estimated basis matrices ( B W and B R ) and activation matrices ( A W and A R ) can be obtained by applying a gradient descent algorithm based on multiplicative update rules. Specifically, the multiplicative update rules to learn those matrices can be computed by taking negative and positive terms of the partial derivative of the global objective function D X | X ^ with respect to B W , B R , A W , and A R , respectively,
B W B W D K L X | X ^ B W + α ψ ( B W ) B W D K L X | X ^ B W + + α ψ ( B W ) B W +
B R B R D K L X | X ^ B R + β ϕ ( B R ) B R D K L X | X ^ B R + + β ϕ ( B R ) B R +
A W A W D K L X | X ^ A W + λ ϕ ( A W ) A W D K L X | X ^ A W + + λ ϕ ( A W ) A W +
A R A R D K L X | X ^ A R D K L X | X ^ A R +
where, for each multiplicative update rule, the division between the negative and positive terms of the partial derivatives is an element-wise division. More details related to the equations of each partial derivative of the multiplication update rules can be found in Appendix A. Finally, the estimated respiratory and wheezing basis ( B W and B R ) and activation matrices ( A W and A R ) are obtained updating the previous rules until the algorithm converges using M iterations. Figure 4 shows the estimated matrices B W and B R decomposing the MP wheezing spectrogram shown in Figure 1B. As can be observed, the matrix B W contains spectral patterns that characterize a typical MP wheezing, which are represented by means of a set of narrowband spectral peaks (or frequency components). In contrast, the estimated matrix B R is composed of a set of wideband spectral patterns that characterize normal respiratory sounds. Therefore, the proposed CL-RNMF approach achieves extracting the wheezing spectral content at the expense of removing normal respiratory sounds.
Experimentally, we found that the proposed CL-RNMF approach tends to compact all the narrowband spectral peaks into a single basis of the matrix B W , as shown in Figure 4A. However, considering that CL-RNMF uses a small set of wheezing components ( K w ), in some cases, the narrowband spectral peaks are divided into several bases of the same matrix B W . To clarify this issue, Figure 5 shows the matrix B W obtained for the different examples of MP and PP wheezing described in Section 1. As shown in Figure 5D, the energy of the narrowband spectral patterns, which characterizes that PP wheezing, are divided into two bases B W ( 1 ) and B W ( 2 ) belonging to the matrix B W . In both bases, B W ( 1 ) and B W ( 2 ) , all narrowband spectral peaks are correctly modeled.
Finally, we propose to obtain the spectral energy distribution ξ ( f ) (see Equation (18)) from the set of bases that compose the matrix B W . This makes it possible to compact the spectral distribution of all narrowband spectral peaks that make up the input MP or PP wheezes to analyze their harmonic structure in Stage II.
ξ ( f ) = k w = 1 K w B W f , k w , f = 1 , , F
Figure 6 shows the spectral energy distribution ξ ( f ) obtained for the four examples of wheezing shown in Section 1. The pseudocode of this Stage I for the modeling and separation of wheezing spectral patterns based on CL-RNMF is detailed in Algorithm 1.
Algorithm 1: CL-RNMF.
Require: x ( t ) , K r , K w , α , β , λ , and M.ss
 1:  Compute the normalized magnitude spectrogram X using Equation (11).ss
 2:  Initialize B W , B R , A W , and A R with random non-negative values.ss
 3:  Update the estimated wheezing basis matrix B W using Equation (14).ss
 4:  Update the estimated respiratory basis matrix B R using Equation (15).ss
 5:  Update the estimated wheezing activations matrix A W using Equation (16).ss
 6:  Update the estimated respiratory activations matrix A R using Equation (17).ss
 7:  Repeat Steps 3–6 until the algorithm converges (or until the maximum number of iterations M is reached).ss
 8:  Compute the spectral energy distribution ξ ( f ) from B W using Equation (18).ss
return ξ ( f )

3.3. Stage II: Harmonic Structure Analysis

The goal of this stage is to classify between MP and PP wheezing by analyzing the spectral energy distribution ξ ( f ) of the different narrowband spectral peaks obtained in the previous stage. Depending on the harmonic structure, wheezing can be classified as MP or PP. Specifically, MP wheezing is composed of a single narrowband spectral peak or the set of harmonically related narrowband spectral peaks. In contrast, PP wheezing is composed of several non-harmonically related narrowband spectral peaks. For this reason, we propose to obtain the number of narrowband spectral peaks η that can be found from ξ ( f ) . Note that the procedure to detect the spectral peaks is a simple task since, as can be seen in Figure 6, the spectral energy distribution ξ ( f ) from CL-RNMF clearly provides a set of narrowband spectral peaks typically found in wheezing sounds. Once the parameter η is obtained, a preliminary classification of the type of wheezing can be performed as follows,
Wheezing category = MP if η = 1 MP or PP if η > 1
Wheezing can only be classified as MP when η = 1 since a wheezing is composed of a single narrowband spectral peak, as can be seen in Figure 6A. However, a wheezing can be classified as MP or PP when η > 1 , depending on the harmonic structure that exists between the different narrowband spectral peaks. Specifically, the wheezing is classified as MP if the set of spectral peaks are harmonically related between them. The wheezing is classified as PP if the spectral peaks are not harmonically related between them. In order to perform the classification between MP and PP wheezing in the case of η > 1 , we propose a two-step procedure, as follows:
  • The objective of the first step is to locate, in terms of frequency, all the narrowband spectral peaks detected in the previous Stage I. For this, we propose to locate the most prominent frequency f p ( z ) in each spectral peak z = 1 , , η . Each value f p ( z ) was calculated using the findpeaks function provided by the MATLAB software [74] due to the satisfactory results obtained in several preliminary analyses performed. Figure 7 shows the location f p ( z ) , in terms of frequency, of each spectral peak for the MP example previously shown in Figure 1B.
  • The objective of the second step is to check if the different spectral peaks z = 1 , , η are harmonically related or not. We assume that the first spectral peak ( z = 1 ) represents the basal peak. Therefore, the wheezing is classified as MP if the rest of spectral peaks ( z = 2 , , η ) are located in the harmonic frequencies (integer multiple) of the basal peak. Otherwise, the wheezing is classified as PP. From the width Δ of the main lobe of the basal peak ( z = 1 ) and the value of its most prominent frequency f p ( 1 ) , the spectral intervals where the possible harmonic frequencies should be located are calculated as follows,
    Λ z = z f p ( 1 ) ( Δ / 2 ) , z f p ( 1 ) + ( Δ / 2 ) , z = 1 , , η
    where i , j denotes the spectral interval comprised between the lower limit i and the upper limit j, in terms of frequency. Specifically, Λ 1 represents the spectral interval associated with the basal peak, and Λ z ( z = 2 , , η ) corresponds to the spectral intervals where the harmonic frequencies should be located. Note that the width of the main lobe Δ was obtained by positioning the reference line beneath the peak at a vertical distance equal to half the peak prominence [74].
Considering the two-step procedure described above, wheezes that are composed of several narrowband spectral peaks ( η > 1 ) can be classified as MP or PP as follows,
Wheezing category = MP if f p ( z ) Λ z , z = 2 , , η PP otherwise
where v V denotes that element v is contained in the interval V. Therefore, when the frequency f p ( z ) of all possible harmonic spectral peaks z = 2 , , η is located in the corresponding spectral intervals Λ z , the wheezing is classified as MP. Otherwise, wheezing is classified as PP because the narrowband spectral peaks that characterize the wheezing are not harmonically related. This occurs when for the frequency f p ( z ) , at least one of the possible harmonic spectral peaks is not located in its corresponding spectral intervals Λ z . Figure 7 shows an example of the procedure described for MP wheezing composed of a basal peak and two harmonics. Figure 8 shows two examples of the procedure described for two PP wheezing with several non-harmonically related spectral peaks. Finally, the pseudocode of this stage for the classification between MP/PP wheezing according to its harmonic structure is detailed in Algorithm 2.
Algorithm 2: Harmonic structure analysis.
Require: ξ ( f ) .
  1:  From ξ ( f ) , detect the number η of narrowband spectral peaks.
  if   η = 1   then
    return        Wheezing category = MP
  else
    2:  Locate the frequency f p ( z ) in each spectral peak z = 1 , , η .ssssss
    3:  Compute the spectral intervals Λ z using Equation (20).
    if   f p ( z ) Λ z , z = 2 , , η   then
      return        Wheezing category = MP
    else
      return        Wheezing category = PP
    end if
  end if

4. Experimental Results and Discussion

4.1. Data Collection

As far as the authors’ knowledge, there is no public wheeze database in which wheezing has been labeled as monophonic or polyphonic. For this reason, we received the collaboration of a pneumologist from the University Hospital of Jaén (Spain) to create and label a database according to the wheezing harmonic structure. The database was created by collecting and categorizing a set of recordings from different subjects of the most widely used Internet pulmonary repositories [75,76,77,78,79,80,81,82,83,84,85,86,87]. Specifically, all previous recordings were collected from subjects with CRDs (asthma or COPD). Note that the set of recordings selected for this assessment was only composed of normal respiratory sounds and wheezing sounds.
The type of wheezing (MP or PP) was labeled by the pneumologist by means of an acoustic inspection and a visual verification of the spectrogram considering the harmonic structure that distinguishes both types of wheezing. The database consisted of 200 MP and 200 PP wheezing segments, where the duration of each segment was at least 100 ms, to be consistent with the literature. As mentioned above, MP wheezing can show two different harmonic structures: Type 1, wheezes with a single peak, that is only the fundamental frequency component is active; and Type 2, wheezes with the harmonics of a single basal peak, that is both the fundamental frequency component and its frequencies that are harmonically related are active. Therefore, to guarantee the maximum variability of the MP wheezing, the 200 MP wheezing segments were divided into 100 MP wheezing segments with a single peak and 100 MP wheezing segments with the harmonics of a single basal peak. Note that all segments were independent of each other, since each segment corresponded to a different wheezing from the rest. Finally, all segments in the database were sampled at 4096 Hz and had a length between 100 and 700 ms. Figure 9 shows the classification performed on the database created.

4.2. Experimental Setup

To be consistent with the literature, we assumed that wheezing sounds were not active below 100 Hz and above 1000 Hz. For this reason, all segments that compose the database were band-limited from 100 Hz–1000 Hz.
The length of the signal frames was set to N = 256 samples (62.5 ms). This frame size was considered large enough to assume a perfect spectral representation of all wheezing frequency components. The overlap between frames was set to 10% (6.25 ms). To obtain the time-frequency representation, windowing with a Hamming window was applied, and the order of the Discrete Fourier Transform (DFT) was set to 2 N frequency bins, similar to [9,50]. This DFT size provides a high enough resolution for modeling the spectral patterns of wheezing sounds and was chosen empirically as a trade-off between achieved quality and complexity. Besides, we empirically observed that the reconstruction error converged after 50 iterations, so the maximum number of iterations for the decomposition was equal to M = 50 .
Finally, note that the performance of the proposed method depends on the initial values with which the basis matrices B W , B R and the activation matrices A W , A R are initialized. Although the obtained results are not dispersed and keep the same behavior, in order to overcome this issue, we ran the proposed method five times for each segment that composed the database, and the results shown in this paper are averaged values.

4.3. Evaluation Metrics

The Accuracy rates (ACC) were used to evaluate the performance of the proposed method, which are commonly used in the field of wheezing classification [7]. In order to provide a fair evaluation of the classification performance obtained by the proposed method and the state-of-the-art algorithms, the following accuracy rates were proposed: (i) A C C G is the ability to correctly classify a wheezing segment as MP or PP; (ii) A C C P represents the ability to correctly classify a wheezing segment as PP; (iii) A C C M corresponds to the ability to correctly classify a wheezing segment as MP; (iv) A C C M 1 indicates the ability to correctly classify a wheezing segment as MP Type 1; and (v) A C C M 2 reports the ability to correctly classify a wheezing segment as MP Type 2. The terms used in Equations (22)–(26) are described in Table 1.
A C C G = T P + T M T P + T M + F P + F M )
A C C P = T P T P + F P
A C C M = T M T M + F M
A C C M 1 = T M 1 T M 1 + F M 1
A C C M 2 = T M 2 T M 2 + F M 2

4.4. State-of-the-Art Method for Comparison

In order to measure the MP/PP classification performance of the proposal, we used the most recent and relevant state-of-the-art algorithm [7], denoted as UPER in this paper. The method UPER was implemented strictly following the instructions specified by the authors in [7]. Firstly, the values of the metric PER were obtained using the 19th parameter set ( p = 10 , q = 11 , s = 7 , and J = 45 ) in the RADWT model. Then, three classifiers, Support Vector Machine (SVM) with the Radial Basis Function kernel (RBF kernel), K-Nearest Neighbor (KNN), and Extreme Learning Machine (ELM) were applied to the PER features. The classification performance of UPER was obtained in Leave-One-Out (LOO) cross-validation schemes with the SVM, KNN, and ELM classifiers. Specifically, LOO cross-validation is a particular case of Leave-p-Out (LPO) cross-validation with p = 1 . Therefore, the LOO scheme involves using one observation as the validation set and the remaining observations as the training set. This is repeated in all ways to cut the database into a validation set of one observation and a training set. Considering the database evaluated in this work (400 segments in total), the LOO cross-validation scheme has 400 possible combinations of validation in which the training set is composed of 399 segments, and only one segment is tested, as can be observed in Figure 10). Results shown in this paper for all classifiers are the average values obtained from the 400 possible validation combinations.

4.5. Accuracy Results

In this section, we evaluate the MP/PP classification performance between the proposed method and UPER [7]. A remarkable distinction between the two methods is that the proposed method is completely unsupervised or blind (no training), but the method UPER depends on the training database.
Table 2 shows the MP/PP classification results, in terms of the accuracy rates, evaluating the database described in Section 4.1. Results provided by UPER, considering the three classifier versions (SVM, KNN, and ELM), were obtained by applying an LOO cross-validation scheme as was previously described in Section 4.4. Results report that the proposed method provides the best overall MP/PP classification results compared to UPER considering all evaluated metrics. Focusing on the different accuracy rates, the following can be observed:
  • the improvement, in terms of A C C G , of the proposed method is about 8.25% UPER (SVM), 12% UPER (KNN), and 10.5% UPER (ELM).
  • the improvement, in terms of A C C P , of the proposed method is about 4% UPER (SVM), 7.1% UPER (KNN), and 5.5% UPER (ELM).
  • the improvement, in terms of A C C M , of the proposed method is about 12.5% UPER (SVM), 17% UPER (KNN), and 15.5% UPER (ELM).
  • the improvement, in terms of A C C M 1 , of the proposed method is about 5% UPER (SVM), 10% UPER (KNN), and 8% UPER (ELM).
  • the improvement, in terms of A C C M 2 , of the proposed method is about 20% UPER (SVM), 24% UPER (KNN), and 23% UPER (ELM).
The main advantage of UPER is that it only uses one feature (PER value) to discriminate between MP and PP wheezing. As shown in Table 2, the SVM classifier obtains the best classification performance in the method UPER. Specifically, the classifier SVM achieves an improvement of 2.25% (KNN) and 1.5% (KNN), in terms of A C C G . These results are consistent with those obtained by the authors in [7], confirming that the SVM classifier with the RBF kernel obtains the best classification performance when the number of features (only one PER value) is small [88].
Performing an empirical analysis of the proposed method and UPER, the following observations were extracted:
(i)
Due to the time-frequency overlapping problem, normal respiratory sounds often mask wheezing sounds, hiding relevant medical information [5]. While the proposed method (based on CL-RNMF) allows removing as much as possible the acoustic interference from normal respiratory sounds, the method UPER is based on a feature PER obtained from the sub-band energy of the wavelet coefficients, so the presence of normal respiratory sounds interferes in the selection of the optimal sub-bands that really belong to the wheezing components.
(ii)
The method UPER has more difficulty in discriminating between PP and MP wheezing composed by a basal peak and its harmonics since it achieves the worst performance in terms of A C C M 2 . The reason is because UPER is based on energy and ignoring the spectral location of the components that model the harmonic behavior of MP wheezing. Results in Table 2 suggest that MP/PP classification based on the spectral location of the harmonic structure as occurs in the proposed method is more reliable than the use of the energy of the wheezing spectral components, as occurs in UPER.
The LOO cross-validation scheme does not show the dependency that classifiers have with the size of the training segments set, since this scheme always uses one segment as the validation set and the remaining segments as the training set. For this reason, we propose to use an LPO cross-validation scheme by varying the size of the training segments set. The LPO scheme requires training and validating the model C p n times, where n is the number of segments that compose the database, p is the number of validation segments, and C p n is the binomial coefficient. As a result, the associated computational cost can be excessive. In order to overcome this issue, we limited the number of iterations of the LPO scheme to 500. Furthermore, the same number of MP and PP wheezes was selected for both training and validation sets in each iteration. Specifically, we used four LPO schemes: (i) p = 80 uses 80% of the total segments as the training set in each iteration; (ii) p = 160 uses 60% of the total segments as the training set in each iteration; (iii) p = 240 uses 40% of the total segments as the training set in each iteration; and (iv) p = 320 uses 20% of the total segments as the training set in each iteration. Considering all the instructions described above, Table 3 shows the MP/PP classification results, in terms of A C C G , obtained by UPER using its three classifier versions (SVM, KNN, and ELM) in order to assess its dependence on the training set size. Comparing the LOO scheme with the LPO scheme ( p = 320 ), the A C C G reduction of the classification performance is about 7.5% (SVM), 8.25% (KNN), and 6.25% (ELM). Results report that the PER feature allows distinguishing between MP and PP wheezing even when the training set size is reduced. In addition, the ELM classifier shows less dependence on the training database size compared to SVM and KNN.

5. Conclusions and Future Work

In this paper, we present a novel Constrained Low-rank Non-negative Matrix Factorization (CL-RNMF) approach to classify monophonic and polyphonic wheezing sounds according to their harmonic structure. The first contribution of this work proposes a CL-RNMF framework that allows extracting the spectral patterns that characterize wheezing sounds with the least possible interference from normal respiratory sounds. Specifically, a low-rank configuration with a reduced number of wheezing bases is presented to compact its frequency components in the least number of bases possible for their posterior analysis. In addition, CL-RNMF uses a set of constraints to model the spectro-temporal behavior of wheezing and normal respiratory sounds. As far as the authors’ knowledge, the non-negative matrix factorization approach has never been applied before to MP/PP wheezing classification. The second contribution analyzes the harmonic structure of the energy distribution from the estimated wheezing spectrogram provided by CL-RNMF to determine the type of wheezing, allowing a more efficient classification based on the location of the wheezing frequency components, rather than the energy of their components.
The most relevant conclusions from the experimental results indicate the following: (i) the proposed method provides the best overall performance related to MP/PP wheezing classification compared to the most relevant method of the state-of-the-art; (ii) unlike most state-of-the-art methods based on classifiers, the proposed method is an unsupervised (blind) approach that does not require any training from wheezing sounds; (iii) the proposed method achieves removing most of the interference from normal respiratory sounds; (iv) specific accuracy rates, A C C M and A C C P , obtained by the proposed method seem to suggest the ability of the proposal to classify both monophonic and polyphonic wheezing sounds correctly.
Future work will be focused on the design of new constraints, to be applied in NMF approaches, that improve the modeling of time-frequency respiratory sound events, analyzing different types of adventitious sounds, such as wheezes and crackles. The objective of this future research line is to perform an early detection and classification among the different types of adventitious sounds active in the auscultation process in order to maximize the reliability of the diagnosis issued by the physician in the case of pathologies of lung diseases caused by the appearance of such adventitious sounds.

Author Contributions

Conceptualization, J.D.L.T.C., F.J.C.Q., N.R.R., and G.P.C.; data curation, J.D.L.T.C. and F.J.C.Q.; formal analysis, J.D.L.T.C., F.J.C.Q., N.R.R., S.G.G., and J.J.C.O.; investigation, J.D.L.T.C. and F.J.C.Q.; methodology, J.D.L.T.C., F.J.C.Q., and N.R.R.; software, J.D.L.T.C., F.J.C.Q., S.G.G., and J.J.C.O.; supervision, F.J.C.Q., N.R.R., S.G.G., and J.J.C.O.; validation, J.D.L.T.C., F.J.C.Q., and G.P.C.; visualization, J.D.L.T.C.; writing, original draft, J.D.L.T.C., F.J.C.Q., N.R.R., S.G.G., J.J.C.O., and G.P.C.; writing, review and editing, J.D.L.T.C., F.J.C.Q., N.R.R., S.G.G., and J.J.C.O. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Programa Operativo FEDER Andalucia 2014–2020 under the project with Reference 1257914 and the Ministry of Economy, Knowledge and University, Junta de Andalucia under Project P18-RT-1994.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Terms of the Multiplicative Update Rules

Here are each of the terms belonging to the multiplicative update rule to obtain the basis matrix B W :
D K L X | X ^ B W = X X ^ A W T
D K L X | X ^ B W + = [ 1 ] A W T
ψ B W B W f , k w = F B W f , k w j = 1 F B W j , k w ( j = 1 F B W j , k w 2 ) 3 2
ψ B W B W f , k w + = 1 1 F j = 1 F B W j , k w 2
Here are each of the terms belonging to the multiplicative update rule to obtain the basis matrix B R :
D K L X | X ^ B R = X X ^ A R T
D K L X | X ^ B R + = [ 1 ] A R T
ϕ B R B R f , k r = 2 F ( B R f 1 , k r + B R f + 1 , k r ) j = 1 F B R j , k r 2 + + 2 F B R f , k r j = 2 F ( B R j , k r B R j 1 , k r ) 2 ( j = 1 F B R j , k r 2 ) 2
ϕ B R B R f , k r + = 4 F B R f , k r j = 1 F B R j , k r 2
Here are each of the terms belonging to the multiplicative update rule to obtain the activations matrix A W :
D K L X | X ^ A W = B W T X X ^
D K L X | X ^ A W + = B W T [ 1 ]
ϕ A ^ W A W k w , t = 2 T ( A W k w , t 1 + A W k w , t + 1 ) i = 1 T A W k w , i 2 + + 2 T A W k w , t i = 2 T ( A W k w , i A W k w , i 1 ) 2 ( i = 1 T A W k w , i 2 ) 2
ϕ A W A W k w , t + = 4 T A W k w , t i = 1 T A W k w , i 2
Here are each of the terms belonging to the multiplicative update rule to obtain the activations matrix A R :
D K L X | X ^ A R = B R T X X ^
D K L X | X ^ A R + = B R T [ 1 ]

References

  1. World Health Organization. Chronic Respiratory Diseases. Available online: https://www.who.int/health-topics/chronic-respiratory-diseases#tab=tab_1 (accessed on 30 December 2020).
  2. World Health Organization, Asthma. Available online: https://www.who.int/news-room/fact-sheets/detail/asthma (accessed on 30 December 2020).
  3. World Health Organization. Chronic Obstructive Pulmonary Disease. Available online: http://www.emro.who.int/health-topics/chronic-obstructive-pulmonary-disease-copd/index.html (accessed on 30 December 2020).
  4. Sarkar, M.; Madabhavi, I.; Niranjan, N.; Dogra, M. Auscultation of the respiratory system. Ann. Thorac. Med. 2015, 10, 158. [Google Scholar] [CrossRef] [PubMed]
  5. Pasterkamp, H.; Kraman, S.S.; Wodicka, G.R. Respiratory sounds: Advances beyond the stethoscope. Am. J. Respir. Crit. Care Med. 1997, 156, 974–987. [Google Scholar] [CrossRef]
  6. Lozano-Garcia, M.; Fiz, J.A.; Martinez-Rivera, C.; Torrents, A.; Ruiz-Manzano, J.; Jane, R. Novel approach to continuous adventitious respiratory sound analysis for the assessment of bronchodilator response. PLoS ONE 2017, 12, e0171455. [Google Scholar] [CrossRef]
  7. Ulukaya, S.; Serbes, G.; Kahya, Y.P. Wheeze type classification using non-dyadic wavelet transform based optimal energy ratio technique. Comput. Biol. Med. 2019, 104, 175–182. [Google Scholar] [CrossRef] [PubMed]
  8. Andrès, E.; Gass, R.; Charloux, A.; Brandt, C.; Hentzler, A. Respiratory sound analysis in the era of evidence-based medicine and the world of medicine 2.0. J. Med. Life 2018, 11, 89. [Google Scholar] [PubMed]
  9. Torre-Cruz, J.; Canadas-Quesada, F.; Carabias-Orti, J.; Vera-Candeas, P.; Ruiz-Reyes, N. A novel wheezing detection approach based on constrained non-negative matrix factorization. Appl. Acoust. 2019, 148, 276–288. [Google Scholar] [CrossRef]
  10. Leng, S.; San Tan, R.; Chai, K.T.C.; Wang, C.; Ghista, D.; Zhong, L. The electronic stethoscope. Biomed. Eng. Online 2015, 14, 1–37. [Google Scholar] [CrossRef] [Green Version]
  11. Sen, I.; Saraclar, M.; Kahya, Y.P. A comparison of SVM and GMM-based classifier configurations for diagnostic classification of pulmonary sounds. IEEE Trans. Biomed. Eng. 2015, 62, 1768–1776. [Google Scholar] [CrossRef]
  12. Salazar, A.J.; Alvarado, C.; Lozano, F.E. System of heart and lung sounds separation for store-and-forward telemedicine applications. Rev. Fac. Ingeniería Univ. Antioq. 2012, 64, 175–181. [Google Scholar]
  13. Douros, K.; Grammeniatis, V.; Loukou, I. Crackles and Other Lung Sounds. In Breath Sounds; Springer International Publishing: Cham, Switzerland, 2018; Chapter 12; pp. 225–236. [Google Scholar]
  14. Lozano, M.; Fiz, J.A.; Jané, R. Automatic differentiation of normal and continuous adventitious respiratory sounds using ensemble empirical mode decomposition and instantaneous frequency. IEEE J. Biomed. Health Inform. 2015, 20, 486–497. [Google Scholar] [CrossRef]
  15. Rao, A.; Huynh, E.; Royston, T.J.; Kornblith, A.; Roy, S. Acoustic methods for pulmonary diagnosis. IEEE Rev. Biomed. Eng. 2018, 12, 221–239. [Google Scholar] [CrossRef] [Green Version]
  16. Pramono, R.X.A.; Bowyer, S.; Rodriguez-Villegas, E. Automatic adventitious respiratory sound analysis: A systematic review. PLoS ONE 2017, 12, e0177926. [Google Scholar] [CrossRef] [Green Version]
  17. Rocha, B.M.; Pessoa, D.; Marques, A.; Carvalho, P.; Paiva, R.P. Automatic Classification of Adventitious Respiratory Sounds: A (Un) Solved Problem? Sensors 2021, 21, 57. [Google Scholar] [CrossRef] [PubMed]
  18. Jin, F.; Krishnan, S.; Sattar, F. Adventitious sounds identification and extraction using temporal–spectral dominance-based features. IEEE Trans. Biomed. Eng. 2011, 58, 3078–3087. [Google Scholar] [PubMed]
  19. Gurung, A.; Scrafford, C.G.; Tielsch, J.M.; Levine, O.S.; Checkley, W. Computerized lung sound analysis as diagnostic aid for the detection of abnormal lung sounds: A systematic review and meta-analysis. Respir. Med. 2011, 105, 1396–1403. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Sakai, T.; Kato, M.; Miyahara, S.; Kiyasu, S. Robust detection of adventitious lung sounds in electronic auscultation signals. In Proceedings of the 21st IEEE International Conference on Pattern Recognition (ICPR2012), Tsukuba, Japan, 11–15 November 2012; pp. 1993–1996. [Google Scholar]
  21. Liu, X.; Ser, W.; Zhang, J.; Goh, D.Y.T. Detection of adventitious lung sounds using entropy features and a 2-D threshold setting. In Proceedings of the 2015 10th IEEE International Conference on Information, Communications and Signal Processing (ICICS), Singapore, 2–4 December 2015; pp. 1–5. [Google Scholar]
  22. Matsutake, S.; Yamashita, M.; Matsunaga, S. Abnormal-respiration detection by considering correlation of observation of adventitious sounds. In Proceedings of the 2015 23rd IEEE European Signal Processing Conference (EUSIPCO), Nice, France, 31 August–4 September 2015; pp. 634–638. [Google Scholar]
  23. Nakamura, N.; Yamashita, M.; Matsunaga, S. Detection of patients considering observation frequency of continuous and discontinuous adventitious sounds in lung sounds. In Proceedings of the 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 16–20 August 2016; pp. 3457–3460. [Google Scholar]
  24. İçer, S.; Gengeç, Ş. Classification and analysis of non-stationary characteristics of crackle and rhonchus lung adventitious sounds. Digit. Signal Process. 2014, 28, 18–27. [Google Scholar] [CrossRef]
  25. Yamashita, M.; Himeshima, M.; Matsunaga, S. Robust classification between normal and abnormal lung sounds using adventitious-sound and heart-sound models. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; pp. 4418–4422. [Google Scholar]
  26. Aykanat, M.; Kılıç, Ö.; Kurt, B.; Saryal, S. Classification of lung sounds using convolutional neural networks. EURASIP J. Image Video Process. 2017, 65, 1–9. [Google Scholar] [CrossRef]
  27. Bardou, D.; Zhang, K.; Ahmad, S.M. Lung sounds classification using convolutional neural networks. Artif. Intell. Med. 2018, 88, 58–69. [Google Scholar] [CrossRef]
  28. Ma, Y.; Xu, X.; Li, Y. LungRN+NL: An improved adventitious lung sound classification using non-local block resnet neural network with mixup data augmentation. In Proceedings of the Interspeech 2020, 21st Annual Conference of the International Speech Communication Association, Virtual Event, Shanghai, China, 25–29 October 2020; pp. 2902–2906. [Google Scholar]
  29. Demir, F.; Ismael, A.M.; Sengur, A. Classification of Lung Sounds With CNN Model Using Parallel Pooling Structure. IEEE Access 2020, 8, 105376–105383. [Google Scholar] [CrossRef]
  30. Sovijarvi, A.; Dalmasso, F.; Vanderschoot, J.; Malmberg, L.; Righini, G.; Stoneman, S. Definition of terms for applications of respiratory sounds. Eur. Respir. Rev. 2000, 10, 597–610. [Google Scholar]
  31. Meslier, N.; Charbonneau, G.; Racineux, J. Wheezes. Eur. Respir. J. 1995, 8, 1942–1948. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Baughman, R.P.; Loudon, R.G. Lung sound analysis for continuous evaluation of airflow obstruction in asthma. Chest 1985, 88, 364–368. [Google Scholar] [CrossRef]
  33. Cortes, S.; Jane, R.; Fiz, J.; Morera, J. Monitoring of wheeze duration during spontaneous respiration in asthmatic patients. In Proceedings of the 27th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Shanghai, China, 17–18 January 2006; pp. 6141–6144. [Google Scholar]
  34. Qiu, Y.; Whittaker, A.; Lucas, M.; Anderson, K. Automatic wheeze detection based on auditory modelling. Proc. Inst. Mech. Eng. Part H J. Eng. Med. 2005, 219, 219–227. [Google Scholar] [CrossRef] [Green Version]
  35. Zhang, J.; Ser, W.; Yu, J.; Zhang, T. A novel wheeze detection method for wearable monitoring systems. In Proceedings of the IEEE International Symposium on Intelligent Ubiquitous Computing and Education, Chengdu, China, 15–16 May 2009; pp. 331–334. [Google Scholar]
  36. Lin, B.S.; Wu, H.D.; Chen, S.J. Automatic wheezing detection based on signal processing of spectrogram and back-propagation neural network. J. Healthc. Eng. 2015, 6, 649–672. [Google Scholar] [CrossRef] [Green Version]
  37. Kochetov, K.; Putin, E.; Azizov, S.; Skorobogatov, I.; Filchenkov, A. Wheeze detection using convolutional neural networks. In EPIA Conference on Artificial Intelligence; Springer: Cham, Switzerland, 2017; pp. 162–173. [Google Scholar]
  38. Kandaswamy, A.; Kumar, C.S.; Ramanathan, R.P.; Jayaraman, S.; Malmurugan, N. Neural classification of lung sounds using wavelet coefficients. Comput. Biol. Med. 2004, 34, 523–537. [Google Scholar] [CrossRef]
  39. Le Cam, S.; Belghith, A.; Collet, C.; Salzenstein, F. Wheezing sounds detection using multivariate generalized Gaussian distributions. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Taipei, Taiwan, 19–24 April 2009; pp. 541–544. [Google Scholar]
  40. Wisniewski, M.; Zielinski, T.P. Tonality detection methods for wheezes recognition system. In Proceedings of the IEEE 19th International Conference on Systems, Signals and Image Processing (IWSSIP), Vienna, Austria, 11–13 April 2012; pp. 472–475. [Google Scholar]
  41. Wisniewski, M.; Zielinski, T.P. Joint application of audio spectral envelope and tonality index in an e-asthma monitoring system. IEEE J. Biomed. Health Inform. 2015, 19, 1009–1018. [Google Scholar] [CrossRef]
  42. Chien, J.C.; Wu, H.D.; Chong, F.C.; Li, C.I. Wheeze detection using cepstral analysis in gaussian mixture models. In Proceedings of the 29th IEEE Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Lyon, France, 22–26 August 2007; pp. 3168–3171. [Google Scholar]
  43. Bahoura, M. Pattern recognition methods applied to respiratory sounds classification into normal and wheeze classes. Comput. Biol. Med. 2009, 39, 824–843. [Google Scholar] [CrossRef]
  44. Bahoura, M.; Pelletier, C. Respiratory sounds classification using Gaussian mixture models. In Proceedings of the IEEE Canadian Conference on Electrical and Computer Engineering, Niagara Falls, ON, Canada, 2–5 May 2004; Volume 3, pp. 1309–1312. [Google Scholar]
  45. Mayorga, P.; Druzgalski, C.; Morelos, R.; Gonzalez, O.; Vidales, J. Acoustics based assessment of respiratory diseases using GMM classification. In Proceedings of the IEEE Annual International Conference of the IEEE Engineering in Medicine and Biology, Buenos Aires, Argentina, 31 August–4 September 2010; pp. 6312–6316. [Google Scholar]
  46. Taplidou, S.A.; Hadjileontiadis, L.J. Wheeze detection based on time-frequency analysis of breath sounds. Comput. Biol. Med. 2007, 37, 1073–1083. [Google Scholar] [CrossRef] [PubMed]
  47. Jain, A.; Vepa, J. Lung sound analysis for wheeze episode detection. In Proceedings of the 30th IEEE Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Vancouver, BC, Canada, 20–25 August 2008; pp. 2582–2585. [Google Scholar]
  48. Mendes, L.; Vogiatzis, I.; Perantoni, E.; Kaimakamis, E.; Chouvarda, I.; Maglaveras, N.; Tsara, V.; Teixeira, C.; Carvalho, P.; Henriques, J.; et al. Detection of wheezes using their signature in the spectrogram space and musical features. In Proceedings of the 37th IEEE Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Milan, Italy, 25–29 August 2015; pp. 5581–5584. [Google Scholar]
  49. Oletic, D.; Bilas, V. Asthmatic wheeze detection from compressively sensed respiratory sound spectra. IEEE J. Biomed. Health Inform. 2018, 22, 1406–1414. [Google Scholar] [CrossRef]
  50. Torre-Cruz, J.; Canadas-Quesada, F.; García-Galán, S.; Ruiz-Reyes, N.; Vera-Candeas, P.; Carabias-Orti, J. A constrained tonal semi-supervised non-negative matrix factorization to classify presence/absence of wheezing in respiratory sounds. Appl. Acoust. 2020, 161, 107188. [Google Scholar] [CrossRef]
  51. De La Torre Cruz, J.; Cañadas Quesada, F.J.; Carabias Orti, J.J.; Vera Candeas, P.; Ruiz Reyes, N. Combining a recursive approach via non-negative matrix factorization and Gini index sparsity to improve reliable detection of wheezing sounds. Expert Syst. Appl. 2020, 147, 113212. [Google Scholar] [CrossRef]
  52. Nagasaka, Y. Lung Sounds in Bronchial Asthma. Allergol. Int. 2012, 61, 353–363. [Google Scholar] [CrossRef] [Green Version]
  53. Mason, R.C.; Murray, J.F.; Nadel, J.A.; Gotway, M.B. Murray & Nadel’s Textbook of Respiratory Medicine E-Book; Elsevier Health Sciences: Amsterdam, The Netherlands, 2015. [Google Scholar]
  54. Taplidou, S.A.; Hadjileontiadis, L.J. Analysis of wheezes using wavelet higher order spectral features. IEEE Trans. Biomed. Eng. 2010, 57, 1596–1610. [Google Scholar] [CrossRef]
  55. Forgacs, P. The functional basis of pulmonary sounds. Chest 1978, 73, 399–405. [Google Scholar] [CrossRef] [Green Version]
  56. Jácome, C.; Oliveira, A.; Marques, A. Computerized respiratory sounds: A comparison between patients with stable and exacerbated COPD. Clin. Respir. J. 2017, 11, 612–620. [Google Scholar] [CrossRef]
  57. Hashemi, A.; Arabalibiek, H.; Agin, K. Classification of wheeze sounds using wavelets and neural networks. In International Conference on Biomedical Engineering and Technology; IACSIT Press: Singapore, 2011; Volume 11, pp. 127–131. [Google Scholar]
  58. Naves, R.; Barbosa, B.H.; Ferreira, D.D. Classification of lung sounds using higher-order statistics: A divide-and-conquer approach. Comput. Methods Programs Biomed. 2016, 129, 12–20. [Google Scholar] [CrossRef]
  59. Ulukaya, S.; Sen, I.; Kahya, Y.P. A novel method for determination of wheeze type. In Proceedings of the 23nd Signal Processing and Communications Applications Conference (SIU), Malatya, Turkey, 16–19 May 2015; pp. 2001–2004. [Google Scholar] [CrossRef]
  60. Ulukaya, S.; Sen, I.; Kahya, Y.P. Feature extraction using time-frequency analysis for monophonic-polyphonic wheeze discrimination. In Proceedings of the 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Milan, Italy, 25–29 August 2015; pp. 5412–5415. [Google Scholar]
  61. Lee, D.D.; Seung, H.S. Learning the parts of objects by non-negative matrix factorization. Nature 1999, 401, 788–791. [Google Scholar] [CrossRef] [PubMed]
  62. Lee, D.D.; Seung, H.S. Algorithms for non-negative matrix factorization. Adv. Neural Inf. Process. Syst. 2001, 23, 556–562. [Google Scholar]
  63. Canadas-Quesada, F.; Ruiz-Reyes, N.; Carabias-Orti, J.; Vera-Candeas, P.; Fuertes-Garcia, J. A non-negative matrix factorization approach based on spectro-temporal clustering to extract heart sounds. Appl. Acoust. 2017, 125, 7–19. [Google Scholar] [CrossRef]
  64. Dia, N.; Fontecave-Jallon, J.; Gumery, P.Y.; Rivet, B. Denoising Phonocardiogram signals with Non-negative Matrix Factorization informed by synchronous Electrocardiogram. In Proceedings of the 2018 26th IEEE European Signal Processing Conference (EUSIPCO), Rome, Italy, 3–7 September 2018; pp. 51–55. [Google Scholar]
  65. Torre-Cruz, J.; Canadas-Quesada, F.; Vera-Candeas, P.; Montiel-Zafra, V.; Ruiz-Reyes, N. Wheezing sound separation based on constrained non-negative matrix factorization. In Proceedings of the 2018 10th International Conference on Bioinformatics and Biomedical Technology, Amsterdam, The Netherlands, 18–24 May 2018; pp. 18–24. [Google Scholar]
  66. De La Torre Cruz, J.; Cañadas Quesada, F.J.; Ruiz Reyes, N.; Vera Candeas, P.; Carabias Orti, J.J. Wheezing Sound Separation Based on Informed Inter-Segment Non-Negative Matrix Partial Co-Factorization. Sensors 2020, 20, 2679. [Google Scholar] [CrossRef]
  67. Févotte, C.; Bertin, N.; Durrieu, J.L. Nonnegative matrix factorization with the Itakura-Saito divergence: With application to music analysis. Neural Comput. 2009, 21, 793–830. [Google Scholar] [CrossRef]
  68. Liutkus, A.; Fitzgerald, D.; Badeau, R. Cauchy nonnegative matrix factorization. In Proceedings of the IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), New Paltz, NY, USA, 18–21 October 2015; pp. 1–5. [Google Scholar]
  69. Canadas-Quesada, F.J.; Vera-Candeas, P.; Ruiz-Reyes, N.; Carabias-Orti, J.; Cabanas-Molero, P. Percussive/harmonic sound separation by non-negative matrix factorization with smoothness/sparseness constraints. EURASIP J. Audio Speech Music Process. 2014, 2014, 26. [Google Scholar] [CrossRef] [Green Version]
  70. Laroche, C.; Kowalski, M.; Papadopoulos, H.; Richard, G. A structured nonnegative matrix factorization for source separation. In Proceedings of the 2015 23rd IEEE European Signal Processing Conference (EUSIPCO), Nice, France, 31 August–4 September 2015; pp. 2033–2037. [Google Scholar]
  71. Eggert, J.; Korner, E. Sparse coding and NMF. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No. 04CH37541), Budapest, Hungary, 25–29 July 2004; Volume 4, pp. 2529–2533. [Google Scholar]
  72. Virtanen, T. Monaural sound source separation by nonnegative matrix factorization with temporal continuity and sparseness criteria. IEEE Trans. Audio Speech Lang. Process. 2007, 15, 1066–1074. [Google Scholar] [CrossRef] [Green Version]
  73. Marxer, R.; Janer, J. Study of regularizations and constraints in NMF-based drums monaural separation. In Proceedings of the International Conference on Digital Audio Effects Conference (DAFx-13), Maynooth, Ireland, 2–6 September 2013. [Google Scholar]
  74. Prominence Criterion of a Peak According to the MATLAB Software. Available online: https://es.mathworks.com/help/signal/ref/findpeaks.html?searchHighlight=findpeak&s_tid=doc_srchtitle#buff2uu (accessed on 30 December 2020).
  75. The r.a.l.e. Repository. Available online: http://www.rale.ca (accessed on 30 December 2020).
  76. Stethographics Lung Sound Samples. Available online: http://www.stethographics.com (accessed on 30 December 2020).
  77. 3m Littmann Stethoscopes. Available online: https://www.3m.com (accessed on 30 December 2020).
  78. East Tennessee State University Pulmonary Breath Sounds. Available online: http://faculty.etsu.edu (accessed on 30 December 2020).
  79. ICBHI 2017 Challenge. Available online: https://bhichallenge.med.auth.gr (accessed on 30 December 2020).
  80. Lippincott NursingCenter. Available online: https://www.nursingcenter.com (accessed on 30 December 2020).
  81. Thinklabs Digital Stethoscope. Available online: https://www.thinklabs.com (accessed on 30 December 2020).
  82. Thinklabs Youtube. Available online: https://www.youtube.com/channel/UCzEbKuIze4AI1523_AWiK4w (accessed on 30 December 2020).
  83. Emedicine/Medscape. Available online: https://emedicine.medscape.com/article/1894146-overview#a3 (accessed on 30 December 2020).
  84. E-Learning Resources. Available online: https://www.ers-education.org/e-learning/reference-database-of-respiratory-sounds.aspx (accessed on 30 December 2020).
  85. Respiratory Wiki. Available online: http://respwiki.com/Breath_sounds (accessed on 30 December 2020).
  86. Easy Auscultation. Available online: https://www.easyauscultation.com/lung-sounds-reference-guide (accessed on 30 December 2020).
  87. Colorado State University. Available online: http://www.cvmbs.colostate.edu/clinsci/callan/breath_sounds.htm (accessed on 30 December 2020).
  88. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. (TIST) 2011, 2, 1–27. [Google Scholar] [CrossRef]
Figure 1. Time-frequency representation of two examples of Monophonic (MP) wheezing: (A) with a single basal peak; (B) with the harmonics of a single basal peak. Note that the frequency components are harmonically related in (B).
Figure 1. Time-frequency representation of two examples of Monophonic (MP) wheezing: (A) with a single basal peak; (B) with the harmonics of a single basal peak. Note that the frequency components are harmonically related in (B).
Sensors 21 01661 g001
Figure 2. Time-frequency representation of two examples of Polyphonic (PP) wheezing: (A) with two non-harmonically related peaks; (B) with five non-harmonically related peaks. Note that the frequency components are not harmonically related in the case of PP wheezing.
Figure 2. Time-frequency representation of two examples of Polyphonic (PP) wheezing: (A) with two non-harmonically related peaks; (B) with five non-harmonically related peaks. Note that the frequency components are not harmonically related in the case of PP wheezing.
Sensors 21 01661 g002
Figure 3. Flowchart of the proposed method.
Figure 3. Flowchart of the proposed method.
Sensors 21 01661 g003
Figure 4. Example of the estimated matrices B W and B R obtained from the proposed CL-RNMF approach, analyzing the MP wheezing spectrogram previously shown in Figure 1B. (A) Although the matrix B W is composed of four spectral bases, the spectral wheezing patterns are compacted into the fourth basis B W ( 4 ) . This spectral basis B W ( 4 ) is composed of three narrowband spectral peaks. (B) The matrix B R is composed of thirty-two wideband spectral bases.
Figure 4. Example of the estimated matrices B W and B R obtained from the proposed CL-RNMF approach, analyzing the MP wheezing spectrogram previously shown in Figure 1B. (A) Although the matrix B W is composed of four spectral bases, the spectral wheezing patterns are compacted into the fourth basis B W ( 4 ) . This spectral basis B W ( 4 ) is composed of three narrowband spectral peaks. (B) The matrix B R is composed of thirty-two wideband spectral bases.
Sensors 21 01661 g004
Figure 5. The estimated basis matrices B W obtained from CL-RNMF in the examples shown in Section 1. (A) B W for the MP wheezing shown in Figure 1A. (B) B W for the MP wheezing shown in Figure 1B. (C) B W for the PP wheezing shown in Figure 2A. (D) B W for the PP wheezing shown in Figure 2B. The wheezing spectral patterns were compacted into a single basis, B W ( 2 ) (in Case (A)), B W ( 4 ) (in Case (B)), and B W ( 2 ) (in Case (C)). However, the energy of the narrowband spectral peaks was divided into two bases B W ( 1 ) and B W ( 2 ) , as can be seen in Case (D).
Figure 5. The estimated basis matrices B W obtained from CL-RNMF in the examples shown in Section 1. (A) B W for the MP wheezing shown in Figure 1A. (B) B W for the MP wheezing shown in Figure 1B. (C) B W for the PP wheezing shown in Figure 2A. (D) B W for the PP wheezing shown in Figure 2B. The wheezing spectral patterns were compacted into a single basis, B W ( 2 ) (in Case (A)), B W ( 4 ) (in Case (B)), and B W ( 2 ) (in Case (C)). However, the energy of the narrowband spectral peaks was divided into two bases B W ( 1 ) and B W ( 2 ) , as can be seen in Case (D).
Sensors 21 01661 g005
Figure 6. The spectral energy distribution ξ ( f ) provided by CL-RNMF from the estimated basis matrix B W shown in Figure 5: (A) Figure 5A. (B) Figure 5B. (C) Figure 5C. (D) Figure 5D.
Figure 6. The spectral energy distribution ξ ( f ) provided by CL-RNMF from the estimated basis matrix B W shown in Figure 5: (A) Figure 5A. (B) Figure 5B. (C) Figure 5C. (D) Figure 5D.
Sensors 21 01661 g006
Figure 7. Example of the proposed two-step procedure to classify between MP and PP wheezing when η > 1 from the example of MP wheezing shown in Figure 1B. Note that the arrows indicate the narrowband spectral peaks that compose the wheezing. In this case, the wheezing is classified as MP because all spectral peaks are harmonically related.
Figure 7. Example of the proposed two-step procedure to classify between MP and PP wheezing when η > 1 from the example of MP wheezing shown in Figure 1B. Note that the arrows indicate the narrowband spectral peaks that compose the wheezing. In this case, the wheezing is classified as MP because all spectral peaks are harmonically related.
Sensors 21 01661 g007
Figure 8. Example of the proposed two-step procedure to classify between MP and PP wheezing when η > 1 , considering the two examples of PP wheezing shown in Figure 2. (A) Two-step procedure applied to the PP wheezing shown in Figure 2A. (B) Two-step procedure applied to the PP wheezing shown in Figure 2B. Note that the arrows indicate the narrowband spectral peaks that compose the wheezing. In this case, both wheezing are classified as PP because not all spectral peaks are harmonically related.
Figure 8. Example of the proposed two-step procedure to classify between MP and PP wheezing when η > 1 , considering the two examples of PP wheezing shown in Figure 2. (A) Two-step procedure applied to the PP wheezing shown in Figure 2A. (B) Two-step procedure applied to the PP wheezing shown in Figure 2B. Note that the arrows indicate the narrowband spectral peaks that compose the wheezing. In this case, both wheezing are classified as PP because not all spectral peaks are harmonically related.
Sensors 21 01661 g008
Figure 9. Scheme of the types of wheezing contained in the database.
Figure 9. Scheme of the types of wheezing contained in the database.
Sensors 21 01661 g009
Figure 10. LOO cross-validation scheme for the database described in this paper.
Figure 10. LOO cross-validation scheme for the database described in this paper.
Sensors 21 01661 g010
Table 1. Definition of the terms that appear in the metrics detailed in Equations (22)–(26).
Table 1. Definition of the terms that appear in the metrics detailed in Equations (22)–(26).
TermsDefinitions
T P (True PP)PP wheezing segments correctly classified
T M (True MP)MP wheezing segments correctly classified
F P (False PP)PP wheezing segments misclassified as MP
F M (False MP)MP wheezing segments misclassified as PP
T M 1 (True MP Type 1)MP Type 1 wheezing segments correctly classified
T M 2 (True MP Type 2)MP Type 2 wheezing segments correctly classified
F M 1 (False MP Type 1)MP Type 1 wheezing segments misclassified as PP
F M 2 (False MP Type 2)MP Type 2 wheezing segments misclassified as PP
Table 2. Comparative ACC results between the proposed method and UPER.
Table 2. Comparative ACC results between the proposed method and UPER.
Algorithm ACC G ACC P ACC M ACC M 1 ACC M 2
Proposed Method92%91.5%92.5%91%94%
UPER (SVM) [7]83.75%87.5%80%86%74%
UPER (KNN) [7]80%84.4%75.5%81%70%
UPER (ELM) [7]81.5%86%77%83%71%
Table 3. Comparative results, in terms of A C C G , between the three classifier versions (SVM, KNN, and ELM) of the method UPER using four Leave-p-Out (LPO) cross-validation schemes.
Table 3. Comparative results, in terms of A C C G , between the three classifier versions (SVM, KNN, and ELM) of the method UPER using four Leave-p-Out (LPO) cross-validation schemes.
SchemeTraining SetValidate SetSVMKNNELM
LOO399 (99.75%)1 (0.25%)83.75%80%81.5%
LPO ( p = 80 )320 (80%)80 (20%)81.5%79.25%80%
LPO ( p = 160 )240 (60%)160 (40%)80.5%77.75%79.5%
LPO ( p = 240 )160 (40%)240 (60%)78.25%74.75%77.25%
LPO ( p = 320 )80 (20%)320 (80%)76.25%71.75%75.25%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

De La Torre Cruz, J.; Cañadas Quesada, F.J.; Ruiz Reyes, N.; García Galán, S.; Carabias Orti, J.J.; Peréz Chica, G. Monophonic and Polyphonic Wheezing Classification Based on Constrained Low-Rank Non-Negative Matrix Factorization. Sensors 2021, 21, 1661. https://doi.org/10.3390/s21051661

AMA Style

De La Torre Cruz J, Cañadas Quesada FJ, Ruiz Reyes N, García Galán S, Carabias Orti JJ, Peréz Chica G. Monophonic and Polyphonic Wheezing Classification Based on Constrained Low-Rank Non-Negative Matrix Factorization. Sensors. 2021; 21(5):1661. https://doi.org/10.3390/s21051661

Chicago/Turabian Style

De La Torre Cruz, Juan, Francisco Jesús Cañadas Quesada, Nicolás Ruiz Reyes, Sebastián García Galán, Julio José Carabias Orti, and Gerardo Peréz Chica. 2021. "Monophonic and Polyphonic Wheezing Classification Based on Constrained Low-Rank Non-Negative Matrix Factorization" Sensors 21, no. 5: 1661. https://doi.org/10.3390/s21051661

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop