Next Article in Journal
A Data-Driven Model for Aerodynamic Loads on Road Vehicles Exposed to Gusty Bora-Like Winds
Next Article in Special Issue
ME-YOLO: Improved YOLOv5 for Detecting Medical Personal Protective Equipment
Previous Article in Journal
Microstructural Variation and Evaluation of Formability According to High-Temperature Compression Conditions of AMS4928 Alloy
Previous Article in Special Issue
Using a Video Device and a Deep Learning-Based Pose Estimator to Assess Gait Impairment in Neurodegenerative Related Disorders: A Pilot Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Progressively Expanded Database for Automated Lung Sound Analysis: An Update

1
Graduate Institute of Biomedical Electronics and Bioinformatics, National Taiwan University, Taipei 10617, Taiwan
2
Cardiovascular Intensive Care Unit, Far Eastern Memorial Hospital, New Taipei 22060, Taiwan
3
Heroic Faith Medical Science Co., Ltd., Taipei 11049, Taiwan
4
Avalanche Computing Inc., Taipei 10687, Taiwan
5
Department of Life Science, College of Life Science, National Taiwan University, Taipei 10617, Taiwan
6
Institute of Biomedical Sciences, Academia Sinica, Taipei 11529, Taiwan
7
HCC Healthcare Group, New Taipei 22060, Taiwan
8
Division of Thoracic Surgery, Department of Surgery, Far Eastern Memorial Hospital, New Taipei 22060, Taiwan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(15), 7623; https://doi.org/10.3390/app12157623
Submission received: 18 May 2022 / Revised: 15 July 2022 / Accepted: 25 July 2022 / Published: 28 July 2022
(This article belongs to the Special Issue Artificial Intelligence in Medicine and Healthcare)

Abstract

:

Featured Application

Auscultatory lung sound analysis in healthcare.

Abstract

We previously established an open-access lung sound database, HF_Lung_V1, and developed deep learning models for inhalation, exhalation, continuous adventitious sound (CAS), and discontinuous adventitious sound (DAS) detection. The amount of data used for training contributes to model accuracy. In this study, we collected larger quantities of data to further improve model performance and explored issues of noisy labels and overlapping sounds. HF_Lung_V1 was expanded to HF_Lung_V2 with a 1.43× increase in the number of audio files. Convolutional neural network–bidirectional gated recurrent unit network models were trained separately using the HF_Lung_V1 (V1_Train) and HF_Lung_V2 (V2_Train) training sets. These were tested using the HF_Lung_V1 (V1_Test) and HF_Lung_V2 (V2_Test) test sets, respectively. Segment and event detection performance was evaluated. Label quality was assessed. Overlap ratios were computed between inhalation, exhalation, CAS, and DAS labels. The model trained using V2_Train exhibited improved performance in inhalation, exhalation, CAS, and DAS detection on both V1_Test and V2_Test. Poor CAS detection was attributed to the quality of CAS labels. DAS detection was strongly influenced by the overlapping of DAS with inhalation and exhalation. In conclusion, collecting greater quantities of lung sound data is vital for developing more accurate lung sound analysis models.

1. Introduction

Respiration is an essential vital sign. Changes in the frequency or intensity of respiratory lung sounds and the manifestation of continuous adventitious sounds (CASs), such as wheezes, stridor, squawk, gasp, and rhonchi, and discontinuous adventitious sounds (DASs), such as crackles and pleural friction rubs, are associated with pulmonary disorders, including pneumonia, asthma, chronic obstructive pulmonary disease (COPD), and cystic fibrosis [1,2]. Detection of adventitious sounds during lung sound auscultation helps health-care professionals make clinical decisions. However, the interpretation of lung sounds auscultated with a traditional stethoscope involves inherent subjectivity and relies on auscultation capability [3]. Electronic auscultation and computerized lung sound analysis can facilitate the interpretation and overcome the above limitations [4].
Computerized lung sound analysis methods based on machine learning have been comprehensively reviewed [5,6,7]. Some previous machine learning studies focused on classifying individuals into a healthy group or groups with specific types of respiratory disorders according to their lung sounds. Some studies developed algorithms to classify normal lung sounds and various types of adventitious lung sounds. Some studies proposed methods to detect breath phases and adventitious sounds in lung sound recordings. Over the last few years, more and more researchers have turned their attention to using deep learning for computerized lung sound analysis. Various architectures of deep neural networks (DNN), such as convolutional neural network (CNN) [8,9,10,11,12,13,14,15,16,17,18], recurrent neural network [9,13,18,19,20,21], residual neural network [22,23], temporal convolutional network [24], regional CNN [25], and DNN with attention mechanism [26,27] have been investigated in developing automated lung sound analysis systems. However, many studies were limited by small data volumes, and some datasets used in these studies were private so a comparison between the proposed models cannot be realized. As data size plays an important role to train a more accurate deep learning model, a larger open-access dataset can greatly help the development of lung sound analysis methods.
Some public databases of respiratory sound have been established. Rocha et al. [28] established an open-access database with which new algorithms for respiratory sound classification can be evaluated and compared. This database, compiled for the first scientific challenge of the International Federation for Medical and Biological Engineering International Conference on Biomedical and Health Informatics (ICBHI) 2017, included 920 audio recordings from 126 subjects and text files showing disease diagnosis and precise location of respiratory cycles, wheezes, and crackles. In addition, Fraiwan and colleagues [29] established a dataset of lung sounds recorded from healthy and unhealthy subjects which included disease labels, such as asthma, heart failure, pneumonia, bronchitis, and COPD. They focused on designing automated machine learning algorithms for the detection of pulmonary disorders. Previously, we reported the development of the world’s largest open-access lung sound database, HF_Lung_V1 (https://gitlab.com/techsupportHF/HF_Lung_V1, accessed on 18 January 2022). Different from the ICBHI 2017 database, we not only labeled respiratory cycles, wheezes, and crackles, but also labeled inhalations, exhalations, stridor, and rhonchi. However, the HF_Lung_V1 database did not include labels of disease diagnosis.
HF_Lung_V1 is benchmarked with several variants of recurrent neural networks [18]. However, while the HF_Lung_V1 database shows high potential for exploiting deep learning for automated inhalation, exhalation, CAS, and DAS detection, it remains to be further improved to achieve better accuracy. Because the performance of deep learning models is positively associated with the size of the training set [30], an increase in the number of sound recordings and labels is expected to improve the database. In addition, given the possible presence of noisy labels, label quality needs to be further optimized. Therefore, the aim of this study was to present an updated database, HF_Lung_V2, which is based on HF_Lung_V1 but expanded to include more high-quality lung sound files and labels. The performance of the detection model trained with HF_Lung_V2 was also evaluated.

2. Materials and Methods

2.1. Participants

We acquired new lung sound recordings from two sources to form a new dataset, HF_Lung_V1_incremental_package (HF_Lung_V1_IP), which can be downloaded at https://gitlab.com/techsupportHF/HF_Lung_V1_IP (accessed on 13 April 2022). By adding the lung sound recordings in HF_Lung_V1_IP with those in HF_Lung_V1, we could establish the lung sound database, HF_Lung_V2.
In HF_Lung_V1_IP, the lung sounds from the first source were collected from seven inpatients receiving long-term mechanical ventilation support in a respiratory care ward (RCW) or respiratory care center (RCC) between October 2019 and December 2019. The lung sounds from the second source were obtained from another 32 inpatients with adventitious sounds at Far Eastern Memorial Hospital (FEMH), Taipei, Taiwan between January 2019 and November 2019. All participants were Taiwanese and aged ≥20 years.
Table 1 summarizes the demographic data and the characteristics of respiratory disease of the subjects in HF_Lung_V2. Note that not every FEMH inpatient was with respiratory disease. We did not have the information about the subjects whose lung sound recordings were acquired and provided by the Taiwan Society of Emergency and Critical Care Medicine (TSECCM). The information about subjects’ other co-occurring diseases can be found in Table S1.

2.2. Lung Sound Recordings

We used two different devices to record the participants’ lung sounds. The first device was a commercial handheld electronic stethoscope, Littmann 3200 (3M, Saint Paul, MN, USA); the second one was a customized auscultation device, HF-Type-1 [18], which supported multichannel continuous sound recording through attachable acoustic sensors. We followed the same recording protocols described in our previous study to acquire the lung sounds from the participants [18]. Eight auscultation locations (Figure 1a) were firstly defined—L1: second intercostal space (ICS) on the right midclavicular line (MCL); L2: fifth ICS on the right MCL; L3: fourth ICS on the right midaxillary line (MAL); L4: tenth ICS on the right MAL; L5: second ICS on the left MCL; L6: fifth ICS on the left MCL; L7: fourth ICS on the left MAL; and L8: tenth ICS on the left MAL. Briefly, when the Littmann 3200 device was used, to complete a standard round of recording, lung sounds must be auscultated sequentially from L1 to L8 (white arrows in the top graph in Figure 1b). When the HF-Type-1 device was used, acoustic sensors were attached at L1, L2, L4, L5, L6, and L8, and lung sounds were continuously recorded at the same time. A standard round of lung sound recording with the HF-Type-1 device consisted of the acquisition of continuous signals from each of these six locations over 30 min (white arrows in the bottom graph in Figure 1b).
The lung sounds of the seven RCW/RCC inpatients were obtained with the Littmann 3200 device for 4–5 rounds and the HF-Type-1 device for 3–4 rounds. The lung sounds of the 32 inpatients at FEMH were recorded using the Littmann 3200 device alone for 1–3 rounds. The sampling rate of the recording devices was 4000 Hz, and the bit depth was 16 bits. We did not ask the participants to take deep breaths during sound recording.

2.3. Data Preparation and Labeling

The length of audio files recorded with the Littmann 3200 device was originally set at 15.8 s; as such, the terminal 0.8-s segment was deleted to obtain 15-s audio files (red blocks in Figure 1b). A standard round of continuous recordings of lung sound generated by the HF-Type-1 device was longer than 30 min; therefore, we truncated the first 15-s segment of every 2-min signal in the continuous recordings for subsequent analysis (red blocks in Figure 1b).
Two licensed respiratory therapists with 8 and 4 years of clinical experience, respectively, and one registered nurse with 13 years of clinical experience performed the labeling. The labelers’ licenses were granted by the Ministry of Examination, Taiwan, and they gained their clinical experience in Taiwan. Each lung sound file was labeled by only one individual, but regular consensus meetings were held to evaluate the labeling and reach a consensus among all labelers. A self-developed software [31] was used to do the labeling. During labeling, the labelers first used the software to load a 15-s lung sound recording (raw signal displayed in Figure 2a), and the corresponding spectrogram (green square in Figure 2b) would be displayed subsequently. The labelers were asked to listen to the sound and observe its corresponding temporal-spectral patterns on the spectrogram once or twice first. Then, they must label all the inhalations and exhalations in the 15-s recording. The labeler can move the yellow lines (indicated by white arrows in Figure 2b) to define a region of interest (ROI) (transparent yellow patch in Figure 2b). By clicking the area of which start and end times were bounded by the ROI on specific label tracks (red arrows in Figure 2b), the labelers can generate corresponding labels. After having the labels of inhalations (I) and exhalations (E), they can proceed to label wheezes, stridor, rhonchi, and DASs. When labeling DASs, they were asked to label the start and end times of a series of crackle sounds but not label each explosive sound; additionally, the sounds of pleural friction rubs were not included in the DAS labels (D) in this study. During the labeling process, they can replay the sound and make a modification to the labels as many times as they wanted. They were instructed not to label if a sound was inaudible. If a sound was cut off at the very beginning or end of a recording which made them have less confidence in labeling, they can skip the labeling of it. After labeling, the software automatically created a label file. The output of the label started with the type of label (I, E, Wheeze, Stridor, Rhonchi, or D), and it was followed by the start and end times of the event, shown in Figure 2c.
Note that during deep-learning processing, labels of wheezes (W), stridor (S), and rhonchi (R) were combined to obtain CAS labels (C).

2.4. Dataset Arrangement

Because we recorded several rounds of the lung sounds from one individual within a short period of time (few hours), the sound patterns collected at the same location may bear similarity. Therefore, to fairly evaluate the performance of the trained models, we put the lung sound recordings collected from the same subject to either training data set or test data set. The ratio of the numbers of recordings in the training and test set was maintained at approximately 4:1.

2.5. Benchmarking of HF_Lung_V2

The convolutional neural network with bidirectional gated recurrent unit (CNN–BiGRU) model is presented in Figure 3. The formulas of ReLU [32] and Sigmoid [32] layers are as follows:
ReLU (x) = max (x, 0),
Sigmoid (x) = 1/(1 + exp (−x)).
The CNN–BiGRU model outperformed all benchmark models used in our previous study [18]. Therefore, the model was used to benchmark HF_Lung_V2 in the present study.
Figure 4 shows the pipeline for preprocessing, deep-learning processing, and postprocessing. All the codes were written using Python 3.7. During the preprocessing, we first used a 10th-order Butterworth high-pass filter with a cutoff frequency of 80 Hz to process the raw 15-s signal. Then, we computed the spectrogram of the filtered signal using short-time Fourier transform (nfft = 256, window length = 256, window type = hann, overlap = 0.75). The output of the short-time Fourier transform was a 938 × 129 array, where 938 was the number of time segments and 129 was the number of frequency bins. Next, Mel-frequency cepstral coefficients (MFCCs) [33], including 20 static, 20 delta, and 20 delta–delta coefficients were calculated for each time segment. Finally, we calculated energy summation in four frequency bands, 0–250, 250–500, 500–1000, and 0–2000 Hz, for each time segment. Min–max normalization was applied to each of the spectrogram, MFCCs, and energy summation. Then, the concatenation of normalized spectrogram, MFCCs, and energy summation, a 938 × 193 array, were fed as input to train the CNN-BiGRU model and generate the results of segment prediction. We used Tensorflow 2.10 framework to construct the layers of neural networks. During training, we used Adam as the optimizer, and batch size was set to 64. We set the initial learning rate to 0.0001 with a step decay (0.2×) when the validation loss did not decrease for 10 epochs. The learning process stopped when no improvement occurred over 50 consecutive epochs. The training was done on an Ubuntu 18.04 server provided by the National Center for High-Performance Computing in Taiwan (Taiwan Computing Cloud (TWCC)). It was equipped with an Intel(R) Xeon(R) Gold 6154 @3.00 GHz CPU with 90 GB RAM. We used the CUDA 10, and CuDNN 7 library to execute the training and inference on an NVIDIA Titan V100 graphics card for GPU acceleration. Finally, we refined the results of segment prediction by merging the connected segments with positive prediction and those with an interval smaller than 0.5 s to generate the detected events. Subsequently, the events with a duration smaller than 0.05 s were seen as burst noises and removed. More details can be found in our previous study [18].
Subject-wise five-fold cross-validation was conducted during the training process. To obtain more stable results, this procedure was repeated thrice. To mitigate data imbalance, we used the audio files containing at least one event of CASs and DASs to train and test the respective models.

2.6. Performance Evaluation

Figure 5 illustrates how we evaluated the performance of segment- and event-detection. First, the ground-truth event labels (red horizontal bars in Figure 5a) were used to generate the ground-truth time segments (red vertical bars in Figure 5b). Next, after we conducted binary thresholding on the output of the model to determine if the probability was high enough to be viewed as a detected segment (a blue vertical bar in Figure 5c), we could obtain the results of segment prediction. By comparing the results of segment prediction (blue vertical bars in Figure 5c) with the ground-truth time segments (Figure 5b), we defined true-positive (TP; orange vertical bars in Figure 5e), true-negative (TN; green vertical bars in Figure 5e), false-positive (FP; black vertical bars in Figure 5e), and false-negative (FN; yellow vertical bars in Figure 5e) time segments.
For event detection, the segment prediction results were post-processed (see Section 2.5) to obtain the event prediction results (Figure 5d). Ground-truth event labels and the results of event prediction were alternately employed as references to evaluate the accuracy of event detection (evaluation using ground-truth labels as reference is displayed in Figure 5f, and evaluation using the results of event prediction as reference is displayed in Figure 5g). The Jaccard index (JI) [34], defined as follows,
Jaccard   Index   ( JI ) = ( target   event referenced   event ) ( target   event referenced   event ) ,
was used to evaluate the extent to which a target event overlapped with a referenced event. We defined a JI of ≥0.5 as a correct match for generating a TP event (orange horizontal bars in Figure 5f,g). If a ground-truth event did not have a correctly matched detected event, it was defined as an FN event (yellow horizontal bars in Figure 5f). If a detected event did not have a correctly matched ground-truth event, it was considered an FP event (black horizontal bars in Figure 5g). A TN event cannot be defined in the event detection evaluation process. Because the TP events were double-counted (orange horizontal bars in Figure 5f,g), a pair of TP events was viewed as a single TP event. However, all FN (Figure 5f) and FP events (Figure 5g) events were taken into computing the model performance even though some FN events overlapped with some FP events, which may have led to slightly biased results.
Accuracy, sensitivity, specificity, positive predictive value (PPV), and F1 score were computed to evaluate the performance of segment detection. Additionally, a receiver operating characteristic (ROC) curve was plotted, and the corresponding 95% confidence interval was displayed to help visualize the performance of segment detection between models. Area under the ROC curve (AUC) was calculated. Event detection was assessed using sensitivity, PPV, and F1 score. We selected a threshold generating the best accuracy of segment detection to calculate all the other performance indexes. The definitions of the performance indexes are described as follows:
accuracy = (TP + TN)/(TP + TN + FP + FN),
sensitivity = TP/(TP + FN),
specificity = (TN)/(TN + FP),
PPV = TP/(TP + FP),
F1 score = 2 × (sensitivity × PPV)/(sensitivity + PPV).
Performance measurements were averaged and reported for 15 models obtained from the repeated five-fold cross-validation. The measures of accuracy, sensitivity, specificity, PPV, and F1 score between the models trained using V1_Train and V2_Train were statistically compared using a Student’s t-test, while the AUC was compared using a Mann–Whitney U test.

2.7. Investigation of Label Quality and Sound Overlapping

Two of the various factors indicated in our previous study may have influenced model performance and thereby required investigation [18]: noisy labels and sound overlapping. First, label quality was reviewed to evaluate what types of patterns may lead to noisy labels. Second, the density of the start and end times of each type of I, E, C, and D labels in the V2_Train and V2_Test was calculated and plotted. Subsequently, the overlap ratios between I, E, C, and D labels were calculated.

3. Results

3.1. Summary of HF_Lung_V1 and HF_Lung_V2 Databases

Statistical data of lung sound files and labels employed in HF_Lung_V1 and HF_Lung_V2 are presented in Table 2. The number of participants increased from 261 in HF_Lung_V1 to 300 in HF_Lung_V2. Moreover, HF_Lung_V2 had 1.43 times the number of 15-s records (13,957) compared to those in HF_Lung_V1 (9765), and the total duration increased from 2441.25 min to 3489.25 min. In addition, the number of I, E, C, and D labels increased from 34,095 to 49,373, from 18,349 to 24,552, from 13,883 to 21,558, and from 15,606 to 19,617, respectively.
The statistics of the lung sound files and labels grouped by data sets can be found in Table 3. The numbers of 15-s files in the training and test data sets in HF_Lung_V1 were 7809 and 1956, respectively; and those in the training and test data sets in HF_Lung_V2 increased to 10554 and 3403, respectively. The number of 15-s files recorded with the Littmann 3200 device increased from 4504 in HF_Lung_V1 to 5156 in HF_Lung_V2, and those recorded with the HF-Type-1 device increased from 5261 in HF_Lung_V1 to 8801 in HF_Lung_V2 (Table S2).

3.2. Performance Benchmark

Performance measurement of the CNN-BiGRU models in segment and event detection on V1_Test and V2_Test is presented in Table 4. We can see that the models trained using V2_Train performed significantly better than those trained using V1_Train in terms of most measures in the segment- and event-detection of inhalation and CAS on V1_Test or V2_Test. The improvement was also significant in the segment- and event-detection of exhalation on V1_Test. However, the models trained using V2_Train only showed slightly improved performance in event detection of DAS on V1_Test. Superior performance of the models trained using V2_Train in inhalation, exhalation, and CAS detection were shown in the ROC curves derived from V1_Test (Figure 6a–c) and V2_Test (Figure 6e–g). However, the ROC curves of the models greatly overlapped with each other in DAS detection (Figure 6d,h).

3.3. Review of Label Quality

Figure 7 illustrates the spectrograms (top graphs in Figure 7a–e) and labels (bottom graphs in Figure 7a–e) of five lung sound recordings. The quality of I and E labels was acceptable. In most cases, inhalations and exhalations can be clearly heard and identified, such as in the example displayed in Figure 7a. However, we can occasionally see wrong labels, such as the exhalation indicated by the white arrow in Figure 7b, or debatable labels, such as the exhalations indicated by the blue arrows in Figure 7b. An exhalation was sometimes not identifiable, such as the sound displayed in Figure 7c.
However, the quality of C labels was unsatisfactory. CASs were often found unlabeled, such as the ones (green arrows) in Figure 7c. In some cases, the labelers were unsure whether to label a borderline case. For example, the labeler did not label the sound indicated by the first green arrow in Figure 7d but labeled the sounds indicated by the second and third green arrows although these sounds did not form clear streak patterns on the spectrogram.
Regarding the quality of D labels, even experts could not arrive at a consensus on the correct labeling of a given DAS. For instance, many noises may have DAS-like patterns and may interfere with the labeling process (purple arrow in Figure 7e). Furthermore, the DAS patterns in Figure 7b are sometimes not as clear as those displayed in Figure 7e.
The density plots of the start and end times of I, E, C, and D labels in the V2_Train and V2_Test are displayed in Figure 8. The density of the start and end times of the I (blue lines in Figure 8a–d), E (orange lines in Figure 8a–d), C (green lines in Figure 8a–d), and D (red lines in Figure 8a–d) labels was lower at the beginning and end of the lung sound recordings. However, the density curves reached a higher plateau roughly between 1–14 s for all I, E, C, and D labels in the V2_Train (blue, orange, red, and green lines in Figure 8a,b), though the density curves of I, E, and D labels fluctuated between 1–14 s in the V2_Test (blue, orange, and green lines in Figure 8c,d).

3.4. Overlap Ratios between Labels

Table 5 lists the ratios that indicate the extent to which specific types of labels overlapped with the other types of labels in the V2_Train and V2_Test; 97.7% (58.4% + 39.3%) of the duration of D labels overlapped with the I and E labels in V2_Train, and 96.7% (66.1% + 30.6%) of the duration of D labels overlapped with the I and E labels in V2_Test.

4. Discussion

The present study reports the process of expanding HF_Lung_V1 to HF_Lung_V2. Performance of the CNN–BiGRU model showed improvement in inhalation, exhalation, CAS, and DAS detection accuracy as the data size increased in HF_Lung_V2, although the improvement in DAS detection may not be significant. The improvement can be clearly observed in Table 4: the models trained on the basis of V2_Train performed better on both V1_Test and V2_Test compared to those trained on the basis of V1_Train. The improvement in segment detection can also be observed in Figure 5: the ROC curves and the 95% confidence intervals of the models trained using V2_Train (red lines and red shades) located closer to the left upper corner than those trained using V1_Train (blue lines and blue shades).
Small data regions, small power-law regions, and irreducible error regions are present in the curve of the power law of learning [30]. The generalization error (log-scale) decreases as the training set size (log-scale) in the power-law region increases [30]. We did not investigate whether the increase in the size of HF_Lung_V2 (which is 1.43 times that of HF_Lung_V1) was in the power-law region. The promising improvement in the performance of inhalation, exhalation, CAS, and DAS detection encourages us to continue collecting more breathing lung sounds and to build a larger dataset.
Notably, the benchmark performance of the CNN-BiGRU models trained using V1_Train and tested on V1_Test differed from those reported in our previous study [18], which is due to retraining the models. Retraining was performed because event detection evaluation was performed somewhat differently. Specifically, unlike in our previous work [18], ground-truth event labels and the results of event prediction were used alternately as references. Thus, an FP event and an FN event could have been counted heedlessly within the same period. Furthermore, five-fold cross-validation was repeated thrice, and the measurements were averaged and reported in this study.
Label quality review revealed that the quality of I and E labels was acceptable. However, the quality of C labels was unsatisfactory. This may have occurred because we asked the labelers to perform the labeling mostly on the basis of what they heard and to use the spectrogram only as an aid. We also asked them not to label a sound without being sure the label was appropriate. Furthermore, some noisy labels may have resulted from mistakes made in operating the labeling software, such as the E label indicated by the white arrow in Figure 7b. In addition, it is especially difficult to establish clear criteria to decide whether a sound can be counted as a CAS or to which types of CAS a sound belongs. Many borderline rhonchi did not form a clear streak pattern; thus, the labelers were unsure whether to label them (green arrows in Figure 7d). Upon the exclusion of R labels from the C labels, the model exhibited superior performance in CAS detection (data not shown). As for D labels, the evaluation quality was strongly influenced by the indifferentiable rubbing-sound noises. The labelers also tended to ignore the sounds located at the very beginning and end of an audio file because the event was cut off or because the full duration of an event was not heard or observed (white squares in Figure 7a,c,d). This phenomenon can be clearly observed in the density plots (Figure 8a–d), which show that the density of the start and end times of labels was lower at the beginning and end of the audio files. This is another source of noisy labels that may cause confusion during CNN training. Additionally, the lower density at the beginning and end of an audio file may cause some false negative events in real use of the models, which is associated with the use of BiGRU networks in our models. The fluctuation of the density curves of the start and end times of the I, E, and D labels between the 1–14 s in the V2_Test (Figure 7c,d) probably resulted from smaller numbers of labels compared to those in the V2_Train.
DAS detection performance was greatly influenced by sound overlapping. We labeled only the start and end times of a series of DASs (mean duration of each series: 0.86 s; Table 2); however, each DAS, i.e., the mean duration of fine and coarse crackle sounds: approximately 5 and 15 ms [6], respectively, was not labeled. This caused the overlapping of more than 96% of D labels with the I and E labels (Table 5). Moreover, because only audio recordings containing D labels were used to train the DAS detection model, the trained model showed a tendency to incorrectly identify inhalations and exhalations as DASs. A new dataset that contains more balanced data or a new training strategy is required to solve this problem in future studies.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/app12157623/s1, Table S1: Characteristics of co-occurring diseases in the subjects in HF_Lung_V2; Table S2: Statistics of lung sound files and labels in the HF_Lung_V1 and HF_Lung_V2 databases grouped by recording devices.

Author Contributions

Conceptualization, F.-S.H., S.-R.H. and Y.-R.C.; methodology, S.-R.H., C.-W.H., Y.-R.C. and C.-C.C.; formal analysis, C.-W.H. and C.-C.C.; resources, J.H. and C.-W.C.; writing—original draft preparation, F.-S.H. and S.-R.H.; writing—review and editing, F.-S.H., S.-R.H. and F.L.; project administration, F.-S.H. and F.L.; funding acquisition, F.-S.H.; investigation, J.H. and C.-W.C. All authors have read and agreed to the published version of the manuscript.

Funding

This study is supported by the Ministry of Science and Technology, Taiwan, R.O.C. under Grant no. MOST 109-EC-17-A-22-I3-0009.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and the lung sound recording was approved by the Institutional Review Board of Far Eastern Memorial Hospital on 19 August 2018 (protocol code 107052-F).

Informed Consent Statement

Written informed consent was obtained from all subjects involved in the study (patients or patients’ legal representatives). Written informed consent has been obtained from the patients to publish this paper.

Data Availability Statement

HF_Lung_V1, including lung sound recordings and labels, can be downloaded at https://gitlab.com/techsupportHF/HF_Lung_V1 (accessed on 18 January 2022). The lung sound recordings in HF_Lung_V1_IP can be downloaded at https://gitlab.com/techsupportHF/HF_Lung_V1_IP (accessed on 13 April 2022). However, the labels in HF_Lung_V1_IP are proprietary property owned by Heroic Faith. Please send an email to [email protected] to request access to the labels.

Acknowledgments

Taiwan’s Raising Children Medical Foundation sponsored the lung sound collection activity. Heroic Faith Medical Science Co., Ltd. sponsored the data labeling and deep learning model training. The authors thank Taiwan’s National Center for High-Performance Computing for providing the computing resources. The authors also thank the employees of Heroic Faith Medical Science Co., Ltd., who contributed to the establishment of the HF_Lung_V2 database.

Conflicts of Interest

F.-S.H., S.-R.H. and Y.-R.C. are employees of Heroic Faith Medical Science Co., Ltd. C.-W.H. and C.-C.C. are with Avalanche Computing Inc., who are commissioned by Heroic Faith Medical Science to train the AI models.

References

  1. Bohadana, A.; Izbicki, G.; Kraman, S.S. Fundamentals of lung auscultation. N. Engl. J. Med. 2014, 370, 744–751. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Fouzas, S.; Anthracopoulos, M.B.; Bohadana, A. Clinical Usefulness of Breath Sounds. In Breath Sounds; Priftis, K.N., Hadjileontiadis, L.J., Everard, M.L., Eds.; Springer: Berlin/Heidelberg, Germany, 2018; pp. 33–52. [Google Scholar]
  3. Kim, Y.; Hyon, Y.; Lee, S.; Woo, S.-D.; Ha, T.; Chung, C. The coming era of a new auscultation system for analyzing respiratory sounds. BMC Pulm. Med. 2022, 22, 119. [Google Scholar] [CrossRef]
  4. Gurung, A.; Scrafford, C.G.; Tielsch, J.M.; Levine, O.S.; Checkley, W. Computerized lung sound analysis as diagnostic aid for the detection of abnormal lung sounds: A systematic review and meta-analysis. Respir. Med. 2011, 105, 1396–1403. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Pramono, R.X.A.; Bowyer, S.; Rodriguez-Villegas, E. Automatic adventitious respiratory sound analysis: A systematic review. PLoS ONE 2017, 12, e0177926. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Muthusamy, P.D.; Sundaraj, K.; Abd Manap, N. Computerized acoustical techniques for respiratory flow-sound analysis: A systematic review. Artif. Intell. Rev. 2020, 53, 3501–3574. [Google Scholar] [CrossRef]
  7. Rocha, B.M.; Pessoa, D.; Marques, A.; Carvalho, P.; Paiva, R.P. Automatic classification of adventitious respiratory sounds: A (un) solved problem? Sensors 2020, 21, 57. [Google Scholar] [CrossRef] [PubMed]
  8. Demir, F.; Sengur, A.; Bajaj, V. Convolutional neural networks based efficient approach for classification of lung diseases. Health Inf. Sci. Syst. 2020, 8, 4. [Google Scholar] [CrossRef] [Green Version]
  9. Fraiwan, M.; Fraiwan, L.; Alkhodari, M.; Hassanin, O. Recognition of pulmonary diseases from lung sounds using convolutional neural networks and long short-term memory. J. Ambient Intell. Humaniz. Comput. 2021, 1–13. [Google Scholar] [CrossRef]
  10. Aykanat, M.; Kılıç, Ö.; Kurt, B.; Saryal, S. Classification of lung sounds using convolutional neural networks. EURASIP J. Image Video Process. 2017, 2017, 65. [Google Scholar] [CrossRef]
  11. Bardou, D.; Zhang, K.; Ahmad, S.M. Lung sounds classification using convolutional neural networks. Artif. Intell. Med. 2018, 88, 58–69. [Google Scholar] [CrossRef] [PubMed]
  12. Jung, S.-Y.; Liao, C.-H.; Wu, Y.-S.; Yuan, S.-M.; Sun, C.-T. Efficiently classifying lung sounds through depthwise separable CNN models with fused STFT and MFCC features. Diagnostics 2021, 11, 732. [Google Scholar] [CrossRef] [PubMed]
  13. Petmezas, G.; Cheimariotis, G.-A.; Stefanopoulos, L.; Rocha, B.; Paiva, R.P.; Katsaggelos, A.K.; Maglaveras, N. Automated Lung Sound Classification Using a Hybrid CNN-LSTM Network and Focal Loss Function. Sensors 2022, 22, 1232. [Google Scholar] [CrossRef] [PubMed]
  14. García-Ordás, M.T.; Benítez-Andrades, J.A.; García-Rodríguez, I.; Benavides, C.; Alaiz-Moretón, H. Detecting respiratory pathologies using convolutional neural networks and variational autoencoders for unbalancing data. Sensors 2020, 20, 1214. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Gairola, S.; Tom, F.; Kwatra, N.; Jain, M. Respirenet: A deep neural network for accurately detecting abnormal lung sounds in limited data setting. In Proceedings of the 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Guadalajara, Mexico, 1–5 November 2021; pp. 527–530. [Google Scholar]
  16. Pham, L.; Phan, H.; Palaniappan, R.; Mertins, A.; McLoughlin, I. Cnn-moe based framework for classification of respiratory anomalies and lung disease detection. IEEE J. Biomed. Health Inform. 2021, 25, 2938–2947. [Google Scholar] [CrossRef]
  17. Minami, K.; Lu, H.; Kim, H.; Mabu, S.; Hirano, Y.; Kido, S. Automatic classification of large-scale respiratory sound dataset based on convolutional neural network. In Proceedings of the 2019 19th International Conference on Control, Automation and Systems (ICCAS), Jeju, Korea, 15–18 October 2019; pp. 804–807. [Google Scholar]
  18. Hsu, F.-S.; Huang, S.-R.; Huang, C.-W.; Huang, C.-J.; Cheng, Y.-R.; Chen, C.-C.; Hsiao, J.; Chen, C.-W.; Chen, L.-C.; Lai, Y.-C.; et al. Benchmarking of eight recurrent neural network variants for breath phase and adventitious sound detection on a self-developed open-access lung sound database-HF_Lung_V1. PLoS ONE 2021, 16, e0254134. [Google Scholar] [CrossRef]
  19. Messner, E.; Fediuk, M.; Swatek, P.; Scheidl, S.; Smolle-Juttner, F.-M.; Olschewski, H.; Pernkopf, F. Crackle and breathing phase detection in lung sounds with deep bidirectional gated recurrent neural networks. In Proceedings of the 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; pp. 356–359. [Google Scholar]
  20. Perna, D.; Tagarelli, A. Deep auscultation: Predicting respiratory anomalies and diseases via recurrent neural networks. In Proceedings of the 2019 IEEE 32nd International Symposium on Computer-Based Medical Systems (CBMS), Cordoba, Spain, 5–7 June 2019; pp. 50–55. [Google Scholar]
  21. Kochetov, K.; Putin, E.; Balashov, M.; Filchenkov, A.; Shalyto, A. Noise masking recurrent neural network for respiratory sound classification. In Proceedings of the International Conference on Artificial Neural Networks, Rhodes, Greece, 4–7 October 2018; pp. 208–217. [Google Scholar]
  22. Chen, H.; Yuan, X.; Pei, Z.; Li, M.; Li, J. Triple-classification of respiratory sounds using optimized s-transform and deep residual networks. IEEE Access 2019, 7, 32845–32852. [Google Scholar] [CrossRef]
  23. Ma, Y.; Xu, X.; Yu, Q.; Zhang, Y.; Li, Y.; Zhao, J.; Wang, G. LungBRN: A smart digital stethoscope for detecting respiratory disease using bi-resnet deep learning algorithm. In Proceedings of the 2019 IEEE Biomedical Circuits and Systems Conference (BioCAS), Nara, Japan, 17–19 October 2019; pp. 1–4. [Google Scholar]
  24. Fernando, T.; Sridharan, S.; Denman, S.; Ghaemmaghami, H.; Fookes, C. Robust and Interpretable Temporal Convolution Network for Event Detection in Lung Sound Recordings. IEEE J. Biomed. Health Inform. 2022, 26, 2898–2908. [Google Scholar] [CrossRef]
  25. Jácome, C.; Ravn, J.; Holsbø, E.; Aviles-Solis, J.C.; Melbye, H.; Ailo Bongo, L. Convolutional neural network for breathing phase detection in lung sounds. Sensors 2019, 19, 1798. [Google Scholar] [CrossRef] [Green Version]
  26. Yang, Z.; Liu, S.; Song, M.; Parada-Cabaleiro, E.; Schuller, B.W. Adventitious respiratory classification using attentive residual neural networks. In Proceedings of the 21st Annual Conference of the International Speech Communication Association (Interspeech 2020), Shanghai, China, 25–29 October 2020; pp. 2912–2916. [Google Scholar]
  27. Zhao, X.; Shao, Y.; Mai, J.; Yin, A.; Xu, S. Respiratory Sound Classification Based on BiGRU-Attention Network with XGBoost. In Proceedings of the 2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Seoul, Korea, 16–19 December 2020; pp. 915–920. [Google Scholar]
  28. Rocha, B.M.; Filos, D.; Mendes, L.; Serbes, G.; Ulukaya, S.; Kahya, Y.P.; Jakovljevic, N.; Turukalo, T.L.; Vogiatzis, I.M.; Perantoni, E. An open access database for the evaluation of respiratory sound classification algorithms. Physiol. Meas. 2019, 40, 035001. [Google Scholar] [CrossRef]
  29. Fraiwan, M.; Fraiwan, L.; Khassawneh, B.; Ibnian, A. A dataset of lung sounds recorded from the chest wall using an electronic stethoscope. Data Brief 2021, 35, 106913. [Google Scholar] [CrossRef] [PubMed]
  30. Hestness, J.; Narang, S.; Ardalani, N.; Diamos, G.; Jun, H.; Kianinejad, H.; Patwary, M.; Ali, M.; Yang, Y.; Zhou, Y. Deep learning scaling is predictable, empirically. arXiv 2017, arXiv:1712.00409. [Google Scholar]
  31. Hsu, F.-S.; Huang, C.-J.; Kuo, C.-Y.; Huang, S.-R.; Cheng, Y.-R.; Wang, J.-H.; Wu, Y.-L.; Tzeng, T.-L.; Lai, F. Development of a Respiratory Sound Labeling Software for Training a Deep Learning-Based Respiratory Sound Analysis Model. arXiv 2021, arXiv:2101.01352. [Google Scholar]
  32. Rasamoelina, A.D.; Adjailia, F.; Sinčák, P. A review of activation function for artificial neural network. In Proceedings of the 2020 IEEE 18th World Symposium on Applied Machine Intelligence and Informatics (SAMI), Herlany, Slovakia, 23–25 January 2020; pp. 281–286. [Google Scholar]
  33. Das, P.P.; Allayear, S.M.; Amin, R.; Rahman, Z. Bangladeshi dialect recognition using Mel frequency cepstral coefficient, delta, delta-delta and Gaussian mixture model. In Proceedings of the 2016 Eighth International Conference on Advanced Computational Intelligence (ICACI), Chiang Mai, Thailand, 14–16 February 2016; pp. 359–364. [Google Scholar]
  34. Nakano, H.; Furukawa, T.; Tanigawa, T. Tracheal sound analysis using a deep neural network to detect sleep apnea. J. Clin. Sleep Med. 2019, 15, 1125–1133. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Lung sound auscultation locations and protocol. (a) Predefined auscultation locations (L1–L8). (b) Littmann 3200 recording protocol. (c) HF-Type-1 recording protocol. In (b,c), white right arrows and red blocks represent original continuous recordings and truncated 15-s recordings, respectively. This figure is adapted from our previous study under a Creative Commons Attribution (CC BY) license [18].
Figure 1. Lung sound auscultation locations and protocol. (a) Predefined auscultation locations (L1–L8). (b) Littmann 3200 recording protocol. (c) HF-Type-1 recording protocol. In (b,c), white right arrows and red blocks represent original continuous recordings and truncated 15-s recordings, respectively. This figure is adapted from our previous study under a Creative Commons Attribution (CC BY) license [18].
Applsci 12 07623 g001
Figure 2. Labeling process. (a) Raw signal of a lung sound recording. (b) Labeling software. (c) Output in a label file. In (b), the green square represents the spectrogram, the yellow lines indicated by the white arrows represent the boundaries of an ROI (the transparent yellow patch), and the red arrows indicate areas on different label tracks of which start and end times bounded by the ROI; when a labeler clicks on one of them, a corresponding label is generated.
Figure 2. Labeling process. (a) Raw signal of a lung sound recording. (b) Labeling software. (c) Output in a label file. In (b), the green square represents the spectrogram, the yellow lines indicated by the white arrows represent the boundaries of an ROI (the transparent yellow patch), and the red arrows indicate areas on different label tracks of which start and end times bounded by the ROI; when a labeler clicks on one of them, a corresponding label is generated.
Applsci 12 07623 g002
Figure 3. Architecture of the CNN–BiGRU model. This figure is adapted from our previous study under a Creative Commons Attribution (CC BY) license [18].
Figure 3. Architecture of the CNN–BiGRU model. This figure is adapted from our previous study under a Creative Commons Attribution (CC BY) license [18].
Applsci 12 07623 g003
Figure 4. Pipeline for preprocessing, deep-learning processing, and postprocessing. Detailed information is described in our previous publication [18]. This figure is adapted from our previous study under a Creative Commons Attribution (CC BY) license [18].
Figure 4. Pipeline for preprocessing, deep-learning processing, and postprocessing. Detailed information is described in our previous publication [18]. This figure is adapted from our previous study under a Creative Commons Attribution (CC BY) license [18].
Applsci 12 07623 g004
Figure 5. Illustration of segment- and event-detection evaluation: (a) ground-truth event labels, (b) ground-truth time segments, (c) segment prediction, (d) event prediction after postprocessing, (e) segment detection evaluation, (f) event evaluation based on ground-truth event labels, (g) event detection evaluation based on prediction, and (h) legend. JI: Jaccard index. This figure is adapted from our previous study under a Creative Commons Attribution (CC BY) license [18].
Figure 5. Illustration of segment- and event-detection evaluation: (a) ground-truth event labels, (b) ground-truth time segments, (c) segment prediction, (d) event prediction after postprocessing, (e) segment detection evaluation, (f) event evaluation based on ground-truth event labels, (g) event detection evaluation based on prediction, and (h) legend. JI: Jaccard index. This figure is adapted from our previous study under a Creative Commons Attribution (CC BY) license [18].
Applsci 12 07623 g005
Figure 6. ROC curves of the V1_Train-based and V2_Train-based models tested using V1_Test for (a) inhalation, (b) exhalation, (c) CAS, and (d) DAS segment detection and using V2_Test for (e) inhalation, (f) exhalation, (g) CAS, and (h) DAS segment detection. Solid lines represent the ROC curves, and the colored shades represent areas between the 95% confidence intervals. The blue lines and blue shades stand for the models trained using V1_Train, and the red lines and red shades stand for the models trained using V2_Train. The values in the parentheses in the legends indicate the mean AUC.
Figure 6. ROC curves of the V1_Train-based and V2_Train-based models tested using V1_Test for (a) inhalation, (b) exhalation, (c) CAS, and (d) DAS segment detection and using V2_Test for (e) inhalation, (f) exhalation, (g) CAS, and (h) DAS segment detection. Solid lines represent the ROC curves, and the colored shades represent areas between the 95% confidence intervals. The blue lines and blue shades stand for the models trained using V1_Train, and the red lines and red shades stand for the models trained using V2_Train. The values in the parentheses in the legends indicate the mean AUC.
Applsci 12 07623 g006
Figure 7. Spectrograms (top) and labels (bottom) of five lung sound recordings. In the label graphs at the bottom, red squares with the letter “I” represent I labels. By the same logic, blue squares with the letter “E” represent E labels, purple squares with the letter “D” represent D labels, and green squares with the letters “S,” “W,” and “R” represent S, W, and R labels, respectively. The blue arrows in (b) indicate two sounds with debatable E labels. The white arrow in (b) indicates that an inhalation was incorrectly labeled as E. The green arrows in (c) indicate CASs that were left without a correct S, W, or R label. The green arrows in (d) indicate the sound recordings that did not form clear streak patterns. The purple arrow in (e) indicates some noises that interfered with DAS labeling. The white dashed squares in (a,c,d) indicate sound events that were left without a label at the beginning and end of a lung sound recording.
Figure 7. Spectrograms (top) and labels (bottom) of five lung sound recordings. In the label graphs at the bottom, red squares with the letter “I” represent I labels. By the same logic, blue squares with the letter “E” represent E labels, purple squares with the letter “D” represent D labels, and green squares with the letters “S,” “W,” and “R” represent S, W, and R labels, respectively. The blue arrows in (b) indicate two sounds with debatable E labels. The white arrow in (b) indicates that an inhalation was incorrectly labeled as E. The green arrows in (c) indicate CASs that were left without a correct S, W, or R label. The green arrows in (d) indicate the sound recordings that did not form clear streak patterns. The purple arrow in (e) indicates some noises that interfered with DAS labeling. The white dashed squares in (a,c,d) indicate sound events that were left without a label at the beginning and end of a lung sound recording.
Applsci 12 07623 g007
Figure 8. Density plots of the start and end times of the I (blue line), E (orange line), C (green line), and D (red line) labels in HF_Lung_V2. (a) Start time in V2_Train. (b) End time in V2_Train. (c) Start time in V2_Test. (d) End time in V2_Test.
Figure 8. Density plots of the start and end times of the I (blue line), E (orange line), C (green line), and D (red line) labels in HF_Lung_V2. (a) Start time in V2_Train. (b) End time in V2_Train. (c) Start time in V2_Test. (d) End time in V2_Test.
Applsci 12 07623 g008
Table 1. Demographic data and the characteristics of respiratory diseases of the subjects in HF_Lung_V2.
Table 1. Demographic data and the characteristics of respiratory diseases of the subjects in HF_Lung_V2.
Subjects from RCC/RCWSubjects from FEMHSubjects in TSECCM Database
Number (n)2532243
Sex (M/F/NA)16/8/119/13NA
Age64.5 ± 18.568.2 ± 11.7NA
Height (cm)162.3 ± 9.1 *160.0 ± 8.8NA
Weight (kg)63.4 ± 12.7 *59.7 ± 12.3NA
BMI (kg/m2)24.0 ± 4.3 *23.6 ± 4.6NA
Recording deviceHF-Type-1 &
Littmann 3200
Littmann 3200Littmann 3200
Respiratory disease
Acute exacerbation of chronic obstructive pulmonary disease1 (4%)1 (3%)NA
Acute respiratory distress syndrome1 (4%)0 (0%)NA
Acute respiratory failure4 (16%)0 (0%)NA
Asthma0 (0%)1 (3%)NA
Bronchitis1 (4%)0 (0%)NA
Chronic respiratory failure12 (48%)2 (6%)NA
Chronic obstructive pulmonary disease2 (8%)5 (16%)NA
Emphysema1 (4%)0 (0%)NA
Pleural effusion0 (0%)1 (3%)NA
Pneumoconiosis0 (0%)1 (3%)NA
Pneumonia6 (24%)7 (22%)NA
Pulmonary embolism0 (0%)1 (3%)NA
* With four missing data. NA: not available.
Table 2. Statistical data of lung sound files in HF_Lung_V1 and HF_Lung_V2 databases.
Table 2. Statistical data of lung sound files in HF_Lung_V1 and HF_Lung_V2 databases.
DatabaseHF_Lung_V1HF_Lung_V2
Subjects261300
No. of 15-s recordings976513,957
Total duration (min)2441.253489.25
InhalationNo.34,09549,373
Duration (min)528.14785.48
Mean (s)0.930.95
ExhalationNo.18,34924,552
Duration (min)292.85374.24
Mean (s)0.960.91
CASNo. C/W/S/R13,883/8457/686/474021,558/13,139/914/7505
Duration (min) C/W/S/R191.16/119.73/9.46/61.98292.85/186.97/12.82/93.06
Mean (s) C/W/S/R 0.83/0.85/0.83/0.780.82/0.85/0.84/0.74
DASNo.15,60619,617
Duration (min)230.87281.55
Mean (s)0.890.86
CAS/C: continuous adventitious sound, DAS: discontinuous adventitious sound. W: wheeze, S: stridor, R: rhonchus.
Table 3. Statistics of lung sound files and labels in the HF_Lung_V1 and HF_Lung_V2 databases grouped by data sets.
Table 3. Statistics of lung sound files and labels in the HF_Lung_V1 and HF_Lung_V2 databases grouped by data sets.
DatabaseHF_Lung_V1HF_Lung_V2
Data SetTrainingTestTrainingTest
No. of 15-s
recordings
7809195610,5543403
Total duration (min)1952.254892638.5850.75
INo.27,223687239,05710,316
Duration (min)422.17105.97623.02162.46
Mean (s)0.930.930.960.94
ENo.15,601274818,3346218
Duration (min)248.0544.81292.8881.37
Mean (s)0.950.980.960.79
CNo. C/W/S/R11,464/7027/657/37802419/1430/29/96017,361/11,453/866/50424197/1686/48/2463
Duration (min) C/W/S/R160.16/100.71/9.10/50.3531.01/19.02/0.36/11.63240.40/163.77/12.34/64.2952.45/23.19/0.48/28.77
Mean (s) C/W/S/R0.84/0.86/0.83/0.800.77/0.80/0.74/0.730.83/0.86/0.85/0.770.75/0.83/0.60/0.70
DNo.13,794181214,2395378
Duration (min)203.5927.29210.9670.59
Mean (s)0.890.900.890.79
I: inhalation labels, E: exhalation labels, C: continuous adventitious sound labels, D: discontinuous adventitious sound labels, W: wheeze labels, S: stridor labels, and R: rhonchus labels.
Table 4. Performance measurement of CNN-BiGRU models trained using V1_Train and V2_Train and tested on V1_Test and V2_Test.
Table 4. Performance measurement of CNN-BiGRU models trained using V1_Train and V2_Train and tested on V1_Test and V2_Test.
Segment Detection Event Detection
AccuracyPPVSensitivitySpecificityF1 ScoreAUC PPVSensitivityF1 Score
Inhalation
Train_V1 on Test_V10.915 ± 0.0010.833 ± 0.0060.781 ± 0.0100.954 ± 0.0030.806 ± 0.0040.962 ± 0.001 0.818 ± 0.0050.855 ± 0.0090.836 ± 0.005
Train_V2 on Test_V10.921 ± 0.001 ***0.849 ± 0.005 ***0.794 ± 0.010 ***0.959 ± 0.002 ***0.821 ± 0.004 ***0.968 ± 0.001 *** 0.836 ± 0.005 ***0.864 ± 0.009 **0.850 ± 0.005 ***
Train_V1 on Test_V20.927 ± 0.0020.834 ± 0.0090.789 ± 0.0100.961 ± 0.0030.811 ± 0.0050.970 ± 0.001 0.819 ± 0.0070.864 ± 0.0090.842 ± 0.005
Train_V2 on Test_V20.931 ± 0.003 †††0.844 ± 0.011 0.799 ± 0.012 0.963 ± 0.0030.821 ± 0.008 †††0.973 ± 0.002 ††† 0.832 ± 0.010 †††0.869 ± 0.0110.851 ± 0.009 ††
Exhalation
Train_V1 on Test_V10.875 ± 0.0030.750 ± 0.0130.536 ± 0.0200.957 ± 0.0040.625 ± 0.0130.901 ± 0.003 0.604 ± 0.0170.629 ± 0.0220.616 ± 0.018
Train_V2 on Test_V10.882 ± 0.004 ***0.763 ± 0.011 **0.573 ± 0.026 ***0.957 ± 0.0030.654 ± 0.018 ***0.912 ± 0.005 *** 0.626 ± 0.026 *0.652 ± 0.025 *0.639 ± 0.025 **
Train_V1 on Test_V20.923 ± 0.0020.782 ± 0.0090.630 ± 0.0180.971 ± 0.0020.698 ± 0.0110.951 ± 0.002 0.751 ± 0.0110.752 ± 0.0200.752 ± 0.015
Train_V2 on Test_V20.924 ± 0.0030.783 ± 0.0110.644 ± 0.0190.970 ± 0.0020.706 ± 0.0120.954 ± 0.002 †† 0.754 ± 0.0170.749 ± 0.0190.751 ± 0.017
CAS
Train_V1 on Test_V10.852 ± 0.0040.683 ± 0.0140.448 ± 0.0310.950 ± 0.0050.540 ± 0.0220.870 ± 0.008 0.470 ± 0.0250.423 ± 0.0290.446 ± 0.026
Train_V2 on Test_V10.869 ± 0.003 ***0.702 ± 0.007 ***0.566 ± 0.030 ***0.942 ± 0.0040.626 ± 0.018 ***0.911 ± 0.005 *** 0.539 ± 0.025 ***0.489 ± 0.030 ***0.514 ± 0.027 ***
Train_V1 on Test_V20.873 ± 0.0020.659 ± 0.0100.383 ± 0.0320.963 ± 0.0040.484 ± 0.0240.890 ± 0.006 0.379 ± 0.0280.339 ± 0.0300.359 ± 0.028
Train_V2 on Test_V20.884 ± 0.003 †††0.675 ± 0.013 †††0.495 ± 0.028 †††0.956 ± 0.004 0.571 ± 0.018 †††0.918 ± 0.005 ††† 0.438 ± 0.023 †††0.403 ± 0.028 †††0.420 ± 0.025 †††
DAS
Train_V1 on Test_V10.828 ± 0.0050.733 ± 0.0160.694 ± 0.0220.888 ± 0.0120.713 ± 0.0090.894 ± 0.004 0.624 ± 0.0170.593 ± 0.0160.608 ± 0.014
Train_V2 on Test_V10.829 ± 0.0060.728 ± 0.0160.706 ± 0.0170.883 ± 0.0110.717 ± 0.0100.895 ± 0.006 0.635 ± 0.0170.605 ± 0.011 *0.620 ± 0.011 *
Train_V1 on Test_V20.877 ± 0.0040.758 ± 0.0120.811 ± 0.0190.901 ± 0.0080.784 ± 0.0080.941 ± 0.003 0.691 ± 0.0140.481 ± 0.0120.586 ± 0.011
Train_V2 on Test_V20.878 ± 0.0060.759 ± 0.0140.819 ± 0.0160.901 ± 0.0080.787 ± 0.0100.942 ± 0.004 0.698 ± 0.0240.485 ± 0.0170.591 ± 0.019
*, **, and *** indicate the models trained using V2_Train had significant improvement (p < 0.05, p < 0.01, and p < 0.001, respectively) on V1_Test compared to those trained using V1_Train. , ††, and ††† indicate the models trained using V2_Train had significant improvement (p < 0.05, p < 0.01, and p < 0.001, respectively) on V2_Test compared to those trained using V1_Train.
Table 5. Overlap ratios between I, E, C, and D labels in HF_Lung_V2.
Table 5. Overlap ratios between I, E, C, and D labels in HF_Lung_V2.
LabelOverlapped withOverlap Ratio
TrainTest
IE0.0010.001
C0.1790.167
D0.1980.287
EI0.0030.002
C0.2340.212
D0.2830.266
CI0.4630.517
E0.2860.329
D0.0060.009
DI0.5840.661
E0.3930.306
C0.0070.006
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hsu, F.-S.; Huang, S.-R.; Huang, C.-W.; Cheng, Y.-R.; Chen, C.-C.; Hsiao, J.; Chen, C.-W.; Lai, F. A Progressively Expanded Database for Automated Lung Sound Analysis: An Update. Appl. Sci. 2022, 12, 7623. https://doi.org/10.3390/app12157623

AMA Style

Hsu F-S, Huang S-R, Huang C-W, Cheng Y-R, Chen C-C, Hsiao J, Chen C-W, Lai F. A Progressively Expanded Database for Automated Lung Sound Analysis: An Update. Applied Sciences. 2022; 12(15):7623. https://doi.org/10.3390/app12157623

Chicago/Turabian Style

Hsu, Fu-Shun, Shang-Ran Huang, Chien-Wen Huang, Yuan-Ren Cheng, Chun-Chieh Chen, Jack Hsiao, Chung-Wei Chen, and Feipei Lai. 2022. "A Progressively Expanded Database for Automated Lung Sound Analysis: An Update" Applied Sciences 12, no. 15: 7623. https://doi.org/10.3390/app12157623

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop