Next Article in Journal
Functionalized Titanium Dioxide Nanoparticle-Based Electrochemical Immunosensor for Detection of SARS-CoV-2 Antibody
Previous Article in Journal
Commercially Available Molecular Approaches to Evaluate Endometrial Receptivity: A Systematic Review and Critical Analysis of the Literature
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Radiomics-Based Detection of COVID-19 from Chest X-ray Using Interpretable Soft Label-Driven TSK Fuzzy Classifier

1
Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Hong Kong, China
2
Department of Medical informatics, Nantong University, Nantong 226007, China
*
Author to whom correspondence should be addressed.
Diagnostics 2022, 12(11), 2613; https://doi.org/10.3390/diagnostics12112613
Submission received: 1 September 2022 / Revised: 16 October 2022 / Accepted: 18 October 2022 / Published: 27 October 2022
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
The COVID-19 pandemic has posed a significant global public health threat with an escalating number of new cases and death toll daily. The early detection of COVID-related CXR abnormality potentially allows the early isolation of suspected cases. Chest X-Ray (CXR) is a fast and highly accessible imaging modality. Recently, a number of CXR-based AI models have been developed for the automated detection of COVID-19. However, most existing models are difficult to interpret due to the use of incomprehensible deep features in their models. Confronted with this, we developed an interpretable TSK fuzzy system in this study for COVID-19 detection using radiomics features extracted from CXR images. There are two main contributions. (1) When TSK fuzzy systems are applied to classification tasks, the commonly used binary label matrix of training samples is transformed into a soft one in order to learn a more discriminant transformation matrix and hence improve classification accuracy. (2) Based on the assumption that the samples in the same class should be kept as close as possible when they are transformed into the label space, the compactness class graph is introduced to avoid overfitting caused by label matrix relaxation. Our proposed model for a multi-categorical classification task (COVID-19 vs. No-Findings vs. Pneumonia) was evaluated using 600 CXR images from publicly available datasets and compared against five state-of-the-art AI models in aspects of classification accuracy. Experimental findings showed that our model achieved classification accuracy of over 83%, which is better than the state-of-the-art models, while maintaining high interpretability.

1. Introduction

Coronavirus disease 2019 (COVID-19) is caused by severe acute respiratory syndrome coronavirus (SARS-CoV-2), which has rapidly spread worldwide and posed significant public health threats. As of 25 January 2021, there have been more than 99.257 million confirmed cases of COVID-19 and 2,131,000 cumulative deaths worldwide, according to the Global COVID-19 Status Report released by Johns Hopkins University (https://coronavirus.jhu.edu/map.html (accessed on 1 August 2021)). The number of confirmed/fatal cases is still on the rise. Currently, the main screening methods for COVID-19 cases include reverse transcription polymerase chain reaction (RT-PCR) or gene sequencing for respiratory or blood specimens [1]. However, the overall RT-PCR positive rate of throat swab samples is reported to be 30 to 60%; thus, subjects with false-negative test results may continue to transmit the virus within a community [2]. CXR is a routine diagnostic tool incorporated in the initial diagnosis and subsequent monitoring for COVID-19 patients. It boasts short examination time, low radiation dose and low cost of examination. CXR images show visual indexes correlated with COVID-19 [3]. Reported CXR features include airspace opacities, which can show multilobar involvement, peripheral dominance, asymmetric or diffuse involvement [4]. Early changes on CXR are often subtle, requiring radiologist expertise. Nevertheless, radiology expertise is scarce in developing countries, which has been further strained by the tremendous case load of COVID-19 infection. Applying artificial intelligence (AI)/machine learning approaches for the automated identification of subtle CXR abnormalities serves as a potential remedy [5,6].
At present, several AI models have been developed for COVID-19 detection using chest radiography images [7,8,9]. Minaee et al. used transfer learning to train four famous convolutional neural networks (ResNet18, ResNet50, SqueezeNet and DenseNet-121) for identifying COVID-19 disease in the analyzed CXR images [10]. The results showed that although most of these networks achieved a sensitivity rate of 98% (±3%), only around a 90% specificity rate was achieved. Ozturk et al. proposed a deep learning model for detecting and classifying COVID-19 cases using CXR images [11]. The proposed model was fully automated with an end-to-end structure without the need for manual feature extraction and produced a classification accuracy of 87.02% for multi-class prediction (COVID-19 vs. No-Findings vs. Pneumonia). Table 1 summarizes part of the recent studies on AI model development for COVID-19 detection using chest radiography images. It can be observed that deep learning models are in great favor due to the promising performance; research in this area is anticipated to grow. In spite of this, almost all deep learning models have been regarded as a black box given the incomprehensible deep features and lack of transparency in the training procedures, raising a concern regarding the model interpretability. Of note, when it comes to adopting the AI models as clinical decision support systems, model interpretability is an indispensable prerequisite [12].
In this regard, Takagi–Sugeno–Kang (TSK) fuzzy systems are also in great favor owing to their capability to maintain outstanding balance between approximation ability and interpretability [24,25,26]. Although fuzzy systems were originally designed for regression or binary tasks (it can be considered as special regression tasks), by virtue of strategic approaches, fuzzy systems have achieved satisfactory performance in clinical diagnosis, mostly in disease classification. For example, Jiang et al. proposed a multi-view TSK fuzzy system for epileptic EEG signal detection [27]. Hsieh et al. combined an adaptive neuro-fuzzy inference system (ANFIS) with greedy forward feature selection to develop an intelligent diagnostic system for differentiating cases with influenza from those with cold [28]. Khayamnia et al. established a Mamdani fuzzy system for migraine headache diagnosis [29]. In each of the above studies, the label vector was transformed into a strict binary matrix in order to apply the fuzzy system for multi-class classification tasks. However, numerous studies indicated that transforming the training samples to a strict binary label matrix is too rigid to learn a discriminative transformation matrix [30,31,32].
Confronted with all these, in this study, we developed a TSK fussy system to enhance the interpretability of our AI model for COVID-19 detection (COVID-19 vs. No-Findings vs. Pneumonia). Meanwhile, we introduced a soft strategy to adaptively relax the binary matrix during the label transformation. Inspired by manifold learning, a class compactness graph was constructed to minimize the risk of model overfitting following the introduction of label relaxation. The core idea lies in that samples sharing the same labels should be kept as close as possible when they are transformed into the label space. The contributions of this study can be summarized from the model and experiment perspectives.
(1)
When classic TSK fuzzy systems are applied to classification tasks, margins between different classes are expected to be as large as possible after they are transformed into their label space. However, this assumption is too rigid to learn a discriminative transformation matrix. Additionally, excessive label fitting may cause overfitting. To address these issues, label softening and the compactness class graph are embedded into the objective function of the classic TSK fuzzy systems, which can bring two advantages: firstly, the new method can enlarge the margins between different classes as much as possible, and it has more freedom to fit the labels better. Secondly, to avoid the problem of overfitting, the new method uses the class compactness graph to guarantee that the samples from the same class can be kept close together in the transformed space.
(2)
Five state-of-the-art algorithms are introduced for comparison studies, and experimental results are reported from the perspectives of classification performance, sensitivity and interpretability. The proposed algorithm significantly performs better than the state-of-the-art algorithm in terms of accuracy and macro F1-score due to the introduced label soften and the compactness class graph. What is more, the improved generalization ability can be observed from the lower accuracy difference between training and testing. In addition, radiomics features have physical meaning; with the help of transparent fuzzy rules generated by the proposed algorithm, the interpretability can be guaranteed.
The rest of the sections are organized as follows: In Section 2, we briefly describe the working mechanism of the zero-order TSK fuzzy system for classification tasks. In Section 3, we introduce the objective function, optimization, and the corresponding algorithm of our proposed TSK fuzzy system. In Section 4, we conduct experiments on radiomics features extracted from CXR images for model evaluation. In Section 5, we discuss the experimental results and conclude our study. The abbreviations and symbols used in this study are given in the Abbreviation section.

2. ZERO-TSK-FS for Classification

The original TSK fuzzy systems were designed for regression tasks. However, the diagnosis of COVID-19 based on radiomics belongs to a classification task, as we know that the zero-order TSK fuzzy system (ZERO-TSK-FS) has very concise fuzzy rules and is widely used due to low model complexity. Therefore, in this section, we re-design ZERO-TSK-FS specifically for the classification task. Suppose we have a training set χ = { ( x i , y i ) } i = 1 N ; each input x i is represented in the d-dimensional feature space [ x i 1 , x i 2 , , x i d ] T R d , and y i R is the corresponding label. Then, the k-th fuzzy rule of ZERO-TSK-FS can be defined as follows,
If   x i 1   is   ϑ 1 k x i 2   is   ϑ 2 k x i d   is   ϑ d k , then   f k ( x i ) = β k , k = 1 , 2 , , K ,
where ϑ j k represents a fuzzy set subscribed by feature x i j for the k-th rule, is a fuzzy conjunction operator, and K is the number of fuzzy rules. Each fuzzy rule is premised on the input feature space [ x i 1 , x i 2 , , x i d ] T R d and transforms the fuzzy sets in the input feature space into a varying singleton represented by f k ( x i ) . After a series of operation and defuzzification steps, we can obtain the output of x i , i.e., f ( x i ) ,
f ( x i ) = k = 1 K f k ( x i ) φ k ( x i , μ k , δ k ) = k = 1 K β k φ k ( x i , μ k , δ k ) ,
where the normalized fuzzy membership function φ k ( x i , μ k , δ k ) of the k-th fuzzy rule can be formulated as
φ k ( x i , μ k , δ k ) = φ ˜ k ( x i , μ k , δ k ) k = 1 K φ ˜ k ( x i , μ k , δ k ) .
In (3), φ ˜ k ( x i , μ k , δ k ) can be computed by j = 1 d φ ˜ ϑ d k k ( x i j ) . If the Gaussian fuzzy membership function is adopted, we have
φ ˜ k ( x i , μ k , δ k ) = j = 1 d exp ( ( x i j μ j k ) 2 δ j k ) ,
where μ k = [ μ 1 k , μ 2 k , , μ d k ] and δ k = [ δ 1 k , δ 2 k , , δ d k ] denote the kernel center vector and kernel width vector, respectively. By denoting ϕ ( x i ) = [ φ 1 ( x i , μ 1 , δ 1 ) , φ 2 ( x i , μ 2 , δ 2 ) , , φ K ( x i , μ K , δ K ) ] , and β = [ β 1 , β 2 , , β K ] T , then f ( x i ) in (3) can be updated as
f ( x i ) = ϕ ( x i ) β .
It can be seen that the optimization of the consequent β can be actually considered as solving a linear regression problem in (5). There are many algorithms that can be used to solve this problem. A representative one proposed by Deng et al. has been demonstrated its promising performance [33], which can be formulated as
min β Θ β y F 2 + λ β F 2 .
where Θ = [ ϕ ( x 1 ) , ϕ ( x 2 ) , , ϕ ( x N ) ] T , and y = [ y 1 , y 2 , , y N ] T R N × 1 is the output vector. Notably, the model obtained from (6) is a regression model that cannot be directly used for our classification task. Usually, when we use a regression model for a multi-class classification task, a label transformation should be conducted [27]. That is to say, we should transform the label space from R N × 1 to R N × C , where C is the number of classes. For example, suppose we have a three-class classification training set χ = { ( x i , y i ) } i = 1 N having four samples, where y = [ 2 , 1 , 3 , 1 ] T R 4 × 1 ; then, after label space transformation, we have the new label space Y R 4 × 3 ,
Y = 0 1 0 1 0 0 0 0 1 1 0 0 ,
which means that the four samples belong to the second, first, third and first class, respectively. Therefore, by substituting the new label space Y R N × C into (6) to replace the original label space y = R N × 1 , we have
min Δ Θ Δ Y F 2 + λ Δ F 2 .
Here, we use Δ R K × C as the consequent matrix to replace β R K × 1 in (6). In this manner of label transformation, ZERO-TSK-FS can be used for the multi-class classification task by transforming the training samples to a strict binary label matrix, such as Y R N × C shown in (7).
However, as we previously stated, transforming the training samples to a strict binary label matrix may not be adequately flexible to learn a discriminative transformation matrix. Therefore, in the following sections, we develop a novel zero TSK fuzzy system for COVID-19 cases detection which can relax the strict binary label matrix to a soft one.

3. LR-ZERO-TSK-FS

3.1. Objective Function

To fit the labels in a soft way, we introduced a non-negative label soften matrix Ω R N × C and a luxury matrix Ξ R N × C to relax the strict binary label matrix. Let us also take the three-class classification training set having four samples as an example; with Ω and Ξ , the strict binary label matrix can be relaxed into the following form:
Y ˜ = Y + Ω Ξ = Ω 11 1 + Ω 12 Ω 13 1 + Ω 21 Ω 22 Ω 23 Ω 31 Ω 32 1 + Ω 33 1 + Ω 41 Ω 42 Ω 43 ,
where the non-negative label matrix Ω R N × C is defined as
Ω = Ω 11 Ω 1 C Ω 21 Ω 2 C Ω N 1 Ω N C s . t .   Ω i j 0 ,
the luxury matrix Ξ R N × C is defined as
Ξ ( i , j ) = + 1   if   Y ( i , j ) = 1 1   if   Y ( i , j ) = 0 ,
and is a Hadamard product operator of matrices.
Therefore, with label matrix relaxation, the objective function of ZERO-TSK-FS in (8) can be updated as
min Δ Θ Δ ( Y + Ω Ξ ) F 2 + λ Δ F 2 .
Although label relaxation can help learn a more discriminant transformation matrix Δ , it may lead to overfitting problems. In manifold learning, the class compactness graph is often used to alleviate overfitting problems. The core idea is that samples sharing the same labels should be kept as close as possible when they are transformed into the label space. More specifically, a weight matrix W is defined as follows to capture the relationship that the samples sharing the same labels should be kept as close as possible,
W ( i , j ) = e x i x j 2 / δ   i f   x i   a n d   x j   have   the   same   labels   0   otherwise .
In the class compactness graph, if sample x i and sample x j have the same labels, they are connected by an undirected edge, and the corresponding weight can be computed by (13). We find that the closer sample x i and sample x j are, the bigger the corresponding weight. Therefore, when all training samples in the training space are transformed into the label space, any two samples having bigger weights in the training space should also be kept as close as possible in the label space. That is to say, it is reasonable to minimize the following objective to achieve our goal,
min f i = 1 N j = 1 N f ( x i ) f ( x j ) 2 W i j ,
where W i j is the i-th row and j-th column element in the weight matrix W , and f ( x i ) = ϕ ( x i ) Δ is the transformation result of x i represented in the label space. By substituting f ( x i ) = ϕ ( x i ) Δ to (14) and performing equivalent mathematical transformation, we have
min Δ tr ( Δ T Θ T L Θ Δ ) ,
where L is the graph Laplacian, which is defined as L = Z W . Z is a diagonal matrix, and the diagonal element can be computed by Z i j = j = 1 N W i j .
Finally, by embedding (15) to (12), we have the following objective function for LR-ZERO-TSK-FS,
min Δ , Ω Θ Δ ( Y + Ω Ξ ) F 2 + λ Δ F 2 + γ tr ( Δ T Θ T L Θ Δ ) .

3.2. Optimization

The objective function shown in (16) has two components that need to be optimized, i.e., the transformation matrix Δ and the non-negative label relaxation matrix Ω . Thus, we introduced an iterative strategy to search for the optimal solution for Δ and Ω [34].
Firstly, suppose that the non-negative label relaxation matrix Ω is fixed; then, the optimization problem becomes
J ( A ) = min Δ Θ Δ U F 2 + λ Δ F 2 + γ tr ( Δ T Θ T L Θ Δ ) ,
where U = Y + Ω Ξ . By setting the partial derivative with respect to Δ to 0, we have
J ( A ) / Δ = 2 Θ T Θ Δ 2 Θ T U + 2 λ Δ + 2 γ Θ T L Θ Δ = 0 Δ = ( Θ T Θ + λ I + γ Θ T L Θ ) 1 Θ T U = ( Θ T Θ + λ I + γ Θ T L Θ ) 1 Θ T ( Y + Ω Ξ )
Secondly, suppose that the transformation matrix Δ is fixed; since the second term in (16) is uncorrelated with Ω , the optimization problem becomes
J ( Ω ) = min Ω V Ω Ξ F 2 s . t .   Ω 0
where V = Θ Δ Y . We have common knowledge that the squared Frobenius norm of the matrix can be decoupled element by element. Therefore, the optimization problem in (19) can be equivalently decoupled into N × C sub-problems. For the i-th row and j-th column element Ω i j in Ω , we have the corresponding sub-problem,
J ( Ω i j ) = min Ω i j V i j Ω i j Ξ i j F 2 s . t .   Ω i j 0
Given that ( Ξ i j ) 2 = 1 , we have ( V i j Ω i j Ξ i j ) 2 = ( Ξ i j V i j Ω i j ) 2 . Considering the non-negative constraint imposed on Ω i j , we can obtain Ω i j = max ( Ξ i j V i j , 0 ) . Accordingly, Ω in (19) can be computed as
Ω = max ( Ξ V , 0 ) .
With the closed-form solution of Δ in (18) and Ω in (21), we can use an iterative strategy to find their optimal values.

3.3. Algorithm

The training steps of LR-ZERO-TSK-FS are shown in Algorithm 1. We can see that the asymptotic time complexity of LR-ZERO-TSK-FS is mainly contributed by two components, i.e., FCM clustering for antecedent learning and the computing of Δ for consequent learning. The time complexity of FCM clustering is O ( d 2 N K ) , where d is the dimension of the training set. The time complexity of the computing of Δ is O ( K 3 ) . Therefore, the asymptotic time complexity of LR-ZERO-TSK-FS is O ( d 2 N K + K 3 ) . Since d and K are relatively small compared with N, the asymptotic time complexity of LR-ZERO-TSK-FS can be considered linear with N.
Algorithm 1 LR-ZERO-TSK-FS
Input: Training set χ = { ( x i , y i ) } i = 1 N , number of fuzzy rules K, regularized parameters λ and γ
Output: Transformation matrix Δ and non-negative label soften matrix Ω
Procedures:
1 Use clustering technique, e.g., FCM (Fuzzy C-Mean) to learn the antecedent parameters, i.e., μ j k and δ j k in (4) of fuzzy rules.
2 Use (3) to get ϕ ( x i ) , and further Θ .
3 Use (13) and Z i j = j = 1 N W i j to compute the graph Laplacian matrix L .
4 Randomize Ω under the constraint Ω 0 .
5 Set t 0 .
Repeat
    6 Update Δ ( t + 1 ) by (19) with current Ω ( t ) .
    7 Update Ω ( t + 1 ) by (21) with current Δ ( t + 1 ) .
    8 t t + 1 .
Until  Δ ( t + 1 ) - Δ ( t ) 2 ε
9 With Δ and Ω , the output can be computed by Y = Θ Δ ( Ω Ξ ) .

4. Experimental Studies

4.1. Data Preprocessing

We collected 600 chest X-Ray (CXR) images from 200 normal cases, 200 COVID-19 cases, and 200 non-COVID-19 pneumonia cases, which are publicly available from online databases (https://github.com/agchung/Figure1-COVID-chestxray-dataset (accessed on 8 August 2021); https://github.com/agchung/Actualmed-COVID-chestxray-dataset (accessed on 8 August 2021); https://www.kaggle.com/tawsifurrahman/covid19-radiography-database (accessed on 8 August 2021); https://www.kaggle.com/c/rsna-pneumonia-detection-challenge/data (accessed on 8 August 2021)). The workflow of data preprocessing is shown in Figure 1.
In data preprocessing, we aimed to extract radiomics features from CXR images. Firstly, a binary lung region mask was generated by a pretrained U-NET segmentation network. With the obtained lung region on each CXR image, we then used the radiomics package to extract different types of radiomics features for downstream modeling.

4.2. Settings

After extracting radiomics features from CXR images, we followed the workflow, as shown in Figure 2, to evaluate the proposed model.
To simulate an independent test, we introduce holdout cross-validation, as shown in Figure 2. To be specific, we firstly independently select a holdout set as a testing set. Then, the rest is cut into K (K = 3 in this study) folds; one is taken as the validation set and the rest are taken as the training set. In the validation phase, minimum-redundancy-maximum-relevance (mRMR) is adopted as a feature-ranking algorithm. Then, the cross-validation (3-CV) strategy was used to determine the optimal feature set and hyper-parameters with respect to the proposed model. In the training phase, with the optimal feature set and hyper-parameters, the best model can be obtained. In the testing phase, with the best model, we can obtain the corresponding testing results.
To highlight the performance of the proposed model, we introduced several benchmarking models for comparison. They are ZERO-TSK-FS [35], L2-TSK-FS [24], FS-FCSVM [36], linear SVM (L-SVM) and Gaussian SVM (G-SVM) [37], respectively. Table 2 gives the experimental settings.
The criteria accuracy and macro F1-score were introduced to quantitatively evaluate the classification performance of all studied models.

4.3. Experimental Results

In this section, we report our evaluation results from three main aspects, i.e., classification performance analysis, sensitivity analysis and interpretability analysis.

4.3.1. Classification Performance Analysis

The testing and training accuracy of the proposed model LR-ZERO-TSK-FS and the studied benchmarking models are shown in Figure 3. The horizontal axis represents the top-k features selected by using the feature selection method, as shown in Figure 2. It can be observed that LR-ZERO-TSK-FS wins the best performance under the most top-k features except top-5 and top-35. Specially, LR-ZERO-TSK-FS always performs better than the baseline ZERO-TSK-FS on the testing set. As indicated in Figure 3b, LR-ZERO-TSK-FS also wins the best from the top-20 to top-50 features. Statistical analysis of the testing performance in terms of t-test is shown in Table 3, where “*” means that there exist significant differences between LR-ZERO-TSK-FS and the state-of-the-art models (p < 0.05). A similar observation can also be obtained by using another metric macro F1-score, as shown in Table 4 and Table 5. In Table 6, we also report the CPU seconds each algorithm consumed.
To further investigate the impact of training sample size on the difference in model accuracy between training and testing sets (i.e., model generalizability), we trained the models under varying training sample sizes. Figure 4 shows the accuracy differences (the absolute error of training accuracy and test accuracy) between the training and testing datasets, where the horizontal axis represents the size of the training samples. We have common knowledge that fewer training samples are more likely to result in model overfitting, which is also demonstrated in Figure 4. It can also be noticed that the proposed model LR-ZERO-TSK-FS is more effective in inhibiting model overfitting compared with the benchmarking models. As we stated in the introduction part, label relaxation can help to learn a more discriminant transformation matrix, but it also tends to result in overfitting. This statement can be demonstrated by the results shown in Figure 5.
When the hyperparameter γ is set to 0, the manifold regularization term in (15) becomes useless. From Figure 5a, it can be observed that LR-ZERO-TSK-FS with γ = 0 has larger accuracy differences than the benchmarking models. When compared with ZERO-TSK-FS, in particular, the result demonstrated that the label relaxation indeed aggravates overfitting in the absence of manifold regularization. When the manifold regularization term was activated, see Figure 5b, the advantageous testing accuracy can basically be maintained. That is to say, the combination of label relaxation and manifold regularization achieve an outstanding balance between classification accuracy and generalization.

4.3.2. Sensitivity Analysis

The proposed model LR-ZERO-TSK-FS has two regularization hyper-parameters λ and γ that needed to be set in advance. In this study, we analyzed the robustness of LR-ZERO-TSK-FS with respect to λ and γ . As illustrated in Figure 6, it can be observed that LR-ZERO-TSK-FS is sensitive to λ and γ , and smaller λ and γ seem to yield better testing accuracy. Therefore, according to Figure 6, λ and γ can be determined by using cross-validation under a small range.
Results from comparative analyses indicated that LR-ZERO-TSK-FS yielded the best overall model performance on the COVID-19 cases detection task (Figure 3). From Figure 5b, we can notice that when the manifold regularization is not activated ( γ = 0 ), LR-ZERO-TSK-FS also performs satisfactorily and better than LR-ZERO-TSK-FS with manifold regularization. That is to say, from Figure 3 and Figure 5b, the capability of our label relaxation strategy for enhancing the classification performance of the TSK fuzzy systems was successfully demonstrated. Moreover, from Figure 4 and Figure 5a, we can see that when the manifold regularization is activated, the differences in accuracy between the training and testing sets remarkably drop. In particular, the comparative analysis with the ZERO-TSK-FS explicitly indicated the superiority of LR-ZERO-TSK-FS in minimizing the risk of model overfitting. It is noteworthy that the present evidence relies on the following particularities of LR-ZERO-TSK-FS:
(1)
Streamlined classification accuracy owing to the introduction of a soft strategy to adaptively relax the binary matrix during label transformation, rendering more flexibility during label transformation and capability in enlarging the margins between different classes.
(2)
Alleviated risk of model overfitting in virtue of the adoption of a class compactness graph during manifold learning, based on the assumption that samples sharing the same labels should be kept as close as possible when they are transformed into the label space.

4.3.3. Interpretability Analysis

The proposed model LR-ZERO-TSK-FS was derived from the zero order TSK fuzzy system, carrying the inherent characteristic of model interpretability. Therefore, in the following, we give an example to illustrate how the proposed model can diagnose a subject with COVID-19 with the use of radiomics features.
The optimal radiomics features selected by step 3 shown in Figure 2 are listed in Table 7, which were used for model training. The training results (five trained fuzzy rules) of LR-ZERO-TSK-FS using the selected radiomics features in terms of antecedent and consequent parameters are listed in Table 8.
Figure 7 shows the fuzzy membership function of one radiomics feature (No. 1: wavelet-HH_glrlm_GrayLevelVariance) trained on five fuzzy rules. According to the center μ k of each fuzzy membership function and domain knowledge, we can assign a linguistic description, e.g., “low”, “a little low”, “medium”, “a little high” and “high”, to each feature. For example, from the fuzzy mapping of “wavelet-HH_glrlm_GrayLevelVariance” in the first fuzzy rule, it can be interpreted that “wavelet-HH_glrlm_GrayLevelVariance” is “low”. According to Yang et al. [38], linguistic meanings are dependent on expert knowledge. This is the beauty of interpretable models because the acquired knowledge by models can be refined and integrated with expert knowledge. Accordingly, the trained five fuzzy rules are described in Table 9.
Under these five trained fuzzy rules, COVID-19 detection of an individual can be made following the procedures illustrated in Figure 8.
From Figure 8, it can be found that the classification of the case depends on the obtained five fuzzy rules. As shown in Table 6, according to the parameters learned in the if-part, the five fuzzy rules can be assigned different linguistic meaning based on domain knowledge. Therefore, each fuzzy rule actually indicates the classification result of one expert (or one kind of knowledge). For example, the first fuzzy rule classifies the case to COVID-19 based on the domain knowledge embedded in the if-part. Finally, by linearly combining all fuzzy rules (combing knowledge of all experts), the case is classified to COVID-19.

5. Conclusions

COVID-19 has posed a significant public health threat globally. The automated detection of CXR abnormality potentially aids identifying patients with significant risk of COVID infection earlier. In this study, we developed an interpretable and soft-label-driven TSK fuzzy system for multi-class COVID-19 detection (COVID-19 vs. No-Findings vs. Pneumonia) using radiomics features extracted from CXR images. The risk of model overfitting is alleviated in virtue of the adoption of a class compactness graph during manifold learning, which is based on the assumption that samples sharing the same labels should be kept as close as possible when they are transformed into the label space. We successfully demonstrated that our model outperformed the comparable state-of-the-art models while maintaining high interpretability.
This study is not without limitations. For example, in the construction of the class compactness graph, we just used Euclidean distance. In addition, in our experiments, multi-center based external validation is not carried out. Therefore, in our future work, we will carry out more in-depth research from the above aspects.

Author Contributions

Data curation, D.Y., X.T., T.Z. and Z.M.; Methodology, Y.Z. and J.Z.; Supervision, T.-C.Y. and J.C.; Validation, B.L.; Writing—original draft, Y.Z.; Writing—review and editing, S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Hong Kong Scholars Program under Grant XJ2019056, Health and Medical Research Fund (HMRF COVID190211), the Food and Health Bureau, The Government of the Hong Kong Special Administrative Regions, Mainland-Hong Kong Joint Funding Scheme (MHKJFS) (MHP/005/20), Shenzhen-Hong Kong-Macau S&T Program (Category C) (SGDX20201103095002019), Shenzhen Basic Research Program (JCYJ20210324130209023) of Shenzhen Science and Technology Innovation Committee, and the Natural Science Foundation of Jiangsu Province (No. BK20201441), and Jiangsu Post-doctoral Research Funding Program (No. 2020Z020).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are available on: https://github.com/agchung/Figure1-COVID-chestxray-dataset (accessed on 8 August 2021); https://github.com/agchung/Actualmed-COVID-chestxray-dataset (accessed on 8 August 2021); https://www.kaggle.com/tawsifurrahman/covid19-radiography-database (accessed on 8 August 2021); https://www.kaggle.com/c/rsna-pneumonia-detection-challenge/data (accessed on 8 August 2021); Code availability: The code can be accessed by sending email to the corresponding author.

Acknowledgments

We would like to thank all the reviewers.

Conflicts of Interest

The authors have declared that no competing interests exist.

Consent to Participate

The authors have declared that they consent to participate.

Consent for Publication

The authors have declared that they consent for publication of this study.

Abbreviations

Abbreviations
COVID-19Corona Virus Disease 2019
CXRChest X-Ray
TSKTakagi–Sugeno–Kang
RT-PCRReverse Transcription Polymerase Chain Reaction
AIArtificial Intelligence
ZERO-TSK-FSZero Order TSK Fuzzy System
ANFISAdaptive Neuro-Fuzzy Inference System
LR-ZERO-TSK-FSLabel Relaxation Zero Order TSK Fuzzy System
mRMRMinimum-Redundancy-Maximum-Relevance
Symbols
χ Training set
{ x i , y i } One training sample (vector) and the corresponding label.
NNumber of training samples.
CNumber of classes.
Ω R N × C Non-negative label soften matrix
Ξ R N × C Luxury matrix
Δ Transformation matrix
W Weight matrix
L Graph Laplacian matrix

References

  1. Wang, W.; Xu, Y.; Gao, R.; Lu, R.; Han, K.; Wu, G.; Tan, W. Detection of SARS-CoV-2 in Different Types of Clinical Specimens. JAMA 2020, 323, 1843–1844. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Yang, Y.; Yang, M.; Yuan, J.; Wang, F.; Wang, Z.; Li, J.; Zhang, M.; Xing, L.; Wei, J.; Peng, L.; et al. Laboratory diagnosis and monitoring the viral shedding of 2019-nCoV infections. Innovation 2020, 1, 100061. [Google Scholar] [CrossRef]
  3. Kanne, J.P.; Little, B.P.; Chung, J.H.; Elicker, B.M.; Ketai, L.H. Essentials for Radiologists on COVID-19: An Update—Radiology Scientific Expert Panel. Radiology 2020, 296, E113–E114. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Rodrigues, J.; Hare, S.; Edey, A.; Devaraj, A.; Jacob, J.; Johnstone, A.; McStay, R.; Nair, A.; Robinson, G.; Rodrigues, J.; et al. An update on COVID-19 for the radiologist-A British society of Thoracic Imaging statement. Clin. Radiol. 2020, 75, 323–325. [Google Scholar] [CrossRef] [Green Version]
  5. Haghanifar, A.; Majdabadi, M.M.; Choi, Y.; Deivalakshmi, S.; Ko, S. COVID-CXNet: Detecting COVID-19 in frontal chest X-ray images using deep learning. Multimed. Tools Appl. 2022, 81, 30615–30645. [Google Scholar] [CrossRef] [PubMed]
  6. Umer, M.; Ashraf, I.; Ullah, S.; Mehmood, A.; Choi, G.S. COVINet: A convolutional neural network approach for predicting COVID-19 from chest X-ray images. J. Ambient Intell. Humaniz. Comput. 2022, 13, 535–547. [Google Scholar] [CrossRef]
  7. Zhang, Y.-D.; Satapathy, S.C.; Liu, S.; Li, G.-R. A five-layer deep convolutional neural network with stochastic pooling for chest CT-based COVID-19 diagnosis. Mach. Vis. Appl. 2021, 32, 14. [Google Scholar] [CrossRef]
  8. Khan, M.A.; Hussain, N.; Majid, A.; Alhaisoni, M.; Bukhari, S.A.C.; Kadry, S.; Nam, Y.; Zhang, Y.-D. Classification of Positive COVID-19 CT Scans using Deep Learning. Comput. Mater. Contin. 2021, 66, 2923–2938. [Google Scholar] [CrossRef]
  9. Wang, S.-H.; Nayak, D.R.; Guttery, D.S.; Zhang, X.; Zhang, Y.-D. COVID-19 classification by CCSHNet with deep fusion using transfer learning and discriminant correlation analysis. Inf. Fusion 2021, 68, 131–148. [Google Scholar] [CrossRef]
  10. Minaee, S.; Kafieh, R.; Sonka, M.; Yazdani, S.; Soufi, G.J. Deep-COVID: Predicting COVID-19 from chest X-ray images using deep transfer learning. Med. Image Anal. 2020, 65, 101794. [Google Scholar] [CrossRef] [PubMed]
  11. Ozturk, T.; Talo, M.; Yildirim, E.A.; Baloglu, U.B.; Yildirim, O.; Acharya, U.R. Automated detection of COVID-19 cases using deep neural networks with X-ray images. Comput. Biol. Med. 2020, 121, 103792. [Google Scholar] [CrossRef] [PubMed]
  12. Jia, X.; Ren, L.; Cai, J. Clinical implementation of AI technologies will require interpretable AI models. Med. Phys. 2020, 47, 1–4. [Google Scholar] [CrossRef] [PubMed]
  13. Xu, X.; Jiang, X.; Ma, C.; Du, P.; Li, X.; Lv, S.; Yu, L.; Chen, Y.; Su, J.; Lang, G. Deep Learning System to Screen novel Coronavirus Disease 2019 Pneumonia. Engineering 2020, 6, 1122–1129. [Google Scholar] [CrossRef] [PubMed]
  14. Zheng, C.; Deng, X.; Fu, Q.; Zhou, Q.; Feng, J.; Ma, H.; Liu, W.; Wang, X. Deep learning-based detection for COVID-19 from chest CT using weak label. MedRxiv 2020. [Google Scholar] [CrossRef] [Green Version]
  15. Wang, S.; Kang, B.; Ma, J.; Zeng, X.; Xiao, M.; Guo, J.; Cai, M.; Yang, J.; Li, Y.; Meng, X.; et al. A deep learning algorithm using CT images to screen for Corona Virus Disease (COVID-19). MedRxiv 2021, 31, 6096–6104. [Google Scholar] [CrossRef]
  16. Song, Y.; Zheng, S.; Li, L.; Zhang, X.; Zhang, X.; Huang, Z.; Chen, J.; Zhao, H.; Jie, Y.; Wang, R. Deep learning Enables Accurate Diagnosis of Novel Coronavirus (COVID-19) with CT images. IEEE/ACM Trans. Comput. Biol. Bioinform. 2020, 18, 2775–2780. [Google Scholar] [CrossRef]
  17. Narin, A.; Kaya, C.; Pamuk, Z. Automatic detection of coronavirus disease (COVID-19) using X-ray images and deep convolutional neural networks. Pattern Anal. Appl. 2021, 24, 1207–1220. [Google Scholar] [CrossRef]
  18. Hemdan, E.E.D.; Shouman, M.A.; Karar, M.E. Covidx-net: A framework of deep learning classifiers to diagnose covid-19 in X-ray images. arXiv 2020, arXiv:2003.11055. [Google Scholar]
  19. Kumar, P.; Kumari, S. Detection of coronavirus disease (COVID-19) based on deep features and support vector machine. Preprints 2020, 2020030300. [Google Scholar] [CrossRef] [Green Version]
  20. Wang, L.; Lin, Z.Q.; Wong, A. COVID-Net: A tailored deep convolutional neural network design for detection of COVID-19 cases from chest X-ray images. Sci. Rep. 2020, 10, 19549. [Google Scholar] [CrossRef]
  21. Apostolopoulos, I.D.; Mpesiana, T.A. Covid-19: Automatic detection from X-ray images utilizing transfer learning with convolutional neural networks. Phys. Eng. Sci. Med. 2020, 43, 635–640. [Google Scholar] [CrossRef] [PubMed]
  22. Karim, M.R.; Döhmen, T.; Cochez, M.; Beyan, O.; Rebholz-Schuhmann, D.; Decker, S. Deepcovidex-plainer: Explainable COVID-19 diagnosis from chest X-ray images. In Proceedings of the 2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Seoul, Korea, 16–19 December 2020; pp. 1034–1037. [Google Scholar]
  23. Brunese, L.; Mercaldo, F.; Reginelli, A.; Santone, A. Explainable Deep Learning for Pulmonary Disease and Coronavirus COVID-19 Detection from X-rays. Comput. Methods Programs Biomed. 2020, 196, 105608. [Google Scholar] [CrossRef]
  24. Deng, Z.; Jiang, Y.; Choi, K.-S.; Chung, F.-L.; Wang, S. Knowledge-Leverage-Based TSK Fuzzy System Modeling. IEEE Trans. Neural Netw. Learn. Syst. 2013, 24, 1200–1212. [Google Scholar] [CrossRef] [PubMed]
  25. Zhang, Y.; Ishibuchi, H.; Wang, S. Deep Takagi–Sugeno–Kang Fuzzy Classifier with Shared Linguistic Fuzzy Rules. IEEE Trans. Fuzzy Syst. 2017, 26, 1535–1549. [Google Scholar] [CrossRef]
  26. Xia, K.; Zhang, Y.; Jiang, Y.; Qian, P.; Dong, J.; Yin, H.; Muzic, R.F. TSK Fuzzy System for Multi-View Data Discovery Underlying Label Relaxation and Cross-Rule & Cross-View Sparsity Regularizations. IEEE Trans. Ind. Inform. 2021, 17, 3282–3291. [Google Scholar] [CrossRef]
  27. Jiang, Y.; Deng, Z.; Chung, F.-L.; Wang, G.; Qian, P.; Choi, K.-S.; Wang, S. Recognition of Epileptic EEG Signals Using a Novel Multiview TSK Fuzzy System. IEEE Trans. Fuzzy Syst. 2016, 25, 3–20. [Google Scholar] [CrossRef]
  28. Lin, C.-L.; Hsieh, S.-T. Work-In-Progress: An intelligent diagnosis influenza system based on adaptive neuro-fuzzy inference system. In Proceedings of the 2015 1st International Conference on Industrial Networks and Intelligent Systems (INISCom), Tokyo, Japan, 2–4 March 2015; pp. 177–180. [Google Scholar] [CrossRef] [Green Version]
  29. Khayamnia, M.; Yazdchi, M.; Vahidiankamyad, A.; Foroughipour, M. The recognition of migraine headache by designation of fuzzy expert system and usage of LFE learning algorithm. In Proceedings of the 2017 5th Iranian Joint Congress on Fuzzy and Intelligent Sys-tems (CFIS), Qazvin, Iran, 7–9 March 2017; pp. 50–53. [Google Scholar] [CrossRef]
  30. Shao, L.; Liu, L.; Li, X. Feature Learning for Image Classification via Multiobjective Genetic Programming. IEEE Trans. Neural Netw. Learn. Syst. 2013, 25, 1359–1371. [Google Scholar] [CrossRef]
  31. Fan, Z.; Xu, Y.; Zhang, D. Local Linear Discriminant Analysis Framework Using Sample Neighbors. IEEE Trans. Neural Netw. 2011, 22, 1119–1132. [Google Scholar] [CrossRef]
  32. Łęski, J. Ho–Kashyap classifier with generalization control. Pattern Recognit. Lett. 2003, 24, 2281–2290. [Google Scholar] [CrossRef]
  33. Deng, Z.; Choi, K.-S.; Jiang, Y.; Wang, S. Generalized Hidden-Mapping Ridge Regression, Knowledge-Leveraged Inductive Transfer Learning for Neural Networks, Fuzzy Systems and Kernel Methods. IEEE Trans. Cybern. 2014, 44, 2585–2599. [Google Scholar] [CrossRef]
  34. Fang, X.; Xu, Y.; Li, X.; Lai, Z.; Wong, W.K.; Fang, B. Regularized Label Relaxation Linear Regression. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 1006–1018. [Google Scholar] [CrossRef]
  35. Liu, J.; Chung, F.-L.; Wang, S. Bayesian zero-order TSK fuzzy system modeling. Appl. Soft Comput. 2017, 55, 253–264. [Google Scholar] [CrossRef]
  36. Juang, C.-F.; Chiu, S.-H.; Shiu, S.-J. Fuzzy System Learned Through Fuzzy Clustering and Support Vector Machine for Human Skin Color Segmentation. IEEE Trans. Syst. Man Cybern.-Part A Syst. Hum. 2007, 37, 1077–1087. [Google Scholar] [CrossRef]
  37. Tsang, I.W.; Kwok, J.T.; Cheung, P.M.; Cristianini, N. Core vector machines: Fast SVM training on very large data sets. J. Mach. Learn. Res. 2005, 6, 363–392. [Google Scholar]
  38. Yang, C.; Deng, Z.; Choi, K.-S.; Wang, S. Takagi–Sugeno–Kang Transfer Learning Fuzzy Logic System for the Adaptive Recognition of Epileptic Electroencephalogram Signals. IEEE Trans. Fuzzy Syst. 2015, 24, 1079–1094. [Google Scholar] [CrossRef]
Figure 1. Workflow of data preprocessing.
Figure 1. Workflow of data preprocessing.
Diagnostics 12 02613 g001
Figure 2. Workflow of model evaluation.
Figure 2. Workflow of model evaluation.
Diagnostics 12 02613 g002
Figure 3. Classification performance in terms of accuracy of all models. (a) Testing accuracy. (b) Training accuracy.
Figure 3. Classification performance in terms of accuracy of all models. (a) Testing accuracy. (b) Training accuracy.
Diagnostics 12 02613 g003
Figure 4. Accuracy differences between training and testing sets.
Figure 4. Accuracy differences between training and testing sets.
Diagnostics 12 02613 g004
Figure 5. Evaluation of manifold regularization. (a) Accuracy difference comparison when manifold regularization is absent. (b) Accuracy comparison between γ = 0 and γ = 0 . 01 .
Figure 5. Evaluation of manifold regularization. (a) Accuracy difference comparison when manifold regularization is absent. (b) Accuracy comparison between γ = 0 and γ = 0 . 01 .
Diagnostics 12 02613 g005
Figure 6. Robustness analysis with respect to λ and γ .
Figure 6. Robustness analysis with respect to λ and γ .
Diagnostics 12 02613 g006
Figure 7. The fuzzy membership function and the corresponding linguistic description of “wavelet-HH_glrlm_GrayLevel Variance” in each fuzzy rule.
Figure 7. The fuzzy membership function and the corresponding linguistic description of “wavelet-HH_glrlm_GrayLevel Variance” in each fuzzy rule.
Diagnostics 12 02613 g007
Figure 8. Interpretable diagnostic procedure.
Figure 8. Interpretable diagnostic procedure.
Diagnostics 12 02613 g008
Table 1. AI models for COVID-19 cases detection on chest radiography images.
Table 1. AI models for COVID-19 cases detection on chest radiography images.
StudiesModalitiesNumber of CasesAI ModelsBest Performance
COVID-19 detection from healthy and pneumonia cases [13].CCT *219 COVID-19 cases (S)
224 Pneumonia cases
175 Healthy cases
ResNet with location attention mechanismOverall accuracy rate was 86.7%.
Severity rating [14].CCT313 COVID-19 cases (S)
229 COVID-19 cases (M)
UNet and 3D Deep NetworkThe model obtained a testing AUC of 0.975.
Severity rating [15].CCT195 COVID-19 cases (S)
258 COVID-19 cases (M)
M-InceptionThe external testing dataset showed a total accuracy of 79.3% with a specificity of 0.83 and sensitivity of 0.67.
COVID-19 detection from healthy cases [16].CCT777 COVID-19 cases (S)
708 Healthy cases
DRE-NetThe model discriminated the COVID-19 patients from the bacteria pneumonia patients with an AUC of 0.95, recall (sensitivity) of 0.96, and precision of 0.79.
Severity rating [17].CXR **50 COVID-19 cases (S)
50 COVID-19 cases (M)
Deep CNN and ResNet-50The model achieved 99.7% accuracy for automatic detection of COVID-19.
COVID-19 detection from healthy cases [18].CXR25 COVID-19 cases (S)
25 Healthy cases
COVIDX-NetThe model achieved f1-scores of 0.89 and 0.91 for normal and COVID-19, respectively.
COVID-19 detection from healthy cases [19].CXR25 COVID-19 cases (S)
25 COVID-19 cases (M)
ResNet50 and SVMThe highest accuracy achieved by ResNet50 plus SVM is 98.66%.
COVID-19 detection from healthy cases [20].CXR53 COVID-19 cases (S)
5526 COVID-19 cases (M)
COVID-NetCOVID-Net achieves good accuracy by achieving 93.3% test accuracy.
COVID-19 detection from healthy and pneumonia cases [21].CXR224 COVID-19 cases (S)
700 Pneumonia cases
504 Healthy cases
VGG-19The best accuracy, sensitivity, and specificity obtained is 96.78%, 98.66%, and 96.46% respectively.
COVID-19 detection from healthy and pneumonia cases [22]CXR15,959 CXR images of 15,854 patients, covering normal, pneumonia, and COVID-19 cases.Deep CNNThe model achieved best positive predictive value of 91.6%, 92.45%, and 96.12%, respectively for normal, pneumonia, and COVID-19 cases, respectively.
COVID-19 detection from healthy and pneumonia cases [23]CXR250 COVID-19 cases
2753 pulmonary cases
3520 healthy cases.
Deep CNNThe best accuracy is 0.99.
*: Chest CT; **: Chest X-Ray. M: mild. S: severity.
Table 2. Experimental settings.
Table 2. Experimental settings.
ModelParameter Settings
ZERO-TSK-FSFCM is used to learn the antecedent parameters. The optimal number of fuzzy rules is determined by cross-validation from [2, 4, 6, …, 30].
L2-TSK-FSFCM is used to learn the antecedent parameters. The optimal number of fuzzy rules is determined by cross-validation from [2, 4, 6, …, 30].
FS-FCSVMFCM is used to learn the antecedent parameters. The optimal number of fuzzy rules is determined by cross-validation from [2, 4, 6, …, 30]. The learning threshold parameter is determined by cross-validation from [0.2, 0.3, …, 0.8]. The regularization parameter is determined by cross-validation from [2−3, 2−2, …, 25, 27].
L-SVMThe regularization parameter C is determined by cross-validation from [10−3, 2−2, …, 103].
G-SVMThe regularization parameter C is determined by cross-validation from [10−3, 2−2, …, 103]. The kernel width σ is determined by cross-validation from [10−3, 10−2, …, 103].
LR-ZERO-TSK-FSFCM is used to learn the antecedent parameters. The optimal number of fuzzy rules is determined by cross-validation from [2, 4, 6, …, 30]. The regularization parameter λ and γ are both determined by cross-validation from [0.01, 0.03, …, 1].
Software and hardware settings
Development platform: Python 3.9.8 (pyRadiomics for radiomics feature extraction), Matlab 2012b (LR-ZERO-TSK-FS coding)
System OS: Windows 10
Hardware: Intel(R) Core (TM) i5-7200U CPU @ 2.50 GHz 2.71 GHz, 16 G RAM
Table 3. Statistical analysis of testing performance in terms of t-test.
Table 3. Statistical analysis of testing performance in terms of t-test.
Number of FeaturesG-SVML-SVMZERO-TSK-FSL2-TSK-FSFCSVM-FS
5*****
10 * *
15*****
20*****
25*****
30*****
35 ****
40****
45*****
50*****
* means that there exist significant differences between LR-ZERO-TSK-FS and the state-of-the-art models (p < 0.05).
Table 4. Classification performance in terms of training macro F1-score.
Table 4. Classification performance in terms of training macro F1-score.
Number of FeaturesG-SVML-SVMZERO-TSK-FSL2-TSK-FSFCSVM-FSLR-ZERO-TSK-FS
50.52890.59470.56740.60880.50980.6145
100.62870.57920.60650.64740.60450.6024
150.61590.68010.65430.66120.61810.6887
200.67010.68100.68030.67040.62510.6912
250.63420.70310.70320.63050.68310.7253
300.78010.76030.74540.65040.64480.7923
350.79560.72510.70020.64050.65230.7664
400.74060.76020.68030.76030.75030.7661
450.75100.76040.77110.73030.78030.8047
500.73010.74020.76000.76000.75890.8232
Table 5. Classification performance in terms of testing macro F1-score.
Table 5. Classification performance in terms of testing macro F1-score.
Number of FeaturesG-SVML-SVMZERO-TSK-FSL2-TSK-FSFCSVM-FSLR-ZERO-TSK-FS
50.55120.57980.59250.59230.59230.5812
100.58230.66120.64870.64870.64870.6256
150.67410.73610.73670.71180.71230.7220
200.70190.70120.70100.70120.71240.7489
250.71890.74900.71020.70010.73020.7491
300.74780.77310.76120.77420.77810.8024
350.74760.78000.78030.77320.79210.8215
400.78030.75860.81370.78030.78030.8510
450.78970.78980.79110.80210.79230.8454
500.80050.80050.80080.80090.80010.8702
Table 6. CPU seconds all algorithms consumed.
Table 6. CPU seconds all algorithms consumed.
Number of FeaturesG-SVML-SVMZERO-TSK-FSL2-TSK-FSFCSVM-FS
51.761.432.111.982.03
102.322.673.332.992.92
152.913.043.503.143.24
203.642.984.114.043.96
257.216.235.555.986.01
307.876.656.047.137.27
358.037.757.568.097.85
4010.439.847.999.438.43
4513.6713.4816.0915.4316.22
5016.4715.9916.7317.4817.78
Table 7. Optimal features for model training.
Table 7. Optimal features for model training.
No.Feature NameNo.Feature Name
1‘wavelet-HH_glrlm_GrayLevelVariance’16‘logarithm_gldm_GrayLevelVariance’
2‘squareroot_glrlm_GrayLevelNonUniformityNormalized’17‘wavelet-HL_glrlm_GrayLevelNonUniformityNormalized’
3‘exponential_firstorder_RobustMeanAbsoluteDeviation’18‘wavelet-HL_ngtdm_Busyness’
4‘gradient_glcm_ClusterProminence’19‘squareroot_glszm_LargeAreaEmphasis’
5‘squareroot_glcm_MCC’20‘squareroot_glrlm_LongRunLowGrayLevelEmphasis’
6‘squareroot_glcm_ClusterProminence’21‘wavelet-LH_glrlm_RunEntropy’
7‘gradient_glszm_GrayLevelVariance’22‘wavelet-LH_glrlm_GrayLevelNonUniformityNormalized’
8‘wavelet-LH_gldm_LowGrayLevelEmphasis’23‘wavelet-HH_gldm_DependenceNonUniformityNormalized’
9‘exponential_firstorder_Range’24‘exponential_glszm_LargeAreaHighGrayLevelEmphasis’
10‘wavelet-LH_glszm_SmallAreaEmphasis’25‘square_glcm_DifferenceEntropy’
11‘square_gldm_DependenceNonUniformityNormalized’26‘original_glcm_SumSquares’
12‘gradient_glcm_JointAverage’27‘squareroot_glcm_Autocorrelation’
13‘logarithm_glrlm_LongRunHighGrayLevelEmphasis’28‘logarithm_glrlm_LongRunEmphasis’
14‘wavelet-LH_glrlm_LongRunLowGrayLevelEmphasis’29‘wavelet-HH_firstorder_90Percentile’
15‘wavelet-LH_glszm_HighGrayLevelZoneEmphasis’30‘wavelet-LH_glrlm_RunEntropy’
Table 8. Five trained fuzzy rules.
Table 8. Five trained fuzzy rules.
LR-ZERO-TSK-FS
k - th   fuzzy   rule :   If   x i 1   is   ϑ 1 k x i 2   is   ϑ 2 k x i d   is   ϑ d k ,   then   f k ( x i ) = β k , k = 1 , 2 , , K .
No.AntecedentConsequent
1[0.3888, 0.4187, 0.0043, 0.3884, 0.2543, …]
[0.0080, 0.0051, 0.0011, 0.0079, 0.0034, …]
[0.0340, 0.0087, 0.0024]
2[0.4791, 0.5709, 0.0008, 0.4801, 0.2732, …]
[0.0214, 0.0105, 0.0005, 0.0213, 0.0068, …]
[0.0096, 0.0148, 0.0119]
3[0.2726, 0.4477, 0.0018, 0.2746, 0.2200, …]
[0.0199, 0.0088, 0.0007, 0.0198, 0.0050, …]
[0.0100, 0.0108, 0.0077]
4[0.3443, 0.2711, 0.0072, 0.3471, 0.1323, …]
[0.0156, 0.0104, 0.0015, 0.0153, 0.0042, …]
[0.0048, 0.0189, 0.0202]
5[0.6174, 0.3593, 0.0018, 0.6172, 0.1463, …]
[0.0283, 0.0094, 0.0007, 0.0279, 0.0048, …]
[0.0026, 0.0109, 0.0165]
Table 9. Five trained fuzzy rules in terms of “If…then…”.
Table 9. Five trained fuzzy rules in terms of “If…then…”.
Rule No.If PartThen Part
1No.1 is “low”, No.2 is “low”, No.3 is “a little low”, No.4 is “a little low”, No.5 is “a litter low”, … f 1 ( x ) = 0 . 0340 for class 1, f 1 ( x ) = 0 . 0087 for class 2 and f 1 ( x ) = 0 . 0024 for class 3.
2No.1 is “low”, No.2 is “low”, No.3 is “a little low”, No.4 is “a little low”, No.5 is “a little low”, … f 2 ( x ) = 0 . 0096 for class 1, f 2 ( x ) = 0 . 0148 for class 2 and f 2 ( x ) = 0 . 0119 for class 3.
3No.1 is “low”, No.2 is “low”, No.3 is “a little low”, No.4 is “a little low”, No.5 is “a little low”, … f 3 ( x ) = 0 . 0100 for class 1, f 3 ( x ) = 0 . 0108 for class 2 and f 3 ( x ) = 0 . 0077 for class 3.
4No.1 is “low”, No.2 is “low”, No.3 is “a little low”, No.4 is “a little low”, No.5 is “a litter low”, … f 4 ( x ) = 0 . 0048 for class 1, f 4 ( x ) = 0 . 0189 for class 2 and f 4 ( x ) = 0 . 0202 for class 3.
5No.1 is “low”, No.2 is “low”, No.3 is “a little low”, No.4 is “a little low”, No.5 is “a little low”, … f 5 ( x ) = 0 . 0026 for class 1, f 5 ( x ) = 0 . 0109 for class 2 and f 5 ( x ) = 0 . 0165 for class 3.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Yang, D.; Lam, S.; Li, B.; Teng, X.; Zhang, J.; Zhou, T.; Ma, Z.; Ying, T.-C.; Cai, J. Radiomics-Based Detection of COVID-19 from Chest X-ray Using Interpretable Soft Label-Driven TSK Fuzzy Classifier. Diagnostics 2022, 12, 2613. https://doi.org/10.3390/diagnostics12112613

AMA Style

Zhang Y, Yang D, Lam S, Li B, Teng X, Zhang J, Zhou T, Ma Z, Ying T-C, Cai J. Radiomics-Based Detection of COVID-19 from Chest X-ray Using Interpretable Soft Label-Driven TSK Fuzzy Classifier. Diagnostics. 2022; 12(11):2613. https://doi.org/10.3390/diagnostics12112613

Chicago/Turabian Style

Zhang, Yuanpeng, Dongrong Yang, Saikit Lam, Bing Li, Xinzhi Teng, Jiang Zhang, Ta Zhou, Zongrui Ma, Tin-Cheung (Michael) Ying, and Jing Cai. 2022. "Radiomics-Based Detection of COVID-19 from Chest X-ray Using Interpretable Soft Label-Driven TSK Fuzzy Classifier" Diagnostics 12, no. 11: 2613. https://doi.org/10.3390/diagnostics12112613

APA Style

Zhang, Y., Yang, D., Lam, S., Li, B., Teng, X., Zhang, J., Zhou, T., Ma, Z., Ying, T.-C., & Cai, J. (2022). Radiomics-Based Detection of COVID-19 from Chest X-ray Using Interpretable Soft Label-Driven TSK Fuzzy Classifier. Diagnostics, 12(11), 2613. https://doi.org/10.3390/diagnostics12112613

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop