Next Article in Journal
Genomic and Transcriptional Profiling of Chinese Melanoma Patients Enhanced Potentially Druggable Targets: A Multicenter Study
Previous Article in Journal
Dose Reduction to Motor Structures in Adjuvant Fractionated Stereotactic Radiotherapy of Brain Metastases: nTMS-Derived DTI-Based Motor Fiber Tracking in Treatment Planning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Habitat Imaging Biomarkers for Diagnosis and Prognosis in Cancer Patients Infected with COVID-19

by
Muhammad Aminu
1,
Divya Yadav
2,
Lingzhi Hong
1,
Elliana Young
3,
Paul Edelkamp, Jr.
3,
Maliazurina Saad
1,
Morteza Salehjahromi
1,
Pingjun Chen
1,
Sheeba J. Sujit
1,
Melissa M. Chen
4,
Bradley Sabloff
5,
Gregory Gladish
5,
Patricia M. de Groot
5,
Myrna C. B. Godoy
5,
Tina Cascone
6,
Natalie I. Vokes
6,7,
Jianjun Zhang
6,7,
Kristy K. Brock
1,
Naval Daver
8,
Scott E. Woodman
7,
Hussein A. Tawbi
9,
Ajay Sheshadri
10,
J. Jack Lee
11,
David Jaffray
12,
D3CODE Team
,
Carol C. Wu
5,†,
Caroline Chung
13,† and
Jia Wu
1,6,*,†
add Show full author list remove Hide full author list
1
Department of Imaging Physics, MD Anderson Cancer Center, 1515 Holcombe Blvd, Houston, TX 77030, USA
2
Department of Radiation Oncology, MD Anderson Cancer Center, Houston, TX 77054, USA
3
Department of Enterprise Data Engineering & Analytics, MD Anderson Cancer Center, Houston, TX 77054, USA
4
Department of Neuroradiology, MD Anderson Cancer Center, Houston, TX 77054, USA
5
Department of Thoracic Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX 77054, USA
6
Department of Thoracic/Head and Neck Medical Oncology, MD Anderson Cancer Center, Houston, TX 77054, USA
7
Department of Genomic Medicine, MD Anderson Cancer Center, Houston, TX 77054, USA
8
Department of Leukemia, MD Anderson Cancer Center, Houston, TX 77054, USA
9
Department of Melanoma Medical Oncology, MD Anderson Cancer Center, Houston, TX 77054, USA
10
Department of Pulmonary Medicine, MD Anderson Cancer Center, Houston, TX 77054, USA
11
Department of Biostatistics, The University of Texas MD Anderson Cancer Center, Houston, TX 77054, USA
12
Office of the Chief Technology and Digital Officer, MD Anderson Cancer Center, Houston, TX 77054, USA
13
Office of the Chief Data Officer, MD Anderson Cancer Center, Houston, TX 77054, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Collaborators/Membership of the D3CODE is provided in the Supplementary Material.
Cancers 2023, 15(1), 275; https://doi.org/10.3390/cancers15010275
Submission received: 8 November 2022 / Revised: 28 December 2022 / Accepted: 29 December 2022 / Published: 31 December 2022
(This article belongs to the Section Cancer Informatics and Big Data)

Abstract

:

Simple Summary

Patients with cancer are often immuno-compromised, and at a high risk of experiencing various COVID-19-associated complications compared to the general population. Additionally, COVID-19 infection and lung toxicities due to cancer treatments can present with similar radiologic abnormalities, such as ground glass opacities or patchy consolidation, which poses further challenges for developing AI algorithms. To fill the gap, we carried out the first imaging AI study to investigate the performance of habitat imaging technique for COVID-19 severity prediction and detection specifically in the cancer patient population, and further tested its performance in the general population based on multicenter datasets. The proposed COVID-19 habitat imaging models trained separately on the cancer cohort outperformed those AI models (including deep learning) trained on the multicenter general population by a significant margin. This suggests that publicly available COVID-19 AI models developed for the general population will not be optimally applied to cancer.

Abstract

Objectives: Cancer patients have worse outcomes from the COVID-19 infection and greater need for ventilator support and elevated mortality rates than the general population. However, previous artificial intelligence (AI) studies focused on patients without cancer to develop diagnosis and severity prediction models. Little is known about how the AI models perform in cancer patients. In this study, we aim to develop a computational framework for COVID-19 diagnosis and severity prediction particularly in a cancer population and further compare it head-to-head to a general population. Methods: We have enrolled multi-center international cohorts with 531 CT scans from 502 general patients and 420 CT scans from 414 cancer patients. In particular, the habitat imaging pipeline was developed to quantify the complex infection patterns by partitioning the whole lung regions into phenotypically different subregions. Subsequently, various machine learning models nested with feature selection were built for COVID-19 detection and severity prediction. Results: These models showed almost perfect performance in COVID-19 infection diagnosis and predicting its severity during cross validation. Our analysis revealed that models built separately on the cancer population performed significantly better than those built on the general population and locked to test on the cancer population. This may be because of the significant difference among the habitat features across the two different cohorts. Conclusions: Taken together, our habitat imaging analysis as a proof-of-concept study has highlighted the unique radiologic features of cancer patients and demonstrated effectiveness of CT-based machine learning model in informing COVID-19 management in the cancer population.

1. Introduction

The global pandemic of the COVID-19 disease forced researchers to swiftly develop effective ways to mitigate the quick spread of the virus. Accurate diagnostic solutions have been developed for COVID-19, among which Reverse Transcription Polymerase Chain Reaction (RT-PCR) is considered as the gold standard in detecting the infection. However, RT-PCR has several limitations with a false-negative rate [1,2,3] that is sufficiently high and multiple tests may be required to confirm the diagnosis of SARS-CoV-2 infection. More importantly, RT-PCR offers no insight on assessing disease severity to guide patient management. Due to these drawbacks, several approaches were investigated to complement the use of the RT-PCR, including clinical symptoms, laboratory findings, and imaging. For imaging, computed tomography (CT) and chest radiography (CXR) are routinely used to guide clinicians in the diagnosis and assess the pulmonary severity of COVID-19 [4,5,6,7,8].
Patients with cancer are often immuno-compromised and at a high risk of experiencing various COVID-19 associated complications compared to the general population [9]. Pilot artificial intelligence (AI) algorithms largely focus on the general population, utilizing features such as clinical symptoms [10,11], lab tests [12,13,14], and CT findings [15,16,17] to diagnose COVID-19 and predict outcome after infection. While these AI algorithms showed initial promise, there were challenges and failures observed in the clinical implementation of these models driven by a failure to validate algorithms in heterogeneous populations as well as a change in disease presentation with SARS-CoV-2 variants [18]. The likelihood for the clinical utility of models built using data from the general population is even smaller in immunocompromised hosts, who may bear distinct risk factors for severe disease. Additionally, COVID-19 infection and lung toxicities due to cancer treatments can present with similar radiologic abnormalities, such as ground glass opacities (GGO) or patchy consolidation [19,20,21] on chest CT, which poses further challenges for AI detection algorithms. Though pilot AI models have been developed to predict COVID-19 severity and deterioration specifically in cancer patients [22,23], they left out the quantitative imaging metrics. Further, it is unclear whether and to what degree these AI models built on the cancer population are different from the ones on the general population, which is what we aim to investigate in this study.
Imaging findings in COVID-19 vary greatly from patient to patient, including in the extent and heterogeneity of involvement and the characteristics of the lung infiltrates. Radiologists score pneumonia severity by assessing the percentage and distribution of distinct infected regions, such as ground-glass opacities (GGO) and consolidation. These standards are known to be coarse, subjective, and not robust enough to characterize complex infection patterns. Many AI approaches (radiomics and deep learning) attempt to automate these measurements through crafting robust computational pipelines. On the other hand, habitat imaging offers an avenue for us to define intrinsic infection patterns. We have shown that habitat imaging quantifies intratumoral heterogeneity in cancer patients [24]. Unlike classical radiomics analysis which treats the heterogeneous tumor as one entity, habitat imaging explicitly partitions the complex tumor into phenotypically distinct subregions, where these intratumoral subregions are termed habitats. These subregions have differing prognostic implications for cancer severity. Similar to genomic sequencing studies that show the clonal diversity of cancer cells [25], habitat imaging offers a new and powerful avenue to investigate how molecular diversity manifests on the radiological scans, and we have demonstrated added value in predicting treatment response [26].
The main goal of this study is to investigate the performance of habitat imaging technique for COVID-19 prognosis (primary aim) as well as detection (secondary aim) in cancer patient population and compare its performance to the general population. Different from existing AI studies that mainly focused on the general population, here we specifically focus on the COVID-vulnerable cancer population.

2. Materials and Methods

2.1. Overall Study Design

The overall goal is to develop and test imaging biomarkers for the diagnosis and prognosis of COVID-19 in the cancer population and further compare its performance in the general population (dataset details in Table 1). To achieve this goal, we proposed a habitat imaging framework that consists of three main steps (Figure 1). First, we applied imaging preprocessing and fusion pipeline to highlight the lung infection patterns. Second, an unsupervised subregion segmentation approach (i.e., habitat analysis) is proposed to reveal these complex infection patterns by partitioning the whole lung regions into phenotypically different habitats. Third, we characterized the spatial arrangement/interaction among these habitats and built machine learning models for COVID-19 detection and severity prediction.

2.2. Patient Population

This retrospective study was approved by the MD Anderson institutional review board and compliant with the Health Insurance Portability and Accountability Act. We enrolled two types of patient population, including a general population set and a cancer population set (Table 1).
Table 1. Demographic and clinical characteristics of general and cancer population. Note, There are three datasets, Stony Brook (N = 275 patients with 304 CT scans), RICORD COVID-19 positive (N = 110 patients with 110 CT scans) and RICORD COVID-19 negative (N = 117 cases with 117 CT scans) that we referred to as the general population. Similarly, we have three datasets, MD Anderson COVID-19 positive (N = 252 patients with 258 CT scans), Leukemia (N = 126 patients with 126 CT scans) and Melanoma (N = 36 patients with 36 CT scans) that we referred to as cancer population. The Leukemia and Melanoma subcohorts are not highlighted in this table as the detailed demographic information for these datasets is not available.
Table 1. Demographic and clinical characteristics of general and cancer population. Note, There are three datasets, Stony Brook (N = 275 patients with 304 CT scans), RICORD COVID-19 positive (N = 110 patients with 110 CT scans) and RICORD COVID-19 negative (N = 117 cases with 117 CT scans) that we referred to as the general population. Similarly, we have three datasets, MD Anderson COVID-19 positive (N = 252 patients with 258 CT scans), Leukemia (N = 126 patients with 126 CT scans) and Melanoma (N = 36 patients with 36 CT scans) that we referred to as cancer population. The Leukemia and Melanoma subcohorts are not highlighted in this table as the detailed demographic information for these datasets is not available.
Datasets CharacteristicsStony Brook
(N = 275)
MD Anderson COVID-19 Positive (N = 252)
No. (%)No. (%)
Age (years)
<18-1 (0.4%)
18–59125 (45%)112 (44%)
60–7483 (30%)99 (39%)
75–9067 (24)37 (15%)
>90-3 (1%)
Sex
Male160 (58%)121 (48%)
Female106 (39%)131 (52%)
NA9 (3%)-
BMI
>30100 (36%)70 (28%)
<30127 (46%)95 (38%)
NA48 (17%)87 (35%)
Smoking status
Current6 (2%)11 (5%)
Former63 (23%)90 (36%)
Never138 (50%)148 (59%)
NA68 (25%)3 (1%)
Major diseases
Malignancy25 (9%)240 (95%)
Hypertension105 (38%)136 (54%)
Diabetes54 (20%)105 (42%)
Coronary artery diseases33 (12%)50 (20%)
Chronic kidney disease19 (7%)75 (30%)
Chronic obstructive pulmonary disease 18 (7%)15 (6%)
Other lung diseases 39 (14%)22 (9%)
Symptoms at onset
Fever 158 (57%)33 (13%)
Shortness of breath 152 (55%)14 (6%)
Cough 157 (57%)39 (15%)
Nausea 49 (18%)17 (7%)
Vomiting 29 (11%)9 (4%)
Diarrhea 65 (24%)17 (7%)
Abdominal pain19 (7%)7 (3%)
Hospitalization status
Inpatient/admitted259 (94%)86 (34%)
Outpatient -166 (66%)
Emergency Visit16 (6%)-
ICU
TRUE83 (30%)42 (17%)
FALSE192 (70%)209 (83%)
Oxygen requirement
TRUE71 (26%)20 (8%)
FALSE204 (74%)232 (92%)
For the general population, a total of 531 CT scan images were obtained from public TCIA database (Stony Brook [27,28] and RICORD [28,29,30] cohorts), including 385 patients with confirmed COVID-19 positive with RT-PCR and 117 COVID-19 negative patients from four international sites. For the cancer population, a total of 420 CT scan images from 414 cancer patients receiving treatment at MD Anderson cancer center were included. Among the 414 cancer patients, 252 patients were confirmed to have COVID-19 infection with RT-PCR while the remaining 162 patients were COVID-19 negative per RT-PCR but had other causes of pneumonia or pneumonitis. We utilized the available CT scan images of same patient at multiple timepoints to build the diagnostic models. For the prognostic model, only the CT scans closest to the PCR test were used to predict the end points (hospitalization, ICU admission and ventilation).

2.3. Imaging Preprocessing and Fusion

We utilized the U-net [31] to segment the right and the left lung from the original CT images. Then, the CT density of lungs was normalized using both the lung and mediastinum windows. Since the infected regions within the lungs manifested as ground glass opacity (GGO), consolidation, and their mixture, which usually corresponded to increased heterogeneity compared to the normal lung parenchyma, we specifically characterized these local density and texture variations using the entropy filter. For a small neighborhood N l with window size n l × m l within an image I , we can compute the local entropy of the neighborhood N l as:
E N l   = r = 0 M 1 p r log p r
where p r = q r / n l × m l denotes the probability of the grayscale r appearing in the neighborhood N l and q r denotes the number of pixels with grayscale r in N l .   M is the maximum grayscale. The filtered image channels were fused together to form the final image on which the habitat analysis is applied. Figure 1 shows an example of the lung window and mediastinum window, the corresponding results of their filtered local entropy maps (Equation (1)), and the fused image. Note that brighter pixels in the filtered maps correspond to neighborhoods with high local entropy in the original images.

2.4. Habitat Image Pipeline

After imaging harmonization, the habitat imaging is an unsupervised (clustering based) approach [26,32,33] containing two clustering modules at both individual and global levels, as illustrated in Step 2 of Figure 1. The individual level module is designed to segregate each patient’s lungs into superpixels by grouping neighboring pixels of similar intensity and texture, whereas the global level module was designed to refine the individual-level over-segmentation by exploiting the similarity of both inter and intra-patient superpixels.
Individual level clustering module
Here, we aimed to partition each of the input 3D lungs into phenotypically similar small pieces. Given the fused maps, the simple linear iterative clustering (SLIC) [34] algorithm was utilized to over-segment them into a cluster of superpixels. The main idea behind the SLIC superpixel algorithm is to adopt the local kmeans algorithm to generate clusters grouped by neighboring voxels of similar imaging patterns. In detail, the SLIC algorithm involves two steps:
Step 1. Cluster center initialization: The algorithm starts by splitting the input image into K regular grids and then chooses the center of each grid as the initial cluster centers. To avoid assigning a cluster center on image edge, the edge pixels with the lowest gradient in a 3 × 3 × 3 neighborhood is chosen as the initial seed location.
Step 2. Local kmeans clustering: Each pixel is then assigned to the nearest cluster center using a distance measure D :
D i , j = d c i , j m 2 + d s i , j S 2
where d c i , j = a i a j and d s i , j = x j x i 2 + y j y i 2 + z j z i 2 , a i denotes the mean density of the voxels, d c denotes the density proximity, d s denotes the spatial proximity, S denotes the initial grid step (sampling interval of the cluster centroids) and m is a parameter that controls the compactness of the superpixels. The distance (Equation (2)) is computed within a confined region around the cluster center and this step is repeated until the maximum number of iterations is reached or the residual error converges.
Global level clustering module
We merged the superpixels obtained for individual patients to study the inter and intra patient similarity, so that superpixels with similar features/characteristics within a lung were fused to form a habitat (subregion). Moreover, corresponding subregions across the entire population were consistently labeled. In particular, each superpixel was represented by first order statistics of its encompassing voxels’ four channels (lung and mediastinum windows, and their entropy), and subsequently, kmeans clustering algorithm was used to define the optimal habitat regions.
Multiregional spatial interaction (MSI) feature extraction
The multiregional spatial interaction (MSI) matrix was used to quantify the intra-lung infection heterogeneity on the habitat maps (Step 3 of Figure 1). In detail, for any voxel within each lung, we scanned its neighbor and the co-occurring pairs were added to the corresponding cell in the MSI matrix (Step 3 of Figure 1). After looping through all lung voxels, the habitats’ spatial distribution and interaction patterns were abstracted in this MSI matrix. Subsequently, a set of quantitative features were extracted from the MSI matrix including first order and second order statistical features. The first order statistical features include the absolute volume and relative proportion of individual subregions as well as their interacting boarders, whereas the second order statistical features summarize the subregion’s spatial heterogeneity, including the contrast, homogeneity, correlation, and energy.

2.5. Habitat Imaging Based Machine Learning Model

To improve their generalizability, we aimed to build a parsimonious model by selecting features with high discriminant power. Two clinical endpoints were studied: (1) COVID-19 diagnosis; (2) prognosis, including admission type: outpatient vs. inpatient, ICU, and ventilation. Given these extracted features, we used a univariate feature ranking approach (chi-square test) to evaluate each feature’s association with individual endpoints. The feature importance score is computed based on the p -value of univariate testing ( log ( p ) ). The features were ranked by computed scores where feature with the largest score was considered as the most important feature. We then iteratively selected the top ranked features to build four types of machine learning models, including logistic regression, generalized additive model, support vector machine, and random forest. 10-fold cross validation was used for parameter tuning and performance evaluation. Of note, the datasets for COVID-19 prognosis and severity are highly imbalanced (see Table 1). To avoid evaluation bias (e.g., assigning all cases to the dominant class), we employed the oversampling approach to augment samples from the underrepresented classes to balance their cases with majority class.

2.6. Model Interpretation

To understand the constructed habitat models, we systematically investigated its meaning. First, we overlapped the habitat subregions with radiologist’s manual annotation of COVID infected regions on CT. Then, we correlated the MSI features to the radiologist’s semantic readings. In particular, the radiologists, blinded from habitat modeling, reported CT features in structured report designed based on the RSNA COVID-19 reporting template (https://radreport.org/home/50830/2020-07-08%2011:50:07 accessed on 1 November 2021). The semantic features include presence of consolidation and/or ground glass opacities (GGO), and if present, the laterality, location, quantity GGO/consolidation, and patterns/morphology of GGO, In addition, presence or absence of centrilobular nodule, discrete solid nodule, lymphadenopathy, bronchial wall thickening, mucoid impaction, pericardial effusion, pleural effusion, pulmonary embolism, smooth septal thickening, endotracheal tube, pulmonary cavities were also noted. COVID classification patterns based on the RSNA Consensus Statement [35] and CT severity score were also recorded. Volcano plot was utilized to visualize the association of the MSI features with individual semantic readings.

2.7. Statistical Analysis

The receiver operating characteristics (ROC) curve analysis as well as the area under the curve (AUC) were used to evaluate the prediction capability of the habitat imaging models. The optimal threshold to separate different classes was defined based on the Youden’s J statistics during training and the same threshold was applied during validation. We also reported the model’s sensitivity, specificity, and accuracy. These metrics were evaluated using 10-fold cross validation scheme.

3. Results

3.1. Habitat Image Analysis

We independently applied the habitat pipeline in both the general and cancer cohorts, and consistently identified six distinct habitats (subregions) at the population level, corroborating prior study of COVID-19 infection [36]. Figure 2 shows the detailed distribution of four input image channels across the six habitat regions as well as their imaging interpretations. In general, these habitat subregions are phenotypically different and that the individual subregions were associated with distinct imaging features. The subregion 4 corresponds to the background normal lung parenchyma, and subregion 2 corresponds to the pulmonary effusion. Subregion 3 corresponds to the dense and homogeneous consolidated area, and subregion 1 corresponds to pure GGO. Interestingly, subregions 5 and 6 correspond to the infected areas from GGO to consolidation at different degrees.
Clinically, it remains a laborious task to manually delineate areas of involvement on each CT image, especially given the heterogeneous infection patterns. We compared the habitat maps with annotation of lung infection by radiologists. As shown in Figure S1, the infected regions segmented by the radiologist were mostly captured and separated from the normal lung parenchyma region using proposed habitat approach. Furthermore, the habitat imaging analysis can partition these heterogeneous infected regions into six COVID-habitat regions (Figure 3).

3.2. Habitat Models for COVID-19 Diagnosis

Given the habitat maps, fifty-eight multiregional spatial interaction features (details in Table 2) were measured from these subregions for COVID-19 detection. We conducted both unsupervised and supervised machine learning analysis to examine the diagnostic values of habitat features for COVID-19 detection.
First, we applied low-rank embedding algorithms to assess the performance of habitat features for differentiating COVID-19 positive from COVID-19 negative cases when mixing the general and cancer patients (Figure 4). The COVID-19 positive cases from the general population (RICORD, Stony Brook) and cancer population (MD Anderson Population) tend to mix and form a tight cluster, indicating high similarity among them as measured by habitat analysis. Since these embedding algorithms are designed to project the inherent local structure of high dimensional data to low dimensional subspace, the clear separation of COVID-19-positive from negative cases indicates that the habitat-derived MSI features can effectively encode discriminant imaging information. Interestingly, the MSI features can also differentiate the non-COVID-associated pneumonia across different cancers, including pneumonia in acute myeloid leukemia or melanoma cases.
Next, we built separate prediction models to detect COVID-19 infection on the general population and cancer population (Figure 5 and Table 3). For the general population, the optimized LR, SVM, RF and GAM models were fitted with the top 23, 10, 13, and 7 features, respectively, and their corresponding confusion matrices were presented in Figure 5A. Similarly, for the cancer population, the performances of these models using different number of features were reported in the accuracy and AUC curves (Figure 5B), where these models achieved high performance for COVID-19 detection.
We compared the performance of our proposed habitat imaging approach to several reported COVID-19 detection approaches from either conventional radiomics or deep learning approaches based on imaging or other data source. As can be seen in Table S1, our habitat imaging approach significantly outperformed the existing approaches.

3.3. Comparison between Diagnostic Models of General Population and Cancer Population

We first examined the feature importance ranked in two models for general or cancer population (see Figure 5) for COVID-19 diagnosis. The top ranked features in the general cohort are significantly different from cancer cohort. For instance, MSI25 which captures interaction between SR2 and SR6 appears to be the most important predictor in the general cohort, while it is not top ranked in cancer. The fluctuation across the two different cohorts is possibly due to their underlying difference in pneumonia patterns as appeared on CT scans.
Next, we tested the performance of machine learning model trained on general population directly on cancer population, where we observed a decreased prediction capacity compare with the dedicated cancer model as shown in Figure S2. In addition, when finetuning the general population model on the cancer population, we observed that the mixture of two heterogenous populations adversely affected model’s performance (Table 4).

3.4. Habitat Models for COVID-19 Prognosis and Severity Prediction

We tested the effectiveness of habitat imaging models in predicting COVID-19 disease severity with three end points include patient’s admission, need for intensive care, and need of mechanical ventilation. The habitat imaging-based classification models achieved high performance for the prognosis analysis, shown in Table 5, Tables S2 and S3, and Figure 6 and Figures S3 and S4. The prognostic models were able to obtain an AUC score of 1 for ICU and ventilation prediction on the cancer population, and AUC scores ranged from 0.96 to 1 on general population. In addition, we found the optimal prediction of the RF and GAM models using fewer number of selected features than LR and SVM models, where the curves started to plateau early on a handful of features. By contrast, the LR and SVM models kept improving with the increase in number of selected features, where they reached optimal performance using almost all the features. For example, the LR and SVM models achieved the best accuracy and AUC scores for admission prediction on cancer cohort using a total of 49 and 51 top-ranked features, respectively. Taken together, the habitat-based machine learning models can accurately infer the severity of a COVID-19 infection in both general and cancer population, by predicting whether the infected patients need to be admitted to hospital and ICU and also if they require ventilation.

3.5. Habitat Models Interpretation

We performed the correlation analysis between habitat imaging features and semantic readings from radiologists. A limited number of habitat features were observed to associate with radiologist’s readings and reached the predefined statistical significance level at a false discovery rate (FDR) of <0.05 as shown in Figure S5. For instance, MSI17 that measures interaction of habitat regions 1 and 2 and MSI19 that measures interaction of habitat regions 1 and 3 were correlated to the presence of ground glass opacities (GGO). Additionally, MSI17, that measures volume of subregion 3, was negatively correlated to presentation of pulmonary embolism, and it was found to be correlated to quantity of GGO consolidation and solid nodules.

3.6. Comparison to Deep Learning Approach

We further examine the performance of deep learning model for COVID-19 diagnosis and prognosis in the general and cancer population. Specifically, we trained a DenseNet121 model using the stochastic gradient descent (SGD) method with momentum of 0.9 and initial learning rate of 0.001. To avoid fluctuation in the later stage of training, we set the learning rate to decay by 0.9 every 10 epochs. We consider the binary cross entropy as the loss function. Inputting the CT scan images, we extracted deep features (N = 1024) at the last layer of the model and used the extracted deep features to train the same classifiers as in the habitat analysis approach.
For COVID-19 diagnosis in the general and cancer population, the best performances in terms of AUC were obtained using the LR and RF models with AUC’s 0.9792 and 0.9750, respectively (Table S4). When training the different classification models on the general population and finetuning on the cancer population, the SVM model obtained the best AUC result of 0.9673 (Table S5). The SVM model performed the best for admission, ventilation and ICU prediction in the cancer population with AUC values of 0.9924, 1.0000 and 1.0000, respectively (Tables S6–S8). Similar performance of the SVM model is observed in the general population with AUC value of 1.0000 for both admission and ventilation prediction, while GAM obtained the best AUC value of 0.9864 for ICU prediction (Tables S6–S8).

4. Discussion

In this study, we developed and validated the habitat imaging approach for COVID-19 severity prediction and detection using chest CT of a cancer population instead of a general population. Our proposed habitat analysis has demonstrated robust performance in partitioning the whole infected lungs into phenotypically distinct subregions. By quantifying the spatial distributions and interactions of these intrinsic subregions using the hand-crafted multi-regional spatial interaction features, machine learning models were built and showed high performance in COVID-19 infection diagnosis. More importantly, the habitat models demonstrated high accuracy for predicting infection severity, including the needs for hospital admission, ICU stay, and ventilation support. Taken together, our habitat imaging analysis as a proof-of-concept study has demonstrated the effectiveness of CT-based machine learning model in informing COVID-19 management in the cancer population, if prospectively validated.
To the best of our knowledge, this is the first reported imaging study that explores the performance of machine learning models on cancer patients who were infected with COVID-19, and further compared it to the general population. Cancer patients were reported to have worse COVID-19 outcomes, greater need for ventilator, and elevated mortality rates [37]. Specifically, we have demonstrated that models trained on a general population will not optimally perform when applying to cancer of phenotypically different radiographic patterns as driven by distinct underlying physiology. Systemic cancer treatment regimens can expose patients to an elevated infection risk and lead to worse COVID-19 outcomes [38]. Thus, our prognostic model can potentially shed light on informing the clinical cancer management. In this study, we have shown that the COVID-19 models trained separately on the cancer cohort outperformed those trained on the multicenter general population by a significant margin. This suggests that publicly available COVID-19 models which are developed and validated in the general population will not be optimally applied to cancer.
The superior performance of the proposed algorithms warrants its further validation for the vulnerable cancer population. If validated, the imaging-based analysis may add clinical value. For COVID-19+ cancers, our prognostic models can be used to identify high-risk patients, who may need urgent and intense medical care such as admission to hospital and ICU, and ventilation. In addition, the routine follow-up and screening CT scans can feed into our diagnostic model for accidental COVID-19 infection detection to prevent nosocomial transmission. In particular, our algorithm has potential clinical values in differentiating the COVID-19-induced pneumonia from the non-COVID-associated pneumonia/pneumonitis.
Currently, chest imaging of COVID-19 positive patient lungs exhibits variable infection patterns across different regions of the lung. Radiologists score their severity through assessing the percentage and distribution of distinct infected regions, such as GGO and consolidation. These standards are known to be qualitative, subjective, and not enough to quantify complex infection patterns. Thus, most existing AI approaches (including radiomics and deep learning) attempt to address these limitations through crafting robust computational pipelines to automate the radiologist’s annotation or semantic readings. For instance, Zhao et al. [39] developed an automatic technique to segment areas of pneumonia on CT scans and extract texture features for COVID-19 diagnosis. Bai et al. [40] developed a federated learning framework of deep learning to advance COVID-19 diagnosis on CT scans obtained from 22 hospitals, which is shown to be correlated with GGO, interlobular septal thickening and consolidation. However, historically these radiographic standards have been developed to assess pneumonia of other causes. A critical question is whether COVID-19 causes a distinct pattern of infection and whether the description of this pattern can be conveyed in the usual terminology used by radiologists. Here, we investigated this question by adapting previously validated habitat imaging pipeline [26,32,33], which has been used to identity intrinsic intra-tumoral subregions with differing imaging phenotypes (i.e., habitats).
This knowledge-discovery approach allows us to reveal more refined delineation of areas affected by pneumonia (i.e., COVID-habitats) by applying the unsupervised clustering algorithm on the CT scans. From the COVID-habitat map, it has demonstrated that these subregions not only overlapped with the areas of pneumonia as 3D annotated by the radiologist, but also divided these areas consistently into intrinsic subregions. Consequently, we have identified more refined six phenotypically distinct subregions within the lungs, which is in line with prior domain knowledge [36]. Of note, our habitat analysis leverages the unsupervised analysis to automate labeling these regions rather than leveraging the manual annotations from radiologists. Moreover, by correlating these habitat-derived features with radiologist’s semantic readings (see Figure S5), interestingly we observe that certain habitat features are significantly correlated to presence and quantity of GGO, quantity of GGO/consolidation, presence of bronchial wall thickening and pulmonary embolism. Of note, our proposed habitat approach does not require manual annotations from the radiologists, which are time consuming to obtain and often complicated with inter-observer variations. Moreover, we used unsupervised learning aims to explore data patterns which are inherently more robust than a supervised approach, less prone to model overfitting, and more likely to identify novel interactions between data that may not be readily apparent. As such, our approach can efficiently capture subtle patterns of infection and may be more robust on small and heterogeneous datasets as compared to deep learning and radiomics approaches. Given that our habitat model outperformed several existing radiomic [41,42,43,44] and deep learning approaches [15,45,46,47,48,49] (Table S1), it is plausible that our approach can improve our understanding of the radiologic manifestation of SARS-CoV-2 pneumonia in cancer patients.
Our study has some limitations. First, we did not study whether our habitat imaging models could predict COVID-19 related deaths because we had limited follow up data. In the future, exploring how patterns of COVID-19 infection affect the overall survival and disease progression in cancer patients will help us better manage this high-risk population. Second, though we have enrolled multi-institutional datasets of different populations (general vs. cancer), these initial encouraging findings need to be further validated. Third, given that the COVID-habitat-based features are fundamentally driven by differing microenvironments in infected lung regions, future efforts are needed to explore the biological drivers of the subregions on CT scans.
In conclusion, we have developed and validated habitat imaging-based CT signatures to diagnose and predict the severity of COVID-19 in cancer and general population. These CT signatures have been developed and validated using data from multiple centers. Thus, the CT signatures showed the potential to help identify cancer patients who may benefit from urgent and intense care. These results warrant further verification in future prospective data to refine such findings and test clinical utility of these imaging biomarkers to manage cancer patients infected with COVID-19.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/cancers15010275/s1, Figure S1: Comparison between radiologist segmentation and habitat subregions; Figure S2: Performance of machine learning model trained on general population and applied on cancer population; Figure S3: Performance of habitat imaging-based classification models for admission prediction; Figure S4: Performance of habitat imaging-based classification models for ICU prediction; Figure S5: Volcano plot showing association between habitat features and radiologist’s readings; Table S1: Performance comparison of our proposed approach with existing approaches; Table S2: Performance comparison of the different models for Admission prediction; Table S3: Performance comparison of the different models for ICU prediction; Table S4: Performance comparison of the different classification models for COVID-19 diagnosis using deep features; Table S5: Performance of the COVID-19 diagnostic models trained with deep features extracted on the general cohort and applied on the cancer cohort; Table S6: Performance comparison of the different classification models for admission prediction using deep feature; Table S7: Performance comparison of the different classification models for ventilation prediction using deep feature; Table S8: Performance comparison of the different classification models for COVID-19 diagnosis using deep features; D3CODE members: D3CODE team author list.

Author Contributions

Conceptualization, J.W.; Methodology, M.A., M.S. (Maliazurina Saad), M.S. (Morteza Salehjahromi), P.C., S.J.S., A.S., C.C.W., C.C. and J.W.; Software, M.A., P.E.J., M.S. (Maliazurina Saad), M.S. (Morteza Salehjahromi), P.C., S.J.S., D.J. and J.W.; Validation, M.A., M.S. (Maliazurina Saad), M.S. (Morteza Salehjahromi) and J.W.; Formal analysis, M.A., L.H., E.Y., P.E.J., M.S. (Maliazurina Saad), M.S. (Morteza Salehjahromi), P.C., S.J.S., M.M.C., B.S., G.G., P.M.d.G., M.C.B.G., T.C., N.I.V., J.Z., K.K.B., N.D., S.E.W., H.A.T., A.S., J.J.L., D.J., C.C.W., C.C. and J.W.; Investigation, N.I.V., J.Z., S.E.W., C.C.W., C.C. and J.W.; Resources, M.A., D.Y., L.H., E.Y., P.E.J., M.S. (Maliazurina Saad), M.S. (Morteza Salehjahromi), S.J.S., N.I.V., J.Z., K.K.B., S.E.W., A.S., D.J., C.C.W., C.C. and J.W.; Data curation, M.A., D.Y., L.H., E.Y., P.E.J., S.J.S., A.S., C.C.W., C.C. and J.W.; Writing—original draft, M.A., A.S. and J.W.; Writing—review & editing, M.A., D.Y., L.H., E.Y., P.E.J., M.S. (Maliazurina Saad), M.S. (Morteza Salehjahromi), P.C., S.J.S., M.M.C., B.S., G.G., P.M.d.G., M.C.B.G., T.C., N.I.V., J.Z., K.K.B., N.D., S.E.W., H.A.T., A.S., J.J.L., D.J., C.C.W., C.C. and J.W.; Visualization, M.A. and J.W.; Supervision, D.J., C.C.W., C.C. and J.W.; Project administration, S.E.W., D.J., C.C.W., C.C. and J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Tumor Measurement Initiative through the MD Anderson Strategic Initiative Development Program (STRIDE). This research was partially supported by the National Institutes of Health (NIH) grants R00 CA218667, R01 CA262425.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board of MD Anderson Cancer Center (IRB ID: 2020-0348_MOD032, approval date 21 September 2021).

Informed Consent Statement

Patient consent was waived due to retrospective nature of the study.

Data Availability Statement

The Stony Brook dataset can be found on The Cancer Imaging Archive (TCIA) at https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=89096912#89096912bcab02c187174a288dbcbf95d26179e8 accessed on 1 November 2021. The RICORD dataset [30] can be found on https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=80969742 accessed on 1 November 2021. Data from MD Anderson is currently not available at public repositories due to privacy and ethical issues but can be made available by the corresponding author upon reasonable request.

Conflicts of Interest

J.Z. reports grants from Merck, grants and personal fees from Johnson and Johnson and Novartis, personal fees from Bristol Myers Squibb, AstraZeneca, GenePlus, Innovent and Hengrui outside the submitted work. T.C. reports speaker fees/honoraria from The Society for Immunotherapy of Cancer, Bristol Myers Squibb, Roche, Medscape, IDEOlogy Health and PeerView; travel, food and beverage expenses from Dava Oncology, IDEOlogy Health and Bristol Myers Squibb; advisory role/consulting fees from MedImmune/AstraZeneca, Bristol Myers Squibb, EMD Serono, Merck & Co., Genentech, Arrowhead Pharmaceuticals, and Regeneron; and institutional research funding from MedImmune/AstraZeneca, Bristol Myers Squibb, Boehringer Ingelheim, and EMD Serono. N.I.V. receives consulting fees from Sanofi, Regeneron, Oncocyte, and Eli Lilly, and research funding from Mirati. The other authors declare no competing interest in the submitted work.

Abbreviations

RT-PCRReverse Transcription Polymerase Chain Reaction
AIartificial intelligence
GGOground glass opacities
MSImultiregional spatial interaction
ROCreceiver operating characteristics
AUCarea under the curve

References

  1. Pecoraro, V.; Negro, A.; Pirotti, T.; Trenti, T. Estimate false-negative RT-PCR rates for SARS-CoV-2. A systematic review and meta-analysis. Eur. J. Clin. Investig. 2022, 52, e13706. [Google Scholar] [CrossRef] [PubMed]
  2. Kucirka, L.M.; Lauer, S.A.; Laeyendecker, O.; Boon, D.; Lessler, J. Variation in false-negative rate of reverse transcriptase polymerase chain reaction–based SARS-CoV-2 tests by time since exposure. Ann. Intern. Med. 2020, 173, 262–267. [Google Scholar] [CrossRef]
  3. West, C.P.; Montori, V.M.; Sampathkumar, P. COVID-19 Testing: The Threat of False-Negative Results. Mayo Clin. Proc. 2020, 95, 1127–1129. [Google Scholar] [CrossRef] [PubMed]
  4. Elie, A.; Ivana, B.; Sally, Y.; Guy, F.; Roger, C.; John, A.; Mansoor, F.; Nicola, F.; Eveline, H.; Hussain, J.; et al. Use of Chest Imaging in the Diagnosis and Management of COVID-19: A WHO Rapid Advice Guide. Radiology 2020, 298, E63–E69. [Google Scholar] [CrossRef]
  5. Li, K.; Wu, J.; Wu, F.; Guo, D.; Chen, L.; Fang, Z.; Li, C. The Clinical and Chest CT Features Associated With Severe and Critical COVID-19 Pneumonia. Investig. Radiol. 2020, 55, 327–331. [Google Scholar] [CrossRef]
  6. Rodriguez-Morales, A.J.; Cardona-Ospina, J.A.; Gutiérrez-Ocampo, E.; Villamizar-Peña, R.; Holguin-Rivera, Y.; Escalera-Antezana, J.P.; Alvarado-Arnez, L.E.; Bonilla-Aldana, D.K.; Franco-Paredes, C.; Henao-Martinez, A.F.; et al. Clinical, laboratory and imaging features of COVID-19: A systematic review and meta-analysis. Travel Med. Infect. Dis. 2020, 34, 101623. [Google Scholar] [CrossRef]
  7. Paterson, R.W.; Brown, R.L.; Benjamin, L.; Nortley, R.; Wiethoff, S.; Bharucha, T.; Jayaseelan, D.L.; Kumar, G.; Raftopoulos, R.E.; Zambreanu, L.; et al. The emerging spectrum of COVID-19 neurology: Clinical, radiological and laboratory findings. Brain 2020, 143, 3104–3120. [Google Scholar] [CrossRef]
  8. Chen, J.; Qi, T.; Liu, L.; Ling, Y.; Qian, Z.; Li, T.; Li, F.; Xu, Q.; Zhang, Y.; Xu, S.; et al. Clinical progression of patients with COVID-19 in Shanghai, China. J. Infect. 2020, 80, e1–e6. [Google Scholar] [CrossRef]
  9. Desai, A.; Mohammed, T.J.; Duma, N.; Garassino, M.C.; Hicks, L.K.; Kuderer, N.M.; Lyman, G.H.; Mishra, S.; Pinato, D.J.; Rini, B.I. COVID-19 and Cancer: A Review of the Registry-Based Pandemic Response. JAMA Oncol. 2021, 7, 1882–1890. [Google Scholar] [CrossRef]
  10. Zoabi, Y.; Deri-Rozov, S.; Shomron, N. Machine learning-based prediction of COVID-19 diagnosis based on symptoms. NPJ Digit. Med. 2021, 4, 3. [Google Scholar] [CrossRef]
  11. Imran, A.; Posokhova, I.; Qureshi, H.N.; Masood, U.; Riaz, M.S.; Ali, K.; John, C.N.; Hussain, M.I.; Nabeel, M. AI4COVID-19: AI enabled preliminary diagnosis for COVID-19 from cough samples via an app. Inform. Med. Unlocked 2020, 20, 100378. [Google Scholar] [CrossRef] [PubMed]
  12. Cabitza, F.; Campagner, A.; Ferrari, D.; Di Resta, C.; Ceriotti, D.; Sabetta, E.; Colombini, A.; De Vecchi, E.; Banfi, G.; Locatelli, M.; et al. Development, evaluation, and validation of machine learning models for COVID-19 detection based on routine blood tests. Clin. Chem. Lab. Med. (CCLM) 2021, 59, 421–431. [Google Scholar] [CrossRef]
  13. Soltan, A.A.; Yang, J.; Pattanshetty, R.; Novak, A.; Yang, Y.; Rohanian, O.; Beer, S.; Soltan, M.A.; Thickett, D.R.; Fairhead, R.; et al. Real-world evaluation of rapid and laboratory-free COVID-19 triage for emergency care: External validation and pilot deployment of artificial intelligence driven screening. Lancet Digit. Health 2022, 4, e266–e278. [Google Scholar] [CrossRef] [PubMed]
  14. Soltan, A.A.; Kouchaki, S.; Zhu, T.; Kiyasseh, D.; Taylor, T.; Hussain, Z.B.; Peto, T.; Brent, A.J.; Eyre, D.W.; Clifton, D.A. Rapid triage for COVID-19 using routine clinical data for patients attending hospital: Development and prospective validation of an artificial intelligence screening test. Lancet Digit. Health 2021, 3, e78–e87. [Google Scholar] [CrossRef] [PubMed]
  15. Mei, X.; Lee, H.C.; Diao, K.Y.; Huang, M.; Lin, B.; Liu, C.; Xie, Z.; Ma, Y.; Robson, P.M.; Chung, M.; et al. Artificial intelligence–enabled rapid diagnosis of patients with COVID-19. Nat. Med. 2020, 26, 1224–1228. [Google Scholar] [CrossRef] [PubMed]
  16. Kurstjens, S.; Van Der Horst, A.; Herpers, R.; Geerits, M.W.; Kluiters-de Hingh, Y.C.; Göttgens, E.L.; Blaauw, M.J.; Thelen, M.H.; Elisen, M.G.; Kusters, R. Rapid identification of SARS-CoV-2-infected patients at the emergency department using routine testing. Clin. Chem. Lab. Med. (CCLM) 2020, 58, 1587–1593. [Google Scholar] [CrossRef] [PubMed]
  17. Shamout, F.E.; Shen, Y.; Wu, N.; Kaku, A.; Park, J.; Makino, T.; Jastrzębski, S.; Witowski, J.; Wang, D.; Zhang, B.; et al. An artificial intelligence system for predicting the deterioration of COVID-19 patients in the emergency department. NPJ Digit. Med. 2021, 4, 80. [Google Scholar] [CrossRef]
  18. Heaven, W.D. Hundreds of AI tools have been built to catch covid. None of them helped. MIT Technol. Review. Retrieved Oct. 2021, 6, 2021. [Google Scholar]
  19. Zhang, Y.J.; Yang, W.J.; Liu, D.; Cao, Y.Q.; Zheng, Y.Y.; Han, Y.C.; Jin, R.S.; Han, Y.; Wang, X.Y.; Pan, A.S.; et al. COVID-19 and early-stage lung cancer both featuring ground-glass opacities: A propensity score-matched study. Transl. Lung Cancer Res. 2020, 9, 1516–1527. [Google Scholar] [CrossRef]
  20. Guarnera, A.; Santini, E.; Podda, P. COVID-19 Pneumonia and Lung Cancer: A Challenge for the Radiologist Review of the Main Radiological Features, Differential Diagnosis and Overlapping Pathologies. Tomography 2022, 8, 513–528. [Google Scholar] [CrossRef]
  21. Dingemans, A.M.C.; Soo, R.A.; Jazieh, A.R.; Rice, S.J.; Kim, Y.T.; Teo, L.L.; Warren, G.W.; Xiao, S.Y.; Smit, E.F.; Aerts, J.G.; et al. Treatment Guidance for Patients With Lung Cancer During the Coronavirus 2019 Pandemic. J. Thorac. Oncol. 2020, 15, 1119–1136. [Google Scholar] [CrossRef] [PubMed]
  22. Xu, B.; Song, K.H.; Yao, Y.; Dong, X.R.; Li, L.J.; Wang, Q.; Yang, J.Y.; Hu, W.D.; Xie, Z.B.; Luo, Z.G.; et al. Individualized model for predicting COVID-19 deterioration in patients with cancer: A multicenter retrospective study. Cancer Sci. 2021, 112, 2522–2532. [Google Scholar] [CrossRef] [PubMed]
  23. Navlakha, S.; Morjaria, S.; Perez-Johnston, R.; Zhang, A.; Taur, Y. Projecting COVID-19 disease severity in cancer patients using purposefully-designed machine learning. BMC Infect. Dis. 2021, 21, 391. [Google Scholar] [CrossRef]
  24. Napel, S.; Mu, W.; Jardim-Perassi, B.V.; Aerts, H.; Gillies, R.J. Quantitative imaging of cancer in the postgenomic era: Radio(geno)mics, deep learning, and habitats. Cancer 2018, 124, 4633–4649. [Google Scholar] [CrossRef]
  25. Zhang, J.; Fujimoto, J.; Zhang, J.; Wedge, D.C.; Song, X.; Zhang, J.; Seth, S.; Chow, C.W.; Cao, Y.; Gumbs, C.; et al. Intratumor heterogeneity in localized lung adenocarcinomas delineated by multiregion sequencing. Science 2014, 346, 256–259. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Wu, J.; Cao, G.; Sun, X.; Lee, J.; Rubin, D.L.; Napel, S.; Kurian, A.W.; Daniel, B.L.; Li, R. Intratumoral Spatial Heterogeneity at Perfusion MR Imaging Predicts Recurrence-free Survival in Locally Advanced Breast Cancer Treated with Neoadjuvant Chemotherapy. Radiology 2018, 288, 26–35. [Google Scholar] [CrossRef]
  27. Saltz, J. Stony Brook University COVID-19 Positive Cases. Cancer Imaging Arch. 2021. [Google Scholar] [CrossRef]
  28. Clark, K.; Vendt, B.; Smith, K.; Freymann, J.; Kirby, J.; Koppel, P.; Moore, S.; Phillips, S.; Maffitt, D.; Pringle, M.; et al. The Cancer Imaging Archive (TCIA): Maintaining and operating a public information repository. J. Digit. Imaging 2013, 26, 1045–1057. [Google Scholar] [CrossRef] [Green Version]
  29. Tsai, E.B.; Simpson, S.; Lungren, M.P.; Hershman, M.; Roshkovan, L.; Colak, E.; Erickson, B.J.; Shih, G.; Stein, A.; Kalpathy-Cramer, J.; et al. Data from medical imaging data resource center (MIDRC)-RSNA international covid radiology database (RICORD) release 1C—Chest X-ray, covid+(MIDRC-RICORD-1C). Cancer Imaging Arch. 2021, 6, 13. [Google Scholar] [CrossRef]
  30. Tsai, E.B.; Simpson, S.; Lungren, M.P.; Hershman, M.; Roshkovan, L.; Colak, E.; Erickson, B.J.; Shih, G.; Stein, A.; Kalpathy-Cramer, J.; et al. The RSNA international COVID-19 open radiology database (RICORD). Radiology 2021, 299, E204–E213. [Google Scholar] [CrossRef]
  31. Hofmanninger, J.; Prayer, F.; Pan, J.; Röhrich, S.; Prosch, H.; Langs, G. Automatic lung segmentation in routine imaging is primarily a data diversity problem, not a methodology problem. Eur. Radiol. Exp. 2020, 4, 50. [Google Scholar] [CrossRef] [PubMed]
  32. Wu, J.; Gensheimer, M.F.; Dong, X.; Rubin, D.L.; Napel, S.; Diehn, M.; Loo, B.W., Jr.; Li, R. Robust intratumor partitioning to identify high-risk subregions in lung cancer: A pilot study. Int. J. Radiat. Oncol. Biol. Phys. 2016, 95, 1504–1512. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Wu, J.; Gensheimer, M.F.; Zhang, N.; Guo, M.; Liang, R.; Zhang, C.; Fischbein, N.; Pollom, E.L.; Beadle, B.; Le, Q.T. Tumor subregion evolution-based imaging features to assess early response and predict prognosis in oropharyngeal cancer. J. Nucl. Med. 2020, 61, 327–336. [Google Scholar] [CrossRef] [PubMed]
  34. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef] [Green Version]
  35. Simpson, S.; Kay, F.U.; Abbara, S.; Bhalla, S.; Chung, J.H.; Chung, M.; Henry, T.S.; Kanne, J.P.; Kligerman, S.; Ko, J.P. Radiological Society of North America expert consensus document on reporting chest CT findings related to COVID-19: Endorsed by the Society of Thoracic Radiology, the American College of Radiology, and RSNA. Radiol. Cardiothorac. Imaging 2020, 2, e200152. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Zhang, K.; Liu, X.; Shen, J.; Li, Z.; Sang, Y.; Wu, X.; Zha, Y.; Liang, W.; Wang, C.; Wang, K. Clinically applicable AI system for accurate diagnosis, quantitative measurements, and prognosis of COVID-19 pneumonia using computed tomography. Cell 2020, 181, 1423–1433.e1411. [Google Scholar] [CrossRef]
  37. Wu, Z.; McGoogan, J.M. Characteristics of and Important Lessons From the Coronavirus Disease 2019 (COVID-19) Outbreak in China: Summary of a Report of 72 314 Cases From the Chinese Center for Disease Control and Prevention. JAMA 2020, 323, 1239–1242. [Google Scholar] [CrossRef]
  38. Zhang, L.; Zhu, F.; Xie, L.; Wang, C.; Wang, J.; Chen, R.; Jia, P.; Guan, H.Q.; Peng, L.; Chen, Y. Clinical characteristics of COVID-19-infected cancer patients: A retrospective case study in three hospitals within Wuhan, China. Ann. Oncol. 2020, 31, 894–901. [Google Scholar] [CrossRef]
  39. Zhao, C.; Xu, Y.; He, Z.; Tang, J.; Zhang, Y.; Han, J.; Shi, Y.; Zhou, W. Lung segmentation and automatic detection of COVID-19 using radiomic features from chest CT images. Pattern Recognit. 2021, 119, 108071. [Google Scholar] [CrossRef]
  40. Bai, X.; Wang, H.; Ma, L.; Xu, Y.; Gan, J.; Fan, Z.; Yang, F.; Ma, K.; Yang, J.; Bai, S. Advancing COVID-19 diagnosis with privacy-preserving collaboration in artificial intelligence. Nat. Mach. Intell. 2021, 3, 1081–1089. [Google Scholar] [CrossRef]
  41. Fang, X.; Li, X.; Bian, Y.; Ji, X.; Lu, J. Radiomics nomogram for the prediction of 2019 novel coronavirus pneumonia caused by SARS-CoV-2. Eur. Radiol. 2020, 30, 6888–6901. [Google Scholar] [CrossRef] [PubMed]
  42. Wang, H.; Wang, L.; Lee, E.H.; Zheng, J.; Zhang, W.; Halabi, S.; Liu, C.; Deng, K.; Song, J.; Yeom, K.W. Decoding COVID-19 pneumonia: Comparison of deep learning and radiomics CT image signatures. Eur. J. Nucl. Med. Mol. Imaging 2021, 48, 1478–1486. [Google Scholar] [CrossRef] [PubMed]
  43. Homayounieh, F.; Ebrahimian, S.; Babaei, R.; Mobin, H.K.; Zhang, E.; Bizzo, B.C.; Mohseni, I.; Digumarthy, S.R.; Kalra, M.K. CT radiomics, radiologists, and clinical information in predicting outcome of patients with COVID-19 pneumonia. Radiol. Cardiothorac. Imaging 2020, 2, e200322. [Google Scholar] [CrossRef] [PubMed]
  44. Saygılı, A. A new approach for computer-aided detection of coronavirus (COVID-19) from CT and X-ray images using machine learning methods. Appl. Soft Comput. 2021, 105, 107323. [Google Scholar] [CrossRef] [PubMed]
  45. Jin, C.; Chen, W.; Cao, Y.; Xu, Z.; Tan, Z.; Zhang, X.; Deng, L.; Zheng, C.; Zhou, J.; Shi, H. Development and evaluation of an artificial intelligence system for COVID-19 diagnosis. Nat. Commun. 2020, 11, 5088. [Google Scholar] [CrossRef]
  46. Aminu, M.; Ahmad, N.A.; Noor, M.H.M. Covid-19 detection via deep neural network and occlusion sensitivity maps. Alex. Eng. J. 2021, 60, 4829–4855. [Google Scholar] [CrossRef]
  47. Song, Y.; Zheng, S.; Li, L.; Zhang, X.; Zhang, X.; Huang, Z.; Chen, J.; Wang, R.; Zhao, H.; Chong, Y. Deep learning enables accurate diagnosis of novel coronavirus (COVID-19) with CT images. IEEE/ACM Trans. Comput. Biol. Bioinform. 2021, 18, 2775–2780. [Google Scholar] [CrossRef]
  48. Zhang, N.; Liang, R.; Gensheimer, M.F.; Guo, M.; Zhu, H.; Yu, J.; Diehn, M.; Loo, B.W., Jr.; Li, R.; Wu, J. Early response evaluation using primary tumor and nodal imaging features to predict progression-free survival of locally advanced non-small cell lung cancer. Theranostics 2020, 10, 11707–11718. [Google Scholar] [CrossRef]
  49. Noor, A.; Muhammad, A.; Mohd, N.M.H. Covid-19 detection via deep neural network and occlusion sensitivity maps. TechRxiv 2021. [Google Scholar] [CrossRef]
Figure 1. Workflow of the proposed approach.
Figure 1. Workflow of the proposed approach.
Cancers 15 00275 g001
Figure 2. Habitat subregions distributions; SR: subregion. Row (A) is a boxplot showing the detailed distribution of the four input image channels across the six habitat regions. Row (B) represent the imaging interpretations of the six habitat regions.
Figure 2. Habitat subregions distributions; SR: subregion. Row (A) is a boxplot showing the detailed distribution of the four input image channels across the six habitat regions. Row (B) represent the imaging interpretations of the six habitat regions.
Cancers 15 00275 g002
Figure 3. Examples of habitat maps of 5 cases from the Stony Brook dataset.
Figure 3. Examples of habitat maps of 5 cases from the Stony Brook dataset.
Cancers 15 00275 g003
Figure 4. Two-dimensional embeddings of general population and cancer population datasets after habitat analysis.
Figure 4. Two-dimensional embeddings of general population and cancer population datasets after habitat analysis.
Cancers 15 00275 g004
Figure 5. Performance comparison of the different diagnostic models based on the multiregional spatial interaction features. Row (A,B) show the performances of the different models on the general and cancer cohorts, respectively. The “Accuracy curves” and “AUC curves” shows the performance (classification accuracies and area under the curve) of the different classification models with respect to different number of selected features. The confusion matrices show the performance of the best models (in terms of the best selected features) and “ROC curves” shows the corresponding receiver operating characteristics curves of the best models.
Figure 5. Performance comparison of the different diagnostic models based on the multiregional spatial interaction features. Row (A,B) show the performances of the different models on the general and cancer cohorts, respectively. The “Accuracy curves” and “AUC curves” shows the performance (classification accuracies and area under the curve) of the different classification models with respect to different number of selected features. The confusion matrices show the performance of the best models (in terms of the best selected features) and “ROC curves” shows the corresponding receiver operating characteristics curves of the best models.
Cancers 15 00275 g005
Figure 6. Comparison of the different models for ventilation prediction. Row (A,B) show the performances of the different models on the general and cancer cohorts, respectively. The “Accuracy curves” and “AUC curves” shows the performance (classification accuracies and area under the curve) of the different classification models with respect to different number of selected features. The confusion matrices show the performance of the best models (in terms of the best selected features) and “ROC curves” shows the corresponding receiver operating characteristics curves of the best models.
Figure 6. Comparison of the different models for ventilation prediction. Row (A,B) show the performances of the different models on the general and cancer cohorts, respectively. The “Accuracy curves” and “AUC curves” shows the performance (classification accuracies and area under the curve) of the different classification models with respect to different number of selected features. The confusion matrices show the performance of the best models (in terms of the best selected features) and “ROC curves” shows the corresponding receiver operating characteristics curves of the best models.
Cancers 15 00275 g006
Table 2. Multiregional spatial interaction (MSI) features interpretation.
Table 2. Multiregional spatial interaction (MSI) features interpretation.
Feature NameFeature Description
MSI 1–MSI 4 2nd order statistics features
MSI 5–MSI 10absolute habitat subregions volume (SR1–SR6)
MSI 11–MSI 16interaction (absolute) between habitat subregions and border
MSI 17–MSI 21interaction (absolute) between SR1 and the remaining subregions, i.e., MSI 17 = SR1 SR2, MSI 18 = SR1 SR3, …, MSI 21 = SR1 SR6.
MSI 22–MSI 25interaction (absolute) between SR2 and SR3, SR4, SR5 and SR6, i.e., MSI 22 = SR2 SR3, MSI 23 = SR2 SR4, MSI 24 = SR2 SR5, MSI 25 = SR2 SR6.
MSI 26–MSI 28interaction (absolute) between SR3 and SR4, SR5 and SR6, i.e., MSI 26 = SR3 SR4, MSI 27 = SR3 SR5, MSI 28 = SR3 SR6.
MSI 29–MSI 30interaction (absolute) between SR4 and SR5 and SR6, i.e., MSI 29 = SR4 SR5, MSI 30 = SR4 SR6.
MSI 31interaction (absolute) between SR5 and SR6, i.e., MSI 30 = SR5 SR6.
MSI 32–MSI 37percentage of habitat subregions volume (SR1–SR6)
MSI 38–MSI 43normalized interaction (percentage) between habitat subregions and border
MSI 44–MSI 48normalized interaction (percentage) between SR1 and the remaining subregions, i.e., MSI 44 = SR1 SR2, MSI 45 = SR1 SR3, …, MSI 48 = SR1 SR6.
MSI 49–MSI 52normalized interaction (percentage) between SR2 and SR3, SR4, SR5 and SR6, i.e., MSI 49 = SR2 SR3, MSI 50 = SR2 SR4, MSI 51 = SR2 SR5, MSI 52 = SR2 SR6.
MSI 53–MSI 55normalized interaction (percentage) between SR3 and SR4, SR5 and SR6, i.e., MSI 53 = SR3 SR4, MSI 54 = SR3 SR5, MSI 55 = SR3 SR6.
MSI 56–MSI 57normalized interaction (percentage) between SR4 and SR5 and SR6, i.e., MSI 56 = SR4 SR5, MSI 57 = SR4 SR6.
MSI 58normalized interaction (percentage) between SR5 and SR6, i.e., MSI 58 = SR5 SR6.
Table 3. Performance comparison of the different models for COVID-19 diagnosis on the general and cancer cohorts. Acc: accuracy; Sen: sensitivity; Spe: specificity; AUC: area under the receiver operating characteristic curve. LR: logistic regression, RF: random forest, SVM: support vector machine, GAM: generalized additive model.
Table 3. Performance comparison of the different models for COVID-19 diagnosis on the general and cancer cohorts. Acc: accuracy; Sen: sensitivity; Spe: specificity; AUC: area under the receiver operating characteristic curve. LR: logistic regression, RF: random forest, SVM: support vector machine, GAM: generalized additive model.
MethodsCohort
GeneralCancer
AccSenSpeAUCAccSenSpeAUC
LR0.99441.00000.99280.99991.00001.00001.00001.0000
RF0.99440.99140.99520.99980.99520.98781.00000.9998
SVM0.99060.99120.99040.99980.99761.00000.99611.0000
GAM0.99251.0000.99040.99970.99520.99380.99611.0000
Table 4. Performance of the COVID-19 diagnostic models trained on the general cohort and applied on the cancer cohort.
Table 4. Performance of the COVID-19 diagnostic models trained on the general cohort and applied on the cancer cohort.
MethodsAll
AccSenSpeAUC
LR0.87380.99100.83170.9807
RF0.85481.00000.80880.9178
SVM0.85481.00000.80880.9586
GAM0.90000.99180.86240.9202
Table 5. Performance comparison of the different models for ventilation prediction.
Table 5. Performance comparison of the different models for ventilation prediction.
MethodsCohort
GeneralCancer
AccSenSpeAUCAccSenSpeAUC
LR0.96160.97960.94570.99610.99121.00000.98211.0000
RF0.91611.00000.85890.99490.95801.00000.92051.0000
SVM0.99281.00000.98611.00000.97571.00000.95241.0000
GAM0.91611.00000.85890.99670.98671.00000.97351.0000
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Aminu, M.; Yadav, D.; Hong, L.; Young, E.; Edelkamp, P., Jr.; Saad, M.; Salehjahromi, M.; Chen, P.; Sujit, S.J.; Chen, M.M.; et al. Habitat Imaging Biomarkers for Diagnosis and Prognosis in Cancer Patients Infected with COVID-19. Cancers 2023, 15, 275. https://doi.org/10.3390/cancers15010275

AMA Style

Aminu M, Yadav D, Hong L, Young E, Edelkamp P Jr., Saad M, Salehjahromi M, Chen P, Sujit SJ, Chen MM, et al. Habitat Imaging Biomarkers for Diagnosis and Prognosis in Cancer Patients Infected with COVID-19. Cancers. 2023; 15(1):275. https://doi.org/10.3390/cancers15010275

Chicago/Turabian Style

Aminu, Muhammad, Divya Yadav, Lingzhi Hong, Elliana Young, Paul Edelkamp, Jr., Maliazurina Saad, Morteza Salehjahromi, Pingjun Chen, Sheeba J. Sujit, Melissa M. Chen, and et al. 2023. "Habitat Imaging Biomarkers for Diagnosis and Prognosis in Cancer Patients Infected with COVID-19" Cancers 15, no. 1: 275. https://doi.org/10.3390/cancers15010275

APA Style

Aminu, M., Yadav, D., Hong, L., Young, E., Edelkamp, P., Jr., Saad, M., Salehjahromi, M., Chen, P., Sujit, S. J., Chen, M. M., Sabloff, B., Gladish, G., de Groot, P. M., Godoy, M. C. B., Cascone, T., Vokes, N. I., Zhang, J., Brock, K. K., Daver, N., ... Wu, J. (2023). Habitat Imaging Biomarkers for Diagnosis and Prognosis in Cancer Patients Infected with COVID-19. Cancers, 15(1), 275. https://doi.org/10.3390/cancers15010275

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop