Next Article in Journal
Percutaneous Superimposed O-Arm-MRI-Navigated Biopsy for Spinal Column Pathologies
Previous Article in Journal
Unravelling Peritoneal Carcinomatosis Using Cross-Sectional Imaging Modalities
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multimodal Spatiotemporal Deep Learning Framework to Predict Response of Breast Cancer to Neoadjuvant Systemic Therapy

by
Monu Verma
1,
Leila Abdelrahman
2,
Fernando Collado-Mesa
3 and
Mohamed Abdel-Mottaleb
1,*
1
Department of Electrical and Computer Engineering, University of Miami, Miami, FL 33146, USA
2
MIT Media Lab, Cambridge, MA 02139-4307, USA
3
Department of Radiology, Miller School of Medicine, University of Miami, Miami, FL 33146, USA
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(13), 2251; https://doi.org/10.3390/diagnostics13132251
Submission received: 4 April 2023 / Revised: 20 June 2023 / Accepted: 27 June 2023 / Published: 3 July 2023
(This article belongs to the Section Medical Imaging and Theranostics)

Abstract

:
Current approaches to breast cancer therapy include neoadjuvant systemic therapy (NST). The efficacy of NST is measured by pathologic complete response (pCR). A patient who attains pCR has significantly enhanced disease-free survival progress. The accurate prediction of pCR in response to a given treatment regimen could increase the likelihood of achieving pCR and prevent toxicities caused by treatments that are not effective. Th early prediction of response to NST can increase the likelihood of survival and help with decisions regarding breast-conserving surgery. An automated NST prediction framework that is able to precisely predict which patient undergoing NST will achieve a pathological complete response (pCR) at an early stage of treatment is needed. Here, we propose an end-to-end efficient multimodal spatiotemporal deep learning framework (deep-NST) framework to predict the outcome of NST prior or at an early stage of treatment. The deep-NST model incorporates imaging data captured at different timestamps of NST regimens, a tumor’s molecular data, and a patient’s demographic data. The efficacy of the proposed work is validated on the publicly available ISPY-1 dataset, in terms of accuracy, area under the curve (AUC), and computational complexity. In addition, seven ablation experiments were carried out to evaluate the impact of each design module in the proposed work. The experimental results show that the proposed framework performs significantly better than other recent methods.

1. Introduction

In the U.S., about 1 in 8 women (13%) are expected to develop invasive breast cancer during their life. Breast cancer is the most diagnosed cancer among U.S. women, accounting for an estimated 30% of newly diagnosed cancers, and breast cancer ranks second amongst the leading causes of cancer-related death in women in the U.S. [1]. Current approaches to breast cancer therapy include neoadjuvant systemic therapy (NST), which has several potential advantages including avoiding mastectomy by reduction in tumor size and downstaging the axilla, which may obviate the need for axillary lymph node dissection and its potential complications. In addition, NST also permits the in vivo assessment of drug efficacy with the possibility to opt for a different treatment approach if the tumor is not responding. The efficacy of NST is measured by pathologic complete response (pCR). A patient who attains pCR has significantly enhanced disease-free progress survival. The accurate prediction of which patients will achieve pCR in response to a given treatment regimen could increase the likelihood of achieving pCR and prevent toxicities caused by treatments that are not effective [2].
The response to NST is traditionally assessed by a combination of physical exams and dedicated breast imaging exams, most commonly dynamic contrast enhance magnetic resonance imaging (DCE-MRI). Radiologists typically assess this response by measuring changes across a limited handful of clinical imaging parameters on DCE-MRI over the course of treatment, such as a tumor’s largest diameter (LD) or its pattern of contrast agent uptake. More recently, hand-crafted radiomic imaging features [3,4,5,6] have been shown to predict pCR from pre-treatment DCE-MRI data by characterizing the texture of the tumor and micro-environment. Nonetheless, radiomics assessment is limited by the need to pre-define meaningful and predictive hand-crafted features to assess the attributes such as lesion shape and image texture, and radio mic models are constrained by the discriminability of a finite and pre-defined pool of features.
Recent advancements in artificial intelligence (AI), especially deep learning (DL) [7,8,9], have shown a significant improvement compared to the conventional approaches [10,11] in the prediction of pathological results through radiological data. DL has become the most effective technique for various applications such as prediction, classification, object detection, and image segmentation. DL makes it possible to automatically extract features from MRI exams instead of engineered features, and it achieves impressive performance for the prediction of the response to NST [7,8]. Data obtained from tumor samples through biopsies can also shed light on cellular biomarkers such as HER2, ER, and PgR. These biomarkers can help guide clinicians’ choices on which NST to prescribe. For example, a patient with ER-positive cancer may respond well to Tamoxifen, a hormone therapy that targets ER-positive cells. On the other hand, if a patient is ER-negative, they will fail to respond to this drug and may require chemotherapy. Combining information from different modalities, e.g., MRI exams and tumor pathological features can help with the development of robust machine/DL algorithms to help predict the response to NST, as shown by some researchers. For example, Ravichandran et al. [7] introduced a deep learning-based unique patch-based response prediction approach that allows for the visualization of specific spatial regions and image patterns associated with poor or favorable responses to NST. They introduced a CNN model to learn the features from the pre- and post-contrast pre-treatment imaging and yielded an AUC of 0.77. In addition, they also validated that the incorporation of the HER2 biomarker features with MRI could yield an AUC of 0.85. Duanmu et al. [12] further showed that convolving the clinical features with the imaging features learned in the CNN instead of concatenating them could yield an accuracy of 0.83 and an AUC of 0.80. However, these studies considered only pre-NST MRI data along with clinical reports. Therefore, these approaches neglected the structural and functional tumor changes on DCI-MRI over the course of NST. The structural and functional changes in the tumor microenvironment at different stages of the NST regimen can better reflect the therapeutic response [13,14].
We propose a multimodal spatiotemporal deep learning framework by incorporating multimodal information for predicting the response of breast cancer to neoadjuvant therapy. The proposed work has the following key contributions:
  • We develop a multimodal spatiotemporal deep learning by integrating the following multi-modalities: imaging data with N-time stamps (pre-treatment, early treatment, inter-regimen, prior to surgery, etc.), molecular data (ER, PgRPos, HRPos, BilateralCa, Laterality, HER2Pos, HR_HER2_Category, and HR_HER2_Status), and demographical data (age and race). We demonstrate the influence of each time point on the predictions made by the network through ablation experiments.
  • We design a novel 3D-CNN-based deep learning framework by introducing a cross-kernel feature fusion (CKFF) module.
  • The CKFF module makes the architecture more learnable at a lower computational cost by paying attention to multiple receptive fields to extract the spatiotemporal features.
  • The efficacy of the proposed framework is tested on a challenging breast cancer data set, ISPY-1 [15], in terms of accuracy and AUC.

2. Methodology

In computer vision applications, the conventional CNN models VGG 16-Net [16], VGG19-Net [16], and ResNet [17] have demonstrated impressive results. On the other hand, a deep, dense network prevents these models from preserving important aspects of breast cancer MRI scans. Progressive convolution and pooling operations may cause the deep, dense CNN architectures to overlook the cancerous regions’ micro-level features [18]. In addition, large data samples are required for deep dense CNN architectures to learn important features. Profound CNN networks fail to learn appropriate features over smaller datasets and endure overfitting [19,20]. Only a few samples in the benchmark datasets of NST are available. Recently, Qu et al. [14] and Ravichandran et al. [7] proposed a deep learning-based solution for NST prediction and overcame the issue of limited data samples through data augmentation. Additionally, before supplying images to the network, they used segmentation. Huynh et al. [21] also found a solution to the problem of overfitting by combinining the LDA classifier with the transfer learning capabilities of VGGNet. However, its two-stage framework makes the presented strategy challenging to implement in real-time applications. We proposed an end-to-end multimodal spatitemporal deep learning framework to extract spatiotemporal features from the MRI scans to predict the pCR response of neoadjuvant treatment.

Proposed Method

The proposed multimodal spatiotemporal deep learning framework comprises four parallel 3D-CNN networks along with clinical features, as shown in Figure 1. Specifically, the 3D-CNN network is introduced to learn spatial and temporal features from the MRI scans at a particular time-stamp. In addition, parallel 3D-CNN networks are used to learn the structural and functional changes in the tumor microenvironment at four-time stamps T 1 , T 2 , T 3 and T 4 , as shown in Figure 1. Further, clinical reports’ features, including molecular and demographical data, are processed to aid imaging features to enhance the generalization of the deep-NST framework. Moreover, we used the LeakyReLU as an activation function over the resultant features of each convolution layer.
The proposed deep-NST framework is trained by using two stages to predict the outcome of the NST, as shown in Figure 1. The proposed deep-NST framework is initially trained using MRI scans captured at four-time stamps ( T 1 , T 2 , T 3 , and T 4 ) and clinical data. Further, knowledge of the 1st 3D-CNN network associated with early-stage MRI scans ( T 1 ) is finetuned again with T 1 MRI scans and clinical features. The motivation behind developing this framework was to utilize the multimodal spatiotemporal features of different time stamps along with clinical data to train deep learning methods. However, the final prediction model only needed pre-NST MRI scans with clinical data and ensured early-stage prediction for NST in breast cancer.

3. Experimental Results and Analysis

In this section, first we present the dataset and the implementation details. Further, the experimental setups and experimental results are discussed. In addition, we explore the importance of each module of the proposed framework in the ablation study. Finally, we compare the computational complexity of the proposed model with state-of-the-art approaches.

3.1. Dataset

To validate the effectiveness of the proposed deep-NST framework, we work with the ISPY-1 [15] public dataset’s dynamic contrast-enhanced (DCE)-MRI images and non-imaging clinical report information. The ISPY-1 dataset comprises a cohort of 207 patients, out of which only 121 had MRI scans at all four-time points (pre-treatment, early treatment, inter-regimen, and prior to surgery). The original dataset exhibited a significant disparity in class distribution, with the majority of samples belonging to the non-responded (pCR0) class, as shown in Figure 2. We noticed instances where certain patients underwent multiple MRI scans at the same timestamp. To address the class imbalance and achieve a balanced distribution for both classes, i.e., non-responded (pCR0) and responded (pCR1) to NST, we included multiple MRI scans captured at the same timestamp for the responding class as shown in Figure 2. However, for the non-responding class, we considered only a single MRI scan. Ultimately, we collected a total of 148 samples from 84 patients in the non-responding class and 105 samples from 37 patients in the responding class for each timestamp. Additionally, we utilized ten non-imaging clinical features (age, race, ERPos, HRPos, PgRpos, Her2MostPos, HR_HER2_CATEGORY, HR_HER2_STATUS, BilaterCa, and Laterality) from the clinical data dictionary of the ISPY-1 trial. In this study, we focused on 121 patients from the ISPY-1 trial, splitting them into a training set consisting of 84 patients with 182 data samples, and a testing set comprising 37 patients with 71 data samples.

3.2. Training and Implementation Details

We used two stages to train the proposed deep-NST framework: training with four-time stamps NST MRI scans and clinical data and fine-tuning with only pre-NST MRI scans and clinical data, as shown in Figure 1. The first stage is trained for 300 epochs, whereas the second stage is finetuned by using 50 epochs. Moreover, final testing is done end-to-end by utilizing pre-treatment NST MRI scans and clinical data captured at the initial stage of the treatment to ensure the early-stage prediction for NST in breast cancer.
The publicly available ISPY-1 breast cancer dataset has MRI scans with different durations. To overcome this issue, we use the temporal interpolation model (TIM) [22] to normalize the length of the scan sequences to 60. The image sequences are normalized to 112 × 112 × 60 before using a spatiotemporal CNN model. To ensure a fair comparison of CNN-based networks, we implemented the conventional deep learning models: VGGNet [16], and ResNet with 3D-CNN [17], and trained them over our experimental settings. Moreover, we tested the effects of cross-entropy loss function over focal loss [23] by evaluating the results of the proposed deep-NST with focal loss and named deep-NST+Focal. All implementations use python 3.6 with Keras 2.3.1 and Tensorflow 2.1.0. We utilized the SGD optimizer for training the models with a 0.001 learning rate, 0.9 momentum, and 0.01 weight decay. The cross-entropy loss function is used for network optimization.

3.3. Experimental Results Analysis

This section presents the experimental results of ISPY-1. We used the prediction accuracy and AUC-ROC performance measures to evaluate the proposed deep-NST with state-of-the-art NST approaches. The AUC-ROC curve represents the degree or measure of separability. The AUC-ROC measures the capability of the model to distinguish between classes by analyzing the true positive response (TPR) against the false-positive response (FPR). The TPR and FPR are calculated by using Equations (1) and (2).
T P R = T r u e P o s i t i v e T o t a l n o . o f d a t a s a m p l e s
F P R = F l a s e P o s i t i v e T o t a l n o . o f d a t a s a m p l e s
Moreover, the prediction accuracy is calculated by the following Equation (3).
A C C = T o t a l n o . o f c o r r e c t l y p r e d i c t e d s a m p l e s T o t a l n o . o f d a t a s a m p l e s
The evaluation results for proposed and conventional CNN models are reported in Table 1. From Table 1, it is validated that the proposed deep-NST framework gains significantly better performance than traditional as well as the current deep learning models for NST. Particularly, the proposed deep-NST achieves 0.20% and 0.38% more AUC than the 3D-VGGNet and 3D-ResNet CNN models, respectively. More detailed results in terms of ROC for 3D-VGGNet, 3D-ResNet, and proposed deep-NST are illustrated in Figure 3. In addition, for detailed class generalization, we have calculated the confusion matrices as shown in Figure 4. From Figure 4b, it is clear that the 3D-ResNet is under-fitted compared to the data and is not suitable for the NST due to limited data samples. In, addition, when we observed the confusion matrices for the existing 3D-VggNet (Figure 4a) and proposed deep-NST (Figure 4b), we saw a skew towards high accuracy in predicting the non-responder patients over responder patients. This may be due to the original skewed dataset distribution. In addition to accuracy and AUC, we also evaluated sensitivity (0.9024) and specificity (0.7666). Sensitivity tells us the proportion of true positives that the model correctly identified. It gives us an idea of how well the model detects the condition when it is actually present. On the other hand, specificity measures the proportion of true negatives that the model correctly identified. It helps us understand how well the model identifies the absence of the condition.

3.3.1. Discussion

In contrast to other methods, we highlight a reduced reliance on feature engineering and manual segmentation (lesion segmentation) during image preprocessing to yield comparable results to current state-of-the-art methods. For example, authors [4,5,6] rely on precalculating the tumor volume or functional tumor volume as a feature before being input into a machine learning classifier. This extra preprocessing step is costly. We rely only on the raw DCE-MRI images and let the network learn the relevant image features through back-propagation. Even compared to existing end-to-end DL methods [7,12,16,17], AUC (0.88) outperforms the prior results. Moreover, our approach uses imaging data from all four-time points in a patient’s NST regimens, which allows the proposed deep-NST to outperform or match the current DL state-of-the-art methods [7,12,24] that rely on data from the only one-time point.

3.3.2. Ablation Study

We have conducted seven ablation experiments to evaluate the importance of each component step-by-step in the proposed deep-NST. First, the CKFF module’s effect is tested by comparing it to the basic 3D convolution layer. The proposed CKFF module is replaced by the 3D convolution of size 3 × 3 × 3 and named deep-NST+3DCNN. The second study has been conducted to examine the CKFF module’s impact over parallel multi-scale convolutional layers and is named deep-NST+3DInception. The other study has been conducted to test the effect of multimodal information on clinical data and MRI data. For this study, we used the information from the MRI imaging and called it a Unimodal ST. The results of Unimodal ST are computed on imaging data only. The comparative results for proposed deep-NST and ablation studies are tabulated in Table 1 and Table 2. The results show that the cross-entropy loss function, the proposed CKFF module, and multimodal data play an important role in the deep-NST framework. Specifically, the proposed framework has 0.23%, 0.23%, and 0.01%, higher AUC than deep-NST+3DCNN, deep-NST+3DInception, and Unimodal ST. Similarly, the proposed framework gain 0.19%, 0.19%, and 0.02% higher accuracy over deep-NST+3DCNN, deep-NST+3DInception, and Unimodal ST.
Moreover, we also investigated the effect of each timestamp MRI scan with the clinical information. The experimental results are tabulated in Table 2. The results show that combining all time stamps’ MRI scans with clinical information allows the network to learn the pertinent features of the breast cancer image sequences and achieve better performance in predicting the outcome of NST. More detailed results in terms of ROC and class-wise confusion matrix are illustrated in Figure 5 and Figure 6, respectively. This figure shows an ablation study by including the time instances during the neoadjuvant regimen (inter-regimen), gradually increasing the model’s performance.

3.3.3. Complexity Analysis

This section represents the comparative study between the traditional 3D-CNN, proposed deep-NST, and its variants: deep-NST+3DCNN, deep-NST+3dInception, and Unimodal ST (spatiotemporal) frameworks in terms of computational complexity. The total number of parameters engaged in each network is presented in Table 3. The proposed deep-NST has significantly fewer learnable parameters; 4 million (M) fewer than other existing NST frameworks such as 3D-VGGNet; 225M, and 3D-ResNet; 311M. Furthermore, deep-NST requires only 37 megabytes (MB) of memory storage, which is minor compared to 3D-VGGNet; 1.8 gigabytes (GB) and 3D-ResNet; 2.5 GB. Additionally, the total number of floating-point operations is much less for the proposed deep-NST than for the 3D-VGGNet and 3D-ResNet models. Moreover, the proposed deep-NST requires 17M fewer parameters than the deep-NST with a parallel multi-scale inception module. Similarly, the proposed deep-NST needs 1363 MB less memory than the deep-NST with a parallel multi-scale inception module.
This validates that the proposed CKFF module is cost-effective. Based on the experimental results and the computational complexity reported in Table 1 and Table 3, respectively, we can conclude that the proposed deep-NST is the most effective and efficient DL-based model for NST outcome prediction.

4. Conclusions

We present a compact and lightweight multimodal spatiotemporal deep learning framework to predict breast cancer response to neoadjuvant therapy by incorporating patients’ MRI, tumor molecular, and demographics features. MRI exams were obtained at different stages of treatment. The proposed framework performs better than state-of-the-art NST approaches. In addition, based on the conducted computational cost analysis, our framework is cost-effective based on the number of parameters and FLOPS compared to other models.

Author Contributions

Conceptualization, M.V. and M.A.-M.; Methodology, M.V.; Software, M.V.; Validation, M.V., L.A., F.C.-M. and M.A.-M.; Investigation, F.C.-M. and M.A.-M.; Resources, F.C.-M. and M.A.-M.; Data curation, F.C.-M. and M.A.-M.; Writing—original draft, M.V.; Writing—review & editing, L.A., F.C.-M. and M.A.-M.; Visualization, F.C.-M. and M.A.-M.; Supervision, F.C.-M. and M.A.-M.; Project administration, F.C.-M. and M.A.-M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by a pilot award from the NCI-supported Sylvester Comprehensive Cancer Center (P30CA240139).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is unavailable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tao, Z.; Shi, A.; Lu, C.; Song, T.; Zhang, Z.; Zhao, J. Breast cancer: Epidemiology and etiology. Cell Biochem. Biophys. 2015, 72, 333–338. [Google Scholar] [CrossRef] [PubMed]
  2. Yankeelov, T.E.; Atuegwu, N.; Hormuth, D.; Weis, J.A.; Barnes, S.L.; Miga, M.I.; Rericha, E.C.; Quaranta, V. Clinically relevant modeling of tumor growth and treatment response. Sci. Transl. Med. 2013, 5, 187ps9. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Braman, N.M.; Etesami, M.; Prasanna, P.; Dubchuk, C.; Gilmore, H.; Tiwari, P.; Plecha, D.; Madabhushi, A. Intratumoral and peritumoral radiomics for the pretreatment prediction of pathological complete response to neoadjuvant chemotherapy based on breast DCE-MRI. Breast Cancer Res. 2017, 19, 1–14. [Google Scholar]
  4. Hylton, N.M.; Blume, J.D.; Bernreuter, W.K.; Pisano, E.D.; Rosen, M.A.; Morris, E.A.; Weatherall, P.T.; Lehman, C.D.; Newstead, G.M.; Polin, S.; et al. Locally advanced breast cancer: MR imaging for prediction of response to neoadjuvant chemotherapy—Results from ACRIN 6657/I-SPY TRIAL. Radiology 2012, 263, 663–672. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Hylton, N.M.; Gatsonis, C.A.; Rosen, M.A.; Lehman, C.D.; Newitt, D.C.; Partridge, S.C.; Bernreuter, W.K.; Pisano, E.D.; Morris, E.A.; Weatherall, P.T.; et al. Neoadjuvant chemotherapy for breast cancer: Functional tumor volume by MR imaging predicts recurrence-free survival—Results from the ACRIN 6657/CALGB 150007 I-SPY 1 TRIAL. Radiology 2016, 279, 44–55. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Li, W.; Arasu, V.; Newitt, D.C.; Jones, E.F.; Wilmes, L.; Gibbs, J.; Kornak, J.; Joe, B.N.; Esserman, L.J.; Hylton, N.M. Effect of MR imaging contrast thresholds on prediction of neoadjuvant chemotherapy response in breast cancer subtypes: A subgroup analysis of the ACRIN 6657/I-SPY 1 TRIAL. Tomography 2016, 2, 378–387. [Google Scholar] [CrossRef] [PubMed]
  7. Ravichandran, K.; Braman, N.; Janowczyk, A.; Madabhushi, A. A deep learning classifier for prediction of pathological complete response to neoadjuvant chemotherapy from baseline breast DCE-MRI. In Medical Imaging 2018: Computer-Aided Diagnosis; SPIE: Bellingham, WA, USA, 2018; Volume 10575, pp. 79–88. [Google Scholar]
  8. Ha, R.; Chin, C.; Karcich, J.; Liu, M.Z.; Chang, P.; Mutasa, S.; Pascual Van Sant, E.; Wynn, R.T.; Connolly, E.; Jambawalikar, S. Prior to initiation of chemotherapy, can we predict breast tumor response? Deep learning convolutional neural networks approach using a breast MRI tumor dataset. J. Digit. Imaging 2019, 32, 693–701. [Google Scholar] [CrossRef] [PubMed]
  9. Braman, N.; Adoui, M.E.; Vulchi, M.; Turk, P.; Etesami, M.; Fu, P.; Bera, K.; Drisis, S.; Varadan, V.; Plecha, D.; et al. Deep learning-based prediction of response to HER2-targeted neoadjuvant chemotherapy from pre-treatment dynamic breast MRI: A multi-institutional validation study. arXiv 2020, arXiv:2001.08570. [Google Scholar]
  10. Marinovich, M.L.; Macaskill, P.; Irwig, L.; Sardanelli, F.; Mamounas, E.; von Minckwitz, G.; Guarneri, V.; Partridge, S.C.; Wright, F.C.; Choi, J.H.; et al. Agreement between MRI and pathologic breast tumor size after neoadjuvant chemotherapy, and comparison with alternative tests: Individual patient data meta-analysis. BMC Cancer 2015, 15, 662. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Lobbes, M.; Prevos, R.; Smidt, M.; Tjan-Heijnen, V.; Van Goethem, M.; Schipper, R.; Beets-Tan, R.; Wildberger, J. The role of magnetic resonance imaging in assessing residual disease and pathologic complete response in breast cancer patients receiving neoadjuvant chemotherapy: A systematic review. Insights Imaging 2013, 4, 163–175. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Duanmu, H.; Huang, P.B.; Brahmavar, S.; Lin, S.; Ren, T.; Kong, J.; Wang, F.; Duong, T.Q. Prediction of pathological complete response to neoadjuvant chemotherapy in breast cancer using deep learning with integrative imaging, molecular and demographic data. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2020, Proceedings of the 23rd International Conference, Lima, Peru, 4–8 October 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 242–252. [Google Scholar]
  13. Bouzón, A.; Acea, B.; Soler, R.; Iglesias, Á.; Santiago, P.; Mosquera, J.; Calvo, L.; Seoane-Pillado, T.; García, A. Diagnostic accuracy of MRI to evaluate tumour response and residual tumour size after neoadjuvant chemotherapy in breast cancer patients. Radiol. Oncol. 2016, 50, 73–79. [Google Scholar] [CrossRef] [PubMed]
  14. Qu, Y.H.; Zhu, H.T.; Cao, K.; Li, X.T.; Ye, M.; Sun, Y.S. Prediction of pathological complete response to neoadjuvant chemotherapy in breast cancer using a deep learning (DL) method. Thorac. Cancer 2020, 11, 651–658. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Newitt, D.; Hylton, N. Multi-center breast DCE-MRI data and segmentations from patients in the I-SPY 1/ACRIN 6657 trials. Cancer Imaging Arch. 2016, 10. [Google Scholar]
  16. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 84–90. [Google Scholar] [CrossRef] [Green Version]
  17. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  18. Verma, M.; Vipparthi, S.K.; Singh, G.; Murala, S. LEARNet: Dynamic imaging network for micro expression recognition. IEEE Trans. Image Process. 2019, 29, 1618–1627. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Schindler, A.; Lidy, T.; Rauber, A. Comparing Shallow versus Deep Neural Network Architectures for Automatic Music Genre Classification. In Proceedings of the FMT, Pölten, Austria, 23–24 November 2016; pp. 17–21. [Google Scholar]
  20. Verma, M.; Vipparthi, S.K.; Singh, G. Hinet: Hybrid inherited feature learning network for facial expression recognition. IEEE Lett. Comput. Soc. 2019, 2, 36–39. [Google Scholar] [CrossRef]
  21. Huynh, B.Q.; Antropova, N.; Giger, M.L. Comparison of breast DCE-MRI contrast time points for predicting response to neoadjuvant chemotherapy using deep convolutional neural network features with transfer learning. In Medical Imaging 2017: Computer-Aided Diagnosis; SPIE: Bellingham, WA, USA, 2017; Volume 10134, pp. 207–213. [Google Scholar]
  22. Verma, M.; Reddy, M.S.K.; Meedimale, Y.R.; Mandal, M.; Vipparthi, S.K. Automer: Spatiotemporal neural architecture search for microexpression recognition. IEEE Transactions on Neural Networks and Learning Systems 2021, 33, 6116–6128. [Google Scholar] [CrossRef] [PubMed]
  23. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  24. Joo, S.; Ko, E.S.; Kwon, S.; Jeon, E.; Jung, H.; Kim, J.Y.; Chung, M.J.; Im, Y.H. Multimodal deep learning models for the prediction of pathologic response to neoadjuvant chemotherapy in breast cancer. Sci. Rep. 2021, 11, 18800. [Google Scholar] [CrossRef] [PubMed]
  25. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  26. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2818–2826. [Google Scholar]
Figure 1. The proposed deep-NST prediction framework architecture. Image sequences with different time stamps (represented by T 1 , T 2 , T 3 , and T 4 ) are processed through parallel 3D-CNN networks. The feature outputs from all time stamps and clinical reports are added and processed by a final soft-max layer. Further trained weights of the first 3D-CNN network are fine-tuned over image sequences captured at the early stage of the treatment ( T 1 ) along with the clinical features. The numbers in the image denote the size of the feature dimension at each layer.
Figure 1. The proposed deep-NST prediction framework architecture. Image sequences with different time stamps (represented by T 1 , T 2 , T 3 , and T 4 ) are processed through parallel 3D-CNN networks. The feature outputs from all time stamps and clinical reports are added and processed by a final soft-max layer. Further trained weights of the first 3D-CNN network are fine-tuned over image sequences captured at the early stage of the treatment ( T 1 ) along with the clinical features. The numbers in the image denote the size of the feature dimension at each layer.
Diagnostics 13 02251 g001
Figure 2. Data sample distribution in pCR0 and pCR1 classes of ISPY-1. In the original dataset, each patient was represented by a single instance of MRI scans. However, in the balanced dataset, we have made updates to include multiple MRI scans at the same time-stamp for patients who responded to the NST.
Figure 2. Data sample distribution in pCR0 and pCR1 classes of ISPY-1. In the original dataset, each patient was represented by a single instance of MRI scans. However, in the balanced dataset, we have made updates to include multiple MRI scans at the same time-stamp for patients who responded to the NST.
Diagnostics 13 02251 g002
Figure 3. ROC curves for pCR prediction models on the ISPY-1 dataset: (a) conventional 3D-VGGNet (AUC = 0.68), (b) conventional 3D-ResNet (AUC = 0.50), and (c) proposed deep-NST (AUC = 0.88).
Figure 3. ROC curves for pCR prediction models on the ISPY-1 dataset: (a) conventional 3D-VGGNet (AUC = 0.68), (b) conventional 3D-ResNet (AUC = 0.50), and (c) proposed deep-NST (AUC = 0.88).
Diagnostics 13 02251 g003
Figure 4. Confusion matrices of (a) 3D-VGGNet, (b) 3D-ResNet, and (c) proposed deep-NST frameworks.
Figure 4. Confusion matrices of (a) 3D-VGGNet, (b) 3D-ResNet, and (c) proposed deep-NST frameworks.
Diagnostics 13 02251 g004
Figure 5. ROC curves for the proposed model on different inputs: (a) only T 1 MRI scans; (b) T 1 MRI scans with clinical information; (c) T 1 and T 2 MRI scans with clinical information; and (d) T 1 , T 2 , and T 3 MRI scans with clinical information, over ISPY-1 dataset.
Figure 5. ROC curves for the proposed model on different inputs: (a) only T 1 MRI scans; (b) T 1 MRI scans with clinical information; (c) T 1 and T 2 MRI scans with clinical information; and (d) T 1 , T 2 , and T 3 MRI scans with clinical information, over ISPY-1 dataset.
Diagnostics 13 02251 g005
Figure 6. Confusion matrices of the proposed model for different inputs: (a) only T1 MRI scans; (b) T 1 MRI scans with clinical information; (c) T 1 and T 2 MRI scans with clinical information; and (d) T 1 , T 2 , and T 3 MRI scans with clinical information, over ISPY-1 dataset.
Figure 6. Confusion matrices of the proposed model for different inputs: (a) only T1 MRI scans; (b) T 1 MRI scans with clinical information; (c) T 1 and T 2 MRI scans with clinical information; and (d) T 1 , T 2 , and T 3 MRI scans with clinical information, over ISPY-1 dataset.
Diagnostics 13 02251 g006
Table 1. The AUC score of various methods using the ISPY-1 dataset. Results above the midline are feature-engineering methods. On the other hand, results below the line are DL-based.
Table 1. The AUC score of various methods using the ISPY-1 dataset. Results above the midline are feature-engineering methods. On the other hand, results below the line are DL-based.
MethodAUCACC
Volume [4]0.73N/A
FTV [5]0.73N/A
FTV and Varying PER and SER [6]0.90N/A
CNN and Feature Convolution [12]0.80N/A
CNN pre-post contrast [7]0.85N/A
3D-VGGNet * [16]0.680.68
3D-ResNet * [17]0.500.42
Deep-NST+Focal0.880.79
Deep-NST+3DCNN0.600.58
Deep-NST+3DInception0.610.58
UniModal ST0.840.72
Deep-NST0.880.85
N/A indicates the information was not available. Here, * with 3D VGGNet and 3D ResNet represents that these models are trained from scratch.
Table 2. The AUC and accuracy score for proposed framework and ablation studies using different inputs of the ISPY-1 dataset.
Table 2. The AUC and accuracy score for proposed framework and ablation studies using different inputs of the ISPY-1 dataset.
InputAUCACC
T 1 MRI Scans0.870.77
T 1 MRI Scans + Clinical Data0.850.82
T 1 + T 2 MRI Scans + Clinical Data0.860.73
T 1 + T 2 + T 3 MRI Scans + Clinical Data0.870.77
T 1 + T 2 + T 3 + T 4 MRI Scans + Clinical Data0.880.85
Table 3. The computational complexity analysis for existing and proposed spatiotemporal deep learning framework.
Table 3. The computational complexity analysis for existing and proposed spatiotemporal deep learning framework.
Method#Param.#Mem#FLOPS
3D-VGGNet [16]225 MB1.8 GB 5.52 × 10 3 G
3D-ResNet [17]311.0 MB2.5 GB 1.95 × 10 2 G
Deep-NST+Focal4 MB37.0 MB30.5 G
Deep-NST+3DCNN3.8 MB31.3 MB2.85 G
Deep-NST+3DInception171.0 MB1.4 GB 1.49 × 10 3 G
Unimodal ST4 MB36.6 MB30.5 G
Deep-NST4 MB37.0 MB30.5 G
Here, 3DInception is the 3D convolutional inception module [25,26].
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Verma, M.; Abdelrahman, L.; Collado-Mesa, F.; Abdel-Mottaleb, M. Multimodal Spatiotemporal Deep Learning Framework to Predict Response of Breast Cancer to Neoadjuvant Systemic Therapy. Diagnostics 2023, 13, 2251. https://doi.org/10.3390/diagnostics13132251

AMA Style

Verma M, Abdelrahman L, Collado-Mesa F, Abdel-Mottaleb M. Multimodal Spatiotemporal Deep Learning Framework to Predict Response of Breast Cancer to Neoadjuvant Systemic Therapy. Diagnostics. 2023; 13(13):2251. https://doi.org/10.3390/diagnostics13132251

Chicago/Turabian Style

Verma, Monu, Leila Abdelrahman, Fernando Collado-Mesa, and Mohamed Abdel-Mottaleb. 2023. "Multimodal Spatiotemporal Deep Learning Framework to Predict Response of Breast Cancer to Neoadjuvant Systemic Therapy" Diagnostics 13, no. 13: 2251. https://doi.org/10.3390/diagnostics13132251

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop