Next Article in Journal
Micropulse Transscleral Cyclophotocoagulation for Glaucoma after Penetrating Keratoplasty
Next Article in Special Issue
Cardiovascular/Stroke Risk Assessment in Patients with Erectile Dysfunction—A Role of Carotid Wall Arterial Imaging and Plaque Tissue Characterization Using Artificial Intelligence Paradigm: A Narrative Review
Previous Article in Journal
Evaluation of a Multiplex Real-Time PCR Assay for Detecting Chlamydia trachomatis in Vaginal Samples
Previous Article in Special Issue
Fully Automated Thrombus Segmentation on CT Images of Patients with Acute Ischemic Stroke
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

End-to-End Deep Learning Approach for Perfusion Data: A Proof-of-Concept Study to Classify Core Volume in Stroke CT

1
Department of Radiology, University Hospital, LMU Munich, 81377 Munich, Germany
2
Comprehensive Pneumology Center (CPC-M), German Center for Lung Research (DZL), 81377 Munich, Germany
3
Department of Diagnostic Imaging, The Hospital for Sick Children, University of Toronto, Toronto, ON M5G 1X8, Canada
4
Department of Neurology, University Hospital, LMU Munich, 81377 Munich, Germany
5
Institute for Stroke and Dementia Research, University Hospital, LMU Munich, 81377 Munich, Germany
*
Author to whom correspondence should be addressed.
Diagnostics 2022, 12(5), 1142; https://doi.org/10.3390/diagnostics12051142
Submission received: 8 March 2022 / Revised: 26 April 2022 / Accepted: 3 May 2022 / Published: 5 May 2022
(This article belongs to the Special Issue Artificial Intelligence in Stroke Imaging)

Abstract

:
(1) Background: CT perfusion (CTP) is used to quantify cerebral hypoperfusion in acute ischemic stroke. Conventional attenuation curve analysis is not standardized and might require input from expert users, hampering clinical application. This study aims to bypass conventional tracer-kinetic analysis with an end-to-end deep learning model to directly categorize patients by stroke core volume from raw, slice-reduced CTP data. (2) Methods: In this retrospective analysis, we included patients with acute ischemic stroke due to proximal occlusion of the anterior circulation who underwent CTP imaging. A novel convolutional neural network was implemented to extract spatial and temporal features from time-resolved imaging data. In a classification task, the network categorized patients into small or large core. In ten-fold cross-validation, the network was repeatedly trained, evaluated, and tested, using the area under the receiver operating characteristic curve (ROC-AUC). A final model was created in an ensemble approach and independently validated on an external dataset. (3) Results: 217 patients were included in the training cohort and 23 patients in the independent test cohort. Median core volume was 32.4 mL and was used as threshold value for the binary classification task. Model performance yielded a mean (SD) ROC-AUC of 0.72 (0.10) for the test folds. External independent validation resulted in an ensembled mean ROC-AUC of 0.61. (4) Conclusions: In this proof-of-concept study, the proposed end-to-end deep learning approach bypasses conventional perfusion analysis and allows to predict dichotomized infarction core volume solely from slice-reduced CTP images without underlying tracer kinetic assumptions. Further studies can easily extend to additional clinically relevant endpoints.

1. Introduction

Acute ischemic stroke occurs when a blood clot interrupts the blood flow (perfusion) to the brain, most commonly in a supplying artery—this causes cell death in the hypoperfused areas [1]. Historically, cerebral perfusion imaging was performed using positron emission tomography using radioactive labeled oxygen to determine oxygen fraction and cerebral metabolic rate for oxygen or single photon emission computed tomography. However, logistics and application of radiotracers made both modalities unfeasible for the emergency setting. Today, computed tomography perfusion (CTP) is the most frequently used method to classify the salvageable brain tissue (penumbra) from the irreversibly damaged core in order to support clinical decision-making [2,3,4].
CTP is based on consecutive sampling of cerebral tissue attenuation after intravenous bolus injection of an iodinated contrast agent. A time-attenuation curve in every voxel represents the passage of the contrast agent through the brain in the reconstructed 4D image. After conversion to concentration, tracer-kinetic analysis aims to quantitatively evaluate the time-concentration curves by estimating perfusion parameters, e.g., cerebral blood volume, cerebral blood flow, time to peak, and mean transit time [5]. The most common approach uses deconvolution: An arterial input function (AIF) is determined in a large feeding artery and the time-concentration curves are deconvolved voxel-wise with the AIF to estimate perfusion parameters [6,7,8].
Radiologists as human experts then examine the generated perfusion parameter maps to detect hypoperfused areas, i.e., penumbra and core, and decide among treatment options. In the setting of acute ischemic stroke, CTP can help to identify patients who have a large penumbra and a small core, as they are likely to have a favorable response to reperfusion therapies [9,10]. Additionally, it was shown that CTP can help identify stroke mimics like epilepsy [11] and improves the detection performance for peripheral ischemia with often minor clinical symptoms, which is paramount for future therapy concepts on medium vessel occlusion [12]. However, availability and usage of advanced stroke imaging methods, including CTP, vary considerably among sites and geographical areas, with only around half of centers using those methods frequently [13].
The value of convolutional neural networks (CNN) has been demonstrated for a variety of medical imaging tasks, e.g., image reconstruction, object detection, segmentation, or classification [14,15,16,17,18,19]. The reason for the success of these systems is based on the capability of CNNs to learn data-driven features from pixels directly. Multiple nonlinear processing layers produce a high-level representation of features in images. Consequently, CNN-based approaches were proposed and applied to perfusion imaging analysis. For dynamic contrast-enhanced (DCE) MRI for example, CNN models were developed to estimate perfusion parameters maps directly from the data without the requirement for a standard deconvolution process [20,21]. A recent study [22] proposed a voxelwise prediction of infarct status from stroke CTP, but with the use of additional clinical and tracer-kinetic related data.
Commercial CTP analysis approaches often require domain expertise—e.g., by localizing, verifying, or correcting vessels for the measurement of arterial input functions. Uncertainties, e.g., induced by partial volume effects, propagate into the calculation of perfusion parameter maps via tracer kinetic modeling and ultimately in the process of clinical decision-making. Deep learning, on the other hand, may enable the direct prediction of clinical endpoints from complex imaging data with minimal user input. Starting from the baseline model developed in this study, more complex models can be adapted to relevant clinical endpoints in acute ischemic stroke, e.g., grade of disability or quality of life. For ischemic stroke, the most widely applied measure for neurological outcome uses the modified Rankin scale at day 90 after stroke. Clinical endpoints are likely associated with subtle patterns in the data and development of deep learning models generally requires large datasets [23]. In a proof-of-concept study, we, therefore, aimed to categorize patients into small or large core with an end-to-end deep learning approach for slice-reduced CT perfusion data.

2. Materials and Methods

2.1. Study Population, Image Acquisition and Core Volumetry

In this retrospective study, we included a training cohort (n = 217) (Figure 1) from among 234 consecutive acute ischemic stroke patients with available raw CTP data from a prospectively acquired cohort (German Stroke Registry, NCT03356392). All patients were treated with endovascular mechanical thrombectomy at our institution. We excluded patients with inconsistent CTP images that did not comply with a standardized time resolution of 1.5 s or a scan duration of 48 s. Patients underwent CTP on admission using a SOMATOM Definition Force, AS+, or Flash CT scanner (Siemens Healthineers, Forchheim, Germany). Automated calculation of ischemic core was performed using the CT vendor’s proprietary software (syngo Neuro Perfusion CT; Siemens Healthineers, Forchheim, Germany), which applies a threshold cerebral blood volume of <1.2 mL/100 mL. Median core volume was calculated and used as threshold value for binary classification: Small core < median core, large core > median core. As a proof-of-principle, we used median core volume to ensure a balanced split of the training data, in contrast to a fixed threshold value of 70 mL, which can sometimes be found in the literature [24].
For independent validation, we included a second external test cohort (n = 23) from among 63 patients of the external, publicly available ISLES 2018 challenge data set [25,26]. We excluded patients with inconsistent CTP images that could not be interpolated to the standardized time resolution of 1.5 s and a scan duration of 48 s. The challenge data include core segmentations, which were used to calculate the ischemic core volume. For this purpose, the number of voxels in the segmentation was multiplied by the voxel dimensions to get an estimated core volume in ml. The median core volume of the training dataset was used as threshold value for binary classification. Figure 1 shows a detailed flowchart of patient selection for both cohorts.

2.2. Preprocessing, Batch Generation, and Data Augmentation

Internal and external datasets were both preprocessed in two steps. First, two axial slices covering the middle cerebral artery territory (basal ganglia and supraganglionic level) based on the Alberta stroke program early CT score (ASPECTS) [27] regions were selected by radiologists. In a second fully automated step, the selected slices were resized and interpolated to 128 × 128 pixels in-plane resolution with 200 × 200 mm2 length. All slices along the time axis were co-registered to the first slice at t = 0 to reduce motion artifacts. All data were processed using custom Python (version 3.8.5) [28] scripts including the publicly available packages SimpleITK (version 2.0.2) [29] and Scikit-learn (version 0.23.2) [30].
During training, a custom batch generator returned a random subset of samples (batch size = 12) from the complete dataset and normalized each batch to zero mean and unit variance. Online data augmentation was applied to each batch in the form of random rotation in the range of (−15°, 15°), xy-shift (−10 pixel, 10 pixel), and vertical flip (True, False) before passing it on to the network.

2.3. Network Architecture

The proposed network architecture was implemented in Python and TensorFlow (version 2.3.0) [31] and is illustrated in Figure 2. It consists of two submodels with identical architecture for each of the selected axial CTP slices. The standardized and augmented 2D+t input images were fed into each submodel and processed through the pipeline to extract spatial and temporal features. The resulting features are concatenated, passed through a fully connected dense layer, and classified (Figure 2 top).
Figure 2 (bottom, zoomed-in) displays a detailed demonstration of the feature extraction part. For spatial feature extraction, each 2D image on the time axis is fed into a VGG19 model [32], pretrained on 2D image net data [33]. The weights are shared across all timepoints within a “TimeDistributed” framework. The resulting 32 × 512 feature matrix is passed on to the temporal feature extraction step. The temporal feature extraction consists of a 1D convolution with three filters followed by a max pooling layer and is divided into a global and a local pathway. In the global pathway, the 1D convolution is performed with kernel size 11, in the local pathway with kernel size 3. This ensures that the model can capture both smaller and larger changes along the time course. The resulting feature vectors are concatenated and passed on to a dense layer with 32 units. Classification is performed using a sigmoid layer. The source code is made publicly available on the development platform Github (https://github.com/AndreasMittermeier/stroke-perfusion-CNN (accessed on 7 March 2022)).

2.4. Training, Validation and Testing

The proposed network was trained, validated, and tested using all included patients from the training cohort within a 10-fold cross-validation (CV). The dataset was randomly split into ten folds according to an 8:1:1 ratio of training, validation, and test. Eight folds were used for training the network. The number of training epochs was set to 500. The validation fold was used to evaluate the model after each epoch and stop training once the validation loss stopped decreasing for 200 epochs (patience = 200). After the last epoch, the test fold was evaluated by the model with the best weights, i.e., the weights which yielded the lowest validation loss. After ten CV iterations, each fold was used for unbiased testing once. The area under the receiver operating characteristic curve (ROC-AUC) was used as evaluation metric. Mean and standard deviation (SD) of ROC-AUC values were reported for the 10-fold CV.
In addition, we performed an ablation study on the effect of the local and global temporal feature extraction. To this end, we trained two reduced models using the (i) local feature extractor alone and using the (ii) global feature extractor alone. Training and evaluation on the test folds was performed in the same CV approach as described above. Mean and SD of the ROC-AUC values were reported and compared with those of the full model.
To evaluate the independent test cohort, an ensemble method was used. The final model was constructed by averaging the predictions from the ten models trained in the CV. The final model was applied to the independent test cohort and the ROC-AUC was reported.

3. Results

Two hundred seventeen patients were included in the training cohort and 23 patients in the independent test cohort. Median core volume for the training cohort was 32.4 mL which yields, per definition, a balanced class split. Applying this threshold to the independent test data resulted in 12 patients with large core volume and 11 patients with small core volume. Training duration for the 10-fold CV was in the range of 24 h on a local workstation (NVIDIA GeForce RTX 2070 Super) with online data augmentation and batch-wise data standardization. Evaluation and prediction of unseen data were of the order of a few seconds.
Figure 3 shows the ROC curves for the test folds within the 10-fold CV. The mean ROC curve is overlaid in blue, an interval of ±1 SD is shaded in grey, and the dashed line represents random guessing. The mean (SD) ROC-AUC over 10 folds was 0.72 (0.10) for the test folds. In comparison, mean (SD) ROC-AUC for the validation folds, which were used to early stop training, was 0.75 (0.11). The averaged ensemble ROC-AUC for the independent test cohort selected from the ISLES 2018 challenge was 0.61, which is close to the ±1 SD interval of the test folds from the training data. All results are summarized in Table 1.
The ablation study showed a decrease in mean (SD) ROC-AUC values for the reduced models, summarized in Table 2. Using the global feature extractor alone resulted in a ROC-AUC of 0.63 (0.14) and using the local feature extractor alone resulted in a ROC-AUC of 0.65 (0.13), compared to the full model with ROC-AUC 0.72 (0.10).

4. Discussion

In this proof-of-concept study, we developed a novel end-to-end deep learning approach to bypass conventional perfusion analysis, which allows to directly categorize patients into small or large core from raw CTP data without tracer-kinetic assumptions. We demonstrated this approach on 217 patients with acute ischemic stroke by directly predicting dichotomized infarct core volume and showed that the model learned relevant spatial and temporal features purely from the data. The results of the ablation study demonstrate the advantage of combined local and global temporal feature extraction, as only the full model yields the best performance. This corroborates our understanding that both short-term effects (e.g., sharp peaks in concentration) and long-term effects (e.g., wash-out) add relevant information and must therefore both be considered in the model architecture. In this proof-of-concept approach, our model cannot be translated to clinical practice immediately, however, achieved good predictive performance on an inhouse dataset in a 10-fold CV approach and generalized to independent test data, showing the potential of end-to-end CT perfusion analysis.
The proposed deep learning model is based on a 2D approach that covers a reduced portion of the middle cerebral artery territory, represented by the ASPECTS regions. Arguably, a 3D approach would contain more relevant information, but whole-brain CT perfusion is not available in all primary stroke centers [34]. Using the 2D approach, we were able to include all possible data, especially the external test data, which consisted of two separate stacks of slices instead of whole-brain perfusion. We believe that this 2D approach is sufficient to prove the concept that spatial and temporal information can be extracted from CTP data to predict dichotomized core volume using deep learning.
In comparison to existing studies using deep learning for perfusion analysis in stroke CT [8], we focused on using the raw perfusion data solely. In contrast to voxelwise prediction of infarct status, the proposed model learned to predict dichotomized infarction core volume without taking additional parameters into account. While additional parameters like treatment information may be beneficial, additional user-provided information, such as a manually selected arterial input function, requires input from expert users and introduces user dependency. Our proposed model learned the link between perfusion input and tissue response purely based on the data and is free from tracer-kinetic assumptions.
Imaging-derived parameters play a crucial role in clinical decision-making in the setting of acute ischemic stroke. Foremost, CTP-derived ischemic core volume has become one of the key parameters in the decision for mechanical thrombectomy in the extended time window [35,36]. Currently, there is no consensus on the use of CTP-parameters for core/penumbra estimation. While software in large clinical trials used relative CBF thresholds (RAPID), other software relies on MTT (Philips Brain CT perfusion) or, in our case, on CBV. As relevant differences have been shown between vendors, our approach needs further validation for other CTP analysis thresholds [37]. In the present proof-of-concept study, we predicted dichotomized ischemic core volume as a simplified endpoint. Given sufficient training data, this approach can easily be generalized to more complex labels in future studies, such as impairment after discharge. The underlying relationship is harder to learn and may require incorporating additional clinical parameters into the model. Such deep-learning-based approaches may, therefore, be used to predict complications and even chronic functional outcomes in order to guide clinical management in and beyond the acute stroke phase.
The present study is not without limitations. First, the sample size of 217 training datasets is small for the complex problem of directly predicting an imaging-derived parameter from raw data and validation with a larger dataset is needed. To this end, dichotomized median core volume was chosen (i) to cast the problem as classification approach and (ii) to provide a balanced group distribution for the training dataset. This preliminary work may lay the groundwork for future studies to examine more clinically relevant endpoints, such as grade of disability or quality of life. Nevertheless, as proof of concept, our model achieved good results, which were validated on external test data. The performance gap between model predictions for in-house and external test data is likely due to differences in data quality and core labeling. Second, the number of perfusion timepoints of the CTP images was fixed for the model input, but appropriate interpolation could solve varying temporal resolutions.

5. Conclusions

In this proof-of-concept study, the proposed end-to-end deep learning approach bypasses conventional perfusion analysis and allows training a model that predicts dichotomized infarction core volume solely from slice-reduced CTP images without underlying tracer kinetic assumptions. Further studies can easily extend to additional clinically relevant endpoints.

Author Contributions

Conceptualization, all authors; methodology, A.M., B.S., P.W. and M.I.; data curation, A.M., P.R., M.P.F., L.K., S.T. and W.G.K.; writing—original draft preparation, A.M., P.R., M.P.F., B.S., P.W. and M.I.; writing—review and editing, A.M., P.R., M.P.F., B.S., P.W., B.E.-W., O.D., J.R., L.K., S.T., W.G.K. and M.I.; All authors have read and agreed to the published version of the manuscript.

Funding

A.M. was supported by the German Research Foundation (DFG) within the Research Training Group GRK 2274.

Institutional Review Board Statement

The retrospective study was approved by the institutional review board LMU Munich, 19-682, 2019-09-26, according to the Declaration of Helsinki of 2013.

Informed Consent Statement

The institutional review board LMU Munich waived the requirement for written informed consent (19-682, 2019-09-26) due to the retrospective analysis of anonymized data.

Data Availability Statement

Validation data available in a publicly accessible repository. The data of the ISLES 2018 challenge used as validation data in this study are openly available at https://www.smir.ch/ISLES/Start2018 (accessed on 2 May 2021).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. van der Worp, H.B.; van Gijn, J. Acute Ischemic Stroke. N. Engl. J. Med. 2007, 357, 572–579. [Google Scholar] [CrossRef] [PubMed]
  2. Chalet, L.; Boutelier, T.; Christen, T.; Raguenes, D.; Debatisse, J.; Eker, O.F.; Becker, G.; Nighoghossian, N.; Cho, T.-H.; Canet-Soulas, E.; et al. Clinical Imaging of the Penumbra in Ischemic Stroke: From the Concept to the Era of Mechanical Thrombectomy. Front. Cardiovasc. Med. 2022, 9, 438. [Google Scholar] [CrossRef] [PubMed]
  3. Allmendinger, A.M.; Tang, E.R.; Lui, Y.W.; Spektor, V. Imaging of Stroke: Part 1, Perfusion CT--Overview of Imaging Technique, Interpretation Pearls, and Common Pitfalls. AJR Am. J. Roentgenol. 2012, 198, 52–62. [Google Scholar] [CrossRef] [PubMed]
  4. Merino, J.G.; Warach, S. Imaging of Acute Stroke. Nat. Rev. Neurol. 2010, 6, 560–571. [Google Scholar] [CrossRef]
  5. Ingrisch, M.; Sourbron, S. Tracer-Kinetic Modeling of Dynamic Contrast-Enhanced MRI and CT: A Primer. J. Pharm. Pharm. 2013, 40, 281–300. [Google Scholar] [CrossRef]
  6. Sourbron, S.; Dujardin, M.; Makkat, S.; Luypaert, R. Pixel-by-Pixel Deconvolution of Bolus-Tracking Data: Optimization and Implementation. Phys. Med. Biol. 2007, 52, 429–447. [Google Scholar] [CrossRef]
  7. Fieselmann, A.; Kowarschik, M.; Ganguly, A.; Hornegger, J.; Fahrig, R. Deconvolution-Based CT and MR Brain Perfusion Measurement: Theoretical Model Revisited and Practical Implementation Details. J. Biomed. Imaging 2011, 2011, 1–20. [Google Scholar] [CrossRef]
  8. Bivard, A.; Levi, C.; Spratt, N.; Parsons, M. Perfusion CT in Acute Stroke: A Comprehensive Analysis of Infarct and Penumbra. Radiology 2013, 267, 543–550. [Google Scholar] [CrossRef] [Green Version]
  9. Lansberg, M.G.; Christensen, S.; Kemp, S.; Mlynash, M.; Mishra, N.; Federau, C.; Tsai, J.P.; Kim, S.; Nogueria, R.G.; Jovin, T.; et al. Computed Tomographic Perfusion to Predict Response to Recanalization in Ischemic Stroke. Ann. Neurol. 2017, 81, 849–856. [Google Scholar] [CrossRef]
  10. Alexandre, A.M.; Pedicelli, A.; Valente, I.; Scarcia, L.; Giubbolini, F.; D’Argento, F.; Lozupone, E.; Distefano, M.; Pilato, F.; Colosimo, C. May Endovascular Thrombectomy without CT Perfusion Improve Clinical Outcome? Clin. Neurol. Neurosurg. 2020, 198, 106207. [Google Scholar] [CrossRef]
  11. Van Cauwenberge, M.G.A.; Dekeyzer, S.; Nikoubashman, O.; Dafotakis, M.; Wiesmann, M. Can Perfusion CT Unmask Postictal Stroke Mimics? A Case-Control Study of 133 Patients. Neurology 2018, 91, e1918–e1927. [Google Scholar] [CrossRef]
  12. Becks, M.J.; Manniesing, R.; Vister, J.; Pegge, S.A.H.; Steens, S.C.A.; van Dijk, E.J.; Prokop, M.; Meijer, F.J.A. Brain CT Perfusion Improves Intracranial Vessel Occlusion Detection on CT Angiography. J. Neuroradiol. 2019, 46, 124–129. [Google Scholar] [CrossRef] [PubMed]
  13. Wintermark, M.; Luby, M.; Bornstein, N.M.; Demchuk, A.; Fiehler, J.; Kudo, K.; Lees, K.R.; Liebeskind, D.S.; Michel, P.; Nogueira, R.G.; et al. International Survey of Acute Stroke Imaging Used to Make Revascularization Treatment Decisions. Int. J. Stroke 2015, 10, 759–762. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  15. LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  16. Szegedy, C.; Toshev, A.; Erhan, D. Deep Neural Networks for Object Detection. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–10 December 2013; Burges, C.J.C., Bottou, L., Welling, M., Ghahramani, Z., Weinberger, K.Q., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2013; Volume 26. [Google Scholar]
  17. Yang, G.; Yu, S.; Dong, H.; Slabaugh, G.; Dragotti, P.L.; Ye, X.; Liu, F.; Arridge, S.; Keegan, J.; Guo, Y.; et al. DAGAN: Deep De-Aliasing Generative Adversarial Networks for Fast Compressed Sensing MRI Reconstruction. IEEE Trans. Med. Imaging 2018, 37, 1310–1321. [Google Scholar] [CrossRef] [Green Version]
  18. Shen, D.; Wu, G.; Suk, H.-I. Deep Learning in Medical Image Analysis. Annu. Rev. Biomed. Eng. 2017, 19, 221–248. [Google Scholar] [CrossRef] [Green Version]
  19. Lai, M. Deep Learning for Medical Image Segmentation. arXiv 2015, arXiv:1505.02000. [Google Scholar]
  20. Ho, K.C.; Scalzo, F.; Sarma, K.V.; El-Saden, S.; Arnold, C.W. A Temporal Deep Learning Approach for MR Perfusion Parameter Estimation in Stroke. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 1315–1320. [Google Scholar]
  21. Ulas, C.; Das, D.; Thrippleton, M.J.; Valdés Hernández, M.D.C.; Armitage, P.A.; Makin, S.D.; Wardlaw, J.M.; Menze, B.H. Convolutional Neural Networks for Direct Inference of Pharmacokinetic Parameters: Application to Stroke Dynamic Contrast-Enhanced MRI. Front. Neurol. 2019, 9, 1147. [Google Scholar] [CrossRef] [Green Version]
  22. Robben, D.; Boers, A.M.M.; Marquering, H.A.; Langezaal, L.L.C.M.; Roos, Y.B.W.E.M.; van Oostenbrugge, R.J.; van Zwam, W.H.; Dippel, D.W.J.; Majoie, C.B.L.M.; van der Lugt, A.; et al. Prediction of Final Infarct Volume from Native CT Perfusion and Treatment Parameters Using Deep Learning. Med. Image Anal. 2020, 59, 101589. [Google Scholar] [CrossRef]
  23. Chartrand, G.; Cheng, P.M.; Vorontsov, E.; Drozdzal, M.; Turcotte, S.; Pal, C.J.; Kadoury, S.; Tang, A. Deep Learning: A Primer for Radiologists. RadioGraphics 2017, 37, 2113–2131. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Hakimelahi, R.; Yoo, A.J.; He, J.; Schwamm, L.H.; Lev, M.H.; Schaefer, P.W.; González, R.G. Rapid Identification of a Major Diffusion/Perfusion Mismatch in Distal Internal Carotid Artery or Middle Cerebral Artery Ischemic Stroke. BMC Neurol. 2012, 12, 132. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Kistler, M.; Bonaretti, S.; Pfahrer, M.; Niklaus, R.; Büchler, P. The Virtual Skeleton Database: An Open Access Repository for Biomedical Research and Collaboration. J. Med. Internet Res. 2013, 15, e2930. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Maier, O.; Menze, B.H.; von der Gablentz, J.; Häni, L.; Heinrich, M.P.; Liebrand, M.; Winzeck, S.; Basit, A.; Bentley, P.; Chen, L.; et al. ISLES 2015—A Public Evaluation Benchmark for Ischemic Stroke Lesion Segmentation from Multispectral MRI. Med. Image Anal. 2017, 35, 250–269. [Google Scholar] [CrossRef] [Green Version]
  27. Barber, P.A.; Demchuk, A.M.; Zhang, J.; Buchan, A.M. Validity and Reliability of a Quantitative Computed Tomography Score in Predicting Outcome of Hyperacute Stroke before Thrombolytic Therapy. ASPECTS Study Group. Alberta Stroke Programme Early CT Score. Lancet 2000, 355, 1670–1674. [Google Scholar] [CrossRef]
  28. Van Rossum, G.; Drake, F.L., Jr. Python Reference Manual; Centrum voor Wiskunde en Informatica Amsterdam: Amsterdam, The Netherlands, 1995. [Google Scholar]
  29. Beare, R.; Lowekamp, B.; Yaniv, Z. Image Segmentation, Registration and Characterization in R with SimpleITK. J. Stat. Softw. 2018, 86, 8. [Google Scholar] [CrossRef] [Green Version]
  30. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-Learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  31. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. arXiv 2016, arXiv:1603.04467. [Google Scholar]
  32. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
  33. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Li, F.-F. Imagenet: A Large-Scale Hierarchical Image Database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  34. Almekhlafi, M.A.; Kunz, W.G.; Menon, B.K.; McTaggart, R.A.; Jayaraman, M.V.; Baxter, B.W.; Heck, D.; Frei, D.; Derdeyn, C.P.; Takagi, T.; et al. Imaging of Patients with Suspected Large-Vessel Occlusion at Primary Stroke Centers: Available Modalities and a Suggested Approach. Am. J. Neuroradiol. 2019, 40, 396–400. [Google Scholar] [CrossRef] [Green Version]
  35. Albers, G.W.; Marks, M.P.; Kemp, S.; Christensen, S.; Tsai, J.P.; Ortega-Gutierrez, S.; McTaggart, R.A.; Torbey, M.T.; Kim-Tenser, M.; Leslie-Mazwi, T.; et al. Thrombectomy for Stroke at 6 to 16 Hours with Selection by Perfusion Imaging. N. Engl. J. Med. 2018, 378, 708–718. [Google Scholar] [CrossRef] [PubMed]
  36. Nogueira, R.G.; Jadhav, A.P.; Haussen, D.C.; Bonafe, A.; Budzik, R.F.; Bhuva, P.; Yavagal, D.R.; Ribo, M.; Cognard, C.; Hanel, R.A.; et al. Thrombectomy 6 to 24 Hours after Stroke with a Mismatch between Deficit and Infarct. N. Engl. J. Med. 2018, 378, 11–21. [Google Scholar] [CrossRef] [PubMed]
  37. Austein, F.; Riedel, C.; Kerby, T.; Meyne, J.; Binder, A.; Lindner, T.; Huhndorf, M.; Wodarg, F.; Jansen, O. Comparison of Perfusion CT Software to Predict the Final Infarct Volume After Thrombectomy. Stroke 2016, 47, 2311–2317. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Flowchart of patient selection for the training and independent test cohort. CTP = CT perfusion.
Figure 1. Flowchart of patient selection for the training and independent test cohort. CTP = CT perfusion.
Diagnostics 12 01142 g001
Figure 2. Model architecture overview and detailed, zoomed-in view of the spatial and temporal feature extraction process. The selected slices A and B are fed into identical submodels for spatial and temporal feature extraction. Spatial feature extraction consists of identical, pretrained VGG19 networks for each timepoint of the input images. The resulting feature vector is passed on to the temporal feature extraction. 1D convolutions with two different kernel sizes are carried out in a global and local pathway. The extracted features A and B for both submodels are concatenated, fully connected (FC), and classified.
Figure 2. Model architecture overview and detailed, zoomed-in view of the spatial and temporal feature extraction process. The selected slices A and B are fed into identical submodels for spatial and temporal feature extraction. Spatial feature extraction consists of identical, pretrained VGG19 networks for each timepoint of the input images. The resulting feature vector is passed on to the temporal feature extraction. 1D convolutions with two different kernel sizes are carried out in a global and local pathway. The extracted features A and B for both submodels are concatenated, fully connected (FC), and classified.
Diagnostics 12 01142 g002
Figure 3. ROC curves for test data in the 10-fold CV. Mean (SD) ROC-AUC for the CV test folds was 0.72 (0.10). CV = cross-validation, ROC-AUC = area under the receiver operator characteristics curve.
Figure 3. ROC curves for test data in the 10-fold CV. Mean (SD) ROC-AUC for the CV test folds was 0.72 (0.10). CV = cross-validation, ROC-AUC = area under the receiver operator characteristics curve.
Diagnostics 12 01142 g003
Table 1. Mean (SD) ROC-AUC of the final model for validation and test folds during CV and for the external test cohort. SD = standard deviation, CV = cross-validation, ROC-AUC = area under the receiver operator characteristics curve.
Table 1. Mean (SD) ROC-AUC of the final model for validation and test folds during CV and for the external test cohort. SD = standard deviation, CV = cross-validation, ROC-AUC = area under the receiver operator characteristics curve.
Validation FoldsTest FoldsIndependent Test Cohort
0.75 (0.11)0.72 (0.10)0.61
Table 2. Mean (SD) ROC-AUC for the test folds during CV of the full model and the reduced models within the ablation study setting. SD = standard deviation, CV = cross-validation, ROC-AUC = area under the receiver operator characteristics curve.
Table 2. Mean (SD) ROC-AUC for the test folds during CV of the full model and the reduced models within the ablation study setting. SD = standard deviation, CV = cross-validation, ROC-AUC = area under the receiver operator characteristics curve.
Full ModelGlobal Feature Extractor AloneLocal Feature Extractor Alone
0.72 (0.10)0.63 (0.14)0.65 (0.13)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mittermeier, A.; Reidler, P.; Fabritius, M.P.; Schachtner, B.; Wesp, P.; Ertl-Wagner, B.; Dietrich, O.; Ricke, J.; Kellert, L.; Tiedt, S.; et al. End-to-End Deep Learning Approach for Perfusion Data: A Proof-of-Concept Study to Classify Core Volume in Stroke CT. Diagnostics 2022, 12, 1142. https://doi.org/10.3390/diagnostics12051142

AMA Style

Mittermeier A, Reidler P, Fabritius MP, Schachtner B, Wesp P, Ertl-Wagner B, Dietrich O, Ricke J, Kellert L, Tiedt S, et al. End-to-End Deep Learning Approach for Perfusion Data: A Proof-of-Concept Study to Classify Core Volume in Stroke CT. Diagnostics. 2022; 12(5):1142. https://doi.org/10.3390/diagnostics12051142

Chicago/Turabian Style

Mittermeier, Andreas, Paul Reidler, Matthias P. Fabritius, Balthasar Schachtner, Philipp Wesp, Birgit Ertl-Wagner, Olaf Dietrich, Jens Ricke, Lars Kellert, Steffen Tiedt, and et al. 2022. "End-to-End Deep Learning Approach for Perfusion Data: A Proof-of-Concept Study to Classify Core Volume in Stroke CT" Diagnostics 12, no. 5: 1142. https://doi.org/10.3390/diagnostics12051142

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop