Next Article in Journal
Improved Cervical Lymph Node Characterization among Patients with Head and Neck Squamous Cell Carcinoma Using MR Texture Analysis Compared to Traditional FDG-PET/MR Features Alone
Previous Article in Journal
An Automatic DWI/FLAIR Mismatch Assessment of Stroke Patients
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sequence-Type Classification of Brain MRI for Acute Stroke Using a Self-Supervised Machine Learning Algorithm

1
Department of Computer Science and Engineering, Konkuk University, Seoul 05029, Republic of Korea
2
Biomedical Research Center, Asan Institute for Life Sciences, Asan Medical Center, Seoul 05505, Republic of Korea
3
Department of Radiology and Research Institute of Radiology, Asan Medical Center, University of Ulsan College of Medicine, Seoul 05505, Republic of Korea
4
Clinical Research Center, Asan Medical Center, Seoul 05505, Republic of Korea
5
Department of Convergence Medicine, University of Ulsan College of Medicine, Seoul 05505, Republic of Korea
6
Trialinformatics Inc., Seoul 05505, Republic of Korea
7
Department of Radiation Science & Technology, Jeonbuk National University, Jeonju 56212, Republic of Korea
8
Shin Poong Pharm. Co., Ltd., Seoul 06246, Republic of Korea
9
Graduate School of Clinical Pharmacy, CHA University, Pocheon-si 11160, Republic of Korea
10
Department of Smart ICT Convergence Engineering, Konkuk University, Seoul 05029, Republic of Korea
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Diagnostics 2024, 14(1), 70; https://doi.org/10.3390/diagnostics14010070
Submission received: 21 November 2023 / Revised: 18 December 2023 / Accepted: 22 December 2023 / Published: 27 December 2023
(This article belongs to the Section Medical Imaging and Theranostics)

Abstract

:
We propose a self-supervised machine learning (ML) algorithm for sequence-type classification of brain MRI using a supervisory signal from DICOM metadata (i.e., a rule-based virtual label). A total of 1787 brain MRI datasets were constructed, including 1531 from hospitals and 256 from multi-center trial datasets. The ground truth (GT) was generated by two experienced image analysts and checked by a radiologist. An ML framework called ImageSort-net was developed using various features related to MRI acquisition parameters and used for training virtual labels and ML algorithms derived from rule-based labeling systems that act as labels for supervised learning. For the performance evaluation of ImageSort-net (MLvirtual), we compare and analyze the performances of models trained with human expert labels (MLhumans), using as a test set blank data that the rule-based labeling system failed to infer from each dataset. The performance of ImageSort-net (MLvirtual) was comparable to that of MLhuman (98.5% and 99%, respectively) in terms of overall accuracy when trained with hospital datasets. When trained with a relatively small multi-center trial dataset, the overall accuracy was relatively lower than that of MLhuman (95.6% and 99.4%, respectively). After integrating the two datasets and re-training them, MLvirtual showed higher accuracy than MLvirtual trained only on multi-center datasets (95.6% and 99.7%, respectively). Additionally, the multi-center dataset inference performances after the re-training of MLvirtual and MLhumans were identical (99.7%). Training of ML algorithms based on rule-based virtual labels achieved high accuracy for sequence-type classification of brain MRI and enabled us to build a sustainable self-learning system.

1. Introduction

Recently, it has become common to collect multi-institutional data to develop and verify artificial intelligence in the field of medical imaging [1,2], and as available medical image data increases, curation work to refine and manage the data becomes more important [3,4]. However, medical imaging data has not been fully standardized, and the metadata of digital imaging and communications in medicine (DICOM) image files often differs across sites, vendors, and acquisition protocols [5,6].
Brain Magnetic Resonance Imaging (MRI) stands as an indispensable tool in the routine diagnosis and ongoing monitoring of individuals affected by various brain-related conditions, encompassing the realm of Brain-Computer Interface (BCI) technology. Its pivotal role lies in its ability to capture a diverse array of sequences, each offering a distinct set of data pertaining to anatomical structures, tissue density, and vascularization patterns within the brain. These sequences serve as a comprehensive means to portray both the anatomical nuances and pathological manifestations within the brain’s intricate landscape. Specifically, the multifaceted information derived from MRI sequences enables a detailed representation of the brain’s structural framework and provides invaluable insights into pathological alterations occurring within its complex network. Brain MRI, in particular, can acquire a variety of sequences, and each sequence provides these data [7].
In the field of stroke research, brain MRI is the most common imaging modality with various sequences, such as T1-weighted image (T1-WI), T2-weighted image (T2-WI), diffusion-weighted image (DWI), fluid-attenuated inversion recovery (FLAIR), perfusion images, susceptibility/gradient images (suscgre), and magnetic resonance angiography (MRA). It is crucial to sort brain MRI according to its specific sequence; however, owing to the variability of DICOM metadata, it is very time-consuming to classify brain MRI sequences according to the standardized sequence types [8]. Therefore, the need for automation for sequence-type classification of the brain MRI has been raised. Indeed, there have been several prior studies on automatic sequence types [9].
Current research on automatic sequence type classification has utilized supervised learning with manual labeling by human experts. However, manual labeling is a time-consuming and tedious task, which may hamper the updates to the automatic sequence type classification technique. A possible solution to this is self-supervised learning, which utilizes unlabeled data for training machine learning (ML) algorithms.
In self-supervised learning, an unlabeled training dataset can be labeled by using supervisory signals embedded in the data, such as metadata, or by using a pre-trained algorithm based on small amounts of human-labeled data [10,11]. Of these, the former method can be easily performed if there are established rules for labeling. In our imaging laboratory, we have established a rule-based labeling system for sequence type classification based on DICOM header information.
Thus, we aimed to develop a self-supervised ML algorithm for sequence-type classification of brain MRI for acute stroke using supervisory signals from DICOM metadata (i.e., rule-based virtual labeling).

2. Materials and Methods

This study was approved by the institutional review board of Asan Medical Center (AMC) (IRB No. 2022-0025). Because this is a retrospective study, informed consent was not required. The results in this study were reported according to the methods and terms in published literature guidance on ML for medical applications [12].

2.1. Data Sources and Dataset

A hospital and a multi-center trial dataset were used in this study (Table 1). For the hospital dataset, we collected from the picture archive and communication system (PACS) of AMC (PetaVision) consecutive brain MRI scans of 1528 patients (704 women; mean age, 59 ± 14.6 years) who underwent brain MRI for evaluation of stroke between January and February 2016. Given that AMC is a tertiary referral center, there were 357 MRI scans taken from lower-level hospitals in the PACS. We included such MRI scans with various manufacturers and sequences in the hospital dataset.
For the multi-center trial dataset, we used brain MRI scans that had been collected and archived for the clinical trial of SP-8203 (NCT 02787278) entitled “Safety and Efficacy of Two Doses of SP-8203 in Patients with Ischemic Stroke Requiring rtPA (SP-8203-2001)” [13]. The multi-center trial dataset contained 59 patients with 256 MRI scans from 8 medical centers in South Korea between June 2016 and August 2017. In the multi-center trial dataset, MRI scans were acquired from three different manufacturers (GE, Philips, and Siemens).

2.2. Generation of Human Expert Labeling

For the generation of human expert labeling (i.e., ground truth [GT]), we labeled eight MRI sequences that were commonly used for stroke evaluation, namely T1, T2, Scout, Diffusion, MRA, FLAIR, suscgre, and Perfusion. T1 refers to common T1-weighted sequences, including T1 spin echo (e.g., T1 SE), fast spin echo (e.g., T1 FSE), gradient (e.g., T1 MPRAGE), or FLAIR (e.g., T1-FLAIR), acquired by either a 2D or 3D scheme. T2 refers to common T2-weighted sequences, including 2D T2 fast spin-echo (e.g., T2 FSE) or 3D T2 fast spin-echo (e.g., T2 SPACE). Diffusion includes various kinds of diffusion-weighted images, including single-shot spin-echo echo-planar imaging (e.g., SE-EPI), multi-shot gradient echo-planar imaging (e.g., RESOLVE), or fast spin-echo imaging (e.g., FSE-DWI). MRA includes time-of-flight images (e.g., TOF-MRA) and contrast-enhanced MRA acquired by spoiled 3D T1 gradient echo sequences. FLAIR refers to the T2-FLAIR either in a 2D or 3D scheme. Suscgre includes susceptibility-weighted imaging acquired by a 3D gradient-echo sequence, echo-planar imaging, or 3D radial sequence, as well as T2*-weighted gradient-echo images. Perfusion includes both dynamic contrast-enhanced (DCE) and dynamic susceptibility contrast (DSC) imaging. All other types of sequences or images, such as MR spectroscopy and screenshots of the outputs of external post-processing software, were categorized as “Others” and excluded from the ML training.
The SeriesDescription and ProtocolName attributes were extracted from each MRI scan based on DICOM metadata attributes ((0008,103 E), (0018,1030), respectively). In each dataset, there were different values of SeriesDescriptions and ProtocolNames across different manufacturers and institutions, as presented in Supplementary Table S1.
Based on these attributes, the GT for the sequence types was generated by two experienced image analysts (S.J.H. and M.H.K.) and checked by a radiologist (K.W.K.). The total number of sequence types in the hospital dataset is presented in Supplementary Table S1. Results from the generation of human expert labels are described here as GT.

2.3. Development of a Rule-Based Labeling System

We developed a rule-based labeling system to generate virtual labels that can simulate and replace human expert labels. Rule-based labeling is a type of semantic-guided virtual label [14]. In addition, the rule-based labeling system is essential for advancing towards a sustainable artificial intelligence (AI) system [15].
Although the SeriesDescription and ProtocolName differ across manufacturers and institutions, they commonly include the same or similar keywords to indicate specific sequences. For example, the sequence type of diffusion-weighted image (hereafter referred to as diffusion) across many different SeriesDescriptions includes several common keywords, such as dw, dwi, diff, adc, and diffusion. We selected keywords that are commonly used in brain MRI for stroke so that they can be applied to data extracted from various hospitals and devices. We exclude ambiguous keywords that can be used in multiple sequence types. For example, the word maximum intensity projection (MIP) is used in the MRA sequence as well as other sequences, such as suscgre.
The rule-based labeling system is composed of two steps (Figure 1). The first step is to identify predefined keywords in Rule Table 1 from the SeriesDescription attribute of the DICOM metadata of an MRI series. If the SeriesDescription attribute shows a null value (i.e., blank) or contains no predefined keywords, then the ProtocolName is searched. The second step is to determine the results. If there is only one keyword, the result is determined by the keyword. If there are two or more keywords in the same SeriesDescription or ProtocolName, a decision is made based on Rule Supplementary Table S1. For example, we classified T1-FLAIR as a T1 and T2-FLAIR as a FLAIR. Labels made through the rule-based labeling system are described here as virtual labels.
Our rule-based labeling system implemented under the above conditions may not cover sequence data with incomplete DICOM metadata. In such a case, the result of the rule-based labeling system is labeled as blank.

2.4. Design of ImageSort-Net Architecture

Figure 2 illustrates the overall architecture of our ImageSort-net, which classifies the MRI sequence type. Our model is developed as a self-sustainable MRI sequence-type classifier algorithm, composed of three steps as follows:
(1)
The pre-processing step extracts various DICOM metadata, including the SeriesDescription attribute, the ProtocolName attribute, and many other attributes associated with MR acquisition parameters.
(2)
Training preparation step, which prepares necessary data for supervised learning of the ML algorithm, including virtual labels derived from our rule-based labeling system and features derived from normalization of DICOM attributes.
(3)
Training step that trains the random forest ML algorithm.
Figure 2. Overall steps to train ImageSort-Net and enable self-sustainable training.
Figure 2. Overall steps to train ImageSort-Net and enable self-sustainable training.
Diagnostics 14 00070 g002

2.5. Pre-Processing Step

In this step, various DICOM metadata, including the SeriesDescription attribute, the ProtocolName attribute, and many other attributes associated with MR acquisition parameters (such as RepetitionTime, EchoTime, and FlipAngle), were extracted. To reduce pre-processing step time, we extracted only information necessary for the rule-based labeling system and ML algorithm training instead of extracting all DICOM attributes (Table 2). The DICOM attributes are extracted in each series of an MRI scan. We additionally obtained the number of slices, not directly derived from the DICOM property, per series using the SeriesInstanceUID (0020,000 E). We used the Python Pydicom package.

2.6. Training Preparation Step

In this step, ImageSort-net prepares virtual labels derived from our rule-based labeling system, which work as labels for supervised learning. Moreover, it prepares various features associated with MRI acquisition parameters for the training of the ML algorithm. These features were selected by two MRI experts (K.W.K., S.C.J.). The normalization of selected DICOM attributes is performed to change the values of numeric columns in the dataset to a common scale without distorting differences in the ranges of values; this is helpful in providing features for ML algorithm training.

2.7. Training Step

The development dataset of ML algorithms was constructed except for blank data that the rule-based labeling system fails (by design) to infer from the dataset. And since the blank data were not used during learning, it is suitable for use as a test dataset. We evaluate and analyze two models learned with human expert labels (MLhuman) and rule-based virtual labeling (MLvirtual) using a blank dataset as a test set (Figure 3).
We chose a random forest classifier for this study, with the random forest algorithm still being the most commonly used method with proven robustness to overfitting. This classifier enables easy handling of missing values and is effective for large dataset processing owing to its fast training and testing speed. Therefore, it is suitable for the MRI header data processing that we are dealing with.
We used the scikit-learn application programming interface (API) for ML for the random forest classifier, and we set all parameters to default values, except the estimator. The estimator is a variable that determines the number of decision trees in a random forest and is an important parameter in relation to overfitting. To find the optimal hyperparameter, we generated 1 to 1000 estimators, conducted repeated experiments, and used 5-fold cross-validation for each training. In cross-validation, the development set is randomly divided into five groups. Each group is used as a validation set at least once for five iterations, and the remaining is used as a training set. Details of the ML learning and evaluation are shown in Figure 4.

2.8. Classification and Performance Evaluation

Using the hospital and multi-center trial datasets, the performances of the rule-based labeling, MLvirtual, and human-expert MLhuman algorithms in classifying MRI sequences were evaluated. For performance measurements, we used the human expert labeling of sequence types as a reference standard.
Classification performance on total sequences (i.e., the sum of all sequences) was evaluated using overall accuracy, and classification performance on each sequence type was evaluated using F1 scores, precision, and recall. The F1-score, precision, recall, and accuracy were defined as:
(a)
F1-score = 2 × (presicion × recall) ÷ (precision + Recall)
(b)
Precision = TP/(TP + FP)
(c)
Recall = TP/(TP + FN)
(d)
Overall accuracy = (TP + TN)/(TP + TN + FP + FN)

2.9. Feasibility Experiment for Sustainable Self-Learning MLvirtual Algorithm

The concept of a sustainable self-learning MLvirtual algorithm is to add new datasets to the preexisting datasets and train the ML algorithm again to increase its classification accuracy. Thus, we performed an experiment to add the multi-center trial dataset to the hospital dataset and re-trained the MLvirtual and MLhuman algorithms using the combined dataset. The training of MLvirtual and MLhuman with the multi-center trial dataset was performed for comparison. Subsequently, the re-training using the combined dataset (hospital dataset + multi-center trial dataset) was performed by following the previously outlined three steps: pre-processing, preparation, and training steps, as illustrated in Figure 2. In the re-training, there was no need for human labeling, given that MLvirtual is an automatic self-learning system. The performance of the ML algorithms was compared between the hospital dataset, the multi-center trial dataset, and the combined dataset (hospital dataset + multi-center trial dataset).

3. Results

3.1. Classification Performance of a Rule-Based Labeling System

Table 3 presents the MRI sequence type classification performance of the rule-based labeling system, which does not use ML. Each sequence classification performance on the hospital and multi-center trial datasets were generally good, with F1-scores ranging from 92.5% to 100% (median 99.0%). Only the Perfusion sequence type showed relatively low F1-scores (92.5%) with the multi-center dataset, while the other sequence types showed high accuracy (94.5–100%). The associated confusion matrix can be found in Supplementary Figure S1.
There were 2075 (12.9%) blank cases for the hospital dataset and 347 (11%) blank cases for the multi-center trial dataset, in which a sequence type cannot be inferred from a rule-based labeling system.

3.2. Classification Performance of the Initial ML Algorithm on the Hospital Dataset

Table 4 demonstrates the performance of the initial MLvirtual and MLhuman algorithms to classify MRI sequence types in the hospital dataset. In total sequence type classification, the overall accuracy of the MLvirtual algorithm was comparable to that of the MLhuman algorithm (98.5% and 99%, respectively). In each sequence type classification, the F1-score of MLvirtual for Scout was relatively low (50%) compared to other sequence types (ranging from 80% to 100%, respectively) (Table 5), which might be attributed to the substantial heterogeneity across vendors and institutions in the DICOM metadata as well as in acquisition parameters. Supplementary Figure S2 shows the confusion matrix associated with the performance of both ML models.

3.3. Classification Performance of Subsequent ML Algorithms with Additional Dataset

The classification performances of MLvirtual and MLhuman trained only on subsequent datasets (multi-center trial datasets) are presented in Table 4. The overall accuracy of the MLvirtual algorithm in total sequence type classification was relatively lower than that of the MLhuman algorithm (95.6% and 99.4%, respectively). In each sequence classification, the F1-score of MLvirtual for suscgre was relatively low (52.6%) compared to other sequence types (ranging from 97.3% to 100%, respectively) (Table 5). Supplementary Figure S3 shows the confusion matrix associated with the performance of both ML models.
After integrating the hospital dataset and multi-center trial dataset into a large, combined dataset, the re-trained MLvirtual model following its sustainable self-learning scheme showed better overall accuracy (99.7%) compared to that with the multi-center trial dataset alone (95.6%), as presented in Table 4. Furthermore, the inference performance on multi-center datasets of MLvirtual and MLhuman after re-training was the same. Supplementary Figure S4 shows the confusion matrix associated with the performance of both ML models. These results indicate that sustainable self-learning ML algorithms using rule-based virtual labeling in the new datasets are feasible.

3.4. Clinical Effectiveness of ImageSort-Net

We applied ImageSort-net in our clinical trial imaging core lab (Asan Image Metrics, www.aim-aicro.com (accessed on 20 November 2023)) in the form of a graphic user interface (GUI) software (v1.0.0) tool (Figure 5), which greatly saved time for image analysts to label MRI sequences. Indeed, the multi-center trial datasets (256 MRI scans with 3146 sequences) were manually labeled by an image analyst (M.H.K. with 15 years of experience in MRI research), which took seven full working days (8 h per day). On the contrary, the actual time spent by ImageSort-net for the overall classification of 256 MRI scans was 305 s for only inference and approximately 3 h for training.

4. Discussion

We created a rule-based labeling system using the metadata of DICOM image files and developed a sustainable self-supervised ML algorithm, named ImageSort-net, for automatic sequence-type classification of brain MRI using supervisory signals from DICOM metadata (i.e., rule-based virtual labeling).
Our rule-based labeling system achieves high performance with an overall accuracy of 99.5% as long as DICOM metadata is enough for classification. However, if the header is empty or the keyword is not in the rule, the rule-based labeling system cannot classify the sequence type and leaves the blanks, which account for more than 11% of the dataset. These blanks were inferred by the ML model in the ImageSort-net, and it achieved high classification performance with an overall accuracy of 98.6%.
Our motive for developing sustainable self-learning machine learning was to reduce the need for human resources and time in the classification of MRI sequences in multi-center clinical trials for acute stroke. Owing to the variability of DICOM metadata, it is very time-consuming to classify imaging sequences according to the standardized sequence types [16]. ImageSort-net can save time and cost by replacing labeling tasks in clinical trials and solving cases where manual labeling may vary depending on the experience of radiologists. In the MRI series of approximately 20,000 used in this study, the virtual labeling process takes 2 to 3 h, which is a task that takes weeks to months if carried out manually. Our proposed algorithm can replace the time-consuming data preparation process in image-based AI algorithm development.
Recently, research on the automatic classification of MRI series has been actively conducted, mainly based on images and text [17]. The image-based deep learning model classifies brain MRI series into eight classes with a high overall accuracy of 98.5% and uses a structure that sorts data according to predictive results. However, the data preparation process for learning early CNNs is essential, and manual labeling is required for model change [18]. For text-based ML models, construct a large dataset (700,000 MRI series) and develop it based on DICOM headers. Approximately 80% of the data sets were labeled based on the Series Description, while the remaining 20% were labeled after image verification. Model accuracy achieved high performance (97.4% to 99.96%).
Compared to these prior studies requiring manual labeling, we propose an MRI series classification system that does not require human interference for data extraction and allows learning by replacing labeling tasks with a rule-based labeling system, achieving performance comparable to models learned by manual labels. We developed ImageSort-net using approximately 20,000 MRI series, but its performance closely approximates that of models using 700,000 MRI series (98.6% and 99.6%, respectively).
In addition, ImageSort-net showed reliable performance by appending a new dataset to an existing dataset without human labeling of the combined dataset. Enhancing external validity is crucial to ensuring the robustness and adaptability of diverse imaging acquisition protocols across different manufacturers. We believe that external validity can be effectively achieved through a retraining process. Our findings, illustrated in Table 5, demonstrated consistent performance when incorporating a new dataset into an existing one without the need for human labeling of the combined dataset. These results underscore the capability of achieving robustness and adaptability to varied imaging conditions through retraining methods. Typically, retraining procedures can be laborious and challenging. However, our proposed algorithm significantly simplifies this process by eliminating the need for manual intervention. This innovation enables us to conduct retraining seamlessly, enhancing ease of implementation and reducing the associated complexities. Consequently, our result indicates that sustainable self-learning ML algorithms using rule-based virtual labeling in new datasets are feasible.
In addition to the automatic sequence type classification, it is also very important to provide detailed information such as the 2D/3D scheme, image plane, and parameter characteristics. For automatic sequence type classification, machine learning is very useful. For instance, we intend to categorize various kinds of MRA images, including raw images (mask) and reconstructed MIP or rotation images, as the “MRA” sequence type. Additionally, for diffusion-weighted images, our algorithm categorizes b0 images, b1000 images, and ADC images as the “DWI” sequence type. For the detailed information, we can extract it directly from the DICOM metadata (i.e., attributes) rather than using machine learning. For example, the DICOM attribute (0018,0023) includes the spatial data encoding scheme, which can be either 2D or 3D. The DICOM attribute (0020,0037) specifies the plane of acquisition for an image, either axial, sagittal, or coronal. In the DWI, detailed information such as the diffusion b-value can be extracted from the DICOM attribute (0018,9087).
Our study has several limitations. First, there were small, unbalanced datasets in the scout and perfusion, which may cause relatively low accuracy in the ML classification algorithm. However, if the dataset is accumulated and becomes large, this problem can be solved automatically. Indeed, we observed that the classification accuracy increased when adding the multi-center trial datasets, mainly because there were fewer incorrect answers due to learning from the combined dataset. Secondly, we trained our ImageSort-net using MRI data acquired from six MRI manufacturers, which may raise the issue of generalizability. However, these six manufacturers have more than 90% of the market share worldwide. In addition, our MRI datasets were acquired from a total of 25 hospitals, which enabled us to develop a generalizable MRI sequence classification algorithm.

5. Conclusions

In summary, the training of ML algorithms utilizing rule-based virtual labels has yielded high accuracy for sequence-type classification of brain MRI. This approach has empowered us to construct a robust, self-learning system with sustained efficacy. We believe that our innovative self-learning system plays a pivotal role in facilitating the seamless integration of changes and advancements in DICOM standards and the evolving landscape of MRI technologies. Future research endeavors should focus on further exploring the potential of developing AI algorithms centered on images leveraging ImageSort-net. Moreover, it remains imperative to investigate the feasibility of extending this methodology to encompass other modalities or anatomical structures.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/diagnostics14010070/s1, Figure S1: Confusion matrix for sequence type classification performance of rule-based labeling system. a hospital dataset. b multi-center dataset.; Figure S2: Confusion matrix for sequence type classification performance in hospital dataset. a MLvirtual b MLhuman. Figure S3: Confusion matrix for sequence type classification performance in multi-center dataset. a MLvirtual. b MLhuman. Figure S4: Multi-center test dataset classification results of models trained with combine dataset. a MLvirtual. b MLhuman. Table S1: Number of SeriesDescription and ProtocolName attributes per sequence type in the hospital dataset.

Author Contributions

K.W.K., K.Y. and S.N. designed this study. S.N. and Y.K. wrote the initial draft of this manuscript. Y.K., S.J.H., Y.S.S., S.C.J., C.J., B.S.K., M.-H.K. and Y.S. were involved in data collection. S.N. and K.Y. performed the statistical analysis. K.W.K., K.Y., S.N., Y.K., Y.S.S. and Y.S. contributed to the interpretation and revision of this article and jointly completed the final draft. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the National Research Foundation of Korea (NRF) grants funded by the Korea government (MSIT) (NRF-2020R1F1A1048267 and NRF-2020M3E5D2A01084578) and a grant of the Korea Health Technology R&D Project through the Korea Health Industry Development Institute (KHIDI), funded by the Ministry of Health & Welfare, Republic of Korea (HI18C1216).

Institutional Review Board Statement

Institutional review board approval was obtained (IRB No. 2022-0025).

Informed Consent Statement

The institutional review board waived the requirement for informed consent in this retrospective study.

Data Availability Statement

The datasets generated and analyzed during the current study are not publicly available due to the security of the data but are available from the corresponding author upon reasonable request.

Conflicts of Interest

Author Mi-Hyun Kim was employed by the company Trialinformatics Inc., author Chung Ju and Byung Su Kim were employed by the company Shin Poong Pharm. Co., Ltd. The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Sui, J.; Jiang, R.; Bustillo, J.; Calhoun, V. Neuroimaging-based Individualized Prediction of Cognition and Behavior for Mental Disorders and Health: Methods and Promises. Biol. Psychiatry 2020, 88, 818–828. [Google Scholar] [CrossRef] [PubMed]
  2. Lundervold, A.S.; Lundervold, A. An overview of deep learning in medical imaging focusing on MRI. Z. Med. Phys. 2019, 29, 102–127. [Google Scholar] [CrossRef] [PubMed]
  3. Prevedello, L.M.; Halabi, S.S.; Shih, G.; Wu, C.C.; Kohli, M.D.; Chokshi, F.H.; Erickson, B.J.; Kalpathy-Cramer, J.; Andriole, K.P.; Flanders, A.E. Challenges Related to Artificial Intelligence Research in Medical Imaging and the Importance of Image Analysis Competitions. Radiol. Artif. Intell. 2019, 1, e180031. [Google Scholar] [CrossRef] [PubMed]
  4. van Ooijen, P.M. Quality and curation of medical images and data. In Artificial Intelligence in Medical Imaging: Opportunities, Applications and Risks; Springer: Cham, Switzerland, 2019; pp. 247–255. [Google Scholar]
  5. Lambin, P.; Leijenaar, R.T.H.; Deist, T.M.; Peerlings, J.; de Jong, E.E.C.; van Timmeren, J.; Sanduleanu, S.; Larue, R.; Even, A.J.G.; Jochems, A.; et al. Radiomics: The bridge between medical imaging and personalized medicine. Nat. Rev. Clin. Oncol. 2017, 14, 749–762. [Google Scholar] [CrossRef] [PubMed]
  6. Gorgolewski, K.J.; Auer, T.; Calhoun, V.D.; Craddock, R.C.; Das, S.; Duff, E.P.; Flandin, G.; Ghosh, S.S.; Glatard, T.; Halchenko, Y.O.; et al. The brain imaging data structure, a format for organizing and describing outputs of neuroimaging experiments. Sci. Data 2016, 3, 160044. [Google Scholar] [CrossRef] [PubMed]
  7. Paszkiel, S. Applications of Brain-Computer Interfaces in Intelligent Technologies; Springer: Berlin/Heidelberg, Germany, 2022. [Google Scholar]
  8. Gauriau, R.; Bridge, C.; Chen, L.; Kitamura, F.; Tenenholtz, N.A.; Kirsch, J.E.; Andriole, K.P.; Michalski, M.H.; Bizzo, B.C. Using DICOM metadata for radiological image series categorization: A feasibility study on large clinical brain MRI datasets. J. Digit. Imaging 2020, 33, 747–762. [Google Scholar] [CrossRef] [PubMed]
  9. van der Voort, S.R.; Smits, M.; Klein, S. DeepDicomSort: An Automatic Sorting Algorithm for Brain Magnetic Resonance Imaging Data. Neuroinformatics 2021, 19, 159–184. [Google Scholar] [CrossRef] [PubMed]
  10. Spathis, D.; Perez-Pozuelo, I.; Marques-Fernandez, L.; Mascolo, C. Breaking away from labels: The promise of self-supervised machine learning in intelligent health. Patterns 2022, 3, 100410. [Google Scholar] [CrossRef] [PubMed]
  11. Alwassel, H.; Mahajan, D.; Korbar, B.; Torresani, L.; Ghanem, B.; Tran, D. Self-supervised learning by cross-modal audio-video clustering. Adv. Neural Inf. Process. Syst. 2020, 33, 9758–9770. [Google Scholar] [CrossRef]
  12. Mongan, J.; Moy, L.; Kahn, C.E., Jr. Checklist for artificial intelligence in medical imaging (CLAIM): A guide for authors and reviewers. Radiology. Artif. Intell. 2020, 2, e200029. [Google Scholar] [CrossRef] [PubMed]
  13. Kim, J.S.; Lee, K.B.; Park, J.H.; Sung, S.M.; Oh, K.; Kim, E.G.; Chang, D.I.; Hwang, Y.H.; Lee, E.J.; Kim, W.K.; et al. Safety and Efficacy of Otaplimastat in Patients with Acute Ischemic Stroke Requiring tPA (SAFE-TPA): A Multicenter, Randomized, Double-Blind, Placebo-Controlled Phase 2 Study. Ann. Neurol. 2020, 87, 233–245. [Google Scholar] [CrossRef] [PubMed]
  14. Long, J.; Sun, L.; Guo, L.; Hua, L.; Yang, Z. Label embedding semantic-guided hashing. Neurocomputing 2022, 477, 1–13. [Google Scholar] [CrossRef]
  15. van Wynsberghe, A. Sustainable AI: AI for sustainability and the sustainability of AI. AI Ethics 2021, 1, 213–218. [Google Scholar] [CrossRef]
  16. Na, S.; Sung, Y.S.; Ko, Y.; Shin, Y.; Lee, J.; Ha, J.; Ham, S.J.; Yoon, K.; Kim, K.W. Development and validation of an ensemble artificial intelligence model for comprehensive imaging quality check to classify body parts and contrast enhancement. BMC Med. Imaging 2022, 22, 87. [Google Scholar] [CrossRef] [PubMed]
  17. Cluceru, J.; Lupo, J.M.; Interian, Y.; Bove, R.; Crane, J.C. Improving the Automatic Classification of Brain MRI Acquisition Contrast with Machine Learning. J. Digit. Imaging 2022, 36, 289–305. [Google Scholar] [CrossRef] [PubMed]
  18. Sugimori, H. Classification of Computed Tomography Images in Different Slice Positions Using Deep Learning. J. Healthc. Eng. 2018, 2018, 1753480. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Rule-based labeling system overview for generating virtual labels.
Figure 1. Rule-based labeling system overview for generating virtual labels.
Diagnostics 14 00070 g001
Figure 3. Development and test set details for the hospital and multi-center trial datasets.
Figure 3. Development and test set details for the hospital and multi-center trial datasets.
Diagnostics 14 00070 g003
Figure 4. Process of training and evaluating random forest classification algorithms with hospital datasets.
Figure 4. Process of training and evaluating random forest classification algorithms with hospital datasets.
Diagnostics 14 00070 g004
Figure 5. ImageSort-net GUI software. (a). ImageSort-net GUI software. (b). Sample of ImageSort-net results for one MRI scan. (c). Incomplete header data were accurately inferred from ImageSort-net.
Figure 5. ImageSort-net GUI software. (a). ImageSort-net GUI software. (b). Sample of ImageSort-net results for one MRI scan. (c). Incomplete header data were accurately inferred from ImageSort-net.
Diagnostics 14 00070 g005
Table 1. Number of patients and MRI scans in datasets.
Table 1. Number of patients and MRI scans in datasets.
DatasetCentersPatientsMRI ScansManufacturers
Hospital dataset *115281531Philips (933), Siemens (507), GE (67), Hitachi (10), Toshiba (12), Medinus (2)
Multi-center trial dataset859256Philips (57), Siemens (110), GE (87)
* Including 16 outside hospitals with 356 patients and 357 MRI scans.
Table 2. Extracted DICOM attribute in the preprocessing step.
Table 2. Extracted DICOM attribute in the preprocessing step.
DICOM AttributeDICOM Standard Attribute Description
Rule-based
system
SeriesDescriptionDescription of the Series
ProtocolNameUser-defined description of the conditions under which the Series was performed.
Machine
Learning
ScanningSequenceDescription of the type of data taken.
SequenceVariantVariant of the scanning sequence.
MRAcquisitionTypeIdentification of a spatial data encoding scheme.
ImageTypeImage identification characteristics.
RepetitionTimeTime in ms between the beginning of a pulse sequence and the beginning of the succeeding (essentially identical) pulse sequence.
EchoTimeTime in ms between the middle of the excitation pulse and the peak of the echo produced
FlipAngleSteady state angle in degrees to which the magnetic vector is flipped from the magnetic vector of the primary field.
ImagingFrequencyPrecession frequency in MHz of the nucleus being addressed
NumberOfPhaseEncodingStepsTotal number of lines in k-space in the ‘y’ direction collected during acquisition.
RowsNumber of rows in the image.
ColumnsNumber of columns in the image.
InversionTimeTime in msec after the middle of inverting the RF pulse to the middle of the excitation pulse to detect the amount of longitudinal magnetization.
SliceThicknessNominal slice thickness, in mm.
ScanOptionParameters of the scanning sequence.
BodyPartExaminedText description of the part of the body examined.
ScliceNum *The number of slices per series instance
* Not directly derived from the DICOM attribute.
Table 3. Sequence type classification performance of a rule-based labeling system.
Table 3. Sequence type classification performance of a rule-based labeling system.
Sequence TypeHospital DatasetMulti-Center Trial Dataset
PrecisionRecallF1-ScoreNo of Blanks *PrecisionRecallF1-ScoreNo of Blanks *
T199.6%99.5%99.6%329/407199.7%95.1%97.3%10/420
T299.2%99.6%99.4%14/1496100%99.3%99.6%0/146
FLAIR99.8%99.4%99.6%3/142299.2%100%99.6%0/266
suscgre99.7%99.6%99.6%83/175299.7%96.3%98%23/354
MRA99.5%99.6%99.6%962/376397.2%99.4%98.3%258/1100
Scout96.5%92.6%94.5%13/30898.4%100%99.2%0/127
Perfusion92.7%100%96.2%4/5586%100%92.5%21/64
Diffusion99.6%100%99.8%667/323999.7%99.8%99.8%35/669
Total 98.3%98.8%98.5%2075/16,10697.5%98.7%98%347/3146
* The number of blanks indicates the number of series that the rule-based labeling system failed to infer (Number of blank/number of series type); Total refers to the sum of all sequence types.
Table 4. Classification performance of MLvirtual and MLhuman in the initial dataset (hospital dataset), subsequent dataset (multi-center trial dataset), and combined dataset (hospital + multi-center dataset).
Table 4. Classification performance of MLvirtual and MLhuman in the initial dataset (hospital dataset), subsequent dataset (multi-center trial dataset), and combined dataset (hospital + multi-center dataset).
PerformancesMLvirtualMLhuman
Only Hospital DatasetOnly Multi-Center Trial DatasetCombined Dataset *Only Hospital DatasetOnly Multi-Center Trial DatasetCombined Dataset *
Overall accuracy98.5%95.6%99.7%99%99.4%99.7%
Precision88.6%86.8%99%89.7%98.2%99%
Recall86.1%98.4%99.9%92.8%99.8%99.9%
F1-score86.3%90%99.5%90%99%99.5%
* Multi-center test dataset results.
Table 5. Performance of each sequence classification of MLvirtual and MLhuman on each dataset.
Table 5. Performance of each sequence classification of MLvirtual and MLhuman on each dataset.
MLvirtualMLhuman
PrecisionRecallF1-ScorePrecisionRecallF1-Score
Initial dataset (hospital dataset only)
Diffusion100%100%100%100%100%100%
FLAIR66.6%100%80%66.6%100%80%
MRA98.4%99.2%98.8%99.1%99.6%99.4%
Perfusion100%66.6%80%100%100%100%
Scout46%54.5%50%54%70%60%
suscgre100%92.2%96%100%91.2%95.4%
T197.8%98.4%98.1%98.4%99%98.7%
T2100%77.8%87.5%100%82.3%90%
Subsequently added dataset (multi-center trial dataset only)
Diffusion100%97.2%98.6%100%100%100%
FLAIR------
MRA100%94.5%97.3%100%99.2%99.6%
Perfusion95.2%100%97.5%95.2%100%97.5%
Scout------
suscgre40%100%56.2%95.6%100%97.7%
T1100%100%100%100%100%100%
T2------
Combined dataset * (hospital dataset + multi-center trial dataset)
Diffusion100%100%100%100%100%100%
FLAIR------
MRA100%99.6%99.8%100%99.6%99.8%
Perfusion95.2%100%97.5%95.2%100%97.5%
Scout------
suscgre100%100%100%100%100%100%
T1100%100%100%100%100%100%
T2------
* Multi-center test dataset results.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Na, S.; Ko, Y.; Ham, S.J.; Sung, Y.S.; Kim, M.-H.; Shin, Y.; Jung, S.C.; Ju, C.; Kim, B.S.; Yoon, K.; et al. Sequence-Type Classification of Brain MRI for Acute Stroke Using a Self-Supervised Machine Learning Algorithm. Diagnostics 2024, 14, 70. https://doi.org/10.3390/diagnostics14010070

AMA Style

Na S, Ko Y, Ham SJ, Sung YS, Kim M-H, Shin Y, Jung SC, Ju C, Kim BS, Yoon K, et al. Sequence-Type Classification of Brain MRI for Acute Stroke Using a Self-Supervised Machine Learning Algorithm. Diagnostics. 2024; 14(1):70. https://doi.org/10.3390/diagnostics14010070

Chicago/Turabian Style

Na, Seongwon, Yousun Ko, Su Jung Ham, Yu Sub Sung, Mi-Hyun Kim, Youngbin Shin, Seung Chai Jung, Chung Ju, Byung Su Kim, Kyoungro Yoon, and et al. 2024. "Sequence-Type Classification of Brain MRI for Acute Stroke Using a Self-Supervised Machine Learning Algorithm" Diagnostics 14, no. 1: 70. https://doi.org/10.3390/diagnostics14010070

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop