Next Article in Journal
Nuclei Detection and Segmentation of Histopathological Images Using a Feature Pyramidal Network Variant of a Mask R-CNN
Previous Article in Journal
Therapeutic Efficacy of Adipose Tissue-Derived Components in Neuropathic Pain: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Detection and Classification Framework for Diagnosing of Cerebral Microbleeds Using Transformer and Language

1
School of Clinical Medicine, College of Medicine, Nanjing Medical University, Nanjing 211166, China
2
Department of Neurology, The First Affiliated Hospital of Nanjing Medical University, Nanjing 210029, China
3
Department of Computer Science and Technology, Shanghai University, 99 Shangda Road, Baoshan District, Shanghai 200444, China
4
Department of Neurology, Nanjing Drum Tower Hospital, Affiliated Hospital of Medical School, Nanjing University, 321 Zhongshan Road, Nanjing 210008, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Bioengineering 2024, 11(10), 993; https://doi.org/10.3390/bioengineering11100993
Submission received: 21 August 2024 / Revised: 24 September 2024 / Accepted: 27 September 2024 / Published: 30 September 2024
(This article belongs to the Section Biosignal Processing)

Abstract

:
The detection of Cerebral Microbleeds (CMBs) is crucial for diagnosing cerebral small vessel disease. However, due to the small size and subtle appearance of CMBs in susceptibility-weighted imaging (SWI), manual detection is both time-consuming and labor-intensive. Meanwhile, the presence of similar-looking features in SWI images demands significant expertise from clinicians, further complicating this process. Recently, there has been a significant advancement in automated detection of CMBs using a Convolutional Neural Network (CNN) structure, aiming at enhancing diagnostic efficiency for neurologists. However, existing methods still show discrepancies when compared to the actual clinical diagnostic process. To bridge this gap, we introduce a novel multimodal detection and classification framework for CMBs’ diagnosis, termed MM-UniCMBs. This framework includes a light-weight detection model and a multi-modal classification network. Specifically, we proposed a new CMBs detection network, CMBs-YOLO, designed to capture the salient features of CMBs in SWI images. Additionally, we design an innovative language–vision classification network, CMBsFormer (CF), which integrates patient textual descriptions—such as gender, age, and medical history—with image data. The MM-UniCMBs framework is designed to closely align with the diagnostic workflow of clinicians, offering greater interpretability and flexibility compared to existing methods. Extensive experimental results show that MM-UniCMBs achieves a sensitivity of 94% in CMBs’ classification and can process a patient’s data within 5 s.

Graphical Abstract

1. Introduction

Cerebral Micro Bleeds (CMBs) are diminutive hemosiderin deposits, a consequence of cerebral microvascular lesions in the brain parenchyma. These lesions, which appear as small, round, homogeneous, and hypointense areas, are detectable through Magnetic Resonance Imaging (MRI), particularly using T2-gradient-recalled echo (T2*-GRE) and susceptibility-weighted imaging (SWI) [1]. The detection of CMBs is a critical step in diagnosing and managing cerebral small-vessel disease (SVD) and is integral to identifying severe neurological conditions such as ischemic stroke, traumatic brain injury, and Alzheimer’s disease [2]. The integration of deep learning with disease diagnosis has become a major focus in medical image recognition tasks. Recently, due to the high costs of resource allocation and sample annotation in deep learning methodologies, researchers have explored Spiking Neural Networks (SNNs) [3,4] for medical applications. SNNs bridge the gap between neuroscience and machine learning by employing biologically inspired neuron models for computational tasks. For instance, Turkson [5] developed an innovative SNN pipeline to classify Alzheimer’s disease (AD) into three binary tasks: AD versus healthy controls (NC), AD versus Mild Cognitive Impairment (MCI), and NC versus MCI, using MRI scans. Additionally, ref. [6] proposed an optimized tumor segmentation approach that integrates a deep spiking neural network (DSNN) with a conditional random field structure to improve segmentation accuracy. To mitigate overfitting in classification models, a deep convolutional spiking neural network (DCSNN) [7] has been introduced for lung disease detection using chest X-ray images. These advancements underscore the growing interest in using deep learning techniques for disease diagnosis.
Due to the small sizes of CMBs in MR images, typically ranging from 2 mm to 10 mm in diameter, CMBs are susceptible to interference from other similar targets, such as calcifications and veins. As a result, accurately identifying genuine CMBs in SWI images poses a significant challenge for clinical practitioners. To address this issue, various automatic detection methods [8] have been proposed to assist clinical physicians in detecting CMBs. These methodologies can be classified into one-stage [9] and two-stage approaches [10]. One-stage methods offer higher integration and faster inference speeds, whereas two-stage approaches can provide superior performance by separating detection from classification tasks. Given the stringent accuracy requirements in CMBs’ detection and classification, current research has focused on the two-stage architecture. In this framework, CMBs’ candidate regions, which include both true CMBs and their mimics, are identified in the detection phase and then classified in the subsequent classification stage [11]. The two-stage pipeline has proven effective in reducing false positives. However, it still presents several unique challenges: (1) balancing performance with speed; (2) reducing annotation costs for training samples; (3) accurately detecting the small and weakly featured CMBs; and (4) distinguishing genuine CMBs from similar-looking targets.
In this paper, we present a decoupled two-stage framework for CMBs’ detection and classification, named MM-UniCMBs, which employs a front-light (detection stage) and back-heavy (classification stage) design. The detection stage is optimized for speed, focusing on rapid processing, while the classification stage is designed to prioritize high accuracy. This decoupled pipeline effectively balances detection speed with classification accuracy, addressing the first challenge. During the labeling process, we observed that classification requires less labeling effort compared to object detection, largely due to the labor-intensive nature of bounding box annotations in the latter. Leveraging this observation, we employ distinct labeling strategies for each phase. In the detection stage, preliminary assessments are made to identify images with potential CMB regions, which does not demand extensive clinical expertise and can involve less experienced intern doctors. In contrast, the classification stage requires clinicians with substantial expertise to make definitive judgments on whether the identified regions contain true CMBs. This bifurcated approach reduces the need for expert involvement in the initial labeling phase, effectively addressing the challenge of annotation costs.
In response to the third challenge of detecting small and weakly featured CMBs, we conducted an analysis of CMB features in SWI images and developed a lightweight 2D Convolutional Neural Network (CNN)-based detection model, named CMBs-YOLO. To strengthen the weak features of CMBs, a novel CMBs-Multi Head Attention (CMBs-MHA) module is proposed, specifically designed to accentuate the distinct disparities between CMB targets and their surrounding background. Simultaneously, to bolster the robustness of modeling small features, we incorporated the Normalized Wasserstein Distance (NWD) [12] method as a supplementary loss, complementing the conventional Intersection over Union (IoU)-based [13] approach.
Faced with the final challenge, we developed an innovative multimodal classification network, named CMBsFormer (CF). This network stands out by integrating text language descriptions with a transformer-based architecture, incorporating the clinical knowledge into the classification modeling process. Our CF framework encapsulates two critical aspects of clinical diagnosis: looking (patient’s MRI images) and asking (patient’s age, gender, past medical history, lifestyle habits, etc.), thus aligning the modeling process more closely with the diagnostic procedures used by clinicians. Meanwhile, the adoption of the transformer structure can achieve an advancement compared to the conventional CNN-based methods, as this structure excels in feature extraction and demonstrates superior proficiency in integrating textual embedding information. Additionally, the proposed CF framework exhibits strong adaptability and flexibility across various configurations.
In summary, our contributions can be summarized as follows:
(1)
We introduce a decoupled two-stage framework named MM-UniCMBs, characterized by the front-light and back-heavy architecture;
(2)
We propose a novel lightweight CMBs detection network called CMBs-YOLO, which incorporates the CMBs-MHA module and NWD loss;
(3)
An innovative multimodal classification network named CMBsFormer is proposed by integrating text language descriptions with a transformer-based architecture;
(4)
The extensive experimental results demonstrate that our proposed framework can achieve satisfactory performance in both the CMB detection and classification tasks.
The subsequent chapters in this paper are organized as follows. Section 2 provides an overall review of CMBs’ detection and classification. In Section 3, we present a description of our CMBs dataset, along with the details pertaining to our detection and classification models. The comparative experimental results and discussions are presented in Section 4. In Section 5, we provide a summary and ablation studies for the field of automated detection and classification of CMBs. The discussion and conclusion are present in Section 6. Finally, the detailed experimental results of our method are presented in the Appendix A, Appendix B and Appendix C.

2. Related Work

In the field of CMB detection and classification, early studies relied on manually predefined features such as intensity thresholds, area size, and other traditional image processing techniques [14,15,16,17,18,19,20]. Dou et al. [21] modeled the detection of CMBs through a three-stage process: initial candidates screening based on intensity values, extraction of compact 3D hierarchical features via a stacked convolutional Independent Subspace Analysis (ISA) network, and elimination of false positives using a Support Vector Machine (SVM) classifier, which leveraged the representation features learned from ISA. Another contribution in [10] employed three distinct machine learning classifiers, SVM, Quadratic Discriminant Analysis (QDA), and an ensemble classifier, to detect and classify CMBs. Fazlollahi et al. [22] introduced a multi-scale Laplacian of Gaussian technique. This technique was designed to detect the potential CMB candidates, along with their corresponding bounding boxes, followed by a cascade of binary random forests to refine detection accuracy. Bian et al. [23] utilized the 2D fast radial symmetry transform for initial detection of candidate CMBs and subsequently applied 3D region growing to eliminate false positives. In another study, Fazlollahi et al. [24] implemented a cascade of random forest classifiers trained on robust radon-based features, particularly addressing the challenge posed by unbalanced sample distribution. Morrison et al. [25] incorporated a 2D fast radial symmetry transform and handcrafted features to automatically detect CMBs with high sensitivity. While these early methodologies laid the groundwork for the evolution of CMB detection and classification, they struggled to achieve satisfactory performance due to the limited feature-learning capabilities and inadequate representational power of manually extracted features.
The advent of deep learning algorithms has significantly revolutionized the field of CMB detection and classification [26,27,28,29,30,31,32,33]. In [34], the researchers employed a fully automated two-stage integrated deep learning strategy by combing a regional-based method called You Only Look Once (YOLO) for initial CMB candidate detection, followed by a 3D-CNN classification stage to reduce false positives. Fang et al. [35] focused on generating heatmaps of CMBs by proposing a 2.5D CNN model to establish a priori knowledge base, addressing the redundant computation in 3D-CNN and the loss of spatial information in 2D-CNN. A novel CMB classification method was proposed in [36], which integrated a deep CNN model with randomized neural networks, a Random Vector Functional Link network (RVFL), and an Extreme Learning Machine (ELM). This combination aimed to balance the image feature learning capabilities with the training speed. Koschmieder et al. [37] developed distinct approaches for the detection of CMBs in traumatic brain injury (TBI) cases, including a classification method (Patch-CNN) and two segmentation methods (U-Net). Ferlin et al. [9] proposed a machine learning approach based on a 2D Faster R-CNN network, achieving a balance between sensitivity and precision. Stanley et al. [38] introduced a novel feature extraction technique for CMB detection, using a CNN extractor combined with Long Short-Term Memory (LSTM) to classify selected Regions of Interest (ROI) as CMBs or non-CMBs. A 3D patch-based deep residual network was implemented in [39,40], specifically designed to differentiate true CMBs from their mimics in SWI images. Compared to traditional machine learning methods, the CNN-based approaches have achieved a significant performance improvement in both detection and classification. However, these methods still have limitations due to their sole reliance on image information. This one-dimensional focus creates a substantial gap between these methods and the diagnostic processes employed by clinical doctors, resulting in a bottleneck in current research.
The emergence of multimodal methods in the medical domain, particularly language–vision models, has demonstrated a remarkable capability to extract and integrate representations from both images and text descriptions [41]. One notable example is the Language meets Vision Transformer (LViT) [42], a language–vision medical image segmentation architecture that utilizes a hybrid CNN–Transformer structure for lung organ segmentation. Liu et al. [43] leveraged the Contrastive Language–Image Pretraining (CLIP) [44] method to effectively integrate text embeddings with medical image segmentation models. Furthermore, Bhalodia et al. [45] introduced a novel pneumonia detection technique based on attention mechanisms, utilizing text information extracted from medical reports. These language–vision methods have attained considerable success across various medical application fields. However, to the best of our knowledge, these approaches have not been employed in the domain of CMBs’ detection or classification. In this article, we will explore the applicability of multimodal methods in the task of CMB classification, focusing on the integration of MRI images with patient descriptors.

3. Materials and Methods

3.1. Data Acquisition

In this study, we collected a dataset of CMBs involving 600 patients from Nanjing Medical University’s Affiliated Jiangsu Province Hospital. The data collection spanned from 1 January 2022 to 1 October 2023. All MRI scans were acquired on the 3.0T GE scanner (Discovery MR750, GE Medical Systems, Milwaukee, WI, USA). Each patient’s data included 130 SWI slices and 130 phase slices, all rendered at a resolution of 768 × 768 pixels. Alongside these images, we incorporated the patient information such as age, gender, past medical history, and lifestyle habits. It is important to note that the inclusion of past medical history and lifestyle habits was applicable only to a limited subset of patients. The diagnostic evaluations of all patients were reviewed by clinical physicians. The identification and annotation of CMB targets in these patients were conducted by neurologists with 15 years of clinical experience. In the case of labeling discrepancies, another senior neuroradiologist was consulted, with the final determination made based on a consensus among these experts. To maintain consistency in analysis, the grayscale values of each SWI and phase image were normalized to a range of 0–255 and then converted into three-channel format. Additionally, we incorporated a supplementary public dataset of CMBs from the VALDO 2021 challenge [46]. For clarity, we refer to these datasets as CMBs-Private and CMBs-Public, respectively. Table 1 and Table 2 provide detailed information on the data partitioning and distribution for training and testing purposes.

3.2. Methods

In this section, we will discuss the implementation details of our MM-UniCMBs. The structure of MM-UniCMBs is illustrated in Figure 1. During the initial detection stage, only SWI images are processed to identify potential candidate regions of CMBs. Subsequently, the classification model takes SWI data and text descriptions as inputs to eliminate false positives.

3.2.1. CMBs-YOLO

At present, prevalent object detection architectures can be classified into CNN-based methods [47] and transformer-based methods [48]. Compared to the transformer structures, CNNs can offer a more favorable balance between performance and speed. An excellent example of CNN methods is Ultralytics YOLOv8, a universal state-of-the-art (SOTA) model in object detection tasks, which builds upon the achievements of earlier YOLO iterations [49]. YOLOv8, recognized for its swiftness and high performance, is an ideal choice for our CMB detection task. However, despite its strengths, the standard YOLOv8 structure still has limitations, particularly in terms of feature modeling capability for weak features and the design of its loss function for small targets. To address these issues, we proposed a novel CMB detection network, termed CMBs-YOLO, as illustrated in Figure 2. By incorporating CMBs-MHA modules in both the feature extraction and fusion stages, the feature learning ability for weak features can be enhanced. Regarding to bounding box regression, we have refined the existing Complete Intersection over Union (CIoU) loss with an additional NWD loss, aiming to augment the stability of the detection network, particularly for small targets.
Regions containing CMB targets exhibit distinguishable differences from the surrounding background areas in SWI. However, this critical characteristic has been overlooked in existing research on the detection of CMBs, leading to a decline in performance. As depicted in Figure 3, the standard YOLOv8 structure struggles to model this difference, leading to missed detections. We should emphasize that incorporating Multi-Head Recurrent Layer Attention (MRLA) [50] in the encoding stage can enhance the interaction ability of layers in the CNN structure, thereby intensifying the network’s attention to diverse features. Integrating MRLA into the YOLOv8 framework, as demonstrated in Figure 3, has indeed improved the detection of CMB targets. However, it is also important to note that the original structure of MRLA was not specifically tailored to accentuate the distinct features of CMB targets. As a result, although the MRLA module brings an improvement in the detection of CMBs, its capability in modeling saliency features remains limited.
To enhance the representation capabilities across various encoding layers and amplify the distinct responses of CMB targets against the background, we introduce the CMBs-MHA module; its structure is shown in Figure 4. The key innovation of CMBs-MHA lies in its utilization of the global max pooling operation to generate query vectors instead of global average pooling, which can help in extracting salient regions from the entire feature map. Meanwhile, it employs a global average pooling operation to create key-value vectors, constructing a background board to model local features. Subsequently, a multi-head attention block is applied on these queries and key vectors to build the saliency relationships among CMBs’ features and their surrounding background. As illustrated in Figure 3, the incorporation of the CMBs-MHA module has significantly enhanced the responses of CMB features on the feature map. Additionally, our CMBs-MHA module is designed for seamless integration as a plugin into any CNN-based detection framework.
Another frequently overlooked aspect in the CMBs detection task relates to the design of the bounding box regression. The small size of CMB targets, typically ranging from 2 mm to 10 mm, corresponds to an occupation of 4–32 pixels on each SWI image. However, the prevalent design of box regression loss, which is based on the IoU theory, has been found to be less efficient for these tiny targets, as indicated in [12]. To illustrate this point, we provide a visual representation of the relationship between the predicted bounding boxes and ground truth boxes in Figure 5, including both small and standard sizes. In this context, A and D denote the ground truth boxes, while B, C, E, and F represent the predicted boxes, with each box being separated by one pixel. This visual representation can help us to understand the challenges associated with solely relying on IoU-based loss functions for the detection of CMB targets.
In Table 3, we present a comparative analysis using different loss functions, focusing on two distinct target sizes: small and standard. Two key issues arise when applying IoU-based methods for the detection of CMBs. Firstly, even a one-pixel discrepancy, as observed between (A, B) and (D, E), leads to a substantial variation in calculated IoU values, with differences reaching up to threefold. Secondly, a size change in a small-sized CMB target from B to C results in an IoU reduction of 0.23, while a similar size change in a normal-sized CMB target from E to F only alters the IoU by 0.07. These observations underscore the limitations of IoU-based methods in processing the small targets. The significant fluctuation in IoU values due to minor positional changes, particularly for small-sized CMB targets, highlights the necessity of introducing an additional loss in the detection of CMBs. To this end, we have integrated the NWD loss into our CMBs-YOLO network.
Given a predicted bounding box A = ( P x 1 , P y 1 , P x 2 , P y 2 ) and a ground truth box B = ( G x 1 , G y 1 , G x 2 , G y 2 ) , the NWD loss can be computed as follows:
W D A , B = e r d i s ( A , B ) + w d i s A , B C
r d i s A , B = P x 1 + P x 2 2 G x 1 + G x 2 2 × P y 1 + P y 2 2 G y 1 + G y 2 2 + δ
w d i s A , B = ( P x 2 P x 1 + δ G x 2 G x 1 + δ ) 2 + ( P y 2 P y 1 + δ G y 2 G y 1 + δ ) 2 4
where C and δ are constants with values set to 12.8 and 1 × 10−7 in our study, respectively. Examination of the results in Table 3 reveals that the NWD can enhance the robustness of small targets in comparison to IoU-based methods.
In the detection stage, we addressed the performance bottlenecks in the detection of CMBs by focusing on two aspects: feature modeling and loss function design. The integration of the CMBs-MHA module during the encoding stage enhances the network’s ability to capture salient responses of CMB targets against the background. Additionally, the implementation of the NWD is essential in improving the training stability of the network, particularly for small targets. The architecture of our CMBs-YOLO is designed to enhance the adaptability and effectiveness of the YOLO framework for the specific demands of detecting CMBs.

3.2.2. CMBsFormer

Following the detection phase, numerous candidate CMB regions can be identified, however, containing a large number of false positives. Unlike other typical image classification tasks, the classification process for CMBs requires integrating multiple slice images. Two primary approaches have been developed for the design of CMB classification networks based on the processing method of input images. The first approach entails processing multiple cropped slice images, which are either stitched together or averaged, to convert them into 2D input formats. Following this transformation, 2D convolutional networks are employed to classify true CMB targets [34]. This approach struggles to achieve satisfactory classification results due to the loss of spatial information in adjacent slice images. To address this, contemporary CMB classification approaches emphasize the use of 3D-CNN structures with 3D image data to enhance the network’s modeling ability. Additionally, some researchers have integrated manually designed features, such as texture or geometric information, to expand the feature set of the classification network [52]. Nevertheless, the above methods still face two notable challenges in classifying CMBs. Firstly, CNN-based methods have limited capacity to differentiate between CMBs and their mimics. Secondly, manually engineered features, which are typically specific to certain datasets, exhibit poor interpretability and flexibility. Meanwhile, the poorly designed manual features can degrade network performance.
In this paper, we explore the inter-relationship between patient’s textual information and SWI data, presenting a novel multimodal CMB classification network. The architecture of CF is visually depicted in Figure 6. The CF architecture consists of four components: the image branch, the text branch, the image–text fusion module, and the transformer classification module. The architecture is designed to accommodate both multimodal (image and text) and single-modal (image) inputs, offering flexibility across different configurations. For clarity, we focus on demonstrating the multimodal input configuration.
In our CF framework, the image branch is designed to extract feature information from 3D SWI and phase images. Assuming the input images are I = { S i 1 ,   S i , S i + 1 , P i } R B × C × T × H × W , the process of feature extraction is articulated through the following equations:
I c o n v _ 1 = G E L U ( B N 3 D ( C o n v 3 D ( I ) ) )
I c o n v _ 2 = B N 3 D ( C o n v 3 D ( I c o n v _ 1 ) )
I p o s = I c o n v _ 2 + p o s _ e m b e d d i n g ( I c o n v _ 2 )
I e = r e s h a p e ( I p o s ) R B × ( C × T ) × ( H 4 × W 4 )
where S i and P i represent the SWI and phase images. Conv3D, BN3D, and GELU refer to 3D convolutional, 3D batch normalization, and the GELU activation functions, respectively. B denotes the batch size, and C, T, H, and W represent the embedding channels, input image slices, height, and width, correspondingly. Through two convolution operations, the width and height of the input images are reduced from 64 to 16, and the number of channels is increased from 3 to 512. Position embedding data are added after the convolution operations to model the positional information of different image patches. Finally, the 3D image data are reshaped into a 1D structure by amalgamating channels and slices as well as width and height, obtaining the image branch output I e R B × ( C × T ) × ( H 4 × W 4 ) .
By analyzing the clinical diagnostic process for CMBs, we observed that their occurrence is closely associated with several factors, including a patient’s age, gender, and lifestyle habits. Additionally, the patient’s medical history—encompassing conditions such as Cerebral Autosomal Dominant Arteriopathy with Sub-cortical Infarcts and Leukoencephalopathy (CADASIL), diabetes, hypertension, and coronary heart disease—plays a critical role in diagnosing CMBs. Currently, no methods are available that incorporate such detailed patient information, leading to a significant gap between model-based processing and clinical diagnostic procedures. In our research, we treat the process of inquiring about patient’s information as a method for modeling patient text data. Of course, due to privacy concerns and the sensitive nature of information related to patients’ lifestyle habits and medical histories, these elements were excluded from our analysis. Nevertheless, we will discuss these aspects in some specific cases in the experimental section. Table 4 presents the structure of our text prompts and their corresponding descriptions, illustrating how these elements were integrated into our research.
Assume the text descriptions in our study are represented as T = { T 1 ,   T 2 , T n } , where T can be replaced or extended to other text statements, such as the patient’s lifestyle habits and past medical history. To process the textual data, we utilize a pretraining model named Bidirectional Encoder Representations from Transformers (BERT) [53]. This model can transform each individual text statement into a 768-dimensional word vector. Subsequently, these individual word vectors are then aggregated to form an n-dimensional tensor. Following this, a permute operation is applied to rearrange this stacked tensor, obtaining the text branch output T e R B × 768 × n .
We execute the fusion of image–text features and the classification of CMBs through the following steps. In the first step, both I e and T e are transformed into 1D data structures and then concatenated together along the channel dimension. A subsequent 1D convolution operation is performed, yielding the image–text fusion features denoted as I _ T e , which will be converted into a data structure of the form (B, C, T, H, W). The second step involves the use of 3D convolution to further process the fusion information derived from both image and text dimensions. This step not only models the combined information but also integrates classification and positional encoding vectors for the following classification stage. The third step involves flattening the 3D tensor into a 1D format, a procedure for compatibility with the following transformer structure. Finally, a multi-head transformer block is utilized to extract features and classify CMBs. In scenarios where the model is provided solely with image inputs, our CF framework can adapt accordingly. It circumvents the feature fusion process, and instead, the results of the 3D convolution are directly channeled into the transformer module for the final classification stage.
By integrating 3D image data and text information into a unified network, our CF framework addresses several key challenges. Firstly, by replacing the traditional 3D-CNN structure with a transformer architecture, CF tackles the problem of feature extraction capability for both CMBs and their mimics. This architectural shift can enhance the network’s ability to discern and extract features more accurately. Secondly, CF incorporates text encoding as an alternative approach to manually designed features, allowing the model to leverage the specialized knowledge of clinical experts to enrich its information base while also improving interpretability. Moreover, the text prompts used in our study are designed to be dataset independent, offering flexibility for incorporating other relevant prompts. Finally, the text embedding structure demonstrates superior compatibility with the transformer architecture compared to CNN-based approaches.

4. Results

4.1. Evaluation Metrics

In this paper, we introduced a two-stage model for the detection and classification of CMBs. To evaluate the model’s performance, we employ different evaluation metrics at each stage. In the detection phase, we utilize precision (P), recall (R), and mean Average Precision (mAP) [48] as evaluation metrics, all based on the IoU method. In the classification stage, we introduced additional evaluation metrics, including sensitivity (also known as the true-positive rate, TPR) [54], specificity (also known as the true-negative rate, TNR) [54], and precision (P) [54]. To assess the model’s accuracy at the patient level, we incorporated FPavg [54] and P p a t i e n t as supplementary evaluation metrics. Here, FPavg represents the average number of false-positive predictions per patient, providing insights into the model’s tendency towards over-diagnosis. P p a t i e n t measures the model’s overall precision at the patient level, reflecting its performance in two key aspects: (1) error-judgment, where the model erroneously identifies a patient as having CMBs when they do not, and (2) missing-judgment, where the model fails to detect the presence of CMBs in a patient who has them. The calculations for these metrics in the classification stage are elaborated upon as follows:
S e n s i t i v i t y = T P T P + F N
S p e c i f i c i t y = T N T N + F P
P r e c i s i o n = T P T P + F P
F P a v g = F P n
P p a t i e n t = i = 0 n p a t i e n t i = 1 , i f j = 0 S T P j > 0   a n d   i f   p a t i e n t i   h a s     C M B s 0 , o t h e r w i s e + i = 0 n p a t i e n t i = 1 , i f j = 0 S T P j = 0   a n d   i f   p a t i e n t i   h a s   n o t   C M B s 0 , o t h e r w i s e n
where TP is the number of correctly classified positive samples, TN is the number of correctly classified negative samples, FP is the number of false positives, and FN is the number of false negatives (missed CMBs), n is the number of patients in the test set, and S is the number of SWI slice images per patient; its value is 130 in this paper. T P j represents the number of true CMBs regions on the j-th slice image.

4.2. Experiment Setup

The hardware and software utilized in this study are as follows. An Intel Xeon Gold 6330 CPU operating at 2.0 GHz, 64GB RAM, and a NVIDIA TITAN RTX with 24GB GPU memory. The software environment includes Python 3.8.5, CUDA 11.1, and cuDNN 8.9.2 for GPU-accelerated libraries.

4.3. Evaluation of Model Performance

4.3.1. Detection Performance of CMBs

In this section, we evaluate the performance of various detection models, focusing on three distinct research directions: CNN-based detection methods, transformer-based methods, and the YOLO series. Our objective is to provide a comprehensive analysis of the strengths and limitations of these different approaches in the context of detecting CMBs. To ensure consistency in our evaluation, all models were trained using input images with a resolution of 512 × 512 . Additionally, data augmentation techniques, including RandomResize, RandomFlip, and RandomCrop, were applied during the training phase for the CMBs-Public dataset. The training configuration details are outlined in Table 5, and the comparative results are presented in Table 6 and Table 7.
The comparative results in Table 6 reveal that our CMBs-YOLO model outperforms other detection methods, with CNN-based methods showing the lowest performance. A notable trend observed across all models is the higher achievement in P metrics compared to R. It indicates a prevalent challenge of missing detections, predominantly attributed to the subtle features of CMB targets. Our CMBs-YOLO model, scaled at X6, achieves the highest precision and recall rates, exhibiting an approximately 5% improvement over transformer-based models. This improvement underscores the efficacy of the CMBs-MHA module and the NWD loss functions implemented in our study. Meanwhile, our proposed method does not entail a substantial increase in learnable parameters and inference time. For example, the smallest variant of the CMBs-YOLO (N) model increases only an additional 0.46 million parameters but achieves a 9% improvement in the m A P 50 95 metric compared to the YOLOv8 model, demonstrating a favorable balance between performance and speed.
Analysis of the CMBs-Public dataset, presented in Table 7, reveals a notable decline in the R metric across all evaluated models. This decrease can be attributed to two factors: first, the limited number of labeled samples of CMBs in the CMBs-Public dataset impairs the models’ ability to effectively model features, leading to underfitting, as exemplified with the Dino method, which uses a transformer structure and experienced a substantial drop in the R index from 0.89 to 0.48. Second, the discrepancy in labeling strategies between the CMBs-Private and CMBs-Public datasets contributes to this decline. Our private dataset includes a broader range of similar targets labeled as CMBs, resulting in a higher number of labeled samples, while the public dataset follows a stricter definition. This difference, combined with the need for models to address both localization and classification of CMBs in a single stage on public datasets, increases network design complexity and limits performance.
In contrast, our study advocates for a decoupled design, where the localization and classification of CMBs are handled in separate stages. This design allows the detection model to focus on identifying candidate CMBs regions in the first stage, without the added complexity of distinguishing false positives, thereby improving the model’s performance, as demonstrated by the differences in results between the public and private datasets.
Additionally, it is noteworthy that the YOLO series detection methods exhibit less pronounced performance degradation on the CMBs-Public dataset compared to transformer-based models. For instance, the CMBs-YOLO (X6) method shows a reduction in the P and R metrics by 0.12 and 0.24, respectively. This suggests that YOLO detectors are better suited for scenarios with limited sample sizes, whereas transformer-based models often require a larger volume of labeled samples to fully leverage their modeling capabilities.
Figure 7 illustrates the relationships between mAP indicators, inference speed, and parameters across different models, providing a clear visual understanding of performance dynamics. Additional comparisons regarding parameters and speed can be found in Figure 8 and Figure 9.
In Figure 10, we present a partial visualization of results obtained from various detection models. Detailed detection results for our CMBs-YOLO model can be found in the Appendix A, Appendix B and Appendix C. It is evident that, except for our CMBs-YOLO, all other models exhibit issues of erroneous and missed detections. As these methods are general-purpose detection networks and are not specifically optimized for detecting CMBs, they demonstrate limited capabilities in identifying CMB targets.

4.3.2. Classification Performance of CMBs

In the detection phase, numerous candidate CMB regions are identified, yet these regions contain a considerable number of false positives. Accurately classifying these regions is crucial for the automated diagnostic process. To ensure a fair evaluation of different classification models, this section assesses their performance within the context of 3D structures. Notably, transformer-based methods have not yet been explored in research on CMBs. To address this gap, we have incorporated the latest advancements from the video classification domain, specifically UniFormer v1 (UFv1) [67] and UniFormer v2 (UFv2) [68], into our comparative analysis. The performance comparison among different classification models is based on results obtained from our CMBs-YOLO (X6) model. Additionally, the methods discussed in this section utilize the CMBs-Private dataset for comparison.
Table 8 presents the results of various classification models, revealing that networks based on transformer structures generally outperform those utilizing CNN architectures. This observation suggests that transformer methods offer more advanced feature representation capabilities, particularly for handling 3D image data. However, CNN-based models retain an advantage in terms of parameters and inference speed. An intriguing observation from Table 8 is that the escalation in the number of parameters for transformer methods does not invariably lead to corresponding improvements in performance. For example, the UFv2 model demonstrated only a 2% enhancement in the P metric when its parameters increased from 108 M to 336 M. This phenomenon implies that once the feature extraction capacity of the transformer structure reaches a certain threshold, network complexity may no longer be the bottleneck for improving the accuracy of CMB classification.
The integration of text description features into our CF model, alongside 3D image features, has significantly improved the accuracy of CMB classification. When comparing models with similar network complexities, such as UFv2_B16, CF_B without text, and CF_B with text, the CF_B with text configuration exhibited a notable performance enhancement. Specifically, it attained a 7% and 12% increase in the P metric compared to UFv2_B16 and CF_B without text models, respectively. This enhancement underscores the critical importance of incorporating text description content into the classification network, highlighting the value of the “asking” step in the clinician’s procedure. For a more intuitive and visual comparison, readers are directed to Figure 11 and Figure 12, which provide clear insights into the comparative strengths and weaknesses of each model.
In Table 9, we present partial visualization results obtained from various classification models. In this table, S i 1 , S i , and S i + 1 represent the SWI images at slices i − 1, i, and i + 1, respectively, while P i represents the corresponding phase image. Notably, in the case of the first and last patients, only the CF_B model, which incorporates additional text information, accurately identified the regions as non-CMBs. Conversely, all other models incorrectly classified these regions as CMBs, due to the high similarity between interfering targets and CMBs. It is noteworthy that the first and last patients are inferred to be relatively young females based on gender and age. Models lacking patient-specific information often struggle with accurate classification in such cases. For a more comprehensive view of results from our method, please refer to the Appendix A, Appendix B and Appendix C.

4.3.3. Total Performance of CMBs

In Section 4.3.1 and Section 4.3.2, our analysis focused on comparing the performance of different models in terms of detection and classification stages. In this section, we shift our attention to evaluate the performance of various model combinations at the patient level. Recognizing the vast array of possible combinations, we have selectively included only the top-performing model from each method category for reference. To guarantee a fair and balanced comparison, especially in terms of inference speed, we have standardized the number of potential CMB regions identified via different detection models to 100. The final total inference time metric can be computed as the average processing time across N patients, using the following formula: i n f e r t i m e = i = 0 N N u m i m a g e × D i m a g e + 100 × C i m a g e N , where N u m i m a g e represents the number of slice images for each patient taking a constant value of 130 in this paper. D i m a g e denotes the time taken by the detection model to process a sliced image, while C i m a g e represents the time required for the classification model to process a CMB region. The dataset used in this section is derived from the CMBs-Private dataset, and comparison results are summarized in Table 10.
Combined with transformer-based classification models, the YOLO series demonstrates superior performance and speed compared to CNN-based architectures. Evaluation of the FPavg index indicates that all models have scores exceeding 3, which signifies that the misjudgment rate of CMBs still remains a considerable challenge in this field. However, our proposed framework, as reflected by the P p a t i e n t index, has attained the highest performance among various combinations. This achievement highlights the effectiveness of our MM-UniCMBs framework in addressing the CMB detection and classification tasks, particularly from a patient-level perspective. Table 11 presents some representative examples of erroneous patient cases. A further analysis of these cases, coupled with clinical diagnostic insights, reveals that the current limitation still lies in the restricted scope of information utilized for identifying challenging cases of CMBs, compared to clinical doctors. Specifically, the absence of detailed patient lifestyle habits and past medical history in the modeling stage forms a significant hurdle in enhancing the accuracy of CMB determinations. This observation underscores the potential values of incorporating a more comprehensive range of patient-specific information into the classification process, suggesting that more holistic input information could improve diagnostic accuracy in the automated detection of CMBs.
In terms of detection models, our CMBs-YOLO model, coupled with text descriptions input, surpasses the YOLOv8 model in performance. Specifically, it improves upon YOLOv8 by 5% in the P p a t i e n t metric and by 0.29 in the FPavg metric, demonstrating that the CMBs-YOLO effectively enhances CMB classification performance through lowering missing detection rates. Additionally, our CF_B model consistently exhibits lower false detection rates compared to other classification models, regardless of the detection model it is paired with. This improvement is attributed to the integration of patient text description information and the transformer architecture in the CMBs classification process. Moreover, the increased speed of our method can be attributed to its ‘front light and back heavy’ structural design, which allocates more computing resources to the classification task. For readers seeking a more visual and intuitive understanding of these performance metrics, Figure 13 provides a comparative illustration of these aspects.

5. Ablation Study

5.1. Ablation Study in Detection Stage

In this section, we discuss the ablation experiments conducted during the detection phase using the CMBs-Private dataset. Table 12 illustrates the impact on performance metrics when the CMBs-MHA and NWD components are removed, respectively.
The results in Table 12 clearly demonstrate that the CMBs-MHA module plays a more critical role than the NWD component in our CMB detection model. The removal of the CMBs-MHA module led to a significant decrease in the m A P 50 95 metric, decreasing from 0.62 to 0.44, representing a substantial reduction of 0.18. In contrast, the exclusion of NWD led to a minor decrease of 0.03 in the same metric. This notable disparity underscores the crucial importance of feature modeling capability in the task of detecting CMBs, directly influencing the performance of detection models. Figure 14 visually represents the impact of these changes, providing an intuitive understanding of their significance.
The incorporation of the NWD loss in detection models is designed to enhance training stability, particularly for small targets. This effect is vividly depicted in Figure 15, where models without NWD loss, such as YOLOv8, exhibit pronounced oscillations in box loss during the training phase. These oscillations lead to increased instability in the network, potentially diminishing the model’s generalization capabilities. In contrast, when NWD is integrated into our CMBs-YOLO model, there is a noticeable improvement in the smoothness and stability of the training process compared to the original configuration. This improvement indicates that the NWD method outperforms IoU-based approaches, especially in dealing with small targets. The increased stability achieved during the training stage is crucial in bolstering the overall robustness and dependability of our detection model, ensuring consistent performance in the CMB detection task.

5.2. Ablation Study of Different Input Resolutions in Detection Stage

In Section 4.3.1, we evaluated detection algorithms using fixed 512 × 512 pixel input resolutions throughout the training and testing phases. In this section, however, we expand the discussion to explore the impact of varying input resolutions on model performance. The dataset used for this analysis is derived entirely from the CMBs-Private dataset.
Table 13 compares the detection performance across different input resolutions. The results show a positive correlation between increased resolution and improvements in the R metric, while reductions in resolution negatively affect both P and R metrics. When compared with the mAP trends in Figure 16, it is evident that transformer-based models exhibit greater robustness to resolution changes compared to other architectures. Notably, in the 768-resolution training mode, even when the validation resolution was reduced from 768 to 256, the DINO model experienced only a slight decrease of 0.03 in mAP, whereas ConvNeXt showed a larger decline of 0.11.
Although our proposed model is sensitive to resolution changes, it demonstrates improved robustness compared to the original YOLOv8 structure. For instance, in the 768-resolution training mode, our model’s mAP decreased by 0.09, whereas YOLOv8 exhibited a drop of 0.13. This improvement can be attributed to the enhanced feature modeling capability in MRI. The saliency modeling module enhances the network’s ability to differentiate between CMBs and their surrounding background. Additionally, the integration of the NWD loss function during training increases the model’s sensitivity to small targets, further mitigating the impact of resolution changes.
It is important to note that inconsistencies between training and testing resolutions can introduce variability in the model’s detection of identical targets, as shown in Figure 17. This occurs because the models are not exposed to randomly resized augmentations during training. Addressing this limitation through strategies such as multi-scale training or data augmentation could enhance the model’s generalization and robustness across varying resolutions.

5.3. Ablation Study of Different Input Resolutions in Classification Stage

During the detection phase, we obtained numerous candidate boxes for CMBs, each varying in size. Section 4.3.2 presents a performance comparison of classification models in fixed-resolution mode. In this section, we will discuss the impact of input resolution on classification models. To facilitate this analysis, we have designed two comparison modes: (1) fixed-resolution training with fixed-resolution testing and (2) dynamic-resolution training with dynamic-resolution testing. In the fixed-resolution mode, the input 3D images are resized to a standard fixed size, whereas in the dynamic-resolution mode, the input image retains the dimensions of the original detection box. The comparison results for these two modes, utilizing the CMBs-YOLO (X6) model during the detection stage, are summarized in Table 14.
In fixed-resolution mode, increasing the input resolution can enhance the model’s performance. Conversely, combining high-resolution training with low-resolution prediction leads to a notable decline in the performance of all methods, indicating a strong coupling between training and prediction. In dynamic-resolution mode, the performance of different models exhibits variability, suggesting that this mode is less suitable for the CMB classification task. However, the incorporation of text information in dynamic mode improves the model’s robustness to variations in resolution. For instance, in the dynamic mode of the CF model, CF-B with text enhanced the P index by 0.16 compared to CF-B without text information. This underscores the importance of incorporating additional feature dimensions in CMB classification to enhance accuracy. Figure 18 provides a more intuitive comparison of the performance changes observed across different models in dynamic mode.

5.4. Ablation Study in Image Branch

In the initial design of our classification model, a 3D CNN structure was employed in the image branch for feature extraction from 3D CMB regions. This section presents an alternative approach, wherein 2D convolution is used alongside 2D images in the image branch. While this method still incorporates text descriptions and transformer structures, it introduces a variation in the treatment of 3D inputs. Specifically, these 3D images are first averaged to create 2D representations, which are then processed through a 2D convolutional network in the image branch. The dataset used in this section is derived from the CMBs-Private dataset, and the results of this comparative approach are detailed in Table 15.
The results presented in Table 15 clearly demonstrate that the classification performance of the 2D method is generally inferior to that of the 3D approach. Notably, this performance decline is more pronounced in CNN-based methods compared to the CF method. For instance, the 2D-R101 model exhibits a 13% decrease in the P indicator relative to its 3D counterpart (3D-R101), whereas the decrease for the 2D-CF_B model is just 6% when compared to the 3D-CF_B model. This suggests that the integration of textual information effectively mitigates the limitations of relying solely on image data. Additionally, the incorporation of a transformer structure enhances the network’s ability to process feature relationships relevant to CMBs and their mimics. Table 16 presents specific cases that illustrate the tendency of 2D methods to produce erroneous results.
As shown in Table 16, the comparative results clearly highlight the limitations of 2D methods when applied in the image branch, particularly their reduced capacity to extract information across different slice images due to the averaging operations involved. This limitation often leads to the misidentification of false CMBs. For instance, in Case 3, cerebral calcification is clearly visible in the phase image P i , but this distinction is obscured in the averaged images, resulting in misclassification. In contrast, 3D methods are more effective at extracting feature information from multiple slice images, which enables them to outperform 2D approaches and deliver more accurate classification results.

5.5. Ablation Study of CLIP Model

In the initial design of our classification model, we utilized BERT to transform our text prompts into word embedding vectors. This section will discuss the application of the CLIP model [44], examining its effectiveness in classifying CMBs. CLIP, renowned for its proficiency in encoding both text and images, inspired us to devise four different comparison strategies. Specifically, in the text branch of our model, we substitute BERT’s text encoding with CLIP’s encoding mechanism. For the image branch, our approach involves incorporating CLIP’s image encoding features into our pre-existing 3D image branch. These alterations to the model’s structure are illustrated in Figure 19. The dataset used in this section is derived from the CMBs-Private dataset. The results of implementing these strategies are documented in Table 17 and are also visually represented in Figure 20.
Table 17 reveals that the implementation of the CLIP method has negatively impacted the performance of classification networks, especially those based on CNN detection models. Among these tested strategies, the most pronounced decline in performance was observed in the third approach, followed by the second, while the first strategy showed the least impact. This trend suggests that CLIP’s image encoding may not be well suited for the classification of CMBs, as it tends to reduce the effectiveness of the original model.
Several factors contribute to these adverse effects. Firstly, the training dataset for CLIP may not include the MRI types, limiting its ability to model the CMBs’ features and their mimics. Secondly, due to the limitations of the CLIP model in processing 3D images, it became necessary to encode each slice image independently and subsequently concatenate them together, akin to 2D convolution. This manner significantly reduces the spatial information of CMBs’ features, consequently diminishing the network’s performance, as discussed in previous section. Moreover, even after fine-tuning the CLIP model with our CMB dataset and employing it as a new pretraining model under the second strategy, we observed only a slight improvement in performance compared to the conventional version of CLIP.
From the perspective of the text branch, CLIP did not outperform BERT. As a multimodal pretraining model, CLIP’s text encoding capabilities are less robust compared to specialized text models like BERT. Consequently, CLIP struggles to accurately model text statements, leading to inferior performance when compared to the BERT approach.
In conclusion, although CLIP is renowned for its versatility and large-scale modeling, it demonstrates limitations in specialized medical tasks, such as the classification of CMBs. Our approach, which combines BERT with a transformer structure, has proven to be more effective in this field.

6. Discussion

The detection of CMBs is a critical step in diagnosing many neurological diseases; however, it poses significant challenges for clinical neurologists [1,2]. CMBs are small, unpredictably located, and often surrounded by structures with similar characteristics, making detection particularly challenging for novice neurologists [69]. Even experienced clinicians find manual detection time-consuming and prone to error. The integration of deep learning techniques into the diagnosis of CMBs has emerged as an important interdisciplinary approach, bridging the fields of medicine and computer science [9,37,39,40].
However, existing research has predominantly focused on the use of image data for the detection and classification of CMBs, while overlooking the role of patient-specific textual information on model results [9,11,14]. These approaches typically employ 3D convolution to extract feature information about cerebral hemorrhage points across different slices, followed by detection and classification. This modeling approach diverges significantly from the actual diagnostic processes employed by clinical practitioners, making it challenging for AI models to overcome existing performance limitations in the detection of CMBs. By researching the actual diagnostic workflow of clinical doctors, we have identified several key processes in CMB detection tasks: 1. “Looking”: the clinician first locates potential bleeding areas based on the patient’s CT scans. 2. “Checking”: physicians must integrate information from multiple slice images to differentiate between bleeding and brain calcification. 3. “Asking”: the final judgment is made considering the patient’s specific circumstances, including age, gender, medical history, and lifestyle habits. Based on this understanding, we propose a novel CMB detection framework designed to simulate the actual diagnostic process of clinical practitioners.
The overall results of this study demonstrate that our proposed framework significantly enhances accuracy compared to existing methods. In the detection stage, as illustrated in Figure 10, the introduction of CMBs-MHA into the detection model enables CMBs-YOLO to strengthen its ability to model the salient features of CMBs. Additionally, Table 6 reveals that our detection approach exhibits strong generalization capabilities, as evidenced by comparisons with various detection models on the CMBs-Public [46] datasets. For classification tasks, we have designed a multimodal network structure that simultaneously integrates image and text data relevant to CMBs. By combining the insights from Table 10 and Figure A1, we conclude that the inclusion of multidimensional features is essential for improving the accuracy of CMBs. Furthermore, a comparison of Table 8 and Table 10 indicates that the multimodal structure achieves superior performance at the patient level, highlighting the advantages of multimodal technology [70,71,72,73] over single-modal model structures [74,75].
Through a series of ablation studies, we validated the hyperparameter optimizations and prediction stability of our framework in the field of CMB detection. Firstly, in the detection phase, we compared the effects of the CMBs-MHA module and the NWD [76,77] module on performance during the training and testing phases. The loss function variations illustrated in Figure 15 indicate that the proposed detection model exhibits strong robustness against underfitting, early convergence, and loss optimization traps. Additionally, we explored the influence of input resolution hyperparameters in the detection networks to further elucidate how different training parameters affect model performance. In the classification network, we first demonstrated the changes in classification performance across various 3D resolutions. As shown in Table 14 and Figure 18, existing classification model theories impose high demands for consistency between training and testing conditions; inconsistent parameter changes can increase the uncertainty of model predictions and consequently reduce performance. We then focused on the impact of text features on CMB classification results. Table 11 presents the role of text information in error cases, highlighting the significance of the text modeling approach in enhancing the accuracy of CMB diagnoses. Building on this, we compared the performance of the CLIP and BERT methods from the perspective of image and text dimensions, as detailed in Table 16 and Figure 19, which underscore the rationality and effectiveness of the methods employed in our framework.
However, our MM-UniCMBs framework exhibits limitations due to its two-stage pipeline design. One concern is that separating the detection and classification stages may disrupt the inherent connection between these tasks, potentially hindering the release of model capabilities. Diwan et al. [78] observed that two-stage detection networks are increasingly being replaced by single-stage structures due to their superior generalization and robustness. Optimization techniques such as Fast Fourier Transform and uncertainty quantification (UQ) [79,80,81] have been explored to accelerate the convergence of object detectors. Abbaszadeh et al. [82] emphasized the importance of UQ methods in assessing model certainty for practical applications and proposed a novel state-of-the-art UQ approach, the Automated Random Deactivating Connective Weights (ARDCWs) ensemble method, for performance validation. In future research, we will investigate the utility of UQ methods in medical recognition tasks to improve the reliability of our models. Another concern is that the decoupled structure of our framework increases training complexity and poses challenges for the automated implementation of CMB detection. In future work, we aim to address these issues by exploring an end-to-end CMB detection method, guided by the principles of multimodal design. Additionally, as more advanced prediction methods are developed, we will investigate the application of technologies such as block combined network structures [83], hybrid block modular models [84], and 3D filtering and block matching [85,86,87] in our medical prediction tasks.

7. Conclusions

In this paper, we present a novel multimodal unified framework, termed MM-UniCMBs, specifically designed for the detection of CMBs using MRI and language data. The MM-UniCMBs framework employs a ‘front-light detection stage’ (CMBs-YOLO) and a ‘back-heavy classification stage’ (CMBsFormer). To enhance the representation of weak features, we propose a multi-head attention module, CMBs-MHA, specifically designed to capture salient features that differentiate CMBs from surrounding tissues. To improve the robustness of box regression for small targets, we introduce the NWD loss function, which enhances stability compared to traditional IoU-based methods. Comparative analysis shows that CMBs-YOLO achieves superior detection performance over existing approaches. In the classification phase, we introduce CMBsFormer, a multimodal classification network designed to address the limitations of previous studies that solely relied on single-image data. CMBsFormer integrates multidimensional patient descriptors during the classification process, combining both visual data and patient-specific information. This framework merges the ‘looking’ step, which analyzes MRI images, with the ‘asking’ step, incorporating patient features such as age and gender. This integration enriches the model’s feature space and enhances interpretability, helping to reduce the disparity between AI-driven methods and clinical diagnostic procedures. Additionally, the incorporation of a transformer structure further strengthens the network’s ability to synthesize information across image and text modalities, providing superior adaptability compared to conventional CNN architectures. CF is flexible and scalable, capable of processing both single-modality image inputs and multimodal image–text inputs. Implemented in the diagnostic process, this framework achieved 82% accuracy and a diagnostic speed of 5 s. Moreover, this framework serves as a valuable tool for less experienced clinicians, offering an auxiliary reference for CMBs diagnosis.

Author Contributions

C.C.: Writing—original draft, Methodology; L.-L.Z.: Methodology and Software; Q.L.: Dataset Collection and Labeling; Y.X.: Writing—Methodology, review and editing, Supervision, Conceptualization. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

https://github.com/ai-application/MM-UniCMBs (accessed on 26 September 2024).

Acknowledgments

The authors would like to thank the editors and the reviewers for their valuable suggestions.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Figure A1 illustrates the detection results from our detection model (CMBs-YOLO-X6). In these visualizations, dashed boxes are color-coded in red, green, and yellow to represent detection scores exceeding 90%, falling between 80 and 90%, and ranging from 70 to 80%, respectively.
Figure A1. The detection results of CMB-YOLO (X6).
Figure A1. The detection results of CMB-YOLO (X6).
Bioengineering 11 00993 g0a1

Appendix B

Figure A2 displays the classification results obtained with our CF-B model with text input, across various patients. In this visualization, red boxes are used to identify true CMBs, while green boxes indicate pseudo CMBs. Specifically, in the first block of Figure A2, the green image illustrates instances of false positives as identified with our model. Conversely, the red image in the second block represents cases where the model has missed classifications.
Figure A2. The classification results of CF (B) with text input.
Figure A2. The classification results of CF (B) with text input.
Bioengineering 11 00993 g0a2

Appendix C

Table A1. The abbreviations and descriptions of detection models.
Table A1. The abbreviations and descriptions of detection models.
AbbreviationDescriptions
R-50 [55]ResNet with 50 convolution depth
ConvNeXt-TConvNext with tiny backbone size, depth = [3, 3, 9, 3]
ConvNeXt-SConvNext with small backbone size, depth = [3, 3, 27, 3]
YOLO-N/CMBs-YOLO-NYOLOv8/CMBs-YOLO methods with N backbone size
depth_multiple = 0.33, width_multiple = 0.25, max_channels = 1024
YOLO-S/CMBs-YOLO-SYOLOv8/CMBs-YOLO methods with S backbone size
depth_multiple = 0.33, width_multiple = 0.5, max_channels = 1024
YOLO-M/CMBs-YOLO-MYOLOv8/CMBs-YOLO methods with M backbone size
depth_multiple = 0.67, width_multiple = 0.75, max_channels = 768
YOLO-X6/CMBs-YOLO-X6YOLOv8/CMBs-YOLO methods with x6 backbone size
depth_multiple = 1.0, width_multiple = 1.25, max_channels = 512
Table A2. The abbreviations and descriptions of classification models.
Table A2. The abbreviations and descriptions of classification models.
AbbreviationDescriptions
R-18ResNet with 18 convolution depth
R-50ResNet with 50 convolution depth
R-101ResNet with 101 convolution depth
UFv1_XXSUniFormer version 1 with XXS backbone size
depth = [2, 5, 8, 2], embed_dim = [56, 112, 224, 448], head_dim = 28
UFv1_XSUniFormer version 1 with XS backbone size
depth = [3, 5, 9, 3], embed_dim = [64, 128, 256, 512], head_dim = 32
UFv2_B16UniFormer version 2 with B16 backbone size
width = 768, layers = 12, heads = 12, output_dim = 512
UFv2_114UniFormer version 2 with 114 backbone size
width = 1024, layers = 24, heads = 16, output_dim = 768
CF_S
w/o text
CMBsFormer with Small backbone size and text inputs (BERT or CLIP)
layers = 6, embed_dims = 512, heads = 4, text_channels = 70
CF_B
w/o text
CMBsFormer with Base backbone size and text inputs (BERT or CLIP)
layers = 12, embed_dims = 768, heads = 8, text_channels = 70
CF_S
w/text
CMBsFormer with Small backbone size
layers = 6, embed_dims = 512, heads = 4
CF_B
w/text
CMBsFormer with Base backbone size
layers = 12, embed_dims = 768, heads = 8

References

  1. Greenberg, S.M.; Vernooij, M.W.; Cordonnier, C.; Viswanathan, A.; Salman, R.A.; Warach, S.; Launer, L.J.; Buchem, M.A.V.B.; Breteler, M.M.B. Cerebral microbleeds: A guide to detection and interpretation. Lancet Neurol. 2009, 8, 165–174. [Google Scholar] [CrossRef]
  2. Charidimou, A.; Werring, D.J. Cerebral microbleeds: Detection, mechanisms and clinical challenges. Future Neurol. 2011, 6, 587–611. [Google Scholar] [CrossRef]
  3. Yamazaki, K.; Vo-Ho, V.K.; Bulsara, D.; Le, N. Spiking neural networks and their applications: A review. Brain Sci. 2022, 12, 863. [Google Scholar] [CrossRef] [PubMed]
  4. Dora, S.; Kasabov, N. Spiking neural networks for computational intelligence: An overview. Big Data Cogn. Comput. 2021, 5, 67. [Google Scholar] [CrossRef]
  5. Turkson, R.E.; Qu, H.; Mawuli, C.B.; Eghan, M.J. Classification of Alzheimer’s disease using deep convolutional spiking neural network. Neural Process. Lett. 2021, 53, 2649–2663. [Google Scholar] [CrossRef]
  6. Ahmadi, M.; Sharifi, A.; Hassantabar, S.; Enayati, S. QAIS-DSNN: Tumor area segmentation of MRI image with optimized quantum matched-filter technique and deep spiking neural network. BioMed Res. Int. 2021, 2021, 6653879. [Google Scholar] [CrossRef] [PubMed]
  7. Rajagopal, R.; Karthick, R.; Meenalochini, P.; Kalaichelvi, T. Deep Convolutional Spiking Neural Network optimized with Arithmetic optimization algorithm for lung disease detection using chest X-ray images. Biomed. Signal Process. Control 2023, 79, 104197. [Google Scholar] [CrossRef]
  8. Nikseresht, G.; Agam, G.; Arfanakis, K. End-to-end task-guided refinement of synthetic images for data efficient cerebral microbleed detection. In Proceedings of the 26th International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 2756–2763. [Google Scholar]
  9. Ferlin, M.A.; Grochowski, M.; Kwasigroch, A.; Mikołajczyk, A.; Szurowska, E.; Grzywińska, M.; Sabisz, A. A comprehensive analysis of deep neural-based cerebral microbleeds detection system. Electronics 2021, 10, 2208. [Google Scholar] [CrossRef]
  10. Ateeq, T.; Majeed, M.N.; Anwar, S.M.; Maqsood, M.; Rehman, Z.; Lee, J.W.; Muhammad, K.; Wang, S.H.; Baik, S.W.; Mehmood, I. Ensemble-classifiers-assisted detection of cerebral microbleeds in brain MRI. Comput. Electr. Eng. 2018, 69, 768–781. [Google Scholar] [CrossRef]
  11. Dou, Q.; Chen, H.; Yu, L.; Zhao, L.; Qin, J.; Wang, D.; Mok, V.; Shi, L.; Heng, P. Automatic detection of cerebral microbleeds from MR images via 3D convolutional neural networks. IEEE Trans. Med. Imaging 2016, 35, 1182–1195. [Google Scholar] [CrossRef]
  12. Xu, C.; Wang, J.; Yang, W.; Yu, H.; Yu, L.; Xia, G. Detecting tiny objects in aerial images: A normalized Wasserstein distance and a new benchmark. ISPRS J. Photogramm. Remote Sens. 2022, 190, 79–93. [Google Scholar] [CrossRef]
  13. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12993–13000. [Google Scholar]
  14. Kaaouana, T.; Bertrand, A.; Ouamer, F.; Ye, B.L.; Pyatigorskaya, N.; Bouyahia, A.; Thiery, N.; Dufouil, C.; Delmaire, C.; Dormont, D.; et al. Improved cerebral microbleeds detection using their magnetic signature on T2*-phase-contrast: A comparison study in a clinical setting. NeuroImage Clin. 2017, 15, 274–283. [Google Scholar] [CrossRef] [PubMed]
  15. Hong, J.; Cheng, H.; Zhang, Y.D.; Liu, J. Detecting cerebral microbleeds with transfer learning. Mach. Vis. Appl. 2019, 30, 1123–1133. [Google Scholar] [CrossRef]
  16. WANG, H.; Gagnon, B. Cerebral microbleed detection by wavelet entropy and naive Bayes classifier. In Proceedings of the 2nd International Conference on Biomedical and Biological Engineering 2017 (BBE 2017), Guilin, China, 26–28 May 2017; pp. 505–510. [Google Scholar]
  17. Bao, F.; Shi, M.; Macdonald, F. Voxelwise detection of cerebral microbleed in CADASIL patients by naïve Bayesian classifier. In Proceedings of the 2018 International Conference on Information Technology and Management Engineering (ICITME 2018), Beijing, China, 26–27 August 2018; pp. 176–180. [Google Scholar]
  18. Sundaresan, V.; Arthofer, C.; Zamboni, G.; Dineen, R.A.; Rothwell, P.M.; Sotiropoulous, S.N.; Auer, D.P.; Tozer, D.J.; Markus, H.S.; Miller, K.L.; et al. Automated detection of candidate subjects with cerebral microbleeds using machine learning. Front. Neuroinform. 2022, 15, 777828. [Google Scholar] [CrossRef] [PubMed]
  19. Chesebro, A.G.; Amarante, E.; Lao, P.J.; Meier, I.B.; Mayeus, R.; Brickman, A.M. Automated detection of cerebral microbleeds on T2*-weighted MRI. Sci. Rep 2021, 11, 4004. [Google Scholar] [CrossRef] [PubMed]
  20. Barnes, S.R.S.; Haacke, E.M.; Ayaz, M.; Boikov, A.S.; Kirsch, W.; Kido, D. Semiautomated detection of cerebral microbleeds in magnetic resonance images. Magn. Reson. Imaging 2011, 29, 844–852. [Google Scholar] [CrossRef]
  21. Dou, Q.; Chen, H.; Yu, L.; Dhi, L.; Wang, D.; Mok, V.; Heng, P. Automatic cerebral microbleeds detection from MR images via independent subspace analysis based hierarchical features. In Proceedings of the 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Milan, Italy, 25–29 August 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 7933–7936. [Google Scholar]
  22. Fazlollahi, A.; Meriaudeau, F.; Giancardo, L.; Villemagne, V.L.; Rowe, C.C.; Yates, P.; Salvado, O.; Bourgeat, P. Computer-aided detection of cerebral microbleeds in susceptibility-weighted imaging. Comput. Med. Imaging Graph 2015, 46, 269–276. [Google Scholar] [CrossRef] [PubMed]
  23. Bian, W.; Hess, C.P.; Chang, S.M.; Nelson, S.J.; Lupo, J.M. Computer-aided detection of radiation-induced cerebral microbleeds on susceptibility-weighted MR images. NeuroImage Clin. 2013, 2, 282–290. [Google Scholar] [CrossRef] [PubMed]
  24. Fazlollahi, A.; Meriaudeau, F.; Villemagne, V.L.; Rowe, C.C.; Yates, P.; Salvado, O.; Bourgeat, P. Efficient machine learning framework for computer-aided detection of cerebral microbleeds using the radon transform. In Proceedings of the 2014 IEEE 11th International Symposium on Biomedical Imaging (ISBI), Beijing, China, 29 April–2 May 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 113–116. [Google Scholar]
  25. Morrison, M.A.; Payabvash, S.; Chen, Y.; Avadiappan, S.; Shah, M.; Zou, X.; Hess, C.P.; Lupo, J.M. A user-guided tool for semi-automated cerebral microbleed detection and volume segmentation: Evaluating vascular injury and data labelling for machine learning. NeuroImage Clin. 2018, 20, 498–505. [Google Scholar] [CrossRef]
  26. Momeni, S.; Fazlollahi, A.; Yates, P.; Rowe, C.; Gao, Y.; Liew, A.; Salvado, O. Synthetic microbleeds generation for classifier training without ground truth. Comput. Methods Programs Biomed. 2021, 207, 106127. [Google Scholar] [CrossRef]
  27. Liu, S.; Utriainen, D.; Chai, C.; Chen, Y.; Wang, L.; Sethi, S.K.; Xia, S.; Haacke, E.M. Cerebral microbleed detection using susceptibility weighted imaging and deep learning. Neuroimage 2019, 198, 271–282. [Google Scholar] [CrossRef]
  28. Wu, R.; Liu, H.; Li, H.; Chen, L.; Wei, L.; Huang, X.; Liu, X.; Men, X.; Li, X.; Ham, L.; et al. Deep learning based on susceptibility-weighted MR sequence for detecting cerebral microbleeds and classifying cerebral small vessel disease. BioMedical Eng. Online 2023, 22, 99. [Google Scholar] [CrossRef] [PubMed]
  29. Li, T.; Zou, Y.; Bai, P.; Li, S.; Wang, H.; Chen, X.; Meng, Z.; Kang, Z.; Zhou, G. Detecting cerebral microbleeds via deep learning with features enhancement by reusing ground truth. Comput. Methods Programs Biomed. 2021, 204, 106051. [Google Scholar] [CrossRef] [PubMed]
  30. Zhang, Y.D.; Hou, X.X.; Chen, Y.; Chen, H.; Yang, M.; Yang, J.Q.; Wang, S.H. Voxelwise detection of cerebral microbleed in CADASIL patients by leaky rectified linear unit and early stopping. Multimed. Tools Appl. 2018, 77, 21825–21845. [Google Scholar] [CrossRef]
  31. Crouzet, C.; Jeong, G.; Chae, R.H.; Lopresti, K.T.; Dunn, C.E.; Xie, D.F.; Agu, C.; Fang, C.; Nunes, A.C.F.; Lau, W.L.; et al. Spectroscopic and deep learning-based approaches to identify and quantify cerebral microhemorrhages. Sci. Rep. 2021, 11, 10725. [Google Scholar] [CrossRef] [PubMed]
  32. Rashid, T.; Abdulkadir, A.; Nasrallah, I.M.; Ware, J.B.; Liu, H.F.; Spincemaille, P.; Romero, J.R.; Bryan, R.N.; Heckbert, S.R.; Habes, M. DEEPMIR: A deep neural network for differential detection of cerebral microbleeds and iron deposits in MRI. Sci. Rep. 2021, 11, 14124. [Google Scholar] [CrossRef]
  33. Lee, H.; Kim, J.H.; Lee, S.; Jung, K.J.; Kim, W.E.; Noh, Y.; Kim, E.Y.; Kang, K.M.; Sohn, C.H.; Lee, D.Y.; et al. Detection of cerebral microbleeds in MR images using a single-stage triplanar ensemble detection network (TPE-Det). J. Magn. Reson. Imaging 2023, 58, 272–283. [Google Scholar] [CrossRef]
  34. Al-Masni, M.A.; Kim, W.R.; Kim, E.Y.; Noh, Y.; Kim, D.H. Automated detection of cerebral microbleeds in MR images: A two-stage deep learning approach. NeuroImage Clin. 2020, 28, 102464. [Google Scholar] [CrossRef]
  35. Fang, Z.; Zhang, R.; Guo, L.; Xia, T.; Zeng, Y.; Wu, X. Knowledge-guided 2.5 D CNN for cerebral microbleeds detection. Biomed. Signal Process. Control 2023, 86, 105078. [Google Scholar] [CrossRef]
  36. Lu, S.Y.; Nayak, D.R.; Wang, S.H.; Zhang, Y.D. A cerebral microbleed diagnosis method via featurenet and ensembled randomized neural networks. Appl. Soft Comput. 2021, 109, 107567. [Google Scholar] [CrossRef]
  37. Koschmieder, K.; Paul, M.M.; Van den Heuvel, T.L.A.; Eerden, A.V.D.; Ginneken, B.V.; Manniesing, R. Automated detection of cerebral microbleeds via segmentation in susceptibility-weighted images of patients with traumatic brain injury. NeuroImage Clin. 2022, 35, 103027. [Google Scholar] [CrossRef] [PubMed]
  38. Stanley, B.F.; Franklin, S.W. Automated cerebral microbleed detection using selective 3D gradient co-occurance matrix and convolutional neural network. Biomed. Signal Process. Control 2022, 75, 103560. [Google Scholar] [CrossRef]
  39. Chen, Y.; Villanueva-Meyer, J.E.; Morrison, M.A.; Lupo, J.M. Toward automatic detection of radiation-induced cerebral microbleeds using a 3D deep residual network. J. Digit. Imaging 2019, 32, 766–772. [Google Scholar] [CrossRef] [PubMed]
  40. Wang, S.; Tang, C.; Sun, J.; Zhang, Y. Cerebral micro-bleeding detection based on densely connected neural network. Front. Neurosci. 2019, 13, 422. [Google Scholar] [CrossRef] [PubMed]
  41. Boecking, B.; Usuyama, N.; Bannur, S.; Castro, D.C.; Schwaighofer, A.; Hyland, S.; Wetscherek, M.; Naumann, T.; Nori, A.; Alvarez-Valle, J.; et al. Making the most of text semantics to improve biomedical vision–language processing. In European Conference on Computer Vision; Springer Nature: Cham, Switzerland, 2022; pp. 1–21. [Google Scholar]
  42. Li, Z.; Li, Y.; Li, Q.; Wang, P.; Guo, D.; Lu, L.; Jin, D.; Zhang, Y.; Hong, Q. Lvit: Language meets vision transformer in medical image segmentation. IEEE Trans. Med. Imaging 2023, 43, 96–107. [Google Scholar] [CrossRef] [PubMed]
  43. Liu, J.; Zhang, Y.; Chen, J.N.; Xiao, J.; Lu, Y.; Landman, B.A.; Yuan, Y.; Yuille, A.; Tang, Y.; Zhou, Z. Clip-driven universal model for organ segmentation and tumor detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 21152–21164. [Google Scholar]
  44. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. Int. Conf. Mach. Learn. PMLR 2021, 8748–8763. [Google Scholar]
  45. Bhalodia, R.; Hatamizadeh, A.; Tam, L.; Xu, Z.; Wang, X.; Turkbey, E.; Xu, D. Improving pneumonia localization via cross-attention on medical images and reports. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Proceedings, Part II 24; Springer International Publishing: Berlin/Heidelberg, Germany, 2021; pp. 571–581. [Google Scholar]
  46. Sudre, C.H.; Van Wijnen, K.; Dubost, F.; Adams, H.; Atkinson, D.; Barkhof, F.; Birhanu, M.A.; Bron, E.E.; Camarasa, R.; Chaturvedi, N.; et al. Where is VALDO? VAscular Lesions Detection and segmentatiOn challenge at MICCAI 2021. Med. Image Anal. 2024, 91, 103029. [Google Scholar] [CrossRef] [PubMed]
  47. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef]
  48. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In European Conference on Computer Vision; Springer International Publishing: Cham, Switzerland, 2020; pp. 213–229. [Google Scholar]
  49. Ultralytics. YOLOv8—Ultralytics Open-Source Research. Available online: https://github.com/ultralytics/ultralytics (accessed on 26 September 2024).
  50. Yu, T.; Li, C.; Huang, J.; Xiao, X.; Zhang, X.; Li, Y.; Fu, B. ReF-DDPM: A novel DDPM-based data augmentation method for imbalanced rolling bearing fault diagnosis. Reliab. Eng. Syst. Saf. 2024, 251, 110343. [Google Scholar] [CrossRef]
  51. Rezatofighi, H.; Tsoi, N.; Gwak, J.Y.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
  52. Suwalska, A.; Wang, Y.; Yuan, Z.; Jiang, Y.; Zhu, D.; Chen, J.; Cui, M.; Chen, X.; Suo, C.; Polanska, J. CMB-HUNT: Automatic detection of cerebral microbleeds using a deep neural network. Comput. Biol. Med. 2022, 151, 106233. [Google Scholar] [CrossRef] [PubMed]
  53. Devlin, J. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  54. Ferlin, M.; Klawikowska, Z.; Grochowski, M.; Grzywinska, M.; Szurowska, E. Exploring the landscape of automatic cerebral microbleed detection: A comprehensive review of algorithms, current trends, and future challenges. Expert Syst. Appl. 2023, 232, 120655. [Google Scholar] [CrossRef]
  55. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  56. Lin, T. Focal Loss for Dense Object Detection. arXiv 2017, arXiv:1708.02002. [Google Scholar]
  57. Cai, Z.; Vasconcelos, N. Cascade R-CNN: High quality object detection and instance segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1483–1498. [Google Scholar] [CrossRef] [PubMed]
  58. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. Centernet: Keypoint triplets for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6569–6578. [Google Scholar]
  59. Zhang, H.; Chang, H.; Ma, B.; Wang, N.; Chen, X. Dynamic R-CNN: Towards high quality object detection via dynamic training. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XV 16; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 260–275. [Google Scholar]
  60. Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.F.; Zhan, W.; Tomizuka, M.; Li, L.; Yuan, Z.H.; Wang, C.H.; et al. Sparse r-cnn: End-to-end object detection with learnable proposals. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 14454–14463. [Google Scholar]
  61. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrel, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11976–11986. [Google Scholar]
  62. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.G.; Dai, J. Deformable detr: Deformable transformers for end-to-end object detection. arXiv 2020, arXiv:2010.04159. [Google Scholar]
  63. Guo, M.H.; Xu, T.X.; Liu, J.J.; Liu, Z.N.; Jiang, P.T.; Mu, T.J.; Zhang, S.H.; Martin, R.R.; Cheng, M.M.; Hu, S.M. Attention mechanisms in computer vision: A survey. Comput. Vis. Media 2022, 8, 331–368. [Google Scholar] [CrossRef]
  64. Zhang, H.; Li, F.; Liu, S.; Zhang, L.; Su, H.; Zhu, J.; Ni, L.M.; Shum, H.Y. Dino: Detr with improved denoising anchor boxes for end-to-end object detection. arXiv 2022, arXiv:2203.03605. [Google Scholar]
  65. Redmon, J. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  66. Ge, Z. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  67. Sun, Z.; Ke, Q.; Rahmani, H.; Bennamoun, M.; Wang, G.; Liu, J. Human action recognition from various data modalities: A review. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 3200–3225. [Google Scholar] [CrossRef]
  68. Fang, Y.; Wang, W.; Xie, B.; Sun, Q.; Wu, L.; Wang, X.; Huang, T.; Wang, X.; Cao, Y. Eva: Exploring the limits of masked visual representation learning at scale. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 19358–19369. [Google Scholar]
  69. Xia, P.; Hui, E.S.; Chua, B.J.; Huang, F.; Wang, Z.; Zhang, Z.; Yu, H.; Lau, K.K.; Mak, H.K.F.; Cao, P. Deep-learning-based MRI microbleeds detection for cerebral small vessel disease on quantitative susceptibility mapping. J. Magn. Reson. Imaging 2024, 60, 1165–1175. [Google Scholar] [CrossRef]
  70. Hong, J.; Wang, S.H.; Cheng, H.; Liu, J. Classification of cerebral microbleeds based on fully-optimized convolutional neural network. Multimed. Tools Appl. 2020, 79, 15151–15169. [Google Scholar] [CrossRef]
  71. Wu, X.; Hong, D.; Chanussot, J. Convolutional neural networks for multimodal remote sensing data classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–10. [Google Scholar] [CrossRef]
  72. Shao, W.; Peng, Y.; Zu, C.; Wang, M.L.; Zhang, D.Q. Hypergraph based multi-task feature selection for multimodal classification of Alzheimer’s disease. Comput. Med. Imaging Graph. 2020, 80, 101663. [Google Scholar] [CrossRef]
  73. Dai, Y.; Gao, Y.; Liu, F. Transmed: Transformers advance multi-modal medical image classification. Diagnostics 2021, 11, 1384. [Google Scholar] [CrossRef] [PubMed]
  74. Kurz, A.; Hauser, K.; Mehrtens, H.A.; Henning, E.K.; Hekler, A.; Kather, J.N.; Frohling, S.; Kalle, C.V.; Brinker, T.J. Uncertainty estimation in medical image classification: Systematic review. JMIR Med. Inform. 2022, 10, e36427. [Google Scholar] [CrossRef]
  75. Raj, R.J.S.; Shobana, S.J.; Pustokhina, I.V.; Pustokhin, D.A.; Gupta, D.; Shankar, K. Optimal feature selection-based medical image classification using deep learning model in internet of medical things. IEEE Access 2020, 8, 58006–58017. [Google Scholar] [CrossRef]
  76. Yang, X.; Yan, J.; Ming, Q.; Wang, W.; Zhang, X.; Tian, Q. Rethinking rotated object detection with gaussian wasserstein distance loss. International conference on machine learning. PMLR 2021, 139, 11830–11841. [Google Scholar]
  77. Han, Y.; Liu, X.; Sheng, Z.; Ren, Y.; Han, X.; You, J.; Liu, R.; Luo, Z. Wasserstein loss-based deep object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 998–999. [Google Scholar]
  78. Diwan, T.; Anirudh, G.; Tembhurne, J.V. Object detection using YOLO: Challenges, architectural successors, datasets and applications. Multimed. Tools Appl. 2023, 82, 9243–9275. [Google Scholar] [CrossRef]
  79. Abdar, M.; Pourpanah, F.; Hussain, S.; Rezazadegan, D.; Liu, L.; Ghavamazdeh, M.; Fieguth, P.; Cao, X.C.; Khosravi, A.; Acharya, U.R.; et al. A review of uncertainty quantification in deep learning: Techniques, applications and challenges. Inf. Fusion 2021, 76, 243–297. [Google Scholar] [CrossRef]
  80. Faghani, S.; Moassefi, M.; Rouzrokh, P.; Khosravi, B.; Baffour, F.I.; Ringler, M.D.; Erickson, B.J. Quantifying uncertainty in deep learning of radiologic images. Radiology 2023, 308, e222217. [Google Scholar] [CrossRef] [PubMed]
  81. Zhao, R.; Wang, K.; Xiao, Y.; Gao, F.; Gao, Z. Leveraging Monte Carlo Dropout for Uncertainty Quantification in Real-Time Object Detection of Autonomous Vehicles. IEEE Access 2024, 12, 33384–33399. [Google Scholar] [CrossRef]
  82. Abbaszadeh Shahri, A.; Shan, C.; Larsson, S. A novel approach to uncertainty quantification in groundwater table modeling by automated predictive deep learning. Nat. Resour. Res. 2022, 31, 1351–1373. [Google Scholar] [CrossRef]
  83. Hosseini, S.A.; Abbaszadeh Shahri, A.; Asheghi, R. Prediction of bedload transport rate using a block combined network structure. Hydrol. Sci. J. 2022, 67, 117–128. [Google Scholar] [CrossRef]
  84. Abbaszadeh Shahri, A.; Maghsoudi Moud, F. Landslide susceptibility mapping using hybridized block modular intelligence model. Bull. Eng. Geol. Environ. 2021, 80, 267–284. [Google Scholar] [CrossRef]
  85. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising with block-matching and 3D filtering. Image processing: Algorithms and systems, neural networks, and machine learning. SPIE 2006, 6064, 354–365. [Google Scholar]
  86. Naveed, K.; Abdullah, F.; Madni, H.A.; Khan, M.A.U.; Khan, T.M.; Naqvi, S.S. Towards automated eye diagnosis: An improved retinal vessel segmentation framework using ensemble block matching 3D filter. Diagnostics 2021, 11, 114. [Google Scholar] [CrossRef] [PubMed]
  87. He, Y.; Zeng, L.; Yu, W.; Gong, C. Noise suppression–guided image filtering for low-SNR CT reconstruction. Med. Biol. Eng. Comput. 2020, 58, 2621–2629. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The architecture of MM-UniCMBs.
Figure 1. The architecture of MM-UniCMBs.
Bioengineering 11 00993 g001
Figure 2. The structure of CMBs-YOLO.
Figure 2. The structure of CMBs-YOLO.
Bioengineering 11 00993 g002
Figure 3. Feature responses of CMBs in MRI.
Figure 3. Feature responses of CMBs in MRI.
Bioengineering 11 00993 g003
Figure 4. The overview of the CMBs-MHA module.
Figure 4. The overview of the CMBs-MHA module.
Bioengineering 11 00993 g004
Figure 5. The relationship between the predicted boxes and ground truth boxes.
Figure 5. The relationship between the predicted boxes and ground truth boxes.
Bioengineering 11 00993 g005
Figure 6. The architecture of CMBsFormer.
Figure 6. The architecture of CMBsFormer.
Bioengineering 11 00993 g006
Figure 7. The evaluation of various models across different metrics in the CMBs-Private testing dataset.
Figure 7. The evaluation of various models across different metrics in the CMBs-Private testing dataset.
Bioengineering 11 00993 g007
Figure 8. Comparing the trade-off between speed and performance among detection models in the CMBs-Private testing dataset.
Figure 8. Comparing the trade-off between speed and performance among detection models in the CMBs-Private testing dataset.
Bioengineering 11 00993 g008
Figure 9. Comparing the trade-off between parameters and performance among different detection models in the CMBs-Private testing dataset.
Figure 9. Comparing the trade-off between parameters and performance among different detection models in the CMBs-Private testing dataset.
Bioengineering 11 00993 g009
Figure 10. Visual results of detection models in CMBs-Private testing dataset.
Figure 10. Visual results of detection models in CMBs-Private testing dataset.
Bioengineering 11 00993 g010
Figure 11. Comparing the trade-off between speed and performance in the CMBs-Private testing dataset.
Figure 11. Comparing the trade-off between speed and performance in the CMBs-Private testing dataset.
Bioengineering 11 00993 g011
Figure 12. Comparing the trade-off between parameters and performance in the CMBs-Private testing dataset.
Figure 12. Comparing the trade-off between parameters and performance in the CMBs-Private testing dataset.
Bioengineering 11 00993 g012
Figure 13. Comparing the trade-off between speed and performance among different models’ combinations in the CMBs-Private testing dataset.
Figure 13. Comparing the trade-off between speed and performance among different models’ combinations in the CMBs-Private testing dataset.
Bioengineering 11 00993 g013
Figure 14. The impact of CMBs-MHA on detection models at various scales in the CMBs-Private testing dataset.
Figure 14. The impact of CMBs-MHA on detection models at various scales in the CMBs-Private testing dataset.
Bioengineering 11 00993 g014
Figure 15. The loss function variation curve during training with the CMBs-Private training dataset.
Figure 15. The loss function variation curve during training with the CMBs-Private training dataset.
Bioengineering 11 00993 g015
Figure 16. Performance variations of different detection models under the 768-resolution training mode.
Figure 16. Performance variations of different detection models under the 768-resolution training mode.
Bioengineering 11 00993 g016
Figure 17. Impact of resolution changes on model prediction scores.
Figure 17. Impact of resolution changes on model prediction scores.
Bioengineering 11 00993 g017
Figure 18. Performance variations of different classification models in dynamic mode.
Figure 18. Performance variations of different classification models in dynamic mode.
Bioengineering 11 00993 g018
Figure 19. Different classification network structures based on CLIP.
Figure 19. Different classification network structures based on CLIP.
Bioengineering 11 00993 g019
Figure 20. Performance comparison of different strategies under CLIP.
Figure 20. Performance comparison of different strategies under CLIP.
Bioengineering 11 00993 g020
Table 1. The distribution of CMBs-Public dataset.
Table 1. The distribution of CMBs-Public dataset.
CMBs-PublicNumber of PatientsNumber of Patients with CMBsAverage Age of Patients with CMBsAverage Age of Patients without CMBsNumber of
Slices per Patient
Slice Image
Resolution
Number of
CMBs
Bounding Boxes
Male
Proportion
Female Proportion
Training set5037××35512 × 512353××
Validation set74××35512 × 51243××
Testing set159××35512 × 512102××
Table 2. The distribution of CMBs-Private dataset.
Table 2. The distribution of CMBs-Private dataset.
CMBs-PrivateNumber of PatientsNumber of Patients with CMBsAverage Age of Patients with CMBsAverage Age of Patients without CMBsNumber of
Slices per Patient
Slice Image
Resolution
Number of
CMBs
Bounding Boxes
Male
Proportion
Female Proportion
Training set4001786657130512 × 512801561%39%
Validation set50176861130512 × 51268754%46%
Testing set150566459130512 × 512243266%34%
Table 3. The comparison of different loss functions.
Table 3. The comparison of different loss functions.
Loss NameValue (A, B)Value (A, C)TrendValue (D, E)Value (D, F)Trend
IoU [12]0.290.06−0.230.920.85−0.07
GIoU [51]0.16−0.26−0.420.920.85−0.07
DIoU [12]0.290.06−0.230.920.85−0.07
CIoU [12]0.290.06−0.230.920.85−0.07
NWD0.90.8−0.10.90.8−0.1
Table 4. The format of text prompts and their corresponding descriptions.
Table 4. The format of text prompts and their corresponding descriptions.
Text PromptsDescriptions
This is a {male/female} patient with {} years old.Describe the patient’s gender and age.
Target length and width are {} and {} pixels.Describe the length and width of the detected CMBs.
The detection status of the {previous/next} frame at the same position is {yes/no}.Describe the detection status of CMBs in the previous and next slice images at the same position.
Table 5. The training configuration of detection models.
Table 5. The training configuration of detection models.
Method
Name
Model NameProposed YearBackboneDropout RateBatch SizeLearning RateEpoches
CNNFaster R-CNN [47]2015R-50 [55]0.080.00130
RetinaNet [56]2017R-500.080.00130
Cascade R-CNN [57]2018R-500.040.00145
CenterNet [58]2019R-500.080.00130
Dynamic R-CNN [59]2020R-500.080.00130
Sparse R-CNN [60]2021R-500.080.00130
ConvNeXt [61]2022T0.580.00130
S0.580.00130
TransformerDETR [48]2020R-500.040.000150
Deformable DETR [62]2021R-500.040.000150
DAB-DETR [63]2022R-500.040.000150
DINO [64]2023R-500.040.000150
YOLOYOLOv3 [65]2018DarkNet-530.080.00130
YOLOX [66]2021S0.080.00130
YOLOv82023N0.3320.00130
S0.3320.00130
M0.2160.00130
X60.280.00130
CMBs-YOLO N0.3320.00130
S0.3320.00130
M0.2160.00130
X60.280.00130
Table 6. The performance comparison in the CMBs-Private testing dataset.
Table 6. The performance comparison in the CMBs-Private testing dataset.
Method
Name
Model NameBackboneParameters
(MB)
PR m A P 50 m A P 50 95 Inference Time/ms
CNNFaster R-CNNR-5041.750.470.380.430.3252.96
RetinaNetR-5037.960.490.430.490.3647.34
Cascade R-CNN R-5069.390.560.490.510.4260.83
CenterNetR-5032.290.470.530.550.3625.32
Dynamic R-CNNR-5041.750.680.570.610.3834.17
Sparse R-CNN R-50106.290.750.630.680.4157.42
ConvNeXtT48.090.790.660.730.4332.95
S106.350.880.750.810.5265.13
TransformerDETRR-5041.580.840.770.840.4954.67
Deformable DETRR-5040.120.860.810.860.5265.36
DAB-DETRR-5044.270.890.830.870.5472.39
DINOR-5047.550.920.910.880.6447.69
YOLOYOLOv3DarkNet-5361.950.570.510.480.3932.54
YOLOXS8.970.810.720.750.4314.61
YOLOv8N2.870.720.630.660.426.2
S10.620.790.680.740.467.3
M24.660.880.710.770.498.9
X692.810.910.780.860.5513.1
CMBs-YOLON3.330.780.690.750.517.6
S12.440.870.790.830.558.4
M27.530.910.820.870.5711.3
X695.430.940.860.910.6214.6
Table 7. The performance comparison in the CMBs-Public testing dataset.
Table 7. The performance comparison in the CMBs-Public testing dataset.
Method
Name
Model NameBackbonePR m A P 50 m A P 50 95
CNNFaster R-CNNR-500.410.220.350.26
RetinaNetR-500.540.290.470.31
Cascade R-CNN R-500.580.510.460.32
CenterNetR-500.640.430.570.34
Dynamic R-CNNR-500.680.570.610.43
Sparse R-CNN R-500.680.390.650.52
ConvNeXtT0.770.470.650.53
S0.820.530.780.59
TransformerDETRR-500.640.410.570.52
Deformable DETRR-500.660.390.630.61
DAB-DETRR-500.680.430.640.61
DINOR-500.810.480.770.65
YOLOYOLOv3DarkNet-530.430.310.380.32
YOLOXS0.720.490.630.54
YOLOv8N0.610.370.580.49
S0.680.480.640.52
M0.750.520.710.57
X60.790.550.740.63
CMBs-YOLON0.660.430.610.55
S0.710.510.680.59
M0.780.570.730.61
X60.830.620.790.68
Table 8. The performance comparison of classification models.
Table 8. The performance comparison of classification models.
Method NameModel NameTPRTNRPParameter (MB)Inference Time/ms
3D-CNNR-180.640.690.1431.67.8
R-500.720.750.2360.510.1
R-1010.850.870.36117.921.8
TransformerUFv1_XXS0.830.840.4610.127.2
UFv1_XS0.840.850.4716.333.1
UFv2_B160.880.860.54108.536.7
UFv2_1140.890.880.56336.951.6
CFCF_S
w/o text
0.860.850.4926.6812.7
CF_B
w/o text
0.890.860.51107.1518.86
CF_S
w/text
0.920.910.5930.118.6
CF_B
w/text
0.940.930.63114.526.12
Table 9. Visual comparison of slice input images and texts in classification.
Table 9. Visual comparison of slice input images and texts in classification.
Input ImagesGenderAgeGround
Truth
R101UFv1_XSUFv2_114CF_B
w/o text
CF_B
w/text
Patient 1Bioengineering 11 00993 i001Female55Not
CMBs
CMBsCMBsCMBsCMBsNot
CMBs
Patient 2Bioengineering 11 00993 i002Male68CMBsCMBsCMBsCMBsCMBsCMBs
Patient 3Bioengineering 11 00993 i003Female57Not
CMBs
CMBsCMBsCMBsCMBsNot
CMBs
Table 10. The comparison of different models’ combinations.
Table 10. The comparison of different models’ combinations.
Detection
Model
Classification
Model
FPavg P p a t i e n t Total
Parameters/MB
Total
Inference Time/s
Cascade R-CNN (R50)R-10110.870.43187.2910.09
UFv1_XS7.980.5985.6910.63
UFv2_1147.060.62406.2913.07
CF_B
w/text
5.490.68183.8610.52
DAB-DETR
(R50)
R-1019.630.51162.1711.59
UFv1_XS7.140.6860.5712.13
UFv2_1145.670.69381.1714.57
CF_B
w/text
3.980.78158.7712.02
YOLOv8
(X6)
R-1018.190.54210.713.88
UFv1_XS6.330.67109.114.42
UFv2_1145.720.71429.716.86
CF_B
w/text
3.660.77207.314.32
CMBs-YOLO
(X6)
R-1018.420.58213.034.08
UFv1_XS6.970.69111.734.62
UFv2_1146.210.73432.337.06
CF_B
w/text
3.370.82209.934.51
Table 11. Instances of incorrect judgments in CMB classification stage.
Table 11. Instances of incorrect judgments in CMB classification stage.
ImageEnlarged ViewAgeGenderLifestyle HabitsMedical HistoryClinical Diagnostic ResultsOur
Results
Patient 1Bioengineering 11 00993 i004Bioengineering 11 00993 i00553Female × CADASILCMBsNot CMBs
Patient 2Bioengineering 11 00993 i006Bioengineering 11 00993 i00755MaleDrinking and SmokingHypertensionCMBsNot
CMBs
Patient 3Bioengineering 11 00993 i008Bioengineering 11 00993 i00978Male × × Cerebral Calcification (Not CMBs)CMBs
Table 12. The impact of CMBs-MHA and NWD on the detection model.
Table 12. The impact of CMBs-MHA and NWD on the detection model.
Model NameBackboneBox LossPR m A P 50 m A P 50 95
YOLOv8N0.490.720.630.660.42
S0.470.790.680.740.46
M0.430.880.710.770.49
X60.400.910.780.860.55
CMBs-YOLO
w/o
NWD
N0.430.760.680.720.52
S0.410.860.780.810.53
M0.360.900.790.850.55
X60.360.920.870.890.59
CMBs-YOLO
w/o
CMBs-MHA
N0.430.730.640.670.44
S0.420.810.670.750.47
M0.370.860.730.780.48
X60.330.880.790.850.52
CMBs-YOLON0.390.780.690.750.54
S0.370.870.790.830.55
M0.320.910.820.870.57
X60.310.940.860.910.62
Table 13. The performance comparison of input resolutions.
Table 13. The performance comparison of input resolutions.
Model NameBackboneTrain ResolutionValidation ResolutionPR m A P 50 m A P 50 95
ConvNeXtS 512 × 512 256 × 256 0.800.660.760.49
512 × 512 0.830.750.810.52
768 × 768 0.880.770.820.54
768 × 768 256 × 256 0.820.570.740.43
512 × 512 0.860.740.820.52
768 × 768 0.890.810.850.57
DINOR50 512 × 512 256 × 256 0.910.850.860.61
512 × 512 0.920.890.880.64
768 × 768 0.920.880.870.64
768 × 768 256 × 256 0.890.880.860.63
512 × 512 0.920.910.870.64
768 × 768 0.930.910.890.66
YOLOv8X6 512 × 512 256 × 256 0.830.740.760.52
512 × 512 0.910.780.860.55
768 × 768 0.920.780.850.54
768 × 768 256 × 256 0.810.720.750.49
512 × 512 0.880.790.850.56
768 × 768 0.920.810.880.57
CMBs-YOLOX6 512 × 512 256 × 256 0.880.810.850.57
512 × 512 0.930.860.910.62
768 × 768 0.940.870.900.61
768 × 768 256 × 256 0.860.830.830.55
512 × 512 0.930.880.900.62
768 × 768 0.950.890.920.62
Table 14. The comparison results of 2D images input in classification.
Table 14. The comparison results of 2D images input in classification.
Method NameModel NameTraining
Mode/Resolution
Testing
Mode/Resolution
TPRTNRP
3D-CNNR-101Fixed 32 × 32 × 32 Fixed 32 × 32 × 32 0.690.710.31
Fixed 64 × 64 × 64 0.830.840.34
Dynamic0.820.860.33
Fixed 64 × 64 × 64 Fixed 32 × 32 × 32 0.790.820.32
Fixed 64 × 64 × 64 0.850.870.36
Dynamic0.840.830.35
DynamicFixed 32 × 32 × 32 0.720.690.34
Fixed 64 × 64 × 64 0.780.760.40
Dynamic0.760.740.38
TransformerUFv2_114Fixed 32 × 32 × 32 Fixed 32 × 32 × 32 0.770.790.39
Fixed 64 × 64 × 64 0.810.830.42
Dynamic0.800.840.41
Fixed 64 × 64 × 64 Fixed 32 × 32 × 32 0.840.830.49
Fixed 64 × 64 × 64 0.890.880.56
Dynamic0.850.810.51
DynamicFixed 32 × 32 × 32 0.780.790.43
Fixed 64 × 64 × 64 0.860.830.52
Dynamic0.850.860.53
CFCF_B
w/o text
Fixed 32 × 32 × 32 Fixed 32 × 32 × 32 0.820.810.41
Fixed 64 × 64 × 64 0.830.820.41
Dynamic0.810.800.39
Fixed 64 × 64 × 64 Fixed 32 × 32 × 32 0.830.820.45
Fixed 64 × 64 × 64 0.890.860.51
Dynamic0.840.810.41
DynamicFixed 32 × 32 × 32 0.790.750.38
Fixed 64 × 64 × 64 0.830.790.43
Dynamic0.820.800.42
CF_B
w/text
Fixed 32 × 32 × 32 Fixed 32 × 32 × 32 0.880.870.56
Fixed 64 × 64 × 64 0.890.880.58
Dynamic0.850.870.55
Fixed 64 × 64 × 64 Fixed 32 × 32 × 32 0.900.890.57
Fixed 64 × 64 × 64 0.940.930.63
Dynamic0.920.910.59
DynamicFixed 32 × 32 × 32 0.870.850.55
Fixed 64 × 64 × 64 0.910.880.56
Dynamic0.910.870.58
Table 15. The comparison results of 2D images input in classification.
Table 15. The comparison results of 2D images input in classification.
Method NameClassification
Model Name
TPRTNRP
2D-CNNR-180.520.490.07
R-500.610.620.15
R-1010.730.760.23
3D-CNNR-180.640.690.14
R-500.720.750.23
R-1010.850.870.36
2D-CFCF_S
w/text
0.870.840.55
CF_B
w/text
0.890.870.57
3D-CFCF_S
w/text
0.920.910.59
CF_B
w/text
0.940.930.63
Table 16. The difference between 2D and 3D images.
Table 16. The difference between 2D and 3D images.
2D Input Image
(Averaging)
3D Input ImagesGround
Truth
2D-R1013D-R1012D-CF_B3D-CF_B
Case 1Bioengineering 11 00993 i010Bioengineering 11 00993 i011CMBsNot
CMBs
CMBsCMBsCMBs
Case 2Bioengineering 11 00993 i012Bioengineering 11 00993 i013Cerebral Calcification
(Not CMBs)
CMBsNot CMBsCMBsNot
CMBs
Case 3Bioengineering 11 00993 i014Bioengineering 11 00993 i015Cerebral Calcification
(Not CMBs)
CMBsNot CMBsCMBsNot
CMBs
Table 17. The comparison of different strategies in the CMBs-Private testing dataset.
Table 17. The comparison of different strategies in the CMBs-Private testing dataset.
Detection ModelClassification ModelComparison StrategyTPRTNRPFPavg P p a t i e n t
Cascade
R-CNN (R50)
CF_B10.760.750.465.920.61
20.710.730.416.310.58
30.720.710.386.870.53
40.790.830.495.490.68
DAB-DETR
(R50)
CF_B10.810.800.524.120.73
20.760.810.494.370.69
30.750.770.464.920.66
40.840.860.533.980.78
YOLOv8
(X6)
CF_B10.850.850.514.070.74
20.820.860.484.220.71
30.810.830.464.560.68
40.880.910.543.660.77
CMBs-YOLO
(X6)
CF_B10.910.860.583.610.78
20.860.870.563.770.76
30.870.840.554.190.73
40.940.930.633.370.82
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, C.; Zhao, L.-L.; Lang, Q.; Xu, Y. A Novel Detection and Classification Framework for Diagnosing of Cerebral Microbleeds Using Transformer and Language. Bioengineering 2024, 11, 993. https://doi.org/10.3390/bioengineering11100993

AMA Style

Chen C, Zhao L-L, Lang Q, Xu Y. A Novel Detection and Classification Framework for Diagnosing of Cerebral Microbleeds Using Transformer and Language. Bioengineering. 2024; 11(10):993. https://doi.org/10.3390/bioengineering11100993

Chicago/Turabian Style

Chen, Cong, Lin-Lin Zhao, Qin Lang, and Yun Xu. 2024. "A Novel Detection and Classification Framework for Diagnosing of Cerebral Microbleeds Using Transformer and Language" Bioengineering 11, no. 10: 993. https://doi.org/10.3390/bioengineering11100993

APA Style

Chen, C., Zhao, L. -L., Lang, Q., & Xu, Y. (2024). A Novel Detection and Classification Framework for Diagnosing of Cerebral Microbleeds Using Transformer and Language. Bioengineering, 11(10), 993. https://doi.org/10.3390/bioengineering11100993

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop