Next Article in Journal
Recent Advancements for the Treatment of Hepatocellular Carcinoma
Next Article in Special Issue
Empowering Vision Transformer by Network Hyper-Parameter Selection for Whole Pelvis Prostate Planning Target Volume Auto-Segmentation
Previous Article in Journal
In Vitro and In Silico Investigation of BCI Anticancer Properties and Its Potential for Chemotherapy-Combined Treatments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Detection of Brain Metastases in T1-Weighted Construct-Enhanced MRI Using Deep Learning Model

1
School of Mechanical, Electrical and Information Engineering, Shandong University, Weihai 264209, China
2
Department of Radiation Oncology and Physics, Shandong Cancer Hospital and Institute, Shandong First Medical University and Shandong Academy of Medical Sciences, Jinan 250117, China
3
Laboratory of Image Science and Technology, School of Computer Science and Engineering, Southeast University, Nanjing 210096, China
4
Department of Oncology, Binzhou People’s Hospital, Binzhou 256610, China
5
Third Clinical Medical College, Xinjiang Medical University, Urumqi 830011, China
*
Authors to whom correspondence should be addressed.
Cancers 2023, 15(18), 4443; https://doi.org/10.3390/cancers15184443
Submission received: 31 July 2023 / Revised: 3 September 2023 / Accepted: 4 September 2023 / Published: 6 September 2023
(This article belongs to the Special Issue Artificial Intelligence and Deep Learning in Radiology Oncology)

Abstract

:

Simple Summary

In this research, we introduced an improved deep learning model for automatic brain metastases detection in MRI. In order to reduce false-positive results while retaining high accuracy, a modified YOLOv5 algorithm with self-attention mechanism is proposed. Our proposed deep learning model showed promising results on the internal testing set, achieving better recall and precision compared to standard algorithms. Furthermore, we also demonstrated the method’s effectiveness and generalizability on the external testing set. The study proved that our proposed framework can be used as a reliable computer-aided diagnosis system for brain metastases detection.

Abstract

As a complication of malignant tumors, brain metastasis (BM) seriously threatens patients’ survival and quality of life. Accurate detection of BM before determining radiation therapy plans is a paramount task. Due to the small size and heterogeneous number of BMs, their manual diagnosis faces enormous challenges. Thus, MRI-based artificial intelligence-assisted BM diagnosis is significant. Most of the existing deep learning (DL) methods for automatic BM detection try to ensure a good trade-off between precision and recall. However, due to the objective factors of the models, higher recall is often accompanied by higher number of false positive results. In real clinical auxiliary diagnosis, radiation oncologists are required to spend much effort to review these false positive results. In order to reduce false positive results while retaining high accuracy, a modified YOLOv5 algorithm is proposed in this paper. First, in order to focus on the important channels of the feature map, we add a convolutional block attention model to the neck structure. Furthermore, an additional prediction head is introduced for detecting small-size BMs. Finally, to distinguish between cerebral vessels and small-size BMs, a Swin transformer block is embedded into the smallest prediction head. With the introduction of the F2-score index to determine the most appropriate confidence threshold, the proposed method achieves a precision of 0.612 and recall of 0.904. Compared with existing methods, our proposed method shows superior performance with fewer false positive results. It is anticipated that the proposed method could reduce the workload of radiation oncologists in real clinical auxiliary diagnosis.

1. Introduction

Brain metastasis (BM) is a severe complication of malignant tumors. Approximately 20% of cancer patients develop BM [1]. BM is common among patients with lung cancer, occurring in up to 50% of patients with advanced non-small-cell lung cancer (NSCLC) [2]. This is primarily attributed to advancements in systemic therapies that enable superior control of extracranial neoplasms, thereby improving survival rates and increasing the opportunities for BM development [2,3]. BM developed from NSCLC is linked to poor prognosis, with a median survival of just a few months if left untreated [4]. The symptoms of BM can range from mild headache and cognitive impairment to seizures, focal neurological deficits, and even coma [5]. These symptoms can significantly impact the quality of life and survival period of NSCLC patients. Therefore, early diagnosis and treatment of NSCLC BM is essential to elevate survival rates and improve prognosis.
Radiotherapy is an effective method of treating BM patients and prolonging their survival [6]. Patients with a limited number of BMs can be treated with stereotactic radiosurgery (SRS), which can improve qualities of life and avoid the cognitive decline caused by whole-brain radiation therapy [7,8]. Prior to SRS treatment, accurate detection of BMs, and subsequent delineation of regions of interest (ROI) are necessary. The detection task is mainly performed by experienced radiation oncologists, and remain challenging due to the blurred boundaries and variable quantity of BMs. Therefore, fast and accurate detection of BMs in MRI images with the assistance of deep learning (DL) is of great clinical importance.
In recent years, DL algorithms with a full convolution structure have been proposed for automatic detection of BMs. Zhou et al. [9] used a single-shot detector (SSD) [10] neural network to automatically detect BMs in T1-weighted MRI datasets, achieving a fast and accurate detection capability with a sensitivity of 81% and a positive predictive value (PPV) of 36%. However, this method produced many false positive results that regarded healthy brain tissue as BMs. Using a multi-level feature-fusion (FF) technique, Amemiya et al. [11] developed a modified SSD structure called FF-SSD. Experimental results showed that FF-SSD achieved an overall sensitivity of 86.0% (an increase of 2.2% compared to SSD) and a PPV of 46.8% (an increase of 1.6% compared to SSD). Although FF-SSD can improve the sensitivity and PPV for small-size BMs, it is relatively limited and produces many false positive results. Other approaches rely on 3D convolutional structures to extract the image’s contextual and spatial information for direct detection and segmentation. Using different kernel sizes with the vanilla UNet, Cao et al. [12] compared the asymmetric 3D UNet and proposed asym-UNet to extract the region features of small BMs and the boundary information of large BMs. This study proves that smaller kernels perform better in detection sensitivity while performing worse in PPV. Li et al. [13] developed a two-stage 3D DL model for automatic detection and simultaneous segmentation of BMs. This novel model obtained a sensitivity of 90% and precision of 56%. Although 3D CNN models have achieved excellent performances, they may be difficult to implement on lower-performance computers used in actual clinical practice.
As a DL algorithm, You Only Look Once version 5 (YOLOv5) [14] has achieved excellent precision and recall, and has been applied in the medical image domain for the automatic detection of stroke lesions [15] and polyps [16] as well as the classification of skin cancer [17], among other applications. The effectiveness of these approaches indicates that YOLOv5 can provide image-guided clinical diagnosis and has great potential for new applications in clinical practice. In order to meet the objective demands of BM detection, the existing published works have obtained balance in the ability of their models by adjusting the confidence threshold parameters. However, certain models exhibit limited detection capabilities, making it challenging to achieve optimal detection without encountering a significant number of false positive results. Due to the varied size and blurred boundaries of BMs, developing an automatic algorithm to reduce false positive results while maintaining excellent sensitivity is quite challenging. In view of this challenge, a modified YOLOv5 called SA-YOLOv5 (self-attention YOLOv5) is proposed in this study.
On the basis of YOLOv5, SA-YOLOv5 adds an attention mechanism and transformer block to enhance the features of the BM region and weaken the features of vascular shadows and gray matter. For the accurate detection of small BMs, an additional prediction head is added to SA-YOLOv5. Meanwhile, the F2-score is introduced to solve the difficulty of determining the confidence threshold to ensure more appropriate detection performance. To demonstrate the detection ability of SA-YOLOv5, we compared it with four existing methods on the internal testing set and validated the method’s effectiveness and generalizability on the external testing set.

2. Materials and Methods

2.1. Study Participants

Approval to carry out this study was received from the institutional review board, and the identification information of the enrolled patients was anonymized; the requirement to obtain written informed consent was waived due to the retrospective nature of the study. In all, 335 lung cancer patients who underwent radiotherapy for BMs at Shandong Cancer Hospital and Institute from August 2018 to December 2020 were enrolled, along with 71 lung cancer patients who underwent radiotherapy for BMs at Binzhou People’s Hospital from April 2022 to December 2022. As shown in Figure 1, after patient exclusion, a total of 240 and 25 patients were finally included in the internal set and external set, respectively. All patients in the internal set were imaged using a GE Discovery MR 750W scanner with six channels of head coils in the same posture. The patients in the external set were all imaged using a Siemens Magentom Avanto scanner.

2.2. Dataset Construction

All these patients had complete single-sequence T1-weighted contrast-enhanced (T1ce). The MRI images were manually annotated by two oncologists with five and six years of experience, respectively, utilizing MIM Maestro 6.8.2 software. Each oncologist annotated all images, while another oncologist with sixteen years of experience modified the mask if necessary and confirmed any inconsistent areas. There were a total of 1030 BMs in the internal dataset, for average of 4.3 BMs per patient. The external set included 130 BMs, for an average of 5.2 BMs per patient. The patients in internal set were randomly divided into a training group and a testing group at a ratio of 8:2 (192:48). To make the SA-YOLOv5 framework focus on the characteristics of BM regions, the slices of patients without BM regions were eliminated in the training group, preserving 2269 images. These preserved images were divided into a training set and a validation set at a ratio of 8:2. The validation set was used to verify the model’s performance and save the best weights of the DL method. The internal testing set used for model evaluation consisted of whole MRI images from the 48 patients in the testing group. The respective distributions of BMs in the training group and internal testing set are shown in Figure 2.
The mask of each BM was projected based on the transverse section direction and the maximum diameter of each BM was determined by calculating the minimum circumcircle of the projection using the opencv_python (version 4.2.0) software package while considering the pixel scale with assistance from the pydicom (version 2.2.0) software package. As shown in Figure 3, the distribution of BM maximum diameters is relatively concentrated. The proportion of BMs with a maximum diameter of 0–10 mm is 55.8%. Small BMs typically have poorly defined boundaries and low contrast. In regular item training, the precision (P) with respect to small targets is generally much lower than that for medium and large targets. Moreover, the distribution of small BMs is not uniform. These two issues complicate BM detection.

2.3. Dataset Ground Truth

The bounding box for each BM had a unique coordinate format produced from the corresponding segmentation masks. In order to make the bounding box properly represent the ROI of BM, an iterative algorithm was used to find the outermost pixel values of the segmentation mask to build the bounding box, making the produced bounding boxes were reliable enough to complete the training and testing of the proposed model.

2.4. Image Preprocessing

The MRI images were cut into 384 × 384 slices and processed via mosaic data enhancement. The mosaic data augmentation technique utilizes random scaling, cropping, and arrangement to splice four images into a single large image [18]. The advantages of this method are that the background of the image is enriched and the batch size is increased through concatenation of the four original images.

2.5. Architecture of SA-YOLOv5

As illustrated in Figure 4, the main architecture of SA-YOLOv5, including the backbone, neck, and multi-head, was implemented by using Pytorch (https://pytorch.io (accessed on 1 October 2022)) [19]. The proposed SA-YOLOv5 utilized in this paper made the following improvements to YOLOv5 [14]. First, an additional prediction head was added for more accurate detection of small BMs. Second, a self-attention block [20] was introduced to enhance the channel of the effective features. Third, a Swin transformer block [21] was added to strengthen the attention to small targets on the feature map. The main modules in the SA-YOLOv5 architecture are described in detail below.

2.5.1. Backbone

The backbone consists of a Focus block, Conv block, C3 block, and SPPF block. The Focus block is implemented using a 6 × 6 convolutional block, while the Conv block is composed of a 1 × 1 convolutional block, batch normalization unit, and SiLU. As shown in Figure 5, the C3 block is constructed via two residual structures composed of several convolutional blocks. The Conv1 block is composed of a 1 × 1 convolutional block, batch normalization unit, and SiLU, while the Conv3 block is composed of a 3 × 3 convolutional block, batch normalization unit, and SiLU. Furthermore, the SPPF block, shown in Figure 6, consists of the Conv block and MaxPool layers. The feature maps output by the MaxPool layers are concatenated together and then reshaped by the Conv block.

2.5.2. Neck

As shown in Figure 4, a path aggregation network (PAN) [24] structure is added to the neck in combination with the feature pyramid network (FPN) [25] structure to form a powerful feature fusion layer in the neck. Furthermore, a convolutional block attention module (CBAM) [20] is added to the neck to capture important information on small BMs. The CBAM contains a Channel Attention Module (CAM) and a Spatial Attention Module (SAM), which can be mathematically expressed as
M c F = σ M L P A v g P o o l F + M L P M a x P o o l F
M s F = σ f 7 × 7 A v g P o o l F ; M a x P o o l F
where M c F and M s F respectively represent the feature maps of the channel attention map and spatial attention map, σ is the sigmoid function, A v g P o o l and M a x P o o l respectively represent the global average pooling and global max pooling, F denotes the input feature map, M L P denotes multi layer perceptron, and f 7 × 7 denotes the convolutional block with a filter size of 7 × 7 .
The CBAM derives the attention map from the given feature map along the channel and space dimensions, then multiplies the attention map by the input feature map for adaptive feature refinement, thereby enhancing the contribution of informative feature channels and weakening the interference of useless channels. The CAM plays a crucial role in integrating relevant information during the merging process of the feature map. As shown in Figure 7, the feature map initially undergoes maximum pooling and average pooling based on its width and height. Subsequently, a multi-layer perceptron (MLP) with shared weights is employed. The outputs from the MLP layer are summed together pixel-wise, resulting in the channel attention map M c F through activation via the sigmoid function. After the CAM, the SAM is used to focus on where the most meaningful features come from. The role of the SAM is to accurately capture and represent the spatial information inherent in the feature map. The SAM takes the output of the CAM as its input and processes it through maximum pooling and average pooling. The feature map is merged into a feature map with two channels, then passes through a 7 × 7 convolutional layer to reduce this to one channel, resulting the spatial attention map M s F being obtained through the sigmoid function.

2.5.3. Multi-Head Attention

The prediction head in the multi-head structure is a decoder that retains significant the spatial structure information of the corresponding original image on the feature map. Its main structure is a classification layer composed of 3 × 3 and 1 × 1 convolutional blocks. As shown in Figure 4, the prediction head 1 uses the 12 × 12 feature map as input; an element in the 12 × 12 feature map corresponds to a pixel region of 32 × 32 in the 384 × 384 input image. Accordingly, the element in the feature map of the prediction head 2 and 3 corresponds to a pixel region of 16 × 16 and 8 × 8 , respectively. In order to effectively detect small BMs, we added an additional small object prediction head combining a Swin transformer block (STB) [21] and 1 × 1 convolutional blocks, which is called a Swin transformer prediction head (STPH). The prediction head 4 employs the 96 × 96 feature map as input, and every element in it corresponds to a 4 × 4 pixel region of the original image.
Unlike the common multi-head self attention module, the STB is formed based on the concept of shifted windows [26]. As shown in Figure 8, two consecutive blocks are in a string connection, with each block is composed of a multilayer perceptron (MLP) with a SiLU, LayerNorm (LN) layer, and multi-head self attention module. The windows multi-head self-attention (W-MSA) module and shifted windows multi-head self-attention (SW-MSA) module are respectively contained in these two blocks. The STB employs W-MSA and SW-MSA to change the regular window partitioning method, instead using the more efficient shifted window partitioning strategy shown in Figure 9. Not only does this maintain efficient computation of non-overlapping windows, it establishes connections between different windows. In the W-MSA module, the feature map is divided into 8 × 8 windows and the self-attention operation is performed within each window. The purpose of the SW-MSA module is to realize information exchange between different windows. The attention window shifts between different attention heads, meaning that each attention head focuses on a different window, thereby increasing the diversity of features. The STB can be mathematically expressed as follows:
a ^ l = W M S A ( L N ( a L 1 ) ) + a L 1
a l = M L P ( L N ( a ^ l ) ) + a ^ l
a ^ l + 1 = S W M S A ( L N ( a l ) ) + a l
a l + 1 = M L P ( L N ( a ^ l + 1 ) ) + a ^ l + 1
where a ^ l , a l represent the outputs of the self-attention module and the MLP module of the l t h block, respectively.
Unlike the normal multi-head self-attention computation, the relative position bias term B is included in the self-attention calculation of each head. The self-attention is computed as follows:
A t t e n t i o n ( Q , K , V ) = S o f t m a x ( Q K T d + B ) V
where Q, K, V, and B respectively denote the query, key, value, and bias matrices in the self-attention mechanism.

2.6. Boundary Loss Function

The most common method for punishing imprecise prediction bounding boxes is to employ the Intersection over Union ( I o U ) loss function [27] or its improved variant. As shown in Figure 10a, the I o U is used to quantify the ratio of the intersection and concatenation of prediction bounding box (P) and ground truth bounding box (G). The I o U and L o s s I o U can be expressed as follows:
I o U = P G P G
L o s s I o U = 1 I o U
where P and G denote the prediction bounding box and the ground truth bounding box, respectively, and L o s s I o U is the IoU loss function.
When P and G do not intersect and I o U ( P , G ) = 0 , the gradient cannot be transferred back and the L o s s I o U cannot optimize this situation. To address this flaw, the Complete IoU ( C I o U ) loss function [28] is utilized for bounding box regression, leading to faster convergence and better performance than L o s s I o U . The C I o U , as illustrated in Figure 10b, introduces a penalty term R = ρ 2 c 2 + α v such that the I o U , center point Euclidean distance, and aspect ratio are included in the consideration range. The L o s s C I o U can be mathematically expressed as follows:
L o s s C I o U = 1 I o U + ρ 2 c 2 + α v
v = 4 π 2 ( a r c t a n w G h G a r c t a n w P h P ) 2
α = v 1 I o U + v
where α denotes the weight parameter of v, v is used to measure the similarity of the aspect ratio of P and G, ρ is the Euclidean distance between P 0 and G 0 , c represents the diagonal length of the minimum outer rectangle between P and G, and w G , w P , h G , and h P respectively denote the widths and heights of G and P.

2.7. Training Configuration and Procedure

Data augmentation was only performed on the training set. The validation set and testing set usd the original data. The weights of the convolutional kernels were initialized by loading the pretraining weight file of the YOLOv5 model. The CIoU loss function [28] was selected as the penalty measurement index of the bounding boxes. An SGD (stochastic gradient descent) optimizer [29] was used to train the network. The initial learning rate was set to 0.01 and reduced by 10% if the validation loss did not improve after three epochs. The training procedure was terminated if the validation loss did not improve within twenty epochs. The model was trained on a Geforce GTX 1080Ti with a memory of 11 GB (NVIDIA, Santa Clara, CA, USA).

2.8. Postprocessing

The output prediction results from the prediction heads are processed by means of non-maximum suppression (NMS) [30]. The basic idea behind NMS is to select the bounding box with the highest confidence score and remove all other bounding boxes that have a significant overlap with it. This is repeated for all remaining bounding boxes until there are no more overlaps. The prediction results after NMS represent the final output of the model.

2.9. Model Evaluation

Each bounding box generated by the detection algorithm is accompanied by a confidence score that ranges from 0% to 100%. The confidence threshold is used to ignore bounding boxes with a confidence value less than the confidence threshold. Previous BM detection studies have used a confidence threshold of 50% [9,31,32] or confidence thresholds ranging from 0.1 to 0.9 [11]; however, these approach may not lead to optimal results for BM detection. Other object detection research has utilized the F1-score, which represents the harmonic mean of precision and recall, to determine the optimal confidence threshold [33,34,35]. As the recall is more important than the precision in BM detection, we introduced the F2-score, which emphasizes the importance of recall by assigning it twice the weight of precision, to determine the optimal confidence threshold.
Although DL detection algorithms for slice-by-slice prediction have been developed, the final performances of these models is evaluated in three-dimensional space. The neighboring bounding boxes of the final prediction are stacked to form a volume, which is considered to be a true positive result if it overlaps with the ground truth. On the other hand, the prediction volume is considered a false positive result if there is no voxel overlap with the ground truth volume.
The evaluation metrics adopted in this study included recall (R), precision (P), and F 2 s c o r e . The precision is the proportion of the number of true positive results to the number of all predicted results in the testing set. The recall is the proportion of the number of true positive results to the number of all BMs in the testing set. These evaluation metrics were defined as follows:
P = T P T P + F P
R = T P T P + F N
F 2 s c o r e = 5 × P × R 4 × P + R
where T P , F P , and F N represent the number of true positives, false positives, and false negatives, respectively.

3. Results

In this section, the prediction results of our proposed model and other state-of-the-art deep learning models (Faster R-CNN [36], SSD [9], FF-SSD [11], and EfficientDet [37]) are compared on the testing set. During our experiments, we strictly followed the principle of randomization when partitioning the dataset. In addition, the dataset was partitioned based on patients rather than on individual BMs, resulting in a more varied distribution of the number and size of BMs in the training set and testing set. This approach produces a testing environment that is closer to the clinical situation.

3.1. Detection Performance of SA-YOLOv5

The representative testing images in Figure 11 and Figure 12 show true positive, false positive, and false negative inferences from the proposed model. The red bounding boxes represent the ground truth, while the yellow bounding boxes represent the model prediction results. As shown in Figure 11 and Figure 12, the proposed SA-YOLOv5 model can detect almost all BMs with regular volumes. However, blood vessels and calcification in MRI images can cause interference during the BM detection process, resulting in false negative and false positive results.

3.2. Comparison with Existing Detection Methods

Table 1 lists the results of the proposed model and four existing models used for comparison. From the table, it can be seen that the recall achieved by Faster R-CNN [36] is 0.690, which is 0.348 higher than its precision. As Faster RCNN is a two-stage detector, its detection speed is relatively slow. Meanwhile, there is a contradiction between semantics and space, making it more difficult to achieve a good balance between deep and shallow feature maps. EfficientDet [37] is the only model with a recall lower than its precision, achieving a precision of 0.669 and recall of 0.614. In EfficientDet, the feature map’s receptive field of feature points is much larger than the downsampling rate. As a result, there are many features from surrounding regions at the points of the feature map; small BMs occupy fewer features, which affects the BM detection results. The performance evaluation metrics of SSD were imbalanced as well, with a recall of 0.822 and precision of 0.369. Although SSD uses multi-layer feature maps, the semantic information of the shallow feature maps is insufficient and feature fusion is not carried out, resulting in poor BM detection results. In contrast to SSD [9], FF-SSD [11] adds a multi-level feature fusion technique to enhance the feature maps of small object, resulting in higher precision and more true positive results. Overall, while SSD and FF-SSD exhibit high accuracy in identifying BMs, they have limitations due to false positive results. Among all the models, our proposed model achieves a recall of 0.904 and precision of 0.612. This suoerior performance may be attributed to the effectiveness of the proposed preprocessing techniques and feature fusion layer. Obviously, the detection results focused more on the recall and slightly sacrifice the precision, as this is more suitable for clinical applications in auxiliary diagnosis.
To more intuitively compare the performance of the different detection models, Figure 13 displays the ground truth and prediction results on the original MRI images of two distinct patients. The predicted results of these two patients are visualized in 2D and 3D, respectively. The red bounding boxes represent the ground truth, while the yellow bounding boxes represent the model prediction results. Notably, EfficientDet [37] and Faster R-CNN [36] do not perform well in detecting BMs. EfficientDet [37] generates fewer prediction bounding boxes, resulting in many undetected lesions. Conversely, Faster R-CNN identifies many vascular regions as BMs, leading to a high number of false positive results. Compared with EfficientDet [37] and Faster RCNN [36], SSD [9], FF-SSD [11], and our proposed model all demonstrate superior detection performance. In particular, our proposed model successfully detects all the BMs distributed in the original images, with arrows indicating possible lesions around the yellow bounding box. A magnified view of this region is visible in the upper right corner of the original image. When re-examined by two oncologists and one physicist, this was revealed to be an undiscovered BM with similar characteristics to those of blood vessels. Compared to other models, SA-YOLOv5 does not misidentify so many vascular regions as BMs, and shows accurate identification ability.

3.3. Effectiveness Analysis of the Improvements Made in SA-YOLOv5

To verify the effectiveness of the improvements made in SA-YOLOv5, the prediction performance of YOLOv5 [14], YOLOv5 + CBAM, YOLOv5 + CBAM + PH, and YOLOv5 + CBAM + STPH are compared in this section. The precision, recall, F2-score, FN per patient, and FP per patient of these models are listed in Table 2. It can be seen from the table that the CBAM block added to the neck structure improved the recall from 0.883 to 0.898, while the precision was decreased slightly. The model produced 177 true positive results and 119 false positive results. These results indicate that CBAM pays more attention to the boundary information of BMs, effectively increasing true positive results while resulting in the misidentification of a small number of vascular areas as BMs. The inclusion of the additional prediction head (STPH without STB) led to an increase in the model’s bounding box output; however, it did not yield improved prediction results. Moreover, it treated a number of small calcification or vascular structures as BMs. The addition of STB enhances the detection ability. The STPH layer is regarded as a detection layer for extremely small BMs, and effectively strengthens the distinction between BMs and the background. YOLOv5 + CBAM + STPH achieved state-of-the-art performance, increasing the recall by 2.1% while ensuring that the precision remained almost unchanged from the baseline. Moreover, it produced the most true positive results. These results indicate that the addition of the CBAM and STPH can effectively improve the performance of YOLOv5 on BM detection.

3.4. Detection Performance on the External Testing Set

To assess the proposed model’s generalizability, we evaluated it on the external testing set. Our proposed model achieved a recall of 0.854 (111/130) and precision of 0.681 (111/163) and had excellent detection results on images obtained with different scanning equipment at different institutions, demonstrating the feasibility of the proposed YOLOv5-based BM detection model. As shown in Figure 14, there were a number of false negative and false positive detection results on the external testing data. In particular, the detection of small BMs partially close to blood vessels is not effective, as these were regarded as negative samples. This may be attributed to the similarity in characteristics between these types of BMs in the external testing data and calcification structures in the training data. In addition, the better precision observed on the external testing data may be attributed to the relatively higher contrast of these images compared to the internal data.

4. Discussion

The improved YOLOv5 algorithm proposed in this paper, called SA-YOLOv5, shows promising results in assisting detection of NSCLC brain metastases. The performance of the proposed model was rigorously evaluated on the internal and external testing sets. Our experimental results show that SA-YOLOv5 can detect nearly all BMs with maximum diameters of 0.5 cm or larger while limiting false positive results. The proposed SA-YOLOv5 achieved a precision of 0.612 and a recall of 0.904 on the internal testing set. Compared to the compared methods, our model demonstrates a significant improvement in both recall and precision. Furthermore, upon evaluating the detection results with two oncologists and one physicist, we discovered three small BMs that were initially overlooked. Our proposed model maintained its BM detection ability on the external testing set, achieving a precision of 0.681 and a recall of 0.854. Overall, SA-YOLOv5 effectively achieves balanced performance in auxiliary diagnosis while avoiding interference from false positive results.
This study presents several limitations that are manifested in the dataset and detection performance. The quality of MRI images varies due to differences in scanning equipment and the parameters used during scanning. Additionally, the sample distribution of patients in different regions may have an impact on the testing results. In future work, continuing to collect MRI data for model training from multiple institutions and scanners represents an effective method to improve the model performance varied data. While the proposed model shows enhanced detection capacity for small BMs, but it can nonetheless generate false negative and false positive results for such lesions. Additionally, the quality of dataset delineation imposes certain restrictions on the model. Even though the delineation outcomes were reviewed by multiple radiologists, BMs may have been missed. The model could be further refined by incorporating additional MRI data on BMs in order to facilitate the detection of exceedingly small BMs and further improve the model’s generalization and robustness.
The current DL models for the segmentation of multiple BMs exhibit considerable scope for improvement in both performance and generalization ability. In addition to automatic detection, future research might focus on developing end-to-end segmentation approaches based on DL. To this end, doctors’ diagnostic experience should be incorporated as prior knowledge to constrain the model and optimize its loss function. Ultimately, an automated BM detection and segmentation model with high accuracy and speed can greatly enhance radiologists’ work efficiency.

5. Conclusions

In this paper, we propose a novel BM detection algorithm based on a self-attention mechanism, which we call SA-YOLOv5. To explore the potential information of the feature map, we employ the CBAM and STPH to capture the features of small BMs. Our experimental results demonstrate that the proposed model attains comparatively well-rounded BM detection performance and that the CBAM and STPH effectively improve the performance of YOLOv5. We believe that our proposed framework can be used as a reliable computer-aided diagnosis system for BM detection using T1ce.

Author Contributions

Conceptualization, Z.Z., Q.Q., T.L. and L.X.; Methodology, Z.Z.; Software, Z.Z.; Validation, Z.Z., H.L., X.G. and R.Y.; Formal analysis, Z.Z. and Q.Q.; Investigation, Z.Z. and Q.Q.; Data curation, Q.Q., H.L., X.G., T.L., L.X. and Y.Y.; Writing—original draft, Z.Z.; Writing—review and editing, Q.Q. and R.Y.; Visualization, Z.Z.; Supervision, R.Y. and Y.Y.; Project administration, Y.Y.; Funding acquisition, R.Y., Q.Q. and Y.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partly supported by the National Natural Science Foundation of China (Grant Nos. 82001902, 82072094, and 12275162), the Natural Science Foundation of Shandong Province (Grant Nos. ZR2020QF037, ZR2020QH198, and ZR2019LZL017), and the Taishan Scholars Project of Shandong Province (Grant No. ts201712098).

Institutional Review Board Statement

Institutional review board approval was received, while the requirement to obtain written informed consent was waived due to the retrospective nature of this study. All MRI images related to this study were anonymized.

Informed Consent Statement

Patient consent was waived due to the retrospective nature of this study.

Data Availability Statement

In order to safeguard the confidentiality of the participants, the data pertaining to this study are currently withheld from public access. The data can be shared upon request.

Acknowledgments

This work was supported by Shandong First Medical University and Shandong Academy of Medical Sciences, which provided the computational resources. We acknowledge the NSCLC patients with BM as well as the clinical help obtained for this project.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
BMBrain metastasis
DLDeep learning
NSCLCNon-small-cell lung cancer
SRSStereotactic radiosurgery
ROIRegion of interest
SSDSingle-shot detector
PPVPositive predictive value
FFFeature fusion
YOLOv5You only look once version 5
SA-YOLOv5Self-attention YOLOv5
T1ceT1-weighted contrast-enhanced
PPrecision
RRecall
CBAMConvolutional block attention module
CAMChannel Attention Module
SAMSpatial Attention Module
STBSwin transformer block
STPHSwin transformer prediction head
SW-MSAShifted windows multi-head self-attention
W-MSAWindows multi-head self-attention
NMSNon-maximum suppression
IoUIntersection over union
SGDStochastic gradient descent

References

  1. Achrol, A.S.; Rennert, R.C.; Anders, C.; Soffietti, R.; Ahluwalia, M.S.; Nayak, L.; Peters, S.; Arvold, N.D.; Harsh, G.R.; Steeg, P.S.; et al. Brain metastases. Nat. Rev. Dis. Prim. 2019, 5, 5. [Google Scholar] [CrossRef] [PubMed]
  2. Page, S.; Milner-Watts, C.; Perna, M.; Janzic, U.; Vidal, N.; Kaudeer, N.; Ahmed, M.; McDonald, F.; Locke, I.; Minchom, A.; et al. Systemic treatment of brain metastases in non-small cell lung cancer. Eur. J. Cancer 2020, 132, 187–198. [Google Scholar] [CrossRef] [PubMed]
  3. Goldberg, S.B.; Contessa, J.N.; Omay, S.B.; Chiang, V. Lung Cancer Brain Metastases. Cancer J. 2015, 21, 398–403. [Google Scholar] [CrossRef] [PubMed]
  4. Sperduto, P.W.; Mesko, S.; Li, J.; Cagney, D.; Aizer, A.; Lin, N.U.; Nesbit, E.; Kruser, T.J.; Chan, J.; Braunstein, S.; et al. Incidence and prognosis of patients with brain metastases at diagnosis of systemic malignancy: A population-based study. J. Clin. Oncol. 2020, 38, 3773–3784. [Google Scholar] [CrossRef] [PubMed]
  5. Cagney, D.N.; Martin, A.M.; Catalano, P.J.; Redig, A.J.; Lin, N.U.; Lee, E.Q.; Wen, P.Y.; Dunn, I.F.; Bi, W.L.; Weiss, S.E.; et al. Incidence and prognosis of patients with brain metastases at diagnosis of systemic malignancy: A population-based study. Neuro-oncology 2017, 19, 1511–1521. [Google Scholar] [CrossRef] [PubMed]
  6. Suh, J.H.; Kotecha, R.; Chao, S.T.; Ahluwalia, M.S.; Sahgal, A.; Chang, E.L. Current approaches to the management of brain metastases. Nat. Rev. Clin. Oncol. 2020, 17, 279–299. [Google Scholar] [CrossRef]
  7. Yamamoto, M.; Serizawa, T.; Shuto, T.; Akabane, A.; Higuchi, Y.; Kawagishi, J.; Yamanaka, K.; Sato, Y.; Jokura, H.; Yomo, S.; et al. Stereotactic radiosurgery for patients with multiple brain metastases (JLGK0901): A multi-institutional prospective observational study. Lancet Oncol. 2014, 15, 387–395. [Google Scholar] [CrossRef]
  8. Niranjan, A.; Monaco, E.; Flickinger, J.; Lunsford, L.D. Guidelines for Multiple Brain Metastases Radiosurgery. Prog. Neurol. Surg. 2019, 34, 100–109. [Google Scholar]
  9. Zhou, Z.; Sanders, J.W.; Johnson, J.M.; Gule-Monroe, M.K.; Chen, M.M.; Briere, T.M.; Wang, Y.; Son, J.B.; Pagel, M.D.; Li, J.; et al. Computer-aided Detection of Brain Metastases in T1-weighted MRI for Stereotactic Radiosurgery Using Deep Learning Single-Shot Detectors. Radiology 2020, 295, 407–415. [Google Scholar] [CrossRef]
  10. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  11. Amemiya, S.; Takao, H.; Kato, S.; Yamashita, H.; Sakamoto, N.; Abe, O. Feature-fusion improves MRI single-shot deep learning detection of small brain metastases. J. Neuroimaging 2022, 32, 111–119. [Google Scholar] [CrossRef]
  12. Cao, Y.; Vassantachart, A.; Ye, J.C.; Yu, C.; Ruan, D.; Sheng, K.; Lao, Y.; Shen, Z.L.; Balik, S.; Bian, S.; et al. Automatic detection and segmentation of multiple brain metastases on magnetic resonance image using asymmetric UNet architecture. Phys. Med. Biol. 2021, 66, 015003. [Google Scholar] [CrossRef] [PubMed]
  13. Li, R.; Guo, Y.; Zhao, Z.; Chen, M.; Liu, X.; Gong, G.; Wang, L. MRI-based two-stage deep learning model for automatic detection and segmentation of brain metastases. Eur. Radiol. 2023, 33, 3521–3531. [Google Scholar] [CrossRef] [PubMed]
  14. Jocher, G.; Chaurasia, A.; Stoken, A.; Borovec, J.; NanoCode012; Kwon, Y.; TaoXie; Michael, K.; Fang, J.; Imyhxy; et al. ultralytics/yolov5: v6.2 - YOLOv5 Classification Models, Apple M1, Reproducibility, ClearML and Deci.ai Integrations. 2022. Available online: https://doi.org/10.5281/zenodo.7002879 (accessed on 1 October 2022).
  15. Chen, S.; Duan, J.; Wang, H.; Wang, R.; Li, J.; Qi, M.; Duan, Y.; Qi, S. Automatic detection of stroke lesion from diffusion-weighted imaging via the improved YOLOv5. Comput. Biol. Med. 2022, 150, 106120. [Google Scholar] [CrossRef] [PubMed]
  16. Wan, J.; Chen, B.; Yu, Y. Polyp Detection from Colorectum Images by Using Attentive YOLOv5. Diagnostics 2021, 11, 2264. [Google Scholar] [CrossRef] [PubMed]
  17. Huang, H.Y.; Hsiao, Y.P.; Mukundan, A.; Tsao, Y.M.; Chang, W.Y.; Wang, H.C. Classification of Skin Cancer Using Novel Hyperspectral Imaging Engineering via YOLOv5. J. Clin. Med. 2023, 12, 1134. [Google Scholar] [CrossRef]
  18. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  19. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. Adv. Neural Inf. Process. Syst. 2019, 32, 8024–8035. [Google Scholar]
  20. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2018; pp. 3–19. [Google Scholar]
  21. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 9992–10002. [Google Scholar]
  22. Ioffe, S.S.C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  23. Elfwing, S.; Uchibe, E.; Doya, K. Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Netw. 2018, 107, 3–11. [Google Scholar] [CrossRef]
  24. Wang, W.; Xie, E.; Song, X.; Zang, Y.; Wang, W.; Lu, T.; Yu, G.; Shen, C. Efficient and Accurate Arbitrary-Shaped Text Detection With Pixel Aggregation Network. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8439–8448. [Google Scholar]
  25. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar]
  26. Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-Unet: Unet-like Pure Transformer for Medical Image Segmentation. In Proceedings of the ECCV Workshops, Tel Aviv, Israel, 23 October 2022. [Google Scholar]
  27. Yu, J.; Jiang, Y.; Wang, Z.; Cao, Z.; Huang, T.S. UnitBox: An Advanced Object Detection Network. In Proceedings of the 24th ACM international conference on Multimedia, Amsterdam, The Netherlands, 15–19 October 2016. [Google Scholar]
  28. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU Loss: Faster and Better Learning for Bounding Box Regression. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January 2019–1 February 2019. [Google Scholar]
  29. Sinha, N.K.; Griscik, M.P. A Stochastic Approximation Method. IEEE Trans. Syst. Man Cybern. 1971, SMC-1, 338–344. [Google Scholar] [CrossRef]
  30. Neubeck, A.; Gool, L.V. Efficient Non-Maximum Suppression. In Proceedings of the 18th International Conference on Pattern Recognition (ICPR’06), Washington, DC, USA, 20–24 August 2006; Volume 3, pp. 850–855. [Google Scholar]
  31. Kato, S.; Amemiya, S.; Takao, H.; Yamashita, H.; Sakamoto, N.; Abe, O. Automated detection of brain metastases on non-enhanced CT using single-shot detectors. Neuroradiology 2021, 63, 1995–2004. [Google Scholar] [CrossRef]
  32. Takao, H.; Amemiya, S.; Kato, S.; Yamashita, H.; Sakamoto, N.; Abe, O. Deep-learning 2.5-dimensional single-shot detector improves the performance of automated detection of brain metastases on contrast-enhanced CT. Neuroradiology 2022, 64, 1511–1518. [Google Scholar] [CrossRef] [PubMed]
  33. Cao, X.; Zhang, Y.; Lang, S.; Gong, Y. Swin-Transformer-Based YOLOv5 for Small-Object Detection in Remote Sensing Images. Sensors 2023, 23, 3634. [Google Scholar] [CrossRef]
  34. Dai, G.; Hu, L.; Fan, J. DA-ActNN-YOLOV5: Hybrid YOLO v5 Model with Data Augmentation and Activation of Compression Mechanism for Potato Disease Identification. Comput. Intell. Neurosci. 2022, 2022, 6114061. [Google Scholar] [CrossRef] [PubMed]
  35. Yang, Z.; Li, L.; Luo, W. PDNet: Improved YOLOv5 Nondeformable Disease Detection Network for Asphalt Pavement. Comput. Intell. Neurosci. 2022, 2022, 5133543. [Google Scholar] [CrossRef] [PubMed]
  36. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  37. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10778–10787. [Google Scholar]
Figure 1. Flowchart of patient inclusion and exclusion criteria.
Figure 1. Flowchart of patient inclusion and exclusion criteria.
Cancers 15 04443 g001
Figure 2. Distributions of BMs in the training group and internal testing set: (a) distribution of BMs in the training group and (b) distribution of BMs in the internal testing set.
Figure 2. Distributions of BMs in the training group and internal testing set: (a) distribution of BMs in the training group and (b) distribution of BMs in the internal testing set.
Cancers 15 04443 g002
Figure 3. The distribution of BM maximum diameters in the training group and internal testing set: (a) distribution of BM maximum diameters in the training group and (b) distribution of BM maximum diameters in the internal testing set.
Figure 3. The distribution of BM maximum diameters in the training group and internal testing set: (a) distribution of BM maximum diameters in the training group and (b) distribution of BM maximum diameters in the internal testing set.
Cancers 15 04443 g003
Figure 4. The architecture of SA-YOLOv5. CBAM: convolutional block attention module, STB: Swin transformer block, SPPF: spatial pyramid pooling–fast. Conv is composed of a 1 × 1 convolutional block, batch normalization unit [22], and SiLU (sigmoid-weighted linear unit) [23].
Figure 4. The architecture of SA-YOLOv5. CBAM: convolutional block attention module, STB: Swin transformer block, SPPF: spatial pyramid pooling–fast. Conv is composed of a 1 × 1 convolutional block, batch normalization unit [22], and SiLU (sigmoid-weighted linear unit) [23].
Cancers 15 04443 g004
Figure 5. The structure of C3 block. The Conv1 block is composed of a 1 × 1 convolutional block, batch normalization unit, and SiLU. The Conv3 block is composed of a 3 × 3 convolutional block, batch normalization unit, and SiLU.
Figure 5. The structure of C3 block. The Conv1 block is composed of a 1 × 1 convolutional block, batch normalization unit, and SiLU. The Conv3 block is composed of a 3 × 3 convolutional block, batch normalization unit, and SiLU.
Cancers 15 04443 g005
Figure 6. The structure of the SPPF block. The Conv block is composed of a 1 × 1 convolutional block, batch normalization unit, and SiLU.
Figure 6. The structure of the SPPF block. The Conv block is composed of a 1 × 1 convolutional block, batch normalization unit, and SiLU.
Cancers 15 04443 g006
Figure 7. The structure of CBAM, consisting of the CAM and SAM.
Figure 7. The structure of CBAM, consisting of the CAM and SAM.
Cancers 15 04443 g007
Figure 8. The architecture of the Swin transformer block. LN: layer normalization, MLP: multilayer perceptron, W-MSA: windows multi-head self-attention, SW-MSA: shifted windows multi-head self-attention.
Figure 8. The architecture of the Swin transformer block. LN: layer normalization, MLP: multilayer perceptron, W-MSA: windows multi-head self-attention, SW-MSA: shifted windows multi-head self-attention.
Cancers 15 04443 g008
Figure 9. Windows used for self-attention computations in the adjacent layers l and l plus one are divided into a regular window and shifted window.
Figure 9. Windows used for self-attention computations in the adjacent layers l and l plus one are divided into a regular window and shifted window.
Cancers 15 04443 g009
Figure 10. (a) I o U schematic and (b) schematic diagram of C I o U penalty items. Here, P 0 and G 0 represent the respective center points of P and G, ρ is the Euclidean distance between P 0 and G 0 , and c denotes the diagonal length of the minimum outer rectangle between P and G.
Figure 10. (a) I o U schematic and (b) schematic diagram of C I o U penalty items. Here, P 0 and G 0 represent the respective center points of P and G, ρ is the Euclidean distance between P 0 and G 0 , and c denotes the diagonal length of the minimum outer rectangle between P and G.
Cancers 15 04443 g010
Figure 11. Detection performance of the proposed model on internal testing MRI images: (ad) examples of true positive inferences. The red bounding boxes represent the ground truth, while the yellow bounding boxes represent the model prediction results.
Figure 11. Detection performance of the proposed model on internal testing MRI images: (ad) examples of true positive inferences. The red bounding boxes represent the ground truth, while the yellow bounding boxes represent the model prediction results.
Cancers 15 04443 g011
Figure 12. Detection performance of the proposed model on internal testing images: (a,b) examples of false negative inferences and (c,d) examples of false positive inferences. The red bounding boxes represent the ground truth, while the yellow bounding boxes represent the model prediction results.
Figure 12. Detection performance of the proposed model on internal testing images: (a,b) examples of false negative inferences and (c,d) examples of false positive inferences. The red bounding boxes represent the ground truth, while the yellow bounding boxes represent the model prediction results.
Cancers 15 04443 g012
Figure 13. Ground truth and prediction results of the different models on original MRI images of two distinct patients. The upper photos show 2D visualizations and the lower photos show 3D visualizations. The red bounding boxes represent the ground truth, while the yellow bounding boxes represent the model prediction results. The arrow indicates the possible lesion with no ground truth.
Figure 13. Ground truth and prediction results of the different models on original MRI images of two distinct patients. The upper photos show 2D visualizations and the lower photos show 3D visualizations. The red bounding boxes represent the ground truth, while the yellow bounding boxes represent the model prediction results. The arrow indicates the possible lesion with no ground truth.
Cancers 15 04443 g013
Figure 14. Prediction results of our proposed model on the external set. The photos show 2D and 3D visualizations of the same patient. The red bounding boxes represent the ground truth, while the yellow bounding boxes represent the model prediction results.
Figure 14. Prediction results of our proposed model on the external set. The photos show 2D and 3D visualizations of the same patient. The red bounding boxes represent the ground truth, while the yellow bounding boxes represent the model prediction results.
Cancers 15 04443 g014
Table 1. Prediction results of different deep learning models on the internal testing set.
Table 1. Prediction results of different deep learning models on the internal testing set.
MethodRecallPrecisionF2-ScoreFN/PatientFP/Patient
Faster R-CNN [36]0.690(136/197)0.342(136/398)0.5731.2715.458
EfficientDet [37]0.614(121/197)0.669(121/181)0.6241.5831.250
SSD [9]0.822(162/197)0.369(162/439)0.6600.7295.771
FF-SSD [11]0.827(163/197)0.397(163/411)0.6800.7085.167
Ours0.904(178/197)0.612(178/291)0.8250.3962.354
Table 2. The prediction results of our proposed model and its variants on the internal testing set.
Table 2. The prediction results of our proposed model and its variants on the internal testing set.
ModelRecallPrecisionF2-ScoreFN/PatientFP/Patient
YOLOv50.883(174/197)0.611(174/285)0.8120.4792.313
YOLOv5+CBAM0.898(177/197)0.598(177/296)0.8160.4172.479
YOLOv5+CBAM+PH 10.898(177/197)0.586(177/302)0.8120.4172.604
YOLOv5+CBAM+STPH0.904(178/197)0.612(178/291)0.8250.3962.354
1 PH denotes the prediction head 4 without STB.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, Z.; Qiu, Q.; Liu, H.; Ge, X.; Li, T.; Xing, L.; Yang, R.; Yin, Y. Automatic Detection of Brain Metastases in T1-Weighted Construct-Enhanced MRI Using Deep Learning Model. Cancers 2023, 15, 4443. https://doi.org/10.3390/cancers15184443

AMA Style

Zhou Z, Qiu Q, Liu H, Ge X, Li T, Xing L, Yang R, Yin Y. Automatic Detection of Brain Metastases in T1-Weighted Construct-Enhanced MRI Using Deep Learning Model. Cancers. 2023; 15(18):4443. https://doi.org/10.3390/cancers15184443

Chicago/Turabian Style

Zhou, Zichun, Qingtao Qiu, Huiling Liu, Xuanchu Ge, Tengxiang Li, Ligang Xing, Runtao Yang, and Yong Yin. 2023. "Automatic Detection of Brain Metastases in T1-Weighted Construct-Enhanced MRI Using Deep Learning Model" Cancers 15, no. 18: 4443. https://doi.org/10.3390/cancers15184443

APA Style

Zhou, Z., Qiu, Q., Liu, H., Ge, X., Li, T., Xing, L., Yang, R., & Yin, Y. (2023). Automatic Detection of Brain Metastases in T1-Weighted Construct-Enhanced MRI Using Deep Learning Model. Cancers, 15(18), 4443. https://doi.org/10.3390/cancers15184443

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop