Next Article in Journal
Synthetic Megavoltage Cone Beam Computed Tomography Image Generation for Improved Contouring Accuracy of Cardiac Pacemakers
Next Article in Special Issue
WindowNet: Learnable Windows for Chest X-ray Classification
Previous Article in Journal
NeuroActivityToolkit—Toolbox for Quantitative Analysis of Miniature Fluorescent Microscopy Data
Previous Article in Special Issue
Placental Vessel Segmentation Using Pix2pix Compared to U-Net
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Lesion Detection in Optical Coherence Tomography with Transformer-Enhanced Detector

1
Department of Electronic Engineering and Computer Science, Queen Mary University of London—QMUL, London E1 4NS, UK
2
Institute of Dentistry at Barts Health, Queen Mary University of London—QMUL, London E1 4NS, UK
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Imaging 2023, 9(11), 244; https://doi.org/10.3390/jimaging9110244
Submission received: 4 September 2023 / Revised: 10 October 2023 / Accepted: 11 October 2023 / Published: 7 November 2023

Abstract

:
Optical coherence tomography (OCT) is an emerging imaging tool in healthcare with common applications in ophthalmology for the detection of retinal diseases and in dentistry for the early detection of tooth decay. Speckle noise is ubiquitous in OCT images, which can hinder diagnosis by clinicians. In this paper, a region-based, deep learning framework for the detection of anomalies is proposed for OCT-acquired images. The core of the framework is Transformer-Enhanced Detection (TED), which includes attention gates (AGs) to ensure focus is placed on the foreground while identifying and removing noise artifacts as anomalies. TED was designed to detect the different types of anomalies commonly present in OCT images for diagnostic purposes and thus aid clinical interpretation. Extensive quantitative evaluations were performed to measure the performance of TED against current, widely known, deep learning detection algorithms. Three different datasets were tested: two dental and one CT (hosting scans of lung nodules, livers, etc.). The results showed that the approach verifiably detected tooth decay and numerous lesions across two modalities, achieving superior performance compared to several well-known algorithms. The proposed method improved the accuracy of detection by 16–22% and the Intersection over Union (IOU) by 10% for both dentistry datasets. For the CT dataset, the performance metrics were similarly improved by 9% and 20%, respectively.

1. Introduction

Optical coherence tomography (OCT) is an emerging medical imaging technique that uses low-coherence infrared light to harmlessly probe into the human body [1]. Low coherence, however, leads to speckle noise in imaging; thus, it gives rise to a poor signal-to-noise ratio (SNR), confounding imaging details and introducing artifacts [2]. OCT has been widely implemented in clinical practice for ophthalmology to detect multiple retinal diseases such as diabetic retinopathy (DR) [3] and age-related macular degeneration (AMD) [4]. Within preliminary dentistry research, OCT has been used for the early detection of carious lesions, but no procedures exist yet for adequately detecting tooth decay. With the advances in deep learning techniques applied to medical image analysis, it is possible to develop automated solutions that detect the disease in OCT imagery and provide quantitative evidence to medical experts. However, speckle noise in OCT presents a great challenge as it hinders clinicians’ ability to diagnose in extensive detail as well as the capability of deep learning models in their detection tasks. In previous decades, numerous image denoising approaches have been created, but they have proved inadequate for removing noise artifacts. Therefore, lesion detection methods have been developed detection in noisy OCT images. Nevertheless, there is not yet an adequate framework to overcome both denoising and lesion detection in OCT images.
Machine learning (ML) models have been utilized for image analysis, including detection, segmentation, and classification, within the medical field over the past decade. Focusing on lesion detection, the main concepts applied have been detection through classification or segmentation but not applying bounding boxes where the lesion is located. Lately, two main ML models have been successfully and widely implemented to output bounding boxes. These are You Only Look Once (YOLO) and Mask-R CNN (MRCNN). However, multiple drawbacks have been presented, including that YOLO struggles with generalizing anomalies in the form of new or uncommon aspect ratios or arrangements. Another drawback is the computational time and power taken for training and testing on new large datasets, which are not clinically efficient.
To address this problem, we propose a computational detection framework based on a Transformer-Enhanced Detection (TED) method with the assistance of denoising [5] It includes a special layout that combines transformers with attention gates (AGs) to suppress the loss of useful data and focus on the different types of anomalies in OCT images. A systematic comparison was conducted with existing state-of-the-art OCT detection algorithms to demonstrate the advantages of TED for clinical practice in multiple medical domains.
In summary, the contributions of this paper are as follows:
  • A novel TED framework is developed, focusing on detecting relevant lesions in noisy OCT images of different organs.
  • In TED, the transformer is adapted to take in images and slide across Regions of Interest (ROIs) provided by AGs. This design aims to adaptively deal with different types of noise artifacts and thus effectively detect a variety of anomalies including tooth decay and numerous lesions across two modalities.
  • A new loss function is proposed along with TED, which combines a sliding box, Intersection Over Union (IOU), and Mean Squared Error (MSE). It compares the IOU and MSE between the predicted and real bounding boxes to evaluate the regions of focus chosen by the AGs.

1.1. Related Work

1.1.1. Detection Methods

Over the past decade, machine learning (ML) models have been utilized for object detection, classification, and segmentation within the medical field [6]. They have provided effective and promising results for pulmonary lesion detection via pulmonary computed tomography (CT) scans, COVID-19 segmentation and diagnosis via X-rays, and many more applications [7,8]. Focusing on OCT image analysis, many convolutional neural networks (CNNs) have been implemented for the classification of retinal diseases such as diabetic retinopathy (DR) and age-related macular degeneration (AMD) [9,10]. In dentistry, they have been applied for caries detection via the transfer learning models VGG16 and ResNet50 or multi-layer perceptron (MLP) models [11,12,13]. These studies have shown relatively high accuracy, sensitivity, and specificity values of above 80%. However, the classification of images does not show the areas that are affected or indicate lesions. This does not aid clinicians in arriving at a faster diagnosis or improve classification where noise artifacts obscure or mislead [14]. A wide range of data is thereby prone to being mislabeled [14]. Researchers have therefore utilized U-Nets and autoencoders for the 2D segmentation of retinal and dental layers and V-Nets for the 3D segmentation of OCT volumes [15,16,17]. Segmentation provided similarly high efficiency through improved accuracy, yet most studies utilized denoised images as inputs rather than noisy images [14]. This is quite impracticable for clinical practice since there is yet to be an agreed denoising method for OCT imaging. Owing to this, many drawbacks of traditional denoising methods revolve around the loss of meaningful detail by their application of smoothing or the limited removal of noise. Additionally, most of these techniques are computationally intensive [14]. Lastly, another computer-aided diagnostic method is object detection, employing either one-stage or two-stage object identifiers. These include You YOLO and recurrent CNNs (R-CNNs), respectively [18,19]. These models apply bounding boxes where lesions are detected. YOLO has produced substantial results owing to the sliding-box feature it implements. The sliding box provides a convolution feature extractor that further provides confidence scores regarding which boxes contain the anomaly [20]. Numerous researchers have proved that YOLO is a suitable architecture for detection, but it struggles with generalizing anomalies in the form of new or uncommon aspect ratios or arrangements [20].

1.1.2. Attention Gates and Transformers

In 2017, Bahdanau et al. first suggested an improvement from long-short term memory (LSTM), which is called attention gates (AGs) [21]. AGs was predominantly applied to natural language processing (NLP) and then later adjusted for computer vision [22]. Within computer vision, AGs are implemented in CNNs to ensure focus on a certain region to allow optimal feature extraction for the classification goal. Throughout training, AGs propose and underline fundamental Regions of Interest (ROIs) and subdue irrelevant background feature activations. Within medical image processing, the main aim of AGs is to perform the segmentation of lesions in image modalities such as breast ultrasound (BUS) [22]. Concentrating on OCT images, segmentation was used for specific diseases such as AMD and DR in OCT images [23]. All studies mentioned have only applied Tversky loss for segmenting images. Yet it has not been successfully applied for object detection, since its main objectives revolve around segmentation and classification.
Later, Dosovitskiy et al. in 2021 further refined AGs and produced attention-based ‘transformers’ that learn at a high effective rate, feature representations, from encoding long-range dependencies [24]. Transformers utilize a ‘multi-head’ attention model to associate long- and short-distance words in both forward and backward directions. Therefore, it outputs positional encoding for any sentence input [24]. Multiple previous research studies noticed the knowledge gap of transformers in computer vision. Therefore, Vision Transformers (ViTs) were developed to replace CNNs [25]. With regard to the medical field, ViT has been widely implemented for MRI, CT and X-rays for image classification, segmentation and reconstruction [25,26,27]. There is a knowledge gap for object detection of tooth decay using OCT images and ViT with effective and relatively accurate results.

2. Materials and Methods

In this paper, the proposed framework for OCT disease detection is described in Figure 1. Initially, data preparation is conducted, then augmentation is used to create more image/bounding pairs of boxes. This is to overcome the main disadvantage of limited clean OCT datasets for the training and validation of a model. The images are also denoised with an advanced ML method called TEAR where needed. The data are then input to a modified transformer that generates predicted outcomes for lesion detection with the help of the proposed L L o s s loss.

2.1. Data Preparation and Augmentation

To ensure a fast and light-weight training, all OCT images are first resized to 250 pixels × 250 pixels. This is displayed in Figure 1 within the Data Augmentation block. The resized images are then augmented with rotation, magnification and inversion to create more image-bounding boxes to improve the training variety and quality. The resized images are flipped about the horizontal axes, using random rotation with a factor of 0.02 and random zoom in both width and height with a common factor of 0.2. Alongside augmentation, patches with a size of 32 pixels × 32 pixels are created from augmented images and passed through an AG.

2.2. Attention Gated Patch Encoder

As illustrated in Figure 1, within the patch encoder, AGs are placed to provide scores from the input, depending on ROIs that have the possible anomalies through the application of a sliding box. This is to aid the network in focusing on foreground detail and ignoring background information based on the content of each image. The patch encoder commences with a few layers that take the raw image ( I i n p u t ) as the first input, and the second input is set with the outputs of a sliding box. The outputs are patch sizes represented as a query value ( I q u e r i e s ) (Equations (1) and (2)). Here, Q is a patch matrix, while K and V are key-value image pairs. Embedding layers ( E ) then process both inputs and convert each patch size into fixed-length vectors of defined size. The embedding layers utilize 64 × 64 units followed by convolution layers ( ) of 3 × 3 with filter size 32 and 64 with a stride of 3 (Equation (3)). It ensures that previous patches are taken into consideration with emphasis focused on areas through the input images and query values to compute matching scores. Weight vectors ( W S ) are created from the score, which are further processed through convolution layers, to calculate the matrices of both query and input patch size. Lastly, the convolution outputs are processed into a single-head attention layer. The number of convolution units is considered as a hyperparameter in all experiments, and it has been tuned accordingly. All the layers mentioned above are trainable and subject to change and adjustment for different types of medical images.
I i n p u t ( Q , K , V ) = n = 1 N s o f t m a x ( Q K ) × V
I q u e r i e s ( Q , K , V ) = n = 1 N s o f t m a x ( Q K ) × V
W S = E ( A i n p u t ) E ( A q u e r i e s )

2.3. Transformer-Enhanced Detection

OCT is an emerging medical imaging modality; however, it is yet to be widely implemented beyond ophthalmology at present. Therefore, the main aim of TED focuses on aiding the detection of anomalies, including lesions and diseases, to simplify the diagnostic procedure for clinicians. As previously mentioned, ViTs are commonly considered as medical imaging processing tools for segmentation and classification tasks. They have yet to be adapted and implemented for object detection in medical images. Hence, the full framework after the patch encoder starts with TEAR. It consists of modifying ViT and placing it as an encoder in the autoencoder for denoising OCT images. TED follows and includes modifying ViT to output possible bounding boxes, indicating the location and label of relevant lesions. Within the TED framework, the original ViT procedure is followed [24], while reshaping the output of the patch encoder y d e n o i s e d (TEAR output, Equation (4)) and flattening the patches to 2D images y d e n o i s e d R N × ( P 2 · C ) . Here, ( P , P ) is the resolution of each image patch, C is the number of channels, D is the latent vector size of all the layers, E is the patch-embedding outputs, and N is the resulting number of patches (Equations (5) and (6)). The input is then submitted to multiple transformer blocks, starting with a normalization layer for the computation of mean and variance along all axes of the encoded patches ( z L ) . Thus,
y d e n o i s e d = T E A R ( P E ( D A ( y i n p u t ) ) )
z 0 = [ y d e n o i s e d ; y p 1 E ; y p 2 E ; ; y p N E ] + E p o s
E R ( P 2 · C ) × D , E p o s R ( N + 1 ) × D
z l = M S A ( L N ( z l 1 ) ) + z l 1 , l = 1 L
z l = M L P ( L N ( z l ) ) + z l , l = 1 L
y p r e d i c t e d = R S ( L N ( z l ) )
Equation (7) shows the encoded patches set to output the same dimension as the query dimension of the ‘multi-head’ attention ( M S A ) layer. Afterwards, we ensure that for each flattened input, y p r e d i c t e d is computed completely separately from other input features (Equation (9)). Next, the normalization layer ( L N ) outputs are passed onto a ‘multi-head’ attention layer for the computation of attention weights from similarity between two patches, as illustrated in Figure 1. Multiple heads process this in parallel, concatenating the cumulative result for one projected output. Equation (8) shows the projected output passed through another normalization layer that computes the mean and variance along the height, channels and width axes of images and then a multi-layer perceptron ( M L P ) block. The computed input features ensure that it is completely independent of other input features of other images in a batch. The MLP block acts as a classification head with Gaussian Error Linear Units (GELUs) featuring non-linearity [28]. The transformer block is repeated eight times, as indicated in Figure 1, which is followed by another MLP block for the final encoding of the image. Lastly, y p r e d i c t e d is reshaped ( R S ) into an array of possible bounding boxes (Equation (9)).
ViT is implemented after the patch encoder shown in Figure 1, in which it re-assembles an array of bounding boxes. The model in the framework is evaluated with numerous learning rates ( 5.0 × 10 3 , 1.0 × 10 3 , 1.0 × 10 4 , 1.0 × 10 1 , 1.0 × 10 2 ) ; epochs (200, 500 and 1000); batch sizes (2 and 4); optimizers (ADAM, ADAMW); and image sizes (500 pixels × 400 pixels, 500 pixels × 500 pixels). The TED model is trained through the process displayed in Figure 2, which utilizes a new proposed loss function. The environment in which the tests were implemented was Tensorflow and Keras, and all models were trained using one NVIDIA P100 GPU with 24 GB memory.

2.4. Loss Function

A new loss function is proposed for the ViT that is a combination of multiple quantitative metrics, where
L L o s s = L M S E + L d i f f e r e n c e + L I O U .
L M S E is the Mean Square Error loss calculated while training to provide the averaged batch size error of the y p r e d i c t e d and y a c t u a l ; L d i f f e r e n c e is the structural similarity difference loss between predicted and actual image pairs; and L I O U is the Intersection Over Union difference between actual and predicted IOU, which is a normalized coefficient [0, 1]. These are calculated as follows:
L M S E = 1 N i = 1 N ( y p r e d i c t e d y a c t u a l ) 2
L I O U = 1 | A B | | A B |
N is the batch size provided in training; y p r e d i c t e d is the predicted bounding box by the CNN and y a c t u a l is the actual bounding box. Equation (12) shows how IOU loss is calculated as a ratio of area-of-overlap over area-of-union. The loss metrics are combined and normalized. The loss function aids the training of the model with a focus on minimizing the distance between predicted and actual bounding boxes. This is conducted with the help of MSE and IOU image metrics to ensure y p r e d i c t e d covers the correct amount of lesion/disease and within a reasonable size.

3. Results

3.1. Datasets

Three datasets from two medical fields were used to train and test the framework; the two dentistry datasets were one collected in the QMUL IDIOT Lab (QueenMary University of London, Institute of Dentistry in which the free space SD-OCT was set up and utilized to scan the teeth and models) and an immense CT dataset called NIH DeepLesion [29]. Dentistry datasets are imaged by a spectral domain OCT (SD-OCT) with an axial resolution of 4.5 μm and consist of ten samples of 500 images of healthy and decayed teeth. Each image is 500 pixels × 412 pixels. Clean images were created for the dentistry dataset 1 by pre-processing it through TEAR [5] The NIH DeepLesion dataset contains 32, 120-axial CT slices from 10,594 studies. All images included in this paper have one to three lesions per image. Regarding the dentistry datasets, lesions were drawn by multiple dentists and further used as ground truth. Each dataset is then randomly split into training, validation and testing with a 60%:10%:30% split, respectively.

3.2. Evaluation Metrics

The proposed framework is examined using conventional detection metrics: IOU, accuracy and a confusion matrix for quantitative evaluation. A confusion matrix is populated with true positive (TP) elements (for which IOU > 0.5) and false positive (FP) elements (for which IOU < 0.5). The numerical metrics consist of accuracy, sensitivity, precision, F1 score and IOU. In these, TN is when the model predicts images as having no lesions correctly, and FN is when it predicts false lesions. Regarding confusion matrix calculations, true negative (TN) and false negative (FN) are set to zero since it is a detection framework with only one class set to identify lesions. Sensitivity is sometimes called true positive rate (TPR) or recall. Both are tests of the ability to identify predictions correctly. Precision is a ratio of accurate predictions against all positive predictions (both true and false). The F1 score is a better measure to provide a sustainable balance between precision and sensitivity as well as provide a metric if there is a class imbalance. IOU is an evaluation metric mainly utilized for measuring the accuracy of an object detector traditional program, i.e., a ratio of area-of-overlap over area-of-union.

3.3. Ablation Study

TED with different settings was evaluated in a study to investigate the optimal learning rates, epochs, batch, optimizer and image sizes. The batch size was set to four, and the optimal image size was 500 pixels × 400 pixels and 500 × 500 pixels. The leading learning-rate and epochs were 5 × 10 4 and 200, respectively, to obtain optimal results in a suitable timely manner.
Furthermore, a comparative study was conducted on the datasets to evaluate the proposed loss function against widely used loss functions for object detection. The loss functions examined against L L o s s consist of Mean Squared Error (MSE) and Mean Absolute Error (MAE). The quantitative evaluations are shown in Table 1 and the leading qualitative evaluation is shown in Figure 3, Figure 4 and Figure 5 for both the dentistry and CT datasets, respectively. The results are averaged over the dataset. The sensitivity is unity for all evaluation tests, since the confusion matrix only counts TP and FP for lesion detection; therefore, TN and FN are zero. Since the datasets utilized in this paper only have bounding boxes for lesions, therefore, there is only one class that is set as a lesion. Hence, accuracy and precision have equivalent metrics for all the tests. Quantitative results display that TED- L L o s s performs 15–30% higher in average accuracy and precision, acquiring an average of 92.5% in both. The F1 score had an increase of 10–15%, demonstrating a more stable lesion detection with L L o s s than MSE and MAE losses. Regarding IOU, TED obtains an average of increase of 25% in dentistry datasets. This is validated with qualitative results in Figure 3a and Figure 4a, in which y p r e d i c t e d (red boxes) overlap the majority of y a c t u a l (green boxes). Additionally, compared to L L o s s , MAE and MSE produced bounding boxes either too small or big, hence the drop in IOU. With regard to the CT dataset, quantitatively, TED- L L o s s obtained the best-performing average accuracy, F1 score and IOU of 97%, 0.98 and 0.81, respectively. That is an increase of 10%, 18% and 15% across the evaluation metrics, respectively. Along with the qualitative results shown in Figure 5a, the best performing results in the test dataset had an IOU of 0.96, accurately overlapping y a c t u a l (green boxes). This is in contrast to MSE and MAE (Figure 5b,c), producing smaller y p r e d i c t e d (red boxes) that partially overlap y a c t u a l (green boxes). This demonstrates the value of implementing L L o s s in TED for lesion detection across two image modalities. This is shown by the increase in all quantitative metrics for all datasets.
Another ablation study was used to compare the TED framework with and without the additional attention gates in the patch encoder and data augmentation. The quantitative evaluations are shown in Table 2 and leading qualitative evaluation in Figure 6, for both dentistry datasets, respectively. Quantitative results display that TED with AGs performs 9% higher on average accuracy and precision, acquiring an average of 92.5% in both. The F1 score had an increase of 8–18%, demonstrating a more stable lesion detection with AGs, than without. Regarding IOU, AGs provides an average of increase of 15% in both dentistry datasets. This is validated with the qualitative results in Figure 6a,c, in which y p r e d i c t e d (red boxes) overlap the majority of y a c t u a l (green boxes). Figure 6 presents that TED with AGs provides a preferable overlap between y a c t u a l and y p r e d i c t e d with a suitable bounding box size. This demonstrates the value for implementing AGs alongside L L o s s in TED for lesion detection across both dentistry datasets. This is shown by the increase in all quantitative and qualitative metrics for all datasets.

3.4. Results and Comparison

In addition, a comparative study was conducted on the three datasets to examine the proposed framework against state-of-the-art DL techniques YOLOv1 [30], YOLOv3 [31] and Mask-RCNN [32]. Quantitative results, with evaluation metrics for the TED method against state-of-the-art detectors for dentistry and CT datasets, are shown in Table 3. The results are averaged over the datasets. Numerous images are shown from each dataset for qualitative comparison of the outputs of the well-known detectors with respect to our proposed framework, as shown in Figure 7, Figure 8 and Figure 9. Quantitative results display that TED- L L o s s performs 16–22% higher in average accuracy and precision, acquiring an average of 92.5% in both dentistry datasets. Mask-RCNN outperformed YOLOv1 and YOLOv3 in dentistry dataset 1 by obtaining an accuracy and average IOU of 76% and 0.79, respectively. This is displayed in Figure 7 in which Mask-RCNN covered a larger amount of the green box than YOLOv1 and YOLOv3. However, Mask-RCNN did not obtain the best quantitative results across dentistry dataset 2 and NIH DeepLesion. This is attributed to a clear misdiagnosis, as shown by the red boxes in Figure 8 and Figure 9 not overlapping with the green boxes in significant proportion. Across all datasets examined, YOLOv3 produced relatively larger y p r e d i c t e d (red boxes), as shown in Figure 7c, Figure 8c and Figure 9c. On the other hand, YOLOv1 (Figure 7b and Figure 9b) produced smaller y p r e d i c t e d (red boxes). This would complicate diagnosis for clinicians, especially for more complex images. With regard to the CT dataset, quantitatively, TED- L L o s s obtained the leading metrics in average accuracy, F1 score and IOU with values of 97%, 0.98 and 0.81, respectively. This is an increase of 9%, 4% and 20% across the evaluation metrics, respectively, indicating the value for utilizing TED- L L o s s for lesion detection across two image modalities. This is shown by the increase in all quantitative metrics for all datasets against well-known object detection CNNs.
Lastly, each DL model was timed for training and testing functionalities for the dentistry datasets and NIH DeepLesion, which consist of 5000 images of size 512 × 500 pixels and 32,120 images of size 512 × 512 pixels, respectively. This is to show and evaluate if they provide results within a timely manner, focusing more on the training time taken and for the adaptation of new and different datasets when required in an efficient process. Therefore, the results are displayed in Table 4 for the CNN models against TED- L L o s s for the three datasets. Regarding dentistry datasets 1 and 2 and the NIH DeepLesion dataset, the proposed method was trained and tested within 9 min 51 s, 10 min 2 s and 3 min 60 s, respectively. This produces an average time of 7 min 57 s taken. That is at least less than 11% over other DL models. Therefore, it implies that TED- L L o s s is a lightweight model that has numerous hyperparameters that are subject to training for different datasets.

4. Discussion

Recently, DL methods have been implemened for many image processing tasks, such as classification, segmentation and detection. Numerous CNN layouts were implemented for low-dose CT and MRI [33,34]. Nonetheless, there is limited research focused on lesion detection in SD-OCT datasets. The leading DL models are RCNN [32] and developments of YOLO (such as YOLOv1 [30], YOLOv3 [31]). These were implemented on public multimodality datasets, such as NIH DeepLesion [29]. DL methods were either not compared against other popular DL methods or proven to improve the next-step analysis of images. Next-step analysis examples include lesion classification as well as dental and abdomen layer segmentation. Furthermore, there is yet to be a DL model implemented for the lesion detection of SD-OCT dental images. Hence, the TED method here is the first work undertaken to show efficacy through comparison against DL detection methods and further analysis against assessment given by a dentist. A key task is to apply detection on more than one dataset of different medical imaging modalities. This was validated here by using both CT and OCT data libraries.
Three datasets were trained and tested: one was a public CT dataset [29] that consisted of CT slices with a table of bounding boxes of lesions in each image; the second and third are dental datasets that consisted of OCT images with a respective table of bounding boxes around lesions. Results show that the TED method generated superior detection capability over multiple modalities even with limited data of OCT images.
The TED method has several advantages that are novel. Firstly it deploys attention gates (AGs) into the data augmentation operation to provide the model with foreground ROIs. This allows the model to focus on creating a larger dataset from the limited data provided from either medical imaging modalities. Additionally, it is expected to aid the network to focus on the foreground and ignore background information based on the content of each image, which later provide specified patches for the CNN to denoise and detect accurately. This also allows the framework to utilize less computational power and time, since it replaces the concept of two function CNNs for feature extraction before medical image analysis.
After the AG model and data augmentation, the patches are fed into the TEAR framework for data preparation to create clean reference images in order to ensure the minimal amount of removal of useful data. TEAR is a new hybrid of ViT, which was implemented as an encoder in an autoencoder to utilize the attention score from AGs and correlate the ROIs to reconstruct the image in the correct manner. Specifically, this did not include the addition of data due to any realignment issue or noise artefact or limiting the removal of useful data between the OCT dental and CT abdominal layers. Instead, the proposed loss function focused on signal restoration, error in data retrieval, thresholding and edge preserving. This creates a robust framework, since different types of noise artefacts and speckle noise were removed to an appropriate limit without removing useful data.
Next, the denoised patches are passed to TED, which is a new hybrid of adapted ViT with additional AGs. An ablation study in Section 3.3, displayed the added value of AGs for detection. The quantitative evaluations are shown in Table 4 and the leading qualitative evaluation is shown in Figure 6 for both dentistry datasets, respectively. Quantitative results display that TED with AGs performs 9% higher in average accuracy and precision, acquiring an average of 92.5% in both. With regard to IOU, AGs provides an average of increase of 15% in both dentistry datasets. Another contribution shown in this papers is a newly implemented loss function, providing the multiple image quality metrics of a sliding box, IOU, and MSE. IOU focuses on bounding boxes placement and MSE in object detection. This creates a robust framework for different types of image modalities in numerous medical fields. In Section 3.3, a comparative study is performed to exhibit the performance of the loss function. Quantitative results display that TED- L L o s s performs 15–30% higher in average accuracy and precision, acquiring an average of 92.5% in both. The F1 score had an increase of 10–15%, demonstrating a more stable lesion detection with L L o s s than MSE and MAE losses. With regard to IOU, TED obtains an average increase of 25% in the dentistry datasets. With regard to the CT dataset, TED- L L o s s obtained the best-performing average accuracy, F1 score and IOU of 97%, 0.98 and 0.81, respectively. That is an increase of 10%, 18% and 15% across the evaluation metrics, respectively.
Within this paper, dentists are queried as to whether the proposed method detects the lesion effectively for them to diagnose. For dentistry dataset 1 (Figure 10a), dentists acquired an average IOU of 0.87 against the actual bounding boxes in which the dentist and predicted boxes overlap completely. Moving onto dentistry dataset 2 (Figure 10b), all bounding boxes overlap in an ideal manner. This produces an average IOU of 0.97, acknowledging that the dentist’s bounding box is relatively larger while the predicted bounding box is narrower.

5. Conclusions

This paper introduces Transformer-Enhanced Detection (TED), which is a novel framework that detects lesions in two medical imaging modalities, optical coherence tomography (OCT) and computed tomography (CT). The OCT images were captured by Spectral Domain OCT (SD-OCT) for dentistry fields and CT over multiple fields. TED delivers substantial advantage to clinicians since it maintains useful information to aid in clear and unambiguous diagnosis. It starts with overcoming the first obstacle of employing OCT by supplementing the limitation of clean OCT datasets through data augmentation. This aids in optimizing the supervised learning within the architecture. The augmented data are then the argument to TED. The workings consist of a layout combining transformers with attention gates to decrease the loss of useful data and to promote focus on the different types of anomalies in OCT images. The new loss function is utilized along with TED, which combines a sliding box, Intersection Over Union (IOU) and Mean Squared Error (MSE). TED improved the accuracy of lesion detection by 16–22% and IOU by 10% in both dentistry datasets and by 9% and 20%, respectively, for the CT dataset. Through testing multiple datasets, it was demonstrated that the TED framework has the ability of being applied to different modalities across multiple medical fields. It has been shown that its framework is capable of automatically adapting to different datasets. This new detection methodology will be next conducted on magnetic resonance imaging (MRI) datasets and OCT cardiology datasets. Further work will also look to simulate the effect of further innovations of the precise segmentation of lesions.

Author Contributions

Conceptualization, H.A. and Q.Z.; methodology, H.A.; software, H.A.; validation, H.A., A.A. and Q.Z.; resources, H.A. and F.W.; data curation, H.A.; writing—original draft preparation, H.A.; writing—review and editing, R.D., Q.Z. and A.A.; visualization, H.A.; supervision, R.D., Q.Z. and A.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All relevant code and data are available upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Aumann, S.; Donner, S.; Fischer, J.; Müller, F. Optical Coherence Tomography (OCT): Principle and Technical Realization. In High Resolution Imaging in Microscopy and Ophthalmology; Springer: Berlin/Heidelberg, Germany, 2019; pp. 59–85. [Google Scholar]
  2. Huang, S.; Tang, C.; Xu, M.; Qiu, Y.; Lei, Z. BM3D-based total variation algorithm for speckle removal with structure-preserving in OCT images. Appl. Opt. 2019, 58, 6233–6243. [Google Scholar] [CrossRef] [PubMed]
  3. Tey, K.Y.; Teo, K.; Tan, A.C.S.; Devarajan, K.; Tan, B.; Tan, J.; Schmetterer, L.; Ang, M. Optical coherence tomography angiography in diabetic retinopathy: A review of current applications. Eye Vis. 2019, 6, 37. [Google Scholar] [CrossRef] [PubMed]
  4. Müller, P.L.; Liefers, B.; Treis, T.; Rodrigues, F.G.; Olvera-Barrios, A.; Paul, B.; Dhingra, N.; Lotery, A.; Bailey, C.; Taylor, P.; et al. Reliability of Retinal Pathology Quantification in Age-Related Macular Degeneration: Implications for Clinical Trials and Machine Learning Applications. Transl. Vis. Sci. Technol. 2021, 10, 4. [Google Scholar] [CrossRef] [PubMed]
  5. Ahmed, H.; Zhanga, Q.; Donnan, R.; Alomainy, A. TEAR cleaning of noisy Optical Coherence Tomography Images. J. Med. Imaging, 2023; submitted. [Google Scholar]
  6. Ker, J.; Wang, L.; Rao, J.; Lim, T. Deep Learning Applications in Medical Image Analysis. IEEE Access 2017, 6, 9375–9389. [Google Scholar] [CrossRef]
  7. Li, D.; Fu, Z.; Xu, J. Stacked-autoencoder-based model for COVID-19 diagnosis on CT images. Appl. Intell. 2021, 51, 2805–2817. [Google Scholar] [CrossRef]
  8. Horry, M.J.; Chakraborty, S.; Paul, M.; Ulhaq, A.; Pradhan, B.; Saha, M.; Shukla, N. COVID-19 Detection through Transfer Learning Using Multimodal Imaging Data. IEEE Access 2020, 8, 149808–149824. [Google Scholar] [CrossRef]
  9. Roy, K.; Chaudhuri, S.S.; Roy, P.; Chatterjee, S.; Banerjee, S. Transfer learning coupled convolution neural networks in detecting retinal diseases using OCT images. In Intelligent Computing: Image Processing Based Applications; Springer: Singapore, 2020; pp. 153–173. [Google Scholar]
  10. Ahmad, M.; Kasukurthi, N.; Pande, H. Deep Learning for Weak Supervision of Diabetic Retinopathy Abnormalities. In Proceedings of the IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), Venice, Italy, 8–11 April 2019. [Google Scholar]
  11. Prajapati, S.A.; Nagaraj, R.; Mitra, S. Classification of dental diseases using CNN and transfer learning. In Proceedings of the International Symposium on Computational and Business Intelligence (ISCBI), Dubai, United Arab Emirates, 11–14 August 2017. [Google Scholar]
  12. Holtkamp, A.; Elhennawy, K.; de Oro, J.E.C.G.; Krois, J.; Paris, S.; Schwendicke, F. Generalizability of deep learning models for caries detection in near-infrared light transillumination images. J. Clin. Med. 2021, 10, 961. [Google Scholar] [CrossRef]
  13. Tripathi, P.; Malathy, C.; Prabhakaran, M. Genetic algorithms based approach for dental caries detection using back propagation neural network. Int. J. Recent Tech. Eng. 2019, 8, 317–319. [Google Scholar]
  14. Rahimi, H.M.; Motamedian, S.R.; Rohban, M.H.; Krois, J.; Schwendicke, F. Deep learning for caries detection: A systematic review. J. Dent. 2022, 122, 104115. [Google Scholar] [CrossRef]
  15. Asgari, R.; Waldstein, S.; Schlanitz, F.; Baratsits, M.; Erfurth, U.S.; Bogunovic, H. U-Net with Spatial Pyramid Pooling for Drusen Segmentation in Optical Coherence Tomography. In Ophthalmic Medical Image Analysis; Springer: Cham, Switzerland, 2019; pp. 77–85. [Google Scholar]
  16. Cantu, A.G.; Gehrung, S.; Krois, J.; Schwendicke, F. Detecting caries lesions of different radiographic extension on bitewings using deep learning. J. Dent. 2020, 100, 103425. [Google Scholar] [CrossRef]
  17. Roy, A.G.; Conjeti, S.; Carlier, S.G.; Houissa, K.; Sheet, D. Multiscale distribution preserving autoencoders for plaque detection in intravascular optical coherence tomography. In Proceedings of the IEEE 13th International Symposium on Biomedical Imaging (ISBI), Prague, Czech Republic, 13–16 April 2016. [Google Scholar]
  18. Zhang, X.; Liang, Y.; Li, W.; Liu, C.; Gu, D.; Sun, W.; Miao, L. Development and evaluation of deep learning for screening dental caries from oral photographs. Oral Dis. 2020, 28, 173–181. [Google Scholar] [CrossRef]
  19. Javid, A.; Rashid, U.; Khattak, A.S. Marking Early Lesions in Labial Colored Dental Images using a Transfer Learning Approach. In Proceedings of the IEEE 23rd International Multitopic Conference (INMIC), Bahawalpur, Pakistan, 5–7 November 2020. [Google Scholar]
  20. Zhao, L.; Wan, Y. A New Deep Learning Architecture for Person Detection. In Proceedings of the IEEE 5th International Conference on Computer and Communications (ICCC), Chengdu, China, 6–9 December 2019. [Google Scholar]
  21. Bahdanau, D.; Cho, K.; Bengio, Y. Neural Machine Translation by Jointly Learning to Align and Translate. arXiv 2014, arXiv:1409.0473. [Google Scholar]
  22. Thirusangu, N.; Almekkawy, M. Segmentation of Breast Ultrasound Images using Densely Connected Deep Convolutional Neural Network and Attention Gates. In Proceedings of the 2021 IEEE UFFC Latin America Ultrasonics Symposium (LAUS), Gainesville, FL, USA, 4–5 October 2021; pp. 1–4. [Google Scholar]
  23. Liu, X.; Wang, S.; Zhang, Y.; Liu, D.; Hu, W. Automatic fluid segmentation in retinal optical coherence tomography images using attention based deep learning. Neurocomputing 2021, 452, 576–591. [Google Scholar] [CrossRef]
  24. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  25. Yang, D.; Myronenko, A.; Wang, X.; Xu, Z.; Roth, H.R.; Xu, D. T-AutoML: Automated Machine Learning for Lesion Segmentation Using Transformers in 3D Medical Imaging. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 3962–3974. [Google Scholar]
  26. Ravishankar, H.; Venkataramani, R.; Anamandra, S.; Sudhakar, P.; Annangi, P. Feature Transformers: Privacy Preserving Lifelong Learners for Medical Imaging. In Proceedings of the MICCAI, Shenzhen, China, 13–17 October 2019; pp. 347–355. [Google Scholar]
  27. Korkmaz, Y.; Dar, S.U.H.; Yurt, M.; Özbey, M.; Çukur, T. Unsupervised MRI Reconstruction via Zero-Shot Learned Adversarial Transformers. IEEE Trans. Med. Imaging 2022, 41, 1747–1763. [Google Scholar] [CrossRef] [PubMed]
  28. Hendrycks, D.; Gimpel, K. Gaussian error linear units (gelus). arXiv 2016, arXiv:1606.08415. [Google Scholar]
  29. Yan, K.; Wang, X.; Lu, L.; Summer, R.M. DeepLesion: Automated Mining of Large-Scale Lesion Annotations and Universal Lesion Detection with Deep Learning. J. Med. Imaging 2018, 5, 036501. [Google Scholar] [CrossRef] [PubMed]
  30. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE CVPR, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  31. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  32. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE CVPR, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  33. Min, L.S.; Beom, S.J.; Jihye, Y.; Young-Hoon, C.; Jens, V.; Mark, L.S.; Warren, B.G.; Namkug, K. Deep Learning Applications in Chest Radiography and Computed Tomography. J. Thorac. Imaging 2019, 34, 75–85. [Google Scholar]
  34. Mazurowski, M.A.; Buda, M.; Saha, A.; Bashir, M.R. Deep learning in radiology: An overview of the concepts and a survey of the state of the art with focus on MRI. J. Magn. Reson. Imaging 2019, 49, 939–954. [Google Scholar] [CrossRef]
Figure 1. The architecture of the Transformer-Enhanced Detection (TED) framework. The denoised images are first augmented to create a larger dataset. The augmented images are then fed into the TED structure that contains a Vision Transformer (ViT) that is managed by the loss function during training.
Figure 1. The architecture of the Transformer-Enhanced Detection (TED) framework. The denoised images are first augmented to create a larger dataset. The augmented images are then fed into the TED structure that contains a Vision Transformer (ViT) that is managed by the loss function during training.
Jimaging 09 00244 g001
Figure 2. Training of Transformer-Enhanced Detection (TED) with the new proposed loss function, consisting of the combination of relevant detection evaluation metrics (MSE and IOU) between an input image and the predicted bounding box computed from TED. Here, the red and green boxes are the predicted and actual bounding boxes.
Figure 2. Training of Transformer-Enhanced Detection (TED) with the new proposed loss function, consisting of the combination of relevant detection evaluation metrics (MSE and IOU) between an input image and the predicted bounding box computed from TED. Here, the red and green boxes are the predicted and actual bounding boxes.
Jimaging 09 00244 g002
Figure 3. Results from comparative study for dentistry dataset 1 where (a) is the Transformer-Enhanced Detection method (TED with L L o s s ), (b) is the MSE loss, and (c) is the MAE loss. Green boxes are actual bounding boxes ( y a c t u a l ), and red boxes are predicted bounding boxes ( y p r e d i c t e d ).
Figure 3. Results from comparative study for dentistry dataset 1 where (a) is the Transformer-Enhanced Detection method (TED with L L o s s ), (b) is the MSE loss, and (c) is the MAE loss. Green boxes are actual bounding boxes ( y a c t u a l ), and red boxes are predicted bounding boxes ( y p r e d i c t e d ).
Jimaging 09 00244 g003
Figure 4. Results from comparative study for dentistry dataset 2 where (a) is the proposed method (TED with L L o s s ), (b) is the MSE loss, and (c) is the MAE loss. Green boxes are actual bounding boxes ( y a c t u a l ), and red boxes are predicted bounding boxes ( y p r e d i c t e d ).
Figure 4. Results from comparative study for dentistry dataset 2 where (a) is the proposed method (TED with L L o s s ), (b) is the MSE loss, and (c) is the MAE loss. Green boxes are actual bounding boxes ( y a c t u a l ), and red boxes are predicted bounding boxes ( y p r e d i c t e d ).
Jimaging 09 00244 g004
Figure 5. Results from the comparative study for NIH DeepLesion where (a) is the Transformer-Enhanced Detection method (TED with L L o s s ), (b) is the MSE loss, and (c) is the MAE loss. Green boxes are actual bounding boxes ( y a c t u a l ), and red boxes are predicted bounding boxes ( y p r e d i c t e d ).
Figure 5. Results from the comparative study for NIH DeepLesion where (a) is the Transformer-Enhanced Detection method (TED with L L o s s ), (b) is the MSE loss, and (c) is the MAE loss. Green boxes are actual bounding boxes ( y a c t u a l ), and red boxes are predicted bounding boxes ( y p r e d i c t e d ).
Jimaging 09 00244 g005
Figure 6. Results from the comparative study for dentistry datasets 1 and 2 where (a) is the Transformer-Enhanced Detection method (TED with AG), (b) is TED without AG for dataset 1, (c) is TED with AG, and (d) is TED without AG for dataset 2. Green boxes are actual bounding boxes ( y a c t u a l ), and red boxes are predicted bounding boxes ( y p r e d i c t e d ).
Figure 6. Results from the comparative study for dentistry datasets 1 and 2 where (a) is the Transformer-Enhanced Detection method (TED with AG), (b) is TED without AG for dataset 1, (c) is TED with AG, and (d) is TED without AG for dataset 2. Green boxes are actual bounding boxes ( y a c t u a l ), and red boxes are predicted bounding boxes ( y p r e d i c t e d ).
Jimaging 09 00244 g006
Figure 7. Results from comparative study for dentistry dataset 1 where (a) TED with L L o s s , (b) YOLOv1,b29, (c) YOLOv3 [31], (d) RCNN [32]. Green boxes are actual bounding boxes ( y a c t u a l ) and red boxes are predicted bounding boxes ( y p r e d i c t e d ).
Figure 7. Results from comparative study for dentistry dataset 1 where (a) TED with L L o s s , (b) YOLOv1,b29, (c) YOLOv3 [31], (d) RCNN [32]. Green boxes are actual bounding boxes ( y a c t u a l ) and red boxes are predicted bounding boxes ( y p r e d i c t e d ).
Jimaging 09 00244 g007
Figure 8. Results from comparative study for dentistry dataset 2 where (a) TED with L L o s s , (b) YOLOv1 [30], (c) YOLOv3 [31], and (d) RCNN [32]. Green boxes are actual bounding boxes ( y a c t u a l ) and red boxes are predicted bounding boxes ( y p r e d i c t e d ).
Figure 8. Results from comparative study for dentistry dataset 2 where (a) TED with L L o s s , (b) YOLOv1 [30], (c) YOLOv3 [31], and (d) RCNN [32]. Green boxes are actual bounding boxes ( y a c t u a l ) and red boxes are predicted bounding boxes ( y p r e d i c t e d ).
Jimaging 09 00244 g008
Figure 9. Results from comparative study for NIH DeepLesion dataset where (a) TED with L L o s s , (b) YOLOv1 [30], (c) YOLOv3 [31], and (d) RCNN [32]. Green boxes are actual bounding boxes ( y a c t u a l ) and red boxes are predicted bounding boxes ( y p r e d i c t e d ).
Figure 9. Results from comparative study for NIH DeepLesion dataset where (a) TED with L L o s s , (b) YOLOv1 [30], (c) YOLOv3 [31], and (d) RCNN [32]. Green boxes are actual bounding boxes ( y a c t u a l ) and red boxes are predicted bounding boxes ( y p r e d i c t e d ).
Jimaging 09 00244 g009
Figure 10. Results from the comparative study for dentistry datasets. (a) is dataset 1 and (b) is dataset 2. Green boxes are actual bounding boxes ( y a c t u a l ), blue boxes are the dentists’ predicted bounding boxes, and red boxes are predicted bounding boxes ( y p r e d i c t e d ).
Figure 10. Results from the comparative study for dentistry datasets. (a) is dataset 1 and (b) is dataset 2. Green boxes are actual bounding boxes ( y a c t u a l ), blue boxes are the dentists’ predicted bounding boxes, and red boxes are predicted bounding boxes ( y p r e d i c t e d ).
Jimaging 09 00244 g010
Table 1. Quantitative results of lesion detection task for dentistry datasets 1 and 2 and NIH DeepLesion dataset of proposed method with different loss functions regarding the averaged accuracy, precision, F1 score and IOU.
Table 1. Quantitative results of lesion detection task for dentistry datasets 1 and 2 and NIH DeepLesion dataset of proposed method with different loss functions regarding the averaged accuracy, precision, F1 score and IOU.
NetAccuracyPrecisionF1 ScoreIOU
Dentistry Dataset 1
TED-MSE0.720.720.760.66
TED-MAE0.680.680.780.68
TED- L L o s s 0.890.890.840.83
Dentistry Dataset 2
TED-MSE0.800.800.840.81
TED-MAE0.920.920.900.71
TED- L L o s s 0.960.960.980.84
NIH DeepLesion Dataset
TED-MSE0.870.870.830.70
TED-MAE0.900.900.850.73
TED- L L o s s 0.970.970.980.81
Table 2. Quantitative results of lesion detection task for dentistry datasets 1 and 2 and NIH DeepLesion dataset of proposed method with and without attention gates (AGs) in averaged accuracy, precision, F1 score and IOU.
Table 2. Quantitative results of lesion detection task for dentistry datasets 1 and 2 and NIH DeepLesion dataset of proposed method with and without attention gates (AGs) in averaged accuracy, precision, F1 score and IOU.
NetAccuracyPrecisionF1 ScoreIOU
Dentistry Dataset 1
TED—without AG0.820.820.780.73
TED—with AG0.890.890.840.83
Dentistry Dataset 2
TED—without AG0.880.880.840.74
TED—with AG0.960.960.980.84
Table 3. Quantitative results of lesion detection task for dentistry datasets 1 and 2 and NIH DeepLesion dataset of the proposed method against well-known CNNs in averaged accuracy, precision, F1 score and IOU.
Table 3. Quantitative results of lesion detection task for dentistry datasets 1 and 2 and NIH DeepLesion dataset of the proposed method against well-known CNNs in averaged accuracy, precision, F1 score and IOU.
NetAccuracyPrecisionF1 ScoreIOU
Dentistry Dataset 1
YOLOv1 [30]0.730.730.840.67
YOLOv3 [31]0.740.740.850.69
Mask-RCNN [32]0.760.760.860.79
TED- L L o s s 0.890.890.840.83
Dentistry Dataset 2
YOLOv1 [30]0.930.930.960.77
YOLOv3 [31]0.870.870.930.79
Mask-RCNN [32]0.740.740.860.73
TED- L L o s s 0.960.960.980.84
NIH DeepLesion Dataset
YOLOv1 [30]0.880.880.940.54
YOLOv3 [31]0.940.940.960.61
Mask-RCNN [32]0.710.710.830.53
TED- L L o s s 0.970.970.980.81
Table 4. Time taken of lesion detection task for training and testing on dentistry datasets 1 and 2 (5000 images of size 500 × 412 pixels) and NIH DeepLesion (32,120 images of size 512 × 512 pixels) of proposed method against well-known CNNs in minutes and seconds.
Table 4. Time taken of lesion detection task for training and testing on dentistry datasets 1 and 2 (5000 images of size 500 × 412 pixels) and NIH DeepLesion (32,120 images of size 512 × 512 pixels) of proposed method against well-known CNNs in minutes and seconds.
NetTime Taken (Min s)
Dentistry Dataset 1
YOLOv1 [30]10 m 36 s
YOLOv3 [31]20 m 3 s
Mask-RCNN [32]9 m 25 s
TED- L L o s s 9 m 51 s
Dentistry Dataset 2
YOLOv1 [30]10 m 33 s
YOLOv3 [31]20 m 13 s
Mask-RCNN [32]10 m 46 s
TED- L L o s s 10 m 2 s
NIH DeepLesion Dataset
YOLOv1 [30]6 m 32 s
YOLOv3 [31]7 m 44 s
Mask-RCNN [32]4 m 25 s
TED- L L o s s 3 m 59 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ahmed, H.; Zhang, Q.; Wong, F.; Donnan, R.; Alomainy, A. Lesion Detection in Optical Coherence Tomography with Transformer-Enhanced Detector. J. Imaging 2023, 9, 244. https://doi.org/10.3390/jimaging9110244

AMA Style

Ahmed H, Zhang Q, Wong F, Donnan R, Alomainy A. Lesion Detection in Optical Coherence Tomography with Transformer-Enhanced Detector. Journal of Imaging. 2023; 9(11):244. https://doi.org/10.3390/jimaging9110244

Chicago/Turabian Style

Ahmed, Hanya, Qianni Zhang, Ferranti Wong, Robert Donnan, and Akram Alomainy. 2023. "Lesion Detection in Optical Coherence Tomography with Transformer-Enhanced Detector" Journal of Imaging 9, no. 11: 244. https://doi.org/10.3390/jimaging9110244

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop