Next Article in Journal
Fog Computing for Control of Cyber-Physical Systems in Industry Using BCI
Previous Article in Journal
SchoolAIR: A Citizen Science IoT Framework Using Low-Cost Sensing for Indoor Air Quality Management
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale-Denoising Residual Convolutional Network for Retinal Disease Classification Using OCT

1
State Key Laboratory of Digital Medical Engineering, School of Biomedical Engineering, Hainan University, Haiko 570228, China
2
Key Laboratory of Biomedical Engineering of Hainan Province, One Health Institute, Hainan University, Haiko 570228, China
3
Research Unit of Multimodal Cross Scale Neural Signal Detection and Imaging, Chinese Academy of Medical Science, HUST-Suzhou Institute for Brainsmatics, Jiangsu Industrial Technology Research Institute (JITRI), Suzhou 215100, China
4
Britton Chance Center for Biomedical Photonics and MoE Key Laboratory for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, Huazhong University of Science and Technology, Wuhan 430074, China
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(1), 150; https://doi.org/10.3390/s24010150
Submission received: 21 November 2023 / Revised: 13 December 2023 / Accepted: 15 December 2023 / Published: 27 December 2023
(This article belongs to the Section Biomedical Sensors)

Abstract

:
Macular pathologies can cause significant vision loss. Optical coherence tomography (OCT) images of the retina can assist ophthalmologists in diagnosing macular diseases. Traditional deep learning networks for retinal disease classification cannot extract discriminative features under strong noise conditions in OCT images. To address this issue, we propose a multi-scale-denoising residual convolutional network (MS-DRCN) for classifying retinal diseases. Specifically, the MS-DRCN includes a soft-denoising block (SDB), a multi-scale context block (MCB), and a feature fusion block (FFB). The SDB can determine the threshold for soft thresholding automatically, which removes speckle noise features efficiently. The MCB is designed to capture multi-scale context information and strengthen extracted features. The FFB is dedicated to integrating high-resolution and low-resolution features to precisely identify variable lesion areas. Our approach achieved classification accuracies of 96.4% and 96.5% on the OCT2017 and OCT-C4 public datasets, respectively, outperforming other classification methods. To evaluate the robustness of our method, we introduced Gaussian noise and speckle noise with varying PSNRs into the test set of the OCT2017 dataset. The results of our anti-noise experiments demonstrate that our approach exhibits superior robustness compared with other methods, yielding accuracy improvements ranging from 0.6% to 2.9% when compared with ResNet under various PSNR noise conditions.

1. Introduction

The macula is the central area of the retina with the highest concentration of visual cells. The status of the macula may be influenced by various pathological factors, such as age-related macular degeneration (AMD) and diabetic macular edema (DME). AMD and DME are prevalent retinal diseases, affecting nearly 8% of the world’s population [1,2]. These diseases can damage patients’ vision and affect their daily lives. Proper screening tools can effectively reduce the harm caused by retinal diseases. Optical coherence tomography (OCT) [3,4,5], a novel optical imaging technology, allows for the three-dimensional imaging of human tissue microstructures and facilitates the detection of retinal disease lesions. OCT is widely used by ophthalmologists due to its non-invasive nature, low cost, and lack of side effects [6]. Identifying retinal lesions in each B-scan of the OCT volume can be time-consuming and onerous for ophthalmologists. There is also the risk of misdiagnosis or missed diagnosis. Thus, automated diagnosis of retinal disease images can assist ophthalmologists in efficiently diagnosing and treating retinal diseases.
Over the past few years, machine learning has achieved significant results in the classification of retinal disease images. Initially, preprocessing methods are applied to the input images, such as image denoising [7,8,9,10] and retinal flattening [9,11]. Subsequently, designed feature descriptors are used to extract features and then put into classifiers to acquire classification results. Albarrak et al. [12] employed preprocessing techniques to extract the ROI and eliminate noise. They combined image decomposition with LBP histograms for feature extraction and utilized the Bayesian classifier for classification. Srinivasan et al. [13] applied various preprocessing methods, such as BM3D, retinal curvature flattening, and image cropping. They then utilized multiscale histograms of gradient descriptors for feature extraction and employed SVM as the classifier. Sun et al. [14] eliminated noise and background areas via retina alignment and image cropping. They obtained image features using sparse coding and a spatial pyramid, employing SVM for classification. While machine learning methods have shown promising results, one limitation is the time-consuming and laborious task of manually extracting features. Meanwhile, moreover, the increasing number of OCT images of the retina show that traditional diagnostic methods are unable to meet the requirements of automated diagnosis.
In recent years, deep learning, specifically convolutional neural networks (CNNs), has made significant advancements in the fields of natural language processing and computer vision [15]. This success has inspired numerous researchers to employ CNNs in the field of retinal OCT images. Several deep CNN models have been applied with great success in the classification of retinal images [16,17,18,19,20]. Huang et al. [21] proposed a layer-guided convolutional neural network (LGCNN). Initially, they utilized ReLayNet to extract information about retinal layers. Subsequently, the LGCNN was employed to integrate this extracted information for the classification of retinal disorders. Fang et al. conducted two studies on retinal image classification. One study [22] adopted an iterative fusion strategy to merge features among different convolutional layers. Another study [23] proposed the lesion-aware convolutional neural network (LACNN). Their approach involves first utilizing a lesion detection network (LDN) to obtain an attention map for retinal lesions, which is then integrated into the classification framework. Sotoudeh-Paima et al. [24] first applied the feature pyramid network (FPN) [25] structure to CNNs to construct multi-scale CNN models. This model can obtain multi-scale features to perform classification well. Ma et al. [26] combined ConvNet and a transformer to extract the local and global features of OCT images, which is superior to a pure vision transformer (ViT). In addition, the transfer learning technique is widely used to classify retinal OCT images and shows promising results [18,19,20,27]. Although these classification methods achieve ideal performance, model optimization remains a significant concern for convolutional neural networks. A residual network is a classical convolutional neural network (CNN) that effectively prevents performance degradation by introducing a skip connection. Some researchers have applied ResNet to the classification of retinal diseases. Serener et al. [16] utilized transfer learning to apply ResNet18 for classifying retinal disease images, demonstrating its superiority over AlexNet. Li et al. [28] proposed a method that combines multiple ResNet50 models to enhance feature extraction. ResNet has also been selected as the backbone for multi-model retinal image classification in [29,30], providing further evidence of the effectiveness of residual networks. Karthik et al. [31] designed the “EdgeEN” block with batch normalization to replace the existing skip connection in ResNet, effectively emphasizing the retinal layers. These studies have investigated the application of residual networks to retinal classification and have demonstrated their reliability and effectiveness. However, the aforementioned deep learning methods fail to consider the impact of noise, especially speckle noise. In clinical applications, retinal images generated by OCT frequently exhibit significant speckle noise [32]. Most existing CNN models are unable to extract sufficiently discriminative features under strong noise conditions. Additionally, the size and morphology of lesion areas in OCT images of retinal diseases vary, posing a challenge for effective feature extraction. Conventional classification models that use the last convolutional layer to classify retinal images appear to be inadequate. Given the aforementioned issues, it is necessary to develop a new deep learning model for classifying retinal disease images.
Based on the above challenges, we propose a multi-scale-denoising residual network (MS-DRCN) for OCT image classification. Initially, the residual network was selected as the backbone. Considering the significant presence of speckle noise in OCT images, we developed a soft-denoising block employing soft thresholding [33,34,35], which effectively reduces the interference features during the training process. In addition, a multi-scale context block was designed to expand the receptive field and enhance the extracted features. The effectiveness of such multi-scale structures has been demonstrated in previous studies [36,37]. Retinopathy displays a wide range of variations and complexities across different scales. Inspired by the feature fusion techniques discussed in [11,22,24], we proposed a feature fusion method to obtain multi-scale information for the precise localization of lesion areas.
Specifically, the main contributions of our study are as follows:
  • The soft-denoising block is designed to effectively remove noise-related features by employing soft thresholding.
  • The multi-scale context block (MCB) aims to express multi-scale features and enhance feature extraction by increasing the receptive field for each subnet.
  • The feature fusion block (FFB) is designed to combine low-resolution and high-resolution features, enabling more accurate identification of retinopathy at various scales in OCT images.
  • Anti-noise experiments proved that the proposed method exhibits superior robustness against various intensities of speckle noise and Gaussian noise.
The subsequent sections are outlined as follows: Section 2 details the framework of our method, the dataset, and the experimental setting. Section 3 presents and analyzes the experimental results, and Section 4 draws the conclusions.

2. Materials and Methods

2.1. OCT Datasets

This section introduces two publicly available OCT datasets of the retina for subsequent experiments.
The first dataset used in our study was the publicly available dataset provided by the University of California San Diego (OCT2017) [20]. This dataset comprises 84,484 B-scans obtained from 4686 patients and includes a training set and a test set. The dataset includes four classes: Drusen, CNV, DME, and Normal. The training set consists of 83,484 retinal OCT images, with 37,205 CNV, 8616 Drusen, 11,348 DME, and 26,315 Normal images. The test set comprises 1000 retinal OCT images, with 250 images per class, obtained from a total of 633 patients.
The second dataset was named OCT-C4, which consists of 25,080 retinal OCT B-scans, including 5540 CNV, 3401 DME, 6765 Drusen, and 9374 Normal scans. This dataset consists of three sub-datasets: NEH [38], Srinivasan2014 [13], and OCT-C8 [39]. The NEH dataset, collected by Noor Eye Hospital in Tehran, Iran, comprises 12,649 retinal OCT B-scans, including 3240 CNV, 3742 Drusen, and 5667 Normal scans. The Srinivasan 2014 dataset, introduced by Duke University, comprises 723 AMD, 1101 DME, and 1407 Normal OCT images. The OCT-C8 dataset consists of eight categories, but we only used four categories (CNV: 2300, Drusen: 2300, DME: 2300, and Normal: 2300).

2.2. Proposed MS-DRCN Method

In this section, the MS-DRCN is proposed to classify OCT images of retinal diseases. The overall architecture of the MS-DRCN is shown in Figure 1, which consists of a residual network, subnets, and a feature fusion block (FFB). The residual network was chosen as the backbone for feature extraction. It is easy to optimize, which can effectively alleviate the problem of performance degradation by using skip connections. The subnets included a soft-denoising block (SDB) and a multi-scale context block (MCB), which can obtain multi-scale features and adaptively eliminate noise-related features. Finally, we used a feature fusion block to merge low-resolution and high-resolution features, and then the output was fed into a classifier for retinal image classification.

2.2.1. Residual Learning

To improve the performance of deep learning models, it is often necessary to increase the depth of the layers. However, this can lead to issues such as gradient vanishing and gradient explosion. To address these issues, batch normalization [40] is employed during the model’s training process for optimization. When the model using batch normalization starts to converge, it may experience performance degradation issues not attributable to overfitting.
A deep residual network can effectively address these challenges. It can enhance model performance by increasing the number of convolutional layers, a benefit derived from the residual block. The key to the residual block is the introduction of a shortcut connection to facilitate the learning of the residual mapping, which is beneficial for model training and optimization. Figure 2 illustrates the residual block. Residual learning is defined as shown in Equation (1).
y = f (x, w) + x
where x refers to the input feature, and y denotes the output feature. The function f (·) is the residual mapping function, while w represents the parameter of the residual block. The residual block consists of two convolutional layers, two BN layers, two ReLu activation functions, and an additive operation. The introduction of shortcut connections in the residual network effectively addresses the issue of performance degradation in deep neural networks.

2.2.2. Soft-Denoising Block

In clinical applications, retinal images generated with OCT often suffer from speckle noise. This type of noise significantly degrades the signal-to-noise ratio and image contrast, thus posing challenges for accurately identifying retinal lesion locations. Consequently, the elimination of speckle-noise-related features has become an urgent problem that must be solved. In this section, we consider soft thresholding to effectively remove these interference features.
The soft-denoising block is shown in Figure 3. The input feature x is transformed into 1D vectors by applying global average pooling (GAP) to its absolute values. Next, the 1D vectors are passed through two fully connected (FC) layers, which are equivalent to the one proposed in [41]. The number of neurons in the FC hidden layer is set to 1/16 of the channel numbers of the input feature. Finally, a sigmoid function is applied to the output of the two FC layers.
F ( z ) = 1 1 + e z
where z is the output of the two FC layers in the soft-denoising block, and F(z) denotes the scaling parameter.
Consequently, the threshold can be calculated by multiplying F(z) with the average value of |x|. By taking the absolute value of the input x, we ensure a positive threshold value. The average value of |x| prevents the threshold from becoming excessively large. The formula can be defined as follows:
τ = F ( z ) * a v e r a g e ( | x | )
where τ represents the threshold, and x is the input feature map. Therefore, Equation (3) not only keeps the threshold in a reasonable range, but it is also positive.
Once the threshold has been determined, a soft threshold function is applied to remove interference information from the extracted features. Finally, the denoised features are added back to the input features. The overall process can be summarized as follows.
y = φ ( x , τ ) + x
where τ is the threshold, φ (·) denotes the soft thresholding, x is the input, and y is the output.
In the past years, soft thresholding has been the core step in signal-denoising methods. Generally, the useful information is transformed into important features far away from zero, and noisy information is converted into unimportant features close to zero. Soft thresholding is used to convert the near-zero features to zero. The function of soft thresholding can be defined using Equation (5).
y = x τ x > τ 0 τ x x + τ x < τ τ
where x is the input feature, y is the output feature, and τ refers to the threshold. Near-zero features are set to zeros by using soft thresholding so that noise-related features can be effectively eliminated.

2.2.3. Multi-Scale Context Block

The size and morphology of lesion areas in OCT images of retinal diseases vary, making it challenging to effectively extract features using a single scale. Therefore, incorporating a multi-scale structure is necessary to enlarge the receptive field of each network layer and enhance the features.
In response to these challenges, we introduce the multi-scale context block (MCB) in this section, as depicted in Figure 4. Bconv denotes the combination of a convolution layer, a batch normalization (BN) layer, and rectified linear unit (ReLu) activation. Initially, a 1 × 1 Bconv is applied to reduce the dimensionality of the input feature channels to one-eighth of the original number. The input features are subsequently divided into four groups, with each group undergoing a 3 × 3 Bconv with a dilation rate of k (where k = 1, 2, 3, and 4). This approach enables the acquisition of multi-scale contextual information by integrating features obtained from various dilation rates. Subsequently, the features from the four groups are concatenated to aggregate the multi-scale information.
C = C o n c a t ( F k ( I i )
where Ii represents the feature of each group, C denotes the concatenated feature, and Fk is the 3 × 3 convolution with a dilation rate of k.
The concatenated feature passes through a 1 × 1 convolution to increase the dimensionality. Finally, an element-wise summation operation is applied to integrate a skip connection into the output feature. The process can be described as follows:
y = x + θ ( C )
where x is the input feature of the MCB, θ represents the 1 × 1 Bconv, and C denotes the concatenated feature. The MCB can acquire multi-scale information and enhance the extracted features.

2.2.4. Feature Fusion Block

Retinopathy exhibits variety and complexity across different scales, and identifying retinal lesions based on single-scale information is ineffective. To better utilize the extracted features, a feature fusion module is introduced to acquire multi-scale information. Previous studies [11,22,24] have shown that feature fusion can significantly enhance the accuracy of retinal disease recognition. Therefore, we propose a feature fusion block that combines low-resolution and high-resolution features to obtain cross-scale information. In a CNN model, earlier feature maps have a high resolution and weak semantics, providing more detailed information, while later feature maps have a low resolution and strong semantics, containing more target feature information. By utilizing the feature fusion block (FFB) to establish a hierarchical pyramid structure, the model can capture strong semantics across all scales.
The FFB is illustrated in Figure 1. The output feature maps of the subnets vary in scales and channels. Initially, a 1 × 1 convolution is employed to adjust the channel of each scale to 256, thereby standardizing the impact of each scale. Subsequently, an upsampling operation is conducted to match the size of the high-resolution feature map with that of the low-resolution feature map. The guide attention (GA) is utilized to merge the high-resolution feature maps with the low-resolution feature maps. The FFB can be expressed as follows:
y i = G A ( x i , u n s a m p l e d ( x i + 1 ) ) i = 1 , 2 , 3 x i i = 4
where xi is the input of each path, yi is the output of the FFB, and GA denotes the guide attention module.
The guide attention (GA) module is shown in Figure 5. First, an upsampling operation is applied to the high-resolution feature to match the size of the low-resolution feature. Next, both features are normalized using a BN layer. Feature attention is obtained using the element-wise product operation. Consequently, a 3 × 3 BConv is employed to generate the modulated feature. Finally, the modulated feature is added to the input feature.
The GA can be expressed as:
f = β ( x i ) β ( u p s a m p l e d ( x i + 1 ) )
y = φ ( f ) + x i
where f denotes the modulated feature, β (·) is the BN operation, φ (·) is the 3 × 3 BConv, upsampled represents the upsampling operation, xi is the low-resolution feature, and xi+1 is the high-resolution feature.

2.3. Experimental Settings

In our study, the model was trained using the OCT2017 and OCT-C4 datasets. For the OCT-C4 dataset, it was divided into a 4:1 ratio for training and testing. In each run, 20% of the training set in both datasets was selected for validation, optimizing the model for improved performance. The training process was repeated five times for each dataset, and the average of the five results was considered as the final result.
The input images were all resized to 224 × 224 for model training. Normalization was performed on the images by subtracting the mean and dividing by the standard deviation. To prevent model overfitting and improve its generalization performance, data augmentation techniques were utilized during the training process to increase the diversity of data features. These techniques included random rotation, scale changes, brightness adjustments, and horizontal flipping. The detailed information for the data augmentation strategy is shown in Table 1.
The cross-entropy loss function is commonly used to classify images. In our study, a weighted cross-entropy loss function was applied to address the problem of class imbalance in the OCT2017 and OCT-C4 datasets. Additionally, we set the number of epochs to 50 and the batch size to 32. Herein, we applied the SGD optimizer with a weight decay of 1 × 10−4 and momentum of 0.9 to optimize the model. The cosine annealing strategy [42] was employed to dynamically adjust the learning rate during initialization. In our study, we initialized the learning rate at 0.001 and set the final learning rate to 1 × 10−5. All model training was based on the transfer learning method.
The experiments were carried out on a workstation equipped with an Intel Xeon Gold 6248R CPU (Intel, Santa Clara, CA, USA) and an NVIDIA RTX A5000 GPU (NVIDIA, Santa Clara, CA, USA). We utilized the PyTorch (v. 2.1.1) deep learning framework to conduct the experiments, leveraging Cuda11.1 and CUDNN10.0 to accelerate the speed of model training.

2.4. Evaluation Criteria

In order to assess the effectiveness of model performance, various metrics including accuracy, precision, sensitivity, specificity, F1-score, and AUC were employed to evaluate the experimental results. These indicators are defined as follows:
A c c u r a c y = T P + T N T P + F P + T N + F N
Pr e c i s i o n = T P T P + F P
S e n s i t i v i t y = T P T P + F N
S p e c i f i c i t y = T N T N + F P
F 1 = 2 × Pr e c i s i o n × S e n s i t i v i t y Pr e c i s i o n + S e n s i t i v i t y
where TP, FP, TN, and FN represent true positive, false positive, true negative, and false negative, respectively.
Based on the multi-class confusion matrix, we needed to calculate the overall accuracy (OA) and overall precision (OP) to evaluate the quantitative results. OA and OP can be defined as:
O A = 1 N i = 1 N T P i
O P = 1 C i = 1 C T P i T P i + F P i
where N is the number of testing samples, and C is the number of total categories.

3. Results and Discussion

3.1. Comparisons on OCT2017 Dataset

The quantitative results of the different methods are shown in Table 2. The methods listed in Table 2 were trained on the OCT2017 dataset using identical experimental conditions as described in Section 2.3. OA and OP refer to overall accuracy and overall precision. The results show the superior performance of the MS-DRCN model compared with well-known CNN models and the retina classification frameworks.
VGG16 [43] and ResNet [44] are the classical convolutional networks. We employed transfer learning to train them on the OCT2017 dataset. Transfer learning [20] was applied by Kermany et al. to fine-tune the Inception v3 model on the OCT2017 dataset. Therefore, the method obtained an accuracy of 94.3%. Fang et al. [22,23] proposed the IFCNN model, which adopted an iterative fusion strategy to accurately identify retinal diseases. FPN-ResNet was proposed by Sotoudeh-Paima et al. [24], who applied a feature pyramid network (FPN) structure to well-known CNN models to acquire multi-resolution information for retinal image classification. For a fair comparison, we implemented the transfer learning, IFCNN, and FPN-ResNet frameworks on the OCT2017 dataset. As shown in Table 2, our model achieved the highest performance in terms of overall accuracy (OA) and precision (OP) on the OCT2017 dataset, with values of 96.4% and 96.6%, respectively. The OA and OP exhibit an enhancement of 2.4% and 2.1% compared with ResNet. The reason may be that MS-DRCN improves the identification of lesion regions by considering multi-scale information and removing noise-related features.
Figure 6 shows the train accuracy, validation accuracy, train loss, and validation loss. In Figure 6b, we can observe that the validation loss set decreased until it reached a plateau. This suggests that there was no overfitting of the MS-DRCN during the training process. Figure 7 displays the correct (first row) and incorrect (second row) classification cases of the OCT images, along with the predicted probability scores for each category. The first row of Figure 7 demonstrates that the MS-DRCN accurately classified the retinal images with high confidence scores. The second row of Figure 7 shows the incorrect classification cases between CNV and DME. This misclassification can be attributed to their similar visual characteristics that involved significant liquid accumulation. However, there were cases where Normal images were misclassified as Drusen images with small lesions, which posed a challenge for the MS-DRCN. Moreover, the majority of misclassifications were observed between Drusen and CNV. Drusen represents the early stage of AMD, characterized by the presence of extracellular material deposits, while CNV corresponds to the later stage of AMD, characterized by the formation of neovascular membranes and significant fluid accumulation. The main reason is similarity between Drusen and CNV, which may confuse the MS-DRCN.

3.2. Comparisons on OCT-C4 Datasets

In order to further validate the effectiveness of the proposed method, we conducted an evaluation of MS-DRCN on the OCT-C4 dataset. The training methods and settings of MS-DRCN on the OCT-C4 dataset were consistent with those on the OCT2017 dataset. We implemented transfer learning, IFCNN, and FPN-ResNet on the OCT-C4 dataset.
Table 3 shows the quantitative analysis results of classification indicators on the OCT-C4 dataset. MS-DRCN generally outperformed other methods in terms of accuracy, precision, sensitivity, OA, and OP. Compared with ResNet, MS-DRCN showed an improvement of 2.5% and 1.7% in OA and OP. Similarly, MS-DRCN exhibited an enhancement of 1.6% and 1.6% in OA and OP compared with FPN-ResNet. The superior performance of MS-DRCN may be attributed to its soft denoising model and multi-scale structure, enabling better identification of the lesion area. As a result, MS-DRCN achieved an OA and OP of 96.5% and 96.9%.

3.3. Robustness of Our Method to Noise

In practical applications, original OCT images often exhibit significant speckle noise due to the interference signal caused by the backscattered light from the biological tissue [32]. Due to the characteristics of OCT imaging, the scattering noise primarily manifests as multiplicative noise. Given the substantial presence of speckle noise in OCT images, it is possible to establish a noise model that accounts for the characteristics of speckle noise. In the logarithmic domain, speckle noise can be effectively approximated as additive Gaussian noise [45]. It can be expressed as follows:
f = x * m + n
where x is the raw image never contaminated by noise, f is the noisy image, and m and n are multiplicative noise and additive noise, respectively. Since additive noise is very small compared with multiplicative noise, it can be ignored.
To make the classification task more challenging and evaluate the robustness of the MS-DRCN, varying intensities of additional noise were introduced into the original OCT images. Since OCT images are primarily affected by speckle noise, and Gaussian noise is commonly employed in anti-noise research [26], we introduced additional speckle noise and Gaussian noise to the test set.
The intensity of noise was determined by the value of the peak signal-to-noise ratio (PSNR). The definition of the PSNR is as follows:
PSNR = 20 log 10 ( M a x M S E )
where Max denotes the maximum pixel value of the OCT image, and MSE refers to the mean squared error. As the value of the PSNR decreases, the severity of the added noise increases.
In our experiments, Gaussian noise and speckle noise were added to the test sets of the OCT2017 and OCT-C4 datasets, resulting in the average PSNR between the original test and its noisy version of 35, 32, 29, 26, 23, and 20 dB, respectively. Notably, we trained the models using the training set of both datasets and evaluated their performance on the test set with Gaussian or speckle noise.
Figure 8 and Figure 9 show the accuracies of the different methods on the test set of the OCT2017 and OCT-C4 datasets. It can be observed that the predicted accuracy of the MS-DRCN was better than that of the other methods under different levels of noise intensity. This means that the MS-DRCN is more robust to different levels of noise. As the PSNR decreases, speckle noise has a greater effect on image quality than Gaussian noise at the same PSNR. The reason may be that multiplicative noise has a greater impact on image quality than additive noise. The experimental results demonstrate the superior robustness of the MS-DRCN to noise.

3.4. Ablation Experiments

This section presents the ablation experiments to assess the effectiveness of various components, including MCB, SDB, and FFB. To evaluate our method, ResNet was selected as the baseline model, which facilitated model optimization. We gradually integrated the FFB, MCB, and SDB modules into the baseline model. After adding the new modules, we retrained the model using the OCT2017 dataset. The effectiveness of the different modules can be known through experimental comparisons.
The quantitative evaluation results on the OCT2017 dataset are shown in Table 4. It is observed that the FFB improved the overall accuracy by 0.9%, indicating the effectiveness of combining low-resolution and high-resolution features for lesion area localization. The MCB captured multi-scale information and enhanced the extracted features by integrating convolution kernels of different sizes. It improved the accuracy by 0.8% compared with the combination of ResNet and FFB. In addition, the soft-denoising block reduced noise-related features by applying soft thresholding during the training process. As a result, the MS-DRCN achieved an overall accuracy (OA), overall precision (OP), and F1-score of 96.4%, 96.6%, and 96.5%, respectively. The MS-DRCN achieved significant performance improvements with a small number of parameters compared with ResNet.

3.5. Visualization

Despite the remarkable success of convolutional neural networks in the field of computer vision, there is still a significant challenge regarding their interpretability. To solve this issue, gradient-weighted class activation mapping (Grad-CAM) [46] is introduced to explain the decision-making process of the CNN model. It highlights significant regions within the image that contribute to accurate predictions. Figure 10 shows the raw image and its Grad-CAM maps of CNV and DME cases. It can be observed that the original image contains a large amount of noise. The activation map of learned features (red regions in Figure 10) is valuable for accurately localizing the lesion area and determining the target type. Retinopathy exhibits a considerable degree of variability and complexity. DME typically presents as fluid-filled cysts and thickening of the retina, while CNV is characterized by retinal fluid accumulation and fibrotic scars. By integrating multi-scale information and reducing noise features, the MS-DRCN enables more accurate lesion localization and prediction.
To further understand the denoising process of the MS-DRCN model, we visualized the learned features. Figure 11 shows the raw image, output of the first residual block, and output after the SDB module in the first residual block of the MS-DRCN, respectively. The feature maps in Figure 11b show that the features exacted by the residual block contain a significant amount of noise features, which are unnecessary for image classification. Figure 11c reveals that the extracted features after soft denoising are more concentrated in the retinal region than in the background region. Clearly, the MS-DRCN is capable of reducing noise-related features during the training process, which further demonstrates the effectiveness of our proposed method.

4. Conclusions

This study introduced a multi-scale denoising residual convolutional network (MS-DRCN) for classifying OCT images of retinal diseases. The key contributions of this research are summarized as follows: (a) designing a soft-denoising block that gradually reduces noise features while preserving useful information; (b) designing a multi-scale context block to capture multi-scale information and enhance extracted features by employing convolutions with varying dilations; and (c) the application of a feature fusion module to integrate features from both low and high resolutions. Distinguishing itself from well-known CNN models and existing multi-scale frameworks, the proposed method not only exploits cross-scale complementarity but also utilizes multi-scale information on a single scale. Furthermore, the impact of speckle noise is effectively mitigated by integrating a soft-denoising block into the MS-DRCN.
The results indicate that the MS-DRCN outperforms well-known models and several recently proposed frameworks for retinal OCT classification. Specifically, the MS-DRCN achieves an accuracy improvement of 2.4% and 2.5% over ResNet50 on the OCT2017 dataset and the OCT-C4 dataset, respectively. Anti-experiments also showed that the MS-DRCN performs better than other methods against various types of noise on the OCT2017 dataset. These results suggest that our method exhibits a superior generalization ability and robustness.

Author Contributions

Conceptualization, J.P. and P.L.; methodology, J.P.; software, J.P.; validation, J.P., J.Z. and P.L.; formal analysis, J.P.; investigation, J.L. and J.Z.; resources, J.P.; data curation, J.P.; writing—original draft preparation, J.P.; writing—review and editing, J.P., J.L. and P.L.; visualization, J.L. and J.Z.; supervision, J.P. and P.L.; project administration, J.P. and P.L.; funding acquisition, P.L. and J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China (NSFC) (61890951, 61890950, 81971659, 62275095, and 82261138559); the Fundamental Research Funds for the Central Universities, HUST (2019kfyXMBZ009); the CAMS Innovation Fund for Medical Sciences (2019-I2M-5-014); Innovation Fund of WNLO; Hainan Provincial Natural Science Foundation of China (grant no. 821QN226); and the Hainan University Research Start-up Fund (grant no. KYQD(ZR)20072 and KYQD(ZR)22074).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used in this study is available in Section 2.1.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gale, R.; Scanlon, P.H.; Evans, M.; Ghanchi, F.; Yang, Y.; Silvestri, G.; Freeman, M.; Maisey, A.; Napier, J. Action on diabetic macular oedema: Achieving optimal patient management in treating visual impairment due to diabetic eye disease. Eye 2017, 31, S1–S20. [Google Scholar] [CrossRef] [PubMed]
  2. Wong, W.L.; Su, X.; Li, X.; Cheung, C.M.; Klein, R.; Cheng, C.Y.; Wong, T.Y. Global prevalence of age-related macular degeneration and disease burden projection for 2020 and 2040: A systematic review and meta-analysis. Lancet Glob. Health 2014, 2, e106–e116. [Google Scholar] [CrossRef] [PubMed]
  3. Puliafito, C.A.; Hee, M.R.; Lin, C.P.; Reichel, E.; Schuman, J.S.; Duker, J.S.; Izatt, J.A.; Swanson, E.A.; Fujimoto, J.G. Imaging of macular diseases with optical coherence tomography. Ophthalmology 1995, 102, 217–229. [Google Scholar] [CrossRef] [PubMed]
  4. Drexler, W.; Fujimoto, J.G. State-of-the-art retinal optical coherence tomography. Prog. Retin. Eye Res. 2008, 27, 45–88. [Google Scholar] [CrossRef] [PubMed]
  5. Huang, D.; Swanson, E.A.; Lin, C.P.; Schuman, J.S.; Stinson, W.G.; Chang, W.; Hee, M.R.; Flotte, T.; Gregory, K.; Puliafito, C.A. Optical coherence tomography. Science 1991, 254, 1178–1181. [Google Scholar] [CrossRef] [PubMed]
  6. van Velthoven, M.E.J.; Faber, D.J.; Verbraak, F.D.; van Leeuwen, T.G.; de Smet, M.D. Recent developments in optical coherence tomography for imaging the retina. Prog. Retin. Eye Res. 2007, 26, 57–77. [Google Scholar] [CrossRef] [PubMed]
  7. Rong, Y.; Xiang, D.; Zhu, W.; Yu, K.; Shi, F.; Fan, Z.; Chen, X. Surrogate-Assisted Retinal OCT Image Classification Based on Convolutional Neural Networks. IEEE J. Biomed. Health Inform. 2019, 23, 253–263. [Google Scholar] [CrossRef]
  8. Rasti, R.; Mehridehnavi, A.; Rabbani, H.; Hajizadeh, F. Convolutional Mixture of Experts Model: A Comparative Study on Automatic Macular Diagnosis in Retinal Optical Coherence Tomography Imaging. J. Med. Signals Sens. 2019, 9, 1–14. [Google Scholar] [CrossRef]
  9. Altan, G. DeepOCT: An explainable deep learning architecture to analyze macular edema on OCT images. Eng. Sci. Technol. Int. J. 2022, 34, 101091. [Google Scholar] [CrossRef]
  10. Wen, H.; Zhao, J.; Xiang, S.; Lin, L.; Liu, C.; Wang, T.; An, L.; Liang, L.; Huang, B. Towards more efficient ophthalmic disease classification and lesion location via convolution transformer. Comput. Methods Programs Biomed. 2022, 220, 106832. [Google Scholar] [CrossRef]
  11. Das, V.; Dandapat, S.; Bora, P.K. Multi-scale deep feature fusion for automated classification of macular pathologies from OCT images. Biomed. Signal Process. Control 2019, 54, 101605. [Google Scholar] [CrossRef]
  12. Albarrak, A.; Coenen, F.; Zheng, Y. Age-related Macular Degeneration Identification In Volumetric Optical Coherence Tomography Using Decomposition and Local Feature Extraction. In Proceedings of the 17th Medical Image, Understanding and Analysis Conference, Aberdeen, UK, 19–21 July 2023; pp. 59–64. [Google Scholar]
  13. Srinivasan, P.P.; Kim, L.A.; Mettu, P.S.; Cousins, S.W.; Comer, G.M.; Izatt, J.A.; Farsiu, S. Fully automated detection of diabetic macular edema and dry age-related macular degeneration from optical coherence tomography images. Biomed. Opt. Express 2014, 5, 3568–3577. [Google Scholar] [CrossRef] [PubMed]
  14. Sun, Y.; Li, S.; Sun, Z. Fully automated macular pathology detection in retina optical coherence tomography images using sparse coding and dictionary learning. J. Biomed. Opt. 2017, 22, 16012. [Google Scholar] [CrossRef] [PubMed]
  15. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  16. Serener, A.; Serte, S. Dry and wet age-related macular degeneration classification using oct images and deep learning. In Proceedings of the 2019 Scientific Meeting on Electrical-Electronics & Biomedical Engineering and Computer Science (EBBT), Istanbul, Turkey, 24–26 April 2019; IEEE: Nicosia, Cyprus, 2019; pp. 1–4. [Google Scholar]
  17. Kaymak, S.; Serener, A. Automated age-related macular degeneration and diabetic macular edema detection on oct images using deep learning. In Proceedings of the 2018 IEEE 14th International Conference on Intelligent Computer Communication and Processing (ICCP), Cluj-Napoca, Romania, 6–8 September 2018; IEEE: Nicosia, Cyprus, 2018; pp. 265–269. [Google Scholar]
  18. Saleh, N.; Abdel Wahed, M.; Salaheldin, A.M. Transfer learning-based platform for detecting multi-classification retinal disorders using optical coherence tomography images. Int. J. Imaging Syst. Technol. 2021, 32, 740–752. [Google Scholar] [CrossRef]
  19. Karri, S.P.; Chakraborty, D.; Chatterjee, J. Transfer learning based classification of optical coherence tomography images with diabetic macular edema and dry age-related macular degeneration. Biomed. Opt. Express 2017, 8, 579–592. [Google Scholar] [CrossRef] [PubMed]
  20. Kermany, D.S.; Goldbaum, M.; Cai, W.; Valentim, C.C.S.; Liang, H.; Baxter, S.L.; McKeown, A.; Yang, G.; Wu, X.; Yan, F.; et al. Identifying Medical Diagnoses and Treatable Diseases by Image-Based Deep Learning. Cell 2018, 172, 1122–1131.e1129. [Google Scholar] [CrossRef]
  21. Huang, L.; He, X.; Fang, L.; Rabbani, H.; Chen, X. Automatic Classification of Retinal Optical Coherence Tomography Images with Layer Guided Convolutional Neural Network. IEEE Signal Process. Lett. 2019, 26, 1026–1030. [Google Scholar] [CrossRef]
  22. Fang, L.; Jin, Y.; Huang, L.; Guo, S.; Zhao, G.; Chen, X. Iterative fusion convolutional neural networks for classification of optical coherence tomography images. J. Vis. Commun. Image Represent. 2019, 59, 327–333. [Google Scholar] [CrossRef]
  23. Fang, L.; Wang, C.; Li, S.; Rabbani, H.; Chen, X.; Liu, Z. Attention to Lesion: Lesion-Aware Convolutional Neural Network for Retinal Optical Coherence Tomography Image Classification. IEEE Trans. Med. Imaging 2019, 38, 1959–1970. [Google Scholar] [CrossRef]
  24. Sotoudeh-Paima, S.; Jodeiri, A.; Hajizadeh, F.; Soltanian-Zadeh, H. Multi-scale convolutional neural network for automated AMD classification using retinal OCT images. Comput. Biol. Med. 2022, 144, 105368. [Google Scholar] [CrossRef] [PubMed]
  25. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2018; pp. 2117–2125. [Google Scholar]
  26. Ma, Z.; Xie, Q.; Xie, P.; Fan, F.; Gao, X.; Zhu, J. HCTNet: A Hybrid ConvNet-Transformer Network for Retinal Optical Coherence Tomography Image Classification. Biosensors 2022, 12, 542. [Google Scholar] [CrossRef] [PubMed]
  27. Wang, J.; Deng, G.; Li, W.; Chen, Y.; Gao, F.; Liu, H.; He, Y.; Shi, G. Deep learning for quality assessment of retinal OCT images. Biomed. Opt. Express 2019, 10, 6057–6072. [Google Scholar] [CrossRef] [PubMed]
  28. Li, F.; Chen, H.; Liu, Z.; Zhang, X.D.; Jiang, M.S.; Wu, Z.Z.; Zhou, K.Q. Deep learning-based automated detection of retinal diseases using optical coherence tomography images. Biomed. Opt. Express 2019, 10, 6204–6226. [Google Scholar] [CrossRef] [PubMed]
  29. He, X.; Deng, Y.; Fang, L.; Peng, Q. Multi-Modal Retinal Image Classification with Modality-Specific Attention Network. IEEE Trans. Med. Imaging 2021, 40, 1591–1602. [Google Scholar] [CrossRef] [PubMed]
  30. Wang, W.; Xu, Z.; Yu, W.; Zhao, J.; Yang, J.; He, F.; Yang, Z.; Chen, D.; Ding, D.; Chen, Y. Two-stream CNN with loose pair training for multi-modal AMD categorization. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2019: 22nd International Conference, Shenzhen, China, 13–17 October 2019; Proceedings, Part I 22, 2019. Springer: Berlin/Heidelberg, Germany, 2019; pp. 156–164. [Google Scholar]
  31. Karthik, K.; Mahadevappa, M. Convolution neural networks for optical coherence tomography (OCT) image classification. Biomed. Signal Process. Control. 2023, 79, 104176. [Google Scholar] [CrossRef]
  32. Ozcan, A.; Bilenca, A.; Desjardins, A.E.; Bouma, B.E.; Tearney, G.J. Speckle reduction in optical coherence tomography images using digital filtering. Int. J. Comput. Vis. 2007, 8, 123–139. [Google Scholar] [CrossRef]
  33. Ruan, F.; Dang, L.; Ge, Q.; Zhang, Q.; Qiao, B.; Zuo, X. Dual-Path Residual “Shrinkage” Network for Side-Scan Sonar Image Classification. Comput. Intell. Neurosci. 2022, 2022, 6962838. [Google Scholar] [CrossRef]
  34. Zhao, M.; Zhong, S.; Fu, X.; Tang, B.; Pecht, M. Deep Residual Shrinkage Networks for Fault Diagnosis. IEEE Trans. Ind. Inform. 2020, 16, 4681–4690. [Google Scholar] [CrossRef]
  35. Isogawa, K.; Ida, T.; Shiodera, T.; Takeguchi, T. Deep Shrinkage Convolutional Neural Network for Adaptive Noise Reduction. IEEE Signal Process. Lett. 2018, 25, 224–228. [Google Scholar] [CrossRef]
  36. Thomas, A.; Harikrishnan, P.M.; Krishna, K.A.; Palanisamy, P.; Gopi, V.P. A novel multiscale convolutional neural network based age-related macular degeneration detection using OCT images. Biomed. Signal Process. Control. 2021, 67, 102538. [Google Scholar] [CrossRef]
  37. Thomas, A.; Harikrishnan, P.M.; Ramachandran, R.; Ramachandran, S.; Manoj, R.; Palanisamy, P.; Gopi, V.P. A novel multiscale and multipath convolutional neural network based age-related macular degeneration detection using OCT images. Comput. Methods Programs Biomed 2021, 209, 106294. [Google Scholar] [CrossRef] [PubMed]
  38. Rasti, R.; Rabbani, H.; Mehridehnavi, A.; Hajizadeh, F. Macular OCT Classification Using a Multi-Scale Convolutional Neural Network Ensemble. IEEE Trans. Med. Imaging 2018, 37, 1024–1034. [Google Scholar] [CrossRef] [PubMed]
  39. Naren, O.S. Retinal OCT-C8. Available online: https://www.kaggle.com/datasets/obulisainaren/retinal-oct-c8 (accessed on 4 August 2022).
  40. Sergey, I.; Christian, S. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  41. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Cluj-Napoca, Romania, 6–8 September 2018; pp. 7132–7141. [Google Scholar]
  42. Loshchilov, I.; Hutter, F. Sgdr: Stochastic gradient descent with warm restarts. arXiv 2016, arXiv:1608.03983. [Google Scholar] [CrossRef]
  43. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  44. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  45. Forouzanfar, M.; Moghaddam, H.A. A Directional Multiscale Approach for Speckle Reduction in Optical Coherence Tomography Images. In Proceedings of the 2007 International Conference Eletrical Engineering, London, UK, 2–4 July 2007; pp. 1–6. [Google Scholar]
  46. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization. IEEE Int. Conf. Comput. Vis. 2020, 128, 336–359. [Google Scholar] [CrossRef]
Figure 1. The overall architecture of MS-DRCN.
Figure 1. The overall architecture of MS-DRCN.
Sensors 24 00150 g001
Figure 2. Residual learning block.
Figure 2. Residual learning block.
Sensors 24 00150 g002
Figure 3. Soft-denoising block.
Figure 3. Soft-denoising block.
Sensors 24 00150 g003
Figure 4. Multi-scale context block.
Figure 4. Multi-scale context block.
Sensors 24 00150 g004
Figure 5. Guide attention module.
Figure 5. Guide attention module.
Sensors 24 00150 g005
Figure 6. (a) Training accuracy and validation accuracy of MS-DRCN on the OCT2017 dataset; (b) training loss and validation loss on the OCT2017 dataset.
Figure 6. (a) Training accuracy and validation accuracy of MS-DRCN on the OCT2017 dataset; (b) training loss and validation loss on the OCT2017 dataset.
Sensors 24 00150 g006
Figure 7. Examples of classification results predicted by the proposed MS-DRCN. The first row shows the cases of correct classification; the second row shows the cases of misclassification.
Figure 7. Examples of classification results predicted by the proposed MS-DRCN. The first row shows the cases of correct classification; the second row shows the cases of misclassification.
Sensors 24 00150 g007
Figure 8. The accuracy of different methods on the test set of OCT2017 dataset with (a) Gaussian noise and (b) speckle noise.
Figure 8. The accuracy of different methods on the test set of OCT2017 dataset with (a) Gaussian noise and (b) speckle noise.
Sensors 24 00150 g008
Figure 9. The accuracy of different methods on the test set of OCT-C4 dataset with (a) Gaussian noise and (b) speckle noise.
Figure 9. The accuracy of different methods on the test set of OCT-C4 dataset with (a) Gaussian noise and (b) speckle noise.
Sensors 24 00150 g009
Figure 10. The raw image and its Grad-CAM of the CNV and DME cases. The red and yellow regions represent the area of interest to the proposed model.
Figure 10. The raw image and its Grad-CAM of the CNV and DME cases. The red and yellow regions represent the area of interest to the proposed model.
Sensors 24 00150 g010
Figure 11. Feature maps visualization. (a) Raw image. (b) The output of first residual block of MS-DRCN. (c) The output after SDB module in the first residual block of MS-DRCN. The feature maps presented in (c) exhibit a higher concentration on the retinal area, specifically in the ones marked by the red boxes, in contrast with (b).
Figure 11. Feature maps visualization. (a) Raw image. (b) The output of first residual block of MS-DRCN. (c) The output after SDB module in the first residual block of MS-DRCN. The feature maps presented in (c) exhibit a higher concentration on the retinal area, specifically in the ones marked by the red boxes, in contrast with (b).
Sensors 24 00150 g011
Table 1. Augmentation strategy in our study.
Table 1. Augmentation strategy in our study.
Augmentation TypeValue
Rotation range±15 degrees
Scale range±20%
Brightness range±20%
Horizontal flipTrue
Table 2. Comparison results for different classification methods on the OCT2017 dataset.
Table 2. Comparison results for different classification methods on the OCT2017 dataset.
MethodClassAccuracy (%)Precision (%)Sensitivity (%)OA (%)OP (%)
Vgg16 [43]CNV94.6 ± 0.482.5 ± 0.699.1 ± 0.293.4 ± 0.594.3 ± 0.5
DME98.9 ± 0.396.6 ± 1.698.3 ± 0.3
Drusen94.6 ± 0.599.3 ± 0.278.6 ± 1.2
Normal98.9 ± 0.298.5 ± 0.297.0 ± 1.0
ResNet [44]CNV95.0 ± 0.585.0 ± 1.497.2 ± 1.294.0 ± 0.694.5 ± 0.5
DME98.7 ± 0.495.5 ± 1.599.5 ± 0.2
Drusen95.1 ± 0.699.3 ± 0.281.0 ± 2.7
Normal98.9 ± 0.298.2 ± 0.698.1 ± 1.0
Transfer learning [20]CNV95.0 ± 0.785.4 ± 2.198.9 ± 0.594.3 ± 0.894.8 ± 0.6
DME99.1 ± 0.498.4 ± 1.397.8 ± 0.6
Drusen95.2 ± 0.697.9 ± 0.282.6 ± 2.4
Normal98.8 ± 0.297.3 ± 0.498.0 ± 0.9
IFCNN [22]CNV95.4 ± 0.384.8 ± 0.799.0 ± 0.594.7 ± 0.295.2 ± 0.3
DME98.8 ± 0.498.4 ± 0.295.5 ± 0.3
Drusen96.0 ± 0.498.0 ± 0.485.1 ± 0.8
Normal98.7 ± 0.398.4 ± 0.398.5 ± 0.4
FPN-ResNet [24]CNV95.9 ± 0.387.2 ± 0.698.2 ± 0.594.6 ± 0.294.9 ± 0.2
DME98.8 ± 0.196.1 ± 0.499.2 ± 0.2
Drusen95.7 ± 0.298.8 ± 0.283.4 ± 0.8
Normal98.9 ± 0.297.8 ± 0.297.3 ± 0.4
MS-DRCNCNV97.4 ± 0.391.1 ± 1.099.2 ± 0.296.4 ± 0.296.6 ± 0.2
DME99.3 ± 0.398.5 ± 1.098.8 ± 0.1
Drusen97.1 ± 0.297.9 ± 0.290.1 ± 0.7
Normal99.0 ± 0.298.7 ± 0.297.5 ± 0.4
Table 3. Comparison results for different classification methods on the OCT-C4 dataset.
Table 3. Comparison results for different classification methods on the OCT-C4 dataset.
MethodClassAccuracy (%)Precision (%)Sensitivity (%)OA (%)OP (%)
Vgg16 [43]CNV97.5 ± 0.597.5 ± 1.191.4 ± 1.493.5 ± 0.694.3 ± 0.3
DME99.2 ± 0.296.2 ± 0.397.6 ± 0.5
Drusen94.5 ± 0.591.8 ± 2.187.5 ± 0.8
Normal95.8 ± 0.391.7 ± 1.297.7 ± 0.5
ResNet [44]CNV98.1 ± 0.297.2 ± 0.694.1 ± 1.294.0 ± 0.495.2 ± 0.4
DME99.3 ± 0.298.4 ± 0.597.0 ± 0.4
Drusen95.0 ± 0.594.0 ± 0.787.0 ± 1.1
Normal96.0 ± 0.591.4 ± 1.498.5 ± 0.6
Transfer learning [20]CNV98.0 ± 0.296.7 ± 0.394.1 ± 0.693.7 ± 0.394.6 ± 0.2
DME99.3 ± 0.397.3 ± 1.397.9 ± 0.6
Drusen94.4 ± 0.394.4 ± 0.584.2 ± 1.1
Normal95.6 ± 0.390.4 ± 0.998.6 ± 0.4
IFCNN [23]CNV98.1 ± 0.297.3 ± 0.294.5 ± 0.594.9 ± 0.395.5 ± 0.3
DME99.2 ± 0.297.0 ± 0.597.4 ± 0.4
Drusen95.7 ± 0.294.1 ± 0.289.7 ± 0.5
Normal96.7 ± 0.293.4 ± 0.398.0 ± 0.2
FPN-ResNet [24]CNV98.4 ± 0.297.0 ± 0.495.8 ± 0.794.9 ± 0.495.3 ± 0.5
DME99.2 ± 0.296.0 ± 1.998.5 ± 0.3
Drusen95.7 ± 0.395.1 ± 0.788.7 ± 0.8
Normal96.5 ± 0.393.2 ± 0.297.6 ± 0.8
MS-DRCNCNV98.7 ± 0.297.4 ± 0.396.8 ± 0.496.5 ± 0.396.9 ± 0.4
DME99.6 ± 0.198.4 ± 0.298.7 ± 0.3
Drusen96.9 ± 0.296.4 ± 0.691.9 ± 0.6
Normal97.0 ± 0.295.3 ± 0.398.7 ± 0.3
Table 4. Comparison results for different classification methods on the OCT-2017 dataset.
Table 4. Comparison results for different classification methods on the OCT-2017 dataset.
MethodOA (%)OP (%)F1 (%)Parameter
ResNet94.0 ± 0.694.5 ± 0.594.2 ± 0.224.5 M
ResNet + FFB94.9 ± 0.495.2 ± 0.395.1 ± 0.426.7 M
ResNet + FFB + MCB95.7 ± 0.295.9 ± 0.295.8 ± 0.226.7 M
MS-DRCN96.4 ± 0.296.6 ± 0.296.5 ± 0.227.3 M
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Peng, J.; Lu, J.; Zhuo, J.; Li, P. Multi-Scale-Denoising Residual Convolutional Network for Retinal Disease Classification Using OCT. Sensors 2024, 24, 150. https://doi.org/10.3390/s24010150

AMA Style

Peng J, Lu J, Zhuo J, Li P. Multi-Scale-Denoising Residual Convolutional Network for Retinal Disease Classification Using OCT. Sensors. 2024; 24(1):150. https://doi.org/10.3390/s24010150

Chicago/Turabian Style

Peng, Jinbo, Jinling Lu, Junjie Zhuo, and Pengcheng Li. 2024. "Multi-Scale-Denoising Residual Convolutional Network for Retinal Disease Classification Using OCT" Sensors 24, no. 1: 150. https://doi.org/10.3390/s24010150

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop