Next Article in Journal
Resection of Colorectal Liver Metastases with Major Vessel Involvement
Next Article in Special Issue
Automated Quantitative Analysis of CT Perfusion to Classify Vascular Phenotypes of Pancreatic Ductal Adenocarcinoma
Previous Article in Journal
Pain Catastrophizing in Cancer Patients
Previous Article in Special Issue
Evaluating Recalibrating AI Models for Breast Cancer Diagnosis in a New Context: Insights from Transfer Learning, Image Enhancement and High-Quality Training Data Integration
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RenseNet: A Deep Learning Network Incorporating Residual and Dense Blocks with Edge Conservative Module to Improve Small-Lesion Classification and Model Interpretation

1
Bionics Research Center, Biomedical Research Division, Korea Institute of Science and Technology (KIST), Seoul 02792, Republic of Korea
2
Department of Urology, Samsung Medical Center (SMC), Seoul 06351, Republic of Korea
*
Author to whom correspondence should be addressed.
Cancers 2024, 16(3), 570; https://doi.org/10.3390/cancers16030570
Submission received: 18 December 2023 / Revised: 16 January 2024 / Accepted: 27 January 2024 / Published: 29 January 2024
(This article belongs to the Special Issue Feature Papers in Section "Methods and Technologies Development")

Abstract

:

Simple Summary

Small-target classification in an image is still challenging in spite of emerging deep learning-based techniques. This study focused on the development of deep learning models for small-lesion classification. The proposed Rense block and edge conservative module enables deep learning models to extract better features of small lesions in images. Our RenseNet was validated with quantitative classification accuracy and a qualitative explanation heatmap for kidney stone and lung tumor computed tomography (CT) image datasets.

Abstract

Deep learning has become an essential tool in medical image analysis owing to its remarkable performance. Target classification and model interpretability are key applications of deep learning in medical image analysis, and hence many deep learning-based algorithms have emerged. Many existing deep learning-based algorithms include pooling operations, which are a type of subsampling used to enlarge the receptive field. However, pooling operations degrade the image details in terms of signal processing theory, which is significantly sensitive to small objects in an image. Therefore, in this study, we designed a Rense block and edge conservative module to effectively manipulate previous feature information in the feed-forward learning process. Specifically, a Rense block, an optimal design that incorporates skip connections of residual and dense blocks, was demonstrated through mathematical analysis. Furthermore, we avoid blurring of the features in the pooling operation through a compensation path in the edge conservative module. Two independent CT datasets of kidney stones and lung tumors, in which small lesions are often included in the images, were used to verify the proposed RenseNet. The results of the classification and explanation heatmaps show that the proposed RenseNet provides the best inference and interpretation compared to current state-of-the-art methods. The proposed RenseNet can significantly contribute to efficient diagnosis and treatment because it is effective for small lesions that might be misclassified or misinterpreted.

1. Introduction

The success of deep learning (DL) in computer vision has led to the increased application of artificial intelligence (AI) in medical image analysis [1,2]. DL-based AI models have been applied to clinical vision applications, such as lesion detection, classification, and segmentation, thereby outperforming conventional machine learning models [3,4,5,6]. Moreover, explainable AI (XAI) allows users to interpret and trust the results predicted by the AI model [7]. Clinical vision applications often require high reliability; therefore, there might be limitations in directly applying DL models in real practice, as the internal process of DL models is a black box type. XAI, an emerging technique in the field of vision, helps to build trust and confidence in model prediction, which can lead to a wider application of DL in the clinical field. Gradient-weighted class activation mapping (Grad-CAM) is the principal XAI algorithm for comprehending the prediction results without modifying the model [8]. Grad-CAM is a generalization of the class activation map (CAM) [9], and recent advanced techniques such as Grad-CAM++ [10], Ablation-CAM [11], and XGrad-CAM [12] are also based on Grad-CAM. Layer-wise relevance propagation (LRP) has been introduced as another method to explain model decisions [13]. However, the explanations are different across DL models, even though the same dataset and XAI algorithm are applied [14]. Therefore, there is room to improve the explanation of results using a DL model design.
Accurate classification of small lesions in medical images for early diagnosis and efficient treatment leads to more effective recovery; however, small-object classification is a challenging and important problem in computer vision research. In general, small objects often exhibit non-evident visual features. Previous studies [3,4] showed that the detection and segmentation accuracy of small lesions is not comparable to that of large targets, despite using DL techniques. This is because information loss of small lesions owing to subsampling (e.g., pooling) is crucial. To overcome the limitations, those studies, which focused on the segmentation task, enhanced performance of a U-Net-based structure by introducing adaptive loss function and task-consistency regularization schemes. In terms of XAI, the natural image dataset (for example, the ImageNet dataset) includes evident and bulky objects; therefore, the heatmap result of Grad-CAM from the DL-based classification model is consistent with human cognition. However, in the medical image dataset, there are many small-lesion scenarios, such as kidney stones, tumors, and blood vessels. The heatmaps estimated from XAI algorithms for small-lesion classification do not provide a reasonable interpretation at times because the prediction results for classification are not propagated backward well in the DL model owing to information loss in the model. Hence, existing DL models exhibit the problem of misprediction and misinterpretation in small-target tasks, which hinders their common use in clinical sites.
In this study, we propose a deep-learning model to improve the classification of small lesions. Our model generates a more reliable explanation heatmap in comparison with state-of-the-art models. The key contributions of this study are as follows.
  • A backbone network, which is composed of Rense blocks merged with residual and dense blocks, was proposed using mathematical analysis such that we can utilize the skip connections more efficiently and provide an insight for model structures.
  • The edge conservative module was designed in the compensation paths, by preserving the image detail lost from the subsampling (pooling) operation. Auxiliary features were then extracted in this module for better inference.
  • The performance of the proposed model was evaluated using a public CT dataset of lung tumors and a private CT dataset of kidney stones, where there are often small-lesion scenarios. Validation results in terms of classification and Grad-CAM-based heatmaps show that the proposed model classifies small lesions better and can provide more trust to clinical users in their decision-making compared with cutting-edge techniques.

2. Related Works

Object classification is a basic task in the field of computer vision. With the development of machine learning, object classification in medical images has also advanced. Before DL was receiving significant attention, classical machine learning algorithms were applied to medical images. Machine learning techniques often require handcrafted features (or manually extracted features), which rely on human intervention, to build a mathematical model in a data-driven manner [15]. For feature engineering in medical image analysis, Haralick features calculated by a gray-level co-occurrence matrix of neighboring gray levels (GLCM) in the image were introduced and compared with other feature descriptors—local binary patterns (LBP), gray-level run-length method (GLRLM), and gray-level difference method (GLDM)—which are well-known statistical texture-based feature descriptors [16]. The auto-correlation Gabor feature (AGF), which is invariant to scale, rotation, and illumination changes in an image, was applied to gastroenterology images [17]. Further, scale-invariant feature transform (SIFT) feature-based representations have achieved gains [18,19]. Over the last few years, aggressive investigations using DL have shifted the paradigm, and convolutional neural networks (CNNs) have been widely applied to object classification with big success. The basic modules of CNNs, such as inception [20], residual [21], and dense [22] blocks, were developed. Recently, EfficientNet [23], which proposed compound scaling for model depth, width, and resolution, and a vision transformer (ViT) [24], which adopted the transformer in natural language processing to the vision field, have been applied to biomedical images [25,26]. In the meantime, various deep learning models that use and combine residual and dense blocks have been studied [27,28]; however, most of them are employed empirically. As such, in this paper, we provide the mathematical analysis and insight for the optimal combination of residual and dense blocks instead of relying on empirical results.

3. Materials and Methods

A full schematic of the proposed network composed of Rense blocks and edge conservative modules is shown in Figure 1. The remainder of this section includes a detailed mathematical and theoretical description of the proposed model.

3.1. Residual and Dense Blocks

Currently, deep neural network (DNN) models often use skip connections to increase learning efficiency. A residual block is defined by a skip connection that learns the residual functions concerning the layer input signal. Then, the main path and skip connection are added before applying nonlinear activation [21]. To easily provide insight, the 2D residual block can be simplified, as shown in Figure 2a. The output of the nth residual block is expressed as:
y = A f x + x W n × H n × C n ,   where   x W n × H n × C n .
In Equation (1), A ( · ) is the activation operator and f ( · ) is a function that includes the convolution, activation, and normalization schemes, where f : W n × H n × C n W n × H n × C n . A dense block also employs skip connections to accumulate layer input signal along the channel direction. Then, the nth dense block with skip connection [1] in Figure 2b results in:
y = f x x W n × H n × 2 C n ,
where ⊚ denotes the concatenation along the channel direction. The residual block effectively reduces the search range in the feature space because it learns only residual signals instead of full output signals. A dense block flexibly utilizes previous features because of their accumulation. Basically, the model can avoid the gradient vanishing problem by skip connections in both blocks such that gradients propagate backward well.

3.2. Rense Block

To take advantage of both blocks, two cases were considered, as shown in Figure 2c,d. Figure 2c shows the case where the dense block is inside the residual block. The output of this block is calculated as follows:
y = A g f x x + x W n × H n × C n ,
where g(∙) is the 1 × 1 convolution operator used to reduce the dimension of the channels. A 1 × 1 convolution generates a channel-wise weighted sum; therefore, concatenation and a 1 × 1 convolution can be demonstrated with linear operations, as shown in Figure 3. In other words, the element of y in the ith channel location, that is, the output feature map in the ith channel, is reformulated as
y i = A k = 0 C n 1 α i , k f k x + β i , k x k + b i + x i , = A k = 0 C n 1 α i , k f k x + k = 0 C n 1 ( β i , k + δ i , k ) x k + b i W n × H n .
Here, f k x W n × H n and x k W n × H n are the elements of f x and x at the kth channel location, respectively; b i is the aggregated bias term of all 1 × 1 convolutions for the ith channel location of the output; and δ i , k = 1 when k = i ; otherwise, δ i , k = 0 . α i , k and β i , k are the kernel weights of f k x and x k for the output element of y in the ith channel location, respectively. Consequently, this block can generalize the residual block in Equation (1) because an element of y in the ith channel location of Equation (1) can be rewritten with that of Equation (4), as follows:
y i = A f i   x + x i W n × H n ,
where both α i , k = 1 and β i , k = 1 when k = i ; otherwise, α i , k = 0 and β i , k = 0 . Here, the additional bias term b i , which can shift the activation function, equals zero. Thus, the original residual block in Figure 2a is a special case of a dense block inside the residual block in Figure 2c because the dense block inside the residual block does not exhibit any limitations for α i , k , β i , k , or b i . Subsequently, this block can obtain a gain from the flexibility of the residual learning, but the original dense effect does not move the needle significantly. In this study, it is defined as a switched Rense (sRense) block.
Equation (5) cannot be further expanded by other linear operations, so we can expect the full effects from addition and concatenation of identity in Equation (5). In other words, the respective skip connections in the residual and dense blocks have their original power, and they are clearly formulated into the sum and concatenation of the input, that is, this block learns the residual function and simultaneously retains the previous features. This block is called Rense, and is the basic block of the proposed RenseNet.

3.3. Edge Conservative Module

Generally, most DNNs comprise a series of layers, including convolution, activation, and pooling operations. The pooling operation is a type of subsampling scheme used to overcome the small receptive field in the convolution kernels. Otherwise, the inference of the model is biased toward the local features of the input data. To extract the global features effectively, the feature maps are resized in two ways—pooling operations and stride factor—in convolution, as follows:
D : W n × H n × C n W n / S n × H n / S n × C n ,
where D is a subsampling operator; S n , C n , W n / S n , and H n / S n are integers; and W n , H n , and C n represent the width, height, and channel dimensions in the nth layer, respectively. After the subsampling operation in the nth layer, the size of the feature maps in the spatial dimension is reduced by a factor of S n in Equation (6), which causes information loss in the original pixel grids and results in blurring: Blurring, owing to max pooling, is also explained in terms of the frequency domain. Subsampling generates aliasing artifacts in the frequency domain [29]. Specifically, max pooling, which is commonly applied in CNNs, is a type of incoherent (pseudo-random) sampling. Its filter does not have a predefined shape because sampling within a 2 × 2 window depends on the input signals. Inspired by compressed-sensing theory, incoherent sampling causes noise-like aliasing artifacts (incoherent interferences) in the frequency domain [30]. Subsequently, noise-like aliased signals are mixed with the original frequency components. In general images, the energy of the high-frequency component is basically lower than that of the low-frequency component; therefore, the high-frequency component is more sensitive to aliasing perturbation, which degrades the image details.
In the computer vision problem, blurring confuses object recognition, and it is a more crucial issue in small object classification because pixels in a small object might merge into neighboring pixels and easily lose their original shape. Therefore, to fully utilize the structural information of the object, we proposed an edge conservative module, as shown in Figure 1 and Figure 4. The degradation of the image detail is compensated by a combination of the residual path and upsampling after the pooling operation, as shown in Figure 4. In the edge conservative module, all compensated features are incorporated into another CNN path to extract auxiliary features. Here, a shallow structure was applied to avoid overfitting because the auxiliary features were sparse.

3.4. RenseNet Details

Our network had two parallel forward paths, as shown in Figure 1: first, the baseline path composed of the Rense blocks in Figure 2d, where the other configurations are the same as those of ResNet50 [21], but all residual blocks are replaced by the proposed Rense blocks and there are transitions before pooling layers; and second, an edge conservative path that has eight 3 × 3 convolution layers and three 1 × 1 layers. Three output layers, then, were used to calculate the classification scores after linear projections (fully connected layers) in this module. The cross-entropy (CE) loss was adopted, and the weights for each CE loss were 1:0.3:0.3 for the baseline, ensemble, and edge conservative paths, respectively. They were empirically set, and the batch size was set to eight. All learning processes were stopped at 150 epochs owing to performance saturation and computation time.

3.5. Dataset and Preparation

In this study, the proposed model was verified using two datasets. The first dataset was composed of CT images obtained from patients with kidney stones. This dataset was acquired from the Samsung Medical Center (SMC), Seoul, South Korea, with Institutional Review Board approval (IRB 2021-04-113). There were 60 patient scans and 1468 slice images (positive stone cases 367 slices and negative cases 1101 slices). Each slice image was 512 × 512, and annotations were made by urologists, radiologists, and clinical experts to determine the presence of kidney stones in the CT images. The dataset was separated into 1236 images of training (positive 309, negative 927) and 232 images of validation (positive 58, negative 174) for fivefold cross-validation. In this dataset, there were only classification labels; therefore, quantitative comparisons were applied to the classification accuracy. The second dataset consisted of CT images of lung cancer tumors. These data were from the Decathlon Challenge [31] in the public domain and included 60 scans with 512 × 512 slice images. There were a total of 3314 slice images (positive tumor cases 1657 slices and negative tumor cases 1657 slices). Specifically, 2814 training images (positive 1407, negative 1407) and 500 validation images (positive 250, negative 250) were prepared for fivefold validation. This public dataset provided segmentation labels such that we could evaluate both classification and explanation accuracy. In our experiments, all data separation for training and validation complied with the patient-level splits. In the training process, random cropping, flip, rotation, and translation were applied for data augmentation.

3.6. Grad-CAM-Based Heatmap

To validate the interpretability, heatmaps were estimated by the Grad-CAM algorithm [9] at the final feature extraction layer of all models. The Grad-CAM results were binarized over an 80% threshold. The public lung dataset provides segmentation labels. Thus, heatmaps can be evaluated using the exact shape and location of the target. The dice similarity coefficient (DSC) between the masks generated by the Grad-CAM-based heatmap and segmentation labels was calculated. Thus, we checked that the models examined the target lesion when they made a decision.

4. Results

4.1. Kidney Stone Dataset

The classification accuracy, precision, recall, specificity, and F1 score for the kidney stone dataset are presented in Table 1, where they are compared against seven models. Their 95% confidence intervals are calculated in Table 2. The proposed RenseNet stands for Rense block with an edge conservative module (W/E).
First, we ensured the effectiveness of the edge conservation module. Most metrics showed improved scores when combined with the edge conservative module (W/E). The maximum gains were 9.05, 26.81, 8.62, 9.20, and 16.60% for accuracy, precision, sensitivity, specificity, and F1 score, respectively. In terms of the basic block, the network based on Rense blocks, instead of sRense blocks, provided the best scores in accuracy, sensitivity, specificity, and F1. The F1 score indicates that the proposed RenseNet (Rense blocks + W/E) has the highest performance in harmonic of precision and sensitivity. A vision transformer (ViT) [24] where pretrained parameters are applied inferred biased results for specificity. The guided attention inference network (GAIN) [32], which was devised for weakly supervised detection with classification labels and general performance improvement, was not effective for small-lesion classification. We conducted t-tests under the null hypothesis H0: μ R e n s e   b l o c k = μ m , where m stands for the compared methods with respect to metrics and μ the mean values in Table 1. We can reject H0 at the significance level 0.05 because all p-values are less than 0.05.

4.2. Lung Tumor Dataset

Similarly, the classification accuracy, precision, recall, specificity, and F1 score for the lung tumor dataset are presented in Table 3. Their 95% confidence intervals are calculated in Table 4.
The trend for the quantitative results with the lung tumor dataset was similar to that with the kidney stone dataset. The edge conservative module improved the scores in most cases.
The maximum gains were 8.92, 8.82, 10.43, 32.39, and 8.47% for accuracy, precision, sensitivity, specificity, and F1 score, respectively. RenseNet (Rense block with an edge conservative module) showed the best scores in terms of accuracy, precision, specificity, and F1 score. Here, ViT and GAIN were not powerful for lung tumor datasets. From the t-test, all p-values are found to be less than 0.05, so score differences in Table 3 are statistically meaningful.
We can quantify the interpretability of the model because the lung dataset provided segmentation labels. From the DSC scores in Table 3, the tendency is similar to results of the other metrics in Table 1 and Table 3. The network composed of Rense blocks outperformed other networks, and an edge conservative module enhanced interpretability. Specifically, the network combined with the Rense block and an edge conservative module (RenseNet) exhibited the best performance. Graphical results for the explainable heatmaps are presented in Figure 5 and Figure 6. For the kidney stone dataset (Figure 5), most models focused on the kidney stone in the CT images, except ResNet50 WO/E, EffieicntNet WO/E, ViT, and GAIN.
However, RenseNet had a more refined and localized heatmap. Heatmaps from the lung dataset accurately indicated the lung tumors in DenseNet121 W/E and the proposed RenseNet. Moreover, RenseNet provided the best localized heatmap, as shown in Figure 6. The visual results of the explainable heatmaps are well aligned with the DSC results in Table 3. Therefore, the Rense block and an edge conservative module contributed to the accurate interpretability of the model. Furthermore, the computational cost for each model was compared. Table 5 implies RenseNet is economical and that there is not a big difference from the best, as transition layers effectively reduce feature size along the channel dimension.

5. Discussion

Many DL models are used to classify targets in images. However, most models are designed for large and evident target objects, for example, the ImageNet dataset. In medical images, the size of the lesions varies, and the boundaries might not be evident [33]. Therefore, it is difficult to match the performance of a model dedicated to large objects with medical image applications. Furthermore, the clinical purposes of DL models require greater reliability and confidence. Thus, XAI can be used to break the bottleneck for widespread DL in real clinical practice. Several algorithms have been developed, and Grad-CAM has become one of the principal algorithms for explaining model decisions. Despite its intuitive analysis for model inference, the explanations from Grad-CAM as well as other XAI algorithms often rely on DL models. Furthermore, model accuracy and interpretability for small lesions might not be appropriate, because most DL models lose their spatial information during inference, which is critical for small-lesion objects [3,33].
To improve both model accuracy and interpretability, we propose a Rense block with an edge conservative module that allows flexible use of skip connection paths in residual and dense blocks, and it compensates for spatial information lost in pooling layers. Through the mathematical analysis of skip connections, we found a way to effectively utilize previous features in feed-forward learning. The compensation module for the edge can prevent distortion or damage to small-lesion features. Basically, the Rense block and an edge conservative module allow the model to fully use information about small lesions.
The overall scores in the kidney stone dataset were better than those in the lung tumor dataset. Kidney stones have brighter HU (Hounsfield unit) than other tissues, which makes the inference easier. The results from both the kidney stone and lung tumor dataset show that RenseNet (Rense blocks with an edge conservative module) performed the best. Specifically, the sRense block is not as powerful as the Rense block, because the sRense block is a generalization of the residual block. We can expect a gain from the flexibility of residual learning, but it weakens the effect of dense learning. However, the Rense block has the merits of both residual and dense blocks, as formulated in Equation (6). This can be observed in the results of the Grad-CAM-based heatmap (Figure 5 and Figure 6) as well. GAIN [32] is the algorithm to refine the result of a Grad-Cam-based heatmap, which can be applied to weakly supervised detection. However, it does not work well in small lesions, because GAIN has a process that hides pixels relevant to the attention regions contributing to classification decision as complete as possible. For small lesions, attention regions are basically small in the heatmap and there are still enough areas not hidden in the image that the model can look over for inference. Thus, the model might be biased if large regions in the image that are not hidden provide other misinformation to the model, as shown in Table 1 and Table 3. The low performance of the ViT can be attributed to the pretrained model that does not describe small lesions well.

6. Conclusions

The proposed network aims to achieve high classification accuracy and reliable explanation of model decisions for small lesions in medical images. The Rense block and edge conservative module satisfied the main goal of this study. The experimental results presented in this study demonstrate the accuracy and reliability of the proposed RenseNet for small lesion lesions. Therefore, the proposed network provides new opportunities for clinical applications. In our network, we introduce an optimized integration of dense and residual blocks; however, we have not yet attained the optimal number of Rense blocks. Future research will focus on determining the ideal number of Rense blocks for improved computational efficiency and analyzing the target lesion size that can be effectively classified by our RenseNet.

Author Contributions

Conceptualization, H.S. and D.H.H.; methodology, H.S. and D.H.H.; software and experiments, S.L. (Seokjun Lee), S.Y., S.L. (Saebom Leem) and S.S.; analysis, H.S., S.L. (Seokjun Lee), S.Y., S.L. (Saebom Leem), S.S. and D.H.H.; writing—review and editing, H.S.; supervision, H.S. and D.H.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the Korea Institute of Science and Technology under grants 2E32211, 2E32271, 2E32341, and 2V10151 and by the SMT-KIST Collaborative Research Program (2E31613).

Institutional Review Board Statement

This dataset was approved by the Institutional Review Board of the Samsung Medical Center (SMC), Seoul, South Korea (IRB 2021-04-113).

Informed Consent Statement

Patient consent was waived due to the retrospective design of this study. As there was no direct patient contact, the Institutional Review Board considered this to be a minimal risk, and so no consent was required.

Data Availability Statement

Individual-level data are not publicly available.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Shen, L.; Zhao, W.; Xing, L. Patient-specific reconstruction of volumetric computed tomography images from a single projection view via deep learning. Nat. Biomed. Eng. 2019, 3, 880–888. [Google Scholar] [CrossRef] [PubMed]
  2. Seo, H.; Badiei Khuzani, M.; Vasudevan, V.; Huang, C.; Ren, H.; Xiao, R.; Jia, X.; Xing, L. Machine learning techniques for biomedical image segmentation: An overview of technical aspects and introduction to state-of-art applications. Med. Phys. 2020, 47, e148–e167. [Google Scholar] [CrossRef]
  3. Seo, H.; Huang, C.; Bassenne, M.; Xiao, R.; Xing, L. Modified U-Net (mU-Net) with incorporation of object-dependent high level features for improved liver and liver-tumor segmentation in CT images. IEEE Trans. Med. Imaging 2019, 39, 1316–1325. [Google Scholar] [CrossRef] [PubMed]
  4. Seo, H.; Bassenne, M.; Xing, L. Closing the gap between deep neural network modeling and biomedical decision-making metrics in segmentation via adaptive loss functions. IEEE Trans. Med. Imaging 2020, 40, 585–593. [Google Scholar] [CrossRef]
  5. Seo, H.; Yu, L.; Ren, H.; Li, X.; Shen, L.; Xing, L. Deep neural network with consistency regularization of multi-output channels for improved tumor detection and delineation. IEEE Trans. Med. Imaging 2021, 40, 3369–3378. [Google Scholar] [CrossRef] [PubMed]
  6. Park, S.; Kim, H.; Shim, E.; Hwang, B.Y.; Kim, Y.; Lee, J.W.; Seo, H. Deep Learning-Based Automatic Segmentation of Mandible and Maxilla in Multi-Center CT Images. Appl. Sci. 2022, 12, 1358. [Google Scholar] [CrossRef]
  7. Gohel, P.; Singh, P.; Mohanty, M. Explainable AI: Current status and future directions. arXiv 2021, arXiv:2107.07045. [Google Scholar]
  8. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization. arXiv 2016, arXiv:1610.02391. [Google Scholar]
  9. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning Deep Features for Discriminative Localization. arXiv 2015, arXiv:1512.04150. [Google Scholar]
  10. Chattopadhay, A.; Sarkar, A.; Howlader, P.; Balasubramanian, V.N. Grad-cam++: Generalized gradient-based visual explanations for deep convolutional networks. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 839–847. [Google Scholar]
  11. Ramaswamy, H.G. Ablation-cam: Visual explanations for deep convolutional network via gradient-free localization. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass, CO, USA, 1–5 March 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 983–991. [Google Scholar]
  12. Fu, R.; Hu, Q.; Dong, X.; Guo, Y.; Gao, Y.; Li, B. Axiom-based grad-cam: Towards accurate visualization and explanation of cnns. arXiv 2020, arXiv:2008.02312. [Google Scholar]
  13. Binder, A.; Montavon, G.; Lapuschkin, S.; Müller, K.-R.; Samek, W. Layer-wise relevance propagation for neural networks with local renormalization layers. In Proceedings of the 25th International Conference on Artificial Neural Networks, Barcelona, Spain, 6–9 September 2016; Springer: Cham, Switzerland, 2016; pp. 63–71. [Google Scholar]
  14. Choi, S.J.; Kim, E.S.; Choi, K. Prediction of the histology of colorectal neoplasm in white light colonoscopic images using deep learning algorithms. Sci. Rep. 2021, 11, 5311. [Google Scholar] [CrossRef]
  15. Ganatra, N. A Comprehensive Study of Applying Object Detection Methods for Medical Image Analysis. In Proceedings of the 2021 8th International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 17–19 March 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 821–826. [Google Scholar]
  16. Kitanovski, I.; Trojacanec, K.; Dimitrovski, I.; Loskovska, S. Modality classification using texture features. In International Conference on ICT Innovations; Springer: Berlin/Heidelberg, Germany, 2011; pp. 189–198. [Google Scholar]
  17. Riaz, F.; Silva, F.B.; Ribeiro, M.D.; Coimbra, M.T. Invariant gabor texture descriptors for classification of gastroenterology images. IEEE Trans. Biomed. Eng. 2012, 59, 2893–2904. [Google Scholar] [CrossRef] [PubMed]
  18. Manivannan, S.; Wang, R.; Trucco, E. Inter-cluster features for medical image classification. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Boston, MA, USA, 14–18 September 2014; Springer: Cham, Switzerland, 2014; pp. 345–352. [Google Scholar]
  19. Khan, S.; Yong, S.-P.; Deng, J.D. Ensemble classification with modified sift descriptor for medical image modality. In Proceedings of the 2015 International Conference on Image and Vision Computing New Zealand (IVCNZ), Auckland, New Zealand, 23–24 November 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1–6. [Google Scholar]
  20. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1–9. [Google Scholar]
  21. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015, arXiv:1512.03385. [Google Scholar]
  22. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 4700–4708. [Google Scholar]
  23. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; PMLR: New York, NY, USA, 2019; pp. 6105–6114. [Google Scholar]
  24. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16×16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  25. Müftüoğlu, Z.; Kizrak, M.A.; Yildlnm, T. Differential privacy practice on diagnosis of COVID-19 radiology imaging using EfficientNet. In Proceedings of the 2020 International Conference on INnovations in Intelligent SysTems and Applications (INISTA), Novi Sad, Serbia, 24–26 August 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–6. [Google Scholar]
  26. Wang, H.; Ji, Y.; Song, K.; Sun, M.; Lv, P.; Zhang, T. ViT-P: Classification of Genitourinary Syndrome of Menopause from OCT Images Based on Vision Transformer Models. IEEE Trans. Instrum. Meas. 2021, 70, 1–14. [Google Scholar] [CrossRef]
  27. Chen, W.F.; Ou, H.Y.; Lin, H.Y.; Wei, C.P.; Liao, C.C.; Cheng, Y.F.; Pan, C.T. Development of Novel Residual-Dense-Attention (RDA) U-Net Network Architecture for Hepatocellular Carcinoma Segmentation. Diagnostics 2022, 12, 1916. [Google Scholar] [CrossRef] [PubMed]
  28. Srivastava, A.; Jha, D.; Chanda, S.; Pal, U.; Johansen, H.D.; Johansen, D.; Riegler, M.A.; Ali, S.; Halvorsen, P. MSRF-Net: A multi-scale residual fusion network for biomedical image segmentation. IEEE J. Biomed. Health Inform. 2021, 26, 2252–2263. [Google Scholar] [CrossRef] [PubMed]
  29. Hénaff, O.J.; Simoncelli, E.P. Geodesics of learned representations. arXiv 2015, arXiv:1511.06394. [Google Scholar]
  30. Lustig, M.; Donoho, D.L.; Santos, J.M.; Pauly, J.M. Compressed sensing MRI. IEEE Signal Process. Mag. 2008, 25, 72–82. [Google Scholar] [CrossRef]
  31. Simpson, A.L.; Antonelli, M.; Bakas, S.; Bilello, M.; Farahani, K.; Van Ginneken, B.; Kopp-Schneider, A.; Landman, B.A.; Litjens, G.; Menze, B.; et al. A large annotated medical image dataset for the development and evaluation of segmentation algorithms. arXiv 2019, arXiv:1902.09063. [Google Scholar]
  32. Li, K.; Wu, Z.; Peng, K.-C.; Ernst, J.; Fu, Y. Guided attention inference network. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 2996–3010. [Google Scholar] [CrossRef] [PubMed]
  33. Seo, H.; So, S.; Yun, S.; Lee, S.; Barg, J. Spatial Feature Conservation Networks (SFCNs) for Dilated Convolutions to Improve Breast Cancer Segmentation from DCE-MRI. In International Workshop on Applications of Medical AI; Springer: Cham, Switzerland, 2022; pp. 118–127. [Google Scholar]
Figure 1. Schematic diagram of the proposed RenseNet (Rense blocks with an edge conservative module).
Figure 1. Schematic diagram of the proposed RenseNet (Rense blocks with an edge conservative module).
Cancers 16 00570 g001
Figure 2. Simplification of basic blocks. (a) Residual block; (b) dense block; (c) switched Rense (sRense) block—dense block inside a residual block; and (d) Rense block—residual block inside a dense block (orange layer: activation, gray layer: 1 × 1 convolution for channel reduction).
Figure 2. Simplification of basic blocks. (a) Residual block; (b) dense block; (c) switched Rense (sRense) block—dense block inside a residual block; and (d) Rense block—residual block inside a dense block (orange layer: activation, gray layer: 1 × 1 convolution for channel reduction).
Cancers 16 00570 g002
Figure 3. Visualization of mathematical analysis for sRense block. This graphical detail shows linear decomposition of Equation (3), which turns into Equation (4).
Figure 3. Visualization of mathematical analysis for sRense block. This graphical detail shows linear decomposition of Equation (3), which turns into Equation (4).
Cancers 16 00570 g003
Figure 4. An example of the feature maps extracted by the baseline path and the edge conservative module.
Figure 4. An example of the feature maps extracted by the baseline path and the edge conservative module.
Cancers 16 00570 g004
Figure 5. Graphical results of explainable heatmaps in kidney stone cases.
Figure 5. Graphical results of explainable heatmaps in kidney stone cases.
Cancers 16 00570 g005
Figure 6. Graphical results of explainable heatmaps in lung tumor cases.
Figure 6. Graphical results of explainable heatmaps in lung tumor cases.
Cancers 16 00570 g006
Table 1. Quantitative comparisons for kidney stone dataset. Here, WO/E and W/E are without and with the edge conservative module, respectively. Our RenseNet is the combined Rense block with edge conservative module.
Table 1. Quantitative comparisons for kidney stone dataset. Here, WO/E and W/E are without and with the edge conservative module, respectively. Our RenseNet is the combined Rense block with edge conservative module.
Kidney Stone
Metrics ResNet50DenseNet121EfficientNet [23]ViT [24]GAIN [32]sRense BlockRense Block
AccuracyWO/E96.9895.2695.6975.8670.3783.1997.80
W/E97.4196.9896.12--92.2497.84
PrecisionWO/E96.3691.2391.3875.0039.9866.6796.41
W/E96.4398.1198.04--93.4896.49
SensitivityWO/E91.3889.5391.385.1781.2565.5294.75
W/E93.1089.6686.21--74.1494.83
SpecificityWO/E96.8297.0997.0595.3785.9489.0897.13
W/E98.8699.4198.85--98.2899.43
F1 scoreWO/E93.8090.3791.389.6753.5966.0995.57
W/E94.7493.6991.75--82.6995.65
Table 2. 95% confidence intervals for kidney stone dataset. Here, WO/E and W/E are without and with the edge conservative module, respectively. Our RenseNet is the combined Rense block with edge conservative module.
Table 2. 95% confidence intervals for kidney stone dataset. Here, WO/E and W/E are without and with the edge conservative module, respectively. Our RenseNet is the combined Rense block with edge conservative module.
Kidney Stone
Metrics ResNet50DenseNet121EfficientNet [23]ViT [24]GAIN [32]sRense BlockRense Block
AccuracyWO/E0.00610.00040.01020.00370.00320.00230.0127
W/E0.00690.01080.0037 0.00480.0153
PrecisionWO/E0.00070.01580.00940.01360.01570.00760.0057
W/E0.00520.00570.0064 0.00530.0071
SensitivityWO/E0.00360.00690.00730.00370.00350.00580.0027
W/E0.00320.01150.0161 0.00140.0024
SpecificityWO/E0.01280.01110.00900.01040.01270.01400.0090
W/E0.01420.01100.0086 0.01670.0029
F1 scoreWO/E0.00370.00450.01190.00650.01460.00670.0071
W/E0.01450.00870.0064 0.00030.0039
Table 3. Quantitative comparisons for lung stone dataset. Here, WO/E and W/E are without and with the edge conservative module, respectively. Our RenseNet is the combined Rense block with edge conservative module.
Table 3. Quantitative comparisons for lung stone dataset. Here, WO/E and W/E are without and with the edge conservative module, respectively. Our RenseNet is the combined Rense block with edge conservative module.
Lung Tumor
Metrics ResNet50DenseNet121EfficientNet [23]ViT [24]GAIN [32]sRense BlockRENSE Block
AccuracyWO/E66.6776.9868.4571.6365.4371.4377.01
W/E75.5977.5873.61--73.4179.58
PrecisionWO/E71.7271.0482.3165.3720.8169.3776.73
W/E80.5479.3984.68--70.6984.97
SensitivityWO/E56.8991.2348.4191.4090.5878.8783.55
W/E58.4184.0058.84--82.0484.23
SpecificityWO/E77.6362.0178.6851.4557.0765.2989.56
W/E94.0494.4076.36--66.1294.85
F1 scoreWO/E63.4579.8860.9676.2233.8473.8279.99
W/E67.7181.6369.43--75.9484.59
DSCWO/E25.4940.8622.8428.0126.6542.3349.13
W/E31.7752.2728.20--58.9763.52
Table 4. 95% confidence intervals for lung stone dataset. Here, WO/E and W/E are without and with the edge conservative module, respectively. Our RenseNet is the combined Rense block with edge conservative module.
Table 4. 95% confidence intervals for lung stone dataset. Here, WO/E and W/E are without and with the edge conservative module, respectively. Our RenseNet is the combined Rense block with edge conservative module.
Lung Tumor
Metrics ResNet50DenseNet121EfficientNet [23]ViT [24]GAIN [32]sRense BlockRense Block
AccuracyWO/E0.01240.00170.00520.00390.01520.01750.0076
W/E0.01440.00780.0029 0.01860.0208
PrecisionWO/E0.00280.00840.00700.01160.00700.01740.0109
W/E0.00840.02020.0197 0.01490.0001
SensitivityWO/E0.01700.01460.01530.01970.00610.00080.0175
W/E0.01680.00920.0160 0.00220.0093
SpecificityWO/E0.00420.00180.01120.02100.01710.01850.0109
W/E0.00510.01540.0202 0.01380.0089
F1 scoreWO/E0.01250.01010.01400.00240.02040.00510.0099
W/E0.00830.00570.0159 0.00530.0195
DSCWO/E2.71903.09033.81341.71932.74622.24371.7087
W/E1.67112.49343.40141.90833.32001.66072.9846
Table 5. Total number of parameters of the models. Here, WO/E and W/E are without and with the edge conservative module, respectively. Our RenseNet is the combined Rense block with edge conservative module.
Table 5. Total number of parameters of the models. Here, WO/E and W/E are without and with the edge conservative module, respectively. Our RenseNet is the combined Rense block with edge conservative module.
ResNet50DenseNet121EfficientNet [23]ViT [24]GAIN [32]sRense BlockRense Block
Total Number of Parameters [ × 106]WO/E23.516.9563.5885.64103.450.130.15
W/E41.9124.6691.39--0.410.43
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Seo, H.; Lee, S.; Yun, S.; Leem, S.; So, S.; Han, D.H. RenseNet: A Deep Learning Network Incorporating Residual and Dense Blocks with Edge Conservative Module to Improve Small-Lesion Classification and Model Interpretation. Cancers 2024, 16, 570. https://doi.org/10.3390/cancers16030570

AMA Style

Seo H, Lee S, Yun S, Leem S, So S, Han DH. RenseNet: A Deep Learning Network Incorporating Residual and Dense Blocks with Edge Conservative Module to Improve Small-Lesion Classification and Model Interpretation. Cancers. 2024; 16(3):570. https://doi.org/10.3390/cancers16030570

Chicago/Turabian Style

Seo, Hyunseok, Seokjun Lee, Sojin Yun, Saebom Leem, Seohee So, and Deok Hyun Han. 2024. "RenseNet: A Deep Learning Network Incorporating Residual and Dense Blocks with Edge Conservative Module to Improve Small-Lesion Classification and Model Interpretation" Cancers 16, no. 3: 570. https://doi.org/10.3390/cancers16030570

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop