Next Article in Journal
Mendelian Randomisation Analysis of Causal Association between Lifestyle, Health Factors, and Keratoconus
Previous Article in Journal
A Comprehensive Review on Synergy of Multi-Modal Data and AI Technologies in Medical Diagnosis
Previous Article in Special Issue
A Critical Assessment of Generative Models for Synthetic Data Augmentation on Limited Pneumonia X-ray Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature Extraction Based on Local Histogram with Unequal Bins and a Recurrent Neural Network for the Diagnosis of Kidney Diseases from CT Images

by
Abdorreza Alavi Gharahbagh
1,
Vahid Hajihashemi
1,
José J. M. Machado
2 and
João Manuel R. S. Tavares
2,*
1
Faculdade de Engenharia, Universidade do Porto, Rua Dr. Roberto Frias, s/n, 4200-465 Porto, Portugal
2
Instituto de Ciência e Inovação em Engenharia Mecânica e Engenharia Industrial, Departamento de Engenharia Mecânica, Faculdade de Engenharia, Universidade do Porto, Rua Dr. Roberto Frias, s/n, 4200-465 Porto, Portugal
*
Author to whom correspondence should be addressed.
Bioengineering 2024, 11(3), 220; https://doi.org/10.3390/bioengineering11030220
Submission received: 8 February 2024 / Revised: 20 February 2024 / Accepted: 22 February 2024 / Published: 25 February 2024
(This article belongs to the Special Issue Artificial Intelligence (AI) for Medical Image Processing)

Abstract

:
Kidney disease remains one of the most common ailments worldwide, with cancer being one of its most common forms. Early diagnosis can significantly increase the good prognosis for the patient. The development of an artificial intelligence-based system to assist in kidney cancer diagnosis is crucial because kidney illness is a global health concern, and there are limited nephrologists qualified to evaluate kidney cancer. Diagnosing and categorising different forms of renal failure presents the biggest treatment hurdle for kidney cancer. Thus, this article presents a novel method for detecting and classifying kidney cancer subgroups in Computed Tomography (CT) images based on an asymmetric local statistical pixel distribution. In the first step, the input image is non-overlapping windowed, and a statistical distribution of its pixels in each cancer type is built. Then, the method builds the asymmetric statistical distribution of the image’s gradient pixels. Finally, the cancer type is identified by applying the two built statistical distributions to a Deep Neural Network (DNN). The proposed method was evaluated using a dataset collected and authorised by the Dhaka Central International Medical Hospital in Bangladesh, which includes 12,446 CT images of the whole abdomen and urogram, acquired with and without contrast. Based on the results, it is possible to confirm that the proposed method outperformed state-of-the-art methods in terms of the usual correctness criteria. The accuracy of the proposed method for all kidney cancer subtypes presented in the dataset was 99.89%, which is promising.

1. Introduction

Two kidneys are in the human body, located behind the abdomen and protected by the thorax. Blood purification is the most critical function of the kidneys. To keep the blood and body cells clean, the kidneys remove toxic substances produced during metabolic processes in the blood. In addition, the kidneys play a fundamental role in vitamin D metabolism [1,2]. Kidney failure can lead to death within a few days if the blood is not cleansed. Despite all the advances, kidney cancer continues to spread worldwide, with over 400,000 new cases diagnosed each year [1]. Unlike most cancers, 59% of the kidney cancer cases occur in developed countries, and the mortality rate of kidney cancer is higher in these countries [1,2]. More than 10% of the world’s population suffers from Chronic Kidney Disease (CKD), which was ranked as the 16th leading cause of death in 2016 and is expected to rise to 5th place by 2040. Kidney cancer is the ninth most common cancer in men and the fourteenth most common cancer in women [3,4]. There is a severe shortage of nephrologists and radiologists in many regions of the world. For example, in South Asia, the ratio of nephrologists to the population is extremely low, with only one nephrologist per one million people. Kidney cancer usually manifests as nodules; although, tumours can form in one or both kidneys. Malignant cells can enter the bloodstream and affect other organs [5,6]. Approximately 90% of the kidney cancers are caused by metastasis. Cyst formation, nephrolithiasis, or kidney stones, and renal cell carcinoma, i.e., kidney tumours, are the most common diseases affecting kidney function after metastasis. Early diagnosis of kidney disorders such as cysts, stones, and tumours is essential to prevent kidney failure. The prevalence of kidney disease, the global shortage of nephrologists and radiologists, and the advent of medical systems based on Deep Learning (DL) paradigms highlight the importance of developing Artificial Intelligence (AI) solutions to diagnose kidney cancer from medical images. Such solutions can help physicians and reduce patient distress [7]. Many recent studies have been conducted to diagnose kidney cancer using medical images. The limitations of these methods include the lack of valid datasets, the diversity of the medical imaging devices, and the structural differences between images acquired using different devices. Compared with long-established Machine Learning (ML) methods, DL methods have led to the development of new medical systems. Concepts such as pre-trained DL models have also been considered in diagnosing kidney cancer abnormalities. Based on these explanations, a DL-based method for kidney cancer classification is proposed in this article. The main innovations of the proposed method are as follows:
  • The use of gradients and histograms with asymmetric intervals to extract features to classify kidney cancer subtypes accurately;
  • The use of feature extraction before applying the DL model to reduce the dimensionality of the input data compared to the conventional methods;
  • The reduction in the dimensions of the input data to increase the training speed and reduce the complexity of the used DL model.
The proposed method, which involves calculating the intensity and gradient histograms of the input kidney Computed Tomography (CT) image at asymmetric intervals, reduces the input dimensions and effectively diagnoses various kidney cancer-related complications. The structure of the remainder of this article is as follows: The following section provides background information and details of applying ML and DL techniques to detect kidney abnormalities. Section 3 explains the methodology and the proposed method. Section 4 describes the obtained results and compares the proposed method with recent studies in this field. Finally, Section 5 provides the conclusion.

2. Literature Review

ML-based methods can accurately segment healthy patients or those with small nodes. However, kidney segmentation is particularly challenging [8,9]. Verma et al. [10] preprocessed kidney ultrasound images using median and Gaussian filters and morphological operations. Using Principal Component Analysis (PCA), the authors extracted image information and classified the kidney images using the K-nearest Neighbour (KNN) algorithm. Khalifa et al. [11] presented a 3D technique for characterising kidneys in abdominal CT images using a level set-based deformable model and Markov–Gibbs random field. Wolz et al. [12] proposed an approach based on a hierarchical atlas registration and weighting scheme to characterise different body organs in abdominal CT images. Yang et al. [13] also proposed a two-step method for kidney segmentation in CT angiographic images based on multi-atlas image registration. Zhao et al. [14] presented a 3D approach to segment kidneys in 3D CT images, which employed an iterative technique to improve resolution. Shehata et al. [15] employed a level set and Markov–Gibbs random field-based framework to discriminate kidneys in diffusion Magnetic Resonance (MR) images. Khalifa et al. [16] used a composite framework that combined non-negative matrix factorisation with a guided active contour model to segment kidneys in 3D images. Skalski et al. [17] proposed a dynamic contour-based method using a level-set structure for kidney segmentation in CT images.
Several ML algorithms have been investigated for kidney image classification, including Decision Trees (DT), Random Forest (RF), Support Vector Machines (SVM), Multilayer Perceptron (MLP), Naive Bayes, and KNN [18]. The best results were obtained using the KNN and Naive Bayes classifiers. Due to the increasing use of DL in image processing and classification, many studies have been conducted to explore its applications. One of the most promising DL applications can be found in the field of medical image processing and analysis. For example, transfer-learning techniques have been used to process medical images using pre-trained DL models. Compared with conventional DL models, pre-trained models have shown better results. ResNet [19], InceptionNet [20], ExceptionNet [21], and EfficientNet [22] are some of the pre-trained models that use transfer learning in medical image classification. Many studies have used attention-based deep pre-trained models for image analysis, including Vision Transformer (ViT) [23], Big Transformer (BiT) [24], External Attention Network (EANet) [25], Compact Convolutional Transformer (CCT) [26], and Shifted Window Transformer (SWT) [27]. Therefore, the number of DL approaches that have been proposed to assist the diagnosis of diseases from image data is continuously increasing. Yang et al. [28], Haghighi et al. [29], and Mehta et al. [30] used fully connected convolutional networks to segment kidney images. Fu et al. [31] used ultrasound images to input an attention module to segment kidney cysts in CT images. Da Cruz et al. [32] presented a fully automated method for segmenting kidneys with and without tumours on CT images. The method consisted of a preprocessing step for histogram normalisation, CNN models, and a post-processing block. The first CNN model for CT image classification and feature reduction was AlexNet, whereas the second model was U-Net, which accurately segmented the kidneys. In [18,33], kidney ultrasound images were analysed using pre-trained DNN models, mainly ResNet-101, ShuffleNet, and MobileNet-v2, for feature extraction, and an SVM, for the classification. The above review indicates that research on kidney images can be divided into two main categories: one focuses on kidney image segmentation, and the other focuses on diagnosing kidney disorders. Due to the limited number of studies addressing the diagnosis of kidney disease using medical images and the complexity of the required processing, this study proposes a new method to identify and categorise kidney cancer subtypes in CT images.

3. Materials and Methods

In this section, the theoretical details of the proposed method are given, along with the selection of its parameter values. Figure 1 shows the block diagram of the adopted methodology. Removing extra borders of the input image is a preprocessing step. The main steps of the methodology include the extraction of the input image histogram and the use of a Recurrent Neural Network (RNN). The selection of the correct asymmetric intervals in the histogram extraction and the use of the RNN model are novelties of the present study. Two blocks, namely for examining correctness and adjusting the DL parameters, were added to maximise the efficiency of the developed methodology. The final step involves the comparison and validation of the obtained results.

3.1. Image Black Border Removal

The border-cutting step seeks to expedite the further feature extraction by shrinking the input image and eliminating extraneous details. Figure 2 shows an original CT image and the corresponding one obtained after applying this procedure. With this processing, the dark margin of the input image is eliminated, and the image is reduced. Furthermore, because CT images contain many pixels with nearly 0 (zero) intensity values, removing these margin pixels decreases the severity of the histogram imbalance.

3.2. Local Histogram with Asymmetric Intervals

An asymmetric interval is taken onto the reference to find the normalised histogram of the regions in the cropped image. The reason for choosing an asymmetric interval for the histogram is the higher pixel density in the darker intervals of the input image compared with the lighter ones. This asymmetry is necessary to achieve an adequate separation in the statistical representation of the image’s pixels. A well-known weakness of the histogram is its global form. To solve this issue, windowing is performed, which involves dividing the input CT image into 100 windows with identical dimensions, i.e., ten intervals in width and height. An asymmetric normalised histogram is then built independently for each window. An eight-point histogram yields a 100 × 8 feature matrix for the input image. Figure 3 shows an example of a windowed image.

3.3. Image Gradient Histogram with Asymmetric Intervals

In addition to extracting the input CT image’s local histogram, the image’s gradient histogram is used to detect the edges associated with the kidney. As all kidney complications are usually visible as edges, the density of these edges can provide a statistical expression of the kidney condition. Similar to the previous step, the input image is first divided into 100 windows with equal dimensions, and then, the asymmetric normalised histogram is built from its gradient. The intervals of the gradient histogram are chosen differently from those of the intensity image because the gradient has a much higher density in the dark regions of the input image; therefore, the intervals are much denser in the 0 regions and more sparse in the lighter areas of the image. The resulting 100 × 8 matrix is merged with the intensity histogram, and the merged matrix is converted into a vector and used as a 1 × 1600 input for a DNN.

3.4. Recurrent Neural Network

Figure 4 illustrates the architecture of the RNN used in this study, which consists of sequential input and output layers and three additional layers. The second layer is a Long Short-Term Memory Network (LSTM), a vital processing layer commonly employed in DL models for sequential data processing.
The LSTM layer can retain sequential input data over long periods, making it useful for various signal processing problems, including of medical imaging. The main advantage of the LSTM layer is that it overcomes the usual memory limitations of the RNNs. In conventional RNNs, the influence of distant samples becomes insignificant and eventually tends toward 0, making them inapplicable for learning long-term dependencies. The LSTM layer solves this problem by updating its weights using gradients while retaining the input data longer. As shown in Figure 5, LSTMs have three main gates, input, forget, and output, which control their output.
Each hidden unit maintains an internal state updated at each time step based on the input and previous state. Increasing the number of hidden units improves the model’s ability to capture complex patterns in the data but also increases the model’s computational complexity and training time. Typically, a combination of LSTM and dropout layers is used in the training phase to prevent overfitting. In the dropout layer, some units of the LSTM layer are randomly disabled during training, which means that they are not updated. However, during the prediction, all units are active. This approach effectively reduces the risk of overfitting. The LSTM layer can operate according to two output modes: sequence mode and last mode. In the first mode, the LSTM generates an output equal to the input length for each input sequence. This mode is often used for time series classification or forecasting tasks. In the second mode, the LSTM produces only one output for the entire input sequence, meaning that only the final response is provided. In addition, an LSTM has several other parameters that control its behaviour, including the gate activation function, weight initialisation, and weight regularisation. Overall, an LSTM is a robust layer for DL models. By incorporating memory cells and information flow-controlling gates, the LSTM layer can effectively capture long-term dependencies in the data while mitigating the challenge of minor sample effects common in traditional RNNs. The mathematical relationships of the LSTM structure shown in Figure 5 are as follows:
f t = σ f W x f x t + W h f h t 1 + b f , g t = σ g W x c x t + W h c h t 1 + b c , i t = σ f W x i x t + W h i h t 1 + b i , c t = f t c t 1 g t i t , o t = σ f W x o x t + W h o h t 1 + b o , h t = o t c t ,
where x t is the input, h t 1 is the previous state, c t is the last state saved in the forget gate, W is the weight matrix, b is the bias of each gate, σ is the activation function, ⊙ is the pointwise multiplication, and h t is the final output. The third layer, the dense layer, is a fully connected layer commonly used in DL models for classification and regression tasks. This layer connects each neuron in the LSTM layer to all neurons in the current layer. The fully connected layer can effectively learn the relationships between the input data and output classes by making such connections. Moreover, the weight matrix is updated during training to minimise errors. In general, the result of the fully connected layer is fed into the output layer through the softmax activation function. This function maps the output to a range of real numbers between 0 (zero) and 1 (one) and creates a probability distribution for the output classes. The softmax activation function is commonly used in DL models involving classification tasks. The output of the softmax layer takes the form of a probability vector, where each element indicates the probability of occurrence of a particular class. The softmax output is calculated by taking the exponential of each input vector element and then normalising the resulting values. This normalisation step ensures that the resultant vector is a probability distribution, a vital characteristic of the softmax activation function. The softmax function is expressed as follows:
softmax ( x i ) = exp ( x i ) / s u m ( exp ( x j ) ) ,
where x i is the input of the ith neuron, and exp is the exponential function. Throughout the training phase, the weights of the LSTM and FC layers are modified to reduce the cross-entropy loss function. The cross-entropy loss function models the dissimilarity between predicted and accurate data. The cross-entropy loss function is described as follows:
L = s u m ( y t r u e y p r e d ) ,
where y t r u e is the actual value, and y p r e d is the predicted value.

4. Results and Discussion

This section presents the used evaluation criteria and dataset, the obtained results, and the comparison of the proposed method with state-of-the-art methods. Thus, first, the evaluation criteria used for the medical image classification methods are presented. Then, the chosen dataset is described. Finally, implementation details, a study on the effects of each method’s parameter on its accuracy, and a comparison between the results of the proposed model and the latest methods in this field are given. All implementations were performed using the MATLAB 2022b software, and the hardware used consisted of an i7-8850H 2.60 GHz (12 CPUs), 16 GB RAM, and an Nvidia Quadro P2000 graphics card.

4.1. Evaluation Metrics

The evaluation of classification models usually relies on metrics such as accuracy, precision, recall, F1 score, and the Matthews Correlation Coefficient (MCC). Among these metrics, accuracy, the most commonly used metric, denotes the proportion of correctly classified samples to the total number of samples. The F1 score measures the balance between precision and recall, as it is the harmonic mean of these two metrics. MCC is a valuable metric when there is an imbalance in the number of classes in the used dataset, such as detecting kidney cysts, where there may be many images without cysts compared to those with cysts. MCC simultaneously considers true and false results in positive and negative samples, allowing for a more accurate performance assessment than accuracy or precision alone. Here, True Positive (TP) is defined as a correctly identified sample belonging to a particular class, whereas True Negative (TN) represents a correctly identified sample that does not belong to the corrected class. On the other hand, False Positive (FP) refers to a sample assigned to a class that does not belong, and False Negative (FN) refers to a sample that belongs to a class but is incorrectly identified as not belonging to one [34,35]. Therefore, one has the following evaluation metrics:
Accuracy = TP + TN TP + TN + FP + FN , Precision = TP TP + FP , Recall = TP TP + FN , F 1 score = 2 × Precision × Recall / Precision + Recall , MCC = TP × TN FP × FN TP + FP × TP + FN × TN + FP × TN + FN .
Additionally, the confusion matrix was used, which is a table containing the values of TP, TN, FP, and FN from which all evaluation criteria can be derived.

4.2. Used Dataset

In this study, a dataset of kidney cysts, stones, and tumours [34] was used, which includes images of healthy, cystic, stone, and tumour-affected kidneys obtained from the Picture Archiving and Communication System (PACS) and workstations of the Dhaka Central International Medical Hospital (DCIMCH), in Dhaka, Bangladesh, where the related data collection and experimental procedures were authorised [34]. According to the dataset description, the image headers, including patient information and various labels, were removed, and the DICOM images were converted to the JPG format. The Philips IntelliSpace Portal software, a standard tool for radiology equipment, was used to label the images. In addition, the DICOM images were converted to JPG format using the Sante DICOM Editor software. Finally, all images were reviewed and labelled by a physician [34]. The used dataset includes 12,446 images, Table 1. Examples of the images included in the dataset are shown in Figure 6, where the red delineations indicate disease issues. For this study, both coronal and axial sections from the CT images of the whole abdomen and urogram, acquired with and without contrast, were selected. The unequal number of samples across categories within the kidney dataset (Table 1) may affect the effectiveness of the method under study. To mitigate this effect on the learning process of the proposed method, approximately 70% (1000 samples) of the smallest category, namely the stone category, was used for training, while the remaining samples were allocated for testing. In addition, an equivalent number of data samples from the other categories was selected for training, which was equal to the training samples from the stone category. This ensured that exactly 1000 samples from each class were used in the training phase, while the remaining samples were assigned for testing (Table 1). Therefore, the used training set included 4000 samples, while 8446 samples were assigned to the test, which means that 32 % of the samples were assigned to the training and the rest to the test. Both training and test data were randomly selected.

4.3. Experimental Results

This section presents the results obtained in each step of the proposed method and the selection scheme of its critical parameters. The results are discussed taking into account the effect of the CT image intensity histogram intervals on the method’s accuracy, the impact of the histogram of the image gradient intervals on the method’s accuracy, and the performance of state-of-the-art methods.

4.3.1. Selection of Asymmetric Intervals of the Intensity Histogram

The proposed method builds an asymmetric histogram using different image windows, which is one of the most important contributions of the current study. The asymmetric interval of the histogram is used to separate the pixels with values near 0, owing to the large number of dark pixels in the input image. This approach provides a more accurate statistical representation of CT images. However, the selection of the most suitable asymmetric interval is crucial for obtaining accurate results. The optimal interval should maximise the intraclass similarity of all windows and minimise the similarity of windows in different classes. The average correlation coefficient between each window and its corresponding window in the same classes, less the average correlation coefficient between each window and its corresponding window in other classes, was used to generate an objective function. The chosen histogram interval is assumed to be the same for all the windows. Thus, the input vector of the cost function consists of seven integers between 0 and 255 values without repetition. A genetic algorithm was used to optimise the cost function. The defined optimal intervals are presented in Table 2. As an example, Figure 7 shows four histograms of a window in four different classes in two symmetric and asymmetric histograms. Based on this figure, one can perceive that, in the conventional histogram, the values of the last three intervals are almost 0. However, all of our defined intervals have values showing that the selected intervals have increased histogram data. In addition, the proposed intervals decrease the similarity of the histogram values in different classes, e.g., two, three, and four samples.

4.3.2. Selection of Asymmetric Intervals for the Gradient Histogram

Another important novelty of the proposed method is the building of the asymmetric gradient histogram of the input CT image. The asymmetric intervals of the histogram for feature extraction from the gradient images are different from the used asymmetric intervals of the CT intensity images because they have more dark pixels. This helps distinguish the high density of pixels with near 0 values and improves the feature dissimilarity in different classes. The process for determining the optimal interval is the same as that used for the intensity images, except that the input is derived from the gradient of the input image. Because the gradient captures only edges, it is generally darker than a standard image, which results in a narrower range near 0 values and a broader range in bright areas, as expected. The defined optimal intervals are presented in Table 3.
Compared with Table 2, the pixels are dense around 0 values and are widely distributed at higher intervals. To illustrate this, Figure 8 shows four gradient histogram examples in two standard and suggested intervals in a window with four different classes.

4.4. Experimental Results

The efficiency of the proposed method was evaluated by training its DL model using four different approaches. First, the histogram features of the CT intensity images with the suggested intervals were used to train the model. Second, the model was trained using the histogram of the CT gradient images. Third, the model was trained using both features. To analyse the effect of asymmetric intervals on the method’s accuracy, a histogram of symmetric intervals was used for feature extraction in both CT intensity and gradient images, and the extracted features were used to train the model. Table 4, Table 5, Table 6 and Table 7 show the confusion matrix built for each of the four classes for the training and test groups. The results confirm that combining the intensity and gradient histograms of the CT images yielded the best accuracy. Table 7 reveals a significant loss of efficiency when standard histograms were used, highlighting the importance of selecting an asymmetric interval for the histogram building.
Table 8 includes state-of-the-art studies that used the same kidney dataset and their corresponding accuracies for each class. In [34,36,37], a better performance than other studies in terms of overall accuracy was found. However, the proposed method outperformed all previous approaches on all correctness criteria. Moreover, regarding the data assigned for training and testing, the proposed method achieved a better response using less training data, indicating its proper efficiency. The proposed method has several unique features that distinguish it from the other methods. One such feature is the use of intensity and gradient histogram features. This reduces the input dimensions and converts the two-dimensional (2D) DL model into a one-dimensional (1D) model. As an advantage, the dimension reduction of input features and conversion to a 1D model reduces the processing and complexity of the model. The second advantage is the feature extraction speed of the proposed method because both the gradient and histogram operators used in the proposed method are computationally fast. Therefore, the proposed method is excellent regarding response time. Based on these findings, one can conclude that the use of histogram and gradient operators in local form can be a simple but effective approach for the classification of medical images in numerous applications.

5. Conclusions

In this study, An innovative method for categorising kidney anomalies, including cysts, healthy kidneys, tumours, and stones, was presented. The method uses CT scans of the kidney as input and builds the images’ intensity and gradient histograms from non-overlapping windows. Then, a five-layer DL model featuring an LSTM layer is trained using these attributes to classify CT images according to the kidney issue. The feature extraction step involves the building of the histograms using asymmetric intervals because the CT images and their gradients usually have many pixels with dark values. The optimum selection of asymmetric histogram intervals is the primary factor contributing to the suggested method’s effectiveness. The histogram intervals that maximised the similarity of the retrieved features across classes and the differences in features between classes were chosen using a cost function. Additionally, the resultant features are combined into a feature vector, which significantly reduces the number of weights and the overall complexity of the used DNN model compared to the related two-dimensional model. Compared to current efforts in this field, testing the suggested method under various situations on the chosen dataset showed that it is highly efficient. The method has demonstrated superior performance compared to existing methods, outperforming them in various evaluation criteria. Specifically, it achieved a precision of 0.984 , a recall of 0.996 , an F1 score of 0.991 , and an MCC of 0.990. The effectiveness of the proposed method can be attributed to its distinctive characteristics, such as the thoughtful selection of relevant features and the design of its DL model comprising five layers. The suggested model structure seamlessly aligns with the identified features, which increases the method’s strength and efficiency. As part of future work, an enhanced version of the proposed method could incorporate additional features, such as texture and time-frequency features. This modification will extend the classification capabilities beyond those offered by the initially proposed method, allowing for the identification of a broader range of classes. Validating the proposed method using different datasets is critical to confirm its performance, mainly in clinical usage, and should be addressed in the near future. Moreover, the explainability of the proposed method should be studied to facilitate its use by clinicians.

Author Contributions

Conceptualization and supervision: J.M.R.S.T.; investigation, data collection, formal analysis, and writing—original draft preparation: A.A.G. and V.H.; writing—review and editing: J.J.M.M. and J.M.R.S.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

A publicly available dataset was used in this study, which can be found here: https://www.kaggle.com/datasets/nazmul0087/ct-kidney-dataset-normal-cyst-tumor-and-stone (accessed on 27 November 2019).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Fund, W.C.R. Kidney Cancer Statistics. 2019. Available online: https://www.wcrf.org/dietandcancer/cancer-trends/kidney-cancer-statistics (accessed on 27 November 2019).
  2. Society, A.C. What Is Kidney Cancer? 2017. Available online: https://www.cancer.org/cancer/kidney-cancer/about/what-is-kidney-cancer (accessed on 27 November 2019).
  3. Rossi, S.H.; Klatte, T.; Usher-Smith, J.; Stewart, G.D. Epidemiology and screening for renal cancer. World J. Urol. 2018, 36, 1341–1353. [Google Scholar] [CrossRef] [PubMed]
  4. Znaor, A.; Lortet-Tieulent, J.; Laversanne, M.; Jemal, A.; Bray, F. International variations and trends in renal cell carcinoma incidence and mortality. Eur. Urol. 2015, 67, 519–530. [Google Scholar] [CrossRef] [PubMed]
  5. Shuch, B.; Amin, A.; Armstrong, A.J.; Eble, J.N.; Ficarra, V.; Lopez-Beltran, A.; Martignoni, G.; Rini, B.I.; Kutikov, A. Understanding pathologic variants of renal cell carcinoma: Distilling therapeutic opportunities from biologic complexity. Eur. Urol. 2015, 67, 85–97. [Google Scholar] [CrossRef] [PubMed]
  6. Seyfried, T.N.; Huysentruyt, L.C. On the origin of cancer metastasis. Crit. Rev. Oncog. 2013, 18, 43–73. [Google Scholar] [CrossRef] [PubMed]
  7. Islam, M. Ct KIDNEY DATASET: Normal-Cyst-Tumor and Stone. Kaggle.com. 2021. Available online: https://www.kaggle.com/datasets/nazmul0087/ct-kidney-dataset-normal-cyst-tumor-and-stone/code (accessed on 18 February 2024).
  8. Bergeron, A. The pulmonologist’s point of view on lung infiltrates in haematological malignancies. Diagn. Interv. Imaging 2013, 94, 216–220. [Google Scholar] [CrossRef] [PubMed]
  9. Candemir, S.; Jaeger, S.; Palaniappan, K.; Musco, J.P.; Singh, R.K.; Xue, Z.; Karargyris, A.; Antani, S.; Thoma, G.; McDonald, C.J. Lung segmentation in chest radiographs using anatomical atlases with nonrigid registration. IEEE Trans. Med. Imaging 2013, 33, 577–590. [Google Scholar] [CrossRef] [PubMed]
  10. Verma, J.; Nath, M.; Tripathi, P.; Saini, K. Analysis and identification of kidney stone using K th nearest neighbour (KNN) and support vector machine (SVM) classification techniques. Pattern Recognit. Image Anal. 2017, 27, 574–580. [Google Scholar] [CrossRef]
  11. Khalifa, F.; Gimel’farb, G.; El-Ghar, M.A.; Sokhadze, G.; Manning, S.; McClure, P.; Ouseph, R.; El-Baz, A. A new deformable model-based segmentation approach for accurate extraction of the kidney from abdominal CT images. In Proceedings of the 2011 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; pp. 3393–3396. [Google Scholar] [CrossRef]
  12. Wolz, R.; Chu, C.; Misawa, K.; Fujiwara, M.; Mori, K.; Rueckert, D. Automated abdominal multi-organ segmentation with subject-specific atlas generation. IEEE Trans. Med. Imaging 2013, 32, 1723–1730. [Google Scholar] [CrossRef] [PubMed]
  13. Yang, G.; Gu, J.; Chen, Y.; Liu, W.; Tang, L.; Shu, H.; Toumoulin, C. Automatic kidney segmentation in CT images based on multi-atlas image registration. In Proceedings of the 2014 36th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Chicago, IL, USA, 26–30 August 2014; pp. 5538–5541. [Google Scholar] [CrossRef]
  14. Zhao, E.; Liang, Y.; Fan, H. Contextual information-aided kidney segmentation in CT sequences. Opt. Commun. 2013, 290, 55–62. [Google Scholar] [CrossRef]
  15. Shehata, M.; Khalifa, F.; Soliman, A.; Alrefai, R.; Abou El-Ghar, M.; Dwyer, A.C.; Ouseph, R.; El-Baz, A. A level set-based framework for 3D kidney segmentation from diffusion MR images. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; pp. 4441–4445. [Google Scholar] [CrossRef]
  16. Khalifa, F.; Soliman, A.; Takieldeen, A.; Shehata, M.; Mostapha, M.; Shaffie, A.; Ouseph, R.; Elmaghraby, A.; El-Baz, A. Kidney segmentation from CT images using a 3D NMF-guided active contour model. In Proceedings of the 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI), Prague, Czech Republic, 13–16 April 2016; pp. 432–435. [Google Scholar] [CrossRef]
  17. Skalski, A.; Heryan, K.; Jakubowski, J.; Drewniak, T. Kidney segmentation in ct data using hybrid level-set method with ellipsoidal shape constraints. Metrol. Meas. Syst. 2017, 24, 101–112. [Google Scholar] [CrossRef]
  18. Aksakalli, I.; Kaçdioğlu, S.; Hanay, Y.S. Kidney X-ray images classification using machine learning and deep learning methods. Balk. J. Electr. Comput. Eng. 2021, 9, 144–151. [Google Scholar] [CrossRef]
  19. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  20. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  21. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  22. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning (PMLR), Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  23. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar] [CrossRef]
  24. Kolesnikov, A.; Beyer, L.; Zhai, X.; Puigcerver, J.; Yung, J.; Gelly, S.; Houlsby, N. Big transfer (bit): General visual representation learning. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020, Proceedings, Part V 16; Springer: Berlin/Heidelberg, Germany, 2020; pp. 491–507. [Google Scholar] [CrossRef]
  25. Guo, M.H.; Liu, Z.N.; Mu, T.J.; Hu, S.M. Beyond self-attention: External attention using two linear layers for visual tasks. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 5436–5447. [Google Scholar] [CrossRef]
  26. Hassani, A.; Walton, S.; Shah, N.; Abuduweili, A.; Li, J.; Shi, H. Escaping the big data paradigm with compact transformers. arXiv 2021, arXiv:2104.05704. [Google Scholar] [CrossRef]
  27. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  28. Yang, G.; Li, G.; Pan, T.; Kong, Y.; Wu, J.; Shu, H.; Luo, L.; Dillenseger, J.L.; Coatrieux, J.L.; Tang, L.; et al. Automatic segmentation of kidney and renal tumor in ct images based on 3d fully convolutional neural network with pyramid pooling module. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 3790–3795. [Google Scholar] [CrossRef]
  29. Haghighi, M.; Warfield, S.K.; Kurugol, S. Automatic renal segmentation in DCE-MRI using convolutional neural networks. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 1534–1537. [Google Scholar] [CrossRef]
  30. Mehta, P.; Sandfort, V.; Gheysens, D.; Braeckevelt, G.J.; Berte, J.; Summers, R.M. Segmenting the kidney on ct scans via crowdsourcing. In Proceedings of the 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), Venice, Italy, 8–11 April 2019; pp. 829–832. [Google Scholar] [CrossRef]
  31. Fu, X.; Liu, H.; Bi, X.; Gong, X. Deep-learning-based CT imaging in the quantitative evaluation of chronic kidney diseases. J. Healthc. Eng. 2021, 2021, 3774423. [Google Scholar] [CrossRef]
  32. da Cruz, L.B.; Araújo, J.D.L.; Ferreira, J.L.; Diniz, J.O.B.; Silva, A.C.; de Almeida, J.D.S.; de Paiva, A.C.; Gattass, M. Kidney segmentation from computed tomography images using deep neural network. Comput. Biol. Med. 2020, 123, 103906. [Google Scholar] [CrossRef] [PubMed]
  33. Sudharson, S.; Kokil, P. An ensemble of deep neural networks for kidney ultrasound image classification. Comput. Methods Programs Biomed. 2020, 197, 105709. [Google Scholar] [CrossRef]
  34. Islam, M.N.; Hasan, M.; Hossain, M.K.; Alam, M.G.R.; Uddin, M.Z.; Soylu, A. Vision transformer and explainable transfer learning models for auto detection of kidney cyst, stone and tumor from CT-radiography. Sci. Rep. 2022, 12, 11440. [Google Scholar] [CrossRef] [PubMed]
  35. Bayram, A.F.; Gurkan, C.; Budak, A.; Karataş, H. A detection and prediction model based on deep learning assisted by explainable artificial intelligence for kidney diseases. Avrupa Bilim Teknol. Derg. 2022, 40, 67–74. [Google Scholar] [CrossRef]
  36. Asif, S.; Yi, W.; Si, J.; Ain, Q.U.; Yi, Y.; Hou, J. Modeling a Fine-Tuned Deep Convolutional Neural Network for Diagnosis of Kidney Diseases from CT Images. In Proceedings of the 2022 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Las Vegas, NV, USA, 6–8 December 2022; pp. 2571–2576. [Google Scholar] [CrossRef]
  37. Bhandari, M.; Yogarajah, P.; Kavitha, M.S.; Condell, J. Exploring the Capabilities of a Lightweight CNN Model in Accurately Identifying Renal Abnormalities: Cysts, Stones, and Tumors, Using LIME and SHAP. Appl. Sci. 2023, 13, 3125. [Google Scholar] [CrossRef]
Figure 1. Diagram of the developed methodology.
Figure 1. Diagram of the developed methodology.
Bioengineering 11 00220 g001
Figure 2. A CT image before (on the left) and after removing its black margin (on the right).
Figure 2. A CT image before (on the left) and after removing its black margin (on the right).
Bioengineering 11 00220 g002
Figure 3. An original CT image (on the left) and the corresponding windowed image (on the right).
Figure 3. An original CT image (on the left) and the corresponding windowed image (on the right).
Bioengineering 11 00220 g003
Figure 4. The architecture of the used RNN.
Figure 4. The architecture of the used RNN.
Bioengineering 11 00220 g004
Figure 5. A common LSTM structure.
Figure 5. A common LSTM structure.
Bioengineering 11 00220 g005
Figure 6. Examples of the images included in the used dataset (delineations in red indicate disease issues) [34].
Figure 6. Examples of the images included in the used dataset (delineations in red indicate disease issues) [34].
Bioengineering 11 00220 g006
Figure 7. Intensity histograms of a sample window in a CT image in four classes with symmetric intervals (on the left) and suggested intervals (on the right).
Figure 7. Intensity histograms of a sample window in a CT image in four classes with symmetric intervals (on the left) and suggested intervals (on the right).
Bioengineering 11 00220 g007
Figure 8. Gradient histograms of a sample window in a CT image with four classes with symmetric intervals (on the left) and suggested intervals (on the right).
Figure 8. Gradient histograms of a sample window in a CT image with four classes with symmetric intervals (on the left) and suggested intervals (on the right).
Bioengineering 11 00220 g008
Table 1. Number of images of each class and number of training and testing samples.
Table 1. Number of images of each class and number of training and testing samples.
TotalTrainTest
Cyst370910002709
Normal507710004077
Stone13771000377
Tumour228310001283
Total12,44640008446
Table 2. Normal intervals and suggested intervals for a CT image intensity histogram.
Table 2. Normal intervals and suggested intervals for a CT image intensity histogram.
Symmetric Interval0326496128159191223255
Suggested Range0816324880100130255
Table 3. Normal intervals and suggested intervals for histogram of the used CT gradient images.
Table 3. Normal intervals and suggested intervals for histogram of the used CT gradient images.
Symmetric Interval0326496128159191223255
Suggested Range0481216244896255
Table 4. Confusion matrix of the proposed method for training and test sets with its model trained with an asymmetric intensity histogram.
Table 4. Confusion matrix of the proposed method for training and test sets with its model trained with an asymmetric intensity histogram.
TrainTest
CystNormalStoneTumourCystNormalStoneTumour
Cyst10000002709000
Normal2998003407400
Stone029980093680
Tumour00799300131271
Table 5. Confusion matrix of the proposed method in the training and test data with its model trained with an asymmetric gradient histogram.
Table 5. Confusion matrix of the proposed method in the training and test data with its model trained with an asymmetric gradient histogram.
TrainTest
CystNormalStoneTumourCystNormalStoneTumour
Cyst10000002707002
Normal129450431240131933
Stone0999100173600
Tumour00899200121271
Table 6. Confusion matrix of the proposed method trained with combined intensity and gradient histograms.
Table 6. Confusion matrix of the proposed method trained with combined intensity and gradient histograms.
TrainTest
CystNormalStoneTumourCystNormalStoneTumour
Cyst10000002709000
Normal2998002407410
Stone039970013600
Tumour0039970051278
Table 7. Confusion matrix of the proposed method trained with symmetric combined intensity and gradient histograms.
Table 7. Confusion matrix of the proposed method trained with symmetric combined intensity and gradient histograms.
TrainTest
CystNormalStoneTumourCystNormalStoneTumour
Cyst80801920213105780
Normal198910004066110
Stone2139850063710
Tumour91229682215591187
Table 8. Comparison of the proposed and latest related methods.
Table 8. Comparison of the proposed and latest related methods.
ModelAccuracyClassPrecisionRecallF1 ScoreMCC
YOLOv7 [35]Cyst0.8920.6330.740.673
Normal
Stone0.8190.8550.8360.816
Tumour0.93610.9660.960
Average0.8820.8290.8540.648
EANet [34]77.02%Cyst0.59310.7450.788
Normal0.8960.8480.8710.616
Stone0.8450.4950.6240.821
Tumour0.930.7770.8470.994
Swin Transformer [34]99.30%Cyst0.9960.9960.9960.981
Normal0.9960.9810.9880.983
Stone0.9810.9890.9850.996
Tumour0.99310.9960.923
CCT [34]96.54%Cyst0.9680.9230.9450.970
Normal0.9890.9750.9820.966
Stone0.9410.9690.956
Tumour0.9640.9640.9640.974
VGG16 [34]98.20%Cyst0.9960.9680.9820.965
Normal0.9850.9730.9790.974
Stone0.9660.9880.9770.986
Tumour0.9820.9960.9890.596
Inception v3 [34]61.60%Cyst0.6450.8260.7240.465
Normal0.5840.8980.7080.459
Stone0.5680.4620.5090.412
Tumour0.760.2950.4250.566
Resnet50 [34]73.80%Cyst0.7350.6410.6850.625
Normal0.770.790.780.684
Stone0.7450.6920.7170.706
Tumour0.7060.8270.7620.673
Deep CNN [36]99.25%Cyst0.9710.981
Normal1110.994
Stone10.9910.988
Tumour10.980.991
Lightweight CNN [37]99.52%Cyst0.9940.9990.9980.995
Normal0.9950.9970.9970.993
Stone0.9970.9790.9880.986
Tumour0.9930.9950.9950.993
Proposed method99.89%Cyst0.999110.999
Normal10.99910.999
Stone0.9840.9970.9910.990
Tumour10.9960.9980.998
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gharahbagh, A.A.; Hajihashemi, V.; Machado, J.J.M.; Tavares, J.M.R.S. Feature Extraction Based on Local Histogram with Unequal Bins and a Recurrent Neural Network for the Diagnosis of Kidney Diseases from CT Images. Bioengineering 2024, 11, 220. https://doi.org/10.3390/bioengineering11030220

AMA Style

Gharahbagh AA, Hajihashemi V, Machado JJM, Tavares JMRS. Feature Extraction Based on Local Histogram with Unequal Bins and a Recurrent Neural Network for the Diagnosis of Kidney Diseases from CT Images. Bioengineering. 2024; 11(3):220. https://doi.org/10.3390/bioengineering11030220

Chicago/Turabian Style

Gharahbagh, Abdorreza Alavi, Vahid Hajihashemi, José J. M. Machado, and João Manuel R. S. Tavares. 2024. "Feature Extraction Based on Local Histogram with Unequal Bins and a Recurrent Neural Network for the Diagnosis of Kidney Diseases from CT Images" Bioengineering 11, no. 3: 220. https://doi.org/10.3390/bioengineering11030220

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop