Next Article in Journal
Self-Medication during and after Cancer: A French Nation-Wide Cross-Sectional Study
Next Article in Special Issue
Automated Detection and Scoring of Tumor-Infiltrating Lymphocytes in Breast Cancer Histopathology Slides
Previous Article in Journal
Adenosine Methylation Level of miR-125a-5p Promotes Anti-PD-1 Therapy Escape through the Regulation of IGSF11/VSIG3 Expression
Previous Article in Special Issue
Henry Gas Solubility Optimization Algorithm based Feature Extraction in Dermoscopic Images Analysis of Skin Cancer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Kidney Cancer Diagnosis and Surgery Selection by Machine Learning from CT Scans Combined with Clinical Metadata

1
Department of Electrical Engineering, Qatar University, Doha 2713, Qatar
2
Urology Division, Surgery Department, Sidra Medicine, Doha 26999, Qatar
3
Department of Surgery, Weill Cornell Medicine-Qatar, Doha 24811, Qatar
4
College of Medicine, Qatar University, Doha 2713, Qatar
5
Clinical Imaging Department, Hamad Medical Corporation, Doha 3050, Qatar
6
Department of Electrical and Electronics Engineering, University of Dhaka, Dhaka 1000, Bangladesh
*
Authors to whom correspondence should be addressed.
Cancers 2023, 15(12), 3189; https://doi.org/10.3390/cancers15123189
Submission received: 1 May 2023 / Revised: 30 May 2023 / Accepted: 7 June 2023 / Published: 14 June 2023
(This article belongs to the Collection Artificial Intelligence and Machine Learning in Cancer Research)

Abstract

:

Simple Summary

Diagnosis is the most important step in treating and managing kidney cancer, requiring accurate identification, localization, and classification of tumor regions. The selection of appropriate surgical procedures for malignant cases is further based on tumor volume and relative severity. In recent years, machine-learning-based approaches have been proposed to localize, quantify, and stratify kidney tumors using contrast-enhanced computed tomography (CT) images. However, previous studies have largely neglected the integration of patient metadata with clinical images to better diagnose and guide surgical interventions. In the current study, we developed a combined clinical and image-based approach to classify kidney cancers using a publicly available dataset. We show that the inclusion of clinical features alongside medical images improves the performance of kidney tumor classification. We further used clinical data together with a machine-learning approach to predict the expected surgical procedure employed in individual kidney cancer patients. In addition to cancer stage and tumor volume, some surprisingly common demographic features were revealed to be key determinants of the surgical procedure later selected for nephrectomy.

Abstract

Kidney cancers are one of the most common malignancies worldwide. Accurate diagnosis is a critical step in the management of kidney cancer patients and is influenced by multiple factors including tumor size or volume, cancer types and stages, etc. For malignant tumors, partial or radical surgery of the kidney might be required, but for clinicians, the basis for making this decision is often unclear. Partial nephrectomy could result in patient death due to cancer if kidney removal was necessary, whereas radical nephrectomy in less severe cases could resign patients to lifelong dialysis or need for future transplantation without sufficient cause. Using machine learning to consider clinical data alongside computed tomography images could potentially help resolve some of these surgical ambiguities, by enabling a more robust classification of kidney cancers and selection of optimal surgical approaches. In this study, we used the publicly available KiTS dataset of contrast-enhanced CT images and corresponding patient metadata to differentiate four major classes of kidney cancer: clear cell (ccRCC), chromophobe (chRCC), papillary (pRCC) renal cell carcinoma, and oncocytoma (ONC). We rationalized these data to overcome the high field of view (FoV), extract tumor regions of interest (ROIs), classify patients using deep machine-learning models, and extract/post-process CT image features for combination with clinical data. Regardless of marked data imbalance, our combined approach achieved a high level of performance (85.66% accuracy, 84.18% precision, 85.66% recall, and 84.92% F1-score). When selecting surgical procedures for malignant tumors (RCC), our method proved even more reliable (90.63% accuracy, 90.83% precision, 90.61% recall, and 90.50% F1-score). Using feature ranking, we confirmed that tumor volume and cancer stage are the most relevant clinical features for predicting surgical procedures. Once fully mature, the approach we propose could be used to assist surgeons in performing nephrectomies by guiding the choices of optimal procedures in individual patients with kidney cancer.

1. Introduction

Chronic kidney diseases (CKDs) are progressive conditions that confer gradual loss of renal function, ultimately leading to kidney failure [1]. CKDs are often associated with malignancy and have been a major cause of death throughout the 21st century: rates vary between 7–12% across regions [1], accounting for a staggering 800 million cases globally in 2017 alone [2]. Kidney cancer itself is the 14th most common malignancy worldwide (9th among men) with more than 430,000 cases diagnosed worldwide in the year 2020 [3]. Based on several recent studies [4,5,6], CKDs are now recognized as the primary reason for many kidney cancers [7,8], but also vice-versa [6]. Kidney cancers are normally diagnosed and followed up using routine blood tests, urinalysis, imaging techniques, and occasionally biopsy in more complex cases [1]. Changes in kidney volume and/or tumor burden have also been identified as features of other CKDs [9], including autosomal dominant polycystic kidney disease (ADPKD) [10] and renal artery atherosclerosis (RAS) [11], which often lead to end-stage kidney disease (ESKD) [12]. However, the detection of CKDs through laboratory tests such as the estimated glomerular filtration rate (eGFR) [13] or albumin-to-creatinine ratio (ACR) [14] is a lengthy and complex process with often unreliable results. Consequently, artificial-intelligence (AI)-based systems that can accurately localize, classify, and quantify kidney tumors using clinical images could greatly improve current diagnosis and patient management.
Kidney scans typically comprise X-rays [15], ultrasonography [16], computed tomography (CT) [17], magnetic resonance imaging (MRI) [18], positron emission tomography (PET) [19], and alternative urography techniques [1]. While CT scans are the most common imaging technique for diagnosing kidney cancer [20], each of these methodologies displays clear strengths and weaknesses, leading to specific technical challenges and specialized applications for each. Traditionally, imaging techniques have been used by clinicians to support biomedical testing or as a method of screening complex patients. In recent years, several studies have attempted to use AI-based methods to diagnose kidney diseases using medical images [19,21,22,23,24,25,26,27,28]. Deep-learning techniques have been employed to detect [21,22], quantify [23,24], and classify [25] different types of kidney pathology including tumors [25], stones [24], glomeruli [23], and hydronephrosis [24], based on analysis of CT [25], Ultrasound (US) [24], MRI [26], and other types of clinical image [19,27,28].
Kidney cancer diagnosis using abdominal CT images is one of the most crucial tasks in clinical management [19,20,21,22,23,24,25,26,27,28,29,30] but this is often performed in isolation of corresponding patient metadata that is likely relevant to tumor subtype and stage [7,31]. Accurately detecting and classifying tumor regions into subclasses, especially malignant versus benign groups, could therefore be improved by integrating clinical metadata into CKD diagnostics. In addition, doctors often face difficulties in selecting operation type (open, robotic, or laparoscopic) [32], optimal procedure (partial or radical nephrectomy) [33], and surgical approach (transperitoneal or retroperitoneal) [34]. Aiding doctors in selecting the most appropriate surgical intervention based on patient demographic data and other preoperative information would represent a major advance for AI-based tools in this domain. We, therefore, sought to develop a smart system able to reliably classify kidney cancers into major subtypes (especially renal cell carcinoma (RCC)), and guide real-time decision-making regarding surgical approach with the ultimate aim of supporting both clinicians and patients to achieve rapid and robust management from the moment of diagnosis.
In this study, we propose a machine-learning-based approach that combines contrast-enhanced CT images with corresponding clinical metadata from individual patients to classify kidney tumors into major subclasses and guide the selection of an optimal surgical procedure. The main contributions of this study are as follows:
  • Computed tomography (CT) images and clinical metadata from the KiTS21 dataset were used to differentiate four major classes of renal cancer: clear cell (ccRCC), chromophobe (chRCC), papillary (pRCC) renal cell carcinoma, and oncocytoma (ONC);
  • Tumor subclass predictions were integrated with clinical metadata to determine the optimal surgical approach in malignant cases (radical versus partial nephrectomy);
  • To the best of our knowledge, this is the first study to determine kidney tumor subclass using a combination of CT images and corresponding clinical features;
  • This pioneering study paves the way for future refinement of tools that can guide surgical interventions in kidney cancer by applying machine-learning algorithms trained on relevant clinical data.
In Section 2, we present a comprehensive review of previous studies dealing with kidney and tumor region of interest (ROI) extraction, segmentation, and classification. We also appraise the current literature on automated classification/surgical decision-making in kidney cancer as well as the associated technical challenges. In Section 3, we describe in detail the various methodologies used by the current study. Section 4 presents the quantitative/qualitative experimental outcomes, as well as supporting visualizations demonstrating the performance of the proposed method. Finally, Section 5 concludes the article and outlines future research directions in this domain.

2. Related Work

According to the American Cancer Society (ACS), kidney tumors rank among the ten most common types of malignancy [35]. In recent years, several studies have sought to diagnose renal tumors using contrast-enhanced abdominal CT images. Early attempts processed CT images using manually defined methods, training classifiers to isolate the kidney, identify tumor ROIs, and diagnose/classify cancer type [36,37,38,39]. Other methods have used deep-learning-based techniques to segment kidneys and tumors from CT images [40,41] or extract features from whole CT images, which are limited by noisy background signals that impact diagnostic performance [42,43,44].
The Kidney and Kidney Tumor Segmentation Challenge (KiTS) was first launched in 2019 [45], continued under the subsequent KiTS21 initiative [46], and has now entered a third iteration under KiTS23 [47]. Over this period, challenge participants have proposed various strategies to identify/segment kidney and tumor regions using contrast-enhanced CT images. The top-five solutions from the KiTS19 grand challenge included manual, semiautomated, and fully automated AI-based techniques to extract and segment an ROI from whole CT volumes (rather than processing the entire image), thereby achieving better diagnostic performance [48]. While the main aim of the KiTS challenges was to segment kidney and tumor volumes from 3D-CT scans [48], many subsequent studies have used KiTS or alternative datasets to generate 2D-CT slices for segmentation and/or deep learning-based detection of renal tumors. Yan et al. [49] proposed 3D-MS-RFCNN, a method of segmenting kidney and tumor volumes using the KiTS19 dataset and evaluated performance against an inhouse dataset collected by their team. Hsiao et al. [50] instead proposed a 2D convolutional neural network (CNN)-based feature pyramid network (FPN) to segment kidneys and tumors using 2D-CT slices extracted from the KiTS dataset. They also developed an adaptive window-selection method to robustly determine appropriate radiodensity ranges for kidney CT slices [51]. Lin et al. [52] performed an extensive study in which the authors segmented kidney, tumor, and cyst volumes using 3D-CT images collected from 441 patients. The Crossbar-Net tool proposed by Yu et al. [53] used horizontal and vertical patches to segment kidney tumors from 2D-CT slices. More recently, investigators have begun to employ advanced AI-based techniques such as graph neural networks (GNNs) [54] and generative adversarial networks (GANs) [55] to process CT images and more accurately segment tumors across diverse datasets. However, very few studies in this domain have proposed end-to-end pipelines for kidney cancer diagnosis, which requires accurate classification and grading of the tumors [56].
Accurate identification of tumor subtype and progression/severity are crucial aspects of kidney cancer diagnosis and key determinants of whether surgery will be required [31]. In addition, the precise grading of malignant tumors critically informs the selection of either partial or radical surgery [57,58]. Alzu’bi et al. [59] previously classified normal and tumorous kidneys using 2D-CT slices, then further distinguished between malignant and benign cases within the cancer cohort. Likewise, Zabihollahy et al. [44] differentiated renal cell carcinoma (RCC) from benign tumors (Oncocytoma (ONC) and Angiomyolipoma (AML)) based on manually annotated regions. Han et al. [60] used a similar approach to classify subtypes of RCC, seeking to distinguish clear cell (ccRCC), from chromophobe (chRCC), and papillary cases (pRCC). Kong et al. conducted an extensive study in which they developed the BKC-Net framework [56], which segments renal tumors from kidney CT images and classifies disease into five distinct subclasses. In this study, the authors were also able to differentiate the most commonly occurring renal cell carcinoma (ccRCC) into benign and malignant classes based on the Fuhrman grade [61]. Uhm et al. [25] also proposed a framework that used 3D-CT slices to sort renal tumors into five distinct classes (ccRCC, chRCC, pRCC, ONC, and AML). However, based on our extensive literature review, none of these studies used clinical metadata either alone or in combination with CT image features to classify tumor subtypes.
Kidney cancers—especially malignant cases—often require surgery to remove the tumorous regions. Depending on case complexity, surgeons might elect to perform open, robotic, or laparoscopic surgery [32], using either a transperitoneal or retroperitoneal approach depending on various patient parameters [34,62]. A key decision is whether to opt for partial or radical nephrectomy [33,63], with several previous studies attempting to determine the optimal procedure for various case types. The outcomes of these studies have been decidedly varied, and the choice of the ‘gold standard’ surgical technique for renal cancer remains unclear. Kunath et al. [33] surveyed multiple studies and concluded that partial nephrectomy was associated with decreased patient-survival time, despite not differing from radical nephrectomy concerning surgery-related mortality, cancer-specific survival, and time to recurrence. Conversely, many urologists and nephrologists [63,64] typically opt for partial nephrectomy except in complex cases where total removal of the kidney is necessary, e.g., where cancer has spread outside the renal hilum boundary (or in elderly patients for whom the remaining kidney is likely to provide a sufficient renal function for remaining lifespan) [64]. A major concern among nephrologists regarding radical surgery in mild cases is the risk of future cancers in the remaining kidney, meaning that the affected patient may eventually require dialysis or an organ transplant. In the current study, we investigated whether artificial intelligence (AI) can be used to assist doctors in determining the suitable surgical procedure for patients with renal cell carcinoma (malignant tumors) using a combination of demographic and preoperative clinical data.

3. Methodology

The overall workflow of this study is presented in Figure 1. In brief, we discuss the KiTS21 dataset, preprocessing steps, scope reduction (removal of non-kidney 2D-CT slices), extraction of regions of interest/ROIs (i.e., kidneys, tumors, and cysts) from CT slices, and tumor classification based on the extracted ROIs (with or without metadata integration). We also discuss the process of determining the optimal surgical procedure for malignant RCC cases via binary classification based on clinical data. The various classical and deep machine-learning tools used in this study have been explained in detail in the respective sections. Finally, we explain the quantitative metrics and qualitative evaluation approaches applied to each experiment.

3.1. Dataset Description

In this study, we used the publicly available “Kidney and Kidney Tumor Segmentation Challenge 2021” (KiTS21) dataset [65]. The contrast-enhanced computed tomography (CT) scan dataset was originally published in 2019 as part of the KiTS19 grand challenge [66]. The KiTS21 dataset includes subjects from M Health Fairview [67] and Cleveland Clinic [68] medical centers who underwent either partial or radical nephrectomy between 2010 and 2018. Each of the 544 original cases was reviewed retrospectively to include only those patients who had also undergone a contrast-enhanced CT scan of the entire kidney(s) and corresponding renal tumors. The final compiled dataset contained 3D-CT slices from n = 300 subjects, each with ground truth masks annotated by a group of experts and trainees [65]. KiTS21 is an extended version of the earlier KiTS19 initiative [65], which provided ground-truth annotations or masks for benign cysts alongside masks for kidneys and tumors. KiTS21 also includes ground-truth masks for 90 test subjects from the original KiTS19 challenge. Baseline patient and tumor characteristics of the KiTS21 dataset are provided in Table 1.

3.2. Preprocessing of Computed Tomography (CT) Images

The KiTS21 dataset is provided in Neuroimaging Informatics Technology Initiative (NIFTI) file format [69] and compressed using the Gzip method from the GNU project [70], such that multistage, full-body 3D-CT images are available for each subject. Original CT images from n = 300 patients are stored in a repository curated by the challenge organizer [71], while annotated ground-truth segmentation masks for the kidney, tumor, and cyst regions are accessed via GitHub [72]. The ground-truth segmentation masks are also provided in a 3D format to match the CT images slice by slice. Annotations use four levels of pixels (0 to 3) to denote kidney (1), tumor (2), cyst (3), and background (0). The KiTS challenge organizers followed three different schemes to combine pixel-wise annotations from multiple experts: in the ‘AND’ scheme only pixels shared by annotations from all labelers were considered for the final masks; in the ‘OR’ scheme pixels present in annotations from any labeler were considered for ground-truth annotation; in the ‘MAJ’ scheme pixels marked by the majority of annotators were used to generate the final masks. For this study, we elected to use masks generated by the MAJ scheme. The 3D-CT images were unzipped and extracted from the NIFTI files into 3D arrays. A total of 64,603 2D slices were then extracted from n = 300 3D-CT images. Next, the Hounsfield Unit (HU) [51] range for each CT slice was determined. Restricting the CT images to a defined HU range assists radiologists (and deep-learning algorithms) by using the variable radiodensity of different organs to focus on a specific target (in this case kidney). Previous studies have also attempted kidney segmentation using fixed HU ranges [41,48,53,73] or adaptive HU range determination [50]. In their official-challenge article, Heller et al. [48] used a fixed HU range of (−200, 500) which was also employed in the current study. To project pixel values within a given HU range, all values exceeding the upper limit (500) were clipped to this cap, while all values below the lower limit (−200) were clipped to this minimum value. In Supplementary Figure S1, a sample CT slice featuring the total HU range and examples of HU-limited slices are shown.

3.3. Scope Reduction through Kidney Instance Classification

As proposed by the challenge organizers [48], the ability of deep-learning models to estimate kidney and tumor regions in CT images depends on five main factors: slice thickness, tumor focality, the field of view (FoV), tumor size, and cancer subtype. In this dataset, FoV varies a lot between cases and scanning sites. Some scans contain full-body captures extending from head to toe, thereby introducing a large amount of ‘non-kidney’ information into the dataset. For slices of this type, the ground-truth masks are completely blank, and it becomes hard for the model to determine the kidney and associated tumor or cyst regions. In this study, we dealt with large FoVs using a deep learning-based ‘Scope Reduction’ technique, for which we divided the slices into kidney and non-kidney classes based on ground-truth masks. A deep-learning-based binary classifier was trained to classify the kidney and non-kidney slices (if there was a single nonzero pixel present for a given slice, that was considered a kidney). In a previous study, Cruz et al. [41] performed scope reduction using AlexNet [74], which is one of the earliest CNN-based 2D classification networks. In the current study, we instead tested various state-of-the-art architectures, including ResNet152 [75], DenseNet201 [76], InceptionV3 [77], and MobileNetV2 [78], with pretrained ImageNet weights for scope reduction. This revealed that a modified DenseNet201-based approach with auxiliary losses (DenseAUXNet201, shown in Figure 2) achieved superior performance metrics to the other networks tested.

DenseAUXNet201 Architecture

The original DenseNet architecture was proposed in 2018 by Huang et al. [76] as a set of four major variants: DenseNet121, DenseNet169, DenseNet201, and DenseNet264. DenseNet121 and DenseNet201 in particular have been applied to many 2D classification tasks [79,80,81]. DenseNet201 contains 6, 12, 48, and 32 convolutional blocks that are each followed by batch normalization and the ‘ReLU’ activation layer. Transition blocks placed after each denseblock down sample the feature maps to lower dimensions. InceptionV3 is another state-of-the-art model that includes variants containing auxiliary losses that enable monitoring and optimizing the flow of features in intermediate layers. Among pretrained state-of-the-art models used in our study, DenseNet201 performed consistently well across all experiments. To further improve performance, we implemented auxiliary losses with independent-classifier blocks after each of the three intermediate dense blocks (Figure 2) and monitored their outputs for optimization. We also took features from each of the intermediate layers, concatenated these with output from the last layer, and passed the concatenated features to a set of densely connected multilayer-perceptron (MLP) blocks for final classification (all four losses were optimized during training). The proposed DenseAUXNet201 framework improved performance across all evaluation criteria, as discussed in the results section. We also employed ‘LogSoftmax’ [82] as the auxiliary and final activation function in DenseAUXNet201 (formulated as shown in Equation (1)).
L o g S o f t m a x ( x i ) = l o g ( e x i i e x i )

3.4. Region of Interest (ROI) Extraction from 2D-CT Slices

Extracted 2D-CT slices containing kidney footprints can be used directly for tumor subclass classification based on labels, but it is important to note that other organs present in the image heavily impact the decision-making process. To mitigate this, a region of interest (ROI) around the kidney can be extracted from whole CT slices so that multiclass classifiers can focus only on tumorous regions in the next stage. Extracting an ROI for kidney detection or segmentation from CT images has been a common practice across a range of published 2D/3D approaches. For example, three of the four top performers ranked by the KiTS19 challenge extracted ROIs from 3D-CT images, using methods extending from coarse segmentation to extracting volume of interest (VOI) (the 3D equivalent of ROI) [48]. In recent years, different generations of the ‘You Only Look Once’ (YOLO) framework has been used for object detection in many complex scenarios. State-of-the-art YOLO versions (YOLOv5 [83], YOLOv7 [84], etc.) have demonstrated excellent performance in object detection, but are only rarely applied to kidney and/or tumor ROI extraction from CT images. Therefore, in the current study, we tested the ability of YOLOv5 and YOLOv7 frameworks to extract kidney and tumor ROIs from whole 2D-CT slices.

3.4.1. Bounding Box Label Generation from Segmentation Masks

The original KiTS21 annotations feature ground-truth masks for segmentation but do not contain bounding-box labels as required for training YOLO frameworks. We, therefore, used the segmentation masks to generate three classes of rectangular bounding boxes for the YOLO framework, corresponding to the three-pixel levels used to denote kidney, tumor, and cyst regions. First, contours were generated and unified around the masks, then bounding boxes were positioned to encompass the four extreme corners of these contours. During the generation of ground-truth bounding boxes, we observed that tumor regions appeared very distorted or blurry if these were either very irregular in shape or small in size (resulting in upsampling/interpolation to fixed dimensions). To avoid this impacting model performance, the bounding boxes were adjusted to form squares based on the highest dimension of length or width. Next, we removed any slices in which the bounding box area for a kidney region was less than 48 pixels height/width or <1% of the total slice area (512 × 512 = 262,144). The overall bounding box generation process is summarized in Supplementary Figure S2. The YOLO framework accepts four parameters alongside the class label (normalized height/width and x/y coordinates of the bounding box center point), whereas contours provide height, width, and x/y coordinates, and define the origin (0, 0) as the left-uppermost corner of the bounding box. We, therefore, generated labels for YOLO using Equations (2)–(5);
N o r m a l i z e d   x c o o r d i n a t e   ( c e n t e r ) = ( x c o o r d i n a t e + b o x   w i d t h 2 ) i m a g e   w i d t h
N o r m a l i z e d   y c o o r d i n a t e   ( c e n t e r ) = ( y c o o r d i n a t e + b o x   h e i g h t 2 ) i m a g e   h e i g h t
N o r m a l i z e d   b o x   w i d t h = b o x   w i d t h i m a g e   w i d t h
N o r m a l i z e d   b o x   h e i g h t = b o x   h e i g h t i m a g e   h e i g h t
For the YOLO framework, labels are saved in text files linked to each image (with matching file names). Since we used YOLO models pretrained on the COCO dataset [85], images were resized to the same (640 × 640) format during training.

3.5. Kidney Tumor Subtype Classification from Extracted CT ROIs

The KiTS metadata file available via GitHub [72] contains demographic and clinical information for each of the n = 300 subjects. These cases include 13 different types of kidney tumors (except for n = 1 individual classed as ‘other’). The predominant classes of renal tumors in this dataset are clear cell (ccRCC—204 cases), papillary (pRCC—28 cases), chromophobe (chRCC—27 cases), oncocytoma (ONC—16 cases), and angiomyolipoma (AML—5 cases). Since the total number of image samples from n = 5 AML cases was inadequate for deep-learning-based image classification, these were removed from the dataset. To further reduce class imbalance among the remaining groups, we augmented the training images in each fold using random rotation (−90° to 90°) and random vertical/horizontal flip (Supplementary Table S1). The final task was therefore a four-group classification problem requiring differentiation between ccRCC, pRCC, chRCC, and ONC. A total of n = 3 ccRCC cases were excluded due to extremely small or highly distorted tumor ROIs (as described in Section 3.4.1) leaving a total of n = 272 cases for use in this study. Figure 3 shows one representative tumor sample per class for visualization (ROI extracted and resized).
The prepared dataset was trained and evaluated using a 2D-CNN-based multiclass classifier. The models used for this task were the same as those employed for scope reduction (ResNet152 [75], DenseNet201 [76], InceptionV3 [77], MobileNetV2 [78], and DenseAUXNet201 with pretrained ImageNet weights). For this application, the last linear layer before final activation included four output neurons to match the number of classes. For subject-wise evaluation, ground truth and predicted image labels were transformed into subject labels through majority voting.

3.6. Tumor Subtype and Surgery Procedure Classification from Clinical or Combined Data

Clinical metadata in the KiTS dataset is provided in JavaScript object notation (JSON) format which was converted into a tabular format for preprocessing. The clinical features can be divided broadly into preoperative, intraoperative, and postoperative categories (with preoperative features also containing demographic information). For this task, the primary aim is to boost the tumor subtype classification performance of machine-learning algorithms by combining clinical data with features extracted from the CT images. Subsequently, these clinical features were also used to classify whether radical or partial nephrectomy was selected when operating on malignant tumors (RCCs).

3.6.1. Clinical Data Pre-Processing

After extracting clinical metadata, we used only preoperative elements to inform our diagnostic framework, since only presurgical features would be available to guide decision-making in real-life scenarios. In addition, intraoperative characteristics and postoperative outcomes arise from surgical decisions taken by a diverse range of clinicians, rather than being objective features of individual patients and kidney tumors. Key preoperative clinical parameters included tumor size (radiographic and pathologic), demographic information (gender, age, and body mass index (BMI), as well as alcohol/smoking/tobacco history, and the presence of 19 common comorbidities (e.g., myocardial infarction, congestive heart failure, etc.). For predicting surgical procedure, we also included tumor subtype and Fuhrman grade [57] since operation choice critically depends on case complexity (including cancer subclass and severity). Within the tabulated dataset, age and BMI were range-normalized between 0 and 1 to remove bias. The only other numerical features present were radiographic and pathologic measures of tumor volume. Categorical features were transformed into numerical values by replacement. KiTS21 metadata does not provide Fuhrman tumor grade but does specify tumor, node, and metastasis (TNM) cancer stages [86], as proposed by the American Cancer Society (ACS) [35]. We, therefore, transformed these values to achieve the four-stage Fuhrman nuclear grade for kidney cancer severity [57] (see Table 2).
In the TNM grading system, T refers to the size and extent of the primary tumor, N indicates the number of malignant lymph nodes nearby, and M denotes whether cancer has metastasized (i.e., severe cases of tumor spreading to remote organs such as lungs, liver, or brain) [87]. Based on Table 2, if the tumor size has reached T-stage 4, cancers with M = 0 and any N value are classified as Fuhrman stage IV. Alternatively, when there are signs of metastasis (M = 1), cancer will always be considered stage IV regardless of T and N values. Stage III cancers depend on both T and N scores unless tumorous tissues are present in one or more nearby lymph nodes (i.e., default stage III disease). Based on these unification criteria, n = 259 patients with malignant tumors included n = 171 stage I cases, n = 17 stage II, n = 56 stage III, and n = 15 stage IV. The n = 16 ONC cases were marked as benign and n = 3 ccRCC cases with small/distorted tumor ROIs were removed from the dataset (as described in Section 3.4.1). Of the remaining cases, n = 163 underwent partial nephrectomy while n = 93 patients were subjected to radical nephrectomy. A total of 29 raw clinical features were used for tumor subtyping and surgical procedure classification.

3.6.2. Feature Engineering

Image features were extracted from the penultimate layer of the best-performing 2D-CNN-based classifier (DenseAUXNet201). The final classifier was removed for feature extraction, and a total of 4480 features were obtained from each image (using the latent layer of DenseAUXNet201). Features from all images for a test subject were concatenated and reduced through principal component analysis (PCA) [88]. For the clinical data, highly correlated features were removed based on a threshold of 0.85 cross-correlations. Finally, the most influential features were ranked using three distinct techniques [89]; XGBoost [90], random forest [91], and extra trees [92] (in addition to combining these with existing clinical features through horizontal concatenation). For surgical procedure selection, we employed the same feature engineering approach to determine the most relevant features before classification.

3.6.3. Classical Machine-Learning Algorithms

For this study, we used different types of ML-based classifiers for both tumor sub-typing (combined approach) and surgical procedure assignment: logistic regression, k-nearest neighbors (kNN), support vector machine (SVM), random forest, extra trees, linear discriminant analysis (LDA), gradient boost (GradBoost) [93], adaptive boost (AdaBoost), XGBoost [90], light GBM (LGBM) [94], ElasticNet, ridge, and multilayer perceptron (MLP).

3.7. Quantitative Evaluation Metrics

We used two broad types of machine-learning applications for classification (deep or standard ML) and object detection (deep ML). Distinct evaluation criteria were applied for each approach.

3.7.1. Classification

The performance of the classifiers (both standard and deep ML) was evaluated based on accuracy, precision, recall/sensitivity, specificity, and F1-score. We initially extracted the number of true positive (TP), true negative (TN), false positive (FP), and false negative (FN) labels from the confusion matrix, then these metrics were formulated for subject-wise evaluation based on Equations (6)–(10);
A c c u r a c y = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F N
R e c a l l   o r   S e n s i t i v i t y = T P T P + F P
S p e i c i f i c i t y = T N T N + F P
F 1 s c o r e 2 P r e c i s i o n R e c a l l P r e c i s i o n + R e c a l l = 2 T P 2 T P + F P + F N
All metrics except accuracy were weighted to deal with class imbalance (a common issue with this type of task). For accuracy, we reported the overall macro value calculated from the confusion matrix for the entire dataset. We also show the confusion matrix for the best-performing model at individual stages of each task.

3.7.2. Object Detection

The metrics most commonly used to assess the accuracy of bounding-box generation by YOLO frameworks are precision, recall, and mean average precision (mAP), at various intersections over union (IoU) levels. Precision and recall metrics have been defined in Equations (7) and (8), respectively (computed based on ground truth and predicted bounding box coordinates). Average precision (AP) can be defined as a technique to summarize the whole precision-recall curve into a single value, as formulated in Equation (11).
A P = k = 0 k = n 1 [ R e c a l l s ( k ) R e c a l l s ( k + 1 ) ] P r e c i s i o n ( k )
AP is the weighted sum of precision at each threshold, where weight is the increase in recall. In this case, ‘n’ is the number of thresholds, hence Recall(n) = 0 and Precision(n) = 1. The mAP value can be derived from AP by calculating the average AP from all the classes. If there are ‘c’ number of classes, mAP can be defined as:
m A P = i = 1 i = c A P k c
The current study includes three distinct classes, namely kidney, tumor, and cyst regions. In this case, the sum of AP in all classes will be divided by three to obtain mAP, which is a key metric of object detection performance (i.e., unbiased generation of accurate bounding boxes). Both large/obvious objects and small/challenging cases are captured by mAP calculation. For YOLO, the common practice is to measure mAP at IoU threshold 0.5 and IoU range 0.5–0.95 [95], which is a robust performance metric for object detection frameworks when considering moderately to highly challenging cases.

3.8. Qualitative Evaluation

To further assess YOLO performance, we plotted ground truth and estimated bounding boxes side-by-side for various types of sample cases. We also plotted the performance curves (F1, P, PR, and R) [96] of the best-performing YOLO model. To illustrate the performance of the deep CNN-based image classifiers, we used class activation mapping (CAM) [97] to generate a weighted activation map for each image based on a trained classifier. Current state-of-the-art CAM techniques include GradCAM [98], GradCAM++ [99], Smooth GradCAM++ [100], and ScoreCAM [101]. We employed ScoreCAM which uses the trained model to generate weighted heatmaps of input test images in each class and visualize the classifier learning process (unlike other versions such as GradCAM, which use generic algorithms for this task). This analysis can improve understanding and validation of model performance when considered alongside quantitative metrics.

4. Experimental Results

Here we present the quantitative and qualitative outcomes of tumor subclass classification by deep learning from kidney CT images (with or without integration of clinical metadata). Assignment of surgical procedures for malignant tumors based on clinical parameters is also assessed.

4.1. Kidney Tumor Classification

As depicted in Figure 1, our proposed method for renal tumor classification from CT images consists of three main steps: scope reduction, kidney/tumor ROI extraction, and tumor subtype classification using deep ML algorithms. In the combined approach, we concatenate both image and clinical features from the same subjects to classify kidney tumors using classical ML techniques. To ensure robust evaluation, we stratified the dataset of n = 300 subjects into five folds in a subject-wise fashion (see Supplementary Table S1). In each fold, the test set contains 60 subjects (20% of the total dataset). The remaining n = 240 cases are then divided into training and validation sets (80:20 ratio). The ranges specified in Supplementary Table S1 were applied in all subsequent experiments.

4.1.1. Scope Reduction through Binary Classification

Scope reduction was performed to remove non-kidney CT slices from the dataset, thus allowing the object detection networks to focus on kidney, tumor, and cyst ROI extraction without any impact of irrelevant slices. We extracted 64,603 2D-CT slices from a total set of 300 CT images. Among these, approximately 41,675 (64.5%) were non-kidney slices, which would represent a major confounder for object detection if not removed. Table 3 shows the combined fivefold performance of the deep-learning-based binary classifiers when processing subject-independent test sets.
The proposed DenseAUXNet201 framework displayed the best performance across all evaluation metrics and achieved the lowest number of missed cases. While only a modest increase in performance was noted, this nonetheless represented a significant improvement in the detection and removal of non-kidney slices (and vice-versa).
As plotted in Figure 4, we observed that 909 slices were misclassified as ‘non-kidney’ and 368 slices were misclassified as ‘kidney’, amounting to just 1277 missed cases among 64,603 total slices. Based on our review of experimental outcomes, most missed cases were extreme, i.e., kidney regions were very small, or the model detected other organs as trace kidney signals.
In Figure 5, the learning capabilities of the trained DenseAUXNet201 model are illustrated using ScoreCAM-generated weighted heatmaps from the final layer (before the last classifier block) as applied to representative regions or features. For slices with one or two kidneys, the model is focusing on the relevant regions, whereas for slices lacking kidneys, the model is learning to classify based on alternative tissues and features (e.g., lungs).

4.1.2. Region of Interest (ROI) Extraction from 2D-CT Slices using YOLO

We experimented with six models from two different generations of YOLO to extract kidney, tumor, and cyst ROIs using the scope-reduced 2D-CT dataset (YOLOv5 [83] variants YOLOv5s, YOLOv5m, YOLOv5l and YOLOv5x, plus YOLOv7 [84] variants YOLOv7 and YOLOv7x). Here, ‘s’, ‘m’, ‘l’, and ‘x’ denote small, medium, large, and extra-large model variants, respectively. We tested YOLO networks on the fivefold split dataset as described in Supplementary Table S1 and the results are provided in Table 4. While older YOLOv5 models achieved better precision than YOLOv7 variants, the more recent models displayed superior mAP at the IoU threshold of 0.5 and range of 0.5–0.95 [95]. Lower precision leads to more false positives from YOLOv7 models compared with YOLOv5 variants, but the higher mAP values indicate more efficient detection of very small and/or challenging regions (crucial in the case of tumor identification). We, therefore, selected basic YOLOv7 as the best-performing model for our experiments.
While YOLO worked indirectly as a classifier for the identification of tumors and cysts, the cyst detection performance was very low due to the tiny size of these structures (and the limited number of examples in the training/test datasets). This was not the case for the processing of tumor ROIs i.e., the primary goal of this study. Example outputs from the YOLOv7 network alongside their corresponding ground-truth values are shown in Figure 6. The performance curves of the YOLOv7 model (e.g., F1, precision-confidence, precision-recall, and recall-confidence) are plotted in Supplementary Figure S3. The plots display curves for each of the three classes as well as the combined outcome.

4.1.3. Kidney Cancer Subtype Classification from Tumor ROIs and Clinical Metadata

We next compared the performance of kidney cancer classification using extracted tumor ROIs (image-based) and the combined approach (image + clinical features). After curation, the dataset contained n = 272 cases for classification. Data were split subject-wise in the same way for all experiments (Supplementary Table S1) so that features from both pipelines could be directly compared. We resized the ROIs and augmented training sets in each fold to minimize class imbalance (due to the relative abundance of real-world ccRCC samples). Here, the model remains biased towards the ccRCC if the class imbalance is not mitigated through proper augmentation. The image-wise outcomes were converted into subject-wise results through majority voting on the outcomes for each patient.
Figure 7a,c show that our proposed DenseAUXNet201 model achieved better performance than the current state-of-the-art classification models across all evaluation criteria. The custom auxiliary losses and feature concatenation from intermediate layers boosted its performance significantly. Next, we proceeded to combine the clinical and engineered image features through concatenation. A total of 4480 features (based on the latent dimension of DenseAUXNet201) were extracted from the tumor ROIs, which were efficiently compressed through PCA to only 20 features and combined with the 29 raw clinical features. The pipeline automatically removed ‘pathologic size’ due to high concordance with the ‘radiographic size’ feature. The remaining 48 features were ranked using the three feature-ranking algorithms described in Section 3.6.2, and only the top 20 features were found to be adequate for the optimum performance, as shown in Figure 8, based on the random forest-based feature selection technique. The outcomes from the XGBoost and the extra trees feature ranking techniques are provided in Supplementary Figure S4. Based on the horizontal bar chart in Figure 8, the tumor class from the images was the most prominent feature affecting the classifier, followed by the tumor malignancy marker stored in the clinical data. Tumor size, demographic features such as age, BMI, gender, and habitual features such as smoking or drinking history were also marked as important factors behind tumor classes by the feature-ranking algorithm. Next, a set of classical ML networks were used to classify cases based on these 20 top-ranked features, the results of which are provided in Figure 7b,d.
These analyses revealed that random forest, XGBoost, and CatBoost algorithms performed far better than the other methods. We selected XGBoost due to its higher F1-score (equivalent to the Sørensen dice coefficient used for segmentation or object detection tasks [102]). From the confusion matrices plotted in Figure 7a,b, we observed that combining image features with clinical ones significantly improved in separating the ONC class due to its nonmalignant or benign nature. In contrast, pRCC cases were classified more effectively after feature combination, ccRCC classification performance remained unchanged overall while chRCC identification notably worsened (confused with ccRCC). This occurred primarily due to the class imbalance in the dataset, which could not be fully mitigated through augmentation.

4.2. Surgical Procedure Determination from Clinical Data for Malignant Tumors

As discussed earlier, the raw clinical features included 29 parameters with the potential to influence the classification of kidney tumor surgical procedures. First, we removed highly correlated features to avoid redundancy and reduce overfitting. As before, the ‘pathologic size’ feature was detected as being highly correlated with ‘radiographic size’ and was therefore removed. The remaining 28 features were then ranked based on their relevance to surgical procedure selection. Among the three feature ranking techniques experimented with in this study (XGBoost, random forest, and extra trees), we opted to use random forest (Figure 9) due to it providing more relevance to practical scenarios.
Feature ranking by random forest (Figure 9) and extra trees (Supplementary Figure S5b) generated similar outputs for both cases (tumor and surgery classification) due to both being tree-based techniques. The most important features ranked by random forest were radiographic size/tumor volume and Fuhrman stage, followed by patient BMI, age, smoking habits, history of CKDs, etc. When using XGBoost (Supplementary Figure S5a), importance was also assigned to the presence of a solid metastatic tumor and additional features such as chronic obstructive pulmonary disease (COPD), which are often linked with kidney pathology [103,104]. Practically, surgeons take a more radical approach for complicated end-stage cancers and large tumors [33,63]. Clinicians also prefer a more radical approach for elderly patients with high BMIs [64]. Long-term smoking habits have been linked with kidney cancers in recent studies [105,106], especially renal cell carcinoma (RCC) [106]. On the other hand, alcohol consumption was found to be a vital cause of ESKDs, especially RCCs [107,108]. So, the top-ranked features from the feature ranking algorithm match closely with relevant literature, which validates our classifier’s performance.
We finally chose the top 20 features as ranked by random forest for the surgery classification task. We removed any feature with <0.1% relevance to the target and then trained the classifiers over 1000 epochs or until convergence (then repeated this process for all five folds). Outcomes were evaluated in terms of overall accuracy and weighted precision, recall/sensitivity, specificity, and F1-score. The combined (averaged) fivefold classification results are provided in Figure 10. Among the 14 classifiers used in this study, logistic regression, linear discriminant analysis (LDA), support vector machines (SVM), ridge, and multilayer perceptron (MLP) performed far better. We selected traditional logistic regression for our study due to this being a simple classifier with high performance for the current application. In Figure 10b–e, we show the overall confusion matrix for surgical procedure classification across the entire dataset, as well as individually for ccRCC, chRCC, and pRCC cases. We observed that the tumor subtype did not bias the performance of the classifiers in selecting the appropriate surgical procedures. Just 7 partial cases were misclassified as radical, whereas 17 radical cases were misclassified as partial. Further refinement of this AI tool can therefore aid clinicians in making prompt surgical decisions based on preoperative clinical data.
According to Cancer Research UK [109], surgeons take objective decisions on the optimal surgical approach for kidney cancers primarily based on parameters such as patient health and fitness, size and location of the tumor, cancer stage, presence of lymph nodes, and metastasis. While these parameters from the KiTS21 metadata have been included as features in our classifier, there can be other uncommon but important factors influencing the surgical procedure. For example, the American Cancer Society (ACS) discusses surgical approaches for recurrent cancers [110]. These factors are out of the scope of this study since the KiTS21 metadata does not contain this information at all or is not adequate for the classifier to make robust decisions. Solving such problems requires collecting a custom, objective-driven dataset from a good number of kidney cancer surgery patients. Nevertheless, with more dedication, our proposed approach has the potential to be an important AI companion to cancer surgeons.

5. Conclusions

AI-based tumor detection and subtype classification can help guide medical and surgical decision-making for kidney cancer patients. Locating kidneys, tumors, and cysts from CT slices using deep-learning approaches is itself a useful application for aiding or training clinicians and radiologists during a cancer diagnosis. Provided acceptable levels of performance, AI-based smart systems can inform the choice of therapy or surgical procedure in kidney cancer, thereby preserving renal function in patients who require only partial nephrectomy or saving the lives of patients with end-stage cancers requiring radical intervention. The primary limitation of this study was the generality of the KiTS dataset. The proposed AI tool could be significantly improved by collecting a large, objective-driven dataset designed to refine this approach. The collected custom dataset should be more balanced in terms of tumor subtypes and cancer stages, ideally containing thousands of diverse cases to make the AI tool more robust for a large population group. More clinical biomarkers, patient medical history, immediate relevant parameters to the surgical table, and other important habits and comorbidities should also be recorded to render this tool more reliable and further improve performance.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/2072-6694/15/12/3189/s1; Figure S1. 2D Kidney CT slices of varying HU. (a) Total HU range, original slice, (b) range between [−200, 500], as used in the current study, and (c) [−30, 200]; Figure S2. Bounding box creation process for training YOLO frameworks; Figure S3. YOLOv7 model performance curves plotted for each object type (kidney, tumor, and cyst), alongside combined performance; Figure S4. XGBoost (a) and Extra Trees (b) classifier-based feature selection outcomes for the kidney tumor classification. Ranked features from the Extra trees and Random Forest algorithms were very similar; Figure S5. XGBoost (a) and Extra Trees (b) classifier-based feature selection outcomes for the surgery procedure classification. Ranked features from the Extra trees and Random Forest algorithms were very similar; Table S1. Subject-wise split five-fold information regarding the subject range.

Author Contributions

Conceptualization S.M. and M.E.H.C.; Data curation S.M. and J.P.; Formal analysis S.M. and A.M.; Funding acquisition M.E.H.C. and T.O.A.; Investigation S.M. and J.P.; Methodology S.M., M.E.H.C., and T.O.A.; Project administration M.E.H.C. and T.O.A.; Resources M.E.H.C. and T.O.A.; Supervision M.E.H.C.; Validation M.E.H.C., A.M. and T.O.A. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by student grant #QUST-1-CENG-2023-795 from Qatar University and partly by Qatar National Library and Sidra Medicine. The statements made herein are solely the responsibility of the authors.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study can be made available upon a reasonable request to the corresponding author.

Conflicts of Interest

The authors have no conflicts of interest to declare.

References

  1. Mayo Clinic. Chronic Kidney Disease. 3 September 2021. Available online: https://www.mayoclinic.org/diseases-conditions/chronic-kidney-disease/symptoms-causes/syc-20354521 (accessed on 25 March 2023).
  2. Kovesdy, C.P. Epidemiology of chronic kidney disease: An update 2022. Kidney Int. Suppl. 2022, 12, 7–11. [Google Scholar] [CrossRef] [PubMed]
  3. WCRF International. Kidney Cancer Statistics: World Cancer Research Fund International. 14 April 2022. Available online: https://www.wcrf.org/cancer-trends/kidney-cancer-statistics (accessed on 26 March 2023).
  4. Ito, C.; Nagata, D. Profound connection between chronic kidney disease and both colorectal cancer and renal cell carcinoma. J. Kidney 2016, 2, 1000118. [Google Scholar]
  5. Hu, M.; Wang, Q.; Liu, B.; Ma, Q.; Zhang, T.; Huang, T.; Lv, Z.; Wang, R. Chronic kidney disease and cancer: Inter-relationships and Mechanisms. Front. Cell Dev. Biol. 2022, 10, 1073. [Google Scholar] [CrossRef] [PubMed]
  6. Webster, A.C.; Nagler, E.V.; Morton, R.L.; Masson, P. Chronic kidney disease. Lancet 2017, 389, 1238–1252. [Google Scholar] [CrossRef] [PubMed]
  7. Saly, D.L.; Eswarappa, M.S.; Street, S.E.; Deshpande, P. Renal cell cancer and chronic kidney disease. Adv. Chronic Kidney Dis. 2021, 28, 460–468. [Google Scholar] [CrossRef]
  8. Lees, J.S.; Elyan, B.M.; Herrmann, S.M.; Lang, N.N.; Jones, R.J.; Mark, P.B. The “other” big complication: How chronic kidney disease impacts on cancer risks and outcomes. Nephrol. Dial. Transplant. 2022, 38, 1071–1079. [Google Scholar] [CrossRef]
  9. Hussain, M.A.; Hamarneh, G.; Garbi, R. Cascaded Regression Neural Nets for Kidney Localization and Segmentation-free Volume Estimation. IEEE Trans. Med. Imaging 2021, 40, 1555–1567. [Google Scholar] [CrossRef]
  10. NHS Choices. Available online: https://www.nhs.uk/conditions/autosomal-dominant-polycystic-kidney-disease-adpkd/ (accessed on 25 March 2023).
  11. Cleveland Clinic. Renal Artery Stenosis: Symptoms, Causes, and Treatment. Available online: https://my.clevelandclinic.org/health/diseases/17422-renal-artery-disease (accessed on 25 March 2023).
  12. Pennmedicine.org. End-Stage Kidney Disease. Available online: https://www.pennmedicine.org/for-patients-and-visitors/patient-information/conditions-treated-a-to-z/end-stage-kidney-disease (accessed on 25 March 2023).
  13. American Kidney Fund. Blood Test: EGFR (Estimated Glomerular Filtration Rate). 24 January 2023. Available online: https://www.kidneyfund.org/all-about-kidneys/tests/blood-test-egfr (accessed on 25 March 2023).
  14. National Kidney Foundation. ACR. 11 May 2017. Available online: https://www.kidney.org/kidneydisease/siemens_hcp_acr (accessed on 25 March 2023).
  15. National Institute of Biomedical Imaging and Bioengineering. X-rays. Available online: https://www.nibib.nih.gov/science-education/science-topics/x-rays (accessed on 25 March 2023).
  16. Cleveland Clinic. Kidney Ultrasound: Procedure Information, Preparation, and Results. Available online: https://my.clevelandclinic.org/health/diagnostics/15764-kidney-ultrasound (accessed on 25 March 2023).
  17. Johns Hopkins Medicine. Computed Tomography (CT or CAT) Scan of the Kidney. 8 August 2021. Available online: https://www.hopkinsmedicine.org/health/treatment-tests-and-therapies/ct-scan-of-the-kidney (accessed on 25 March 2023).
  18. Tomboc, K.; Ezra. MRI of the Kidney: Your Guide for Preventative Screening. 29 July 2022. Available online: https://ezra.com/kidney-mri/ (accessed on 26 March 2023).
  19. Krishnan, N.; Perazella, M.A. The role of PET scanning in the evaluation of patients with kidney disease. Adv. Chronic Kidney Dis. 2017, 24, 154–161. [Google Scholar] [CrossRef]
  20. Tests for Kidney Cancer. Kidney Cancer Diagnosis. Available online: https://www.cancer.org/cancer/kidney-cancer/detection-diagnosis-staging/how-diagnosed.html (accessed on 25 March 2023).
  21. Shubham, S.; Jain, N.; Gupta, V.; Mohan, S.; Ariffin, M.M.; Ahmadian, A. Identify glomeruli in human kidney tissue images using a deep learning approach. Soft Comput. 2021, 27, 2705–2716. [Google Scholar] [CrossRef]
  22. Elton, D.C.; Turkbey, E.B.; Pickhardt, P.J.; Summers, R.M. A deep learning system for Automated Kidney Stone Detection and volumetric segmentation on noncontrast CT scans. Med. Phys. 2022, 49, 2545–2554. [Google Scholar] [CrossRef]
  23. Abdelrahman, A.; Viriri, S. Kidney tumor semantic segmentation using Deep Learning: A Survey of state-of-the-art. J. Imaging 2022, 8, 55. [Google Scholar] [CrossRef]
  24. Smail, L.C.; Dhindsa, K.; Braga, L.H.; Becker, S.; Sonnadara, R.R. Using deep learning algorithms to grade hydronephrosis severity: Toward a clinical adjunct. Front. Pediatr. 2020, 8, 1. [Google Scholar] [CrossRef]
  25. Uhm, K.-H.; Jung, S.-W.; Choi, M.H.; Shin, H.-K.; Yoo, J.-I.; Oh, S.W.; Kim, J.Y.; Kim, H.G.; Lee, Y.J.; Youn, S.Y.; et al. Deep learning for end-to-end kidney cancer diagnosis on multi-phase abdominal computed tomography. NPJ Precis. Oncol. 2021, 5, 54. [Google Scholar] [CrossRef]
  26. Mo, X.; Chen, W.; Chen, S.; Chen, Z.; Guo, Y.; Chen, Y.; Wu, X.; Zhang, L.; Chen, Q.; Jin, Z.; et al. MRI texture-based machine learning models for the evaluation of renal function on different segmentations: A proof-of-concept study. Insights Imaging 2023, 14, 28. [Google Scholar] [CrossRef] [PubMed]
  27. Alnazer, I.; Bourdon, P.; Urruty, T.; Falou, O.; Khalil, M.; Shahin, A.; Fernandez-Maloigne, C. Recent advances in medical image processing for the evaluation of chronic kidney disease. Med. Image Anal. 2021, 69, 101960. [Google Scholar] [CrossRef]
  28. Yao, L.; Zhang, H.; Zhang, M.; Chen, X.; Zhang, J.; Huang, J.; Zhang, L. Application of artificial intelligence in renal disease. Clin. eHealth 2021, 4, 54–61. [Google Scholar] [CrossRef]
  29. Rini, B.I.; Campbell, S.C.; Escudier, B. Renal cell carcinoma. Lancet 2009, 373, 1119–1132. [Google Scholar] [CrossRef] [PubMed]
  30. Hsieh, J.J.; Purdue, M.P.; Signoretti, S.; Swanton, C.; Albiges, L.; Schmidinger, M.; Heng, D.Y.; Larkin, J.; Ficarra, V. Renal cell carcinoma. Nat. Rev. Dis. Prim. 2017, 3, 17009. [Google Scholar] [CrossRef]
  31. Tanaka, T.; Huang, Y.; Marukawa, Y.; Tsuboi, Y.; Masaoka, Y.; Kojima, K.; Iguchi, T.; Hiraki, T.; Gobara, H.; Yanai, H.; et al. Differentiation of small renal masses on multiphase contrast-enhanced CT by deep learning. Am. J. Roentgenol. 2020, 214, 605–612. [Google Scholar] [CrossRef] [PubMed]
  32. Schneider, R.; Franciscan Health. Open, Laparoscopic and Robotic Surgery: What to Know. 22 January 2021. Available online: https://www.franciscanhealth.org/community/blog/minimally-invasive-surgery (accessed on 26 March 2023).
  33. Krabbe, L.-M.; Kunath, F.; Schmidt, S.; Miernik, A.; Cleves, A.; Walther, M.; Kroeger, N. Partial nephrectomy versus radical nephrectomy for clinically localized renal masses. Cochrane Database Syst. Rev. 2017, 2017, CD012045. [Google Scholar]
  34. Mittakanti, H.R.; Heulitt, G.; Li, H.-F.; Porter, J.R. Transperitoneal vs. Retroperitoneal robotic partial nephrectomy: A matched-paired analysis. World J. Urol. 2019, 38, 1093–1099. [Google Scholar] [CrossRef] [PubMed]
  35. American Cancer Society. Key Statistics about Kidney Cancer. Available online: https://www.cancer.org/cancer/kidney-cancer/about/key-statistics.html (accessed on 11 March 2023).
  36. Nazari, M.; Shiri, I.; Hajianfar, G.; Oveisi, N.; Abdollahi, H.; Deevband, M.R.; Oveisi, M.; Zaidi, H. Noninvasive Fuhrman grading of clear cell renal cell carcinoma using computed tomography radiomic features and machine learning. Radiol. Med. 2020, 125, 754–762. [Google Scholar] [CrossRef]
  37. Yu, L.; Chen, H.; Dou, Q.; Qin, J.; Heng, P.-A. Automated melanoma recognition in dermoscopy images via very deep residual networks. IEEE Trans. Med. Imaging 2016, 36, 994–1004. [Google Scholar] [CrossRef] [PubMed]
  38. Zhou, Y.; He, X.; Huang, L.; Liu, L.; Zhu, F.; Cui, S.; Shao, L. Collaborative learning of semi-supervised segmentation and classification for medical images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 2079–2088. [Google Scholar]
  39. Xie, Y.; Xia, Y.; Zhang, J.; Song, Y.; Feng, D.; Fulham, M.; Cai, W. Knowledge-based collaborative deep learning for benign-malignant lung nodule classification on chest CT. IEEE Trans. Med. Imaging 2018, 38, 991–1004. [Google Scholar] [CrossRef]
  40. Corbat, L.; Henriet, J.; Chaussy, Y.; Lapayre, J.-C. Fusion of multiple segmentations of medical images using ov2assion and deep learning methods: Application to CT-scans for tumoral kidney. Comput. Biol. Med. 2020, 124, 103928. [Google Scholar] [CrossRef] [PubMed]
  41. da Cruz, L.B.; Araujo, J.D.L.; Ferreira, J.L.; Diniz, J.O.B.; Silva, A.C.; de Almeida, J.D.S.; de Paiva, A.C.; Gattass, M. Kidney segmentation from computed tomography images using Deep Neural Network. Comput. Biol. Med. 2020, 123, 103906. [Google Scholar] [CrossRef]
  42. Pedersen, M.; Andersen, M.B.; Christiansen, H.; Azawi, N.H. Classification of a renal tumor using convolutional neural networks to detect oncocytoma. Eur. J. Radiol. 2020, 133, 109343. [Google Scholar] [CrossRef]
  43. McGillivray, P.D.; Ueno, D.; Pooli, A.; Mendhiratta, N.; Syed, J.S.; Nguyen, K.A.; Schulam, P.G.; Humphrey, P.A.; Adeniran, A.J.; Boutros, P.C.; et al. Distinguishing benign renal tumors with an oncocytic gene expression (ONEX) classifier. Eur. Urol. 2021, 79, 107–111. [Google Scholar] [CrossRef]
  44. Zabihollahy, H.F.; Schieda, N.; Krishna, S.; Ukwatta, E. Automated classification of solid renal masses on contrast-enhanced computed tomography images using a convolutional neural network with decision fusion. Eur. Radiol. 2020, 30, 5183–5190. [Google Scholar] [CrossRef] [PubMed]
  45. Grand Challenge. Kits19—Grand Challenge Homepage. Available online: https://kits19.grand-challenge.org/ (accessed on 11 March 2023).
  46. KiTS21 Challenge. The 2021 Kidney Tumor Segmentation Challenge. Available online: https://kits21.kits-challenge.org/ (accessed on 11 March 2023).
  47. The 2023 Kidney Tumor Segmentation Challenge. KiTS23. 14 April 2001. Available online: https://kits-challenge.org/kits23/ (accessed on 24 April 2023).
  48. Heller, N.; Isensee, F.; Maier-Hein, K.H.; Hou, X.; Xie, C.; Li, F.; Nan, Y.; Mu, G.; Lin, Z.; Han, M.; et al. The state of the art in kidney and kidney tumor segmentation in contrast-enhanced CT imaging: Results of the kits19 challenge. Med. Image Anal. 2021, 67, 101821. [Google Scholar] [CrossRef]
  49. Yang, E.; Kim, C.K.; Guan, Y.; Koo, B.-B.; Kim, J.-H. 3D multi-scale residual fully convolutional neural network for segmentation of extremely large-sized kidney tumor. Comput. Methods Programs Biomed. 2022, 215, 106616. [Google Scholar] [CrossRef]
  50. Hsiao, C.-H.; Lin, P.-C.; Chung, L.-A.; Lin, F.Y.-S.; Yang, F.-J.; Yang, S.-Y.; Wu, C.-H.; Huang, Y.; Sun, T.-L. A deep learning-based precision and automatic kidney segmentation system using efficient feature pyramid networks in computed tomography images. Comput. Methods Programs Biomed. 2022, 221, 106854. [Google Scholar] [CrossRef]
  51. Gaillard, F. Hounsfield Unit: Radiology Reference Article. Radiopaedia Blog RSS. 23 June 2022. Available online: https://radiopaedia.org/articles/hounsfield-unit (accessed on 11 March 2023).
  52. Lin, Z.; Cui, Y.; Liu, J.; Sun, Z.; Ma, S.; Zhang, X.; Wang, X. Automated segmentation of kidney and renal mass and automated detection of renal mass in CT urography using 3D u-net-based deep convolutional neural network. Eur. Radiol. 2021, 31, 5021–5031. [Google Scholar] [CrossRef]
  53. Yu, Q.; Shi, Y.; Sun, J.; Gao, Y.; Zhu, J.; Dai, Y. Crossbar-Net: A Novel Convolutional Neural Network for Kidney Tumor Segmentation in CT Images. IEEE Trans. Image Process. 2019, 28, 4060–4074. [Google Scholar] [CrossRef]
  54. Xuan, P.; Cui, H.; Zhang, H.; Zhang, T.; Wang, L.; Nakaguchi, T.; Duh, H.B. Dynamic graph convolutional autoencoder with node-attribute-wise attention for kidney and tumor segmentation from CT volumes. Knowl.-Based Syst. 2022, 236, 107360. [Google Scholar] [CrossRef]
  55. Ruan, Y.; Li, D.; Marshall, H.; Miao, T.; Cossetto, T.; Chan, I.; Daher, O.; Accorsi, F.; Goela, A.; Li, S. MB-FSGAN: Joint segmentation and quantification of kidney tumor on CT by the multi-branch feature sharing generative Adversarial Network. Med. Image Anal. 2020, 64, 101721. [Google Scholar] [CrossRef] [PubMed]
  56. Kong, J.; He, Y.; Zhu, X.; Shao, P.; Xu, Y.; Chen, Y.; Coatrieux, J.-L.; Yang, G. BKC-net: Bi-knowledge contrastive learning for renal tumor diagnosis on 3D CT images. Knowl.-Based Syst. 2022, 252, 109369. [Google Scholar] [CrossRef]
  57. Lin, F.; Cui, E.-M.; Lei, Y.; Luo, L.-P. CT-based machine learning model to predict the Fuhrman nuclear grade of clear cell renal cell carcinoma. Abdom. Radiol. 2019, 44, 2528–2534. [Google Scholar] [CrossRef]
  58. Zhao, Y.; Chang, M.; Wang, R.; Xi, I.L.; Chang, K.; Huang, R.Y.; Vallières, M.; Habibollahi, P.; Dagli, M.S.; Palmer, M. Deep learning based on MRI for differentiation of low-and high-grade in low-stage renal cell carcinoma. J. Magn. Reson. Imaging 2020, 52, 1542–1549. [Google Scholar] [CrossRef]
  59. Alzu’bi, D.; Abdullah, M.; Hmeidi, I.; AlAzab, R.; Gharaibeh, M.; El-Heis, M.; Almotairi, K.H.; Forestiero, A.; Hussein, A.M.; Abualigah, L. Kidney tumor detection, and classification based on Deep Learning Approaches: A new dataset in CT scans. J. Healthc. Eng. 2022, 2022, 3861161. [Google Scholar] [CrossRef]
  60. Han, S.; Hwang, S.I.; Lee, H.J. The classification of renal cancer in 3-phase CT images using a deep learning method. J. Digit. Imaging 2019, 32, 638–643. [Google Scholar] [CrossRef] [PubMed]
  61. Liu, N.; Gan, W.; Qu, F.; Wang, Z.; Zhuang, W.; Agizamhan, S.; Xu, L.; Yin, J.; Guo, H.; Li, D. Does the Fuhrman or World Health Organization/International Society of Urological Pathology grading system apply to the XP11.2 translocation renal cell carcinoma? Am. J. Pathol. 2018, 188, 929–936. [Google Scholar] [CrossRef]
  62. Campi, R.; Sessa, F.; Rivetti, A.; Pecoraro, A.; Barzaghi, P.; Morselli, S.; Polverino, P.; Nicoletti, R.; Marzi, V.L.; Spatafora, P.; et al. Case report: Optimizing pre- and intraoperative planning with hyperaccuracy three-dimensional virtual models for a challenging case of robotic partial nephrectomy for two complex renal masses in a horseshoe kidney. Front. Surg. 2021, 8, 665328. [Google Scholar] [CrossRef]
  63. Temple Health. Partial or Total Nephrectomy. Available online: https://www.templehealth.org/services/treatments/nephrectomy (accessed on 26 March 2023).
  64. Nash, K. How Do You Decide between Partial and Radical Nephrectomy? Urology Times. 17 August 2020. Available online: https://www.urologytimes.com/view/how-do-you-decide-between-partial-and-radical-nephrectomy (accessed on 26 March 2023).
  65. license.umn.edu. Kidney and Kidney Tumor Segmentation Data Available from Technology Commercialization. Available online: https://license.umn.edu/product/kidney-and-kidney-tumor-segmentation-data (accessed on 24 April 2023).
  66. Grand-challenge.org. KiTS19—Grand Challenge. Available online: https://kits19.grand-challenge.org/data/ (accessed on 24 April 2023).
  67. M Health Fairview. Homepage. Available online: https://www.mhealthfairview.org/ (accessed on 11 March 2023).
  68. Cleveland Clinic. Access Anytime Anywhere. Available online: https://my.clevelandclinic.org/ (accessed on 11 March 2023).
  69. Knipe, H. NIfTI (File Format): Radiology Reference Article. Radiopaedia Blog RSS. 6 December 2019. Available online: https://radiopaedia.org/articles/nifti-file-format (accessed on 11 March 2023).
  70. The GNU Operating System and the Free Software Movement. Available online: https://www.gnu.org/home.en.html (accessed on 11 March 2023).
  71. KiTS19_digitaloceanspaces. DigitalOcean Documentation. Available online: https://docs.digitalocean.com/products/spaces/ (accessed on 11 March 2023).
  72. Neheller. Neheller/Kits21: The Official Repository of the 2021 Kidney and Kidney Tumor Segmentation Challenge. GitHub. Available online: https://github.com/neheller/kits21 (accessed on 11 March 2023).
  73. da Cruz, L.B.; Júnior, D.A.; Diniz, J.O.; Silva, A.C.; de Almeida, J.D.; de Paiva, A.C.; Gattass, M. Kidney tumor segmentation from computed tomography images using deeplabv3+ 2.5D model. Expert Syst. Appl. 2022, 192, 116270. [Google Scholar] [CrossRef]
  74. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  75. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  76. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  77. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  78. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. MobileNetV2: Inverted residuals and linear bottlenecks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  79. Tahir, A.M.; Chowdhury, M.E.; Khandakar, A.; Rahman, T.; Qiblawey, Y.; Khurshid, U.; Kiranyaz, S.; Ibtehaz, N.; Rahman, M.S.; Al-Maadeed, S.; et al. COVID-19 infection localization and severity grading from chest X-ray images. Comput. Biol. Med. 2021, 139, 105002. [Google Scholar] [CrossRef] [PubMed]
  80. Qiblawey, Y.; Tahir, A.; Chowdhury, M.E.H.; Khandakar, A.; Kiranyaz, S.; Rahman, T.; Ibtehaz, N.; Mahmud, S.; Al Maadeed, S.; Musharavati, F.; et al. Detection and severity classification of COVID-19 in CT images using Deep Learning. Diagnostics 2021, 11, 893. [Google Scholar] [CrossRef]
  81. Rahman, T.; Ibtehaz, N.; Khandakar, A.; Hossain, S.A.; Mekki, Y.M.S.; Ezeddin, M.; Bhuiyan, E.H.; Ayari, M.A.; Tahir, A.; Qiblawey, Y.; et al. QUCoughScope: An intelligent application to detect COVID-19 patients using cough and breath sounds. Diagnostics 2022, 12, 920. [Google Scholar] [CrossRef]
  82. Logsoftmax. LogSoftmax—PyTorch 1.13 Documentation. Available online: https://pytorch.org/docs/stable/generated/torch.nn.LogSoftmax.html (accessed on 12 March 2023).
  83. Jocher, G.; Chaurasia, A.; Stoken, A.; Borovec, J.; Kwon, Y.; Michael, K.; Tao, X.; Fang, J.; Imyhxy, L.; Zeng, Y.; et al. Ultralytics/yolov5: V7.0—yolov5 Sota Realtime Instance Segmentation. Zenodo. 22 November 2022. Available online: https://zenodo.org/record/7347926 (accessed on 12 March 2023).
  84. Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M. Yolov7: Trainable Bag-of-Freebies Sets New State-of-the-Art for Real-Time Object Detectors. arXiv.org. 6 July 2022. Available online: https://arxiv.org/abs/2207.02696 (accessed on 12 March 2023).
  85. COCO. Common Objects in Context. Available online: https://cocodataset.org/#home (accessed on 12 March 2023).
  86. National Cancer Institute. Cancer Staging. Available online: https://www.cancer.gov/about-cancer/diagnosis-staging/staging (accessed on 1 April 2023).
  87. Renal Cell Carcinoma TNM Staging. Available online: https://reference.medscape.com/calculator/573/renal-cell-carcinoma-tnm-staging (accessed on 1 April 2023).
  88. Brems, M. A One-Stop Shop for Principal Component Analysis. Medium. 26 January 2022. Available online: https://towardsdatascience.com/a-one-stop-shop-for-principal-component-analysis-5582fb7e0a9c (accessed on 1 April 2023).
  89. Hafizh, M.; Badri, Y.; Mahmud, S.; Hafez, A.; Choe, P. COVID-19 vaccine willingness and hesitancy among residents in Qatar: A quantitative analysis based on machine learning. J. Hum. Behav. Soc. Environ. 2022, 32, 899–922. [Google Scholar] [CrossRef]
  90. Brownlee, J. Feature Importance and Feature Selection with XGBoost in Python. MachineLearningMastery.com. 27 August 2020. Available online: https://machinelearningmastery.com/feature-importance-and-feature-selection-with-xgboost-in-python/ (accessed on 3 April 2023).
  91. Malato, G. Feature Selection with Random Forest. Your Data Teacher. 27 May 2022. Available online: https://www.yourdatateacher.com/2021/10/11/feature-selection-with-random-forest/ (accessed on 3 April 2023).
  92. GeeksforGeeks. ML: Extra Tree Classifier for Feature Selection. 1 July 2020. Available online: https://www.geeksforgeeks.org/ml-extra-tree-classifier-for-feature-selection/ (accessed on 3 April 2023).
  93. CatBoost. State-of-the-Art Open-Source Gradient Boosting Library with Categorical Features Support. Available online: https://catboost.ai/ (accessed on 3 April 2023).
  94. Welcome to LightGBM’s Documentation! LightGBM 3.3.2 Documentation. Available online: https://lightgbm.readthedocs.io/en/v3.3.2/ (accessed on 3 April 2023).
  95. Hui, J. MAP (Mean Average Precision) for Object Detection. Medium. 3 April 2019. Available online: https://jonathan-hui.medium.com/map-mean-average-precision-for-object-detection-45c121a31173 (accessed on 12 March 2023).
  96. He, Y.; Huang, C.-W.; Wei, X.; Li, Z.; Guo, B. TF-YOLO: An improved incremental network for real-time object detection. Appl. Sci. 2019, 9, 3225. [Google Scholar] [CrossRef]
  97. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning Deep Features for Discriminative Localization. arXiv.org. 14 December 2015. Available online: https://arxiv.org/abs/1512.04150 (accessed on 12 March 2023).
  98. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. Int. J. Comput. Vis. 2019, 128, 336–359. [Google Scholar] [CrossRef]
  99. Chattopadhay, A.; Sarkar, A.; Howlader, P.; Balasubramanian, V.N. Grad-cam++: Generalized gradient-based visual explanations for deep convolutional networks. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018. [Google Scholar]
  100. Omeiza, D.; Speakman, S.; Cintas, C.; Weldermariam, K. Smooth Grad-Cam++: An Enhanced Inference Level Visualization Technique for Deep Convolutional Neural Network Models. arXiv.org. 3 August 2019. Available online: https://arxiv.org/abs/1908.01224 (accessed on 12 March 2023).
  101. Wang, H.; Wang, Z.; Du, M.; Yang, F.; Zhang, Z.; Ding, S.; Mardziel, P.; Hu, X. Score-cam: Score-weighted visual explanations for Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  102. Mudgalvivek. Machine Learning: Precision, Recall, F1-Score. Medium. 24 June 2020. Available online: https://medium.com/@mudgalvivek2911/machine-learning-precision-recall-f1-score-c0a064ea6008 (accessed on 3 April 2023).
  103. Suzuki, N.; Matsuki, E.; Araumi, A.; Ashitomi, S.; Watanabe, S.; Kudo, K.; Ichikawa, K.; Inoue, S.; Watanabe, M.; Ueno, Y.; et al. Association among chronic kidney disease, airflow limitation, and mortality in a community-based population: The Yamagata (Takahata) study. Sci. Rep. 2020, 10, 5570. [Google Scholar] [CrossRef]
  104. Trudzinski, F.C.; Alqudrah, M.; Omlor, A.; Zewinger, S.; Fliser, D.; Speer, T.; Seiler, F.; Biertz, F.; Koch, A.; Vogelmeier, C.; et al. Consequences of chronic kidney disease in chronic obstructive pulmonary disease. Respir. Res. 2019, 20, 151. [Google Scholar] [CrossRef]
  105. Chow, W.-H.; Dong, L.M.; Devesa, S.S. Epidemiology and risk factors for kidney cancer. Nat. Rev. Urol. 2010, 7, 245–257. [Google Scholar] [CrossRef] [PubMed]
  106. Ljungberg, B.; Campbell, S.C.; Cho, H.Y.; Jacqmin, D.; Lee, J.E.; Weikert, S.; Kiemeney, L.A. The epidemiology of renal cell carcinoma. Eur. Urol. 2011, 60, 615–621. [Google Scholar] [CrossRef]
  107. Lew, J.Q.; Chow, W.-H.; Hollenbeck, A.R.; Schatzkin, A.; Park, Y. Alcohol consumption and risk of renal cell cancer: The NIH-AARP diet and health study. Br. J. Cancer 2011, 104, 537–541. [Google Scholar] [CrossRef] [PubMed]
  108. Reynolds, K.; Gu, D.; Chen, J.; Tang, X.; Yau, C.; Yu, L.; Chen, C.-S.; Wu, X.; Hamm, L.; He, J. Alcohol consumption and the risk of end-stage renal disease among Chinese men. Kidney Int. 2008, 73, 870–876. [Google Scholar] [CrossRef]
  109. Cancerresearchuk.org. Treatment Options. Available online: https://www.cancerresearchuk.org/about-cancer/kidney-cancer/treatment/decisions (accessed on 30 May 2023).
  110. Cancer.org. Treatment of Kidney Cancer by Stage. Available online: https://www.cancer.org/cancer/types/kidney-cancer/treating/by-stage.html (accessed on 30 May 2023).
Figure 1. Schematic of the proposed framework for kidney cancer subtype classification using a combination of computed tomography (CT) images and clinical metadata (pipeline begins on the left and extends to the right-hand side of the figure). Surgical procedure classification for malignant renal cell carcinoma (RCC) tumors is a key output from this framework. Major methodological steps are highlighted in blue text.
Figure 1. Schematic of the proposed framework for kidney cancer subtype classification using a combination of computed tomography (CT) images and clinical metadata (pipeline begins on the left and extends to the right-hand side of the figure). Surgical procedure classification for malignant renal cell carcinoma (RCC) tumors is a key output from this framework. Major methodological steps are highlighted in blue text.
Cancers 15 03189 g001
Figure 2. (a) The original (a) DenseNet201 architecture depicted against the proposed novel (b) DenseAUXNet201 having two auxiliary outputs and a modified header. Deeper CNN layers are represented by darker blue shades. MLP layers are shown as yellow blocks.
Figure 2. (a) The original (a) DenseNet201 architecture depicted against the proposed novel (b) DenseAUXNet201 having two auxiliary outputs and a modified header. Deeper CNN layers are represented by darker blue shades. MLP layers are shown as yellow blocks.
Cancers 15 03189 g002
Figure 3. Example kidney tumor subtype visualization. From left, (a) clear cell (ccRCC), (b) Chromophobe (chRCC), (c) Papillary (pRCC) renal cell carcinoma, and (d) Oncocytoma (ONC).
Figure 3. Example kidney tumor subtype visualization. From left, (a) clear cell (ccRCC), (b) Chromophobe (chRCC), (c) Papillary (pRCC) renal cell carcinoma, and (d) Oncocytoma (ONC).
Cancers 15 03189 g003
Figure 4. Confusion Matrix of scope reduction by DenseAUXNet201 binary classifier.
Figure 4. Confusion Matrix of scope reduction by DenseAUXNet201 binary classifier.
Cancers 15 03189 g004
Figure 5. ScoreCAM-generated heatmaps of the (ac) kidney and (d) non-kidney slices.
Figure 5. ScoreCAM-generated heatmaps of the (ac) kidney and (d) non-kidney slices.
Cancers 15 03189 g005
Figure 6. Example ground-truth labels vs. prediction plots from the best performing YOLOv7 model. Shown is the detection of kidney, tumor, and cyst ROIs from CT slices (displayed for qualitative evaluation).
Figure 6. Example ground-truth labels vs. prediction plots from the best performing YOLOv7 model. Shown is the detection of kidney, tumor, and cyst ROIs from CT slices (displayed for qualitative evaluation).
Cancers 15 03189 g006
Figure 7. Subject-wise renal tumor subtype classification outcomes. Classification evaluation metrics for (a) CT-based classification using deep ML (the best performance for each metric has been made bold), and (b) combined features approach using classical ML. Confusion matrix for (c) CT-based classification, and (d) combined features approach.
Figure 7. Subject-wise renal tumor subtype classification outcomes. Classification evaluation metrics for (a) CT-based classification using deep ML (the best performance for each metric has been made bold), and (b) combined features approach using classical ML. Confusion matrix for (c) CT-based classification, and (d) combined features approach.
Cancers 15 03189 g007
Figure 8. Key determinants of kidney tumor classification through the combined approach as identified by random forest-based feature ranking. Here, more reddish bars represent higher magnitudes while more greenish bars represent lower magnitudes.
Figure 8. Key determinants of kidney tumor classification through the combined approach as identified by random forest-based feature ranking. Here, more reddish bars represent higher magnitudes while more greenish bars represent lower magnitudes.
Cancers 15 03189 g008
Figure 9. Key determinants of surgical procedure selection as identified by random forest-based feature ranking. Here, more reddish bars represent higher magnitudes while more greenish bars represent lower magnitudes.
Figure 9. Key determinants of surgical procedure selection as identified by random forest-based feature ranking. Here, more reddish bars represent higher magnitudes while more greenish bars represent lower magnitudes.
Cancers 15 03189 g009
Figure 10. Subject-wise classification results for surgical procedure selection in malignant cases (partial or radical nephrectomy decision based on clinical data). (a) Performance evaluation metrics for all classical ML-based classifiers, confusion matrix for (the best performance for each metric have been made bold) (b) all 256 malignant cases, (c) ccRCC cases, (d) chRCC cases, and © pRCC cases.
Figure 10. Subject-wise classification results for surgical procedure selection in malignant cases (partial or radical nephrectomy decision based on clinical data). (a) Performance evaluation metrics for all classical ML-based classifiers, confusion matrix for (the best performance for each metric have been made bold) (b) all 256 malignant cases, (c) ccRCC cases, (d) chRCC cases, and © pRCC cases.
Cancers 15 03189 g010
Table 1. Baseline patient and tumor characteristics of the KiTS21 dataset.
Table 1. Baseline patient and tumor characteristics of the KiTS21 dataset.
AttributeValues (N = 300)
Demographic
 Age (years)60 (51, 68)
 BMI ( kg / m 2 )29.82 (26.16, 35.28)
 Gender (M/F)(180/120) (60%/40%)
Tumor Size
 Tumor Diameter 1 (cm)4.2 (2.600, 6.125)
 Tumor Volume 1 ( cm 3 )34.93 (9.587, 109.7)
Subtype
 Clear Cell RCC (ccRCC 2)203 (67.7%)
 Papillary RCC (pRCC)28 (9.3%)
 Chromophobe RCC (chRCC)27 (9%)
 Oncocytoma (ONC)16 (5.3%)
 Angiomyolipoma (AML)5 (1.7%)
 Other21 (7%)
Malignancy
 Benign25 (8.3%)
 Malignant275 (91.7%)
Pathological T-stage
 025 (8.3%)
 1 (1a, 1b)180 (60%)
 2 (2a, 2b)20 (6.7%)
 370 (23.3%)
 45 (1.7%)
Pathological N-stage
 0122 (40.7%)
 111 (3.7%)
 21 (0.3%)
 X 3166 (55.3%)
Pathological M-stage
 0130 (43.3%)
 132 (10.7%)
 X 3138 (46%)
Renal Anatomy
 Normal293 (97.7%)
 Solitary5 (1.67%)
 Horseshoe2 (0.67%)
Surgery Type
 Open79 (26.3%)
 Robotic172 (57.3%)
 Laparoscopic49 (16.3%)
Surgical Procedure
 Radical Nephrectomy112 (37.3%)
 Partial Nephrectomy188 (62.6%)
Surgical Approach
 Transperitoneal244 (81.3%)
 Retroperitoneal56 (18.7%)
1 In cases of multiple tumor instances, only the largest measured tumor has been reported. 2 RCC: Renal Cell Carcinoma. 3 X: Unknown or unsure.
Table 2. Fuhrman cancer stage determination using the TNM grading system.
Table 2. Fuhrman cancer stage determination using the TNM grading system.
Cancer StageTNM
I100
II200
III300
1–31–30
IV4Any0
AnyAny1
Table 3. Scope reduction performance of deep-CNN-based binary classifiers. The best performance for each metric has been made bold.
Table 3. Scope reduction performance of deep-CNN-based binary classifiers. The best performance for each metric has been made bold.
ModelsPerformanceMissed Cases
AccuracyPrecisionRecallSpecificityF1-Score
ResNet15297.53%97.55%97.54%96.28%97.54%1592
DenseNet20197.95%97.95%97.95%97.07%97.95%1325
InceptionV397.75%97.75%97.75%96.66%97.75%1457
MobileNetV297.64%97.65%97.64%96.46%97.65%1524
DenseAUXNet20198.02%98.03%98.02%97.13%98.03%1277
Table 4. YOLO performance during kidney, tumor, and cyst ROI extraction from 2D-CT slices. The best performance for each metric has been made bold.
Table 4. YOLO performance during kidney, tumor, and cyst ROI extraction from 2D-CT slices. The best performance for each metric has been made bold.
ModelClassPrecisionRecallmAP at 0.5mAP at [0.5:0.95]Inference Time (ms)Number of Parameters (Millions)LayersGIGA-FLOPS
YOLOv5sKidney0.9820.9650.9860.920147.0215715.8
YOLOv5m0.9830.9680.9870.92217.320.8621247.9
YOLOv5l0.9830.9670.9870.92331.246.12267107.7
YOLOv5x0.9810.9690.9870.92420.986.2322203.8
YOLOv70.9750.9740.9880.91813.936.50314103.2
YOLOv7x0.9650.9750.9880.91713.370.80362188.0
YOLOv5sTumor0.7550.6390.7040.428147.0215715.8
YOLOv5m0.7630.6560.7120.45017.320.8621247.9
YOLOv5l0.7980.6640.7260.47631.246.12267107.7
YOLOv5x0.7580.6550.7130.47020.986.2322203.8
YOLOv70.7730.7180.7560.52513.936.50314103.2
YOLOv7x0.7610.7190.7550.51013.370.80362188.0
YOLOv5sCyst0.440.1840.1830.113147.0215715.8
YOLOv5m0.5130.2140.1820.11317.320.8621247.9
YOLOv5l0.4830.1810.1920.12031.246.12267107.7
YOLOv5x0.4820.2250.2220.13720.986.2322203.8
YOLOv70.4140.340.2780.18213.936.50314103.2
YOLOv7x0.4050.360.2770.17913.370.80362188.0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mahmud, S.; Abbas, T.O.; Mushtak, A.; Prithula, J.; Chowdhury, M.E.H. Kidney Cancer Diagnosis and Surgery Selection by Machine Learning from CT Scans Combined with Clinical Metadata. Cancers 2023, 15, 3189. https://doi.org/10.3390/cancers15123189

AMA Style

Mahmud S, Abbas TO, Mushtak A, Prithula J, Chowdhury MEH. Kidney Cancer Diagnosis and Surgery Selection by Machine Learning from CT Scans Combined with Clinical Metadata. Cancers. 2023; 15(12):3189. https://doi.org/10.3390/cancers15123189

Chicago/Turabian Style

Mahmud, Sakib, Tariq O. Abbas, Adam Mushtak, Johayra Prithula, and Muhammad E. H. Chowdhury. 2023. "Kidney Cancer Diagnosis and Surgery Selection by Machine Learning from CT Scans Combined with Clinical Metadata" Cancers 15, no. 12: 3189. https://doi.org/10.3390/cancers15123189

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop