Next Article in Journal
Ethnicity-Specific Molecular Alterations in MAPK and JAK/STAT Pathways in Early-Onset Colorectal Cancer
Previous Article in Journal
The Impact of miRNA Expression on Colon Cancer Severity, Invasiveness, and Localization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning: A Heuristic Three-Stage Mechanism for Grid Searches to Optimize the Future Risk Prediction of Breast Cancer Metastasis Using EHR-Based Clinical Data

1
Department of Biomedical Informatics, University of Pittsburgh, Pittsburgh, PA 15206, USA
2
Division of Hematology/Oncology, University of Pittsburgh School of Medicine, Pittsburgh, PA 15213, USA
3
UPMC Hillman Cancer Center, Pittsburgh, PA 15232, USA
4
Department of Pathology, University of Pittsburgh and Pittsburgh VA Health System, Pittsburgh, PA 15261, USA
*
Author to whom correspondence should be addressed.
Cancers 2025, 17(7), 1092; https://doi.org/10.3390/cancers17071092
Submission received: 24 January 2025 / Revised: 14 March 2025 / Accepted: 21 March 2025 / Published: 25 March 2025
(This article belongs to the Section Cancer Metastasis)

Simple Summary

Breast cancer metastasis is a major cause of death for women, and predicting its occurrence is vital for better treatment planning. In this study, the authors developed deep learning models to predict the risk of breast cancer metastasis over five, ten, and fifteen years. One challenge in improving the performance of these models is selecting the best set of adjustable parameters, which can be time-consuming and costly. The study introduces a three-stage grid search method to optimize these parameters while managing limited computational time. The first stage narrows down reasonable ranges for parameters, the second finds the optimal values, and the third refines the models using specific search strategies. The results showed that this method improved the accuracy of predictions by 18.6%, 16.3%, and 17.3% for the different time frames compared to random searches. This approach is valuable for clinicians, as it offers a manageable way to develop better predictive models despite limited resources. Additionally, the study highlights key factors influencing breast cancer metastasis prediction and identifies important parameters for improving model performance. These insights can contribute to more personalized and effective treatment plans for patients.

Abstract

Background: A grid search, at the cost of training and testing a large number of models, is an effective way to optimize the prediction performance of deep learning models. A challenging task concerning grid search is time management. Without a good time management scheme, a grid search can easily be set off as a “mission” that will not finish in our lifetime. In this study, we introduce a heuristic three-stage mechanism for managing the running time of low-budget grid searches with deep learning, sweet-spot grid search (SSGS) and randomized grid search (RGS) strategies for improving model prediction performance, in an application of predicting the 5-year, 10-year, and 15-year risk of breast cancer metastasis. Methods: We develop deep feedforward neural network (DFNN) models and optimize the prediction performance of these models through grid searches. We conduct eight cycles of grid searches in three stages, focusing on learning a reasonable range of values for each of the adjustable hyperparameters in Stage 1, learning the sweet-spot values of the set of hyperparameters and estimating the unit grid search time in Stage 2, and conducting multiple cycles of timed grid searches to refine model prediction performance with SSGS and RGS in Stage 3. We conduct various SHAP analyses to explain the prediction, including a unique type of SHAP analyses to interpret the contributions of the DFNN-model hyperparameters. Results: The grid searches we conducted improved the risk prediction of 5-year, 10-year, and 15-year breast cancer metastasis by 18.6%, 16.3%, and 17.3%, respectively, over the average performance of all corresponding models we trained using the RGS strategy. Conclusions: Grid search can greatly improve model prediction. Our result analyses not only demonstrate best model performance but also characterize grid searches from various aspects such as their capabilities of discovering decent models and the unit grid search time. The three-stage mechanism worked effectively. It not only made our low-budget grid searches feasible and manageable but also helped improve the model prediction performance of the DFNN models. Our SHAP analyses not only identified clinical risk factors important for the prediction of future risk of breast cancer metastasis, but also DFNN-model hyperparameters important to the prediction of performance scores.

1. Introduction

Electronic health record (EHR) systems have been used in clinical settings for many years. With the help of these systems, researchers, or medical practitioners who work in a clinical environment often have opportunities to access and curate a clinical dataset concerning a group of patients, and sometimes they may consider using such a dataset to build a patient outcome prediction model. For instance, one can use the EHR data collected from a breast cancer patient care center to develop deep learning models that can be used to predict for a patient the risk of future occurrences of breast cancer metastasis.
Breast cancer is a major cancer-related cause of death for women worldwide. Based on the report updated on 13 March 2024 by the World Health Organization (WHO), “Breast cancer was also the most common cancer in women in 157 countries out of 185 in 2022”, and it is also one of the main causes of cancer related death in women worldwide, and “caused 670,000 deaths globally in 2022”. Again, according to the WHO, as of the end of 2020, “there were 7.8 million women alive who were diagnosed with breast cancer in the past 5 years, making it the world’s most prevalent cancer”. Women do not die of breast cancer, rather, they die mainly due to breast cancer metastasis, which can occur years after the initial treatment of breast cancer [1,2]. Predicting a late metastatic occurrence of breast cancer for a patient is important because the prediction can help make more suitable treatment plans for the patient, which may help prevent breast cancer metastasis. Improving our capability of predicting breast cancer metastasis is an important task in breast cancer patient care. Even just a small percentage of improvement can help greatly improve patient quality of life and save lives and care-related costs.
The field of machine learning (ML) and deep learning [3,4,5,6,7,8] has provided us with various AI-based computational methods for conducting predictions. Using these ML methods, we can learn a prediction model automatically from a dataset. However, a prediction model that is developed in such a manner does not always predict well [9]. There are often multiple factors that can affect model performance. For example, the model performance is usually dataset-dependent, that is, the same machine learning method can perform totally differently when it is applied to different datasets [4,9,10]. This phenomenon is perhaps partly because different datasets contain different levels of “signals” that are critical in making correct predictions. When a dataset contains very weak signals, even an advanced method can fail learning a good prediction model. By “signals” we mean the information usable for making predictions, contained in data. A good example of a signal is the so-called correlation between two variables. If two variables are correlated, then one of them can be used to predict the other, and in that case, the former is often called the predictor and the latter is often called the outcome. Sometimes, to curate a dataset that contains sufficient information for learning a good prediction model, we need to collect a lot of data. Generally speaking, the more datapoints (cases) a dataset contains, the more likely it provides sufficient information for learning a good prediction model. That perhaps explains in an aspect why the applications of ML methods are often associated with the term “big data” and a field called data science.
Other than the dataset itself, another important factor for model performance is the value used for an adjustable hyperparameter of a prediction model. All machine learning methods that we used so far have adjustable hyperparameters, but a difference is some methods have more and some have less [4,11,12,13,14]. During the early years of using machine learning methods to carry on real-life prediction tasks, we paid little attention to the selection of a value for an adjustable hyperparameter that is built into a machine learning method [9,10,15]. A normal practice is to use the default value recommended by the developers of the machine learning method or by a machine learning textbook, or at most try a few values that are close to the default value. This may have contributed significantly to the fact that some of the ML methods such as the first generation of the Neural Network reportedly had poor prediction performance [9,10,16].
Deep learning is a machine learning method that has quite some adjustable hyperparameters [17,18,19]. For example, we identified 13 adjustable hyperparameters (see Table 1 and a more detailed description of these hyperparameters in Section 2.3) for the Deep Feedforward Neural Network (DFNN) models that we developed for predicting later occurrences of breast cancer metastasis [4,19]. In a previous study, by conducting machine learning experiments, we found that different value assignments of the adjustable hyperparameters can lead to models with significantly different prediction performance [4]. We used a method called grid search [4,20,21,22,23,24] to systematically train and test different DFNN models by changing the values of the set of adjustable hyperparameters [4]. Therefore, in order to do a grid search, we normally preselect a range of ranges for each of the hyperparameters as an input to the grid search.
We now use an example to explain what a grid search does and, in the meantime, introduce a major challenge of conducting a grid search. In this example, we use the number of values given to each of the 13 hyperparameters as shown in the second row of Table 1 below. The number of hidden layers and the number of hidden nodes (neurons) in a hidden layer are the two structural hyperparameters that together determines the structure of a DFNN model. In this example grid search, the number of hidden layers has four different values, that is, the model can contain up to four hidden layers. The number of hidden nodes per hidden layer has 22 different values. Therefore, there could be 22, 484, 10,648, or 234,256 different model structures when the number of hidden layers is configured to be 1, 2, 3, or 4. Thus, we can make a total of 245,410 different models by considering the two structural hyperparameters alone. The total number of possible unique value assignments to the set of the 11 remaining non-structural hyperparameters are the product of the number of values given to each of these hyperparameters, which arrives at 1.7424 × 109. Therefore, considering all 13 hyperparameters, there are in total 4.276 × 1014 unique value assignments. What the example grid search does is to train and test 4.276 × 1014 DFNN models determined by the 4.276 × 1014 different value assignments, one at a time. We call a unique value assignment to the set of hyperparameters of a grid search a hyperparameter setting. Note that under each of the hyperparameter settings, there would be k different models trained and tested if the k-fold cross validation (CV) procedure (see the Methods section) is applied. Since we use a 5-fold CV procedure in our grid searches, the number of models trained and tested is five times the number of hyperparameters settings used in these grid searches.
A grid search can be very costly. Based on the grade search experiments we conducted in our previous study [4], the average running time per hyperparameter setting is 117 s for a particular dataset we used. By using this average unit running time, the estimated total running time for the example grid search, as described above, would be 1,586,424,369 years. Apparently, this grid search is not feasible for us unless we use billions of computers to run it parallelly. A grid search is, in general, very time consuming, but computation time can sometimes be resolved by using high-speed computing, which can be bought by money. Therefore, the feasibility issue of a grid search essentially boils down to a financial budget issue. Going back to the scenario that we mentioned at the beginning of this introduction, in which medical researchers or practitioners want to learn a prediction model using their own EHR-based datasets, a normal situation is that the time and funds available are both very limited for conducting a grid search to optimize prediction. We call a grid search for which the money allocation is very limited a low-budget grid search. It is not uncommon to encounter a low-budget grid search in the real-life applications of machine learning methods.
Moreover, the example grid search we described above can only be called a small-scale grid search, in which only a small number of values is given to each of the hyperparameters. But the estimated running time is already hard to manage. Based on this example, we see that the number of values allowed for each of the hyperparameters can only be very small for a low-budget grid search to finish in a foreseeable time. But note that some of the hyperparameters of our DFNN models can take a very large number of different values (in Section 2.3). For example, based on our previous studies [4,19], the range of values we consider for a hyperparameter called epochs is from 5 to 2000, which means for each grid search we need to select a very small set of values from the 1996 different values for epochs. As a matter of fact, most of the 13 hyperparameters can take a very large number of values, and some can even take an infinite number of different values within a normal range. It is often a challenging task to select a very small set of values from a large number of values available for each of the hyperparameters like epochs, and in the meantime to ensure the feasibility of a low-budget grid and meet the goal of digging out a better prediction model through the grid search. To our knowledge, there is no standard and good way of doing this. In this study, we have limited time to run grid searches for optimizing our DFNN models that predict the risk of breast cancer metastasis. We therefore introduce what we call a three-stage heuristic grid-search mechanism that we use to manage this challenging task. We describe this mechanism and the experiments we conduct, in which we apply this mechanism, in Section 2, and present and analyze the results of the experiments in Section 3 and Section 4.

2. Methods

2.1. The DFNN Models

Deep learning and deep feedforward neural network (DFNN): An Artificial Neural Network (ANN) is a machine learning framework, which is designed to recognize patterns using a model loosely resembling the human brain [25,26]. ANNs can be used for clustering (unsupervised) on unlabeled data or classification (supervised) on labeled data [4]. Deep Neural Networks (DNNs), called deep learning, refers to the use of neural networks composed of more than one hidden layer [3,5,6,7,27]. The DNN has obtained significant success in commercial applications such as voice and pattern recognition, computer vision, and image-based processing [28,29,30,31,32,33,34,35,36,37,38]. However, its power has not been fully explored or demonstrated in applications that are not image-based, such as the prediction of breast cancer metastasis using non-image clinical data. This is due in part to the sheer magnitude of the number of variables involved in these problems, which presents formidable computational and modeling challenges [17,18,39]. We developed the deep feedforward neural network (DFNN) models that predict the risk of a future occurrence of breast cancer metastasis for a patient [4,19] and conduct grid searches to hand these challenges. The DFNN models we developed are fully connected neural networks that do not contain cycles. Figure 1 illustrates, as an example, the structure and the inner connections of the DFNNs that we have developed. The example shown in Figure 1 is a six-layer neural network that contains one input layer, four hidden layers, and one output layer. The 31 input nodes to this neural network represent the 31 clinical features contained in the patient data that we use, and the output layer contains two nodes representing the binary status of 5-year, 10-year, or 15-year breast cancer metastasis. Each node in this model has an activation function (in Section 2.3), represented by f(x), which decides the node’s individual output value established by the current value of the node. In such a DFNN model, each hidden layer has a certain number of hidden nodes that can be different from the other layers. Both the number of hidden layers and the number of hidden nodes in a hidden layer are two of the set of hyperparameters whose values are subjected to changes during grid searches. These two hyperparameters together with other adjustable hyperparameters for our grid searchers are described in Section 2.3 below.

2.2. Datasets

The MBIL method is a Bayesian Network-based method for identifying risk factors (RFs) for an outcome feature, which was applied to three EHR-based clinical datasets concerning breast cancer metastasis, that is, the LSM-5-year, 10-year, and 15-year datasets [40]. These LSM datasets were preprocessed through previous research concerning a clinical decision support system [2] and therefore do not contain missing values. The procedure for preprocessing the EHR-based dataset can be found in this previous study [2]. We use MBIL to retrieve all RFs concerning breast cancer metastasis from the LSM datasets and develop three new datasets according to the RFs: the LSM_RF-5-year, 10-year, and 15-year datasets. Using the LSM_RF-5Year dataset as an example, the 2-step procedure for developing this dataset is as follows: Step 1, applying the MBIL method to the LSM-5Year dataset to retrieve the RFs of 5-year breast cancer metastasis. The original LSM-5-year dataset contains 32 features including a feature called metastasis, which represents the state of having or not having breast cancer metastasis by the 5th year post the initial treatment [40]. In this study, metastasis is the outcome feature, which we also call the target feature, because we are interested in predicting the value of these features using the other features; the remaining 31 features are called predictors. Step 2: remove all predictors that do not belong to the set of RFs found in Step 1 from the LSM-5Year dataset, and all data points of the remaining features form the LSM_RF-5-year dataset. We also follow this 2-step procedure to obtain the LSM_RF-10-year dataset from the original LSM-10-year dataset and the LSM_RF-15-year dataset from the LSM-15-year dataset. All experiments conducted in this study are based on the RF datasets. Table 2 below shows the counts of the cases and predictors included in the three RF datasets. A detailed description of the predictors is included in the Tables S1–S3 of the supplement.

2.3. The Three-Stage Grid Search Mechanism and the Experiments

In this study, we tune 13 hyperparameters (Table 3) following a heuristic three-stage grid search mechanism for learning the DFNN models. In Stage 1, we focus on learning performance trends and a set of proper values for each of the hyperparameters that can take a very large or an infinite number of different values, such as the epochs and learning rate. The extreme values that can result in performance outliers, that is, the models that perform very poorly, are usually removed from the set of proper values. Stage 2 and Stage 3 grid searches are guided by the set of proper values. In Stage 2, we attempt to estimate the average running time per hyperparameter setting (RTPS) and identify performance sweet spots by conducting a grid search that is guided by the results of the Stage 1 grid searches. The main difference between the Stage 1 and the Stage 2 grid searches is that, in a Stage 1 grid search, we only change the values of a single hyperparameter, and this allows the Stage 1 grid search to train and test models using a large number of different values of the hyperparameter and still finish in an acceptable timeframe. But in a Stage 2 grid search or a Stage 3 grid search, we allow all hyperparameters to take multiple different values. The Stage 2 experiments are important because, based on the results of such a grid search, we can do better in managing the running time of the subsequent Stage 3 grid searches. Specifically, we can estimate the total running time based on the RTPS if we know the input number of hyperparameter settings, and on the other hand, we can computer the input number of hyperparameter settings based the total running time allowed. This ensures that a grid search finishes within an acceptable timeframe and, therefore, is critical to a low-budget and time-sensitive grid search. In Stage 3, we conduct multiple cycles of grid searches to further refining mode prediction performance. A cycle means the entire process of running a grid search from preselecting a set of values for each hyperparameter till finishing all experiments concerning model training and testing scheduled for the grid search. In this study, we conduct 6-cycles of Stage 3 grid searches for each of the three datasets, labeled as Stage3-c1, c2, and up to c6, each, respectively. The hyperparameter settings of the stage3-c1 are determined based on the results of the Stage 2 grid searches, that is, when conducting hyperparameter value selection, we focus on a small number of values that are close to the value used by the best model obtained from the Stage 2 grid searches. We call such a grid search a sweet spot grid search (SSGS), in which the sweet spot is determined by a previous grid search. Stage 3 cycle 1 through 5 are all SSGSs. In Stage 3, we also use our out of the local optimal (OLO) strategy for preselecting the input hyperparameter values. This is because in an SSGS we only use a small range of values that are in the neighborhood of the value used in the best model produced from a previous gird search, then the next best value of a new SSGS will still be within this small range of values, forming a situation that resembles the “local optima” phenomenon. Specifically, in Stage3-c4 and Stage3-c5, we identify a new sweet spot that is outside the “neighborhood” of the previous sweet spot to get out of the potential local optimal formed during the SSGSs. Finally, in Stage3-c6, we apply both OLO and a strategy called the randomized grid search (RGS) that we created. Once the proper set of values for each of the hyperparameters is determined via Stage 1, we will be able to determine all possible unique hyperparameter settings, which we call the pool of hyperparameter settings (PHS). A RGS (Randomized Grid Search) is a grid search that trains and tests models at a set of hyperparameters settings that are randomly picked from a PHS. Specifically, we first randomly selected a set of values for each of the adjustable hyperparameters to produce the input set of hyperparameter settings. This is because the number of possible hyperparameter settings in the pool can be very large and cannot be searched exhaustively with the limited time and low budget that we have. The key point of our RGS strategy is to randomly generate a fixed number of hyperparameter settings following uniform distributions, so that the grid search can be finished within a predictable timeframe and in the meantime every hyperparameter setting in the pool has a chance to be selected. Figure 2 below is a flowchart for demonstrating the three-stage grid search mechanism. Table 3 below contains more detailed information about the 13 hyperparameters of our DFNN models. The specific range of values used in each cycle of our three-stage grid searches are shown in Table S4–S6 of the supplement.

2.4. Prediction Performance Metrics and the 5-Fold Cross Validation Process

Our grid searches follow the 5-fold cross validation (CV) mechanism to train and test models at each hyperparameter setting and use an AUC score to measure the prediction performance of a model. AUC stands for the area under the curve of a receiver operator characteristic (ROC) curve that plots the true positive rate against the false positive rate for all possible cutoff values [41]. An AUC score measures the discrimination performance of the model.
To conduct a 5-fold CV, we need to split a dataset prior to grid searches. We use the following procedure to split our datasets: (1) split the entire dataset into a train–test set that contains 80% of the cases and a validation set that contains 20% of the cases. The train–test set will be given to a grid search as the input dataset, and the validation set will be kept aside for a later validation test; (2) divide a train–test set evenly into 5 portions for conducting the 5-fold CV. The division is mostly performed randomly except that each portion should have approximately 20% of the positive cases and 20% of the negative cases to ensure that it is a representative fraction of the dataset. We conduct a 5-fold CV at each hyperparameter setting of a grid search. During a 5-fold CV, 5 different models are generated and tested, each is trained using a unique combination of 4 portions and tested with the remaining portion. Then, 5 AUC scores are produced based on the tests, and the average of these scores is called the mean_test_AUC. The mean_test_AUC metric is used by grid searches to measure the model discrimination performance. A top hyperparameter setting selection at the end of a grid search is also based on this metric. A top model is developed by refitting the entire train–test set using the top hyperparameter setting selected by a grid search.

2.5. The SHAP Values and Plots

We use the SHAP (Shapley Additive Explanations) values to explain the prediction results of our DFNN models and identify the important features for the predicted future risk of breast cancer metastasis. The SHAP values are established based on the Shapley values, which distribute the payoff by measuring the marginal contribution of individual team members to the outcome of a cooperation game [42,43].
We use the Kernel Explainer provided by the SHAP package [44] to compute the SHAP values for the DFNN models. In order to manage computation time while preserving the integrity of the information contained in data, we process the training dataset using the k-means clustering method. We therefore identify k representative cluster centroids, where k is equal to the number of features in the dataset. These centroids, which epitomize the typical characteristics of the training dataset, are then employed as the background data for the SHAP Kernel Explainer. The background value of a feature is the mean of the corresponding feature values of the k centroids, identified using the k-means clustering.
The formula used to compute SHAP values is as follows:
ϕ i   ( p ) = S F \ { i } S ! F S 1 ! F !   [ p ( S { i } ) p ( S ) ]
In this formula, F denotes the complete set of features, i represents the ith feature, and S represents a subset of F not including the ith feature, that is, S F \ { i } . |S| is the size of a subset S, and |F| is the size of the set F. Let p(x) represents a model’s prediction outcome, then p ( S { i } ) represents the model’s prediction output when using the features in a subset S together with the ith feature as the predictors, while p ( S ) denotes the model’s prediction output when using only the features in S as the predictors. The term p ( S { i } ) p ( S ) therefore reflects the contribution to the model’s prediction output made by the ith feature with respect to a subset S.
Inspired by LIME [45], the Kernel Explainer generates synthetic samples, which are used as the test cases for a model of interest to compute the SHAP value of a feature. Let us call the feature the ith feature. Each of the synthetic samples contains the real values of a subset S, taken from the 20% set-aside validation dataset (see Section 2), and the background values for the remaining features in F, excluding the ith feature. The ith feature takes its real value for the synthetic sample that is created for obtaining the p ( S { i } ) , and takes its corresponding background value in the synthetic sample that is created to obtain the p ( S ) . Recall that the background values for all features in F are generated using the k-means method as described above.
We also use the SHAP package to generate a SHAP bar plot which ranks the feature importance values of all predictors from high to low. A SHAP feature importance value is the mean absolute SHAP value of all the test cases for the feature of interest. Another type of SHAP plots we show are the so called SHAP summary plots, which not only rank the features by their feature importance values but also show the SHAP value of each individual case that is tested. The SHAP heatmap plots show patterns of groups of instances or features, and the dependence plots explore potential interactive features which jointly affect a SHAP value.

3. Results

We followed the three-stage mechanism, described in Section 2, and ran grid search experiments in eight cycles: 1 cycle for Stage 1, 1 cycle for stage 2, and 6 cycles for stage 3, named as Stage3-c1, Stage3-c2, and so on and so forth. For each cycle, we identified groups of top performing models, that is, the top 1, top 5, top 10, top 50, and top 100 models, out of all models that were trained. We then computed the average mean_test_AUC for each group as a measurement of group prediction performance. We also obtained the average mean_test_AUC for all models trained during a cycle of grid searches as the largest group. We compare the group prediction performance side by side of the eight cycles of grid searches for predicting 5-year breast cancer metastasis in Table 4, 10-year in Table 5, and 15-year in Table 6 below.
In addition, to compare the top-performing groups of models among the eight cycles of grid searches, we are also interested in knowing the performance of grid searches in identifying “decent” prediction models. An AUC of 0.5 is often treated as the worst prediction performance score, because it indicates that a model’s prediction capability is equivalent to that of a random guess. What we mean by a “decent” prediction model is a model that scores at least as high as what we call a mid-point score, which is the average of 0.5 and the highest score of all models trained during the corresponding grid searches. So, a “decent” prediction model found in a cycle is a model that has an AUC no less than the mid-point score of the cycle of grid searches. A mid-point group is the group of “decent” prediction models found in a cycle of grid searches. Table 7 below compares the results concerning the mid-point groups of the eight cycles of grid searches.
As previously mentioned, managing the running time of a grid search is a challenging task. Knowing the RTPS, the average time it takes to train models at one hyperparameter setting, can be very useful in various aspects such as pre-estimating the total running time of a grid search, and estimating number of hyperparameter settings that a grid search take in when time we have to run a grid search is limited. During each of the eight cycles of grid searches, we periodically updated the corresponding RTPS based on the results we already obtained at the time and then use it to guide upcoming experiments. Table 8 below shows the final RTPSs in seconds, the total number of hyperparameter settings (TNS), and the total running time (TRT) in hours for each cycle of grid searches we conducted, corresponding to the 5-year, 10-year, and 15-year models each, respectively.
Figure 3, Figure 4 and Figure 5 each contains eight SHAP feature importance plots generated using the eight best DFNN models, each obtained from one of the eight cycles of grid searches, for predicting 5-year, 10-year, and 15-year breast cancer metastasis, respectively. In such a plot, each feature’s mean absolute SHAP value is shown on the Y-axis, and the feature names are shown on the X-axis. Figure 6 contains two types of SHAP plots, the SHAP heatmaps (Figure 6a,c,e) and SHAP summary plots (Figure 6b,d,f), produced using the three best DFNN models for predicting 5-year, 10-year, and 15-year breast cancer metastasis, each, respectively. Each of the three best Models is selected out of all models trained in the eight cycles of grid searches. A SHAP heatmap details the SHAP value distribution for all features, allowing the revelations of value patterns of grouped cases and features. In such a plot, features are ordered by their SHAP feature importance values from high to low as shown by a black bar plot to the right of the heatmap, and cases are grouped by the similarity of their SHAP impact based on hierarchical clustering. The notation f(x) represents the distribution of the model’s predictions. The heatmap plots as shown in Figure 6 were produced using the default hierarchical clustering method included in the SHAP package [44]. In companion with a SHAP heatmap, a SHAP summary plot not only shows the ordered average impact of all features but also shows the specific SHAP values of the test cases with respect to each value of a feature. Therefore, such a plot reflects how a value of a feature influences the prediction. The summary plots in Figure 6 were also generated using the SHAP package.
We further analyze and demonstrate the interpretation of a model’s prediction using the SHAP dependence plots with interaction visualization, as shown in Figure 7. To generate each of the subfigures in Figure 7, we used the best model found from all models trained during the corresponding eight cycles of grid searches. Figure 7a,c,e are the correlation bar plots for the 5-year, 10-year, and 15-year predictions each, respectively, which rank the association strength between each feature and the SHAP values of the most important feature identified using the SHAP feature importance. Using Figure 7e, the figure for the 15-year prediction as an example, since the most import feature based on the SHAP feature importance value (see Figure 6e or Figure 6f) is AGE (age_at_diagonosis), Figure 7e shows that the feature that is mostly correlated with the SHAP impact of AGE is LYS (lymph_node_status). Figure 7b,d,f are the corresponding dependence plots with interaction visualization, which demonstrate the interaction between the most import feature and the feature that is mostly correlated with the impact of the most important feature. Again, using the 15-year as an example, since LYS is mostly correlated with the SHAP impact of the most important feature AGE, as shown in Figure 7e, Figure 7f visualizes the joint effect of AGE and LYS. In Figure 7f, with respect to each individual case, represented as a colored dot, the value of the most important feature, AGE, is plotted on the X axis, the SHAP value for the most important feature is shown on the Y-axis, and the color shows the value of LYS, the feature that is mostly correlated with the SHAP impact of AGE.
In the stage3-c6 grid searches, we used the RGS strategy (see Section 2), which allows us to use a broad range of values for each of the hyperparameters (see Tables S4–S6); therefore a large number of different hyperparameter settings were used in these grid searches. Specifically, in this cycle, 70,568, 105,500, and 93,000 different hyperparameter settings were used to train models for predicting 5-year, 10-year, and 15-year breast cancer metastasis, each, respectively. Since the information about each hyperparameter setting and the corresponding mean_test_AUC from the 5-fold CV process are both recorded in our grid-search output, we were able to conduct SHAP analyses using these output as new data to identify the important hyperparameters to the predicted mean_test_AUC scores. In these SHAP analyses, each of the DFNN model hyperparameters was treated as a predictor (a feature), and the mean_test_AUC was treated as the outcome variable.
Our procedure for conducting this unique application of the SHAP analyses is as follows: (1) process the grid search results so that they can be used as the input dataset of the SHAP analyses. For example, we recoded free-text-like values of a categorical hyperparameter such as kinit (kernel initializer) with digits. Using kinit as an example, we converted the value constant to 0, glorot_normal to 1, glorot_uniform to 2, he_normal to 3, and he_uniform to 4; (2) use the k-means algorithm to cluster the input dataset into n clusters, n is corresponding to the number of hyperparameters. Then, use the centroid of each cluster to serve as the background data required by the SHAP Kernel Explainer (see the Methods); (3) train a prediction model using 80% of the input data. For our analysis specifically, we trained a prediction model using the Random Forest method with 100 estimators; (4) calculate the SHAP values using the 20% saved data and the prediction model obtained in (3), and then compute SHAP importance values for the features (the hyperparameters).
The results of the SHAP analyses concerning DFNN model hyperparameters are shown in Figure 8 below. Figure 8a,c,e are the SHAP feature importance plots concerning the hyperparameters, generated using the DFNN model grid-search output for predicting 5-year, 10-year, and 15-year breast cancer metastasis, each, respectively, and Figure 8b,d,f are the corresponding SHAP summary plots.

4. Discussion

Based on Table 4, for predicting 5-year breast cancer metastasis, the best performing model comes from Stage3-c6, with a mean_test_AUC that is higher than the average mean_test_AUC of all models trained in this cycle by 18.6%. Since we applied the RGS strategy in the Stage3-c6 grid searches, the average mean_test_AUC of all models trained in this cycle reflects the expected model performance when a hyperparameter value is randomly selected from a set of proper values. Table 5 shows that for predicting 10-year breast cancer metastasis, the best performing model comes also from stage 3 but is in cycle 3, with a mean_test_AUC that is higher than the average mean_test_AUC of all models trained in Stage3-c6 by 16.3%. For predicting 15-year breast cancer metastasis, the best performing model is found in stage3-c1, with a mean_test_AUC that is higher than the average mean_test_AUC of all models trained in Stage3-c6 by 17.3%.
Table 4, Table 5 and Table 6 also show that the three-stage grid search mechanism overall works as expected due to the following reasons: Firstly, all best models are found in the Stage 3 grid searches. This is consistent with our goal of further refining grid searches in Stage 3 by building upon the preparation work from the Stage 1 and Stage 2 grid searches. Secondly, we conducted SSGSs in Stage3-c1 through Stage3-c3 grid searches, focusing on refining prediction perform based on the “sweet spot” derived from the results of its previous cycle. As expected, both Table 4 and Table 5 show a steady increase in terms of both the best and group-average model performance across these three cycles of grid searches. Thirdly, we used the RGS strategy in Stage3-c6, in which we allowed a hyperparameter to have a chance of taking any value in its proper set of values, determined based on Stage 1 grid searches. We reckon that such a grid search may accidently identify a best model because it has more “freedom” in choosing its hyperparameter value combinations, but its group model performance may not be as good because the pure randomness in selecting hyperparameter values can also result in very bad models. This is indeed reflected in our results as shown in Table 4 through Table 6.
In Table 4, Table 5 and Table 6, we not only compare the prediction performance of the best models but also compare group model performance among the five groups including top 5, 10, 50, 100, and “All”. The group performance comparisons reflect somewhat the characteristics of the grid searches of the different stages, in which different strategies were applied. As seen in Table 4 through Table 6, Stage 1 grid searches tend to have a low “All” model performance. This is perhaps because in Stage 1 we ran grid searches that focused on one hyperparameter at a time and, therefore, we were able to test a broad range of hyperparameter values. The extreme values used for a hyperparameter could lead to very bad models, which help drag down the average score of all models. Stage 2 overall does better than Stage 1 in group model performance, this is perhaps because the hyperparameter values used in stage 2 were selected from the set of proper values resulting from Stage 1. Table 4, Table 5 and Table 6 also show that Stage 3 Cycle 1 through Cycle 5 tend to do better than both Stage 1 and Stage 2, and the explanation mostly lies in the fact that we conducted SSGSs in these cycles, in which the hyperparameter values used are “in the neighborhood” of some best values found in previous grid searches. The results seem to suggest that, in terms of model performance, if one model does well, then its neighbors tend to do well also. In addition, Cycle 4 and 5 tend to do worse than Cycle 1 through 3 perhaps due to the OLO strategy. Our results also show that Stage 3 Cycle 6 grid searches are competitive in terms of group model performance except for the “All” model group. This indicates that the RGS strategy is competitive in terms of identifying top performing models, but the average performance of all models is brought down by performance outliers resulted from the pure randomness of hyperparameter value selection.
We trained hundreds of thousands of models at each of the eight cycles (see Table 8, and recall that 5 models were trained at each hyperparameter setting due to the 5-fold CV). Out of such a large number of models, how many of them are decent prediction models? An answer to this question may help further characterize the performance of grid searches. Table 7 shows the summary data that we derived from the output of our grid searches to answer this question. We described what we meant by a decent model in the Result section. Based on the CHS/TNS ratios in Table 7, Stage 3 Cycle 1 through Cycle 5 grid searches, that is, the so called SSGSs, did the best in terms of searching for a decent prediction model. The models trained by each of these five cycles are 100% or close to 100% decent, and this is true regardless which dataset was used. We also notice that the Stage 3 Cycle 6 grid searches produced the worst results in terms of identifying decent models, perhaps also due to the randomness of the hyperparameter value selection, required by the RGS strategy.
As described in the Introduction, time management has always been a critical issue in grid searches, especially in low-budget ones. Based on Table 8, Stage 1 and the SSGSs in Stage 3, especially the Cycle 1 through Cycle 4, tend to link to a relatively low RTPS, while the Stage 3 Cycle 6 for which we used the RGS strategy and Stage 2 tend to give a relatively high RTPS. By consulting Table S4 through Figure S6, we found that the values of a hyperparameter called the number of hidden nodes in a hidden layer seem to positively correlate with the RTPS, because the value changes in this hyperparameter can well explain the changes in RTPS. For example, the highest RTPS that we see comes from Stage 3 Cycle 6, which can be explained by the range of high values (550–800) of this hyperparameter used in this cycle. In another example, we notice that Stage 1 allows this hyperparameter to take the highest value 1005 but only has the second highest RTPS. This can be explained by the broadest range of values (1 to 1005) that this hyperparameter takes in Stage 1, which should give an average value that is lower than the values used in Stage 3 Cycle 6. From Table 8, we also notice that within the same cycle of grid searches, where the same search strategy was used, the RTPSs for the three different datasets are also significantly different. The RTPS is the highest for the 5-year dataset, and lowest for the 15-year dataset. Based on Table 2, a significant difference among the three datasets is the size of a dataset, that is, the number of cases (also called data points in machine learning) contained in a dataset. The 5-year dataset contains the largest number of cases, while the 15-year dataset contains the lowest number. Table 2 and Table 8 together reveal a positive correlation between the number of data points contained in a dataset and the RTPS of a grid search.
Based on Figure 3, Figure 4 and Figure 5, the important features in the best models are quite consistent among the eight cycles. A variable called LYP (lymph_nodes_positive) is found to be the most important feature for predicting 5-year breast cancer metastasis in all eight best models shown in Figure 3. In addition, Figure 3 also demonstrates that LYP is far more important than the other features, and this is consistently demonstrated by the SHAP importance plots of almost all the best models. According to our knowledge about cancer, it seems to make sense that the number of positive lymph nodes is a good indicator for breast cancer metastasis at diagnosis and, therefore, a good indicator for the risk of cancer recurrence in the near future. Based on Figure 3, the top three important features for predicting 5-year breast cancer metastasis also include ER (estrogen_receptor_expression) and STA (stage), with ER to be the second most important feature voted by five out of the eight best models, and STA to be the third most important feature voted by six out of the eight best models.
According to Figure 4, the top three most important features for predicting the risk of 10-year breast cancer include STA, identified as the most important in five out of the eight best models, GRA (grade), identified as the most important by three out of the eight best models, and AGE (age_at_diagnosis), identified as the number three important by five out of the eight best models. Unlike the situation where one feature mostly dominates as seen in the 5-year DFNN models, we see that more than one feature dominate in most of the importance plots. Based on Figure 5, the top three important features for predicting the risk of 15-year breast cancer metastasis are AGE, LYS (lymph_nodes_status) and MEN (menopausal_status). We also notice that in seven out of the eight best models, these three top features together dominate the overall predicted risk of breast cancer metastasis.
Lymph node status, estrogen positivity, and age are known clinical predictors for 5-year BCSS (breast cancer specific survival) [46,47]. Stage, grade, and age are also known clinical predictors for 10-year BCSS [46,48]. Finally, age, lymph node stage, and menopausal status are predictors of 15-year BCSS [46,48].
Although the feature importance figures (Figure 3, Figure 4 and Figure 5) identify the top feature(s) that dominate(s) the overall importance of the predicted results, they do not show the relationships among the features in terms of their impact. The SHAP heatmaps as shown in Figure 6 perhaps do better in this regard. For example, Figure 6e, a SHAP heatmap plot, reveals color patterns in terms of both features and cases, which indicate similar influences of the grouped members on the model’s prediction. However, the heatmap plots do not show whether the features interact to have a joint effect on prediction.
An interactive effect is an additional effect after subtracting the individual effect from each of the interactive features. A SHAP dependence plot with interaction visualization such as the ones shown in Figure 7 carries more information in terms of a joint contribution between two features. For example, Figure 7e shows that LYS is strongly correlated with the impact from AGE, which is the most important feature for predicting the risk of 15-year breast cancer metastasis (see Figure 5f and Figure 6e). Figure 7f demonstrates that even at the same value of AGE, the SHAP values distribute differently at different values of LYS, confirming that AGE and LYS, the top two most influential features for predicting the risk of 15-year breast cancer metastasis, indeed interact to contribute to the prediction. More specifically, based on Figure 7f, when the value AGE is 0 (for patients who are 0 to 49 years old), it has overall a negative impact on the predicted risk, when the value of AGE is 1 (for patients who are 50 to 69 years old), it has an overall positive but low impact on the predicted risk, when AGE is 2 (for patients who are above 70 years old), it has a significantly increased positive impact on the predicted risk. In addition, Figure 7f reveals that at low ages, LYS tends to be positively correlated with the SHAP impact of AGE, while as age increases, LYS tends to be negatively correlated with the SHAP impact of AGE. In terms of AGE and LYS, LYS is known to clinically affect DFMS (distant metastasis free survival) as above. Age may be important as older women either (1) die of causes other than breast cancer over a 15-year period, and (2) younger women have a longer lifespan to develop distant disease [49]. Additionally older women tend to be undertreated with adjuvant therapy as opposed to younger women [50].
As shown by Figure 8a,c,e, the hyperparameter that has the highest overall impact on the predicted mean_test_AUC is kinit, which is used to assign the initial weights of some regression functions used in gradient descent, a key process in deep learning. This is a bit out of our expectation, because we originally thought for a deep learning, the depth of the network, which is reflected in a hyperparameter called mstruct should be the most important feature. The importance of kinit perhaps comes from its influence on gradient descent. An inappropriate value assignment of kinit can cause divergent learning, which should have a negative impact on model performance. The summary plots as shown in Figure 8b,d,f, reveal that the value of kinit is positively correlated with its SHAP value, which seems to suggest that he_uniform, the highest value of kinit, has the highest positive impact on the predicted mean_test_AUC of a DFNN model. This may help explain that he_uniform was often associated with a good DFNN model, as seen in some preliminary experiments that we conducted [4].
Figure 8 also demonstrates that the second most important feature in terms of the predicted mean_test_AUC is L1, a coefficient of the LASSO regularization, which can help select features and therefore control model sparsity during training [51]. The summary plots in Figure 8 suggest a negative correlation between the values of L1 and its SHAP values. Mstruct ranks number three in terms of its importance to the predicted mean_test_AUCs. It is a two-dimensional hyperparameter for our DFNN models, consisting of two hyperparameters: the number of hidden layers and the number of hidden nodes in a hidden layer. The summary plots in Figure 8 reveal a positive correlation between mstruct and its SHAP values, indicating that the complexity of a neural network structure may have a positive impact on the predicted mean_test_AUCs. This may help explain the observed success of deep neural networks in contrast to the “shallow” neural networks, namely, the first-generation one-hidden layer neural networks. Finally, we also notice from Figure 8 that the rankings of the top three most important features are quite consistent in all plots. This may indicate that the most important adjustable hyperparameters for the DFNN models are not so dependent on a dataset. However, since this is the first time, as far as we know, that a SHAP analysis is used to explain the role of a hyperparameter as a feature in a predicted mean_test_AUC, the implications of the kind of results shown in Figure 8 are subjected to further explorations.

5. Conclusions

Our grid searches improve the risk prediction of 5-year, 10-year, and 15-year breast cancer metastasis by 18.6%, 16.3%, and 17.3%, relatively to the average mean_test_AUC of all corresponding models, for which the value of each hyperparameter was randomly picked from a set of reasonable values. Our result analyses characterize grid searches in terms of their capabilities of identifying groups of top performers and “decent” models. The heuristic three-stage mechanism worked effectively. It enables us to finish the low-budget grid searches within the expected timeframe. The results of the grid searches are consistent with what we expected, that is, all the best performing models are found in Stage 3, the final model-refining stage. The local cycles of SSGSs in Stage 3 overall show steadily increased performance not only in terms of the best models, but also in different groups of models. The SSGS Strategy significantly beats the RGS strategy in terms of the average performance of all models and percentage of “decent” models found in all models, but the RGS is quite competitive with the SSGS in identifying best models. The unit grid search time is positively correlated with both the number of hidden nodes in a hidden layer and the number of data points contained in a dataset. The results also suggest that a model with a set of hyperparameter values “close to” the set of values of a good prediction model tends to be a good model. The SHAP analyses not only reveal the features with high impact on the predicted risk of breast cancer metastasis but also help identify pairs of interactive features with regard to SHAP values. Finally, a unique SHAP application demonstrates that the kernel initializer, L1, and mstruct are the top three most important hyperparameters to the predicted mean_test_AUCs.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/cancers17071092/s1, Table S1: Predictors in the LSM_RF-5Year Dataset; Table S2: Predictors in the LSM_RF-10Year Dataset; Table S3: Predictors in the LSM_RF-15 Year Dataset; Table S4: Range of hyperparameter values used in the three-stage grid searches for predicting 5 year breast cancer metastasis; Table S5: Range of hyperparameter values used in the three-stage grid searches for predicting 10 year breast cancer metastasis; Table S6: Range of hyperparameter values used in the three-stage grid searches for predicting 15 year breast cancer metastasis.

Author Contributions

X.J. originated the study and designed the methods. X.J. and Y.Z. wrote the first draft of the manuscript. X.J., C.X. and Y.Z. implemented the methods. C.X. and Y.Z. conducted the experiments. X.J., Y.Z. and C.X. prepared and analyzed the results. All authors contributed to the preparation and revision of the manuscript. All work was conducted at the University of Pittsburgh. All authors have read and agreed to the published version of the manuscript.

Funding

Research reported in this paper was supported by the U.S. Department of Defense through the Breast Cancer Research Program under Award No. W81XWH1910495 (to XJ). Other than supplying funds, the funding agencies played no role in the research.

Institutional Review Board Statement

This study was approved by the University of Pittsburgh Institutional Review Board (IRB # 196003, 6/12/2019) and the U.S. Army Human Research Protection Office (HRPO # E01058.1a, 7/31/2020). The need for patient consent was waived by the ethics committees because the data consists only of de-identified data that are publicly available.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study are available at datadryad.org (DOI 10.5061/dryad.64964m0).

Conflicts of Interest

The authors declare that they have no competing interests.

References

  1. Gupta, G.P.; Massagué, J. Cancer Metastasis: Building a Framework. Cell 2006, 127, 679–695. [Google Scholar]
  2. Jiang, X.; Wells, A.; Brufsky, A.; Neapolitan, R. A Clinical Decision Support System Learned from Data to Personalize Treatment Recommendations towards Preventing Breast Cancer Metastasis. PLoS ONE 2019, 14, e0213292. [Google Scholar] [CrossRef]
  3. Schmidhuber, J. Deep Learning in Neural Networks: An Overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [PubMed]
  4. Jiang, X.; Xu, C. Deep Learning and Machine Learning with Grid Search to Predict Later Occurrence of Breast Cancer Metastasis Using Clinical Data. J. Clin. Med. 2022, 11, 5772. [Google Scholar] [CrossRef] [PubMed]
  5. Chen, L.C.; Schwing, A.G.; Yuille, A.L.; Urtasun, R. Learning Deep Structured Models. In Proceedings of the 32nd International Conference on Machine Learning, ICML 2015, Lille, France, 6–11 July 2015; Volume 3. [Google Scholar]
  6. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M.; et al. TensorFlow: A System for Large-Scale Machine Learning. In Proceedings of the 12th USENIX Symposium on Operating Systems Design and Implementation, OSDI 2016, Savannah, GA, USA, 2–4 November 2016. [Google Scholar]
  7. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT: Cambridge, MA, USA, 2016. [Google Scholar]
  8. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning Bible. Healthc. Inform. Res. 2016, 22, 351–354. [Google Scholar]
  9. Cai, B.; Jiang, X. Computational Methods for Ubiquitination Site Prediction Using Physicochemical Properties of Protein Sequences. BMC Bioinform. 2016, 17, 116. [Google Scholar] [CrossRef] [PubMed]
  10. Jiang, X.; Cai, B.; Xue, D.; Lu, X.; Cooper, G.F.; Neapolitan, R.E. A Comparative Analysis of Methods for Predicting Clinical Outcomes Using High-Dimensional Genomic Datasets. J. Am. Med. Inform. Assoc. 2014, 21, e312–e319. [Google Scholar]
  11. Goodfellow, I.; Bengio, Y.; Courville, A.; RegGoodfellow, I.; Bengio, Y.; Courville, A. Regularization for Deep Learning. In Deep Learning; MIT Press: Cambridge, MA, USA, 2016; pp. 216–261. [Google Scholar]
  12. Brownlee, J. How to Choose Loss Functions When Training Deep Learning Neural Networks; Machine Learning Mastery: Santa Monica, CA, USA, 2019. [Google Scholar]
  13. Hoerl, A.E.; Kennard, R.W. Ridge Regression: Biased Estimation for Nonorthogonal Problems. Technometrics 1970, 12, 55–67. [Google Scholar] [CrossRef]
  14. Ghojogh, B.; Crowley, M. The Theory Behind Overfitting, Cross Validation, Regularization, Bagging, and Boosting: Tutorial. arXiv 2019, arXiv:1905.12787. [Google Scholar]
  15. Jiang, X.; Neapolitan, R.E. Evaluation of a Two-Stage Framework for Prediction Using Big Genomic Data. Brief. Bioinform. 2015, 16, 912–921. [Google Scholar] [CrossRef]
  16. Cai, B.; Jiang, X. A Novel Artificial Neural Network Method for Biomedical Prediction Based on Matrix Pseudo-Inversion. J. Biomed. Inform. 2014, 48, 114–121. [Google Scholar] [CrossRef] [PubMed]
  17. Glorot, X.; Bengio, Y. Understanding the Difficulty of Training Deep Feedforward Neural Networks. In Proceedings of the 13th International Conference on Artificial Intelligence and Statistics (AISTATS), Sardinia, Italy, 13–15 May 2010; Volume 9. [Google Scholar]
  18. Kim, L.S. Understanding the Difficulty of Training Deep Feedforward Neural Networks Xavier. In Proceedings of the International Joint Conference on Neural Networks, Nagoya, Japan, 25–29 October 1993; Volume 2. [Google Scholar]
  19. Xu, C.; Coen-Pirani, P.; Jiang, X. Empirical Study of Overfitting in Deep Learning for Predicting Breast Cancer Metastasis. Cancers 2023, 15, 1969. [Google Scholar] [CrossRef] [PubMed]
  20. Ngoc, T.T.; van Dai, L.; Phuc, D.T. Grid Search of Multilayer Perceptron Based on the Walk-Forward Validation Methodology. Int. J. Electr. Comput. Eng. 2021, 11, 1742–1751. [Google Scholar] [CrossRef]
  21. Siji George, C.G.; Sumathi, B. Grid Search Tuning of Hyperparameters in Random Forest Classifier for Customer Feedback Sentiment Prediction. Int. J. Adv. Comput. Sci. Appl. 2020, 11, 173–178. [Google Scholar] [CrossRef]
  22. Michel, A.; Ro, V.; McGuinness, J.E.; Mutasa, S.; Terry, M.B.; Tehranifar, P.; May, B.; Ha, R.; Crew, K.D. Breast cancer risk prediction combining a convolutional neural network-based mammographic evaluation with clinical factors. Breast Cancer Res. Treat. 2023, 200, 237–245. [Google Scholar] [CrossRef] [PubMed]
  23. Hamedi, S.Z.; Emami, H.; Khayamzadeh, M.; Rabiei, R.; Aria, M.; Akrami, M.; Zangouri, V. Application of machine learning in breast cancer survival prediction using a multimethod approach. Sci. Rep. 2024, 14, 30147. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  24. Akselrod-Ballin, A.; Chorev, M.; Shoshan, Y.; Spiro, A.; Hazan, A.; Melamed, R.; Barkan, E.; Herzel, E.; Naor, S.; Karavani, E.; et al. Predicting Breast Cancer by Applying Deep Learning to Linked Health Records and Mammograms. Radiology 2019, 292, 331–342. [Google Scholar] [CrossRef] [PubMed]
  25. McCulloch, W.S.; Pitts, W. A Logical Calculus of the Ideas Immanent in Nervous Activity. Bull. Math. Biophys. 1943, 5, 115–133. [Google Scholar] [CrossRef]
  26. Farley, B.G.; Clark, W.A. Simulation of Self-Organizing Systems by Digital Computer. IRE Prof. Group Inf. Theory 1954, 4, 76–84. [Google Scholar] [CrossRef]
  27. Neapolitan, R.E.; Jiang, X. Neural Networks and Deep Learning. In Artificial Intelligence; Chapman and Hall/CRC: Boca Raton, FL, USA, 2018; pp. 389–411. [Google Scholar]
  28. Ran, L.; Zhang, Y.; Zhang, Q.; Yang, T. Convolutional Neural Network-Based Robot Navigation Using Uncalibrated Spherical Images. Sensors 2017, 17, 1341. [Google Scholar] [CrossRef]
  29. Deng, L.; Tur, G.; He, X.; Hakkani-Tur, D. Use of Kernel Deep Convex Networks and End-to-End Learning for Spoken Language Understanding. In Proceedings of the 2012 IEEE Workshop on Spoken Language Technology, SLT 2012—Proceedings, Miami, FL, USA, 2–5 December 2012. [Google Scholar]
  30. Fernández, S.; Graves, A.; Schmidhuber, J. An Application of Recurrent Neural Networks to Discriminative Keyword Spotting. In Artificial Neural Networks—ICANN 2007; Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2007; Volume 4669 LNCS. [Google Scholar]
  31. Naik, N.; Madani, A.; Esteva, A.; Keskar, N.S.; Press, M.F.; Ruderman, D.; Agus, D.B.; Socher, R. Deep Learning-Enabled Breast Cancer Hormonal Receptor Status Determination from Base-Level H&E Stains. Nat. Commun. 2020, 11, 5727. [Google Scholar] [CrossRef] [PubMed]
  32. Min, S.; Lee, B.; Yoon, S. Deep Learning in Bioinformatics. Brief. Bioinform. 2017, 18, 851–869. [Google Scholar]
  33. Lundervold, A.S.; Lundervold, A. An Overview of Deep Learning in Medical Imaging Focusing on MRI. Z. Med. Phys. 2019, 29, 102–127. [Google Scholar] [CrossRef] [PubMed]
  34. Lee, Y.W.; Huang, C.S.; Shih, C.C.; Chang, R.F. Axillary Lymph Node Metastasis Status Prediction of Early-Stage Breast Cancer Using Convolutional Neural Networks. Comput. Biol. Med. 2021, 130, 104206. [Google Scholar] [CrossRef]
  35. Papandrianos, N.; Papageorgiou, E.; Anagnostis, A.; Feleki, A. A Deep-Learning Approach for Diagnosis of Metastatic Breast Cancer in Bones from Whole-Body Scans. Appl. Sci. 2020, 10, 997. [Google Scholar] [CrossRef]
  36. Zhou, L.Q.; Wu, X.L.; Huang, S.Y.; Wu, G.G.; Ye, H.R.; Wei, Q.; Bao, L.Y.; Deng, Y.B.; Li, X.R.; Cui, X.W.; et al. Lymph Node Metastasis Prediction from Primary Breast Cancer US Images Using Deep Learning. Radiology 2020, 294, 19–28. [Google Scholar] [CrossRef]
  37. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A Survey on Deep Learning in Medical Image Analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar]
  38. Mohanty, S.P.; Hughes, D.P.; Salathé, M. Using Deep Learning for Image-Based Plant Disease Detection. Front. Plant Sci. 2016, 7, 1419. [Google Scholar] [CrossRef]
  39. Shen, H. Towards a Mathematical Understanding of the Difficulty in Learning with Feedforward Neural Networks. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  40. Jiang, X.; Wells, A.; Brufsky, A.; Shetty, D.; Shajihan, K.; Neapolitan, R.E. Leveraging Bayesian Networks and Information Theory to Learn Risk Factors for Breast Cancer Metastasis. BMC Bioinform. 2020, 21, 298. [Google Scholar] [CrossRef]
  41. Fawcett, T. An Introduction to ROC Analysis. Pattern Recognit. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
  42. Shapley, L.S. Notes on the N-Person Game II: The Value of an n-Person Game; Rand Corporation Research Memoranda; RM-670; Rand Corporation: Santa Monica, CA, USA, 1951. [Google Scholar]
  43. Shapley, L.S. The Value of an N-Person Game. Contributions to the Theory of Games (AM-28), Volume II. In Classics in Game Theory; Princeton University Press: Princeton, NJ, USA, 1953. [Google Scholar]
  44. Lundberg, S.M.; Lee, S.I. A Unified Approach to Interpreting Model Predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  45. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why Should I Trust You?” Explaining the Predictions of Any Classifier. In Proceedings of the NAACL-HLT 2016—2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Demonstrations Session, San Diego, CA, USA, 12–17 June 2016. [Google Scholar]
  46. Wishart, G.C.; Bajdik, C.D.; Dicks, E.; Provenzano, E.; Schmidt, M.K.; Sherman, M.; Greenberg, D.C.; Green, A.R.; Gelmon, K.A.; Kosma, V.M.; et al. PREDICT Plus: Development and validation of a prognostic model for early breast cancer that includes HER2. Br. J. Cancer 2012, 107, 800–807. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  47. Baade, P.D.; Fowler, H.; Kou, K.; Dunn, J.; Chambers, S.K.; Pyke, C.; Aitken, J.F. A prognostic survival model for women diagnosed with invasive breast cancer in Queensland, Australia. Breast Cancer Res. Treat. 2022, 195, 191–200. [Google Scholar] [CrossRef] [PubMed]
  48. Soerjomataram, I.; Louwman, M.W.; Ribot, J.G.; Roukema, J.A.; Coebergh, J.W. An overview of prognostic factors for long-term survivors of breast cancer. Breast Cancer Res. Treat. 2008, 107, 309–330. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  49. Pan, H.; Gray, R.; Braybrooke, J.; Davies, C.; Taylor, C.; McGale, P.; Peto, R.; Pritchard, K.I.; Bergh, J.; Dowsett, M.; et al. 20-year risks of breast-cancer recurrence after stopping endocrine therapy at 5 years. N. Engl. J. Med. 2017, 377, 1836–1846. [Google Scholar] [CrossRef] [PubMed]
  50. Sang, Y.; Yang, B.; Mo, M.; Liu, S.; Zhou, X.; Chen, J.; Hao, S.; Huang, X.; Liu, G.; Shao, Z.; et al. Treatment and survival outcomes in older women with primary breast cancer: A retrospective propensity score-matched analysis. Breast 2022, 66, 24–30. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  51. Tibshirani, R. Regression Shrinkage and Selection Via the Lasso. J. R. Stat. Soc. Ser. B (Methodol.) 1996, 58, 267–288. [Google Scholar] [CrossRef]
Figure 1. The structure of an example DFNN model we developed [4,19], a fully connected deep feedforward neural network that consists of an input layer, 4 hidden layers, and an output layer.
Figure 1. The structure of an example DFNN model we developed [4,19], a fully connected deep feedforward neural network that consists of an input layer, 4 hidden layers, and an output layer.
Cancers 17 01092 g001
Figure 2. A flowchart for the 3-stage grid search procedure.
Figure 2. A flowchart for the 3-stage grid search procedure.
Cancers 17 01092 g002
Figure 3. Feature importance plots of the best cycle-wise DFNN-5-year models. (a), (b), (c), (d), (e), (f), (g), and (h) were generated using the best DFNN models found in Stage 1, Stage 2, Stage 3-c1, …, Stage3-c6, each respectively. AGE: age at diagnosis of the disease; ALC: alcohol usage; DCI: type of ductal carcinoma in situ; ER: estrogen receptor expression; ERP: percent of cell stain pos for ER receptors; ETH: ethnicity; FAM: family history of cancer; GRA: grade of disease; HER: HER2 expression; HI1: tumor histology; HI2: tumor histology subtypes; INL: where invasive tumor is located; INV: whether tumor is invasive; LYP: number of positive lymph nodes; LYR: number of lymph nodes removed; LYS: patient had any positive lymph nodes; MEN: inferred menopausal status; MRI: MRIs within 60 days of surgery; NTN: number of nearby cancerous lymph nodes; PR: progesterone receptor expression; PRP: percent of cell stain pos for PR receptors; P53: whether P53 is mutated; RAC: race; REE: removal of an additional margin of tissue; SID: side of tumor; SIZ: size of tumor in mm; SMO: smoking; STA: composite of size and # positive nodes; SUR: whether residual tumor; TNE: triple negative status in terms of patient being ER, PR, and HER2 negative; TTN: prime tumor stage in TNM system.
Figure 3. Feature importance plots of the best cycle-wise DFNN-5-year models. (a), (b), (c), (d), (e), (f), (g), and (h) were generated using the best DFNN models found in Stage 1, Stage 2, Stage 3-c1, …, Stage3-c6, each respectively. AGE: age at diagnosis of the disease; ALC: alcohol usage; DCI: type of ductal carcinoma in situ; ER: estrogen receptor expression; ERP: percent of cell stain pos for ER receptors; ETH: ethnicity; FAM: family history of cancer; GRA: grade of disease; HER: HER2 expression; HI1: tumor histology; HI2: tumor histology subtypes; INL: where invasive tumor is located; INV: whether tumor is invasive; LYP: number of positive lymph nodes; LYR: number of lymph nodes removed; LYS: patient had any positive lymph nodes; MEN: inferred menopausal status; MRI: MRIs within 60 days of surgery; NTN: number of nearby cancerous lymph nodes; PR: progesterone receptor expression; PRP: percent of cell stain pos for PR receptors; P53: whether P53 is mutated; RAC: race; REE: removal of an additional margin of tissue; SID: side of tumor; SIZ: size of tumor in mm; SMO: smoking; STA: composite of size and # positive nodes; SUR: whether residual tumor; TNE: triple negative status in terms of patient being ER, PR, and HER2 negative; TTN: prime tumor stage in TNM system.
Cancers 17 01092 g003
Figure 4. Feature importance plots of the best cycle-wise DFNN-10-year models. (a), (b), (c), (d), (e), (f), (g), and (h) were generated using the best DFNN models found in Stage 1, Stage 2, Stage 3-c1, …, Stage3-c6, each respectively. AGE: age at diagnosis of the disease; ALC: alcohol usage; DCI: type of ductal carcinoma in situ; ER: estrogen receptor expression; ERP: percent of cell stain pos for ER receptors; ETH: ethnicity; FAM: family history of cancer; GRA: grade of disease; HER: HER2 expression; HI1: tumor histology; HI2: tumor histology subtypes; INL: where invasive tumor is located; INV: whether tumor is invasive; LYP: number of positive lymph nodes; LYR: number of lymph nodes removed; LYS: patient had any positive lymph nodes; MEN: inferred menopausal status; MRI: MRIs within 60 days of surgery; NTN: number of nearby cancerous lymph nodes; PR: progesterone receptor expression; PRP: percent of cell stain pos for PR receptors; P53: whether P53 is mutated; RAC: race; REE: removal of an additional margin of tissue; SID: side of tumor; SIZ: size of tumor in mm; SMO: smoking; STA: composite of size and # positive nodes; SUR: whether residual tumor; TNE: triple negative status in terms of patient being ER, PR, and HER2 negative; TTN: prime tumor stage in TNM system.
Figure 4. Feature importance plots of the best cycle-wise DFNN-10-year models. (a), (b), (c), (d), (e), (f), (g), and (h) were generated using the best DFNN models found in Stage 1, Stage 2, Stage 3-c1, …, Stage3-c6, each respectively. AGE: age at diagnosis of the disease; ALC: alcohol usage; DCI: type of ductal carcinoma in situ; ER: estrogen receptor expression; ERP: percent of cell stain pos for ER receptors; ETH: ethnicity; FAM: family history of cancer; GRA: grade of disease; HER: HER2 expression; HI1: tumor histology; HI2: tumor histology subtypes; INL: where invasive tumor is located; INV: whether tumor is invasive; LYP: number of positive lymph nodes; LYR: number of lymph nodes removed; LYS: patient had any positive lymph nodes; MEN: inferred menopausal status; MRI: MRIs within 60 days of surgery; NTN: number of nearby cancerous lymph nodes; PR: progesterone receptor expression; PRP: percent of cell stain pos for PR receptors; P53: whether P53 is mutated; RAC: race; REE: removal of an additional margin of tissue; SID: side of tumor; SIZ: size of tumor in mm; SMO: smoking; STA: composite of size and # positive nodes; SUR: whether residual tumor; TNE: triple negative status in terms of patient being ER, PR, and HER2 negative; TTN: prime tumor stage in TNM system.
Cancers 17 01092 g004
Figure 5. Feature importance plots of the best cycle-wise DFNN-15-year models. (a), (b), (c), (d), (e), (f), (g), and (h) were generated using the best DFNN models found in Stage 1, Stage 2, Stage 3-c1, …, Stage3-c6, each respectively. AGE: age at diagnosis of the disease; ALC: alcohol usage; DCI: type of ductal carcinoma in situ; ER: estrogen receptor expression; ERP: percent of cell stain pos for ER receptors; ETH: ethnicity; FAM: family history of cancer; GRA: grade of disease; HER: HER2 expression; HI1: tumor histology; HI2: tumor histology subtypes; INL: where invasive tumor is located; INV: whether tumor is invasive; LYP: number of positive lymph nodes; LYR: number of lymph nodes removed; LYS: patient had any positive lymph nodes; MEN: inferred menopausal status; MRI: MRIs within 60 days of surgery; NTN: number of nearby cancerous lymph nodes; PR: progesterone receptor expression; PRP: percent of cell stain pos for PR receptors; P53: whether P53 is mutated; RAC: race; REE: removal of an additional margin of tissue; SID: side of tumor; SIZ: size of tumor in mm; SMO: smoking; STA: composite of size and # positive nodes; SUR: whether residual tumor; TNE: triple negative status in terms of patient being ER, PR, and HER2 negative; TTN: prime tumor stage in TNM system.
Figure 5. Feature importance plots of the best cycle-wise DFNN-15-year models. (a), (b), (c), (d), (e), (f), (g), and (h) were generated using the best DFNN models found in Stage 1, Stage 2, Stage 3-c1, …, Stage3-c6, each respectively. AGE: age at diagnosis of the disease; ALC: alcohol usage; DCI: type of ductal carcinoma in situ; ER: estrogen receptor expression; ERP: percent of cell stain pos for ER receptors; ETH: ethnicity; FAM: family history of cancer; GRA: grade of disease; HER: HER2 expression; HI1: tumor histology; HI2: tumor histology subtypes; INL: where invasive tumor is located; INV: whether tumor is invasive; LYP: number of positive lymph nodes; LYR: number of lymph nodes removed; LYS: patient had any positive lymph nodes; MEN: inferred menopausal status; MRI: MRIs within 60 days of surgery; NTN: number of nearby cancerous lymph nodes; PR: progesterone receptor expression; PRP: percent of cell stain pos for PR receptors; P53: whether P53 is mutated; RAC: race; REE: removal of an additional margin of tissue; SID: side of tumor; SIZ: size of tumor in mm; SMO: smoking; STA: composite of size and # positive nodes; SUR: whether residual tumor; TNE: triple negative status in terms of patient being ER, PR, and HER2 negative; TTN: prime tumor stage in TNM system.
Cancers 17 01092 g005
Figure 6. SHAP heatmap and summary plot for the best DFNN-5-year, 10-year, and 15-year model, respectively. (a,c,e) are the SHAP heatmaps produced using the best three DFNN models for predicting 5-year, 10-year, and 15-year breast cancer metastasis, each respectively. (b,d,f) are the SHAP summary plots produced using the same three best models. AGE: age at diagnosis of the disease; ALC: alcohol usage; DCI: type of ductal carcinoma in situ; ER: estrogen receptor expression; ERP: percent of cell stain pos for ER receptors; ETH: ethnicity; FAM: family history of cancer; GRA: grade of disease; HER: HER2 expression; HI1: tumor histology; HI2: tumor histology subtypes; INL: where invasive tumor is located; INV: whether tumor is invasive; LYP: number of positive lymph nodes; LYR: number of lymph nodes removed; LYS: patient had any positive lymph nodes; MEN: inferred menopausal status; MRI: MRIs within 60 days of surgery; NTN: number of nearby cancerous lymph nodes; PR: progesterone receptor expression; PRP: percent of cell stain pos for PR receptors; P53: whether P53 is mutated; RAC: race; REE: removal of an additional margin of tissue; SID: side of tumor; SIZ: size of tumor in mm; SMO: smoking; STA: composite of size and # positive nodes; SUR: whether residual tumor; TNE: triple negative status in terms of patient being ER, PR, and HER2 negative; TTN: prime tumor stage in TNM system.
Figure 6. SHAP heatmap and summary plot for the best DFNN-5-year, 10-year, and 15-year model, respectively. (a,c,e) are the SHAP heatmaps produced using the best three DFNN models for predicting 5-year, 10-year, and 15-year breast cancer metastasis, each respectively. (b,d,f) are the SHAP summary plots produced using the same three best models. AGE: age at diagnosis of the disease; ALC: alcohol usage; DCI: type of ductal carcinoma in situ; ER: estrogen receptor expression; ERP: percent of cell stain pos for ER receptors; ETH: ethnicity; FAM: family history of cancer; GRA: grade of disease; HER: HER2 expression; HI1: tumor histology; HI2: tumor histology subtypes; INL: where invasive tumor is located; INV: whether tumor is invasive; LYP: number of positive lymph nodes; LYR: number of lymph nodes removed; LYS: patient had any positive lymph nodes; MEN: inferred menopausal status; MRI: MRIs within 60 days of surgery; NTN: number of nearby cancerous lymph nodes; PR: progesterone receptor expression; PRP: percent of cell stain pos for PR receptors; P53: whether P53 is mutated; RAC: race; REE: removal of an additional margin of tissue; SID: side of tumor; SIZ: size of tumor in mm; SMO: smoking; STA: composite of size and # positive nodes; SUR: whether residual tumor; TNE: triple negative status in terms of patient being ER, PR, and HER2 negative; TTN: prime tumor stage in TNM system.
Cancers 17 01092 g006
Figure 7. Correlation bar plot and independence plot with interaction visualization for the best DFNN-5Year (a,b), DFNN-10-year (c,d), and DFNN-15-year (e,f) model, each, respectively. AGE: age at diagnosis of the disease; ALC: alcohol usage; DCI: type of ductal carcinoma in situ; ER: estrogen receptor expression; ERP: percent of cell stain pos for ER receptors; ETH: ethnicity; FAM: family history of cancer; GRA: grade of disease; HER: HER2 expression; HI1: tumor histology; HI2: tumor histology subtypes; INL: where invasive tumor is located; INV: whether tumor is invasive; LYP: number of positive lymph nodes; LYR: number of lymph nodes removed; LYS: patient had any positive lymph nodes; MEN: inferred menopausal status; MRI: MRIs within 60 days of surgery; NTN: number of nearby cancerous lymph nodes; PR: progesterone receptor expression; PRP: percent of cell stain pos for PR receptors; P53: whether P53 is mutated; RAC: race; REE: removal of an additional margin of tissue; SID: side of tumor; SIZ: size of tumor in mm; SMO: smoking; STA: composite of size and # positive nodes; SUR: whether residual tumor; TNE: triple negative status in terms of patient being ER, PR, and HER2 negative; TTN: prime tumor stage in TNM system.
Figure 7. Correlation bar plot and independence plot with interaction visualization for the best DFNN-5Year (a,b), DFNN-10-year (c,d), and DFNN-15-year (e,f) model, each, respectively. AGE: age at diagnosis of the disease; ALC: alcohol usage; DCI: type of ductal carcinoma in situ; ER: estrogen receptor expression; ERP: percent of cell stain pos for ER receptors; ETH: ethnicity; FAM: family history of cancer; GRA: grade of disease; HER: HER2 expression; HI1: tumor histology; HI2: tumor histology subtypes; INL: where invasive tumor is located; INV: whether tumor is invasive; LYP: number of positive lymph nodes; LYR: number of lymph nodes removed; LYS: patient had any positive lymph nodes; MEN: inferred menopausal status; MRI: MRIs within 60 days of surgery; NTN: number of nearby cancerous lymph nodes; PR: progesterone receptor expression; PRP: percent of cell stain pos for PR receptors; P53: whether P53 is mutated; RAC: race; REE: removal of an additional margin of tissue; SID: side of tumor; SIZ: size of tumor in mm; SMO: smoking; STA: composite of size and # positive nodes; SUR: whether residual tumor; TNE: triple negative status in terms of patient being ER, PR, and HER2 negative; TTN: prime tumor stage in TNM system.
Cancers 17 01092 g007
Figure 8. The SHAP importance and summary plots concerning important hyperparameters for explaining predicted mean_test_AUCs based on results of DFNN-5-year (a,b), 10-year (c,d), and 15-year (e,f) models, respectively. AGE: age at diagnosis of the disease; ALC: alcohol usage; DCI: type of ductal carcinoma in situ; ER: estrogen receptor expression; ERP: percent of cell stain pos for ER receptors; ETH: ethnicity; FAM: family history of cancer; GRA: grade of disease; HER: HER2 expression; HI1: tumor histology; HI2: tumor histology subtypes; INL: where invasive tumor is located; INV: whether tumor is invasive; LYP: number of positive lymph nodes; LYR: number of lymph nodes removed; LYS: patient had any positive lymph nodes; MEN: inferred menopausal status; MRI: MRIs within 60 days of surgery; NTN: number of nearby cancerous lymph nodes; PR: progesterone receptor expression; PRP: percent of cell stain pos for PR receptors; P53: whether P53 is mutated; RAC: race; REE: removal of an additional margin of tissue; SID: side of tumor; SIZ: size of tumor in mm; SMO: smoking; STA: composite of size and # positive nodes; SUR: whether residual tumor; TNE: triple negative status in terms of patient being ER, PR, and HER2 negative; TTN: prime tumor stage in TNM system.
Figure 8. The SHAP importance and summary plots concerning important hyperparameters for explaining predicted mean_test_AUCs based on results of DFNN-5-year (a,b), 10-year (c,d), and 15-year (e,f) models, respectively. AGE: age at diagnosis of the disease; ALC: alcohol usage; DCI: type of ductal carcinoma in situ; ER: estrogen receptor expression; ERP: percent of cell stain pos for ER receptors; ETH: ethnicity; FAM: family history of cancer; GRA: grade of disease; HER: HER2 expression; HI1: tumor histology; HI2: tumor histology subtypes; INL: where invasive tumor is located; INV: whether tumor is invasive; LYP: number of positive lymph nodes; LYR: number of lymph nodes removed; LYS: patient had any positive lymph nodes; MEN: inferred menopausal status; MRI: MRIs within 60 days of surgery; NTN: number of nearby cancerous lymph nodes; PR: progesterone receptor expression; PRP: percent of cell stain pos for PR receptors; P53: whether P53 is mutated; RAC: race; REE: removal of an additional margin of tissue; SID: side of tumor; SIZ: size of tumor in mm; SMO: smoking; STA: composite of size and # positive nodes; SUR: whether residual tumor; TNE: triple negative status in terms of patient being ER, PR, and HER2 negative; TTN: prime tumor stage in TNM system.
Cancers 17 01092 g008
Table 1. The number of values for each of the adjustable hyperparameters of a DFNN model, used in the example grid search.
Table 1. The number of values for each of the adjustable hyperparameters of a DFNN model, used in the example grid search.
Hyperparameter
Name
Acronym *
NHLNHNAFFIOLRMIDDREBSL1L2
Number of Values42245510411312111010
* NHL: number of hidden layers; NHN: number of hidden nodes; AF: activation function; KI: kernel initializer; O: optimizer; LR: learning rate; M: momentum; ID: iteration-based decay; DR: dropout rate; E: epochs; BS: batch_size.
Table 2. Case and predictor counts of the three LSM_RF datasets.
Table 2. Case and predictor counts of the three LSM_RF datasets.
Total # of Cases# Positive Cases# Negative Cases# of Predictors
LSM-RF-5year4189437375220
LSM-RF-10year1827572125518
LSM-RF-15year75160814317
Table 3. The set of adjustable hyperparameters and their ranges of values considered for our grid searches.
Table 3. The set of adjustable hyperparameters and their ranges of values considered for our grid searches.
HyperparameterDescriptionValues
Number of hidden layersThe depth of a DNN1, 2, 3, 4
Number of hidden nodes in a hidden layerNumber of neurons in a hidden layer1–1005
Activation functionIt determines the value to be passed to the next node based on the value of the current nodeRelu, Sigmoid, Softmax, Tanh
Kernel initializerAssigning initial values to the internal model parameters Constant, Glorot_normal, Glorot_uniform, He_normal, He_uniform
OptimizerOptimizes internal model parameters towards minimizing the lossSGD, Adam, Adagrad, Nadam, Adamax
Learning rateUsed by both SGD and Adagrad0.001~0.3
MomentumSmooths out the curve of gradients by moving average. Used by SGD.0~0.9
Iteration-based decayIteration-based decay; updating learning rate by a decreasing factor in each epoch 0~0.1
Dropout rateManage overfitting and training time by randomly selecting nodes to ignore0~0.5
EpochsNumber of times model is trained by each of the training set samples exactly one5~2000
Batch_sizeUnit number of samples fed to the optimizer before updating weights 1 to the # of datapoints in a dataset
L1 Sparsity regularization; 0~0.03
L2 Weight decay regularization: it penalizes large weights to adjust the weight updating step0~0.2
Table 4. Comparison of group prediction performance in average mean_test_AUCs for the DFNN-5-year models.
Table 4. Comparison of group prediction performance in average mean_test_AUCs for the DFNN-5-year models.
GroupResultsStage1Stage2Stage3-c1Stage3-c2Stage3-c3Stage3-c4Stage3-c5Stage3-c6
Top1best0.753010.748120.750240.750660.754180.753860.752500.75825
Top5Avg0.752250.747250.748940.749800.753440.752600.752190.75532
Top 10Avg0.751440.746820.746580.749230.752740.751790.751940.75435
Top 50Avg0.748520.74560.746580.747670.751140.750310.750910.75206
Top 100Avg0.746920.74490.745840.747040.750440.749560.750380.75122
AllAvg0.664560.68390.720960.735050.736570.720520.719970.63960
Table 5. Comparison of group prediction performance in average mean_test_AUCs for the DFNN-10-year models.
Table 5. Comparison of group prediction performance in average mean_test_AUCs for the DFNN-10-year models.
GroupResultsStage1Stage2Stage3-c1Stage3-c2Stage3-c3Stage3-c4Stage3-c5Stage3-c6
Top1best0.777890.782800.788020.790230.791520.788850.780590.78564
Top5Avg0.775930.781410.786960.788970.790300.787620.779360.78272
Top10Avg0.775110.780900.786260.788200.789800.787030.778890.78189
Top50Avg0.773480.779420.784750.786440.788650.785530.777310.77992
Top100Avg0.772570.778660.784020.785510.787970.784830.776520.77900
AllAvg0.713440.733600.761240.759010.767020.761940.743360.68075
Table 6. Comparison of group prediction performance in average mean_test_AUCs for the DFNN-15-year models.
Table 6. Comparison of group prediction performance in average mean_test_AUCs for the DFNN-15-year models.
GroupResultsStage1Stage2Stage3-c1Stage3-c2Stage3-c3Stage3-c4Stage3-c5Stage3-c6
Top1best0.865020.865480.880230.874630.873590.875510.872550.86657
Top5Avg0.862320.864030.876320.873970.873370.874760.870950.86578
Top 10Avg0.861050.862940.874260.873530.872800.874270.870280.86510
Top 50Avg0.858060.860690.870730.871490.870820.872810.868720.86271
Top 100Avg0.856780.859660.869500.870550.869540.871990.867890.86170
AllAvg0.765120.788530.831670.845350.833990.837820.837310.75042
Table 7. Mid-point group results of grid searches.
Table 7. Mid-point group results of grid searches.
Stage1Stage2Stage3-c1Stage3-c2Stage3-c3Stage3-c4Stage3-c5Stage3-c6
5 YBest-mean0.753010.748120.750240.750660.754180.753860.752500.75825
Mid-point0.626500.624060.625120.625330.627090.626930.626250.62913
CHS39,112510,494203,991101,250202,500404,726408,62543,035
TNS52,042582,471202,500101,250202,500405,000409,05070,568
CHS/TNS0.751550.876430.999461.01.00.999320.998960.60984
Avg-mean0.711160.702780.721020.735050.736570.720590.720100.70415
10 YBest_mean0.777890.782800.788020.790230.791520.788850.780590.78564
Mid_point0.638950.641400.644010.645120.645760.644420.640290.64282
CHS48,517635,770202,50060,750202,495405,000404,99483,896
TNS52,416638,244202,50060,750202,500405,000405,000105,500
CHS/TNS0.925610.996121.01.00.999981.00.999990.79522
Avg_mean0.721160.734000.761240.759010.767020.761940.743370.72642
15 YBest_mean0.865020.865480.880230.874630.873590.875510.872550.86657
Mid_point0.682510.682740.690110.687310.686790.687760.686280.68328
CHS41,065952,677337,492162,000270,0002,058,750405,00073,007
TNS52,0901,048,575337,500162,000270,0002,058,750405,00093,000
CHS/TNS0.788350.908540.999981.01.01.01.00.78502
Avg_mean0.8006150.802890.831670.845350.833990.837820.837310.81619
Best-mean: the highest mean_test_AUC of all corresponding models; mid-point: the average of 0.5 and the best-mean AUC; CHS: the count of hyperparameter settings at which the mean_test_AUCs are at least as high as the corresponding mid-point AUC; TNS: the total number of hyperparameter settings tested during the corresponding cycle of grid searches; CHS/TNS: the ratio of CHS over TNS; avg-mean: the average mean_test_AUC of all models no worse than the mid-point AUC; Y: year.
Table 8. Running time and count of hyperparameter settings concerning grid searches.
Table 8. Running time and count of hyperparameter settings concerning grid searches.
Stage5-Year10-Year15-Year
RTPSTNSTRTRTPSTNSTRTRTPSTNSTRT
Stage142.4252,042613.2326.5752,416386.8720.0652,090290.25
Stage25.65582,471914.723.36638,244594.862.491,048,575725.06
Stage3-c110.65202,500599.013.55202,500199.422.43337,500227.39
Stage3-c211.97101,250336.545.4960,75092.661.63162,00073.17
Stage3-c311.11202,500624.886.50202,500365.771.25270,00093.60
Stage3-c425.46405,0002863.737.66405,000861.751.892,058,7501080.01
Stage3-c566.80409,0507590.0729.44405,0003311.7612.22398,2521352.36
Stage3-c6116.4070,5682281.7751.05105,5001495.9127.3693,000706.71
RTPS: running time per hyperparameter setting in seconds; TNS: total number of hyperparameter settings; TRT: total running time in hours.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jiang, X.; Zhou, Y.; Xu, C.; Brufsky, A.; Wells, A. Deep Learning: A Heuristic Three-Stage Mechanism for Grid Searches to Optimize the Future Risk Prediction of Breast Cancer Metastasis Using EHR-Based Clinical Data. Cancers 2025, 17, 1092. https://doi.org/10.3390/cancers17071092

AMA Style

Jiang X, Zhou Y, Xu C, Brufsky A, Wells A. Deep Learning: A Heuristic Three-Stage Mechanism for Grid Searches to Optimize the Future Risk Prediction of Breast Cancer Metastasis Using EHR-Based Clinical Data. Cancers. 2025; 17(7):1092. https://doi.org/10.3390/cancers17071092

Chicago/Turabian Style

Jiang, Xia, Yijun Zhou, Chuhan Xu, Adam Brufsky, and Alan Wells. 2025. "Deep Learning: A Heuristic Three-Stage Mechanism for Grid Searches to Optimize the Future Risk Prediction of Breast Cancer Metastasis Using EHR-Based Clinical Data" Cancers 17, no. 7: 1092. https://doi.org/10.3390/cancers17071092

APA Style

Jiang, X., Zhou, Y., Xu, C., Brufsky, A., & Wells, A. (2025). Deep Learning: A Heuristic Three-Stage Mechanism for Grid Searches to Optimize the Future Risk Prediction of Breast Cancer Metastasis Using EHR-Based Clinical Data. Cancers, 17(7), 1092. https://doi.org/10.3390/cancers17071092

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop