Next Article in Journal
Artificial Intelligence for Personalized Genetics and New Drug Development: Benefits and Cautions
Next Article in Special Issue
Robust Removal of Slow Artifactual Dynamics Induced by Deep Brain Stimulation in Local Field Potential Recordings Using SVD-Based Adaptive Filtering
Previous Article in Journal
The COVID-19 Pandemic: How Technology Is Reshaping Public Health and Medicine
Previous Article in Special Issue
Epileptic Tissue Localization through Skewness-Based Functional Connectivity in the High-Frequency Band of Intracranial EEG
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Explainable Machine-Learning Model for Compensatory Reserve Measurement: Methods for Feature Selection and the Effects of Subject Variability

by
Carlos N. Bedolla
1,†,
Jose M. Gonzalez
1,†,
Saul J. Vega
1,
Víctor A. Convertino
1,2,3,4 and
Eric J. Snider
1,*
1
U.S. Army Institute of Surgical Research, JBSA Fort Sam Houston, San Antonio, TX 78234, USA
2
Department of Medicine, Uniformed Services University, Bethesda, MD 20814, USA
3
Department of Emergency Medicine, University of Texas Health, San Antonio, TX 78229, USA
4
Department of Biomedical Engineering, University of Texas Health, San Antonio, TX 78249, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Bioengineering 2023, 10(5), 612; https://doi.org/10.3390/bioengineering10050612
Submission received: 20 April 2023 / Revised: 8 May 2023 / Accepted: 10 May 2023 / Published: 19 May 2023
(This article belongs to the Special Issue Advances of Biomedical Signal Processing)

Abstract

:
Tracking vital signs accurately is critical for triaging a patient and ensuring timely therapeutic intervention. The patient’s status is often clouded by compensatory mechanisms that can mask injury severity. The compensatory reserve measurement (CRM) is a triaging tool derived from an arterial waveform that has been shown to allow for earlier detection of hemorrhagic shock. However, the deep-learning artificial neural networks developed for its estimation do not explain how specific arterial waveform elements lead to predicting CRM due to the large number of parameters needed to tune these models. Alternatively, we investigate how classical machine-learning models driven by specific features extracted from the arterial waveform can be used to estimate CRM. More than 50 features were extracted from human arterial blood pressure data sets collected during simulated hypovolemic shock resulting from exposure to progressive levels of lower body negative pressure. A bagged decision tree design using the ten most significant features was selected as optimal for CRM estimation. This resulted in an average root mean squared error in all test data of 0.171, similar to the error for a deep-learning CRM algorithm at 0.159. By separating the dataset into sub-groups based on the severity of simulated hypovolemic shock withstood, large subject variability was observed, and the key features identified for these sub-groups differed. This methodology could allow for the identification of unique features and machine-learning models to differentiate individuals with good compensatory mechanisms against hypovolemia from those that might be poor compensators, leading to improved triage of trauma patients and ultimately enhancing military and emergency medicine.

1. Introduction

Monitoring vital signs, often via means of physiological sensors, is critical throughout patient care, but there remain shortcomings with triage decision support. Accurate patient status assessment can only begin after sensor placement, which for invasive line placement can preclude data collection in the pre-hospital setting. This challenge extends particularly to military settings, where remote, austere environments with less skilled medical personnel can further compromise efficacious sensor placement [1,2]. An ongoing research and development goal is to develop wearable healthcare devices capable of improving these shortcomings by allowing watches, adhesive e-tattoos, or ingestible sensors to record data at the point of injury and continue through higher echelons of care [3,4,5].
Another shortcoming of current physiological monitors is the inability to provide the caregiver with the most complete picture of a patient’s clinical status which, in turn, can lead to inadequate, inappropriate, or ineffective medical intervention. This problem can extend to wearable healthcare devices as some physiological parameters can be influenced by fear or pain, as well as the high variability of other factors, such as degrees of injury severity and choice of interventions. The complexity of physiology of a trauma patient is further complicated by the high inter- and intra-patient variability with compensatory mechanisms that can mask shock symptoms until life-saving interventions are less effective [6]. Physiological compensation mechanisms protect oxygen delivery to vital organs, primarily by redistributing blood flow to the brain and heart via sympathetic vasoconstriction and increased extraction of oxygen from blood [7]. To address the development of monitoring capabilities for accurate assessment of the complex nature of physiological compensation, recent efforts have focused on the application of artificial intelligence (AI) or machine-learning (ML) algorithms to better interpret sensor data. This has been demonstrated for estimating core body temperature [8], blood pressure [9,10,11,12], and compensatory reserve [6,13,14]. While these methods allow for tracking and estimating metrics and values that are otherwise not possible from raw data streams, oftentimes decisions and methods employed specifically via AI models cannot be easily explained due to the large number of parameters (sometimes millions) that are required by these algorithms [15,16]. This is a critical challenge from a regulatory perspective but also for medical adoption of the technology, as recent studies have shown that explainable AI models can improve trust and reliance on AI by medical personnel [17].
In an effort to avoid the creation of a “black box” AI algorithm, we assessed extracting measurable features of arterial waveforms to calculate a triaging metric known as the Compensatory Reserve Measurement (CRM), which aims to represent the biophysics that underly cardiac and peripheral vascular mechanisms of compensation for clinical conditions of central hypovolemia. Signal processing with the extraction of specific waveform characteristics has been used to identify individual parameters, such as traditional systolic and diastolic pressure, as well as millions of complex combinatorial features [18]. The feature extraction approach used in this present study allows for a more explainable algorithm, as the key waveform features that are responsible for tracking CRM can be specifically identified. Therefore, through extracting waveform features from arterial waveform data sets, we hypothesized that the compensatory reserve measurement can be tracked via a classical machine-learning model just as well as more complex “black box” AI models. Furthermore, feature extraction methods have the potential to reveal sub-populations as determined by heterogeneity in the identified waveform features needed to predict CRM.

1.1. Overview of Feature Extraction

Feature extraction is the process of identifying specific repeated characteristics or patterns that are observed in a signal. Features can be absolute magnitudes, relative time, or percentage differences between features, algebraic combinations of multiple features, as well as many other approaches. Extracted features can lessen the data input into an algorithm as a single feature can represent thousands of raw waveform data points of the original signal [19]. These techniques have been used extensively to develop machine-learning algorithms capable of analyzing a variety of biomedical signals. For example, a previous study used wavelet-based feature extraction methods in developing algorithms that can differentiate between a subject’s “resting” and “thinking” status from electroencephalography signals with 98% accuracy [20]. In another study, researchers extracted 3022 individual features from arterial waveform data and employed combinatory techniques to produce more than 2 million features to successfully develop an ML model for detecting hypotension during intensive care and surgical procedures [18]. Finally, a previous study developed linear regression models using nine features extracted from arterial waveform data to estimate CRM. In that study, all of these features were not required, as ultimately, a single feature was able to linearly correlate to CRM [21]. However, this study used a limited dataset of 13 human subjects exposed to low levels of lower body negative pressure, resulting in less subject variability in the data set.

1.2. Compensatory Reserve Measurement

The Compensatory Reserve Measurement is a triaging metric previously developed at the US Army Institute of Surgical Research that estimates the level of physiological decompensation in an individual experiencing hypovolemia [22]. This compensatory index provides a value on a scale of 0 to 1 (or 0% to 100%) that reflects the overall physiological compensation status, where 0 represents the threshold of decompensated shock. The idea behind this measurement is to provide a capability for a clinical caregiver to predict with sufficient time the imminent onset of overt shock in a patient who might not otherwise show signs of hemodynamic decompensation based on standard vital signs. The CRM was previously defined by Moulton et al. as follows [13]:
C R M = 1 B L V B L V H D D ,
where BLV is the volume of blood a patient has lost, and BLVHDD is the volume of blood loss required for the same patient to reach hemodynamic decompensation (HDD), i.e., to exhaust the body’s compensatory reserve. Since neither of these volumes can be known with certainty in real-world scenarios, mathematical algorithms have been developed instead to estimate a CRM value from features found in the patient’s arterial waveform [13,22].
The development of these algorithms requires physiological data obtained during hemorrhage to the point of HDD. Since experimenting on a human subject by hemorrhaging them to HDD would be dangerous and unethical, a more practical method has been used since the 1960s to safely simulate the experience of hemorrhaging large volumes of blood. In this model, the lower section of the body of a healthy research subject is placed inside a negative pressure chamber sealed around the waist. By applying increasing levels of lower-body negative pressure (LBNP) to the subject, their blood volume is progressively redistributed to the lower portions of the body, which results in central hypovolemia, thus simulating the physiological effects of hemorrhage. Using this technique, the subject can be brought to the point of hemodynamic instability safely and in a controlled fashion [23].
Using LBNP as a model for simulated blood loss, the CRM can then be estimated by the following formula [13]:
C R M = 1 L B N P L B N P H D D ,
where LBNP is the level of negative pressure a subject is experiencing at any given time, and LBNPHDD is the level of negative pressure at which the same subject reaches HDD.
Algorithms developed to estimate a CRM value from LBNP data attempt to match features found on a relevant physiological waveform, such as pulse oximetry and/or arterial pressure, to a subject’s level of HDD [13,14]. In this technique, a reference CRM of 1.0 is assigned to the baseline LBNP level (i.e., zero negative pressure), and a CRM of 0.0 is assigned to the LBNP level at which a subject experiences HDD. A machine-learning model is then trained to match patterns and features found in the physiological waveform to each LBNP level found between those two reference points. The goal is for the resulting algorithm to be able to predict the instantaneous relative LBNP level of an individual and thus estimate their level of HDD on a scale from 0.0 to 1.0 based on a sample of a relevant physiological waveform, without any additional knowledge of that individual’s physiological status.

2. Materials and Methods

2.1. Retrospective Analysis of Lower Body Negative Pressure Datasets

LBNP datasets were previously generated during human research studies performed at the US Army Institute of Surgical Research [13]. These prior studies were approved by the required Institutional Review Board committees and followed the guidelines of the Declaration of Helsinki.
Briefly, during an LBNP experimental session, the lower portion of the body of a subject laying in the supine position was placed inside a sealed vacuum chamber. During the protocol, negative pressure was safely and progressively applied to the subject’s lower body from baseline pressure of 0 mmHg down to a minimum of −100 mmHg, in 9 incremental steps that lasted 5 min each (Figure 1A). Each experiment ended when the pressure reached the aforementioned minimum limit, or when the subject reached their own individual LBNPHDD, at which point the chamber’s pressure was returned to baseline.
Throughout the LBNP experiments photoplethysmography (PPG) waveform data sampled at 500 Hz were recorded using a Finometer PRO blood pressure monitor (Finapres Medical Systems, Amsterdam, The Netherlands) from 218 subjects. The data were further processed and analyzed to develop new machine-learning models to estimate CRM; these models will be referred to as CRM-ML throughout.

2.2. Pre-Processing Datasets

De-identified data from the LBNP subjects were used for analysis in this study and were the foundation for creating an ML algorithm for the prediction of CRM at different hypovolemic states. Subjects reached varying final pressure levels of LBNP exposure, so each dataset was classified by its maximum LBNP step reached prior to HDD to determine the sample distribution (Figure 1B). Due to these varying levels of LBNP steps reached, data were equally sampled from each LBNPHDD step to equally account for differences between each LBNPHDD sub-group in the CRM-ML algorithm. However, due to limited subject data in the distribution extremes, datasets for LBNPHDD at steps 2, 3, and 9 were removed from consideration. Equal number of randomly sampled 4, 5, 6, 7, and 8 LBNPHDD step subjects were then used. While a total of 218 subjects were available in the dataset, step group 4 had only 12 subjects available, limiting the number of subjects for all other steps to maintain an 8:4 training-to-blind testing ratio. Of these selected subjects, 62% were male and 38% were female, with heights ranging between 152 and 193 cm, weights ranging between 44 and 117 kg, and ages ranging between 18 and 54 years of age. A total of 8 subjects were selected from each LBNPHDD step group for the training of the CRM-ML algorithm (40 total subjects), and 4 subjects for each LBNPHDD were held out for blind testing (20 total subjects).

2.3. Feature Extraction Methodology

Features were extracted from the arterial pressure vs. time LBNP datasets using MATLAB (v2022b) MathWorks, Natick, MA, USA). First, the wavelength was filtered using a finite impulse response (FIR) window lowpass filter [24]. The filtered data were then analyzed to calculate the first and second derivatives, which were then used to identify the peaks and troughs of the waveform that could assist in locating important landmarks in the signal (Figure 2). For simplicity, the start of the systolic phase of the waveform, or pulse foot, was identified as point “A” and the systolic peak as point “C”. A halfway point between “A” and “C” was identified as “B” to represent the “half-rise” [25]. Lastly, the inflection point after the systolic peak was labeled as point “D”. Inflection points were not always identified in some segments of the arterial waveforms. For consistency, if inflection points were not identified, these arterial waveform segments were excluded from the analysis. In other waveform segments, two or more inflection points were identified after the systolic peak. In these situations, only the first inflection point was tracked for consistency.
These points were then used to calculate additional waveform features selected from previous research efforts by Gupta et al. [21] and Hatib et al. [18]. A total of 54 features were extracted from each waveform as summarized in Table 1. More detailed definitions for these features, including their equations are detailed in Supplementary Table S1. These features were collected from each subject dataset for training an ML model and for blind test data to later assess model performance.

2.4. Machine-Learning Models

The Regression Learner Toolbox in MATLAB was used for evaluating a wide range of machine-learning models for estimating CRM. Training was performed on 40 total LBNP subjects (n = 8 for each LBNPHDD step). For feature selection, features were ranked using the minimal-redundancy–maximal-relevance (MRMR) criterion, which balances selection of a feature between maximum correlation to the signal and least correlation to other features [26]. To evaluate different ML models, Linear Regression, Fine Tree, Medium Tree, Coarse Tree, Ensemble Bagged Trees, and Ensemble Boosted Trees models were constructed using MATLAB’s Regression Learner Toolbox. The two ensemble models have an eight-leaf size with 30 learners. To find the optimal feature selection, all models were trained with the top 15, 10, 5, and 1 feature(s). Blind testing was conducted using 4 subjects, separate from the training subjects, from each LBNP step group (20 subjects total). To identify differences among subject subpopulations, a second approach to model training was used, in which only data from each single LBNP step group were used to train a single CRM-ML model, i.e., one model per step group. For these individual models, the same methods for training the general model were used, except only the 8 subjects for that specific LBNPHDD step group were used during training. Testing was still conducted using all the datasets.
In order to compare and down-select the optimal machine-learning model and number of features, the training and testing root-mean-squared error (RMSE) and coefficient of determination (R2) values were computed. It is important to note that the R2 and RMSE values were collected from the Regression Learner Toolbox. The methodology for calculating R2 and RMSE values in this toolbox was conducted via fitting CRM predictions vs. CRM calculated (Equation (2)) values to a perfect regression ( y = x ), as opposed to the more conventional linear regression ( y = m x + b ). Thus, throughout, these values obtained from the toolbox will be referred to as perfect R2 (or P-R2) and perfect RMSE (or P-RMSE) to differentiate them from the traditional regression (R2) and RMSE values for the two-parameter regression fit.

2.5. Deep-Learning Model

To compare the performance of the new machine-learning models developed in this study to more complex algorithms, we estimated CRM values on the same LBNP datasets using a deep-learning model (CRM-DL) with a one-dimensional Convolutional Neural Network [14]. For its development, arterial waveform data sampled at 100 Hz was normalized to a 0–1 range and then segmented into 20 s long samples for model regression learning. The model then learned to perform a linear regression on each data segment to estimate a corresponding CRM.
To evaluate the CRM-DL model, we again used the same LBNP subject datasets used for testing the CRM-ML models (n = 4 per each LBNPHDD step group). The result of this process was a series of CRM values sampled at 10 s intervals that spanned the full length of the LBNP sessions for each experimental subject. Both conventional and perfect linear regression methods were used to calculate RMSE and R2 to evaluate the performance of CRM-DL in predicting CRM against the calculated CRM value (Equation (2)).
The likelihood that during our evaluation of its performance the CRM-DL may have processed instances already learned is noteworthy given that the model was originally trained on the same LBNP datasets. This would give the CRM-DL model an unfair advantage when comparing its performance against that of our new classical ML models. However, the exact number of records that overlap the DL training set and our evaluation, if any, is unknown.

3. Results

3.1. Machine-Learning Model and Number of Features Selection

Different machine-learning model types were trained to predict CRM using different numbers of extracted arterial waveform features (Table 2). The ensemble bagged trees model performed the best, with lower P-RMSE and highest P-R2 in training, regardless of the number of features used, while linear regression was consistently one of the lowest-performing models for both P-RMSE and P-R2. During blind testing, the simpler decision trees performed worse than other ML models, likely indicating an overfitting of training data. Training performance was worsened when only a single feature was considered by the model, but blind testing performance remained similar for most CRM-ML models. During blind testing, the ensemble bagged tree and boosted tree ML model performed well. In general, P-RMSE and P-R2 results showed that a higher number of features in ML models were associated with better model performance, especially for training performance. However, the improvement in performance between 15 and 10 features was relatively small, so it was decided to continue testing with the ensemble bagged trees ML model using 10 features.

3.2. Effect of LBNP Final Step Reached on Model Performance

Using the 10-feature ensemble bagged-tree ML model, we examined the effect of subjects’ LBNP performances (i.e., tolerance) on model accuracy. Representative plots of theoretical CRM (Equation (2)) compared to the predicted CRM-ML are displayed in Figure 3 for the different LBNPHDD step groups. Both perfect and conventional linear regression are highlighted for each. Predictions of CRM at specific central blood volume levels (i.e., LBNP levels) tracked the general theoretical CRM trend but demonstrated a bias toward certain steps in the LBNP process and subgroups. In general, CRM-ML predictions were more accurate for conventional regression approaches compared to perfect regression.
To highlight the difference in model performance for the different subgroups, R2 and RMSE were calculated for both regression methods for the blind testing at each step for the all-step training model (n = 4 subjects each). The average results are summarized in Table 3. Overall, the testing on the 6-step subgroup resulted in higher coefficients of determination and lower root mean square errors, using both perfect and standard linear regression methodologies, while the tests on the 4-step subgroup resulted in worse scores on both accounts.

3.3. Differences in Features for Separate Models

To further highlight differences in subject subgroups based on pressure step at LBNPHDD, five additional CRM-ML models were developed, each one trained on a single sub-group, i.e., one model for each of the 4, 5, 6, 7, and 8 LBNPHDD subjects. This training approach resulted in ML models with different extracted features being selected as “top 10”. This is detailed in Table 4, where the top ten features, ranked using MRMR, are shown for the different CRM-ML models. The movement of key features in ranking for each single-step trained ML model is evident compared to the all-step CRM-ML model. Certain features, such as HRIP, remained among the top 5 for all individual models, while others traveled significantly, such as sys_rise_area_norm, compared to the model trained with all steps. We postulate that features that persistently ranked among the top ten used in training the models hold significant weight in the prediction of a CRM, while some features that only appear at certain step models could be used as a differentiator or predictor of the level of LBNP exposure at which an individual may decompensate.
Each single-step trained ML model was blind tested against subjects from within its own LBNP group as well as from other step subgroups. For example, the 4-step training model (i.e., the ML model trained only on 4-step subjects) was blind tested on the 4-, 5-, 6-, 7-, and 8-step test data. This was repeated for all models. P-RMSE and P-R2 results were plotted in a heat map for comparison purposes, as shown in Figure 4.
The training model with the highest average P-R2 (0.81) and the lowest average P-RMSE value (0.13) was the model trained on the 5-step LBNPHDD datasets. The blind test data that had the highest average P-R2 (value range = 0.83–0.92) and the lowest average P-RMSE (value range = 0.09 to 0.13) was the test data at step 6. In general, all the individualized models performed better against their specific datasets at that step.

3.4. Deep-Learning Results

CRM-ML model results were compared against the CRM-DL model; the all-step trained CRM-ML model was used for comparison as opposed to the subject-specific models. Calculated “true” CRM values and the predictions via both the CRM-DL and the CRM-ML models were plotted against each other to visually assess the differences in their performances. In general, the CRM-ML model had higher accuracy for higher CRM values than the CRM-DL model, while both reach similar values of CRM in the lower bound. The ML model also tended to overshoot the theoretical CRM, while the DL model tended to undershoot it. The plots in Figure 5 show representative blind testing at different steps of both the DL and ML models.
Both the perfect regression and conventional linear regression were compared using both R2 and RMSE metrics, as shown in Figure 6. Overall, the CRM-ML model had an average P-R2 value higher than CRM-DL (CRM-ML = 0.70 vs. CRM-DL = 0.50), but relatively same P-RMSE values (CRM-ML = 0.171 vs. CRM-DL = 0.165). With conventional linear regression, both the CRM-ML model and CRM-DL had similar average R2 (CRM-ML = 0.85 vs. CRM-DL = 0.88) and RMSE values (CRM-ML = 0.168 vs. CRM-DL = 0.165). Performance differences for the different LBNPHDD steps varied with the 4-step data having the worse P-R2 metrics and the 6-step having the strongest.

4. Discussion

Advanced monitoring is critical for conducting accurate triage of a patient presenting with significant hemorrhage, especially in scenarios where invasive sensor placement may not be possible, such as in pre-hospital and military settings. To simplify regulatory hurdles and improve medical adoption, algorithms for this purpose should be as simple as possible to meet the needs of emergency medical personnel. In this study, we focused on compensatory reserve measurement, a metric derived from arterial waveform data for accurately tracking physiological decompensation and early detection of hemorrhagic shock. To develop a simpler model as an alternative to the current deep-learning models employed in CRM estimation, regression-based machine-learning models were developed using extracted features from the arterial waveform. Simpler ML methods such as this are more explainable, as the individual waveform features critical to calculating CRM are identified and may further understand the underlying physiological compensatory mechanisms.
Overall, the feature extraction methodology paired with classical ML models allowed for accurately tracking CRM. This approach required extracting ten features from each arterial waveform segment which was performed as a pre-processing setup prior to ML model implementation. However, this feature extraction signal processing could be easily performed during data collection with minimal computing requirements to enable real-time reporting of CRM values. Compared to deep-learning models that require processing millions of parameters, the CRM-ML models developed here should perform more computationally efficiently, which may be critical in a resource-limited environment. Feature extraction methodologies have been recently used by Gupta et al. to develop ML models that estimate CRM [21]. They used a single parameter model, time from half-rise to dicrotic notch (HRDN), which resulted in R2 values of approximately 0.67, lower than our average R2 value of 0.85. Different LBNP data sets were used by Gupta et al., with different simulated hemorrhage magnitudes and a lower subject number of only n = 13 subjects. The differences in data sets, reduced number of subjects, and the increase from a single extracted feature to 10 features in CRM-ML likely account for our increased R2 performance compared to Gupta. It is worth noting that the time from half-rise to inflection point (HRIP) feature used in this study was the top correlative feature in the CRM-ML model and is analogous to the top-performing feature (HRDN) identified by Gupta.
It is important to note that bagged decision tree models have advantages and disadvantages compared to other ML/DL approaches. The crucial advantage of bagged decision tree models is that multiple models are bootstrapped so that not one model drives predictions, allowing for a more robust ensemble. This removes bias since there are multiple models that can agree or disagree creating a more generalized predictive algorithm [27,28]. However, bagging multiple models into a more robust ensemble can lead to losing some interpretability compared to a single model. This is exacerbated in training situations where there are rare, underrepresented events that may be lost in an ensemble, generalized model approach. However, we think this is less of a concern for a use case, such as measuring compensatory status, which needs to be applicable to a large subject population. Another possible disadvantage, depending on the size of the training data, is that training ensemble bagged trees may require a high computational burden [27,28,29]. However, the computational time required to calculate a single prediction is much lower than that consumed for model training and is still much less intense compared to most deep-learning models.
Next, CRM-ML performance was contrasted against the CRM-DL model for predicting CRM [14]. Perfect and conventional regression RMSE results were very comparable between the CRM-ML and CRM-DL models, highlighting that the simpler approach used here may be sufficient for tracking CRM. However, larger differences were more evident between the two models when evaluating P-R2 values (0.70 for CRM-ML vs. 0.50 for CRM-DL). That said, the CRM-DL fit the data worse when subjects did not reach lower pressure steps (LBNPHDD step 4 P-R2 = 0.37 vs. step 8 P-R2 = 0.56), just like CRM-ML struggled in these same situations (LBNPHDD step 4 P-R2 = 0.52 vs. step 8 P-R2 = 0.74). Of note, when the goodness of fit was evaluated via standard linear regression, the CRM-DL model performed more comparable to CRM-ML (DL R2 = 0.88 vs. ML R2 = 0.85). This suggests that the deep-learning model can be adjusted to better match calculated CRM, but it is not predicting at a 1:1 correlation without a slope, intercept adjustment. One major limitation of this part of this study, however, is that the CRM-DL model was trained previously [14] using all of the LBNP step data, meaning the test data used in this study was not completely blind to the CRM-DL algorithm.
An unexpected trend was detected based on the maximum hemodynamic decompensation threshold reached by the subject, presented as the final LBNP pressure step reached. The generalized CRM-ML models struggled to perform accurately across each LBNP step sub-population, so models were trained for each individual LBNP step sub-population. In doing so, the top 10 features selected for each CRM-ML model differed greatly across the individual LBNP step models. While features such as HRIP were prevalent in the top features for all CRM-ML LBNPHDD models, others were specific to certain models, as highlighted in Table 3. Subpopulations have been previously reported from these LBNP datasets, referred to as high- and low-tolerant subjects based on exceeding a certain LBNP exposure level [7,30]. We chose to not focus on these two sub-groups but instead break the data into more granular subsets to highlight potential differences across subgroups. This may suggest there are more subgroups in these datasets. Furthermore, the variations in features extracted via the individualized models may identify physiologically pertinent differences that can be used to understand variability in the patient’s compensatory reserve capacity. For instance, if these different CRM-ML models are utilized to predict CRM in real time, tracking which model best fits the patient may allow for the prediction of the compensatory capacity before hemodynamic decompensation is reached. More analysis and follow-up studies are needed for further teasing out these differences, including a more thorough understanding of the underlying physiology behind the extracted features.
The next steps for this work will be three-fold. First, the existing CRM-ML models will be used with additional data sets for measuring CRM compared to the CRM-DL model to further assess model performance. This includes datasets for patients undergoing surgery and animal data with deeper levels of hemorrhagic shock to evaluate performance outside the scope of current CRM-DL training methodologies. Second, the CRM-ML model will be further improved by evaluating additional extracted features, combinations of features, and trends in features as inputs to the model. This may reveal additional, arterial waveform characteristics distinctive of each subject population, and subgroups, and/or further improve model accuracy. Lastly, this feature extraction methodology will be applied to other medical applications, including tracking CRM with novel wearable sensors, and identifying other metrics for tracking shock after decompensation and as inputs for guiding hemorrhage fluid resuscitation.

5. Conclusions

In conclusion, algorithms for advanced monitoring have the potential to improve medical care at the point of injury by enabling medical providers to have more information to diagnose and provide treatment when it is more effective. Here, we focused on feature extraction methodologies for tracking CRM using 10 waveform features as opposed to millions of tuned AI model parameters. Overall, model performance was similar to a more complex deep-learning model, and by tracking extracted features, differences were identified in subgroups in the data sets. Overall, this machine-learning approach can create a more explainable model, a critical step towards the adoption of advanced monitoring technologies in emergency and military medicine.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/bioengineering10050612/s1, Supplementary Table S1: Summary of all 54 extracted features with their abbreviated name and description of their calculation. Features calculations were based on [18,21].

Author Contributions

Conceptualization, E.J.S. and S.J.V.; original LBNP data collection, V.A.C.; methodology, C.N.B., J.M.G. and S.J.V.; data curation, C.N.B., J.M.G. and S.J.V.; data analysis, C.N.B. and J.M.G.; writing—original draft preparation, C.N.B., J.M.G., S.J.V., V.A.C. and E.J.S.; writing—review and editing, C.N.B., J.M.G., S.J.V., V.A.C. and E.J.S.; visualization, C.N.B. and J.M.G.; supervision, E.J.S.; funding acquisition, E.J.S. and V.A.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the US Army Medical Research and Development Command (IS220008, E.J.S.) and the Congressionally Directed Medical Research Program (DM180240, V.A.C.). This project was supported in part by an appointment to the Science Education Programs at the National Institutes of Health (NIH), administered by ORAU through the US Department of Energy Oak Ridge Institute for Science and Education (C.N.B. and J.M.G.).

Institutional Review Board Statement

The data analyzed in this research effort was from a prior research study which was conducted in accordance with the Declaration of Helsinki, and all experimental procedures were conducted in accordance with protocols reviewed and approved by the Brooke Army Medical Institutional Review Board (Protocol C.2004.161, Approval date: 7 July 2004).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the prior study after receiving a verbal and written briefing of the experimental procedures.

Data Availability Statement

The data presented in this study are not publicly available because they have been collected and maintained in a government-controlled database that is located at the US Army Institute of Surgical Research. As such, these data can be made available through the development of a Cooperative Research & Development Agreement (CRADA) with the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

DoD Disclaimer

The views expressed in this article are those of the authors and do not reflect the official policy or position of the US Army Medical Department, Department of the Army, DoD, or the US Government.

References

  1. Townsend, S.; Lasher, W. The U.S. Army in Multi-Domain Operations 2028; U.S. Army: Arlington, VA, USA, 2018. [Google Scholar]
  2. Carius, B.M.; Naylor, J.F.; April, M.D.; Fisher, A.D.; Hudson, I.L.; Stednick, P.J.; Maddry, J.K.; Weitzel, E.K.; Convertino, V.A.; Schauer, S.G. Battlefield Vital Sign Monitoring in Role 1 Military Treatment Facilities: A Thematic Analysis of After-Action Reviews from the Prehospital Trauma Registry. Mil. Med. 2020, 187, e28–e33. [Google Scholar] [CrossRef]
  3. Ates, H.C.; Nguyen, P.Q.; Gonzalez-Macia, L.; Morales-Narváez, E.; Güder, F.; Collins, J.J.; Dincer, C. End-to-End Design of Wearable Sensors. Nat. Rev. Mater. 2022, 7, 887–907. [Google Scholar] [CrossRef] [PubMed]
  4. Castaneda, D.; Esparza, A.; Ghamari, M.; Soltanpur, C.; Nazeran, H. A Review on Wearable Photoplethysmography Sensors and Their Potential Future Applications in Health Care. Int. J. Biosens. Bioelectron. 2018, 4, 195–202. [Google Scholar] [CrossRef] [PubMed]
  5. Dias, D.; Paulo Silva Cunha, J. Wearable Health Devices—Vital Sign Monitoring, Systems and Technologies. Sensors 2018, 18, 2414. [Google Scholar] [CrossRef] [PubMed]
  6. Convertino, V.A.; Wirt, M.D.; Glenn, J.F.; Lein, B.C. The Compensatory Reserve for Early and Accurate Prediction of Hemodynamic Compromise: A Review of the Underlying Physiology. Shock 2016, 45, 580–590. [Google Scholar] [CrossRef]
  7. Convertino, V.A.; Koons, N.J.; Suresh, M.R. Physiology of Human Hemorrhage and Compensation. Compr. Physiol. 2011, 11, 1531–1574. [Google Scholar]
  8. Looney, D.P.; Buller, M.J.; Gribok, A.V.; Leger, J.L.; Potter, A.W.; Rumpler, W.V.; Tharion, W.J.; Welles, A.P.; Friedl, K.E.; Hoyt, R.W. Estimating Resting Core Temperature Using Heart Rate. J. Meas. Phys. Behav. 2018, 1, 79–86. [Google Scholar] [CrossRef]
  9. Daimiwal, N.; Sundhararajan, M.; Shriram, R. Respiratory Rate, Heart Rate and Continuous Measurement of BP Using PPG. In Proceedings of the 2014 International Conference on Communication and Signal Processing, Chennai, India, 3–5 April 2014; pp. 999–1002. [Google Scholar]
  10. Jeong, I.; Jun, S.; Um, D.; Oh, J.; Yoon, H. Non-Invasive Estimation of Systolic Blood Pressure and Diastolic Blood Pressure Using Photoplethysmograph Components. Yonsei Med. J. 2010, 51, 345–353. [Google Scholar] [CrossRef]
  11. Kurylyak, Y.; Lamonaca, F.; Grimaldi, D. A Neural Network-Based Method for Continuous Blood Pressure Estimation from a PPG Signal. In Proceedings of the 2013 IEEE International Instrumentation and Measurement Technology Conference (I2MTC), Minneapolis, MN, USA, 6–9 May 2013; pp. 280–283. [Google Scholar]
  12. Sun, S.; Bezemer, R.; Long, X.; Muehlsteff, J.; Aarts, R.M. Systolic Blood Pressure Estimation Using PPG and ECG during Physical Exercise. Physiol. Meas. 2016, 37, 2154. [Google Scholar] [CrossRef]
  13. Moulton, S.L.; Mulligan, J.; Grudic, G.Z.; Convertino, V.A. Running on Empty? The Compensatory Reserve Index. J. Trauma Acute Care Surg. 2013, 75, 1053–1059. [Google Scholar] [CrossRef]
  14. Techentin, R.W.; Felton, C.L.; Schlotman, T.E.; Gilbert, B.K.; Joyner, M.J.; Curry, T.B.; Convertino, V.A.; Holmes, D.R.; Haider, C.R. 1D Convolutional Neural Networks for Estimation of Compensatory Reserve from Blood Pressure Waveforms. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; IEEE: New York, NY, USA, 2019; pp. 2169–2173. [Google Scholar]
  15. Shin, D. The Effects of Explainability and Causability on Perception, Trust, and Acceptance: Implications for Explainable AI. Int. J. Hum.-Comput. Stud. 2021, 146, 102551. [Google Scholar] [CrossRef]
  16. Amann, J.; Blasimme, A.; Vayena, E.; Frey, D.; Madai, V.I.; the Precise4Q consortium. Explainability for Artificial Intelligence in Healthcare: A Multidisciplinary Perspective. BMC Med. Inform. Decis. Mak. 2020, 20, 310. [Google Scholar] [CrossRef] [PubMed]
  17. Folke, T.; Yang, S.C.-H.; Anderson, S.; Shafto, P. Explainable AI for Medical Imaging: Explaining Pneumothorax Diagnoses with Bayesian Teaching. In Proceedings of the Artificial Intelligence and Machine Learning for Multi-Domain Operations Applications III, Online Only, 12–16 April 2021; SPIE: Bellingham, WA, USA, 2021; Volume 11746, pp. 644–664. [Google Scholar]
  18. Hatib, F.; Jian, Z.; Buddi, S.; Lee, C.; Settels, J.; Sibert, K.; Rinehart, J.; Cannesson, M. Machine-Learning Algorithm to Predict Hypotension Based on High-Fidelity Arterial Pressure Waveform Analysis. Anesthesiology 2018, 129, 663–674. [Google Scholar] [CrossRef] [PubMed]
  19. Krishnan, S.; Athavale, Y. Trends in Biomedical Signal Feature Extraction. Biomed. Signal Process. Control 2018, 43, 41–63. [Google Scholar] [CrossRef]
  20. Amin, H.U.; Malik, A.S.; Ahmad, R.F.; Badruddin, N.; Kamel, N.; Hussain, M.; Chooi, W.-T. Feature Extraction and Classification for EEG Signals Using Wavelet Transform and Machine Learning Techniques. Australas. Phys. Eng. Sci. Med. 2015, 38, 139–149. [Google Scholar] [CrossRef]
  21. Gupta, J.F.; Arshad, S.H.; Telfer, B.A.; Snider, E.J.; Convertino, V.A. Noninvasive Monitoring of Simulated Hemorrhage and Whole Blood Resuscitation. Biosensors 2022, 12, 1168. [Google Scholar] [CrossRef]
  22. Convertino, V.A.; Schauer, S.G.; Weitzel, E.K.; Cardin, S.; Stackle, M.E.; Talley, M.J.; Sawka, M.N.; Inan, O.T. Wearable Sensors Incorporating Compensatory Reserve Measurement for Advancing Physiological Monitoring in Critically Injured Trauma Patients. Sensors 2020, 20, 6413. [Google Scholar] [CrossRef]
  23. Goswami, N.; Blaber, A.P.; Hinghofer-Szalkay, H.; Convertino, V.A. Lower Body Negative Pressure: Physiological Effects, Applications, and Implementation. Physiol. Rev. 2019, 99, 807–851. [Google Scholar] [CrossRef]
  24. Eleti, A.A.; Zerek, A.R. FIR Digital Filter Design by Using Windows Method with MATLAB. In Proceedings of the 14th International Conference on Sciences and Techniques of Automatic Control & Computer Engineering-STA’2013, Sousse, Tunisia, 20–22 December 2013; IEEE: New York, NY, USA, 2013; pp. 282–287. [Google Scholar]
  25. Gupta, J.F.; Telfer, B.A.; Convertino, V.A. Feature Importance Analysis for Compensatory Reserve to Predict Hemorrhagic Shock. In Proceedings of the 2022 44th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Glasgow, UK, 11–15 July 2022; IEEE: New York, NY, USA; pp. 1747–1752. [Google Scholar]
  26. Ding, C.; Peng, H. Minimum Redundancy Feature Selection from Microarray Gene Expression Data. J. Bioinform. Comput. Biol. 2005, 3, 185–205. [Google Scholar] [CrossRef]
  27. Kern, C.; Klausch, T.; Kreuter, F. Tree-Based Machine Learning Methods for Survey Research. Surv. Res. Methods 2019, 13, 73–93. [Google Scholar]
  28. Kotsiantis, S.B.; Tsekouras, G.E.; Pintelas, P.E. Bagging Model Trees for Classification Problems. In Advances in Informatics, Proceedings of the 10th Panhellenic Conference On Informatics (Hci’99), Volas, Greece, 11–13 November 2005; Bozanis, P., Houstis, E.N., Eds.; Springer: Berlin/Heidelberg, Germany, 2005; pp. 328–337. [Google Scholar]
  29. Hayes, T.; Usami, S.; Jacobucci, R.; McArdle, J.J. Using Classification and Regression Trees (CART) and Random Forests to Analyze Attrition: Results from Two Simulations. Psychol. Aging 2015, 30, 911–929. [Google Scholar] [CrossRef] [PubMed]
  30. Convertino, V.A.; Koons, N.J. The Compensatory Reserve: Potential for Accurate Individualized Goal-Directed Whole Blood Resuscitation. Transfusion 2020, 60, S150–S157. [Google Scholar] [CrossRef] [PubMed]
Figure 1. (A) The Lower Body Negative Pressure (LBNP) step profile and subject distribution. A maximum of nine negative steps were used, including the initial baseline starting pressure. Subjects were removed from LBNP at hemodynamic decompensation or after the final pressure step (B) Sample distribution (N = 218 subjects) for the LBNP pressure where hemodynamic decompensation was reached.
Figure 1. (A) The Lower Body Negative Pressure (LBNP) step profile and subject distribution. A maximum of nine negative steps were used, including the initial baseline starting pressure. Subjects were removed from LBNP at hemodynamic decompensation or after the final pressure step (B) Sample distribution (N = 218 subjects) for the LBNP pressure where hemodynamic decompensation was reached.
Bioengineering 10 00612 g001
Figure 2. Representative arterial waveform segment from LBNP dataset with key landmarks identified. Landmarks include the pulse foot (“A”), systolic peak (“C”), half-rise between pulse foot and systolic peak (“B”), and the post-systolic inflection point (“D”).
Figure 2. Representative arterial waveform segment from LBNP dataset with key landmarks identified. Landmarks include the pulse foot (“A”), systolic peak (“C”), half-rise between pulse foot and systolic peak (“B”), and the post-systolic inflection point (“D”).
Bioengineering 10 00612 g002
Figure 3. Representative plots comparing perfect fit and linear fit regression models for theoretical vs. predicted CRM at (A) 4, (B) 5, (C) 6, (D) 7, and (E) 8 final LBNP step subjects.
Figure 3. Representative plots comparing perfect fit and linear fit regression models for theoretical vs. predicted CRM at (A) 4, (B) 5, (C) 6, (D) 7, and (E) 8 final LBNP step subjects.
Bioengineering 10 00612 g003
Figure 4. Heat map matrices of the different trained ML models (rows) vs. the blind test data (columns) showing results for (A) perfect coefficient of determinations (P-R2) and (B) perfect root mean squared errors (P-RMSE).
Figure 4. Heat map matrices of the different trained ML models (rows) vs. the blind test data (columns) showing results for (A) perfect coefficient of determinations (P-R2) and (B) perfect root mean squared errors (P-RMSE).
Bioengineering 10 00612 g004
Figure 5. Representative data for (A) 4, (B) 5, (C) 6, (D) 7, and (E) 8 final LBNP step subjects comparing the CRM-ML model (green line), CRM-DL model (red line), and calculated CRM (orange line).
Figure 5. Representative data for (A) 4, (B) 5, (C) 6, (D) 7, and (E) 8 final LBNP step subjects comparing the CRM-ML model (green line), CRM-DL model (red line), and calculated CRM (orange line).
Bioengineering 10 00612 g005
Figure 6. (A) P-R2 values and (B) P-RMSE of CRM-ML and CRM-DL models using the perfect regression method. (C) R2 and (D) RMSE values of CRM-ML and CRM-DL models using the conventional linear regression method.
Figure 6. (A) P-R2 values and (B) P-RMSE of CRM-ML and CRM-DL models using the perfect regression method. (C) R2 and (D) RMSE values of CRM-ML and CRM-DL models using the conventional linear regression method.
Bioengineering 10 00612 g006
Table 1. Summary of the 54 distinct features extracted from each LBNP waveform. Supplementary Table S1 contains additional descriptions for each individual feature. Features calculated by subtracting the pressure value at the inflection point are referred to as “NODIA”.
Table 1. Summary of the 54 distinct features extracted from each LBNP waveform. Supplementary Table S1 contains additional descriptions for each individual feature. Features calculated by subtracting the pressure value at the inflection point are referred to as “NODIA”.
Feature TypesDescriptionNumber of Features
Individual FeaturesFeatures from the arterial waveform consist of standard waveform measurements (such as pulse pressure [PP] and peak-to-peak interval [PPI]).7
Time Duration FeaturesDuration of certain phases of the arterial waveform (such as time from half-rise to inflection point [HRIP] and duration of the systolic phase [t_sys]).6
Average Pressure FeaturesAverage pressures of different arterial waveform phases.5
Area Under the Curve FeaturesArea under the curve of different waveform phases.5
Normalized FeaturesArea features normalized by the number of samples in the waveform and phases.10
NODIA FeaturesArea under the curve and normalized features subtracted by the waveform value of the inflection point.18
Slope FeaturesAverage slope of different phases of the arterial waveform.3
Table 2. Training and testing results for different classical machine-learning models trained using the top 1, 5, 10, or 15 extracted features. Performance results are shown for perfect RMSE and R2 values for each model. Green heat map overlay is set between the minimum and maximum value to indicate better performance.
Table 2. Training and testing results for different classical machine-learning models trained using the top 1, 5, 10, or 15 extracted features. Performance results are shown for perfect RMSE and R2 values for each model. Green heat map overlay is set between the minimum and maximum value to indicate better performance.
15 Features
TrainingTesting
ModelP-RMSEP-R2P-RMSEP-R2
Linear Regression0.178650.740.177660.73
Fine Tree0.101330.920.232740.54
Medium Tree0.101760.920.216630.6
Coarse Tree0.109570.90.200970.66
Boosted Tree0.143220.830.159660.78
Bagged Tree0.077620.950.171330.75
10 Features
TrainingTesting
ModelP-RMSEP-R2P-RMSEP-R2
Linear Regression0.178650.740.184310.71
Fine Tree0.101330.920.223120.58
Medium Tree0.103060.910.208330.63
Coarse Tree0.111490.90.199780.66
Boosted Tree0.145820.830.162020.78
Bagged Tree0.0790940.950.174420.74
5 Features
TrainingTesting
ModelP-RMSEP-R2P-RMSEP-R2
Linear Regression0.19480.690.179420.73
Fine Tree0.125010.870.22250.58
Medium Tree0.122290.880.209160.63
Coarse Tree0.127540.870.19250.68
Boosted Tree0.157040.80.165730.77
Bagged Tree0.10350.910.177080.73
1 Feature
TrainingTesting
ModelP-RMSEP-R2P-RMSEP-R2
Linear Regression0.199570.680.184410.71
Fine Tree0.206070.660.195940.67
Medium Tree0.189540.710.179260.73
Coarse Tree0.181420.730.169950.75
Boosted Tree0.179210.740.173510.74
Bagged Tree0.187960.710.17740.73
Table 3. Coefficients of determination and root mean squared error performance metrics using the CRM-ML model to predict CRM for each LBNPHDD sub-group. Results are shown for perfect regression (P-R2 and P-RMSE) and traditional linear regression (R2 and RMSE).
Table 3. Coefficients of determination and root mean squared error performance metrics using the CRM-ML model to predict CRM for each LBNPHDD sub-group. Results are shown for perfect regression (P-R2 and P-RMSE) and traditional linear regression (R2 and RMSE).
Step SubgroupP-R2P-RMSER2RMSE
40.520.200.830.20
50.690.190.840.19
60.840.130.890.13
70.710.170.830.17
80.740.160.850.15
Table 4. Summary of ten highest ranked features for an ML model trained with all data or exclusively the 8, 7, 6, 5, or 4 final LBNP step subgroup. Superscript values for each feature indicate how much higher (positive/green), lower (negative/red), or no change (equal sign) the specific extracted feature’s importance in ranking shifted relative to the all-steps CRM-ML model. More information on how each of the features was calculated is described in Supplementary Table S1.
Table 4. Summary of ten highest ranked features for an ML model trained with all data or exclusively the 8, 7, 6, 5, or 4 final LBNP step subgroup. Superscript values for each feature indicate how much higher (positive/green), lower (negative/red), or no change (equal sign) the specific extracted feature’s importance in ranking shifted relative to the all-steps CRM-ML model. More information on how each of the features was calculated is described in Supplementary Table S1.
All Steps8 Steps7 Steps6 Steps5 Steps4 Steps
RankFeatureFeatureFeatureFeatureFeatureFeature
1HRIPsys_rise_area_norm+52sys_rise_area_norm+52HRIP=sys_rise_area_norm+52HRIP=
2dec_area_norSI+16sys_rise_area_nodia+18avg_sys_rise+35slope_desc_sys+5avg_dia+33
3t_sys_riset_sys_rise=t_sys_rise=avg_sys_rise_nodia+10avg_sys_nodia+11avg_sys_dec_nodia+1
4avg_sys_dec_nodiaavg_sys_dec_nodia=HRIP−3PP+6dec_area_nodia+2sys_rise_area_norm+49
5avg_dia_nodiaHRIP−4PP+5t_sys_rise−2HRIP−4slope_sys+12
6dec_area_nodiadia_area_nodia+9dec_area_nodia=sys_area_norm+16avg_sys_dec_nodia−2t_sys_rise−3
7slope_desc_sysslope_dia+1sys_area+2sys_dec_area+16t_sys_rise−4t_sys+23
8slope_diaslope_desc_sys−1avg_sys_dec_nodia−4slope_dia=slope_dia=slope_desc_sys−1
9sys_areasys_area_nodia+11slope_sys+8pp_area_nodia+10PP+1sys_area=
10PPpp_area_nodia+9slope_desc_sys−3avg_sys_nodia+4PPI+1dec_area_nodia−4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bedolla, C.N.; Gonzalez, J.M.; Vega, S.J.; Convertino, V.A.; Snider, E.J. An Explainable Machine-Learning Model for Compensatory Reserve Measurement: Methods for Feature Selection and the Effects of Subject Variability. Bioengineering 2023, 10, 612. https://doi.org/10.3390/bioengineering10050612

AMA Style

Bedolla CN, Gonzalez JM, Vega SJ, Convertino VA, Snider EJ. An Explainable Machine-Learning Model for Compensatory Reserve Measurement: Methods for Feature Selection and the Effects of Subject Variability. Bioengineering. 2023; 10(5):612. https://doi.org/10.3390/bioengineering10050612

Chicago/Turabian Style

Bedolla, Carlos N., Jose M. Gonzalez, Saul J. Vega, Víctor A. Convertino, and Eric J. Snider. 2023. "An Explainable Machine-Learning Model for Compensatory Reserve Measurement: Methods for Feature Selection and the Effects of Subject Variability" Bioengineering 10, no. 5: 612. https://doi.org/10.3390/bioengineering10050612

APA Style

Bedolla, C. N., Gonzalez, J. M., Vega, S. J., Convertino, V. A., & Snider, E. J. (2023). An Explainable Machine-Learning Model for Compensatory Reserve Measurement: Methods for Feature Selection and the Effects of Subject Variability. Bioengineering, 10(5), 612. https://doi.org/10.3390/bioengineering10050612

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop