Next Article in Journal
The Role of ctDNA for Diagnosis and Histological Prediction in Early Stage Non-Small-Cell Lung Cancer: A Narrative Review
Next Article in Special Issue
EM-DeepSD: A Deep Neural Network Model Based on Cell-Free DNA End-Motif Signal Decomposition for Cancer Diagnosis
Previous Article in Journal
Clinicians’ Agreement on Extrapulmonary Radiographic Findings in Chest X-Rays Using a Diagnostic Labelling Scheme
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Electrocardiogram Abnormality Detection Using Machine Learning on Summary Data and Biometric Features

1
Department of Computer Science, College of Engineering and Computing Sciences, New York Institute of Technology, 1855 Broadway, New York, NY 10023, USA
2
Entrepreneurship and Technology Innovation Center, College of Engineering and Computing Sciences, New York Institute of Technology, Old Westbury, NY 11568, USA
3
Department of Osteopathic Manipulative Medicine, College of Osteopathic Medicine, New York Institute of Technology, Old Westbury, NY 11568, USA
4
Catholic Health Service of Long Island, 245 Old Country Rd, Melville, NY 11747, USA
*
Authors to whom correspondence should be addressed.
Diagnostics 2025, 15(7), 903; https://doi.org/10.3390/diagnostics15070903
Submission received: 29 January 2025 / Revised: 25 March 2025 / Accepted: 27 March 2025 / Published: 1 April 2025
(This article belongs to the Special Issue Deep Learning in Biomedical Signal Analysis)

Abstract

:
Background/Objectives: Electrocardiogram data are widely used to diagnose cardiovascular diseases, a leading cause of death globally. Traditional interpretation methods are manual, time-consuming, and prone to error. Machine learning offers a promising alternative for automating the classification of electrocardiogram abnormalities. This study explores the use of machine learning models to classify electrocardiogram abnormalities using a dataset that combines clinical features (e.g., age, weight, smoking status) with key electrocardiogram measurements, without relying on time-series data. Methods: The dataset included demographic and electrocardiogram-related biometric data. Preprocessing steps addressed class imbalance, outliers, feature scaling, and the encoding of categorical variables. Five machine learning models—Gaussian Naive Bayes, support vector machines, random forest trees, extremely randomized trees, gradient boosted trees, and an ensemble of top-performing classifiers—were trained and optimized using stratified k-fold cross-validation. Model performance was evaluated on a reserved testing set using metrics such as accuracy, precision, recall, and F1-score. Results: The extremely randomized trees model achieved the best performance, with a testing accuracy of 66.79%, recall of 66.79%, and F1-score of 62.93%. Ventricular rate, QRS duration, and QTC (Bezet) were identified as the most important features. Challenges in classifying borderline cases were noted due to class imbalance and overlapping features. Conclusions: This study demonstrates the potential of machine learning models, particularly extremely randomized trees, in classifying electrocardiogram abnormalities using demographic and biometric data. While promising, the absence of time-series data limits diagnostic accuracy. Future work incorporating time-series signals and advanced deep learning techniques could further improve performance and clinical relevance.

1. Introduction

Electrocardiogram (ECG) data plays a critical role in diagnosing cardiovascular diseases, which are a leading cause of mortality worldwide [1]. Traditional ECG interpretation relies on trained medical professionals to visually inspect the ECG waveforms and identify abnormalities [2]. However, this manual process can be time-consuming, subjective, and prone to human error [2,3,4]. In recent years, machine learning (ML) techniques have shown great promise in automating ECG diagnosis, enabling more efficient and accurate detection of cardiovascular conditions [5,6,7]. However, existing studies primarily rely on time-series ECG data, which, while highly feature-rich, may not always be available in all clinical settings. This study focuses on the practical application of ML models in scenarios where access to such time-series signals is unavailable. This reflects real-world challenges faced by many healthcare facilities, particularly those with limited resources or infrastructure, where only summary ECG data and demographic features are accessible. By demonstrating the feasibility of using limited data for ECG abnormality classification, our work highlights the potential of ML to extend diagnostic support to a broader range of clinical environments. Such approaches are particularly valuable in resource-constrained contexts, where even incremental improvements in clinical workflows can have significant impacts on patient care.
The features of ECG biometrics are derived from the characteristic waveforms of the heart’s electrical activity, including the P-R interval, QRS duration, Q-T interval, and axes such as P, R, and T. These features are influenced by physiological and demographic factors such as age, weight, and heart rate, making them highly individualized. ECG-related biometrics are explored not only for their diagnostic potential but also for their utility in classification tasks using ML algorithms. By combining clinical features (e.g., age, weight, and smoking status) with key ECG measurements, this study aims to classify ECG abnormalities without relying on time-series data.
Recent advancements in ML have advanced ECG analysis by improving diagnostic precision and interpretability through novel architectures and optimization techniques [8]. A foundational review underscores the transformative role of AI in automating cardiovascular disease diagnosis, particularly through deep learning models that enhance accuracy and efficiency [9]. Building on this, innovative architectures like xECGArch leverage both short- and long-term ECG features to achieve high interpretability and performance in atrial fibrillation detection [10], while portable models such as CardioAttentionNet combine BiLSTM networks and attention mechanisms for robust arrhythmia classification [11]. Feature engineering also plays a critical role, as evidenced by metaheuristic algorithms that optimize feature selection to attain near-perfect accuracy in arrhythmia classification [12]. Despite these advancements, challenges like data scarcity and class imbalance persist, prompting the systematic exploration of data augmentation methods to improve model generalizability under constrained datasets [13]. Together, these studies highlight a growing emphasis on hybrid methodologies that integrate architectural innovation, interpretability, and data-centric approaches, setting the stage for clinically adaptable solutions in ECG abnormality detection.
This study investigates the use of various shallow ML models, including extremely randomized trees (EXT), gradient boosted trees (GBT), support vector machines (SVM), and others, to classify a subject’s ECG status based on a unique dataset containing both clinical features (e.g., age, weight, heart rate) and key ECG measurements (such as the P-R interval and QRS duration). Using this combination of demographic and ECG-related biometric data, trained models aim to accurately identify ECG abnormalities without relying on the actual ECG time series signal.
The study addresses several challenges throughout the research process, including exploratory data analysis, class imbalance, data preprocessing, model selection, and hyperparameter tuning. Class imbalance, in particular, posed a significant challenge, as the majority of samples were labeled as “Abnormal ECG”, which could lead to biased models that struggle to correctly classify minority classes. To mitigate this, techniques such as stratified sampling and the use of evaluation metrics like the F1-score were employed. The methodology for this study involved preprocessing the dataset to handle outliers, scaling features, and encoding categorical variables. Multiple ML models, including Gaussian Naive Bayes, support vector machines, random forest trees, extremely randomized trees, and gradient boosted trees, were trained and optimized using stratified k-fold cross-validation. Model performance was evaluated on a reserved testing set using metrics such as accuracy, precision, recall, and F1-score. The study also explored feature importance and reframed the classification problem to address challenges in distinguishing borderline ECG cases.
The contribution of this study lies in applying the foundational steps of ML (ranging from data analysis, preprocessing, training, and hyperparameter optimization) to address the challenges inherent to working with an incomplete dataset: specifically, the absence of time-series ECG signals, the presence of outlier values, and a relatively small sample size present significant obstacles. However, this study demonstrates that clinically useful and realistic results can still be achieved by focusing on real-world scenarios where access to complete datasets, such as time-series ECG signals, may not always be possible. By addressing these challenges in a methodical and transparent manner, the study highlights the potential of ML models to provide value in resource-constrained settings without inflating performance expectations.

2. Methods

This section details the processes of data preparation, analysis, and ML model development. The data collection process is described, including the sources and types of data utilized. Exploratory data analysis is conducted to examine the dataset’s structure, distributions, and potential issues such as outliers and imbalances. Data preprocessing steps are outlined, including cleaning, transforming, and preparing the data for ML, with a focus on handling missing values, scaling features, and addressing class imbalance. The section also explains the training and optimization of ML models, covering the selection of algorithms, hyperparameter tuning, and evaluation methods to ensure robust and reliable performance.

2.1. Data Collection

An anonymized dataset of N = 4466 ECG summary records was acquired for this study from a regional healthcare system in the New York metro area. The dataset included a diverse range of demographic, clinical, and biometric features, such as age, weight, smoking status, ventricular rate, atrial rate, and key ECG measurements (e.g., P-R interval, QRS duration, Q-T interval, and QTC (Bezet)). To ensure patient privacy, all personally identifiable information was removed prior to data access, and the dataset was provided in compliance with applicable data protection regulations. The data represented a heterogeneous population, reflecting the variability in patient characteristics and cardiovascular health statuses. This diversity was critical for developing and evaluating machine learning models capable of generalizing across different patient populations. However, the dataset also presented challenges, such as class imbalance, with a majority of samples labeled as “Abnormal ECG”, and the presence of outliers, such as implausible weight or heart rate values, likely resulting from data entry errors. These issues were addressed during preprocessing to ensure the dataset was suitable for training robust and reliable machine learning models.

2.2. Exploratory Data Analysis

Data visualizations and analysis of summary statistics, e.g., the feature mean and its distribution, presence of outliers, etc., reveal additional information about the data [14]. This guided the authors on what data preprocessing methods are necessary or relevant for each feature, such as value encoding, normalizing/standardization to some distribution, outlier handling, etc.
Table 1 shows some examples of the dataset (after the removal of nuisance columns such as IDs that are unique to each observation—which does not provide predictive value).
Figure 1 shows the target distribution, i.e., the frequency of value occurrences of the DIAGNOSIS_LINE. The majority of the data has Abnormal ECG values, indicating an imbalanced dataset. This imbalance is addressed in Section 2.3 Data Preprocessing via stratified sampling.
The distributions of the numerical features are shown in Figure 2. A cursory visual analysis of the researchers leads to the conclusion that most of the features follow a Gaussian distribution, while some have a long right-tailed distribution.
The outlier analysis of the numerical columns WEIGHT, VENTRICULAR RATE, ATRIAL RATE is shown in Figure 3. Outlier values above 95% of the cumulative distribution function are expected to be realistic, e.g., subjects with obesity or poor cardiovascular health. However, outliers on the lower end are unrealistic and are considered absurd outlier values, e.g., a person having a weight value of less than 50 lbs or heart rate of 10 beats/min: This may be due to some flaw that occurred in the data collection process of the source. Again, this issue is addressed in Section 2.3.

2.3. Data Preprocessing

In this subsection, we focus on the significant preprocessing steps performed such as target value curation, treatment of columns that had values originating from non-standardized data collection, and handling of categorical, ordinal, and numerical features. Some examples were preprocessed (The examples in Table 2 are not necessarily the same examples from Table 1).

2.3.1. Target Creation

Target features were created by consolidating the multi-line format of the DIAGNOSIS_LINE_* columns into one column (DIAGNOSIS) via concatenation, separated by a space. Then, by case-insensitive string matching via the substring operator ∋, target values of the observations containing the keys outlined in the map in Equation (1) are replaced with their corresponding ordinal value:
d DIAGNOSIS , f ( d ) = 0 , if Normal ECG d 1 , else if Borderline ECG d 2 , else if Abnormal ECG d
Finally, those observations that did not contain any of the keys were dropped from the dataset.

2.3.2. Smoker Column

Redundant and synonymous smoker values, e.g., Some Days and Light Smoker, were consolidated. This column is coded as an ordinal variable ranging from observations that individuals did not smoke to those that they had heavy smoke exposure. The re-labeling of every possible smoker value into its corresponding ordinal is outlined in the map in Equation (2):
s SMOKING _ STATUS , f ( s ) = 0 , if never s 1 , if never assessed s unknown s 2 , if former s 3 , if smoker , current status unknown s light smoker s some days s passive smoke exposure - never smoker s 4 , if every day s heavy smoker s

2.3.3. Other Nominal Columns

Other nominal columns, SEX and DIABETES, are encoded as 0s and 1s. N.B., the DIABETES column may be considered as a categorical feature, i.e., having no order in the values, since it only has binary values. Finally, examples with missing values of these features are imputed using their mode.

2.3.4. Numerical Columns

The remaining numerical columns, including those with long right-tailed distributions, were scaled to a standard normal distribution (the choice to standardize long-tailed distributions, rather than apply a log transformation or min-max scaling, was guided by cross-validation results), with careful attention to outlier values. To address these outliers, these features are scaled into a robust standard normal that involved removal of the median and scaling of the data based on the interquartile range [15]. Standardizing these columns also offers the added advantage of improving the training efficiency of gradient-based methods and SVMs, which are particularly sensitive to data scaling [16]. Finally, examples with missing values of these features are imputed using their mean.

2.3.5. Dataset Split

Permutation of the dataset was performed—see the Source Code and Scripts section that guarantees deterministic/reproducible results. Then, the dataset was split into an 80:20 ratio, where 80% ( N train = 3188 ) of the data is for the training set and the other 20% ( N test = 798 ) is for the test set. A further 80:20 stratified cross-validation split on the training set produced a partial-training set and validation set for the purposes of hyperpameter optimization.

2.4. Model Training and Hyperparameter Tuning

2.4.1. The Baseline (Null) Classifier

Given the multiclass classification at hand (3 classes), a naive or null model would predict any sample using the dataset’s target mode, which is DIAGNOSIS = “Abnormal ECG” or 2, which comprises (52.36%) majority of the full dataset. That is, a higher accuracy than that will be a successful classifier.
Considering the practicality of the model (specifically, its value in predicting patient ECG data) and the unbalanced nature of the dataset, the primary metric of interest will be the F1 score. This metric, which is the harmonic mean of precision and recall, is robust to class imbalance [17].

2.4.2. Model Training and Optimization

Classifiers based on different priors are considered, e.g., tree-based, gradient-based, etc., to enable various exploration of the unknown dataset hypothesis. All classification models trained are as follows: K-nearest neighbor (KNN), Gaussian Naive Bayes (GNB) [18], support vector machine (SVM) [19], random forest (RFT) [20], extremely randomized trees (EXT) [21], and gradient boosted trees (GBT) [22], Finally, an ensemble of the top 4 fine-tuned weak learners (ENS) is used, leveraging a meta-learning approach. This ensemble predicts classes using a soft-voting scheme that averages the predictions and confidence scores of the individual weak learners. Soft voting, in contrast to majority (hard) voting, helps handle ties in the overall predictions. Equation (3) represents the soft-voting mechanism, where the final prediction y ^ for a given diagnosis d is determined by selecting the class c that maximizes the weighted sum of probabilities P m ( c ) predicted by each weak learner m, with weights w m . This approach incorporates the confidence of each model, making it more robust than hard voting.
d DIAGNOSIS , y ^ = arg max c C m = 1 M w m P m ( c )
Model variety and the No Free Lunch Theorem guided the researchers to explore and choose the best performing model on different metrics [23,24,25].
Stratified k-fold (K = 5) validation was used to train the models to ensure robustness in validation metrics, achieve partial train-validation split, and avoid any bottleneck of the lowest frequency class.
A grid search of plausible hyperparameters for every model is explored, e.g., number of neighbors for each sample, depth for tree-based methods, penalty parameter C for SVM, etc., again using stratified k-fold validation. This allows for the hyperparameter exploration of each model to avoid overfitting the training data, i.e., for the purposes of model regularization.
Finally, all models are parameterized with their best hyperparameters and retrained using the entire training dataset, that is, including the validation set, to maximize the available data for learning and enhance the model’s generalization performance [23].

3. Results

The results of this study are presented in the following subsections. First, the performance of the hyperparameter-optimized models is evaluated using the reserved testing set, with particular emphasis on the SVM model, which outperformed most models across several metrics. Although the ensemble of multiple classifiers (SVM, EXT, GNB, and GBT) showed superior performance compared to the individual SVC model, the authors selected the most parsimonious model based on Occam’s razor. The confusion matrix analysis provides a detailed view of the model’s classification accuracy and highlights challenges in distinguishing borderline ECG cases. To address these challenges, the problem is reframed as a binary classification task and the trade-offs between precision and recall are explored using a precision–recall curve. Finally, feature importance analysis identifies the most influential predictors that contribute to the model performance, providing valuable information on the clinical and biometric parameters critical to ECG classification.

3.1. Model Evaluation on the Testing Set

The hyperparameter-optimized models were evaluated using the reserved test set. Table 3 summarizes the test metrics for all models, highlighting the best-performing metrics for each. The results indicate that the extremely randomized trees (EXT) model consistently outperformed all other models in most metrics of interest, including testing accuracy, recall, and F1 score. Figure 4 shows a visualization of the results in Table 3.
The computational efficiency of the models was assessed during the training and evaluation phases. All models were trained and tested on standard computational hardware, with training times ranging from a few seconds to several minutes depending on the model complexity and hyperparameter tuning. Prediction times for new data were near-instantaneous, demonstrating the practicality of the approach for real-world clinical applications. The overall computational requirements are minimal, making the proposed solution feasible for deployment in resource-constrained settings.

3.2. Overview of the SVM Model

The SVM model demonstrated superior performance [19]. For the given dataset, the optimal hyperparameters for the SVM model were determined to be a tolerance C of 1 and using the radial basis function as the kernel. These hyperparameters were identified through grid search and cross-validation, ensuring the model’s robustness and ability to generalize effectively.

3.3. Confusion Matrix Analysis

Figure 5 presents the confusion matrix for the EXT model’s predictions on the testing set. The matrix highlights the distribution of true versus predicted classes for normal, borderline, and abnormal ECG diagnoses.
The confusion matrix reveals that the EXT model struggled to classify borderline ECG cases accurately. These cases were often misclassified as either normal or abnormal ECGs, reflecting the inherent difficulty in distinguishing borderline cases due to their overlap with the other two categories.

3.4. Binary Classification Reframing

A naive solution to address the high and dangerous cost of a false negative diagnosis of a patient (and also false positives, depending on the situation and goal) is to reframe the problem from multiclass to binary classification by forming an equivalence between the target values: Borderline ECG ≡ Abnormal ECG. The modified target map in Equation (4) replaces the DIAGNOSIS values with its corresponding ordinals:
d DIAGNOSIS , f ( d ) = 0 , if Normal ECG d 1 , else if Borderline ECG d Abnormal ECG d
This conversion from multiclass to binary classification allows the model to have a precision–recall (PR) trade-off curve with a dynamic threshold, enabling the selection of a specific precision value along with its corresponding recall, and vice versa [26,27,28].
Figure 6 shows the PR curve of the EXT on the testing set with 95% target values for both precision and recall. Depending on the situation and goal, to minimize either false negatives or false positives, the decision threshold for prediction can be adjusted to 0.7967 (for 95% precision) and 0.3883 (for 95% recall), respectively. This reframing allowed for the generation of a precision–recall (PR) curve, as shown in Figure 6. The PR curve demonstrates the trade-off between precision and recall at varying decision thresholds, enabling the selection of thresholds based on specific clinical priorities.
Key thresholds were identified, including one achieving 95% precision (with recall at 33.21%) and another achieving 95% recall (with precision at 68.09%). These thresholds provide flexibility for optimizing the model based on the desired balance between minimizing false positives and minimizing false negatives, depending on the clinical context.

3.5. Feature Importance Analysis via the EXT

Although EXT did not outperform SVC, it is still valuable to know the decision function made by this bagging-trees model (EXT outperformed RFT). Feature importance analysis for the EXT model revealed that ventricular rate, QRS duration, and QTC (Bezet) were the most influential features in predicting ECG classifications. Figure 7 illustrates the relative importance of each feature, along with the variability in importance scores across different subsets of the data.
These findings highlight the critical role of specific clinical and biometric parameters in distinguishing among normal, borderline, and abnormal ECG statuses. The variability in feature importance scores underscores the robustness of these key predictors across different subsets of the data.

4. Discussion

The implications of these misclassifications are critical in a healthcare context. False negatives, where borderline ECGs are misclassified as normal, may lead to complacency in patients, delaying necessary medical interventions. Conversely, false positives, where borderline ECGs are misclassified as abnormal, could result in unnecessary treatments, causing undue stress, financial burden, and potential harm to patients. These risks underscore the importance of minimizing both false negatives and false positives in medical diagnostic models to ensure accurate and actionable predictions.

4.1. Challenges in Classifying Borderline ECG Cases

The classification of borderline ECG cases presented a significant challenge in this study. These cases, which lie between normal and abnormal ECGs in terms of risk, were the most difficult to classify accurately. This difficulty is likely due to the inherent overlap in features between the borderline category and the other two classes. From a clinical perspective, this overlap reflects the continuum of cardiovascular health, where borderline cases may share characteristics with both normal and abnormal ECGs. Consequently, the model’s misclassifications often involved predicting borderline cases as either normal or abnormal.
The values in the multiple grouped distributions represent different clinical and biometric features (e.g., ventricular rate, QRS duration, weight, etc.) and their respective frequencies or measurements across the dataset. These distributions are important because they provide insights into the variability and patterns within the data, which directly influence the performance of the machine learning models used to classify ECG outcomes. For example, the distributions reveal key challenges such as class imbalance (e.g., the majority of samples being labeled as “Abnormal ECG”) and the presence of outliers (e.g., unrealistic weights or heart rates). These issues must be addressed during preprocessing to ensure the models are trained on clean, representative data. Additionally, the differences in distributions highlight the heterogeneity of the dataset, which reflects the diversity of patient characteristics and conditions.
Relating these distributions back to the use of ML in evaluating ECG outcomes, they underscore the complexity of the classification task. AI models, particularly machine learning algorithms, rely on patterns in the data to make predictions. Understanding the distributions helps identify which features are most informative for distinguishing between “Normal”, “Borderline”, and “Abnormal” ECGs. For instance, features like ventricular rate and QRS duration, which exhibit distinct patterns in their distributions, were found to be among the most important predictors in the study. These distributions are critical as they guide the preprocessing steps (e.g., scaling, outlier removal, and stratified sampling) and influence model selection and optimization. By addressing the challenges revealed in the distributions, the study ensures that the AI models are robust and capable of accurately classifying ECG abnormalities, ultimately improving diagnostic efficiency and reducing the risk of human error in clinical practice.

4.2. Reframing the Problem as Binary Classification

To address the challenges associated with borderline ECG classification, the problem was reframed as a binary classification task. By combining the borderline and abnormal ECG categories into a single “at-risk” class, the model’s focus shifted to distinguishing between normal and at-risk cases. This reframing allowed for the generation of a precision–recall (PR) curve, which provided a dynamic threshold for optimizing the trade-off between precision and recall.
The PR curve enabled the identification of thresholds tailored to specific clinical priorities. For instance, a threshold achieving high precision minimizes false positives, ensuring that patients classified as at-risk are highly likely to require medical attention. Conversely, a threshold achieving high recall minimizes false negatives, ensuring that nearly all at-risk patients are identified. This flexibility is particularly valuable in medical applications, where the cost of misclassification varies depending on the clinical context. For example, in screening scenarios, high recall may be prioritized to ensure no at-risk patients are missed, while in treatment planning, high precision may be preferred to avoid unnecessary interventions.

4.3. Feature Importance and Clinical Relevance

The feature importance analysis of the extremely randomized trees (EXT) model provided valuable insights into the predictors most influential in ECG classification. Ventricular rate, QRS duration, and QTC (Bezet) were identified as the top contributors to the model’s predictions. These features are well-established indicators of cardiovascular health, with abnormalities in these parameters often associated with arrhythmias, conduction disorders, and other cardiac conditions. The robustness of these features across different subsets of the data highlights their clinical relevance and supports their use in automated diagnostic models.
The variability in feature importance scores, as indicated by the standard deviations across out-of-bag samples, suggests that while these features are consistently influential, their relative contributions may vary depending on the specific characteristics of the dataset [12]. This variability underscores the importance of incorporating diverse and representative data in model training to ensure generalizability across different patient populations.

4.4. Limitations and Future Directions

Despite the promising results, this study faced several limitations that warrant further investigation. First, the dataset contained unrealistic outliers, such as subjects with weights below 50 lbs or heart rates below 10 beats per minute. These anomalies likely resulted from errors during data collection and were addressed during preprocessing. However, a cleaner and more standardized dataset would likely improve model performance and reliability.
Second, the absence of time-series ECG signal data limited the scope of analysis to static demographic and biometric features. ECG data are inherently time-series in nature, and the inclusion of raw signal data could enable the application of advanced deep learning techniques, such as long short-term memory (LSTM) networks or convolutional neural networks (CNNs). These methods are well-suited for capturing temporal patterns and could significantly enhance the model’s diagnostic accuracy. Future studies should prioritize access to time-series ECG data to explore these possibilities.
Finally, the class imbalance in the dataset posed a challenge, particularly for the borderline ECG category. While stratified sampling and evaluation metrics such as the F1-score were employed to mitigate this issue, alternative approaches, such as synthetic data generation or cost-sensitive learning, could further improve the model’s ability to handle imbalanced datasets.
Hence, to improve the diagnostic capabilities and practical application of ML models for ECG abnormality detection, future research should prioritize several key directions. First, incorporating time-series ECG signal data into the analysis is essential, as such data captures temporal patterns critical for accurate diagnosis. Other advanced deep learning techniques could be leveraged to process these time-series signals to enhance model performance [10]. Additionally, efforts should be made to obtain larger and more diverse datasets that reflect a wide range of patient demographics and clinical conditions, thereby improving the generalizability of the models. Another critical area for exploration is the development of explainable AI (XAI) systems, which can provide clinicians with interpretable insights into the decision-making process, fostering greater trust and adoption in clinical settings. Lastly, standardized validation frameworks, such as those aligned with regulatory benchmarks like the FDA guidance [29], should be adopted to ensure robust model evaluation and facilitate meaningful cross-study comparisons.
Furthermore, field-programmable gate array (FPGA)-accelerated architectures achieve 8–10× higher energy efficiency than GPUs through reduced-bitwidth computations [30]. These hardware optimizations could enhance model portability in resource-constrained clinical settings where low latency and energy efficiency are critical [31]. Future work should investigate dedicated hardware implementations, such as FPGA-based inference accelerators, to improve computational throughput for real-time deployment. Techniques like 8-bit quantization [32] could reduce model size and processing latency without compromising diagnostic accuracy, aligning with energy-efficient edge computing advancements.

4.5. Implications for Clinical Practice

The findings of this study demonstrate the potential of ML models, particularly the EXT model, in automating the classification of ECG abnormalities using demographic and biometric data. The ability to dynamically adjust decision thresholds based on clinical priorities offers a practical advantage, allowing healthcare providers to tailor the predictions of the model to specific diagnostic or screening scenarios [33]. However, the challenges associated with borderline ECG classification highlight the need for careful consideration of the clinical context and the potential consequences of misclassification. Either way, as of now, no ML model is intended to be used exclusively for medical diagnosis. These models are designed to serve as supplementary tools to assist physicians by providing potential insights or predictions, rather than making definitive clinical decisions. Their primary purpose is to guide clinicians on what to expect and to support, not replace, human expertise.
Hence, while the EXT model showed promise in classifying ECG abnormalities, further improvements in data quality, feature representation, and model architecture are necessary to fully realize the potential of machine learning in this domain. Future research should focus on addressing these limitations and exploring the integration of time-series data and advanced deep learning techniques to enhance diagnostic accuracy and clinical utility.

4.6. Challenges in Cross-Study Comparability

A significant challenge in the field of ML for healthcare care, particularly in studies such as this, is the lack of standardized validation frameworks, which makes comparisons between studies of model performance unreliable. Although some studies report impressively high accuracy metrics, these often result from inadequate validation practices, such as improper data splitting, a lack of external testing, or overfitting to specific datasets. This issue is particularly relevant when comparing our rigorously validated accuracy of 66.79% with reported in other studies that claim higher but potentially less reliable metrics.
The absence of unified methods for evaluating ML models leads to inflated and misleading performance claims. Many studies do not adopt robust practices such as three-way data splitting (training, validation, and testing) or external validation on independent datasets. Without these practices, reported metrics may reflect overfitting or data leakage rather than true generalizability. For example, a model achieving 98% accuracy on a poorly validated dataset may appear superior to a rigorously validated model with lower accuracy, but its clinical utility is practically meaningless.
In addition, the variability in the datasets, preprocessing techniques, and evaluation metrics between studies further complicates the comparisons. Factors such as differences in patient demographics, data collection methods, and feature engineering can significantly influence model performance. Without standardized reporting and validation protocols, it becomes impossible to discern whether a model’s reported accuracy reflects genuine clinical utility or methodological artifacts.
In this study, rigorous validation practices were prioritized, including stratified k-fold cross-validation and careful handling of class imbalance, to ensure the reliability of the results. Although our accuracy may appear modest compared to some studies, it reflects a realistic and reproducible performance level, free from overfitting or data leakage. Until the field adopts standardized validation frameworks, such as those outlined in recent FDA guidance [29], comparing ML studies remains speculative and potentially misleading. Therefore, we caution against equating high reported accuracies with clinical reliability and emphasize the importance of robust validation in evaluating ML models.

5. Conclusions

This study demonstrates the potential of ML models in detecting abnormalities in ECGs using summary ECG data and biometric features. By addressing the challenges of missing time-series ECG signals, outliers, and class imbalance in a relatively small dataset, the research highlights the feasibility of applying ML in real-world, resource-constrained settings. The EXT model emerged as the most effective classifier, achieving the highest accuracy and recall, with ventricular rate, QRS duration, and QTC (Bezet) identified as critical predictive features. The results underscore the practicality of leveraging summary ECG and biometric data for ECG abnormality detection, making this approach particularly suitable for healthcare environments where access to time-series data is limited. This study’s findings demonstrate deploying ML-based diagnostic tools as scalable solutions in diverse clinical environments. Future work could explore refining the models using larger datasets, incorporating additional biometric parameters, and validating the models in varied clinical contexts to ensure broader applicability and improved diagnostic outcomes.

Author Contributions

All authors (K.J.B., A.S., D.N., C.H., M.R., J.T., M.N. and M.T.) contributed to all aspects of this work. This includes but is not limited to conceptualization, methodology, software management, validation, formal analysis, investigation, resources management, data curation, writing—original draft preparation, writing—review and editing, and visualization. Supervision and project administration were handled by M.N. and M.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ECG/EKGElectrocardiogram
ENSEnsemble
EXTExtremely Randomized Trees
GBTGradient Boosted Trees
SVMSupport Vector Machines
GNBGaussian Naive Bayes
OOBOut-of-Bag
FPGAField-Programmable Gate Array

References

  1. Centers for Disease Control and Prevention (CDC). Heart Disease Facts and Statistics. 2024. Available online: https://www.cdc.gov/heart-disease/data-research/facts-stats/ (accessed on 29 March 2025).
  2. Kadish, A.H.; Buxton, A.E.; Kennedy, H.L.; Knight, B.P.; Mason, J.W.; Schuger, C.D.; Tracy, C.M.; Winters, W.L.; Boone, A.W.; Elnicki, M.; et al. ACC/AHA Clinical Competence Statement on Electrocardiography and Ambulatory Electrocardiography. Circulation 2001, 104, 3169–3178. [Google Scholar] [CrossRef] [PubMed]
  3. Compiet, S.; Willemsen, R.; Konings, K.; Stoffers, H. Competence of general practitioners in requesting and interpreting ECGs—A case vignette study. Neth. Heart J. 2018, 26, 377–384. [Google Scholar] [CrossRef] [PubMed]
  4. Cook, D.A.; Oh, S.Y.; Pusic, M.V. Accuracy of Physicians’ Electrocardiogram Interpretations: A Systematic Review and Meta-analysis. JAMA Intern. Med. 2020, 180, 1461–1471. [Google Scholar] [CrossRef] [PubMed]
  5. Ziz, S.; Ahmed, S.; Alouini, M.S. ECG-based machine-learning algorithms for heartbeat classification. Sci. Rep. 2021, 11, 18738. [Google Scholar] [CrossRef]
  6. Mincholé, A.; Camps, J.; Lyon, A.; Rodríguez, B. Machine learning in the electrocardiogram. J. Electrocardiol. 2019, 57, S61–S64. [Google Scholar] [CrossRef]
  7. Attia, Z.I.; Harmon, D.M.; Behr, E.R.; Friedman, P.A. Application of artificial intelligence to the electrocardiogram. Eur. Heart J. 2021, 42, 4717–4730. [Google Scholar] [CrossRef]
  8. Chong, L.; Husain, G.; Nasef, D.; Vathappallil, P.; Matalia, M.; Toma, M. Machine Learning Strategies for Improved Cardiovascular Disease Detection. Med. Res. Arch. 2025, 13, 1–16. [Google Scholar] [CrossRef]
  9. Denysyuk, H.V.; Pinto, R.J.; Silva, P.M.; Duarte, R.P.; Marinho, F.A.; Pimenta, L.; Gouveia, A.J.; Gonçalves, N.J.; Coelho, P.J.; Zdravevski, E.; et al. Algorithms for automated diagnosis of cardiovascular diseases based on ECG data: A comprehensive systematic review. Heliyon 2023, 9, e13601. [Google Scholar] [CrossRef]
  10. Goettling, M.; Hammer, A.; Malberg, H.; Schmidt, M. xECGArch: A trustworthy deep learning architecture for interpretable ECG analysis considering short-term and long-term features. Sci. Rep. 2024, 14, 13122. [Google Scholar] [CrossRef]
  11. He, Y.; Zhou, Y.; Qian, Y.; Liu, J.; Zhang, J.; Liu, D.; Wu, Q. Cardioattentionnet: Advancing ECG beat characterization with a high-accuracy and portable deep learning model. Front. Cardiovasc. Med. 2025, 11, 1473482. [Google Scholar] [CrossRef]
  12. Darmawahyuni, A.; Nurmaini, S.; Tutuko, B.; Rachmatullah, M.N.; Firdaus, F.; Sapitri, A.I.; Islami, A.; Marcelino, J.; Isdwanta, R.; Perwira, M.I. An improved electrocardiogram arrhythmia classification performance with feature optimization. BMC Med. Inform. Decis. Mak. 2024, 24, 412. [Google Scholar] [CrossRef]
  13. Rahman, M.M.; Rivolta, M.W.; Badilini, F.; Sassi, R. A Systematic Survey of Data Augmentation of ECG Signals for AI Applications. Sensors 2023, 23, 5237. [Google Scholar] [CrossRef] [PubMed]
  14. Tukey, J.W. Exploratory Data Analysis; Addison-Wesley: Boston, MA, USA, 1977. [Google Scholar]
  15. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  16. Juszczak, P.; Tax, D.M.J.; Duin, R.P.W. Feature scaling in support vector data description. In Proceedings of the International Conference on Pattern Recognition (ICPR), Quebec City, QC, Canada, 11–15 August 2002; Available online: https://api.semanticscholar.org/CorpusID:14498563 (accessed on 29 March 2025).
  17. Bishop, C.M. Pattern Recognition and Machine Learning; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  18. Jahromi, A.H.; Taheri, M. A non-parametric mixture of Gaussian naive Bayes classifiers based on local independent features. In Proceedings of the 2017 Artificial Intelligence and Signal Processing Conference (AISP), Shiraz, Iran, 25–27 October 2017; pp. 209–212. [Google Scholar] [CrossRef]
  19. Stitson, M.; Weston, J.; Gammerman, A.; Vovk, V.; Vapnik, V. Theory of support vector machines. Univ. Lond. 1996, 117, 188–191. [Google Scholar]
  20. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar]
  21. Geurts, P.; Ernst, D.; Wehenkel, L. Extremely randomized trees. Mach. Learn. 2006, 63, 3–42. [Google Scholar]
  22. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. Lightgbm: A highly efficient gradient boosting decision tree. Adv. Neural Inf. Process. Syst. 2017, 30, 1–9. Available online: https://proceedings.neurips.cc/paper_files/paper/2017/file/6449f44a102fde848669bdd9eb6b76fa-Paper.pdf (accessed on 29 March 2025).
  23. Chollet, F. Deep Learning with Python, 2nd ed.; Manning: Merseyside, UK, 2021. [Google Scholar]
  24. Moreno-Ibarra, M.A.; Villuendas-Rey, Y.; Lytras, M.D.; Yáñez-Márquez, C.; Salgado-Ramírez, J.C. Classification of Diseases Using Machine Learning Algorithms: A Comparative Study. Mathematics 2021, 9, 1817. [Google Scholar] [CrossRef]
  25. Barth, J.; Klucken, J.; Kugler, P.; Kammerer, T.; Steidl, R.; Winkler, J.; Hornegger, J.; Eskofier, B. Biometric and mobile gait analysis for early diagnosis and therapy monitoring in Parkinson’s disease. In Proceedings of the 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Boston, MA, USA, 30 August–30 September 2011; pp. 868–871. [Google Scholar] [CrossRef]
  26. Boyd, K.; Eng, K.H.; Page, C.D. Area under the Precision-Recall Curve: Point Estimates and Confidence Intervals. In Proceedings of the Machine Learning and Knowledge Discovery in Databases; Blockeel, H., Kersting, K., Nijssen, S., Železný, F., Eds.; Springer: Berlin/Heidelberg, Germany, 2013; pp. 451–466. [Google Scholar]
  27. Buckland, M.; Gey, F. The relationship between Recall and Precision. J. Am. Soc. Inf. Sci. 1994, 45, 12–19. [Google Scholar] [CrossRef]
  28. Flach, P.; Kull, M. Precision-Recall-Gain Curves: PR Analysis Done Right. In Proceedings of the Advances in Neural Information Processing Systems; Cortes, C., Lawrence, N., Lee, D., Sugiyama, M., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2015; Volume 28. [Google Scholar]
  29. Food, U.; Administration, D. Considerations for the Use of Artificial Intelligence to Support Regulatory Decision-Making for Drug and Biological Products: Draft Guidance for Industry and Other Interested Parties; U.S. Department of Health and Human Services: Washington, DC, USA; Center for Drug Evaluation and Research (CDER), Center for Biologics Evaluation and Research (CBER), Center for Devices and Radiological Health (CDRH): Silver Spring, MD, USA, 2025. Available online: https://www.fda.gov/media/184830/download (accessed on 29 March 2025).
  30. Guo, K.; Zeng, S.; Yu, J.; Wang, Y.; Yang, H. [DL] A Survey of FPGA-based Neural Network Inference Accelerators. ACM Trans. Reconfig. Technol. Syst. 2019, 12, 1–26. [Google Scholar] [CrossRef]
  31. Liu, Y.; Du, H.; Wu, Y.; Mo, T. FPGA Accelerated Deep Learning for Industrial and Engineering Applications: Optimal Design Under Resource Constraints. Electronics 2025, 14, 703. [Google Scholar] [CrossRef]
  32. Dawwd, S. FPGA based Flexible Implementation of Light Weight Inference on Deep Convolutional Neural Networks. Int. Arab. J. Inf. Technol. 2024, 21, 408–417. [Google Scholar] [CrossRef]
  33. Nasef, D.; Nasef, D.; Basco, K.J.; Singh, A.; Hartnett, C.; Ruane, M.; Tagliarino, J.; Nizich, M.; Toma, M. Clinical Applicability of Machine Learning Models for Binary and Multi-Class Electrocardiogram Classification. AI 2025, 6, 59. [Google Scholar] [CrossRef]
Figure 1. Frequency distribution of ECG diagnosis categories in the dataset ‘DIAGNOSIS’ column, illustrating the proportion of samples classified as ‘Abnormal ECG’, ‘Normal ECG’, and ‘Borderline ECG’. The chart highlights a significant class imbalance, with ‘Abnormal ECG’ constituting the majority of the dataset, emphasizing the challenges this imbalance poses for effective model training and classification.
Figure 1. Frequency distribution of ECG diagnosis categories in the dataset ‘DIAGNOSIS’ column, illustrating the proportion of samples classified as ‘Abnormal ECG’, ‘Normal ECG’, and ‘Borderline ECG’. The chart highlights a significant class imbalance, with ‘Abnormal ECG’ constituting the majority of the dataset, emphasizing the challenges this imbalance poses for effective model training and classification.
Diagnostics 15 00903 g001
Figure 2. Distribution of ECG-related numerical features before preprocessing, including clinical parameters such as weight, ventricular rate, atrial rate, P-R interval, QRS duration, Q-T interval, and axes (P, R, T). Most features exhibit a Gaussian distribution, while others display long right-tailed distributions. These visualizations guided preprocessing steps, such as outlier analysis and standardization, to optimize the dataset for machine learning model training.
Figure 2. Distribution of ECG-related numerical features before preprocessing, including clinical parameters such as weight, ventricular rate, atrial rate, P-R interval, QRS duration, Q-T interval, and axes (P, R, T). Most features exhibit a Gaussian distribution, while others display long right-tailed distributions. These visualizations guided preprocessing steps, such as outlier analysis and standardization, to optimize the dataset for machine learning model training.
Diagnostics 15 00903 g002
Figure 3. Boxplots of selected numerical features (weight, ventricular rate, and atrial rate) illustrating the results of outlier analysis. The analysis highlights the presence of extreme and unrealistic outliers, such as weights below 50 lbs, ventricular rates below 30 bpm, and atrial rates below 50 bpm. These outliers likely result from errors during the data collection process. The findings informed preprocessing steps, which included the removal of implausible values to improve data quality and model performance.
Figure 3. Boxplots of selected numerical features (weight, ventricular rate, and atrial rate) illustrating the results of outlier analysis. The analysis highlights the presence of extreme and unrealistic outliers, such as weights below 50 lbs, ventricular rates below 30 bpm, and atrial rates below 50 bpm. These outliers likely result from errors during the data collection process. The findings informed preprocessing steps, which included the removal of implausible values to improve data quality and model performance.
Diagnostics 15 00903 g003
Figure 4. The results in Table 3 of all models evaluated on the test set visualized to compare which model outperforms on all or most of the metrics.
Figure 4. The results in Table 3 of all models evaluated on the test set visualized to compare which model outperforms on all or most of the metrics.
Diagnostics 15 00903 g004
Figure 5. The confusion matrix of predictions of the SVM illustrates the distribution of true versus predicted classes for normal, borderline, and abnormal ECG diagnoses. Borderline ECG predictions were the least frequent, with most misclassifications occurring between borderline and the other two categories. The darker diagonal cells indicate higher accuracy for normal and abnormal ECG classifications, whereas the lighter shading in the borderline column highlights the model’s challenge in correctly identifying this category.
Figure 5. The confusion matrix of predictions of the SVM illustrates the distribution of true versus predicted classes for normal, borderline, and abnormal ECG diagnoses. Borderline ECG predictions were the least frequent, with most misclassifications occurring between borderline and the other two categories. The darker diagonal cells indicate higher accuracy for normal and abnormal ECG classifications, whereas the lighter shading in the borderline column highlights the model’s challenge in correctly identifying this category.
Diagnostics 15 00903 g005
Figure 6. The PR curve highlights the trade-off between precision and recall across varying decision thresholds. Key points include a threshold yielding 95% precision (with recall at 33.21%) and another yielding 95% recall (with precision at 68.09%).
Figure 6. The PR curve highlights the trade-off between precision and recall across varying decision thresholds. Key points include a threshold yielding 95% precision (with recall at 33.21%) and another yielding 95% recall (with precision at 68.09%).
Diagnostics 15 00903 g006
Figure 7. The bar heights represent the relative importance scores, indicating that features such as ventricular rate, QRS duration, and QTC (Bezet) are among the most influential in determining ECG classifications. The error bars illustrate the standard deviation of the importance scores across out-of-bag samples, reflecting variability in feature contributions.
Figure 7. The bar heights represent the relative importance scores, indicating that features such as ventricular rate, QRS duration, and QTC (Bezet) are among the most influential in determining ECG classifications. The error bars illustrate the standard deviation of the importance scores across out-of-bag samples, reflecting variability in feature contributions.
Diagnostics 15 00903 g007
Table 1. Sample rows from the original patient dataset showcasing ECG-related biometric data and clinical parameters. The table includes features such as sex, weight, smoking status, ventricular rate, atrial rate, and key ECG measurements (e.g., P-R interval, QRS duration, Q-T interval, and QTC (Bezet)), along with target labels for ECG diagnosis (normal, borderline, or abnormal ECG). These raw data entries highlight the diversity of patient characteristics and serve as the foundation for preprocessing and machine learning model training.
Table 1. Sample rows from the original patient dataset showcasing ECG-related biometric data and clinical parameters. The table includes features such as sex, weight, smoking status, ventricular rate, atrial rate, and key ECG measurements (e.g., P-R interval, QRS duration, Q-T interval, and QTC (Bezet)), along with target labels for ECG diagnosis (normal, borderline, or abnormal ECG). These raw data entries highlight the diversity of patient characteristics and serve as the foundation for preprocessing and machine learning model training.
SEXWEIGHTDIABETESSMOKEVENT. RATEATRIAL RATEP-R INTERVALQRS DURATIONQ-T INTERVALQTC (BEZET)P AXISR AXIST AXISDIAGNOSIS_LINE_1...DIAGNOSIS_LINE_14
Male204.59YFormer7474266.08843848637.0−2827.0Abnormal ECG...NaN
Male162.48NNever5858135.09241040345.02829.0Normal ECG...NaN
Female119.00NNever7562172.014643248249.0−21118.0Abnormal ECG...Synus Rhythm
Table 2. Preprocessed patient dataset highlighting key features used for machine learning model training and evaluation. The table includes standardized and encoded numerical features (e.g., P-R interval, QRS duration, Q-T interval, QTC (Bezet), ventricular rate, and weight), as well as one-hot encoded categorical features (e.g., sex and diabetes).
Table 2. Preprocessed patient dataset highlighting key features used for machine learning model training and evaluation. The table includes standardized and encoded numerical features (e.g., P-R interval, QRS duration, Q-T interval, QTC (Bezet), ventricular rate, and weight), as well as one-hot encoded categorical features (e.g., sex and diabetes).
SMOKEP-R INTERVALQRS DURATIONQ-T INTERVALQTC (BEZET)SEX_FemaleSEX_MaleDIABETESATRIAL RATEP AXISR AXIST AXISVENT. RATEWEIGHT
31.24−0.59−0.380.00101.001.360.350.320.520.214.23
00.50−1.120.10−1.28011.00−0.22−0.85−0.77−2.32−0.12−0.30
00.570.53−0.910.91010.001.001.751.590.770.724.11
Table 3. The table compares the performance of Gaussian Naive Bayes (GNB), support vector machines (SVM), random forest trees (RFT), extremely randomized trees (EXT), gradient boosted trees (GBT), K-nearest neighbors (KNN), and the ensemble (ENS) consisting of GNB, GBT, EXT, & SVM on the testing dataset. Metrics include testing accuracy, precision, recall, and F1-score. The EXT model achieved the highest overall performance, demonstrating its robustness in handling class imbalance and optimizing the bias-variance trade-off.
Table 3. The table compares the performance of Gaussian Naive Bayes (GNB), support vector machines (SVM), random forest trees (RFT), extremely randomized trees (EXT), gradient boosted trees (GBT), K-nearest neighbors (KNN), and the ensemble (ENS) consisting of GNB, GBT, EXT, & SVM on the testing dataset. Metrics include testing accuracy, precision, recall, and F1-score. The EXT model achieved the highest overall performance, demonstrating its robustness in handling class imbalance and optimizing the bias-variance trade-off.
ModelTesting AccuracyTesting PrecisionTesting RecallTesting F1-Score
GNB65.29%65.08%65.29%63.03%
SVM66.17%59.69%66.17%62.63%
RFT64.79%65.29%64.79%61.60%
EXT65.54%62.07%65.54%61.99%
GBT64.16%60.10%64.16%61.13%
KNN62.53%61.64%62.53%60.74%
ENS67.04%60.79%67.04%63.53%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Basco, K.J.; Singh, A.; Nasef, D.; Hartnett, C.; Ruane, M.; Tagliarino, J.; Nizich, M.; Toma, M. Electrocardiogram Abnormality Detection Using Machine Learning on Summary Data and Biometric Features. Diagnostics 2025, 15, 903. https://doi.org/10.3390/diagnostics15070903

AMA Style

Basco KJ, Singh A, Nasef D, Hartnett C, Ruane M, Tagliarino J, Nizich M, Toma M. Electrocardiogram Abnormality Detection Using Machine Learning on Summary Data and Biometric Features. Diagnostics. 2025; 15(7):903. https://doi.org/10.3390/diagnostics15070903

Chicago/Turabian Style

Basco, Kennette James, Alana Singh, Daniel Nasef, Christina Hartnett, Michael Ruane, Jason Tagliarino, Michael Nizich, and Milan Toma. 2025. "Electrocardiogram Abnormality Detection Using Machine Learning on Summary Data and Biometric Features" Diagnostics 15, no. 7: 903. https://doi.org/10.3390/diagnostics15070903

APA Style

Basco, K. J., Singh, A., Nasef, D., Hartnett, C., Ruane, M., Tagliarino, J., Nizich, M., & Toma, M. (2025). Electrocardiogram Abnormality Detection Using Machine Learning on Summary Data and Biometric Features. Diagnostics, 15(7), 903. https://doi.org/10.3390/diagnostics15070903

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop