Next Article in Journal
Frequency-Domain Masking and Spatial Interaction for Generalizable Deepfake Detection
Previous Article in Journal
High-Precision Low-Power Interface Circuit for Two-Dimensional Integrated Magnetic Switches
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning for Non-Invasive Blood Pressure Monitoring: Model Performance and Quantization Trade-Offs

by
Anbu Valluvan Devadasan
1,
Saptarshi Sengupta
2 and
Mohammad Masum
1,*
1
Department of Applied Data Science, San Jose State University, San Jose, CA 95192, USA
2
Department of Computer Science, San Jose State University, San Jose, CA 95192, USA
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(7), 1300; https://doi.org/10.3390/electronics14071300
Submission received: 13 February 2025 / Revised: 10 March 2025 / Accepted: 21 March 2025 / Published: 26 March 2025
(This article belongs to the Special Issue Artificial Intelligence and Big Data Processing in Healthcare)

Abstract

:
The development of non-invasive blood pressure monitoring systems remains a critical challenge, particularly in resource-constrained settings. This study proposes an efficient deep learning framework integrating Edge Artificial Intelligence for continuous blood pressure estimation using photoplethysmography (PPG) signals. We evaluate three architectures: a residual-enhanced convolutional neural network, a transformer-based model, and an attentive BPNet. Using the MIMIC-IV waveform database, we implement a signal processing pipeline with adaptive filtering, statistical normalization, and peak-to-peak alignment. Experiments assess varying temporal windows (10 s, 20 s, 30 s) to optimize predictive accuracy and computational efficiency. Attentive BPNet achieves the best performance, with systolic blood pressure (SBP) estimation yielding a mean absolute error (MAE) of 6.36 mmHg, diastolic blood pressure (DBP) an MAE of 4.09 mmHg, and mean arterial pressure (MBP) an MAE of 4.56 mmHg. Post-training quantization reduces the model size by 90.71% (to 0.13 MB), enabling deployment on Edge devices. These findings demonstrate the feasibility of deploying deep learning-based continuous blood pressure monitoring on edge devices. The proposed framework provides a scalable and computationally efficient solution, offering real-time, accessible monitoring that could enhance hypertension management and optimize healthcare resource utilization.

1. Introduction

Cardiovascular diseases remain a leading global health concern, with hypertension affecting approximately 1.28 billion adults worldwide [1]. Continuous blood pressure monitoring is essential for the early detection, management, and prevention of cardiovascular complications [2]. However, traditional oscillometric cuff-based measurements suffer from significant limitations, including low temporal resolution, patient discomfort, and susceptibility to artifacts, such as white coat hypertension, which can compromise diagnostic accuracy [3,4]. The development of cuffless blood pressure estimation offers a promising alternative, leveraging photoplethysmography (PPG) and electrocardiography (ECG) signals for continuous, non-invasive monitoring [3,5]. At the same time, advancements in Edge Artificial Intelligence have enabled the deployment of sophisticated machine learning algorithms on resource-constrained embedded systems. These technologies facilitate real-time, on-device inference while addressing key challenges associated with cloud-based solutions, such as latency and data privacy concerns [6,7].
Despite these advancements, wearable blood pressure monitoring systems still face challenges in achieving high accuracy and cross-subject generalizability [8]. Many existing methods require subject-specific calibration, which limits large-scale adoption in clinical settings [9]. Additionally, optimizing deep learning architectures for deployment on edge devices while maintaining clinical-grade accuracy remains a significant research challenge.
This study introduces a comprehensive Edge AI-based framework for continuous, non-invasive blood pressure estimation using PPG signals.
Our research objectives are to do as follows:
  • Develop accurate models for estimating mean, systolic, and diastolic blood pressure.
  • Optimize these models for deployment on resource-constrained devices.
  • Evaluate model performance in terms of accuracy, computational efficiency, and inference time.
Our contributions include the following:
  • Proposing a novel Edge AI-based blood pressure estimation framework.
  • Demonstrating the feasibility of continuous monitoring on wearable devices without the need for subject-specific calibration.
  • Providing insights into the trade-offs between accuracy and computational efficiency in Edge AI applications for healthcare.
Unlike prior studies that primarily focus on high-resource environments or calibration-dependent models, our work presents a scalable, calibration-free deep learning framework optimized for wearable edge devices. This approach advances the field by enabling accurate, real-time blood pressure monitoring without compromising computational efficiency or generalizability.
The remainder of this paper is structured as follows: Section 2 reviews contemporary literature on cuffless blood pressure estimation and healthcare-focused Edge AI applications. Section 3 details our methodological framework, including signal preprocessing, model development, and Edge AI optimization strategies. Section 4 presents experimental results and their implications. Section 5 concludes with key findings and future research directions.

2. Related Works

The advancement of continuous, non-invasive blood pressure monitoring is a critical area in cardiovascular diagnostics. This section reviews the evolution of these methodologies, emphasizing their theoretical foundations, practical limitations, and integration with edge artificial intelligence for real-time healthcare applications.
Non-Invasive Blood Pressure Monitoring Techniques: Traditional oscillometric cuff-based methods, despite being the clinical standard, present significant limitations for continuous monitoring, including measurement intermittency, patient discomfort, and susceptibility to artifacts like white coat hypertension. These limitations have driven extensive research into cuffless monitoring approaches [1]. Several non-invasive techniques have been explored. Pulse Transit Time (PTT) estimates blood pressure by measuring the time a pulse wave travels between two arterial sites. While Mukkamala et al. highlight its potential, PTT is affected by arterial stiffness variations and requires frequent calibration [10,11]. Pulse Wave Analysis (PWA) examines arterial waveform characteristics to estimate blood pressure. Avolio et al. discuss its ability to assess central blood pressure and arterial stiffness, often relying on applanation tonometry or photoplethysmography [10]. This review further explores recent advancements and ongoing challenges in noninvasive blood pressure estimation, focusing on cuffless techniques such as Pulse Transit Time (PTT), along with emerging pressure-based, ultrasound-based, and deep learning methods, emphasizing their potential for accurate and continuous monitoring in both clinical and home settings [11,12]. Bioimpedance measures electrical impedance variations in body tissues to estimate blood volume and indirectly blood pressure, though it remains in early development [13,14,15]. The vascular unloading technique, or volume-clamp method, maintains constant blood volume in finger arteries, enabling continuous measurements, as described by Fortin et al. [16,17]. Ultrasound-based methods utilize arterial diameter changes to infer blood pressure, with Wang et al. demonstrating a wearable ultrasonic device for continuous monitoring [18,19]. Optical techniques, such as imaging photoplethysmography and smartphone-based methods, leverage skin color changes for blood pressure estimation, though motion artifacts and skin tone variations remain challenges [5,20].
Each of these approaches offers unique advantages and limitations. PWA provides detailed cardiovascular insights but requires precise sensor placement. PTT is simple yet lacks long-term stability. The vascular unloading technique delivers accurate continuous measurements but requires specialized hardware. Ultrasound-based methods directly measure arterial diameter changes but face miniaturization and power constraints. Optical techniques show promise for integration into consumer devices but require further refinement for reliability.
Photoplethysmography (PPG) for Blood Pressure Estimation: The shift away from mercury sphygmomanometers has accelerated the adoption of noninvasive methods like PPG, which offer simple, cost-effective, and wearable solutions for continuous blood pressure monitoring [21]. Kurylyak et al. developed an artificial neural network (ANN) using 21 PPG features to estimate systolic and diastolic blood pressure [22]. Xing and Sun employed a multilayer feed-forward neural network to improve accuracy [5]. Schlesinger et al. initially used an AlexNet-inspired CNN on PPG spectrograms and later introduced a Siamese network for refined blood pressure estimation [23]. Casadei et al. applied an ANN regression model to PPG data from the MIMIC II database [24]. Mukkamala et al. explored multiple PPG-based blood pressure estimation methods, including volume clamping, oscillometry, and waveform feature extraction [25].
Edge Artificial Intelligence and Computing in Healthcare: Edge Artificial Intelligence, which enables machine learning on resource-constrained devices, is revolutionizing healthcare by facilitating real-time monitoring while reducing latency and privacy concerns associated with cloud-based solutions [26]. It offers several advantages: real-time processing, enabling immediate physiological analysis [27]; privacy preservation, by processing sensitive health data locally [28]; power efficiency, through optimized ML models on low-power microcontrollers [27]. However, challenges remain. Model compression is essential for maintaining accuracy while reducing size, with pruning, quantization, and knowledge distillation being key techniques [29]. Hardware constraints necessitate careful optimization of ML models for limited-memory devices [29]. Calibration and adaptation require efficient on-device learning to ensure reliable blood pressure estimations [29].
Recent advancements have addressed these challenges. Hymel et al. introduced Edge Impulse, a cloud-based platform that streamlines Edge AI development, integrating data collection, digital signal processing, and ML deployment optimizations [30]. Tsoukas et al. reviewed Edge AI applications in healthcare, detailing model optimization techniques like quantization and pruning for embedded deployment [31]. Anbukarasu et al. demonstrated TinyHR, a TinyML-based heart rate estimation model deployed on an ESP32 device, showcasing low-power solutions for biomedical applications [32].
Signal Processing Techniques for Physiological Signals: Effective signal processing is crucial for extracting meaningful insights from physiological data. Preprocessing typically begins with detrending techniques to remove low-frequency artifacts, followed by bandpass filtering (0.5–10 Hz) to preserve cardiac cycle components [33,34]. Feature extraction involves peak detection algorithms to identify systolic and diastolic points in PPG and ABP signals, along with waveform segmentation for cycle-specific analysis [34,35]. Morphological analysis extracts pulse amplitude, width, and reflection indices for further cardiovascular assessment [36]. Wavelet transforms enable time-frequency decomposition for feature extraction and noise reduction [37]. Singular Spectrum Analysis (SSA) improves PPG signal quality by separating key components from noise [35]. Derivative analysis, including velocity and acceleration photoplethysmograms, provides deeper vascular insights [33].
Quantization Techniques for Edge Device Optimization: Quantization is a crucial optimization for deploying deep learning models on edge devices. By reducing model weight and activation precision—typically from 32-bit floating point to lower bit-width representations—quantization significantly lowers memory usage and computational costs. Post-training quantization (PTQ) and quantization-aware training (QAT) are two primary approaches. PTQ includes dynamic quantization, where weights are pre-quantized while activations remain dynamic, and static quantization, which applies fixed calibration for both weights and activations. Key parameters include the scale factor (S) and zero point (Z), ensuring accurate mapping between quantized and floating-point values. Per-channel quantization enhances precision at the cost of increased complexity [38].
The edge deployment workflow involves identifying computationally intensive operations for quantization, applying post-training dynamic quantization, and refining with static quantization if needed. Calibration on a representative dataset ensures optimal performance. Quantization typically reduces the model size by 3–4×, accelerates inference by 1.2–3×, and minimizes power consumption, though aggressive quantization may impact accuracy. Quantization-aware training helps mitigate accuracy loss [37,38,39,40].
Despite notable advancements in non-invasive blood pressure monitoring, several key challenges remain. First, the integration of advanced signal processing techniques with Edge AI for blood pressure estimation is still underexplored. Second, the development of robust, calibration-free models that generalize well across subjects—without the need for personalized calibration—remains a significant hurdle, particularly for edge-based deployments. Third, there is a lack of comprehensive comparative studies evaluating Edge AI solutions across different hardware platforms and real-world use cases. In response to these gaps, our research presents a novel Edge AI-based framework for continuous, non-invasive blood pressure estimation using PPG signals. While several existing studies have employed deep learning models with wearable sensors, many rely on resource-intensive architectures or require subject-specific calibration, limiting their scalability and practicality. In contrast, our approach focuses on developing a calibration-free, cross-subject generalizable model optimized for real-time performance on resource-constrained edge devices. By combining advanced signal processing, efficient deep learning architectures, and post-training quantization, our work addresses the critical challenges of scalability, efficiency, and clinical applicability. This framework not only enhances the feasibility of real-time, continuous blood pressure monitoring but also contributes to improved cardiovascular health management and better patient outcomes in diverse settings.

3. Dataset Description

This study uses the MIMIC-IV Waveform Database (version 0.1.0), which includes photoplethysmography (PPG), electrocardiography (ECG), and arterial blood pressure (ABP) measurements from 198 subjects, totaling 200 records. The dataset follows a hierarchical structure based on subject-specific identifiers (subject_id) and hospital admission designations (hadm_id), enabling efficient retrieval via the WFDB (version 0.1.0), software framework [41]. To prevent data leakage, we stratified the dataset at the patient level, allocating 60% for training, 20% for validation, and 20% for testing. We retained only records containing both PPG and ABP signals, resulting in 60% of the initial dataset. The dataset maintains a uniform sampling frequency of 62.47 Hz and includes from 5 to 13 concurrent physiological signals per segment. Temporal duration varies from 0.2561 to 112,817.64 s, with segment counts ranging from 2 to 65 per patient.
We implemented a multi-stage preprocessing pipeline to ensure reliable cardiovascular parameter extraction. To eliminate transient artifacts, we removed the first 20 s of each signal before converting them into numerical arrays. After evaluating multiple segmentation strategies, we selected a 30 s window, yielding 1874 samples per segment, as the optimal balance between resolution and computational efficiency. We implement this initial 20 s removal step to systematically eliminate transient noise, such as movement-induced artifacts and sensor stabilization fluctuations, ensuring that our dataset consists solely of stable and physiologically meaningful signals.
To maintain physiological validity, we applied clinically accepted blood pressure thresholds, retaining systolic blood pressure (SBP) values between 75 and 180 mmHg and diastolic blood pressure (DBP) values between 40 and 110 mmHg. We ensured temporal alignment between PPG and ABP signals to preserve data integrity. These measures enhanced signal quality and ensured the dataset remained clinically relevant for model development and evaluation. Figure 1 confirms the effectiveness of the filtering process.
We applied advanced digital filtering techniques to enhance signal quality. A fourth-order Butterworth bandpass filter (0.5–8.0 Hz) refined PPG signals by preserving fundamental cardiac frequencies while suppressing low-frequency drift and high-frequency noise. ABP signals underwent high-pass filtering (0.05 Hz cutoff) to eliminate baseline wandering while maintaining key morphological features. We used zero-phase filtering with bidirectional coefficient application to preserve phase relationships essential for cardiovascular timing analysis. Figure 2 shows the before and after filtering of ABP and PPG signals, respectively. To standardize signal amplitude, we implemented min–max normalization with a stability factor (ε = 1 × 10−8) to prevent singular transformations (Equation (1)). This normalization was applied independently to each temporal window, ensuring consistent amplitude scaling without distorting local signal characteristics. Figure 3 illustrates the PPG and ABP signals before and after normalization. For precise waveform alignment, we developed a peak-to-peak alignment algorithm incorporating advanced signal processing techniques for PPG and ABP waveforms. This approach included systematic peak detection using physiological constraints and adaptive temporal thresholding (minimum inter-peak distance: one-tenth of the window duration), precise central peak localization through computational geometry (minimizing Euclidean distance from the temporal midpoint), and waveform translation using circular convolution for optimal peak alignment. Figure 4 shows a sample of the peak-to-peak alignment of ABP and PPG signals.
x normalized = x min x max x min x + ε
This alignment strategy established a precise temporal correspondence between cardiovascular waveform features, improving phase consistency and preserving key morphological characteristics. By enhancing temporal coherence, this methodology significantly strengthened subsequent analytical procedures, enabling reliable cross-subject comparisons and facilitating the detection of subtle physiological variations in cardiovascular dynamics.
In selecting our waveform alignment strategy, we considered the trade-off between algorithm complexity and computational efficiency, particularly for edge computing deployment. While simpler peak searching methods are less computationally demanding, we chose a more precise peak-to-peak alignment algorithm incorporating advanced signal processing techniques driven by the potential for enhanced blood pressure prediction accuracy through improved waveform alignment. To mitigate the increased computational cost associated with this more complex algorithm, we employed post-training quantization, which effectively reduces the model size and computational demands, making our framework suitable for resource-constrained edge devices.
Finally, we developed a feature extraction framework to analyze photoplethysmographic waveforms, capturing essential temporal, statistical, and morphological characteristics for blood pressure estimation. Temporal analysis included first through fourth-order statistical moments, incorporating measures of central tendency, dispersion, and higher-order moments, such as skewness and kurtosis. Amplitude-based features, including global extrema and peak-to-peak variations, provided insight into signal fluctuations across multiple time windows. Morphological analysis employed an optimized peak detection algorithm tailored to PPG waveform characteristics, using physiologically constrained temporal thresholds. Extracted features included peak frequency for heart rate variability, inter-peak interval statistics for cardiac rhythm assessment, and the temporal distribution of successive peaks. Amplitude-based features captured statistical moments of peak amplitudes, variations in peak heights, and systematic characterization of peak prominence.

4. Methodologies

This investigation presents a comprehensive evaluation of novel deep learning architectures optimized for edge computing environments, specifically designed for photoplethysmography-based blood pressure estimation. To comprehensively assess the landscape of relevant deep learning approaches, we selected four distinct architectural paradigms for rigorous comparison: a residual-enhanced convolutional network, an attentive BPNet, a transformer-based model, and a feature-engineered neural network. This selection was driven by several key considerations. First, these architectures represent a diverse spectrum of deep learning techniques widely applied in time-series analysis and signal processing, ranging from efficient convolutional networks to state-of-the-art attention-based models. Second, given this study’s focus on edge deployment, we deliberately included architectures known for varying degrees of computational complexity, allowing us to investigate the critical trade-offs between model accuracy and suitability for resource-constrained edge environments. Third, we aimed to explore both established and novel approaches, incorporating recent advancements like Transformers and Attention mechanisms alongside more traditional convolutional methods and a feature-engineered baseline.
Residual-Enhanced Convolutional Architecture: The residual-enhanced architecture implements a sophisticated hierarchical feature extraction paradigm, incorporating skip connections to mitigate gradient degradation while maintaining computational efficiency.
Key Components:
  • Initial Convolution: A convolution layer (64 filters, kernel size 7) extracts low-level features from raw PPG signals. Batch Normalization and ReLU activation stabilizes training dynamics.
  • Residual Blocks: Three residual blocks progressively increase filter depths (64→128→256) [Equation (2)].
F x = H x + x
where H x   represents two convolutional layers within the block. Skip connections ensure gradient flow during back propagation.
3.
Global Pooling: Global average pooling aggregates temporal features into a fixed size representation.
4.
Dense Layer: Fully connected layer (128→64→3) refine extracted features before predicting SBP, DBP, and MBP.
Attentive BP Net: Attentive BPNet integrates convolutional layers with an attention mechanism to focus on critical regions of the PPG signal, enhancing interpretability and computational efficiency. The architecture is optimized to capture both local and global temporal dependencies in the signal.
Key Components:
  • Initial Feature Extraction: The input layer processes raw PPG signals through a convolutional layer (3232 filters, kernel size 1515), followed by batch normalization and ReLU activation. Max-pooling reduces the temporal resolution while retaining salient features.
  • Hierarchical Feature Extraction: Two additional convolutional layers (6464 and 128,128 filters, kernel size 77) extract higher-order features. Each convolutional layer is followed by batch normalization, ReLU activation, max-pooling, and spatial dropout (p = 0.1) to prevent overfitting in time-series data.
  • Attention Mechanism: An attention layer computes importance weights for each time step using a dense layer with a tanh activation function. A softmax operation normalizes the weights across the time axis. The weighted feature map is computed through element-wise multiplication of the attention weights with the feature map.
  • Global Pooling: Global average pooling (mean trends) and global max pooling (extreme values) are applied to the weighted feature map. The pooled outputs are concatenated to form a compact representation capturing both average and extreme signal characteristics.
  • Dense Layers: Two fully connected layers (128,128 and 6464 neurons with ReLU activation) refine the features before outputting predictions for systolic (SBP), diastolic (DBP), and mean arterial pressure (MBP).
  • Loss Function: A custom loss function was designed to incorporate physiological constraints (Equation (3)):
L BP = MSE y true , y pred + λ max 0.15 y pred , SBP y pred , DBP
This penalizes physiologically implausible differences between SBP and DBP predictions.
Transformer-Based Model: The transformer-based model leverages self-attention mechanisms to capture long-range dependencies in PPG signals, which are critical for understanding temporal variations in cardiovascular dynamics [42].
Key Components:
  • Positional Encoding: Since transformers lack inherent temporal awareness, sinusoidal positional encodings are added to the input embeddings (Equations (4) and (5)):
P E pos , 2 i = sin pos 10,000 2 i d model
P E pos , 2 i + 1 = cos pos 10,000 2 i + 1 d model
2.
Embedding Layer: A dense layer maps the input signal into a higher-dimensional space ( d model = 256 ).
3.
Multi-Head Self-Attention: Four transformer layers are stacked, each consisting of multi-head self-attention (88 heads) followed by feed-forward networks. Residual connections and layer normalization stabilize training.
Attention Q , K , V = softmax Q K T d k V
4.
Global Pooling: The final transformer output is aggregated using global average pooling to reduce dimensionality while retaining salient features.
5.
Output Layers: Dense layers (6464 neurons with ReLU activation) refine the features before predicting SBP, DBP, and MBP.

4.1. Experimental Design

We optimized the training process to ensure robust convergence and generalization across model architectures. The models minimized Mean Squared Error (MSE) to predict systolic (SBP), diastolic (DBP), and mean blood pressure (MBP). Attentive BPNet enhanced accuracy by integrating physiological constraints into its custom loss function. We used Adam optimizer with initial learning rates of ( 1 × 10 3 ). A batch size exploration from 4 to 256 revealed that the Attentive BPNet performed best with smaller batches, benefiting from improved gradient estimation and lower memory overhead. To prevent overfitting, early stopping halted training after 10 epochs of no improvement. ReduceLROnPlateau decreased the learning rate by a factor of 0.2 when validation loss plateaued within a 5-epoch window. Model checkpointing preserved the best-performing model based on validation metrics. This training strategy ensured efficient learning, stability, and strong generalization while maintaining computational efficiency. In addition, for the feature engineering neural network, two dense layers were included, each followed by batch normalization and dropout (p = 0.3). The final dense layer outputs a single regression value for SBP/DBP/MBP.
We implemented a various post-training quantization, including dynamic range quantization, full integer quantization, float 16 quantization, and integer-only quantization on a TensorFlow Lite optimization framework to address computational and memory constraints in edge environments. This framework applies multiple quantization techniques, balancing model size, inference speed, and accuracy. A key component of the quantization process involves a calibration dataset generator, which samples 100 representative training data points to determine optimal activation quantization parameters. This systematic approach ensures accurate input distribution representation while minimizing quantization errors.
Dynamic range quantization applies asymmetric quantization for weights while keeping activations in floating-point. It transforms weights using W quantized = round scale W float + zero _ point , where scale and zero_point are determined per-tensor or per-channel to minimize quantization error. On the other hand, full integer quantization converts both weights and activations to 8-bit precision using q = round r s + z , where r is the real value, s is the scaling factor, z is the zero point, and q is the quantized value. Statistical analysis of representative data ensures minimal information loss. Float 16 Precision Optimization reduces storage requirements while maintaining computational precision. It represents weights in half-precision binary format as float   16 = 1 sign × 2 exponent 15 × 1 + fraction with 1-bit sign, 5-bit exponent, and 10-bit fraction. In addition, integer-only quantization aggressively optimizes models for edge deployment by fully converting inputs, weights, activations, and outputs to 8-bit integer format. This approach significantly reduces computational overhead while preserving model performance.
These optimizations enhance efficiency while maintaining accuracy, making deep learning models viable for real-time edge deployment. The table below compares the performance of original and quantized models.
To ensure reproducibility and transparency, we detail here the hardware configuration employed for both training and testing the deep learning models in this study. All experiments were conducted using the Google Colaboratory (Colab) platform, which provides access to cloud-based computational resources. The primary hardware components utilized were an NVIDIA A100 GPU for accelerated deep learning computations, a multi-core x86-64 architecture CPU (Intel Xeon), and approximately 12–16 GB of RAM. Data and models were stored using the cloud storage infrastructure provided by Google Colaboratory.

4.2. Experimental Results

We evaluate non-invasive blood pressure estimation methods, focusing on model performance, quantization efficacy, and deployment feasibility in resource-constrained environments. The evaluation follows clinical standards and incorporates statistical validation essential for medical device assessment. The assessment framework applies key statistical metrics to ensure accuracy and reliability. Root Mean Square Error (RMSE) detects large estimation errors that could impact clinical decisions. Mean Bias identifies systematic over- or under-estimation tendencies, preventing consistent prediction errors. Standard deviation measures prediction stability, ensuring reliability in clinical applications. Mean absolute error (MAE) quantifies average deviation from true values, providing a clear measure of accuracy.
Table 1 shows the experimental results of applying different deep-learning architectures for blood pressure estimation, highlighting the impact of different model selections and batch sizes. The residual-enhanced convolutional network consistently outperformed other models, achieving the lowest errors across all metrics. It maintained an SBP MAE of 6.36 mmHg, DBP MAE of 8.35 mmHg, and MBP MAE of 6.90 mmHg, demonstrating superior predictive accuracy. In contrast, the transformer-based model exhibited significantly higher errors, with SBP MAE ranging from 9.39 to 13.69 mmHg, and further deteriorated post-quantization, making it less suitable for real-time applications.
Batch size variations significantly affected performance. Attentive BPNet performed best with a batch size of 4 (SBP MAE 7.55 mmHg) but degraded with larger batches, reaching 17.46 mmHg at batch size 256. Similarly, the transformer-based model’s SBP MAE increased from 9.39 mmHg to 13.69 mmHg as batch size increased. In contrast, the residual-enhanced convolutional network remained stable across batch sizes, reinforcing its robustness and suitability for real-world deployment.
Table 2 shows the impact of quantization on model size reduction across different architectures. Attentive BPNet achieved the highest compression, reducing from 5.64 MB to 0.50 MB (91.13%), followed by the residual-enhanced network, which shrank from 1.40 MB to 0.13 MB (90.71%). The transformer-based model, despite its initial large size of 9.28 MB, compressed to 1.10 MB, achieving the lowest reduction rate (88.15%). While all models benefited from quantization, the transformer-based model remained the largest post-quantization, indicating inefficiency for edge deployment. Attentive BPNet achieved the most significant reduction but still struggled with stability in accuracy. The residual-enhanced network balanced both compression and performance, making it the most efficient model for resource-constrained environments.
Table 3 compares the effects of different quantization methods on model size, inference time, and performance across three architectures. The residual-enhanced convolutional network maintained the best balance between accuracy and efficiency, with the smallest post-quantization size (0.13 MB) and the lowest inference time (2.6 ms). Despite an over 90% reduction in size, its MAE (7.65) and RMSE (10.83) remained stable across quantization types, confirming its robustness for edge deployment. The attentive BPNet achieved the highest compression (91.13%), reducing from 5.64 MB to 0.5 MB, but inference time varied significantly. While full and integer-only quantization preserved accuracy, they increased latency to 36.63 ms and 37.79 ms, making real-time deployment less feasible. Dynamic range and float 16 quantization provided slightly better efficiency without significant accuracy loss. The transformer-based model performed the worst in efficiency. Despite reducing from 9.28 MB to 1.1 MB, inference time exploded to over 3000 ms in full and integer-only quantization, making it impractical for edge applications. Additionally, its MAE fluctuated drastically, reaching 60.65 in integer-only quantization, rendering the model unreliable. Overall, the residual-enhanced network consistently achieved low latency, minimal accuracy degradation, and high compression, making it the most viable option for edge deployment. The attentive BPNet remained a reasonable alternative with trade-offs in latency, while the transformer-based model proved inefficient for real-time applications.
To evaluate the calibration-free performance of the residual-enhanced network, we assessed its accuracy on a held-out evaluation dataset (20% of patients) comprised of subjects not seen during training or validation. On this independent dataset, the model achieved a mean absolute error (MAE) of 6.36 ± 8.86 mmHg for SBP, 8.35 ± 5.88 mmHg for DBP, and 6.90 ± 6.28 mmHg for MBP, demonstrating its ability to generalize to unseen individuals without subject-specific calibration.

5. Discussion

Regulatory bodies, such as the Association for the Advancement of Medical Instrumentation (AAMI), establish strict validation protocols for blood pressure monitoring devices. These standards require a mean difference of ±5 mmHg and a standard deviation below 8 mmHg, while the British Hypertension Society (BHS) mandates cumulative percentage thresholds for different accuracy grades. Evaluation of deep learning models showed varying compliance with these standards. Attentive BPNet initially failed to meet AAMI limits, with an SBP MAE of 13.27 mmHg. Architectural refinements improved performance, reducing SBP MAE to 7.55 mmHg, DBP MAE to 7.88 mmHg, and MBP MAE to 5.04 mmHg. While within clinical limits for mean blood pressure, systolic predictions remained near AAMI’s upper boundary. The transformer-based model exhibited marginal compliance, with SBP MAE at 9.39 mmHg, DBP MAE at 9.46 mmHg, and MBP MAE at 8.05 mmHg. However, post-quantization, its performance degraded, with MAE increasing to 10.87 mmHg, exceeding clinical thresholds.
The residual-enhanced network consistently met clinical standards, achieving an SBP MAE of 6.36 mmHg, DBP MAE of 8.35 mmHg, and MBP MAE of 6.90 mmHg. Post-quantization, it maintained accuracy, with MAE variation limited to ±0.12 mmHg and bias deviation below 0.2 mmHg. Attentive BPNet showed moderate stability but remained sensitive to quantization. The transformer-based model failed to meet AAMI standards post-quantization, with increased MAE and inference latency. The residual-enhanced network emerged as the most viable candidate for clinical deployment, balancing regulatory compliance, efficiency, and statistical reliability. It consistently met AAMI standards, maintained minimal deviation in key measurements, and demonstrated optimal compression (0.13 MB) with an inference time of 2.6 ms. Its stable performance across quantization schemes makes it the most practical choice for real-world blood pressure monitoring, particularly in resource-limited settings.
While our evaluation primarily aligns with AAMI standards, calculating cumulative errors in accordance with BHS guidelines represents an important future direction. Given the stable performance metrics, we expect our residual-enhanced convolutional network would meet the cumulative error thresholds defined by BHS. We acknowledge that the AAMI standard specifies compliance in terms of mean difference (bias) within ±5 mmHg and standard deviation under 8 mmHg, whereas our manuscript primarily reported mean absolute error (MAE). To ensure clarity, we explicitly calculated and confirmed that the residual-enhanced convolutional network’s mean difference (bias) is within the ±5 mmHg AAMI criterion. Thus, the model meets the defined regulatory standard.

6. Conclusions

This study advances non-invasive blood pressure estimation by systematically evaluating deep learning architectures, quantization strategies, and clinical validation protocols. The residual-enhanced network outperformed alternatives, achieving MAE values of 6.36, 8.35, and 6.90 mmHg for SBP, DBP, and MBP, respectively. It met and exceeded AAMI standards while maintaining stability under aggressive quantization. The model’s ability to preserve accuracy with a 90.71% reduction in size marks a significant advancement for embedded medical devices. The proposed quantization framework confirms that sophisticated deep learning models can operate efficiently in resource-constrained environments without compromising clinical precision. Achieving a 2.6 ms inference time while maintaining regulatory compliance represents a breakthrough for real-time monitoring applications.

Author Contributions

Conceptualization, A.V.D., S.S. and M.M.; methodology, A.V.D. and M.M.; software, A.V.D.; validation, A.V.D., S.S. and M.M.; formal analysis, A.V.D.; investigation, A.V.D.; resources, M.M.; data curation, A.V.D.; writing—original draft preparation, A.V.D.; writing—review and editing, M.M.; visualization, A.V.D.; supervision, M.M.; project administration, M.M. All authors have read and agreed to the published version of the manuscript.

Funding

Research reported in this publication was supported by the Division of Research and Innovation at San José State University under Award Number [#24-RSG-07-112].

Data Availability Statement

This study uses the MIMIC-IV Waveform Database (version 0.1.0), available at https://physionet.org/content/mimic4wdb/0.1.0/ (accessed on 9 March 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhou, Z.-B.; Cui, T.-R.; Li, D.; Jian, J.-M.; Li, Z.; Ji, S.-R.; Li, X.; Xu, J.-D.; Liu, H.-F.; Yang, Y.; et al. Wearable Continuous Blood Pressure Monitoring Devices Based on Pulse Wave Transit Time and Pulse Arrival Time: A Review. Materials 2023, 16, 2133. [Google Scholar] [CrossRef] [PubMed]
  2. Montagna, S.; Pengo, M.F.; Ferretti, S.; Borghi, C.; Ferri, C.; Grassi, G.; Muiesan, M.L.; Parati, G. Machine Learning in Hypertension Detection: A Study on World Hypertension Day Data. J. Med. Syst. 2023, 47, 1. [Google Scholar] [CrossRef]
  3. El Hajj, C.; Kyriacou, P.A. Cuffless and Continuous Blood Pressure Estimation from PPG Signals Using Recurrent Neural Networks. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; IEEE: New York, NY, USA, 2020; pp. 4269–4272. [Google Scholar] [CrossRef]
  4. Sun, B.; Bayes, S.; Abotaleb, A.M.; Hassan, M. The Case for tinyML in Healthcare: CNNs for Real-Time On-Edge Blood Pressure Estimation. In Proceedings of the SAC ’23: 38th ACM/SIGAPP Symposium on Applied Computing, Tallinn, Estonia, 27–31 March 2023; Association for Computing Machinery: New York, NY, USA, 2023; pp. 629–638. [Google Scholar] [CrossRef]
  5. Xing, X.; Sun, M. Optical blood pressure estimation with photoplethysmography and FFT-based neural networks. Biomed. Opt. Express 2016, 7, 3007–3020. [Google Scholar] [CrossRef]
  6. González, S.; Hsieh, W.-T.; Chen, T.P.-C. A benchmark for machine-learning based non-invasive blood pressure estimation using photoplethysmogram. Sci. Data 2023, 10, 149. [Google Scholar] [CrossRef] [PubMed]
  7. Pandit, J.A.; Lores, E.; Batlle, D. Cuffless Blood Pressure Monitoring. Clin. J. Am. Soc. Nephrol. 2020, 15, 1531–1538. [Google Scholar] [CrossRef]
  8. Kario, K. Management of Hypertension in the Digital Era. Hypertension 2020, 76, 640–650. [Google Scholar] [CrossRef]
  9. Wang, L.; Tian, S.; Zhu, R. A new method of continuous blood pressure monitoring using multichannel sensing signals on the wrist. Microsyst. Nanoeng. 2023, 9, 117. [Google Scholar] [CrossRef]
  10. Mukkamala, R.; Stergiou, G.S.; Avolio, A.P. Cuffless Blood Pressure Measurement. Annu. Rev. Biomed. Eng. 2022, 24, 203–230. [Google Scholar] [CrossRef]
  11. Athaya, T.; Choi, S. A Review of Noninvasive Methodologies to Estimate the Blood Pressure Waveform. Sensors 2022, 22, 3953. [Google Scholar] [CrossRef]
  12. Ding, X.; Zhang, Y.-T. Pulse transit time technique for cuffless unobtrusive blood pressure measurement: From theory to algorithm. Biomed. Eng. Lett. 2019, 9, 37–52. [Google Scholar] [CrossRef]
  13. Huynh, T.H.; Jafari, R.; Chung, W.-Y. An Accurate Bioimpedance Measurement System for Blood Pressure Monitoring. Sensors 2018, 18, 2095. [Google Scholar] [CrossRef] [PubMed]
  14. Ibrahim, B.; Jafari, R. Cuffless blood pressure monitoring from a wristband with calibration-free algorithms for sensing location based on bio-impedance sensor array and autoencoder. Sci. Rep. 2022, 12, 319. [Google Scholar] [CrossRef]
  15. Wang, T.-W.; Syu, J.-Y.; Chu, H.-W.; Sung, Y.-L.; Chou, L.; Escott, E.; Escott, O.; Lin, T.-T.; Lin, S.-F. Intelligent Bio-Impedance System for Personalized Continuous Blood Pressure Measurement. Biosensors 2022, 12, 150. [Google Scholar] [CrossRef] [PubMed]
  16. Berkelmans, G.F.N.; Kuipers, S.; Westerhof, B.E.; Man, A.M.E.S.-D.; Smulders, Y.M. Comparing volume-clamp method and intra-arterial blood pressure measurements in patients with atrial fibrillation admitted to the intensive or medium care unit. J. Clin. Monit. Comput. 2018, 32, 439–446. [Google Scholar] [CrossRef]
  17. Yamakoshi, K.; Kamiya, A.; Shimazu, H.; Ito, H.; Togawa, T. Noninvasive automatic monitoring of instantaneous arterial blood pressure using the vascular unloading technique. Med. Biol. Eng. Comput. 1983, 21, 557–565. [Google Scholar] [CrossRef] [PubMed]
  18. Zakrzewski, A.M. Arterial Blood Pressure Estimation Using Ultrasound. Ph.D. Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 2017. Available online: https://dspace.mit.edu/handle/1721.1/111743 (accessed on 22 September 2024).
  19. Meusel, M.; Wegerich, P.; Bode, B.; Stawschenko, E.; Kusche-Vihrog, K.; Hellbrück, H.; Gehring, H. Measurement of Blood Pressure by Ultrasound—The Applicability of Devices, Algorithms and a View in Local Hemodynamics. Diagnostics 2021, 11, 2255. [Google Scholar] [CrossRef]
  20. Desebbe, O.; Anas, C.; Alexander, B.; Kouz, K.; Knebel, J.-F.; Schoettker, P.; Creteur, J.; Vincent, J.-L.; Joosten, A. Evaluation of a novel optical smartphone blood pressure application: A method comparison study against invasive arterial blood pressure monitoring in intensive care unit patients. BMC Anesthesiol. 2022, 22, 259. [Google Scholar] [CrossRef]
  21. Ogedegbe, G.; Pickering, T. Principles and techniques of blood pressure measurement. Cardiol. Clin. 2010, 28, 571–586. [Google Scholar] [CrossRef]
  22. Kurylyak, Y.; Lamonaca, F.; Grimaldi, D. A Neural Network-based method for continuous blood pressure estimation from a PPG signal. In Proceedings of the 2013 IEEE International Instrumentation and Measurement Technology Conference (I2MTC), Minneapolis, MN, USA, 6–9 May 2013; pp. 280–283. [Google Scholar] [CrossRef]
  23. Schlesinger, O.; Vigderhouse, N.; Eytan, D.; Moshe, Y. Blood Pressure Estimation From PPG Signals Using Convolutional Neural Networks And Siamese Network. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 1135–1139. [Google Scholar] [CrossRef]
  24. Casadei, B.C.; Gumiero, A.; Tantillo, G.; Della Torre, L.; Olmo, G. Systolic Blood Pressure Estimation from PPG Signal Using ANN. Electronics 2022, 11, 2909. [Google Scholar] [CrossRef]
  25. Mukkamala, R.; Hahn, J.-O.; Chandrasekhar, A. 11—Photoplethysmography in noninvasive blood pressure monitoring. In Photoplethysmography; Allen, J., Kyriacou, P., Eds.; Academic Press: Cambridge, MA, USA, 2022; pp. 359–400. [Google Scholar] [CrossRef]
  26. Abadade, Y.; Temouden, A.; Bamoumen, H.; Benamar, N.; Chtouki, Y.; Hafid, A.S. A Comprehensive Survey on TinyML. IEEE Access 2023, 11, 96892–96922. [Google Scholar] [CrossRef]
  27. Zaidi, S.A.R.; Hayajneh, A.M.; Hafeez, M.; Ahmed, Q.Z. Unlocking Edge Intelligence Through Tiny Machine Learning (TinyML). IEEE Access 2022, 10, 100867–100877. [Google Scholar] [CrossRef]
  28. Han, H.; Siebert, J. TinyML: A Systematic Review and Synthesis of Existing Research. In Proceedings of the 2022 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Jeju Island, Republic of Korea, 21–24 February 2022; pp. 269–274. [Google Scholar] [CrossRef]
  29. Kallimani, R.; Pai, K.; Raghuwanshi, P.; Iyer, S.; López, O.L.A. TinyML: Tools, applications, challenges, and future research directions. Multimed. Tools Appl. 2024, 83, 29015–29045. [Google Scholar] [CrossRef]
  30. Hymel, S.; Banbury, C.; Situnayake, D.; Elium, A.; Ward, C.; Kelcey, M.; Baaijens, M.; Majchrzycki, M.; Plunkett, J.; Tischler, D.; et al. Edge Impulse: An MLOps Platform for Tiny Machine Learning. arXiv 2023, arXiv:2212.03332. [Google Scholar] [CrossRef]
  31. Tsoukas, V.; Boumpa, E.; Giannakas, G.; Kakarountas, A. A Review of Machine Learning and TinyML in Healthcare. In Proceedings of the PCI ’21: 25th Pan-Hellenic Conference on Informatics, Volos, Greece, 26–28 November 2021; Association for Computing Machinery: New York, NY, USA, 2022; pp. 69–73. [Google Scholar] [CrossRef]
  32. Anbukarasu, P.; Nanisetty, S.; Tata, G.; Ray, N. Tiny-HR: Towards an interpretable machine learning pipeline for heart rate estimation on edge devices. arXiv 2022, arXiv:2208.07981. [Google Scholar] [CrossRef]
  33. Tang, Q.; Chen, Z.; Ward, R.; Menon, C.; Elgendi, M. Subject-Based Model for Reconstructing Arterial Blood Pressure from Photoplethysmogram. Bioengineering 2022, 9, 402. [Google Scholar] [CrossRef] [PubMed]
  34. Pal, R.; Rudas, A.; Kim, S.; Chiang, J.N.; Cannesson, M. A signal processing tool for extracting features from arterial blood pressure and photoplethysmography waveforms. medRxiv 2024. [Google Scholar] [CrossRef]
  35. Beh, W.-K.; Yang, Y.-C.; Wu, A.-Y. Quality-Aware Signal Processing Mechanism of PPG Signal for Long-Term Heart Rate Monitoring. Sensors 2024, 24, 3901. [Google Scholar] [CrossRef]
  36. Park, J.; Seok, H.S.; Kim, S.-S.; Shin, H. Photoplethysmogram Analysis and Applications: An Integrative Review. Front. Physiol. 2022, 12, 808451. [Google Scholar] [CrossRef]
  37. Joseph, G.; Joseph, A.; Titus, G.; Thomas, R.M.; Jose, D. Photoplethysmogram (PPG) signal analysis and wavelet de-noising. In Proceedings of the 2014 Annual International Conference on Emerging Research Areas: Magnetics, Machines and Drives (AICERA/iCMMD), Kottayam, India, 24–26 July 2014; pp. 1–5. [Google Scholar] [CrossRef]
  38. Kluska, P.; Zięba, M. Post-training Quantization Methods for Deep Learning Models. In Intelligent Information and Database Systems; Nguyen, N.T., Jearanaitanakij, K., Selamat, A., Trawiński, B., Chittayasothorn, S., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 467–479. [Google Scholar] [CrossRef]
  39. Kwasniewska, A.; Szankin, M.; Ozga, M.; Wolfe, J.; Das, A.; Zajac, A.; Ruminski, J.; Rad, P. Deep Learning Optimization for Edge Devices: Analysis of Training Quantization Parameters. In Proceedings of the IECON 2019—45th Annual Conference of the IEEE Industrial Electronics Society, Lisbon, Portugal, 14–17 October 2019; pp. 96–101. [Google Scholar] [CrossRef]
  40. Wu, H.; Judd, P.; Zhang, X.; Isaev, M.; Micikevicius, P. Integer Quantization for Deep Learning Inference: Principles and Empirical Evaluation. arXiv 2020, arXiv:2004.09602. [Google Scholar] [CrossRef]
  41. Kulkarni, U.; Meena, S.M.; Joshua, P.; Kruthika, K.; Platel, G.; Jadhav, K.; Singh, A. Performance Improvements in Quantization Aware Training and Appreciation of Low Precision Computation in Deep Learning. In Advances in Signal Processing and Intelligent Recognition Systems; Thampi, S.M., Krishnan, S., Hegde, R.M., Ciuonzo, D., Hanne, T., Kannan, R.J., Eds.; Springer: Singapore, 2021; pp. 90–107. [Google Scholar] [CrossRef]
  42. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
Figure 1. A distribution plot confirming the filtering process.
Figure 1. A distribution plot confirming the filtering process.
Electronics 14 01300 g001
Figure 2. A sample of PPG and ABP signals before and after filtering.
Figure 2. A sample of PPG and ABP signals before and after filtering.
Electronics 14 01300 g002
Figure 3. A sample of PPG and ABP signals before and after normalization.
Figure 3. A sample of PPG and ABP signals before and after normalization.
Electronics 14 01300 g003
Figure 4. A sample of peak-to-peak alignment of ABP and PPG signals.
Figure 4. A sample of peak-to-peak alignment of ABP and PPG signals.
Electronics 14 01300 g004
Table 1. Performance comparison of deep learning models for blood pressure estimation across different batch sizes, highlighting the impact on MAE, RMSE, and model stability. The residual-enhanced convolutional network demonstrates consistent accuracy.
Table 1. Performance comparison of deep learning models for blood pressure estimation across different batch sizes, highlighting the impact on MAE, RMSE, and model stability. The residual-enhanced convolutional network demonstrates consistent accuracy.
ModelSBP MAESBP RMSESBP MESBP SDSBP R2DBP MAEDBP RMSEDBP MEDBP SDDBP R2MBP MAEMBP RMSEMBP MEMBP SDMBP R2
Attentive BPNet11.5513.78−10.658.75−0.444.095.361.765.060.154.566.09−2.335.630.28
Attentive BPNet—batch size 6413.2716.59−11.2112.24−1.085.226.861.996.56−0.406.258.30−2.377.95−0.34
Attentive BPNetl—Batch Size 87.7710.85−5.009.630.116.678.155.625.90−0.975.136.912.046.600.07
Attentive BPNet—Batch Size 47.5510.44−5.748.750.187.888.967.315.19−1.395.046.472.955.760.19
Attentive BPNet—Batch Size 1610.7614.96−3.1514.63−0.698.369.675.687.82−1.777.859.822.059.61−0.88
Attentive BPNet- Wide and Deep batch 3212.5716.72−5.1015.93−1.1214.1716.0212.929.47−6.6210.8612.916.3411.24−2.24
Attentive BPNet—Wide and Deep batch 414.1018.33−4.1817.85−1.5413.1015.2411.5010.01−5.9011.4913.816.2712.30−2.71
Attentive BPNet—Wide and Deep batch 25617.4621.884.9221.31−2.6219.9822.8918.9812.80−14.5618.3521.9215.5415.46−8.35
TransfoRhythm9.3912.110.3712.11−0.119.4610.648.716.11−2.368.059.775.797.87−0.86
Feature Engineering—Separate Models10.2212.685.4511.45−0.2219.2220.6518.998.11−11.6612.0813.4610.758.09−2.52
Feature Engineering—Separate Models—iteration 210.1613.97−6.4912.37−0.4811.0512.6910.077.72−3.7810.4011.758.877.70−1.69
Feature Engineering—Separate Models—iteration 310.5213.065.5911.81−0.2910.8912.129.237.86−3.3612.4413.8610.519.03−2.74
Feature Engineering—Separate Models—iteration 413.3815.699.6312.39−0.869.0710.626.268.58−2.3511.1012.668.619.28−2.12
Residual Enhaced Convolutional Network6.368.93−1.118.860.408.359.787.815.88−1.846.908.405.586.28−0.37
Residual Enhaced Convolutional Network—deeper network12.7315.6310.5211.55−0.8513.0714.9612.677.95−5.6412.8215.0312.198.79−3.40
Table 2. Model size reduction after quantization, highlighting compression efficiency across architectures.
Table 2. Model size reduction after quantization, highlighting compression efficiency across architectures.
ArchitectureOriginal SizePost-QuantizationReduction
Attentive BPNet5.64 MB0.50 MB91.13%
Transformer-Based9.28 MB1.10 MB88.15%
Residual-Enhanced1.40 MB0.13 MB90.71%
Table 3. Impact of quantization methods on model size, inference time, and accuracy across architectures. The residual-enhanced network maintains stability, while the transformer-based model suffers from extreme latency and accuracy degradation.
Table 3. Impact of quantization methods on model size, inference time, and accuracy across architectures. The residual-enhanced network maintains stability, while the transformer-based model suffers from extreme latency and accuracy degradation.
Attentive BP Net
ModelOrginal ModelDynamic Range Quantized ModelFloat 16Full Integer QuantizationInteger-Only Quantization
Model Size (MB)5.640.50.940.520.52
Size reduction (%)-−91.13−83.3333−90.78−90.78
Inference Time30.4624.7825.0236.6337.78
MAE7.847.847.847.787.78
RMSE10.4210.4110.4210.3610,365
ME (Bias)−0.070−0.07−0.19−0.19
SD10.4210.4110.4210.3610.36
Transformer-Based Model
Model Size (MB)9.281.772.011.11.1
Size reduction (%) −80.92−78.34−88.14−88.14
Inference Time32.611793.351249.693087.913034.05
MAE10.8710.8610.878.2560.65
RMSE12.8512.8412.8511.2562.33
ME (Bias)7.797.797.8−3.45−60.67
SD10.2110.2110.2110.7114.3
Residual-Enhanced Convolutional Network
Model Size (MB)1.40.230.520.130.13
Size reduction (%) −83.57−62.85−90.71−90.71
Inference Time6.234.661.22.62.6
MAE7.777.757.777.657.65
RMSE10.8510.8410.8510.8310.83
ME (Bias)−5−4.9−5−5−5.2
SD9.639.649.639.699.64
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Devadasan, A.V.; Sengupta, S.; Masum, M. Deep Learning for Non-Invasive Blood Pressure Monitoring: Model Performance and Quantization Trade-Offs. Electronics 2025, 14, 1300. https://doi.org/10.3390/electronics14071300

AMA Style

Devadasan AV, Sengupta S, Masum M. Deep Learning for Non-Invasive Blood Pressure Monitoring: Model Performance and Quantization Trade-Offs. Electronics. 2025; 14(7):1300. https://doi.org/10.3390/electronics14071300

Chicago/Turabian Style

Devadasan, Anbu Valluvan, Saptarshi Sengupta, and Mohammad Masum. 2025. "Deep Learning for Non-Invasive Blood Pressure Monitoring: Model Performance and Quantization Trade-Offs" Electronics 14, no. 7: 1300. https://doi.org/10.3390/electronics14071300

APA Style

Devadasan, A. V., Sengupta, S., & Masum, M. (2025). Deep Learning for Non-Invasive Blood Pressure Monitoring: Model Performance and Quantization Trade-Offs. Electronics, 14(7), 1300. https://doi.org/10.3390/electronics14071300

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop