Using Data Augmentation to Improve the Accuracy of Blood Pressure Measurement Based on Photoplethysmography

: Convenient and accurate blood pressure (BP) measurement is of great importance in both clinical and daily life. Although deep learning (DL) can achieve cuff-less BP measurement based on Photoplethysmography (PPG), the performance of DL is affected by few-shot data. Data augmentation becomes an effective way to enhance the size of the training data. In this paper, we use cropping, flipping, DTW barycentric averaging (DBA), generative adversarial network (GAN) and variational auto-encoder (VAE) for the data augmentation of PPG. Furthermore, a PE–CNN–GRU model is designed for cuff-less BP measurement applying position encoding (PE), convolutional neural networks (CNNs) and gated recurrent unit (GRU) networks. Experiment results based on real-life datasets show that VAE is the most suitable method for PPG data augmentation, which can reduce the mean absolute error (MAE) of PE–CNN–GRU when measuring systolic blood pressure (SBP) and diastolic blood pressure (DBP) by 18.80% and 19.84%. After the data augmentation of PPG, PE–CNN–GRU achieves accurate and cuff-less BP measurement, thus providing convenient support for preventing cardiovascular diseases.


Introduction
Blood pressure (BP) is one of the most important indicators of the cardiovascular system.Studies show that controlling BP within a reasonable range can effectively prevent cardiovascular diseases [1,2].Therefore, accurate BP measurement is of great significance in both clinical and daily health.However, conventional cuff-based BP measurement methods, such as Korotkoff sound auscultation [3] and the oscillometric method [4], are quite inconvenient for patients due to the requirement of wearing a cuff on the arm.
Recently, cuff-less BP measurement based on pulse wave has attracted widespread research interest.The pulse wave is caused by the cyclical change of the contraction and relaxation of the ventricles [5] and contains rich physiological information, such as BP.Furthermore, pulse waves can be collected with Photoplethysmography (PPG) thanks to the advancement of hardware.Smartwatches or fingertip oximeters can achieve continuous PPG signal collection to support cuff-less BP measurement.However, the nonlinear relationship between PPG and BP is difficult to calculate directly [6].Deep learning provides a data-driven method for BP measurement based on PPG [7].
There are mainly two kinds of data-driven BP measurement methods based on PPG: the feature-based method and the data-based method.The main idea of the featurebased method is to extract hand-crafted features, then explore the relationship between features and BP values [8][9][10].The data-based method directly uses neural networks to fit the relationship between PPG and BP, without the need for hand-crafted feature extraction [11][12][13][14][15][16][17].Generally, the accuracy of the data-based method is usually lower than the feature-based methods.
The data-driven BP measurement method, whether based on feature or data, requires the support of massive data for training.Most existing works are based on a Multiparameter Intelligent Monitoring in Intensive Care (MIMIC) dataset, which currently has four open-source versions [18][19][20][21].Although MIMIC contains physiological data such as BP and PPG from tens of thousands of patients, there exist issues of noise and missing values, resulting in a limited amount of available high-quality data.Considering the high cost of data collection, we can use data augmentation to expand the number of training samples to improve the performance of data-driven BP measurement methods.
The main idea of data augmentation is to generate synthetic data that cover the unexplored input space while maintaining correct data labels.Data augmentation performs well in computer vision [22] and natural language processing [23], but less attention has been paid to data augmentation for time series data [24].
Iwana et al. [25] introduced data augmentation methods of time series based on transformation, pattern mixing, generative models, and decomposition.Experiments on 128 datasets show that data augmentation can improve the accuracy of time series classification.Gao et al. [26] designed a time-frequency data augmentation scheme to improve the performance of time series anomaly recognition.In the time domain, they generated synthetic data through flipping and downsampling.In the frequency domain, they generated synthetic data through making perturbations in magnitude and phase.
There is currently little research on PPG data augmentation.Kiyasseh et al. [27] proposed a data augmentation method based on conditional generative adversarial networks (CGANs), which can lead to an improvement of medical diagnosis tasks by up to 29%.Similarly, Ding et al. [28] designed log-spectral matching GAN (LSM-GAN) to alleviate the class imbalance in a PPG dataset.Experiments show that using LSM-GAN for PPG data augmentation can improve the accuracy of atrial fibrillation detection.Mazumder et al. [29] proved that generating synthetic PPG data to train classifiers can improve performance for coronary artery disease.Variational autoencoder (VAE) can outperform a baseline GAN architecture.
Although some research shows that generating synthetic time series data can improve the performance of regression models [30,31], there is a lack of research on data augmentation in the regression problem of BP measurement based on PPG.Song et al. [32] paid attention to data augmentation in BP measurement, but they used multivariate Gaussian distribution to generate PPG features rather than raw PPG data.Wu et al. [33] designed GAN to generate synthetic remote photoplethysmography (rPPG), which can reduce errors in BP measurement.rPPG is extracted from face image and differs from fingertip PPG.In this paper, we use the data augmentation of PPG to improve the accuracy of cuff-less BP measurements.The contributions of this paper are as follows: 1.
In order to improve the accuracy of BP measurement based on few-show PPG data, we design five PPG data augmentation methods: flipping, cropping, DTW barycentric averaging (DBA), generative adversarial network (GAN) and variational auto-encoder (VAE).Each method can generate synthetic PPG data to expand the size of dataset.

2.
We design a data-driven model applying position encoding (PE), a convolutional neural network (CNN) and a gated recurrent unit (GRU) network to fully utilize the sequence information in PPG.The well-trained model end-to-end outputs BP values based on PPG data as inputs, thus realizing cuff-less BP measurement.

3.
We perform valid experiments based on a MIMIC II dataset.Results show that VAE improves the measurement accuracy most significantly.PE-CNN-GRU after PPG data augmentation performs well in cuff-less BP measurement compared to existing works.
The remainder of this paper is organized as follows.In Section 2, we introduce the preprocessing of PPG and BP data.In Section 3, we introduce all the PPG data augmentation methods.In Section 4, we introduce the PE-CNN-GRU model for BP measurement based on PPG.The numerical results and performance analysis are shown in Section 5. Finally, we make brief conclusions in Section 6.

Data Preparation
In this paper, we use PPG and BP data from the processed sub-dataset of MIMIC II.Although Kachuee et al. [34] smoothed signals using average filtering and removed abnormal and irregular segments to form this sub-dataset, there still exist artifacts in PPG.It is necessary to design preprocessing methods based on characteristics of PPG data.

Detection of Artifacts in PPG
Artifacts are unwanted or erroneous components that distort the original signal, which can be caused by various factors such as motion and sensor displacement, etc. Artifacts can seriously affect the quality of PPG, resulting in the PPG data being unacceptable for model training; thus, it is important to detect and remove artifacts in PPG [35].However, existing works often focus on the denoising of PPG signals but neglect the detection and removal of artifacts.
Due to the uncertainty of collection devices and test subjects, artifacts are often randomly embedded in PPG signals.In order to detect the range of artifacts accurately, we first divide PPG signals into non-overlapping segments for every 1000 data points.Pulse waves change with the regular contraction and relaxation of the heart, so PPG is a kind of quasi-periodic signal [36].In detail, PPG shows certain periodicity, but the waveform of each period is not completely consistent.There is even no fixed length of each period.However, there is a strong correlation between periods of PPG, and artifacts break the correlation between normal periods.Therefore, we use the Pearson correlation coefficient to evaluate the correlation between periods of PPG.
Considering that the principal wave of PPG is the most obvious and easy to capture, we use PeakUtils package in Python to capture the principal wave.Then, there is a period of PPG between the two principal waves.For each segment of PPG, assuming N + 1 principal waves are obtained, the index is represented as {i 1 , i 2 , • • • , i N+1 }.There are N periods between these N + 1 principal waves, which are denoted by {S 1 , S 2 , • • • , S N }.
The Pearson correlation coefficient is the quotient of the covariance and standard deviation of two sets of data.Generally, the length of the two sets of data needs to be equal when calculating the Pearson correlation coefficient.Considering that the length of the PPG period in this paper is not fixed, we design a calculation method of correlation coefficient for PPG: Cov(S i (1:l i ),S j (1+n:l i +n)) where Cov(•) denotes the covariance of two variables, and σ denotes the standard deviation of two variables.Then, we can construct a correlation coefficient matrix for each PPG segment, which is defined as: where r i,j represents the correlation coefficient between S i and S j .Obviously, r i,i = 1, r i,j = r j,i .Furthermore, we designed the quality index of pulse wave (QIPW) to evaluate the state of artifacts in PPG segments, which is calculated as: The value range of QIPW is [−1, 1].If the value of QIPW is close to 1, there are less artifacts within the PPG segment.Figure 1 displays PPG signal segments in different QIPW.The QIPW in the left figure is 0.9963.It can be seen that there are no artifacts in this segment, and the signal quality is good.The QIPW in the middle figure is 0.8502, and it can be clearly seen that there are artifacts in this segment.The QIPW in the right figure is 0.669.This segment has relatively severe artifacts.As a result, QIPW is very sensitive to artifacts and has strong robustness.

Formation of PPG-BP Samples
After calculating QIPW for all segments, we can set a reasonable threshold.In this paper, we set the threshold value to 0.99; then, we discard PPG segments below the threshold and only select PPG segments above the threshold for further processing.About 59% of PPG segments are discarded.These PPG segments with high QIPW, i.e., highquality PPG segments, are used to form the PPG-BP samples.
Inspired by [37], we segment PPG and reference BP signals into non-overlapping windows with respect to the position of the cardiac periods.We adopt a variable PPG window, i.e., each window contains a fixed number of periods, rather than setting a fixed length of window.The collection frequency in MIMIC II is 125 Hz, while the heart rate of normal people is between 60 and 100 beats per second.Therefore, the length of each period is within 1 s, which is less than 125 data points.We set up a PPG window for every four principal waves.Each window contains three complete periods, i.e., approximately 200 to 400 data points.The physiological information contained in these three complete periods of PPG can effectively support BP measurements during this window.Considering that BP values fluctuate over time, dividing the PPG into windows can achieve more accurate BP measurements than directly using the PPG segment with a long-time scale.
The BP measurement method based on DL requires fixed-length input, so it is necessary to unify the length of PPG windows.We use Fast Fourier Transform (FFT) to uniformly resample the PPG window to 256 data points.Examples of resampled PPG windows are shown in Figure 2. It can be seen that we obtain high-quality PPG windows after data preprocessing.The SBP and DBP associated with each PPG window are calculated as follows.For each PPG segment with 1000 data points, the corresponding ABP window is located with the same starting and ending time steps.The maximum value and minimum value of this ABP window are used as the reference SBP and DBP values.

Data Augmentation of PPG
Due to the amount of preprocessed data being relatively small, it is prone to over-fitting when training DL models for BP measurements.In this paper, we introduce different data augmentation methods to generate synthetic PPG data to improve the accuracy of BP measurements.The illustration of the PPG data augmentation pipeline is shown in Figure 3.

Cropping and Flipping
Cropping and flipping are both basic data augmentation methods in time domains [24].In the cropping method, we crop each PPG window in a certain proportion to generate new PPG windows.The deep learning model needs a fixed length of the input, and it is necessary to use zero-padding to unify the newly formed PPG window length to 256 data points.The processing of cropping is denoted as follows: where N is the length of PPG windows.k i ∈ (0, 1) is the random cropping proportion.round() is the rounding operation.An example of synthetic PPG windows generated by cropping is shown in Figure 4b.In the flipping method, we flip the PPG window left and right or up and down to form new PPG windows.Compared to the clipping method, flipping does not change the length of the PPG windows, so there is no need for further zero-padding.The processing of cropping is denoted as follows: where mean(P i ) is the mean value of P i , which means we flip the PPG windows up and down on the mean value axis.An example of synthetic PPG windows generated by flipping is shown in Figure 4c.

DTW Barycentric Averaging
We introduce DTW barycentric averaging in [25] to achieve the data augmentation of PPG.Dynamic time warping (DTW) is a method of elastic measures, which is commonly used to measure the distance between two temporal sequences.Compared to the lock-step measures, DTW linearly scales one of the sequences (i.e., warping).Each point from this sequence can be matched with one or more points from the other sequence when calculating distances, and the mapping of points must be monotonically increasing.DTW can find the nearest distance as much as possible, thus better measuring the similarity between two temporal sequences.
The main idea of DBA is to calculate the weighted average time series of a set of time series based on DTW, which minimizes where DTW means calculated distance based on DTW.w i is the weight, which means each time series contributes differently to the final average.P can be used as a new synthetic sample.
In this paper, we randomly select two samples each time to calculate a synthetic PPG window using DBA.An example of synthetic PPG windows generated by DBA is shown in Figure 4d.We can find that the generated PPG window has a similar trend to the original data.However, there are morphological differences between the synthetic data and the original data in peaks and valleys.

Generative Adversarial Network
GAN is one of the most representative data augmentation methods in the field of computer vision [38].Considering that PPG can be seen as image data, we use GAN to generate synthetic PPG data to explore whether GAN can achieve valid PPG data augmentation.
The structure of GAN is shown in Figure 5. GAN consists of a generator and a discriminator.The generator G takes random noise as input to generate synthetic data.The discriminator D discriminates between the synthetic data and the real data.During the training of GAN, the generator hopes that synthetic data is as close to the real data as possible, which requires the accuracy of the discriminator to become worse, but the discriminator requires the accuracy to become better.Therefore, the generator and the discriminator form a two-player minimax game [39].The well-trained generator in GAN that can generate data has exactly the same probability distribution as the real data, i.e., the generator generates synthetic data that are the same as the real data.
As shown in Table 1, we design the generator and discriminator in GAN based on the characteristics of PPG data.We set the input of the generator as a one-dimensional random noise with a length of 256, which follows a normal distribution.The input noise is encoded with a fully connected (FC1) layer containing 128 neurons.Through a fully connected (FC2) layer containing 128 neurons and a fully connected layer (FC3) containing 256 neurons, the encoded vector is decoded, and synthetic PPG data is achieved.We also use batch normalization layers between fully connected layers.We design the discriminator as a Multi-Layer Perception (MLP).In detail, the input passes through fully connected layers consisting of 512 neurons (FC1), 256 neurons (FC2), and 1 neuron (FC3) to output the classification results.We also used batch normalization layers in the discriminator.As shown in Figure 4e, the synthetic PPG window generated by GAN has a similar morphology and trend to the original data.Although we have fine-tuned the structure and training epochs of GAN, there still exist burrs in generated PPG data.

Variational Auto-Encoder
We design and train a variational auto-encoder (VAE) network for data augmentation.VAE is considered one of the most important methods in the field of deep generative models.Unlike a traditional auto-encoder (AE), which describes latent spaces through numerical means, VAE describes observations of latent spaces as probability distribution.
As shown in Figure 6, VAE consists of an encoder and a decoder.During training, the input of VAE is real PPG data.The encoder maps the input data into a Gaussian probability density function z in latent space.New samples can be drawn from the Gaussian probability for generating synthetic data via the decoder.The goal of training is to minimize the reconstruction error between the encoded-decoded data and the initial data.The loss function of VAE is shown as follows: In Equation (7), the first component ∥x − x∥ 2 is the reconstruction loss, which ensures the output x closely matches the original input x.The second component KL(N(µ x , σ x ), N(0, 1)) is the regularization term, where KL means the Kullback-Leibler Divergence.The regularization term ensures that the distribution of latent space conforms to a standard Gaussian distribution, which can control latent variability and encourage probabilistic modeling.The loss function of VAE balances accurate reconstruction with regularization, making it more powerful for data representation and generation.We design the structure of the VAE based on the characteristics of PPG data, which is shown in Table 2.The input of the encoder is set to 256.The input layer is followed by a hidden layer (FC1) with 128 neurons.The length of latent space is set to 64.In detail, two fully connected layers (FC1 and FC2) are used as the mean vector and standard deviation (SD) vector in latent space.Then, a sampling layer is designed to achieve sampling operation.The hidden layer (FC1) of the decoder has 128 neurons.Finally, a fully connected layer (FC2) with 256 neurons is used to output the synthetic PPG windows.The synthetic PPG windows generated by VAE is shown in Figure 4f.We can find that the generated data are very close to the original data in both morphology and trend.
The process of using DL models to measure BP based on PPG is to fit the approximate function f , which maps P i to the approximate measured BP value Bi .The closer the approximate value Bi is to the real value B i , the better the performance of the model.

Position Encoding
In our previous works [16,17], we have proved that combining convolutional neural networks with recurrent neural networks (LSTM or GRU) can achieve good performance in BP measurement.However, the receptive fields of the first few layers in a CNN are relatively small, and they are not enough to effectively extract the features of PPG solely based on CNN.In this paper, we introduce the position encoding (PE) technology in the Transformer [40] to capture the sequence information in PPG.
The main idea of PE is to encode the sequence information through the position index, which is defined as follows: where P s is the sequence information, and p is the position index.Then, the information is mapped into high-dimensional space using the position encoding basis p b , which is defined as follows: where j is the dimension index.Based on p b , position encoding vectors P′ ∈ R d in highdimensional spaces can be defined as follows: In order to enhance the ability of the proposed model for nonlinear problems, we need to introduce the nonlinear function F to further calculate the position encoding vector: where P(p, j) is the final position encoding vector, and F is defined as follows:

The Receptive Field of CNN
When designing a CNN, we need to consider the relationship of the length of input data and the receptive field of CNN.The receptive field is the region in the input space that a CNN feature is affected by.The size of the receptive field can represent the ability of the feature map in capturing information.
In general, the convolutional kernel is much smaller than the input size; thus, the kernel can only capture information of the input for a short period of time rather than capturing information in a broader time dimension.However, increasing the size of the convolutional kernel results in more parameters and a higher computational complexity of neural networks.In fact, the ability of a CNN to capture information on time sequence is not solely determined by the size of the convolutional kernel.Instead, the receptive field size of the final output feature map has a significant impact on the capture ability.
Given a CNN, the receptive field of the output feature map is calculated as follows: where k l is the kernel size of the l-th layer, and s i is the stride of kernels in the i-th layer.Specially, s 0 is the stride of the input, so we set s 0 = 1.

The Architecture of PE-CNN-GRU Model
The architecture of the proposed PE-CNN-GRU model is shown in Figure 7. Due to the fact that the dimensionality of the input and the dimensionality of the position encoding vector may be different, we need to use two functions ϕ and ψ to map input and position encoding vector to the same dimension [41].ϕ and ψ are implemented via two convolutional layers.In each layer, the number of channels is 16, the kernel size is 1 and the stride is 1 (denoted by (16, 1, 1) in Figure 7).The input and position encoding vector are inputted into the convolutional layer separately.Then, they are fused using element wise add and further inputted into CNN.In order to capture the global feature from the input PPG data, we hope that the receptive field of the final output feature map r f is greater than or equal to the length of the input.Therefore, we set up seven convolutional modules in the proposed model.Each module consists of a convolutional layer, an activation function layer (ReLU layer), and a batch normalization layer.Referring to classic CNN such as VGG16 [42] and ResNet [43], as the number of layers deepens, the size of the kernels shows a decreasing trend, and the number of channels shows an increasing trend.Taking the first convolutional module as an example, (16,5,2) indicates that the number of channels is 16, the kernel size is 5, and the stride is 2. According to Equation ( 13), the receptive field of the CNN in the PE-CNN-GRU model is 261, which can effectively capture the information of the input PPG data with a length of 256.
Feature maps extracted by a CNN are then inputted into the GRU module for further processing.As one of the variants of recurrent neural networks, a GRU can exploit the long-term dependence and sequence information of PPG data with the help of the reset gate and the update gate inside its units.The GRU module contains two GRU hidden layers, each containing 64 units.The GRU module outputs tensors to a fully connected (FC) layer containing two neurons.The fully connected layer finally outputs bi , which contains the predicted SBP and DBP values.
In summary, the proposed model uses position encoding to fully capture the sequence information in PPG, which can avoid the neglect of sequential information in a CNN.The proposed model further combines the ability of a CNN in feature extraction and the advantages of a GRU in processing time series, making it more efficient and more accurate for BP measurement through better utilization of features in PPG.

Dataset Description
After the pre-processing of the PPG data, a total of 18,023 pieces of data were formed for training and testing, and the distribution of BP is shown in Figure 8.The red line is the mean value, and the blue line is mean±std.The mean value of SBP is 140.52 mmHg, with a standard deviation of 23.92 mmHg.The mean value of DBP is 58.83 mmHg, with a standard deviation of 9.39 mmHg.The data used in this paper were collected from patients in an Intensive Care Unit (ICU), and BP values were measure via arterial puncture method in the brachial/radial artery.We can find that the average SBP is higher than SBP measured in daily life.Generally, SBP is higher in the brachial/radial artery than in the aorta, while DBP remains unchanged [44].However, most of the DBP values are concentrated around 50 mmHg, with a low proportion exceeding 90 mmHg, which is caused by unstable vital signs of patients in ICU.We control the number of windows for each piece of test data to avoid an uneven distribution of training data.In detail, for each piece of the measured data of each user, 80% of the windows are used for training, and the remaining 20% are used for testing.This results in a total of 14,211 pieces of training data and 3812 pieces of test data.All pieces of PPG data are non-overlapping.

Performance of Data Augmentation Methods
In this paper, we use data augmentation methods to double the size of the training set, without augmentation of the test set.For augmented PPG windows, we label the same SBP and DBP values as the original PPG windows.As a result, the augmented training set contains 28,422 pieces of data, which does not change the BP distribution of the original data.
The goal of data augmentation is to improve the accuracy of a BP measurement.Therefore, we train the PE-CNN-GRU model using the augmented training set of each method and then test the model performance based on the same test set to evaluate whether the method achieves valid data augmentation.
We choose mean absolute error (MAE), root mean square error (RMSE), mean error (ME) and standard deviation of error (SD) to evaluate the BP measurement performance of the PE-CNN-GRU model.The impact of data augmentation methods on the performance of SBP measurement and the impact of data augmentation methods on the performance of DBP measurement are shown in Tables 3 and 4, respectively.In the clipping method, we cut each window with a certain ratio.Although the generated data maintain the same distribution as the original data, cropping still causes the loss of information in the PPG windows.As a result, cropping can reduce all the metrics of DBP measurement and reduce the MAE of the SBP measurement, but it increases the RMSE and SD of the SBP measurement.Therefore, using clipping cannot effectively improve the accuracy of a BP measurement.In the field of image data augmentation, flipping does not significantly change data distribution.However, PPG is a kind of one-dimensional temporal signal.Flipping the data left and right or up and down destroys the original data distribution; thus, using generated data for training has a negative impact on the PE-CNN-GRU model.As a result, the flipping method increases the measurement error of SBP and DBP.
The most significant advantage of DBA is using the warping operation to align PPG windows in morphology before calculating the weighted average series.As mentioned in Section 2, we use QIPW to ensure that periods have strong correlation and different PPG windows have similar distributions.As a result, DBA improves the performance of SBP and DBP measurements and achieves valid data augmentation due to generated synthetic data with a distribution similar to the original data.
In GAN, the generator takes random noise as an input to generate synthetic PPG data.In other words, the synthetic data comes from noise, which causes burs in the synthetic data.The discriminator is also unable to effectively distinguish synthetic data containing burrs from real data.The synthetic data are equivalent to data with poor quality, which affects the training of the PE-CNN-GRU model.Therefore, this method has deteriorated in all metrics of SBP and DBP measurements compared to the method without augmentation.
VAE represents the latent space in a probabilistic manner, which is very suitable for representing PPG signals with certain regularity.When training VAE, the latent space becomes similar to the prior distribution of input PPG.The decoder draws random samples from the prior distribution and then converts them to the input space; thus, the generated PPG data have the same distribution as the real data.Therefore, we can find that VAE significantly improves the measurement accuracy of the PE-CNN-GRU model for SBP and DBP.In terms of SBP, VAE causes the most significant decrease in MAE, reaching 18.80%.Considering that RMSE, which is more sensitive to outliers, also decreases significantly, VAE achieves an overall improvement in SBP measurement accuracy.In terms of DBP, VAE also performs best, which can reduce MAE and RMSE by 19.84% and 10.39%.Therefore, VAE is the most suitable data augmentation method in this paper.In addition, most researchers use signal denoising methods to improve accuracy of BP measurements from PPG. Kachuee et al. [34] used denoising methods such as an averaging filter to form the high-quality dataset used in this paper, but signal denoising cannot solve the problem of a limited amount of PPG data.The superior performance of data augmentation by VAE shows that data augmentation can further improve the accuracy of BP measurements based on signal denoising.

Performance Comparison of PE-CNN-GRU
We select some representative deep learning BP measurement methods based on PPG data directly as baselines.Due to the lack of open-source code or a detailed description of the model structure, we directly use the performance of ANN, CycleGAN and Transformer in their papers.We reproduce ResNet+LSTM, RDAE and PPG2ABP based on our data.In addition, we redesign the CNN-GRU model proposed in our previous work [16] as a baseline to verify the effectiveness of the PE.The performance comparison between PE-CNN-GRU after data augmentation with baselines is shown in Table 5. N/A means the metrics are not available in related references.When comparing with CNN-GRU, PE-CNN-GRU reduces the MAE of SBP and DBP measurements by 36.13% and 28.11%.The CNN neglects the sequence information of the original data when extracting features.PE effectively solves this problem by calculating the extra position encoding vector as part of the input.Therefore, the use of PE can significantly improve the accuracy of a BP measurement.
After the data augmentation of PPG, PE-CNN-GRU achieves the second best performance for DBP measurement, only worse than Transformer.For SBP measurement, only Transformer and Cycle GAN are more accurate than PE-CNN-GRU.Compared to PE-CNN-GRU, Transformer and Cycle GAN have more parameters and more complex structures.The performance of PE-CNN-GRU in RMSE and SD means there exit outliers, which is consistent with the Bland-Altman analysis in Figure 9. Compared with existing data-driven models, CNN has significant advantages for the feature extraction of PPG, which can help the model better fit the relationship between PPG and BP.The only problem of CNN is that extracting features can cause a loss of sequence information, but this can be compensated by PE.GRU further utilizes long-term dependence and sequence information through the reset gate and update gate with each GRU unit.The proposed PE-CNN-GRU can realize accurate BP measurement.However, we need to conduct further research on post-processing methods for measurement results to reduce the RMSE and SD of SBP measurements.

Conclusions
In this paper, we focus on using the data augmentation of PPG to improve the accuracy of BP measurements.After the detection and removal of artifacts in PPG, high-quality PPG windows are formed from the MIMIC II dataset.Considering that few-shot PPG data limit the performance of BP measurements, we use cropping, flipping, DBA, GAN and VAE for the data augmentation of PPG.Then, a PE-CNN-GRU model is proposed to achieve cuff-less BP measurements based on PPG directly.The generated data of VAE are closest to the original data in both morphology and trend thorough observation.Experiment results also show that the improvement in BP measurement accuracy by VAE is the most significant.In detail, VAE can reduce the MAE of SBP and DBP by 18.80% and 19.84%, respectively.In addition, we verify that PE can significantly improve the accuracy of BP measurements on the basis of CNN-GRU.After PPG data augmentation, the MAE of SBP and DBP measurements of PE-CNN-GRU achieve 3.80 mmHg and 2.02 mmHg; thus, PE-CNN-GRU performs well compared with existing cuff-less BP measurement works.

Figure 3 .
Figure 3. Illustration of PPG data augmentation pipeline.

Figure 4 .
Figure 4. Examples of synthetic PPG windows generated by different methods.

Table 1 .
Network architecture of GAN based on PPG data.

Table 2 .
Network architecture of VAE based on PPG data.

. Blood Pressure Measurement Based on PE-CNN-GRU Model In
The SBP and DBP values related to PPG window P i can be represented by this paper, each PPG window can be represented as P i = [pw 1 i , pw 2 i , . . ., pw 256 i ].

Table 3 .
Impact of data augmentation methods on the performance of SBP measurement.

Table 4 .
Impact of data augmentation methods on the performance of DBP measurement.

Table 5 .
Comparison between PE-CNN-GRU and existing works.