Next Article in Journal
Recreating the Trabecular Outflow Tissue on Implantable, Micropatterned, Ultrathin, Porous Polycaprolactone Scaffolds
Previous Article in Journal
Advances in Biomechanics-Based Motion Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning Techniques for the Performance Enhancement of Multiple Classifiers in the Detection of Cardiovascular Disease from PPG Signals

by
Sivamani Palanisamy
1,* and
Harikumar Rajaguru
2
1
Department of Electronics and Communication Engineering, Jansons Institute of Technology, Coimbatore 641659, India
2
Department of Electronics and Communication Engineering, Bannari Amman Institute of Technology, Sathyamangalam 638402, India
*
Author to whom correspondence should be addressed.
Bioengineering 2023, 10(6), 678; https://doi.org/10.3390/bioengineering10060678
Submission received: 28 March 2023 / Revised: 11 May 2023 / Accepted: 28 May 2023 / Published: 2 June 2023
(This article belongs to the Section Biosignal Processing)

Abstract

:
Photoplethysmography (PPG) signals are widely used in clinical practice as a diagnostic tool since PPG is noninvasive and inexpensive. In this article, machine learning techniques were used to improve the performance of classifiers for the detection of cardiovascular disease (CVD) from PPG signals. PPG signals occupy a large amount of memory and, hence, the signals were dimensionally reduced in the initial stage. A total of 41 subjects from the Capno database were analyzed in this study, including 20 CVD cases and 21 normal subjects. PPG signals are sampled at 200 samples per second. Therefore, 144,000 samples per patient are available. Now, a one-second-long PPG signal is considered a segment. There are 720 PPG segments per patient. For a total of 41 subjects, 29,520 segments of PPG signals are analyzed in this study. Five dimensionality reduction techniques, such as heuristic- (ABC-PSO, cuckoo clusters, and dragonfly clusters) and transformation-based techniques (Hilbert transform and nonlinear regression) were used in this research. Twelve different classifiers, such as PCA, EM, logistic regression, GMM, BLDC, firefly clusters, harmonic search, detrend fluctuation analysis, PAC Bayesian learning, KNN-PAC Bayesian, softmax discriminant classifier, and detrend with SDC were utilized to detect CVD from dimensionally reduced PPG signals. The performance of the classifiers was assessed based on their metrics, such as accuracy, performance index, error rate, and a good detection rate. The Hilbert transform techniques with the harmonic search classifier outperformed all other classifiers, with an accuracy of 98.31% and a good detection rate of 96.55%.

Graphical Abstract

1. Introduction

PPG has demonstrated that it is an effective device for the early diagnosis of cardiac disorders. Using PPG, blood volume fluctuations in tissues are continuously measured. PPG serves as a substantial promising technique for the thorough screening of cardiovascular diseases (CVDs). Blood circulation from the heart to the toes and fingertips is measured using PPG signals [1]. PPG sensors typically have an infrared-operating avalanche photo diode as a detector and a light-emitting diode (LED) as the source [2]. Both the European Committee for Standardization and the International Standards Organization (ISO) have recognized PPG as a standard noninvasive procedure for measuring and analyzing blood oxygen saturation levels. CVD leads to changes in the heart rate. PPG is skin-friendly, since it does not require direct skin-to-surface contact. Cardiac functions, such as blood flow, heart rate, and mean circulation time, are measured using PPG signals [3], since heart rate is associated with multiple physiological variables connected with hormonal and neuronal disturbances, pumping mechanisms, and a mean blood circulation time. A variation in the heart rate is viewed as a frequency shift over time and is specified as quasi-periodic in nature. A spectral analysis of PPG is used to determine the Eigen structure approaches as established by Al-Fahoum et al. [4]. The methods described in [4] are more appropriate to localize the spectral content of the PPG signal, which will identify the variations in the heart rate. Sukor et al. [5] discussed the presence and removal of signal artifacts in PPG signals using the waveform morphology method and achieved an accuracy of 83 ± 11% in the detection of CVD. CVDs are becoming one of the top causes of death worldwide. Approximately 16.7 million deaths worldwide, or 29.2% of all deaths, have occurred as a result of the different types of CVDs over the past ten years. All things considered, doctors and wellbeing experts advocate for the early discovery and treatment of potential side effects to guarantee the shirking of an out-and-out heart-related disease [6]. The main disadvantage of PPG signals is that they are highly corrupted by noise components, such as skin artifacts and motion artifacts. Hence, the preprocessing of PPG signals to attain cleaning signals is, itself, a broad area of research. Tun [7] designed an FIR filter for a heart rate detection system, which helps to minimize motion artifacts in a light-based measuring system. The reduction in artifacts in PPG signals using an AS-LMS adaptive filter was discussed by Ram et al. [8], and Luke et al. [9] presented a methodology using efficient signal processing algorithms for artifact removal in PPG signals and attained an SNR value of 41.52 db. From the analysis of PPG signals, the common parameters extracted are the pulse transit time, heart rate, stiffness index, and pulse wave velocity. Large artery stiffness and pulse width, pulse interval, systolic amplitude, augmentation index, and peak-to-peak interval are certain characteristics extracted from PPG signals [10]. Shintomi et al. [11] used mobile and wearable sensor equipment to measure the effectiveness of the heartbeat error and compensation strategies on heart rate variability (HRV). PPG is frequently used instead of an electrocardiogram (ECG) to assess heart rate in wearable devices. However, there are inherent differences between PPG and ECG due to the fact that PPG is affected by body motions, vascular stiffness, and blood pressure variations. Using ECG and PPG readings obtained from 28 people, the methods described in [11] determined how these errors affected the analysis of HRV. The assessment’s findings demonstrate that the error compensation method enhances the precision of HRV analysis in both the time and frequency domains as well as in nonlinear analysis. When compared to recent ECG observing systems, PPG signal measurements are more accessible and require less hardware for signal acquisition [12]. PPG does not require a reference signal, making it possible to integrate PPG sensors with wristbands. As a result, it can be employed in a variety of studies for the investigation and diagnosis of CVDs [13]. Moshawrab et al. [14] discussed how CVDs can be detected and predicted using smart wearable devices. In addition, the authors indicated that a review of the evaluation of the development and usage of smart wearable devices for the management of CVD demonstrates the high efficacy of such smart wearable devices.
The main objective of this research was to enhance the performance of multiple classifiers in the detection of cardiovascular disease (CVD) from PPG signals. The PPG signals obtained from the Capno database consisted of various features and occupied a large amount of memory, so it was necessary to extract the useful features by dimensionality reduction techniques. The main contributions of this work are that, by utilizing heuristic- and transformation-based dimensionality reduction techniques, the dimensions of the PPG data were reduced. The dimensionality-reduced PPG data were processed with twelve different classifiers, such as PCA, EM, logistic regression, GMM, Bayesian LDC, firefly, harmonic search, DFA, PAC Bayesian learning, KNN-PAC Bayesian, SDC, and detrend SDC. Machine learning techniques were used to improve the classification performance of the classifiers in the detection of CVDs from PPG signal.
The performance of the classifiers in terms of the classification accuracy, GDR, and computational complexity were examined and analyzed as the outcomes of this research. This article is organized as follows: The introduction is presented in Section 1, followed by the methodology in Section 2. Section 3 discusses the five techniques for reducing dimensionality. The twelve classifiers used to classify the normal and CVD segments of the PPG signals are discussed in Section 4. In Section 5, the obtained results are discussed. In Section 6, the article is concluded.

Review of Previous Work

Certain works associated with PPG signal exploration and their use in various biomedical areas are provided below. For approximately three decades, research on PPG signals with respect to CVDs has been very interesting, and many research outcomes have been reported. Allen [15] examined PPG and how it could be used in clinical physiological measurements through the clinical monitoring of the human body, autonomic function, and vascular assessment. Sunil Kumar and Harikumar [16] used parameters, such as independent component analysis (ICA), principal component analysis (PCA), entropy, and mutual information (MI) to analyze the PPG signals. PPG has proven to be the most capable method for the early screening of heart-related diseases. Almarshad et al. [17] disclosed the diagnostic properties of PPG signals in addition to their prospective clinical applications in healthcare and assessed the possible effects of PPG signals on the screening, monitoring, diagnosis, and fitness of inpatients and outpatients. Yousefi et al. [18] developed a technique to automatically detect premature ventricular contraction (PVC) from the extracted features of PPG signals using higher-order statistics (HOS) and the chaotic method for the KNN classifier, with a classification accuracy of 95.5%. Cardiac health supervision based on PPG through smart phones or wearable sensors has the potential to attain high accuracy with fewer false alarms, as mentioned by Ukil et al. [19]. With the help of the ensemble method, Almanifi et al. [20] used a computer-vision approach to find human activities using PPG signals. The accuracy of the PPG was 88.91%, which shows that wrist PPG data could be used with the ensemble method in human activity recognition (HAR) to make accurate detections. The time domain analysis of a PPG signal and its second derivative was performed by Paradkar et al. [21] to extract the features from the PPG signal, and these extracted features were classified by the SVM classifier to detect coronary artery disease (CAD) with a sensitivity of 85% and a specificity of 78%. Neha et al. [22] investigated the detection of arrhythmias using PPG signals. In this analysis, a low-pass Butterworth filtering method was used to remove artifacts, and the extracted features were applied to various machine learning algorithms to classify normal and abnormal pulses. The results show that the SVM classifier had a high accuracy of 97.674% in identifying arrhythmia pulses. Prabhakar et al. [23] analyzed metaheuristic-based optimization algorithms as dimensionality reduction techniques and then applied various classifiers to these dimensionally reduced values for the classification of CVD. The results in [23] show that, for chi-squared probability density function (PDF) optimized values, the artificial neural network (ANN) classifier attained a maximum classification accuracy of 99.48% for normal subjects, and the logistic regression classifier produced a maximum classification accuracy of 99.48% for CVD cases. In order to analyze CVD patients utilizing PPG signals, Sadad et al. [24] applied various machine learning techniques and a deep learning model to create a system that helps doctors with continuous monitoring, and they obtained an accuracy of 99.5%.
In this article, a PPG signal was dimensionally reduced using five dimensionality reduction techniques, namely the Hilbert transform (HT), nonlinear regression (NLR), ABC PSO, cuckoo search, and dragonfly methods. These dimensionally reduced PPG signals were further classified as CVD or normal using twelve classifiers. Principle component analysis (PCA), expectation maximization (EM), logistic regression (LR), Gaussian mixture model (GMM), Bayesian linear discriminant analysis classifier (BLDC), firefly clusters, harmonic search, detrend fluctuation analysis (DFA), PAC Bayesian learning, KNN-PAC Bayesian, softmax discriminant classifier (SDC), and detrend with SDC were the classifiers used for classification purposes.

2. Methodology

The PPG data recordings with various waveform morphologies utilized in this work were obtained from the IEEE TMBE pulse oximeter standard CapnoBase database dataset, which is available online [25]. The dataset comprises raw PPG signal recordings with a duration of 8 min. This database consists of annotated respiratory signals, such as pressure, respiratory flow, and inhaled and exhaled carbon dioxide (capnogram). All 41 records were considered in this investigation. Twenty-one of the forty-one cases are normal, while the other twenty have cardiovascular disease. The dataset for the PPG signals was sampled at a rate of 300 Hz, and 144,000 samples of data were extracted from the PPG, with 720 segments overall. Each of these segments had 200 samples at equal intervals.
Independent component analysis (ICA) was used to remove the noise components in the PPG signals. The classification of a PPG signal involved two steps: First, dimensionality reduction (DR) was achieved with the help of heuristic- and transformation-based techniques. Second, these dimensionally reduced values were classified using various classifiers to detect whether the corresponding PPG signal was associated with a person with CVD or a normal subject. After the implementation of the dimensionality reduction techniques, the original PPG samples of a patient (200 × 720) were reduced to 100 × 720. These dimensionally reduced samples (100 × 720) were input into the classifiers for further classification. The organization of the CVD detection from the PPG signals is depicted in Figure 1.

3. Dimensionality Reduction Methods

Dimensionality reduction (DR) is a preprocessing technique used to eliminate irrelevant data and redundant features in order to reduce the training time of PPG signals [26]. All machine learning techniques and models become increasingly challenging as the dimensions of the input dataset increase. Dimensions are a problem in PPG signals with large amounts of data. As the number of features increases, the number of samples also increases proportionately, and the chances of overflow also increase. When a machine learning model is trained on large datasets, it becomes superabundant and results in a mediocre performance. As a result, it is necessary to decrease the number of features, which can be accomplished by dimensionality reduction. In this way, the DR technique is used to prevent data overfitting and select the most informative characteristics for classification purposes [27]. In this research, five dimensionality reduction techniques, divided into two categories, were utilized. First, transformation-based optimization techniques include Hilbert transformation and NLR optimization. Second, heuristic-based optimization includes ABC PSO, cuckoo search, and dragonfly optimization. These methods are discussed in the following.

3.1. Hilbert Transform

The Hilbert transform (HT) is a mathematical operation that is used to obtain the analytical representation of a real-valued signal.
The Hilbert transform [28] of signal y(t) is given by
y ^ t = H y t = 1 π y τ t τ d τ  
It can be observed from the above equation that this transformation has no effect on the independent variable; therefore, the output y ^ t is also a function that changes over time. Furthermore, the output y ^ t is a linear function of input y t . It is produced by applying convolution to y t with π t 1 , as shown in the equation below:
y ^ t = 1 π t y t
By applying Fourier transform (FT) to the above equation, we obtain
F y ^ t = 1 π F 1 t F y t
A phase shift of −90 degrees is produced on all positive frequency components of the real-valued signal y(t) and +90 degrees on all negative frequency components when a Hilbert transform is applied to y(t). The domain of the signal is not changed by HT. When a Hilbert transform is applied to two different signals with the same amplitude but different phases, the magnitude spectrum is the same because the transform does not change the magnitude spectrum but changes the phase spectrum. This small phase response is obtained from spectral analysis using the Hilbert transform. In signal processing, the Hilbert transform is frequently used to generate the analytical representation of the real-valued signal y(t). All Fourier transformable signals are Hilbert transformable [29].

3.2. Nonlinear Regression

Nonlinear regression (NLR) is a statistical technique that involves developing a regression model to represent a nonlinear relationship between independent and dependent variables. The fundamental concept of linear regression and nonlinear regression is the same, that is, to connect a response,   R , to a set of predictors, Z = z 1 , z 2 . z n T . The prediction equation for nonlinear regression varies nonlinearly on one or more unidentified parameters. Typically, nonlinear regression occurs when there is a specific functional shape in the relationship between the predictors and the response. The main goal of this model is to provide low sums of squares. The sum of squares parameter is connected to the number of observations that deviate from the dataset’s mean value. The variance between the mean and the individual points in the dataset is used to estimate the mean or average parameter. The collected variants are squared before being added together in the final step. The objective best matches the dataset points if the sum of the squared variations is found to be low. The least squares method, the equations of which comprise nonlinear elements, is used to compute the parameters in a nonlinear model. The steepest descent, Taylor series, and Levenberg–Marquardt methods can be used to solve these kinds of equations. The least squares in a nonlinear model is calculated by applying the Levenberg–Marquardt algorithm, which is its most extensive use. The advantages of this strategy include the optimum feature selection and deserving model convergence through iterations.
The structure of a nonlinear regression model as shown in [30] is
R i = f z i , φ + e i   for   i = 1 , 2 , .. n
where R i are responses;   f is a known function of the covariate vector Z i = z i 1 , z i 2 . z i n T and the vector parameter φ = φ 1 , φ 2 . φ n T ; and e i represents the random error values. Typically, it is assumed that random errors have an uncorrelated mean of zero and constant variance.
The formula for calculating the residual sum of squares is expressed as
S φ = i = 1 n R i f z , φ 2

3.3. ABC-PSO

ABC-PSO stands for artificial bee colony–particle swarm optimization. It is a hybrid optimization algorithm that combines the artificial bee colony (ABC) algorithm and the particle swarm optimization (PSO) algorithm to enhance the search capability and convergence speed of the optimization process. The ABC method has the advantages of simplicity, flexibility, and fast achievement of good results for multidimensional datasets. It is obvious that, in order to locate the locations of new food sources, the ABC algorithm does not necessarily need to apply the population’s best global solution. In the meantime, it may be concluded that PSO particles may not be able to escape from the local minima by a performing random search as scout bees in ABC. In addition, the update equation in ABC updates a single variable instead of all variables as in PSO. In order to achieve the best results, the ABC algorithm has been combined with the PSO search algorithm. In ABC-PSO, three ABC phases are used, and for the employed bee phase, it uses velocity and the PSO’s method of locating new food sources. The best location currently visited by the person is updated after the position of a new food supply is updated. The food source’s trial counter is reset if the current best position is altered; otherwise, its value is raised by 1. Onlooker bees memorize their positions throughout the employed bee phase and search for new food sources based on their knowledge of the best food source locations that the employed bees have visited. The new candidate solution “z” is generated by utilizing the ABC update equation, as follows:
z m n = p b e s t m n   if   n k p b e s t m k + m k . p b e s t m k p b e s t l m ,   if   n = k
where z m n is the m th dimension of the n th employed bee selected; k is the random index; l is the index of a randomly selected individual; and m k is a random number between −1 and 1. The food source trial counter will reset if the new food source location has a better value; otherwise, it is raised by 1. The scout bee phase follows the ABC algorithm [31].
ABC-PSO Algorithm:
1.
Initialization of the swarm.
2.
Velocity and position of the particle are updated by performing the employed bee phase.
3.
The local best position of a particle is updated by finding its new position by performing the onlooker bee phase.
4.
If the highest value of the trial counter for any food source is higher than the limit, a scout bee will search for a new food source site.
5.
At this point, instead of using scout bees, the PSO algorithm is used to look for new food sources.
6.
Particles with random placements are used to initialize the population of new food sources.
7.
The fitness value is determined for all particles for the specific objective function.
8.
The fitness function is used to select the optimal set of features. The expression for the fitness function is as follows:
F = a φ k + b T K T
where φ k is the classifier performance in subset K ; b is the feature subset length; T is the total number of features; and a is the classification quality.
9.
The number of particles that are currently present is set as p b e s t .
10.
A new set of particles is created by adding velocity to the initial particle, and a fitness value is calculated for the same.
11.
A new p b e s t is discovered between the two particle sets by comparing the fitness values of each particle to one another.
12.
The least fitness value is determined by comparing the two sets of particles, and the corresponding particle is then referred to as the G b e s t .
13.
Simultaneously, in the next iteration, the update in the velocity ( v q + 1 ) and position x q + 1   is conducted as follows:
v q + 1 = v q + a p b e s t x q + b G b e s t x q
x q + 1 = x q + v q + 1
The maximum step size that a particle can take in each iteration is influenced by the acceleration coefficients a and b .
14.
The PSO iterations are continued until convergence is reached.
15.
The finest food source is identified and remembered.
16.
The process can be performed as many times as necessary to fully satisfy the stopping criteria.
The main goal of this combined hybridization method is to combine the elements of ABC and PSO so that separate problems may be readily addressed by ABC and PSO simultaneously addressing rotationally invariant behavior.

3.4. Cuckoo Search

The cuckoo search (CS) algorithm is a nature-inspired metaheuristic optimization algorithm, which is based on the brood parasitism of some cuckoo species, along with Levy flight random walks established by Xin-She Yang and Suash Deb [32]. This optimization method depends on the incubate parasitism activities of certain cuckoo types along with the behavior of the Levy flights of certain fruit flies and birds. The three common rules of the CS algorithm can be given as follows:
  • At a certain time, every cuckoo bird lays one egg and dumps it in an arbitrary selected host nest.
  • The subsequent generation will carry the top-quality eggs, which are present in the best host nests.
  • There are only fixed quantities of host nests available, and a host bird can realize a cuckoo’s egg with a probability of P a 0   1 . In this instance, a cuckoo’s egg in the host nest may be thrown away by the host bird or it abandons the nest and creates an entirely new nest in a different location.
The fundamental steps of the cuckoo search algorithm can be summed up as follows using the above three rules.
  • Create a population of N host nests at the beginning.
  • Randomly select the host nest X.
  • Lay the egg in the selected host nest X.
  • Compare the fitness of the cuckoo’s egg with the host egg’s fitness.
  • If the fitness of the cuckoo’s egg is better than the host egg’s fitness, replace the egg in nest X with the cuckoo’s egg.
  • Abandon the nest if the host bird notices and build a new one.
  • Repeat steps 2–6 until the termination criteria are met.
Levy flight is the term used to describe the random flight patterns that birds use to find their next position, z i t + 1 , based on the present position, z i t . Levy flights are used to build new hosts in new places. Consider cuckoo “i”; a Levy flight is performed to produce the new results z t + 1 :
z i t + 1 = z i t + β levy Υ
where ⊕ denotes the entry-wise multiplication, and β > 0 is a scaling factor that denotes the step size. Here, the β value is considered the one to optimize. A random walk is provided by the Levy flight, and random step sizes are calculated from the Levy distributions as follows:
Levy ~ v = t Υ ( 1 < Υ 3 )
This has an infinite variance and mean. The symbol “ ~ ” indicates that the numbers being generated are pseudorandom and are drawn from a probability distribution. The consecutive jumps of a CS are basically a random walk procedure that follows a step-length distribution of a power-law through a heavy tail. Utilizing the Levy walk approach, several new solutions around the current best solution should be produced by wide-area randomization. The locations of the new solutions should be far from the current best solution. This will prevent the system from becoming stuck at a local optimum.

3.5. Dragonfly

The dragonfly algorithm (DA) is a swarm intelligence algorithm, and it is inspired by the dynamic and static swarming behaviors of dragonflies. The dragonfly algorithm is a modern heuristic optimization technique created by Mirjalili in 2016 [33]. The static and dynamic swarming behaviors of dragonflies in nature serve as the primary source of inspiration for the dragonfly algorithm. In the dynamic or exploitation phase, a huge number of dragonflies form swarms and travel over long distances in one particular direction to distract their enemies. In the static or exploration phase, dragonflies form groups and move frontward and backward in a small zone for hunting and attract their prey. The five fundamental principles of DA are separation, alignment, cohesiveness, attraction, and diversion. In the following equations, Q and Q j denote the current and j th   positions of the individual dragon flies, respectively, and the total number of neighboring flies is denoted by K .
1.
Separation: This indicates the static avoidance of flies colliding with other flies in the area. It is calculated as
S i = j = 1 K Q Q j
where S i denotes the separation motion of the i th individual.
2.
Alignment: This signifies the velocity matching among individual flies within the same group. It is denoted as
A i = j = 1 K V j K
where V j denotes the velocity of the j th individual.
3.
Cohesiveness: This denotes the tendency of individual flies to move to the center of swarms. The estimation of this is given by
C i = j = 1 k Q j K Q
4.
Attraction towards the nourishment source is estimated as
F i = Q + Q
where F i denotes the nourishment source of the i th individual and Q + is the position of the nourishment source.
5.
Diversion: This represents the distance outwards to the enemy. It is calculated as
E i = Q + Q
where E i denotes the i th individual enemy’s position and Q denotes the enemy’s position.
Within the search space, the locations of artificial dragonflies are updated by the step vector, Δ Q , and the current position vector, Q . The direction of the dragonfly’s movement is indicated by the step vector, Δ Q , and it is evaluated as
Δ Q i t + 1 = s S i + a A i + c C i + f F i + e E i + ω Δ Q i t
where s , a , c , f , and e are the separation weight, alignment weight, cohesion weight, attraction weight, and enemy weight, respectively. The inertia weight is denoted by ω and “t” denotes the iteration number. The exploration and exploitation phases can be obtained by changing the weights.
At t + 1 iterations, the position of the i th dragonfly is calculated as follows:
Q i t + 1 = Q i t + Δ Q i t + 1

3.6. Statistical Analysis of Dimensionally Reduced PPG Signals

The dimensionally reduced PPG signals were analyzed through the extraction of statistical parameters and sample entropy for ascertaining the nonvariation in PPG signal characteristics. The statistical features [34], such as the mean, variance, skewness, kurtosis, Pearson correlation coefficient (PCC), and sample entropy [35], were extracted from the dimensionally reduced PPG samples among the CVD and normal classes. This reduced dataset provides the appropriate information from the above features.
Table 1 shows that the statistical analysis of the parameters was carried out with DR techniques for the PPG signals. It is observed from Table 1 that, for normal cases, lower mean values were obtained across the various optimization techniques. For cases of CVD, higher mean values, except for ABC-PSO and the negative mean, were obtained under dragonfly optimization. The skewness and kurtosis were highly skewed values for normal as well as CVD cases. It is inferred from Table 1 that the sample entropy values were the same across all classes, except for the NLR DR method in normal cases and the cuckoo search for the CVD cases. In addition, from Table 1, it is shown that the PCC values were low, which indicates that the optimized features were nonlinear and uncorrelated within the classes. Therefore, it is better to apply nonlinear classifiers to detect the CVD and normal segments of the PPG signals. If the CCA values are greater than 0.5, then there will be high correlation across the classes. From Table 1, it can be noticed that the Hilbert transform optimization was highly correlated across the classes. It also exhibits that the ABC-PSO optimization method was the least correlated among the classes.
Therefore, the above analyses of the dimensionally reduced PPG signals make a strong case for the usage of better classifiers. In order to identify the presence of nonlinearity in the dimensionally reduced signals, a normal probability plot for Hilbert transform-based dimensionally reduced values of the PPG signals in cases of CVD for Patient 2 is shown in Figure 2. From Figure 2, it can be observed that the normal plots exhibit the presence of nonlinearity and the overlapping of the Hilbert transformed values of the PPG signals.
Figure 3 shows a normal probability plot for the Hilbert transform-based dimensionally reduced values of the PPG signals in normal cases for Patient 30. The presence of outliers and nonlinear features of the Hilbert transform values can be observed. As a result, nonlinearity and outliers are preserved in the PPG signals of both classes after the DR methods.

4. Classifiers for Detection of CVD

Classifiers play a vital role in classifying data. An ideal classifier is one that provides high accuracy with a low error rate for a given computational complexity. The following sections of this paper discuss the classifiers that were used for this purpose.

4.1. PCA as a Classifier

Principal component analysis (PCA) is a technique that can be utilized for both data compression and classification purposes. The original m predictor variables are reduced to a smaller number of derived variables, n , by PCA. The derived variables are obtained by transforming the m original predictor variables, Y = y 1 , y 2 . y m , into a new variable set (the principal components) n , Z = z 1 , z 2 . z n . The new variables are linear combinations of the original variables. Mathematically, the eigenvalues and eigenvectors of the data covariance matrix are calculated to obtain principal components. The direction of the largest variation is identified from the eigenvector that has the largest eigenvalue [36].

4.2. Expectation Maximization as a Classifier

The expectation maximization (EM) algorithm is a method used to compute the maximum likelihood estimation in the presence of latent variables. Consider Z as the observed data, statistical parameters as φ , and the missing data as δ . The aim is to maximize the function.
p Z | φ = p Z , δ | φ d δ
This equation cannot be systematically solved. It is assumed that the whole likelihood parameter or the posterior distribution p Z , δ | φ can be dealt easily with by applying the EM algorithm [37].
To reach convergence, this algorithm iterates between the E and M phases, as follows:
  • E step (expectation): calculate the Q function:
Q   φ | φ t = E δ | φ t , Z [ log p Z , δ | φ ]
  • M step (maximization): compute the maximum:
    φ t + 1 = arg max φ Q φ | φ t  
    where “ t “ denotes the iteration number. In the E-step, for each test point, the likelihood is computed from the individual cluster, and the calculation of the respective probability is carried out by assigning the test point to the corresponding cluster based on the maximum probability. All parameters are updated in the M step. This algorithm is repeated until it reaches convergence.

4.3. Logistic Regression as a Classifier

Logistic regression (LR) is a type of supervised machine learning algorithm that is utilized for predicting the probability of a target variable. The inputs are applied through a prediction function, which yields a probability value between 0 and 1, where 1 indicates CVD and 0 indicates normal [38]. To classify the positive and negative classes, a hypothesis, H θ x = θ P X , is designed. The threshold value H θ x for the classifier is 0.5. If H θ x 0.5 , then classify the values into the CVD class, and if   H θ x < 0.5 , then classify the values into the normal class. A classifier with at least a 50% predicted chance of classifying a CVD will be classified as class 1. Therefore, the threshold value H θ x ) for the logistic regression classifier was set to 0.5.
The LR function is given below:
H θ x = g θ P X

4.4. Gaussian Mixture Model (GMM) as a Classifier

The Gaussian mixture model (GMM) is a machine learning technique utilized to classify data into different groups according to the probability distribution. Combinations of the number of Gaussian distributions are referred to as GMM. Given the data vector y , the GMM is defined as [39]
p y | θ = q = 1 Z π q M y | μ q , q
where   q ,   μ q , π q are the covariance, mean, and mixture components of the GMM, respectively. In addition,
q Z π q = 1   , π q 0   Additionally ,   θ = μ 1 , 1 , π 1 μ q , q , π q
The R -dimensional Gaussian distribution is represented by M :
M y | μ , = 1 2 π R 2 1 2 exp 1 2 y μ T 1 y μ
The EM algorithm is used to calculate the GMM’s parameters by applying the E and M steps.
  • E step: the posterior probability, p i q t , is evaluated at t iterations as
    p i q t = π q t p y i | μ q t , q t q = 1 z π q t p y i | μ q t , q t
  • M step: utilizing the probabilities evaluated from the E step, the parameters q ,   μ q , π q are updated at t + 1 iterations:
    π q t + 1 = 1 N i = 1 N p i q t
    μ q t + 1 = i = 1 N p i q t y i i = 1 N p i q t
    q t + 1 = q = 1 z p i q t y i μ q t y i μ q t T i = 1 N p i q t
To stabilize the parameters at a specific value, these two steps are repeated.

4.5. Bayesian Linear Discriminant Analysis as a Classifier

The Bayesian linear discriminant analysis classifier (BLDC) is a generative model that estimates the probability distribution of data for each class and uses the Bayes’ theorem to predict the class of new data. In order to maximize the class posterior probability, it is chosen from observation “ k ”, or in the case of two classes, x and   y . Choose class   x ,   if   q x k q y k D , where D denotes the decision threshold, and q x k is the discriminant function ,   q x k = ln P x | k   [40].
Assume that every class observation is taken from the multivariate normal distribution and that the covariance matrix is identical for all classes. Now, by applying the Bayes rule, the discriminant function is given as follows:
q x k = 1 2 k μ x T 1 k μ x + ln P x
where μ x is the mean feature vector for class x , Ʃ denotes the pooled covariance matrix for all classes, and P x denotes class x ’s prior probability. The prior probability of all classes is considered constant; then, the decision boundary, D , is given as follows:
k μ y T 1 k μ y k μ x T 1 k μ x  
It is very clear that the mean vectors μ x and μ y are further away from each other in the feature space. If the term 1 μ x μ y is larger, then the classes are more separable.

4.6. Firefly Algorithm as a Classifier

The firefly algorithm is a metaheuristic approach used for solving optimization problems, and it is inspired by the flashing patterns exhibited by fireflies, which was first developed by Yang [41]. The firefly algorithm employs three idealized rules:
  • All fireflies considered here are unisex in nature, and along these lines, one firefly will be attractive to other fireflies irrespective of sex.
  • The attractive feature of a particular firefly varies with respect to its intelligence. Thus, for any two fireflies, the brighter firefly effectively pulls in the darker firefly. Assuming there are no fireflies brighter than a particular firefly, at that point, that specific firefly will move arbitrarily.
  • When the distance from the firefly increases, the brightness or light intensity of a firefly will decrease because the light is captured as it passes through the air. Subsequently, the engaging quality or intelligence of a particular firefly, s , seen by firefly “ t ” is characterized as
    α s r = α s 0 e β r 2
    where β is the light ingestion coefficient of a particular medium, α s 0 signifies the brightness of firefly s ” at r = 0 , and r indicates the Euclidean distance between firefly “ t   and firefly s :
    r = | | z t z s | | = i = 1 d z t i z s i 2
    where z t   and z s   are the individual areas of fireflies “ t   and s , respectively. If firefly s is the brighter one, then its amount of attractiveness directs the movement of the specific firefly “ t ” as per the accompanying condition:
    z t = z t + α s r z s z t + γ rand
    where γ is the randomization parameter, and rand denotes a random number taken from a uniform distribution that lies in the range between −1 and +1, inclusively. Firefly t   can effectively move towards firefly s based on the second term in the above equation.

4.7. Harmonic Search as a Classifier

Harmony search (HS) is a music-based metaheuristic algorithm inspired by the evolution of music and the pursuit of the ideal harmony. Geem et al. [42] proposed that HS imitates the improvisational method used by musicians. The steps to be followed in the HS algorithm are:
1.
Problem Definition and HS Parameter Initialization
An unconstrained optimization problem is described as the minimization or maximization of the objective function, f Y , given as follows:
U y i y i   L y i
where Y denotes the decision variable set; y i is the set of all possible values of every decision variable; and U y i and L y i represent the upper and lower bounds of the i th decision variable.
2.
Initialization of the Harmony Memory
In this stage, the harmony memory (HM) is initialized. All decision variables in the HM are kept as matrices. The opening harmony memory is created from a uniform random distribution of values that are constrained by the parameters U y i and L y i .
3.
Improvisation of a New Harmony
The HM is utilized in this process to create a new harmony.
4.
Updating the Harmony Memory
The HM is updated with the new harmony vector and the minimal harmony vector is deleted from the HM if the new improvised harmony vector is superior to the minimum harmony vector in the HM.
5.
Verification of the Terminating Criterion
When the termination criterion is satisfied, the iterations are terminated. If not, steps 3 and 4 are repeated until the allotted number of iterations has been reached.

4.8. Detrend Fluctuation Analysis as a Classifier

Detrended fluctuation analysis (DFA) is a mathematical method used to analyze the presence of long-term correlations, or persistence, in a time series dataset. The main purpose of DFA is to scale long-range correlations in a time series. DFA is very similar to the Hurst exponent analysis, which is an enhancement of the standard fluctuation analysis [43]. DFA heavily relies on the random walk theory.
The cumulative profile, M t , is obtained from the bounded time series, m q ,   of length K , which is given by
M t = q = 1 t m q m
A local least squares straight-line fit is calculated by minimizing the squared errors inside each time window after dividing K into time windows, each with the length of n samples. The average fluctuation function is given by
F n = 1 K q = 1 K M t N t 2
where N t is the piecewise sequence of the straight-line fits.

4.9. Probably Approximately Correct (PAC) Bayesian Learning Method as a Classifier

PAC Bayes is a generic framework to efficiently rethink generalization for numerous machine learning algorithms. It influences the flexibility of Bayesian learning and allows deriving new learning algorithms [44]. In PAC Bayesian learning theory, the hypothesis space is denoted as “S”, which is the variation between the empirical error and expected error of a hypothesis. The high-probability bounds of the deviation of the weighted averages of independent random variables are provided by the PAC Bayesian analysis of s , which is derived by the function θ s .   Prior distribution over the hypothesis space is denoted as Π , and the randomized classifier is defined as γ. Each time the game is played, the randomized classifier selects a hypothesis s from “S” in accordance with γ and uses it to predict the outcome of the subsequent sample.

4.10. KNN-PAC Bayesian Learning Method as a Classifier

KNN (K-nearest neighbors)-PAC (probably approximately correct)-Bayesian learning method is a machine learning algorithm used to make more accurate predictions by finding the nearest neighbors and updating the probability distribution of the data. The PAC Bayesian classifier is used to evaluate the divergence and training error of the finite data sample. For greater divergence, the risk factor will be high. The PAC Bayesian classifier output is fed to the KNN classifier to improve the accuracy of the classification. The KNN algorithm simply stores the dataset during the training phase and then classifies incoming data into a category that is very close to the previously stored dataset. A nonspecific sample can be classified considering the training data and samples. The selection of the K value is a critical stage in the KNN algorithm [45].

4.11. Softmax Discriminant Classifier (SDC) as a Classifier

The softmax discriminant classifier (SDC) is a supervised machine learning algorithm used for multiclass classification problems. It is based on the concept of the discriminant function, which maps input variables to a class label. SDC properly identifies the testing sample to which a particular class belongs by weighing the distance between the training and testing samples from that particular class [46]. Considering the training set   X = X 1 , X 2 , . , X k R m × n , chosen from “k” different classes, and X k = X 1 k , X 2 k , . , X n k k R m × n k represents n k samples from the k t h class, where j = 1 k n j = n , and the testing samples are assumed to be X R m × 1 .
Here, the SDC is defined as
f x = arg   max j q x j
f x = arg max j log i = 1 n j exp λ | | x x i j | | 2
where q x j   , f x   denotes the separation between the testing and j th class sample. A relative penalty cost is given when λ > 0 . Where x and x i have similar characteristics, if x belongs to the j th class, | | x x i j | | 2 is almost taken to zero and q x j can asymptotically reach the maximum value.

4.12. Detrend with SDC as a Classifier

Detrend fluctuation analysis (DFA) with the softmax discriminant classifier (SDC) is a machine learning algorithm that combines the DFA and SDC techniques to classify time series data with long-range correlations by removing the trend and identifying the long-term correlations before classification. The correlation qualities of PPG signals can be determined over a long duration in DFA. SDC is used to determine and identify the class to which a particular test sample belongs.

5. Results and Discussion

This section explores the performances of different classifiers based on their benchmark parameters. A better classification accuracy with a lower error rate leads to a good classifier performance. Therefore, the classifiers were trained and tested for the dimensionally reduced values in the CapnoBase PPG signal dataset.

5.1. Training and Testing of the Classifiers

The training and testing of classifiers are very important steps for all of the classification processes. The training allows a classifier to learn the patterns associated with the given DR data. In this study, we chose 90% of the data for training and 10% for testing. The mean square error (MSE) was maintained as the stopping criteria for the training and testing of the classifiers. The mathematical expression for MSE is given below:
MSE = 1 M i = 1 M O i T k 2
where O i is the observed value at a definite time; T k   indicates the model k ’s target value, where “ k ” varies from 1 to 15; and M is assumed as 1000 and denotes the number of observations per case.

5.2. Selection of the Optimal Parameters for the Classifiers

Consider that the PPG dataset had two classes, namely CVD and normal, when determining the target values, and the target T C V D was carefully selected with higher values in the range from 0 to 1. The condition used for selecting T C V D is as follows:
1 X i = 1 X μ i   T C V D
The features of the total ( X ) CVD PPG data were normalized, and their mean is signified by μ i , as mentioned in Equation (40), which can be applied for the classification.
For normal subjects, the target T N o r m a l with lower values between 0 and 1 was preferred when implementing the condition:
1 Y j = 1 Y μ j   T N o r m a l
The features of the total ( Y ) normal PPG data were normalized, and their mean is signified by μ j , as mentioned in Equation (41), which can be applied for the classification.
The   T C V D value should be greater than that estimated for μ i and μ j . It must be determined whether the difference between T C V D and T N o r m a l is zero or greater than 0.5.
T C V D T N o r m a l 0.5
Depending on the condition given in (42), the T C V D and T N o r m a l values were set as 0.85 and 0.1, respectively. The classifiers were trained with a 10-fold training and testing method, along with an MSE value of (10)−5 or a maximum operation of 1000, whichever was achieved earlier, as the stopping criterion. Table 2 demonstrates the selection of the optimal parameters for the classifiers.
Table 3 illustrates the analysis of the testing MSE values for the CVD and normal cases across the various classifiers with different DR techniques. It is perceived from Table 3 that, for the CVD cases, the ABC-PSO DR method with the DFA classifier resulted in the overall minimum MSE value of 4.00 × 10−8. The cuckoo search DR technique with the PCA classifier resulted in an overall maximum MSE of 6.60 × 10−4. Similarly, for normal classes, the overall minimum MSE of 9.00 × 10−8 was obtained when the Hilbert transform DR values were classified with the harmonic search classifier. The overall maximum MSE of 4.84 × 10−4 was obtained when the cuckoo search DR values were classified with the logistic regression classifier.

5.3. Performance Metrics of the Classifiers

In order to analyze the performance of the classifiers, the parameters, namely the performance index (PI), sensitivity, specificity, accuracy, good detection rate (GDR), and error rate, were calculated from the confusion matrix. Table 4 depicts the general confusion matrix for the detection of CVD.
True positive (TP): An output where the model accurately predicted the positive class, indicating that the person has cardiovascular disease.
True negative (TN): An output where the model accurately predicted the negative class, which shows that it is a healthy person.
False positive (FP): An output in which the positive class was incorrectly predicted by the model, which indicates that the healthy person is incorrectly classified as having CVD.
False negative (FN): An output in which the negative class was incorrectly predicted by the model, which indicates that a person with CVD is incorrectly classified as a healthy person.
PPG signals are sampled at 200 samples per second. Therefore, 144,000 samples per patient are available. There are 41 patients with 20 labeled has having CVD and 21 patients labeled as normal cases. Now, a one-second-long PPG signal is considered as a segment. Hence, there will be 720 such segments available per patient. The total number of segments for CVD cases is [20 × 720 = 14,400] and for normal cases is [21 × 720 = 15,120]. Therefore, the overall available segments for a total of 41 cases are 29,520 segments of one second duration only. The PPG signals are analyzed across the patients based on the signal segments. The beat to beat analysis is not included in this study. As a sample, the confusion matrix attained for the Hilbert transform DR method for different classifiers was undertaken, as shown in Table 5. As indicated in Table 5, the harmonic search classifier attained higher classification capability than the other eleven classifiers with low MSE values. At the same time, the firefly classifier well indented with more false negatives and less true positives categories.
These performance measures were calculated as follows:
The performance index (PI) is calculated as
PI = T P + T N F N F P T P + T N × 100
The sensitivity, specificity, accuracy, good detection rate (GDR), and error rate were calculated as shown below [19]:
Sensitivity   = T P T P + F N × 100
Specificity   = T N T N + F P × 100
Accuracy   = T P + T N T P + T N + F P + F N × 100
GDR: this represents the ability of a detector in fruitful detection, and it is given as
GDR = T P + T N F P T P + T N + F N × 100
Error   Rate   = F P + F N T P + T N + F P + F N × 100
As a sample for the classifier performance analysis, the cuckoo search DR method was undertaken, as shown in Table 6. As mentioned in Table 6, the harmonic search classifier performed better than all other classifiers in terms of the parametric values, such as an accuracy of 96.095%, performance index (PI) of 91.29%, sensitivity of 92.185%, specificity of 100%, highest GDR of 92.19%, and the lowest error rate of 7.81%. On the other hand, the firefly classifier showed the lowest performance with respect to all of the parametric values, such as an accuracy of 78.275%, a performance index (PI) of 20.76%, a sensitivity of 56.15%, a specificity of 100%, a GDR of 56.15%, and a high error rate of 43.855%. It is also observed from Table 6 that the GMM classifier reached a high sensitivity of 100% and ebbed at a low specificity of 84.38% due to the low number of true negative subjects. Even though there was 100% specificity for the PCA, logistic regression, BLDC, firefly, DFA, and PAC Bayesian learning classifiers, this did not assure good sensitivity values, except in the case of the harmonic classifier.
Table 7 exhibits the consolidated classifiers’ performance analysis across the different DR techniques. We can deduct from Table 7 that the Hilbert transformation DR approach with the harmonic search classifier retained its number one position with the highest PI of 96.485%, the highest accuracy of 98.31%, the lowest error rate of 3.38%, and the highest GDR of 96.55%, whereas the logistic regression classifier produced the lowest PI of 17.07%, the lowest accuracy of 77.38%, and the highest error rate of 45.245% under the Hilbert transformation DR technique. The lowest GDR of 40.755% was produced by the firefly classifier under the ABC-PSO DR method.
Figure 4 displays the performance analysis of the classifiers for the different DR techniques with respect to the error rate and GDR parameters. The harmonic search classifier achieved the highest GDR of 96.55%, with the lowest average error rate of 3.38% for the Hilbert transform DR technique, and the logistic regression classifier achieved the highest error rate of 45.245% for the Hilbert transform DR method, while the firefly classifier achieved the lowest GDR of 40.755% for the ABC-PSO DR technique.
Figure 5 demonstrates the performance analysis of the classifiers for the different DR techniques with respect to the accuracy. According to Figure 5, the harmonic search classifier clearly produced the highest accuracy of 98.31%, and the logistic regression classifier had the lowest accuracy of 77.38% for the Hilbert transformation dimensionality reduction method.
Next, we extensively examined the classifier accuracy as follows: for the PCA classifier, the higher accuracy of 87.45% was attained for the ABC-PSO DR method, and a lower accuracy of 80.36% was limited to the cuckoo search DR method. Figure 5 shows that the EM classifier maintained a high accuracy of 89.715% when using the cuckoo search DR method, while the NLR DR technique maintained a low accuracy of 82.255%. The high accuracy value for the logistic regression classifier was limited to 84.015% for the ABC-PSO DR method and a low accuracy of 77.38% for the Hilbert transformation DR technique. The GMM classifier placed a high accuracy of 95.12% with the dragonfly DR method and achieved a low accuracy of 86.79% with the NLR DR technique. The BLDC classifier secured a high accuracy of 90.63% with the cuckoo search DR method and retained a low accuracy of 80.21% with the ABC-PSO DR technique. The firefly classifier settled at a high accuracy of 94.145% with the NLR DR method and a low accuracy of 78.275% for the cuckoo search DR technique. For the harmonic search classifier, it showed a remarkable performance with a high accuracy of 98.31% with the Hilbert transformation DR method and maintained a low accuracy of 91.805% for the ABC-PSO DR technique. The harmonic search classifier maintained a high accuracy across the different DR methods, which was due to the better segregation and learning ability of the classifier. In the case of the DFA classifier, a high accuracy of 95.575% was achieved with the ABC-PSO DR method and a low accuracy of 88.38% was achieved with the NLR DR technique. The DFA classifier exhibited the second-best classification accuracy performance across the DR techniques. In the case of the PAC Bayesian learning classifier with the Hilbert transformation DR method, it retained a good accuracy of 83.525% and a low accuracy of 78.7% with the dragonfly DR method. The KNN-PAC Bayesian learning hybrid classifier reached a high accuracy of 89.26% for the Hilbert transformation DR method and maintained a low accuracy of 84.765% with the NLR DR method. For the SD classifier, it achieved a high accuracy of 94.99% with the cuckoo search DR technique, and it displayed a drastically low accuracy of 83.615% for the dragonfly DR technique. In the case of the DFA SDC hybrid classifier with the NLR DR method, it remained at a high accuracy at 92.585% and reached a low accuracy of 82.215% for the dragonfly DR method.
The robustness of the classifiers is reflected in the accuracy of the classifiers across the five dimensionality reduction techniques, namely Hilbert transform (HT), nonlinear regression (NLR), artificial bee colony–particle swarm optimization (ABC-PSO), cuckoo search, and dragonfly. All other classifiers, except the logistic regression classifier, settled at a higher accuracy of the maximum 80%, as shown in Table 7. This is due to the fact that the classifiers are trained and the optimal parameters for the classifiers are attained after the tuning process. The k-fold training and testing of the classifiers caused the classifiers to be more robust for the detection of CVD in the submitted PPG signal.

5.4. Summary of Previous Works on the Detection of CVD Classes

A summary of previous works on the detection of CVD classes is listed in Table 8. The time and frequency domain features, SVD, and stochastic features were extracted from the PPG signals, and these features were classified with various classifiers, such as ANN, KNN, ELM, GMM, softmax regression model, DNN, SDC, SVM, and harmonic search, to detect cases of CVD.
It is observed in Table 8 that Soltane et al. [47] proposed the artificial neural network (ANN) method to divide the PPG signal into two different classes. The input signal was smoothed to reduce the dimensionality, and the smoothing accuracy was used to explore the features in the multilayer feed-forward networks that were highly parallelized (MFN), and this achieved a classification rate for testing datasets of 94.7% and training datasets of 100%. Hosseini et al. [48] utilized finger PPG, a noninvasive optical signal collected before and after reactive hyperemia, to distinguish between people with various CVDs, with a maximum accuracy of 81.5% for the KNN classifier. Shobitha et al. [49] used the extreme learning machine (ELM), a supervised learning algorithm, to classify PPG signals as normal or affected by cardiovascular illness and compared its performance with backpropagation and support vector machine (SVM) techniques. These algorithms were validated by testing healthy and pathological signals from each of the 30 patients. In addition, with only five features as input, ELM produced the best accuracy, with a specificity of 90.33% and a sensitivity of 89.33%, and it also took less computational time to determine the risk of CVD. Prabhakar et al. [50] considered PPG signals obtained from a single patient. They extracted the statistical features, and the annotation of the PPG signals was conducted by using SVD. The annotated features of the class labels were verified and classified by the GMM and achieved an accuracy of 98.97%. This may be due to the smaller class vector size and overfitting condition for the GMM classifier. In the research work, heuristic- and transformation-based dimensionally reduced PPG data samples of 21 normal and 20 CVD cases were considered. The GMM classifier reached a maximum accuracy of 95.12% with the dragonfly dimensionality reduction technique. Based on patient diagnostic results for coronary heart disease, the classification and prediction models using deep neural networks (DNNs) were created and tested by Miao and Miao [51]. The created DNN learning model consisted of a classification model based on training data, and 303 clinical instances from patients with coronary heart disease at the Cleveland Clinic Foundation were used to create a prediction model for diagnosing new patient cases. The results of the tests indicate that the created classification and prediction model had an 83.67% diagnosis accuracy for heart disease. Hao et al. [52] proposed the softmax regression model, which employs neural networks for training and learning, and calculates the probability that reclassified data will fall into each category. This method classified CVD with an accuracy of 94.44%. Divya et al. [53] proposed a computer-aided diagnostic system that uses PPG signals to determine the different levels of CVD risk. From the PPG signals, statistical characteristics, wavelets, and singular value decomposition features were retrieved. By utilizing the SDC and GMM classifiers, the extracted feature vectors were classified to indicate the various risk levels of CVD, the results show that a classification accuracy of 97.88%, specificity of 99.09%, and a sensitivity of 97.24% were obtained by incorporating the SDC with value decomposition (SVD) and statistical features. In addition, a classification accuracy of 96.64%, specificity of 99.65%, and sensitivity of 93.80% were obtained by incorporating the GMM with SVD and statistical features. Prabhakar et al. [54] used a fuzzy-based approach to optimize the extracted parameters from PPG signals. The statistical features were extracted from the PPG signals, and fuzzy-based modeling was utilized to predict the CVD risk levels from the PPG signals. To optimize the fuzzy model levels, four types of optimization were performed. In order to produce the best results, the optimized values were categorized using the appropriate classifiers, and the support vector machine–radial basis function (SVM–RBF) classifier produced a maximum classification accuracy of 95.05% when the fuzzy model-based levels were optimized with animal migration optimization (AMO). A deep convolutional neural network was developed by Liu et al. [55] to classify multiple rhythms of 23,384 PPG waveforms from 45 patients and achieved an accuracy of 85%. Ihsan et al. [56] studied feature extraction algorithms, such as the respiratory rate (RR) interval, HRV features, and time domain features for detecting coronary heart disease using PPG and achieved an accuracy of 94.4% for HRV features using the decision tree classifier. Al Fahoum et al. [57] extracted the time domain features and health status information from PPG signals and applied feature selection-based classifiers in order to identify the difference between healthy persons and CVD patients. Seven distinct classifiers were utilized to classify the dataset and apply the feature selection. In the first stage, the naïve Bayes classifier achieved the highest accuracy of 94.44%, and in the second stage, an accuracy of 89.37% was attained. Rajaguru et al. [58] extracted the statistical features from the CapnoBase PPG signals of a single CVD patient, and the extracted features were classified with linear regression which produced a better accuracy of 65.85%.
In this research, the harmonic search classifier yielded the best classification accuracy of 98.31% for the HT DR values. The Hilbert transform is a linear operator that causes a 90-degree phase shift in a signal to obtain the desired separation, which is required in the exploration of phase in a harmonic search classifier. The harmonic search classifier is a pitch adjustment of the harmonics; therefore, more such phase exploration is possible to provide better classification. Hilbert transform segregates the signals at the first level itself, which reduces the burden of the classifiers; hence, the harmonic search classifier yielded a better classification accuracy. In order to identify a good classifier, the computational complexity performance measure plays a tradeoff role, as discussed below.

5.5. Computational Complexity Analysis of the Classifiers

The computational complexity may also be a performance metric for a classifier. Computational complexity is analyzed by utilizing an input size of m. If the size of the input is   O 1 , the computational complexity will be very low. There will be an increase in computational complexity if there is an increase in the number of inputs. Computational complexity is denoted as O ( log m ) when it increases log m times with respect to the increase in m.
Table 9 indicates the computational complexity of the classifiers among the various dimensionality reduction techniques. Under the Hilbert transformation DR technique, the logistic regression and firefly classifiers had the lowest computational complexity of O ( m log m ) . The highest computational complexity of O m 7 was reached by the KNN-PAC Bayesian classifier with the ABC-PSO optimization technique. Even though ABC-PSO and DFA had the highest computational complexity of O m 5 , a higher accuracy of 95.58% was exhibited. The higher accuracy of this classifier was due to the DFA’s characteristics feature. DFA identifies the peak value of the features, and the ABC-PSO will smooth the features and place them in the labeled classes without any outliers.

6. Conclusions

This study intended to detect cardiovascular disease (CVD) from PPG signals. The dimensionally reduced features obtained from the PPG signal were stored as datasets. Then, classifiers were used to detect CVD in the patients. The objective was to classify CVD with a high classification rate and a low rate of false positives and false negatives. Even though it is difficult to obtain perfect classification with classifiers, a compromise was made. As a high number of false positives decreases a classifier’s accuracy, therefore, a low number of false positives is the most important. The main limitation of this work is that the PAC Bayesian learning and logistic regression classifiers failed to achieve a higher classification accuracy across all five dimensionality reduction techniques, and thus the second-to-second detection of PPG classes will result in more false alarms. At the same time, 30 s segmented epochs of PPG signals were considered for the better classification accuracy of the classifiers. Under this circumstance, the classifiers will be overfitted with the training process and may end in higher accuracy. A compromise is made by taking the segment of a one-minute duration of raw PPG signals to attain a better classification accuracy. The results show that a high classification accuracy of 98.31% was attained when the Hilbert transform optimized values were classified with the harmonic search classifier, and a second highest accuracy of 97.79% was obtained when nonlinear regression optimized values were classified with the harmonic search classifier. A third highest accuracy of 96.095% was obtained when the cuckoo search optimized values were classified with the harmonic search classifier. It was also observed that the harmonic search classifier outperformed across all dimensionality reduction techniques. The convenience and real-time nature of a PPG-based method make it an attractive option for large-scale screening, which has the potential to be helpful in the long-term and real-time monitoring of CVD. PPG-based approaches could potentially be performed remotely without direct patient contact and with minimal patient training by wearable devices, such as fitness bands and smartwatches. As a result, the use of PPG-based methods could play a significant role in detecting CVD at an early stage and continuously measuring risk factors, leading to timely clinical evaluation. The further enhancement of the classifiers’ performance will be in the direction of the hyper-parameters’ selection through heuristic methods. The future research is toward CNNs and deep neural networks for the detection of CVD, with a minimum time lapse. Because CNNs are good at extracting features from PPG signals and identifying relevant patterns for CVD detection, deep neural networks can identify the most relevant risk factors and develop accurate models for CVD detection; by combining these two types of artificial intelligence, healthcare providers can more accurately diagnose and treat patients with CVD.

Author Contributions

Conceptualization, S.P.; Methodology, S.P.; Software, S.P.; Validation, H.R.; Formal analysis, H.R.; Investigation, S.P.; Resources, S.P. and H.R.; Data curation, H.R.; Writing—original draft, S.P.; Writing—review and editing, H.R.; Visualization, S.P.; Supervision, H.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Elgendi, M. On the analysis of fingertip photoplethysmogram signals. Curr. Cardiol. Rev. 2012, 8, 14–25. [Google Scholar] [CrossRef] [PubMed]
  2. Pilt, K.; Meigas, K.; Ferenets, R.; Kaik, J. Photoplethysmographic signal processing using adaptive sum comb filter for pulse delay measurement. Est. J. Eng. 2010, 16, 78–94. [Google Scholar] [CrossRef] [Green Version]
  3. Giannetti, S.M.; Dotor, M.L.; Silveira, J.P.; Golmayo, D.; Miguel-Tobal, F.; Bilbao, A.; Galindo, M.; Martín-Escudero, P. Heuristic algorithm for photoplethysmographic heart rate tracking during maximal exercise test. J. Med. Biol. Eng. 2012, 32, 181–188. [Google Scholar] [CrossRef]
  4. Al-Fahoum, A.S.; Al-Zaben, A.; Seafan, W. A multiple signal classification approach for photoplethysmography signals in healthy and athletic subjects. Int. J. Biomed. Eng. Technol. 2015, 17, 1–23. [Google Scholar] [CrossRef]
  5. Sukor, J.A.; Redmond, S.J.; Lovell, N.H. Signal quality measures for pulse oximetry through waveform morphology analysis. Physiol. Meas. 2011, 32, 369–384. [Google Scholar] [CrossRef]
  6. Di, U.; Te, A.; De, J. Awareness of Heart Disease Prevention among Patients Attending a Specialist Clinic in Southern Nigeria. Int. J. Prev. Treat. 2012, 1, 40–43. [Google Scholar]
  7. Tun, H.M. Photoplethysmography (PPG) Scheming System Based on Finite Impulse Response (FIR) Filter Design in Biomedical Applications. Int. J. Electr. Electron. Eng. Telecommun. 2021, 10, 272–282. [Google Scholar] [CrossRef]
  8. Ram, M.R.; Madhav, K.V.; Krishna, E.H.; Komalla, N.R.; Reddy, K.A. A novel approach for motion artifact reduction in PPG signals based on AS-LMS adaptive filter. IEEE Trans. Instrum. Meas. 2011, 61, 1445–1457. [Google Scholar] [CrossRef]
  9. Luke, A.; Shaji, S.; Menon, K.U. Motion artifact removal and feature extraction from PPG signals using efficient signal processing algorithms. In Proceedings of the International Conference on Advances in Computing, Communications and Informatics (ICACCI), Bangalore, India, 19–22 September 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 624–630. [Google Scholar] [CrossRef]
  10. Otsuka, T.; Kawada, T.; Katsumata, M.; Ibuki, C. Utility of second derivative of the finger photoplethysmogram for the estimation of the risk of coronary heart disease in the general population. Circ. J. 2006, 70, 304–310. [Google Scholar] [CrossRef] [Green Version]
  11. Shintomi, A.; Izumi, S.; Yoshimoto, M.; Kawaguchi, H. Effectiveness of the heartbeat interval error and compensation method on heart rate variability analysis. Healthc. Technol. Lett. 2022, 9, 9–15. [Google Scholar] [CrossRef]
  12. Moraes, J.L.; Rocha, M.X.; Vasconcelos, G.G.; Vasconcelos Filho, J.E.; De Albuquerque, V.H.; Alexandria, A.R. Advances in photopletysmography signal analysis for biomedical applications. Sensors 2018, 18, 1894. [Google Scholar] [CrossRef] [Green Version]
  13. Hwang, S.; Seo, J.; Jebelli, H.; Lee, S. Feasibility analysis of heart rate monitoring of construction workers using a photoplethysmography (PPG) sensor embedded in a wristband-type activity tracker. Autom. Constr. 2016, 71, 372–381. [Google Scholar] [CrossRef]
  14. Moshawrab, M.; Adda, M.; Bouzouane, A.; Ibrahim, H.; Raad, A. Smart Wearables for the Detection of Cardiovascular Diseases: A Systematic Literature Review. Sensors 2023, 23, 828. [Google Scholar] [CrossRef] [PubMed]
  15. Allen, J. Photoplethysmography and its application in clinical physiological measurement. Physiol. Meas. 2007, 28, R1. [Google Scholar] [CrossRef] [Green Version]
  16. Kumar, P.S.; Harikumar, R. Performance Comparison of EM, MEM, CTM, PCA, ICA, entropy and MI for photoplethysmography signals. Biomed. Pharmacol. J. 2015, 8, 413–418. [Google Scholar] [CrossRef] [Green Version]
  17. Almarshad, M.A.; Islam, S.; Al-Ahmadi, S.; BaHammam, A.S. Diagnostic Features and Potential Applications of PPG Signal in Healthcare: A Systematic Review. Healthcare 2022, 10, 547. [Google Scholar] [CrossRef]
  18. Yousefi, M.R.; Khezri, M.; Bagheri, R.; Jafari, R. Automatic detection of premature ventricular contraction based on photoplethysmography using chaotic features and high order statistics. In Proceedings of the 2018 IEEE International Symposium on Medical Measurements and Applications (MeMeA), Rome, Italy, 11–13 June 2018; pp. 1–5. [Google Scholar] [CrossRef]
  19. Ukil, A.; Bandyoapdhyay, S.; Puri, C.; Pal, A.; Mandana, K. Cardiac condition monitoring through photoplethysmogram signal denoising using wearables: Can we detect coronary artery disease with higher performance efficacy? In Proceedings of the Computing in Cardiology Conference (CinC), Vancouver, BC, Canada, 11–14 September 2016; pp. 281–284. [Google Scholar] [CrossRef]
  20. Almanifi, O.R.A.; Khairuddin, I.M.; Razman, M.A.M.; Musa, R.M.; Majeed, A.P.A. Human activity recognition based on wrist PPG via the ensemble method. ICT Express 2022, 8, 513–517. [Google Scholar] [CrossRef]
  21. Paradkar, N.; Chowdhury, S.R. Coronary artery disease detection using photoplethysmography. In Proceedings of the 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Jeju, Republic of Korea, 11–15 July 2017; pp. 100–103. [Google Scholar] [CrossRef]
  22. Neha; Kanawade, R.; Tewary, S.; Sardana, H.K. Neha; Kanawade, R.; Tewary, S.; Sardana, H.K. Photoplethysmography based arrhythmia detection and classification. In Proceedings of the 6th International Conference on Signal Processing and Integrated Networks (SPIN), Noida, India, 7–8 March 2019; pp. 944–948. [Google Scholar] [CrossRef]
  23. Prabhakar, S.K.; Rajaguru, H.; Lee, S.W. Metaheuristic-based dimensionality reduction and classification analysis of PPG signals for interpreting cardiovascular disease. IEEE Access 2019, 7, 165181–165206. [Google Scholar] [CrossRef]
  24. Sadad, T.; Bukhari, S.A.C.; Munir, A.; Ghani, A.; El-Sherbeeny, A.M.; Rauf, H.T. Detection of Cardiovascular Disease Based on PPG Signals Using Machine Learning with Cloud Computing. Comput. Intell. Neurosci. 2022, 2022, 1672677. [Google Scholar] [CrossRef]
  25. Karlen, W.; Turner, M.; Cooke, E.; Dumont, G.; Ansermino, J.M. CapnoBase: Signal database and tools to collect, share and annotate respiratory signals. In Proceedings of the 2010 Annual Meeting of the Society for Technology in Anesthesia, West Palm Beach, FL, USA, 13–16 January 2010; Society for Technology in Anesthesia: Milwaukee, WI, USA, 2010; p. 27. [Google Scholar]
  26. Velliangiri, S.; Alagumuthukrishnan, S.J. A review of dimensionality reduction techniques for efficient computation. Procedia Comput. Sci. 2019, 165, 104–111. [Google Scholar] [CrossRef]
  27. Van Der Maaten, L.; Postma, E.; Van den Herik, J. Dimensionality reduction: A comparative. J. Mach. Learn Res. 2009, 10, 13. [Google Scholar]
  28. Rajaguru, H.; Prabhakar, S.K. PPG signal analysis for cardiovascular patient using correlation dimension and Hilbert transform based classification. In New Trends in Computational Vision and Bioinspired Computing: ICCVBIC 2018; Springer: Cham, Switzerland, 2020; pp. 1103–1110. [Google Scholar] [CrossRef]
  29. Benitez, D.; Gaydecki, P.A.; Zaidi, A.; Fitzpatrick, A.P. The use of the Hilbert transform in ECG signal analysis. Comput. Biol. Med. 2001, 31, 399–406. [Google Scholar] [CrossRef] [PubMed]
  30. Smyth, G.K. Nonlinear regression. Encycl. Environ. Metr. 2002, 3, 1405–1411. [Google Scholar]
  31. Khuat, T.T.; Le, M.H. A novel hybrid ABC-PSO algorithm for effort estimation of software projects using agile methodologies. J. Intell. Syst. 2018, 27, 489–506. [Google Scholar] [CrossRef]
  32. Gandomi, A.H.; Yang, X.S.; Alavi, A.H. Cuckoo search algorithm: A metaheuristic approach to solve structural optimization problems. Eng. Comput. 2013, 29, 17–35. [Google Scholar] [CrossRef]
  33. Mirjalili, S. Dragonfly algorithm: A new meta-heuristic optimization technique for solving single-objective, discrete, and multi-objective problems. Neural Comput. Appl. 2016, 27, 1053–1073. [Google Scholar] [CrossRef]
  34. Esmael, B.; Arnaout, A.; Fruhwirth, R.; Thonhauser, G. A statistical feature-based approach for operations recognition in drilling time series. Int. J. Comput. Inf. Syst. Ind. Manag. Appl. 2012, 4, 100–108. [Google Scholar]
  35. Richman, J.S.; Moorman, J.R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiol. -Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef] [Green Version]
  36. Dai, J.J.; Lieu, L.; Rocke, D. Dimension reduction for classification with gene expression microarray data. Stat. Appl. Genet. Mol. Biol. 2006, 5, 1–19. [Google Scholar] [CrossRef]
  37. Fan, X.; Yuan, Y.; Liu, J.S. The EM algorithm and the rise of computational biology. Stat. Sci. 2010, 25, 476–491. [Google Scholar] [CrossRef] [Green Version]
  38. Li, G. Application of finite mixture of logistic regression for heterogeneous merging behavior analysis. J. Adv. Transp. 2018, 2018, 1436521. [Google Scholar] [CrossRef]
  39. Li, R.; Perneczky, R.; Yakushev, I.; Foerster, S.; Kurz, A.; Drzezga, A.; Kramer, S. Alzheimer’s Disease Neuroimaging Initiative. Gaussian mixture models and model selection for [18F] fluorodeoxyglucose positron emission tomography classification in Alzheimer’s disease. PLoS ONE 2015, 10, e0122731. [Google Scholar] [CrossRef] [Green Version]
  40. Fonseca, P.; Den Teuling, N.; Long, X.; Aarts, R.M. Cardiorespiratory sleep stage detection using conditional random fields. IEEE J. Biomed. Health Inform. 2016, 21, 956–966. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Yang, X.S. Firefly algorithm, stochastic test functions and design optimization. Int. J. Bioinspired Comput. 2010, 2, 78–84. [Google Scholar] [CrossRef]
  42. Bharanidharan, N.; Rajaguru, H. Classification of dementia using harmony search optimization technique. In Proceedings of the IEEE Region 10 Humanitarian Technology Conference (R10-HTC), Malambe, Sri Lanka, 6–8 December 2018; pp. 1–5. [Google Scholar] [CrossRef]
  43. Berthouze, L.; Farmer, S.F. Adaptive time-varying detrended fluctuation analysis. J. Neurosci. Methods 2012, 209, 178–188. [Google Scholar] [CrossRef]
  44. Guedj, B. A primer on PAC-Bayesian learning. arXiv 2019, arXiv:1901.05353. [Google Scholar]
  45. Aci, M.; Inan, C.; Avci, M. A hybrid classification method of k nearest neighbor, Bayesian methods and genetic algorithm. Expert Syst. Appl. 2010, 37, 5061–6067. [Google Scholar] [CrossRef]
  46. Rajaguru, H.; Prabhakar, S.K. Softmax discriminant classifier for detection of risk levels in alcoholic EEG signals. In Proceedings of the International Conference on Computing Methodologies and Communication (ICCMC), Erode, India, 18–19 July 2017; pp. 989–991. [Google Scholar] [CrossRef]
  47. Soltane, M.; Ismail, M.; Rashid, Z.A. Artificial Neural Networks (ANN) approach to PPG signal classification. Int. J. Comput. Inf. Sci. 2004, 2, 58–65. [Google Scholar]
  48. Hosseini, Z.S.; Zahedi, E.; Attar, H.M.; Fakhrzadeh, H.; Parsafar, M.H. Discrimination between different degrees of coronary artery disease using time-domain features of the finger photoplethysmogram in response to reactive hyperemia. Biomed. Signal Process. Control 2015, 18, 282–292. [Google Scholar] [CrossRef]
  49. Shobitha, S.; Sandhya, R.; Ali, M.A. Recognizing cardiovascular risk from photoplethysmogram signals using ELM. In Proceedings of the Second International Conference on Cognitive Computing and Information Processing (CCIP), Mysore, India, 12–13 August 2016; pp. 1–5. [Google Scholar] [CrossRef]
  50. Prabhakar, S.K.; Rajaguru, H. Performance analysis of GMM classifier for classification of normal and abnormal segments in PPG signals. In Proceedings of the 16th International Conference on Biomedical Engineering: ICBME 2016, Singapore, 7–10 December 2016; Springer: Singapore, 2017; pp. 73–79. [Google Scholar] [CrossRef]
  51. Miao, K.H.; Miao, J.H. Coronary heart disease diagnosis using deep neural networks. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 1–8. [Google Scholar] [CrossRef]
  52. Hao, L.; Ling, S.H.; Jiang, F. Classification of cardiovascular disease via a new softmax model. In Proceedings of the 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; pp. 486–489. [Google Scholar] [CrossRef]
  53. Ramachandran, D.; Thangapandian, V.P.; Rajaguru, H. Computerized approach for cardiovascular risk level detection using photoplethysmography signals. Measurement 2020, 150, 107048. [Google Scholar] [CrossRef]
  54. Prabhakar, S.K.; Rajaguru, H.; Kim, S.H. Fuzzy-inspired photoplethysmography signal classification with bioinspired optimization for analyzing cardiovascular disorders. Diagnostics 2020, 10, 763. [Google Scholar] [CrossRef] [PubMed]
  55. Liu, Z.; Zhou, B.; Jiang, Z.; Chen, X.; Li, Y.; Tang, M.; Miao, F. Multiclass Arrhythmia Detection and Classification from Photoplethysmography Signals Using a Deep Convolutional Neural Network. J. Am. Heart Assoc. 2022, 11, e023555. [Google Scholar] [CrossRef] [PubMed]
  56. Ihsan, M.F.; Mandala, S.; Pramudyo, M. Study of Feature Extraction Algorithms on Photoplethysmography (PPG) Signals to Detect Coronary Heart Disease. In Proceedings of the International Conference on Data Science and Its Applications (ICoDSA), Bandung, Indonesia, 6–7 July 2022; pp. 300–304. [Google Scholar] [CrossRef]
  57. Al Fahoum, A.S.; Abu Al-Haija, A.O.; Alshraideh, H.A. Identification of Coronary Artery Diseases Using Photoplethysmography Signals and Practical Feature Selection Process. Bioengineering 2023, 10, 249. [Google Scholar] [CrossRef]
  58. Rajaguru, H.; Shankar, M.G.; Nanthakumar, S.P.; Murugan, I.A. Performance analysis of classifiers in detection of CVD using PPG signals. In AIP Conference Proceedings; AIP Publishing LLC: Melville, NY, USA, 2023; Volume 2725, p. 020002. [Google Scholar] [CrossRef]
Figure 1. Organization of the CVD detection from PPG signals.
Figure 1. Organization of the CVD detection from PPG signals.
Bioengineering 10 00678 g001
Figure 2. Normal probability plot for Hilbert transform-based dimensionally reduced values for the PPG signals in cases of CVD.
Figure 2. Normal probability plot for Hilbert transform-based dimensionally reduced values for the PPG signals in cases of CVD.
Bioengineering 10 00678 g002
Figure 3. Normal probability plot for Hilbert transform-based dimensionally reduced values for PPG signals in normal cases.
Figure 3. Normal probability plot for Hilbert transform-based dimensionally reduced values for PPG signals in normal cases.
Bioengineering 10 00678 g003
Figure 4. Performance of the classifiers in terms of the error rate and GDR metrics for the different dimensionality reduction methods.
Figure 4. Performance of the classifiers in terms of the error rate and GDR metrics for the different dimensionality reduction methods.
Bioengineering 10 00678 g004
Figure 5. Performance of the classifiers’ accuracy for the different dimensionality reduction (DR) techniques.
Figure 5. Performance of the classifiers’ accuracy for the different dimensionality reduction (DR) techniques.
Bioengineering 10 00678 g005
Table 1. Average statistical parameters of different dimensionality reduction approaches for normal and CVD cases.
Table 1. Average statistical parameters of different dimensionality reduction approaches for normal and CVD cases.
Statistical
Parameters
Dimensionality Reduction Techniques
Hilbert
Transform
NLRABC PSOCuckoo SearchDragonfly
NormalCVDNormalCVDNormalCVDNormalCVDNormalCVD
Mean2.06115.4572.126117.4690.088860.79390.517611.6220.7803−5.461
Variance0.07061.16223.056225,124.70.01050.34120.046393.9237411.1297274.941
Skewness−1.118−0.81513.27895.769−0.1046−0.08880.1411−0.1172−0.0015−0.0084
Kurtosis5.6261.63814.5946.050.22260.0991−0.4971−1.6769−1.0065−0.74424
PCC0.36870.23320.010.0063−0.02150.01340.23190.24−0.15410.0775
Sample Entropy9.76959.93286.91427.39599.94949.94739.94944.99199.94999.9522
CCA0.54250.21980.10660.36740.4621
Table 2. The selection of the optimal parameters for the classifiers.
Table 2. The selection of the optimal parameters for the classifiers.
ClassifiersOptimal Parameters of the Classifiers
Principal component analysis (PCA)Decorrelated Eigen vector w k , and a threshold value of 0.72 with the training of trial and error method with MSE of (10)−5 or a maximum iteration of 1000—whichever happens first.
Expectation maximizationTest point likelihood probability 0.15, cluster probability of 0.6, with a convergence rate of 0.6. Criterion: MSE
Logistic regressionThreshold H θ x = 0.5 . Criterion: MSE
Gaussian mixture model (GMM)Mean, covariance of the input samples and tuning
parameter is EM steps. Criterion: MSE
Bayesian linear discriminant analysis (BLDC)Prior probability P(x): 0.5, class mean μ x = 0.8   and
μ y = 0.1 . Criterion: MSE
Firefly algorithm classifier γ = 0.1,   α s 0 = 0.65 initial conditions, with an MSE of (10)−5 or maximum iteration of 1000—whichever happens first. Criterion: MSE
Harmonic searchClass harmony is fixed at target values for the classes 0.85 and 0.1. The upper and lower bounds are adjusted with a step size of ∆w of 0.004. The final harmony aggregation is attained with an MSE of (10)−5 or a maximum iteration of 1000—whichever happens first. Criterion: MSE
Detrend fluctuation analysis (DFA)The initial values for K = 10000, M = 1000, and n = 100 are set to find F(n). Criterion: MSE
Probably approximately correct (PAC)
Bayesian learning method
P(z), class probability: 0.5, class mean: 0.8,0.1; γ = 0.13,
Criterion: MSE
KNN-PAC Bayesian learning methodNumber of clusters = 2 with PAC Bayesian variables.
Criterion: MSE
Softmax discriminant classifier (SDC) λ = 0.5 along with mean of each class target values as 0.1 and 0.85.
Detrend with SDCCascaded condition of DFA with SDC classifiers with
parameters as mentioned above.
Table 3. Analysis of the testing MSE for classifiers in CVD and normal cases.
Table 3. Analysis of the testing MSE for classifiers in CVD and normal cases.
ClassifiersCategoryHilbert
Transform
NLRABC PSOCuckooDragonfly
PCACVD2.81 × 10−57.84 × 10−62.50 × 10−76.60 × 10−41.44 × 10−6
Normal3.06 × 10−42.79 × 10−42.40 × 10−43.72 × 10−51.21 × 10−4
EMCVD1.60 × 10−53.97 × 10−51.44 × 10−43.24 × 10−67.74 × 10−5
Normal2.28 × 10−46.24 × 10−58.10 × 10−73.03 × 10−52.50 × 10−7
Logistic regressionCVD9.03 × 10−58.27 × 10−51.09 × 10−53.61 × 10−56.24 × 10−5
Normal3.23 × 10−41.84 × 10−52.56 × 10−44.84 × 10−41.33 × 10−4
GMMCVD1.44 × 10−64.41 × 10−66.40 × 10−54.84 × 10−66.76 × 10−6
Normal2.70 × 10−56.56 × 10−53.60 × 10−71.09 × 10−53.60 × 10−7
Bayesian LDCCVD2.92 × 10−52.50 × 10−55.49 × 10−51.00 × 10−62.22 × 10−5
Normal1.45 × 10−54.00 × 10−67.57 × 10−53.03 × 10−53.06 × 10−4
FireflyCVD4.41 × 10−46.25 × 10−68.84 × 10−56.40 × 10−58.41 × 10−6
Normal3.36 × 10−51.44 × 10−63.25 × 10−54.20 × 10−49.61 × 10−6
Harmonic searchCVD1.60 × 10−74.90 × 10−73.61 × 10−63.24 × 10−65.29 × 10−6
Normal9.00 × 10−83.24 × 10−66.76 × 10−61.60 × 10−74.90 × 10−7
DFA (weighted)CVD1.60 × 10−72.56 × 10−64.00 × 10−81.30 × 10−52.50 × 10−7
Normal6.25 × 10−65.04 × 10−57.84 × 10−62.03 × 10−56.76 × 10−6
PAC Bayesian learningCVD4.90 × 10−73.35 × 10−43.06 × 10−41.68 × 10−55.78 × 10−5
Normal2.89 × 10−45.04 × 10−52.70 × 10−51.19 × 10−43.03 × 10−4
KNN-PAC BayesianCVD7.84 × 10−68.41 × 10−66.25 × 10−61.00 × 10−68.41 × 10−6
Normal2.40 × 10−51.32 × 10−41.96 × 10−41.10 × 10−44.76 × 10−5
SDCCVD2.89 × 10−64.84 × 10−61.21 × 10−41.44 × 10−63.69 × 10−4
Normal8.41 × 10−61.15 × 10−51.31 × 10−52.57 × 10−61.16 × 10−5
Detrend SDCCVD2.40 × 10−48.10 × 10−73.06 × 10−41.02 × 10−54.41 × 10−4
Normal1.69 × 10−51.75 × 10−51.76 × 10−51.85 × 10−51.86 × 10−5
Table 4. Confusion matrix for the detection of CVD.
Table 4. Confusion matrix for the detection of CVD.
Actual Classification
Class Output
Predicted Classification Class Output
CVDNormal
CVDTPFN
NormalFPTN
Table 5. Confusion Matrix for Classifiers based on PPG Signal Segments for Hilbert Transform.
Table 5. Confusion Matrix for Classifiers based on PPG Signal Segments for Hilbert Transform.
ClassifiersTPTNFPFN
PCA10,080792072004320
EM8640792072005760
Logistic regression11,520792072002880
GMM12,96010,80043201440
Bayesian LDC10,08012,24028804320
Firefly720010,80043207200
Harmonic search14,40014,4007200
DFA (weighted)13,68012,9602160720
PAC Bayesian learning11,520792072002880
KNN-PAC Bayesian7920792072006480
SDC12,96012,96021601440
Detrend SDC720011,52036007200
Table 6. Performance analysis of cuckoo search DR method for different classifiers.
Table 6. Performance analysis of cuckoo search DR method for different classifiers.
ClassifiersPI
(%)
Sensitivity
(%)
Specificity
(%)
Accuracy
(%)
GDR (%)
Error Rate (%)
PCA30.97560.72510080.3660.7239.28
EM72.694.0185.4289.71573.42520.57
Logistic regression30.17560.12510080.06560.12539.875
GMM81.39510084.3892.1981.41515.625
Bayesian LDC75.20581.25510090.6381.25518.745
Firefly20.7656.1510078.27556.1543.855
Harmonic search91.2992.18510096.09592.197.81
DFA (weighted)71.91577.9951008977.99522.005
PAC Bayesian learning48.60567.110082.8467.08532.915
KNN-PAC Bayesian56.25577.9395.83586.0973.38526.235
SDC89.01594.66595.31594.9989.49510.02
Detrend SDC75.63588.5591.1589.8577.7920.315
Table 7. Consolidated classifiers’ performance analysis across the different DR techniques.
Table 7. Consolidated classifiers’ performance analysis across the different DR techniques.
ClassifiersPerformance
Metrics
HilbertNLRABC PSOCuckooDragonfly
PCAPI35.6947.19555.14530.97555.215
Error rate37.7831.0725.139.2826.69
Accuracy81.1184.26587.4580.3686.655
GDR62.2268.93574.960.7273.31
EMPI47.144.28555.4172.660.925
Error rate33.3535.54526.1120.5723.3
Accuracy83.33582.22586.9589.71588.35
GDR44.48564.45555.5773.42561.05
Logistic
regression
PI17.0748.95546.19530.17527.255
Error rate45.24532.8631.9839.87541.93
Accuracy77.3883.57584.01580.06579.035
GDR54.75551.1366.12560.12545.555
GMMPI75.44559.20564.68581.39588.815
Error rate18.6826.4322.1315.6259.76
Accuracy90.6686.7988.93592.1995.12
GDR81.3272.61564.90581.41588.945
Bayesian LDCPI68.63573.28534.14575.20537.65
Error ate24.34520.31539.5818.74536.74
Accuracy87.8389.84580.2190.6381.63
GDR75.65578.79560.4281.25563.26
FireflyPI31.7286.85540.71520.7680.075
Error rate39.2311.71536.52543.85516.595
Accuracy80.3994.14581.7478.27591.7
GDR60.7787.0440.75556.1581.85
Harmonic searchPI96.48595.3582.6591.2989.405
Error rate3.384.42516.4057.819.375
Accuracy98.3197.7991.80596.09595.315
GDR96.5595.57583.59592.1990.625
Detrend fluctuation analysis (weighted)PI89.5766.1289.6571.91587.29
Error rate9.1123.2358.8522.00512.495
Accuracy95.44588.3895.5758993.76
GDR90.8976.76591.1577.99587.505
PAC Bayesian
learning
PI44.9526.88535.8248.60525.285
Error rate32.9641.63537.71532.91542.39
Accuracy83.52579.2281.1482.8478.7
GDR64.7458.36562.28567.08546.055
KNN-PAC
Bayesian
PI71.87549.75550.00556.25563.69
Error rate21.4830.4729.9526.23525.45
Accuracy89.2684.76585.02586.0987.275
GDR78.5269.53570.0573.38574.55
SDCPI84.21581.17548.4189.01543.41
Error rate13.53515.75531.7710.0232.915
Accuracy93.2392.12584.11594.9983.615
GDR86.46583.2368.2389.49567.09
Detrend SDCPI45.63583.39542.9375.63539.775
Error rate33.82514.8434.65520.31535.585
Accuracy83.09592.58582.67589.8582.215
GDR66.17584.87565.34577.7964.42
Table 8. Summary of previous works on the detection of CVD classes.
Table 8. Summary of previous works on the detection of CVD classes.
Sl.noAuthorsFeaturesClassifierAccuracy (%)
1Soltane et al. [47]
2004
Time and frequency domain featuresArtificial neural network94.70%
2Hosseini et al. [48]
2015
Time domain featuresK-nearest neighbor81.50%
3Shobita et al. [49]
2016
Time domain featuresExtreme learning machine82.50%
4Prabhakaret al. [50]
2017
Statistical
features +SVD
GMM98.97%
5Miao and Miao [51]
2018
Time domain featuresDeep neural networks83.67%
6Hao et al. [52]
2018
Statistical featuresSoftmax regression model94.44%
7Divya et al. [53]
2019
SVD + statistical features + waveletsSDC97.88%
GMM96.64%
8Prabhakar et al. [54]
2020
Fuzzy-inspired statistical featuresSVM–RBF (kernel) for CVD95.05%
RBF neural network-for normal94.79%
9Liu et al. [55]
2022
Time domain featuresDeep convolutional neural network85%
10Ihsan et al. [56]
2022
HRV features and time domain featuresDecision tree classifier94.4%
11Al Fahoum et al. [57]
2023
Time domain featuresNaive Bayes94.44% in first stage
89.37% in second stage
12Rajaguru et al. [58]
2023
Statistical featuresLinear regression65.85%
13As reported in this paperHilbert transformHarmonic search
classifier
98.31%
Table 9. Classifiers’ computational complexity among various dimensionality reduction techniques.
Table 9. Classifiers’ computational complexity among various dimensionality reduction techniques.
ClassifiersOptimization Techniques
HilbertNLRABC PSOCuckooDragonfly
PCA O m 2 O ( m 2 log m ) O m 5 O ( 2 m 2 log m ) O ( 4 m 2 log m )
EM O m 2 O ( m 2 log m ) O m 5 O ( 2 m 2 log m ) O ( 4 m 2 log m )
Logistic regression O ( m log m ) O ( 2 m log m ) O ( m 4 log m ) O ( 4 m log m ) O ( 8 m log m )
GMM O m 2 O ( m 2 log m ) O m 5 O ( 2 m 2 log m ) O ( 4 m 2 log m )
Bayesian LDC O m 3 O ( m 3 log m ) O m 6 O ( 2 m 3 log m ) O ( 4 m 3 log m )
Firefly O ( m log m ) O ( 2 m 2 log m ) O ( m 4 log m ) O ( m log m 2 ) O ( 8 m log m )
Harmonic search O m 3 O ( m 3 log m ) O m 6 O ( 2 m 3 log m ) O ( 4 m 3 log m )
DFA (weighted) O m 2 O ( m 2 log m ) O m 5 O ( 2 m 2 log m ) O ( 4 m 2 log m )
PAC Bayesian l earning O m 3 O ( m 3 log m ) O m 6 O ( 2 m 3 log m ) O ( 4 m 3 log m )
KNN-PAC Bayesian O m 4 O ( m 4 log m ) O m 7 O ( 2 m 4 log m ) O ( 4 m 4 log m )
SDC O m 2 O ( m 3 log m ) O m 5 O ( 2 m 2 log m ) O ( 4 m 2 log m )
Detrend SDC O m 3 O ( m 4 log m ) O m 6 O ( 2 m 3 log m ) O ( 4 m 3 log m )
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Palanisamy, S.; Rajaguru, H. Machine Learning Techniques for the Performance Enhancement of Multiple Classifiers in the Detection of Cardiovascular Disease from PPG Signals. Bioengineering 2023, 10, 678. https://doi.org/10.3390/bioengineering10060678

AMA Style

Palanisamy S, Rajaguru H. Machine Learning Techniques for the Performance Enhancement of Multiple Classifiers in the Detection of Cardiovascular Disease from PPG Signals. Bioengineering. 2023; 10(6):678. https://doi.org/10.3390/bioengineering10060678

Chicago/Turabian Style

Palanisamy, Sivamani, and Harikumar Rajaguru. 2023. "Machine Learning Techniques for the Performance Enhancement of Multiple Classifiers in the Detection of Cardiovascular Disease from PPG Signals" Bioengineering 10, no. 6: 678. https://doi.org/10.3390/bioengineering10060678

APA Style

Palanisamy, S., & Rajaguru, H. (2023). Machine Learning Techniques for the Performance Enhancement of Multiple Classifiers in the Detection of Cardiovascular Disease from PPG Signals. Bioengineering, 10(6), 678. https://doi.org/10.3390/bioengineering10060678

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop