Next Article in Journal
Kinetics and Fluid-Specific Behavior of Metal Ions After Hip Replacement
Previous Article in Journal
Towards Wideband Characterization and Modeling of In-Body to On-Body Intrabody Communication Channels
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Pragmatic Models for Detection of Hypertension Using Ballistocardiograph Signals and Machine Learning

Department of Artificial Intelligence Convergence, Hallym University, Chuncheon 24252, Republic of Korea
*
Author to whom correspondence should be addressed.
Bioengineering 2026, 13(1), 43; https://doi.org/10.3390/bioengineering13010043
Submission received: 21 November 2025 / Revised: 23 December 2025 / Accepted: 27 December 2025 / Published: 30 December 2025
(This article belongs to the Section Biosignal Processing)

Abstract

To identify hypertension, Ballistocardiograph (BCG) signals can be primarily utilized. The BCG signal must be thoroughly understood and interpreted so that its application in the classification process could become clearer and more distinct. Various unhealthy habits such as excess consumption of alcohol and tobacco, accompanied by a lack of good diet and a sedentary lifestyle, lead to hypertension. Common symptoms of hypertension include chest pain, shortness of breath, blurred vision, mood swings, frequent urination, etc. In this work, two pragmatic models are proposed for the detection of hypertension using BCG signals and machine learning models. The first model uses K-means clustering, the maximum overlap discrete wavelet transform (MODWT) and the Empirical Wavelet Transform (EWT) techniques for feature extraction, followed by the Binary Tunicate Swarm Algorithm (BTSA) and Information Gain (IG) for feature selection, as well as two efficient hybrid classifiers such as the Hybrid AdaBoost–-Maximum Uncertainty Linear Discriminant Analysis (MULDA) classifier and the Hybrid AdaBoost–Random Forest (RF) classifier for the classification of BCG signals. The second model uses Principal Component Analysis (PCA), Kernel Principal Component Analysis (KPCA) and the Random Feature Mapping (RFM) technique for feature extraction, followed by IG and the Aquila Optimization Algorithm (AOA) for feature selection, as well as two versatile hybrid classifiers such as the Hybrid AutoRegressive Integrated Moving Average (ARIMA)–AdaBoost classifier and the Time-weighted Hybrid AdaBoost–Support Vector Machine (TW-HASVM) classifier for the classification of BCG signals. The proposed methodology was tested on a publicly available BCG dataset, and the best results were obtained when the KPCA feature extraction technique was used with the AOA feature selection technique and classified using the Hybrid ARIMA–AdaBoost classifier, reporting a good classification accuracy of 96.89%.

1. Introduction

The BCG is an estimate of ballistic potencies produced and initiated by the heart [1]. With every heartbeat, when the blood is ejected suddenly into the great vessels, the motions of the human body have a repetitive format and are represented graphically with the help of the BCG [2]. The utilization of the BCG lies in the use of a ballistographic scale, where the recoil of the body of the patient can be measured and analyzed. The heart rate of a person can be exhibited very well with the help of a BCG scale [3]. The pumping action of the heart and other body movements can be apprehended well with the help of non-invasive sensors, and some mechanical forces are generated and recorded, known as BCG signals [4]. With every heartbeat, the ejection of the blood is performed into the aorta when the recoil of the body is generated, and this is represented as an actual form of BCG. The BCG signals help to capture the heartbeat forces, vital sign monitoring and other body movements. Non-invasive sensors are utilized to record BCG signals. Some of the sensor types used to analyze the BCG signal include fiber optic sensors, strain gauge, piezoelectric, gyroscope, pneumatic, hydraulic, polyvinylidene and electromechanical film sensors, etc. [5]. The BCG signals are subject to data processing such as filtering and amplification before being processed for full-fledged analysis. Some of the wonderful applications of BCG include non-obtrusive monitoring, sleep tracking, infant monitoring and other clinical applications, including stroke volume estimation and blood pressure estimation [6]. As with other biomedical signals, BCG signals also are subject to a lot of problems, such as signal noise, standardization issues and data interpretation problems.
The heartbeat-prompted motions of the human body, which are continuously repeated, can be graphically represented with the help of this non-invasive technique. During the systole and diastole periods of heartbeat, the blood accelerates rapidly, leading to these motions being repeated incessantly. The overall execution of the circulatory system can be assessed well with the help of BCG. The mass movements of the cardiac cycle are measured and analyzed by this technique. When the systolic and diastolic phases occur, a lot of components can become shifted due to the changes happening in the body movements, cardiac activity and respiratory nodes [7]. Because of these movements, the BCG waveform is generated, as there would be a significant change in the distribution of blood during this cardiac cycle. Earlier BCG did not have a proper and standardized measurement protocol; moreover, the origin of this waveform was not understood well [8]. For clinical diagnosis of cardiovascular disorders such as myocardial infarction and coronary artery diseases, a high level of reliability and specificity is required, which was not reached by BCG at that time. Moreover, more advanced techniques such as echocardiography, photoplethysmography (PPG) and ultrasound proved to be a success, as they provided a comparatively high level of reliability and sensitivity to medical researchers who work in this domain [9]. Currently, due to the advent of both software and hardware technologies, a lot of interest has been received by BCG. BCG sensors can be embedded in medical environments easily, even without the involvement of medical professionals or the general public. Nowadays, the stress associated with medical checkup for patients has been greatly reduced. With the help of multiple technological leads, the concept of BCG has been received well in the past two decades [10]. Therefore, the concept of BCG is highly useful for tracking sleep-related disorders, managing respiratory disorders and monitoring cardiac functions. Moreover, the easy availability and accessibility of BCG allow the public to employ it in their homes for tracking the daily activities of the patient.
Genetics and other environmental factors play a vital role in hypertension [11]. Reducing stress and improving one’s lifestyle by following a good diet and exercise pattern can improve the symptoms of hypertension. When the heart pumps, the BCG signals can be generated, and there is a good correlation in detecting hypertension or any cardiovascular disorder associated with it [12]. The vital information about the blood volume pumped by the heart per 60 s can be obtained with the help of BCG signals. The elevated blood pressure can be described well by the BCG signals, and solid data can be provided about the blood vessel stiffness too [13]. A lot of important challenges are faced by the healthcare systems these days. To improve healthcare quality, a lot of invasive and non-invasive techniques have been developed by researchers, largely considering the effectiveness and cost factors, and in this aspect, BCG plays quite an important role in improving the health quality of the patients to a great extent. The morphology of the BCG signal varies greatly within the subjects, and so BCG signal processing is quite a challenging task [14].
An in-depth study of BCG covering its theory and developments was discussed in detail by Pinheiro et al. [15]. For the analysis of a medical cardiovascular physiology system, a BCG system was developed by Antonio [16], and a comparative study was conducted by Sadek and Biswas [17] based on the non-intrusive heart rate measurement utilizing the BCG signals. The BCG peak detection techniques were discussed in detail by Suliman et al. [18], and various clustering techniques utilized to detect the heartbeat in the BCG signals were elaborated by Paalasmaa and Ranta [19]. In a moving wheelchair equipped with embedded sensors, the acquisition of BCG was studied by Pinheiro et al. [20], and a maximum dispersion technique was used for the detection of a simple real-time heartbeat in BCG signals by Choe et al. [21]. A versatile BCG acquisition system was developed for home monitoring purposes by Inan et al. [22], and the motion artifacts were detected automatically in the BCG based on a modified bathroom scale by Wiard et al. [23]. The heart rate estimation based on an adaptive beat-to-beat basis for the BCG signals was developed by Bruser et al. [24], and a novel technique for BCG measurements utilizing fiber Bragg grating-based sensors was developed by Lukasz et al. [25]. The Phonocardiogram (PCG), Electrocardiogram (ECG) and BCG signals were thoroughly compared for the application of heart rate monitoring of the human body by Nedoma et al. [26].
A lot of the recent important works for the detection of hypertension using BCG signals are discussed as follows. Using ensemble empirical mode decomposition (EMD), the detection of hypertension was performed using BCG signals, and an accuracy of 92.3% was reported [27]. The mining class association, along with many features was used to identify the hypertension by Liu et al., achieving an accuracy of 85.20% [28]. By means of integrating association rule mining and classification rules, hypertension could be identified by Liu et al. with an accuracy of 84.40% [29]. A deep CNN architecture with two stages was used for the classification of low-risk versus high-risk hypertension using ECG signals, and a higher accuracy of 99.68% was obtained [30]. The EMD with wavelet decomposition and ensemble gentle-boost classifier was used by Rajput et al., achieving an accuracy of 89% [31]. A continuous wavelet transform with deep neural networks (DNN) was used by Rajput et al., achieving an accuracy of 86.14% [32]. A Gabor transform-based deep convolutional neural network (CNN) was used for the automated detection of hypertension by Gupta et al., achieving an accuracy of 97.65% [33]. Depending on the time–frequency domain and deep learning features, pulmonary hypertension was detected using a PCG dataset, achieving a classification accuracy of 88.61% [34]. The power-normalized cepstral coefficients, along with CNN, were used for the analysis of congenital heart disease-related pulmonary arterial hypertension using heart sound signals, achieving an accuracy of 88.61% [35]. A deep CNN technique was used for hypertension on pregnancy disorders achieving an accuracy of 93.10% [36]. A novel odd–even pattern was utilized for the detection of hypertension using BCG signals, achieving an accuracy of 97.78% [37]. A CNN was used with fine-tuned spectrogram images and ConvMixer model, achieving an accuracy of 97.69% [38]. An integrated tunable Q-factor wavelet transform along with the concept of multi-verse optimization was proposed by Gupta et al., achieving a classification accuracy of 92.21% [39]. In this work two interesting pragmatic models were proposed for the analysis and classification of hypertension from BCG signals. The organization of this work is as follows:
Once the basic pre-processing of the BCG signals is performed by means of implementing a simple Independent Component Analysis (ICA), it is then processed with the help of two proposed pragmatic models.
The first model uses the idea of K-means clustering, MODWT and EWT for feature extraction followed by the usage of BTSA and IG for feature selection and two hybrid classifiers such as Hybrid AdaBoost–MULDA classifier and Hybrid AdaBoost–Random Forest classifier.
The second model uses the PCA, KPCA and the RFM technique for feature extraction, followed by the usage of AOA and IG for feature selection and two hybrid classifiers, such as the Hybrid ARIMA–AdaBoost classifier and the TW-HASVM classifier.
The workflow implemented in this paper is the first of its kind, as no work in the past has adopted such a workflow employing such a hybrid combination to test its versatility on BCG signal analysis.
The overall pictorial representation of the work is illustrated in Figure 1 below.

2. Proposed Pragmatic Model 1

The first model uses the idea of K-means clustering, MODWT and EWT for feature extraction, followed by the usage of BTSA and IG for feature selection and two hybrid classifiers, such as the Hybrid AdaBoost–MULDA classifier and the Hybrid AdaBoost–Random Forest classifier.

2.1. Feature Extraction Using K-Means Clustering, MODWT and EWT

2.1.1. K-Means Clustering

For the categorization of unlabeled data, K-means clustering is used so that the cluster centers can be located easily [40]. In an iterative manner, the cluster centers are changed by this K-means algorithm, and so the overall variation can be mitigated amongst these clusters. The distance between the center of a cluster and its respective data can be reduced drastically by assigning observations to the K-clusters. When there is a change in the group center, the redistribution of the data points happens and then the computation of group center can be performed easily. The grouping of the entire data can be carried out into specific numbers of K categories. The basic procedure for utilizing the K-means algorithm is as follows:
In a random manner, the creation of the K data points is performed so that the primary cluster centers can be created.
The computation of the distance of every data point from the cluster centers is performed.
The assignment of the data point to any of the neighboring clusters is performed depending on a proximity basis.
Assume a set of q -dimensional vectors with the notation P i = p i 1 , p i 2 , , p i q . The square of the Euclidean distance usually assesses the dissimilarity measure and is expressed as follows:
d p i , p j = j = 1 q p i j p i j 2 = p i p j 2
Within a particular cluster, the data is assigned, and the establishment of a new cluster center is determined.
For every cluster, the process is repeated until the progress is quite satisfactory.

2.1.2. Maximal Overlap Discrete Wavelet Transforms (MODWT)

With the help of three unique stages, the MODWT can be explained as Discrete Wavelet Transform (DWT), Discrete Wavelet Packet Transform (DWPT) and, finally, MODWT [41]. An incoming signal can be evaluated easily with this transformation. Initially, DWT helps to decompose the input signal so that the estimates can be obtained correctly. An incoming signal is split by DWT into scaling coefficients and wavelet coefficients. Secondly, when a gradual refinement of DWT happens, it is projected as DWPT [42]. The modification of the DWT pyramid technique is performed as if it was an orthonormal transformation. At a specific level, both the wavelets and scaling coefficients are decomposed by DWPT, and so the prediction of parameters becomes quite easy. Some of the decomposition coefficients can be easily lost with the help of DWT and DWPT methods, as they are highly time-variant transforms. A time-invariant transformation is therefore created so that the wavelet values are aligned correctly with the input parameters and it is performed without the aid of down-sampling method. In MODWT, the input signal is split with the help of high- and low-pass filters, and the coefficients are obtained, which would have a good pass-band period during the decomposition stage [41]. The input signal length becomes similar to the coefficient’s length when the down-sampling is eliminated. Every produced decomposed constituent finally has a corresponding time series, which is strongly linked to the others.

2.1.3. Empirical Wavelet Transform (EWT)

In different areas of data sciences, one of the most powerful signal processing tools implemented is EWT [43]. The implementation of fixed basis functions is restricted mostly in wavelet transformations. The signal decomposition can be performed by EMD to a certain extent, where the oscillatory components could be utilized to highlight the properties of the input signal. When using the EMD, the derivation of the basis function is quite easy from the input signal. Though EMD has proved its mettle in various applications, it still has some drawbacks due to the lack of a strong mathematical foundation, so EWT came into existence. EWT uses the properties of both EMD and WT. A time-series signal is transformed by EWT decomposition into a collection of time series, which exhibits a stable variance when compared with the original signal, so the ultimate result would be more accurate and precise. Fourier segments can be assessed easily with the help of EWT, and so several wavelet filters can be constructed easily so that different modes can be extracted from the time-series signal. When the specification of Fourier segments is given, the application of EWT occurs. The local maxima of the spectrum are located and designed by this empirical rule. Finally, the Fourier segments boundaries are used and projected as the midpoint of the adjacent local maxima.

2.2. Feature Selection Using Binary Tunicate Swarm Algorithm (BTSA)

One of the famous swarm-based optimization techniques is TSA, where the search mechanism of food source of tunicates is mimicked in the deep ocean [44]. With the help of a jet propulsion scheme, all the tunicates progress towards food sources, and an extreme level of intelligence is exhibited. To solve an optimization issue, the above behavior of tunicates is modelled mathematically. The tunicates’ position is generally assigned to a binary vector in the case of BTSA. The tunicates exhibit a wonderful swarm behavior, which enables them to move towards the best tunicate. A good proximity is maintained among the tunicates and, at the same time, collision is avoided among them in a careful manner. The mathematical modelling of the tunicate behavior is performed as follows.
When the tunicates search for food, they try to avoid a lot of collision, and certain parameters are utilized to control this behavior and are represented as follows:
T = G S
where the gravitational force is expressed by G , and the social force is expressed by S , which largely helps to affect the tunicate movement. With the aid of water-flow advection W , the control of the gravitational force is performed as follows:
G = k 1 + k 2 W
W = 2 k 3
where k 1 , k 2 , k 3 are represented as random parameters represented in the range of [0, 1]. The parameter S is expressed as follows:
S = w i + k 3 w m w i
where the w i expresses minimum speed, and w m expresses maximum speed so that a large number of social interactions are made. The tunicate helps to find the best position in the process. To avoid collision, a distance is maintained among the remaining tunicates, and the distance is expressed as follows:
D h = B q r 1 . B h ( i )
where D h represents the distance of current location of the h t h tunicate B h ( i ) , and r 1 represents a random number in the range of [0, 1]. Around the best tunicate, the remaining tunicates remain, and they trace the optimal position of the food source. This mechanism is expressed as follows:
B h ( i ) = B q + T D h , r a n d 0.5 B q T D h , r a n d 0.5
The updating of the h t h tunicates position is expressed as follows:
B h ( i + 1 ) = B h ( i ) + B q 2 + k 1
To match the feature selection issues, the conversion of these positions into a binary value is quite mandatory. By means of implementing various transfer functions, such as V-shaped and S-shaped, this conversion occurs. Generally, it is observed that the efficiency is much higher in the V-shaped transfer function when compared to the S-shaped transfer function. With the help of the V-shaped transfer function, the design of BTSA is performed. The positions are updated in the subsequent iteration from 0 to 1 with the help of the transfer function. The updating of the tunicate position in the BTSA is expressed as follows:
T . F ( t ) = 2 π arctan π 2 t
where T . F represents a transfer function.
B h ( i + 1 ) = B h ( i ) , r a n d   < T . F B h ( i + 1 ) B h ( i ) , r a n d   T . F B h ( i + 1 )
where B h ( i ) indicates the complement of B h ( i ) .
The fitness function here depends on Feature Selection Ratio (FSR) and classification error, and is designed as follows:
F i t n e s s = k 1 e r r o r + k 2 F S R
The original classification error is denoted by e r r o r . The weighting factors are indicated by k 1 and k 2 , falling in the range of [0, 1]. To have a lower classification burden, a lower FSR is generally required.

2.3. Classification Using Two Hybrid Models

Two hybrid classifiers are proposed, such as the Hybrid AdaBoost–MULDA and the Hybrid AdaBoost–RF classifiers.

2.3.1. Hybrid AdaBoost–MULDA Classifier

(A) MULDA:
For the non-dominant values, the estimation errors are higher and moreover the impact of the principal components largely affects the classification accuracy. To address this issue, MULDA is usually used in literature [45]. In MULDA, the smaller eigen values can be enlarged well so that the larger eigen values do not change. The instabilities can be reduced greatly by MULDA so that the identity matrix can be stabilized well thereby managing the within-class covariance matrix in an efficient manner. For stabilizing the identity matrix, choosing the maximum entropy technique is highly essential. Assuming a data comprising of n samples and is split into ‘c’ number of classes. Assume P = p 1 , p 2 , , p n d indicate the sample mean, p ¯ specifies the sample mean in a class. The MULDA procedure is implemented as follows:
Compute the within class matrix and between class matrix M w and M b .
Calculate M q = M w / n c .
Compute the eigenvector matrix φ and eigenvalue diagonal matrix Λ .
Assess the eigenvalues λ of M q .
Assess the average eigenvalues λ ¯ .
New eigenvalues matrix Λ * is generated by replacing λ with λ ¯ .
Compute the within-class scatter matrix as: M w * = M q * ( n c ) = φ * φ T n c .
The discriminant transformation vector is obtained ultimately by M w * 1 M b .
For the training samples, c 1 discriminant transformation vectors are calculated. Then the projection of test samples on the training samples is performed in the subsequent stage. Using a KNN algorithm, the performance of classification is assessed and finally the effectiveness of this validation is evaluated to serve the classification accuracy.
(B) Hybrid AdaBoost–MULDA classification:
AdaBoost helps to transform many weak classifiers into one strong classifier [46]. MULDA is hybrid with this concept so that better classification accuracy could be obtained. A lot of weight adjustments are performed iteratively and misclassified samples are given more importance in this hybrid model. The new weak classifiers are trained iteratively and so the misclassified samples are prioritized greatly, so that a best predictive classifier can be obtained. For diverse classification tasks, AdaBoost classifier is used as it could very well handle high dimensional data in a versatile manner. For these reasons, MULDA is hybrid with AdaBoost so that it could develop as a strong classification model. Across all the samples, the weight distribution is split equally to all the combinations. The classification error is then computed so that the weights can be adjusted normally. Increased weightage is given to the misclassified samples and less weightage is given to the correctly classified samples. So, the primary focus is on error-prone samples by means of iteratively adjusting the weights leading to multiple MULDA classifiers. Each sample must be concentrated keenly so that the effectiveness of every classifier is projected clearly as the performance output measure. In this Hybrid AdaBoost-MULDA classification model, the effect of every MULDA classifier is assessed by the classification accuracy. This model ensures and strives to bring out the best outcome in terms of classification accuracy.

2.3.2. Hybrid AdaBoost–RF Classifier

An ensemble machine learning technique can be modelled with the help of AdaBoost classifier and RF technique [47]. The reassignment of every weight of the value can be applied to each instance which has a higher weight. Various weak classifiers can be merged so that a robust classifier can be constructed easily. There is a less necessity for tweaking parameters and it is easy to implement as it is not prone to over fitting. The simple pseudocode for the AdaBoost algorithm is expressed in Algorithm 1 as follows:
Algorithm 1: AdaBoost Classification
Assume: a i 1 , b i 1 a i z , b i z , a i i A i , b i i 1 , 1
Start D i i ( i i ) = 1 z i
For t i = 1 , , T i
Weak classifier training phase using the distribution D i
Obtain weak hypothesis h i i : A i 1 , 1
Compute error Σ i i : h i t ( a i i ) b i i D i t a i i
Choose α i t = 1 2 log 1 ε t i ε t i
Updating all the parameters
D i t + 1 ( i i ) = D i t ( i i ) N t = e α t   i f   i n s t a n c e   i i   i s   c l a s s i f i e d   c o r r e c t l y e α t   i f   i n s t a n c e   i i   i s   n o t   c l a s s i f i e d   c o r r e c t l y
where N t denotes the normalization factors and it is presumed as i i = 1 m i D i t + 1 = 1
Output the hypothesis H i ( a i ) = s i g n t i 1 T α i t h i t a i i
(A) RF Implementation:
One of the most used ML algorithms is RF. A single outcome can be obtained by the hybridization of different decision trees in this concept of RF. Both classification issues and regression problems can be tackled well with the help of RF. The performance of any model can be enhanced by merging different classifiers so that more complex problems can be solved efficiently. The prediction of the correct output cannot be performed by decision trees alone, as generally a high number of trees are required to build the model, resulting in high computational complexity. The pseudocode for RF is expressed in Algorithm 2.
Algorithm 2: RF Classification
Input: Training samples
Weight initialization by random values; W i i = V a l u e r a n d o m
Compute the sum of inputs A i 1 , A i 2 , A i 3 , , A i n : S h i : = i = 1 n A i i . W i i + c i
Multiply with weights W i 1 , W i 2 , W i 3 , , W i n for the hidden layer node
Project the hidden output for hidden layer node
The input for the output layer nodes is expressed as
S o u t : = i = 1 n A i i . W i i + c i
Project the output at output nodes
The difference between the predicted output and actual output is computed so that the error rate is obtained
The error rate is detected from hidden layer to output layer nodes
The updating of the network weights is performed
The final output is the difference between the predicted output and actual output
(B) Hybrid AdaBoost–RF implementation:
The weak learners can be converted to strong learners easily with the help of AdaBoost and RF algorithms. To enhance accuracy and to improve the stability, RF is hybridized with AdaBoost so that the high weight instances are easily chosen. The training of RF algorithm is easier, but once the training process is complete, it becomes slow; therefore, to enhance speed, the AdaBoost algorithm is hybridized with RF. By hybridizing AdaBoost and RF, training time can be reduced easily so that better accuracy can be obtained. The RF estimation is widely used, leading to an accurate classification by largely minimizing the losses. Both the weak points and strong points of the hybrid algorithm are examined so that the obtained error rate is much lower. The algorithm for the Hybrid AdaBoost–RF algorithm is expressed in Algorithm 3 as follows:
Algorithm 3: Hybrid AdaBoost–RF classification
Given a i 1 , b i 1 a i z , b i z , a i i A i , b i i B i = 1 , 1
Assign D i i ( i i ) = 1 z i
For t i = 1 , , T i
Train the weak classifiers by using distribution D i
Weak hypothesis is fetched h i i : A i 1 , 1
Compute error i i : h i t a i i b i i D i t a i i
Choose α i t = 1 2 log 1 ε t i ε t i
Increment the hypothesis if required
Compute the sum of the inputs as S h i : = i = 1 n A i i . W i i + c i
Multiply A i 1 , A i 2 , A i 3 , , A i n with W i 1 , W i 2 , W i 3 , , W i n for the hidden layer nodes
Compute the hidden output for hidden layer node
Input for the output layer node is given as S o u t : = i = 1 n A i i . W i i + c i
Compute the distribution as
D i t + 1 ( i i ) = S o u t ( i i ) N t = e α t   i f   i n s t a n c e   i i   i s   c l a s s i f i e d   a c c u r a t e l y e α t   i f   i n s t a n c e   i i   i s   n o t   c l a s s i f i e d   a c c u r a t e l y
where the normalization factor N t is expressed as i i = 1 z i D i t + 1 = 1
Output the final hypothesis H i ( a i ) = s i g n t i 1 T α i t h i t a i i

3. Proposed Pragmatic Model 2

The second model uses the PCA, KPCA and RFM techniques for feature extraction, followed by the usage of AOA and IG for feature selection and two hybrid classifiers such as the Hybrid ARIMA–AdaBoost classifier and the TW-HASVM classifier.

3.1. Feature Extraction Using PCA, KPCA and RFM Techniques

3.1.1. PCA and KPCA

To handle nonlinear relationships efficiently, the traditional PCA can be extended with the aid of kernels. It is widely used in multiple applications such as remote sensing, Natural Language Processing (NLP), image processing etc. As raw data has a lot of redundant features, KPCA is utilized to largely remove them [48]. Initially, the ideal kernel function must be selected. Various kernels, such as Gaussian, Polynomial, Radial Basis, sigmoid, etc., are available, and in this work, a Gaussian kernel is chosen. Then, the kernel matrix is computed so that the pairwise similarity is specified correctly between the data points with the help of kernel function. The kernel matrix is a simple n × n symmetric matrix. If the total number of data points is represented as n , the kernel function is represented as follows:
K p i , p j = p i , p j
Then, eigendecomposition is implemented so that the eigenvectors and eigenvalues are obtained. In the high-dimensional space, the principal components are the simple eigenvectors. Eigenvalues help to assess the total variance of every component. The eigenvector E v is indicated as follows:
E v = j = 1 n β j μ p j
where the training samples are represented by p j , and the eigenvectors are represented by E v . The K eigenvectors related to the K eigenvalues are chosen. The vital principal components are specified by the eigenvectors so that a majority of variance is captured in the entire data. For extracting the entire features, the chosen principal components are the original projection of the entire data. Thus, the essential structure of the data is preserved and, at the same time, the overall dimension of the data is greatly mitigated. The derivation of the high dimensional feature space is expressed as follows:
e v l . μ p i = j = 1 n β j l K p j , p i
where the principal components are expressed by β j ; the feature vectors are represented as e v l , and the mapping coefficient is denoted by μ . The inherent patterns are captured, and the extracted features are effectively obtained by implementing the KPCA model. From these extracted features, the most suitable features can be selected and fed for classification to obtain a high classification accuracy.

3.1.2. Random Feature Mapping Technique

A dataset of various dimensions must be processed, and so RFM is utilized in this work [49]. On the input samples, the linear transformation is incorporated so that feature nodes are formed. The stacking of feature nodes is performed so that a feature mapping layer is formed which aids in feature extraction. A random feature mapping technique usually has an input layer and a feature mapping layer. The input data features equal the total number of neurons in the input layer. If ‘K’ number of features are present in the input data, the vector g i = g i 1 , g i 2 , , g i k is utilized to specify the i t h input data. The matrix G = g 1 , g 2 , , g i , , g n T helps to identify the n   K -dimensional input vectors from the input data. The batch processing of the input data G is performed and then for transformation purposes, it is progressed to the feature mapping context zone. The normalization of the original input data is performed so that it has a variety of properties. Assuming ‘q’ group nodes are present in the feature nodes of the feature mapping, they are larger and represented as Z i   ( i = 1 , 2 , , q ) , which has a collection of ‘n’ numbers of neurons. Once the feature mapping of Z i is performed successfully, G is expressed as follows:
Z i = φ i G W e + β i
where the activation function is activated by φ i ; in this work, the ReLU function is used. The weight matrix is indicated by W i , and the bias matrix is specified by β i . A random generation of the initial values can be performed successfully in the weight matrix and the bias matrix, and it is represented as follows:
W i M × n
β i Q × n
The random features must be fine-tuned slightly into a new set of sparse features, and this is enabled by sparse autoencoders. For the generated new feature nodes, the linear correlation is reduced drastically. From the input data, the sparse features must be extracted, and the lasso technique is implemented so that the optimization problem can be solved easily. To fine-tune W i and β i , a sparse matrix is assigned, and the objective function for optimization purposes is assigned as follows:
arg min W i : Z W i G 2 2 + γ W i 1
A lot of restrictions can be imposed on the weight matrix by γ . Thus, the important features are explored by sparse autoencoders, and the randomness can be easily overcome. For the data, the distribution specifications are fully reflected and identified by the generated feature subsets so that the generalization aspects of the model are largely improved.

3.2. Feature Selection Using Information Gain (IG)

A famous technique to evaluate the significance of features with respect to target variables is IG [50]. It has multiple advantages, including independence, robustness, speed and simplicity. The uncertainty is tried and removed by the information entropy and when the categorization occurs, each category is indicated by some random variable. Assuming the category to be P = p 1 , p 2 , , p i , then for the classification task, the information entropy is mathematically expressed as follows:
H ( P ) = i = 1 n Pr p i h 2 1 Pr p i = i = 1 n Pr p i h 2 Pr p i
where the number of classes is represented by n , and the probability is represented by Pr . For each feature, the IG needs to be computed. Assuming the feature as F and its respective values as f 1 , f 2 , , f k , for a class P , the conditional entropy is expressed as follows:
H P F = k = 1 l Pr f k i = 1 n Pr p i f k log 2 Pr p i f k
where the average uncertainty of class p is expressed by H P F . When a feature F is fixed, the conditional probability of class p is indicated by Pr p i f i . For a feature F , the IG is indicated as the difference between a priori entropy H P and conditional entropy H P F and is expressed as follows:
I G ( F ) = H ( P ) H P F

3.3. Feature Selection Using AOA

For certain parameters, optimal values can be easily determined by the process of optimization. The best solutions cannot be assessed exactly with the help of mathematical programming methods, as the problem might display unrestricted specifications sometimes. Therefore, to trace the unknown search spaces and to seek the local optima, optimization techniques are usually proposed, and over the past few decades, meta-heuristic algorithms have gained considerable attention. Exploration and exploitation are the two important stages that all the meta-heuristic algorithms come across and a good balance should be framed between these two aspects always. The hunting behavior of the Aquila bird is mimicked by this nature-inspired metaheuristic algorithm [51]. Aquila has excellent hunting abilities and it targets animals such as hares, squirrels, rats, etc. These birds have very large territories, and they build their nest in very high places so that they can raise their young ones peacefully. When present in an isolated environment, a male Aquila can become more aggressive and can attack even an adult deer successfully. All kinds of rodents are generally a staple part of their diet. Four specific hunting techniques are employed by Aquila and depending on the situation, they can adopt and change their hunting methodology. The initial technique involves flying high in sky with a faster movement to attack the flying birds. The second technique involves hunting the target by flying at a low-level descent or altitude. The third technique involves the targeted attack of the prey, which are relatively slow-moving such as tortoises and snails. The final techniques involve the grabbing method of attack, where they usually pull the target out from hiding and then kill it. Based on these techniques, the AOA was developed. As Aquila is a relatively skillful hunter, the modelling of its different strategies can be expressed as follows:
Technique 1: This phase could be called as expanded exploration phase. At the start, the information about its target can be identified by the Aquila, and then an exhaustive search strategy is used to accurately locate the position and is represented as follows:
Y i ( t + 1 ) = Y b e s t ( t ) × 1 t T + Y p ( t ) Y b e s t ( t ) r
The individual positions i is expressed by Y i ( t + 1 ) in the next ( t + 1 ) iteration. In the current iteration, the best optimal location is identified by Y b e s t ( t ) . In the current iteration, for every individual, the average position is projected by Y p ( t ) and is expressed as follows:
Y p ( t ) = 1 N i = 1 N Y i ( t )
At an iteration t , the position of the i t h individual is expressed by Y i ( t ) . The entire swarm population size is expressed by N . Here, a random number is assigned, which is a Gaussian distribution and is within the range [0, 1]. The highest iteration number is depicted by T .
Technique 2: This phase could be called a focused exploration phase. When this phase occurs, the prey is located by aquila, and so it circles around the target and prepares to land to attack the prey, and it can be mathematically expressed as follows:
Y i ( t + 1 ) = Y b e s t ( t ) × L e v y ( D ) + Y R ( t ) + z y r
Here, r is expressed as the random number, the dimensionality is indicated by D and L e v y ( D ) helps to compute the Levy flights and is expressed as follows:
L e v y ( D ) = c × μ × σ v 1 β
where a constant parameter is represented by c and has a value of 0.1 in our experiment. The random numbers are again initiated to μ and v , and are assigned within the range of [0, 1]. The computation of σ is performed as follows:
σ = γ 1 + α × sin π α 2 γ 1 + α × α × 2 α 1 2
α is, again, a constant parameter and is set as 0.5 in our experiment, while the gamma factor is represented by γ .
Y R ( t ) indicates a randomly chosen candidate present at the current iteration. With the help of variables z and y , the indication of the spiraling shape is performed and is computed as follows:
z = r × cos θ
y = r × sin θ
r = r 1 × U × D 1
θ = w × D 1 + θ 1
θ 1 = 3 π 2
r 1 denotes a constant number within the range of [1, 10]. D 1 denotes an integer parameter range of [1, 10] again in the experiment. A constant parameter is again denoted by w and is fixed with a value of 0.01 in our experiment.
Technique 3: This phase could be called the expanded exploitation phase. Sometimes during the exploitation phase, the target cannot be determined by the Aquila, and so complete reinitialization of the process must be performed. Then, the position must be updated with the help of the following equation:
Y i ( t + 1 ) = λ × Y b e s t ( t ) Y p ( t ) + σ × U B L B × r + L B
where the lower-bound domain is indicated by L B , and the upper-bound domain is indicated by U B . The two fixed numbers or constants can be considered as λ and σ here.
Technique 4: This phase could be called the forwarded exploitation phase. When the Aquila targets the prey by approaching it, it utilizes the focused exploitation technique by means of updating the exact position as follows:
Y i ( t + 1 ) = Q F × Y b e s t ( t ) G 1 × Y i ( t ) × r G 2 × L e v y D + r × G 1
where the quality factor, denoted by Q F , helps to stabilize the search process and is calculated as follows:
Q F = t 2 × r 1 ( 1 T ) 2
Once the position is updated, and the search process is stabilized and balanced, the best features are obtained and it is fed to the machine learning classifiers. As far as the fitness function is concerned here, it depends on FSR and classification errors such as the AOA.

3.4. Classification Using Two Hybrid Models

3.4.1. ARIMA–AdaBoost Hybrid Model

With the help of the AdaBoost regressor, the regression inaccuracies can be easily tackled by AdaBoost techniques. The ARIMA regression model is hybrid with AdaBoost model and the classification is carried out [52]. The training instances can be reweighed successfully with the help of AdaBoost regressor so that weak regressors can be produced which could have a low variance error and high bias error. To obtain the final prediction, the hybridization of every weak regressor is performed so that a model can be obtained, which would have a low bias error and low variance.
Assuming a sample set such as V = a i , b i i = 1 , 2 , , N is present initially.
At the t t h iteration, the samples have a specific weight distribution and is specified by D t ( i ) , t = 1 , 2 , , T .
The weight distribution is initialized during the initial iteration, i.e., when t = 1 , so that D 1 ( i ) = 1 N .
The forecasting error is computed as e t ( i ) = f t ( a i ) , b i , e t ( i ) 0 , 1 .
The expected output of f t ( a i ) is initially computed based on the weight distribution.
The proportional error computation is performed as ε t = i = 1 N D t ( i ) e t ( i ) .
The connection weights are traced out and are expressed as w t = 1 2 log 1 β t .
The weighing schemes are adapted accordingly and are expressed as
D t + 1 , ( i ) = D t , ( i ) × β t ε t / W t + 1 and W t + 1 = i = 1 N D t + 1 , ( i ) .
After a particular number of T iterations, the end predictions can be obtained as F ( a ) = t = 1 T w t f t ( a ) .
A famous time-series forecasting method is the ARIMA model. Here, the ARIMA model is projected as k , d , l , where the autoregressive term is indicated by k , the moving average term is indicated by l , and the degree of difference is denoted by d , which helps to assess the stability of data. Structural models that are relatively complicated can be easily tackled by the ARIMA model. The future value of a variable is a simple linear combination of its errors, combined with its historical values and is expressed as
P t = ϕ 0 + ϕ 1 Q t 1 + ϕ 2 Q t 2 + + ϕ k Q t k + ε t θ 1 ε t 2 , θ l ε t l
where the actual value is represented by P t .
The random error at t is indicated by ε t , and the coefficients are indicated by ϕ i and ϕ j . The autoregressive integer is specified by k , and the moving average integer is specified by l . The ARIMA k , d , l has the important constituents of the model such as autoregression, integrated model and average motion. The time-series models between AR k and MA l have their own set of specifications. A simple AR model of order k , projected as AR k , is expressed as a linear process projected as follows:
Q t = c + i = 1 K ϕ i Q t i + ε t
where the constant is denoted as c . The coefficient of Q t i is represented as ϕ i , and the white noise is represented as ε t . The relationship between Q t values is the present period and the Q values in the past k periods are expressed by the AR k model. For an order l , the moving average model is specified by the notation MA l and is represented as
p t = μ + i = 0 l θ i ε t i
where the expected value of p t is denoted by μ . The Gaussian white noise is represented by ε t and the weight prescribed to the time period is expressed by θ 0 = 1 . The integer values help to replace the ARIMA k , d , l model with the aid of three arguments of k , d , l . With changes in k , d and l , the generation of countless models can be achieved. The time attribute of the present and past data is used mainly for prediction purposes. To predict the future values successfully, a lot of patterns can be changed gradually during the process. With the help of an averaging technique, the classification of time series can be performed using this exponential smoothing technique. The null hypothesis can be investigated well with the help of the time-series data to analyze its stationary property.
For regression and classification tasks, AdaBoost can be used, as it is a versatile framework. Based on a specific number of rules, many weak estimators are used in this framework. During the training process of the AdaBoost classifier, decision trees and regression trees are used as a single learning technique. The important steps for the ARIMA–AdaBoost classification are as follows:
The collection of data and assimilation, performed together so that the train and test sets can be decomposed easily;
The implementation of the stationarity test;
Regression model training using AdaBoost;
The validation of the generated model using a test dataset.
This model can be applied to other datasets and real-world issues as well. The simplified flowchart of the classification model of the ARIMA–AdaBoost hybrid classifier is shown in Figure 2 as follows:

3.4.2. Time-Weighted Hybrid AdaBoost–Support Vector Machine (TW-HASVM) Model

Based on time, the data batches are given more importance, and so this approach is termed the time-weighting approach [53]. The formulation of time-weighing function is expressed as follows:
T W t = exp η t , t = 0 , 1 , 2 , , H 1
The batch number of the data is represented as t , and the highest number of data batch is represented as H , respectively. The value of η is defined from 0 to 0.5 in this work. The sample weighting model is improved by the TW-HASVM classification model. The sample weighting functions are greatly improvised and hybridized by using the average sample weights and are represented as follows:
T W 1 k = 1 n , k = 1 , 2 , , N
T W u + 1 k = T W u k exp γ u l u k exp γ t k l u k , k = 1 , 2 , , n ; u = 1 , 2 , , u 1
l u k = 1 i f b k = a k 1 i f b k a k
The highest number of iterations are indicated by u , and the total number of samples is represented by n . The weight factor is represented by γ u . T W u + 1 k represents the weight of the u + 1 t h iteration. The term l u k represents the control parameter of the sample weight direction. For the misclassified samples, the value is expressed as −1, and for correctly classified samples, the value is expressed as +1. For the classifier, the voting weight is expressed as follows:
γ u = 0.5 ln 1 ε u ε u
This time weighted concept is incorporated with the SVM–AdaBoost hybrid classifier as follows. SVM always tries to trace the optimal hyperplane so that the data points can be separated well in various classes [54]. With the aid of kernel functions, the data is transferred to high dimensional spaces by the SVM. In AdaBoost, the weak classifiers are trained iteratively, and the data points are constantly adjusted so that the performance of the weak classifier is improved. Higher weights are assigned to the misclassified instances so that the preceding weak classifiers have a higher focus on other weak training samples. By hybridization of these two techniques, the design of the TW-HASVM classifier is performed so that the data imbalance problem can be solved well. In this hybrid model, multiple SVM classifiers are trained by the AdaBoost algorithm on various subsets. To the specific data points, the time weights are introduced so that the higher weights are received by the hybrid form. The distribution of the data becomes relatively sensitive in the classification process, and the assignment of time weights largely stabilizes the process. The hybrid weighted model of the SVM classifiers, when trained with the AdaBoost algorithm, yields the hybrid TW-HASVM classifier. Handling the imbalance data becomes much easier with his hybrid classifier rather than the traditional one, as a versatile performance could be maintained in the majority class.
The training of the TW-HASVM hybrid classifier is performed with the help of the chosen features. Multiple SVM classifiers are trained with the help of AdaBoost algorithm on various subsets of the data. The assignment of the various time weights is performed on the weight points. Weights which are higher are received by the recent instances, and weights which are lower are received by the later instances. In the data distribution, a lot of recent changes can be captured by this hybrid classifier. In the training dataset, for each data point, the time weights w i ( 1 ) are initialized. The weight assignment process is carried out successfully and is represented as follows:
w i ( 1 ) = 1 n f o r i = 1 , 2 , , n
In the training dataset, every data point i is prioritized with an initial weight of 1 n , where the data point number is indicated by n . For the AdaBoost algorithm, the number of iterations must be determined. This helps to assess the training of weak SVM classifiers. For all the iterations t = 1 , 2 , , T , the time weights are normalized as follows:
w i ( t ) = w i ( t ) / N t , where the normalization factor is represented as N t and is indicated as follows:
N t = s u m w i ( t ) f o r i = 1 , 2 , , n
Depending on the normalized time weights, the training data subset is sampled so that the weak SVM classifiers can be trained well in every iteration. With the help of the kernel function, the SVM classifier is trained C t , and in this work, a polynomial kernel is used. On the entire training dataset, the performance of the weak classifier is assessed, and the classification error rate is computed. When computing the classification error rate, the time weights are taken into consideration and are represented as follows:
e t = s u m w i ( t ) I b i C t a i / s u m w i ( t )
where the indicator function is represented by I , and a misclassification is represented by b i C t ( a i ) . The classifier weight α t is calculated as a projection of error rate e t . The computation of weight is as follows:
α t = 0.5 log 1 e t / e t
To the classifiers, the assignment of the classifier weight is performed so that a low error rate is obtained. The updating of time weights of instances is performed in the AdaBoost algorithm at every iteration so that the performance of the weak classifiers is upgraded well. The updating of time weights is expressed as follows:
w i ( t + 1 ) = w i ( t ) exp α t I b i C t ( a t )   f o r           i = 1 , 2 , , n
The data point i has a particular time weight at a specific iteration t and is represented as w i ( t ) . The data point i is now updated with a new time weight and ready for the upcoming iteration t + 1 , being represented as w i ( t + 1 ) . For the upcoming weak classifiers, the training data is sampled by the updated weight. At a specific iteration t , based on a weak classifier C t , a particular weight is assigned to improve its performance and is expressed by α t . I b i C t ( a t ) represents the indicator function to analyze the difference between time class label b i and the predicted class label C t a t . The indicator function expresses the output as 0 if the prediction is correct, and the indicator function expresses the output as 1 if the prediction is incorrect. The following expression is utilized to compute the product of the indicator function I and the weight of the classifier α t .
exp α t I b i C t a t
There will be a large exponential term if there is an incorrect prediction, so that in the next iteration the time weight will be increased. The exponential term equals 1 if there is a correct prediction, so that in the next iteration the time weight remains steady.
Hybridization classifiers.
Due to the weighted amalgamation of the independent SVM classifiers trained with the aid of the AdaBoost algorithm, the TW-HASVM classifier is obtained. During the training process, the performance measure obtained is assessed, and then the determination of every weight for the SVM classifier is performed. The higher weights received by the best classifier are assessed well in this model. As a weighted hybridization of all the weak SVM classifiers along with their weights, the ultimate TW-HASVM classifier is obtained after finishing all the iterations and is represented as follows:
F ( a ) = s u m α t C t ( a )   f o r t = 1 , 2 , , T
Classification process by TW-HASVM:
Once the training of the TW-HASVM is performed, it is utilized to assess and classify the features. The output of every SVM classifier is computed when the features are presented to it, and it is performed in the weighted combination. With the help of a weighted vote, the assessment of a final class label is performed. The final prediction is nothing but the class which has obtained the largest combined weight in the classification model. Assuming the data points as a , the TW-HASVM hybrid classifier aims to classify and predict its class label. The output for the data point a is computed for all the weak SVM classifiers C t . The prediction of the class label by the weak classifier C t is the respective output and is represented as follows:
h t ( a ) = C t ( a )
For every class label, the weighted output sum is calculated with the help of classifier weights α t and is represented as follows:
S c l ( a ) = s u m α t h t ( a ) f o r c l 1 , 2 , , c l
where the number of classes is expressed by c l . Depending on the number of classes, the weighted sum for every class can be computed. For the data point a , the final class label is analyzed by means of choosing the class which has the highest weighted sum. The hybrid classifier’s prediction is this class label.
b p r e d ( a ) = arg max c l S c l a
where each class is represented by the label c l . The weighted output sum is indicated by S c l ( a ) and is computed as follows:
S c l ( a ) = s u m α t h t ( a ) f o r c l 1 , 2 , , c l
The number of classes is represented by c l . The term h t ( a ) represents the output of the weak SVM classifier C t . At an iteration t and for the data point a , the classifier weight is expressed as α t . The overall simplified illustration of the TW-HASVM classification model is expressed in Figure 3 as follows.

4. Results and Discussion

A publicly available BCG database was used for the analysis in this work [29]. The explanation of the dataset is provided in detail as follows. For the acquisition of BCG signals, a standard BCG signal acquisition termed RS-611 was used in this study [29,55]. The RS-611 comprises three important components, such as a sensitive mattress which can track the micro-movements, an analog-to-digital converter (ADC), and a personal computer attached to it. There are two main hydraulic pressure sensors attached to the sensitive mattress, where the upper part of the sensor is attached to the chest and the lower part of sensor is attached to the leg. Both components are placed in a parallel manner so that the BCG signal can be captured easily from the subjects. The original pressure caused due to the various cardiac activities is recorded, amplified and then, with the help of the ADC device, it is converted to a 16-bit resolution digital signal which had a sample rate of 100 Hz. The main advantage of using RS-611 in the experimentation study [29,55] is because it does not have any complexity in terms of handling and maintenance.
A total of 175 people were initially considered as subjects in this study. Before recruiting them for this study, their informed consent was obtained for this study, and it was performed according to the standard medical experimental norms of the Helsinki Declaration standards. To obtain the basic physiological index, preliminary routine examinations were conducted. With the help of a mercury sphygmomanometer, the measurement of blood pressure was performed in the left arm consecutively with short intervals. Based on the standard set by the American Society of Hypertension referred to in [56], the measurement was performed when the subjects were seated in a relaxed position. The recording of the blood pressure was performed in different sessions in a time span of three weeks, and then the values obtained were arranged so that the final systolic and diastolic values of the blood pressure were obtained. If the subjects had a blood pressure reading greater than 140/90 mm Hg, it was categorized as hypertensive; otherwise, it was categorized as normal. The subjects with the following criteria were excluded from the experimentation procedure: subjects using anti-hypertensive medication when the blood pressure was measured, subjects with a clinical history of myocardial infarction, subjects with a Body Mass Index (BMI) > 30, subjects with diabetes mellitus, subjects with smoking and chronic alcoholism and subjects with atrial fibrillation conditions. Once these exclusions were incorporated, a total of 68 hypertensive classes and 76 normal classes were shortlisted. Again, after further scrutiny, some subjects were eliminated. One day before the experiment was conducted, the subjects were asked to sleep well by means of adjusting the room temperature, light intensity and noise levels so that they could sleep peacefully for a period of about 7 to 8 h. The next morning, these subjects were asked some questions about their sleep pattern. The BCG recordings of the subjects who reported that they did not sleep well or had a disturbed sleep the night before the experimentation day were deemed unfit for research and were discarded. A manual investigation was conducted into the BCG recordings, and some recordings, which showed drastic fluctuations, were eliminated. The final BCG dataset was obtained from a total of 128 participants, among whom there were 67 normal subjects and 61 subjects with hypertension. At a frequency of 100 Hz, the BCG signals were sampled in this dataset. The initial segmentation of the BCG signals was performed for a period of 5 min, and in this work the basic preprocessing step of utilizing ICA to the BCG signals was implemented before proceeding with the proposed pragmatic models. The dataset can be accessed at the following URL: https://figshare.com/articles/dataset/Unobtrusive_Mattressbased_Identification_of_Hypertension_by_Integrating_Classification_and_Association_Rule_Mining/7594433. The details of the dataset are provided in Table 1 and Table 2 below.
Using the MATLAB R2022a platform, the experiment was conducted on a system which had a 64 bit operating system, 16 GB of RAM and a 1.60 GHz processor. A 10-fold cross-validation scheme was utilized thoroughly for all the hybrid classifiers used in this work. For the KNN algorithm used in this work, the value of k was chosen to be 10 in our experiment. For the KPCA algorithm used in this work, the kernel type was set to linear, and the total number of components used was four. For the TW-HASVM classifier, the current data batch number was set to 1, and the weights were adjusted accordingly. For the BTSA, the hyperparameter settings were set as follows: the search agents were set to 100, and the number of generations was set to 1000. The minimum parameters and maximum parameters were assigned in the range of 1 to 20 for this algorithm. The exploitation fixed parameters in AOA were set to 0.5, respectively. The total number of iterations used in the AdaBoost module was set to 1000, and the kernel type used when engaging the SVM classifier was a polynomial kernel. For comparison purposes, other prominent feature selection techniques such as the chi-square test, Pearson’s correlation coefficient (PCC), Particle Swarm Optimization (PSO) and Ant Colony Optimization (ACO) were considered. In addition, for the comparison of machine learning performance, the standard machine learning classifiers such as RF, Decision Trees (DTs), Naïve Bayesian Classifier (NBC) and the standard AdaBoost were considered apart from the concentrated four hybrid classifiers in this work. The main performance metric analyzed was classification accuracy and was represented in terms of True Positive ( T P ) , True Negative T N , False Positive F P , False Negative ( F N ) and computed as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N
On examining Table 3, a high classification accuracy of 96.48% is observed when the MODWT feature extraction technique is combined with the BTSA feature selection technique and classified with the Hybrid AdaBoost–MULDA classifier. On careful analysis of Table 4, a high classification accuracy of 93.87% is observed when the MODWT feature extraction technique is combined with the IG feature selection technique and classified with the Hybrid AdaBoost–MULDA classifier. When Table 5 is examined, a high classification accuracy of 87.36% is observed when the MODWT feature extraction technique is combined with the chi-squared feature selection technique and classified with the Hybrid AdaBoost–MULDA classifier. If Table 6 is studied, a high classification accuracy of 89.9% is observed when the EWT feature extraction technique is combined with the PCC feature selection technique and classified with the Hybrid AdaBoost–MULDA classifier. If Table 7 is carefully studied, then, a high classification accuracy of 91.23% is observed when the MODWT feature extraction technique is combined with the PSO feature selection technique and classified with the Hybrid AdaBoost–MULDA classifier. On examining Table 8, a high classification accuracy of 89.67% is observed when the MODWT feature extraction technique is combined with the ACO feature selection technique and classified with the Hybrid AdaBoost–MULDA classifier. On careful analysis of Table 9, a high classification accuracy of 96.89% is observed when the KPCA feature extraction technique is combined with the AOA feature selection technique and classified with the Hybrid ARIMA–AdaBoost classifier. When Table 10 is examined, a high classification accuracy of 92.37% is observed when the KPCA feature extraction technique is combined with the IG feature selection technique and classified with the TW-HASVM classifier. On examining Table 11, a high classification accuracy of 89.30% is observed when the RFM feature extraction technique is combined with the chi-square feature selection technique and classified with the TW-HASVM classifier. If Table 12 is studied, a high classification accuracy of 88.88% is observed when the KPCA feature extraction technique is combined with the PCC feature selection technique and classified with the TW-HASVM classifier. On examining Table 13, a high classification accuracy of 91.48% is observed when the RFM feature extraction technique is combined with the PSO feature selection technique and classified with the Hybrid ARIMA–AdaBoost classifier. On careful analysis of Table 14, a high classification accuracy of 89.84% is observed when RFM feature extraction technique is combined with the ACO feature selection technique and classified with the Hybrid ARIMA–AdaBoost classifier.
Figure 4, Figure 5, Figure 6 and Figure 7 are plotted here, as they yield comparatively higher classification accuracies than the remaining methods. The overall higher classification accuracy is yielded by the AOA feature selection technique and the Hybrid ARIMA–AdaBoost classification model. The second highest classification accuracy is achieved by the BTSA feature selection technique and the Hybrid AdaBoost–MULDA classification model. The third highest classification accuracy is achieved by the IG feature selection technique employed with the MODWT feature extraction model and classified with the Hybrid AdaBoost–MULDA classification model. The fourth highest classification accuracy is achieved by the IG feature selection technique combined with the MODWT feature extraction scheme and the TW-HASVM classification model.
When the standard Cohen’s Kappa coefficient test was applied to the extracted features, the values indicated good agreement, ranging from 0.5 to 1, respectively. A good variation was found for the feature values when Friedman test analysis was conducted, and so it was deemed fit for the classification. The extracted and selected features passed the standard Kruskal–Wallis test, with statistically significant results ( p < 0.01 ), indicating that the features were correctly extracted and retained the original information; therefore, the features can be passed through the hybrid classifiers. When a two-sided Wilcoxon test was applied, all features showed statistically significant effects. As far as computational complexity was concerned, all the models exhibited low complexity and were therefore analyzed as O n 3 log n .

4.1. Comparison with Previous Works

For a clearer comparison, the results of the current study are compared with previous work and tabulated in Table 15.
In this work, an exhaustive analysis of BCG signals was conducted for the classification of hypertension using multiple feature extraction techniques, multiple feature selection schemes and multiple machine learning classifiers, including a few hybrid machine learning models. The results obtained yielded comparatively good results when compared to a majority of previous studies. Though some previous works have obtained a slightly higher classification accuracy than the present work, it must be noted that the current work has proposed plenty of methods, and a comprehensive analysis was conducted, which many previous studies failed to perform, focusing only on a single proposed method. Considering this aspect and the workflow of the paper, this work is considered good in terms of an exhaustive analysis performed on the BCG signals for the classification of hypertension.
The overall computational time for the best performing models was computed and is shown in Table 16 below:
On analysis of Table 16, it is evident that a low computational time of 5.112 s was produced by the model which employs the KPCA feature extraction technique, the AOA feature selection scheme and the Hybrid ARIMA–AdaBoost classifier. The next best computational time of 6.012 s was achieved by the model which employs the RFM feature extraction technique, the AOA feature selection scheme and the Hybrid ARMIA–AdaBoost classifier.

4.2. Limitations of This Work

For a research framework like this, no exact physiological rationale can be provided as to why only some feature extraction/selection techniques along with certain machine learning classifiers were chosen in the experiment. With hundreds of well-known feature extraction methods and thousands of feature selection techniques available in the literature, it is very difficult to justify why only a particular subset of methods was considered for this research. During the analysis, when starting to build the experiment, various factors were considered in the research, as follows: the combination should not have been employed by other researchers in the past, the workflow should be novel to some extent and it should be the first of its kind to have been applied to the BCG dataset so that the results can be analyzed well. Based on these factors, the experimentation was performed by choosing some random algorithms and evaluating their performance on the BCG dataset.
The two pragmatic models are not two distinct models but rather two large baskets of arbitrarily grouped techniques. In this research, the flow is presented as a mechanical pipeline, and the major insight here is very simple: just an ordinary flow of feature extraction and feature selection, followed by classification. The results can be attributed to some factors, as they lie in the effectiveness of the algorithms used and it is always interesting to observe how a particular set of feature extraction techniques, when selected with conventional feature selection techniques and metaheuristic optimizers, followed by classification with traditional and hybrid classifiers, perform for a particular dataset to yield a specific set of results. The reason for the classification accuracy may be attributed to the intrinsic properties of the wonderful algorithms used, as these algorithms work independently and depend on the hyperparameters set by the authors. Based on the natural design of the algorithm and the hyperparameters set by the authors, the results can be obtained.
Moreover, another limitation is that the discrimination improvement obtained by adding BCG-derived features to a model including all clinical characteristics alone, compared with the clinical model by itself, was not thoroughly investigated in this work, and this will be considered as possible future work. The addition of BCG-derived parameters to patient-factor–based models improved the prediction of hypertension, but the incremental predictive performance, with discrimination analysis and the benefit of BCG parameters for identifying individuals with hypertension, was not analyzed in detail in this work, and it could be considered as possible future work. In this work, only classification accuracy is given a high priority, and in the future, more biomedical metrics are planned to be computed, such as sensitivity, specificity, precision, F1-score, geometric mean, etc.

5. Conclusions and Future Works

In the tele-healthcare paradigm, an important achievement has been made in the field of Internet of Medical Things and medical signal processing. When the blood circulates in the body, a ballistic force is generated, and it can be assessed as a BCG signal. Various sensors were utilized in the past for extracting the BCG signal, as it is more flexible and provides greater convenience to use. In this work, two interesting pragmatic models using efficient feature extraction, feature selection and hybrid machine learning models are used for the analysis and classification of BCG signals for the assessment of hypertension. The best results are shown when the KPCA feature extraction technique is used with the AOA feature selection technique and classified with the Hybrid ARIMA–AdaBoost classifier, reporting a classification accuracy of 96.89%. The second-best classification accuracy of 96.48% is obtained when the MODWT feature extraction technique is used with the BTSA feature selection technique and classified with the Hybrid AdaBoost–MULDA classifier. The third-best classifier classification accuracy of 96.23% is obtained when the PCA feature extraction technique is used with the AOA feature selection technique and classified with the Hybrid ARIMA–AdaBoost classifier. Future work plans to use a variety of other efficient feature extraction techniques, other novel metaheuristic and swarm-based algorithms for feature selection and more efficient hybridization of machine learning classifiers so that the intrinsic properties of all the classifiers can be utilized well, and hence the classification accuracy can be improved to a much higher extent. Future work also plans to incorporate the concept of explainable Artificial Intelligence (EAI), along with the pragmatic models to test its efficacy in terms of classification accuracy. In addition, the work is planned to be implemented on real-time test beds, and it can be applied to telemedicine applications as well in future work.

Author Contributions

S.K.P.—conceptualization, methodology, software, validation, formal analysis, investigation, resources, data curation; D.-O.W.—validation, formal analysis, writing—review and editing, visualization, supervision, project administration, and funding acquisition. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2022R1A5A8019303), and partly supported by the Hallym University Research Fund, 2024 (HRF-202411-007).

Data Availability Statement

Conflicts of Interest

The authors declare no competing interest.

References

  1. Starr, I.; Schroeder, H.A. Ballistocardiogram. II. Normal standards, abnormalities commonly found in diseases of the heart and circulation, and their significance. J. Clin. Investig. 1940, 19, 437–450. [Google Scholar] [CrossRef]
  2. Vogt, E.; MacQuarrie, D.; Neary, J.P. Using ballistocardiography to measure cardiac performance: A brief review of its history and future significance. Clin. Physiol. Funct. Imaging 2012, 32, 415–420. [Google Scholar] [CrossRef] [PubMed]
  3. Scarborough, W.R.; Talbot, S.A.; Braunstein, J.R.; Rappaport, M.B.; Dock, W.; Hamilton, W.; Smith, J.E.; Nickerson, J.L.; Starr, I. Proposals for ballistocardiographic nomenclature and conventions: Revised and extended. Circulation 1956, 14, 435–450. [Google Scholar] [CrossRef] [PubMed]
  4. Ashouri, H.; Hersek, S.; Inan, O.T. Universal pre-ejection period estimation using seismocardiography: Quantifying the effects of sensor placement and regression algorithms. IEEE Sensors J. 2017, 18, 1665–1674. [Google Scholar] [CrossRef]
  5. Javaid, A.Q.; Ashouri, H.; Dorier, A.; Etemadi, M.; Heller, J.A.; Roy, S.; Inan, O.T. Quantifying and reducing motion artifacts in wearable seismocardiogram measurements during walking to assess left ventricular health. IEEE Trans. Biomed. Eng. 2016, 64, 1277–1286. [Google Scholar] [CrossRef] [PubMed]
  6. Hwang, S.H.; Lee, H.J.; Yoon, H.N.; Jung, D.W.; Lee, Y.J.G.; Lee, Y.J.; Jeong, D.U.; Park, K.S. Unconstrained sleep apnea monitoring using polyvinylidene fluoride film-based sensor. IEEE Trans. Biomed. Eng. 2014, 61, 2125–2134. [Google Scholar] [CrossRef]
  7. Wang, F.; Zou, Y.; Tanaka, M.; Matsuda, T.; Chonan, S. Unconstrained cardiorespiratory monitor for premature infants. Int. J. Appl. Electromagn. Mech. 2007, 25, 469–475. [Google Scholar] [CrossRef]
  8. Kortelainen, J.M.; van Gils, M.; Pärkkä, J. Multichannel bed pressure sensor for sleep monitoring. In Proceedings of the 2012 Computing in Cardiology, Krakow, Poland, 9–12 September 2012; pp. 313–316. [Google Scholar]
  9. Katz, Y.; Karasik, R.; Shinar, Z. Contact-free piezo electric sensor used for real-time analysis of inter beat interval series. In Proceedings of the 2016 Computing in Cardiology Conference (CinC), Vancouver, BC, Canada, 11–14 September 2016; pp. 769–772. [Google Scholar]
  10. Junnila, S.; Akhbardeh, A.; Barna, L.C.; Defee, I.; Varri, A. A wireless ballistocardiographic chair. In Proceedings of the 2006 International Conference of the IEEE Engineering in Medicine and Biology Society, New York, NY, USA, 30 August–3 September 2006; pp. 5932–5935. [Google Scholar]
  11. Aubert, X.L.; Brauers, A. Estimation of vital signs in bed from a single unobtrusive mechanical sensor: Algorithms and real-life evaluation. In Proceedings of the 2008 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Vancouver, BC, Canada, 20–25 August 2008; pp. 4744–4747. [Google Scholar]
  12. Brüser, C.; Winter, S.; Leonhardt, S. Robust inter-beat interval estimation in cardiac vibration signals. Physiol. Meas. 2013, 34, 123–138. [Google Scholar] [CrossRef]
  13. Zink, M.D.; Brüser, C.; Stüben, B.O.; Napp, A.; Stöhr, R.; Leonhardt, S.; Marx, N.; Mischke, K.; Schulz, J.B.; Schiefer, J. Unobtrusive nocturnal heartbeat monitoring by a ballistocardiographic sensor in patients with sleep disordered breathing. Sci. Rep. 2017, 7, 13175. [Google Scholar] [CrossRef]
  14. Watanabe, K.; Watanabe, T.; Watanabe, H.; Ando, H.; Ishikawa, T.; Kobayashi, K. Noninvasive measurement of heartbeat, respiration, snoring and body movements of a subject in bed via a pneumatic method. IEEE Trans. Biomed. Eng. 2005, 52, 2100–2107. [Google Scholar] [CrossRef]
  15. Pinheiro, E.; Postolache, O.; Girão, P. Theory and developments in an unobtrusive cardiovascular system representation: Ballistocardiography. Open Biomed. Eng. J. 2010, 4, 201. [Google Scholar] [CrossRef] [PubMed]
  16. Antonio, E.-Z. A Simple Ballistocardiographic System for a Medical Cardiovascular Physiology Course. Adv. Physiol. Educ. 2003, 27, 224–229. [Google Scholar] [CrossRef]
  17. Sadek, I.; Biswas, J. Nonintrusive heart rate measurement using ballistocardiogram signals: A comparative study. Signal Image Video Process. 2018, 13, 475–482. [Google Scholar] [CrossRef]
  18. Suliman, A.; Carlson, C.; Ade, C.J.; Warren, S.; Thompson, D.E. Performance comparison for ballistocardiogram peak detection methods. IEEE Access 2019, 7, 53945–53955. [Google Scholar] [CrossRef]
  19. Paalasmaa, J.; Ranta, M. Detecting heartbeats in the ballistocardiogram with clustering. In Proceedings of the ICML/UAI/COLT 2008 Workshop on Machine Learning for Health-Care Applications, Helsinki, Finland, 5–9 July 2008; Volume 9. [Google Scholar]
  20. Pinheiro, E.; Postolache, O.; Girão, P. Study on ballistocardiogram acquisition in a moving wheelchair with embedded sensors. Metrol. Meas. Syst. 2012, 19, 739–750. [Google Scholar] [CrossRef]
  21. Choe, S.T.; Cho, W.D. Simplified real-time heartbeat detection in ballistocardiography using a dispersion-maximum method. Biomed. Res. 2017, 28, 3974–3985. [Google Scholar]
  22. Inan, O.T.; Etemadi, M.; Wiard, R.M.; Giovangrandi, L.; Kovacs, G.T.A. Robust ballistocardiogram acquisition for home monitoring. Physiol. Meas. 2009, 30, 169–185. [Google Scholar] [CrossRef]
  23. Wiard, R.M.; Inan, O.T.; Argyres, B.; Etemadi, M.; Kovacs, G.T.A.; Giovangrandi, L. Automatic detection of motion artifacts in the ballistocardiogram measured on a modified bathroom scale. Med. Biol. Eng. Comput. 2010, 49, 213–220. [Google Scholar] [CrossRef]
  24. Bruser, C.; Stadlthanner, K.; de Waele, S.; Leonhardt, S. Adaptive beat-to-beat heart rate estimation in ballistocardiograms. IEEE Trans. Inf. Technol. Biomed. 2011, 15, 778–786. [Google Scholar] [CrossRef] [PubMed]
  25. Dziuda, Ł.; Skibniewski, F.W. A new approach to ballistocardiographic measurements using fibre Bragg grating-based sensors. Biocybern. Biomed. Eng. 2014, 34, 101–116. [Google Scholar] [CrossRef]
  26. Nedoma, J.; Fajkus, M.; Martinek, R.; Kepak, S.; Cubik, J.; Zabka, S.; Vasinek, V. Comparison of bcg, pcg and ecg signals in application of heart rate monitoring of the human body. In Proceedings of the 40th International Conference on Telecommunications and Signal Processing (TSP), Barcelona, Spain, 5–7 July 2017; IEEE: New York, NY, USA, 2017; pp. 420–424. [Google Scholar]
  27. Song, Y.; Ni, H.; Zhou, X.; Zhao, W.; Wang, T. Extracting features for cardiovascular disease classification based on ballistocardiography. In Proceedings of the 2015 IEEE 12th Intl Conf on Ubiquitous Intelligence and Computing and 2015 IEEE 12th Intl Conf on Autonomic and Trusted Computing and 2015 IEEE 15th Intl Conf on Scalable Computing and Communications and its Associated Workshops (UIC-ATC-ScalCom), Beijing, China, 10–14 August 2015; IEEE: New York, NY, USA, 2015; pp. 1230–1235. [Google Scholar]
  28. Liu, F.; Zhou, X.; Wang, Z.; Wang, T.; Zhang, Y. Identification of hypertension by mining class association rules from multi-dimensional features. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; IEEE: New York, NY, USA, 2018; pp. 3114–3119. [Google Scholar]
  29. Liu, F.; Zhou, X.; Wang, Z.; Cao, J.; Wang, H.; Zhang, Y. Unobtrusive mattress-based identification of hypertension by integrating classification and association rule mining. Sensors 2019, 19, 1489. [Google Scholar] [CrossRef]
  30. Jain, P.; Gajbhiye, P.; Tripathy, R.; Acharya, U.R. A two-stage deep CNN architecture for the classification of low-risk and high-risk hypertension classes using multi-lead ECG signals. Informatics Med. Unlocked 2020, 21, 100479. [Google Scholar] [CrossRef]
  31. Rajput, J.S.; Sharma, M.; Kumbhani, D.; Acharya, U.R. Automated detection of hypertension using wavelet transform and nonlinear techniques with ballistocardiogram signals. Informatics Med. Unlocked 2021, 26, 100736. [Google Scholar] [CrossRef]
  32. Rajput, J.S.; Sharma, M.; Kumar, T.S.; Acharya, U.R. Automated detection of hypertension using continuous wavelet transform and a deep neural network with Ballistocardiography signals. Int. J. Environ. Res. Public Health 2022, 19, 4014. [Google Scholar] [CrossRef] [PubMed]
  33. Gupta, K.; Bajaj, V.; Ansari, I.A.; Acharya, U.R. Hyp-Net: Automated detection of hypertension using deep convolutional neural network and Gabor transform techniques with ballistocardiogram signals. Biocybern. Biomed. Eng. 2022, 42, 784–796. [Google Scholar] [CrossRef]
  34. Ge, B.; Yang, H.; Ma, P.; Guo, T.; Pan, J.; Wang, W. Detection of pulmonary hypertension associated with congenital heart disease based on time-frequency domain and deep learning features. Biomed. Signal Process. Control 2023, 81, 104316. [Google Scholar] [CrossRef]
  35. Ma, P.; Ge, B.; Yang, H.; Guo, T.; Pan, J.; Wang, W. Application of time-frequency domain and deep learning fusion feature in non-invasive diagnosis of congenital heart disease-related pulmonary arterial hypertension. MethodsX 2023, 10, 102032. [Google Scholar] [CrossRef]
  36. Sun, H.; Jiao, J.; Ren, Y.; Guo, Y.; Wang, Y. Multimodal fusion model for classifying placenta ultrasound imaging in pregnancies with hypertension disorders. Pregnancy Hypertens. 2023, 31, 46–53. [Google Scholar] [CrossRef]
  37. Dogan, S.; Barua, P.D.; Tuncer, T.; Acharya, U.R. An accurate hypertension detection model based on a new odd-even pattern using ballistocardiograph signals. Eng. Appl. Artif. Intell. 2024, 133, 108306. [Google Scholar] [CrossRef]
  38. Ozcelik, S.T.A.; Uyanık, H.; Deniz, E.; Sengur, A. Automated hypertension detection using ConvMixer and spectrogram techniques with ballistocardiograph signals. Diagnostics 2023, 13, 182. [Google Scholar] [CrossRef]
  39. Gupta, K.; Bajaj, V.; Ansari, I.A. A support system for automatic classification of hypertension using BCG signals. Expert Syst. Appl. 2022, 214, 119058. [Google Scholar] [CrossRef]
  40. Wahyuningrum, T.; Khomsah, S.; Suyanto, S.; Meliana, S.; Yunanto, P.E.; Al Maki, W.F. Improving Clustering Method Performance Using K-Means, Mini Batch K-Means, BIRCH and Spectral. In Proceedings of the 2021 4th International Seminar on Research of Information Technology and Intelligent Systems (ISRITI), Yogyakarta, Indonesia, 16–17 December 2021; pp. 206–210. [Google Scholar]
  41. Ali, H.H.S.M.; Sharif, S.M. Comparison Between Discrete Wavelet Transform and Maximal Overlap Discrete Wavelet Transform as an Analysis tool for H.264/AVC Video. In Proceedings of the 2018 International Conference on Computer, Control, Electrical, and Electronics Engineering (ICCCEEE), Khartoum, Sudan, 12–14 August 2018; pp. 1–5. [Google Scholar]
  42. Hidayat, R.; Kristomo, D.; Togarma, I. Feature extraction of the Indonesian phonemes using discrete wavelet and wavelet packet transform. In Proceedings of the 2016 8th International Conference on Information Technology and Electrical Engineering (ICITEE), Yogyakarta, Indonesia, 5–6 October 2016; pp. 1–6. [Google Scholar]
  43. Liu, W.; Cao, S.; Chen, Y. Seismic Time–Frequency Analysis via Empirical Wavelet Transform. IEEE Geosci. Remote Sens. Lett. 2015, 13, 28–32. [Google Scholar] [CrossRef]
  44. Prasad, K.D.; Khan, M.R.; Das, B. MUD in an Underwater Network Using Bioinspired Binary Tunicate Swarm Optimizer. In Proceedings of the 2025 3rd IEEE International Conference on Industrial Electronics: Developments & Applications (ICIDeA), Bhubaneswar, India, 21–22 February 2025; pp. 1–5. [Google Scholar]
  45. Thomaz, C.E.; Gillies, D.F. A Maximum Uncertainty LDA-Based Approach for Limited Sample Size Problems—With Application to Face Recognition. In Proceedings of the XVIII Brazilian Symposium on Computer Graphics and Image Processing (SIBGRAPI’05), Natal, Brazil, 9–12 October 2005; pp. 89–96. [Google Scholar]
  46. Wu, S.; Nagahashi, H. Parameterized AdaBoost: Introducing a Parameter to Speed Up the Training of Real AdaBoost. IEEE Signal Process. Lett. 2014, 21, 687–691. [Google Scholar] [CrossRef]
  47. Martins, W.; Bagesteiro, L.B.; Weber, T.O.; Balbinot, A. FPGA-based Implementation of Random Forest Classifier for sEMG Signal Classification. In Proceedings of the 2024 46th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 15–19 July 2024; pp. 1–4. [Google Scholar]
  48. Ge, W.; Hongzhe, X.; Weibin, Z.; Weilu, Z.; Baiyang, F. Multi-kernel PCA based high-dimensional images feature reduction. In Proceedings of the 2011 International Conference on Electric Information and Control Engineering, Wuhan, China, 15–17 April 2011; pp. 5966–5969. [Google Scholar]
  49. Chung, A.G.; Shafiee, M.J.; Wong, A. Random feature maps via a Layered Random Projection (LARP) framework for object classification. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 246–250. [Google Scholar]
  50. Yang, Y.Z.; Liu, P.Y.; Zhu, Z.F.; Qiu, Y. The research of an improved information gain method using distribution information of terms. In Proceedings of the 2009 IEEE International Symposium on IT in Medicine & Education, Jinan, China, 14–16 August 2009; pp. 938–941. [Google Scholar]
  51. Li, X.; Ma, Y.; Li, Y.; Li, H.; Zuo, H.; Sun, W. Improved Aquila Optimizer Optimization Algorithm Based on Multi-strategy Fusion. In Proceedings of the 2022 2nd International Conference on Computer Science, Electronic Information Engineering and Intelligent Control Technology (CEI), Nanjing, China, 23–25 September 2022; pp. 757–760. [Google Scholar]
  52. Fei, W.; Bai, L. Auto-regressive models of non-stationary time series with finite length. Tsinghua Sci. Technol. 2005, 10, 162–168. [Google Scholar] [CrossRef]
  53. Shang, Y. Finite-Time Weighted Average Consensus and Generalized Consensus Over a Subset. IEEE Access 2016, 4, 2615–2620. [Google Scholar] [CrossRef]
  54. Bernardini, M.; Romeo, L.; Misericordia, P.; Frontoni, E. Discovering the Type 2 Diabetes in Electronic Health Records Using the Sparse Balanced Support Vector Machine. IEEE J. Biomed. Health Inform. 2019, 24, 235–246. [Google Scholar] [CrossRef]
  55. Li, W.; Wang, R.; Huang, D. Assessment of Micro-movement Sensitive Mattress Sleep Monitoring System (RS611) in the detection of obstructive sleep apnea hypopnea syndrome. Chin. J. Gerontol. 2015, 35, 1160–1162. [Google Scholar]
  56. Park, S.H.; Reyes, J.A.; Gilbert, D.R. Prediction of protein-protein interaction types using association rule-based classification. BMC Bioinform. 2009, 10, 36. [Google Scholar] [CrossRef]
Figure 1. Illustration of the two proposed pragmatic models.
Figure 1. Illustration of the two proposed pragmatic models.
Bioengineering 13 00043 g001
Figure 2. Simplified illustration of the ARIMA–AdaBoost Hybrid model.
Figure 2. Simplified illustration of the ARIMA–AdaBoost Hybrid model.
Bioengineering 13 00043 g002
Figure 3. Simplified illustration of the TW-HASVM classification model.
Figure 3. Simplified illustration of the TW-HASVM classification model.
Bioengineering 13 00043 g003
Figure 4. Comparison of performance analysis (accuracy) of the BTSA feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Figure 4. Comparison of performance analysis (accuracy) of the BTSA feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Bioengineering 13 00043 g004
Figure 5. Comparison of performance analysis (accuracy) of the AOA feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Figure 5. Comparison of performance analysis (accuracy) of the AOA feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Bioengineering 13 00043 g005
Figure 6. Comparison of performance analysis (accuracy) of the IG feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Figure 6. Comparison of performance analysis (accuracy) of the IG feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Bioengineering 13 00043 g006
Figure 7. Comparison of performance analysis (accuracy) of the IG feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Figure 7. Comparison of performance analysis (accuracy) of the IG feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Bioengineering 13 00043 g007
Table 1. Hypertensive group.
Table 1. Hypertensive group.
Number61
Gender (male)33
Gender (female)38
Age group55.6 ± 7.9
Body Mass Index (BMI)24.3 ± 3.6
Heart rate (bpm)77.1 ± 9.2
Systolic blood pressure (mmHg)155.6 ± 11.2
Diastolic blood pressure (mmHg)103.6 ± 8.2
Table 2. Control group.
Table 2. Control group.
Number67
Gender (male)35
Gender (female)32
Age group53.2 ± 9.2
Body Mass Index (BMI)23.4 ± 3.3
Heart rate (bpm)73.6 ± 8.3
Systolic blood pressure (mmHg)112.1 ± 15.7
Diastolic blood pressure (mmHg)74.4 ± 6.3
Table 3. Performance analysis (accuracy) of the BTSA feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Table 3. Performance analysis (accuracy) of the BTSA feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
RFDTNBCAdaBoostHybrid AdaBoost–MULDAHybrid AdaBoost–RF
K-means89.4589.5889.0591.4595.5195.45
MODWT91.6791.1490.6492.2496.4895.48
EWT90.9890.5789.8790.6994.3793.54
Table 4. Performance analysis (accuracy) of the IG feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Table 4. Performance analysis (accuracy) of the IG feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
RFDTNBCAdaBoostHybrid AdaBoost–MULDAHybrid AdaBoost–RF
K-means88.3487.2188.0990.2293.0991.98
MODWT90.5689.3688.9791.2493.8792.23
EWT88.6388.7887.4590.5792.4291.58
Table 5. Performance analysis (accuracy) of the chi-square feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Table 5. Performance analysis (accuracy) of the chi-square feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
RFDTNBCAdaBoostHybrid AdaBoost–MULDAHybrid AdaBoost–RF
K-means85.0982.2183.8784.3386.2185.02
MODWT84.9783.2484.7481.3787.3686.34
EWT83.5884.6785.3583.8986.9885.67
Table 6. Performance analysis (accuracy) of the PCC feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Table 6. Performance analysis (accuracy) of the PCC feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
RFDTNBCAdaBoostHybrid AdaBoost–MULDAHybrid AdaBoost–RF
K-means84.2380.0984.2186.0988.1188.05
MODWT83.4581.8782.2583.9889.2689.04
EWT83.6882.3484.7885.4289.9087.37
Table 7. Performance analysis (accuracy) of the PSO feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Table 7. Performance analysis (accuracy) of the PSO feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
RFDTNBCAdaBoostHybrid AdaBoost–MULDAHybrid AdaBoost–RF
K-means86.3284.8982.2184.0990.2189.33
MODWT87.6685.9783.2586.9891.2390.46
EWT88.7885.5582.7885.7790.5888.78
Table 8. Performance analysis (accuracy) of the ACO feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Table 8. Performance analysis (accuracy) of the ACO feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
RFDTNBCAdaBoostHybrid AdaBoost–MULDAHybrid AdaBoost–RF
K-means85.2184.0385.2187.0389.2187.01
MODWT85.3386.4583.2384.3389.6788.21
EWT85.4886.6882.6985.5987.9987.08
Table 9. Performance analysis (accuracy) of the AOA feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Table 9. Performance analysis (accuracy) of the AOA feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
RFDTNBCAdaBoostHybrid ARIMA–AdaBoostTW-HASVM
PCA90.2190.4489.1192.0996.2396.02
KPCA90.4592.5691.2493.9896.8995.34
RFM91.7891.7990.7892.5695.8996.71
Table 10. Performance analysis (accuracy) of the IG feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Table 10. Performance analysis (accuracy) of the IG feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
RFDTNBCAdaBoostHybrid ARIMA–AdaBoostTW-HASVM
PCA87.2386.2189.9890.0292.1192.12
KPCA89.4488.2288.8791.3491.2692.37
RFM87.5787.5789.5891.6891.8791.98
Table 11. Performance analysis (accuracy) of the chi-square feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Table 11. Performance analysis (accuracy) of the chi-square feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
RFDTNBCAdaBoostHybrid ARIMA–AdaBoostTW-HASVM
PCA84.2180.7884.0985.2187.2284.22
KPCA83.2281.9885.8782.3588.3785.37
RFM81.4682.3683.3684.7886.6589.30
Table 12. Performance analysis (accuracy) of the PCC feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Table 12. Performance analysis (accuracy) of the PCC feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
RFDTNBCAdaBoostHybrid ARIMA–AdaBoostTW-HASVM
PCA82.8982.1282.0984.2186.8787.09
KPCA80.4480.3481.8485.2585.7788.88
RFM80.5781.7883.2284.6787.6387.34
Table 13. Performance analysis (accuracy) of the PSO feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Table 13. Performance analysis (accuracy) of the PSO feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
RFDTNBCAdaBoostHybrid ARIMA–AdaBoostTW-HASVM
PCA87.2186.8785.0988.2390.7789.21
KPCA88.2287.7586.8785.4590.6391.36
RFM89.3687.4984.3487.6191.4889.89
Table 14. Performance analysis (accuracy) of the ACO feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
Table 14. Performance analysis (accuracy) of the ACO feature selection technique combined with the feature extraction modules and the hybrid AdaBoost-based classifiers.
RFDTNBCAdaBoostHybrid ARIMA–AdaBoostTW-HASVM
PCA83.2185.0982.2385.0988.2489.09
KPCA82.3484.3684.6786.9788.6787.88
RFM84.9883.7885.3384.4589.8489.21
Table 15. Performance summary of the current results compared with previous studies.
Table 15. Performance summary of the current results compared with previous studies.
YearAuthorsMethodsPerformance Metrics
2018Liu et al. [28]Mining class association from multi-dimensional featuresAcc = 85.20%
2019Liu et al. [29]Integrated classification and association rule miningAcc = 84.40%
2021Rajput et al. [31]EMD, wavelet decomposition and nonlinear techniquesAcc = 89.00%
2022Rajput et al. [32]Continuous wavelet transform combined with deep neural networksAcc = 86.14%
2022Gupta et al. [33]Gabor transform and deep convolutional neural networkAcc = 97.65%
2023Gupta et al. [39]Integrated tunable Q-factor wavelet transform and multiverse optimization conceptAcc = 92.21%
2023 Ozcelik et al. [38]Spectrogram tasks + ConvMixerAcc = 97.69%
2024Dogan et al. [37]Odd–even pattern with KNN algorithm and majority votingAcc = 97.78%
2025Proposed works
(a)
K-means + BTSA and Hybrid AdaBoost–MULDA
(b)
MODWT + BTSA + Hybrid AdaBoost–MULDA
(c)
EWT + BTSA + Hybrid AdaBoost–MULDA
Acc = 95.51%
Acc = 96.48%
Acc = 94.37%
(a)
K-means + IG and Hybrid AdaBoost–MULDA
(b)
MODWT + IG + Hybrid AdaBoost–MULDA
(c)
EWT + IG + Hybrid AdaBoost–MULDA
Acc = 93.09%
Acc = 93.87%
Acc = 92.42%
(a)
PCA + AOA and Hybrid ARIMA–AdaBoost
(b)
KPCA + AOA + Hybrid ARIMA–AdaBoost
(c)
RFM + AOA + Hybrid ARIMA–AdaBoost
Acc = 96.23%
Acc = 96.89%
Acc = 95.89%
(a)
PCA + IG + TW-HASVM
(b)
KPCA + IG + TW-HASVM
(c)
RFM + IG + TW-HASVM
Acc = 92.12%
Acc = 92.37%
Acc = 91.98%
Table 16. Computational time analysis for the best performing models.
Table 16. Computational time analysis for the best performing models.
YearAuthorsMethodsComputational Time
2025Proposed works
(d)
K-means + BTSA and Hybrid AdaBoost–MULDA
(e)
MODWT + BTSA + Hybrid AdaBoost–MULDA
(f)
EWT + BTSA + Hybrid AdaBoost–MULDA
10.563 s
11.342 s
11.291 s
(d)
K-means + IG and Hybrid AdaBoost–MULDA
(e)
MODWT + IG + Hybrid AdaBoost–MULDA
(f)
EWT + IG + Hybrid AdaBoost–MULDA
10.247 s
9.298 s
9.037 s
(d)
PCA + AOA and Hybrid ARIMA–AdaBoost
(e)
KPCA + AOA + Hybrid ARIMA–AdaBoost
(f)
RFM + AOA + Hybrid ARIMA–AdaBoost
6.567 s
5.112 s
6.012 s
(d)
PCA + IG + TW-HASVM
(e)
KPCA + IG + TW-HASVM
(f)
RFM + IG + TW-HASVM
8.378 s
9.001 s
8.981 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Prabhakar, S.K.; Won, D.-O. Pragmatic Models for Detection of Hypertension Using Ballistocardiograph Signals and Machine Learning. Bioengineering 2026, 13, 43. https://doi.org/10.3390/bioengineering13010043

AMA Style

Prabhakar SK, Won D-O. Pragmatic Models for Detection of Hypertension Using Ballistocardiograph Signals and Machine Learning. Bioengineering. 2026; 13(1):43. https://doi.org/10.3390/bioengineering13010043

Chicago/Turabian Style

Prabhakar, Sunil Kumar, and Dong-Ok Won. 2026. "Pragmatic Models for Detection of Hypertension Using Ballistocardiograph Signals and Machine Learning" Bioengineering 13, no. 1: 43. https://doi.org/10.3390/bioengineering13010043

APA Style

Prabhakar, S. K., & Won, D.-O. (2026). Pragmatic Models for Detection of Hypertension Using Ballistocardiograph Signals and Machine Learning. Bioengineering, 13(1), 43. https://doi.org/10.3390/bioengineering13010043

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop