Next Article in Journal
A Survey of Machine and Deep Learning Methods for Privacy Protection in the Internet of Things
Next Article in Special Issue
Competition between Direct Detection Mechanisms in Planar Bow-Tie Microwave Diodes on the Base of InAlAs/InGaAs/InAlAs Heterostructures
Previous Article in Journal
Suitability Study of Using UAVs to Estimate Landfilled Fly Ash Stockpile
Previous Article in Special Issue
Packaging and Optimization of a Capacitive Biosensor and Its Readout Circuit
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing System Performance through Objective Feature Scoring of Multiple Persons’ Breathing Using Non-Contact RF Approach

1
Department of Electrical Engineering, HITEC University, Taxila 47080, Pakistan
2
Department of Electrical and Computer Engineering, COMSATS University Islamabad, Attock Campus, Attock 43600, Pakistan
3
College of Information Technology, United Arab Emirates University (UAEU), Abu Dhabi 15551, United Arab Emirates
4
Research Centre for Intelligent Healthcare, Coventry University, Coventry CV1 5FB, UK
5
School of Electronic Engineering and Computer Science, Queen Mary University of London, London E1 4NS, UK
6
School of Electronic Engineering, Xidian University, Xi’an 710071, China
7
School of Engineering, University of Glasgow, Glasgow G12 8QQ, UK
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(3), 1251; https://doi.org/10.3390/s23031251
Submission received: 2 December 2022 / Revised: 16 January 2023 / Accepted: 19 January 2023 / Published: 21 January 2023
(This article belongs to the Special Issue Feature Papers in Physical Sensors 2022)

Abstract

:
Breathing monitoring is an efficient way of human health sensing and predicting numerous diseases. Various contact and non-contact-based methods are discussed in the literature for breathing monitoring. Radio frequency (RF)-based breathing monitoring has recently gained enormous popularity among non-contact methods. This method eliminates privacy concerns and the need for users to carry a device. In addition, such methods can reduce stress on healthcare facilities by providing intelligent digital health technologies. These intelligent digital technologies utilize a machine learning (ML)-based system for classifying breathing abnormalities. Despite advances in ML-based systems, the increasing dimensionality of data poses a significant challenge, as unrelated features can significantly impact the developed system’s performance. Optimal feature scoring may appear to be a viable solution to this problem, as it has the potential to improve system performance significantly. Initially, in this study, software-defined radio (SDR) and RF sensing techniques were used to develop a breathing monitoring system. Minute variations in wireless channel state information (CSI) due to breathing movement were used to detect breathing abnormalities in breathing patterns. Furthermore, ML algorithms intelligently classified breathing abnormalities in single and multiple-person scenarios. The results were validated by referencing a wearable sensor. Finally, optimal feature scoring was used to improve the developed system’s performance in terms of accuracy, training time, and prediction speed. The results showed that optimal feature scoring can help achieve maximum accuracy of up to 93.8% and 91.7% for single-person and multi-person scenarios, respectively.

1. Introduction

Human breathing monitoring is essential and has a significant role in various healthcare applications [1]. Breathing monitoring helps in differentiating between normal and abnormal breathing. In normal breathing, a human takes 12 to 24 breaths per minute (bpm), while in abnormal breathing, bpm can be above or below this range. If breathing is above 24 bpm, this is called fast breathing, while in slow breathing, it is less than 12 bpm [2]. There is substantial proof that breathing monitoring is a beneficial vital sign. It predicts potentially severe adverse events [3] and a timely, clear indication of physiological deterioration [4]. Furthermore, breathing monitoring is not only an important indicator of cardiac arrest and intensive care units (ICU) admission but also an independent prognostic indicator for risk assessment after an acute heart attack [5]. In addition, it is critical in detecting the risk of dangerous conditions such as sleep apnea [6], respiratory deterioration in post-surgical patients [7], and sudden infant death syndrome [8]. Likewise, breathing monitoring detects various pathological conditions, such as diabetic toxicological issues, ketoacidosis, allergic reactions, pain, shock, and dehydration [9]. Therefore, the technological development of accurate breathing measurement is essential for real-time deployment in domestic and clinical settings.
An abundance of literature is present in the area of human breathing monitoring. Traditional solutions typically require a wearable device that can bring discomfort to patients and may cause the spread of viruses such as coronavirus (COVID) [10]. Camera technologies [11] are also considered a practical solution, but they have limitations with regard to privacy in the case of imaging cameras, and in line-of-sight (LOS) scenarios. Some acoustic-based solutions [12] can also accomplish high accuracy. Still, they have a low sensing range and are more sensitive to environmental noise. Various RF-based solutions are discussed in the literature, including radar, Wi-Fi, and SDR. Radar-based RF solutions are reasonable but require dedicated, expensive hardware and are not readily available in home and hospital settings [13]. Wi-Fi-based RF sensing has recently gained considerable research interest, as Wi-Fi devices are readily available. The CSI retrieved from commodity Wi-Fi devices can provide helpful knowledge about breathing activities [14]. However, these Wi-Fi-based solutions are not scalable and flexible [15]. Various authors have also previously studied SDR-based RF solutions for breathing monitoring [16,17,18]. SDR-based RF sensing is considered the most efficient and effective breathing monitoring among all RF-based solutions. It provides flexibility, scalability, and portability by offering user-specific configuration of various parameters, such as operating frequency range and transmit/received power. Despite the fact that all previous work on SDR-based RF sensing has been proposed for single-person breathing scenarios, none of the previous approaches proposed a solution for analyzing real-time breathing monitoring for multi-person scenarios. From the literature, this is the first research work conducted for multi-person breathing using SDR-based RF sensing.
In the case of RF-based sensing, two types of information are obtained and widely used in the time domain. One piece of information is the breathing pattern, which is essentially a detailed process of inhalation and exhalation over time [19]. The other piece of information is the breathing rate, which is the frequency or number of breaths taken over time. This data can be derived from breathing patterns, such as counting the number of complete breathing cycles. In a single-person scenario, both pieces of information can be extracted from time-domain CSI amplitude. In a multi-person scenario, however, it is impossible to extract an individual person’s breathing information from entangled CSI because received CSI is influenced by independent chest movements of all persons at the same time. Based on this observation, it is possible to extract breathing information for multi-person scenarios by transferring time-domain CSI amplitude information into the frequency domain. The Fourier transformation is used for this purpose, which aids in extracting individual breathing information in a multi-person scenario.
This experimental study attempted to address multiple research gaps in RF sensing for breathing monitoring and, in doing so, the significant contributions are:
This study addressed the existing RF-based sensing limitations [11,13,15] by developing a non-contact SDR-based RF sensing platform for monitoring breathing abnormalities.
The developed system can adopt multiple wireless standards compared to Wi-Fi-based RF-sensing.
The performance of classical feature extracting approaches was improved by introducing the optimal feature scoring methods for building ML models.
This study is the first to consider multi-person breathing monitoring using RF sensing by exploiting the SDR technology to offer a portable and adaptable solution.
The rest of the paper is structured as follows. Section 2 presents the related work. Section 3 discusses design methodology. Section 4 is all about the experimental setup. Section 5 presents the result and discussion. Section 6 provides the conclusions and future recommendations.

2. Related Work

This section discusses the work related to human breathing monitoring human using various RF technologies. These technologies can be classified into three groups:

2.1. Radar-Based RF Sensing

Radar-based RF sensing includes a wide range of literature for breathing monitoring. Frequency-modulated continuous wave (FMCW) radar is utilized to monitor breathing, and it works at a wide bandwidth of 1–2 GHz by radiating signals modulated with a linear frequency [20]. In vital radio [21], FMCW radar detects multi-person breathing by separating signals based on propagation time. In [22], exploit the feasibility of single- and multi-person scenario breathing monitoring using automotive FMCW radar (76–81 GHz). In [23], millimetre-wave 122 GHz radar detects humans and non-humans by detecting their vital breathing signs. The automatic detection of human breathing separation is demonstrated using two different radars operating at frequencies (24 and 122 GHz). This approach identifies humans in complex, cluttered environments [24]. However, this system does not work when multiple persons are located nearby. “DeepBreath” [25] used the independent component analysis (ICA) method to measure multiple persons breathing. However, dedicated hardware is costly and makes it impractical for daily home usage. Ultra-wideband (UWB) pulse radar, and continuous wave (CW) doppler radar are also used for breathing monitoring. UWB pulse radar works by transmitting and receiving short-duration pulses [26]. This reduces interference and multipath by having a large bandwidth of 1–2 GHz [27]. However, such large bandwidths demand peak signal strength and accurate pulse width control, resulting in hardware complexity [28]. CW doppler radar radiates a signal having fixed amplitude and frequency. This radar has a simple design but is more susceptible to multipath reflections and environmental noise.

2.2. Wi-Fi-Based Breathing Sensing

Wi-Fi-based RF sensing for human breathing is exploited for single-person or multi-person scenarios. For a single-person scenario, [29] developed a system for sleep monitoring by obtaining breathing information through Wi-Fi signals. In [30], the author improved the performance by considering abnormal breathing patterns and sleeping postures. These systems extracted a sinusoidal pattern from the time domain CSI for a single person’s breathing scenario. The Fresnel zone theory is introduced by [31] to find the reason behind the blind spots issue, which occurs in the single-person scenario. “FullBreathe” [32] utilized phase and amplitude information of CSI to eliminate the blind spots problem. [33] expanded the range of breathing sensing up to 9 m by gathering information from both antennas of the Wi-Fi device. In [32,33], the authors also suggested solutions to counter the blind spot problem in a one-person scenario. However, the proposed system fails for multi-person scenarios because the theory applied for developing the system lies on the assumption of a one-person scenario. For multi-person scenarios, various authors contributed and obtained valuable results. [34] made the first attempt to find the breathing rate of multi-person by evaluating frequency domain CSI. Wang et al. [35] highlighted that the blind spot issue significantly lessens the method’s performance presented by [34]. Apart from these systems, “PhaseBeat” [36] applied the root-MUSIC algorithm [37] to the phase difference of CSI between multiple antennas to obtain the breathing information for the multi-person case. Similarly, “TensorBeat” [38] detected multi-person breathing through CSI phase difference. In comparison, “TR-BREATH” [39] applied a root-MUSIC algorithm to find the breathing rates of multiple persons. [40] obtained the breathing information by optimal placement of Wi-Fi transceivers so that each transceiver pair is only impacted by one person breathing.

2.3. SDR-Based Breathing Sensing

Diverse systems have been proposed in the literature for SDR-based RF monitoring of single-person breathing. A non-contact SDR-based system is proposed by [16] for estimating breathing and heart rate due to minute movements of the chest. For this purpose, directional antennas are exploited, and a vector network analyzer (VNA) is used for results comparison. The system performance is also evaluated by varying distances between the human body and antennas. Additionally, a through-wall monitoring scenario is also considered. In [17], a continuous wave multi-frequency radar system (MFCW) is deployed using SDR and breathing patterns monitoring is done at target distances. In [41], channel frequency response (CFR) is exploited to detect minor variations in OFDM subcarriers caused by various human movements over wireless channels. The developed platform accurately captured hand waving movement, abnormal coughing, and different breathing patterns. In [42], a contactless breathing pattern detection system is developed using universal software radio peripheral (USRP). This platform utilized CSI to record the tiny movements generated due to breathing activity and detected three breathing patterns, while in [43], SDR-based system is developed, and system design is validated by first analyzing the CFR for various simulated channels. Finally, several breathing patterns are successfully classified using ML algorithms. In [44], SDR-based breathing pattern sensing detects and classifies six abnormal breathing patterns, while in [18], this work is further extended by classifying up to eight breathing patterns.

3. Design Methodology

The design methodology exploits the wireless communication system for sensing purposes by detecting human motion by capturing CSI through the transmission of electromagnetic (EM) waves. The RF signal is transmitted and received via multiple paths, thus regenerating the signal with multi-path superposition. The received signal contains information such as characteristics from the physical space environment. The characteristics include distance, power, the human body’s motion, and environmental influences in signal propagation. RF sensing precisely adds the environmental factors’ influence on the signal to clearly understand the characteristics of the environment to realize human body motion. The extra signal path is added to the human body’s diffraction or reflection whenever a human is present in the physical space environment. As a result, the influence of human body motion on the signal propagation contributes to the wireless communication. Figure 1 illustrates the design methodology consisting of four main blocks: wireless signal sensing, signal preprocessing, breathing monitoring, and breathing classification. The description of each block is given below:

3.1. Wireless Signal Sensing

As shown in Figure 2, the wireless signal sensing block consists of three subsections: the transmitter, the real-time wireless channel, and the receiver. The details of each subsection are given below:

3.1.1. Transmitter

Initially, in the transmitter, random data bits are generated and converted into symbols using quadrature amplitude modulation (QAM). These QAM symbols are then used to generate parallel streams. Then, reference data symbols are inserted to estimate the channel at the receiver side. Nulls and DC symbols are also inserted in each frame. The signal is then converted to time domain using Inverse FFT. The cyclic prefix (CP) is then appended to each frame. Finally, using an ethernet cable, the data generated by the host transmitter PC are transferred to the universal software radio peripheral (USRP) kit. The USRP performs a variety of operations, including digital up-conversion (DUC) and digital-to-analog conversion (DAC). After this, low pass filtering is applied, and the resulting signal is mixed with a carrier frequency specified by the user. The signal is then amplified using a transmitting amplifier before being transmitted via an omnidirectional antenna.

3.1.2. Wireless Channel

A real-time wireless channel was considered to monitor human breathing activities in this research work. The real-time wireless channel contains a wealth of information about the environment, and various techniques are used in the literature to extract this valuable information. However, the focus of this study was on the analysis of wireless CSI using CFR, which is calculated using Equation (1):
H ( k ) = Y ( k ) X ( k )  
Here H ( k ) represents CFR, while X ( k ) and Y ( k )   represents frequency domain transmitted and received signals, respectively. Since H ( k )   is a complex value so that we can extract the amplitude response given in Equation (2):
| H ( k ) | = H R e 2 + H I m 2
Here H R e 2 and H I m 2 represent the real as well as the imaginary part of the CFR.
For a single experimental measurement E, the CFR amplitude information recorded in time history using multiple OFDM frames can be given by Equation (3).
| H ( k ) | E = [ | H ( k ) | 1 , 1 | H ( k ) | 1 , 2 | H ( k ) | 1 , F | H ( k ) | 2 , 1 | H ( k ) | 2 , 2 | H ( k ) | 2 , F | H ( k ) | K , 1 | H ( k ) | K , 2 | H ( k ) | K , F ]
where K represents the total number of OFDM subcarriers, and F represents the total number of OFDM frames received during the single E .

3.1.3. Receiver

On the receiver side, the signal is initially received via an omnidirectional antenna before passing through low noise amplifier (LNA) and drive amplifier (DA). Following that, a baseband complex signal is obtained via the mixing process and a direct conversion receiver (DCR). After that, low pass filtering (LPF) is used, followed by analog to digital conversion (ADC), and finally, digital down-conversion (DDC). Finally, the down-converted signal is routed to the host PC via the ethernet cable. The time and frequency offsets are included in the received signal when it arrives at the host PC. The time offset t O S results in the rotation of data symbols and can be modeled as a delay in the channel impulse response, while frequency offset f O S results in shifting of all subcarriers and can be modeled as complex multiplicative distortion e j 2 π f O S n N . The received time-domain signal r ( n ) after including both these offsets can be written as Equation (4):
r ( n ) = c ( n ) × s ( n t O S ) e j 2 π f O S n N + N ( n )
where c ( n )   is channel response and s ( n ) is the transmitted signal in time domain. In this research work, the van de Beek algorithm was used to estimate the time offset t O S and frequency offset f O S   using Equations (5) and (6):
t O S = arg   max {   | γ ( t O S ) | ρ Φ ( t O S ) }
f O S = 1 2 π γ ( t O S )
The above equation | γ ( t O S ) | is the correlation between two pairs of L samples of OFDM frame that are N samples apart and represented in Equation (7). Φ ( t O S ) is the energy part, and ρ is the magnitude of the correlation coefficients, and both are represented by Equations (8) and (9), respectively. γ ( t O S ) is used to estimate time offset t O S and f O S . The magnitude of γ ( t O S ) is compensated by energy term Φ ( t O S ) and peaks at time instant, which provides f O S , while its phase at this time instant is proportional to f O S .
γ ( m ) = n = m m + L 1 r ( n ) r * ( n + N )
Φ ( m ) = 1 2 n = m m + L 1 | r ( n ) | 2 + | r ( n + N ) | 2
ρ = | E { r ( k ) r * ( k + N ) } | E { | r ( k ) | } 2   E { | r ( k + N ) | } 2  
After the removal of time and frequency offsets, CP is also removed. Later, FFT is applied to acquire frequency-domain signal. Subsequently, reference symbols are removed, followed by nulls and DC. The reference symbols retrieved here are used for channel estimation. Next, equalized data is obtained based on the estimated channel and fed to QAM demodulation block, which converts symbols into the bitstream.

3.2. Signal Preprocessing

During signal processing, various steps are taken, and detail of each stage is given below:

3.2.1. Subcarrier Selection

Initially, for data preprocessing, the subcarrier selection is performed. The purpose of subcarrier selection is to remove such subcarriers having less sensitivity to breathing activity. For this purpose, subcarriers’ variance is measured, after which subcarriers having minor variance below 0.001 are eliminated, as shown in Figure 3a.

3.2.2. Outlier Removal

After subcarrier selection, outliers are removed through wavelet filtering. As shown in Figure 3b, outlier removal eliminates all outliers but keeps a sharp data transition.

3.2.3. Smoothening

The moving average filter is applied to eliminate high-frequency noise for data smoothening. The moving average filter of window size 8 is applied [14], using Equation (10), and output can be seen in Figure 3c:
y [ n ] = 1 M k = 0 M 1 x [ n k ]
Here y [ n ] ,   x [ n ] represent the current output, and input respectively, and M is the size of the window of the moving average filter.

3.2.4. Normalization

The final step of data preprocessing is normalization, and the purpose is to normalize the data to maximum and minimum values between 1 and −1. First, the data are normalized using Equation (11). Figure 3d shows the output of data normalisation for a single subcarrier.
y [ n ] ´ = y [ n ] o f f s e t s c a l e
Here, y [ n ] ´ represents the normalized data, and y [ n ]   represents the input data. The normalized data is acquired by adjusting scaling and offset values.

3.3. Breathing Monitoring

Following data preprocessing, the next step is breathing monitoring, which extracts breathing patterns and rates for single and multiple-person scenarios. CSI amplitude in the time history recorded data can be used to extract breathing patterns. However, in order to calculate the breathing rate, the CSI amplitude data is converted into the frequency domain using FFT. As a result, a strong frequency peak in the frequency domain is obtained, which corresponds to breathing rate. The breathing rate from the frequency peak value is calculated by using Equation (12) as:
B r e a t h i n g   r a t e = f m a x   × s
where f m a x   represent maximum frequency peak value and s represents the number of seconds for breathing activity. The number of maximum peaks is proportional to the number of persons doing the breathing activity. Therefore, one maximum frequency peak is obtained for the single-person scenario. Similarly, multiple frequency peaks are obtained for the multiple-person scenarios depending on the number of persons.

3.4. Breathing Classification

The last block in the methodology is breathing classification. This block first performs optimal feature scoring to obtain optimal features for the ML classification model. Later, various ML models are applied to classify different breathing patterns for a single-person and multi-person scenario using these features.

3.4.1. Optimal Feature Scoring

Optimal feature scoring is a two-step process. The first step extracts various features from preprocessed time-domain CSI amplitude information. Afterward, feature selection methods are applied to select only relevant features, which helps improve the ML algorithms’ performance.
(a)
Features extraction
Feature extraction is a significant step in the classification of breathing patterns. Developing a classification model having high dimensionality is time-consuming. Therefore, extracting only the useful features can improve the performance of the classification model [45]. For this purpose, various statistical features were calculated, and their details are shown in Table 1. Similar features must be eliminated for optimal feature scoring. In this research work, boxplots were used to compare the similarity of various features. Boxplots summarize feature datasets, such as the minimum, maximum, median, first (lower) quartile, and third (upper) quartile. If the boxplots for multiple features overlap, this indicates that there is no difference between these data, and these features can be removed. Figure 4 shows that some features are the same and overlapping and thus can be removed using feature selection methods.
(b)
Features Selection
Feature selection reduces data dimensionality by selecting only the most relevant and useful features. The feature selection methods are employed in order to discover features that improve prediction efficiency [46]. In addition, even using a large number of relevant features can reduce prediction efficiency. Various feature selection methods are discussed in the literature, but this study used only two methods to improve ML classification model performance.
I.
Minimum redundancy maximum relevance (MRMR) algorithm
The MRMR algorithm (Algorithm 1) is used as a filter type method for feature selection. It measures feature importance based on the characteristics of the features, such as feature variance and feature relevance to the output. The MRMR algorithm [47] works on the principle of finding an optimal set of maximally and mutually unique features. The benefit of such features is that they can effectively represent output variables. This algorithm not only maximizes the relevance but also minimizes the redundancy of optimal feature set to the output variable. The algorithm uses pairwise mutual information of features and the output to measure redundancy and relevance. This algorithm obtains optimal features set F that maximizes M F   , and minimizes N F with respect to output variable y, where M F   and N F are the relevance and redundancy of F respectively. Both M F   and N F are defined through mutual information I in Equations (13) and (14) as:
M F = 1 | F | x F I ( x , y )
N F = 1 | F | 2 x , z F I ( x , z )
Here | F | represents total features in F . To find N F , 2 | | combinations are considered, while the whole feature set is denoted by . Alternatively, MRMR algorithm applies a forward addition scheme for ranking features through employing the mutual information quotient (MIQ) value as in Equation (15), but this requires O ( | | . | S | ) computations.
M I Q x = M x N x
Here M x is the relevance and N x is the redundancy of a feature, as shown in Equations (16) and (17), respectively:
M x = I ( x , y )
N x = 1 | F | z F I ( x , z )
This MRMR algorithm ranks features in by returning features indices based on importance. This results in a computation cost of O ( | | 2 ) . The function returns the score of feature importance using a heuristic algorithm. A feature with a large value score shows the importance of the feature and vice versa. The feature importance for various features is shown in Figure 5. It can be seen that out of 10 features, some features have very low importance and can be removed.
II.
Principle component analysis (PCA)
PCA is a dimensionality reduction technique that aids in feature reduction by transforming data from the original high-dimensional feature space to a new space with reduced dimensionality [48]. PCA projects large features into subspaces with fewer features while retaining the essence of the original data. In this study, for example, all features with 95% variance were kept, while others were removed. This aids in feature selection, saving storage and computational time, and improving comprehension.
Algorithm 1 MRMR Algorithm—Pseudocode for optimal features selection
  • Choose the feature with the highest relevance,   x m a x   M x   and then add that chosen feature to an empty set F ;
  • Find the features with nonzero relevance and zero redundancy in the complement of F ,   F c ;
    • If F c   does not contain a feature with nonzero relevance and zero redundancy, then move to step 4;
    • Otherwise, choose the feature with the highest relevance,   x F c , N x = 0 m a x     M x   . Then add the chosen feature into the set F ;
  • Repeat step 2 unless the redundancy is not zero for all the features in   F c ;
  • Choose the feature that has the largest MIQ value with nonzero relevance and nonzero redundancy in F c , and add the selected feature to the set F ;
         M x F c m a x I Q x =   x F c m a x I ( x , y ) 1 | F | z F I ( x , y )
  • Step 4 is repeated unless the relevance of each feature in F c is zero;
  • Features having zero relevance are included in a random sequence in F .

3.4.2. Breathing Patterns Classification

The next step is breathing pattern classification. In this step, various ML algorithms are used for breathing pattern classification. Each algorithm is evaluated based on accuracy, training time, and prediction speed for the scenarios of a single person and multi-person. First, each algorithm is evaluated without feature selection methods. Then after applying feature selection methods, all algorithms are re-evaluated.

4. Experimental Setup

The experimental setup for breathing data collection comprises of pair of USRPs and PCs, as shown in Figure 6. For transmitting and receiving, each USRP is outfitted with an omnidirectional antenna. The main advantage of an omnidirectional antenna is that it can detect human breathing in both LOS and non-LOS scenarios. The transmitter PC is linked to the USRP via an ethernet cable, and the function of each PC on the transmitter side is to generate OFDM subcarrier data. Simultaneously, preprocess and classify the raw breathing data on the receiver side. To avoid the blind spot problem, the experimental setup is installed based on empirical experience. USRP kits are kept parallel to the participants’ abdomens.
A total of ten male participants are asked to do breathing experiments. Table 2 lists each participant’s information. For the scenario of a single person, all participants are asked to perform breathing at normal, slow, and fast rates. For a two-person scenario, both participants are asked to perform breathing at normal, slow, and fast rates, resulting in six different cases. For a three-person scenario, all three participants are asked to perform breathing at normal, slow, and fast rates, resulting in nine different cases. Ten data sets are collected for each breathing pattern activity, and the duration of each activity is 30 s. Moreover, extensive experimentation is performed to achieve high accuracy.

5. Results and Discussions

This section is divided further into three subsections. The first subsection displays the results of several breathing patterns. The results of breathing rate extraction are shown and discussed in the second subsection. Finally, the results of breathing patterns classification are shown in the final subsection, in which various ML algorithms are evaluated with and without feature selection methods.

5.1. Breathing Pattern Extraction

For a single-person scenario, three breathing patterns are detected: normal, slow, and fast, as shown in Figure 7. In this study, a wearable sensor is used as a reference to compare the performance of non-contact sensing. There are nine breaths in 30 s, as shown in Figure 7a, indicating that this is a case of normal breathing. Similarly, in Figure 7b, there are six breaths in 30 s, confirming this is a case of slow breathing. There are thirteen breaths in 30 s for fast breathing, as shown in Figure 7c, confirming that this is the fast-breathing case. All non-contact sensing results for normal, slow, and fast breathing match wearable sensor results, validating the proposed non-contact sensing system in this study.
Both subjects in the two-person scenario are asked to perform different breathing patterns, including normal, slow, and fast. This yields six distinct cases for the two-person scenario. Figure 8a shows the result for only one case for illustration purposes, in which one subject is doing normal breathing and the other is doing slow breathing. It is impossible to distinguish breathing patterns for each person from this Figure 8a because, in the multi-person scenario, variation in CSI is obtained due to independent chest movements of all persons doing breathing activity simultaneously.
For the three-person scenario, all three subjects are asked to perform different breathing patterns, including normal, slow, and fast. For the three-person scenario, this results in nine different cases of breathing. For illustration purposes, Figure 8b shows the outcome of a single case in which all subjects are breathing normally. Here again, it is impossible to distinguish each person’s breathing pattern from time-domain CSI amplitude information.

5.2. Breath Rate Extraction

For breathing rate extraction, frequency domain CSI amplitude information is used. The breathing rate extraction results for the single and multi-person scenarios are shown below:
For a single-person scenario, frequency domain CSI amplitude information is obtained by applying a Fourier transformation, which results in only one frequency peak f m a x . Then by using Equation (12), the breathing rate is calculated by multiplying the f m a x   value with 60 s. Figure 9a–c shows a single maximum frequency peak for the cases of normal, slow, and fast breathing rates. Furthermore, it can also be observed from Figure 9a–c that normal breathing rate has high f m a x   value than slow breathing rate but has less f m a x   value than fast breathing rate.
For the two-person scenario, six different cases are considered in which both persons are either doing breathing at the same or at a different rate. For each case, it can be seen from Figure 10a–f that two frequency peaks   f m a x     are obtained. Then again, using Equation (12), the individual breathing rate for each person can be calculated.
For a three-person scenario, only two cases are observed. In the first case, all three persons are requested to do the breathing at the same rate, i.e., normal, and in the second case, all three subjects are asked to do breathing at a different rate. It can be observed from Figure 11a,b that three frequency peaks   f m a x     are obtained, and using Equation (12), individual breathing rates for all three persons can be calculated.

5.3. Comparison with Wearable Sensor

To compare the performance of non-contact SDR-based sensing with the wearable sensor, mean square error (MSE) is calculated for single-person, two-person, and three-person scenarios. It is observed from Figure 12 that MSE for a single-person scenario is 0.27, while 0.36 for two person-scenario, and the three-person scenario MSE is increased to 0.42.

5.4. Breathing Patterns Classification

Different ML algorithms are used to classify various breathing patterns in single and multi-person scenarios. ML algorithms’ accuracy, prediction speed, and training time are used to evaluate their performance. The performance of all algorithms is first evaluated without using feature selection methods, and then various feature selection methods are used to improve performance.
Table 3 shows the results for a single-person scenario, and all three algorithms show a significant improvement in terms of accuracy, training time, and prediction speed. Similarly, feature selection methods for multi-person scenarios result in noticeable performance improvements in accuracy, training time, and prediction speed, as shown in Table 4. As shown in Table 3 and Table 4, feature selection methods improve accuracy and prediction speed while decreasing training time because fewer relevant features are available for the training ML model.

5.5. Comparison with Previous Approaches

This SDR-based research work is compared in terms of mean absolute error (MAE) with previous approaches for multiple-person scenarios in this section, as shown in Figure 13 [31] used CSI amplitude to calculate breathing rate and obtained MAE of 2.04 bpm. In contrast [33] measured the phase difference of CSI rather than the amplitude and attained an MAE of 1.33 bpm. Similarly [38] used CSI phase difference to obtain an MAE of 1.19 bpm. [36] used calibrated CSI to obtain an MAE of 0.71. Furthermore [49] uses multiple antennas to exploit CSI amplitude variation, and breathing rate is measured with an MAE of 0.42 bpm. In this study, MAE was calculated to be 0.42 bpm, comparable to [49], but the results were obtained using only a single pair of antennas.

6. Conclusions and Future Recommendations

This study presented the first SDR-based RF sensing system capable of detecting various breathing pattern abnormalities in multiple scenarios, even when they are close. In contrast to previous SDR work, the system can detect and classify various breathing patterns for single-person scenarios only. However, this research work was extended for multiple-person scenarios. Extensive testing confirmed that this system is reliable. However, system performance was further enhanced by applying optimal feature scoring a maximum accuracy of 93.8% and 91.7% for single-person and multi-person scenarios, respectively. Nevertheless, there are some limitations to this study. First, experiments were carried out in a static and controlled laboratory setting. The second limitation is that actual patients were not used in the research. The third limitation is that the experiments were limited to three persons. As a result, future recommendations would be to include breathing monitoring for more than three subjects in a dynamic environment. In future, this research work can be extended to identify the person having breathing abnormality in a multi-person scenario for timely medical interventions. Furthermore, data for actual patients will be collected to develop a realistic model.

Author Contributions

Conceptualization, M.R., R.A.S. and M.B.K.; methodology, M.R. and M.B.K.; software M.R., M.B.K.; validation, S.A.S., X.Y. and M.H.; formal analysis, S.A.S., X.Y. and M.H.; investigation, N.A.A.A., A.A. and M.H.; resources, N.A.A.A. and X.Y.; data curation, A.A.; writing—original draft, M.R. and M.B.K.; writing—review and editing, S.A.S. and Q.H.A.; visualization, A.A.; supervision, R.A.S. and N.A.A.A.; project administration, N.A.A.A. and M.A.I.; funding acquisition, N.A.A.A. and Q.H.A. All authors have read and agreed to the published version of the manuscript.

Funding

Zayed Health Center at UAE University supports this work in parts under Fund code G00003476, EPSRC grant numbers EP/T021063/1 and EP/T021020/1.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Ethics Committee of COMSATS University Islamabad, Attock Campus.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The datasets used and analyzed during the current experimental study are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Nandakumar, R.; Gollakota, S.; Watson, N. Contactless Sleep Apnea Detection on Smartphones. In Proceedings of the 13th Annual International Conference on Mobile Systems, Applications, and Services, Florence, Italy, 18–22 May 2015; pp. 45–57. [Google Scholar]
  2. Rehman, M.; Shah, R.A.; Khan, M.B.; Ali, N.A.A.; Alotaibi, A.A.; Althobaiti, T.; Ramzan, N.; Shah, S.A.; Yang, X.; Alomainy, A. Contactless Small-Scale Movement Monitoring System Using Software Defined Radio for Early Diagnosis of COVID-19. IEEE Sens. J. 2021, 21, 17180–17188. [Google Scholar] [CrossRef] [PubMed]
  3. Cretikos, M.A.; Bellomo, R.; Hillman, K.; Chen, J.; Finfer, S.; Flabouris, A. Respiratory Rate: The Neglected Vital Sign. Med. J. Aust. 2008, 188, 657–659. [Google Scholar] [CrossRef] [PubMed]
  4. Parkes, R. Rate of Respiration: The Forgotten Vital Sign. Emerg. Nurse 2011, 19, 12–17. [Google Scholar] [CrossRef] [PubMed]
  5. Barthel, P.; Wensel, R.; Bauer, A.; Müller, A.; Wolf, P.; Ulm, K.; Huster, K.M.; Francis, D.P.; Malik, M.; Schmidt, G. Respiratory Rate Predicts Outcome after Acute Myocardial Infarction: A Prospective Cohort Study. Eur. Heart J. 2013, 34, 1644–1650. [Google Scholar] [CrossRef]
  6. Helfenbein, E.; Firoozabadi, R.; Chien, S.; Carlson, E.; Babaeizadeh, S. Development of Three Methods for Extracting Respiration from the Surface ECG: A Review. J. Electrocardiol. 2014, 47, 819–825. [Google Scholar] [CrossRef]
  7. Gupta, K.; Prasad, A.; Nagappa, M.; Wong, J.; Abrahamyan, L.; Chung, F.F. Risk Factors for Opioid-Induced Respiratory Depression and Failure to Rescue: A Review. Curr. Opin. Anaesthesiol. 2018, 31, 110–119. [Google Scholar] [CrossRef]
  8. Rantonen, T.; Jalonen, J.; Grönlund, J.; Antila, K.; Southall, D.; Välimäki, I. Increased Amplitude Modulation of Continuous Respiration Precedes Sudden Infant Death Syndrome: Detection by Spectral Estimation of Respirogram. Early Hum. Dev. 1998, 53, 53–63. [Google Scholar] [CrossRef]
  9. Lovett, P.B.; Buchwald, J.M.; Stürmann, K.; Bijur, P. The Vexatious Vital: Neither Clinical Measurements by Nurses nor an Electronic Monitor Provides Accurate Measurements of Respiratory Rate in Triage. Ann. Emerg. Med. 2005, 45, 68–76. [Google Scholar] [CrossRef]
  10. Wiesner, S.; Yaniv, Z. Monitoring Patient Respiration Using a Single Optical Camera. In Proceedings of the 29th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Lyon, France, 22–26 August 2007; pp. 2740–2743. [Google Scholar]
  11. Massaroni, C.; Lopes, D.S.; Lo Presti, D.; Schena, E.; Silvestri, S. Contactless Monitoring of Breathing Patterns and Respiratory Rate at the Pit of the Neck: A Single Camera Approach. J. Sens. 2018, 2018, 4567213. [Google Scholar] [CrossRef]
  12. Wang, T.; Zhang, D.; Zheng, Y.; Gu, T.; Zhou, X.; Dorizzi, B. C-FMCW Based Contactless Respiration Detection Using Acoustic Signal. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 1, 1–20. [Google Scholar] [CrossRef]
  13. Yue, S.; He, H.; Wang, H.; Rahul, H.; Katabi, D. Extracting Multi-Person Respiration from Entangled RF Signals. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 2, 1–22. [Google Scholar] [CrossRef] [Green Version]
  14. Fan, D.; Ren, A.; Zhao, N.; Haider, D.; Yang, X.; Tian, J. Small-Scale Perception in Medical Body Area Networks. IEEE J. Transl. Eng. Health Med. 2019, 7, 2700211. [Google Scholar] [CrossRef]
  15. Qian, K.; Wu, C.; Yang, Z.; Liu, Y.; Jamieson, K. Widar: Decimeter-Level Passive Tracking via Velocity Monitoring with Commodity Wi-Fi. In Proceedings of the 18th ACM International Symposium on Mobile Ad Hoc Networking and Computing, Chennai, India, 10–14 July 2017; pp. 1–10. [Google Scholar]
  16. Al-Wahedi, A.; Al-Shams, M.; Albettar, M.A.; Alawsh, S.; Muqaibel, A. Wireless Monitoring of Respiration and Heart Rates Using Software-Defined-Radio. In Proceedings of the 2019 16th International Multi-Conference on Systems, Signals & Devices (SSD), Istanbul, Turkey, 21–24 March 2019; pp. 529–532. [Google Scholar]
  17. Praktika, T.O.; Pramudita, A.A. Implementation of Multi-Frequency Continuous Wave Radar for Respiration Detection Using Software Defined Radio. In Proceedings of the 2020 10th Electrical Power, Electronics, Communications, Controls and Informatics Seminar (EECCIS), Malang, Indonesia, 26–28 August 2020; pp. 284–287. [Google Scholar]
  18. Rehman, M.; Shah, R.A.; Khan, M.B.; Shah, S.A.; AbuAli, N.A.; Yang, X.; Alomainy, A.; Imran, M.A.; Abbasi, Q.H. Improving Machine Learning Classification Accuracy for Breathing Abnormalities by Enhancing Dataset. Sensors 2021, 21, 6750. [Google Scholar] [CrossRef]
  19. Walker, H.K.; Hall, W.D.; Hurst, J.W. Clinical Methods: The History, Physical, and Laboratory Examinations. JAMA 1990, 264, 2808–2809. [Google Scholar]
  20. Liu, C.; Xiong, J.; Cai, L.; Feng, L.; Chen, X.; Fang, D. Beyond Respiration: Contactless Sleep Sound-Activity Recognition Using RF Signals. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2019, 3, 1–22. [Google Scholar] [CrossRef]
  21. Adib, F.; Mao, H.; Kabelac, Z.; Katabi, D.; Miller, R.C. Smart Homes That Monitor Breathing and Heart Rate. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems, Soul, Republic of Korea, 18–23 April 2015; pp. 837–846. [Google Scholar]
  22. Islam, S.M.; Motoyama, N.; Pacheco, S.; Lubecke, V.M. Non-Contact Vital Signs Monitoring for Multiple Subjects Using a Millimeter Wave FMCW Automotive Radar. In Proceedings of the 2020 IEEE/MTT-S International Microwave Symposium (IMS), Los Angeles, CA, USA, 4–6 August 2020; pp. 783–786. [Google Scholar]
  23. Cardillo, E.; Li, C.; Caddemi, A. Empowering Blind People Mobility: A Millimeter-Wave Radar Cane. In Proceedings of the 2020 IEEE International Workshop on Metrology for Industry 4.0 & IoT, Online, 3–5 June 2020; pp. 213–217. [Google Scholar]
  24. Cardillo, E.; Caddemi, A. Radar Range-Breathing Separation for the Automatic Detection of Humans in Cluttered Environments. IEEE Sens. J. 2020, 21, 14043–14050. [Google Scholar] [CrossRef]
  25. Yue, S.; Yang, Y.; Wang, H.; Rahul, H.; Katabi, D. BodyCompass: Monitoring Sleep Posture with Wireless Signals. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2020, 4, 1–25. [Google Scholar] [CrossRef]
  26. Ultra-Wideband Radar Technology. Available online: https://www.routledge.com/Ultra-wideband-Radar-Technology/Taylor/p/book/9780849342677 (accessed on 1 December 2022).
  27. Li, C.; Lubecke, V.M.; Boric-Lubecke, O.; Lin, J. A Review on Recent Advances in Doppler Radar Sensors for Noncontact Healthcare Monitoring. IEEE Trans. Microw. Theory Tech. 2013, 61, 2046–2060. [Google Scholar] [CrossRef]
  28. Lai, J.C.Y.; Xu, Y.; Gunawan, E.; Chua, E.C.-P.; Maskooki, A.; Guan, Y.L.; Low, K.-S.; Soh, C.B.; Poh, C.-L. Wireless Sensing of Human Respiratory Parameters by Low-Power Ultrawideband Impulse Radio Radar. IEEE Trans. Instrum. Meas. 2010, 60, 928–938. [Google Scholar] [CrossRef]
  29. Liu, X.; Cao, J.; Tang, S.; Wen, J. Wi-Sleep: Contactless Sleep Monitoring via Wifi Signals. In Proceedings of the 2014 IEEE Real-Time Systems Symposium, Rome, Italy, 2–5 December 2014; pp. 346–355. [Google Scholar]
  30. Liu, X.; Cao, J.; Tang, S.; Wen, J.; Guo, P. Contactless Respiration Monitoring via Off-the-Shelf WiFi Devices. IEEE Trans. Mob. Comput. 2015, 15, 2466–2479. [Google Scholar] [CrossRef]
  31. Wang, H.; Zhang, D.; Ma, J.; Wang, Y.; Wang, Y.; Wu, D.; Gu, T.; Xie, B. Human Respiration Detection with Commodity Wifi Devices: Do User Location and Body Orientation Matter? In Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing, Heidelberg, Germany, 12–16 September 2016; pp. 25–36. [Google Scholar]
  32. Zeng, Y.; Wu, D.; Gao, R.; Gu, T.; Zhang, D. FullBreathe: Full Human Respiration Detection Exploiting Complementarity of CSI Phase and Amplitude of WiFi Signals. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 2, 1–19. [Google Scholar] [CrossRef]
  33. Zeng, Y.; Wu, D.; Xiong, J.; Yi, E.; Gao, R.; Zhang, D. FarSense: Pushing the Range Limit of WiFi-Based Respiration Sensing with CSI Ratio of Two Antennas. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2019, 3, 1–26. [Google Scholar] [CrossRef] [Green Version]
  34. Liu, J.; Wang, Y.; Chen, Y.; Yang, J.; Chen, X.; Cheng, J. Tracking Vital Signs during Sleep Leveraging Off-the-Shelf Wifi. In Proceedings of the 16th ACM International Symposium on Mobile ad hoc Networking and Computing, Hangzhou, China, 22–25 June 2015; pp. 267–276. [Google Scholar]
  35. Niu, K.; Zhang, F.; Chang, Z.; Zhang, D. A Fresnel Diffraction Model Based Human Respiration Detection System Using COTS Wi-Fi Devices. In Proceedings of the 2018 ACM International Joint Conference and 2018 International Symposium on Pervasive and Ubiquitous Computing and Wearable Computers, Singapore, 8–12 October 2018; pp. 416–419. [Google Scholar]
  36. Wang, X.; Yang, C.; Mao, S. PhaseBeat: Exploiting CSI Phase Data for Vital Sign Monitoring with Commodity WiFi Devices. In Proceedings of the 2017 IEEE 37th International Conference on Distributed Computing Systems (ICDCS), Atlanta, GA, USA, 5–8 June 2017; pp. 1230–1239. [Google Scholar]
  37. Rao, B.D.; Hari, K.S. Performance Analysis of Root-MUSIC. IEEE Trans. Acoust. Speech Signal Process. 1989, 37, 1939–1949. [Google Scholar] [CrossRef]
  38. Wang, X.; Yang, C.; Mao, S. TensorBeat: Tensor Decomposition for Monitoring Multiperson Breathing Beats with Commodity WiFi. ACM Trans. Intell. Syst. Technol. (TIST) 2017, 9, 1–27. [Google Scholar] [CrossRef]
  39. Chen, C.; Han, Y.; Chen, Y.; Lai, H.-Q.; Zhang, F.; Wang, B.; Liu, K.R. TR-BREATH: Time-Reversal Breathing Rate Estimation and Detection. IEEE Trans. Biomed. Eng. 2017, 65, 489–501. [Google Scholar] [CrossRef] [PubMed]
  40. Yang, Y.; Cao, J.; Liu, X.; Xing, K. Multi-Person Sleeping Respiration Monitoring with COTS WiFi Devices. In Proceedings of the 2018 IEEE 15th International Conference on Mobile Ad Hoc and Sensor Systems (MASS), Chengdu, China, 9–12 October 2018; pp. 37–45. [Google Scholar]
  41. Rehman, M.; Ali, N.A.A.; Shah, R.A.; Khan, M.B.; Shah, S.A.; Alomainy, A.; Yang, X.; Imran, M.A.; Abbasi, Q.H. Development of an Intelligent Real-Time Multiperson Respiratory Illnesses Sensing System Using SDR Technology. IEEE Sens. J. 2022, 22, 18858–18869. [Google Scholar] [CrossRef]
  42. Ashleibta, A.M.; Abbasi, Q.H.; Shah, S.A.; Khalid, M.A.; AbuAli, N.A.; Imran, M.A. Non-Invasive RF Sensing for Detecting Breathing Abnormalities Using Software Defined Radios. IEEE Sens. J. 2020, 21, 5111–5118. [Google Scholar] [CrossRef]
  43. Khan, M.B.; Rehman, M.; Mustafa, A.; Shah, R.A.; Yang, X. Intelligent Non-Contact Sensing for Connected Health Using Software Defined Radio Technology. Electronics 2021, 10, 1558. [Google Scholar] [CrossRef]
  44. Rehman, M.; Shah, R.A.; Khan, M.B.; AbuAli, N.A.; Shah, S.A.; Yang, X.; Alomainy, A.; Imran, M.A.; Abbasi, Q.H. RF Sensing Based Breathing Patterns Detection Leveraging USRP Devices. Sensors 2021, 21, 3855. [Google Scholar] [CrossRef]
  45. Marković, I.; Stojanović, M.; Stanković, J.; Stanković, M. Stock Market Trend Prediction Using AHP and Weighted Kernel LS-SVM. Soft Comput. 2017, 21, 5387–5398. [Google Scholar] [CrossRef]
  46. Guyon, I.; Elisseeff, A. An Introduction to Variable and Feature Selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. Available online: https://dl.acm.org/doi/10.5555/944919.944968 (accessed on 1 December 2022).
  47. Ding, C.; Peng, H. Minimum Redundancy Feature Selection from Microarray Gene Expression Data. J. Bioinform. Comput. Biol. 2005, 3, 185–205. [Google Scholar] [CrossRef] [PubMed]
  48. How to Choose a Feature Selection Method for Machine Learning—MachineLearningMastery.Com. Available online: https://machinelearningmastery.com/feature-selection-with-real-and-categorical-data/ (accessed on 1 December 2022).
  49. Zeng, Y.; Wu, D.; Xiong, J.; Liu, J.; Liu, Z.; Zhang, D. MultiSense: Enabling Multi-Person Respiration Sensing with Commodity Wifi. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2020, 4, 1–29. [Google Scholar] [CrossRef]
Figure 1. Design methodology.
Figure 1. Design methodology.
Sensors 23 01251 g001
Figure 2. OFDM transceiver for wireless signal sensing.
Figure 2. OFDM transceiver for wireless signal sensing.
Sensors 23 01251 g002
Figure 3. Data preprocessing: (a) subcarrier selection; (b) outlier removal; (c) smoothening; and (d) normalization.
Figure 3. Data preprocessing: (a) subcarrier selection; (b) outlier removal; (c) smoothening; and (d) normalization.
Sensors 23 01251 g003
Figure 4. Boxplots to check similarity between all features.
Figure 4. Boxplots to check similarity between all features.
Sensors 23 01251 g004
Figure 5. Features importance score to select most relevant features only.
Figure 5. Features importance score to select most relevant features only.
Sensors 23 01251 g005
Figure 6. Experimental setup for single-person and multi-person breathing monitoring.
Figure 6. Experimental setup for single-person and multi-person breathing monitoring.
Sensors 23 01251 g006
Figure 7. Breathing patterns detection for single person-scenario (non-contact sensor (Red) vs. Wearable sensor (Black): (a) normal breathing; (b) slow breathing; and (c) fast breathing.
Figure 7. Breathing patterns detection for single person-scenario (non-contact sensor (Red) vs. Wearable sensor (Black): (a) normal breathing; (b) slow breathing; and (c) fast breathing.
Sensors 23 01251 g007
Figure 8. Breathing patterns detection for two-person and three persons-scenarios: (a) two persons-scenario (normal-slow); and (b) three persons scenario (normal-normal-normal).
Figure 8. Breathing patterns detection for two-person and three persons-scenarios: (a) two persons-scenario (normal-slow); and (b) three persons scenario (normal-normal-normal).
Sensors 23 01251 g008
Figure 9. Breathing rate extraction for one person-scenario: (a) normal breathing; (b) slow breathing; and (c) fast breathing.
Figure 9. Breathing rate extraction for one person-scenario: (a) normal breathing; (b) slow breathing; and (c) fast breathing.
Sensors 23 01251 g009
Figure 10. Breathing rate extraction for two persons-scenario: (a) normal-normal; (b) slow-slow; (c) fast-fast; (d) normal-slow; (e) normal-fast; and (f) slow-fast.
Figure 10. Breathing rate extraction for two persons-scenario: (a) normal-normal; (b) slow-slow; (c) fast-fast; (d) normal-slow; (e) normal-fast; and (f) slow-fast.
Sensors 23 01251 g010aSensors 23 01251 g010b
Figure 11. Breathing rate extraction for three persons-scenario: (a) normal–normal–normal; and (b) normal-slow-fast.
Figure 11. Breathing rate extraction for three persons-scenario: (a) normal–normal–normal; and (b) normal-slow-fast.
Sensors 23 01251 g011
Figure 12. Mean absolute error vs. number of persons.
Figure 12. Mean absolute error vs. number of persons.
Sensors 23 01251 g012
Figure 13. Comparison with previous approaches [34,36,38,39,49].
Figure 13. Comparison with previous approaches [34,36,38,39,49].
Sensors 23 01251 g013
Table 1. Statistical Features.
Table 1. Statistical Features.
Sr. #Statistical FeaturesDetailEquation
1.MeanData mean y ¯ = i = 1 n y i n
2.Standard deviationData dispersion relative to mean y S D = 1 n i = 1 n ( y i y ¯ ) 2
3.Peak-to-peakMax. to min. value difference y p p = y m a x y m i n
4.RMSRoot mean square y R M S = 1 n i = 1 n y i 2 2
5.KurtosisFrequency peaks distribution y K u r = 1 n i = 1 n ( | y i | y ¯ ) 4 y S D 4
6.SkewnessSymmetry in data distribution y S k e w = 1 n i = 1 n ( | y i | y ¯ ) 3 y S D 3
7.Shape FactorSquare root of variance y S F = y R M S 1 n i = 1 n | y i |
8.Crest FactorPeak height value to RMS value y C F = m a x i | y i | y R M S
9.Impulse FactorPeak height value to mean value y I F = m a x i | y i | y ¯
10.EntropyMeasure of randomness of data y E n t = i = n n h i s t ( y i )   l n 2 ( h i s t ( y i ) )
Table 2. Participants detail.
Table 2. Participants detail.
Sr. #Age (Y)Height (In)Weight (Kg)BMI
1.24687023.5
2.26687625.5
3.28706520.6
4.31695216.9
5.31705116.1
6.31686521.8
7.32708326.3
8.33619137.9
9.35628835.5
10.37688428.2
Table 3. Single-person scenario.
Table 3. Single-person scenario.
ML AlgorithmsParametersWithout Feature SelectionUsing MRMR AlgorithmUsing PCA
Fine Gaussian SVMAccuracy (%)92.793.293.7
Training Time (s)43.5340.94941.04
Prediction Speed (obs/s)~49,000~7400~3700
Medium KNNAccuracy (%)89.892.792.3
Training Time (s)81.46164.5562.086
Prediction Speed (obs/s)~17,000~48,000~12,000
Wide Neural NetworkAccuracy (%)91.793.893.6
Training Time (s)392.32324.43329.13
Prediction Speed (obs/s)~99,000~260,000~82,000
Table 4. Multi-persons scenario.
Table 4. Multi-persons scenario.
ML AlgorithmsParametersWithout Feature SelectionUsing MRMR AlgorithmUsing PCA
Fine Gaussian SVMAccuracy (%)92.793.293.7
Training Time (s)43.5340.94941.04
Prediction Speed (obs/s)~49,000~7400~3700
Medium KNNAccuracy (%)89.892.792.3
Training Time (s)81.46164.5562.086
Prediction Speed (obs/s)~17,000~48,000~12,000
Wide Neural NetworkAccuracy (%)91.793.893.6
Training Time (s)392.32324.43329.13
Prediction Speed (obs/s)~99,000~260,000~82,000
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rehman, M.; Shah, R.A.; Ali, N.A.A.; Khan, M.B.; Shah, S.A.; Alomainy, A.; Hayajneh, M.; Yang, X.; Imran, M.A.; Abbasi, Q.H. Enhancing System Performance through Objective Feature Scoring of Multiple Persons’ Breathing Using Non-Contact RF Approach. Sensors 2023, 23, 1251. https://doi.org/10.3390/s23031251

AMA Style

Rehman M, Shah RA, Ali NAA, Khan MB, Shah SA, Alomainy A, Hayajneh M, Yang X, Imran MA, Abbasi QH. Enhancing System Performance through Objective Feature Scoring of Multiple Persons’ Breathing Using Non-Contact RF Approach. Sensors. 2023; 23(3):1251. https://doi.org/10.3390/s23031251

Chicago/Turabian Style

Rehman, Mubashir, Raza Ali Shah, Najah Abed Abu Ali, Muhammad Bilal Khan, Syed Aziz Shah, Akram Alomainy, Mohammad Hayajneh, Xiaodong Yang, Muhammad Ali Imran, and Qammer H. Abbasi. 2023. "Enhancing System Performance through Objective Feature Scoring of Multiple Persons’ Breathing Using Non-Contact RF Approach" Sensors 23, no. 3: 1251. https://doi.org/10.3390/s23031251

APA Style

Rehman, M., Shah, R. A., Ali, N. A. A., Khan, M. B., Shah, S. A., Alomainy, A., Hayajneh, M., Yang, X., Imran, M. A., & Abbasi, Q. H. (2023). Enhancing System Performance through Objective Feature Scoring of Multiple Persons’ Breathing Using Non-Contact RF Approach. Sensors, 23(3), 1251. https://doi.org/10.3390/s23031251

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop