Next Article in Journal
A Heterogeneity-Aware Semi-Decentralized Model for a Lightweight Intrusion Detection System for IoT Networks Based on Federated Learning and BiLSTM
Previous Article in Journal
Measurement and Decoupling of Hygrothermal-Mechanical Effects with Optical Fibers: Development of a New Fiber Bragg Grating Sensor
Previous Article in Special Issue
Current Diagnostic Techniques for Pneumonia: A Scoping Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced Human Activity Recognition Using Wi-Fi Sensing: Leveraging Phase and Amplitude with Attention Mechanisms

1
Department of Computer Science and Information Engineering, National Central University, Taoyuan City 320317, Taiwan
2
Department of Mechanical Engineering, National Central University, Taoyuan City 320317, Taiwan
*
Authors to whom correspondence should be addressed.
Sensors 2025, 25(4), 1038; https://doi.org/10.3390/s25041038
Submission received: 30 December 2024 / Revised: 26 January 2025 / Accepted: 7 February 2025 / Published: 9 February 2025
(This article belongs to the Special Issue Advancing Healthcare: Integrating AI and Smart Sensing Technologies)

Abstract

:
Wi-Fi-based human activity recognition (HAR) is a non-intrusive and privacy-preserving method that leverages Channel State Information (CSI) for identifying human activities. However, existing approaches often struggle with robust feature extraction, especially in dynamic and multi-environment scenarios, and fail to effectively integrate amplitude and phase features of CSI. This study proposes a novel model, the Phase–Amplitude Channel State Information Network (PA-CSI), to address these challenges. The model introduces two key innovations: (1) a dual-feature approach combining amplitude and phase features for enhanced robustness, and (2) an attention-enhanced feature fusion mechanism incorporating multi-scale convolutional layers and Gated Residual Networks (GRN) to optimize feature extraction. Experimental results demonstrate that the proposed model achieves state-of-the-art performance on three datasets, including StanWiFi (99.9%), MultiEnv (98.0%), and the MINE lab dataset (99.9%). These findings underscore the potential of the PA-CSI model to advance Wi-Fi-based HAR in real-world applications.

1. Introduction

Human activity recognition (HAR) [1] is a significant area of research in computer vision and pattern recognition that focuses on identifying and classifying human activities from multimedia environments. HAR can be achieved using various technologies, including camera-based [2], sensor-based [1], and radar-based technologies [3], as well as Wi-Fi [1,4]. Camera-based HAR relies on video data to analyze and classify actions. While it offers high-resolution and detailed visual information, it raises privacy concerns and is sensitive to lighting conditions and occlusions. Sensor-based HAR utilizes wearable devices like accelerometers and gyroscopes to capture movement data, providing accuracy and robustness but requiring user compliance, while also potentially being intrusive [1,5]. Radar-based HAR uses radio waves to detect motion, offering privacy advantages and working well in low-light conditions, but it can be affected by interference and requires complex signal processing. In contrast, Wi-Fi-based HAR leverages the analysis of CSI to recognize actions by detecting changes in signal propagation caused by human movement. This method is advantageous for its non-intrusiveness, cost-effectiveness, and ability to work through walls and in various lighting conditions. However, Wi-Fi-based HAR may face challenges in environments with high signal noise and requires sophisticated algorithms to interpret the data accurately [6,7]. Wi-Fi-based HAR stands out due to its balance of privacy, convenience, and practicality in real-world applications.
In Wi-Fi sensing, two critical metrics are commonly used to represent the characteristics of a Wi-Fi signal: the Received Signal Strength Indicator (RSSI) and Channel State Information (CSI) [5,8]. These signals can propagate and synthesize through either line-of-sight (LOS) or non-line-of-sight (NLOS) scenarios (Figure 1). RSSI is a measure of the power level that a receiver detects from a transmitted signal, often used to estimate the distance between the transmitter and receiver. It provides a coarse measure of signal strength but lacks detailed information about the multipath effects and the fine-grained behavior of the signal [5]. On the other hand, CSI offers a more comprehensive representation by capturing the amplitude and phase of the signal across multiple subcarriers within a Wi-Fi channel. This allows CSI to provide detailed insights into the propagation environment, including reflections, scattering, and the impact of obstacles [8]. While RSSI is easier to obtain and requires less computational processing, CSI is preferred in advanced Wi-Fi sensing applications, such as human activity recognition and indoor localization, due to its ability to capture rich spatial and temporal information about the signal’s interactions with the environment.
Channel State Information (CSI) comprises multiple extractable features that reveal characteristics of the wireless environment, such as Time of Flight (ToF), amplitude, phase, Direction of Arrival (DoA), Angle of Arrival (AoA), and phase shift [8]. Each feature provides unique information; for example, ToF enables precise distance estimation, while DoA and AoA are critical for identifying the signal’s spatial direction. In human activity recognition (HAR) applications, amplitude and phase are two particularly important features. Amplitude reflects the magnitude of signal variations, whereas phase provides detailed insights into signal propagation, including phase shifts caused by reflections and scattering. However, Wi-Fi sensing research has largely focused on the amplitude component of CSI, often neglecting phase information in the received signal. Additionally, CSI data are represented as a series of time-varying values across subcarriers. Most recognition models, however, concentrate on the temporal dimension and tend to overlook inter-subcarrier relationships, which could capture variations in body posture or actions, such as standing up and sitting down.
To address these gaps, this study seeks to utilize relevant features in conjunction with attention mechanisms across both temporal and channel dimensions, enhancing the accuracy and robustness of Wi-Fi-based applications for human activity recognition (HAR). In summary, the primary contributions of this paper to Wi-Fi sensing for HAR are as follows:
  • We propose an attention-based model that effectively utilizes both phase and amplitude components to improve HAR performance. While prior studies primarily focused on amplitude, our approach incorporates phase data, which provides complementary insights into human activity. By optimizing feature extraction for both amplitude and phase, our method significantly enhances recognition accuracy and robustness across a wide range of activities and environments.
  • We introduce an attention-based feature fusion mechanism that integrates spatial and temporal features. This includes the implementation of multi-scale convolutional layers, enabling the network to efficiently capture both local and global patterns. In addition, a Gated Residual Network (GRN) is incorporated into our framework as part of the feature fusion process. GRNs can selectively retain or discard information, improving learning efficiency and reducing unnecessary complexity. This adaptation enhances classification accuracy and addresses the limitations of traditional networks. Our focus lies in how the GRN is specifically adapted and utilized in conjunction with attention mechanisms to maximize the utility of multi-scale features.
  • Our model is rigorously evaluated using three datasets, including two publicly available datasets and a custom dataset collected for this study. The results demonstrate that our approach achieves superior accuracy and performance compared to existing state-of-the-art (SOTA) models.

2. Literature Review

The Received Signal Strength Indicator (RSSI), a metric used to measure transmission channel power, has been extensively studied and applied across various domains, including human activity recognition (HAR) [9,10,11,12,13,14]. Despite its utility, RSSI’s efficiency is constrained. In contrast, Channel State Information (CSI) offers a more comprehensive and detailed representation, providing a four-dimensional matrix that captures the characteristics of transmission and reception channels, subcarriers, and temporal packets. Due to its complexity, numerous studies have explored the application of CSI in HAR using advanced deep learning techniques, such as Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTM) networks, and Transformer models.
As interest in Wi-Fi CSI methods continues to grow, numerous datasets containing CSI data have been developed for diverse applications. These datasets have become foundational for many researchers, serving as the basis for experimental studies documented in a wide range of published literature. Table 1 provides an overview of commonly used public datasets and their descriptions, facilitating an understanding of their utility and scope.

2.1. Wi-Fi CSI Based on CNN and LSTM Approaches

The intricate features of CSI data, including phase and amplitude, have been effectively utilized in CNN models, yielding significant advancements. Wang et al. [21] introduced a CSI-based human activity recognition and monitoring system (CARM), a CSI-based method for HAR and monitoring. This model integrated two modules: the CSI-speed module, which captured the relationship between CSI dynamics and human motion, and the CSI-activity module, which correlated movement speed with activity. By leveraging the phase component of CSI, the model was tested on their dataset and achieved 96% accuracy in HAR while demonstrating robustness across different environments. Although the method was accurate, cost-effective, and privacy-preserving when using standard Wi-Fi devices, it was sensitive to noise, impacted Wi-Fi performance, and required extensive training.
Similarly, Alsaify et al. [15] proposed the MultiEnv dataset, evaluating system performance in office and hall environments under line-of-sight (LOS) and non-line-of-sight (NLOS) conditions. Their robust five-stage preprocessing and feature extraction pipeline achieved 91.27% accuracy on their dataset [15,22]. However, environmental factors such as distance and interference, as well as similar movement patterns (e.g., falling versus sitting), led to higher misclassification rates.
While CNN-based models demonstrated promising results, they lacked the ability to exploit the temporal characteristics of CSI data. Consequently, temporal-based approaches utilizing LSTM networks have gained attention. For example, Chen et al. [23] developed the ABLSTM model (attention-based bidirectional LSTM), which was designed to extract representative temporal features from sequential CSI data in both directions. The ABLSTM framework achieved state-of-the-art accuracy (≥95%) in Wi-Fi CSI-based HAR, evaluated on the StanWiFi dataset and their own dataset, by integrating bidirectional LSTM and attention mechanisms. However, this model required significant computational resources, faced challenges in cross-environment scenarios, and did not address multi-user activity recognition.
Yadav et al. [24] proposed CSITime, an enhanced version of the InceptionTime network, which achieved state-of-the-art accuracies across three public datasets: 98.2% on ARIL, 98.88% on StanWiFi, and 99.09% on SignFi. CSITime incorporated data augmentation and advanced optimizations within a streamlined InceptionTime-based architecture. However, the model faced challenges in high-interference environments and multi-user scenarios, while demanding high computational resources.
To address spatial and temporal aspects of CSI data, several studies combined CNN and LSTM architectures. For instance, Moshiri et al. [25] employed Raspberry Pi devices to extract amplitude CSI signals for seven activities, converting them into 2D images using pseudo-color plots. Among four evaluated models (1D-CNN, 2D-CNN, LSTM, and bidirectional LSTM), the 2D-CNN achieved 95% accuracy on their collected data. Similarly, Salehinejad and Valaee [26] introduced LiteHAR, which uses randomly initialized convolution kernels for feature extraction, achieving 93% accuracy on the StanWiFi dataset. Shalaby et al. [27] compared deep learning models on the StanWiFi dataset, with a CNN-GRU model achieving 99.31% accuracy and its attention-based variant reaching 99.16%. Recently, Islam et al. [28] proposed the STC-NLSTMNet model (spatio-temporal convolution with nested long short-term memory), which integrated spatial–temporal convolution and nested LSTMs, achieved 99.88% and 98.20% accuracy on public datasets (StanWiFi and MultiEnv). This model was a robust and efficient method for HAR, achieving SOTA performance by integrating spatial and temporal features. However, the model had limitations in NLOS environments and multi-user settings.

2.2. Attention-Based Approaches

Traditional models such as CNNs and LSTMs were limited in capturing interdependencies within the same data dimension. In recent years, Transformer models have gained prominence due to their multi-head attention mechanism [29]. Researchers have applied these models to Wi-Fi-based HAR systems with notable success.
Ding et al. (2022) introduced a Channel–Time–Subcarrier Attention Mechanism (CTS-AM) to enhance location-independent HAR. The model CTS-AM demonstrated innovation in feature extraction, resource efficiency, and flexibility, outperforming existing methods like CNN and Wi-Hand. However, the system was limited to single-actor scenarios, assumed consistent motion during data collection, and recognized only predefined activities. This model achieved more than 90% average accuracy on their dataset across various locations with limited training samples [30]. Yang et al. (2023) [31] proposed WiTransformer, which adapted two Transformer architectures—the United Spatiotemporal Transformer (UST) and the Separated Spatiotemporal Transformer (SST)—to improve recognition accuracy and robustness in complex environments [31]. The UST integrated spatial and temporal features through early fusion in a single transformer encoder, resulting in high accuracy, computational efficiency, and robustness across tasks of varying complexity. In contrast, the SST processed spatial and temporal features separately using two transformer encoders, offering flexibility, but at the cost of increased complexity and resource demands. The SST struggled with generalization on smaller datasets, exhibited sensitivity to imbalanced data, and lacked the inductive biases needed for effective cross-domain adaptation. The UST’s efficient and unified approach made it more practical and robust compared to the SST, particularly for real-world applications. They experimented with two models on the Widar 3.0 dataset and achieved an overall recognition accuracy of 86.16%.
Some studies also explored two-way relationships within CSI data. Li et al. (2021) [32] developed the Two-stream Convolution Augmented Human Activity Transformer (THAT) model, which uses a dual convolution-augmented HAR layer to capture channel and temporal structures. This model effectively accommodated variations in activity speed and blank intervals using Gaussian encoding. Yang et al. (2023) [19] applied Vision Transformers (ViT) to Wi-Fi sensing tasks, leveraging spatial and temporal features to handle complex data relationships. However, ViT models required substantial training data for optimal performance.
Although a few studies have utilized phase information for HAR, the majority of existing works predominantly focus on amplitude features, leaving phase information underexplored or insufficiently integrated with amplitude. While existing attention-based methods primarily focused on either spatial or temporal dimensions, our proposed PA-CSI model incorporates attention across both. By leveraging multi-head attention mechanisms and multi-scale convolutional neural networks, the PA-CSI model effectively integrates amplitude and phase features to achieve superior performance. Moreover, the use of Gaussian Range Encoding (GRE) ensures robust temporal feature representation, addressing limitations observed in CTS-AM and THAT models.

3. Materials and Methods

3.1. Channel State Information (CSI)

Channel State Information (CSI) refers to the detailed properties of a communication channel in wireless communication systems. CSI encompasses information about the channel’s characteristics, including path loss, fading, delay spread, and interference [33]. CSI is typically obtained through channel estimation techniques, which involve measuring the channel response to known pilot signals transmitted between the sender and receiver. Mathematically, CSI can be represented by the channel matrix H , which characterizes the effect of the channel on the transmitted signal. If x is the transmitted signal and y is the received signal, N is the total number of subcarriers, and η represents the noise in the system, the relationship between x and y for each i-th subcarrier can be expressed as follows (1):
y i = H i x i + η ,   i = 1 ,   2 , , N
The matrix H i presents the relations from transmitter (T) and receiver (R) as (2):
H i = h 11 h 1 R h 1 T h T R i ,
where each h r t contains complex values, represented as shown in (3):
h r t = h r t + h r t j ,   j 2 = 1
Each hrt value consists of both the real h r t and imaginary h r t parts, which encapsulate the amplitude and phase changes imposed by the channel. The amplitude ( A r t ) and phase ( ϕ r t ) are calculated as given in Equations (4) and (5):
A r t = h r t 2 + h r t 2 ,
ϕ r t = a r c t a n h r t h r t ,
An example visualization of falling and sitting down activities, represented by the averaged amplitude and phase values of Wi-Fi Channel State Information (CSI), is shown in Figure 2.

3.2. Datasets

In our experiment, we selected and evaluated three datasets: MultiEnv (three environments) [15], StanWiFi [16], and a dataset collected by our research team. The rationale for selecting these datasets was their inclusion of both CSI phase and amplitude values, which are critical for the model’s ability to effectively utilize these inputs.

3.2.1. StanWiFi

The StanWiFi dataset was collected by Yousefi et al. (2017) [16] in a line-of-sight (LOS) environment. The setup included a single Wi-Fi router with one antenna as the transmitter and three antennas as the receiver, installed on a laptop equipped with an Intel® 5300 NIC. The transmitter and receiver were three meters apart. Each session, recorded at a sampling rate of 1000 Hz, lasted 20 s. This dataset consists of six activities—fall, run, lie down, walk, sit down, and stand up—performed by six participants, each repeating them 20 times. In total, the dataset comprises 577 samples.

3.2.2. Multiple Environment (MultiEnv)

The MultiEnv dataset was collected by Alsaify et al. (2020) [15] across three indoor environments: an office and a hall for line-of-sight (LOS) scenarios, and a non-line-of-sight (NLOS) scenario with a wooden barrier separating the transmitter and receiver. The distance between the transmitter and receiver was set at 3.7 m, with an 8-centimeter wall as the barrier. Data were collected using two computers: one serving as a single-antenna transmitter and the other as a three-antenna receiver, both equipped with an Intel® 5300 NIC. As listed in Table 2, the dataset includes six activity classes, with a total of 3000 samples (30 subjects × 5 experiments × 20 trials).

3.2.3. Our Research Team Dataset (MINE Lab Dataset)

This dataset was collected in an office (LOS) within our laboratory (5.5 × 3 m). The setup utilized Intel® 5300 NIC devices, each equipped with three antennas for both the transmitter and receiver, placed 4.0 m apart (Figure 3). The experiment involved five participants performing six activities: standing up and squatting, raising and lowering the right hand, opening and closing the arms, kicking the right and the left leg. In total, the dataset comprises 216 collected samples.

3.3. Methods

Figure 4 illustrates a pipeline of the proposed HAR system in this paper. It begins with raw CSI data, which undergoes preprocessing steps including Kalman filtering [34], sliding windows (optional), time alignment, and feature extraction. After preprocessing, the CSI data are extracted into two components: amplitude and phase. The amplitude data are normalized, while the phase data are unwrapped, and both features are passed through four Multi-scale Convolution Augmented Transformer (MCAT) layers [33]. Each MCAT layer processes data along two streams: the data are kept in their original format for the temporal stream, while it is transposed to feed into the channel stream. Gaussian Range Encoding is applied to preserve temporal order. The outputs of the MCAT layers are combined using a CNN with max-pooling layers. The combined outputs are then subsequently fed into a Gated Residual Network (GRN) [35,36], which includes dense layers with exponential linear unit (ELU) activations, layer normalization, and a final sigmoid activation. The GRN integrates the processed amplitude and phase features to classify human actions. This system effectively enables the classification of human activities from CSI Wi-Fi signals as they interact with their environment.

3.3.1. Preprocessing: Kalman Filter

The Kalman filter is a recursive algorithm that efficiently estimates the state of a dynamic system from a series of noisy measurements [34], to minimize η, as described in Equation (1). It operates in two main phases: prediction and update. During the prediction phase, the filter uses the system’s previous state and a mathematical model to predict the next state, while also estimating the associated uncertainty. In the update phase, it updates the estimate by incorporating new measurements and reducing uncertainty based on the discrepancy between the predicted and observed data. The Kalman filter assumes that the errors in both the system and the measurements are normally distributed, following a Gaussian distribution, making it optimal for linear systems affected by Gaussian noise.

3.3.2. Preprocessing: Sliding Windows

In this phase, appropriate preprocessing techniques are applied to the collected CSI data depending on the characteristics of each dataset. For both the StanWiFi dataset and our own dataset, we use a sliding window approach to generate additional samples, which is also essential for real-time action prediction when the exact point at which the action will occur within the signal is unknown (Figure 5). Specifically, each instance is shifted by 150, and a window size of 1000 is used for each sample. As a result, the data augmentation process leads to an expansion of the StanWiFi dataset from 577 records to 3119 records, while our dataset increases from 216 samples to 2111 samples.

3.3.3. Preprocessing: Time Alignment

In the datasets, CSI signal sample lengths vary, and the signals exhibit high density with similar characteristics. To address this issue, we propose a time alignment algorithm designed to standardize the sample sizes using collection times and average signal values. This algorithm processes CSI data based on timestamps, normalizing them to a specified maximum length. For example, in the StanWiFi dataset, sample lengths include 19,995, 19,980, 19,990, and so on. By setting the maximum length to 2000, the algorithm reduces all signal lengths to 2000 (Figure 6).

3.3.4. Preprocessing: Feature Extraction

After the Wi-Fi signal is extracted using the sliding window method or through the time alignment technique, the Channel State Information (CSI) data are decomposed into amplitude and phase components as described in Equations (4) and (5).

3.3.5. Normalization

CSI data collected from various environments or devices often exhibit variations in size. Normalization ensures that all data samples are standardized to a uniform size, to ensure compatibility with models requiring fixed input dimensions. In this context, normalization is applied to the amplitude across the datasets using the widely utilized min–max normalization technique for scaling data within a specific range. The formula for min–max normalization is shown in Equation (6).
x = x min x max x min x ,
where x represents the original data point, min(x) is the minimum value in the dataset, max(x) is the maximum value in the dataset, and x′ is the normalized value of x.

3.3.6. Phase Unwrapping

The CSI phase provides critical information about wireless signal propagation, including path delays, multipath effects, and Doppler shifts. However, the phase information derived from CSI is often wrapped, meaning it is confined to a specific range, such as [−π, π] or [0, 2π]. This wrapping occurs because phase is typically expressed as an angle, and angles exceeding these bounds are wrapped back into the range.
Phase unwrapping is the process of resolving these phase ambiguities to create a continuous phase representation. This is achieved by identifying and correcting discontinuities introduced by wrapping, ensuring smooth phase progression over time or frequency. The visualization of phase unwrapping for falling action is illustrated in Figure 7, while the formula for phase unwrapping is provided in Equation (7).
Φ ^ i = Φ i + 2 π Φ ^ i 1 Φ i 2 π ,
where Φ ^ i ,   Φ i are the unwrapped and raw phases at index i, respectively; Φ ^ i 1 is the previously unwrapped phase and the floor function denotes rounding to the nearest integer.

3.3.7. Gaussian Range Encoding

Gaussian Range Encoding (GRE) is introduced in [32,37] as a mechanism to preserve order information from CSI data, which is crucial for recognizing sequential activities such as “sitting down” or “standing up”. Unlike traditional positional encoding methods, which assign unique encodings to individual time steps, GRE focuses on encoding ranges. This approach preserves temporal continuity and improves robustness against subtle variations in activity speeds or blank intervals within the sequence. The final representation of GRE is obtained by multiplying the K, learnable range of vector embeddings, with the probability density function (PDF) vector. The PDF is normalized across K Gaussian distributions. Figure 8 shows how data are mapped into Gaussian distributions and how these distributions are transformed into learnable embeddings for integration into the neural network.

3.3.8. Multi-Scale Convolution Augmented Transformer (MCAT) Layer

The MCAT layer [32] integrates two key components: a multi-head self-attention mechanism [29] and a multi-scale convolutional neural network. These components are connected using residual connections and layer normalization [38], as shown in Figure 9.
In the multi-head attention stage, the input X   L × d i n , representing amplitude or phase—is transformed into three distinct vectors: query (Q), key (K), and value (V)—via three linear projections. Self-attention computes a weighted sum of the input values, where the weights are determined by the dot product between the query and the corresponding key. The process is defined mathematically as follows:
Q = X W Q ,       W Q   d i n × d k
K = X W K ,         W K   d i n × d k
V = X W V ,         W v   d i n × d v
A t t e n t i o n   Q , K , V = s o f t m a x Q K T d k V ,
here, d k represents the dimensionality of both the queries and keys, d k is the scaling factor used to stabilize gradients and enhance computational efficiency.
To jointly attend to information from different representation subspaces, the vectors Q, K, and V are projected h times (referred to as h-heads) using distinct projection parameters. The outputs from these projections are then concatenated and passed through a final projection to generate the output:
M u l t i H e a d   Q , K , V = h e a d 1 ; h e a d 2 ; ; h e a d h W O ,
where each h e a d i = A t t e n t i o n X W Q i ,   X W K i , X W V i and W O h d v × d 0 is the final projection matrix.
Subsequently, the multi-scale CNN stage processes the output features Y from the multi-head self-attention module by applying learnable weights W j i   d 0 × j i × d 0 . Here, Wij consists of do filters, where each filter sized ji × do. The output of the multi-cale CNN is presented as P = {P1,P2,…,Pj}, where P i   L × d 0 is the feature matrix corresponding to the kernel size ji at index i. Pi is computed as follows:
P i = R e L U ( D r o p o u t B N C o n v W j i ; Y ,
Equation (12) involves a series of operations, including convolution, batch normalization (BN), dropout, and a rectified linear unit (ReLU) activation function.

3.3.9. Gated Residual Network (GRN)

Gated Residual Networks (GRNs) [35] are a type of neural network architecture designed to improve information flow via residual connections while incorporating a gating mechanism to regulate the flow of features. The gating mechanism, typically implemented using a sigmoid function, selectively controls which parts of the input are passed through, allowing the network to prioritize the most relevant features. This combination of residual connections and gating mechanisms enables a GRN to more effectively capture complex patterns and improve model performance, especially in tasks that involve high-dimensional data or deep architectures. GRNs have been widely applied in various domains, including time series forecasting [36] and sequence modeling [39], which require learning long-term dependencies.
The structure of the GRN in our model is illustrated in Figure 10. This figure demonstrates how amplitude and phase data are processed through dense layers with exponential linear unit (ELU) activation functions, after which their outputs are summed using a residual connection (denoted by the summation symbol “+”). The resulting feature matrix undergoes layer normalization before being passed through another dense layer and a sigmoid gating mechanism. The final output is computed by integrating the gated outputs through an additional summation operation. The exponential linear unit (ELU) activation functions are defined as shown in Equation (14).
E L U x = x , i f   x > 0 α exp x 1 , a n d   i f   x 0 ,

4. Experimental Evaluation

In this section, we present the experimental results for the three datasets introduced in Section 3.2 (MultiEnv, StanWiFi, and the MINE lab dataset). The evaluation metrics employed include accuracy (Acc), precision (Pre), recall, and F1-score. Additionally, we provide a comparative analysis of our results against the baseline and the state-of-the-art model.

4.1. Hyperprameters

In this subsection, we provide a concise summary of the symbols for the variables and hyperparameters utilized in the experiment, and their corresponding values, chosen to optimize performance (Table 3).
The StanWiFi and MultiEnv datasets were divided into training and validation sets using an 80:20 split ratio, with shuffling applied to ensure randomness. This approach aligns with common practices in the literature for utilizing the StanWiFi and MultiEnv datasets, as referenced in previous studies such as [16,22,24,26,27,31]. For the MINE lab dataset, a 5-fold cross-validation approach was employed to ensure robust evaluation, given its smaller size. The dataset was divided into five equal parts, with each fold serving as the test set once, while the remaining folds were used for training and validation.
The loss function deployed in this experiment is sparse categorical cross-entropy, which is defined in Equation (15):
L o s s y , y ^ = 1 N i = 1 N log y ^ i , y i ,
where N represents the number of samples, yi is the true label, and y ^ i , y i is the predicted probability for the true class yi for the i-th example.

4.2. Experimental Results on StanWiFi and MINE Lab Datasets

Table 4 compares models on the StanWiFi dataset based on accuracy, precision, recall, and F1-score. The models are listed chronologically, with the earliest model being THAT (2021) and CSITime (2022), both of which achieved over 98% accuracy but did not report precision, recall, or F1-score. Several other models, such as LiteHAR (2022) and CNN-GRU (2022), report high precision and recall, though not all models provide complete metric details. Notably, our model achieves the highest reported metrics: 99.93% accuracy, 99.86% precision, 99.95% recall, and 99.95% F1-score. This outperforms all other models in the table, demonstrating a significant improvement over previous works in the field.
Table 5 compares the performance of two models, THAT (2021), by Bing et al., and our model on the MINE lab dataset. The THAT (2021) model achieves consistent values across all metrics, scoring 97% for accuracy, precision, recall, and F1-score. Our model significantly outperforms the THAT model, achieving 99.24% across all evaluated metrics. This demonstrates a notable improvement on the MINE lab dataset, highlighting the superior capability of our model in accuracy, precision, recall, and F1-score. We selected the THAT model for comparison because it is publicly available online.
Figure 11 illustrates the model’s performance on six activities from the StanWiFi dataset on the left side: fall, run, lie down, walk, sit down, and stand up. The right side visualizes the model’s performance on additional activities: stand up and squat down, raise and lower the right hand, open and close arms, kick the right leg, and kick the left leg. The confusion matrices indicate that the model achieved a near-perfect classification for most activities, demonstrating high true positive rates across both datasets.

4.3. Experimental Results on MultiEnv Dataset

Table 6 summarizes the performance of various models on the MultiEnv dataset, divided into three different environments: E1: Office (LOS), E2: Hall (LOS), and E3: Room and hall (NLOS). In the E1 environment, various models such as SVM (2021, 2022), THAT (2021), STC-NLSTMNet (2023), and AAE+RF (2023) are compared. Our model achieves the highest accuracy (99.47%) and outperforms others in all metrics, including precision (99.48%), recall (99.47%), and F1-score (99.47%). In the E2 environment, comparisons include models such as SVM (2021, 2022), THAT (2021), and STC-NLSTMNet (2023). Again, our model achieves the best performance with an accuracy of 98.43%, precision of 98.01%, recall of 97.90%, and F1-score of 97.90%. In the E3 environment, our model once again excels with an accuracy of 98.78%, precision, recall, and F1-score all at 98.78%. This is superior to other models, such as THAT (2021), STC-NLSTMNet (2023), and AAE+RF (2023).
Figure 12 visualizes the model’s performance across six activities in the MultiEnv dataset running on E1, E2, and E3. These matrices show that the model achieved high accuracy across all activities.

5. Conclusions

In this study, we presented a novel approach for human activity recognition (HAR) using Wi-Fi sensing, leveraging both the phase and amplitude components of Channel State Information (CSI). Our proposed model, which combines attention-based mechanisms with a Gated Recurrent Network (GRN) architecture, outperformed state-of-the-art models in HAR. The model’s efficacy was rigorously tested across multiple datasets, including StanWiFi, MultiEnv, and the MINE lab dataset, showcasing its adaptability and robustness in both line-of-sight (LOS) and non-line-of-sight (NLOS) environments. The source code is available at https://github.com/thai-duy-quy/PA-CSI-HAR (access on 26 January 2025).
By utilizing both phase and amplitude signals, we addressed the limitations of prior models that predominantly focused on a single feature, leading to improved accuracy in recognizing complex human activities. Our experimental results indicate that the multi-feature approach of combining amplitude and phase information significantly enhances accuracy, precision, recall, and F1-score metrics compared to existing methods.
Despite these advancements, our method has certain limitations. Experiments conducted in controlled laboratory settings differ from real-world environments, where Wi-Fi signals are influenced by factors such as stationary and moving objects. Moreover, our current study does not include cross-environment testing, even though real-world Wi-Fi signals are significantly affected by the surrounding environment. In the future, we aim to address these challenges by developing solutions to improve the robustness and adaptability of our method for diverse and dynamic conditions.

Author Contributions

Conceptualization, T.K.S. and C.-Y.L.; methodology, C.-Y.L. and T.D.Q.; software, T.D.Q. and C.-Y.L.; validation, T.D.Q.; resources, T.K.S.; data curation, T.D.Q.; writing—original draft preparation, T.D.Q.; writing—review and editing, C.-Y.L. and T.D.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The StanWiFi dataset is available at https://github.com/ermongroup/Wifi_Activity_Recognition (accessed on 6 February 2025). The MultiEnv dataset is available at https://data.mendeley.com/datasets/v38wjmz6f6/1 (accessed on 6 February 2025). The MINE lab dataset is provided upon request by contacting the corresponding author due to author privacy concerns.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kaseris, M.; Kostavelis, I.; Malassiotis, S. A Comprehensive Survey on Deep Learning Methods in Human Activity Recognition. Mach. Learn. Knowl. Extr. 2024, 6, 842–876. [Google Scholar] [CrossRef]
  2. Aggarwal, J.K.; Ryoo, M.S. Human Activity Analysis: A Review. ACM Comput. Surv. 2011, 43, 1–43. [Google Scholar] [CrossRef]
  3. Li, X.; He, Y.; Jing, X. A Survey of Deep Learning-Based Human Activity Recognition in Radar. Remote Sens. 2019, 11, 1068. [Google Scholar] [CrossRef]
  4. Zhuravchak, A.; Kapshii, O.; Pournaras, E. Human Activity Recognition Based on Wi-Fi CSI Data -A Deep Neural Network Approach. In Proceedings of the Procedia Computer Science; Elsevier B.V.: Amsterdam, The Netherlands, 2021; Volume 198, pp. 59–66. [Google Scholar]
  5. Jannat, M.K.A.; Islam, M.S.; Yang, S.H.; Liu, H. Efficient Wi-Fi-Based Human Activity Recognition Using Adaptive Antenna Elimination. IEEE Access 2023, 11, 105440–105454. [Google Scholar] [CrossRef]
  6. Chen, C.; Zhou, G.; Lin, Y. Cross-Domain WiFi Sensing with Channel State Information: A Survey. ACM Comput. Surv. 2023, 55, 231. [Google Scholar] [CrossRef]
  7. Lin, C.-Y.; Lin, C.-Y.; Liu, Y.-T.; Chen, Y.-W.; Shih, T.K. WiFi-TCN: Temporal Convolution for Human Interaction Recognition Based on WiFi Signal. IEEE Access 2024, 12, 126970–126982. [Google Scholar] [CrossRef]
  8. Yang, Z.; Qian, K.; Wu, C.; Zhang, Y. Smart Wireless Sensing: From IoT to AIoT; Springer Nature: Berlin/Heidelberg, Germany, 2021; ISBN 9789811656583. [Google Scholar]
  9. Sigg, S.; Blanke, U.; Troster, G. The Telepathic Phone: Frictionless Activity Recognition from WiFi-RSSI. In Proceedings of the IEEE International Conference on Pervasive Computing and Communications (PerCom), Budapest, Hungary, 24–28 March 2014. [Google Scholar]
  10. Gu, Y.; Ren, F.; Li, J. PAWS: Passive Human Activity Recognition Based on WiFi Ambient Signals. IEEE Internet Things J. 2016, 3, 796–805. [Google Scholar] [CrossRef]
  11. Gu, Y.; Quan, L.; Ren, F. WiFi-Assisted Human Activity Recognition. In Proceedings of the Wireless and Mobile, 2014 IEEE Asia Pacific Conference, Bali, Indonesia, 28–30 August 2014; pp. 60–65. [Google Scholar]
  12. Sigg, S.; Shi, S.; Buesching, F.; Ji, Y.; Wolf, L. Leveraging RF-Channel Fluctuation for Activity Recognition: Active and Passive Systems, Continuous and RSSI-Based Signal Features. In Proceedings of the MoMM 2013: The 11th International Conference on Advances in Mobile Computing and Multimedia, Vienna, Austria, 2–4 December 2013; Association for Computing Machinery: New York, NY, USA, December, 2013; p. 599. [Google Scholar]
  13. Youssef, M.; Mah, M.; Agrawala, A. Challenges: Device-Free Passive Localization for Wireless Environments. In Proceedings of the 13th Annual ACM International Conference on Mobile Computing and Networking, Montréal, QC, Canada, 9–14 September 2007; ACM Digital Library: New York, NY, USA, 2007; pp. 222–229. [Google Scholar]
  14. Rathnayake, R.M.M.R.; Maduranga, M.W.P.; Tilwari, V.; Dissanayake, M.B. RSSI and Machine Learning-Based Indoor Localization Systems for Smart Cities. Eng 2023, 4, 1468–1494. [Google Scholar] [CrossRef]
  15. Alsaify, B.A.; Almazari, M.M.; Alazrai, R.; Daoud, M.I. A Dataset for Wi-Fi-Based Human Activity Recognition in Line-of-Sight and Non-Line-of-Sight Indoor Environments. Data Brief 2020, 33, 106534. [Google Scholar] [CrossRef] [PubMed]
  16. Yousefi, S.; Narui, H.; Dayal, S.; Ermon, S.; Valaee, S. A Survey on Behavior Recognition Using WiFi Channel State Information. IEEE Commun. Mag. 2017, 55, 98–104. [Google Scholar] [CrossRef]
  17. Zhang, Y.; Zheng, Y.; Qian, K.; Zhang, G.; Liu, Y.; Wu, C.; Yang, Z. Widar3.0: Zero-Effort Cross-Domain Gesture Recognition With Wi-Fi. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 8671–8688. [Google Scholar] [CrossRef]
  18. Guo, L.; Guo, S.; Wang, L.; Lin, C.; Liu, J.; Lu, B.; Fang, J.; Liu, Z.; Shan, Z.; Yang, J. Wiar: A Public Dataset for Wifi-Based Activity Recognition. IEEE Access 2019, 7, 154935–154945. [Google Scholar] [CrossRef]
  19. Yang, J.; Chen, X.; Zou, H.; Lu, C.X.; Wang, D.; Sun, S.; Xie, L. SenseFi: A Library and Benchmark on Deep-Learning-Empowered WiFi Human Sensing. Patterns 2023, 4, 100703. [Google Scholar] [CrossRef]
  20. Meneghello, F.; Fabbro, N.D.; Garlisi, D.; Tinnirello, I.; Rossi, M. A CSI Dataset for Wireless Human Sensing on 80 MHz Wi-Fi Channels. IEEE Commun. Mag. 2023, 61, 146–152. [Google Scholar] [CrossRef]
  21. Wang, W.; Liu, A.X.; Shahzad, M.; Ling, K.; Lu, S. Device-Free Human Activity Recognition Using Commercial WiFi Devices. IEEE J. Sel. Areas Commun. 2017, 35, 1118–1131. [Google Scholar] [CrossRef]
  22. Alsaify, B.A.; Almazari, M.; Alazrai, R.; Alouneh, S.; Daoud, M.I. A CSI-Based Multi-Environment Human Activity Recognition Framework. Appl. Sci. 2022, 12, 930. [Google Scholar] [CrossRef]
  23. Chen, Z.; Zhang, L.; Jiang, C.; Cao, Z.; Cui, W. WiFi CSI Based Passive Human Activity Recognition Using Attention Based BLSTM. IEEE Trans. Mob. Comput. 2019, 18, 2714–2724. [Google Scholar] [CrossRef]
  24. Yadav, S.K.; Sai, S.; Gundewar, A.; Rathore, H.; Tiwari, K.; Pandey, H.M.; Mathur, M. CSITime: Privacy-Preserving Human Activity Recognition Using WiFi Channel State Information. Neural Netw. 2022, 146, 11–21. [Google Scholar] [CrossRef]
  25. Moshiri, F.P.; Shahbazian, R.; Nabati, M.; Ghorashi, S.A. A CSI-Based Human Activity Recognition Using Deep Learning. Sensors 2021, 21, 7225. [Google Scholar] [CrossRef]
  26. Salehinejad, H.; Valaee, S. LiteHAR: Lightweight Human Activity Recognition from WiFi Signals with Random Convolution Kernels. In Proceedings of the ICASSP 2022—2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022. [Google Scholar] [CrossRef]
  27. Shalaby, E.; ElShennawy, N.; Sarhan, A. Utilizing Deep Learning Models in CSI-Based Human Activity Recognition. Neural Comput. Appl. 2022, 34, 5993–6010. [Google Scholar] [CrossRef]
  28. Islam, M.S.; Jannat, M.K.A.; Hossain, M.N.; Kim, W.S.; Lee, S.W.; Yang, S.H. STC-NLSTMNet: An Improved Human Activity Recognition Method Using Convolutional Neural Network with NLSTM from WiFi CSI. Sensors 2023, 23, 356. [Google Scholar] [CrossRef]
  29. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  30. Ding, X.; Jiang, T.; Zhong, Y.; Wu, S.; Yang, J.; Zeng, J. Wi-Fi-Based Location-Independent Human Activity Recognition with Attention Mechanism Enhanced Method. Electronics 2022, 11, 642. [Google Scholar] [CrossRef]
  31. Yang, M.; Zhu, H.; Zhu, R.; Wu, F.; Yin, L.; Yang, Y. WiTransformer: A Novel Robust Gesture Recognition Sensing Model with WiFi. Sensors 2023, 23, 2612. [Google Scholar] [CrossRef]
  32. Li, B.; Cui, W.; Wang, W.; Zhang, L.; Chen, Z.; Wu, M. Two-Stream Convolution Augmented Transformer for Human Activity Recognition. Proc. AAAI Conf. Artif. Intell. 2021, 35, 286–293. [Google Scholar] [CrossRef]
  33. Zeng, Y.; Wu, D.; Gao, R.; Gu, T.; Zhang, D. FullBreathe. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 2, 1–19. [Google Scholar] [CrossRef]
  34. Li, C.; Brauer, J.; Sezgin, A.; Zenger, C. Kalman Filter Based MIMO CSI Phase Recovery for COTS WiFi Devices. In Proceedings of the ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canad, 6–11 June 2021; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2021; pp. 4820–4824. [Google Scholar]
  35. Savarese, P.; Figueiredo, D. Residual Gates: A Simple Mechanism for Improved Network Optimization. 2017. Available online: https://home.ttic.edu/~savarese/savarese_files/Residual_Gates.pdf (accessed on 6 February 2025).
  36. Lim, B.; Arik, S.O.; Loeff, N.; Pfister, T. Temporal Fusion Transformers for Interpretable Multi-Horizon Time Series Forecasting. Int. J. Forecast. 2019, 37, 1748–1764. [Google Scholar] [CrossRef]
  37. Stragapede, G.; Delgado-Santos, P.; Tolosana, R.; Vera-Rodriguez, R.; Guest, R.; Morales, A. Mobile Keystroke Biometrics Using Transformers. In Proceedings of the 2023 IEEE 17th International Conference on Automatic Face and Gesture Recognition (FG), Waikoloa Beach, HI, USA, 5–8 January 2022. [Google Scholar]
  38. Ba, J.L.; Kiros, J.R.; Hinton, G.E. Layer Normalization. arXiv 2016, arXiv:1607.06450. [Google Scholar]
  39. Qin, Z.; Yang, S.; Zhong, Y. Hierarchically Gated Recurrent Neural Network for Sequence Modeling. arXiv 2023, arXiv:2311.04823. [Google Scholar]
Figure 1. The propagation of Wi-Fi signals from the transmitter to the receiver in LOS/NLOS scenarios.
Figure 1. The propagation of Wi-Fi signals from the transmitter to the receiver in LOS/NLOS scenarios.
Sensors 25 01038 g001
Figure 2. Visualization of Wi-Fi CSI amplitude (above) and phase (below). (a) Signal description of the falling action across 3 antennas; (b) Signal description of the sitting down action across 3 antennas.
Figure 2. Visualization of Wi-Fi CSI amplitude (above) and phase (below). (a) Signal description of the falling action across 3 antennas; (b) Signal description of the sitting down action across 3 antennas.
Sensors 25 01038 g002aSensors 25 01038 g002b
Figure 3. MINE lab environment.
Figure 3. MINE lab environment.
Sensors 25 01038 g003
Figure 4. The architecture of the proposed HAR system (CA-CSI).
Figure 4. The architecture of the proposed HAR system (CA-CSI).
Sensors 25 01038 g004
Figure 5. Sliding window applied to the StanWiFi dataset and our dataset.
Figure 5. Sliding window applied to the StanWiFi dataset and our dataset.
Sensors 25 01038 g005
Figure 6. An illustrate of time alignment on three samples in the StanWiFi dataset.
Figure 6. An illustrate of time alignment on three samples in the StanWiFi dataset.
Sensors 25 01038 g006
Figure 7. An example of phase unwrapping in CSI using for falling action.
Figure 7. An example of phase unwrapping in CSI using for falling action.
Sensors 25 01038 g007
Figure 8. An example of GRE with K univariate Gaussian distributions.
Figure 8. An example of GRE with K univariate Gaussian distributions.
Sensors 25 01038 g008
Figure 9. The structure of MCAT layer.
Figure 9. The structure of MCAT layer.
Sensors 25 01038 g009
Figure 10. The structure of the GRN implemented in our model.
Figure 10. The structure of the GRN implemented in our model.
Sensors 25 01038 g010
Figure 11. Model’s accuracy performance on StanWiFi datasets (left) and MINE lab dataset (right).
Figure 11. Model’s accuracy performance on StanWiFi datasets (left) and MINE lab dataset (right).
Sensors 25 01038 g011
Figure 12. Confusion matrices for the MultiEnv datasets (E1: left; E2: middle; E3: right).
Figure 12. Confusion matrices for the MultiEnv datasets (E1: left; E2: middle; E3: right).
Sensors 25 01038 g012
Table 1. Summary of public datasets.
Table 1. Summary of public datasets.
SourceDataset NameDescription
Alsaify et al. (2020) [15] MultiEnvThis dataset was collected in three scenarios: line-of-sight (LOS) in both the office and hall, and non-line-of-sight (NLOS).
Yousefi et al. (2017) [16]StanWiFiThis dataset contains continuous CSI data for six activities without precise segmentation timestamps for each sample.
Yang et al. (2022) [17]Widar 3.0This large dataset, collected using Intel 5300 NIC with 30 subcarriers and containing 258 K Wi-Fi-based hand gesture instances spanning 8620 min across 75 domains.
Guo et al. (2019) [18]WiARThis dataset includes 16 activities, comprising coarse-grained activities and gestures, performed 30 times each, by ten volunteers.
Yang et al. (2023) [19]NTU-FiCollected using the Atheros CSI tool, this dataset features 114 subcarriers per antenna pair, and it includes 6 human activities and 14 gait patterns.
Meneghello et al. (2023) [20]WiFi-80 MHzCollected using two Netgear X4S AC2600 IEEE 802.11ac routers with 256 subcarriers (242 usable), this dataset features ten subjects and three applications
Table 2. Activities and class descriptions in the MultiEnv dataset.
Table 2. Activities and class descriptions in the MultiEnv dataset.
ClassActivityDescription
0No movementSitting, standing, or lying on the ground
1FallingFalling from a standing position or from a chair
2Sitting down or standing upSitting down on a chair or standing up from a chair
3WalkingWalking between the transmitter and receiver
4TurningTurning at the transmitter’s or receiver’s location
5Picking upPicking up an object such as a pen from the ground
Table 3. Hyperparameters and values chosen.
Table 3. Hyperparameters and values chosen.
HyperparameterValues
Window sizeStanWiFi dataset: 2000; our own dataset: 1000
Stride sizeStanWiFi dataset: 200; our own dataset: 100
K-Gaussian encoding10
Input data dimensions ( d i n × d k )StanWiFi dataset: (2000, 90); MultiEnv dataset: (850, 90); MINE lab dataset: (1000, 90)
Filter size in Multi-scale CNN (horizontal, vertical)Horizontal: {10, 40}; Vertical: {2, 4}
Number of heads in the multi-head self-attention mechanismh-head: 9; v-head: 50
Dropout rate0.1
Number of dense layers in GRN256
OptimizerAdam (learning rate = 0.001; decay rate = 0.9)
Batch size8
Epochs 200
Training environment NVIDIA GeForce RTX 3060 with CUDA v. 12.4, Python 3.11, TensorFlow 2.16
Table 4. Results of the experiments on the StanWiFi dataset.
Table 4. Results of the experiments on the StanWiFi dataset.
SourceModelAccPreRecallF1-Score
Li et al. [32]THAT (2021)98.20---
Yadav et al. [24]CSITime (2022)98.00---
Salehinejad et al. [26]LiteHAR (2022)98.0099.1698.8799.01
Salaby et al. [27]CNN-GRU (2022)99.3199.599.43-
Islam et al. [28]STC-NLSTMNet (2023)99.8899.7299.73-
Jannat et al. [5] AAE+RF (2023)99.8499.8299.8399.81
Ours PA-CSI (2024)99.9399.8699.9599.95
Table 5. The experiments on MINE lab dataset.
Table 5. The experiments on MINE lab dataset.
SourceModelAccPreRecallF1-Score
Li et al. [32]THAT (2021)97.0097.0097.0097.00
OursPA-CSI (2024)99.2499.2499.2499.24
Table 6. The experiments on MultiEnv dataset.
Table 6. The experiments on MultiEnv dataset.
EnvironmentSourceModelAccPreRecallF1-Score
E1: Office
(LOS)
Alsaify et al. [15]SVM (2020)94.03---
Li et al. [32]THAT (2021)98.9598.2898.2698.26
Alsaify et al. [22]SVM (2022)91.27---
Islam et al. [28]STC-NLSTMNet (2023)98.2098.1098.0898.09
Jannat et al. [5]AAE+RF (2023)97.6596.4296.4194.40
OursPA-CSI (2024)99.4799.4899.4799.47
E2: Hall
(LOS)
Alsaify et al. [15]SVM (2020)94.03- - -
Li et al. [32]THAT (2021)97.3997.2497.2297.22
Alsaify et al. [22]SVM (2022)91.27 - - -
Islam et al. [28]STC-NLSTMNet (2023)96.6596.5496.4196.48
OursPA-CSI (2024)98.4398.0197.9097.90
E3: Room
and hall
(NLOS)
Li et al. [32]THAT (2021)97.5697.0497.0497.03
Islam et al. [28]STC-NLSTMNet (2023)94.6894.5794.5594.56
Jannat et al. [5]AAE+RF (2023)93.3393.1293.0793.14
OursPA-CSI (2024)98.7898.7998.7898.78
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Quy, T.D.; Lin, C.-Y.; Shih, T.K. Enhanced Human Activity Recognition Using Wi-Fi Sensing: Leveraging Phase and Amplitude with Attention Mechanisms. Sensors 2025, 25, 1038. https://doi.org/10.3390/s25041038

AMA Style

Quy TD, Lin C-Y, Shih TK. Enhanced Human Activity Recognition Using Wi-Fi Sensing: Leveraging Phase and Amplitude with Attention Mechanisms. Sensors. 2025; 25(4):1038. https://doi.org/10.3390/s25041038

Chicago/Turabian Style

Quy, Thai Duy, Chih-Yang Lin, and Timothy K. Shih. 2025. "Enhanced Human Activity Recognition Using Wi-Fi Sensing: Leveraging Phase and Amplitude with Attention Mechanisms" Sensors 25, no. 4: 1038. https://doi.org/10.3390/s25041038

APA Style

Quy, T. D., Lin, C.-Y., & Shih, T. K. (2025). Enhanced Human Activity Recognition Using Wi-Fi Sensing: Leveraging Phase and Amplitude with Attention Mechanisms. Sensors, 25(4), 1038. https://doi.org/10.3390/s25041038

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop