You are currently viewing a new version of our website. To view the old version click .
Bioengineering
  • Article
  • Open Access

21 August 2025

A High-Fidelity mmWave Radar Dataset for Privacy-Sensitive Human Pose Estimation

,
,
and
1
Department of Building Environment and Energy Engineering, The Hong Kong Polytechnic University, Hong Kong SAR, China
2
Department of Biomedical Engineering, The Hong Kong Polytechnic University, Hong Kong SAR, China
3
Research Institute for Sports Science and Technology, The Hong Kong Polytechnic University, Hong Kong SAR, China
4
Research Institute for Smart Ageing, The Hong Kong Polytechnic University, Hong Kong SAR, China
This article belongs to the Special Issue Biomechanics and Motion Analysis

Abstract

Human pose estimation (HPE) in privacy-sensitive environments such as healthcare facilities and smart homes demands non-visual sensing solutions. Millimeter-wave (mmWave) radar emerges as a promising alternative, yet its development is hindered by the scarcity of high-fidelity datasets with accurate annotations. This paper introduces mmFree-Pose, the first dedicated mmWave radar dataset specifically designed for privacy-preserving HPE. Collected through a novel visual-free framework that synchronizes mmWave radar with VDSuit-Full motion-capture sensors, our dataset covers 10+ actions, from basic gestures to complex falls. Each sample provides (i) raw 3D point clouds with Doppler velocity and intensity, (ii) precise 23-joint skeletal annotations, and (iii) full-body motion sequences in privacy-critical scenarios. Crucially, all data is captured without the use of visual sensors, ensuring fundamental privacy protection by design. Unlike conventional approaches that rely on RGB or depth cameras, our framework eliminates the risk of visual data leakage while maintaining high annotation fidelity. The dataset also incorporates scenarios involving occlusions, different viewing angles, and multiple subject variations to enhance generalization in real-world applications. By providing a high-quality and privacy-compliant dataset, mmFree-Pose bridges the gap between RF sensing and home monitoring applications, where safeguarding personal identity and behavior remains a critical concern.

1. Introduction

Human pose estimation (HPE) has emerged as a foundational technology enabling transformative applications across healthcare monitoring [1,2,3], human–computer interaction [4,5,6], and smart environment systems [7,8,9]. In clinical contexts, continuous pose tracking facilitates early detection of critical events such as falls in elderly populations [10,11], while in residential settings, it supports intuitive gesture-based control of domestic appliances [5,12]. However, the predominant reliance on visual sensing modalities—particularly RGB cameras [13,14]—introduces significant limitations in privacy-sensitive environments. These systems inherently capture personally identifiable information, raising serious ethical and regulatory concerns in contexts like patient rooms, bathrooms, or private residences, where confidentiality is paramount. Moreover, their performance degrades substantially under suboptimal lighting conditions, limiting reliability in nocturnal care scenarios.
Millimeter-wave (mmWave) radar technology [15,16] presents a promising alternative, operating through radio frequency signals rather than optical capture. This modality offers two distinct advantages: immunity to lighting variations and inherent privacy preservation through non-visual data acquisition. By generating sparse 3D point clouds representing spatial coordinates without capturing biometric identifiers, mmWave systems fundamentally avoid the privacy violations associated with visual sensors. Despite these benefits, advancement in mmWave-based HPE has been constrained by the absence of high-fidelity datasets that simultaneously satisfy two critical requirements: precise annotations and coverage of wide daily movements.
Existing datasets exhibit notable limitations in this domain. The HuPR benchmark [17] leverages synchronized RGB cameras to generate pseudo-labels via pre-trained vision models, inheriting annotation inaccuracies while compromising privacy through visual data capture. Similarly, the dataset introduced by Sengupta et al. [18] relies on Microsoft Kinect for ground truth, introducing depth-sensing errors estimated at >50 mm positional deviation. More fundamentally, current resources lack comprehensive coverage of medically critical scenarios such as falls—events that demand high detection reliability but occur predominantly in privacy-sensitive contexts. This gap between technological potential and available training resources has impeded the development of robust mmWave HPE systems for real-world applications where privacy and accuracy are significant.
To address these challenges, we present mmFree-Pose, a novel dataset specifically designed for privacy-sensitive human pose estimation using mmWave radar. The specific comparison is shown in Figure 1. Our acquisition framework employs synchronized mmWave radar and inertial motion-capture sensors, eliminating optical sensors entirely from the data collection pipeline. This approach ensures fundamental privacy protection while achieving medical-grade annotation precision. The dataset comprises 4534 temporally aligned samples capturing 10+ actions performed by three subjects, encompassing both basic movements (e.g., arm waving and leg lifts) and complex scenarios (e.g., forward/backward falls) in multiple orientations relative to the radar. Each sample provides five-dimensional point cloud data (x, y, z, velocity, and intensity) paired with precise 23-joint skeletal annotations.
Figure 1. Comparative framework of human pose dataset acquisition: visual vs. non-visual paradigms. (a) Traditional visual/semi-visual methods (MS COCO and UCH-Thermal-Pose). (b) Semi-visual pseudo-labeling (HuPR). (c) Proposed privacy-centric framework with visual-free annotation.
The primary contributions of this work are threefold:
  • We establish a privacy-by-design data acquisition paradigm, enabling high-fidelity pose estimation in privacy-sensitive environment.
  • We establish the first mmWave radar dataset with high-fidelity full-body pose annotations generated without any optical or visual data.
  • Experimental validation demonstrates executive performance on complex poses, with testing on several benchmark models, achieving a significant performance.
Collectively, this resource bridges a critical gap in sensing infrastructure for applications where privacy constraints preclude visual monitoring, such as healthcare facilities or smart homes.
The remainder of this paper is structured as follows: Section 2 examines related work on privacy-preserving human pose estimation, mmWave-based pose estimation, and identifies specific limitations in existing datasets and inertial motion capture for annotation. Section 3 details the sensor synchronization methodology and data collection protocol. Section 4 presents benchmark evaluations and comparative analysis. Section 5 discusses practical applications, limitations, and future research directions. Concluding remarks are provided in Section 6.

3. Privacy-First Data Construction

3.1. Privacy-by-Design Framework

The core architecture of our data collection framework implements a strict privacy-by-design paradigm through the complete exclusion of optical sensors. Millimeter-wave radar captures motion data exclusively as sparse 3D point clouds, P = { p i | p i = x , y , z , v , I } R 5 , inherently incapable of reconstructing visual identifiers such as facial features or body morphology. Simultaneously, the VDSuit-Full motion-capture system employs several inertial measurement units (IMUs), directly generating joint coordinates without recording any visual representations. Additionally, the point cloud representation provides information-theoretic anonymity through dimensionality reduction. Crucially, the system operates without supplemental RGB, infrared, or depth cameras that could compromise privacy, establishing a new standard for ethically compliant human sensing.

3.2. Action Design

Action sequences in the dataset were carefully curated to address critical gaps in privacy-sensitive monitoring scenarios, particularly those relevant to healthcare, eldercare, and smart home environments. A total of 11 representative human actions were selected, spanning a spectrum from low-dynamic daily tasks to high-risk emergency motions. These actions were classified into three functional categories:
  • Basic gestures, such as single-hand waving and shoulder touching, reflect common interactive movements that may be used in gesture-controlled interfaces or passive monitoring of user activity.
  • Activities of daily living (ADLs), including sitting-to-standing transitions and simulated dressing motions, are essential for assessing autonomy and physical function—especially in bathroom or bedroom settings, where visual sensors are categorically prohibited due to privacy regulations.
  • Critical fall scenarios, including forward, backward, and lateral falls, were deliberately designed and validated by referencing the established fall-detection literature. These sequences simulate the most frequent and injurious fall patterns observed in elderly care environments.
Additional movements, such as controlled leg raises, static balancing poses, and transitional weight shifts, were incorporated to increase diversity in joint kinematics and center-of-mass dynamics. Importantly, to model viewpoint and orientation variance, each action was performed at three different yaw angles relative to the radar’s boresight (0°, 90°, and 180°), ensuring robustness of the dataset against rotation-dependent occlusion and sensor placement biases.
Each sequence spans approximately 5 to 15 s, capturing the complete temporal evolution of the motion from initiation to resolution. This full-trajectory coverage is crucial for downstream learning tasks such as sequence-based pose estimation, activity classification, and time-series fall prediction. The design ensures that mmFree-Pose reflects realistic movement complexity under privacy-preserving conditions, providing a benchmark that is both ethically compliant and technically challenging.

3.3. Data Preprocessing Pipeline

Using the IWR6843 ISK radar sensor from Texas Instruments, raw data underwent rigorous preprocessing to enhance quality while preserving privacy attributes. Radar point clouds were filtered to remove non-human points and environmental noise, retaining only clusters corresponding to body segments. Moreover, precise alignment between radar and motion-capture systems was achieved through temporal signal. Additionally, spatial calibration employed a hip-centered coordinate system where J hip = 0 , 0 , 0 , with Kabsch algorithm optimization minimizing residual error between radar centroids and mocap joint positions. Skeletal data was normalized to the hip joint origin through affine transformation, J k ^ = J k J h i p , ensuring consistent coordinate frames despite participant movement. To compensate for radar sparsity, we applied a simple temporal fusion technique by averaging five consecutive frames to enhance point density. While this approach helps reduce transient noise and improve annotation alignment with IMU data, it may also introduce temporal smoothing effects that can blur rapid motions. In future work, we aim to adopt more advanced temporal denoising methods, such as Kalman filtering or learned temporal smoothers, to better preserve dynamic motion characteristics. Final anonymization scrubbed all metadata, including timestamps, locations, and anthropometric records, assigning de-identified participant codes (P01–P03). The resulting dataset comprises 4534 samples with matched point clouds and 23-joint annotations, partitioned into training (80%), validation (10%), and test sets (10%) while maintaining temporal continuity within sequences. A data exemplar is provided in Figure 2a.
Figure 2. Temporal synchronization and feature Analysis of mmFree-Pose data: (a) time-aligned data flow—synchronized mmWave radar point clouds (top) and motion-capture keypoints (bottom); and (b) t-SNE visualization of pose feature embeddings (colored by participant).
Furthermore, to visually observe the impact of subjects on data distribution and action categories, we performed t-SNE visualization on the data samples, as shown in Figure 2b. The visualization results revealed that the inter-subject variability in point cloud distributions is relatively small compared to the variability introduced by different poses and actions. This suggests that the model’s performance is more influenced by pose diversity than by individual anthropometric differences.

4. Data Value Validation

4.1. Experimental Protocol

The validation framework employed standardized evaluation metrics and baseline models to objectively quantify dataset efficacy. Point Transformer [40], PointNet [41], and DGCNN [42] architectures served as reference models, representing fundamental point cloud processing approaches widely adopted in mmWave HPE research. The models are evaluated under three different settings—11, 16, and 23 joints—representing varying levels of model complexity and granularity in estimating human poses; more details are illustrated in Figure 3.
Figure 3. Three different joint settings are used in our experiments to assess the model’s robustness under varying joint counts. (a) displays a full human body skeleton with 23 labeled joint points. (bd) show point sets of body joints with 11, 16, and 23 joints, respectively. The incremental number of joints allows us to evaluate the model’s performance across different levels of complexity.
Performance was assessed through three complementary metrics: Percentage of Correct Keypoints within a 10 cm threshold (PCK@5 cm), Mean per Joint Position Error (MPJPE), and Average Precision across Object Keypoint Similarity thresholds from 0.5 to 0.95 (AP). Statistical significance against existing datasets was established via Wilcoxon signed-rank tests with α = 0.01 confidence level. The experiment setup is illustrated in Figure 4.
Figure 4. (Left) Hardware setup: mmWave radar sensor and laptop terminal for real-time data processing. (Right) Motion-capture interface displaying synchronized point cloud visualization.

4.2. Benchmark Performance

As shown in Table 1, PointNet demonstrates superior performance across all joint configurations and metrics, achieving State-of-the-Art results on the mmFree-Pose dataset. For the critical 23-joint evaluation, PointNet reaches the lowest Mean Localization Error (12.39 cm vs. DGCNN’s 13.01 cm) while improving PCK (47.18% vs. Point Transformer’s 40.89%) and AP (95% vs. DGCNN’s 89%). Notably, its advantage widens with simpler skeletons (16 joints: 44.17% PCK higher than DGCNN), suggesting exceptional efficiency in core joint tracking. Surprisingly, the attention-based Point Transformer underperforms despite its theoretical capacity, indicating sparse mmWave point clouds (∼50 points/frame) may hinder complex architectural learning. These results establish PointNet as the optimal baseline for privacy-sensitive HPE applications requiring balance between accuracy and computational efficiency. Visualization results, provided in Figure 5, further emphasize the robustness of deep learning approaches in accurately estimating human poses across different keypoint configurations, regardless of complexity.
Table 1. Quantitative comparison of 3D deep HPE methods on curated datasets. Arrows show optimization directions: ↑ for higher-better (PCK, AP); ↓ for lower-better (MLE).
Figure 5. Qualitative comparison of pose estimation models on privacy-sensitive actions.
Error distribution: As detailed in Table 2, joint-wise error analysis reveals a consistent hierarchy in prediction accuracy across all evaluated models. Core joints—notably the hip and spine—demonstrate substantially lower MPJPEs compared to peripheral joints. For example, hip localization errors range from 0.001 cm (DGCNN) to 0.202 cm (Point Transformer), while spine-related joints maintain sub-2 cm accuracy. In contrast, extremities such as hands and feet exhibit much higher error rates, ranging from ~18 cm to over 27 cm, with the left hand reaching 26.27 cm error in Point Transformer and 23.24 cm in PointNet.
Table 2. Mean per Joint Position Error (MPJPE) is used to measure the average Euclidean distance between the predicted keypoints and the ground truth for each joint, where a lower MPJPE indicates higher accuracy. The comparison of several benchmarks is conducted on the full dataset, using a setting of 23 joints for evaluation.
This disparity is primarily attributable to point cloud sparsity and radar reflectivity variance: the torso typically generates denser and more stable radar returns due to its larger surface area and proximity to the radar boresight, whereas smaller appendages, like wrists, toes, and fingers, contribute weaker, noisier returns that may intermittently disappear from the radar’s field of view during rapid motion or occlusion. Moreover, limb extremities exhibit greater range of motion, increasing prediction difficulty under non-uniform sampling.
Efficiency trade-offs: The further comparison for real-world deployment feasibility is summarized in Table 3. Computational benchmarks reveal that PointNet achieves an optimal trade-off between efficiency and accuracy, making it well-suited for latency-sensitive and resource-constrained environments such as smart home hubs, mobile health monitors, or embedded fall detection units.
Table 3. The computational comparison of model components at FLOPs, parameters, and inference time per frame.
Specifically, PointNet requires only 11.45 million floating-point operations (FLOPs)—largely fewer than DGCNN (54.69 M)—while maintaining a low inference latency of 2 milliseconds per frame. This low computational footprint ensures high responsiveness even on non-GPU devices or low-power processors, which are common in medical wearables or IoT edge systems. While Point Transformer exhibits superior parameter efficiency (298.18 K parameters vs. PointNet’s 349.12 K), its higher FLOPs (24.48 M) indicate more intensive real-time processing demands, which—combined with its significantly lower fall detection PCK—limit its viability in clinical applications, where both accuracy and reaction time are essential.
Collectively, these findings establish PointNet as the most deployable architecture under practical constraints. Its balance of accuracy, low latency, and moderate complexity makes it ideal for real-world HPE systems deployed in privacy-sensitive, real-time, and resource-limited environments.

4.3. Performance on Complex Motions

In challenging fall scenarios—one of critical but technically demanding tasks for human pose estimation—PointNet demonstrates superior robustness compared to competing architectures (Table 4). Specifically, it achieves an MLE of 16.73 cm, representing a 16.2% improvement over DGCNN (18.99 cm) and a 24.4% improvement over Point Transformer (22.13 cm). Furthermore, it records the highest AP of 90%, indicating a reliable capacity to localize joints even under rapid, chaotic movement conditions common in falls.
Table 4. Quantitative comparison of 3D deep HPE methods on curated datasets in complicated fall patterns. Arrows show optimization directions: ↑ for higher-better (PCK, AP); ↓ for lower-better (MLE).
This performance is particularly significant given the inherent difficulties in radar-based estimation of high-velocity limb trajectories. Fall events typically involve sudden, multi-joint rotations and ground contact within a short time span, often accompanied by partial occlusion or transient disappearance of extremity reflections in the radar signal. Such dynamics challenge model inference due to sparse and noisy point clouds, and degrade spatial coherence across frames.
Despite these constraints, PointNet achieves a Peak Correct Keypoint (PCK) of 43.93%, narrowly outperforming DGCNN (43.66%) and clearly surpassing Point Transformer (33.34%). The margin suggests that simpler architectural designs may generalize better to data-sparse, high-variance scenarios than transformer-based networks, which tend to rely on dense attention patterns ill-suited for radar sparsity (typically <200 points/frame). The reliability of PointNet in fall detection extends its potential for real-time applications.

5. Reproducibility and Applications

The mmFree-Pose dataset enables transformative applications in privacy-sensitive domains where visual monitoring is ethically or legally prohibited. In clinical settings, its high-fidelity fall sequences support developing non-invasive emergency alert systems for elderly care facilities, operating reliably in darkness without compromising patient dignity. For smart homes, gesture recognition during activities like sitting-to-standing control of bathroom appliances, while low-latency inference (e.g., PointNet at 2 ms) facilitates deployment on resource-constrained edge devices. Indus-trial safety monitoring further benefits from the inherent anonymity of point clouds, enabling worker activity tracking in hazardous zones.
Despite these advances, three limitations warrant consideration. First, peripheral joint localization remains challenged during dynamic motions, with high hand errors (Table 4) due to radar shortcomings—a fundamental constraint of mmWave sparsity. Second, demographic coverage is currently limited to three subjects (age 25–32, BMI 18.5–24.9), restricting generalizability to diverse populations such as pediatric or obese individuals. Third, while the dataset includes multiple orientations, it lacks arbitrary view angles and continuous transitions between poses. This discrete setup limits the ecological realism of natural human behavior, where orientation and pose evolve fluidly. Finally, data collection occurred in controlled lab environments, omitting real-world interference like moving furniture or pets, potentially inflating performance metrics versus field deployments.
Future work will address these constraints through targeted innovations. Demographic diversity will be expanded via recruiting more representative subjects without additional privacy-sensitive captures. Finally, we will initiate real-world validation in assisted living facilities, incorporating environmental clutter to bridge the lab-to-field gap. These efforts will solidify mmFree-Pose as a foundational resource for privacy-centric human sensing, advancing technologies that harmonize safety with uncompromising ethical standards.

6. Conclusions

This paper presents mmFree-Pose, the first dedicated mmWave radar dataset explicitly designed to address the critical challenge of privacy preservation in human pose estimation (HPE). By establishing a novel visual-free acquisition framework, we eliminate the ethical and technical risks associated with optical sensors while achieving medical-grade annotation accuracy through synchronized inertial motion capture. The dataset covers 11 safety-critical and socially relevant actions—from basic gestures to high-risk falls—captured under multiple orientations relative to the radar. Each sample provides raw 3D point clouds (with Doppler velocity and intensity) paired with precise 23-joint skeletal annotations, enabling robust model training without compromising subject anonymity.
Our experimental validation demonstrates that mmFree-Pose bridges the gap between RF sensing and privacy-critical applications. Benchmark evaluations reveal that PointNet achieves a State-of-the-Art performance on complex motions, outperforming transformer-based and graph convolutional counterparts. This underscores its suitability for latency-sensitive deployments in eldercare monitoring or smart home systems, where low computational overhead (11.45 M FLOPs, 2 ms inference time) is essential. Crucially, the dataset’s design directly resolves the longstanding trade-off between annotation fidelity and privacy compliance, setting a new standard for ethically grounded human sensing.
Despite its contributions, limitations remain: peripheral joint localization errors persist due to radar sparsity, demographic diversity is currently limited, and real-world environmental interference was not modeled. Future work will expand subject diversity (e.g., age and BMI), integrate multi-radar configurations to mitigate occlusion, and validate the framework in operational settings such as assisted-living facilities. We aim to catalyze advancements in privacy-preserving HPE, ultimately enabling technologies that safeguard dignity while enhancing safety in sensitive environments.

Author Contributions

Y.S. contributed to data collection and analysis, data interpretation, draft of the manuscript, and editing; H.H. contributed to concept development, draft of the manuscript, project management, student supervision, and funding acquisition; H.L. contributed to concept development and data collection; C.Z.-H.M. contributed to draft of the manuscript and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by grants from the Hong Kong Polytechnic University: Start-Up Fund for New Recruits (Project ID: P0040305) and departmental fund by the Department of Building Environment and Energy Engineering (Project ID: P0052446).

Institutional Review Board Statement

The ethics of this study was approved by the Hong Kong Polytechnic University (reference number: HSEARS20240926009, 7 October 2024).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding authors.

Acknowledgments

All the authors would like to thank the anonymous participants who volunteered to participate in this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Stenum, J.; Cherry-Allen, K.M.; Pyles, C.O.; Reetzke, R.D.; Vignos, M.F.; Roemmich, R.T. Applications of Pose Estimation in Human Health and Performance across the Lifespan. Sensors 2021, 21, 7315. [Google Scholar] [CrossRef]
  2. Zhu, Y.; Xiao, M.; Xie, Y.; Xiao, Z.; Jin, G.; Shuai, L. In-Bed Human Pose Estimation Using Multi-Source Information Fusion for Health Monitoring in Real-World Scenarios. Inf. Fusion 2024, 105, 102209. [Google Scholar] [CrossRef]
  3. Chen, K.; Gabriel, P.; Alasfour, A.; Gong, C.; Doyle, W.K.; Devinsky, O.; Friedman, D.; Dugan, P.; Melloni, L.; Thesen, T. Patient-Specific Pose Estimation in Clinical Environments. IEEE J. Transl. Eng. Health Med. 2018, 6, 1–11. [Google Scholar] [CrossRef]
  4. Huo, R.; Gao, Q.; Qi, J.; Ju, Z. 3d Human Pose Estimation in Video for Human-Computer/Robot Interaction. In Proceedings of the International Conference on Intelligent Robotics and Applications, Hangzhou, China, 5–7 July 2023; Springer: Singapore, 2023; pp. 176–187. [Google Scholar]
  5. Neto, P.; Simão, M.; Mendes, N.; Safeea, M. Gesture-Based Human-Robot Interaction for Human Assistance in Manufacturing. Int. J. Adv. Manuf. Technol. 2019, 101, 119–135. [Google Scholar] [CrossRef]
  6. Liu, H.; Nie, H.; Zhang, Z.; Li, Y.-F. Anisotropic Angle Distribution Learning for Head Pose Estimation and Attention Understanding in Human-Computer Interaction. Neurocomputing 2021, 433, 310–322. [Google Scholar] [CrossRef]
  7. Cosma, A.; Radoi, I.E.; Radu, V. Camloc: Pedestrian Location Estimation through Body Pose Estimation on Smart Cameras. In Proceedings of the 2019 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Pisa, Italy, 30 September–3 October 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–8. [Google Scholar]
  8. Palermo, M.; Moccia, S.; Migliorelli, L.; Frontoni, E.; Santos, C.P. Real-Time Human Pose Estimation on a Smart Walker Using Convolutional Neural Networks. Expert Syst. Appl. 2021, 184, 115498. [Google Scholar] [CrossRef]
  9. Wu, C.; Aghajan, H. Real-Time Human Pose Estimation: A Case Study in Algorithm Design for Smart Camera Networks. Proc. IEEE 2008, 96, 1715–1732. [Google Scholar] [CrossRef]
  10. Beddiar, D.R.; Oussalah, M.; Nini, B. Fall Detection Using Body Geometry and Human Pose Estimation in Video Sequences. J. Vis. Commun. Image Represent. 2022, 82, 103407. [Google Scholar] [CrossRef]
  11. Juraev, S.; Ghimire, A.; Alikhanov, J.; Kakani, V.; Kim, H. Exploring Human Pose Estimation and the Usage of Synthetic Data for Elderly Fall Detection in Real-World Surveillance. IEEE Access 2022, 10, 94249–94261. [Google Scholar] [CrossRef]
  12. Coronado, E.; Villalobos, J.; Bruno, B.; Mastrogiovanni, F. Gesture-Based Robot Control: Design Challenges and Evaluation with Humans. In Proceedings of the 2017 IEEE international conference on robotics and automation (ICRA), Singapore, 29 May–3 June 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 2761–2767. [Google Scholar]
  13. Mehta, D.; Sridhar, S.; Sotnychenko, O.; Rhodin, H.; Shafiei, M.; Seidel, H.-P.; Xu, W.; Casas, D.; Theobalt, C. Vnect: Real-Time 3d Human Pose Estimation with a Single Rgb Camera. ACM Trans. Graph. (TOG) 2017, 36, 1–14. [Google Scholar] [CrossRef]
  14. Zheng, C.; Wu, W.; Chen, C.; Yang, T.; Zhu, S.; Shen, J.; Kehtarnavaz, N.; Shah, M. Deep Learning-Based Human Pose Estimation: A Survey. ACM Comput. Surv. 2023, 56, 1–37. [Google Scholar] [CrossRef]
  15. Rao, S.; Texas Instruments. Introduction to Mmwave Sensing: FMCW Radars; Texas Instruments: Dallas, TX, USA, 2017. [Google Scholar]
  16. Bi, X.; Bi, X. Millimeter Wave Radar Technology. In Environmental Perception Technology for Unmanned Systems; Springer: Berlin/Heidelberg, Germany, 2021; pp. 17–65. [Google Scholar]
  17. Lee, S.P.; Kini, N.P.; Peng, W.H.; Ma, C.W.; Hwang, J.N. HuPR: A Benchmark for Human Pose Estimation Using Millimeter Wave Radar. In Proceedings of the Proceedings—2023 IEEE Winter Conference on Applications of Computer Vision, WACV 2023, Waikoloa, HI, USA, 2–7 January 2023. [Google Scholar]
  18. Sengupta, A.; Jin, F.; Zhang, R.; Cao, S. Mm-Pose: Real-Time Human Skeletal Posture Estimation Using MmWave Radars and CNNs. IEEE Sens. J. 2020, 20, 10032–10044. [Google Scholar] [CrossRef]
  19. Liu, H.; Liu, T.; Zhang, Z.; Sangaiah, A.K.; Yang, B.; Li, Y. ARHPE: Asymmetric Relation-Aware Representation Learning for Head Pose Estimation in Industrial Human–Computer Interaction. IEEE Trans. Ind. Inf. 2022, 18, 7107–7117. [Google Scholar] [CrossRef]
  20. Newell, A.; Yang, K.; Deng, J. Stacked Hourglass Networks for Human Pose Estimation. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part VIII 14. Springer: Cham, Switzerland, 2016; pp. 483–499. [Google Scholar]
  21. Sharma, A.M.; Venkatesh, K.S.; Mukerjee, A. Human Pose Estimation in Surveillance Videos Using Temporal Continuity on Static Pose. In Proceedings of the 2011 International Conference on Image Information Processing, Shimla, India, 3–5 November 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 1–6. [Google Scholar]
  22. Li, M.; Hu, H.; Xiong, J.; Zhao, X.; Yan, H. TSwinPose: Enhanced Monocular 3D Human Pose Estimation with JointFlow. Expert Syst. Appl. 2024, 249, 123545. [Google Scholar] [CrossRef]
  23. Zhao, M.; Li, T.; Abu Alsheikh, M.; Tian, Y.; Zhao, H.; Torralba, A.; Katabi, D. Through-Wall Human Pose Estimation Using Radio Signals. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7356–7365. [Google Scholar]
  24. Li, J.; Bian, S.; Zeng, A.; Wang, C.; Pang, B.; Liu, W.; Lu, C. Human Pose Regression with Residual Log-Likelihood Estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 11025–11034. [Google Scholar]
  25. Manzo, G.; Serratosa, F.; Vento, M. Online Human Assisted and Cooperative Pose Estimation of 2D Cameras. Expert Syst. Appl. 2016, 60, 258–268. [Google Scholar] [CrossRef]
  26. Zimmermann, C.; Welschehold, T.; Dornhege, C.; Burgard, W.; Brox, T. 3d Human Pose Estimation in Rgbd Images for Robotic Task Learning. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, QLD, Australia, 21–25 May 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1986–1992. [Google Scholar]
  27. Chen, S.; Xu, Y.; Zou, B. Prior-Knowledge-Based Self-Attention Network for 3D Human Pose Estimation. Expert Syst. Appl. 2023, 225, 120213. [Google Scholar] [CrossRef]
  28. Guo, Y.; Chen, Y.; Deng, J.; Li, S.; Zhou, H. Identity-Preserved Human Posture Detection in Infrared Thermal Images: A Benchmark. Sensors 2022, 23, 92. [Google Scholar] [CrossRef]
  29. Smith, J.; Loncomilla, P.; Ruiz-Del-Solar, J. Human Pose Estimation Using Thermal Images. IEEE Access 2023, 11, 35352–35370. [Google Scholar] [CrossRef]
  30. Xie, C.; Zhang, D.; Wu, Z.; Yu, C.; Hu, Y.; Sun, Q.; Chen, Y. Accurate Human Pose Estimation Using RF Signals. In Proceedings of the 2022 IEEE 24th International Workshop on Multimedia Signal Processing (MMSP), Shanghai, China, 26–28 September 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–6. [Google Scholar]
  31. Yu, C.; Zhang, D.; Wu, Z.; Xie, C.; Lu, Z.; Hu, Y.; Chen, Y. Fast 3D Human Pose Estimation Using RF Signals. In Proceedings of the ICASSP 2023—2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–5. [Google Scholar]
  32. Li, G.; Zhang, Z.; Yang, H.; Pan, J.; Chen, D.; Zhang, J. Capturing Human Pose Using MmWave Radar. In Proceedings of the 2020 IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops), Austin, TX, USA, 23–27 March 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–6. [Google Scholar]
  33. Cui, H.; Zhong, S.; Wu, J.; Shen, Z.; Dahnoun, N.; Zhao, Y. Milipoint: A Point Cloud Dataset for Mmwave Radar. Adv. Neural Inf. Process. Syst. 2023, 36, 62713–62726. [Google Scholar]
  34. Ahmad, N.; Ghazilla, R.A.R.; Khairi, N.M.; Kasi, V. Reviews on Various Inertial Measurement Unit (IMU) Sensor Applications. Int. J. Signal Process. Syst. 2013, 1, 256–262. [Google Scholar] [CrossRef]
  35. Roetenberg, D.; Luinge, H.; Slycke, P. Xsens MVN: Full 6DOF Human Motion Tracking Using Miniature Inertial Sensors. Xsens Motion Technol. BV Tech. Rep. 2009, 1, 1–7. [Google Scholar]
  36. Felius, R.A.W.; Geerars, M.; Bruijn, S.M.; van Dieën, J.H.; Wouda, N.C.; Punt, M. Reliability of IMU-Based Gait Assessment in Clinical Stroke Rehabilitation. Sensors 2022, 22, 908. [Google Scholar] [CrossRef] [PubMed]
  37. Felius, R.A.W.; Geerars, M.; Bruijn, S.M.; Wouda, N.C.; Van Dieën, J.H.; Punt, M. Reliability of IMU-Based Balance Assessment in Clinical Stroke Rehabilitation. Gait Posture 2022, 98, 62–68. [Google Scholar] [CrossRef] [PubMed]
  38. Kobsar, D.; Charlton, J.M.; Tse, C.T.F.; Esculier, J.-F.; Graffos, A.; Krowchuk, N.M.; Thatcher, D.; Hunt, M.A. Validity and Reliability of Wearable Inertial Sensors in Healthy Adult Walking: A Systematic Review and Meta-Analysis. J. Neuroeng. Rehabil. 2020, 17, 62. [Google Scholar] [CrossRef] [PubMed]
  39. Marano, D.; Cammarata, A.; Fichera, G.; Sinatra, R.; Prati, D. Modeling of a Three-Axes MEMS Gyroscope with Feedforward PI Quadrature Compensation. In Proceedings of the Advances on Mechanics, Design Engineering and Manufacturing: Proceedings of the International Joint Conference on Mechanics, Design Engineering & Advanced Manufacturing (JCM 2016), Catania, Italy, 14–16 September 2016; Springer: Cham, Switzerland, 2016; pp. 71–80. [Google Scholar]
  40. Zhao, H.; Jiang, L.; Jia, J.; Torr, P.H.S.; Koltun, V. Point Transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 16259–16268. [Google Scholar]
  41. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep Learning on Point Sets for 3d Classification and Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  42. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic Graph Cnn for Learning on Point Clouds. ACM Trans. Graph. (TOG) 2019, 38, 1–12. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.