DaRA Dataset: Combining Wearable Sensors, Location Tracking, and Process Knowledge for Enhanced Human Activity and Human Context Recognition in Warehousing
Abstract
1. Introduction
- Sensor data, for example, positional data of subjects and objects;
- Class labels, including the location of a subject and its tools, the subject’s process steps, or an order ID;
- Knowledge, such as order composition or an ideal process flow.
2. Related Work
2.1. HAR and Context
2.2. HAR Method in Production and Logistics
2.3. HAR Datasets
2.4. Research Gaps
3. DaRA Dataset
3.1. Experimental Setup
3.1.1. Introduction to the Laboratory Picking Lab
- Small items (from g), such as screws, locknuts, washers, or bits;
- Medium items (approximately 50 to 800 g), such as softshell jackets, ties, gloves, hoodies, bags, shirts, or notebooks;
- Large items (up to 5149 g), such as palm soil, axes, and hacksaws.
3.1.2. Logistics Scenarios
- Scenario 1: Paper list with pen.
- Scenario 2: Portable Data Terminal (PDT).
- Scenario 3: Paper list with glove scanner.
3.1.3. Sensor Configuration


3.2. Subjects
3.3. Data Recording
3.3.1. Preliminaries
3.3.2. Recording Process
3.3.3. Recording Results
3.4. Class Categories and Class Labels
3.5. Annotation and Revision
3.5.1. Annotation Methodology
3.5.2. Annotation Sessions
3.5.3. Annotation Revision
3.5.4. Time Effort
3.6. Available Data and Dataset Utilization
- A total of 216 annotations for the synchronized cameras (18 subjects × 12 class categories);
- A total of 420 annotations each for the IMU data and the Beacon data ([18 subjects × 2 wearable sets—one faulty wearable set from subject S10] × 12 class categories).
4. Evaluation—Dataset Quality
4.1. Annotation and Revision Quality
4.2. Sensor Data Quality
4.3. Quality of Revised Annotations and Sensors Combined—Deploying DaRA for HAR
5. Discussion and Future Works
5.1. Discussion
5.2. Future Works
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
| ADL | Activities of Daily Living |
| BLE | Bluetooth Low Energy |
| BPMN | Business Process Model and Notation |
| CAARL | Context-Aware Activity Recognition in Logistics |
| CC | Class Category |
| CL | Class Label |
| CNN | Convolutional Neural Network |
| DaRA | Data Fusion for advanced Research in industrial Applications |
| FN | False Negative |
| FP | False Positive |
| fps | Frame per Second |
| FPV | First-Person View |
| HAR | Human Activity Recognition |
| HCR | Human Context Recognition |
| HMM | Hidden Markov Model |
| Hz | Herz |
| ID | Identification |
| IMU | Inertial Measurement Unit |
| IT | Information Technology |
| LARa | Logistic Activity Recognition Challenge |
| LSTM | Long Short-Term Memory |
| MoCap | Motion Capture |
| Nr. | Number |
| P | Precision |
| PDT | Portable Data Terminal |
| PH | Person-Hours |
| R | Recall |
| RGB | Red–Green–Blue (refering to colored video) |
| RGB-D | Red–Green–Blue and Depth (refering to colored video with depth information) |
| RSSI | Received Signal Strength Indicator |
| TN | True Negative |
| tcnn | Temporal Convolutional Neural Network |
| TP | True Positive |
| TPV | Third-Person View |
| WMS | Warehouse Management System |
Appendix A







References
- Chen, K.; Zhang, D.; Yao, L.; Guo, B.; Yu, Z.; Liu, Y. Deep learning for sensor-based human activity recognition: Overview, challenges, and opportunities. ACM Comput. Surv. CSUR 2021, 54, 77. [Google Scholar]
- Reining, C.; Niemann, F.; Moya Rueda, F.; Fink, G.A.; ten Hompel, M. Human activity recognition for production and logistics—A systematic literature review. Information 2019, 10, 245. [Google Scholar] [CrossRef]
- Ordóñez, F.J.; Roggen, D. Deep convolutional and lstm recurrent neural networks for multimodal wearable activity recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef] [PubMed]
- Moya Rueda, F.; Grzeszick, R.; Fink, G.A.; Feldhorst, S.; Ten Hompel, M. Convolutional neural networks for human activity recognition using body-worn sensors. Informatics 2018, 5, 26. [Google Scholar] [CrossRef]
- Yoshimura, N.; Morales, J.; Maekawa, T.; Hara, T. OpenPack: A Large-Scale Dataset for Recognizing Packaging Works in IoT-Enabled Logistic Environments. In Proceedings of the 2024 IEEE International Conference on Pervasive Computing and Communications (PerCom), Biarritz, France, 11–15 March 2024; pp. 90–97. [Google Scholar] [CrossRef]
- OpenPack OpenPack Dataset. 2022. Available online: https://open-pack.github.io/ (accessed on 18 January 2026).
- Maurice, P.; Malaisé, A.; Amiot, C.; Paris, N.; Richard, G.J.; Rochel, O.; Ivaldi, S. Human movement and ergonomics: An industry-oriented dataset for collaborative robotics. Int. J. Robot. Res. 2019, 38, 1529–1537. [Google Scholar] [CrossRef]
- Maurice, P.; Malaisé, A.; Ivaldi, S.; Rochel, O.; Amiot, C.; Paris, N.; Richard, G.J.; Fritzsche, L. AndyData-Lab-onePerson. 2019. Available online: http://zenodo.org/record/3254403#.XmDpQahKguV (accessed on 10 October 2025).
- Reining, C.; Schlangen, M.; Hissmann, L.; ten Hompel, M.; Moya, F.; Fink, G.A. Attribute representation for human activity recognition of manual order picking activities. In Proceedings of the 5th international Workshop on Sensor-Based Activity Recognition and Interaction, Berlin, Germany, 20–21 September 2018; pp. 1–10. [Google Scholar]
- Niemann, F.; Reining, C.; Moya Rueda, F.; Nair, N.R.; Steffens, J.A.; Fink, G.A.; Ten Hompel, M. LARa: Creating a Dataset for Human Activity Recognition in Logistics Using Semantic Attributes. Sensors 2020, 20, 4083. [Google Scholar] [CrossRef]
- Niemann, F.; Lüdtke, S.; Bartelt, C.; Ten Hompel, M. Context-Aware Human Activity Recognition in Industrial Processes. Sensors 2021, 22, 134. [Google Scholar] [CrossRef] [PubMed]
- Dourish, P. What we talk about when we talk about context. Pers. Ubiquitous Comput. 2004, 8, 19–30. [Google Scholar] [CrossRef]
- Bordel, B.; Alcarria, R.; Robles, T. Recognizing human activities in Industry 4.0 scenarios through an analysis-modeling- recognition algorithm and context labels. Integr. Comput.-Aided Eng. 2021, 29, 83–103. [Google Scholar] [CrossRef]
- Schmidt, A.; Beigl, M.; Gellersen, H.W. There is more to context than location. Comput. Graph. 1999, 23, 893–901. [Google Scholar] [CrossRef]
- Niemann, F.; Rueda, F.M.; Nair, N.R.; Orth, A.; Kfari, M.K.A.; Frichert, M.; Abdulaal, A.; Abu Seer, M.; Almatalka, H.; Asskar, H.; et al. Data Fusion for advanced Research in industrial Applications (DaRA)—A Multi-Sensor, Multi-Level Annotated Dataset for Human Activity and Human Context Recognition in Warehousing. 2026. Available online: https://zenodo.org/records/10468175 (accessed on 18 January 2026).
- Lüdtke, S.; Rueda, F.M.; Ahmed, W.; Fink, G.A.; Kirste, T. Human Activity Recognition using Attribute-Based Neural Networks and Context Information. arXiv 2021, arXiv:2111.04564. [Google Scholar] [CrossRef]
- Diete, A.; Sztyler, T.; Weiland, L.; Stuckenschmidt, H. Recognizing grabbing actions from inertial and video sensor data in a warehouse scenario. Procedia Comput. Sci. 2017, 110, 16–23. [Google Scholar] [CrossRef]
- Moya Rueda, F.; Fink, G.A. Learning Attribute Representation for Human Activity Recognition. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 523–528. [Google Scholar] [CrossRef]
- Cheng, H.T.; Sun, F.T.; Griss, M.; Davis, P.; Li, J.; You, D. Nuactiv: Recognizing unseen new activities using semantic attribute-based learning. In Proceedings of the 11th Annual International Conference on Mobile Systems, Applications, and Services, Taipei, Taiwan, 25–28 June 2013; pp. 361–374. [Google Scholar]
- Riboni, D.; Sztyler, T.; Civitarese, G.; Stuckenschmidt, H. Unsupervised recognition of interleaved activities of daily living through ontological and probabilistic reasoning. In Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing, Heidelberg, Germany, 12–16 September 2016; pp. 1–12. [Google Scholar]
- Krüger, F.; Nyolt, M.; Yordanova, K.; Hein, A.; Kirste, T. Computational state space models for activity and intention recognition. A feasibility study. PLoS ONE 2014, 9, e109381. [Google Scholar] [CrossRef]
- Rueda, F.M.; Lüdtke, S.; Schröder, M.; Yordanova, K.; Kirste, T.; Fink, G.A. Combining symbolic reasoning and deep learning for human activity recognition. In Proceedings of the 2019 IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops), Kyoto, Japan, 11–15 March 2019; pp. 22–27. [Google Scholar]
- Dönnebrink, R.; Moya Rueda, F.; Grzeszick, R.; Stach, M. Miss-placement Prediction of Multiple On-body Devices for Human Activity Recognition. In Proceedings of the 8th International Workshop on Sensor-Based Activity Recognition and Artificial Intelligence, iWOAR ’23, Lübeck, Germany, 21–22 September 2023. [Google Scholar] [CrossRef]
- Bassani, G.; Avizzano, C.A.; Filippeschi, A. Deep Learning Algorithms for Human Activity Recognition in Manual Material Handling Tasks. Sensors 2025, 25, 6705. [Google Scholar] [CrossRef]
- Kaczmarek, S.; Fiedler, M.; Bongers, A.; Wibbeling, S.; Grzeszick, R. Dataset and Methods for Recognizing Care Activities. In Proceedings of the 7th International Workshop on Sensor-Based Activity Recognition and Artificial Intelligence, iWOAR ’22, Rostock, Germany, 19–20 September 2022. [Google Scholar] [CrossRef]
- Al Farid, F.; Bari, A.; Miah, A.S.M.; Mansor, S.; Uddin, J.; Kumaresan, S.P. A Structured and Methodological Review on Multi-View Human Activity Recognition for Ambient Assisted Living. J. Imaging 2025, 11, 182. [Google Scholar] [CrossRef]
- Pabón, J.; Gómez, D.; Cerón, J.D.; Salazar-Cabrera, R.; López, D.M.; Blobel, B. A Comprehensive Dataset for Activity of Daily Living (ADL) Research Compiled by Unifying and Processing Multiple Data Sources. J. Pers. Med. 2025, 15, 210. [Google Scholar] [CrossRef]
- Sztyler, T.; Carmona, J.; Völker, J.; Stuckenschmidt, H. Self-tracking Reloaded: Applying Process Mining to Personalized Health Care from Labeled Sensor Data. In Transactions on Petri Nets and Other Models of Concurrency XI; Koutny, M., Desel, J., Kleijn, J., Eds.; Series Title: Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2016; Volume 9930, pp. 160–180. [Google Scholar] [CrossRef]
- Pires, I.; Garcia, N.M. Raw Dataset with Accelerometer, Gyroscope, Magnetometer, Location and Environment Data for Activities Without Motion; Mendeley Data: Aveiro, Portugal, 2022; Volume 3. [Google Scholar] [CrossRef]
- Cerón, J. Jesusceron/DataPreProcess, 2021. Original-Date: 2020-01-14T15:03:34Z. Available online: https://github.com/jesusceron/DataPreProcess (accessed on 18 January 2026).
- Ceron, J.D.; Kluge, F.; Küderle, A.; Eskofier, B.M.; López, D.M. Simultaneous Indoor Pedestrian Localization and House Mapping Based on Inertial Measurement Unit and Bluetooth Low-Energy Beacon Data. Sensors 2020, 20, 4742. [Google Scholar] [CrossRef] [PubMed]
- Cerón, J. Jesusceron/SLAM_HAR_IL, 2023. Original-Date: 2022-02-01T21:40:27Z. Available online: https://github.com/jesusceron/SLAM_HAR_IL (accessed on 18 January 2026).
- Ceron, J.D.; López, D.M.; Kluge, F.; Eskofier, B.M. Framework for Simultaneous Indoor Localization, Mapping, and Human Activity Recognition in Ambient Assisted Living Scenarios. Sensors 2022, 22, 3364. [Google Scholar] [CrossRef]
- Sztyler, T. Human Activity Recognition. Available online: https://www.uni-mannheim.de/dws/research/projects/activity-recognition/dataset/dataset-realworld/ (accessed on 10 October 2025).
- Sztyler, T.; Baur, H. On-Body Localization of Wearable Devices: An Investigation of Position-Aware Activity Recognition. Available online: http://publications.wim.uni-mannheim.de/informatik/lski/Sztyler2016Localization.pdf (accessed on 20 March 2025).
- Casilari, E.; Santoyo-Ramón, A.J. UMAFall: Fall Detection Dataset (Universidad de Malaga). Available online: http://figshare.com/articles/UMA_ADL_FALL_Dataset_zip/4214283 (accessed on 10 October 2025).
- Casilari, E.; Santoyo-Ramón, J.A.; Cano-García, J.M. UMAFall: A Multisensor Dataset for the Research on Automatic Fall Detection. Procedia Comput. Sci. 2017, 110, 32–39. [Google Scholar] [CrossRef]
- Chereshnev, R.; Kertész-Farkas, A. Romanchereshnev/HuGaDB. Available online: http://github.com/romanchereshnev/HuGaDB (accessed on 10 October 2025).
- Chereshnev, R.; Kertész-Farkas, A. HuGaDB: Human Gait Database for Activity Recognition from Wearable Inertial Sensor Networks. In Analysis of Images, Social Networks and Texts; Van Der Aalst, W.M., Ignatov, D.I., Khachay, M., Kuznetsov, S.O., Lempitsky, V., Lomazova, I.A., Loukachevitch, N., Napoli, A., Panchenko, A., Pardalos, P.M., et al., Eds.; Series Title: Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2018; Volume 10716, pp. 131–141. [Google Scholar] [CrossRef]
- Wiki Datasets. wiki:dataset [Human Activity Recognition Datasets]. Available online: http://har-dataset.org/doku.php?id=wiki:dataset (accessed on 18 January 2026).
- Forster, K.; Roggen, D.; Troster, G. Unsupervised Classifier Self-Calibration through Repeated Context Occurences: Is there Robustness against Sensor Displacement to Gain? In Proceedings of the 2009 International Symposium on Wearable Computers, Linz, Austria, 4–7 September 2009; pp. 77–84. [Google Scholar] [CrossRef]
- Bulling, A.; Blanke, U.; Schiele, B. A tutorial on human activity recognition using body-worn inertial sensors. ACM Comput. Surv. 2014, 46, 33. [Google Scholar] [CrossRef]
- Bulling, A.; Blanke, U.; Schiele, B. Andreas-Bulling/ActRecTut. Available online: http://github.com/andreas-bulling/ActRecTut (accessed on 10 October 2025).
- Hsiao, Y.S.; Sanchez-Riera, J.; Lim, T.; Hua, K.L.; Cheng, W.H. LaRED: A large RGB-D extensible hand gesture dataset. In Proceedings of the 5th ACM Multimedia Systems Conference, Singapore, 19–21 March 2014; pp. 53–58. [Google Scholar] [CrossRef]
- HaGRID HaGRID—HAnd Gesture Recognition Image Dataset. Available online: https://github.com/hukenovs/hagrid?tab=readme-ov-file (accessed on 18 January 2026).
- Kapitanov, A.; Kvanchiani, K.; Nagaev, A.; Kraynov, R.; Makhliarchuk, A. HaGRID–HAnd Gesture Recognition Image Dataset. arXiv 2022, arXiv:2206.08219. [Google Scholar] [CrossRef]
- Essid, S.; Lin, X.; Gowing, M.; Kordelas, G.; Aksay, A.; Kelly, P.; Fillon, T.; Zhang, Q.; Dielmann, A.; Kitanovski, V.; et al. 3DLife ACM MM Grand Challenge 2011–Realistic Interaction in Online Virtual Environments. Available online: http://perso.telecom-paristech.fr/essid/3dlife-gc-11/ (accessed on 10 October 2025).
- Essid, S.; Lin, X.; Gowing, M.; Kordelas, G.; Aksay, A.; Kelly, P.; Fillon, T.; Zhang, Q.; Dielmann, A.; Kitanovski, V.; et al. A multi-modal dance corpus for research into interaction between humans in virtual environments. J. Multimodal User Interfaces 2012, 7, 157–170. [Google Scholar] [CrossRef]
- Vögele, A.; Krüger, B. HDM12 Dance–Documentation on a Data Base of Tango Motion Capture. Available online: https://cg.cs.uni-bonn.de/publication/voegele-2016-hdm12 (accessed on 18 January 2026).
- Zhang, W.; Liu, Z.; Zhou, L.; Leung, H.; Chan, A.B. Martial Arts, Dancing and Sports dataset: A challenging stereo and multi-view dataset for 3D human pose estimation. Image Vis. Comput. 2017, 61, 22–39. [Google Scholar] [CrossRef]
- Zhang, W.; Liu, Z.; Zhou, L.; Leung, H.; Chan, A.B. Martial Arts, Dancing and Sports Dataset|VISAL. Available online: http://visal.cs.cityu.edu.hk/research/mads/ (accessed on 10 October 2025).
- Tits, M.; Laraba, S.; Caulier, E.; Tilmanne, J.; Dutoit, T. UMONS-TAICHI. Available online: http://github.com/numediart/UMONS-TAICHI (accessed on 10 October 2025).
- Tits, M.; Laraba, S.; Caulier, E.; Tilmanne, J.; Dutoit, T. UMONS-TAICHI: A multimodal motion capture dataset of expertise in Taijiquan gestures. Data Brief 2018, 19, 1214–1221. [Google Scholar] [CrossRef]
- CRCV|Center for Research in Computer Vision at the University of Central Florida. Available online: https://www.crcv.ucf.edu/data/UCF_Sports_Action.php (accessed on 18 January 2026).
- Rodriguez, M.D.; Ahmed, J.; Shah, M. Action MACH a spatio-temporal Maximum Average Correlation Height filter for action recognition. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar] [CrossRef]
- Hoelzemann, A.; Romero, J.L.; Bock, M.; Van Laerhoven, K.; Lv, Q. Hang-Time HAR: A Benchmark Dataset for Basketball Activity Recognition using Wrist-worn Inertial Sensors. Available online: https://zenodo.org/record/7920485 (accessed on 18 January 2026).
- Hoelzemann, A.; Romero, J.L.; Bock, M.; Van Laerhoven, K.; Lv, Q. Hang-Time HAR: A Benchmark Dataset for Basketball Activity Recognition Using Wrist-Worn Inertial Sensors. Sensors 2023, 23, 5879. [Google Scholar] [CrossRef] [PubMed]
- Medrano, C.; Igual, R.; Plaza, I.; Castro, M. Fall ADL Data|EduQTech. Available online: https://www.unizar.es/ (accessed on 10 October 2025).
- Medrano, C.; Igual, R.; Plaza, I.; Castro, M. Detecting Falls as Novelties in Acceleration Patterns Acquired with Smartphones. PLoS ONE 2014, 9, e94811. [Google Scholar] [CrossRef] [PubMed]
- Sucerquia, A.; López, J.D.; Vargas-Bonilla, J.F. SisFall|SISTEMIC. Available online: https://www.kaggle.com/datasets/nvnikhil0001/sis-fall-original-dataset (accessed on 10 October 2025).
- Sucerquia, A.; López, J.; Vargas-Bonilla, J. SisFall: A Fall and Movement Dataset. Sensors 2017, 17, 198. [Google Scholar] [CrossRef]
- Martinez-Villaseñor, L.; Ponce, H.; Brieva, J.; Moya-Albor, E.; Núñez Martínez, J.; Peñafort Asturiano, C. HAR-UP. Available online: http://sites.google.com/up.edu.mx/har-up/ (accessed on 10 October 2025).
- Martínez-Villaseñor, L.; Ponce, H.; Brieva, J.; Moya-Albor, E.; Núñez-Martínez, J.; Peñafort-Asturiano, C. UP-Fall Detection Dataset: A Multimodal Approach. Sensors 2019, 19. [Google Scholar] [CrossRef]
- Forkan, A.; Jayaraman, P.P.; Antonmeryl, C.; Montori, F.; Banerjee, A.; Fizza, K.; Georgakopoulos, D. Internet of Things Dataset for Human Operator Activity Recognition in Industrial Environment. In Proceedings of the 34th ACM International Conference on Information and Knowledge Management, Seoul, Republic of Korea, 10–14 November 2025; pp. 6356–6360. [Google Scholar]
- Forkan, A.; Jayaraman, P.P.; Antonmeryl, C.; Montori, F.; Banerjee, A.; Fizza, K.; Georgakopoulos, D. A Dataset for Assessing Worker Activities in Industrial Settings. Available online: https://digitalinnovationlab.github.io/mppdataset (accessed on 18 January 2026).
- Sturm, F.; Hergenroether, E.; Reinhardt, J.; Vojnovikj, P.S.; Siegel, M. Challenges of the creation of a dataset for vision based human hand action recognition in industrial assembly. In Proceedings of the Science and Information Conference, Chengdu, China, 25–27 August 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 1079–1098. [Google Scholar]
- Iodice, F.; De Momi, E.; Ajoudani, A. Hri30: An action recognition dataset for industrial human-robot interaction. In Proceedings of the 2022 26th International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022; pp. 4941–4947. [Google Scholar]
- Iodice, F.; Momi, E.D.; Ajoudani, A. HRI30: An Action Recognition Dataset for Industrial Human-Robot Interaction. 2022. Available online: https://zenodo.org/records/5833411 (accessed on 18 January 2026).
- Lagamtzis, D.; Schmidt, F.; Seyler, J.R.; Dang, T. Coax: Collaborative action dataset for human motion forecasting in an industrial workspace. In Proceedings of the ICAART (3), Online, 3–5 February 2022; pp. 98–105. [Google Scholar]
- Lagamtzis, D.; Schmidt, F.; Seyler, J.; Dang, T. CoAx–Collaborative Action Dataset. Available online: https://dlgmtzs.github.io/dataset-coax/ (accessed on 18 January 2026).
- Dallel, M.; Havard, V.; Dupuis, Y.; Baudry, D. Digital twin of an industrial workstation: A novel method of an auto-labeled data generator using virtual reality for human action recognition in the context of human–robot collaboration. Eng. Appl. Artif. Intell. 2023, 118, 105655. [Google Scholar] [CrossRef]
- Dallel, M.; Havard, V.; Baudry, D.; Savatier, X. InHARD-DT–Industrial Human Action Recognition Dataset–Digital Twin. 2022. Available online: https://zenodo.org/records/7644247 (accessed on 18 January 2026).
- Cicirelli, G.; Marani, R.; Romeo, L.; Domínguez, M.G.; Heras, J.; Perri, A.G.; D’Orazio, T. The HA4M dataset: Multi-Modal Monitoring of an assembly task for Human Action recognition in Manufacturing. Sci. Data 2022, 9, 745. [Google Scholar] [CrossRef]
- Cicirelli, G.; Marani, R.; Romeo, L.; Domínguez, M.G.; Heras, J.; Perri, A.G.; D’Orazio, T. Human Action Multi-Modal Monitoring in Manufacturing (HA4M) Dataset. Available online: https://baltig.cnr.it/ISP/ha4m (accessed on 18 January 2026).
- Sener, F.; Chatterjee, D.; Shelepov, D.; He, K.; Singhania, D.; Wang, R.; Yao, A. Assembly101: A Large-Scale Multi-View Video Dataset for Understanding Procedural Activities. arXiv 2022, arXiv:2203.14712. [Google Scholar] [CrossRef]
- Assembly101. Assembly101: A Large-Scale Multi-View Video Dataset for Understanding Procedural Activities. Available online: https://assembly-101.github.io/ (accessed on 18 January 2026).
- Munasinghe, C.; Amin, F.M.; Scaramuzza, D.; van de Venn, H.W. Covered, collaborative robot environment dataset for 3d semantic segmentation. In Proceedings of the 2022 IEEE 27th International Conference on Emerging Technologies and Factory Automation (ETFA), Stuttgart, Germany, 6–9 September 2022; pp. 1–4. [Google Scholar]
- Munasinghe, C.; Amin, F.M.; Scaramuzza, D.; van de Venn, H.W. COVERED, CollabOratiVE Robot Environment Dataset for 3D Semantic Segmentation. Available online: https://github.com/Fatemeh-MA/COVERED (accessed on 18 January 2026).
- Niemann, F.; Bas, H.; Steffens, J.A.; Nair, N.R.; ten Hompel, M. Context-Aware Activity Recognition in Logistics (CAARL)—A optical marker-based Motion Capture Dataset. 2021. Available online: https://zenodo.org/records/5680951 (accessed on 18 January 2026).
- Tamantini, C.; Cordella, F.; Lauretti, C.; Zollo, L. The WGD—A dataset of assembly line working gestures for ergonomic analysis and work-related injuries prevention. Sensors 2021, 21, 7600. [Google Scholar] [CrossRef]
- Mohammadi Amin, F.; Rezayati, M.; van de Venn, H.W.; Karimpour, H. A mixed-perception approach for safe human–robot collaboration in industrial automation. Sensors 2020, 20, 6347. [Google Scholar] [CrossRef]
- Rezayati, M.; van de Venn, H.W. Physical Human-Robot Contact Detection. Available online: https://data.mendeley.com/datasets/ctw2256phb/2 (accessed on 18 January 2026).
- Alia, S.S.; Adachi, K.; Nahid, N.; Kaneko, H.; Lago, P.; Inoue, S. Bento Packaging Activity Recognition Challenge. 2021. Available online: https://abc-research.github.io/bento2021/data/ (accessed on 18 January 2026).
- Alia, S.S.; Adachi, K.; Nahid, N.; Kaneko, H.; Lago, P.; Inoue, S. Bento Packaging Activity Recognition Challenge, IEEE DataPort. 2021. Available online: https://ieee-dataport.org/competitions/bento-packaging-activity-recognition-challenge (accessed on 18 January 2026).
- Dallel, M.; Havard, V.; Baudry, D.; Savatier, X. InHARD–Industrial Human Action Recognition Dataset in the Context of Industrial Collaborative Robotics. Available online: https://zenodo.org/records/4003541 (accessed on 18 January 2026).
- Dallel, M.; Havard, V.; Baudry, D.; Savatier, X. InHARD–Industrial Human Action Recognition Dataset in the Context of Industrial Collaborative Robotics. In Proceedings of the 2020 IEEE International Conference on Human-Machine Systems (ICHMS), Rome, Italy, 7–9 September 2020; pp. 1–6. [Google Scholar] [CrossRef]
- Niemann, F.; Reining, C.; Moya Rueda, F.; Bas, H.; Altermann, E.; Nair, N.R.; Steffens, J.A.; Fink, G.A.; ten Hompel, M. Logistic Activity Recognition Challenge (LARa Version 02)—A Motion Capture and Inertial Measurement Dataset. 2022. Available online: https://zenodo.org/records/5761276 (accessed on 18 January 2026).
- Niemann, F.; Reining, C.; Moya Rueda, F.; Nair, N.R.; Oberdiek, P.; Bas, H.; Spiekermann, R.; Altermann, E.; Steffens, J.A.; Fink, G.A.; et al. Logistic Activity Recognition Challenge (LARa Version 03)—A Motion Capture and Inertial Measurement Dataset. 2023. Available online: https://zenodo.org/records/8189341 (accessed on 18 January 2026).
- Niemann, F.; Reining, C.; Moya Rueda, F.; Nair, N.R.; Steffens, J.A.; Fink, G.A.; ten Hompel, M. Logistic Activity Recognition Challenge (LARa)—A Motion Capture and Inertial Measurement Dataset. 2020. Available online: https://zenodo.org/records/3862782 (accessed on 18 January 2026).
- Ragusa, F.; Furnari, A.; Livatino, S.; Farinella, G.M. The meccano dataset: Understanding human-object interactions from egocentric videos in an industrial-like domain. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 1569–1578. [Google Scholar]
- Ragusa, F.; Furnari, A.; Livatino, S.; Farinella, G.M. The Meccano Dataset: Understanding Human-Object Interactions from Egocentric Videos in an Industrial-like Domain. 2021. Available online: https://iplab.dmi.unict.it/MECCANO/ (accessed on 18 January 2026).
- Ben-Shabat, Y.; Yu, X.; Saleh, F.; Campbell, D.; Rodriguez-Opazo, C.; Li, H.; Gould, S. The IKEA ASM dataset: Understanding people assembling furniture through actions, objects and pose. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 847–859. [Google Scholar]
- Ben-Shabat, Y.; Yu, X.; Saleh, F.; Campbell, D.; Rodriguez-Opazo, C.; Li, H.; Gould, S. The IKEA ASM Dataset: Understanding People Assembling Furniture through Actions, Objects and Pose. 2020. Available online: https://ikeaasm.github.io/ (accessed on 18 January 2026).
- Attila Reiss, I.I. PPG-DaLiA, 2019. UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu/dataset/495/ppg+dalia (accessed on 18 January 2026).
- Reiss, A.; Indlekofer, I.; Schmidt, P.; Van Laerhoven, K. Deep PPG: Large-Scale Heart Rate Estimation with Convolutional Neural Networks. Sensors 2019, 19, 3079. [Google Scholar] [CrossRef]
- Ashry, S.; Elbasiony, R.; Gomaa, W. An LSTM-based Descriptor for Human Activities Recognition using IMU Sensors. In Proceedings of the 15th International Conference on Informatics in Control, Automation and Robotics, Porto, Portugal, 29–31 July 2018; pp. 494–501. [Google Scholar] [CrossRef]
- Mohammed, S.; Gomaa, W. HAD-AW Data-Set Benchmark for Human Activity Recognition Using Apple Watch. 2018. Available online: http://www.researchgate.net/publication/324136132_HAD-AW_Data-set_Benchmark_For_Human_Activity_Recognition_Using_Apple_Watch (accessed on 10 October 2025).
- Nath, N.D.; Chaspari, T.; Behzadan, A.H. Automated ergonomic risk monitoring using body-mounted sensors and machine learning. Adv. Eng. Inform. 2018, 38, 514–526. [Google Scholar] [CrossRef]
- Vaizman, Y.; Ellis, K.; Lanckriet, G. Recognizing Detailed Human Context in the Wild from Smartphones and Smartwatches. IEEE Pervasive Comput. 2017, 16, 62–74. [Google Scholar] [CrossRef]
- ExtraSensory. The ExtraSensory Dataset. Available online: http://extrasensory.ucsd.edu/ (accessed on 18 January 2026).
- Zappi, P.; Lombriser, C.; Stiefmeier, T.; Farella, E.; Roggen, D.; Benini, L.; Tröster, G. Activity Recognition from On-Body Sensors: Accuracy-Power Trade-Off by Dynamic Sensor Selection. In Wireless Sensor Networks; Verdone, R., Ed.; Series Title: Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2008; Volume 4913, pp. 17–33. [Google Scholar] [CrossRef]
- Kwapisz, J.R.; Weiss, G.M.; Moore, S.A. WISDM Lab: Dataset. Available online: http://www.cis.fordham.edu/wisdm/dataset.php (accessed on 10 October 2025).
- Niemann, F.; Rueda, F.M.; Al Kfari, M.K.; Nair, N.R.; Lüdtke, S.; Kirchheim, A. Towards Standardized Dataset Creation for Human Activity Recognition: Framework, Taxonomy, Checklist, and Best Practices. In Annotation of Real-World Data for Artificial Intelligence Systems; Tonkin, E.L., Tourte, G.J.L., Yordanova, K., Eds.; Series Title: Communications in Computer and Information Science; Springer Nature: Cham, Switzerland, 2026; Volume 2706, pp. 74–93. [Google Scholar] [CrossRef]
- Gebru, T.; Morgenstern, J.; Vecchione, B.; Vaughan, J.W.; Wallach, H.; Iii, H.D.; Crawford, K. Datasheets for datasets. Commun. ACM 2021, 64, 86–92. [Google Scholar] [CrossRef]
- Wilkinson, M.D.; Dumontier, M.; Aalbersberg, I.J.; Appleton, G.; Axton, M.; Baak, A.; Blomberg, N.; Boiten, J.W.; Da Silva Santos, L.B.; Bourne, P.E.; et al. The FAIR Guiding Principles for scientific data management and stewardship. Sci. Data 2016, 3, 160018. [Google Scholar] [CrossRef]
- Rueda, F.M.; Al Kfari, M.K. Data4Sim, 2025. Version 1. Available online: https://github.com/wilfer9008/Data4Sim (accessed on 18 January 2026).
- Elke, L.; Braun, C.; Krooß, A.; Wings, L.M.; Niemann, F.; Kretschmer, V. Bewegungsdaten als Planungsgrundlage. Logist. J. Proc. 2024, 20, 1–10. [Google Scholar] [CrossRef]
- Wings, L.M.; Mazur, E.; Kretschmer, V. Light it Up! Integrationsleitfaden von Pick-by-Light für die Kommissionierung. 2024. Available online: https://publica.fraunhofer.de/entities/publication/13b962a2-5130-4436-aa0d-f28e5f5e00ce (accessed on 18 January 2026).
- Moya Rueda, F.; Ravi Nair, N.; Spiekermann, R.; Altermann, E.; Oberdiek, P.; Reining, C.; Fink, G.A. Retrieval-Based Annotation for Multi-Channel Time Series Data of Human Activities. In Annotation of Real-World Data for Artificial Intelligence Systems; Tonkin, E.L., Tourte, G.J.L., Yordanova, K., Eds.; Series Title: Communications in Computer and Information Science; Springer Nature: Cham, Switzerland, 2026; Volume 2706, pp. 53–73. [Google Scholar] [CrossRef]
- Papadatou-Pastou, M.; Ntolka, E.; Schmitz, J.; Martin, M.; Munafò, M.R.; Ocklenburg, S.; Paracchini, S. Human handedness: A meta-analysis. Psychol. Bull. 2020, 146, 481–524. [Google Scholar] [CrossRef]
- Aggarwal, J.; Ryoo, M. Human activity analysis: A review. ACM Comput. Surv. 2011, 43, 16. [Google Scholar] [CrossRef]
- Tran, H.; Potter, V.; Mazzucchelli, U.; John, D.; Intille, S. Towards Practical, Best Practice Video Annotation to Support Human Activity Recognition. In Annotation of Real-World Data for Artificial Intelligence Systems; Tonkin, E.L., Tourte, G.J.L., Yordanova, K., Eds.; Series Title: Communications in Computer and Information Science; Springer Nature: Cham, Switzerland, 2026; Volume 2706, pp. 94–118. [Google Scholar] [CrossRef]
- Niemann, F.; Rueda, F.M.; Nair, N.R.; Orth, A.; Kfari, M.K.A.; Frichert, M.; Abdulaal, A.; Abu Seer, M.; Almatalka, H.; Asskar, H.; et al. Test Annotations for Quality Evaluation of the DaRA Dataset—Annotated and Revised Video Data with Activity and Context Labels. 2026. Available online: https://zenodo.org/records/15118022 (accessed on 18 January 2026).
- Cohen, J. A Coefficient of Agreement for Nominal Scales. Educ. Psychol. Meas. 1960, 20, 37–46. [Google Scholar] [CrossRef]
- Light, R.J. Measures of response agreement for qualitative data: Some generalizations and alternatives. Psychol. Bull. 1971, 76, 365–377. [Google Scholar] [CrossRef]
- Landis, J.R.; Koch, G.G. The Measurement of Observer Agreement for Categorical Data. Biometrics 1977, 33, 159. [Google Scholar] [CrossRef]
- Monserud, R.A.; Leemans, R. Comparing global vegetation maps with the Kappa statistic. Ecol. Model. 1992, 62, 275–293. [Google Scholar] [CrossRef]
- Fleiss, J.L.; Levin, B.; Paik, M.C. Statistical Methods for Rates and Proportions, 1st ed.; Wiley Series in Probability and Statistics; Wiley: Hoboken, NJ, USA, 2003. [Google Scholar] [CrossRef]
- Zhong, R.; Hu, B.; Liu, Z.; Qin, Q.; Feng, Y.; Wang, X.V.; Wang, L.; Tan, J. A two-stage framework for learning human-to-robot object handover policy from 4D spatiotemporal flow. Robot. Comput.-Integr. Manuf. 2026, 98, 103171. [Google Scholar] [CrossRef]
- Castellani, C.; Turco, E.; Bo, V.; Malvezzi, M.; Prattichizzo, D.; Costante, G.; Pozzi, M. Soft Human-Robot Handover Using a Vision-Based Pipeline. IEEE Robot. Autom. Lett. 2025, 10, 891–898. [Google Scholar] [CrossRef]




shaped path. On the left side of the photograph, the Base, Cross Aisle Paths, and the Picking Lab with its eight rack complexes and five Aisle Paths are visible.
shaped path. On the left side of the photograph, the Base, Cross Aisle Paths, and the Picking Lab with its eight rack complexes and five Aisle Paths are visible.






| General | Download DaRA Dataset | [15] |
| Recording Environment | semi-controlled laboratory (Section 3.1.1) | |
| Scenario | warehousing: order picking, packaging, unpacking, storage (Section 3.1.2) | |
| BPMN (Section 3.3.2) | ||
| Dataset Size | 31:55:26 h of recording time (Section 3.3.3) | |
| Data Availability/Usage | Section 3.6 | |
| Sensor (Section 3.1.3) | Action Cameras | 1 camera per subject, fps, 32 h |
| Fixed Cameras | 6 cameras, fps, 77 h | |
| IMUs | 6 IMUs per subject (2 sets), 100 Hz | |
| Beacons | 57 beacons, 10 Hz | |
| Subjects (Section 3.2) | Number | 18 (4 female, 14 male) |
| Age | 21 to 67 years (avg. years) | |
| Weight | 62 to 103 kg (avg. kg) | |
| Height | 160 to 187 cm (avg. cm) | |
| Annotation | Class Categories (Section 3.4) | 12 categories with human movements and context |
| Class Labels (Section 3.4) | 207 labels, label representations | |
| Annotation (Section 3.5.2) | 1572 h manual annotated by 15 domain experts and trained internal annotators | |
| Revision (Section 3.5.3) | 361 h manual revision by 8 experts and automated plausibility checks | |
| Label Quality (Section 4.1) | Light’s Kappa from % to % depending on the class category |
| Dataset | Sensors | Subjects | Recording | Labels | Annotation | ||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| Name | Ref. | Year | Public | Size | Nr. | Type | Nr. | Environment | Category | Nr. and Type | |
| MPP Dataset | [64,65] | 2025 | ✓ | 3:23 h | 2 | inertial | 4 | real-world | human–object interactions | 7 activity classes | domain expert |
| IHADv 1 | [66] | 2023 | - | 459,180 images | 1 | visual (RGB) | - | controlled | human–object interaction | 12 activity classes | not mentioned |
| HRI30 | [67,68] | 2022 | ✓ | 15 GB | 1 | visual | 11 | controlled | body pose, human–object and human–robot interactions | 30 actions | manually annotated |
| CoAx | [69,70] | 2022 | ✓ | 1:58 h | 1 | visual (RGB-D) | 6 | controlled | human–object and human–robot interactions | 10 action and 8 object annotations | action and object annotation |
| OpenPack | [5,6] | 2022 | ✓ | 53.8 h | 20 | visual, inertial, physiological/biosensors, other | 16 | controlled | human-to-object interactions | 11 activity classes | expert |
| InHARD-DT | [71,72] | 2022 | ✓ | 25.8 GB | 34 | visual (RGB, MoCap), inertial | 12 | virtual | human–object and human–robot interactions | 18 event/action classes | auto-labelled |
| HA4M | [73,74] | 2022 | ✓ | 4.1 TB | 1 | visual (RGB, RGB-D, Infrared) | 41 | controlled | human–object interaction | 12 actions | manual annotation |
| Assembly101 | [75,76] | 2022 | ✓ | 513 h | 13 | visual | 53 | controlled | human-to-object interactions | 1380 fine-grained, 202 coarse actions | trained annotators |
| COVERED | [77,78] | 2022 | ✓ | 860 MB | 1 | visual | - | real-world | postures, human–robot interactions | 6 semantic segmentation classes | |
| CAARL | [11,79] | 2021 | ✓ | 2:33 h | 46 | visual (RGB, MoCap), inertial | 2 | controlled | postures/static activities, human-to-object interaction, locomotion | 8 activity classes, 19 attributes | annotation tool SARA |
| WGD | [80] | 2021 | - | - | 8 | visual (MoCap, RGB) | 8 | controlled | posture, human–object interactions | - | - |
| Physical Human–Robot Contact Detection | [81,82] | 2021 | ✓ | 79.9 MB | 2 | visual (RGB-D) | - | controlled | human–robot interactions, postures | 5 actions | - |
| ABC Bento | [83,84] | 2021 | ✓ | 499 MB | 20 | visual (MoCap) | 4 | controlled | human-to-object interaction | 10 labels | participants are designing methods |
| InHARD | [85,86] | 2020 | ✓ | 51.6 GB | 35 | visual (MoCap, RGB) | 16 | semi-controlled | human–object interaction | 14 low-level, 74 high-level action classes | annotation tool Anvil |
| LARa | [10,87,88,89] | 2020 | ✓ | 12:6 h | 54 | visual (RGB, MoCap), inertial | 16 | controlled | postures/static activities, human–object interaction, locomotion | 8 activity classes, 19 attributes | annotation tool SARA |
| MECCANO | [90,91] | 2020 | ✓ | 10.5 MB | 1 | visual | 20 | controlled | human–object interaction | 61 action classes with verb and object/s and bounding box annotations | manual |
| IKEA ASM | [92,93] | 2020 | ✓ | 35:26 h | 3 | visual (RGB, RGB-D) | 48 | controlled | human–object interaction | 33 verb-object | Amazon Turk manual annotators |
| AndyData-lab-onePerson | [7,8] | 2019 | ✓ | 5 h | 31 | visual (MoCap, RGB), inertial, tactile/force | 13 | controlled | postures/static activities, human–object interaction | 6 general, 5 detailed posture, 8 action | annotation tool Anvil |
| PPG-DaLiA | [94,95] | 2019 | ✓ | 36 h | 2 | inertial, physiological/biosensors | 15 | semi-controlled | postures/static activities, ADL, sports | 9 activity labels | protocol-defined |
| HAD-AW | [96,97] | 2018 | ✓ | 102 MB | 1 | inertial | 16 | real-world | ADL, sports | 8 ADLs consisting of 31 motion primitives | not explicitly mentioned |
| Nath et al. | [98] | 2018 | - | 40 min | 2 | inertial | 2 | semi-controlled, real-world | human–object interaction | 5 activity labels | manually |
| ExtraSensory Dataset | [99,100] | 2017 | ✓ | 5000 h | 1 | inertial, positioning, acoustic, environmental, other (phone state) | 60 | real-world | ADL | 116 original labels, 51 cleaned labels | by the user |
| Skoda Mini Checkpoint | [40,101] | 2008 | ✓ | – | 20 | inertial | 1 | controlled | human–object interaction | 10 gesture, 70 instances of each gesture | experimenters |
| Scenario | |||||||||
|---|---|---|---|---|---|---|---|---|---|
| 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | ||
| High-Level Processes | Retrieval (picking and packing) | X | X | X | X | ||||
| Storage (unpacking and storing) | X | X | X | X | |||||
| Picking Strategies | Single-order picking (serial) | X | X | X | X | X | X | ||
| Multi-order picking (parallel) | X | X | |||||||
| Information Technologies | Picking list and pen | X | X | X | X | X | X | ||
| Portable data terminal | X | ||||||||
| Picking list and glove scanner | X | ||||||||
| Customer Order | 2904 | X | X | X | X | ||||
| 2905 | X | X | X | X | |||||
| 2906 | X | X | |||||||
| Errors in Picking List | With intentional errors | X | X | ||||||
| Without intentional errors | X | X | X | X | X | X | |||
| ID | Sex | Age | Weight | Height | Handedness | Employment | Experience [from 1 = Extensive to 6 = None] | ||
|---|---|---|---|---|---|---|---|---|---|
| [F/M] | [years] | [kg] | [cm] | [L/R] | Status | Order Picking | Packaging | Similar Studies | |
| S01 | F | 32 | 68 | 171 | R | Student | 2 | 3 | 6 |
| S02 | M | 27 | 76 | 167 | R | Student | 3 | 6 | 6 |
| S03 | M | 64 | 69 | 171 | R | Employee | 6 | 5 | 5 |
| S04 | M | 31 | 85 | 183 | L | Employee | 5 | 4 | 6 |
| S05 | M | 67 | 100 | 177 | R | Retiree | 6 | 3 | 6 |
| S06 | M | 24 | 82 | 178 | R | Student | 4 | 6 | 6 |
| S07 | M | 41 | 70 | 180 | R | Employee | 6 | 5 | 6 |
| S08 | F | 29 | 62 | 163 | R | Student | 6 | 6 | 6 |
| S09 | M | 21 | 85 | 180 | R | Student | 6 | 6 | 6 |
| S10 | M | 28 | 85 | 160 | R | Student | 3 | 3 | 6 |
| S11 | M | 59 | 85 | 178 | R | Employee | 3 | 2 | 6 |
| S12 | M | 43 | 103 | 186 | R | Job seeker | 6 | 6 | 4 |
| S13 | F | 52 | 66 | 175 | R | Employee | 5 | 4 | 6 |
| S14 | M | 32 | 80 | 176 | R | Employee | 6 | 5 | 5 |
| S15 | M | 43 | 88 | 177 | R | Employee | 6 | 5 | 6 |
| S16 | M | 29 | 100 | 175 | R | Student | 6 | 3 | 6 |
| S17 | F | 25 | 75 | 180 | R | Employee | 6 | 5 | 6 |
| S18 | M | 26 | 80 | 187 | R | Student | 6 | 6 | 6 |
| Min. | 21 | 62 | 160 | ||||||
| Avg. | 37.4 | 81.1 | 175.8 | ||||||
| Max. | 67 | 103 | 187 | ||||||
| ID | Recording | Scope of the Scenarios 1–8 [hh:mm:ss] | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| Session | Retrieval (Scenario 1–3) | Storage (Scenario 4–6) | Perfect Run | Other | Total | ||||||
| 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | ||||
| S01 | 1 | 00:18:15 | 00:19:20 | 00:18:39 | - | - | 00:15:51 | 00:23:42 | 00:14:34 | 00:10:59 | 02:01:19 |
| S02 | 1 | 00:19:43 | 00:16:36 | 00:22:16 | - | 00:23:56 | - | - | - | 00:15:43 | 01:38:14 |
| S03 | 1 | 00:24:41 | 00:25:07 | 00:09:34 | 00:27:04 | - | - | - | - | 00:03:11 | 01:29:37 |
| S04 | 2 | 00:16:22 | 00:16:09 | 00:17:57 | - | 00:32:17 | - | 00:26:00 | 00:14:28 | 00:13:17 | 02:16:30 |
| S05 | 2 | 00:25:47 | 00:20:05 | 00:19:11 | - | - | 00:26:36 | - | - | 00:08:42 | 01:40:22 |
| S06 | 2 | 00:22:08 | 00:16:45 | 00:17:27 | 00:25:27 | - | - | - | - | 00:02:29 | 01:24:16 |
| S07 | 3 | 00:20:13 | 00:23:38 | 00:16:16 | - | 00:26:40 | - | - | - | 00:15:16 | 01:42:02 |
| S08 | 3 | 00:19:47 | 00:20:10 | 00:15:49 | - | - | 00:21:29 | - | - | 00:03:57 | 01:21:11 |
| S09 | 3 | 00:18:18 | 00:16:33 | 00:18:05 | 00:27:40 | - | - | 00:23:47 | 00:15:57 | 00:05:03 | 02:05:24 |
| S10 | 4 | 00:25:18 | 00:24:02 | 00:21:07 | - | - | 00:26:50 | - | - | 00:13:37 | 01:50:54 |
| S11 | 4 | 00:17:13 | 00:34:10 | - | 00:33:30 | - | - | - | - | 00:08:20 | 01:33:13 |
| S12 | 4 | 00:24:24 | 00:26:29 | 00:28:18 | - | 00:31:17 | - | - | - | 00:10:33 | 02:01:00 |
| S13 | 5 | 00:22:28 | 00:19:11 | 00:20:07 | - | - | 00:24:08 | - | - | 00:02:59 | 01:28:53 |
| S14 | 5 | 00:13:27 | 00:16:07 | 00:15:44 | 00:28:18 | - | - | 00:26:57 | 00:19:23 | 00:35:15 | 02:35:11 |
| S15 | 5 | 00:27:55 | 00:24:44 | 00:25:14 | - | 00:29:57 | - | - | - | 00:07:26 | 01:55:17 |
| S16 | 6 | 00:23:11 | 00:17:25 | 00:20:22 | - | - | 00:20:17 | - | - | 00:16:24 | 01:37:38 |
| S17 | 6 | 00:18:42 | 00:19:59 | 00:15:45 | 00:24:08 | - | - | - | - | 00:02:08 | 01:20:41 |
| S18 | 6 | 00:20:02 | 00:20:53 | 00:20:56 | - | 00:37:01 | - | - | - | 00:14:51 | 01:53:43 |
| Min. | 00:13:27 | 00:16:07 | 00:09:34 | 00:24:08 | 00:23:56 | 00:15:51 | 00:23:42 | 00:14:28 | 00:02:08 | 01:20:41 | |
| Avg. | 00:21:00 | 00:20:58 | 00:18:59 | 00:27:41 | 00:30:11 | 00:22:32 | 00:25:06 | 00:16:05 | 00:10:34 | 01:46:25 | |
| Max. | 00:27:55 | 00:34:10 | 00:28:18 | 00:33:30 | 00:37:01 | 00:26:50 | 00:26:57 | 00:19:23 | 00:35:15 | 02:35:11 | |
| Sum | 06:17:54 | 06:17:22 | 05:22:46 | 02:46:07 | 03:01:08 | 02:15:11 | 01:40:26 | 01:04:21 | 03:10:10 | 31:55:26 | |
| Class Categories [CC] | M | C | Class Labels [CL] | |||
|---|---|---|---|---|---|---|
| Icon | ID | Name | Nr. | List | ||
![]() | Main Activity | X | 15 | CL001|Synchronization;
CL002|Confirming with Pen;
CL003|Confirming with Screen;
CL004|Confirming with Button; CL005|Scanning; CL006|Pulling Cart; CL007|Pushing Cart; CL008|Handling Upwards; CL009|Handling Centered; CL010|Handling Downwards; CL011|Walking; CL012|Standing; CL013|Sitting; CL014|Another Main Activity; CL015|Main Activity Unknown | ||
![]() | Sub-Activity–Legs | X | 8 |
CL016|Gait Cycle;
CL017|Step;
CL018|Standing Still;
CL019|Sitting;
CL020|Squat;
CL021|Lunges; CL022|Another Leg Activity; CL023|Leg Activity Unknown | ||
![]() | Sub-Activity–Torso | X | 6 |
CL024|No Bending;
CL025|Slightly Bending;
CL026|Strongly Bending;
CL027|Torso Rotation; CL028|Another Torso Activity; CL029|Torso Activity Unknown | ||
![]() | Sub-Activity–Left Hand | X | 35 | Primary Position:
CL030|Upwards;
CL031|Centered;
CL032|Downwards;
CL033|Position Unknown Type of Movement: CL034|Reaching, Grasping, Moving, Positioning and Releasing; CL035|Manipulating; CL036|Holding; CL037|No Movement; CL038|Another Movement; CL039|Movement Unknown Object: CL040|No Object; CL041|Large Item; CL042|Medium Item; CL043|Small Item; CL044|Tool; CL045|Cart; CL046|Load Carrier; CL047|Cardboard Box; CL048|On Body; CL049|Another Logistic Object; CL050|No Logistic Object; CL051|Object Unknown Tool: CL052|Portable Data Terminal; CL053|Glove Scanner; CL054|Plastic Bag; CL055|Picking List; CL056|Pen; CL057|Button; CL058|Computer; CL059|Bubble Wrap; CL060|Tape Dispenser; CL061|Knife; CL062|Shipping/Return Label; CL063|Elastic Band; CL064|Another Tool | ||
![]() | Sub-Activity–Right Hand | X | 35 | Primary Position:
CL065|Upwards;
CL066|Centered;
CL067|Downwards;
CL068|Position Unknown Type of Movement: CL069|Reaching, Grasping, Moving, Positioning and Releasing; CL070|Manipulating; CL071|Holding; CL072|No Movement; CL073|Another Movement; CL074|Movement Unknown Object: CL075|No Object; CL076|Large Item; CL077|Medium Item; CL078|Small Item; CL079|Tool; CL080|Cart; CL081|Load Carrier; CL082|Cardboard Box; CL083|On Body; CL084|Another Logistic Object; CL085|No Logistic Object; CL086|Object Unknown Tool: CL087|Portable Data Terminal; CL088|Glove Scanner; CL089|Plastic Bag; CL090|Picking List; CL091|Pen; CL092|Button; CL093|Computer; CL094|Bubble Wrap; CL095|Tape Dispenser; CL096|Knife; CL097|Shipping/Return Label; CL098|Elastic Band; CL099|Another Tool | ||
![]() | Order | X | 5 | CL100|2904; CL101|2905; CL102|2906; CL103|No Order; CL104|Order Unknown | ||
![]() | Information Technology | X | 5 | CL105|List and Pen; CL106|List and Glove Scanner; CL107|Portable Data Terminal; CL108|No Information Technology; CL109|Information Technology Unknown | ||
![]() | High-Level Process | X | 4 | CL110|Retrieval; CL111|Storage; CL112|Another High-Level Process; CL113|High-Level Process Unknown | ||
![]() | Mid-Level Process | X | 10 |
CL114|Preparing Order;
CL115|Picking–Travel Time;
CL116|Picking–Pick Time;
CL117|Unpacking; CL118|Packing; CL119|Storing–Travel Time; CL120|Storing–Store Time; CL121|Finalizing Order; CL122|Another Mid-Level Process; CL123|Mid-Level Process Unknown | ||
![]() | Low-Level Process | X | 31 | CL124|Collecting Order and Hardware; CL125|Collecting Cart; CL126|Collecting Empty Cardboard Boxes; CL127|Collecting Packed Cardboard Boxes; CL128|Transporting a Cart to the Base; CL129|Transporting to the Packaging/Sorting Area; CL130|Handing Over Packed Cardboard Boxes; CL131|Returning Empty Cardboard Boxes; CL132|Returning Cart; CL133|Returning Hardware; CL134|Waiting; CL135|Reporting and Clarifying the Incident; CL136|Removing Cardboard Box/Item from the Cart; CL137|Moving to the Next Position; CL138|Placing Items on a Rack; CL139|Retrieving Items; CL140|Moving to a Cart; CL141|Placing Cardboard Box/Item on a Table; CL142|Opening Cardboard Box; CL143|Disposing of Filling Material or Shipping Label; CL144|Sorting; CL145|Filling Cardboard Box with Filling Material; CL146|Printing Shipping Label and Return Slip; CL147|Preparing or Adding Return Label; CL148|Attaching Shipping Label; CL149|Removing Elastic Band; CL150|Sealing Cardboard Box; CL151|Placing Cardboard Box/Item in a Cart; CL152|Tying Elastic Band Around Cardboard; CL153|Another Low-Level Process; CL154|Low-Level Process Unknown | ||
![]() | Location–Human | X | 26 | Main Area:
CL155|Office;
CL156|Cart Area;
CL157|Cardboard Box Area;
CL158|Base;
CL159|Packing/Sorting Area;
CL160|Issuing/Receiving Area;
CL161|Path;
CL162|Cross Aisle Path;
CL163|Aisle Path Path: CL164|Path (Office); CL165|Path (Cardboard Box Area); CL166|Path (Cart Area); CL167|Path (Issuing Area) Cross Aisle Path: CL168|1–2; CL169|2–3; CL170|3–4; CL171|4–5 Aisle Path: CL172|1; CL173|2; CL174|3; CL175|4; CL176|5; CL177|Front; CL178|Back Other: CL179|Another Location; CL180|Location Unknown | ||
![]() | Location–Cart | X | 27 | Main Area:
CL181|Transition between Areas;
CL182|Office;
CL183|Cart Area;
CL184|Cardboard Box Area;
CL185|Base;
CL186|Packing/Sorting Area;
CL187|Issuing/Receiving Area;
CL188|Path;
CL189|Cross Aisle Path;
CL190|Aisle Path Path: CL191|Path (Office); CL192|Path (Cardboard Box Area); CL193|Path (Cart Area); CL194|Path (Issuing Area) Cross Aisle Path: CL195|1–2; CL196|2–3; CL197|3–4; CL198|4–5 Aisle Path: CL199|1; CL200|2; CL201|3; CL202|4; CL203|5; CL204|Front; CL205|Back Other: CL206|Another Location; CL207|Location Unknown | ||
| Class Category | Annotation | Revision | |||
|---|---|---|---|---|---|
| Total | Ratio | Total | Ratio | ||
| [hh:mm:ss] | |||||
| CC01 | Main Activity | 172:35:30 | 0:05:24 | – | – |
| CC02 | Sub-Activity–Legs | 278:44:27 | 0:08:44 | 68:12:31 | 0:02:08 |
| CC03 | Sub-Activity–Torso | 108:12:37 | 0:03:23 | 76:13:01 | 0:02:23 |
| CC04 | Sub-Activity–Left Hand | 384:24:34 | 0:12:02 | 71:15:00 | 0:02:14 |
| CC05 | Sub-Activity–Right Hand | 378:12:41 | 0:11:51 | 73:46:51 | 0:02:19 |
| CC06 | Order | 3:13:20 | 0:00:06 | 1:01:40 | 0:00:02 |
| CC07 | Information Technology | ||||
| CC08 | High-Level Process | ||||
| CC09 | Mid-Level Process | 129:15:32 | 0:04:03 | 39:20:41 | 0:01:14 |
| CC10 | Low-Level Process | ||||
| CC11 | Location–Human | 92:08:38 | 0:02:53 | 22:51:42 | 0:00:43 |
| CC12 | Location–Cart | 25:19:00 | 0:00:48 | 8:30:55 | 0:00:16 |
| Total | All Categories | 1572:06:19 | 0:49:15 | 361:12:21 | 0:11:19 |
| Class Category | Cohen’s/Light’s Kappa [%] | ||
|---|---|---|---|
| ID | Name | Annotation | Revision |
| CC01 | Main Activity | ||
| CC02 | Sub-Activity–Legs | ||
| CC03 | Sub-Activity–Torso | ||
| CC04 | Sub-Activity–Left Hand | 71.32 | 78.35 |
| CC05 | Sub-Activity–Right Hand | ||
| CC06 | Order | ||
| CC07 | Information Technology | ||
| CC08 | High-Level Process | ||
| CC09 | Mid-Level Process | ||
| CC10 | Low-Level Process | ||
| CC11 | Location–Human | ||
| CC12 | Location–Cart | ||
| Main Activity | Metric | |
|---|---|---|
| Recall | Precision | |
| Confirm with Pen | ||
| Confirm with Screen | ||
| Confirm with Button | ||
| Scan | ||
| Pull | ||
| Push | ||
| Handling Upwards | ||
| Handling Centered | ||
| Handling Downwards | ||
| Walking | ||
| Standing | ||
| Metric | Softmax | Attributes |
|---|---|---|
| Acc [%] | 72.12 | 74.62 |
| wF1 [%] | 70.40 | 73.70 |
| Main Activity | Confusion Matrix | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| Confirm with | Scan | Pull | Push | Handling | Walk. | Stand. | |||||
| Pen | Screen | Button | Up. | Cen. | Down. | ||||||
| Confirm with Pen | 124 | 0 | 0 | 32 | 7 | 43 | 1948 | 1623 | 83 | 121 | 87 |
| Confirm with Screen | 0 | 0 | 5 | 0 | 1 | 12 | 45 | 367 | 42 | 106 | 107 |
| Confirm with Button | 0 | 0 | 46 | 3 | 0 | 3 | 161 | 609 | 167 | 155 | 2 |
| Scan | 0 | 63 | 1 | 291 | 6 | 9 | 340 | 1924 | 534 | 156 | 600 |
| Pull | 0 | 0 | 0 | 0 | 2347 | 775 | 6 | 374 | 3 | 3 | 1 |
| Push | 0 | 0 | 0 | 0 | 249 | 6276 | 0 | 411 | 0 | 3 | 5 |
| Handling Upwards | 11 | 1 | 12 | 191 | 19 | 58 | 7899 | 4513 | 14 | 119 | 29 |
| Handling Centered | 1 | 26 | 4 | 502 | 319 | 883 | 3251 | 64,222 | 2079 | 3157 | 1805 |
| Handling Down. | 0 | 1 | 1 | 51 | 0 | 2 | 63 | 4858 | 6711 | 467 | 190 |
| Walking | 0 | 0 | 6 | 2 | 15 | 303 | 155 | 5398 | 62 | 20,004 | 598 |
| Standing | 0 | 0 | 5 | 462 | 40 | 93 | 714 | 4906 | 404 | 714 | 15,473 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Niemann, F.; Rueda, F.M.; Al Kfari, M.K.; Nair, N.R.; Schauten, D.; Kretschmer, V.; Lüdtke, S.; Kirchheim, A. DaRA Dataset: Combining Wearable Sensors, Location Tracking, and Process Knowledge for Enhanced Human Activity and Human Context Recognition in Warehousing. Sensors 2026, 26, 739. https://doi.org/10.3390/s26020739
Niemann F, Rueda FM, Al Kfari MK, Nair NR, Schauten D, Kretschmer V, Lüdtke S, Kirchheim A. DaRA Dataset: Combining Wearable Sensors, Location Tracking, and Process Knowledge for Enhanced Human Activity and Human Context Recognition in Warehousing. Sensors. 2026; 26(2):739. https://doi.org/10.3390/s26020739
Chicago/Turabian StyleNiemann, Friedrich, Fernando Moya Rueda, Moh’d Khier Al Kfari, Nilah Ravi Nair, Dustin Schauten, Veronika Kretschmer, Stefan Lüdtke, and Alice Kirchheim. 2026. "DaRA Dataset: Combining Wearable Sensors, Location Tracking, and Process Knowledge for Enhanced Human Activity and Human Context Recognition in Warehousing" Sensors 26, no. 2: 739. https://doi.org/10.3390/s26020739
APA StyleNiemann, F., Rueda, F. M., Al Kfari, M. K., Nair, N. R., Schauten, D., Kretschmer, V., Lüdtke, S., & Kirchheim, A. (2026). DaRA Dataset: Combining Wearable Sensors, Location Tracking, and Process Knowledge for Enhanced Human Activity and Human Context Recognition in Warehousing. Sensors, 26(2), 739. https://doi.org/10.3390/s26020739













