Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,322)

Search Parameters:
Keywords = robot architecture

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
39 pages, 1100 KB  
Article
Generalized Kinematic Modeling of Wheeled Mobile Robots: A Unified Framework for Heterogeneous Architectures
by Jesús Said Pantoja-García, Alejandro Rodríguez-Molina, Miguel Gabriel Villarreal-Cervantes, Andrés Abraham Palma-Huerta, Mario Aldape-Pérez and Jacobo Sandoval-Gutiérrez
Mathematics 2026, 14(3), 415; https://doi.org/10.3390/math14030415 - 25 Jan 2026
Abstract
The increasing heterogeneity of wheeled mobile robot (WMR) architectures, including differential-drive, Ackermann, omnidirectional, and reconfigurable platforms, poses a major challenge for defining a unified, scalable kinematic representation. Most existing formulations are tailored to specific mechanical layouts, limiting analytical coherence, cross-platform interoperability, and the [...] Read more.
The increasing heterogeneity of wheeled mobile robot (WMR) architectures, including differential-drive, Ackermann, omnidirectional, and reconfigurable platforms, poses a major challenge for defining a unified, scalable kinematic representation. Most existing formulations are tailored to specific mechanical layouts, limiting analytical coherence, cross-platform interoperability, and the systematic reuse of modeling, odometry, and motion-related algorithms. This work introduces a generalized kinematic modeling framework that provides a mathematically consistent formulation applicable to arbitrary WMR configurations. Wheel–ground velocity relationships and non-holonomic constraints are expressed through a concise vector formulation that maps wheel motions to chassis velocities, ensuring consistency with established models while remaining independent of the underlying mechanical structure. A parameterized wheel descriptor encodes all relevant geometric and kinematic properties, enabling the modular assembly of complete robot models by aggregating wheel-level relations. The framework is evaluated through numerical simulations on four structurally distinct platforms: differential-drive, Ackermann, three-wheel omnidirectional (3,0), and 4WD. Results show that the proposed formulation accurately reproduces the expected kinematic behavior across these fundamentally different architectures and provides a coherent and consistent representation of their motion. The unified representation further provides a common kinematic backbone that is directly compatible with odometry, motion-control, and simulation pipelines, facilitating the systematic retargeting of algorithms across heterogeneous robot platforms without architecture-specific reformulation. Additional simulation studies under realistic physics-based conditions show that the proposed formulation preserves coherent kinematic behavior during complex trajectory execution and supports the explicit incorporation of geometric imperfections, such as wheel mounting misalignments, when such parameters are available. By consolidating traditionally separate derivations into a single coherent formulation, this work establishes a rigorous, scalable, and architecture-agnostic foundation for unified kinematic modeling of wheeled mobile robots, with particular relevance for modular, reconfigurable, and cross-architecture robotic systems. Full article
(This article belongs to the Special Issue Mathematical Modelling and Applied Statistics)
23 pages, 2388 KB  
Article
Action-Aware Multimodal Wavelet Fusion Network for Quantitative Elbow Motor Function Assessment Using sEMG and Robotic Kinematics
by Zilong Song, Pei Zhu, Cuiwei Yang, Daomiao Wang, Jialiang Song, Daoyu Wang, Fanfu Fang and Yixi Wang
Sensors 2026, 26(3), 804; https://doi.org/10.3390/s26030804 - 25 Jan 2026
Abstract
Accurate upper-limb motor assessment is critical for post-stroke rehabilitation but relies on subjective clinical scales. This study proposes the Action-Aware Multimodal Wavelet Fusion Network (AMWFNet), integrating surface electromyography (sEMG) and robotic kinematics for automated Fugl-Meyer Assessment (FMA-UE)-aligned quantification. Continuous Wavelet Transform (CWT) converts [...] Read more.
Accurate upper-limb motor assessment is critical for post-stroke rehabilitation but relies on subjective clinical scales. This study proposes the Action-Aware Multimodal Wavelet Fusion Network (AMWFNet), integrating surface electromyography (sEMG) and robotic kinematics for automated Fugl-Meyer Assessment (FMA-UE)-aligned quantification. Continuous Wavelet Transform (CWT) converts heterogeneous signals into unified time-frequency scalograms. A learnable modality gating mechanism dynamically weights physiological and kinematic features, while action embeddings encode task contexts across 18 standardized reaching tasks. Validated on 40 participants (20 post-stroke, 20 healthy), AMWFNet achieved 94.68% accuracy in six-class classification, outperforming baselines by 9.17% (Random Forest: 85.51%, SVM: 85.30%, 1D-CNN: 91.21%). The lightweight architecture (1.27M parameters, 922ms inference) enables real-time assessment-training integration in rehabilitation robots, providing an objective, efficient solution. Full article
(This article belongs to the Special Issue Advances in Robotics and Sensors for Rehabilitation)
11 pages, 5421 KB  
Article
Underground Multi-Robot Systems at Work: A Revolution in Mining
by Victor Vigara Puche, Kashish Verma and Matteo Fumagalli
Appl. Sci. 2026, 16(3), 1212; https://doi.org/10.3390/app16031212 - 24 Jan 2026
Viewed by 41
Abstract
The growing global demand for critical raw materials has highlighted the need for autonomous systems in abandoned underground mines. We propose a multi-robot coordination architecture using Hierarchical Finite State Machines (HFSMs) for sequential task execution in GPS-denied, infrastructure-less environments. Unlike existing centralized approaches, [...] Read more.
The growing global demand for critical raw materials has highlighted the need for autonomous systems in abandoned underground mines. We propose a multi-robot coordination architecture using Hierarchical Finite State Machines (HFSMs) for sequential task execution in GPS-denied, infrastructure-less environments. Unlike existing centralized approaches, our system enables each robot to execute its own HFSM behavior triggered through inter-robot communication, eliminating dependency on persistent connectivity. We implemented and validated this architecture using a Deployer robot and a Stinger robot within the EU Horizon PERSEPHONE project. Experimental validation demonstrated successful coordination both with persistent connectivity and during network interruptions, proving the system’s fault tolerance capabilities. The system successfully executed sequential deployment and anchoring tasks, demonstrating that this coordination approach enables multi-robot coordination without requiring persistent connectivity, thereby addressing critical limitations for autonomous operations in underground environments. Full article
(This article belongs to the Special Issue Intelligent Drilling Technology: Modeling and Application)
Show Figures

Figure 1

22 pages, 995 KB  
Review
Stroke Rehabilitation, Novel Technology and the Internet of Medical Things
by Ana Costa, Eric Schmalzried, Jing Tong, Brandon Khanyan, Weidong Wang, Zhaosheng Jin and Sergio D. Bergese
Brain Sci. 2026, 16(2), 124; https://doi.org/10.3390/brainsci16020124 - 24 Jan 2026
Viewed by 48
Abstract
Stroke continues to impose an enormous morbidity and mortality burden worldwide. Stroke survivors often incur debilitating consequences that impair motor function, independence in activities of daily living and quality of life. Rehabilitation is a pivotal intervention to minimize disability and promote functional recovery [...] Read more.
Stroke continues to impose an enormous morbidity and mortality burden worldwide. Stroke survivors often incur debilitating consequences that impair motor function, independence in activities of daily living and quality of life. Rehabilitation is a pivotal intervention to minimize disability and promote functional recovery following a stroke. The Internet of Medical Things, a network of connected medical devices, software and health systems that collect, store and analyze health data over the internet, is an emerging resource in neurorehabilitation for stroke survivors. Technologies such as asynchronous transmission to handle intermittent connectivity, edge computing to conserve bandwidth and lengthen device life, functional interoperability across platforms, security mechanisms scalable to resource constraints, and hybrid architectures that combine local processing with cloud synchronization help bridge the digital divide and infrastructure limitations in low-resource environments. This manuscript reviews emerging rehabilitation technologies such as robotic devices, virtual reality, brain–computer interfaces and telerehabilitation in the setting of neurorehabilitation for stroke patients. Full article
Show Figures

Figure 1

19 pages, 1007 KB  
Review
Machine Learning-Powered Vision for Robotic Inspection in Manufacturing: A Review
by David Yevgeniy Patrashko and Vladimir Gurau
Sensors 2026, 26(3), 788; https://doi.org/10.3390/s26030788 - 24 Jan 2026
Viewed by 84
Abstract
Machine learning (ML)-powered vision for robotic inspection has accelerated with smart manufacturing, enabling automated defect detection and classification and real-time process optimization. This review provides insight into the current landscape and state-of-the-art practices in smart manufacturing quality control (QC). More than 50 studies [...] Read more.
Machine learning (ML)-powered vision for robotic inspection has accelerated with smart manufacturing, enabling automated defect detection and classification and real-time process optimization. This review provides insight into the current landscape and state-of-the-art practices in smart manufacturing quality control (QC). More than 50 studies spanning across automotive, aerospace, assembly, and general manufacturing sectors demonstrate that ML-powered vision is technically viable for robotic inspection in manufacturing. The accuracy of defect detection and classification frequently exceeds 95%, with some vision systems achieving 98–100% accuracy in controlled environments. The vision systems use predominantly self-designed convolutional neural network (CNN) architectures, YOLO variants, or traditional ML vision models. However, 77% of implementations remain at the prototype or pilot scale, revealing systematic deployment barriers. A discussion is provided to address the specifics of the vision systems and the challenges that these technologies continue to face. Finally, recommendations for future directions in ML-powered vision for robotic inspection in manufacturing are provided. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Graphical abstract

24 pages, 6152 KB  
Article
Adaptive Realities: Human-in-the-Loop AI for Trustworthy XR Training in Safety-Critical Domains
by Daniele Pretolesi, Georg Regal, Helmut Schrom-Feiertag and Manfred Tscheligi
Multimodal Technol. Interact. 2026, 10(1), 11; https://doi.org/10.3390/mti10010011 - 22 Jan 2026
Viewed by 54
Abstract
Extended Reality (XR) technologies have matured into powerful tools for training in high-stakes domains, from emergency response to search and rescue. Yet current systems often struggle to balance real-time AI-driven personalisation with the need for human oversight and calibrated trust. This article synthesizes [...] Read more.
Extended Reality (XR) technologies have matured into powerful tools for training in high-stakes domains, from emergency response to search and rescue. Yet current systems often struggle to balance real-time AI-driven personalisation with the need for human oversight and calibrated trust. This article synthesizes the programmatic contributions of a multi-study doctoral project to advance a design-and-evaluation framework for trustworthy adaptive XR training. Across six studies, we explored (i) recommender-driven scenario adaptation based on multimodal performance and physiological signals, (ii) persuasive dashboards for trainers, (iii) architectures for AI-supported XR training in medical mass-casualty contexts, (iv) theoretical and practical integration of Human-in-the-Loop (HITL) supervision, (v) user trust and over-reliance in the face of misleading AI suggestions, and (vi) the role of interaction modality in shaping workload, explainability, and trust in human–robot collaboration. Together, these investigations show how adaptive policies, transparent explanation, and adjustable autonomy can be orchestrated into a single adaptation loop that maintains trainee engagement, improves learning outcomes, and preserves trainer agency. We conclude with design guidelines and a research agenda for extending trustworthy XR training into safety-critical environments. Full article
Show Figures

Figure 1

45 pages, 15149 KB  
Review
A New Era in Computing: A Review of Neuromorphic Computing Chip Architecture and Applications
by Guang Chen, Meng Xu, Yuying Chen, Fuge Yuan, Lanqi Qin and Jian Ren
Chips 2026, 5(1), 3; https://doi.org/10.3390/chips5010003 - 22 Jan 2026
Viewed by 84
Abstract
Neuromorphic computing, an interdisciplinary field combining neuroscience and computer science, aims to create efficient, bio-inspired systems. Different from von Neumann architectures, neuromorphic systems integrate memory and processing units to enable parallel, event-driven computation. By simulating the behavior of biological neurons and networks, these [...] Read more.
Neuromorphic computing, an interdisciplinary field combining neuroscience and computer science, aims to create efficient, bio-inspired systems. Different from von Neumann architectures, neuromorphic systems integrate memory and processing units to enable parallel, event-driven computation. By simulating the behavior of biological neurons and networks, these systems excel in tasks like pattern recognition, perception, and decision-making. Neuromorphic computing chips, which operate similarly to the human brain, offer significant potential for enhancing the performance and energy efficiency of bio-inspired algorithms. This review introduces a novel five-dimensional comparative framework—process technology, scale, power consumption, neuronal models, and architectural features—that systematically categorizes and contrasts neuromorphic implementations beyond existing surveys. We analyze notable neuromorphic chips, such as BrainScaleS, SpiNNaker, TrueNorth, and Loihi, comparing their scale, power consumption, and computational models. The paper also explores the applications of neuromorphic computing chips in artificial intelligence (AI), robotics, neuroscience, and adaptive control systems, while facing challenges related to hardware limitations, algorithms, and system scalability and integration. Full article
14 pages, 4270 KB  
Article
Dual-Arm Coordination of a Tomato Harvesting Robot with Subtask Decoupling and Synthesizing
by Binhao Chen, Liang Gong, Shenghan Xie, Xuhao Zhao, Peixin Gao, Hefei Luo, Cheng Luo, Yanming Li and Chengliang Liu
Agriculture 2026, 16(2), 267; https://doi.org/10.3390/agriculture16020267 - 21 Jan 2026
Viewed by 58
Abstract
Robotic harvesters have the potential to substantially reduce the physical workload of agricultural laborers. However, in complex agricultural environments, traditional single-arm robot path planning methods often struggle to accomplish fruit harvesting tasks due to the presence of collision avoidance requirements and orientation constraints [...] Read more.
Robotic harvesters have the potential to substantially reduce the physical workload of agricultural laborers. However, in complex agricultural environments, traditional single-arm robot path planning methods often struggle to accomplish fruit harvesting tasks due to the presence of collision avoidance requirements and orientation constraints during grasping. In this work, we design a dual-arm tomato harvesting robot and propose a reinforcement learning-based cooperative control algorithm tailored to the dual-arm system. First, a deep learning-based semantic segmentation network is employed to extract the spatial locations of tomatoes and branches from sensory data. Building upon this perception module, we develop a reinforcement learning-based cooperative path planning approach to address inter-arm collision avoidance and end-effector orientation constraints during the harvesting process. Furthermore, a task-driven policy network architecture is introduced to decouple the complex harvesting task into structured subproblems, thereby enabling more efficient learning and improved performance. Simulation and experimental results demonstrate that the proposed method can generate collision-free harvesting trajectories that satisfy dual-arm orientation constraints, significantly improving the tomato harvesting success rate. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

30 pages, 965 KB  
Article
Guarded Swarms: Building Trusted Autonomy Through Digital Intelligence and Physical Safeguards
by Uwe M. Borghoff, Paolo Bottoni and Remo Pareschi
Future Internet 2026, 18(1), 64; https://doi.org/10.3390/fi18010064 - 21 Jan 2026
Viewed by 102
Abstract
Autonomous UAV/UGV swarms increasingly operate in contested environments where purely digital control architectures are vulnerable to cyber compromise, communication denial, and timing faults. This paper presents Guarded Swarms, a hybrid framework that combines digital coordination with hardware-level analog safety enforcement. The architecture builds [...] Read more.
Autonomous UAV/UGV swarms increasingly operate in contested environments where purely digital control architectures are vulnerable to cyber compromise, communication denial, and timing faults. This paper presents Guarded Swarms, a hybrid framework that combines digital coordination with hardware-level analog safety enforcement. The architecture builds on Topic-Based Communication Space Petri Nets (TB-CSPN) for structured multi-agent coordination, extending this digital foundation with independent analog guard channels—thrust clamps, attitude limiters, proximity sensors, and emergency stops—that operate in parallel at the actuator interface. Each channel can unilaterally veto unsafe commands within microseconds, independently of software state. The digital–analog interface is formalized via timing contracts that specify sensor-consistency windows and actuation latency bounds. A two-robot case study demonstrates token-based arbitration at the digital level and OR-style inhibition at the analog level. The framework ensures local safety deterministically while maintaining global coordination as a best-effort property. This paper presents an architectural contribution establishing design principles and interface contracts. Empirical validation remains future work. Full article
(This article belongs to the Special Issue Intelligent Agents and Their Application)
Show Figures

Graphical abstract

19 pages, 5815 KB  
Article
MonoPrior-Fusion: Monocular-Prior-Guided Multi-Frame Depth Estimation with Multi-Scale Geometric Fusion
by Zhiwei Lin, Bohan Sun, Zhan Zhang, Linrui Qian and Nianyu Yi
Sensors 2026, 26(2), 712; https://doi.org/10.3390/s26020712 - 21 Jan 2026
Viewed by 137
Abstract
Precise 3D perception is critical for indoor robotics, augmented reality, and autonomous navigation. However, existing multi-frame depth estimation methods often suffer from significant performance degradation in challenging indoor scenarios characterized by weak textures, non-Lambertian surfaces, and complex layouts. To address these limitations, we [...] Read more.
Precise 3D perception is critical for indoor robotics, augmented reality, and autonomous navigation. However, existing multi-frame depth estimation methods often suffer from significant performance degradation in challenging indoor scenarios characterized by weak textures, non-Lambertian surfaces, and complex layouts. To address these limitations, we propose MonoPrior-Fusion (MPF), a novel framework that integrates pixel-wise monocular priors directly into the multi-view matching process. Specifically, MPF modulates cost-volume hypotheses to disambiguate matches and employs a hierarchical fusion architecture across multiple scales to propagate global and local geometric information. Additionally, a geometric consistency loss based on virtual planes is introduced to enhance global 3D coherence. Extensive experiments on ScanNetV2, 7Scenes, TUM RGB-D, and GMU Kitchens demonstrate that MPF achieves significant improvements over state-of-the-art multi-frame baselines and generalizes well across unseen domains. Furthermore, MPF yields more accurate and complete 3D reconstructions when integrated into a volumetric fusion pipeline, proving its effectiveness for dense mapping tasks. The source code will be made publicly available to support reproducibility and future research. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

11 pages, 2695 KB  
Proceeding Paper
Automatic Control of a Flywheel Actuator for Mobile Platform Stabilization
by Alina Fazylova, Kuanysh Alipbayev, Nazgul Kaliyeva, Yerkin Orazaly and Teodor Iliev
Eng. Proc. 2026, 122(1), 25; https://doi.org/10.3390/engproc2026122025 - 20 Jan 2026
Viewed by 68
Abstract
This paper presents the design, modeling and control of a flywheel actuator for mobile platform stabilization. A Lagrangian-based model couples platform mechanics with DC-motor electromechanics. Analytical calculations estimate natural frequencies, damping and actuator limits. Numerical simulations in Python 3.12 evaluate cascade and state-feedback [...] Read more.
This paper presents the design, modeling and control of a flywheel actuator for mobile platform stabilization. A Lagrangian-based model couples platform mechanics with DC-motor electromechanics. Analytical calculations estimate natural frequencies, damping and actuator limits. Numerical simulations in Python 3.12 evaluate cascade and state-feedback controllers for suppressing free oscillations and rejecting external disturbances. Additional studies examine filtering to improve measurement quality and unloading strategies to avoid actuator saturation. The results validate the proposed control architecture and demonstrate its applicability to robotic and energy systems operating under dynamic loads. Full article
Show Figures

Figure 1

24 pages, 69667 KB  
Article
YOLO-ELS: A Lightweight Cherry Tomato Maturity Detection Algorithm
by Zhimin Tong, Yu Zhou, Changhao Li, Changqing Cai and Lihong Rong
Appl. Sci. 2026, 16(2), 1043; https://doi.org/10.3390/app16021043 - 20 Jan 2026
Viewed by 74
Abstract
Within the domain of intelligent picking robotics, fruit recognition and positioning are essential. Challenging conditions such as varying light, occlusion, and limited edge-computing power compromise fruit maturity detection. To tackle these issues, this paper proposes a lightweight algorithm YOLO-ELS based on YOLOv8n. Specifically, [...] Read more.
Within the domain of intelligent picking robotics, fruit recognition and positioning are essential. Challenging conditions such as varying light, occlusion, and limited edge-computing power compromise fruit maturity detection. To tackle these issues, this paper proposes a lightweight algorithm YOLO-ELS based on YOLOv8n. Specifically, we reconstruct the backbone by replacing the bottlenecks in the C2f structure with Edge-Information-Enhanced Modules (EIEM) to prioritize morphological cues and filter background redundancy. Furthermore, a Large Separable Kernel Attention (LSKA) mechanism is integrated into the SPPF layer to expand the effective receptive field for multi-scale targets. To mitigate occlusion-induced errors, a Spatially Enhanced Attention Module (SEAM) is incorporated into the decoupled detection head to enhance feature responses in obscured regions. Finally, the Inner-GIoU loss is adopted to refine bounding box regression and accelerate convergence. Experimental results demonstrate that compared to the YOLOv8n baseline, the proposed YOLO-ELS achieves a 14.8% reduction in GFLOPs and a 2.3% decrease in parameters, while attaining a precision, recall, and mAP@50% of 92.7%, 83.9%, and 92.0%, respectively. When compared with mainstream models such as DETR, Faster-RCNN, SSD, TOOD, YOLOv5s, and YOLO11n, the mAP@50% is improved by 7.0%, 4.7%, 11.4%, 8.6%, 3.1%, and 3.2%. Deployment tests on the NVIDIA Jetson Orin Nano Super edge platform yield an inference latency of 25.2 ms and a detection speed of 28.2 FPS, successfully meeting the real-time operational requirements of automated harvesting systems. These findings confirm that YOLO-ELS effectively balances high detection accuracy with lightweight architecture, providing a robust technical foundation for intelligent fruit picking in resource-constrained greenhouse environments. Full article
(This article belongs to the Section Agricultural Science and Technology)
Show Figures

Figure 1

25 pages, 12600 KB  
Article
Underwater Object Recovery Using a Hybrid-Controlled ROV with Deep Learning-Based Perception
by Inés Pérez-Edo, Salvador López-Barajas, Raúl Marín-Prades and Pedro J. Sanz
J. Mar. Sci. Eng. 2026, 14(2), 198; https://doi.org/10.3390/jmse14020198 - 18 Jan 2026
Viewed by 311
Abstract
The deployment of large remotely operated vehicles (ROVs) or autonomous underwater vehicles (AUVs) typically requires support vessels, crane systems, and specialized personnel, resulting in increased logistical complexity and operational costs. In this context, lightweight and modular underwater robots have emerged as a cost-effective [...] Read more.
The deployment of large remotely operated vehicles (ROVs) or autonomous underwater vehicles (AUVs) typically requires support vessels, crane systems, and specialized personnel, resulting in increased logistical complexity and operational costs. In this context, lightweight and modular underwater robots have emerged as a cost-effective alternative, capable of reaching significant depths and performing tasks traditionally associated with larger platforms. This article presents a system architecture for recovering a known object using a hybrid-controlled ROV, integrating autonomous perception, high-level interaction, and low-level control. The proposed architecture includes a perception module that estimates the object pose using a Perspective-n-Point (PnP) algorithm, combining object segmentation from a YOLOv11-seg network with 2D keypoints obtained from a YOLOv11-pose model. In addition, a Natural Language ROS Agent is incorporated to enable high-level command interaction between the operator and the robot. These modules interact with low-level controllers that regulate the vehicle degrees of freedom and with autonomous behaviors such as target approach and grasping. The proposed system is evaluated through simulation and experimental tank trials, including object recovery experiments conducted in a 12 × 8 × 5 m test tank at CIRTESU, as well as perception validation in simulated, tank, and harbor scenarios. The results demonstrate successful recovery of a black box using a BlueROV2 platform, showing that architectures of this type can effectively support operators in underwater intervention tasks, reducing operational risk, deployment complexity, and mission costs. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

27 pages, 11232 KB  
Article
Aerokinesis: An IoT-Based Vision-Driven Gesture Control System for Quadcopter Navigation Using Deep Learning and ROS2
by Sergei Kondratev, Yulia Dyrchenkova, Georgiy Nikitin, Leonid Voskov, Vladimir Pikalov and Victor Meshcheryakov
Technologies 2026, 14(1), 69; https://doi.org/10.3390/technologies14010069 - 16 Jan 2026
Viewed by 229
Abstract
This paper presents Aerokinesis, an IoT-based software–hardware system for intuitive gesture-driven control of quadcopter unmanned aerial vehicles (UAVs), developed within the Robot Operating System 2 (ROS2) framework. The proposed system addresses the challenge of providing an accessible human–drone interaction interface for operators in [...] Read more.
This paper presents Aerokinesis, an IoT-based software–hardware system for intuitive gesture-driven control of quadcopter unmanned aerial vehicles (UAVs), developed within the Robot Operating System 2 (ROS2) framework. The proposed system addresses the challenge of providing an accessible human–drone interaction interface for operators in scenarios where traditional remote controllers are impractical or unavailable. The architecture comprises two hierarchical control levels: (1) high-level discrete command control utilizing a fully connected neural network classifier for static gesture recognition, and (2) low-level continuous flight control based on three-dimensional hand keypoint analysis from a depth camera. The gesture classification module achieves an accuracy exceeding 99% using a multi-layer perceptron trained on MediaPipe-extracted hand landmarks. For continuous control, we propose a novel approach that computes Euler angles (roll, pitch, yaw) and throttle from 3D hand pose estimation, enabling intuitive four-degree-of-freedom quadcopter manipulation. A hybrid signal filtering pipeline ensures robust control signal generation while maintaining real-time responsiveness. Comparative user studies demonstrate that gesture-based control reduces task completion time by 52.6% for beginners compared to conventional remote controllers. The results confirm the viability of vision-based gesture interfaces for IoT-enabled UAV applications. Full article
(This article belongs to the Section Information and Communication Technologies)
Show Figures

Figure 1

24 pages, 2343 KB  
Article
Design and Implementation of a Low-Water-Consumption Robotic System for Cleaning Residential Balcony Glass Walls
by Maria-Alexandra Mielcioiu, Petruţa Petcu, Dumitru Nedelcu, Augustin Semenescu, Narcisa Valter and Ana-Maria Nicolau
Appl. Sci. 2026, 16(2), 945; https://doi.org/10.3390/app16020945 - 16 Jan 2026
Viewed by 122
Abstract
Manual window cleaning in high-rise urban buildings is labor-intensive, risky, and resource-inefficient. This study addresses these challenges by investigating a resource-aware mechatronic architecture through the design, development, and experimental validation of a modular Automated Window Cleaning System (AWCS). Unlike conventional open-loop solutions, the [...] Read more.
Manual window cleaning in high-rise urban buildings is labor-intensive, risky, and resource-inefficient. This study addresses these challenges by investigating a resource-aware mechatronic architecture through the design, development, and experimental validation of a modular Automated Window Cleaning System (AWCS). Unlike conventional open-loop solutions, the AWCS integrates mechanical scrubbing with a closed-loop fluid management system, featuring precise dispensing and vacuum-assisted recovery. The system is governed by a deterministic finite state machine implemented on an ESP32 microcontroller, enabling low-latency IoT connectivity and autonomous operation. Two implementation variants—integrated and retrofit—were validated to ensure structural adaptability. Experimental results across 30 cycles demonstrate a cleaning efficiency of ~2 min/m2, a water consumption of <150 mL/m2 (representing a >95% reduction compared to manual methods), and an optical cleaning efficacy of 96.9% ± 1.4%. Safety protocols were substantiated through a calculated mechanical safety factor of 6.12 for retrofit applications. This research establishes the AWCS as a sustainable, safe, and scalable solution for autonomous building maintenance, contributing to the advancement of resource-circular domestic robotics and smart home automation. Full article
Show Figures

Figure 1

Back to TopTop