Next Article in Journal
Running-Time Analysis of Brain Storm Optimization Based on Average Gain Model
Next Article in Special Issue
Whole-Body Dynamics for Humanoid Robot Fall Protection Trajectory Generation with Wall Support
Previous Article in Journal
A Multi-Objective Sine Cosine Algorithm Based on a Competitive Mechanism and Its Application in Engineering Design Problems
Previous Article in Special Issue
Integrating Egocentric and Robotic Vision for Object Identification Using Siamese Networks and Superquadric Estimations in Partial Occlusion Scenarios
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Teleoperated Grasping Using Data Gloves Based on Fuzzy Logic Controller

1
School of Automotive Engineering, Harbin Institute of Technology (Weihai), Weihai 264209, China
2
China North Vehicle Research Institute, Beijing 100072, China
3
Yangtze River Delta HIT Robot Technology Research Institute, Wuhu 241060, China
*
Author to whom correspondence should be addressed.
Biomimetics 2024, 9(2), 116; https://doi.org/10.3390/biomimetics9020116
Submission received: 23 November 2023 / Revised: 29 January 2024 / Accepted: 6 February 2024 / Published: 15 February 2024
(This article belongs to the Special Issue Intelligent Human-Robot Interaction: 2nd Edition)

Abstract

:
Teleoperated robots have attracted significant interest in recent years, and data gloves are one of the commonly used devices for their operation. However, existing solutions still encounter two challenges: the ways in which data gloves capture human operational intentions and achieve accurate mapping. In order to address these challenges, we propose a novel teleoperation method using data gloves based on fuzzy logic controller. Firstly, the data are collected and normalized from the flex sensors on data gloves to identify human manipulation intentions. Then, a fuzzy logic controller is designed to convert finger flexion information into motion control commands for robot arms. Finally, experiments are conducted to demonstrate the effectiveness and precision of the proposed method.

1. Introduction

Recently, with human expansion into outer space and the deep sea, teleoperated robots capable of substituting humans in dangerous and harmful environments have become a popular technology [1,2,3]. Teleoperation is a human–robot collaboration system. Operators control distant robots through devices like joysticks and the robots leverage human cognition, creativity, and experience to complete tasks in complex environments. Human intervention in the system can compensate for deficiencies in robot control, sensing, and artificial intelligence, among others [4,5,6]. As of today, many scholars have researched teleoperation technology. Teleoperation robots are usually implemented using vision cameras and force feedback devices such as Geomagic Touch [7].
The vision-based teleoperation method uses visual cameras to capture real-time images of the operator’s gestures. These gesture images are then mapped into command information for the robot arm [8,9,10,11]. Du et al. [12] proposed a markerless hand tracking algorithm for real-time teleoperation of robot arms. They used depth images from Kinect to estimate the position and posture of the index finger and thumb in three-dimensional space. The calculated gestures controlled the posture of end effectors in real time. This method has been validated effectively through pick-and-place object experiments. Handa et al. [13] developed a depth camera-based teleoperation system called DexPilot, where two depth cameras visually enable experimenters to control the movements of robot arms and fingers barehandedly.
However, the vision-based teleoperation method needs to combine camera images to calculate the position and movement of human hands. This method needs to be conducted in a well-lit, obstacle-free environment. Thus, it has high requirements for the experimental environment and is limited by distance.
Robot teleoperation based on force feedback devices utilizes devices like Geomagic Touch to provide haptic feedback [14,15,16]. Operators transmit commands to the robot through the handle and simultaneously receive the interaction information from the robot’s interaction with the external environment. Zhu et al. [17] designed a dual manipulator robotic teleoperation system structure equipped with modular hybrid grippers. Its master end was composed of two Geomagic Touches and one leap motion controller, while the slave end encompassed two mechanical arms fitted with two grippers. The leap motion controller was tasked with tracking human hand movements. Asad et al. [18] presented a novel state convergence control algorithm for bilateral tele-operation of robotic systems, emphasizing a model in state space and a design procedure to compute control gains. In 2022, Asad et al. [19] proposed an extended composite state convergence scheme for multi-primary, multi-secondary teleoperation systems, incorporating a disturbance observer-based architecture to handle uncertainties. Yang et al. [20] introduced a synchronization control algorithm for a networked nonlinear master-slave teleoperation system, addressing operator fatigue in continuous operations.
However, force-feedback teleoperation based on Geomagic Touch requires professional equipment and operating skills. This makes it inconvenient to use, especially in scenarios such as assisting the elderly and disabled.
To overcome these issues, researchers began to focus on more natural, intuitive, and precise teleoperation methods. Data glove technology emerged, which uses built-in sensors to detect hand movement and posture information [21,22,23,24] and transmits these data to the robot arm system. By wearing a data glove, operators can manipulate mechanical arms in real time by changing hand movements. This offers higher operation precision and freedom of freedom, particularly during tasks such as grasping or placement, demonstrating the immense potential of data gloves [25,26,27,28].
Colasanto et al. [29] proposed a method using data gloves to capture finger motion information, mapping human hand movements onto robotic bionic hands. They developed a hybrid mapping combining joint space mapping and fuzzy-based posture mapping, which can achieve both joint mapping based on human hand movements and posture mapping for grasping tasks. Tyagi et al. [30] proposed using a fuzzy inference system to control the movement of an intelligent wheelchair. By wearing gloves equipped with flex sensors, they used their fingers to control wheelchair movement.
However, the above-mentioned data glove-based method faces the following problems: (1) The data glove-based teleoperation control system involves multi-joint motion control. Owing to the kinematic differences between human hands and robot arm systems, teleoperation cannot be simply mapped. (2) Due to varying human skeletal structures and different operator behavior habits, determining ways to process the sensor information from data gloves to accurately capture the manipulation intentions of different operators is key.
Inspired by the above literature review, a teleoperation method using data gloves based on fuzzy logic controller is proposed. Firstly, the human hand’s skeletal structure and movements are analyzed to obtain finger flexion information through flex sensor data from the data gloves. Subsequently, a fuzzy logic controller is designed to implement teleoperation control. The fuzzy logic controller captures human manipulation intentions from finger flexion information and sends motion control commands to the robot arm through fuzzy logic rules. Through wireless communication, real-time data transmission and operation are realized. Finally, experiments verify the effectiveness of teleoperation based on data gloves. The main contributions of this paper are summarized as follows:
  • We propose a teleoperation method using data gloves and normalize the flex sensor data to identify human manipulation intentions.
  • To achieve control of teleoperation, a fuzzy logic controller is designed. The control commands for the joint velocity of the robot arm are obtained through fuzzy logic rules based on finger flexion information.
  • The experimental results indicated that compared to the joint velocity output of the PID controller, the ability of the data gloves to use a fuzzy logic algorithm has better non-linearity and enhances the stability of teleoperated robot arm control.
The remainder of this paper is organized as follows: Section 2 offers an overview of the system and the details of the proposed method. Section 3 presents the experimental evaluation results. Section 4 concludes the paper.

2. Method

2.1. System Overview

As shown in Figure 1, the framework of the teleoperation system based on data gloves consists of two parts: Transmitter and Receiver. The operator wears data gloves and controls the movement of the robot arm in real time by flexing their fingers. The flex sensors and inertial sensors on the data gloves detect the degree of finger flexion. The Transmitter filters and standardizes information about finger flexion; the fuzzy logic controller processes input standard values of finger flexion degrees and outputs control commands for robot arms. The Receiver receives command information for controlling joint velocity on robot arms to control its motion state in real time. Data from flex sensors and robot control commands are transmitted wirelessly. The desired joint velocities given by the fuzzy controller are transmitted to the bottom controller for each joint embedded in the robot arm, which can track the desired commands based on the PID control law.

2.2. The Data Glove

2.2.1. Hand Modeling

Informed by anatomical analysis [31,32], the human hand skeletal model typically exhibits over 20 degrees of freedom, as shown in Figure 2. Fingers are composed of phalanges and finger joints, and they have flexion, extension, adduction, and abduction movement patterns. Each finger has four DOFs: the Meta-Carpophalangeal (MCP) joint has two DOFs, the Proximal–Interphalangeal (PIP) joint has one DOF, and the Distal Interphalangeals (DIP) joint has one DOF.
There are four methods for mapping human hand movements: fingertip mapping, joint mapping, key point mapping, and object-based mapping. Compared with direct manipulation in the end-effector space, joint angle mapping can not only regulate the configuration of the robot arm to ensure its safety, but also effectively complete the given tasks in the complex environments. We employ joint angle mapping. To streamline motion analysis, we simplify the human hand skeletal structure into a linkage structure, as depicted in Figure 3.
In the figure, l1 represents the length of the metacarpal bone, l2 represents the length of the proximal phalanx, l3 represents the length of the distal phalanx, θMCP represents the flexion angle of the Meta-Carpophalangeal joint, θPIP represents the flexion angle of the Proximal-Inter-phalangeal joint, θDIP represents the flexion angle of the Distal Inter-phalangeal joint and β indicates abduction and adduction angles at the Meta-Carpophalangeal joint.

2.2.2. Data Filtering

Due to environmental noise and uncertainty, instability and fluctuations of sensor measurements may occur. To eliminate interference from environmental noise, obtain more accurate and reliable hand bending angles, and enhance control stability, it is necessary to filter the data from the data glove. Figure 4 illustrates the raw data and filtered data of the index finger flexion angle. In this paper, a digital Butterworth low-pass filter is employed to eliminate high-frequency noise from the raw data, smoothing the data and reducing instability. The sensor sample rate of the data glove used in this paper is 90 Hz, with a cutoff frequency set at 5 Hz.

2.2.3. Normalization

In the process of hand modeling, the flexion angle of each finger joint represents its current posture. Flex sensors on data gloves measure the degree of finger flexion. However, due to variations in human skeletal structures, finger sizes, and user habits, individuals might exhibit different physical ranges of finger flexion. To ensure the robustness of gesture mapping across diverse operators, this paper normalizes the degree of finger flexion within a specific range. This standardization facilitates the comparison of finger data across individuals and simplifies data complexity for more direct comparison and analysis.
θ ^ j = 2 θ c θ j 1 ,   j MCP , PIP , DIP ,
where θ c represents the current human finger joint flexion angle outputted by the flex sensor in the data glove, and θ j represents the limit value of finger joint constraints. To eliminate unnatural gestures caused by sensor noise, upper and lower limits of finger joint motion are defined through joint constraints.
Human finger joint movement constraints [33] can be expressed as
0 θ MCP 90 0 θ PIP 110 0 θ DIP 90 15 β 15 .
The data gloves used in this study are equipped with five 2-DOF flex sensors and a 9-DOF IMU for each finger. The direction sensor has an accuracy of ±2.5 degrees, enabling precise tracking of finger movements with a signal delay of less than 5 ms, thus offering excellent real-time performance. The sensor frequency is 90 Hz, which is sufficient for tracking human finger movements. When the fingers bend towards the back of the hand, the standard value gradually changes from 0 to −1; when they bend towards the palm, it changes from 0 to 1; when fingers are straightened, their standard value is set at 0. Using the right index finger as an example, schematic diagrams of three states at standard values −1, 0 and 1 are shown in Figure 5.
The selection of gestures is guided by a thorough consideration of user intuitiveness. We conducted an analysis of common hand movements and gestures to identify that gestures were chosen to ensure users could easily associate them with the desired robot movements. The mapping for gestures to robotic commands is explained in Table 1 below. For the gesture depicted in Table 1, bending and straightening of the left ring finger controls the forward and reverse rotation of each joint.

2.3. Fuzzy Logic Controller

The dynamic characteristics of the teleoperated robot arm system based on data gloves are complex and difficult to represent with precise mathematical models. Fuzzy logic control introduces fuzzy logic into the control system, allowing the use of fuzzy, imprecise information to describe the behavior of teleoperated robot arm systems, thereby enhancing their nonlinearity. Therefore, this paper proposes a method based on fuzzy logic control to control remotely the robot arm.
In Figure 6, left_ring represents the standard value of the left hand’s ring finger flexion; right_x represents the standard values of the right hand’s finger flexions, including the right hand’s thumb PIP joint (right_thumb_PIP), right hand’s index finger (right_index), right hand’s middle finger (right_middle), right hand’s ring finger (right_ring), and right hand’s little finger (right_little); rt_mcp represents the right hand’s thumb MCP joint; velj represents the joint velocity (rad/s) sent to Joint j of the robot arm, j ∈ [1,2,3,4,5].
Fuzzy logic controller is an intelligent control method that simulates artificial behavior based on human experience. Its process includes three steps: fuzzification, fuzzy inference, and defuzzification. This paper uses the standard value of left ring finger flexion and right finger flexion as two inputs for fuzzy logic controller. Joint velocities are considered as the output of the fuzzy controller instead of joint position, because the motion space of the fingers is relatively smaller compared to the motion space of the robot arm joints. Using joint position as the output may result in lower precision due to this kind of mapping.
For data gloves, mapping with the end-effector position of the robot arm is a challenging task byits finger joints. Unlike systems with clear end effectors in series or parallel structure of the general haptic devices, data gloves do not have a distinct end effector in the task space, making direct mapping complicated. When using data gloves for teleoperation, it becomes easier to conduct tasks by mapping the joint of the glove to the corresponding joint of the robot arm. Meanwhile, teleoperating the robot arm in the joint space can well optimize the joint configuration, contributing to safer and more complex movements.
As shown in Figure 7, the 1~6 joints of the UR3e robot arm are remotely controlled, respectively, through the PIP joint of the right thumb, right index finger, right middle finger, right ring finger, right little finger and MCP joint of the right thumb. The magnitude of the standardized finger flexion value indicates the magnitude of the robot arm joint velocity, and, according to left ring finger flexion standard value, changes the direction of rotation for robot arm joints. Based on the above analysis, fuzzy control rule base designed in this paper is shown in Table 2. The inputs are fuzzified into seven subsets: Very Small (VS), Small (S), Moderately Small (MS), Moderate (M), Moderately Large (ML), Large (L), Very Large (VL). The joint velocity of the mechanical arm is fuzzified into five subsets: NB, NS, ZE, PS, PB.
A triangular membership function is used to implement the fuzzification of inputs and outputs in the fuzzy control system. The fuzzy control input membership function designed is shown in Figure 8, and the output membership function is shown in Figure 9. The fuzzy control surface is shown in Figure 10.
The above controller output value is defuzzified to derive an accurate inference result. This paper employs centroid-based defuzzification on robot arm joint velocity to procure precise output values [34].
v j = v e l j μ v e l j ( v e l j ) d v e l j μ v e l j ( v e l j ) d v e l j ,
where velj represents the fuzzy value of the joint velocity of the robot arm; μvelj represents the membership function value of the joint velocity of the robot arm; vj represents the precise value of joint velocity for robot arm.

3. Experiments

The experiments are conducted in two parts: the first experiment is carried out to test the effect of the fuzzy controller on the control of the robot arm joints, and the second experiment is conducted to observe the whole system’s performance in robot arm grasping.

3.1. Experimental Setup

As shown in Figure 11, the experimental platform consists of Prime X Haptic VR data gloves from MANUS, a Lenovo Savior Y9000P laptop, Easy Gripper stepper motor mechanical claw, and UR3e robot arm with a force/torque sensor from Universal Robots company. The data glove is equipped with flex sensors and inertial sensors. The flex sensors can detect the degree of finger flexion, as shown in Figure 12. The experiment communicates between ROS and robot arm wirelessly, realizing cross-platform communication between Windows and Linux based on the distributed operating system. It achieves real-time data transmission between the data glove and the robot arm.
In our experimental setup, we utilize the wireless LAN (Local Area Network) to ensure the finger flexion data collected by the gloves could be transmitted to the controller on Linux.

3.2. Fuzzy Logic Control Testing

In order to validate the effectiveness of the proposed method, we designed two teleoperation robot arm methods: one is based on a PID controller for remotely controlling the robot arm, and the other is based on a fuzzy logic controller for remotely controlling the robot arm. The UR3e robot arm’s Joints 1 to 6 are remotely manipulated through the PIP joint of the right thumb, right index finger, right middle finger, right ring finger, right little finger and MCP joint of the right thumb. Using one-dimensional velocity as an example, we conducted an experimental comparison between the two controllers.
By combining the bending of the left ring finger and the right fingers, remote control of Joints 1–5 of the robot arm was achieved. Specifically, the right thumb controlled the movement speed of Joint 1, the right index finger controlled the movement speed of Joint 2, the right middle finger controlled the movement speed of Joint 3, the right ring finger controlled the movement speed of Joint 4, and the right little controlled the movement speed of Joint 5. The bending and straightening of the left ring finger controlled the forward and reverse motion of each joint, thereby completing the control of the robot arm based on the data glove. The mapping of the standard values of bending in the fingers of the left and right hands to the joint velocities of the robot arm using a PID controller is as follows:
v e l j = r i g h t _ x 0 . 5 ,   l e f t _ r i n g < 0 . 5   and   0 . 5 < r i g h t _ x ( r i g h t _ x 0 . 5 ) ,   0 . 5 < l e f t _ r i n g   and   0 . 5 < r i g h t _ x 0 ,   else
where left_ring represents the standard value of left hand’s ring finger flexion; right_x represents the standard values of right hand’s finger flexions, including right thumb (right_thumb), right index finger (right_index), right middle finger (right_middle), right ring finger (right_ring), and right little finger (right_little); velj represents the joint velocity (rad/s) sent to Joint j of the robot arm, j ∈ [1,2,3,4,5].
The experimental results are shown in Figure 13 and Figure 14. It can be seen that when left hand’s ring finger flexion standard value is less than 0.5 and right hand’s thumb flexion standard value gradually increases from a smaller value to approach 0, joint velocity of Joint 1 of robot arm gradually increases when right hand’s thumb flexion standard value is greater than 0.5 with movement direction towards left. When the right hand’s ring finger flexion standard value is less than 0.5, the velocity of Joint 1 remains at zero, thus indicating no movement. The movement of the robot arm matches the control expectations.
From Figure 14, it appears that compared to the joint velocity output of a PID controller, the data glove’s ability to control the robot arm using a fuzzy logic algorithm leads to better results than that of a PID controller. This is shown by its sensitivity to finger flexion changes. Moreover, its higher non-linearity prevents abrupt changes in joint velocity due to abnormal fluctuations in the data glove’s control values, enhancing the stability of teleoperated robot arm control. According to the experimental results, the measured delay time of the robot arm control system using a fuzzy controller is less than 0.05 s.
In order to validate the teleoperation of the robot arm predicated on fuzzy logic control, the left ring finger governs the robot arm’s rotational direction, while the right hand independently controls the movement velocity of the robot arm’s joints. The standard values of each finger’s flexion and the robot arm’s joint velocity are shown in Figure 15. It can be seen from Figure 15 that the joint motion of the robot arms can follow the flexion motion of the fingers.

3.3. Teleoperated Grasping Testing

Figure 16 shows an object grasping experiment designed to substantiate the comprehensive functionality of the remote operation system in this study. The experimental procedure comprises the following steps: (1) Moving to the initial position. (2) Grasping the object. (3) Shifting to the target position. (4) Releasing the object. (5) Reverting to the initial position.
In the experiment of remotely operating a robot arm to grasp objects, the robot arm first rotates left to above the object. During its descent, it makes fine adjustments based on the relative position between the object and end effector of robot arm, ensuring that the end effector can smoothly grab the object. After grasping the object, the end effector of the robot arm rises to a certain height, then turns right until it reaches above the designated location. Then, the end effector slowly descends until the object completely touches table surface. Finally, the robot arm lifts up, completing the experiment of remotely operating a robot arm to grab an object. The position of the end effector is shown in Figure 17. The green point in Figure 17 represents the starting position of robot arm’s endpoint and the red point represents the final position.
Figure 18 shows the information of the data gloves for both hands. During this process, the MCP joints of right index finger, right little finger, and right thumb always maintain a standard flexion value of zero. The movements of the robot arm mainly depend on Joint 1, Joint 3, and Joint 4, while Joints 2, 5, and 6 remain relatively stationary.
In the aforementioned movements, the joint velocity and joint position of the robot arm are shown in Figure 19. The change range of the joint position of robot arm is large, but the motion curve is relatively smooth, without abnormal jitter and jump, indicating that the fuzzy controller proposed in this paper is quite stable and reliable. The joint velocity of robot arm can follow finger flexion movement, change rotation direction according to changes in standard flexion value of left ring finger, and balance flexibility and stability at same time. According to the experimental results, using data gloves for teleoperation based on fuzzy logic controller produces higher non-linearity, which is sensitive to the finger flexion changes. The measured delay time of the control system using a fuzzy controller is less than 0.05 s due to the communication delay.

4. Conclusions

In this paper, a data glove-based teleoperation using fuzzy logic controller method was proposed and validated. We collected hand posture through data gloves and standardized the data to interpret human operational intentions. A fuzzy logic controller was designed to achieve precise control of teleoperation. The experimental results indicated that compared to the joint velocity output of the PID controller, the ability of the data glove to use a fuzzy logic algorithm for mobile robot control leads to better results. This is shown by its sensitivity to finger flexion changes. Moreover, its higher non-linearity prevents abrupt changes in joint velocity due to abnormal fluctuations in the data glove’s control values, enhancing the stability of teleoperated robot arm control.
To further improve the performance of the proposed teleoperation method, our future work will develop a more complete teleoperation system and consider the case of a robot interacting with the external environment.

Author Contributions

Conceptualization, C.L. and L.J.; methodology, C.L. and L.J.; software, L.J.; validation, C.L. and L.J.; formal analysis, C.L.; investigation, C.L.; resources, J.W.; data curation, J.W.; writing—original draft preparation, C.L. and L.J.; writing—review and editing, Y.L. and W.L.; visualization, C.L. and W.L.; supervision, W.L.; funding acquisition, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant 52175007/52175012, the China Postdoctoral Science Foundation under Grant 2018M630348/2019T120259, and the Yangtze River Delta HIT robot technology research institute under Grant HIT-CXY-CMP2-ADTIL-21-01.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Batty, T.; Ehrampoosh, A.; Shirinzadeh, B.; Zhong, Y.; Smith, J. A Transparent Teleoperated Robotic Surgical System with Predictive Haptic Feedback and Force Modelling. Sensors 2022, 22, 9770. [Google Scholar] [CrossRef] [PubMed]
  2. Doisy, G.; Ronen, A.; Edan, Y. Comparison of Three Different Techniques for Camera and Motion Control of a Teleoperated Robot. Appl. Ergon. 2017, 58, 527–534. [Google Scholar] [CrossRef]
  3. Shiroma, N.; Miyauchi, R.; Matsuno, F. Mobile Robot Teleoperation through Virtual Robot. In Proceedings of the RO-MAN 2008-The 17th IEEE International Symposium on Robot and Human Interactive Communication, Munich, Germany, 1–3 August 2008; IEEE: New York, NY, USA, 2008; pp. 477–482. [Google Scholar]
  4. Zou, R.; Liu, Y.; Li, Y.; Chu, G.; Zhao, J.; Cai, H. A Novel Human Intention Prediction Approach Based on Fuzzy Rules through Wearable Sensing in Human–Robot Handover. Biomimetics 2023, 8, 358. [Google Scholar] [CrossRef] [PubMed]
  5. Bauer, A.; Wollherr, D.; Buss, M. Human–Robot collaboration: A survey. Int. J. Hum. Robot. 2008, 5, 47–66. [Google Scholar] [CrossRef]
  6. Xia, Y.; Li, W.; Duan, S.; Lei, W.; Wu, J. Low-Cost, Light-Weight Scalable Soft Data Glove for VR Applications. In Proceedings of the 2022 5th International Conference on Circuits, Systems and Simulation (ICCSS), Nanjing, China, 13–15 May 2022; IEEE: Nanjing, China, 2022; pp. 202–205. [Google Scholar]
  7. Alves de Oliveira, T.E.; Cretu, A.-M.; Petriu, E.M. Multimodal Bio-Inspired Tactile Sensing Module. IEEE Sens. J. 2017, 17, 3231–3243. [Google Scholar] [CrossRef]
  8. Li, S.; Ma, X.; Liang, H.; Görner, M.; Ruppel, P.; Fang, B.; Sun, F.; Zhang, J. Vision-Based Teleoperation of Shadow Dexterous Hand Using End-to-End Deep Neural Network. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 416–422. [Google Scholar]
  9. Li, S.; Jiang, J.; Ruppel, P.; Liang, H.; Ma, X.; Hendrich, N.; Sun, F.; Zhang, J. A Mobile Robot Hand-Arm Teleoperation System by Vision and Imu. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October 2020–24 January 2021; IEEE: New York, NY, USA, 2020; pp. 10900–10906. [Google Scholar]
  10. Gomez-Donoso, F.; Orts-Escolano, S.; Cazorla, M. Accurate and Efficient 3D Hand Pose Regression for Robot Hand Teleoperation Using a Monocular RGB Camera. Expert Syst. Appl. 2019, 136, 327–337. [Google Scholar] [CrossRef]
  11. Yang, W.; Paxton, C.; Cakmak, M.; Fox, D. Human Grasp Classification for Reactive Human-to-Robot Handovers. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October 2020–24 January 2021; pp. 11123–11130. [Google Scholar]
  12. Du, G.; Zhang, P.; Mai, J.; Li, Z. Markerless Kinect-Based Hand Tracking for Robot Teleoperation. Int. J. Adv. Robot. Syst. 2012, 9, 36. [Google Scholar] [CrossRef]
  13. Handa, A.; Van Wyk, K.; Yang, W.; Liang, J.; Chao, Y.-W.; Wan, Q.; Birchfield, S.; Ratliff, N.; Fox, D. DexPilot: Vision-Based Teleoperation of Dexterous Robotic Hand-Arm System. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 9164–9170. [Google Scholar]
  14. Chotiprayanakul, P.; Liu, D.K. Workspace Mapping and Force Control for Small Haptic Device Based Robot Teleoperation. In Proceedings of the 2009 International Conference on Information and Automation, Zhuhai/Macau, China, 22–24 June 2009; IEEE: New York, NY, USA, 2009; pp. 1613–1618. [Google Scholar]
  15. Farkhatdinov, I.; Ryu, J.-H.; An, J. A Preliminary Experimental Study on Haptic Teleoperation of Mobile Robot with Variable Force Feedback Gain. In Proceedings of the 2010 IEEE Haptics Symposium, Waltham, MA, USA, 25–26 March 2010; IEEE: New York, NY, USA, 2010; pp. 251–256. [Google Scholar]
  16. Tang, Y.; Liu, S.; Deng, Y.; Zhang, Y.; Yin, L.; Zheng, W. Construction of Force Haptic Reappearance System Based on Geomagic Touch Haptic Device. Comput. Methods Programs Biomed. 2020, 190, 105344. [Google Scholar] [CrossRef] [PubMed]
  17. Zhu, G.; Xiao, X.; Li, C.; Ma, J.; Ponraj, G.; Prituja, A.V.; Ren, H. A Bimanual Robotic Teleoperation Architecture with Anthropomorphic Hybrid Grippers for Unstructured Manipulation Tasks. Appl. Sci. 2020, 10, 2086. [Google Scholar] [CrossRef]
  18. Asad, M.U.; Farooq, U.; Gu, J.; Abbas, G.; Liu, R.; Balas, V.E. A Composite State Convergence Scheme for Bilateral Teleoperation Systems. IEEE/CAA J. Autom. Sin. 2019, 6, 1166–1178. [Google Scholar] [CrossRef]
  19. Asad, M.U.; Gu, J.; Farooq, U.; Balas, V.E.; Balas, M.M.; Abbas, G. An Improved Composite State Convergence Scheme with Disturbance Compensation for Multilateral Teleoperation Systems. Stud. Inform Control 2022, 31, 43–52. [Google Scholar] [CrossRef]
  20. Yang, Y.; Jiang, H.; Hua, C.; Li, J. Practical Preassigned Fixed-Time Fuzzy Control for Teleoperation System under Scheduled Shared-Control Framework. IEEE Trans. Fuzzy Syst. 2024, 32, 470–482. [Google Scholar] [CrossRef]
  21. Fang, B.; Guo, D.; Sun, F.; Liu, H.; Wu, Y. A Robotic Hand-Arm Teleoperation System Using Human Arm/Hand with a Novel Data Glove. In Proceedings of the 2015 IEEE International Conference on Robotics and Biomimetics (ROBIO), Zhuhai, China, 6–9 December 2015; IEEE: Zhuhai, China, 2015; pp. 2483–2488. [Google Scholar]
  22. Caeiro-Rodríguez, M.; Otero-González, I.; Mikic-Fonte, F.A.; Llamas-Nistal, M. A Systematic Review of Commercial Smart Gloves: Current Status and Applications. Sensors 2021, 21, 2667. [Google Scholar] [CrossRef] [PubMed]
  23. Basjaruddin, N.C.; Sutjiredjeki, E.; Akbar, H.W.C. Developing an Electronic Glove Based on Fuzzy Logic for Mobile Robot Control. J. Intell. Fuzzy Syst. 2019, 36, 1639–1645. [Google Scholar] [CrossRef]
  24. Rodriguez, L.; Przedworska, Z.; Obidat, O.; Parron, J.; Wang, W. Development and Implementation of an AI-Embedded and ROS-Compatible Smart Glove System in Human-Robot Interaction. In Proceedings of the 2022 IEEE 19th International Conference on Mobile Ad Hoc and Smart Systems (MASS), Denver, CO, USA, 19–23 October 2022; IEEE: Denver, CO, USA, 2022; pp. 699–704. [Google Scholar]
  25. Ali, M.; Malik, A.; Yusof, Z.M.; Kushairy, A.K.; Zaharah, H.F.; Ismail, A. Development of Smart Glove System for Therapy Treatment. In Proceedings of the 2015 International Conference on BioSignal Analysis, Processing and Systems (ICBAPS), Kuala, Lumpur, 26–28 May 2015; IEEE: Kuala, Lumpur, 2015; pp. 67–71. [Google Scholar]
  26. Li, F.; Chen, J.; Zhou, Z.; Xie, J.; Gao, Z.; Xiao, Y.; Dai, P.; Xu, C.; Wang, X.; Zhou, Y. Lightweight Soft Robotic Glove with Whole-Hand Finger Motion Tracking for Hand Rehabilitation in Virtual Reality. Biomimetics 2023, 8, 425. [Google Scholar] [CrossRef] [PubMed]
  27. Zhu, Q.; Da Fonseca, V.P.; Lima, B.M.R.; Welyhorsky, M.; Goubran, M.; De Oliveira, T.E.A.; Petriu, E.M. Teleoperated Grasping Using a Robotic Hand and a Haptic-Feedback Data Glove. In Proceedings of the 2020 IEEE International Systems Conference (SysCon), Montreal, QC, Canada, 24 August 2020; IEEE: Montreal, QC, Canada, 2020; pp. 1–7. [Google Scholar]
  28. Dhepekar, P.; Adhav, Y.G. Wireless Robotic Hand for Remote Operations Using Flex Sensor. In Proceedings of the 2016 International Conference on Automatic Control and Dynamic Optimization Techniques (ICACDOT), Pune, India, 9–10 September 2016; IEEE: Pune, India, 2016; pp. 114–118. [Google Scholar]
  29. Colasanto, L.; Suarez, R.; Rosell, J. Hybrid Mapping for the Assistance of Teleoperated Grasping Tasks. IEEE Trans. Syst. Man Cybern. Syst. 2013, 43, 390–401. [Google Scholar] [CrossRef]
  30. Tyagi, V.; Gupta, N.K.; Tyagi, P.K. Smart Wheelchair Using Fuzzy Inference System. In Proceedings of the 2013 IEEE Global Humanitarian Technology Conference: South Asia Satellite (GHTC-SAS), Trivandrum, India, 23–24 August 2013; IEEE: New York, NY, USA, 2013; pp. 175–180. [Google Scholar]
  31. ElKoura, G.; Singh, K. Handrix: Animating the Human Hand. In Proceedings of the 2003 ACM SIGGRAPH/Eurographics Symposium on Computer Animation, San Diego, CA, USA, 26–27 July 2003; pp. 110–119. [Google Scholar]
  32. Li, K.; Chen, I.-M.; Yeo, S.H.; Lim, C.K. Development of Finger-Motion Capturing Device Based on Optical Linear Encoder. J. Rehabil. Res. Dev. 2011, 48, 69. [Google Scholar] [CrossRef] [PubMed]
  33. Lin, J.; Wu, Y.; Huang, T.S. Modeling the Constraints of Human Hand Motion. In Proceedings of the Proceedings Workshop on Human Motion, Austin, TX, USA, 7–8 December 2000; IEEE: New York, NY, USA, 2000; pp. 121–126. [Google Scholar]
  34. Tanaka, K.; Sugeno, M. Stability Analysis and Design of Fuzzy Control Systems. Fuzzy Sets Syst. 1992, 45, 135–156. [Google Scholar] [CrossRef]
Figure 1. System overview.
Figure 1. System overview.
Biomimetics 09 00116 g001
Figure 2. Human skeletal model.
Figure 2. Human skeletal model.
Biomimetics 09 00116 g002
Figure 3. Simplified model of a human finger.
Figure 3. Simplified model of a human finger.
Biomimetics 09 00116 g003
Figure 4. Filtered index finger flexion angle.
Figure 4. Filtered index finger flexion angle.
Biomimetics 09 00116 g004
Figure 5. Schematic diagrams of three states: (a) Flexion standard value is −1; (b) Flexion standard value is 0; (c) Flexion standard value is 1.
Figure 5. Schematic diagrams of three states: (a) Flexion standard value is −1; (b) Flexion standard value is 0; (c) Flexion standard value is 1.
Biomimetics 09 00116 g005
Figure 6. Design of fuzzy logic controller.
Figure 6. Design of fuzzy logic controller.
Biomimetics 09 00116 g006
Figure 7. Gesture diagram of robot arm joint control: (a) Move joint 1 forward or reverse; (b) Move joint 2 forward or reverse. (c) Move joint 3 forward or reverse; (d) Move joint 4 forward or reverse; (e) Move joint 5 forward or reverse; (f) Move joint 6 forward or reverse.
Figure 7. Gesture diagram of robot arm joint control: (a) Move joint 1 forward or reverse; (b) Move joint 2 forward or reverse. (c) Move joint 3 forward or reverse; (d) Move joint 4 forward or reverse; (e) Move joint 5 forward or reverse; (f) Move joint 6 forward or reverse.
Biomimetics 09 00116 g007
Figure 8. Input membership functions: (a) Membership function of left_ring/right_x; (b) Membership function of rt_mcp.
Figure 8. Input membership functions: (a) Membership function of left_ring/right_x; (b) Membership function of rt_mcp.
Biomimetics 09 00116 g008
Figure 9. Input membership functions.
Figure 9. Input membership functions.
Biomimetics 09 00116 g009
Figure 10. Fuzzy control surfaces: (a) Joint 1~5 velocity output surface; (b) Joint 6 velocity output surface.
Figure 10. Fuzzy control surfaces: (a) Joint 1~5 velocity output surface; (b) Joint 6 velocity output surface.
Biomimetics 09 00116 g010
Figure 11. Experimental platform.
Figure 11. Experimental platform.
Biomimetics 09 00116 g011
Figure 12. The flex sensors.
Figure 12. The flex sensors.
Biomimetics 09 00116 g012
Figure 13. PID Controller Test Results.
Figure 13. PID Controller Test Results.
Biomimetics 09 00116 g013
Figure 14. Fuzzy Logic Controller Test Results.
Figure 14. Fuzzy Logic Controller Test Results.
Biomimetics 09 00116 g014
Figure 15. Motion test results using fuzzy logic controller: (a) Finger flexion standard value; (b) Joint velocity of robot arm.
Figure 15. Motion test results using fuzzy logic controller: (a) Finger flexion standard value; (b) Joint velocity of robot arm.
Biomimetics 09 00116 g015
Figure 16. Robot arm grasping experiment.
Figure 16. Robot arm grasping experiment.
Biomimetics 09 00116 g016
Figure 17. Robot arm end trajectory.
Figure 17. Robot arm end trajectory.
Biomimetics 09 00116 g017
Figure 18. Glove data: (a) Finger curvature of the left ring; (b) Finger curvature of the right hand.
Figure 18. Glove data: (a) Finger curvature of the left ring; (b) Finger curvature of the right hand.
Biomimetics 09 00116 g018
Figure 19. Robot arm joint velocity and position: (a) Robot arm joint velocity; (b) Robot arm joint position.
Figure 19. Robot arm joint velocity and position: (a) Robot arm joint velocity; (b) Robot arm joint position.
Biomimetics 09 00116 g019
Table 1. The mapping for Gestures to Robotic.
Table 1. The mapping for Gestures to Robotic.
Manipulator ActionGesture
Return to positionNo finger bend
Move robotic joint 1Bend left ring finger (left_ring) and bend right thumb PIP joint (right_x)
Move robotic joint 2Bend left ring finger (left_ring) and bend right index fin-ger (right_x)
Move robotic joint 3Bend left ring finger (left_ring) and bend right middle fin-ger (right_x)
Move robotic joint 4Bend left ring finger (left_ring) and bend right ring fin-ger (right_x)
Move robotic joint 5Bend left ring finger (left_ring) and bend right little fin-ger (right_x)
Move robotic joint 6Bend left ring finger (left_ring) and bend right thumb MCP joint (rt_mcp)
Table 2. Fuzzy control rule table.
Table 2. Fuzzy control rule table.
Joint Velocity (velj)Left_Ring
VSSMSMMLLVL
right fingers’ flexion standard valueVSZZZZZZZ
SZZZZZZZ
MSNSNSNSZSSS
MNSNSNSZSSS
MLNSNSNSZSSS
LNBNBNSZSPBPB
VLNBNBNBZPBPBPB
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lu, C.; Jin, L.; Liu, Y.; Wang, J.; Li, W. Teleoperated Grasping Using Data Gloves Based on Fuzzy Logic Controller. Biomimetics 2024, 9, 116. https://doi.org/10.3390/biomimetics9020116

AMA Style

Lu C, Jin L, Liu Y, Wang J, Li W. Teleoperated Grasping Using Data Gloves Based on Fuzzy Logic Controller. Biomimetics. 2024; 9(2):116. https://doi.org/10.3390/biomimetics9020116

Chicago/Turabian Style

Lu, Chunxiao, Lei Jin, Yufei Liu, Jianfeng Wang, and Weihua Li. 2024. "Teleoperated Grasping Using Data Gloves Based on Fuzzy Logic Controller" Biomimetics 9, no. 2: 116. https://doi.org/10.3390/biomimetics9020116

Article Metrics

Back to TopTop