Next Article in Journal
Evaluation of the Undrained Shear Strength in Preconsolidated Cohesive Soils Based on the Seismic Dilatometer Test
Next Article in Special Issue
Design of a Laparoscopic Robot System Based on Spherical Magnetic Field
Previous Article in Journal
An Intelligent Hybrid Energy Management System for a Smart House Considering Bidirectional Power Flow and Various EV Charging Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Human–Robot Cooperative Control Based on Virtual Fixture in Robot-Assisted Endoscopic Sinus Surgery

1
Shenzhen Key Laboratory of Minimally Invasive Surgical Robotics and System, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
2
Shenzhen College of Advanced Technology, University of Chinese Academy of Sciences, Shenzhen 518055 China
3
Technical Aspects of Multimodal Systems (TAMS), University of Hamburg, 22527 Hamburg, Germany
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2019, 9(8), 1659; https://doi.org/10.3390/app9081659
Submission received: 31 March 2019 / Revised: 15 April 2019 / Accepted: 18 April 2019 / Published: 22 April 2019
(This article belongs to the Special Issue Next-Generation Surgical Robotics)

Abstract

:
In endoscopic sinus surgery, the robot assists the surgeon in holding the endoscope and acts as the surgeon’s third hand, which helps to reduce the surgeon’s operating burden and improve the quality of the operation. This paper proposes a human–robot cooperative control method based on virtual fixture to realize accurate and safe human–robot interaction in endoscopic sinus surgery. Firstly, through endoscopic trajectory analysis, the endoscopic motion constraint requirements of different surgical stages are obtained, and three typical virtual fixtures suitable for endoscopic sinus surgery are designed and implemented. Based on the typical virtual fixtures, a composite virtual fixture is constructed, and then the overall robot motion constraint model is obtained. Secondly, based on the obtained robot motion constraint model, a human–robot cooperative control method based on virtual fixture is proposed. The method adopts admittance control to realize efficient human–robot interaction between the surgeon and robot during the surgery; the virtual fixture is used to restrain and guide the motion of the robot, thereby ensuring motion safety of the robot. Finally, the proposed method is evaluated through a robot-assisted nasal endoscopy experiment, and the result shows that the proposed method can improve the accuracy and safety of operation during endoscopic sinus surgery.

1. Introduction

Traditional endoscopic sinus surgery is usually performed by a surgeon with a left-handed nasal endoscope and right-handed surgical instruments. The surgeon’s hands are prone to fatigue and shaking due to the long operation, which may cause misoperation. Additionally, for some complicated operations, the assistant surgeon is required to hold the endoscope while the chief surgeon operates with both hands. Surgical operations between surgeons are easily interfered, and coordination is difficult. Introducing a robot to assist the surgeon to hold the endoscope, acting as the surgeon’s third hand, can combine the advantages of the robot with the surgeon’s experience, so that the surgeon can change the one-hand operation mode to a two-hand operation mode, thus the stability of the operation can be enhanced and the quality and safety of the operation can be improved.
Currently available robots for endoscopic sinus surgery can be divided into two categories: Passive surgical robots and active surgical robots. Active surgical robots include single-arm surgical robots, multi-arm surgical robots, and single-port type surgical robots. From the perspective of control, there are mainly three control modes. The first mode is the passive control mode, in which the surgeon locks or loosens the robot joints by operating buttons, pedals, etc., and constantly drags and adjusts the posture of the robot. This passive control mode is simple and reliable, and can meet the requirements of minimally invasive surgery to a certain extent, for example, the surgical robot Tiska [1] developed by Karl Storz, the Point Setter and UniARM [2,3] developed by Mitaka, and the EndoArm [4,5,6] developed by Olympus. Although the above passive control mode can basically realize the function of assisting surgeon in holding the endoscope in surgery, the robot still needs the surgeon to manually operate the button, pedal, etc. to loosen or lock all the joints, and the control has low efficiency and low intelligence. The second mode is the active control mode, in which the surgeon’s head movement and voice are used to control the motion of the robot. For example, the active surgical robot, Aesop [7,8,9,10], can use the voice control mode to drive the endoscope at the end of the robot to the desired spatial pose; the surgical robot, Endo Assist [11,12], can control the motion of the endoscope at the end of the robot with the movement of the surgeon’s head. The third mode is the master-slave control mode, which is often used in multi-arm surgical robot systems, such as Laprotek [13,14], Zeus [15,16,17], Da Vinci [18,19], Miro Surge [20,21,22,23], and Raven [24,25]. The surgeon controls the slave robot arms at the main console, and switches the control of each slave arm by pedals. The ratio of the surgeon’s input motion to the output motion of the robot can be adjusted in the master-slave control mode, achieving finer operation than the human hand. At the same time, the signal-filtering algorithm can eliminate the influence of the surgeon’s hand tremor, improving the quality of surgery. According to the above analysis, the passive control mode is simple and reliable, but the surgeon needs to manually adjust the posture of the nasal endoscope during the operation; the adjustment is awkward and has a low level of automation. The master-slave control mode is generally applicable on active surgical robots in the form of a multiple-arm, such as the Da Vinci robot. Considering that the end-effector of the multi-arm form robot is generally large in size, it can only be applied to the types of surgery that have a large operation space, such as abdominal cavity surgery, and it cannot be applied to types of surgery with a small operation space, such as the nasal cavity. For the active control mode based on the voice and head movement, the human–robot interaction efficiency is low, and the training time is long, which requires further research.
For robot-assisted endoscopic sinus surgery, due to the narrow and complex anatomy of the nasal cavity, the endoscope easily collides and interferes with the nasal tissue. To ensure the safety of the operation, it is necessary to restrain the movement of the endoscope at the end of the robot. Presently, there are two main motion constraint methods for robots: Hardware constraint based on robot structure design and software constraint based on algorithms. Hardware-based constraint is generally considered to be more safe, but often less flexible, such as RCM (remote center of motion) constraint. VF (virtual fixture) represents one of the most important software-based constraints, which was proposed by Professor Rosenberg [26,27] in 1993. VF is a constraint method that uses an algorithm to generate invisible spatial motion constraint, thereby limiting robot motion. Current VF can be divided into GVF (guiding virtual fixture) and FRVF (forbidden regions virtual fixture) [28]. A GVF is generally used to guide the end-effector of the robot to move along a predefined trajectory to avoid interference; an FRVF is generally used to prevent the end-effector of the robot from entering a specific area and can be used to protect critical tissues and organs during surgery. Park et al. [29] prevented the robot end-effector from entering a specific area by constructing a VF "virtual wall" in robot-assisted cardiac surgery and verified the effectiveness of the VF through simulation. Prada et al. [30] proposed a VF library design method by introducing VF into the virtual reality system. Ren et al. [31] proposed a dynamic VF design method for minimally invasive cardiac surgery, in which the interpolation algorithm was used to calculate the dynamic VF at different moments in the heartbeat cycle to avoid the probe entering the dangerous area during operation. Hu et al. [32] restricted the movement of the end-effector of a spinal surgical robot by introducing a GVF (straight type) and an FRVF (cone type) to improve the safety of the operation. Li et al. [33,34,35] used an optimization constraint algorithm to generate VFs to solve the problem of a narrow operating space in endoscopic sinus surgery, but the algorithm is too complicated and does not consider nasal soft tissue. In summary, the existing VF research is mainly based on regular geometric shapes. Due to the complex and narrow anatomical structure of the human nasal cavity, the use of regular geometric shapes as a motion constraint will result in the loss of substantial valuable operation space, which is not allowed in the surgical operation. Therefore, it is urgent that the problem regarding how to properly fit the VF model with a 3D anatomical structure of the human nasal cavity is solved and the safety constraint of the nasal endoscope’s movement is realized.
Based on the above control mode and safety constraint analysis, considering the current clinical requirements of endoscopic sinus surgery, this paper proposes a human–robot cooperative control method based on VFs in robot-assisted endoscopic sinus surgery. This method uses admittance-based cooperative control to realize human–robot interaction between the surgeon and robot during surgery; the VFs are used to restrain and guide the movement of the robot, thereby ensuring the movement safety of the nasal endoscope. The experimental results showed that the robot can assist the surgeon in holding the endoscope in the endoscopic sinus surgery with the proposed method, and the surgical operation is more accurate and safer during surgery.
The remainder of this paper is organized as follows. The robot motion constraint model based on VF is presented in Section 2. The human–robot cooperative control method is detailed in Section 3. The experiment and discussion are presented in Section 4, and the conclusions are summarized in Section 5.

2. Robot Motion Constraint Model Based on Virtual Fixture

The narrow and complex anatomical structure of the nasal sinus makes it easy for the endoscope at the end of the robot to interfere with the nasal tissue, and the entire surgical operation space is very crowded. To ensure the safety of the surgical operation, it is necessary to restrain the movement of the endoscope. The VF can construct invisible "clamps" (e.g., spatial curves and surfaces) to assist the operator to improve operational efficiency and accuracy. In this section, the robot motion constraint requirements for each motion stage during endoscopic sinus surgery are determined based on the endoscope’s trajectory analysis. Next, based on the obtained motion constraint requirements, three typical VF models suitable for endoscopic sinus surgery are constructed. Finally, based on the three typical VF models, a composite VF model is constructed to achieve the robot constraint for the whole endoscopic sinus surgery.

2.1. Motion Constraint Requirements Analysis

Generally, the trajectory of the endoscope at the end of the robot in endoscopic sinus surgery is shown in Figure 1.
It can be seen from Figure 1 that the movement of the endoscope can be mainly divided into four stages:
The first stage is from point A to point B: The endoscope moves from a safe area away from the patient (point A) to the vicinity of the patient’s nasal inlet (point B). In this stage, since the endoscope is far away from the patient, it is not easy to collide with the patient. Therefore, it is only necessary to construct a plane-based FRVF to prevent the endoscope from colliding with the patient.
The second stage is from point B to point C: When the endoscope moves further from point B to point C, the endoscope gradually approaches point C at the patient’s nasal inlet. To prevent the endoscope from colliding with the patient during this stage, a single-leaf hyperboloid based FRVF is constructed to guide and constrain the movement of the endoscope to prevent it from colliding with the patient’s nasal inlet.
The third stage is from point C to point D: When the endoscope moves from point C at the inlet of the nose to point D in the surgical area, the endoscope movement path is in the nasal cavity of the patient. Due to the narrow and complex anatomy of the nasal cavity, the endoscope easily collides with the nasal tissue during surgery; thus, a VF based on a spatial curve is constructed to guide the motion of the robot.
The fourth stage: After the endoscope reaches the surgical area, the surgeon constantly adjusts the endoscope posture to obtain a suitable surgical vision field according to the needs of the operation. To prevent collision and interference between the endoscope and nasal tissue, an FRVF based on the anatomical structure of the nasal cavity is constructed to constrain the motion of the endoscope.

2.2. Virtual Fixtures for Endoscopic Sinus Surgery

Based on the above robot motion constraint requirement analysis, the main types of VFs in endoscopic sinus surgery can be obtained as follows: GVF based on the spatial curve, FRVF based on the plane and hyperboloid, and FRVF based on the anatomical structure of the nasal cavity. Their action mechanisms and implementation processes are shown below.

2.2.1. Guiding Virtual Fixture Based on Spatial Curve

For the geometry of GVFs, there are mainly straight line, two-dimensional curve, three-dimensional curve, etc. However, for the robot system, considering the smoothness and computational complexity of the motion trajectory, the cubic polynomial spatial curve is mostly used.
As shown in Figure 2a, the schematic diagram of the GVF based on a spatial curve is shown. Assuming that the coordinates of the robot end are p ( x , y , z ) , the velocity of the robot end is V i n , and the curve, S, is the guiding spatial curve, the velocity of the robot end can be decomposed into the velocity along the tangential direction and velocity in the direction perpendicular to the tangential direction:
{ V i n = V i n + V i n V i n = T · V i n V i n = V i n V i n
Introducing a guiding stiffness coefficient, k [ 0 , 1 ] , to scale the velocity in the perpendicular tangential direction, the following is obtained:
V o u t = k · V i n
Thus, the velocity of the robot end after the action of the GVF can be obtained as follows:
V o u t = V i n + k · V i n
For the above GVF, different guiding requirements can be achieved by setting different values of the guiding stiffness coefficient. When k = 1 , the velocity of the robot end remains unchanged before and after guidance, the VF does not work, and the robot is in a free motion state. When 0 < k < 1 , the velocity in the perpendicular tangential direction of the robot end is reduced, so the robot will be directed to the curve, S. When k = 0 , the velocity of the robot end in the perpendicular tangential direction is completely removed and the robot end will be forced to move in the tangential direction of the curve, S. From the above, the smaller the guiding stiffness coefficient is, the stronger the motion guidance to the robot end obtained.
The above situation is the velocity guidance under ideal conditions. In practical applications, the end of the robot is generally not on the spatial curve, so that the motion trajectory of the robot end will be a curve parallel to the guiding curve.
Figure 2b shows the action of the GVF when the robot end is not on the guiding curve. Assume P s is the closest point on the curve to the robot end, then:
P s = Mindistance ( S , P )
Adding a regression coefficient, λ , then the regression velocity can be calculated as follows (direction is pointed to P s from P ):
V s = λ · PP s ¯
By adjusting the regression coefficient, the velocity of the robot end returning to the curve can be adjusted, and then the final velocity can be obtained:
V o u t s = V o u t + V s
From the above analysis, it can be seen that different guiding requirements of the GVF can be realized by adjusting the guiding stiffness coefficient, k, and the regression coefficient, λ .
The above derivation process of the GVF based on the spatial curve can be applied to the straight line, the two-dimensional curve, and so on because they can be regarded as a special case of the spatial curve.

2.2.2. Forbidden Virtual Fixture Based on Hyperboloid

The FRVF can limit the operating range of the robot and prevent the robot from entering a specific area, thereby avoiding collision between the instrument and tissue during operation. For the FRVF, how to construct appropriate geometries to constrain the workspace of the robot is an important issue to be considered in the application of the surgical robot. Considering that the endoscope and instrument swing around the nostril to form a workspace similar to the "funnel" shape [36], a single-leaf hyperboloid is used as the geometry of the FRVF.
The schematic diagram of the FRVF based on a single-leaf hyperboloid is shown in Figure 3. Assuming that the position of the robot end is P ( x , y , z ) , and the velocity of the robot end is V i n (direction as PD ¯ ), the equation of the single-leaf hyperboloid, S, is shown below:
x 2 a 2 + y 2 b 2 z 2 c 2 = d
The action process of the FRVF based on a single-leaf hyperboloid is as follows:
1) According to the position of the robot end and Equation (7), the spatial relationship between P ( x , y , z ) and the single-leaf hyperboloid is judged.
2) If x 2 a 2 + y 2 b 2 d + z 2 c 2 , the robot end, P ( x , y , z ) , is inside the single-leaf hyperboloid, the FRVF does not work, and the robot end is in the free motion state, its velocity, V i n , remains unchanged.
3) If x 2 a 2 + y 2 b 2 > d + z 2 c 2 , the robot end is outside the single-leaf hyperboloid, the velocity of the robot end will be changed by the FRVF. If V i n · P s B ¯ > 0 , the velocity of the robot end is in the same direction as the normal vector, P s B ¯ , of the tangent plane, P s AC ¯ , indicating that the robot end is leaving the prohibited area of the VF; if V i n · P s B ¯ < 0 , the velocity of the robot end is opposite to the normal vector, P s B ¯ , of the tangent plane, P s AC ¯ , indicating that the robot is entering the prohibited area of the VF. In the above two cases, the velocity of the robot needs to be adjusted.
By projecting the velocity vector, V i n , onto the normal, V i n , and tangent planes, V i n (the V i n direction is along the direction of P s B ¯ , the V i n direction is along the direction of P s D ¯ ), the following can be obtained:
{ V i n = V i n + V i n V i n = V i n · n o r m a l ( P s B ¯ ) V i n = V i n V i n
Introducing the forbidden stiffness coefficient, k [ 1 , 1 ] , the input velocity can be expressed as follows:
V i n = V i n + k · V i n
When V i n and P s B ¯ is in the same direction, if k < 0 , the robot end will continue to enter the prohibited area; if k 0 , the robot end will leave the prohibited area. When V i n and P s B ¯ is in the reverse direction, if k < 0 , the robot end will leave the prohibited area; if k 0 , the robot end will continue to enter the prohibited area. In the whole process, the velocity of the robot can be adjusted to leave the forbidden area by applying different values of k .
The above derivation process of the FRVF based on the single-leaf hyperboloid is applicable to other three-dimensional surface geometries, such as the plane, except that the method of solving the tangent plane of the nearest point is different.

2.2.3. Forbidden Virtual Fixture Based on the Anatomical Structure of the Nasal Cavity

When the robot end enters the inlet of the nasal cavity, due to the complex anatomy of the nasal cavity, the VF based on regular geometry (e.g., cylinder and cone) is used to constrain the movement of the robot, which will lose substantial precious operation space, and is not allowed during endoscopic sinus surgery. Therefore, in this section, a point cloud model is obtained based on the CT image sequence of the patient’s nasal tissue, based on which an FRVF suitable for the nasal cavity is constructed. The boundary of the FRVF coincides with the boundary of the nasal tissue, thereby maximally retaining the operation space while protecting the nasal tissue.
Figure 4 shows the schematic diagram of the FRVF based on the anatomical structure of the nasal cavity.
The FRVF works as follows:
1) Assuming that the position of the endoscope tip is P ( x , y , z ) , the corresponding pixel point, Q ( i , j , k ) , of the endoscope tip in the binarized point cloud model can be obtained as follows:
i = INT ( ( x o r i g i n . x ) / s p a c e . x )
j = INT ( ( y o r i g i n . y ) / s p a c e . y )
k = INT ( ( z o r i g i n . z ) / s p a c e . z )
where o r i g i n . x , o r i g i n . y , and o r i g i n . z represent the origin coordinates of the binarized point cloud model of nasal tissue; s p a c e . x , s p a c e . y , and s p a c e . z represent the pixel spacing in the x, y, and z coordinate directions of the nasal tissue binarized point cloud model; and INT() represents the rounding function.
2) Assuming that the search radius of the endoscope tip is R, it can be determined that the local search range of the collision detection is:
i ¯ = [ INT ( R / s p a c e . x ) + i , INT ( R / s p a c e . x ) + i ]
j ¯ = [ INT ( R / s p a c e . y ) + j , INT ( R / s p a c e . y ) + j ]
k ¯ = [ INT ( R / s p a c e . z ) + k , INT ( R / s p a c e . z ) + k ]
3) For the pixels in the local search range, the Euclidean distance from the pixels to the endoscope tip is calculated as follows:
D i s t a n c e = s q r t ( ( ( i ¯ i ) · s p a c e . x ) 2 + ( ( j ¯ j ) · s p a c e . y ) 2 + ( ( k ¯ k ) · s p a c e . z ) 2 )
4) If the pixel value of the local pixel is 0, it indicates the pixel is a free spatial point and is shown with the black color in Figure 4. There is no collision in this case, the endoscope is in a free motion state, and the FRVF does not work. If the pixel value of the local pixel is 255, it indicates the pixel is a nasal tissue point and is shown with the white color in Figure 4. There is a collision in this case. Since the surface of the human nasal sinus is a thin layer of mucous membrane, considering the safety of the operation, the velocity of the endoscope is set to zero, which means the robot immediately stops moving, protecting the nasal tissue of the patient.

2.3. Overall Robot Motion Constraint Model Based on Virtual Fixture

Through the analysis in the above section, several typical VF types suitable for endoscopic sinus surgery have been obtained. Considering that it is difficult for a single VF to meet the diverse constraint requirements during different stages of robot-assisted endoscopic sinus surgery, a composite VF is constructed based on the above typical VFs, and the overall robot motion constraint model is obtained, as shown in Figure 5.
After multiple VFs are combined, the working condition and scope of each basic VF have been changed to avoid conflicts among the basic VFs. Based on the robot-assisted endoscopic sinus surgery procedure, the process of the composite VF system is shown in Figure 6.
The detailed process of the composite VF is as follows:
1) Enter the coordinates, P ( x , y , z ) , and velocity, V i n , of the endoscope tip in the composite VF system.
2) Calculate the directed distance, D P 2 P l a n e , from the endoscope tip to the constraint plane (plane normal direction is positive). If D P 2 P l a n e 0 , indicating the endoscope tip is above the constraint plane, the robot can move freely in all directions, and go to step (7); if D P 2 P l a n e < 0 , indicating the endoscope tip is below the constraint plane, go to step (3).
3) Calculate the spatial relationship between the endoscope tip and nasal entrance point (RCM point) (the direction of the single-leaf hyperboloid opening is positive). If D P 2 RCM 0 , it means that the endoscope tip is above the nasal entrance point, go to step (4); if D P 2 RCM < 0 , it means that the endoscope tip is below the nasal entrance point, go to step (5).
4) Calculate the spatial relationship between the endoscope tip and single-leaf hyperboloid (if the endoscope tip is outside the single-leaf hyperboloid, the D P 2 S u r f a c e is positive). If D P 2 S u r f a c e 0 , it means the endoscope tip is outside the single-leaf hyperboloid, go to step (6); if D P 2 S u r f a c e < 0 , it means the endoscope tip is inside the single-leaf hyperboloid, the robot can move freely in all directions, go to step (7).
5) The endoscope enters the RCM restraint state, and the movement of the endoscope will be constrained to two rotational degrees of freedom around the RCM point, go to step (7).
6) Compare the distance from the endoscope tip to the constraint plane and distance from the endoscope tip to the single-leaf hyperboloid. If D P 2 S u r f a c e D P 2 P l a n e , it means the endoscope tip is close to the constraint plane, and the plane constraint is used to deal with the velocity of the endoscope tip; if D P 2 S u r f a c e < D P 2 P l a n e , it means the endoscope tip is close to the single-leaf hyperboloid, and the velocity of the endoscope tip is processed by the single-leaf hyperboloid constraint. Go to step (7).
7) Calculate the distance between the endoscope tip and the point cloud model of the nasal tissue. If D P 2 C l o u d > 0 , it means that there is no interference between the endoscope tip and nasal tissue, and the robot can move freely in all directions; if D P 2 C l o u d 0 , it means that the endoscope tip and nasal tissue are interfering. The robot stops moving.

3. Human–Robot Cooperative Control Based on Virtual Fixture

3.1. Interactive Force Acquisition and Filtering

Figure 7 shows the schematic diagram of human–robot cooperative control based on VFs. The surgeon adjusts the endoscope to the desired posture by dragging the operating handle at the end of the robot. During the dragging process, the robot is restrained and guided by the VFs to prevent the endoscope at the end of the robot from interfering with the nasal tissue. The human–robot interaction process of the control method is the same as that when no robot is involved in the surgical operation, but the advantage is that after the endoscope of the robot is dragged to the proper posture, the robot can be maintained in the current posture without operating buttons or pedals. With the control method, the surgeon can free his left hand and carry out complex surgical operations with both hands, effectively reducing the surgeon’s operational burden.
Firstly, the surgeon drags the handle of the robot to move, and the control system obtains the surgeon’s drag force through the six-dimensional force sensor (ATI Industrial Automation Company, USA):
F = [ f x   f y   f z   m x   m y   m z ] T
Considering the manual operation is easy to shake, the obtained force signal is generally not smooth and needs to be filtered. Considering the computational efficiency, the force signal can be smoothed with the moving average filtering as follows:
F ¯ = f ( F 1 , F 2 , , F n )
where n is the length of the intercepted window; the larger the n is, the smoother the force signal obtained. However, excessive n is likely to cause a control response delay. Assuming that the system control period is T control , the force sensor sampling period is T sample , then n can be determined by n = T control / T sample .

3.2. Admittance Mapping Design Based on Piecewise Function

After obtaining the interaction force between the surgeon and robot end-effector, the admittance control method is used to map the interaction force to the velocity of the endoscope, thereby establishing the spatial admittance relationship between the surgeon’s drag force and the velocity of the robot:
V F = [ v w ] = K · F ¯
v = [ v x , v y , v z ] T
w = [ w x , w y , w z ] T
Since the drag force contains spatial force and moment, the admittance coefficient, K, is a square matrix of 6 × 6. For admittance control, the design of admittance mapping is key. In general, there are two types of admittance mappings: Linear admittance mapping and nonlinear admittance mapping. Linear admittance mapping is simple to calculate, but it is easy to produce a sudden change in velocity at the boundary point, affecting the drag performance and resulting in unstable robot motion. Nonlinear admittance mapping can easily adjust the admittance coefficient at different regions, but the calculation is relatively complicated. Considering the motion stability requirements of medical robots and computational efficiency, a mixed admittance mapping based on the piecewise function is adopted in this paper. The mapping combines the advantages of linear admittance mapping and nonlinear admittance mapping, and smooths the inflection point by the cubic polynomial function. The detailed mapping function is shown below:
V ( F ) = { V max ,   F < k 4 V max + C ( F + k 4 ) 3 ,   k 4 F < k 3 3 C ( k 2 k 1 ) 2 ( F + k 2 ) C ( k 2 k 1 ) 3 ,   k 3 F < k 2 C ( F + k 1 ) 3 ,   k 2 F < k 1 0 ,   k 1 F k 1 C ( F k 1 ) 3 ,   k 1 < F k 2 3 C ( k 2 k 1 ) 2 ( F k 2 ) + C ( k 2 k 1 ) 3 ,   k 2 < F k 3 V max + C ( F k 4 ) 3 ,   k 3 < F k 4 V max ,   F > k 4
Figure 8 shows the admittance mapping relationship based on the piecewise function, and the coefficients, k 1 , k 2 , k 3 , k 4 , are positive. For the drag force at the end of the robot, when k 1 F k 1 , the drag force is in the stopping zone. The setting of the stopping zone can effectively prevent the robot from moving under the action of a small input force (such as mechanical vibration and sensor zero drift), so that the robot can smoothly maintain the original posture after the drag of the robot is stopped. When k 1 < F k 2 or k 2 F < k 1 , the drag force is in the transition zone from the stopping zone to the working zone. The transition zone allows the robot motion admittance mapping to smoothly change from the stopping zone to the working zone. When k 2 < F k 3 or k 3 < F k 2 , the drag force is in the main working zone. The sensitivity of the robot’s dragging motion can be dynamically adjusted by adjusting the slope of the zone. When k 3 < F k 4 or k 4 < F k 3 , the drag force is in the transition zone from the working zone to the saturation zone. The transition zone allows the robot motion admittance mapping to smoothly change from the working zone to the saturation zone. When F > k 4 or F < k 4 , the drag force is in the saturation zone. In this zone, even if the drag force continues to increase, the response velocity of the robot remains unchanged. The saturation zone is used to limit the maximum drag response velocity of the robot, preventing the robot from moving too fast and losing control.

3.3. Human–Robot Cooperative Control Based on Virtual Fixture

The velocity, F V F = [ v x   v y   v z   ω x   ω y   ω z ] T , obtained based on Equation (22) is the velocity vector in the force sensor coordinate system. Considering the VF is generally designed in the image coordinate system, it is necessary to transform the velocity of the endoscope from the force sensor coordinate system to the image coordinate system before performing the VF calculation. Assuming that after the navigation system is registered, the expression of the force sensor coordinate system in the image coordinate system is:
T Robot Image = [ R Robot Image P Robot Image O r g 0 1 ]
The expression of the endoscope velocity in the image coordinate system can be obtained as follows:
I V F = [ R Robot Image 0 0 R Robot Image ] F V F
For I V F , using the composite VFs in Figure 6 for constraint calculation, the endoscope velocity after VF restraint can be obtained in the image coordinate system as I V o u t . Then, the endoscope velocity is transformed back from the image coordinate system to the robot base coordinate system:
Robot V o u t = [ R Robot Image 1 0 0 R Robot Image 1 ] I V o u t
According to the robot Jacobian matrix, J 0 , the robot joint velocity, q ˙ , can be obtained as follows:
q ˙ = J 0 1 · R o b o t V o u t

4. Experiments and Discussion

4.1. Experiment Setup

The experimental platform is shown below in Figure 9, which includes a nasal endoscope subsystem (Shenda Endoscope Company, China), a self-developed image navigation subsystem based on an optical measurement device, Polaris Vicra (Northern Digital Company, Canada), a human head model (Phacon Company, Germany), and an endoscopic sinus surgery assistant robot. The robot is developed by our team on the basis of its previous version [37]; it has 7 degrees of freedom and a ring-shaped workspace, which can cover the operation range of endoscopic sinus surgery of 200 × 200 × 150 mm (size of a typical human head).

4.2. Guiding Virtual Fixture Experiment

Figure 10 shows the experiment scene of the GVF, which is a cubic polynomial curve. The purpose of this experiment is to evaluate the constraint performance of the GVF. The starting and ending coordinates of the curve in the figure are (147.3, 49.2, –2.5), (–117.8, 90.6, –30.3). The robot works in the admittance-based drag control mode, and is dragged from the starting point along the curve to the ending point, without GVF constraint and with GVF constraint. During the experiment, the trajectory of the endoscope tip is obtained through the navigation system.
The experimental results are shown in Figure 11. It can be seen that the maximum deviation of the trajectory of the endoscope tip from the ideal trajectory is approximately 5.2 mm without the GVF constraint. After loading the GVF constraint, the maximum deviation from the ideal trajectory is approximately 1.0 mm. The comparison results show that the GVF can effectively reduce the trajectory tracking error of the robot and improve the drag operation precision.

4.3. Forbidden Virtual Fixture Experiment

Figure 12 shows the experiment scene of the FRVF. Considering FRVF generally contains multiple VF geometries, the spatial plane-based FRVF and the single-leaf hyperboloid-based FRVF are composites to be used in this experiment. While evaluating the constraint performance of the single FRVF, the constraint performance of the composite FRVF is also evaluated. In the experiment, the robot works in the admittance-based drag control mode. The endoscope of the robot is dragged from the starting point to the ending point, without FRVF constraint and with composite FRVF constraint. During the experiment, the trajectory of the endoscope tip is obtained through the navigation system.
The experimental results are shown in Figure 13. It can be seen that without the FRVF constraint, the trajectory of the endoscope tip fluctuates drastically, and the maximum fluctuation distance is approximately 4.9 mm. To avoid collision, the drag velocity is slowed down when the endoscope tip is close to the nasal entrance (the density of the sampling points is increased). With the composite FRVF constraint, the trajectory of the endoscope tip is basically above the plane and inside the hyperboloid, the maximum constraint surface penetration distance is approximately 1.3 mm. At the intersection of the plane and the hyperboloid, the trajectory of the endoscope tip is smooth, verifying the effectiveness of the composite FRVF constraint.

4.4. Robot-Assisted Endoscopic Nasal Examination Experiment

In endoscopic nasal examination surgery, the surgeon inserts the endoscope into the small, narrow nasal cavity and carefully examines the anatomy, such as the shape of the turbinate and the opening of the sinus. This surgery is a typical surgical procedure in nasal surgery, so it is selected in this experiment.
Experiment model. Considering that a real human corpse head is very rare and an animal’s nasal cavity is not similar to a human’s cavity, a high-imitation head model with fine anatomical structure (Phacon Company, Germany) is adopted in this experiment, as shown in Figure 14.
Surgical path planning. A three-dimensional model of the nasal cavity is obtained by reconstruction of the nasal CT image sequence, and then the surgical path from the entrance point of the nasal cavity to the opening of each sinus (taking maxillary sinus as an example) is obtained by the surgical planning module.
Interactive endoscopic nasal examination. The surgical scene of the interactive endoscopic nasal examination is shown in Figure 15. The upper left corner of the figure shows the real-time image of the endoscope during examination, and the upper right corner of the figure shows the control system interface and the virtual robot system.
Based on the proposed human–robot cooperative control method, the operator drags the endoscope at the end of the robot and performs interactive nasal sinus examination, browsing along the planned path that is a GVF based on a spatial curve. During this process, the motion of the endoscope tip is constrained by the robot motion constraint model, and the operator constantly adjusts the pose of the endoscope to see the small and deep anatomy inside the nasal cavity from different angles of view. Figure 16 shows the endoscope tip trajectory together with the preoperative planned trajectory during interactive endoscopic nasal examination. It can be seen that due to the effect of the VF constraint, the trajectory of the endoscope tip is substantially coincident with the preoperative planned trajectory, and the maximum deviation is approximately 1.1 mm. Figure 17 shows the endoscopic image during the interactive endoscopic nasal examination. As shown in the figure, the endoscopic field of view is clear throughout the whole examination.
Through the above experiment, it can be seen that the proposed human–robot cooperation control method based on VFs can effectively realize the motion guidance and constraint of the robot, improving the accuracy and safety of surgical operation during surgery.

5. Conclusions

This paper proposed a human–robot cooperative control method based on VF to improve the accuracy and safety of surgical operation in robot-assisted endoscopic sinus surgery. Firstly, based on the analysis of endoscopic motion constraint requirements in different surgical stages, three typical VFs suitable for endoscopic sinus surgery were designed and implemented. The GVF based on a spatial curve was designed to achieve motion guidance from the nasal cavity entrance to the operating area. Considering the funnel-shaped workspace of the endoscope, the FRVF based on the single-leaf hyperboloid was designed to achieve motion constraints at the entrance to nasal cavity. The FRVF based on the anatomical structure of the nasal cavity was designed to constrain the motion of the endoscope inside the nasal cavity, the boundary of which coincides with the boundary of the nasal tissue, thereby maximally retaining the precious operation space while protecting the nasal tissue. Thereafter, a composite VF was constructed based on the above typical VFs, and an overall robot motion constraint model was obtained. Secondly, based on the obtained robot motion constraint model, a human–robot cooperative control method based on VF was proposed. The method adopts admittance-based control to realize efficient human–robot interaction between the surgeon and robot during surgery, and the surgeon adjusts the endoscope to the desired posture by dragging the end of the robot. During the dragging process, the motion of the robot is restrained and guided by the VFs to prevent the endoscope at the end of the robot from interfering with the nasal tissue. With the proposed control method, the surgeon can free his left hand and carry out complex surgical operations with both hands, thus improving operation efficiency. Finally, validation experiments were conducted. The GVF experiment based on the cubic polynomial curve was performed and the results showed that the proposed GVF can effectively reduce the trajectory tracking error of the robot and improve the drag operation precision. The composite FRVF based on the spatial plane and a single-leaf hyperboloid was tested and the results showed that the proposed FRVFs can prevent the endoscope from colliding with the nasal tissue and improve the safety of the surgical operation. Thereafter, robot-assisted endoscopic nasal examination experiment was performed using a high-imitation head model, and the results showed that the proposed human–robot cooperative control method based on VF can effectively realize the motion guidance and constraint of the robot, and help to improve the accuracy and safety of operation during endoscopic sinus surgery. Considering clinical applicability, the project team is now working on clinical experiments to further optimize the robot system.

Author Contributions

Conceptualization, Y.H. (Ying Hu) and J.W.Z.; methodology, Y.H. (Yucheng He), P.Z. and Y.H. (Ying Hu); experiments, Y.H. (Yucheng He), P.Z., B.L.Z., and X.Z.Q.; writing—original draft preparation, Y.H. (Yucheng He), and P.Z.; writing—review and editing, B.L.Z., X.Z.Q. and Y.H. (Ying Hu); project administration, Y.H. (Ying Hu) and J.W.Z.

Funding

This research was funded by the National Natural Science Foundation of China (Grant No. 61603374, U1813213 and U1713218), Shenzhen Fundamental Research Funds (Grant Nos. JCYJ20160608153218487 and JCYJ20180507182415428), Shenzhen Peacock Plan (Grant No. KQTD2016113010571019) and Shenzhen Key Laboratory Project (Grant No. ZDSYS201707271637577).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Buess, G.F.; Schurr, M.O.; Fischer, S.C. Robotics and Allied Technologies in Endoscopic Surgery. Arch. Surg. 2000, 135, 229–235. [Google Scholar] [CrossRef]
  2. Arnholt, J.L.; Mair, E.A. A ‘third hand’for endoscopic skull base surgery. Laryngoscope 2002, 112, 2244–2249. [Google Scholar] [CrossRef]
  3. Morita, A.; Okada, Y.; Kitano, M.; Hori, T.; Taneda, M.; Kirino, T. Development of hybrid integrated endoscope-holder system for endoscopic microneurosurgery. Neurosurgery 2004, 55, 926–932. [Google Scholar] [CrossRef]
  4. Eskandari, R.; Amini, A.; Yonemura, K.; Couldwell, W. The use of the Olympus EndoArm for spinal and skull-based transsphenoidal neurosurgery. min-Minim. Invasive Neurosurg. 2008, 51, 370–372. [Google Scholar] [CrossRef]
  5. Nishiyama, Y.; Kinouchi, H.; Horikoshi, T. Surgery on intracranial aneurysms under simultaneous microscopic and endoscopic monitoring. Neurosurgery 2011, 58, 84–92. [Google Scholar] [CrossRef]
  6. Taniguchi, M.; Kato, A.; Taki, T.; Tsuzuki, T.; Yoshimine, T.; Kohmura, E. Microsurgical maneuvers under side-viewing endoscope in the treatment of skull base lesions. Skull Base 2011, 21, 115–122. [Google Scholar] [CrossRef] [PubMed]
  7. Mettler, L.; Ibrahim, M.; Jonat, W. One year of experience working with the aid of a robotic assistant (the voice-controlled optic holder AESOP) in gynaecological endoscopic surgery. Hum. Reprod. 1998, 13, 2748–2750. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Ballester, P.; Jain, Y.; Haylett, K.; McCloy, R.F. Comparison of task performance of robotic camera holders EndoAssist and Aesop. In Proceedings of the International Congress Series, Berlin, Germany, 27–30 June 2001. [Google Scholar]
  9. Alessandrini, M.; De Padova, A.; Napolitano, B.; Camillo, A.; Bruno, E. The AESOP robot system for video-assisted rigid endoscopic laryngosurgery. Eur. Arch. Oto-Rhino-Laryngol. 2008, 265, 1121–1123. [Google Scholar] [CrossRef]
  10. Jacobs, L.; Shayani, V.; Sackier, J. Determination of the learning curve of the AESOP robot. Surg. Endosc. 1997, 11, 54–55. [Google Scholar] [CrossRef]
  11. Aiono, S.; Gilbert, J.; Soin, B.; Finlay, P.; Gordan, A. Controlled trial of the introduction of a robotic camera assistant (Endo Assist) for laparoscopic cholecystectomy. Surg. Endosc. Other Interv. Tech. 2002, 16, 1267–1270. [Google Scholar] [CrossRef] [PubMed]
  12. Gilbert, J. The EndoAssist™ robotic camera holder as an aid to the introduction of laparoscopic colorectal surgery. Ann. R. Coll. Surg. Engl. 2009, 91, 389–393. [Google Scholar] [CrossRef]
  13. Franzino, R.J. The Laprotek surgical system and the next generation of robotics. Surg. Clin. 2003, 83, 1317–1320. [Google Scholar] [CrossRef]
  14. Peddamatham, S.; Peine, W.; Tan, H.Z. Assessment of vibrotactile feedback in a needle-insertion task using a surgical robot. In Proceedings of the 2008 Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems, Reno, NE, USA, 13–14 March 2008; pp. 93–99. [Google Scholar]
  15. Butner, S.E.; Ghodoussi, M. A real-time system for tele-surgery. In Proceedings of the 21st international Conference on Distributed Computing Systems, Mesa, AZ, USA, 16–19 April 2001; pp. 236–243. [Google Scholar]
  16. Butner, S.E.; Ghodoussi, M. Transforming a surgical robot for human telesurgery. IEEE Trans. Robot. Autom. 2003, 19, 818–824. [Google Scholar] [CrossRef]
  17. Ghodoussi, M.; Butner, S.E.; Wang, Y. Robotic surgery-the transatlantic case. In Proceedings of the 2002 IEEE International Conference on Robotics and Automation (Cat. No. 02CH37292), Washington, DC, USA, 11–15 May 2002; pp. 1882–1888. [Google Scholar]
  18. Dwivedi, J.; Mahgoub, I. Robotic surgery: A review on recent advances in surgical robotic systems. In Proceedings of the 2012 Florida Conference on Recent Advances in Robotics, Boca raton, FL, USA, 10–11 May 2012; pp. 16–22. [Google Scholar]
  19. Guthart, G.S.; Salisbury, J.K. The Intuitive/sup TM/telesurgery system: Overview and application. In Proceedings of the 2000 ICRA. Millennium Conference. IEEE International Conference on Robotics and Automation. Symposia Proceedings (Cat. No. 00CH37065), San Francisco, CA, USA, 24–28 April 2000; pp. 618–621. [Google Scholar]
  20. Konietschke, R.; Hagn, U.; Nickl, M.; Jorg, S.; Tobergte, A.; Passig, G.; Seibold, U.; Le-Tien, L.; Kubler, B.; Groger, M. The DLR MiroSurge-A robotic system for surgery. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 1589–1590. [Google Scholar]
  21. Tobergte, A.; Konietschke, R.; Hirzinger, G. Planning and control of a teleoperation system for research in minimally invasive robotic surgery. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 4225–4232. [Google Scholar]
  22. Hagn, U.; Konietschke, R.; Tobergte, A.; Nickl, M.; Jörg, S.; Kübler, B.; Passig, G.; Gröger, M.; Fröhlich, F.; Seibold, U. DLR MiroSurge: A versatile system for research in endoscopic telesurgery. Int. J. Comput. Assist. Radiol. Surg. 2010, 5, 183–193. [Google Scholar] [CrossRef]
  23. Hagn, U.; Nickl, M.; Jörg, S.; Passig, G.; Bahls, T.; Nothhelfer, A.; Hacker, F.; Le-Tien, L.; Albu-Schäffer, A.; Konietschke, R. The DLR MIRO: A versatile lightweight robot for surgical applications. Ind. Robot: Int. J. 2008, 35, 324–336. [Google Scholar] [CrossRef]
  24. Lum, M.J.; Friedman, D.C.; Sankaranarayanan, G.; King, H.; Fodero, K.; Leuschke, R.; Hannaford, B.; Rosen, J.; Sinanan, M.N. The RAVEN: Design and validation of a telesurgery system. Int. J. Robot. Res. 2009, 28, 1183–1197. [Google Scholar] [CrossRef]
  25. Hannaford, B.; Rosen, J.; Friedman, D.W.; King, H.; Roan, P.; Cheng, L.; Glozman, D.; Ma, J.; Kosari, S.N.; White, L. Raven-II: An open platform for surgical robotics research. IEEE Trans. Biomed. Eng. 2013, 60, 954–959. [Google Scholar] [CrossRef]
  26. Rosenberg, L.B. Virtual fixtures: Perceptual tools for telerobotic manipulation. In Proceedings of the IEEE Virtual Reality Annual International Symposium, Seattle, WA, USA, 18–22 September 1993; pp. 76–82. [Google Scholar]
  27. Rosenberg, L.B. Virtual fixtures as tools to enhance operator performance in telepresence environments. In Proceedings of the Telemanipulator Technology and Space Telerobotics, Boston, MA, USA, 21 December 1993; pp. 10–22. [Google Scholar]
  28. Bowyer, S.A.; Davies, B.L.; y Baena, F.R. Active constraints/virtual fixtures: A survey. IEEE Trans. Robot. 2014, 30, 138–157. [Google Scholar] [CrossRef]
  29. Park, S.; Howe, R.D.; Torchiana, D.F. Virtual fixtures for robotic cardiac surgery. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2001; pp. 1419–1420. [Google Scholar]
  30. Prada, R.; Payandeh, S. On study of design and implementation of virtual fixtures. Virtual Real. 2009, 13, 117–129. [Google Scholar] [CrossRef]
  31. Ren, J.; Patel, R.V.; McIsaac, K.A.; Guiraudon, G.; Peters, T.M. Dynamic 3-D virtual fixtures for minimally invasive beating heart procedures. IEEE Trans. Med Imaging 2008, 27, 1061–1070. [Google Scholar]
  32. Jin, H.; Hu, Y.; Tian, W.; Zhang, P.; Zhang, J.; Li, B. Safety analysis and control of a robotic spinal surgical system. Mechatronics 2014, 24, 55–65. [Google Scholar] [CrossRef]
  33. Li, M.; Taylor, R.H. Optimum robot control for 3D virtual fixture in constrained ENT surgery. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2003; pp. 165–172. [Google Scholar]
  34. Li, M.; Taylor, R.H. Spatial motion constraints in medical robot using virtual fixtures generated by anatomy. In Proceedings of the IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA’04, New Orleans, LA, USA, 26 April–1 May 2004; pp. 1270–1275. [Google Scholar]
  35. Li, M.; Ishii, M.; Taylor, R.H. Spatial motion constraints using virtual fixtures generated by anatomy. IEEE Trans. Robot. 2007, 23, 4–19. [Google Scholar] [CrossRef]
  36. Eichhorn, K.W.; Westphal, R.; Last, C.; Rilk, M.; Bootz, F.; Wahl, F.M.; Jakob, M. Workspace and pivot point for robot-assisted endoscope guidance in functional endonasal sinus surgery (FESS). Int. J. Med Robot. Comput. Assist. Surg. 2015, 11, 30–37. [Google Scholar] [CrossRef] [PubMed]
  37. He, Y.; Zhao, B.; Hou, X.; Gao, P.; Hu, Y.; Zhang, P. An assistant robot system for sinus surgery. J. Med. Devices 2016, 10, 030925. [Google Scholar] [CrossRef]
Figure 1. Trajectory of the endoscope at the end of the robot during endoscopic sinus surgery (taking the right maxillary sinus as an example).
Figure 1. Trajectory of the endoscope at the end of the robot during endoscopic sinus surgery (taking the right maxillary sinus as an example).
Applsci 09 01659 g001
Figure 2. Schematic diagram of GVF based on the spatial curve (a) The robot end is on the guiding curve, (b) the robot end is not on the guiding curve.
Figure 2. Schematic diagram of GVF based on the spatial curve (a) The robot end is on the guiding curve, (b) the robot end is not on the guiding curve.
Applsci 09 01659 g002
Figure 3. Schematic diagram of the FRVF based on a single-leaf hyperboloid. (a) The funnel-shaped endoscope workspace, (b) the FRVF based on a single-leaf hyperboloid.
Figure 3. Schematic diagram of the FRVF based on a single-leaf hyperboloid. (a) The funnel-shaped endoscope workspace, (b) the FRVF based on a single-leaf hyperboloid.
Applsci 09 01659 g003
Figure 4. Schematic diagram of the FRVF based on the anatomical structure of the nasal cavity. (a) The point cloud model acquisition process, (b) the FRVF based on the anatomical structure of the nasal cavity.
Figure 4. Schematic diagram of the FRVF based on the anatomical structure of the nasal cavity. (a) The point cloud model acquisition process, (b) the FRVF based on the anatomical structure of the nasal cavity.
Applsci 09 01659 g004
Figure 5. Overall robot motion constraint model based on VFs.
Figure 5. Overall robot motion constraint model based on VFs.
Applsci 09 01659 g005
Figure 6. The process of the composite VF.
Figure 6. The process of the composite VF.
Applsci 09 01659 g006
Figure 7. Schematic diagram of human–robot cooperative control based on VF.
Figure 7. Schematic diagram of human–robot cooperative control based on VF.
Applsci 09 01659 g007
Figure 8. Admittance mapping based on the piecewise function.
Figure 8. Admittance mapping based on the piecewise function.
Applsci 09 01659 g008
Figure 9. Endoscopic sinus surgery assistant robot system.
Figure 9. Endoscopic sinus surgery assistant robot system.
Applsci 09 01659 g009
Figure 10. The experiment scene of the GVF.
Figure 10. The experiment scene of the GVF.
Applsci 09 01659 g010
Figure 11. Comparison of experimental results of the GVF. (a) Without GVF constraint; (b) with GVF constraint.
Figure 11. Comparison of experimental results of the GVF. (a) Without GVF constraint; (b) with GVF constraint.
Applsci 09 01659 g011
Figure 12. Experiment scene of the FRVFs.
Figure 12. Experiment scene of the FRVFs.
Applsci 09 01659 g012
Figure 13. Comparison of the experimental results of the composite FRVF. (a) Endoscopic tip trajectory without FRVF constraint; (b) endoscopic tip trajectory with the composite FRVF constraint.
Figure 13. Comparison of the experimental results of the composite FRVF. (a) Endoscopic tip trajectory without FRVF constraint; (b) endoscopic tip trajectory with the composite FRVF constraint.
Applsci 09 01659 g013
Figure 14. High-imitation head model with fine anatomical structure of the nasal sinus. (a) Nasal cavity module; (b) high-imitation head model.
Figure 14. High-imitation head model with fine anatomical structure of the nasal sinus. (a) Nasal cavity module; (b) high-imitation head model.
Applsci 09 01659 g014
Figure 15. Surgical scene of the interactive endoscopic nasal examination.
Figure 15. Surgical scene of the interactive endoscopic nasal examination.
Applsci 09 01659 g015
Figure 16. The endoscope tip trajectory and the preoperative planned trajectory.
Figure 16. The endoscope tip trajectory and the preoperative planned trajectory.
Applsci 09 01659 g016
Figure 17. Endoscopic image during the interactive endoscopic nasal examination (sort by time).
Figure 17. Endoscopic image during the interactive endoscopic nasal examination (sort by time).
Applsci 09 01659 g017

Share and Cite

MDPI and ACS Style

He, Y.; Hu, Y.; Zhang, P.; Zhao, B.; Qi, X.; Zhang, J. Human–Robot Cooperative Control Based on Virtual Fixture in Robot-Assisted Endoscopic Sinus Surgery. Appl. Sci. 2019, 9, 1659. https://doi.org/10.3390/app9081659

AMA Style

He Y, Hu Y, Zhang P, Zhao B, Qi X, Zhang J. Human–Robot Cooperative Control Based on Virtual Fixture in Robot-Assisted Endoscopic Sinus Surgery. Applied Sciences. 2019; 9(8):1659. https://doi.org/10.3390/app9081659

Chicago/Turabian Style

He, Yucheng, Ying Hu, Peng Zhang, Baoliang Zhao, Xiaozhi Qi, and Jianwei Zhang. 2019. "Human–Robot Cooperative Control Based on Virtual Fixture in Robot-Assisted Endoscopic Sinus Surgery" Applied Sciences 9, no. 8: 1659. https://doi.org/10.3390/app9081659

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop