Next Article in Journal
Influence of Process Energy on the Formation of Imperfections in Body-Centered Cubic Cells with Struts in the Vertical Orientation Produced by Laser Powder Bed Fusion from the Magnesium Alloy WE43
Previous Article in Journal
Batch Fabrication of Microelectrode Arrays with Glassy Carbon Microelectrodes and Interconnections for Neurochemical Sensing: Promises and Challenges
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Stereo Visual Servoing Control of a Soft Endoscope for Upper Gastrointestinal Endoscopic Submucosal Dissection †

1
School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
2
Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
3
Centre of AI and Robotics, Hong Kong Institute of Science and Innovation, Chinese Academy of Sciences, Hong Kong
4
Department of Biomedical Engineering, City University of Hong Kong, Hong Kong
5
School of Biomedical Engineering and Imaging Sciences, King’s College London, London SE1 7EU, UK
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published in the 45th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC 2023), Sydney, Australia, 24–27 July 2023.
Micromachines 2024, 15(2), 276; https://doi.org/10.3390/mi15020276
Submission received: 4 January 2024 / Revised: 7 February 2024 / Accepted: 14 February 2024 / Published: 15 February 2024

Abstract

:
Quickly and accurately completing endoscopic submucosal dissection (ESD) operations within narrow lumens is currently challenging because of the environment’s high flexibility, invisible collision, and natural tissue motion. This paper proposes a novel stereo visual servoing control for a dual-segment robotic endoscope (DSRE) for ESD surgery. Departing from conventional monocular-based methods, our DSRE leverages stereoscopic imaging to rapidly extract precise depth data, enabling quicker controller convergence and enhanced surgical accuracy. The system’s dual-segment configuration enables agile maneuverability around lesions, while its compliant structure ensures adaptability within the surgical environment. The implemented stereo visual servo controller uses image features for real-time feedback and dynamically updates gain coefficients, facilitating rapid convergence to the target. In visual servoing experiments, the controller demonstrated strong performance across various tasks. Even when subjected to unknown external forces, the controller maintained robust performance in target tracking. The feasibility and effectiveness of the DSRE were further verified through ex vivo experiments. We posit that this novel system holds significant potential for clinical application in ESD surgeries.

1. Introduction

Gastrointestinal (GI) cancer is a prevalent and highly malignant tumor in clinical practice [1]. Stomach cancer, in particular, ranks third in terms of mortality due to its low survival rate, with 782,685 deaths reported in 2018 [2]. Improving the survival rate of GI cancer relies heavily on early-stage treatment [3,4]. Owing to the high compliance, control precision, and cost-effectiveness, flexible endoscopes are increasingly being utilized in a variety of minimally invasive surgical procedures (see Figure 1a). Since the first application in 1988 [5], endoscopic submucosal dissection (ESD) has become the gold standard for the early treatment of GI cancer [6]. ESD enables the en bloc resection of submucosal lesions, resulting in a lower recurrence rate [7]. Currently, most ESD procedures are dependent on systems that utilize a single flexible arm outfitted with a monocular endoscope and various electrosurgical knives designed for specific surgical tasks [8]. However, due to the limited degrees of freedom (DoFs) of existing ESD endoscopes and the high skills required, improper operations during the procedure may lead to some complications such as tissue perforation and bleeding [9]. Additionally, the narrow visual field offered by traditional endoscopes hinders the surgeon’s ability to perform precise, multi-angle excisions, emphasizing the necessity for an ESD system with improved visual capabilities and instrument control [10].
To simplify the procedure and improve maneuverability, researchers have proposed different robot arm designs by adapting conventional surgical tools, enabling novices to perform the surgery easily. Chiu et al. [11] compared the feasibility and proficiency of ESD using an insulation-tipped (IT) diathermic knife, demonstrating that the robotic ESD device MASTER is more effective in terms of time cost and exhibits a good grasping and cutting efficiency [12]. Other similar ESD devices include ESD+AWC [13], K-FLEX [14], and TURBT [15]. However, the monocular vision system on traditional ESD endoscopes limits depth perception and lacks the stereoscopic perception of the lesion, which may increase the surgical operation time [16] and compromise the accuracy of resection [17,18]. The application of stereoscopic vision in various miniaturized surgical instruments, such as the da Vinci surgical system [19] and Versius surgical robot [20], enhances surgeons’ hand–eye coordination and depth judgment capabilities in minimally invasive surgeries. Figure 1b illustrates a dual-segment robotic endoscope (DSRE) [21] with a stereo vision system, which is used to reach the stomach lesion through the esophagus and perform the excision. As shown in Figure 1c, the DSRE consists of two air-driven soft segments that can be actively bent by tuning the air pressure.
Challenges remain in controlling the end effector to achieve a desired pose in a flexible continuum robot arm, especially given the hyper-redundancy and infinite DoFs [22]. Contact with the target also introduces disturbances at the tip and the circumferential body. Therefore, accurate modeling of the relationship between the actuation inputs, tip pose, and compensation for disturbances is essential. Although modeling in continuum robotics is a well-explored topic [23], the design of an eye-in-hand system like a flexible endoscope is not common. The stereo vision, embedded at the robot tip, contributes to image-based visual servoing (IBVS), enhancing the system’s compatibility with the eye-in-hand setup [24]. Table 1 summarizes the related research on multi-segment soft robot control, highlighting the superiority of visual servoing for robot operation in unstructured environments. The classical IBVS has been widely used to solve the tracking [25], shape control [26], and depth estimation [27] problems of flexible endoscopes. During endoscopic operations, external interference, such as inserting internal instruments, may lead to potential issues with the proportional controller. These issues could manifest as slow convergence [28] and decreased tracking performance [25]. To overcome these uncertainties, control algorithms such as sliding mode control [29], model predictive control [30], and multi-sensor fusion [31] are deployed in the IBVS framework to improve system robustness. The camera not only provides a view of the lesion but also senses the actual motion of the robot arm, serving as a proprioception mechanism for disturbance compensation. In [31], the depth of the endoscopic image is extracted for camera position adjustment in laparoscopic surgeries. In this study, stereo vision is employed to capture a wide view of the lesion and estimate the depth between the robot tip and the desired target, providing sensing information for disturbance compensation.
This study presents a dual-segment robotic endoscope with a control algorithm based on binocular visual servoing, and preliminarily discusses the application value of stereo vision in ESD surgery. To address the extant clinical challenges, the design requirements of the DSRE are introduced in detail in Section 2, and the fabrication procedure of the robot and the design of the pneumatic control unit are explained, respectively. Section 3 explains the modeling of the relationship between the actuation inputs of the robot and the tip pose, as well as the approach to sensing an object in 3D form. Additionally, Section 3 introduces an adaptive stereo visual servoing (ASVS)-based control scheme. The experiment results in Section 4 demonstrate the effectiveness of the robot and the proposed methods through controller tests and an ex vivo ESD trial. Finally, Section 5 concludes this work. Compared with a previously published conference paper [21], the main improvements in this expanded study include a smaller and more flexible robot for ESD with segments calibration, the ASVS controller with adaptive gain and PD control, and a more comprehensive experimental validation.

2. System Design

2.1. Robot Design

The DSRE is designed to function the same as the distal tip [37] on a traditional endoscope; thus, it was designed with the following three goals in mind: (i) to allow for multiple orientations along a resection path, as a single segment proves inadequate; (ii) to fit within the adult esophagus, which measures 25 cm in length and 2 cm in diameter [38]; and (iii) to facilitate a wide viewing angle during surgery and the efficient operation of a T-type electrosurgical cutting knife.
To this end, both the proximal and distal segment of the DSRE were designed to have three actuators longitudinally aligned with the main body and have a 2 mm through working channel for the electrosurgical knife or biopsy grasper, as shown in Figure 2b. When fabricated with appropriate pretension, this construction allows for approximately 170° and 88° of curvature when the proximal and distal are fully actuated, respectively. More detailed dimensions of the DSRE are shown in Figure 2b. As shown in Table 2, compared with the previous design [21] and conventional endoscopes, the DSRE’s robotic actuator can provide more precise control, while its smaller size and increased bending capabilities expand the operational area within the gastrointestinal tract. Additionally, the stereo vision provides supplementary depth information.
Two commercial cameras (OV6946, OmniVision, Santa Clara, CA, USA) with a distribution distance of 4.4 mm were attached at the tip, providing a wide view angle during surgery, as shown in Figure 2c. Similarly, a T-type electrosurgical cutting knife was fixed at the tip, acting as the end effector of the robot arm. The resection power and mode could be set on an electrosurgical generator (DGD-300B-2, Beilin, Beijing, China).

2.2. Robot Fabrication

As shown in Figure 2a, by casting the prototypes on 3D-printed molds, the soft segments were constructed of two low-stiffness silicones, Ecoflex 0050 and Dragon Skin 10 AF (Smooth-On, Macungie, PA, USA) mixed in a 1:1 ratio. To restrict elongation and limit radial expansion when pressurized, a nylon thread was attached at the end of each segment and another thread was coiled spirally. The DSRE of two serially connected segments was then affixed to a linear stage to facilitate insertion into the stomach.

2.3. Pneumatic Control Design

Pneumatic actuation is used in the DSRE system for its high precision control, quick response, safety for medical use, and the adaptability of in-body applications. Each air chamber of the robotic arm was linked to a pneumatic regulator (ITV0030, SMC, Tokyo, Japan) by a 60 cm silicone tube (0.3 mm inner, 0.8 mm outer diameter), and command individual pressures continuously to bend the DSRE with a direction angle and a bending angle. These valves were chosen for the following reasons: fast response times (0.1 s with no load), high stability (within 0.5% full span repeatability error), and accurate response (within 1% full span linearity error).

2.4. Robot Calibration

The calibration of soft robots is an essential process that helps determine the mechanical behavior under different loads and conditions, enabling better control and manipulation. To verify the mechanical performance and chamber consistency of the proposed dual-segment soft robot, we initially examined its bending behavior corresponding to each independent air chamber.
Figure 3a illustrates the configurations of the electromagnetic (EM) trackers located at the proximal and distal segments of the DSRE. Notably, to accurately obtain the position and orientation of each segment center, two trackers were symmetrically affixed to a 3D-printed part mounted at the proximal/distal segment’s end. We subsequently processed the average of the sensed data to calculate the positions and Euler angles. During testing, the pneumatic regulator was controlled to gradually increase the air pressure with a constant increment of 5000 Pa, while simultaneously recording the bending angles. As shown in Figure 3b, the curves of the bending angles corresponding to each segment’s three chambers were roughly equivalent. When reaching the maximum air pressure (0.14 MPa for the proximal segment and 0.1 MPa for the distal segment), the standard deviations of the bending angles across the different chambers were remarkably low, which were only 0.1012 rad and 0.0428 rad, respectively. These results illustrate that the soft robot has a consistent and stable bending performance in all directions.

3. Methodology

Through geometric analysis and an optimization algorithm, the modeling was first established, and the two endoscopes placed in an embedded sensing system to compensate for the external load effect.

3.1. Kinematics

3.1.1. Shape to Tip Pose

For soft segments with a large slenderness ratio, when the eccentrically fixed chamber is subjected to changing air pressure, the robot will bend and deform. Referring to the PCC model proposed in [42], it is assumed that the bending shape of each soft robot on the DSRE is a circular arc. As shown in Figure 4a, the DSRE can be geometrically parameterized by Φ = ( z s θ s φ s θ e φ e ) T in the configuration space, where z s is the overall insert distance provided by a linear stage, θ s and θ e are the bending angles, φ s and φ e are the direction angles between the bending plane and the oxz plane, l s and l e are the length of the proximal and distal segment, and z e and t d are the length of the middle cap and tip cap.
Taking the proximal segment as an example, the coordinate transformation sequence from the O s b X s b Y s b Z s b to O s t X s t Y s t Z s t coordinate system is as follows: first rotate the φ s angle around the z-axis, then translate l s / θ s along the x-axis, then rotate the θ s angle around the y-axis, and, finally, translate l s / θ s along the negative direction of the x-axis. The corresponding homogeneous transformation matrix is calculated as follows:
T s t s b = R z ( φ s ) τ x ( l s θ s ) R y ( θ s ) τ x ( l s θ s )
where τ * ,   R * 4 × 4 , respectively, denote the translation and rotation about the { x , y , z } axis.
Considering the proximal segment and the displacements z s , z e , and t d , the overall transformation matrix T t b 4 × 4 from the base frame O b X b Y b Z b to the robot tip frame O t X t Y t Z t of the DSRE system is as follows:
T t b = τ z ( z s ) R z ( φ s ) τ x ( l s θ s ) R y ( θ s ) τ x ( l s θ s ) R z ( π 3 + φ e ) τ z ( z e ) R z ( φ e ) τ x ( l e θ e ) R y ( θ e ) τ x ( l e θ e ) τ z ( t d )
where the shift angle between the two segments is 60°
The Jacobian matrix J r 6 × 5 is used to analytically establish the approximate relationship between the robot tip velocity v t 6 and joint velocity Φ ˙ :
v t = J r Φ ˙
where J r can be derived through forward kinematics T t b . A more detailed calculation of the forward kinematics and the Jacobian matrix can be found in Appendix A.
To reach a given target position of the end of the robot, we need to inversely solve the appropriate joint configuration.

3.1.2. Air Pressure to Shape

After finding the shape parameters, the robot system should set proper air pressures for each segment. The joint angles θ s , θ e , φ s , and φ e can be calculated from the actuator space variables q = ( z s p s , 1 p s , 2 p s , 3 p e , 1 p e , 2 p e , 3 ) T :
θ i = 2 p i , 1 2 + p i , 2 2 + p i , 3 2 p i 1 p i 2 p i 2 p i 3 p i 1 p i 3 3 ρ i φ i = tan 1 ( p i 1 + p i 3 2 p i 2 3 ( p i 3 p i 1 ) )
where i { e , s } , and the subscripts e and s are used to represent the proximal and distal segment, respectively, ρ i is the distance between the center of the air chamber and the center of the segment, p i , m ,   m { 1 , 2 , 3 } are the air pressure of the chambers in each segment.

3.2. Depth Estimation

The two cameras mounted at the robot tip could not only view the lesion with a wide angle but also could be employed to calculate the depth of a target (vertical distance from camera to the target). In this study, we employed the semi-global matching (SGM) algorithm [43], renowned for its efficacy in disparity estimation. From the disparity and stereo camera intrinsic information, the targets’ 3D position relative to camera frames can be estimated. The SGM algorithm encompasses a pixelwise matching technique based on mutual information ( M I ) [44] and a semi-global 2D constraint, composed of multiple 1D constraints. It is detailed as follows.

3.2.1. Mutual Information-Based Cost function

Initially, SGM computes the similarity between pixels in two rectified stereo images using a mutual information (MI)-based cost function, which is less sensitive to photometric changes. Thus, the MI value can be used to evaluate the suitability of pixels. The higher the value, the better the match, and vice versa. For the left and right images I l , I r from the end effector of the DSRE, the MI value is M I I l , I r = H I l + H I r H I l , I r ( H is the entropy of image, H I l , I r is the joint entropy of images). The MI-based cost function would be computed for each pixel and disparity, resulting in a cost volume that the SGM method would use in subsequent steps to compute the final disparity map.

3.2.2. Aggregate Cost

After the initial cost computation, these costs need to be aggregated across multiple paths through the image to enforce the smoothness constraints while preserving disparity discontinuities, which are often present at object boundaries. A path-wise dynamic programming approach was applied in the SGM algorithm. The cost aggregation step involves accumulating the costs for each pixel along several paths through the image. Typically, 8 paths are considered to cover all directions: left-to-right, right-to-left, top-to-bottom, bottom-to-top, and the four diagonals.
For a given pixel h and disparity d , the aggregated cost E s ( h , d ) along a path s is calculated as follows:
E s ( h , d ) = C ( h , d ) + min { E s ( h s , d ) E s ( h s , d 1 ) + K 1 E s ( h s , d + 1 ) + K 1 min i E s ( h s , i ) + K 2 } min k E s ( h s , k )
where C ( h , d ) is the matching cost at pixel h for disparity d computed from the mutual information, E s ( h s , d ) is the cost for pixel h at disparity d accumulated from the previous pixel along path s , K 1 and K 2 are the penalty for small disparity changes and for large disparity changes, respectively, to allow for disparity discontinuities. The last term is the minimum cost for the previous pixel h s over all disparities. After the cost aggregation step, each pixel will have 8 different accumulated costs, 1 for each path. These costs are then combined to produce a single, aggregated cost for each pixel and disparity:
E ( h , d ) = s E s ( h , d )
The final disparity for each pixel is chosen as the disparity that minimizes this aggregated cost.

3.2.3. Disparity Refinement

Once the initial disparity map is computed, it often requires further refinement to improve the accuracy, especially in occluded or low-texture areas. Refinement steps may include subpixel enhancement, occlusion handling, disparity map filtering, and left–right consistency check. Figure 5 illustrates the application of the SGM algorithm on the DSRE system, which was conducted in the MATLAB R2022a (Mathworks Inc. Portola Valley, CA, USA) with the disparity SGM and reconstruct Scene function. The depth information would be further used in the interaction matrix L in the ASVS controller. This matrix is essential for providing accurate and reliable positional information regarding the target.

3.3. Adaptive Stereo Visual Servoing Control

The visual feedback is directly utilized in the control loop of IBVS. The goal is to match the current image with the desired image. The error signal e 2 for the controller is derived from the difference between the current image features s 2 and the desired image features s d , and e ˙ = s ˙ .

3.3.1. Monocular Vision Model

Take the left camera as an example to demonstrate the camera model for brevity. The midpoint of the connection line between the two cameras coincides with O t .
As shown in Figure 4b, as a given point Q 3 in O c l X c l Y c l Z c l , its coordinates in the image frame O l x l y l and pixel frame O p l u v are q l = ( x l y l ) T and s l = ( u l v l ) T , respectively. According to the pinhole camera model, the perspective equation can be obtained from the relationship on similar triangles, i.e., as follows:
u l = λ x l x l λ l + c x l v l = λ y l y l λ l + c y l
where λ x l , λ y l are the focal length in pixels, c x l , c y l are the optical center in pixels, and λ l is the focal length in millimeters.
The motion of the features s ˙ l on the image plane can be predicted using the interaction matrix and velocity of camera v c l 6 :
s ˙ l = L l v c l
with (7), L l 2 × 6 is given by
L l = [ λ l z c l e 0 x l z c l e x l y l λ λ l 2 + x l 2 λ y l 0 λ l z c l e y l z c l e λ l 2 + y l 2 λ l x l y l λ l x l ]
where z c l e is the depth of point Q , which can be estimated online with the stereo vision. Similarly, we can obtain the right camera model with s ˙ r = L r v c r .

3.3.2. Stereo Vision Model

Considering the distance d between cameras, the transformation matrix T t c l 4 × 4 from the left camera frame to the robot tip frame is
T t c l = τ x ( d / 2 ) = [ I 3 p t c l 0 1 ] ,
while the transformation matrix for the right camera is T t c r = τ x ( d / 2 ) .
Since the left–right camera is fixed on the tip of the DSRE, the velocity transformation between the camera frame and robot tip frame is given by
v c l = M t c l v t v c r = M t c r v t
where the transformation matrix M t c * 6 × 6 can be calculated from [45] with (10), i.e.,
M t c * = [ I 3 [ p t c * ] × O 3 I 3 ]
where [ p t c * ] × is the skew-symmetric matrix of p t c * .
Let s ˙ = ( s ˙ l s ˙ r ) T represent the image features motion extracted from the left and right cameras simultaneously, and the overall velocity transformation from the robot tip frame to the pixel frame can be derived from (8), (11), and (12):
s ˙ = [ L l M t c l L r M t c r ] v t = L v t
where L 4 × 6 is the overall interaction matrix of the stereo vision system.
Inserting (3) in (13), we obtain the relationship of the joint velocity and image feature:
e ˙ = L J r Φ ˙
L J r 4 × 5 is a singular matrix without an inverse matrix. Commonly used pseudo-inverse methods often have stability problems. When the robot approaches a singular configuration, this method will cause excessive joint velocities and cause safety problems. The damped least squares method [46] not only improves the image feature tracking effect, but also avoids the singularity problem by limiting joint velocities
min ( L J r Φ ˙ e ˙ 2 + σ Φ ˙ 2 )
where σ is a non-zero damping constant.
The inverse problem could be solved by minimizing the object (14). The joint velocities could be derived further as
Φ ˙ = L a e ˙ L a = ( L J r ) T ( L J r ( L J r ) T + σ I ) 1

3.3.3. Adaptive PD Controller

The classic IBVS tasks usually use a proportional control law [28]. Let the error of the feature satisfy the first-order equation e ˙ = η e , so that the error value can decrease rapidly in an exponential speed. Thus,
Φ ˙ = η L a e
Although a proportional controller responds quickly and is simple to deploy, it cannot eliminate the steady-state error (SSE). In this paper, we adopt a PD controller to reduce the SSE and increase the system stability. Adding a derivative term to (17), we obtain
Φ ˙ = η L a e K d e ˙
where K d is the non-negative derivative gain.
To enhance the efficiency of the IBVS convergence process, we propose the implementation of an adaptive proportional gain. η is dynamically adjusted in real time based on the infinity norm of the feature error, aiming to reduce the number of iterations necessary for convergence and enhance the system’s performance. η is set as follows:
η = ( η u η l ) e e 0 + η l
where η u and η l are the preset maximum and minimum gains, respectively, e 0 is the initial error to a new target. The detailed flow chart of the ASVS controller is shown in Figure 6.

4. Experimental Validation

4.1. Depth Estimation Validation

To validate the efficacy of the proposed camera depth estimation algorithm, a comparison was made between the ground truth, derived from EM trackers, model 10001742 (NDI Aurora, Waterloo, ON, Canada), and the values estimated by the algorithm. The specific setup used for the testing process is illustrated in Figure 7a. Three trackers were fixed on a piece of tissue while two base trackers were mounted on the tip of the distal segment. For this experiment, the ground truth was defined as the distance between the base tracker and the trackers placed within the task space. The stereo camera in Figure 7b was first well calibrated with the method in [47], resulting in a notably low mean reprojection error of merely 0.2 pixels. A comparison was then made between the estimated depth, as calculated by the SGM, and the actual values across different altitudes. As depicted in the error band curves in Figure 8, the RMSE of the estimated depth parameters of different markers were 1.5334 mm, 1.5449 mm, and 1.5468 mm, respectively, when compared with the ground truth. This illustrates that the proposed algorithm is highly accurate, and it could obtain reliable depth estimations in various surgical applications. The estimated depth was further used in the visual servoing control.

4.2. Controller Performance

The performance of the ASVS was evaluated through the execution of four distinct manipulation tasks in this section. The visual servo controller’s parameters during these tests were established as follows: q 0 = ( 0 0.2 0.2 0.2 0.2 0.2 0.2 ) T , η u = 1 , η l = 0.5 , σ = 0.5 , K d = diag { 0.02 , 0.06 , 0.02 , 0.06 } T . In this study, the tracking error was quantified by the Euclidean distance between the current coordinates and target coordinates of the extracted image feature in the left and right frames. The control goal was to stabilize the error below 15 pixels. The frame rate of the left and right cameras was 30 FPS to ensure that the calculated air pressures have been applied to the segments to generate the desired robot shape.
As indicated in Figure 9, to simplify the feature detection and extraction process, the stereo vision system was programmed to track the AprilTags [48] in the workspace. Tags with ID 1, 2, 8, and 9 were selected as the tracking targets. The soft robot was controlled to track the target tags to the desired position on the image plane under the air pressure from the pneumatic regulators. A straight beam equipped with a force sensor was fixed on a linear rail guide slide actuator, and the tip caused contact disturbance to the robot. More results can be found in the Supplementary Video S1.

4.2.1. Static Target Tracking

In this experiment, the regions of interest (ROIs) within the left and right images were delineated as circular areas centered at points s d l = ( 210 200 ) T and s d r = ( 190 200 ) T with a radius of 15 pixels, as shown by the red circle and the green dashed circle in Figure 10b. The initial orientation of the DSRE robot was positioned to face the initial target with the AprilTag ID 3. After initialization, the robot was controlled to bring tags with different IDs to the ROI areas. These tags were equidistantly distributed at 90-degree intervals along the circumference of a circle. In each static tag tracking experiment, the adaptively adjusted gains of the ASVS controller were compared with the classic visual servoing method [28] employing fixed gains of η = 0.5 , η = 0.75 , and η = 1 . Figure 10b illustrates the trajectories of various tags in the left and right images, and all ultimately converged within the ROIs. Due to the inherent limitation of robots with inconsistent stiffness in all directions, the DSRE cannot reach the target at some positions, and it will overshoot after passing through these positions, which resulted in the controller’s tracking effect on tag 1 being inferior to several others, as depicted by the blue curve in Figure 10b. Figure 10c and Table 3 detail the pixel errors and convergence times for tracking different tags, respectively. As indicated by (17), when the gain was low, the robot’s end effector moved at a reduced velocity, resulting in excessive time consumption. Conversely, with high gain, the end effector’s velocity increased, potentially leading to overshoot and oscillation. With the gain set at constant values of 5, 0.75, and 1, the classic controller required an average of 19.4 s, 8.29 s, and 9.04 s, respectively, to achieve the tracking of a static target, reducing the error to below 15 pixels. Meanwhile, the proposed control algorithm necessitated only 4.99 s, corresponding to a reduction in time consumption of 74.28%, 39.81%, and 37.94%, respectively. This result illustrates the robustness and reliability of this method in achieving the accurate control of the soft robot arm and precise tracking of the moving target.

4.2.2. Dynamic Target Tracking

This part introduces a tracking experiment designed to evaluate the control performance of the robot under dynamic target changes. As demonstrated in the previous experiment, comparisons were made between the ASVS and classical algorithms using different fixed gains. Figure 11a shows the experimental setup, with tags affixed to an electric linear slide rail, which moved back and forth between two points, A and B, 15 mm apart, at a specified velocity of v t a g = 0.75 mm / s . The endoscopic images captured during the experiment and the corresponding tracking errors are presented in Figure 11b,c, respectively. The robot was tasked to follow tag 9, while the linear guide remained stationary until the initial stable tracking of the target was accomplished by the robot. As shown in Figure 11b, a significant surge in error occurred when the tag’s movement direction changed, but the controller quickly re-captured the tag and subsequently reduced the tracking difference to less than 15 pixels. Table 4 enumerates the tracking errors’ root mean square error (RMSE), mean absolute value error (MAE), standard deviation (SD), and maximum error e max for different controllers. The ASVS outperformed the classic controller across various metrics, maintaining the RMSE and MAE beneath 15 pixels, with an SD of merely 5.14 pixels. This rapid error correction is indicative of the agility and precision of the real-time tracking of dynamic targets.

4.2.3. Tracking under Unknown External Disturbance

The performance of dealing with external disturbance was also evaluated. During the test, the robot utilized endoscopic image feedback exclusively. An external force of unknown magnitude and location was applied to the soft robot using a slender beam affixed with a single-axial force sensor, which caused a mutable robot configuration. The control objective was to bring the target tag into the predefined ROIs on the image planes. The robot configuration and experimental results are shown in Figure 12. During the experiment, at 11.6 s, the robot first received an external force of 120 mN, and the tracking error increased from 5.9 pixels to 93.4 pixels. The controller subsequently engaged to resist the external force and tracked the tag to the ROIs within approximately 3.5 s, reducing the error to less than 15 pixels. After the external force was maintained for 5 s, it was removed, and the controller repeated the above error correction process. During the entire test process, the robot was affected by dynamic external forces a total of three times. The maximum external force was 150 mN. The robot could maintain the tracking of the target and the error could converge to five pixels. This shows that image-based feedback control is efficacious in enabling the DSRE robot to adjust its configuration responsively to external environmental conditions, thereby enhancing the tracking precision.

4.2.4. Trajectory Tracking

Marking the lesion is an important early step in ESD surgery [7]. The end effector of the soft robot carrying the electrosurgical knife should always be perpendicular to the working surface, marking along the periphery of the target lesion by coagulation. This experiment aims to evaluate the feasibility of robots in the marking process, with the DSRE tracking a triangular reference trajectory. This trajectory guided the tag moving along the pre-defined path in the captured images. The experimental trajectories of the pixel positions on the left and right image planes are shown in Figure 13. The overall RMSE of the left and right image errors was 13.51 pixels, which demonstrates that the controller has a good tracking performance for the trajectories.

4.3. Resection in an Ex Vivo Porcine Stomach

This test was designed to assess the practical efficacy of robotic systems in ESD procedures. As shown in Figure 14, an ex vivo porcine stomach was affixed within a receptacle to simulate surgical conditions. The operator used a gamepad to perform a remote operation program on the DSRE and controlled the robot to complete the ESD surgery in the stomach. Figure 15 shows the entire process of the ESD surgery. According to the DSRE configuration shown in Figure 2c, the angle between the end effector and the horizontal plane was only 12.5°, which illustrates that the robot can cut tissue in multiple directions. Such a small operating angle can improve the cutting efficiency and safety of traditional ESD. Figure 15a,b show the left endoscopic images during ESD marking and cutting, respectively. The operator operated the DSRE to perform electrocoagulation around the lesion and successfully completed a series of marking points. During the cutting process, the operator used tweezers to assist in pulling the target tissue up to expose the cutting point to the robot and used the joystick of the gamepad to adjust the cutting angle of the robot. After the procedure, as illustrated in Figure 15c, the DSRE has completely removed the target tissue mass without causing perforation in the porcine stomach wall. Finally, a piece of tissue measuring approximately 5 × 4 mm was successfully removed. Therefore, this ex vivo experiment illustrates the feasibility of applying the DSRE in ESD surgery.

5. Conclusions

This work presents a novel dual-segment soft robotic endoscope with stereo vision designed to enhance reachability and dexterity during ESD surgery. The proposed system can execute multi-angle cutting operations at a small angle relative to the lesion surface, allowing for efficient en bloc resection. Additionally, the system incorporates two calibrated RGB cameras and a depth estimation algorithm to provide real-time 3D information of the tumor, which is also used to guide the control framework. Based on the adaptively updated gain and PD control laws, the stereo vision servo controller improves the convergence speed and path tracking performance during surgery. The experimental results indicate that the proposed system improves the motion stability and precision. Ex vivo testing further demonstrates its significant potential in endoscopic surgery.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/mi15020276/s1, Video S1: the experimental results of the dual-segment robotic endoscope (DSRE).

Author Contributions

Conceptualization, J.C. and H.L.; funding acquisition, H.L.; methodology, J.C., S.W. and Q.Z.; project administration, H.L.; resources, W.H.; software, J.C. and M.C.; supervision, H.L.; validation, J.C., S.W. and Y.W.; visualization, J.C., S.W. and Q.Z.; writing—original draft, J.C. and S.W.; writing—review and editing, Q.Z., J.H. and H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Centre of AI and Robotics, Hong Kong Institute of Science and Innovation, Chinese Academy of Sciences, sponsored by InnoHK Funding, HKSAR, partially supported by the Institute of Automation, Chinese Academy of Sciences, and partially supported by the Sichuan Science and Technology Program (Grant number: 2023YFH0093).

Data Availability Statement

Data are contained within the article.

Acknowledgments

Parts of Figure 1 were created using templates from Servier Medical Art (http://smart.servier.com/), accessed on 14 December 2023, licensed under a Creative Common Attribution 3.0 Generic License.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

The translation and rotation matrices in (2) are given by
R y ( θ ) = [ c θ 0 s θ 0 0 1 0 0 s θ 0 c θ 0 0 0 0 1 ] , R z ( φ ) = [ c φ s φ 0 0 s φ c φ 0 0 0 0 1 0 0 0 0 1 ] , τ x ( t ) = [ 1 0 0 t 0 1 0 0 0 0 1 0 0 0 0 1 ] , τ z ( t ) = [ 1 0 0 0 0 1 0 0 0 0 1 t 0 0 0 1 ]
In order to simplify the formula, use s and c instead of sine and cosine functions, and the forward kinematics in (2) can be extended to
T t b = [ R t b P t b 0 1 ] = [ T 11 T 12 T 13 T 14 T 21 T 22 T 23 T 24 T 31 T 32 T 33 T 34 0 0 0 1 ] = [ c ( θ e ) σ 4 σ 1 s ( φ e ) σ 8 c ( φ e ) σ 9 σ 2 s ( θ e ) σ 4 l e ( c ( θ e ) σ 4 + σ 1 ) θ e t d ( s ( θ e ) σ 4 σ 2 ) + l s c ( φ s ) θ s + z e c ( φ s ) s ( θ s ) l e σ 4 θ e l s c ( θ s ) c ( φ s ) θ s σ 5 c ( φ e ) σ 12 s ( φ e ) σ 13 σ 6 t d σ 6 l e σ 5 θ e + l s s ( φ s ) θ s + z e s ( θ s ) s ( φ s ) + l e σ 10 θ e l s c ( θ s ) s ( φ s ) θ s σ 3 c ( θ e ) σ 11 c ( φ e ) s ( θ s ) σ 14 + s ( θ s ) σ 15 s ( φ e ) σ 7 z s + t d σ 7 + z e c ( θ s ) l e σ 11 θ e + l s s ( θ s ) θ s + l e ( σ 3 + c ( θ e ) σ 11 ) θ e 0 0 0 1 ]
where
σ 1 = c ( φ s ) s ( θ e ) s ( θ s ) ,   σ 2 = c ( θ e ) c ( φ s ) s ( θ s ) ,   σ 3 = c ( θ s ) s ( θ e ) ,   σ 4 = c ( φ e ) σ 8 + s ( φ e ) σ 9 ,   σ 5 = c ( θ e ) σ 10 s ( θ e ) s ( θ s ) s ( φ s ) ,   σ 6 = s ( θ e ) σ 10 + c ( θ e ) s ( θ s ) s ( φ s ) ,   σ 7 = c ( θ e ) c ( θ s ) s ( θ e ) σ 11 σ 8 = s ( φ s ) σ 14 c ( θ s ) c ( φ s ) σ 15 ,   σ 9 = σ 15 s ( φ s ) + c ( θ s ) c ( φ s ) σ 14 ,   σ 10 = c ( φ e ) σ 13 + s ( φ e ) σ 12 ,   σ 11 = c ( φ e ) s ( θ s ) σ 15 s ( θ s ) s ( φ e ) σ 14 ,   σ 12 = c ( φ s ) σ 15 c ( θ s ) s ( φ s ) σ 14 ,   σ 13 = c ( φ s ) σ 14 + c ( θ s ) σ 15 s ( φ s ) ,   σ 14 = s ( φ e + π 3 ) .
The pose Θ t 6 of the robot end effector can be calculated from (A2)
Θ t = [ P t b Ω t b ] T = [ P x P y P z Ω x Ω y Ω z ] T P x = T 14 ,   P y = T 24 ,   P z = T 34 Ω x = a tan 2 ( T 32 ,   T 33 ) ,   Ω y = a tan 2 ( T 31 ,   T 32 2 + T 33 2 ) ,   Ω z = a tan 2 ( T 21 ,   T 11 )
The velocity v t 6 of the robot end effector is
v t = Θ ˙ t = [ τ x τ y τ z ω x ω y ω z ] T
Thus, the Jacobian matrix J r 6 × 5 in (3) is used to represent the approximate linear relationship between the robot tip velocity v t and joint velocity Φ ˙ , which can be derived from the forward kinematics as follows:
v t = J r Φ ˙ [ τ x τ y τ z ω x ω y ω z ] = [ τ x z s τ x θ s τ x φ s τ x θ e τ x φ e τ y z s τ y θ s τ y φ s τ y θ e τ y φ e ω z z s ω z θ s ω z φ s ω z θ e ω z φ e ] [ z s θ s φ s θ e φ e ]

References

  1. Asawa, S.; Nüesch, M.; Gvozdenovic, A.; Aceto, N. Circulating Tumour Cells in Gastrointestinal Cancers: Food for Thought? Br. J. Cancer 2023, 128, 1981–1990. [Google Scholar] [CrossRef]
  2. Feng, R.-M.; Zong, Y.-N.; Cao, S.-M.; Xu, R.-H. Current Cancer Situation in China: Good or Bad News from the 2018 Global Cancer Statistics? Cancer Commun. 2019, 39, 22. [Google Scholar] [CrossRef]
  3. Lei, Z.-N.; Teng, Q.-X.; Tian, Q.; Chen, W.; Xie, Y.; Wu, K.; Zeng, Q.; Zeng, L.; Pan, Y.; Chen, Z.-S.; et al. Signaling Pathways and Therapeutic Interventions in Gastric Cancer. Signal Transduct. Target. Ther. 2022, 7, 358. [Google Scholar] [CrossRef]
  4. Huang, R.J.; Hwang, J.H. Improving the Early Diagnosis of Gastric Cancer. Gastrointest. Endosc. Clin. N. Am. 2021, 31, 503–517. [Google Scholar] [CrossRef]
  5. Hirao, M.; Masuda, K.; Asanuma, T.; Naka, H.; Noda, K.; Matsuura, K.; Yamaguchi, O.; Ueda, N. Endoscopic Resection of Early Gastric Cancer and Other Tumors with Local Injection of Hypertonic Saline-Epinephrine. Gastrointest. Endosc. 1988, 34, 264–269. [Google Scholar] [CrossRef]
  6. Maple, J.T.; Dayyeh, B.K.A.; Chauhan, S.S.; Hwang, J.H.; Komanduri, S.; Manfredi, M.; Konda, V.; Murad, F.M.; Siddiqui, U.D.; Banerjee, S. Endoscopic Submucosal Dissection. Gastrointest. Endosc. 2015, 81, 1311–1325. [Google Scholar] [CrossRef]
  7. Esaki, M.; Ihara, E.; Gotoda, T. Endoscopic Instruments and Techniques in Endoscopic Submucosal Dissection for Early Gastric Cancer. Expert Rev. Gastroenterol. Hepatol. 2021, 15, 1009–1020. [Google Scholar] [CrossRef] [PubMed]
  8. Kume, K. Flexible Robotic Endoscopy: Current and Original Devices. Comput. Assist. Surg. 2016, 21, 150–159. [Google Scholar] [CrossRef] [PubMed]
  9. Osawa, K.; Bandara, D.S.V.; Nakadate, R.; Nagao, Y.; Akahoshi, T.; Eto, M.; Arata, J. Stress Dispersion Design in Continuum Compliant Structure toward Multi-DOF Endoluminal Forceps. Appl. Sci. 2022, 12, 2480. [Google Scholar] [CrossRef]
  10. Zorn, L.; Nageotte, F.; Zanne, P.; Legner, A.; Dallemagne, B.; Marescaux, J.; de Mathelin, M. A Novel Telemanipulated Robotic Assistant for Surgical Endoscopy: Preclinical Application to ESD. IEEE Trans. Biomed. Eng. 2018, 65, 797–808. [Google Scholar] [CrossRef] [PubMed]
  11. Chiu, P.W.; Phee, S.J.; Bhandari, P.; Sumiyama, K.; Ohya, T.; Wong, J.; Poon, C.C.; Tajiri, H.; Nakajima, K.; Ho, K.Y. Enhancing Proficiency in Performing Endoscopic Submucosal Dissection (ESD) by Using a Prototype Robotic Endoscope. Endosc. Int. Open 2015, 3, E439–E442. [Google Scholar] [CrossRef]
  12. Ho, K.-Y.; Phee, S.J.; Shabbir, A.; Low, S.C.; Huynh, V.A.; Kencana, A.P.; Yang, K.; Lomanto, D.; So, B.Y.J.; Wong, Y.Y.J.; et al. Endoscopic Submucosal Dissection of Gastric Lesions by Using a Master and Slave Transluminal Endoscopic Robot (MASTER). Gastrointest. Endosc. 2010, 72, 593–599. [Google Scholar] [CrossRef] [PubMed]
  13. Knoop, R.F.; Wedi, E.; Petzold, G.; Bremer, S.C.B.; Amanzada, A.; Ellenrieder, V.; Neesse, A.; Kunsch, S. Endoscopic Submucosal Dissection with an Additional Working Channel (ESD+): A Novel Technique to Improve Procedure Time and Safety of ESD. Surg. Endosc. 2021, 35, 3506–3512. [Google Scholar] [CrossRef] [PubMed]
  14. Hwang, M.; Kwon, D.-S. K-FLEX: A Flexible Robotic Platform for Scar-Free Endoscopic Surgery. Int. J. Med. Robot. Comput. Assist. Surg. MRCAS 2020, 16, e2078. [Google Scholar] [CrossRef]
  15. Sarli, N.; Del Giudice, G.; Herrell, D.S.; Simaan, N. A Resectoscope for Robot-Assisted Transurethral Surgery1. J. Med. Devices 2016, 10, 020911. [Google Scholar] [CrossRef]
  16. Wagner, O.J.; Hagen, M.; Kurmann, A.; Horgan, S.; Candinas, D.; Vorburger, S.A. Three-Dimensional Vision Enhances Task Performance Independently of the Surgical Method. Surg. Endosc. 2012, 26, 2961–2968. [Google Scholar] [CrossRef]
  17. Sachdeva, R.; Traboulsi, E.I. Performance of Patients With Deficient Stereoacuity on the EYESi Microsurgical Simulator. Am. J. Ophthalmol. 2011, 151, 427–433.e1. [Google Scholar] [CrossRef] [PubMed]
  18. Bogdanova, R.; Boulanger, P.; Zheng, B. Depth Perception of Surgeons in Minimally Invasive Surgery. Surg. Innov. 2016, 23, 515–524. [Google Scholar] [CrossRef]
  19. Douissard, J.; Hagen, M.E.; Morel, P. The Da Vinci Surgical System. In Bariatric Robotic Surgery: A Comprehensive Guide; Domene, C.E., Kim, K.C., Vilallonga Puy, R., Volpe, P., Eds.; Springer International Publishing: Cham, Switzerland, 2019; pp. 13–27. ISBN 978-3-030-17223-7. [Google Scholar]
  20. Puntambekar, S.P.; Rajesh, K.N.; Goel, A.; Hivre, M.; Bharambe, S.; Chitale, M.; Panse, M. Colorectal Cancer Surgery: By Cambridge Medical Robotics Versius Surgical Robot System-a Single-Institution Study. Our Experience. J. Robot. Surg. 2022, 16, 587–596. [Google Scholar] [CrossRef]
  21. Chen, J.; Wang, S.; Zhao, Q.; Chen, M.; Liu, H. A Robotized Soft Endoscope with Stereo Vision for Upper Gastrointestinal Endoscopic Submucosal Dissection (ESD). In Proceedings of the 2023 45th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Sydney, Australia, 24–27 July 2023; pp. 1–6. [Google Scholar]
  22. Zhao, Q.; Lai, J.; Hu, X.; Chu, H.K. Dual-Segment Continuum Robot With Continuous Rotational Motion Along the Deformable Backbone. IEEE/ASME Trans. Mechatron. 2022, 27, 4994–5004. [Google Scholar] [CrossRef]
  23. Wockenfuß, W.R.; Brandt, V.; Weisheit, L.; Drossel, W.-G. Design, Modeling and Validation of a Tendon-Driven Soft Continuum Robot for Planar Motion Based on Variable Stiffness Structures. IEEE Robot. Autom. Lett. 2022, 7, 3985–3991. [Google Scholar] [CrossRef]
  24. Chen, M.; Huang, Y.; Chen, J.; Zhou, T.; Chen, J.; Liu, H. Fully Robotized 3D Ultrasound Image Acquisition for Artery. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 2690–2696. [Google Scholar]
  25. Li, Y.; Ng, W.Y.; Li, W.; Huang, Y.; Zhang, H.; Xian, Y.; Li, J.; Sun, Y.; Chiu, P.W.Y.; Li, Z. Towards Semi-Autonomous Colon Screening Using an Electromagnetically Actuated Soft-Tethered Colonoscope Based on Visual Servo Control. IEEE Trans. Biomed. Eng. 2023, 71, 77–88. [Google Scholar] [CrossRef]
  26. Xu, F.; Zhang, Y.; Sun, J.; Wang, H. Adaptive Visual Servoing Shape Control of a Soft Robot Manipulator Using Bézier Curve Features. IEEE/ASME Trans. Mechatron. 2023, 28, 945–955. [Google Scholar] [CrossRef]
  27. Fallah, M.M.H.; Norouzi-Ghazbi, S.; Mehrkish, A.; Janabi-Sharifi, F. Depth-Based Visual Predictive Control of Tendon-Driven Continuum Robots. In Proceedings of the 2020 IEEE/ASME International Conference on Advanced Intelligent Mechatronics (AIM), Boston, MA, USA, 6–9 July 2020; pp. 488–494. [Google Scholar]
  28. Chaumette, F.; Hutchinson, S. Visual Servo Control. I. Basic Approaches. IEEE Robot. Autom. Mag. 2006, 13, 82–90. [Google Scholar] [CrossRef]
  29. Jiang, J.; Wang, Y.; Jiang, Y.; Xie, H.; Tan, H.; Zhang, H. A Robust Visual Servoing Controller for Anthropomorphic Manipulators With Field-of-View Constraints and Swivel-Angle Motion: Overcoming System Uncertainty and Improving Control Performance. IEEE Robot. Autom. Mag. 2022, 29, 104–114. [Google Scholar] [CrossRef]
  30. Ke, F.; Li, Z.; Xiao, H.; Zhang, X. Visual Servoing of Constrained Mobile Robots Based on Model Predictive Control. IEEE Trans. Syst. Man Cybern. Syst. 2017, 47, 1428–1438. [Google Scholar] [CrossRef]
  31. Zhang, C.; Zhu, W.; Peng, J.; Han, Y.; Liu, W. Visual Servo Control of Endoscope-Holding Robot Based on Multi-Objective Optimization: System Modeling and Instrument Tracking. Measurement 2023, 211, 112658. [Google Scholar] [CrossRef]
  32. Roshanfar, M.; Taki, S.; Sayadi, A.; Cecere, R.; Dargahi, J.; Hooshiar, A. Hyperelastic Modeling and Validation of Hybrid-Actuated Soft Robot with Pressure-Stiffening. Micromachines 2023, 14, 900. [Google Scholar] [CrossRef]
  33. Lau, K.C.; Leung, E.Y.Y.; Chiu, P.W.Y.; Yam, Y.; Lau, J.Y.W.; Poon, C.C.Y. A Flexible Surgical Robotic System for Removal of Early-Stage Gastrointestinal Cancers by Endoscopic Submucosal Dissection. IEEE Trans. Ind. Inform. 2016, 12, 2365–2374. [Google Scholar] [CrossRef]
  34. Li, C.; He, S.; Xu, Y.; Li, D.; Guan, Y. Model and Control of Hybrid Hard-Soft Robots Using Model Predictive Control. In Proceedings of the 2022 41st Chinese Control Conference (CCC), Hefei, China, 25–27 July 2022; pp. 3705–3710. [Google Scholar]
  35. Abdulhafiz, I.; Nazari, A.A.; Abbasi-Hashemi, T.; Jalali, A.; Zareinia, K.; Saeedi, S.; Janabi-Sharifi, F. Deep Direct Visual Servoing of Tendon-Driven Continuum Robots. In Proceedings of the 2022 IEEE 18th International Conference on Automation Science and Engineering (CASE), Mexico City, Mexico, 20–24 August 2022; pp. 1977–1984. [Google Scholar]
  36. Greer, J.D.; Morimoto, T.K.; Okamura, A.M.; Hawkes, E.W. Series Pneumatic Artificial Muscles (sPAMs) and Application to a Soft Continuum Robot. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 5503–5510. [Google Scholar]
  37. Kadhim, I.; Choi, J. Extension Mechanism of a Flexible End for Colonoscopy. Med. Devices Evid. Res. 2020, 13, 245–258. [Google Scholar] [CrossRef]
  38. Ferhatoglu, M.F.; Kıvılcım, T.; Ferhatoglu, M.F.; Kıvılcım, T. Anatomy of Esophagus. In Esophageal Abnormalities; IntechOpen: London, UK, 2017; ISBN 978-953-51-3632-3. [Google Scholar]
  39. Colonoscopes—Gastroenterology. Available online: https://www.olympus.co.uk/medical/en/Products-and-solutions/Products/Gastroenterology/Colonoscopes.html (accessed on 6 February 2024).
  40. ELUXEO 700 Series Colonoscopes. Available online: https://healthcaresolutions-us.fujifilm.com/products/endoscopy/diagnostic-gastroenterology/eluxeo-700-series-colonoscopes (accessed on 6 February 2024).
  41. Ehrlich, D.; Muthusamy, V.R. Device Profile of the EXALT Model D Single-Use Duodenoscope for Endoscopic Retrograde Cholangiopancreatography: Overview of Its Safety and Efficacy. Expert Rev. Med. Devices 2021, 18, 421–427. [Google Scholar] [CrossRef] [PubMed]
  42. Webster, R.J.; Jones, B.A. Design and Kinematic Modeling of Constant Curvature Continuum Robots: A Review. Int. J. Robot. Res. 2010, 29, 1661–1683. [Google Scholar] [CrossRef]
  43. Hirschmuller, H. Stereo Processing by Semiglobal Matching and Mutual Information. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 328–341. [Google Scholar] [CrossRef] [PubMed]
  44. Viola, P.; Wells III, W.M. Alignment by Maximization of Mutual Information. Int. J. Comput. Vis. 1997, 24, 137–154. [Google Scholar] [CrossRef]
  45. Murray, R.M.; Li, Z.; Sastry, S.S. A Mathematical Introduction to Robotic Manipulation, 1st ed.; CRC Press: Boca Raton, FL, USA, 2017; ISBN 978-1-315-13637-0. [Google Scholar]
  46. Buss, S. Introduction to Inverse Kinematics with Jacobian Transpose, Pseudoinverse and Damped Least Squares Methods. IEEE Trans. Robot. Autom. 2004, 17, 16. [Google Scholar]
  47. Zhang, Z. A Flexible New Technique for Camera Calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef]
  48. Wang, J.; Olson, E. AprilTag 2: Efficient and Robust Fiducial Detection. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Republic of Korea, 9–14 October 2016; pp. 4193–4198. [Google Scholar]
Figure 1. (a) Flexible endoscopes provide effective minimally invasive treatment through different naturally narrow orifices in the human body. (b) DSRE promotes dexterity and flexibility for ESD, and stereo vision contributes to evaluation before and after surgery. (c) Schematic diagram of the DSRE silicone-body soft segment, which would bend when the air pressure in the chamber increases.
Figure 1. (a) Flexible endoscopes provide effective minimally invasive treatment through different naturally narrow orifices in the human body. (b) DSRE promotes dexterity and flexibility for ESD, and stereo vision contributes to evaluation before and after surgery. (c) Schematic diagram of the DSRE silicone-body soft segment, which would bend when the air pressure in the chamber increases.
Micromachines 15 00276 g001
Figure 2. (a) Fabrication process of the DSRE. (b) Detailed dimension. (c) DSRE with two cameras and an electrosurgical knife.
Figure 2. (a) Fabrication process of the DSRE. (b) Detailed dimension. (c) DSRE with two cameras and an electrosurgical knife.
Micromachines 15 00276 g002
Figure 3. (a) Configuration of the EM trackers on proximal and distal segments. (b) Bending angles of proximal and distal segments corresponding to the air pressure, the red lines correspond to the maximum air pressure exerted on each segment.
Figure 3. (a) Configuration of the EM trackers on proximal and distal segments. (b) Bending angles of proximal and distal segments corresponding to the air pressure, the red lines correspond to the maximum air pressure exerted on each segment.
Micromachines 15 00276 g003
Figure 4. (a) Illustration of coordinate frames. (b) The binocular stereo vision on the tip.
Figure 4. (a) Illustration of coordinate frames. (b) The binocular stereo vision on the tip.
Micromachines 15 00276 g004
Figure 5. The application of the SGM algorithm on the DSRE system.
Figure 5. The application of the SGM algorithm on the DSRE system.
Micromachines 15 00276 g005
Figure 6. Flow chart of the ASVS controller.
Figure 6. Flow chart of the ASVS controller.
Micromachines 15 00276 g006
Figure 7. Depth estimation validation setup. (a) A base tracker was used to measure the ground truth of depth. (b) The stereo vision on the tip of DSRE. (c) Image of left camera. (d) Image of right camera.
Figure 7. Depth estimation validation setup. (a) A base tracker was used to measure the ground truth of depth. (b) The stereo vision on the tip of DSRE. (c) Image of left camera. (d) Image of right camera.
Micromachines 15 00276 g007
Figure 8. Comparative error band analysis of the camera depth estimation results.
Figure 8. Comparative error band analysis of the camera depth estimation results.
Micromachines 15 00276 g008
Figure 9. (a) The actuator configuration. (b) Experimental setup for the ASVS controller performance verification.
Figure 9. (a) The actuator configuration. (b) Experimental setup for the ASVS controller performance verification.
Micromachines 15 00276 g009
Figure 10. Static target tracking results. (a) Distribution configuration of target markers. (b) Target movement trajectories in the left and right images, the tags moved from the start points and converged to the end points. (c) Tracking errors of different controllers.
Figure 10. Static target tracking results. (a) Distribution configuration of target markers. (b) Target movement trajectories in the left and right images, the tags moved from the start points and converged to the end points. (c) Tracking errors of different controllers.
Micromachines 15 00276 g010
Figure 11. Dynamic target tracking results. (a) Experiment setup. (b) Endoscopic view when the tag moved. (c) Tracking error of different controllers.
Figure 11. Dynamic target tracking results. (a) Experiment setup. (b) Endoscopic view when the tag moved. (c) Tracking error of different controllers.
Micromachines 15 00276 g011
Figure 12. Tracking results under unknown external disturbance. The upper pictures show the robot configurations before, during, and after the external force application. The figure below shows the changes in external force and tracking error results.
Figure 12. Tracking results under unknown external disturbance. The upper pictures show the robot configurations before, during, and after the external force application. The figure below shows the changes in external force and tracking error results.
Micromachines 15 00276 g012
Figure 13. Triangular trajectory tracking results. (a) Target movment in the left image. (b) Target movment in the right image.
Figure 13. Triangular trajectory tracking results. (a) Target movment in the left image. (b) Target movment in the right image.
Micromachines 15 00276 g013
Figure 14. Experimental setup of ex vivo trial in a porcine stomach.
Figure 14. Experimental setup of ex vivo trial in a porcine stomach.
Micromachines 15 00276 g014
Figure 15. ESD test results in ex vivo porcine stomach. (a) Marking points. (b) Cutting with assistance. (c) Cutting results. (d) Target tissue after dissection.
Figure 15. ESD test results in ex vivo porcine stomach. (a) Marking points. (b) Cutting with assistance. (c) Cutting results. (d) Target tissue after dissection.
Micromachines 15 00276 g015
Table 1. Comparison of the performance of the DSRE with related robot system.
Table 1. Comparison of the performance of the DSRE with related robot system.
StudyBackboneActuatorModelControllerTimeError
Zhao et al. [22]ContinuousPneumaticPCCJacobian estimation0.18 s0.62 mm
Zhang et al. [31]Rigid-GeometricVS with optimization-15.4629 Pixel
Roshanfar et al. [32]ContinuousPneumaticCosserat--MAE 5.79%
Chen et al. [21]ContinuousPneumaticPCCClassic IBVS13 iterationsRMSE 1.199 mm
Lau et al. [33]DiscreteCablePCCFeedforward0.25 s
Li et al. [34]ContinuousCablePCCMPC--
Abdulhafiz et al. [35]DiscreteCableCNNCNN-based VS85 iterationsSAD 0.058 mm
Greer et al. [36]ContinuoussPAMsPCCClassic IBVS-<5% in 2 s
DSRE in this studyContinuousPneumaticPCCASVS0.19 s13.51 Pixel
Table 2. Comparison of the DSRE system with previous design [21] and conventional endoscopes.
Table 2. Comparison of the DSRE system with previous design [21] and conventional endoscopes.
SystemsActuation
Type
SegmentsBody Length
(mm)
Diameter
(mm) 1
Max. Angles
(°)
CameraWorking Channel (mm)
DSRERobotized pneumatic272P 9, D 7P170 + D8822
Chen et al. [21]Robotized pneumatic285P 10, D 8P100 + D882s2
CF-XZ1200L/I [39]Manual cable-driven1-13.218013.7
EC-760S-V/L [40]Manual cable-driven1-12.818013.8
EXALT Model D [41]Manual cable-driven1-15.112014.2
1 P and D denote proximal and distal, respectively.
Table 3. Convergence times of ASVS and the classic control law with different constant η .
Table 3. Convergence times of ASVS and the classic control law with different constant η .
ControllersTag #1 (s)Tag #2 (s)Tag #8 (s)Tag #9 (s)Avg. (s)Dec. (%)
ASVS5.894.784.934.364.99-
η = 0.5
35.8713.8013.8714.0519.4074.28
η = 0.75
14.786.866.225.318.2939.81
η = 1
11.215.336.299.338.0437.94
Table 4. Control performance of ASVS and the classic control law with different constant η in [28].
Table 4. Control performance of ASVS and the classic control law with different constant η in [28].
ControllerRMSEMAESD
e max
Data (pixel)Dec.
(%)
Data (pixel)Dec.
(%)
Data (pixel)Dec.
(%)
Data (pixel)Dec.
(%)
ASVS13.48-12.46-5.14-27.24-
η = 0.5
57.6376.5752.0676.0724.7479.2286.2068.40
η = 0.75
19.3730.4117.8830.317.4430.9134.9121.97
η = 1
18.6527.7213.8810.2312.4758.7891.9570.38
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, J.; Wang, S.; Zhao, Q.; Huang, W.; Chen, M.; Hu, J.; Wang, Y.; Liu, H. Stereo Visual Servoing Control of a Soft Endoscope for Upper Gastrointestinal Endoscopic Submucosal Dissection. Micromachines 2024, 15, 276. https://doi.org/10.3390/mi15020276

AMA Style

Chen J, Wang S, Zhao Q, Huang W, Chen M, Hu J, Wang Y, Liu H. Stereo Visual Servoing Control of a Soft Endoscope for Upper Gastrointestinal Endoscopic Submucosal Dissection. Micromachines. 2024; 15(2):276. https://doi.org/10.3390/mi15020276

Chicago/Turabian Style

Chen, Jian, Shuai Wang, Qingxiang Zhao, Wei Huang, Mingcong Chen, Jian Hu, Yihe Wang, and Hongbin Liu. 2024. "Stereo Visual Servoing Control of a Soft Endoscope for Upper Gastrointestinal Endoscopic Submucosal Dissection" Micromachines 15, no. 2: 276. https://doi.org/10.3390/mi15020276

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop