Optimal Image-Based Guidance of Mobile Manipulators Using Direct Visual Servoing

: This paper presents a direct image-based controller to perform the guidance of a mobile manipulator using image-based control. An eye-in-hand camera is employed to perform the guidance of a mobile differential platform with a seven degrees-of-freedom robot arm. The presented approach is based on an optimal control framework and it is employed to control mobile manipulators during the tracking of image trajectories taking into account robot dynamics. The direct approach allows us to take both the manipulator and base dynamics into account. The proposed image-based controllers consider the optimization of the motor signals sent to the mobile manipulator during the tracking of image trajectories by minimizing the control force and torque. As the results show, the proposed direct visual servoing system uses the eye-in-hand camera images for concurrently controlling both the base platform and robot arm. The use of the optimal framework allows us to derive different visual controllers with different dynamical behaviors during the tracking of image trajectories.


Introduction
This paper proposes a direct image-based visual servoing system for the guidance of a mobile manipulator.As described throughout the paper, the visual information is employed to guide both the robot manipulator and the base platform for the tracking of image trajectories.A common classification for visual servoing systems based on position and image is established [1].Within the field of position-based visual servoing systems, it is worth mentioning some classic works [2] as well as some more recent ones such as the works of Huang et al. [3].Besides these two works, it is worth mentioning other remarkable ones such as the works of Park et al. [4], which describes the use of vision for a robot's positioning with regard to an object whose position was previously determined.Position-based systems are those where the reference of the control system is in a desired tridimensional position and orientation.In contrast to position-based approaches, image-based visual servoing systems determine the reference in the image space.This strategy is the one used in this paper.Although the literature regarding image-based visual servoing is quite extensive, works, such as the presented in [1,5], perform a review of the classical approaches and problems related to indirect image-based controllers.One of the elements which most affects the behavior of image-based visual servoing controllers is the interaction matrix, which is employed by the controller and depends on several parameters.The interaction matrix J s is employed by the image-based visual servoing controllers for relating velocities in a tridimensional point with velocities of a corresponding point in the image space.Deep learning approaches are currently in development for image-based control of robot manipulators (see e.g.[6,7]).When this kind of approaches has to be applied to a mobile manipulator, it is necessary to consider not only the manipulator but also the mobile platform, having a redundant system.When dealing with stability problems, it is worth mentioning those that stem from dealing with a redundant robot [8].In this case, it is not only necessary to correctly perform the end effector guidance, but also to solve the redundancy by establishing the most appropriate joint configuration for each situation.
Another useful classification depending on the control strategy of visual servoing systems divides them in direct and indirect visual servoing.In indirect visual servoing, the control action is specified as velocities to be applied at the robot's end effector and does not take robot dynamics into account.In this case, the internal robot controller is used and translates these velocities into torques and forces for the joints.When applying an indirect controller to a task only kinematic aspects of the robot are taken into account.However, in direct controllers, the control action is usually given as forces and torques applied directly to the manipulator joints.By using this approach, the internal feedback control loop for the servomotors is removed, and the visual information is used to directly generate the currents or torques to be applied to the motors of the robot's every joint.Some direct visual control systems for redundant robots with chaos compensation have been developed to avoid robotic chaotic behaviors with new architectures to improve systems maintainability and traceability [9,10].Within the field of direct image-based visual servoing systems it is worth mentioning works such as [11] that considers a FPGA (Field Programmable Gate Arrays) implementation, or [12] that extends the previous direct visual servoing for the optimal control of robot manipulators.
Within the field of indirect visual servoing, several approaches can be found for the guidance of mobile robots or mobile manipulators.In [13], a homography-based 2D approach for visual control of a mobile manipulator that does not need any measure of the 3D structure of the observed target has been proposed.A task-space sensing and control system designed to control the end effector motion of a mobile manipulator in the presence of dynamic and unknown base motion is proposed in [14].In [15], an image-based approach with redundancy resolution is proposed for the guidance of a mobile manipulator.An indirect position-based visual control approach is described in [16] for the guidance of a mobile manipulator.In [17], a control scheme that uses image-based visual servoing was developed along with a functional mobile manipulator platform, where a hybrid camera configuration composed of monocular and stereo vision cameras was integrated into the system.
In contrast with previous approaches, the one proposed in this paper employs a direct visual servoing system for guiding mobile manipulators.In this case, the controller directly generates the torques, forces, and moments to be applied to the manipulator and base platform taking into account the robot dynamics.In the previous indirect approaches, the controller assumes that the guided device is a perfect positioning system, and therefore its dynamics is not considered.However, the use of a direct visual servo approach allows for the manipulator and base dynamics to be taken into account.The result is a faster and more accurate control that reacts more quickly to abrupt changes in the image trajectories.Additionally, an optimal control approach is used for the definition of the proposed controllers.This control approach considers the optimization of the motor signals sent to the mobile manipulator during the tracking of image trajectories.Using this optimal approach, different visual controllers with different dynamic properties can be derived.
The paper is divided into the following sections.First, Section 2 describes the robotic system and main coordinate frames.The system kinematics is described in Section 3 and the dynamics in Section 4. Section 5 describes the optimal controller and the visual servoing of the mobile manipulator.Finally, in Section 6, the main results obtained in the application of the proposed optimal controller are described.

System Architecture and Nomenclature
Figure 1 represents the main components of the mobile manipulator (TIAGo robot from PAL Robotics).The robot is composed of a mobile differential platform with a 7-DOF (degrees-of-freedom) arm, a RGB-D camera which is not employed in this paper, and an additional eye-in-hand camera.Frame F p is attached to the mobile robot platform and F m is the manipulator arm coordinate system attached to the base of the robot arm.The origin of F m is translated with respect to the frame F p a vector (x pm , y pm , z pm ).The world coordinate frame is called F o .The end effector frame F e is attached to the manipulator end effector, and frame F c is the eye-in-hand camera frame.In this paper we consider an eye-in-hand camera system where F c = F e .The orientation of the mobile platform is considered as α, i.e., the angle between de X-axis of the frame F p and the X-axis of the frame F o and the origin of F p is at the position (x op , y op , z op ) with respect to F o .From the manipulator kinematics we can obtain the position of F e with respect to F m as (x me , y me , z me ) and (φ xme , φ yme , φ zme ), respectively.
modules that compose the computer vision system.The first step preprocesses the image in order to obtain a single format of the image no matter what camera is employed (e.g., image conversion from RGB to gray scale, image resizing, noise filtering, etc.).In the second stage, the image is binarized through a thresholding process.The result is a binary image with all the essential data of the pattern.The third step is an erosion operation.This is a morphological operation on the image that removes the pixels located on the edge of any object in the image.This step is mainly used to reduce the noise of the image.The fourth phase is the most important stage of the vision module: blob detection.The purpose of this module is to detect the connected pixels which form an object in the image and then assign a label to them.Finally, the vision module has to provide the pixel coordinates of the center of gravity of the visual features.The fifth submodule (area and centroid computation) accomplishes this task.In [12], a detailed description of the computer vision components required to determine the set of visual features is presented.

System Kinematics
The main components and coordinate frames of the robotic system are presented in Section 2. This section describes the kinematic formulation required for the definition of the proposed controllers. ∈ ℜ m represent the generalized joint coordinates of the manipulator (in our case, m  Regarding the visual system, an eye-in-hand camera observes a set of four visual features located on the surface of the tracked object.As Figure 1 shows, a black pattern with four white points is located in the tracked object.Therefore, the computer vision extracts four visual feature points, T that correspond to the center of these white points.Figure 2 shows the different modules that compose the computer vision system.The first step preprocesses the image in order to obtain a single format of the image no matter what camera is employed (e.g., image conversion from RGB to gray scale, image resizing, noise filtering, etc.).In the second stage, the image is binarized through a thresholding process.The result is a binary image with all the essential data of the pattern.The third step is an erosion operation.This is a morphological operation on the image that removes the pixels located on the edge of any object in the image.This step is mainly used to reduce the noise of the image.The fourth phase is the most important stage of the vision module: blob detection.The purpose of this module is to detect the connected pixels which form an object in the image and then assign a label to them.Finally, the vision module has to provide the pixel coordinates of the center of gravity of the visual features.The fifth submodule (area and centroid computation) accomplishes this task.In [12], a detailed description of the computer vision components required to determine the set of visual features is presented.
Electronics 2019, 8, 374 3 of 13 effector frame ℱ is attached to the manipulator end effector, and frame ℱ is the eye-in-hand camera frame.In this paper we consider an eye-in-hand camera system where ℱ ℱ .The orientation of the mobile platform is considered as α, i.e., the angle between de X-axis of the frame ℱ and the X-axis of the frame ℱ and the origin of ℱ is at the position ( ,  ,  ) with respect to ℱ .From the manipulator kinematics we can obtain the position of ℱ with respect to ℱ as ( ,  ,  ) and ( ,  ,  ), respectively.Regarding the visual system, an eye-in-hand camera observes a set of four visual features located on the surface of the tracked object.As Figure 1 shows, a black pattern with four white points is located in the tracked object.Therefore, the computer vision extracts four visual feature points,  =  ,  ,  ,  , T that correspond to the center of these white points.Figure 2 shows the different modules that compose the computer vision system.The first step preprocesses the image in order to obtain a single format of the image no matter what camera is employed (e.g., image conversion from RGB to gray scale, image resizing, noise filtering, etc.).In the second stage, the image is binarized through a thresholding process.The result is a binary image with all the essential data of the pattern.The third step is an erosion operation.This is a morphological operation on the image that removes the pixels located on the edge of any object in the image.This step is mainly used to reduce the noise of the image.The fourth phase is the most important stage of the vision module: blob detection.The purpose of this module is to detect the connected pixels which form an object in the image and then assign a label to them.Finally, the vision module has to provide the pixel coordinates of the center of gravity of the visual features.The fifth submodule (area and centroid computation) accomplishes this task.In [12], a detailed description of the computer vision components required to determine the set of visual features is presented.

System Kinematics
The main components and coordinate frames of the robotic system are presented in Section 2. This section describes the kinematic formulation required for the definition of the proposed controllers. ∈ ℜ m represent the generalized joint coordinates of the manipulator (in our case, m

System Kinematics
The main components and coordinate frames of the robotic system are presented in Section 2. This section describes the kinematic formulation required for the definition of the proposed controllers.q T m ∈ m represent the generalized joint coordinates of the manipulator (in our case, m = 7) and q T p ∈ np are the generalized coordinates of the platform.The generalized coordinates of both manipulator and platform can be represented as q = q T p q T m .In a general case, the computer vision system extracts k visual feature points from the observed object s = f 1, f 2, . . ., f k T ∈ R k (k = 4 in the experiments, see Figure 1).Therefore, the image-based direct visual controller must perform the guidance of the mobile manipulator to track the desired trajectory in the image space, s d (t).The visual servoing control approach allows the control of the manipulator joints and base using k visual features, using an eye-in-hand configuration, where a camera is held by the end effector.Considering an image feature point, f i, , the relationship between velocities in the camera image space, .f i , and the velocity twist .
of the camera (expressed in its own frame), is given by: .
where Z is the depth of the image feature and the value of J fi can be found in [1].From a kinematic point of view, the motor commands are considered as a set of platform command velocities, u p ∈ R p , and manipulator command velocities, u m ∈ R m , therefore In the above equation the manipulator motor commands are the joint velocities, i.e., . q m = u m , and from the kinematic model of the mobile platform the following relation can be considered: .
where, usually, u p = v ω T is the platform linear and angular velocities and the function ψ spans the admissible velocity space at each mobile platform configuration.In this paper, an image-based approach will be employed for the guidance of the mobile manipulator, therefore, an image Jacobian, J i , must be defined which relates the differential mapping between the motor commands u and the time derivative of the set of image features s: where J i will be defined throughout the paper as the product of the matrices J s and J c , therefore, J i = J s J c .The interaction matrix J s ∈ R 2k×6 for the set of image features points, s, is the stack of k matrices J f i for each feature: .
where Z = Z 1 . . .Z k T ∈ R k is a depth vector for each image feature.The matrix J c relates the end effector velocity with respect to the camera coordinate frame v c ω c T and the motor command u: As previously described, the mobile manipulator is composed of a differential drive platform and a 7-DOF robot arm with an eye-in-hand camera.In order to derive the value of the matrix J c , first the position and orientation of the camera with respect to the world coordinate F o is defined as where t T (q) represents the position and φ T (q) is the minimal parametrization of the camera orientation by Euler angles.From the robot kinematics and the system architecture presented in Section 2, it is possible to obtain t y = x me + x pm sinα + y me + y pm cosα + y op (9) From the previous equations we can obtain the time derivatives of r, i.e., v T ω T T .First, the components of the translation velocity, .

t
T (q), can be defined as .
Therefore, the motion of the manipulator end depends not only on the joint positions and velocities of the manipulator arm but also on the motion of the mobile platform and its orientation α.
Additionally, the components of the angular speed, .φ T (q) of the robot can be computed as [15]   . . .
From Equations ( 11)-( 16 Another useful kinematic relationship for the definition of the proposed controllers is the image acceleration or second time derivative of s.This relationship is obtained by the definition of J i (Equation ( 4)) and differentiating w.r.t.

System Dynamics
This section describes the main dynamic equations for the robotic system.The dynamic equations of motion of the mobile manipulator can be written as Electronics 2019, 8, 374 6 of 13 where .. q m ∈ m is the set of joint accelerations of the robot manipulator, .
the platform linear and angular accelerations, M bb ∈ p×p is the inertia matrix of the platform, M bm ∈ p×m is the coupled inertia matrix of the base and the manipulator, and M mm ∈ m×m is the inertia matrix of the manipulator; c b and c m are a velocity/displacement-dependent, nonlinear terms for the base and manipulator, respectively, F b is the force and moment exerted on the base, and τ is the applied joint torque on the robot manipulator.From a dynamic point of view, the motion of the mobile manipulator is governed by the applied torques on the manipulator joints and by the force and moment exerted on the base.Hence, Equation ( 18) can be written in the following compact form. where q m , and C = c b c m .

Optimal Control Definition
In complex high dimensional systems, several tasks can be accomplished simultaneously taking advantage of the robot redundancy.This paper considers a system with µ constrains that represents the task for the mobile manipulator to be executed.These constraints can be represented as where A (t) ∈ µ×(p+m) and b (t) ∈ µ×1 .An advantage of this task formulation is that nonholonomic constraints can be treated in the same general way.In order to reduce the energy required for performing the visual servoing task, the proposed optimal controller is designed to minimize the control torque for the mobile manipulator, considering the following cost function.
where W(t) is a time-dependent weight matrix.Let us assume that the robot model presented in the previous section and a constraint formulation are given.In this case, the optimal controller has to guarantee that the task is perfectly achieved, i.e., that A (t) .u = b(t) holds at all times.Additionally, the minimization of the control command with respect to some given metric, i.e., Ω(t) = τ T W(t) τ, is intended at each instant of time.The solution to this pointwise optimal control problem [18] can be derived from a generalization of Gauss' principle, as originally suggested in [19].It is also a generalization of the propositions in [20,21], where the following control action is obtained considering the general expression for the dynamic model expressed in Equation ( 19) (for the sake of clarity the time dependences are not indicated): where the symbol + denotes the pseudo inverse of a general matrix.As it can be seen in Equation ( 22), the matrix W is an important factor in the control law that can be used to determine how the control efforts are distributed over the joints and the robot base.

Optimal Direct Visual Servoing
Once the optimal control framework is presented in Section 5.1, this section describes its application for the optimal control of the mobile platform.First, we define .. s r as the reference image accelerations that will be employed by the optimal controller.The value of these image accelerations can be obtained from Equation (17).This last expression can be expressed into the form of Equation ( 20) as This way, the visual servoing task is defined by the following relationships.
Therefore, with the given definition of A and b, the optimal control will minimize the motor commands while performing the tracking in the image space.The final control law can be obtained replacing these variables into the function that minimizes the motor signals described by Equation ( 22): As it can be seen, the visual controller represented by (25) implicitly depends on the weighting matrix W, and different values of this matrix can be used to coordinate the motion of the base and of the manipulator.The diagram of the considered control loop is depicted in Figure 3. Considering  M -2 and replacing this value in the control framework expressed in Equation ( 25), the result yields This controller represents a direct visual servo control using inversion of the dynamic model.Additionally, new controllers can be obtained using different values of W. For example, an important value for W, due to its physical interpretation, is  M -1 , since it is consistent with the principle of d'Alembert: Now, the definition of the reference control, s r , is described considering the eye-in-hand camera system that extracts a set of k image feature points.The task description as a constraint is given by the following equation in the image space.
where  ,  , and  are the desired image space accelerations, velocities, and positions, respectively.K and K are proportional and derivative gain matrices, respectively.This equation can be expressed in regards to image error in the following way.
where  and  are the image error and the time derivative of the image error, respectively.As stated, the variable  denotes the reference image accelerations of the proposed image space based controller.Substituting this variable into the dynamic visual servo controller in Equation ( 25), the control law is set by the following relationship.
In order to demonstrate the asymptotic tracking of the control law, some operations must be Considering W = M −2 and replacing this value in the control framework expressed in Equation ( 25), the result yields This controller represents a direct visual servo control using inversion of the dynamic model.Additionally, new controllers can be obtained using different values of W. For example, an important value for W, due to its physical interpretation, is W = M −1 , since it is consistent with the principle of d'Alembert: Now, the definition of the reference control, .. s r , is described considering the eye-in-hand camera system that extracts a set of k image feature points.The task description as a constraint is given by the following equation in the image space. .. .
s d , and s d are the desired image space accelerations, velocities, and positions, respectively.K P and K D are proportional and derivative gain matrices, respectively.This equation can be expressed in regards to image error in the following way. ..
where e s and .e s are the image error and the time derivative of the image error, respectively.As stated, the variable .. s r denotes the reference image accelerations of the proposed image space based controller.Substituting this variable into the dynamic visual servo controller in Equation (25), the control law is set by the following relationship.
In order to demonstrate the asymptotic tracking of the control law, some operations must be done.First, the closed-loop behavior is computed from Equation (19) as Equation ( 31) can be simplified by premultiplying its left and right sides by Finally, using the relationship expressed in (23), it can be concluded that . .
Therefore, when J i is full-rank, an asymptotic tracking is achieved by the visual servo controller expressed in Equation (30) for the tracking of an image trajectory.

Results
This section presents the system behavior during the tracking of different trajectories.The robot is guided by an eye-in-hand camera system.The parameters of a Gigabit Ethernet TM6740GEV camera is considered, which acquires 200 images every second with a resolution of 1280 × 1024 pixels.The eye-in-hand camera extracts four visual features from the four corners of the target.
Three different experiments are presented in this section.In these experiments the value of W = M −1 is considered.Figure 4a,b represents the initial configuration employed in all of the experiments.Figure 4a represents the target and the mobile manipulator with the eye-in-hand camera system.The initial position of the visual features is represented in Figure 4b.The desired trajectory employed in the first experiment is a linear trajectory that represents a lateral translation of 200 px. in x and y directions in the image from the initial position of the visual features.As it can be seen in Figure 4d, the image-based controller correctly carries out the tracking and a straight line is obtained in the image space.Figure 4c represents the final pose of the mobile manipulator.The second experiment, represented in Figure 4e,f, evaluates the tracking when a displacement in depth must be performed.As it can be seen in Figure 4e, a displacement in depth is carried out and, once the experiment ends, the mobile manipulator is closer to the target.Figure 4f represents the obtained image trajectory.This trajectory corresponds to the desired trajectory specified as a linear increase in depth from the initial image configuration.Finally, the experiment represented in Figure 4g,h requires a change in orientation of the eye-in-hand camera with respect to the observed target.To do this, the desired image trajectories for the visual features are represented by a straight line between the initial and the final ones defined as s * = [f 1d = (503, 584), f 2d = (612, 600), f 3d = (523, 472), f 4d = (630, 477)] T px.As in the previous experiments, Figure 4g represents the final robot pose once the experiment ends and Figure 4h represents the image trajectory.As the experiments show, the proposed controller allows the concurrent guidance of both the manipulator and robotic platform during the tracking of the desired image trajectory.Figure 5 represents the joint torques applied to the manipulator (m = 1 . . .7) and force and moments applied to the robot platform during the first and second experiments.As it can be seen in Figure 5a (first experiment), the torques, force, and moment remain low, and a smooth behavior is obtained during the tracking.For the second experiment, Figure 5b represent the joint torques applied to the manipulator and force and moment applied to the base platform during the robot motion in depth.It is worth noting that the reduction of the distance between the two bodies (target and robot) is obtained by modifying both the base platform pose and the robot manipulator joint configuration.Finally, as in the previous experiments, Figure 6 represents the joint torques applied to the manipulator and force and moment applied to the base platform during the third tracking experiment.The desired relative orientation is achieved by controlling both the orientation of the base platform and the robot manipulator joint configuration.One of the main advantages of the proposed controller is the possibility to define new direct image-based controllers by modifying the value of the matrix W of the optimal controller (see Equation ( 25)).A different controller with different dynamic behavior during the tracking of image trajectories can be obtained by modifying this matrix.In order to evaluate the proposed framework, in Table 1 we represent the obtained image error during the tracking of a repetitive image trajectory   One of the main advantages of the proposed controller is the possibility to define new direct image-based controllers by modifying the value of the matrix W of the optimal controller (see Equation ( 25)).A different controller with different dynamic behavior during the tracking of image trajectories can be obtained by modifying this matrix.In order to evaluate the proposed framework, in Table 1 we represent the obtained image error during the tracking of a repetitive image trajectory defined by the following equation.One of the main advantages of the proposed controller is the possibility to define new direct image-based controllers by modifying the value of the matrix W of the optimal controller (see Equation ( 25)).A different controller with different dynamic behavior during the tracking of image trajectories can be obtained by modifying this matrix.In order to evaluate the proposed framework, in Table 1 we represent the obtained image error during the tracking of a repetitive image trajectory defined by the following equation.
Additionally, the performance of the controllers is compared with respect to previous controllers only based on inverse dynamics [22] (indicated as "previous" in Table 1).

Conclusions
This paper presents a direct image-based visual servoing system for the guidance of mobile manipulators.This approach is based on an optimal framework that considers the optimization of the motor signals sent to the mobile manipulator during the tracking of image trajectories.The stability of the controller has been proved analytically, and different dynamical behavior can be obtained by using different values of the matrix W.
As a result, an adaptive and flexible tool has been obtained, allowing for concurrent control of both the base platform and manipulator during the tracking of trajectories.With the given definition of A and b, the optimal control can minimize the motor commands while performing the tracking of trajectories in the image space.The viability of the controller has been tested in a variety of test case experiments, including approaches to an object, as well as different positioning tasks that require modifying the position and/or the orientation of the mobile manipulator.In all such cases, the experiments and trajectories have been accomplished.The relation between the previous Cartesian velocities and v T ω T T can be obtained by using Equations ( 11)-( 16): Therefore, the value of J c is equal to

.
) the values of v T ω T T to the motor commands.Using the rotation matrix R c , which represents the orientation frame F c with respect to F o , the values of v T ω T T can be expressed with respect to the frame F c , v T Appendix A describes the computation of the numerical values of matrix J c .

Figure 3 .
Figure 3. Diagram of the control scheme for the mobile manipulator robot.

Figure 3 .
Figure 3. Diagram of the control scheme for the mobile manipulator robot.

Figure 4 .
Figure 4. Robot pose and image trajectories in the experiments.(a) 3D initial pose of the mobile manipulator and target.(b) Initial visual features extracted by the eye-in-hand camera.First Figure 4. Robot pose and image trajectories in the experiments.(a) 3D initial pose of the mobile manipulator and target.(b) Initial visual features extracted by the eye-in-hand camera.First

Figure 4 .
Figure 4. Robot pose and image trajectories in the experiments.(a) 3D initial pose of the mobile manipulator and target.(b) Initial visual features extracted by the eye-in-hand camera.First experiment: (c) 3D final pose.(d) Image trajectory.Second experiment: (e) 3D final pose.(f) Image trajectory.Third experiment: (g) 3D Final pose.(h) Image trajectory.

Figure 5 .
Figure 5. (a) Experiment 1: Joint torque and moment and force exerted on the base platform.(b) Experiment 2: Joint torque, moment, and force exerted on the base platform.

Figure 6 .
Figure 6.Experiment 3: Joint torque, moment, and force exerted on the base platform.

Figure 5 .
Figure 5. (a) Experiment 1: Joint torque and moment and force exerted on the base platform.(b) Experiment 2: Joint torque, moment, and force exerted on the base platform.

Figure 5 .
Figure 5. (a) Experiment 1: Joint torque and moment and force exerted on the base platform.(b) Experiment 2: Joint torque, moment, and force exerted on the base platform.

Figure 6 .
Figure 6.Experiment 3: Joint torque, moment, and force exerted on the base platform.

Figure 6 .
Figure 6.Experiment 3: Joint torque, moment, and force exerted on the base platform.
sinα − x pm sinα x me cosα + x pm cosα 0 using the rotation matrix R c which represents the orientation frame F c with respect to the fame F o , the values of v T ω T T can be expressed with respect to the frame F c ,

Table 1 .
Mean image error in pixels (mm) during the tracking of the image trajectory.

Table 1
represents the mean error in pixels and mm during the tracking of the image trajectory considering different tracking velocities.As it can be seen in this table, different tracking errors are obtained and the best behavior at high velocities is obtained considering W