Abstract
For the constrained mobile robot automatic parking system, the hybrid model predictive control with a penalty factor based on image-based visual servoing (IBVS) is proposed to address the problem of feature point loss and emergency braking in dynamic obstacle scenarios caused by excessive target bias gain when using traditional IBVS control methods. The traditional IBVS control is transformed into an optimization problem with constraints in the finite time domain, by defining the optimization function based on the mobile robot’s positional deviation and image feature point deviation, while using actuator saturation and speed limit as constraints. Based on this, a convex optimization function with penalty factors is defined and combined with incremental model predictive control. This control strategy could ensure the emergency braking performance of the mobile robot when the image feature points are massively obscured by obstacles in dynamic scenes, while improving the accuracy and real-time of its trajectory tracking control. Finally, simulation comparisons are conducted to verify the effectiveness of the proposed control method.
1. Introduction
With the development of science and technology, automatic parking has become a research hotspot, and the trajectory tracking control of mobile robots is a key technology related to the accuracy of automatic parking [1]. In particular, the research of visual servo-based automatic parking and emergency braking for mobile robots constrained by feature points has received increasing attention [2].
Classical visual servo control is mainly divided into position-based visual servoing control (PBVS), image-based visual servoing control (IBVS) and hybrid visual servoing control [3]. The PBVS control needs to define the error signal in the three-dimensional Euclidean space for three-dimensional reconstruction. The direct error control in the three-dimensional space can ensure its convergence [4]. However, this method is vulnerable to system calibration errors and visual measurement noise [5]. The IBVS is controlled by defining the error signal in the two-dimensional image plane [6]. It does not need the 3D pose information of the mobile robot and is robust to the system disturbances [7]. The IBVS has been combined with advanced control algorithms to realize automatic parking for mobile robots. In [8], for the problem of lack of depth information and parameter estimation in the visual servo control of mobile robots, the sliding mode control method based on IBVS three-view pair-pole geometry is proposed, to solve the depth correction problem and ensure the closed-loop stability of the system by expanding the pair-pole geometry to three pictures. In [9], for the limitation that the camera plane must be parallel to the motion plane in the visual servo control of mobile robots, the IBVS-based adaptive controller is proposed, which has a simple structure and high efficiency to achieve good trajectory tracking control performance. Nonetheless, the target deviation gain of the mobile robot visual servo control system is prone to excessive selection, when the camera parameters are not accurately calibrated. Then, the above control methods are difficult to deal with the problem of feature point loss and system instability caused by excessive target deviation gain.
For the problem of losing feature points and system instability due to excessive target deviation gain in the vision servo control system of mobile robots, it is necessary to constrain image feature points to maintain them within the field of view at all times [10]. The existing studies are divided into two main approaches: algorithm design and hardware adjustment. In [11], for the visual servo control of mobile robots, a time-varying continuous hybrid visual servo controller is designed to realize consistent tracking and pose correction for mobile robots, to solve the problem that the target feature points are difficult to maintain within the field of view of the camera, and to ensure the stability of feature point tracking; in [12], for the image feature point constraint problem, a two-stage active visual servo scheme for mobile robots is proposed. By using an external mobile platform camera device, it could keep the feature points around the center of the image plane to achieve mechanical tracking of the image feature points and ensure that the image feature points are always maintained within the field of view. However, none of the above methods are calibrated with camera specific parameters. In [13], for the problem of constrained mobile robot visual servo control when the camera parameters are not accurately calibrated, the specific parameters of the camera are estimated and an adaptive continuous visual servo controller is designed to realize the trajectory tracking control of the mobile robot under the image feature point constraint. The internal constraints such as actuator saturation and speed limitation are not considered in the above methods, and these methods are not universally applicable.
Model predictive control (MPC) is a model-based optimal control strategy [14], which owns superiorities such as rolling optimization, constraint handling, and state prediction [15]. It is suitable for handling the internal constraints of the control system and predicting the system state [16]. The model predictive control algorithm has been introduced to the IBVS-based mobile robot control system. In [17], for the mobile robot automatic parking control system, the constrained IBVS-based model predictive control method is proposed to ensure the accuracy of the mobile robot trajectory tracking control, by considering the internal constraints such as actuator saturation and speed limitation. But the computational burden is large and it is difficult to be applied practically. In [18], for the problem of poor real-time image processing and rolling optimization of the combination of IBVS and MPC, the IBVS-based model predictive control self-triggering switching scheme is designed to reduce the computation time, by eliminating the need for visual measurement and computation of control inputs at two consecutive triggering moments. But it is difficult to meet the high accuracy requirements of mobile robot trajectory tracking control. To solve the problem of difficulty in ensuring both real-time and accuracy of the IBVS-based mobile robot model predictive control system mentioned above, the IBVS-based incremental MPC scheme is proposed in [19]. This method reconstructs incremental control quantities, while considering internal constraints such as actuator saturation and speed limits, to improve the real-time and accuracy of mobile robot trajectory tracking. However, this method only considers the image feature point deviation when defining the optimization function, and does not incorporate the positional deviation of the mobile robot. When the image feature points are intermittently lost due to external disturbances, it is easy to cause a decrease in the accuracy of trajectory tracking control for mobile robots.
In addition, the automatic parking process of mobile robots is prone to emergency braking problems when dynamic obstacles emerge. In [20], for the mobile robot system in narrow environments, the MPC hierarchical obstacle avoidance framework based on scene search is presented. The upper-level scene search scheme is used to plan the obstacle avoidance path, and the lower-level MPC is designed to control the mobile robot to park automatically. Then, the emergency braking control of the mobile robot in dynamic obstacle scenarios is realized. In [21], for the emergency braking problem in the presence of a sudden dynamic obstacle, the predictive obstacle avoidance control scheme is proposed. It is used to avoid dynamic obstacles by using pre-sighting point search methods for local path planning. Nonetheless, all of the above obstacle avoidance algorithms use the search-before-braking scheme, which is easy to lead to emergency braking lag and difficult to ensure the safety and real-time of automatic parking for mobile robots.
In response to the above issues, the main contributions of this paper are reflected in the following aspects:
- (1)
- For the mobile robot system subject to feature point motion constraints, the IBVS-based incremental model predictive control algorithm is designed, to solve the problem of feature point loss and system instability due to excessive target deviation gain when the traditional IBVS control method is applied to an automatic parking control system. The traditional IBVS control is transformed into an optimization problem with constraints in the finite time domain, by defining the optimization function based on the mobile robot’s positional deviation and image feature point deviation, while using actuator saturation and speed limit as constraints. Then, the accuracy and real-time of the mobile robot tracking control during automatic parking is improved simultaneously.
- (2)
- For the problem of emergency braking of mobile robot automatic parking in dynamic obstacle scenes, by defining the convex optimization function with penalty factor, the hybrid model predictive control with a penalty factor based on IBVS (IBVS-PF-HMPC) is proposed. Then, it could guarantee the emergency braking performance of the mobile robot automatic parking when the image feature points are massively obstructed by obstacles in dynamic scenes.
This work is organized as follows: Section 2 presents the problem formulation, such as the kinematic model and the IBVS model. The formulation of the mobile robot controller is presented in Section 3 with an IBVS-PF-HMPC scheme. Section 4 shows the comparative results through simulations. The conclusions of this work are presented in Section 5.
2. Problem Formulation
Based on the kinematic model and the IBVS model of the mobile robot, the IBVS-PF-HMPC is designed by considering the position deviation and the image feature point deviation for automatic parking trajectory tracking of the mobile robot. The designed controller still operates based on the position deviation to guarantee the safety and accuracy of the mobile robot system, when the feature points are intermittently lost due to a disturbance at a certain time. For example, the mobile robot will brake urgently when the feature points are obstructed in dynamic environments.
2.1. Model of Mobile Robots
Due to the low-speed characteristics of automated parking, the kinematic characteristics of the mobile robot are much greater than the kinetic characteristics. The mobile robot satisfies the Ackermann steering principle, and the trajectory of its rear axis center point can characterize the lateral motion characteristics, and the front axis as the steering axis can reflect the lateral motion characteristics. The motion model in the world coordinate system is shown in Figure 1, by using the center coordinates of the rear axis and the equivalent deflection angle of the front axis as the objects.
Figure 1.
Motion model of the mobile robot.
In Figure 1, and represent the coordinates of the rear and front axle axes of the mobile robot, represents the transverse sway angle, represents the front wheel deflection angle, represents the rear axle center velocity, represents the front axle center velocity, represents the axle distance, represents the rear wheel steering radius, and represents the instantaneous center of vehicle rotation, assuming that the lateral eccentricity of the center of mass of the mobile robot remains constant during the steering process, i.e., the instantaneous steering radius is the same as the radius of curvature of the road, and the mobile robot is a rigid body.
At the rear axle center , the rear axle center velocity is
The kinematic constraints of the front and rear axles are
Combining (1) and (2) we have
According to the geometric relationship between the front and rear wheels, we have
Then, the angular velocity can be solved as
where is the angular velocity. Then, we can have rear wheel steering radius and front wheel deflection angle :
Then, the kinematic equation of the mobile robot in the geodesic coordinate system is given as
where and are the velocities in the axis and axis, and is the acceleration of the transverse pendulum angle. Due to the large calculation of , the angular velocity is introduced, and is the control quantity.
The kinematic model of the mobile robot can be transformed into the following equation:
where is the position of the mobile robot in the world coordinate system, is the linear and angular velocities.
2.2. Model of IBVS System
It is necessary to establish suitable expressions to change the coordinates of a three-dimensional (3D) space into a two-dimensional (2D) plane. The relationship between image change and mobile robot motion of the calibration-free servo system can be modeled directly from the known image information, i.e., the image Jacobi matrix can be obtained to establish the transformation between the 3D space and 2D plane. This method could reduce the amount of coordinate conversion and system calculations, and does not need to calculate the internal and external parameters of the camera. Thus, it can avoid the impact of errors generated during the calibration process. It is defined as
where is the image feature point coordinates, is the velocity of the image feature point change, is the mobile robot position, is the velocity of the mobile robot position change, and is the image Jacobi matrix.
Figure 2 is the camera imaging principle. is the world coordinate system. is the camera coordinate system. is the image coordinate system.
Figure 2.
The camera imaging principle.
In Figure 2, the coordinate point of the camera coordinate system is , and of the plane image coordinate system is . According to the camera imaging principle, we have
where is the camera focal length, is the image plane origin, and are the physical sizes of a unit pixel on the horizontal and vertical axes, respectively. Taking the derivative of (10),
Let be the linear velocity of the mobile robot, and be the angular velocity of the mobile robot. The velocity of a point in space with respect to the camera coordinate system is
According to (12),
Then we have
where the velocity of the image feature point change is , the velocity of the mobile robot position change is , and the image Jacobi matrix is
Combining (14) with the kinematic model of the mobile robot, the velocity of the restricted mobile robot’s position change is
The relationship between the change speed of 2D image feature points and the mobile robot control quantity can be obtained:
where is the velocity of the 2D image feature point change, and are the linear and angular velocity control quantities of the mobile robot, respectively.
3. Controller Design
Figure 3 is the control block diagram of the IBVS-PF-HMPC system. are the desired positions of the mobile robot, are the predicted positions correction of the mobile robot, are the desired coordinate inputs of the image feature points, are the predicted correction outputs of the image coordinate points, is the state deviation input, is the predicted deviation, is the actual value output, is the model prediction output, and are the control quantities.
Figure 3.
The control block diagram of IBVS-PF-HMPC system.
3.1. Design of the Hybrid Model Predictive Control Based on IBVS
Model predictive control (MPC) is an iterative optimization technique. The optimal control volume prediction sequence is obtained by substituting the current state measured or estimated at each sampling time into the optimization function. And the first of these control quantities is taken as the control input at the current moment.
The general form of the nonlinear system can be abstracted as
where is the state, is the rate of the state change, is the control input.
Let the current moment be , and the control period be . (18) is discretized by Euler’s method:
in which,
where and represent the prediction range and the control range . The state quantities in the prediction time domain are
Define the state quantity deviation as , in which, is the current state, is the desired state. Then, the optimization function of MPC is defined as follows:
where represents the state deviation of , which is predicted at the time of , represents the input deviation of , which is predicted at the time of, and and are the weighting matrices of the deviation of the state quantity and the input quantity, respectively.
Rewrite both the state and input quantities as increments:
Combining (8) and (17), define the mobile robot’s position as , the image feature point coordinate as , and the control quantity of the mobile robot as . The deviation in (22) can be divided into two parts: the deviation of the position and the deviation of the feature points.
Then, the hybrid model predictive control based on IBVS (IBVS-HMPC) can be obtained:
where are the minimum and maximum values of control quantity and control increment, respectively.
3.2. Design of the IBVS-PF-HMPC
The emergency braking performance of the mobile robot needs to be guaranteed when the image feature points are massively obscured by obstacles in dynamic scenes. So it is necessary to introduce the penalty factor into the above IBVS-HMPC.
The number of feature points in the image is , and the feature points . The number of obstacle-obscuring feature points is
Introducing the penalty factor , the penalty term of the optimization function with respect to the loss of feature points is obtained from . Then, (25) can be rewritten as
The sequence of control increment predictions for the current moment can be obtained:
The first element is taken as the actual control increment acting on the controlled system at the current moment:
The predicted state quantity is obtained by modifying the actual state quantity and transmitted to the next control cycle for optimization and a solution. Repeat the above process to complete rolling optimization within the control time domain and the automatic parking trajectory tracking control for mobile robots is realized.
When the feature points are massively lost, is large, and the optimization function (28) has no solution. Thus, the control input is 0, and the emergency braking of the mobile robot in dynamic obstacle scenarios could be realized.
The steps of the proposed control methodology of IBVS-PF-HMPC are shown in Table 1.
Table 1.
The steps of the proposed control methodology of IBVS-PF-HMPC.
4. Simulation Results
In order to verify the effectiveness of the proposed control method, four groups of comparison experiments based on MATLAB are conducted. According to the motion characteristics of the mobile robot, the system limitations during simulation tests are as follows: the line velocity limitation of the mobile robot is ; the angular velocity limitation of the mobile robot is ; the linear velocity increment limitation of the mobile robot is ; and the angular velocity increment limitation of the mobile robot is . The physical parameters of the mobile robot and the parameters of the IBVS-PF-HMPC when it reaches steady-state are given in Table 2. The control quantity predictive sequence is used as an optimization variable in the optimal control process of the automatic parking for mobile robots. The optimization criterion is minimizing the respective sums of the position deviation increment, the image feature point deviation increment and the control quantity predictive sequence. Then, the controller could take into account the state quantity output and control quantity input to ensure the good trajectory tracking control of the constrained mobile robot in the automatic parking process.
Table 2.
The parameters of the mobile robot and IBVS-PF-HMPC.
4.1. Parking Trajectory Planning
Both the arc-tangent function curve and the higher polynomial curve fit better with the automatic parking trajectory. Considering that the higher polynomial with more than four times is more complicated in the calculation, therefore, the arc-tangent function is selected design the reference trajectory in Figure 4, and the defined parallel parking trajectory model is as follows:
where .
Figure 4.
The reference trajectory for mobile robot automatic parking.
4.2. Stability Performance
In the first simulation scenario, the proposed hybrid model predictive control with a penalty factor based on IBVS (IBVS-PF-HMPC) is compared with the traditional IBVS control (T-IBVSC), to verify the tracking performance under the image feature point loss scenario. Figure 5 and Figure 6 exhibit the trajectory tracking curves of the mobile robot under the control of T-IBVSC and IBVS-PF-HMPC, respectively.
Figure 5.
The tracking curves of the mobile robot under the T-IBVSC.
Figure 6.
The tracking curves of the mobile robot under the IBVS-PF-HMPC.
As shown in Figure 5 and Figure 6, T-IBVSC and IBVS-PF-HMPC could maintain good trajectory tracking control until 2.5 m in the X axis. But the T-IBVSC fails to complete trajectory tracking control at 2.5 m due to the sudden increase in the control quantity , which causes the image feature points to be lost in the field of view and is out of control in the following time.
4.3. Tracking Accuracy Performance
In the second simulation scenario, the proposed IBVS-PF-HMPC is compared with the model predictive control based on IBVS (IBVS-MPC), to verify the tracking performance under the image feature point loss scenario. Although the IBVS-MPC has a constraint limit on the image feature points, it only considers the image feature deviation without considering the position deviation of the mobile robot at the same time. Then, the image feature points are lost intermittently by adding interference. Figure 7 is the trajectory tracking curves of the mobile robot under the control of IBVS-MPC and IBVS-PF-HMPC. Figure 8 and Figure 9 exhibit the control quantity and the position deviation of the mobile robot under the control of IBVS-MPC and IBVS-PF-HMPC, respectively. And the root means square errors (RMSE) of the deviation of each component of the mobile robot position under the control of IBVS-MPC and IBVS-PF-HMPC are shown in Table 3.
Figure 7.
The tracking curves of the mobile robot under the IBVS-MPC and IBVS-PF-HMPC.
Figure 8.
The control quantity under the control of IBVS-MPC and IBVS-PF-HMPC.
Figure 9.
The position deviation under the control of IBVS-MPC and IBVS-PF-HMPC.
Table 3.
The RMSE of the deviation of each component of the mobile robot position under the control of IBVS-MPC and IBVS-PF-HMPC.
As shown in Figure 7, the maximum tracking error of IBVS-PF-HMPC is m, and the maximum tracking error of IBVS-MPC is m. The proposed IBVS-PF-HMPC could achieve higher trajectory tracking accuracy than the IBVS-MPC. This is due to the fact that the proposed controller considers both image feature point deviation and mobile robot position deviation. When the image feature points are intermittently lost due to external perturbations, the position deviation still acts on the mobile robot controller.
As shown in Figure 8 and Figure 9, the average values of each position deviation of the mobile robot are m, m and rad under the control of IBVS-PF-HMPC, and the average values of each position deviation of the mobile robot are m, m and rad under the control of IBVS-MPC. It shows that the position tracking accuracy of IBVS-PF-HMPC is 35.80%, 64.82% and 31.75% higher than that of IBVS-MPC, respectively. This is due to the loss of image feature points, as the control quantity of IBVS-MPC jittered more significantly and reached the limit constraint several times.
As shown in Table 3, the RMSE of the deviation of each component of the mobile robot position under the IBVS-PF-HMPC is smaller than the RMSE of IBVS-MPC. Therefore, the proposed control algorithm owns better trajectory tracking performance in the case of image feature point loss.
4.4. Real-Time Performance
In the third simulation scenario, the proposed IBVS-PF-HMPC is compared with the IBVS-PF-HMPC without increment (NI-IBVS-PF-HMPC), to verify the real-time and tracking accuracy of the mobile robot. Figure 10 is the trajectory tracking curves of the mobile robot under the control of NI-IBVS-PF-HMPC and IBVS-PF-HMPC. Figure 11 and Figure 12 exhibit the control quantity and the position deviation of the mobile robot under the control of NI-IBVS-PF-HMPC and IBVS-PF-HMPC, respectively. And the RMSE of the deviation of each component of the mobile robot position under the control of NI-IBVS-PF-HMPC and IBVS-PF-HMPC are shown in Table 4.
Figure 10.
The tracking curves of the mobile robot under the NI-IBVS-PF-HMPC and IBVS-PF-HMPC.
Figure 11.
The control quantity under the control of NI-IBVS-PF-HMPC and IBVS-PF-HMPC.
Figure 12.
The position deviation under the control of NI-IBVS-PF-HMPC and IBVS-PF-HMPC.
Table 4.
The RMSE of the deviation of each component of the mobile robot position under the control of NI-IBVS-PF-HMPC and IBVS-PF-HMPC.
As shown in Figure 10, the maximum tracking error of IBVS-PF-HMPC is m, and the maximum tracking error of NI-IBVS-PF-HMPC is m. The simulation results show that the proposed control method owns better tracking accuracy.
As shown in Figure 11 and Figure 12, the average values of each position deviation of the mobile robot are m, m and rad under the control of BVS-PF-HMPC, and the average values of each position deviation of the mobile robot are m, m and rad under the control of IBVS-PF-HMPC. It shows that the position tracking accuracy of IBVS-PF-HMPC is 64.26%, 73.86% and 75.06% higher than that of IBVS-MPC, respectively. Meanwhile, it can be seen that the angular velocity of the mobile robot has a significant lag, which leads to poor real-time when the state and control quantities are not reconfigured. And the position deviation of NI-IBVS-PF-HMPC is significantly larger than the position deviation of IBVS-PF-HMPC.
As shown in Table 4, the RMSE of the deviation of each component of the mobile robot position under the IBVS-PF-HMPC is smaller than the RMSE of NI-IBVS-PF-HMPC. Therefore, the proposed control algorithm owns better real-time performance and tracking accuracy.
4.5. Emergency Braking Performance
In the fourth simulation scenario, the emergency braking performance of the proposed IBVS-PF-HMPC is verified, by adding more than 10 image feature points that are obscured by dynamic obstacles during the automatic parking of the mobile robot. Figure 13 is the trajectory tracking curve of the mobile robot when the image feature points are massively occluded by obstacles. Figure 14 and Figure 15 are the control quantity and the position deviation when the mobile robot is obscured by the dynamic obstacle, respectively.
Figure 13.
The trajectory tracking curves of the mobile robot facing the massive obstacles.
Figure 14.
The control quantity facing the massive obstacles.
Figure 15.
The position deviation facing the massive obstacles.
As shown in Figure 13, when the image feature points are massively occluded by obstacles, the mobile robot control system no longer tracks the desired trajectory. It can be seen from Figure 14 and Figure 15 that the control quantities tend to 0 when . Consequently, the emergency braking of the mobile robot in the dynamic obstacle scene could be realized in time.
5. Conclusions
For the constrained mobile robot automatic parking system, the IBVS-PF-HMPC is proposed, to address the problem of feature point loss and emergency braking in dynamic obstacle scenarios caused by excessive target bias gain when using traditional IBVS control methods. The main conclusions are as follows:
- (1)
- The IBVS-based incremental model predictive control algorithm is designed. The traditional IBVS control is transformed into an optimization problem with constraints in the finite time domain, by defining the optimization function based on the mobile robot’s positional deviation and image feature point deviation, while using the actuator saturation and speed limit as constraints. Then, the accuracy and real-time of the mobile robot tracking control during automatic parking is improved simultaneously.
- (2)
- The convex optimization function with penalty factor is defined. Then, the IBVS-PF-HMPC is proposed, to guarantee the emergency braking performance of the mobile robot automatic parking when the image feature points are massively obstructed by obstacles in dynamic scenes.
- (3)
- Several simulation comparisons further verify the correctness and effectiveness of the proposed IBVS-PF-HMPC.
Author Contributions
Conceptualization, H.G., Q.Q. and J.M.; software, Q.Q. and H.G.; validation, J.M. and H.G.; formal analysis, H.G. and J.M.; investigation, J.M. and Y.H.; data curation, H.G. and Q.Q.; writing—original draft preparation, H.G.; writing—review and editing, H.G. and X.S.; supervision, Q.Q.; project administration, J.M.; funding acquisition, Q.Q. and J.M. All authors have read and agreed to the published version of the manuscript.
Funding
This work is supported in part by the National Natural Science Foundation of China under Grant No. 52177051, the Natural Science Research Program of Jiangsu Colleges and Universities under Grant No. 20KJA470002, the Excellent Teaching Team of “Qinglan Project” of Jiangsu Colleges and Universities and the Talent Introduction Startup Fund of Nantong University under Grant No. 135437612076.
Data Availability Statement
The data used to support the findings of this study are available from the corresponding author upon request.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Yu, Y. Smart parking system based on edge-cloud-dew computing architecture. Electronics 2023, 12, 2801. [Google Scholar] [CrossRef]
- Chen, W.; Xu, T.; Liu, J.; Wang, M.; Zhao, D. Picking robot visual servo control based on modified fuzzy neural network sliding mode algorithms. Electronics 2019, 8, 605. [Google Scholar] [CrossRef]
- He, S.; Xu, Y.; Guan, Y.; Li, D.; Xi, Y. Synthetic robust model predictive control with input mapping for constrained visual servoing. IEEE Trans. Ind. Electron. 2023, 70, 9270–9280. [Google Scholar] [CrossRef]
- Xu, D. A tutorial for monocular visual servoing. Acta Automatica Sin. 2018, 44, 1729–1746. [Google Scholar]
- Zhong, H.; Wang, Y.; Miao, Z.; Li, L.; Fan, S.; Zhang, H. A homography-based visual servo control approach for an underactuated unmanned aerial vehicle in GPS-denied environments. IEEE T. Intell. Veh. 2023, 8, 1119–1129. [Google Scholar] [CrossRef]
- Chan, W.; Srigrarom, S. Image-based visual-servoing for air-to-air drone tracking & following with model predictive control. In Proceedings of the 2023 SICE International Symposium on Control Systems (SICE ISCS), Kusatsu, Japan, 9–11 March 2023. [Google Scholar]
- Aaron, M.; Marwen, J.; Peter, C. Image-based visual servoing with unknown point feature correspondence. IEEE Robot. Autom. Lett. 2017, 2, 601–607. [Google Scholar]
- Becerra, H.; Lopez-Nicolas, G.; Saguees, C. A sliding-mode-control law for mobile robots based on epipolar visual servoing from three views. IEEE Trans. Robot. 2011, 27, 175–183. [Google Scholar] [CrossRef]
- Wang, F.; Qin, Y.; Guo, F.; Ren, B.; John, T. Adaptive visually servoed tracking control for wheeled mobile robot with uncertain model parameters in complex environment. Complexity 2020, 3, 8836468. [Google Scholar] [CrossRef]
- Sunhyo, K.; Se-Young, O. Hybrid position and image based visual servoing for mobile robots. J. Intell. Fuzzy Syst. 2007, 18, 73–82. [Google Scholar]
- Li, B.; Fang, Y.; Hu, G.; Zhang, X. Model-free unified tracking and regulation visual servoing of wheeled mobile robots. IEEE Trans. Control Syst. Technol. 2016, 24, 1328–1339. [Google Scholar] [CrossRef]
- Fang, Y.; Liu, X.; Zhang, X. Adaptive active visual servoing of nonholonomic mobile robots. IEEE Trans. Ind. Electron. 2012, 59, 486–497. [Google Scholar] [CrossRef]
- Zhang, K.; Chen, J.; Li, Y.; Gao, Y. Unified visual servoing tracking and regulation of wheeled mobile robots with an uncalibrated camera. IEEE-ASME Trans. Mechatron. 2018, 23, 1728–1739. [Google Scholar] [CrossRef]
- Yuan, W.; Liu, Y.; Su, C.; Zhao, F. Whole-body control of an autonomous mobile manipulator using model predictive control and adaptive fuzzy technique. IEEE Trans. Fuzzy Syst. 2023, 31, 799–809. [Google Scholar] [CrossRef]
- Kang, E.; Qiao, H.; Chen, Z.; Gao, J. Tracking of uncertain robotic manipulators using event-triggered model predictive control with learning terminal cost. IEEE Trans. Autom. Sci. Eng. 2022, 19, 2801–2815. [Google Scholar] [CrossRef]
- Heshmati-alamdari, S.; Karavas, G.; Eqtami, A.; Drossakis, M.; Kyriakopoulos, K. Robustness analysis of model predictive control for constrained image-based visual servoing. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014. [Google Scholar]
- Bai, H.; Gao, J.; Sun, X.; Yan, W. Model predictive visual trajectory-tracking control of wheeled mobile robots. In Proceedings of the 2019 IEEE 28th International Symposium on Industrial Electronics (ISIE), Vancouver, BC, Canada, 12–14 June 2019. [Google Scholar]
- Heshmati-alamdari, S.; Karras, G.; Eqtami, A.; Kyriakopoulos, K. A robust self-triggered image based visual servoing model predictive control scheme for small autonomous robots. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015. [Google Scholar]
- Li, Z.; Yang, C.; Su, C.; Deng, J.; Zhang, W. Vision-based model predictive control for steering of a nonholonomic mobile robot. IEEE Trans. Control Syst. Technol. 2016, 24, 553–564. [Google Scholar] [CrossRef]
- Chi, X.; Liu, Z.; Huang, J.; Hong, F.; Su, H. Optimization-based motion planning for autonomous parking considering dynamic obstacle: A hierarchical framework. In Proceedings of the 2022 34th Chinese Control and Decision Conference (CCDC), Hefei, China, 15–17 August 2022. [Google Scholar]
- Yu, R.; Guo, H.; Chen, H. Predictive obstacle-avoidance control for autonomous vehicle information and control. Inf. Contrl. 2015, 44, 117–124. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).