Next Article in Journal
Minimal Gluten Exposure Alters Urinary Volatile Organic Compounds in Stable Coeliac Disease
Previous Article in Journal
Crossing Frequency Method Applicable to Intermediate Pressure Plasma Diagnostics Using the Cutoff Probe
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robotic Manipulation Planning for Automatic Peeling of Glass Substrate Based on Online Learning Model Predictive Path Integral

1
College of Mechanical and Electrical Engineering, Central South University, Changsha 410083, China
2
State Key Laboratory for High Performance Complex Manufacturing, Central South University, Changsha 410083, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(3), 1292; https://doi.org/10.3390/s22031292
Submission received: 8 January 2022 / Revised: 1 February 2022 / Accepted: 5 February 2022 / Published: 8 February 2022
(This article belongs to the Topic Robotics and Automation in Smart Manufacturing Systems)

Abstract

:
Autonomous planning robotic contact-rich manipulation has long been a challenging problem. Automatic peeling of glass substrates of LCD flat panel displays is a typical contact-rich manipulation task, which requires extremely high safe handling through the manipulation process. To this end of peeling glass substrates automatically, the system model is established from data and is used for the online planning of the robot motion in this paper. A simulation environment is designed to pretrain the process model with deep learning-based neural network structure to avoid expensive and time-consuming collection of real-time data. Then, an online learning algorithm is introduced to tune the pretrained model according to the real-time data from the peeling process experiments to cover the uncertainties of the real process. Finally, an Online Learning Model Predictive Path Integral (OL-MPPI) algorithm is proposed for the optimal trajectory planning of the robot. The performance of our algorithm was validated through glass substrate peeling tasks of experiments.

1. Introduction

The glass substrate is the base material of LCD flat panel displays, and its thinning process is necessary for the end product being lighter, thinner and smoother [1]. At present, the thinning process mainly adopts polishing machines with chemical mechanical processing. The workpiece of glass substrate, say 0.73 m in width, 0.92 m in length and 2 mm in thickness, needs to be fixed on the surface of a rotating rigid workbench plate, on which an adsorption pad is attached. The adsorption pad (a polyurethane foam pad about 1 mm thick) is generally used for the fixing task, and the glass substrate is absorbed into the pad with the vacuum suction force in between after pressing the glass substrate evenly on the pad. Separating the glass substrate from the pad is needed after the thinning process, and the process of unloading the glass substrate is usually done by human operators. They use their fingernails to “peel” the glass substrate off the pad from a corner of the substrate (Figure 1a). This manual process is labor-consuming, apt to glass damage, and therefore also expensive. Automating the process of unloading the glass substrate has long been expected, and we presented a scheme (Figure 1b) in [2]. It should be pointed out that the original intention of introducing Figure 1b here is to better show the task we need to accomplish. The actual experiment platform used can be seen in Figure 7, which is different from Figure 1b.
We invented a wedge blade (Figure 1b) as the end-effector of a robotic arm to imitate the “peeling” motion of human operators, but the action and reaction of human operators in the “peeling” process is still a challenge to imitate by robots, and this paper reports our work of the effort to make the robotic arm learn adaptively in the process to the end of replacing the human operator in the unloading process of glass substrate.
Our challenge is a kind of contact-rich manipulation in unstructured environments, which is one of the biggest challenges in robotics applications [3]. The external forces that the robotic arm faces during the peeling process is affected by the dynamics of the multi-interface contact, which shows a high degree of nonlinearity. The salient feature of the contact-rich manipulation task is that the robot or the tool at the end of the robot will remain in contact with the environment during the process. Such tasks are very common, such as assembly tasks (such as peg-in-hole task [4]) and material removal tasks (such as grinding [5], digging [6]) in the production environment, as well as glass cleaning [7], massage [8] and other tasks in life. There are mainly two kinds of ways for contact-rich manipulation tasks, namely, planning-based and learning-based. Learning-based methods such as reinforcement learning and imitation learning have been successfully applied in contact-rich tasks in recent years [9,10,11]. This type of method describes the manipulation tasks as a regression problem of machine learning, and establishes the mapping relationship from perceptual information to motion control signals. The advantage of learning-based methods is that they do not rely on prior knowledge, so as to improve the manipulation ability of robot in cross-task and cross-scene situations. However, the applications of learning-based methods have been almost accomplished in laboratory environments, and the robustness of the algorithms is still not enough for actual industrial application scenarios. This paper turns to planning-based method, which is another way of representing contact-rich manipulation tasks. The planning-based methods consider the manipulation planning as a motion planning problem with force constraints. The realization of planning-based methods includes two main stages: forward model establishment and planning.
There are many ways to derive the forward model, which can be divided into two types: mechanism model and data-driven model. The mechanism model relies on a large amount of prior information, which has a strong interpretability for the behavior of the robot [12]. The lack of generalization of the mechanism model makes it only suitable for specific task. For the glass substrate peeling process, it is almost impossible to establish an accurate mechanism model of the system, although we did make some effort on the process modeling. The pure data-driven method is to embed the physical system into the parameterized model without any prior information [13], but it needs a large number of training samples to make the model converge. As a result of this, pure data-driven modeling methods are not suitable for scenarios where sample collection costs are high, such as glass substrate peeling. Combining the mechanism model and the data-driven model is expected to bring out the advantages of both. The most classical method is parameter identification [14], but it still requires high modeling accuracy of mechanism models. Another classical method is to use a machine learning model to compensate the mechanism model [15]. When the mechanism model is inaccurate, this type of method requires the collection of sufficient experimental data to correct the mechanism model, which is also not suitable for the application scenarios of this paper. Both of the above two methods adopt the idea of mechanism modeling as the main part and data modeling as the supplement. Reference [16], on the other hand, takes the data modeling as the main part and mechanism modeling as the supplement. This modeling idea integrates the mechanism modeling as a priori information into the training process of the data-driven model, which significantly improves the training efficiency of the data-driven model. Reference [17] introduces the above modeling ideas into dynamic system modeling. The constraint of the methods in [16,17] is its only applicability to Newtonian rigid body systems, which is not the case of this paper that includes elastic objects such as glass substrate and adsorption pad.
Inspired by reference [17], this paper embeds the evolution of the physical system dynamics of peeling process of the glass substrate into the parameter model implicitly, and proposes a practical model structure and model training method. The specific description of the method will be introduced in Section 3.
When we complete the establishment of a forward model, it is needed to plan a control sequence that meets the task requirements according to the process. The planning of the control sequence for robot manipulations usually uses Linear Quadratic Regulator (LQR) [18,19] and Model Predictive Control (MPC) [20,21]. However, these methods require the differentiable overall loss function (or cost function) respect to the control variable, which makes they are not suitable for our cost function with contact force-related constraints, or to suit these methods we can only use force thresholds for safe operation which is a risky way to limit control inputs for safety.
Fortunately, path integral framework [22,23] allows more flexible design of cost function, and supports the planning directly with the nonlinear system forward model, rather than after linearization of the nonlinear model like LQR. The Model Predictive Path Integral (MPPI) planner proposed in reference [23] provides a suitable start for our manipulation planning of glass substrate peeling robot, but its offline nature [24] of system modeling cannot cover the uncertainty nature of our system. The contact dynamics of the glass substrate peeling process is greatly affected by the adsorption pad, the adhesive condition, and other working conditions, and the model from offline design may deviate greatly from the real process. This paper innovatively combines the meta-learning with MPPI to realize the online update of the forward model, thereby making the planner easily accommodate the uncertainty of the real process of the glass substrate peeling task while avoiding catastrophic forgetting of the model [25].
The main contributions of this paper are: (1) proposed a forward model modeling method that combines deep learning models and prior physical knowledge together; (2) proposed an algorithm based on Online Learning Model Predictive Path Integral (OL-MPPI), which is suitable for manipulation planning in tasks with large dynamical uncertainties.
The paper is organized as follows. Section 2 is a mathematical description of the research problem. Section 3 introduces the main algorithms. Section 4 is the experimental verification results of the algorithms. Finally, the conclusions are given in Section 5.

2. Problem Formulation

The automatic process of the robotic arm in Figure 1b can be outlined in Figure 2: (1) From the initial point A to point B is an auxiliary motion, and at B, the wedge blade comes to the surface of the adsorption pad with a certain downward pressing force on the pad; (2) From B to C, the wedge blade keeps the downward pressing while moving horizontally until point C where it touches the edge of the glass substrate; (3) From C to D is the peeling process we are concerned about in this paper, after that other process and mechanical devices will be involved to separate the entire glass substrate from the adsorption pad (that is, to complete the workpiece unloading), which is beyond the scope of this paper.
This paper mainly considers the motion planning problem of the blade driven by the robotic arm from point C to point D. During this process, the wedge blade goes through the adhesive surface between the pad and the glass substrate to release the vacuum suction force. Our attention is mainly on the movement of the blade rather than the joints of the robotic arm, and the following movement descriptions will be in the task space.
Consider the following discrete system:
x t + 1 = f θ ( x t , v t )
where v t is the wedge blade velocity at discrete time t, x t is the system state vector at time t which is defined in Equation (2), f parameterized by θ is the map of system state transition from time t to the next time point t + 1 . Equation (1) means that with the control input v t the system state transits from x t to x t + 1 according to the model defined by f θ ( x t , v t ) with the set of model parameter θ .
Generally, the control variable v t R m . In our special case we only consider the horizontal movement of the blade from C to D, so m = 1. Here we model the peeling process as a discrete system, and actually this should be a top-level system model corresponding to kinematic control level, We will not involve bottom level of dynamical control in this paper, which should be our further experimental study about this project. We simply put v t ~ N ( u t , Σ ) to model the motion with acceleration and deceleration process of the real robot, where u t is the target velocity of the movement at control cycle t , and v t is the observation of u t ,with normal distribution, and Σ is the variance.
x t is defined as a triplet of position q , velocity v , and forces F R :
x t = q t h : t , v t h : t , F t h : t R R n
where, q t h : t , v t h : t and F t h : t R respectively represent the time series of position, velocity and contact force observations from time (th) to time t. h determines how much historical information is used to create the system model. The contact force of the wedge blade in the peeling process, F t R , is nonlinear and non-stationary in which only horizontal and vertical components F t R X and F t R Z are counted here.
A movement from point C to point D in Figure 2 is expressed as a trajectory τ , that is, τ = ( x 0 , u 0 , x 1 , u 1 , , x t , u t , x T ) , which is a time sequence of states and control inputs. where, x 0 and x T represent the initial state and the terminal state, and the sampling period of time between successive subscripts is Δ t . In order to evaluate whether a trajectory τ i meets the requirements of the glass substrate peeling task, we define the following cost function J τ i :
J ( τ i ) = ϕ x t T , t T + t 0 t T L t d t
where, t 0 is the initial time; t T is the terminal time; the cost of the trajectory includes two parts: cost obtained in the terminal state ϕ , and the cost integrated through all the other steps; L t is called the immediate cost at time t, which is defined in Equation (4).
L t = w d i s q t q t 1 q t e r q 0 + w f X F R X F lim R X + w f Z F R Z F lim R Z
where, q t e r and q 0 represent the end position and start position respectively; F lim R X and F lim R Z represent the maximum limit of horizontal and vertical force. It can be seen that the immediate cost function is composed of three items. The first one intends to encourage the robot to complete the task faster. The second and third items are used to punish the actions that cause excessive contact force and to reduce the wear or damage of the glass and the adsorption pad. Different task requirements can be achieved by adjusting coefficients w d i s , w f X and w f Z .
When the robot does not reach the end position q t e r within an upper bound time t max , the robot will get an additional terminal cost ϕ :
ϕ = 0 , i f   ( t < t max a n d   q t = q t e r ) ϕ = 1 , i f   ( t > t max   a n d   q t = q t e r ) ϕ = 2 , i f   ( t > t max   a n d   q t < q t e r )
According to the above definition, the manipulation planning problem of the glass substrate peeling robot can be regarded as a stochastic optimal control problem. The ultimate goal is to find the optimal control input sequence [ u 0 * , u 1 * , u T * ] to minimize the value function:
V ( x 0 ) = min u 0 : u T E τ i J τ i
where, E τ i is the expectation of the cost function of all trajectories whose initial state is x 0 .

3. Method

Figure 3 shows our algorithm framework to solve the robot manipulation planning problem of automated glass substrate peeling formulated in Equation (6).
The two boxes of left-most in Figure 3 gives pretrained model, which is a deep learning-based neural network. In Section 3.1, we will introduce the design of deep learning model suitable for the glass substrate peeling task; simulation environment will be introduced (in Section 3.2) to collect samples for learning to avoid expensive and time-consuming collection of real data; the algorithm of offline learning of the deep neural network will be introduced in Section 3.3.
The loop in Figure 3 includes: (1) online learning (in Section 3.4) to tune the pretrained model according to the real data from the peeling process experiments to cover the uncertainties of the real process. (2) optimal trajectory planning based on the MPPI algorithm with the feature of online learning, which will be discussed in Section 3.5.

3.1. Model Design

We use deep neural network to model the discrete mapping function f θ ( ) in Equation (1), and sets the prediction range as one sampling period, that is, single-step prediction. Considering the observation range h > 1 (in Equation (2)), the model of f θ ( ) should capture the nature of time sequence. The Recurrent Neural Network (RNN) and its variant Long Short-Term Memory (LSTM) [26] should be a reasonable choice for the model structure. The difficulties of the model training for our case comes from, (1) the noises of the input forces in Equation (2) which are from the real force sensor readings mounted on the end of the robotic arm, and (2) the uncertainties of the dynamic parameters of the peeling process.
The structure of the model proposed here in this paper can be depicted as in Figure 5, and we take several tricks to tackle the difficulties mentioned above and to improve the robustness and representation ability of the model. These tricks include the local feature extraction, the attention mechanism, and the time series feature extraction, and we connect these modules in series to obtain the final system model.

3.1.1. Local Feature Extraction Based on Multi-Scale Convolution Kernel

The motion of wedge blade during the peeling process of the glass substrate takes the fashion of “stick-slip” motion. The proceeding forward of the wedge blade between the adsorption pad and the glass substrate needs the robot to drive the blade harder enough to overcome the resistance ahead. This resistance fluctuates very sharply, causing continuous stick-slip phenomenon in the movement of the blade. “Sticky” behavior is often concentrated on a small-time scale, while “slip” behavior is distributed in a large time scale. This feature can be extracted from time series data according to different time scales. Based on this, we designed a multi-scale 1D convolutional neural network (1D-CNN) to extract local features from time series data.
The convolutional layer uses the convolution kernel to traverse the input sequence at a certain step length. At each traversed position, use the following matrix convolution formula,
y ( i , j ) = w c x ( i , j ) = m n x ( i + m , j + n ) w c ( m , n )
where, wc is the convolution kernel, and x ( i , j ) is the input series data whose starting point is (i,j).
For our application, we use three convolution kernels of different sizes, say 3 × 2, 3 × 4 and 3 × 6 respectively. The convolution kernels traverse the input sequence with stride = 1, and the paddings of the three convolution kernels are 0, 1, and 2, respectively to make the output sequence lengths consistent. The schematic diagram of extracting the local features of the input sequence with the multi-scale convolution kernel is shown in Figure 4.

3.1.2. Attention Mechanism

Attention mechanism is used on the output sequences S 1 , S 2 , and S 3 in Figure 4 after the local feature extraction module. The attention mechanism is widely used in the fields of image and natural language processing. It can automatically assign different weights to the input vectors, so that the more valuable information for the prediction task can be fully used. The motivation for designing the attention mechanism is that when the blade is in different motion states, the importance of motion information and force information for interpreting contact dynamics should be automatically weighted. Intuitively, small changes in position or velocity should be weighted more than force information to infer the occurrence of the state switch from “stick” to “slide”, while from “slide” to “stick” the force information should count more. Considering the different importance of these three time series for force prediction, we design the following attention mechanism using MLP (a fully connected neural network) and SoftMax activation function:
o t = M L P ( [ S 1 , S 2 , S 3 ] )
[ w 1 , w 2 , w 3 ] = s o f t max ( o t )
The output of the attention mechanism module can be calculated as follows:
[ S 1 , S 2 , S 3 ] = [ w 1 S 1 , w 2 S 2 , w 3 S 3 ]

3.1.3. Time Series Feature Extraction Based on LSTM

The temporal dependence of physical variable is crucial for modeling contact dynamics, and the changing trends of different physical variables over time imply the next motion of the blade. We use LSTM neurons for this purpose.
The LSTM neuron contains a memory cell state and three gates: forget gate, input gate and output gate. In the time step of t, the formulas are as follows:
f t = σ ( W f [ h t 1 , x t ] + b f )
i t = σ ( W i [ h t 1 , x t ] + b i )
g t = tanh ( W g [ h t 1 , x t ] + b g )
C t = f t C t 1 + i t g t
o t = s o f t max ( W o [ h t 1 , x t ] + b o )
h t = o t tanh C t
where, f , i, g, C, and o are the forget gate, input gate, candidate update cell state, updated cell state, and output gate. tanh is the hyperbolic tangent activation function.
After concatenating the obtained feature vector with the velocity at the future time, then connect with the four-layer fully connected layer, so that we can get a complete system model based on deep learning. Here, the function of the fully connected layer is to realize the regression task. The number of neurons in each fully connected layer is 128, 64, 32, and 16, respectively, and the activation function uses the ReLu function. The overall structure of the model is shown in Figure 5.

3.2. Simulation Environment

Simulation can facilitate the process of our model learning. Based on the dynamics we studied previously [27], a simulation environment is established using Pybullet [28] as the physics engine to simulate the peeling process of the glass substrate.
There are uncertainties in the dynamic parameters of the glass substrate, the adsorption pad and the readings of force sensor, which have great influence on the accuracy of the simulation results. In order to make the simulation data more realistically reflect the real contact behavior, this paper adopts the following two measures:

3.2.1. Dynamic Parameters Randomization

In the actual peeling process of the glass substrate, even if the same batch of adsorption pads are used, the contact behaviors between the blade and the adsorption pad may still be different due to the slight variations in the dynamic parameters. To improve the robustness of the offline training system model, this paper uses the method of dynamic parameters randomization to collect training samples containing different contact behaviors in the simulation environment. In this way, real contact dynamics information can be included in a part of the samples. Pk = {p1, p2, …, pN}k∈RN represents a set of dynamic parameters, where N is the number of variable dynamic parameters. Different parameters combinations are sampled in the dynamic parameter space to construct different simulation environments, and each dynamic parameter pi(i = 1, 2,…, N) is uniformly sampled in the corresponding interval [pilow, pihigh]. The variable dynamic parameters include the sliding friction coefficient between the glass substrate and the adsorption pad, as well as their contact stiffness and contact damping.

3.2.2. Force Compensation

The contact forces in the peeling process are far from rigid. For example, the glass substrate is an elastic body, which will bend during the peeling process, and there is an adsorption force between the glass substrate and the adsorption pad. To reduce the gap between the simulation environment and the real environment, the influence of the above factors should be considered.
The force of the blade on the glass substrate, on the one hand, generates a shear force fbend that causes the glass substrate to bend, and on the other hand, it also provides the load fpeel required for the desorption of the glass substrate.
The fbend can be calculated according to the deflection curve formula of the cantilever beam:
f b e n d = 2 E I φ L p
where ϕ is the desorption angle, Lp is the length of the glass substrate that has been peeled. the above two quantities can be observed in the simulation environment, and then calculated based on geometric relationship. E is the elastic modulus, and I is the bending moment of inertia.
Since the cross section of the glass substrate is rectangular, the thickness is h, and the width is b. Therefore, we have:
I = b h 3 12
Substituting Formula (18) into Formula (17), we have:
f b e n d = E b h 3 φ 6 L p
Our previous studies have shown that the critical peel force f ¯ p e e l for peeling the glass substrate from the adsorption pad is a function of velocity and peeling angle [28], and its physical unit is N. Combined with the measured data, the binary quadratic polynomial is used to approximate f ¯ p e e l , and the following expression is obtained:
f ¯ p e e l = 0.1623 + 0.01834 v 0.001579 ϕ 0.0003526 v 2 6.707 × 10 6 v ϕ + 9.326 × 10 6 ϕ 2
We assume that each joint of the robotic arm that performs the task can provide a large enough torque to peel the glass substrate, so the robot can provide a driving force that meets the critical peeling force f ¯ p e e l at any running speed, so f p e e l f ¯ p e e l .
Assuming that the inclination angle of the blade is α, the F R collected in the simulation environment can be compensated according to following equations:
F ˜ R X = f b e n d sin α
F ˜ R Z = f b e n d cos α + f p e e l
where F ˜ R Z and F ˜ R X are the compensations for vertical and horizontal components of F R , respectively.

3.3. Offline Learning of the Model

3.3.1. Dataset Construction

The data obtained through simulation is stored in the form of trajectories. We adopt an experience replay mechanism [29] to obtain datasets that can be used for model training from these trajectory data. The motivation for adopting the experience replay mechanism is to allow the stored samples to be resampled multiple times to calculate the gradient. Compared with discarding after use, this mechanism of reusing past training data helps to improve sample efficiency.
Our experience replay mechanism is an adaptation of replay mechanism that is a one-step state transition { x t , v t , x t + 1 } and helps improve the exploration of reinforcement learning. To accommodate the time series signals in our application, our experience replay mechanism is based on trajectory slices. The experience pool stores a complete trajectory each time, and a fixed-length trajectory slice is intercepted from a random position of a random trajectory during sampling. This not only allows the model to learn the time series characteristics of the sample but also indirectly expands the sample size by randomizing the sampling trajectory and slice position.
The training data sampling method used in the experience replay mechanism is shown in Figure 6, in which the trajectory slice is taken as the unit to extract data from the replay buffer to construct a batch of data, and then, the Mini-batch training method is used to train the system model.

3.3.2. Loss Function

The establishment of the system model is through regression process, and the loss function is mainly the mean square error (MSE):
L M S E = i = 1 n F x , v F ^ θ x , v 2
where F ^ θ x , v is the prediction of the model, and F x , v is the ground truth.
An additional part of loss function is designed as follows:
L P E N E = i n min ( 0 , Ψ lim ψ p , n , i ) 2
where ψ p , n , i is the distance that the blade is pressed into the adsorption pad, and ψ lim is the limit distance that the blade is allowed to be pressed into. L P E N E is added to avoid some unreasonable calculation results, which imply the wedge blade penetrating into the adsorption pad too deep.
The final regression of the system modeling is expressed as the optimization problem:
θ ^ = a g r min θ ( L M S E + κ L P E N E )
where κ is the adjustment weight coefficient.

3.4. Online Learning of the Model

Since the training data collected in the simulation environment is inevitably different from the real data, the model through offline learning cannot be directly applied to online planning. Although we can fine-tune the model by collecting some samples online, it is impossible to directly construct a complete data set covering all the situations. The manipulation planning needs the model to be adjustable online.
The ordinary deep neural network has obvious shortcomings in solving online learning problems [30], for example, the inevitably catastrophic forgetting phenomenon [31]. When using new data for training, the updated neural network parameters would make the network tend to ignore the impact of past tasks, leading to the learning system catastrophically forgetting past knowledge.
A straightforward idea is to store all past training samples [32]. Every time new samples are collected, the new samples and all past samples are used to update the model parameters. This will undoubtedly increase the data storage budget and increase the computation burden. More importantly, the new samples have a smaller proportion in the dataset, and the latest features of the system model are difficult to incorporate into the model parameter updates.
This paper introduces the idea of model agnostic meta learning (MAML) [33] for model training, trying to train a model with strong generalization ability, which can generally adapt to new peeling tasks. The algorithm is described below.
(1)
A new trajectory τ n e w is collected from an experiment, from which a dataset D n e w is sampled with the method shown in Figure 6. The parameters of the model are updated from θ to θ n e w on the training set D n e w , as follows with the gradient update step size α :
θ n e w = θ α θ L D n e w ( F θ )
(2)
From the experience pool (replay buffer) M trajectories, τ i ( i [ 1 , M ] ) are sampled corresponding to the M training set D i ( i [ 1 , M ] ) , and M sets of parameters θ i ( i [ 1 , M ] ) are obtained also through Equation (26).
(3)
Another M + 1 dataset is sampled independently with D N e w from τ n e w and D i ( i [ 1 , M ] ) from τ i ( i [ 1 , M ] ) . Now, the parameters are updated in a new way that sums up all the lost function together on this M + 1 dataset as follows. The purpose of this stage is to learn the commonality of all the trajectories.
θ θ β θ L D n e w f θ n e w + i = 1 M L D i f θ i
where β is the gradient update step size. In the field of meta learning, D n e w and D i are also called support sets, and D N e w and D i are called query sets.
The pseudocode of the model online learning algorithm (Algorithm 1) based on meta learning is as follows:
Algorithm 1. Online learning of the model based on meta learning
InputReplay Buffer D, Gradient update step size α and β
1While not converge:
2  Build data set D n e w from new trajectory τ n e w collected in real time
3  Obtain θ n e w based on formula (26)
4  Sample M trajectories { τ 1 , τ 2 , , τ M } from D
5  For i = 1 To M:
6    Build data set D i from trajectory τ i
7    Obtain θ i based on the formula similar to formula (26)
8  Build data set D i from trajectory τ i
9  Build data set D n e w from trajectory τ n e w
10  Update θ according to formula (27)
Output θ
In step 4 of Algorithm 1, different sampling methods can be used. We use a task-oriented experience replay mechanism. First, use the trajectory scoring mechanism proposed in [34] to score and sort the trajectories. Then, we extract the M trajectories with the highest scores. This sampling strategy can introduce the task goal as a priori information into the model training.

3.5. Manipulation Planning Based on Online Learning Model Predictive Path Integral

After the establishment of the system model x t + 1 = F θ ( x t , v t ) , the next step is to plan the robot motion according to the system model F θ to complete the glass substrate peeling. Suppose the optimal distribution of the velocity for completing the task is represented by Q * ( V ) . The optimal control input sequence satisfies:
U t * = E Q * v t = q * ( V t ) v t d V
Q * ( V ) is unknown, and we can use importance sampling to estimate the optimal control sequence:
q * ( V t ) v t d V = q * ( V t ) q ( V t U , Σ ) q ( V t U , Σ ) v t d V
We define ω ( V ) as the importance sampling weight:
ω ( V ) = q * ( V t ) q ( V t U , Σ )
Then Equation (28) can be written as:
U t * = E Q U , Σ ω V v t
ω ( V ) can be split into two parts through the following transformation:
ω ( V ) = q * V p V p V q V U , Σ
According to the information theory and optimal control principle, J ( τ i ) and the optimal velocity distribution q * ( V ) , satisfy the following formula [35]:
q * ( V ) = 1 η exp 1 λ J ( τ ) p V
η = exp 1 λ J ( τ ) p V d V
Substituting Formulas (33) and (34) into Formula (32):
ω ( V ) = 1 η exp 1 λ J ( τ ) p V q V U , Σ
According to the probability density formula of Gaussian distribution, we have:
p V p V U , Σ = exp 1 2 t = 0 T 1 v t T Σ 1 v t exp 1 2 t = 0 T 1 v t u t T Σ 1 v t u t = exp 1 2 t = 0 T 1 v t T Σ 1 v t v t T Σ 1 v t + 2 v t T Σ 1 u t u t T Σ 1 u t T = exp 1 2 t = 0 T 1 2 v t T Σ 1 u t u t T Σ 1 u t T
Since v t ~ N ( u t , Σ ) , we can make v t = u t + ε t . ε t is sampled from the normal distribution N ( 0 , Σ ) , and the sampling algorithm used in this paper is Box-Muller transform [36]. Therefore, the above formula can be written as:
p V p V U , Σ = exp 1 2 t = 0 T 1 2 u t T Σ 1 u t + 2 ε t T Σ 1 u t u t T Σ 1 u t T = exp 1 2 t = 0 T 1 u t T Σ 1 u t + 2 ε t T Σ 1 u t
Therefore, ω ( V ) can be written as:
ω ( V ) = 1 η exp 1 λ J ( τ ) + λ 2 t = 0 T 1 u t T Σ 1 u t + 2 ε t T Σ 1 u t
In this way, the optimal control input sequence can be calculated according to Formula (31). Since the glass substrate peeling task is safety critical, a model predictive control framework is introduced; that is, only the first element of the optimal control input sequence is executed each time. The robot manipulation planning algorithm (Algorithm 2) for the glass substrate peeling process is summarized as follows:
Algorithm 2. Online Learning Model Predictive Path Integral (OL-MPPI)
InputSystem model Fθ, Number of sampled trajectories N, rollout steps H, λ
InitialControl input sequence U, Variance Σ, Timing variable t, DΣ
1While tasks not completed:
2  While not (t > tmax or qt > =qter):
3    x ← GetStateFromSenors()
4    For k ← 0 To N-1:
5      Jk ← 0
6      Sample ε k = { ε 0 k , ε 1 k , , ε H 1 k } ,   ε k N(0,Σ)
7      For t← 0 To H-1:
8        vt = ut + εt
9        xx + Fθ(x,vt)∆t
10        Obtain Lt according to Formula (4), JkJk + Lt
11      If qt > =qter:
12        Obtain ϕ according to Formula (5), JkJk + ϕ
13        Obtain ω k according to Formula (38)
14      Update U based on formula (31)
15    Execute u* = U0 and Measure v*, DΣ ←{v*-u*}
16     t t + Δ t
17    Update Σ with DΣ
18    For t ← 0 To H-1:
19     ut-1ut
20  Update θ with Algorithm 1, t←0

4. Experiments

4.1. The Experiment Platform

(1)
The experimental platform is shown in Figure 7. Figure 7a is the Nachi 6DOF robot MZ04 with about ±0.1 mm of position tracking accuracy. It should be noted that the position tracking accuracy may deteriorate in the contact task environment. A six-dimensional force/torque sensor WEF-6A is installed at the end of the robot. The rated load of the sensor in three directions is ±500 N, and its measurement accuracy is 0.01 N. Figure 7b,c respectively show that the blade is located at the starting point and the end point of the peeling task. The thickness of the glass substrate we used is 0.5 mm, and the total distance that the robotic arm needs to move forward is 25 mm, that is, q ter = 25   mm .
(2)
We conducted simulation and offline model training on a computer server equipped with a GTX 1080Ti GPU.
(3)
A laptop computer was used for all the online algorithms calculation of experiments, which is connected to the robot controller with Ethernet TCP/IP protocol. The laptop CPU is Intel Core i7 8550U, and the GPU is NVIDIA Geforce MX150.
(4)
In the experiments, the robot works in speed control mode. The laptop calculates the optimal control speed and sends it to the robot controller for execution. The layout of the system is shown in Figure 8.

4.2. Performance Measures

(1)
System Model Evaluation
We use the root mean square error (RMSE) and correlation coefficient (Corr) as the evaluation index of the system model. The former can be used to judge the deviation of the predicted results, and the latter can be used to test whether the predicted trend of the model is consistent with the real trend. RMSE and Corr can be calculated according to the following formula:
R M S E ( y ) = 1 n i n y i y ^ i 2
C o r r ( y , y ^ ) = C o v ( y , y ^ ) V a r ( y ) V a r ( y )
where y and y ^ represent the real measured value and the model predicted value of the variable, respectively, and V a r ( y ) and V a r ( y ^ ) represent the variance of the real value and the predicted value of the variable, respectively.
(2)
Planning Results Evaluation
Two merits, namely the speed and the contact forces, are used to measure the performance of the peeling task, from C to D (Figure 1). The higher the speed to complete the task and the smaller the contact forces during the peeling, the better the performance is. The total cost function calculated according to Formula (3) can comprehensively reflect the above two aspects. Therefore, the cost function J ( τ i ) is used as an index to evaluate the overall performance of the planning result.
In addition, the task completion time tcpl is also used to evaluate the task completion speed, and the sum of the contact forces F R X and F R Z of all sampling points is used to evaluate the wear of the blade and pad. In the experiments, the sampling period is 100 ms, and the control period is 500 ms.
To avoid robot jamming at a certain position, set the maximum task completion time t max = 60   s . To avoid irreversible damage to the workpiece caused by the motion of the manipulator, the safety prewarning value of the horizontal and vertical contact force is set to 20 N during the peeling process. In all experiments, when the contact force exceeds the prewarning value, or the time exceeds t max , the experiment will be immediately stopped.

4.3. Model Offline Learning Process and Results

A simulation environment was constructed in Pybullet, and a total of 1000 sets of dynamic parameters were set to carry out simulation experiments. According to each group of dynamic parameters, 50 trajectories were sampled, and finally, a total of 50,000 trajectories were collected. Among them, 40,000 trajectories were used as the training set, 5000 trajectories were used as the verification set and the remaining 5000 trajectories were used for testing. Obtain the triple ( x t , v t , x t + 1 ) from the trajectory as a training sample, where x t is also a triple; see Formula (2). Among them, let the speed v t be equal to the ratio of the robot movement distance to the sampling period from time t to time (t + 1) and other quantities can be directly observed. The deep learning model is constructed by Pytorch. The optimizer used to solve Formula (25) is Adam, and the learning rate is equal to 0.0001.
Figure 9 is the RMSE curve under different epochs. The dotted line in the figure is the case where only LMSE is used, and the solid line is the case where both LMSE and LPENE are used. It can be found that the model performance jitter is obvious when only LMSE is used, while training is more stable when LMSE and LPENE. The reason for this phenomenon is that the former will fit some simulation data that is contrary to the actual physical situation, and the distribution between this type of data and the real data is very different, so the model training is unstable.
After offline training of the model based on simulation data, the RMSE is 1.21 after the model converges. To verify the superiority of the system model designed in this paper, the model is compared with the MLP model [37] and CNN-LSTM model [38], which are commonly used in data-driven system modeling. In the experiment, the MLP model and the regressor of our model adopt the same structure. The CNN-LSTM model only uses a single-size convolution kernel, and other settings (including the attention mechanism module) are consistent with the model we propose. The comparison results of the three models are shown in Figure 10. It can be seen that the prediction performance of the model in this paper is better than that of the MLP model and the CNN-LSTM model. The multi-scale convolution kernel used in this model can extract different levels of local information. Compared with the single-size convolution kernel model, it has obvious advantages in the representation of contact dynamics.

4.4. Manipulation Planning Results

In each peeling experiment, the planner will simultaneously simulate N = 50 trajectories based on the observation after each step of the control command and calculate their weights. Then, a new control sequence is obtained based on the weights, and the first step of the control sequence is executed. To test the robustness of the method proposed, 20 consecutive peeling experiments were carried out, during which there was no glass substrate breakage. As a comparison, we compared two other baseline methods:
  • Baseline 1: The system model is not updated, as in the original MPPI.
  • Baseline 2: Compared with OL-MPPI, each update of the system model uses all previous historical experience.
The total cost of the three methods for 20 consecutive peeling experiments is shown in Figure 11. According to the results of Baseline1, it can be seen that the total cost of the 20 peeling experiments is relatively stable, which shows that the experiments in this paper have good consistency. From the results of OL-MPPI, it can be seen that, as the number of times of performing tasks increases, the total cost decreases significantly, reduced from 42.22 in the 1st task to 15.98 in the 11th task. It shows that, as more trajectory data containing real system dynamics is accumulated, the system model can better characterize the real manipulation environment. In contrast, if the system model is trained with all previous experience each time, the total cost fluctuates greatly. Although the cost was reduced to 18.46 in the sixth experiment, it was difficult to maintain good performance due to catastrophic forgetting. In summary, the OL-MPPI algorithm can gradually improve the quality of manipulation planning.
Table 1 shows the task completion time and total contact force of the 20th peeling task. It can be seen that the algorithm proposed in this paper can complete the task in the shortest time and can significantly reduce the total wear on the glass substrate.
Figure 12 shows the curve of the robot velocity and the force in the two directions over time in the 20th task.
When the peeling starts, the horizontal and vertical forces fluctuate greatly. At this time, the velocity is slow to avoid glass breakage due to excessive contact force. As the vacuum between the glass substrate and the adsorption pad is broken, the horizontal force becomes the main factor that determines the velocity of the robot. At 3.0 s, 4.0 s and 5.0 s, the increase in horizontal contact force is accompanied by a decrease in the planned velocity. It can be seen that the planned control input sequence allows the robot motion to balance efficiency and wear, and its motion behavior is very similar to that of experienced workers.

4.5. Hyperparameter Analysis

The OL-MPPI algorithm can meet the needs of automatic peeling of glass substrates. However, in the experiment, we found that the hyperparameters in the OL-MPPI algorithm have a certain impact on the performance, which we will discuss here. We mainly focus on the influence of the three hyperparameters of N, H and λ. Using the system model used in the 20th peeling experiment, different hyperparameters were selected to carry out comparative experiments. The results are shown in Table 2.
  • N represents the number of simulations based on the system model in a single plan. Since the system model is highly nonlinear, it may not be possible to explore the state with low cost with a small number of simulations. When the computing resource is sufficient enough, sampling more trajectories can theoretically achieve better results. The experimental results show that the total cost decreases as N increases, which is consistent with our expectations. However, it should be noted that, when the value of N exceeds a certain threshold, the reduction in cost becomes smaller. Although the experiments were carried out with limited computing resources, choosing the right N can still achieve good performance.
  • H determines the length of the forward simulation. A higher H means that the agent has a “broader vision” of the task. However, as H increases, it will inevitably spend more time and may even fail to complete the calculation within the control period. It can be seen from the experimental result that with the increase of H, the total cost shows a trend of first decreasing and then increasing. This is because if H is selected too small, then the agent will not know the future loss, and the decision is too short-sighted. Setting H too large may amplify the error of the system model and then formulate unreasonable planning results based on the simulation results with large deviations.
  • λ determines the importance of the trajectory with the lowest cost. When λ is extremely small, the final trajectory will almost be the lowest cost trajectory, which easily makes the control sequence a local optimal solution. If λ is too large, the algorithm will be similar to random shooting. The experimental results show that there is a nonlinear relationship between the value of λ and the total cost. However, in general, when λ is within a certain range, the manipulation performance is not sensitive to λ.

5. Conclusions

Contact-rich manipulation is a challenging problem in robotic engineering applications. This paper takes the automated peeling of glass substrates as an example and proposes the OL-MPPI algorithm.
The proposed algorithm can perform manipulation planning in real time by updating a forward system model online. We carried out experiments to verify that the control sequence planned can meet the demand for automated peeling of glass substrates. The OL-MPPI algorithm can also be applied to other application scenarios involving contact-rich manipulations, because only a small amount of prior information about the task needs to be used to accelerate the offline model learning process, and the rest of the algorithm is scenario-independent. The advantage of the OL-MPPI algorithm is not only that it can support flexible cost function design but also can adapt to changes in the environment. This paper also provides some tricks that combine analytical mechanics and data-driven methods to build complex system models, which can be applied to various nonlinear dynamical systems in engineering applications. The forward model in the OL-MPPI algorithm can also use the pure mechanism model. In the online learning part of the OL-MPPI model, the mechanism model can be corrected online by adding a compensation term.
In the future, we will further improve the OL-MPPI algorithm and make efforts from the following aspects: (1) improve the efficiency of trajectory sampling in the planning process, so that the performance of the algorithm will no longer depend on the GPU used and (2) theoretically prove the influence mechanism of hyperparameters on the performance of the algorithm and provide a basis for the hyperparameters setting.

Author Contributions

Conceptualization, L.H.; methodology, L.H.; software, L.H.; validation, L.H., H.Z. and Y.Z.; formal analysis, L.H.; investigation, L.H.; resources, H.W.; data curation, L.H.; writing—original draft preparation, L.H.; writing—review and editing, H.W.; visualization, L.H.; supervision, H.W.; project administration, H.W. and funding acquisition, H.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Basic Research Program of China (grant No. 2013CB035504), Shenzhen Jiade Equipment Technology Co., Ltd. (Shenzhen, China), the Fundamental Research Funds for the Central Universities of Central South University (grant No. 2019zzts252) and National Natural Science Foundation of China (grant No. 51975587).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Choi, Y.S.; Yun, J.U.; Park, S.E. Flat panel display glass: Current status and future. J. Non-Cryst. Solids 2016, 431, 2–7. [Google Scholar] [CrossRef]
  2. Wang, H.; Zhang, Z.; Yang, H.; Liu, R.; Qiu, W.; Hou, L.; Liu, Y. Design and Use of the Glass Shoveling Device, Robotic Arm, Automatic Loading and Unloading Device. CN108726174A, 2 November 2018. (In Chinese). [Google Scholar]
  3. Kemp, C.C.; Edsinger, A.; Torres-Jara, E. Challenges for robot manipulation in human environments. IEEE Robot. Autom. Mag. 2007, 14, 20–29. [Google Scholar] [CrossRef]
  4. Qin, L.; Wang, H.; Yuan, Y.; Qin, S. Multi-Sensor Perception Strategy to Enhance Autonomy of Robotic Operation for Uncertain Peg-in-Hole Task. Sensors 2021, 11, 3818. [Google Scholar] [CrossRef]
  5. Lee, M.C.; Go, S.J.; Lee, M.H.; Jun, C.S.; Kim, D.S.; Cha, K.D.; Ahn, J.H. A robust trajectory tracking control of a polishing robot system based on CAM data. Robot. Comput. -Integr. Manuf. 2011, 17, 177–183. [Google Scholar] [CrossRef]
  6. Son, B.; Kim, C.U.; Kim, C.; Lee, D. Expert-emulating excavation trajectory planning for autonomous robotic industrial excavator. In Proceedings of the 2020 International Conference on Intelligent Robots and Systems, Las Vegas, NV, USA, 25 October 2020–24 January 2021; pp. 2656–2662. [Google Scholar]
  7. Zhou, Q.; Li, X. Experimental comparison of drag-wiper and roller-wiper glass-cleaning robots. Ind. Robot. 2016, 43, 409–420. [Google Scholar] [CrossRef]
  8. Li, C.; Fahmy, A.; Li, S.; Sienz, J. An enhanced robot massage system in smart homes using force sensing and a dynamic movement primitive. Front. Neurorobot. 2020, 14, 30. [Google Scholar] [CrossRef] [PubMed]
  9. Levine, S.; Finn, C.; Darrell, T.; Abbeel, P. End-to-end training of deep visuomotor policies. J. Mach. Learn. Res. 2016, 17, 1334–1373. [Google Scholar]
  10. Zhang, H.; Peng, Q.; Zhang, J.; Gu, P. Planning for automatic product assembly using reinforcement learning. Comput. Ind. 2021, 130, 103471. [Google Scholar] [CrossRef]
  11. Masahide, O.; Kusakabe, T.; Kutsuzawa, K.; Sakaino, S.; Tsuji, T. Reinforcement learning for robotic assembly using non-diagonal stiffness matrix. IEEE Robot. Autom. Lett. 2021, 6, 2737–2744. [Google Scholar]
  12. Zhou, J.; Mason, M.T.; Paolini, R.; Bagnell, D. A convex polynomial model for planar sliding mechanics: Theory, application, and experimental validation. Int. J. Robot. Res. 2018, 37, 249–265. [Google Scholar] [CrossRef]
  13. Bauza, M.; Alberto, R. A probabilistic data-driven model for planar pushing. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation, Marina Bay Sands, Singapore, 29 May–3 June 2017; pp. 3008–3015. [Google Scholar]
  14. Fazeli, N.; Kolbert, R.; Tedrake, R.; Rodriguez, A. Parameter and Contact Force Estimation of Planar Rigid-bodies Undergoing Frictional contact. Int. J. Robot. Res. 2017, 36, 1437–1454. [Google Scholar] [CrossRef] [Green Version]
  15. Ajay, A.; Wu, J.; Fazeli, N.; Bauza, M.; Kaelbling, L.P.; Tenenbaum, J.B.; Rodriguez, A. Augmenting Physical Simulators with Stochastic Neural Networks: Case Study of Planar Pushing and Bouncing. In Proceedings of the 2018 International Conference on Intelligent Robots and Systems, Madrid, Spain, 1–5 October 2018; pp. 3066–3073. [Google Scholar]
  16. Golts, A.; Freedman, D.; Elad, M. Unsupervised Single Image Dehazing Using Dark Channel Prior Loss. IEEE Trans. Image Process. 2019, 29, 2692–2701. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Cranmer, M.; Greydanus, S.; Hoyer, S.; Battaglia, P.; Spergel, D.; Ho, S. Lagrangian Neural Networks. arXiv 2020, arXiv:2003.04630. [Google Scholar]
  18. Luo, J.; Solowjow, E.; Wen, C.; Ojea, J.A.; Agogino, A.M.; Tamar, A.; Abbeel, P. Reinforcement learning on variable impedance controller for high-precision robotic assembly. In Proceedings of the 2019 IEEE International Conference on Robotics and Automation, Montreal, QC, Canada, 20–24 May 2019; pp. 3080–3087. [Google Scholar]
  19. Terra, M.H.; Cerri, J.P.; Ishihara, J.Y. Optimal robust linear quadratic regulator for systems subject to uncertainties. IEEE Trans. Autom. Control 2014, 59, 2586–2591. [Google Scholar] [CrossRef]
  20. Hogan, F.R.; Grau, E.R.; Rodriguez, A. Reactive planar manipulation with convex hybrid mpc. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation, Brisbane, Australia, 21–26 May 2018; pp. 247–253. [Google Scholar]
  21. Sleiman, J.; Farshidian, F.; Minniti, M.V.; Hutter, M. A unified mpc framework for whole-body dynamic locomotion and manipulation. IEEE Robot. Autom. Lett. 2021, 6, 4688–4695. [Google Scholar] [CrossRef]
  22. Kappen, K.J. Path integrals and symmetry breaking for optimal control theory. J. Stat. Mech. Theory Exp. 2005, 11, 11011. [Google Scholar] [CrossRef] [Green Version]
  23. Williams, G.; Aldrich, A.; Theodorou, E.A. Model predictive path integral control: From theory to parallel computation. J. Guid. Control Dyn. 2017, 40, 344–357. [Google Scholar] [CrossRef]
  24. Arruda, E.; Mathew, M.J.; Kopicki, M.; Mistry, M.; Azad, M.; Wyatt, J.L. Uncertainty averse pushing with model predictive path integral control. In Proceedings of the 2017 IEEE RAS International Conference on Humanoid Robotics, Birmingham, UK, 15–17 November 2017; pp. 497–502. [Google Scholar]
  25. French, R.M. Catastrophic forgetting in connectionist networks. Trends Cogn. Sci. 1999, 3, 128–135. [Google Scholar] [CrossRef]
  26. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  27. Hou, L.; Wang, H.; Liu, R. Research on Modeling and Experiment of Glass Substrate Peeling Based on Adhesion Theory. Teh. Vjesn. 2019, 26, 1827–1832. [Google Scholar]
  28. Pybullet, a Python Module for Physics Simulation in Robotics, Games and Machine Learning. Available online: https://github.com/bulletphysics/bullet3 (accessed on 8 January 2022).
  29. Adam, S.; Lucian, B.; Robert, B. Experience replay for real-time reinforcement learning control. IEEE Trans. Syst. Man Cybern. Part C 2011, 42, 201–212. [Google Scholar] [CrossRef]
  30. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef] [PubMed]
  31. Thrun, S.; Mitchell, T.M. Lifelong robot learning. Robot. Auton. Syst. 1995, 15, 25–46. [Google Scholar] [CrossRef]
  32. Goodfellow, I.J.; Mirza, M.; Xiao, D.; Courville, A.; Bengio, Y. An empirical investigation of catastrophic forgetting in gradient-based neural networks. arXiv 2013, arXiv:1312.6211. [Google Scholar]
  33. Finn, C.; Abbeel, P.; Levine, S. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 11–15 August 2017; Volume 70, pp. 1126–1135. [Google Scholar]
  34. Hou, L.; Wang, H.; Zou, H.; Wang, Q. Efficient Robot Skills Learning with Weighted Near-Optimal Experiences Policy Optimization. Appl. Sci. 2021, 11, 1131. [Google Scholar] [CrossRef]
  35. Theodorou, E.; Todorov, E. Relative entropy and free energy dualities: Connections to Path Integral and KL control. In Proceedings of the 51st IEEE Conference on Decision and Control, Maui, HI, USA, 10–13 December 2012; pp. 1466–1473. [Google Scholar]
  36. Box, G.E.P.; Muller, M.E. A note on the Generation of Random Normal Deviates. Ann. Math. Stat. 1958, 29, 610–611. [Google Scholar] [CrossRef]
  37. Janner, M.; Fu, J.; Zhang, M.; Levine, S. When to trust your model: Model-based policy optimization. In Proceedings of the 33rd Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; pp. 12519–12530. [Google Scholar]
  38. Funabashi, S.; Ogasa, S.; Isobe, T.; Ogata, T.; Schmitz, A.; Tomo, T.P.; Sugano, S. Variable In-Hand Manipulations for Tactile-Driven Robot Hand via CNN-LSTM. In Proceedings of the 2020 International Conference on Intelligent Robots and Systems, Las Vegas, NV, USA, 25 October 2020–24 January 2021; pp. 9472–9479. [Google Scholar]
Figure 1. Unloading processes of glass substrate. (a) Manual unloading. (b) Automatic process setup.
Figure 1. Unloading processes of glass substrate. (a) Manual unloading. (b) Automatic process setup.
Sensors 22 01292 g001
Figure 2. Schematic diagram of glass substrate peeling process.
Figure 2. Schematic diagram of glass substrate peeling process.
Sensors 22 01292 g002
Figure 3. Manipulation planning framework for automated peeling of glass substrate.
Figure 3. Manipulation planning framework for automated peeling of glass substrate.
Sensors 22 01292 g003
Figure 4. Local feature extraction based on multiscale convolution kernel.
Figure 4. Local feature extraction based on multiscale convolution kernel.
Sensors 22 01292 g004
Figure 5. Structure of forward system model.
Figure 5. Structure of forward system model.
Sensors 22 01292 g005
Figure 6. Sampling datasets based on a trajectory slice.
Figure 6. Sampling datasets based on a trajectory slice.
Sensors 22 01292 g006
Figure 7. Experiment platform. (a) Layout of the system. (b) Starting point of peeling. (c) End of peeling.
Figure 7. Experiment platform. (a) Layout of the system. (b) Starting point of peeling. (c) End of peeling.
Sensors 22 01292 g007
Figure 8. Experiment system layout.
Figure 8. Experiment system layout.
Sensors 22 01292 g008
Figure 9. RMSE curve under different epochs.
Figure 9. RMSE curve under different epochs.
Sensors 22 01292 g009
Figure 10. Performance comparison of different models.
Figure 10. Performance comparison of different models.
Sensors 22 01292 g010
Figure 11. Total cost of different planning methods.
Figure 11. Total cost of different planning methods.
Sensors 22 01292 g011
Figure 12. Velocity and the force in the two directions over time in the 20th task.
Figure 12. Velocity and the force in the two directions over time in the 20th task.
Sensors 22 01292 g012
Table 1. Duration and wear of the peeling task.
Table 1. Duration and wear of the peeling task.
MethodTcpl (s) F R X ( N ) F R Z ( N )
Baseline113.5 s1388.48 N836.37 N
Baseline210.4 s1057.54 N612.41 N
OL-MPPI7.1 s784.64 N486.45 N
Table 2. The impact of hyperparameters on performance.
Table 2. The impact of hyperparameters on performance.
HyperparameterValueTotal Cost
N1018.614
3015.448
5012.834
8012.226
10012.098
H124.691
317.541
512.834
716.657
1020.228
λ0.0213.178
0.0512.421
0.112.834
0.212.086
0.513.551
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hou, L.; Wang, H.; Zou, H.; Zhou, Y. Robotic Manipulation Planning for Automatic Peeling of Glass Substrate Based on Online Learning Model Predictive Path Integral. Sensors 2022, 22, 1292. https://doi.org/10.3390/s22031292

AMA Style

Hou L, Wang H, Zou H, Zhou Y. Robotic Manipulation Planning for Automatic Peeling of Glass Substrate Based on Online Learning Model Predictive Path Integral. Sensors. 2022; 22(3):1292. https://doi.org/10.3390/s22031292

Chicago/Turabian Style

Hou, Liwei, Hengsheng Wang, Haoran Zou, and Yalin Zhou. 2022. "Robotic Manipulation Planning for Automatic Peeling of Glass Substrate Based on Online Learning Model Predictive Path Integral" Sensors 22, no. 3: 1292. https://doi.org/10.3390/s22031292

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop