Abstract
This chapter deals with Iterative Learning Control ILC schemes to solve the trajectory tracking problem of affine and non-affine nonlinear systems performing repetitive tasks. Two ILC laws are presented; the first law is a simple on-line 2D-type learning control for affine nonlinear systems. In addition, an initial condition algorithm is generated to provide the initial state value at each iteration automatically. To prove the asymptotic stability of the closed loop system over the whole finite time interval when the iteration number tends to infinity, \(\lambda \)-norm is used, as the topological measure. The second law is the on-line P-type ILC applied to non affine nonlinear systems. The asymptotic stability of the closed loop system is guaranteed upon the use of a Lyapunov-like positive definite sequence, which is shown to be monotonically decreasing under the proposed control scheme. Finally, simulation results on nonlinear system are provided to illustrate the effectiveness of the two controllers.
Access provided by Autonomous University of Puebla. Download chapter PDF
Similar content being viewed by others
Keywords
1 Introduction
The design of a controller for dynamical system is typically divided into two different design problems: The first design problem is a regulation problem which consists of finding a control law that manipulates the input variable so that the system automatically holds the output at a constant value even when unknown disturbances try to move output away from this constant set point. The second one is the trajectory tracking problem which consists of forcing the output response to follow a desired trajectory as close as possible, for example Proportional-Integral-Derivative PID control [17], adaptive control [20], variable structure control [21], fuzzy control [26] and passivity-based control [2, 3]. In many cases, this desired trajectory is repeated over a given operation time, i.e., these nonlinear systems are used for repetitive tasks. The use of conventional control algorithms with such systems will result in the same level of tracking error being repeated time and time again. Iterative Learning Control ILC is a relatively new addition to the toolbox of control algorithms. It is concerned with the performance of systems that operate in a repetitive manner and includes examples such as robot arm manipulators. The basic idea of this method is to use information from previous operation in an attempt to improve performance from repetition to repetition in the sense that the tracking error (between the output and a specified reference trajectory) is sequentially reduced to zero. Since the early works of Arimoto et al. [1], Casalino and Bartolini [12], Craig [13], this technique has been the centre of interest of many researchers over the last decade (see, for instance, Bouakrif [4–6, 9] and Tayebi [25]).
In these works, the structures of ILC appeared as D-type, P-type, PD-type and PID-type. In addition, and for the sake of convergence speed, the forgetting factors have been introduced in ILC by Bouakrif [7]. From these works, it should be noted that ILC can be further classified into two kinds: off-line learning and on-line learning. In the case of off-line learning control, information in the controlled torque in the current iteration comes from the previous iteration. Philosophically, the learning in this case is shifted to the off-line mode. In the case of the on-line learning control, information in the controlled torque in the current iteration comes from the current iteration. Thus, the feedback control decision incorporates ILC at real-time.
Although, using ILC, the system should be started with the same initial condition at the beginning of each iteration [15, 16, 19], there have many works on ILC without identical initial condition [8, 18, 23, 27]. The first control law presented in this chapter is an ILC scheme without identical condition.
The most useful and general approach for studying the stability of nonlinear control systems is the theory introduced by the Russian mathematician Alexander Mikhailovich Lyapunov. Lyapunov’s work includes two methods for stability analysis, the linearization method and direct method. The first one draws conclusions about a nonlinear system’s local stability around an equilibrium point from the stability properties of its linear approximation. The second method is not restricted to local motion, and determines the stability properties of a nonlinear system by constructing a scalar energy-like function for the system and examining the function’s time variation. This last method has become the most important tool for nonlinear system analysis and design. Recently, another type of ILC algorithms has been developed using a positive definite Lyapunov-like sequence which is made monotonically decreasing along the iteration axis via a suitable choice of the control input. In fact, [14] utilized Lyapunov-based techniques to develop an ILC that is combined with a robust control design to achieve global uniformly ultimately bounded link position tracking for robot manipulators. Using Lyapunov-like function, [10, 24] derived an adaptive ILC and a velocity observer based ILC, respectively, to solve the trajectory tracking problem of robot manipulators.
Using another proof of the stability for such controller (iterative learning controller), \(\lambda \)-norm defined firstly by Arimoto et al. [1], is used as the topological measure, to prove the asymptotic stability of the closed loop system over the whole finite time interval when the iteration number tends to infinity. Using this norm and to solve the trajectory tracking problem for nonlinear systems with arbitrary relative degree and no state measurement, a state observer based iterative learning controller has been presented by Bouakrif [11].
In this chapter, we present two ILC schemes to solve the trajectory tracking problem of affine and non affine nonlinear systems performing repetitive tasks. The first law is a simple on-line 2D-type learning control for affine nonlinear systems with an initial condition algorithm to provide the initial state value at each iteration automatically. Using \(\lambda \)-norm, the asymptotic stability of such controller is guaranteed, over the whole finite time interval when the iteration number tends to infinity. The second law is an on-line P-type ILC of non affine nonlinear systems. To prove the asymptotic stability of the closed loop system, a Lyapunov-like positive definite sequence is used. It is shown to be monotonically decreasing under the proposed control scheme. Finally, simulation results on nonlinear system are provided to illustrate the effectiveness of the two controllers.
2 Iterative learning control design
2.1 General Form of Iterative Learning Controller
In general case, the ILC scheme is presented as follows
We note that f depends on errors, and/or on derivative errors, and/or on integral errors obtains from different cycles. The fundamental problem resides to determine a simple recursive form of f ensuring the error convergence and a satisfactory rate convergence. Indeed, the Eq. (1) can be written as
Figure 1 shows the block diagram of Eq. (2).
From this block, the following can be easily obtained
\(u_{ff} (t)\) : is the feedforword controller.
\(u_{fb} (t)\) : is the current cycle feedback controller.
2.2 Off-Line Iterative Learning Control Scheme
The off-line ILC scheme is obtained when \(\phi _{k+1} \) equals to zero. It is important to note that this scheme in open loop via iterations domain. The off-line ILC scheme is given by:
The original off-line ILC scheme was proposed firstly by Arimoto et al. [1], and it applied mainly in robotics. This scheme is as follows:
where \(\phi \) is the learning gain and \(\frac{d}{dt}e_k (t)\) is the derivative error.
2.3 On-Line Iterative Learning Control Scheme
Adding a current cycle feedback controller to open loop ILC scheme, the on-line ILC scheme is obtained. Thus, the form of this scheme is given by:
3 On-Line 2D-Type Iterative Learning Control for Affine Non Linear Systems
3.1 System Description
Consider the following affine non linear system
where k denotes the iteration index or the operation number, \(x_k \in R^{n} \) , \(u_k \in R^{r} \) and \(y_k \in R^{m} \) are the state, control input and output of the system, respectively. The function \(f({*})=[f_1 ({*}),\ldots .f_n ({*})]\in R^{n}\) is strictly unknown. The functions \(f({*})=[f_1 ({*}),\ldots .f_n ({*})]\in R^{n}\) and \(h({*})=[h_1 ({*})... h_m ({*})]^{T}\in R^{m}\) are smooth in their domain of definition, and let the associated time interval be \(t\in [0, T]\).
The Jacobi of h(x(t), t) is given by
The following assumption is needed.
Assumption A1
The function f(x(t), t) satisfies the Lipschitz condition with respect x over the time interval t\(\in \)[0, T]. Thus \(\exists \alpha \) \(_{ }\)for \(x_1 (t),x_2 (t)\in R\times [0, T]\) such as
The following lemmas are used.
Lemma 1
Let \(z(t)=[z_1 (t),z_2 (t),\ldots .z_n (t)]^{T}\in R^{n}\) is defined for t\(\in \)[0, T], then we have
Lemma 2
(Gronwall-Bellman) [22] Suppose that f(t) and \(g(t)\ge 0\) are real and locally integrable scalar functions in [a, b], and L is a constant. If the real scalar function f(t) satisfies the integral equation
Then, on the same interval, f(t) satisfies
Lemma 3
Letting \(\xi (t),\eta _1 (t),\eta _2 (t)\) the continuous functions in [0, T]. If
then
with \(\Gamma \ge 0\) , \(\Gamma _1 \) and \(\Gamma _2 \) are constants.
Proof
Multiplying (13) by \(e^{-\lambda t}\), we have
Applying the Lemma 1, we obtain
Using Lemma 2, it comes
Thus
3.2 Iterative Learning Control Design
Objective
Our objective is to design a controller that is updated iteratively such that the output trajectory \(y_k (t)\) follows a desired trajectory \(y_d (t)\), for \(t\in [0,T]\) \(_{ }\)and \(k\rightarrow \infty \).
In this section, we consider a simple on-line 2D-type learning control and an initial condition algorithm. This algorithm provides the initial state value at each iteration automatically. Using \(\lambda \)-norm defined firstly by Arimoto et al. [1], as the topological measure, the asymptotic stability of the closed loop system is guaranteed over the whole finite time interval when the iteration number tends to infinity. More specifically, the output system is proved to be convergence to the desired output by showing that the inequality
holds if \(\varpi \prec 1. \text { With } e_{k} (t)=y_{d} (t)-y_{k} (t)_{.}\)
The formal definition of the \(\lambda \)-norm for a function \(f:[0,T]\rightarrow R^{n}\) is given by
Throughout the paper, we will use the following norms.
The asymptotic stability conditions for such controller are given in the following theorem.
Theorem 1
Given the affine nonlinear system (7), and let assumption A1 be satisfied. For any initial state \(x_1 (0)\), any admissible control \(u_1 (t)\), and applying the following on-line 2D-type iterative learning control and initial state algorithm
If
1—\(h_x (x,t)\) is bounded in \(R^{n}\times [0,T]\),
2—\(I_m +h_x (x(t),t)B(t)K(t)\) is non-singular,
3—\(\mathop {\sup }\limits _{(x,t)\in R^{n}\times [0,T]} \left\| {(I_m -h_x (x,t)B(t)L(t)).(I_m +h_x (x,t)B(t)K(t))^{-1}} \right\| <1.\)
Then
where \(e_k (t)=y_d (t)-y_k (t)\). \(K(t)\in R^{r\times m}\) and \(L(t)\in R^{r\times m}\) are the gain matrices, with \(I_n +B(0)K(0)h_x (x(0))\) is non singular.
Proof
From (7), (21) and (22), we have
consequently
Using assumption A1, Lemmas 1 and 2, we find
with \(a=\mathop {\max }\limits _{t\in [0, T]} \left\| {B(t)L(t)} \right\| \), \(b=\mathop {\max }\limits _{t\in [0, T]} \left\| {\frac{dB(t)L(t))}{dt}} \right\| \), \(\gamma _1 =a+\frac{b}{\lambda }\), \(\gamma _2 =c+\frac{d}{\lambda }\), \(c=\mathop {\max }\limits _{t\in [0, T]} \left\| {B(t)K(t)} \right\| \) and \(d=\mathop {\max }\limits _{t\in [0, T]} \left\| {\frac{dB(t)K(t))}{dt}} \right\| \).
Using the theorem of differential mean value, it exists \(\xi _k (t)\), such that
From (26), (28) and using assumption A1, we have
Choosing \(\lambda (\lambda >0)\) widely great, we obtain
If
Then \(y_k (t)\) converges uniformly to \(y_d (t)\), when \(k\rightarrow \infty \) and t\(\in \)[0, T].
3.3 Simulation Results
Consider the dynamic model of robot manipulator given by
g is the gravitational acceleration, u(t) is the input control, q(t) is the rotation angle of the robot.
We takes \(x_1 =q\) and \(x_2 =\dot{q}\), thus we obtain
The output is \(y=\frac{2}{5}x_2 \). \(J_m =14\) kgm\(^{2}\), \(S=6\,\mathrm{kg} \cdot \mathrm{m}\), \(g=9.8\) m/s\(^{2}\), T\(=\)2 s, \(h_x =\left[ {0,\frac{2}{5}} \right] \).
The ILC laws is given by
with \(L(t)=10\) and \(K(t)=14\).
For a continuous function \(y_d (t)\) in [0, T], it exists \(u(t)\in U\), such that \(y_d (t)=h(x_d (t),t)\), with \(x_d (t)\) is generated by \(u_d (t)\), given as
with \(x_1 (0)=[0,1]\), and \(u_1 (t)=1\).
Applying the ILC law, we obtain the following results.
The simulation results are given in Figs. 2, 3, 4 and 5. These figures present the simulation results of the real and desired trajectories for 1st, 3rd, 7th and 10th iteration. We can see that the real trajectory follows the desired one through learning iteration. Thus, the system executes 10 iterations so that the real output system follows the desired trajectory without error.
4 On-Line P-Type Iterative Learning Control for a Class of Non affine Nonlinear Systems
4.1 System Description
Consider the non-affine time-varying nonlinear systems described by
where k denotes the iteration index or the operation number, \(x_k \in R^{n} \) , \(u_k \in R^{r} \) and \(y_k \in R^{m} \) are the state, control input and output of the system, respectively. \(w_{k}(t)\) is the external disturbance of the system. The functions \(f({*})=[f_1 ({*}),\ldots .f_n ({*})]\in R^{n}\) and \(h({*})=[h_1 ({*})...h_m ({*})]^{T}\in R^{m}\) are smooth in their domain of definition, and let the associated time interval be \(t\in [0,T]\).
4.2 Iterative Learning Control Design
In order to design a controller that is updated iteratively such that the system output (36) can follow the desired output, one presents the following on-line P-type ILC law:
where \(e_k (t)=y_d (t)-y_k (t)\) is the trajectory tracking error, \(L_1 \) is diagonal positive definite matrix.
It is important to note here that the asymptotic stability is proved using a positive definite Lyapunov-like sequence which is made monotonically decreasing along the iteration axis.
Theorem 2
Given the unknown nonlinear systems (36) under the iterative learning control law (37), the closed loop system is asymptotically stable, i.e.,
-
(i)
\(\mathop {\lim }\limits _{k\rightarrow \infty }e_k (t)=0 \qquad t\in [0,T].\)
-
(ii)
\(u_k (t)\) is uniformly convergent on [0, T].
Proof
Letting
where \(\Delta {\begin{array}{l} {u_k (t)} \\ \end{array} }=u_k (t)-u_{k-1} (t)\).
It is clear that
Thus
From (37), the control law at \(k^{th}\) iteration is as follows
From (37), (40) and (41), we have
where \(B=\left( {{\begin{array}{l} {-I_n } \\ {-L_2 } \\ \end{array} }} \right) \) and \(F_k (t)=\left( {h(x_{k+1} (t),t)-h(x_k (t),t)} \right) \).
Stability analysis: The proof of the theorem 2 is in three parts. The first part consists of taking a positive definite Lyapunov-like composite energy function, namely \( {\sum ^{f}}V_k (t)\), and show that this sequence is non-increasing with respect to k and hence bounded if \(V_0 (t)\) is bounded. In the second part, one shows that \(V_0 (t)\) is bounded for all \(t\in [0,T]\). In the third part, one shows that \(\mathop {\lim }\limits _{k\rightarrow \infty } e_k (t)=0\) and \(u_k (t)\) is uniformly convergent \(\forall t\in [0,T]\).
Part 1: Let us consider the following Lyapunov-like sequence
where P is symmetric, positive definite matrix.
At \((k+1)\)th iteration, we have
In the sequel, since the time t does not have any impact on the stability analysis, it is removed for the sake of simplicity.
Let’s define
From (43), (44), (45) and (46), we can write
Since P is symmetric, we can apply the theorem of Rayleigh-Ritz. It becomes
where \(\lambda _{\min }(P)\) denotes the minimum eigenvalue of P.
It is clear that
Hence \(V_k \) is non-increasing sequence. Thus if \(V_0 \) is bounded, we can conclude that \(V_k \) is bounded.
Part 2: Now, we will show that \(V_0 \) is bounded over the time interval [0, T]. In fact, from (44) \(V_0 \) is given by
with \(\Phi _j =\left( {{\begin{array}{l} {e_j } \\ {\Delta {\begin{array}{l} {u_j } \\ \end{array} }} \\ \end{array} }} \right) \).
From (40) and (42), \(\Delta {\begin{array}{l} {u_0 } \\ \end{array} }\) and \(\Delta {\begin{array}{l} {u_1 } \\ \end{array} }\) are given by
and
Before applying the control at the first iteration, it is logical to suppose that \(e_{-1} =e_0 =0\) and \(u_{-1} (t)=u_0 (t)=0\). Hence, it becomes
and
Since the control at the first iteration is bounded, we can conclude from (54) that \(e_1 \) is bounded.
We note that
Knowing that \(V_{k+1} \le V_k \) for \(k\ge 1\), we can use the following reasoning:
Since \(e_1 \) is bounded and using this reasoning, we can conclude that \(e_\infty \) is bounded. This implies that all \(\Phi _j \), \(j=1,\ldots ,\infty \) are bounded. Finally and knowing that \(\Phi _0 \) is bounded, we can conclude from (50) that \(V_0 \) is bounded over [0, T].
Part 3: We note that \(V_k \) can be written as follows
Thus, from (48), we have
This implies that
Since \(V_k (t)\) is bounded \(\forall k\in N\) and \(\forall t\in [0,T]\), this implies that \(\Phi _k \) disappears when \(k\rightarrow \infty \). Therefore, \(\mathop {\lim }\limits _{k\rightarrow \infty } e_{k}(t)=0\) for \(t\in [0,T]\) and \(u_k (t)\) is uniformly convergent. This completes the proof.
4.3 Simulation Results
Consider the dynamical model of a nonlinear system given as follows
where \(t\in [0,2](s)\), \(J_m =14\,\mathrm{kgm}^{2}\), \(S=6\,\mathrm{kg} \cdot \mathrm{m}\), \(g=9.8\,\mathrm{{m/s}}^{2}\) (gravitational acceleration), and \(u_k (t)\) is the control.
Adding the disturbances, we have
where the disturbances are given by
with \(f_0 =1/(20h)\) and \(a=0.1\). The output is chosen as: \(y=\frac{1}{4}x_2 \). The desired trajectory is chosen as: \(y_d (t)=t^{2}-t\), and \(u_1 (t)=1\).
Simulation parameters: \(L_1 =diag\left\{ {90,90} \right\} \).
Applying the control law (37), the simulation results for real and desired trajectories for 2\(\mathrm{nd}\), 5\(\mathrm{th}\), 10\(\mathrm{th}\), 30\(\mathrm{th}\) and 40\(\mathrm{th}\) iteration are shown in Figs. 6, 7, 8, 9 and 10. We can see that the real trajectory follows the desired trajectory through learning iteration. Thus, the system executes 40 iterations so that the real trajectory follows the desired trajectory without error. Figure 11 shows the trajectory tracking error for different number of iterations. It is clear that the trajectory tracking error decrease through the iterations. Therefore, we can conclude that the control algorithm works well.
5 Conclusion
In this chapter, two ILC schemes have been presented to solve the trajectory tracking problem of affine and non affine nonlinear systems performing repetitive tasks. The first law is a simple on-line 2D-type learning control applied to affine nonlinear systems. In this control scheme, the initial state value at each iteration is provided automatically via an initial condition algorithm. Using \(\lambda \)-norm, the asymptotic stability of such controller is guaranteed, over the whole finite time interval when the iteration number tends to infinity. The second law is an on-line P-type ILC of non affine nonlinear systems. To prove the asymptotic stability of the closed loop system, a Lyapunov-like positive definite sequence is used. It is shown to be monotonically decreasing under the proposed control scheme. Finally, simulation results on nonlinear system are provided to illustrate the effectiveness of the two controllers. It is important to note that, the control laws here are very simple in the sense that there is no dependence with the system.
References
Arimoto S, Kawamura S, Miyazaki F (1984) Bettering operation of robots by learning. J Robot Syst 1:123–140
Berghuis H, Nijmeijer H (1993) A passivity approach to controller observer design for robots. IEEE Trans Robot Autom 9:740–754
Bouakrif F, Boukhetala D, Boudjema F (2010) Passivit based controller-observer for robot manipulators. Int J Robot Autom 25:1–8
Bouakrif F (2010a) Reference model iterative learning control for nonlinear systems. Int J Adv Manuf Tech 51:1159–1169
Bouakrif F (2010b) Reference model iterative learning control for linear systems. Proceedings, MED’2010, IEEE mediterranean conference on control and automation, 18:1637–1642
Bouakrif F (2011a) Iterative learning control for strictly unknown nonlinear systems subject to external disturbances. Int J Control Autom Syst 9:642–648
Bouakrif F (2011b) Iterative learning control with forgetting factor for robot manipulators with strictly unknown model. Int J Robot Autom 26:264–271
Bouakrif F (2011c) D-type iterative learning control without resetting condition for robot manipulators. Robotica 29:975–980
Bouakrif F (2012) Commande par apprentissage itératif des robots manipulateurs, pour résoudre le problème de poursuite de trajectoires. Editions Universitaires Européennes, Germany
Bouakrif F, Boukhetala D, Boudjema F (2013) Velocity observer based iterative learning control for robot manipulators. Int J Syst Sci 42:214–222
Bouakrif F (2013) Iterative learning control for MIMO nonlinear systems with arbitrary relative degree and no states measurement. Complexity 19:37–45
Casalino G, Bartolini G (1984) A Learning Procedure for the Control of Movements of Robotic Manipulators. IASTED Symposium on Robotics and Automation, Amsterdam, pp 108–111
Craig J. J. (1984). Adaptive Control of Manipulators Through Repeated Trials. Proceedings of the American Control Conference, San Diego, pp. 1566–1574
Ham C, Qu Z, Johnson R (2000) A nonlinear iterative learning control for robot manipulators in the presence of actuator dynamics. Int J Robot Autom 15:119–130
Jang T, Choi C, Ahn H (1995) Iterative learning control in feedback systems. Automatica 31:243–248
Lee KH, Bien Z (1991) Initial condition problem of learning control. IEE Proceedings D (Control Theory and Applications) 138:525–528
Li Y, Ang KH, Chong GCY (2006) PID control system analysis and design. IEEE Control Syst Mag 26:32–41
Park KH, Bien Z (2000) A generalized iterative learning controller against initial state error. Int J Control 73:871–881
Porter B, Mohamed SS (1991) Iterative learning control of partially irregular multivariable plants with initial state shifting. Int J Syst Sci 22:229–235
Sastry S, Bodson M (1989) Adaptive control: stability convergence and robustness. Prentice Hall, New Jersey
Sira-Ramirez H (1992) On the sliding mode control of nonlinear systems. Syst Control Lett 19:303–312
Sontag ED (1998) Mathematical control theory, 2nd edn. Springer, New York
Sun M, Wang D (2002) Iterative learning control with initial rectifying action. Automatica 38:1177–1182
Tayebi A (2004) Adaptive iterative learning control for robot manipulators. Automatica 40:1195–1203
Tayebi A (2007) Analysis of two particular iterative learning control schemes in frequency and time domains. Automatica 43:1565–1572
Wang HO, Tanaka K, Griffin MF (1996) An approach to fuzzy control of nonlinear systems: stability and design issues. IEEE Trans Fuzzy Syst 4:14–23
Xu JX, Yan R (2005) On initial conditions in iterative learning control. IEEE Trans Autom Control 50:1349–1354
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Bouakrif, F. (2016). Iterative Learning Control for Affine and Non-affine Nonlinear Systems. In: Vaidyanathan, S., Volos, C. (eds) Advances and Applications in Nonlinear Control Systems. Studies in Computational Intelligence, vol 635. Springer, Cham. https://doi.org/10.1007/978-3-319-30169-3_24
Download citation
DOI: https://doi.org/10.1007/978-3-319-30169-3_24
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-30167-9
Online ISBN: 978-3-319-30169-3
eBook Packages: EngineeringEngineering (R0)