Escolar Documentos
Profissional Documentos
Cultura Documentos
4, 527540
FINITE HORIZON NONLINEAR PREDICTIVE CONTROL BY THE TAYLOR
APPROXIMATION: APPLICATION TO ROBOT TRACKING TRAJECTORY
RAMDANE HEDJAR
, REDOUANE TOUMI
, PATRICK BOUCHER
, DIDIER DUMUR
(),
[t, t + h] the optimal control vector for the above
problem. The currently applied control u(t) is set equal
to u
T
(T)Q(T) dT,
V(x, x
ref
, h) =
_
h
0
T
(T)Q
_
Z(x, T)
d(t, T)
_
dT.
R. Hedjar et al.
530
Tracking performance. We will assume that the matrix
W(x) has a full rank. This assumption is needed for the
stability analysis, but is not necessary for the control law
to be applicable, since one can always choose R > 0, and
then the inverse matrix in (7) will still exist. If R = 0,
then (7) becomes
u
op
= W(x)
1
1
(h) (K(h)e(t) + V(x, x
ref
, h)) .
This optimal control signal u
op
, used in (2), leads to the
closed-loop equation of the i-th component of the track-
ing error vector e(t) which is given by
h
ri
(2r
i
+ 1)r
i
!
(ri)
e
i
+
h
ri1
2r
i
(r
i
1)!
(ri1)
e
i
+
+
h
3
3!(r
i
+ 4)
(3)
e
i
+
h
2
2!(r
i
+ 3)
e
i
+
h
(r
i
+ 2)
e
i
+
1
(r
i
+ 1)
e
i
= 0,
or, in a compact form,
ri
j=0
h
j
j!(r
i
+ j + 1)
(j)
e
i
= 0, (8)
where
(j)
e
i
= L
j1
f
f
i
(j)
x
ref
i
, 0 < j r
i
,
(0)
e = x
i
x
ref
i
, j = 0.
The error dynamics (8) are linear and time-invariant.
Thus, the proposed controller that minimizes the pre-
dicted tracking error naturally leads to a special case of
input/state linearization. The advantage of this controller
compared with the linearization method is a clear physical
meaning of a maximum and a minimum when saturation
occurs. Note that, by using the Routh-criterion, we can
show that the tracking error dynamics (8) are stable only
for systems with r
i
4.
For most mechanical systems with actuator dynamics
neglected, the relative degree is r
i
= 1 or r
i
= 2. In this
case, the eigenvalues of the characteristic equations of the
error dynamics are:
Pings method (Ping, 1995):
s
1
=
1
h
(r
i
= 1)
or s
1,2
=
1
h
(1 j) (r
i
= 2).
the proposed method:
s
1
=
3
2h
(r
i
= 1)
or s
1,2
=
5
4h
_
1 j
_
17
15
_
(r
i
= 2).
Thus, the proposed controller achieves higher tracking er-
ror dynamics compared with Pings method (Ping, 1995).
2.2. Particular Afne Nonlinear Systems
To overcome the stability restriction of the relative degree
r
i
4, we will consider a special form of nonlinear sys-
tems that are modelled by the equations
_
x
1
= x
2
,
x
2
= f(x) + g(x)u(t),
(9)
where x =
_
x
1
x
2
T
R
2n
and u(t) R
n
. We
note that many physical systems can be modeled by the
above equations. For example, in mechanical systems, x
1
can represent a position vector and x
2
a velocity vector.
In this case, the objective function to minimize is
J
2
(e, u, t)
=
1
2
_
h
0
_
e
1
(t + T)
e
2
(t + T)
_
T
_
Q
1
0
0 T
2
Q
2
_
_
e
1
(t + T)
e
2
(t + T)
_
dT +
1
2
u
T
(t)R u(t).
(10)
The tracking error is given by
e(t) = x(t) x
ref
(t) =
e
1
(t)
e
2
(t)
x
1
x
ref1
x
2
x
ref2
.
By using the Taylor approximation, the tracking error
is then predicted as a function of u(t) by
_
_
e
1
(t + T) = e
1
(t) + T e
1
+
T
2
2!
(f(x) x
ref1
)
+
T
2
2!
g(x) u(t),
e
2
(t + T) = e
2
(t) + T (f(x) x
ref2
)
+Tg(x) u(t),
and the minimization of the cost function J
2
gives
u(t) = M(x)P
1
_
h
3
6
Q
1
e
1
+
h
4
8
(Q
1
+ 2Q
2
)e
2
+
h
5
20
(Q
1
+ 4Q
2
)(f(x) x
ref2
)
_
, (11)
where
P =
h
5
20
(Q
1
+ 4Q
2
) + M(x)RM(x)
is a positive-denite matrix, x
ref
1
= x
ref
2
and M(x) =
g
1
(x).
Finite horizon nonlinear predictive control by the Taylor approximation . . .
531
2.3. Stability Issues
Dynamic performance. To obtain the tracking error dy-
namic, one substitutes the control signal (11) in (9), to
have
_
_
e
1
= e
2
,
e
2
=
h
3
6
P
1
Q
1
e
1
h
4
8
P
1
(Q
1
+ 2Q
2
)e
2
+P
1
M(x)R M(x)(f(x) x
ref2
).
(12)
Let Q
1
= q
1
I
n
and Q
2
= q
2
I
n
. The tracking error equa-
tion (12) can be written in a compact form:
e = (x, h)e + B S
l
(x, x
ref
), (13)
where
(x, h) =
0 I
n
h
3
q1
6
P
1
h
4
(q1+2q2)
8
P
1
,
B =
0
I
n
,
and the perturbed term is given by
S
l
(x, x
ref
) = P
1
M(x)R M(x)(f(x) x
ref2
).
Assumptions (A1)(A4) insure the boundedness of this
additional term.
Lemma 1. The matrix (x, h) is Hurwitz.
Proof. Both the matrix P and its inverse are symmetric
and positive denite. Let x R
n
and R
+
be the
eigenvector and the correspondent eigenvalue of the ma-
trix P
1
, respectively. Thus, for R we have the
equalities
(x, h)
x
x
h
3
q1
6
x
h
4
(q1+2q2)
8
x
2
x
x
x
,
where is the solution of the equation
2
+
h
4
(q
1
+ 2q
2
)
8
+
h
3
q
1
6
= 0. (14)
Therefore, is the eigenvalue of the matrix (x, h) and
|
x
x
| is the correspondent eigenvector. Setting
1
and
2
as the solutions of (14), we have the relations
1
+
2
=
h
4
8
(q
1
+ 2q
2
),
2
=
h
3
6
q
1
.
Since the eigenvalue is positive, then both
1
and
2
have negative real parts. Thus, the matrix
(x, h) is Hurwitz. Consequently, for any symmetric
positive-denite matrix Q
a
(x, h), there exists a symmet-
ric positive-denite matrix P
a
(x, h) being a solution of
the Lyapunov equation
P
a
(x, h) +
T
(x, h)P
a
(x, h) + P
a
(x, h)(x, h)
= Q
a
(x, h).
Theorem 1. The solution e(t) of the system (12) is
uniformly ultimately bounded (Khalil, 1992) for all
t t
0
> 0.
Proof. Consider the Lyapunov function candidate
V(e) = e
T
P
a
e. (15)
The differentiation of V along the trajectories of the sys-
tem (12) leads to
V (e) = e
T
Q
a
e + 2S
T
l
B
T
P
a
e, (16)
which can be bounded by using Assumptions (A1)(A4)
as
V (e)
min
(Q
a
) e
2
+2
max
(P
a
)rm
2
e ,
where r = R and m = M(x) .
We will use the well-known inequality
ab za
2
+
b
2
4z
for any real a, b and z > 0. With z =
min
(Q
a
) and
0 < < 1, we obtain
V (e) (1 )
min
(Q
a
)e
2
+
max
(P
a
)
2
r
2
2
m
4
min
(Q
a
)
.
(17)
The solution of this inequality is
V (t)
_
V (0)
_
exp
_
t +
_
,
where
= (1 )
min
(Q
a
)
max
(P
a
)
and
=
max
(P
a
)
2
r
2
2
m
4
min
(Q
a
)
.
As t , the tracking error is bounded by
e
max
(P
a
)
min
(Q
a
)
rm
2
_
(1 )
max
(P
a
)
min
(P
a
)
. (18)
R. Hedjar et al.
532
It can be easily shown that, as R tends to a null matrix
by reducing the penalty on control, the bound of the per-
turbed term decreases for large t and the equilibriumpoint
tends to the origin. Setting R = 0 in (12), the time deriva-
tive of the Lyapunov function becomes
V (e) = e
T
Q
a
e,
which is negative denite for all e. By LaSalles invari-
ance theorem, the solution e(t) of (12) tends to the invari-
ant set S =
_
e | e
2
= 0, P
1
e
1
= 0
_
. Since the matrix
P
1
has a full rank, we have that e
1
= 0. So, the origin
e = 0 is globally asymptotically stable.
2.4. Robustness Issues
In the real world, model uncertainties are frequently en-
countered in nonlinear control systems. These model un-
certainties may decrease signicantly the performance of
the method in terms of tracking accuracy. Therefore, one
should inspect the robustness of the closed-loop system
with respect to uncertainties. The model of the nonlinear
system (9) with uncertainties can be written as
_
x
1
= x
2
,
x
2
= f(x) + f(x) + (g(x) + g(x)) u(t).
(19)
To estimate the worst-case bound of the uncertainties, we
make the following assumptions:
(A5) x(t) X, > 0, f(x) < .
(A6) x(t) X, = max
g(x)
g(x)
, the uncertainties in
the matrix g(x) can be bounded by g(x) = g(x)
with 0 < < 1.
Let R = 0. The dynamics of the tracking error in
a mismatched case in a closed loop with the optimal con-
trol (11) is given by
_
_
e
1
= e
2
,
e
2
=
h
3
6
(1 + )P
1
Q
1
e
1
h
4
8
(1 + )P
1
(Q
1
+ 2Q
2
)e
2
+ f(x) (f(x) x
ref2
).
(20)
Note that here, even though R = 0, the origin is not
an equilibrium point of the system (20). However, we can
use the steps of the Lemma 1 and Theorem 1 to show that
the tracking error e(t) is ultimately bounded in this mis-
matched case and the equilibrium point is given by the set
S= {e/e
1
= 0, e
2
= 0}. Hence, the uncertainties will in-
troduce only a short steady-state error in the tracking po-
sition error. The bound of this steady state error depends
on the magnitude of the uncertainties.
Integral action. It is known in the literature that the in-
tegral action increases the robustness of the closed-loop
system against low frequency disturbances as long as the
closed-loop system is stable. In this part, we shall incor-
porate an integral action into the loop to enhance the ro-
bustness of the proposed control scheme with respect to
model uncertainties and disturbances. The price to be paid
is an increase in the system dimension. Thus, the nonlin-
ear system (9) is augmented with the differential equation
x
0
= x
1
and the tracking error vector becomes
e(t) =
e
0
(t) e
1
(t) e
2
(t)
T
,
with e
0
=
_
t
0
e
1
() d and e
0
(t + h) given by
e
0
(t+T) = e
0
+Te
1
+
T
2
2
e
2
+
T
3
6
(f(x) x
ref2
)+
T
3
6
g(x)u(t). (21)
The cost function to be minimized becomes
J
3
(e, u, t) =
1
2
_
h
0
e(t + T)
T
Qe(t + T) dT
+
1
2
u(t)
T
R u(t), (22)
where Q = diag (Q
0
, Q
1
, T
2
Q
2
) and Q
0
R
nn
is
a positive-denite matrix. Following the same steps as in
the previous section, the optimal control u(t) that mini-
mizes the cost function J
3
(e, u, t) is
u(t) = M(x)P
1
_
0
(h)e
0
+
1
(h)e
1
+
2
(h)e
2
+
3
(h)(f(x) x
ref2
)
_
, (23)
where
0
(h) =
h
4
12
Q
0
,
1
(h) =
h
5
15
Q
0
+
h
3
6
Q
1
,
2
(h) =
h
6
36
Q
0
+
h
4
8
Q
1
+
h
4
4
Q
2
,
3
(h) =
h
7
63
Q
0
+
h
5
20
Q
1
+
h
5
5
Q
2
,
P =
3
(h) + M(x)RM(x).
Finite horizon nonlinear predictive control by the Taylor approximation . . .
533
Dynamic performance. Let R = 0. Then the dynamics
of the tracking error are given in a compact form by
e = (x, h)e + BS
l
, (24)
where
(x, h)
=
0 In 0
0 0 In
(1+)P
1
0(h) (1+)P
1
1(h) (1+)P
1
2(h)
,
S
l
= f(x) (f(x) x
ref2
)
and
B =
0
0
I
n
.
Note that also the perturbed term S
l
is bounded.
Lemma 2. Let the parameters Q
0
, Q
1
, Q
2
and h > 0
satisfy the inequality
max
(P) <
(1 + )
1
(h)
2
(h)
0
(h)
.
Hence the matrix (x, h) is Hurwitz.
Proof. Let Q
0
= q
0
I
n
, Q
1
= q
1
I
n
, Q
2
= q
2
I
n
, and as-
sume that x and are an eigenvector and the correspond-
ing eigenvalue of the matrix P
1
, respectively. Hence we
have
(x, h)
x
x
2
x
2
x
3
x
x
x
2
x
,
where R is the solution of
3
+ (1 + )
2
(h)
2
+ (1 + )
1
(h)
+ (1 + )
0
(h) = 0. (25)
The roots of this equation are stable if
>
0
(h)
(1 + )
1
(h)
2
(h)
.
Since is the eigenvalue of the matrix P
1
, the pre-
vious inequality becomes
min
(P
1
) >
0
(h)
(1 + )
1
(h)
2
(h)
or
max
(P) <
(1 + )
1
(h)
2
(h)
0
(h)
,
and this ensures that all poles of the matrix (x, h) lie in
the stable domain. It is to be noted that is the eigenvalue
of the matrix (x, h) and the vector
x x
2
x
T
is the corresponding eigenvector.
Theorem 2. Under the assumptions of Lemma 2, the so-
lution of the tracking error (24) is ultimately uniformly
bounded.
Proof. Since the matrix (x, h) is Hurwitz, we know that
for any symmetric positive-denite matrix Q
a
(x, h), the
solution P
a
(x, h) of the Lyapunov equation
P
a
(x, h) +
T
(x, h)P
a
(x, h) + P
a
(x, h)(x, h)
= Q
a
(x, h) (26)
is a positive-denite matrix. We use V = e
T
P
a
e as a
Lyapunov function candidate for the augmented nonlinear
system (24). Following the same steps as in the proof of
Theorem1, we can showthat the tracking error is bounded
by
e
max
(P
a
)
min
(Q
a
)
( + )
_
(1 )
max
(P
a
)
min
(P
a
)
. (27)
The tracking error in the mismatched case with
integral action is bounded. Here also the bound de-
pends on the magnitude of the uncertainties. However,
the equilibrium point of the augmented system is S =
{e | e
0
= 0, e
1
= 0, e
2
= 0, }, the position tracking error
in this case converges to zero. Consequently, the steady
error induced by uncertainties is eliminated by the integral
action. Note that the price to be paid is the control signal
that will not vanish as the time t goes towards innity.
3. Simulation Examples
In this section, the reference trajectory tracking problem
is simulated to show the validity and the achieved perfor-
mance of the proposed method.
3.1. Nonlinear Predictive Control of a Nonholonomic
Mobile Robot
A kinematic model of a wheeled mobile robot with two
degrees of freedom is given by (Kim et al., 2003):
x = v cos() d sin(),
y = v sin() + d cos(),
= ,
(28)
where the forward velocity v and the angular velocity
are considered as the inputs, (x, y) is the centre of the rear
axis of the vehicule, is the angle between the heading
direction and the x-axis, and d is the distance from the
R. Hedjar et al.
534
coordinate of the origin of the mobile robot to the axis of
the driving wheel.
The nonholonomic constraint is written as
y cos() x sin() = d
.
The nonlinear model of the mobile robot can be rewritten
as
Z = G()U,
where
Z =
_
x y
_
T
,
G() =
cos() d sin()
sin() d cos()
0 1
,
U =
_
v
_
T
.
Note that the above model matches the rst general
multi-variable afne nonlinear system given by (2) with
f(x) = 0.
Consider the problem of tracking a reference trajec-
tory given by
x
ref
= v
ref
cos(
ref
),
y
ref
= v
ref
sin(
ref
),
ref
=
ref
,
(29)
or, in a compact form,
Z
ref
= G(
ref
)U
ref
. The optimal
control that minimizes the objective function (6) subject
to (28) is
U =
_
h
3
3
G
T
()Q G() + R
_
1
G
T
()Q
_
h
2
2
e(t)
h
3
3
G(
ref
)U
ref
_
, (30)
where e(t) = Z(t) Z
ref
.
In the simulation, the control parameters are
h = 0.01, Q = 10
4
I
3
, R = 10
7
I
2
.
The reference model and initial conditions are
ref
= 4 [rad/s], d = 0.5 [m],
v
ref
= 15 [m/s], x(0) = 0,
y(0) = 4 [m], (0) = [rad].
Figures 1 and 2 show the resulting trajectory and position
tracking error
e
r
(t) =
_
(x x
ref
)
2
+ (y y
ref
)
2
,
when the nonlinear predictive controller (30) is applied to
the system (28). We can see that the mobile robot tracks
the reference trajectory successfully. Figure 3 depicts the
manipulated variables v(t) and w(t).
4 3 2 1 0 1 2 3 4
1
0
1
2
3
4
5
6
7
8
y
(
t
)
x(t)
Position of the mobile robot and the refrence trajectory
Z(t)
Zref(t)
Fig. 1. Tracking performance.
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
0
0.5
1
1.5
2
2.5
3
3.5
4
Time(s)
Tracking error:
Fig. 2. Tracking error dynamics.
Consequently, the proposed approach, which can be
viewed as an extension to nonlinear systems of the CGPC
developed by Demircioglu et al. (1991), was success-
fully applied to control a nonlinear system with a non-
holonomic constraint. On the other hand, the CGPC ap-
proach (Demircioglu et al., 1991), can be applied only to
linear systems. Moreover, with the proposed algorithm,
the stability of the closed-loop system is guaranteed and
asymptotic tracking performances are achieved.
3.2. Nonlinear Predictive Control of a Rigid-Link
Robot
To illustrate the conclusions of this paper, we have simu-
lated the nonlinear predictive scheme (11) on a two-link
Finite horizon nonlinear predictive control by the Taylor approximation . . .
535
robot arm used in (Lee et al., 1997; Spong et al., 1992)
with the parameters given in Table 1.
Table 1. Physical parameters of a two-link robot manipulator.
Link1 m1 = 10 kg l1 = 1 m lc1 = 0.5 m I1 =
10
12
kgm
2
Link
2
m2 = 5 kg l2 = 1 m lc2 = 0.5 m I2 =
5
12
kgm
2
The kinetic energy of a robot manipulator with n de-
grees of freedomcan be calculated as (Spong et al., 1989):
K(q, q) =
1
2
q
T
(t)M(q) q(t),
where q(t) R
n
is the link position vector, M(q) is
the inertia matrix, and U(q) stands for the potential en-
ergy generating gravity forces. Applying Euler-Lagrange
equations (Spong et al., 1989), we obtain the model
M(q) q + C(q, q) q + G(q) + F
r
= , (31)
where
G(q) =
U(q)
q
R
n
,
C(q, q) q is the vector of the Coriolis and centripetal
torques, R
n
stands for the applied torque, and F
r
represents friction torques acting on the joints. These fric-
tion are unknown and are modeled by
F
r
= f q(t) +f sign( q(t))
with f =diag(f, f, . . . , f) R
nn
and f =diag (f, f, . . . , f) R
nn
.
A state representation. The dynamic equation of an n-
link robot manipulator (31) can be written in the state
space representation as
_
_
x
1
= x
2
,
x
2
= f(x
1
, x
2
) + P(x
1
) (t),
y = x
1
,
(32)
where x =
_
x
1
x
2
T
=
_
q q
T
R
2n
is
the state vector, (t) R
n
represents the control torque
vector and y(t) is the output vector (angular position).
Here f(x) = f(x
1
, x
2
) = M(q)
1
(C(q, q) q +
G(q)) R
n
and P(x
1
) = M(q)
1
R
nn
are a
bounded vector under the assumption of the bounded-
ness of joints velocities and a bounded matrix, respec-
tively. We note that both the symmetric positive-denite
matrix M(q) and its inverse are uniformly bounded with
respect to the joint angular position q(t). Thus Assump-
tions (A1)(A4) are satised by the nonlinear model of
the robot given by (32), c.f. (Spong et al., 1989).
The dynamic model is described by (31) with (see
Lee et al., 1997; Spong et al., 1992):
M
11
(q) = m
1
l
2
c1
+ m
2
l
2
c2
+ m
2
l
2
1
+ 2m
2
l
1
l
c2
cos(q
2
) + I
1
+ I
2
,
M
21
(q) = M
12
(q) = m
2
lc2
2
+ m
2
L
1
l
c2
cos(q
2
) + I
2
,
M
22
(q) = m
2
l
2
c2
+ I
2
,
C
11
(q, q) = m
2
l
1
l
c2
sin(q
2
) q
2
,
C
12
(q, q) = m
2
l
1
l
c2
sin(q
2
)( q
1
+ q
2
),
C
21
(q, q) = m
2
l
1
l
c2
sin(q
2
) q
1
,
C
22
(q, q) = 0,
G
1
(q) = (m
1
l
c1
+ m
2
l
1
) g cos(q
1
)
+ m
2
l
c2
g cos(q
1
+ q
2
),
G
2
(q) = m
2
l
c2
g cos(q
1
+ q
2
).
Nonlinear observer. A drawback of the previous non-
linear predictive controller is that it requires at least the
measurement of the velocity on the link side. Therefore,
a nonlinear observer proposed in (Gauthier et al., 1992) is
used in this paper. Dene the new state vector as
z(t) = T x(t) =
_
q
i
(t) q
i
(t)
R
2n
,
where q
i
(t)and q
i
(t) are the link position and the velocity
of the i-th arm, respectively. T R
2n2n
is the trans-
formation matrix. With the assumption that the control
torque (t) is uniformly bounded, the observer described
in (Gauthier et al., 1992) can be used to estimate the an-
gular positions and angular velocities of the n-link rigid
robot manipulator (32). The dynamic nonlinear observer
is given by
_
z = A z + H f(q,
q) + H P(q)(t)
S
1
C
T
(y y),
y = C z,
x = T
1
z,
(33)
where
A = diag(A
i
), A
i
=
_
0 1
0 0
_
,
C = diag(C
i
), C
i
=
_
1 0
_
,
and
H = diag(H
i
), H
i
=
_
0 1
_
with i = 1, n.
The observer gain S
() = diag(S
i
()) is given by
the solution of the following Ricatti equation with the real
positive factor :
S
i
() + A
T
i
S
i
() + S
i
()A
i
= C
T
i
C
i
. (34)
R. Hedjar et al.
536
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
100
0
100
200
300
400
500
600
Control signal: v(t)
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
0
50
100
150
200
Time(s)
Control signal: W(t)
Fig. 3. Manipulated variables.
0 1 2 3 4
0
0.5
1
1.5
2
(
r
a
d
)
q
1
and q
ref1
0 1 2 3 4
0.5
0
0.5
1
1.5
2
(
r
a
d
)
q
2
and q
ref2
0 1 2 3 4
0.4
0.3
0.2
0.1
0
0.1
0.2
0.3
Time(s)
(
r
a
d
)
Tracking error: e
1
(t)
0 1 2 3 4
0.4
0.2
0
0.2
0.4
0.6
Time(s)
(
r
a
d
)
Tracking error: e
2
(t)
Fig. 4. Position tracking performance.
Finite horizon nonlinear predictive control by the Taylor approximation . . .
537
0 1 2 3 4
2
0
2
4
6
8
(
r
a
d
/
s
)
qv
1
and qv
ref1
0 1 2 3 4
10
5
0
5
10
15
(
r
a
d
/
s
)
qv
2
and qv
ref2
0 1 2 3 4
4
2
0
2
4
Time(s)
(
r
a
d
/
s
)
Tracking error: de
1
0 1 2 3 4
10
5
0
5
10
Time(s)
(
r
a
d
/
s
)
Tracking error: de
2
Fig. 5. Velocity tracking performance.
Fig. 6. Estimation error.
R. Hedjar et al.
538
0 0.5 1 1.5 2 2.5 3 3.5 4
2000
1000
0
1000
2000
Time(s)
(
N
m
)
Torque:
1
0 0.5 1 1.5 2 2.5 3 3.5 4
500
0
500
Time(s)
(
N
m
)
Torque:
2
Fig. 7. Applied control signals.
Fig. 8. Performance in the mismatched case.
Finite horizon nonlinear predictive control by the Taylor approximation . . .
539
According to (Gauthier et al., 1992), a Lyapunov analy-
sis shows that for a real positive factor and the uniform
observability assumption of the nonlinear system (32), the
solution (34) guarantees an exponential decay of the ob-
servation error.
The reference models chosen in continuous time are
x
ref1
=
q
ref1
q
ref2
x
ref1
x
ref2
.
They are smoothed by means of second-order polynomials
that are respectively given by
q
ref1
(s) =
2
1
s
2
+ 2
1
s +
2
1
r
1
(s)
and
q
ref2
(s) =
2
2
s
2
+ 2
2
s +
2
2
r
2
(s).
The nonlinear predictive controller (11) is used to
force the joint positions to track the desired trajectory (Lee
et al., 1997):
r
1
(t) = r
2
(t) = 1.5(1 exp(5t)(1 + 5t)) [rad].
For this simulation, the parameter values of the two refer-
ence models are chosen as = 1 and
1
=
2
= 10 and
the initial conditions are
x(0) =
_
q
1
(0) q
2
(0) q
1
(0) q
2
(0)
=
_
0.1 0.1 0 0
,
x(0) =
_
0 0 0 0
.
Note that the initial estimation errors are different from
zero. Thus, with the proposed feedback nonlinear predic-
tive controller, one does not need to constrain the initial
estimation errors in the joint position to be zero to en-
sure the convergence of the tracking error to zero as in
(Canudas De Wit et al., 1992).
The nite horizon predictive controller (11) has been
tested by simulation with the following control parame-
ters: Q
1
= 2 10
2
I
2
, Q
2
= 2 10
2
I
2
, R = 10
8
I
2
and h = 0.01. The resulting position and speed track-
ing error are depicted in Figs. 4 and 5. The behaviour of
the state x(t) is close to the reference trajectory x
ref
(t).
Observing these results, the state x(t) tracks tightly the
reference trajectories x
ref
(t). Figure 6 displays the ob-
servation tracking error achieved by Gauthiers observer.
Figure 7 illustrates the torque signals applied to the robot
manipulator, which are inside the saturation limits (Lee et
al., 1997; Spong et al., 1992).
In the mismatched case, the uncertainties used are
Parameter variations in Link 2 due to an unknown
load are m
2
= 5 kg; l
2
= 0.3 m and I
2
=
1/6.
The friction is added to the robot manipulator (31)
with the values f
1
= f
2
= 10 Nm and f
1
= f
2
=
10 N/ms
2
.
Case (a) in Fig. 8 shows the tracking performance of
Link 2 without integral action. It observed that the un-
certainties induce a short steady-state error in the second
link position q
2
(t), and this was expected by the robust-
ness analysis. When the integral action is introduced in
the control loop with Q
0
= 10
4
I
2
, it is observed from
the same gure (Case (b)) that the position reference tra-
jectory is closely tracked and the tracking error converges
towards the origin. Thus the torque frictions and parame-
ters uncertainties have no effect on the state tracking error.
4. Conclusion
In this paper, a nite-horizon nonlinear predictive con-
troller using the Taylor approximation is presented and
applied to two kinds of nonlinear systems. Minimizing
a quadratic cost function of the predicted tracking error
and the control input, we derived the control law. One
of the main advantages of these control schemes is that
they do not require on-line optimization and asymptotic
tracking of the smooth reference signal is guaranteed. The
stability was shown by using the Lyapunov method. Ac-
cording to a suitable choice of the control parameters, we
showed that all variables of the state tracking error are
bounded. The boundedness can be made small by reduc-
ing the penalty on the control torque signal. Moreover,
to increase the robustness of the proposed scheme to vari-
ations and uncertainties in parameters, an integral action
was incorporated into the loop. The proposed controllers
are applied to both the planning motion problem of a mo-
bile robot under nonholomonic constraints and the prob-
lem of tracking trajectories of a rigid link robot manip-
ulator. Finally, we expect that the results presented here
can be explored and extended to a discrete implementation
of these continuous-time predictive controllers through ei-
ther computers or specialized chips that can run at a higher
speed.
Acknowledgment
The authors are grateful to the anonymous referees for
their constructive and helpful observations and sugges-
tions that improved the quality of this paper.
R. Hedjar et al.
540
References
Boucher P. and Dumur D. (1996): La commande prdictive.
Paris: Technip.
Canudas De Wit C. and Fixot N. (1992): Adaptive control
of robot manipulators via velocity estimated feedback.
IEEE Trans. Automat. Contr., Vol. 37, No. 8, pp. 1234
1237.
Chen W.H., Balance D.J. and Gawthrop P.J. (2003): Optimal
control of nonlinear systems: A predictive control ap-
proach. Automatica, Vol. 39, No. 4, pp. 633641.
Chun Y.S. and Stepanenko Y. (1996): On the robust control
of robot manipulators including actuator dynamics. J.
Robot. Sys., Vol. 13, No. 1, pp. 110.
Clarke D.W, Mohtadi C. and Tuffs P.S. (1987a): Generalized
predictive control, Part I: The basic algorithm. Auto-
matica, Vol. 23, No. 2, pp. 137148.
Clarke D.W, Mohtadi C. and Tuffs P.S. (1987b): Generalized
predictive control, Part II. Extension and interpretations.
Automatica, Vol. 23, No. 2, pp. 149160.
Demircioglu H. and Gawthrop P.J. (1991): Continuous-time gen-
eralized predictive control (GPC). Automatica, Vol. 27,
No. 1, pp. 5574.
Gauthier J.P., Hammouri H. and Othman S. (1992): A simple
observer for nonlinear systems: Application to bioreactor.
IEEE Trans. Automat. Contr., Vol. 37, No. 6, pp. 875
880.
Henson M.A. and Seborg D.E. (1997): Nonlinear Process Con-
trol. Englewood Cliffs, NJ: Prentice Hall.
Henson M.A. (1998): Nonlinear model predictive control: Cur-
rent status and future directions. Comput. Chemi. Eng.,
Vol. 23, No. 2, pp. 187202.
Khalil H.K. (1992): Nonlinear Systems. New York: Macmil-
lan.
Kim M.S., Shin J.H., Hong S.G. and Lee J.J. (2003): De-
signing a robust adaptive dynamic controller for nonholo-
nomic mobile robots under modelling uncertainty and dis-
turbances. Mechatron. J., Vol. 13, No. 5, pp. 507519.
Lee K.W. and Khalil H.K. (1997): Adaptive output feedback
control of robot manipulators using high gain observer.
Int. J. Contr., Vol. 67, No. 6, pp. 869886.
Mayne D.Q., Rawlings J.B., Rao C.V. and Scokaert P.O.M.
(2000): Constrained model predictive control: stability
and optimality. Automatica, Vol. 36, No. 6, pp. 789
814.
Michalska H. and Mayne D.Q. (1993): Robust receding horizon
control of constrained nonlinear systems. IEEE Trans.
Automat. Contr., Vol. 38, No. 11, pp. 16231633.
Morari M., Lee J.H. (1999): Model predictive control: past,
present and future. Comput. Chem. Eng., Vol. 23,
No. 45, pp. 667682.
Ping L. (1995): Optimal predictive control of continuous nonlin-
ear systems. Int. J. Contr., Vol. 62, No. 2, pp. 633649.
Singh S.N., Steinberg M. and DiGirolamo R.D. (1995): Nonlin-
ear predictive control of feedback linearizable systems and
ight control system design. J. Guid. Contr. Dynam.,
Vol. 18, No. 5, pp. 10231028.
Souroukh M. and Kravaris C. (1996): A continuous-time for-
mulation of nonlinear model predictive control. Int. J.
Contr., Vol. 63, No. 1, pp. 121146.
Spong M.W. and Vidyasagar M. (1989): Robot Dynamics and
Control. New York: Wiley.
Spong M.W. (1992): Robust control of robot manipulators.
IEEE Trans. Automat. Contr., Vol. 37, No. 11, pp. 1782
1786.
Received: 24 August 2004
Revised: 8 March 2005
Re-revised: 8 September 2005