Unreachable Setpoints in MPC
Unreachable Setpoints in MPC
9, OCTOBER 2008
I. INTRODUCTION
Model predictive control (MPC) has been employed extensively for
constrained, multivariable control in chemical plants [1]. The main moManuscript received February 25, 2008; revised April 23, 2008. Current version published October 8, 2008. This work was supported in part by the industrial members of the TexasWisconsin Modeling and Control Consortium
and the National Science Foundation under Grant CTS-0456694, by the Computer Aided Process and Products Center, Technical University of Denmark,
and by the Danish Department of Energy under Grant EFP 273/01-0012. Recommended by Associate Editor M. Fujita.
J. B. Rawlings is with the Department of Chemical and Biological Engineering, University of Wisconsin, Madison, WI 53706-1691 USA (e-mail: [email protected]).
D. Bonn is with the CAPEC, Department of Chemical and Biochemical Engineering, Technical University of Denmark, DK 2800 Lyngby, Denmark.
J.B. Jrgensen is with the DTU Informatics, Technical University of Denmark, DK 2800 Lyngby, Denmark.
A.N. Venkat is with the Shell Global Solutions (U.S.), Inc., Houston, TX
77084 USA.
S.B. Jrgensen is with the CAPEC, Department of Chemical and Biochemical
Engineering, Technical University of Denmark, DK 2800 Lyngby, Denmark.
Digital Object Identifier 10.1109/TAC.2008.928125
2209
tivation for choosing MPC is to handle constrained systems. Essentially all current MPC theory is predicated on the assumption that the
setpoint has been transformed to a desired and reachable steady-state
target in a separate steady-state optimization [2], [3]. This assumption
allows current theory to address only transient active constraints that
occur when solving the controllers dynamic optimization to drive the
system to the desired and reachable steady state. In practice, the setpoint often becomes unreachable due to a disturbance and the choice
to transform the problem to one with a desired and reachable steady
state is a significant decision that affects controller performance. We
show in this paper that when the setpoint is unreachable, this two-level
optimization is suboptimal and does not minimize the tracking error.
To improve the performance of MPC when the setpoint is unreachable,
we propose defining system performance relative to the unreachable
setpoint rather than the reachable steady-state target. As we show in
the examples, a consequence of the unreachable setpoint formulation
is a desirable fast/slow asymmetry in the system tracking response that
depends on the system initial condition, speeding up approaches to the
unreachable setpoint, but slowing down departures from the unreachable setpoint. One technical difficulty that arises is that the proposed
MPC cost function is unbounded on the infinite horizon. A more significant problem is that the existing MPC theory for stability and convergence no longer applies because this theory is based on establishing
the controller cost function as a Lyapunov function for the closed-loop
system, but the controller cost function is not strictly decreasing in this
case.
Optimal control problems with unbounded cost are not new to control theory. The first studies of this class of problems arose in the economics literature in the 1920s [4], in which the problem of interest was
to determine optimal savings rates to maximize capital accumulation.
Since this problem has no natural final time, it was considered on the
infinite horizon. A flurry of activity in the late 1950s and 1960s led
to generalizations regarding future uncertainty, scarce resources, expanding populations, multiple products and technologies, and many
other economic considerations. Much of this work focused on establishing existence results for optimal control with infinite horizon and
unbounded cost, and the famous turnpike theorems [5] that characterize the optimal trajectory. Refs. [6] and [7] provides a comprehensive and readable overview of this research.
This class of problems was transferred to and further generalized in
the control literature. For infinite horizon optimal control of continuous
time systems, [8] established the existence of overtaking optimal trajectories. Convergence of these trajectories to an optimal steady state
is also demonstrated. Ref. [9] extended the results of [8] to infinite
horizon control of discrete time systems. Reduction of the unbounded
cost, infinite horizon optimal control problem to an equivalent optimization problem with finite costs is established. Ref. [10] provide a
comprehensive overview of these infinite horizon results.
For feedback control subject to hard input constraints, we often
employ a receding horizon implementation of a suitably chosen finite
horizon optimal control problem. The controller is implemented by
solving the open-loop control problem at each sample time as the state
becomes available and injecting only the first move of the open-loop
optimal solution. Ref. [11] provides a review of the methods for analyzing the resulting closed-loop behavior of these controllers. In this
formulation, the existence of the optimal control is not at issue because
the horizon is finite and the cost therefore bounded, but the stability
and convergence of the resulting receding horizon implementation of
the control law is the interesting question. In this paper, we address
the closed-loop properties of this receding horizon implementation for
the case of unreachable setpoints. In the formulation introduced here,
2210
the stage cost of the optimal control problem measures distance from
the setpoint even if the setpoint is unreachable at steady state due to
the input constraints.
The paper is organized as follows. In Section II, basic results required to establish properties of the new controllers are given. The
new controller, termed setpoint tracking MPC, is defined in Section III.
For simplicity of exposition, a terminal state equality constraint is employed. Convergence and nominal asymptotic stability are established
for the closed-loop system. Two examples demonstrating the benefit of
the proposed setpoint tracking MPC over current approaches are presented in Section IV. Extending the asymptotic stability result to terminal penalty MPC is discussed in Section V, and conclusions of this
study are presented in Section VI.
II. PRELIMINARIES
( ) = f jk 0 k g
x
min ( ) subject to
x
( ) = 12 0
Consider a sequence ( ) 2
0
x Qx + q x + c
L x
x i
xP
Q >
= 1 ...
1
=
()
,i
=1
( )=
=1
( ( )) 0 21
=1
( + 1) =
( ) 1
P
=1
k ( ) 0 k2
x i
xP
(1)
( )=
P L xP
=1
( ( ))
the
()
x i
not
all
= 1 ...
;
;P
equal,
01
are
2
gives
the
=1 k ( ) 0 k
(
(
))
0
(
)
.
Knowing
=1
P
i
x i
xP
min ( ) subject to 2
x
L x ;
2
2
()
(2)
Bu k
L x; u
) = 21 ( 0 sp )0 ( 0 sp ) + ( 0 sp )0 ( 0 sp )
0
0
x
Q >
R >
Q x
R u
=f j
u Hu
g
L x; u
x;u
( ( )) if and only if ( )
x i
Ax k
min (
L x i
L x i
= ( + 1)
Definition 1 (Optimal Steady State): The optimal steady state, denoted (x3 ; u3 ), is the solution to the following optimization problem
L xP
( )+
x k
x i :
L x i
p;
with mean
;P
Ax
L x ;
1
2
Bb a
If
subject to
x
Ax
Bu
We assume the model and constraints admit a steady state, i.e., the
steady-state problem is feasible. If A has no eigenvalues at unity, then
any admits a steady state. If A has eigenvalues at unity, then containing an element of the null space of B , for example, is sufficient for
a steady state to exist.
Remark 3: Positive definiteness of Q and R implies that L(x; u) is
strictly convex. Therefore (x3 ; u3 ) is unique according to Lemma 2.
A. Terminal Constraint MPC
Define the controller cost function
x;
f ( )g =001 =
u i
N
i
( + 1) =
x k
01
=0
( ( ) ( ))
L x k ;u k
( )+
Ax k
()
Bu k ;
(0) =
x:
u=
min
8(x; u)
u
subject to x(0) = x
x(k + 1) = Ax(k ) + Bu(k )
k = 0; . . . ; N 0 1
3
x (N ) = x
u (k ) 2
k = 0; . . . ; N 0 1:
The set is nonempty because it contains x3 . The set is bounded because N is finite and is bounded. All closed-loop sequences are therefore bounded because the set N is positive invariant under the MPC
control law.1
For singular A, the steerable set is not bounded. The following
lemma, however, shows that the closed-loop sequence starting in the
(unbounded) steerable set evolves in a bounded set.
Lemma 5: For singular A, there exists 0 r n + N such that a
closed-loop sequence x(k) under terminal constraint MPC starting in
the steerable set remains in a nonempty, bounded, convex subset of n
for times k r .
Lemma 6: Every closed-loop sequence has at least one accumulation point.
Proof: Remark 4 and Lemma 5 ensure that every closed-loop
sequence x(k) is bounded after finite k , and the Bolzano-Weierstrass
property ensures that every infinite sequence in a closed, bounded
subset of n has at least one accumulation point [14].
Let fx(k)g be a closed-loop sequence with an accumulation point
a. Say the closed-loop sequence enters B (a) at time index E +1. Define the sequence fw(k)g to be a portion of the closed-loop sequence,
w (k ) = x(k + E ), k = 1; . . . P . Because a is an accumulation point
of the closed-loop sequence, for every > 0, we can choose E and P
large enough so that w(1); w(P ) 2 B (a) as depicted in Fig. 1. Define
the means over the closed-loop sequence to be
wP
= P1
()
w i uP
i=1
= P1
u
i=1
(w(i)):
f x
Proof: For x(0) = x3 in the MPC control problem, (3), both the initial
and terminal states are equal to x3 . Strict convexity of the cost function
implies the entire trajectory remains at x3 so u0 (x3 ) = u3 .
Definition 2 (Steerable Set): The steerable set N is the set of states
steerable to x3 in N steps
N
2211
wP
(4)
Now consider the MPC optimal open-loop trajectory from each w(k)
for k = 1; . . . ; P . The optimal open-loop input trajectories are
which is created by dropping the first input u0 (0; w), shifting all the
other inputs forward in the sequence, and appending u3 for the final
input. The corresponding trajectory for w(2) is labeled Trajectory 1
in Fig. 1. This input evaluated at state w(k) is admissible for state
w (k + 1) because it satisfies the model, terminal constraint, and input
constraints.2 Its cost from state w(k + 1) is directly related to the optimal cost from w(k) by
2212
(6)
( ()
L w i ;u
i=1
dP
(w(i))) P L(x3 ; u3 )
1 P kw(i) 0 w k2 + u0 (w(i)) 0 u 2
P Q
P R
2 i=1
KL kw(1) 0 w(P + 1)k
+ dP + L(w(P ); u0 (w(P ))) 0 L(x3 ; u3 )
1 P kw(i) 0 w k2 + u0 (w(i)) 0 u 2
P Q
P R
2 i=1
P (L(x3; u3 ) 0 L(wP ; uP ))
+ dP + L(w(P ); u0 (w(P ))) 0 L(x3 ; u3 )
(7)
min
L(x; u)
x;u
x
= Ax + Bu + p
subject to
(9)
kw(i) 0 wP kQ2 +
(w(i)) 0 uP
2
R
2b ( 0 )2 :
and denote this solution as (x3 (p); u3 (p)). From its construction we
know L(x3 (p); u3 (p)) L(wP ; uP ) and therefore
1 P kw(i) 0 w k2 + u0 (w(i)) 0 u 2
P Q
P R
2 i=1
P (L(x3; u3 ) 0 L(x3(p); u3 (p)))
+ dP + L(w(P ); u0 (w(P ))) 0 L(x3 ; u3 )
1 P kw(i) 0 w k2 + u0 (w(i)) 0 u 2
P Q
P R
2 i=1
KL kw(1) 0 w(P + 1)k
+ dP + L(w(P ); u0 (w(P ))) 0 L(x3 ; u3 )
b
2
4 ( 0 ) KL kw(1) 0 w(P + 1)k
+ dP + L(w(P ); u0 (w(P ))) 0 L(x3 ; u3 )
(8)
Two examples are presented to illustrate the advantages of the proposed setpoint tracking MPC framework (sp-MPC) compared to tradi-
2213
tional target tracking MPC (targ-MPC). The regulator cost function for
the new controller, setpoint-tracking MPC (sp-MPC), is
= 12
sp-MPC
01
k ( )0
x j
j =0
xsp
TABLE I
COMPARISON OF CONTROLLER PERFORMANCE (EXAMPLE 1)
2
Q
+ k ( ) 0 k + k ( + 1) 0 ( )k (10)
0, 0 and at least one of
0. This system
in which
can be put in the standard form defined for terminal constraint MPC by
( 0 1)] [3]. The regulator
augmenting the state as ~( ) = [ ( )
u j
Q >
usp
u j
R; S
u j
R; S >
x k
x k
u k
targ-MPC
= 12
01
k ( ) 0 3k
x j
j =0
+ k ( ) 0 3 k + k ( + 1) 0 ( )k
u j
u j
u j
(11)
In the examples, the controller performance is assessed using the following three closed-loop control performance measures:
0
8 ( ) = 1 21 k ( ) 0 k + k ( + 1) 0 ( )k
0
8 ( ) = 1 21 k ( ) 0 k
8( ) = 8 ( ) + 8 ( )
1
u k
kT
u j
y k
kT
usp
u j
u j
j =0
x j
xsp
j =0
u k
y k
in which T is the process sample time. For each of the indices defined
above, we define the percentage improvement of sp-MPC compared to
targ-MPC by
1% = 8
targ-MPC
08
sp-MPC
targ-MPC
2 100
( ) = 60 0+0 592623
2 +1
G s
= 10
0 953
: s
=1
( )=
(12)
1
= 10 = 0
= 80
sampled with T
s. The input u is constrained as juj . The de:
which corresponds to a steady-state
sired output setpoint is ysp
. The regulator parameters are Qy
,R
,
input value of 0 :
0
S
, Q
C Qy C
:
I2 . A horizon length of N
is used. Between times 50130, 200270 and 360430, a state dis0 causes the input to saturate at its lower
: ; :
turbance dx
limit. The output setpoint is unreachable under the influence of this
state disturbance dx . The closed-loop performance of sp-MPC and
targ-MPC under the described disturbance scenario are shown in Fig. 2.
= 0 25
=
+ 0 01
= [17 1 1 77]
( )
The closed-loop performance of the two control formulations are compared in Table I.
2) Comments: In the targ-MPC framework, the controller tries to
reject the state disturbance and minimize the deviation from the new
steady-state target. This requires a large, undesirable control action that
forces the input to move between the upper and lower limits of operation. The sp-MPC framework, on the other hand, attempts to minimize
the deviation from setpoint and subsequently the input just rides the
lower limit input constraint. The benefit here is that the sp-MPC controller slows down the departure from setpoint, but speeds up the approach to setpoint. The traditional targ-MPC can be tuned to be fast or
slow through relative choice of tuning parameters Q and R, but it is fast
or slow from all initial conditions, some of which lead to an approach
setpoint, but others of which lead to a departure from setpoint.
The greater cost of control action in targ-MPC is shown by the cost
index u in Table I. The cost of control action in targ-MPC exceeds
that of sp-MPC by nearly 100%. The control in targ-MPC causes the
output of the system to move away from the (unreachable) setpoint
faster than the corresponding output of sp-MPC. Since the control objective is to be close to the setpoint, this undesirable behavior is eliminated by sp-MPC.
3) Example 2: The second example is the 2 input2 output system
G s
1:5
(s+2)(s+1)
0:5
(s+0:5)(s+1)
0:75
(s+5)(s+2)
2
(s+2)(s+3)
= 0 25
(13)
:
s. The inputs u1 and
The system is sampled at the rate T
are constrained between 0 : and 0.5. The desired output setpoint
0 which corresponds to a steady-state input of
:
; :
is ysp
0 . The state and measurement evolution of the
usp
:
;0 :
system are corrupted by noise. The state and measurement noise covari04 I2 , respectively. In addition to
04 I6 and Rv
ance are Qx
random state noise, a zero-mean square-wave disturbance of magnitude
6 : ; 0 : 0 also affects the state evolution of the system. For this
u2
= [0 337 0 34]
= [0 496 0 468]
= 10
[0 03 0 03]
05
= 10
2214
TABLE II
COMPARISON OF CONTROLLER PERFORMANCE (EXAMPLE 2)
terms of the size of the set of admissible initial states, and the undesirable difference between open-loop prediction and closed-loop behavior
[11]. Here we briefly outline how the previous analysis can be applied
to prove asymptotic stability of terminal penalty MPC. First define the
rotated cost Lr [9]
and compute the infinite horizon rotated cost-to-go under control law
uc (x) = K (x 0 x3 ) + u3 with K chosen so that A = A + BK is
asymptotically stable. A simple calculation gives
L1
r (x ) =
k=0
Lr (x(k); u(k))
5 = A0 5A + Q + K 0 RK
0
= (I 0 A )01 Q(xsp 0 x3 ) + K 0 R(usp 0 u3 ) :
Note that the rotated cost-to-go satisfies Lr1 (x) = Lr1 (Ax +
Buc (x)) + Lr (x; uc (x)). Next define the terminal penalty MPC cost
8r
1
x; fu(i)gN0
i=0
N01
k=0
Notice this inequality is the same as (5) if we replace the cost function
8 in (5) by the rotated cost function 8r . The remaining steps after (5) in
the previous development can then be followed to establish asymptotic
stability of (x3 ; u3 ) for terminal penalty MPC.
If the system is unstable and feasible K does not exist to make (A +
BK ) stable given the active constraints at u3 , constraints to zero the
unstable modes at stage N are added to (14) [15]. The set of admissible
initial states is chosen to ensure positive invariance and feasibility of
control law uc . These plus compactness of ensure system evolution
in a compact set. The user has some flexibility in choosing K , such as
the simple choice K = 0. A procedure to choose feasible K that more
closely approximates the infinite horizon solution is given in [3].
VI. CONCLUSION
1
Lr1 (x) = (x 0 x3 )0 5(x 0 x3 ) 0 0 (x 0 x3 )
2
function as
2215
In this work, a new MPC controller was presented that handles unreachable setpoints better than standard MPC methods. The new controller was based on a cost function measuring distance from the unreachable setpoint, and, for simplicity of exposition, a finite horizon
with a terminal constraint. Nominal asymptotic stability of the optimal
steady state was established for this terminal constraint MPC. Two examples were presented that demonstrate the advantages of the proposed
setpoint-tracking MPC over the existing target-tracking MPC. Cost improvements between 50% and 70% were obtained for the two examples. It was next shown how to extend the asymptotic stability result to
the case of terminal penalty MPC by introducing rotated cost.
This new unreachable setpoint approach should also prove useful in
applications where optimization of a systems economic performance
is a more desirable goal than simple setpoint tracking.
and controller
ACKNOWLEDGMENT
min
8r (x; u)
u
subject to x(0) = x
x(k + 1) = Ax(k) + Bu(k)
k = 0; . . . ; N 0 1
k = 0 ; . . . ; N 0 1:
u (k ) 2
(14a)
(14b)
(14c)
REFERENCES
(14d)
u~ (w) = fu0 (1; w); u0 (2; w); . . . ; u0 (N 0 1; w); uc (x0 (N; w))g
in which x0 (N; w) is the state value at stage N using the optimal input
sequence. Because we use the control law uc (x) in place of the constant
u3 as we did in terminal constraint MPC, and u3 is on the boundary of
, we must restrict the gain matrix K and initial state w such that the
control law is feasible [3] and the system is positive invariant. Computing the change in cost for this candidate sequence gives
The last three terms cancel and noting the optimization at stage k + 1,
we have the inequality