Optimal control
Simplified case : Linear system, quadratic performance index, fixed horizon and final state
Contents
- Problem :
- Solution using Pontryagin’s minimum principle :
- Exercises
Problem :
Let us consider the linear system :
From a given initial state , the objective is to bring back the state to 0 within a given time horizon
(
) while minimizing the quadratic performance index :
where and
are given weighting matrices with
and
.
Solution using Pontryagin’s minimum principle :
- The Hamiltonian reads :
where is the costate vector.
- the optimal control minimizes
:
- Costate dynamics :
- State-costate dynamics : (1), (2) and (3) leads to :
with
is the
Hamiltonian matrix associated to such a control problem. (4) can be intregrated taken into account boundary conditions on the state-costate augmented vector
:
- initial conditions on
:
(5),
- terminal conditions on
:
(6).
The set of equations (4), (5) and (6) is also called a two point boundary-value problem.
- Integration of the two point boundary-value problem :
where ,
are the 4
submatrices partionning
(WARNING !! :
).
Then one can easily derive the initial value of the costate :
where depends only on the problem data :
,
,
,
,
and not on
.
- Optimal control initial value : from equation (2) :
- Closed-loop optimal control at any time
: at time
, assuming that the current state
is known (using a measurement system), the objective is still to bring back the final state to
(
) but the time horizon is now
. The calculus of the current optimal control
is the same problem than the previous one, just changing
by
and
by
. Thus :
with :
the time-varying state feedback to be implemented in closed-loop according to the following Figure :
Remark : is not defined since
and is not invertible.
- Optimal state trajectories : The integration of equation (4) between 0 and
(
) leads to (first
row) :
where :
is called the transition matrix.
- Optimal performance index :
For any and a current state
one can define the cost-to-go function (or value-function)
as :
and the optimal cost-to-go function as :
From equation (4) : one can derive that :
Thus (after simplification) :
Thus :
From this last equation, on can find again the definition of the costate used to solve the Hamilton–Jacobi–Bellman equation ; i.e. : the gradient of the optimal cost-to-go function w.r.t.
:
The optimal performance index is : .
Exercises
- Exo #1 : show that
is the solution of the matrix Riccati differential equation :
also written as :
- Exo #2 : considering now that
, compute the time-variant state feedback gain
and the time-variant feedforward gain
of the optimal closed-loop control law to be implemented according to the following Figure.