The Maximum Principle: Continuous Time Main purpose: to introduce the maximum principle as a necessary condition that must be satisfied by any optimal.

Download Report

Transcript The Maximum Principle: Continuous Time Main purpose: to introduce the maximum principle as a necessary condition that must be satisfied by any optimal.

The Maximum Principle: Continuous Time Main purpose

: to introduce the maximum principle as a necessary condition that must be satisfied by any optimal control.

--Necessary conditions for optimization of dynamic systems --General derivation by Pontryagin in 1956-60 --A simple (but not completely rigorous) proof using dynamic programming -- Examples -- Statement of sufficiency conditions -- Computational method

2.1 Statement of the problem

Optimal control theory deals with the problem of optimization of dynamic systems 2.1.1

The Mathematical Model

State equation where state variables,

x

(

t

) 

u

(

t

) 

E m

, and

f

:

E n E n

,

control variables

,

x E m x E 1

E n

.

f

is assumed to be continuously differentiable. The

t

path

x

(t),

t

[0,T]

[0,T]

is called a is called a

state trajectory control trajectory

.

and

u

(

t

),

2.1.2 Constraints

Admissible control

u(t), t

[0,T]

is piecewise continuous and satisfies, in addition,

2.1.3 The Objective Function

Objective function is defined as follows Where

F

:

E n x E m x E 1

E 1

and

S: E n x E 1

E 1

.

F

and

S

are assumed to be continuously differentiable.

2.1.4 The Optimal Control Problem

The problem is to find an admissible control

u

* , which maximizes the objective function (2.3) subject to (2.1) and (2.2). We restate the optimal control problem as:

Control u

* is called

optimal control

,

x

* is called the

optimal trajectory

optimal value of

J

.

under

u

*,

J

(

u

*) or

J

* will denote

Case 1. The optimal control problem (2.4) is said to be in

Bolza form

.

Case 2. When

S

 0, it is said to be in

Lagrange form

.

Case 3. When

F F

 0 and

S

 0, it is said to be in is linear, it is in

Mayer form linear Mayer form

, i.e., , If where

c=(c 1 ,c 2 ,…,c n )

E n

.

The Bolza form can be reduced to the linear Mayer form , we define a new state vector

y=

(

y 1 ,y 2 ,…,y n+1

)

,

having n+1 components defined as follows:

y i =x i i=1,2…,n

and for so that the objective function is

J=cy(T)=y n+1 (T).

If we now integrate (2.6) from 0 to

T

, we have which is the same as the objective function

J

in (2.4)

Statement of Bellman’s Optimality Principle

“An optimal policy has the property that whatever the initial state and initial decision are, the remaining decisions must constitute an optimal policy with regard to the state resulting from the first decision.”

PRINCIPLE OF OPTIMALITY

J ab J be

a b

J bce

e c

Assertion: If

abe

is the optimal path from

a

to

e

, then be is the optimal path from b to

e

.

Proof: Suppose it is not. Then there is another path (note that existence is assumed here)

bce

which is optimal from

b

to

e

, i.e.

J

bce but then >

J

be

J

abe =

J

ab +

J

be <

J

ab +

J

bce =

J

abce This contradicts the hypothesis that

abe

is the optimal path from

a

to

e

.

A Dynamic Programming Example Stagecoach Problem Costs:

Solution:

Let 1-

u 1 -u 2 -u 3

-10 be the optimal path.

Let

f n

(

s,u n

) be the minimal cost path given that current state is

s

and the decision taken is

u n

.

f

n *(

s

) = min

f

(s,

u n

) =

f n

(

s,u n

*)

u n

This is the Recursion Equation of D.P. It can be solved by a backward procedure – which starts at the terminal stage and stops at the initial statge.

Note: 1-2-6-9-10 with cost=13 is a greedy path that minimizes cost at each stage. This may not be minimal cost solution, however, E.g. 1-4-6 is cheaper overall than 1-2-6.

2.2 Dynamic Programming and the Maximum Principle 2.2.1 The Hamilton – Jacobi- Bellman Equation

where for

s

t ,

Principle of Optimality

An optimal policy has the property that, whatever the initial state and initial decision are, the remaining decision must constitute an optimal policy with regard to the outcome resulting from the first decision.

Figure 2.1 An Optimal Path in the State-Time Space

The change in the objective function consists of two parts:

1. The incremental changes in

J

from t to

t

+ 

t

, which is given by the integral of

F

(

x,u,t

) from t to

t

+ 

t

, 2. The value function

V(x+

x, t+

t)

at time

t

+ 

t

.

In equation form we have Since

F

is continuous, the integral in (2.9) is approximately

F

(

x,u,t

) 

t

so that we rewrite (2.9) as

Assume that Taylor series expansion of obtain

V V

with respect to 

t

and Substituting for

x

from (2.1) canceling

V

(

x,t

) on both sides and then dividing by 

t

we get

Let 

t

 0 with the boundary condition

V x

(

x,t

) can be interpreted as the marginal contribution of the state variable

x

function. Denote it by to the maximized objective  (

t

) 

E n

called the a

djoint

(row) vector i.e.,

We introduce the so-called

Hamiltonian

or The (2.14) can rewritten as the following (2.19) will be called the

Hamilton-Jacobi-Bellman equation

or, HJB equation.

From (2.19) we can get Hamiltonian maximizing condition of the maximum principle canceling the term

V t

on both sides, we obtain for all

u

  (

t

).

Remark: H

decouples the problem over time by means of  (

t

), which is analogous to dual variables or shadow prices in Linear Programming.

2.2.2 Derivation of the Adjoint Equation Let where for a small positive  .

Fix

t

and use H-J-B equation (2.19) LHS=0 from (2.19) since

u*

maximizes [

H+V t

]. RHS will be zero if

u*(t)

also maximizes

H+V t

the state. In general

x(t)

x*(t),

with thus RHS 

x

(

t

) as 0, But then RHS|

x(t)=x*(t) =

0  RHS is maximized at

x(t)=x*(t).

Since

x(t)

is unconstrained, we have 

RHS

/  x |

x(t)=x*(t)

= 0 or, By definition of

H

Note: (2.25) assumes

V

to be twice continuously differentiable.

By definition (2.16) of  (

t

)

Using (2.25) and (2.26) we have Using (2.16), we have From (2.18), we have

Terminal boundary condition or transversality condition:

(2.28) and (2.29) can determine the adjoint variables.

From (2.28), we can rewrite the state equation as From (2.28), (2.29), (2.30) and (2.1), we get (2.31) is called

a canonical system of equation or canonical adjoints

.

Free and Fixed End Point Problems

S

 0   (

T

) = 0 no salvage value

x

(

T

) fixed   (

T

) = some constant (given)

2.2.3 The Maximum Principle

The necessary conditions for

u

* to be an optimal control are:

2.2.4 Economic Interpretation of the Maximum Principles

where

F

is considered to be the instantaneous profit rate measured in dollars per unit of time, and

S

[

x,T

] is the salvage value, in dollars.

Multiplying (2.18) formally by

dt

and from (2.1), we have

F

(

x,u,t

)

dt

: direct contribution to

J

dx

: indirect contribution to

J

in

$

from

t

in dollars.

to

t+dt

.

Hdt

: total contribution to

J

from time t to t+dt when

x

(

t

)=

x

and

u

(

t

) =

u

in the interval [

t

,

t+dt

].

By (2.28) and (2.29) we have Rewriting the first equation as

-d

 : marginal cost of holding capital from

t

to

t+dt Hxdt

: marginal revenue of investing the capital

Fxdt

: direct marginal contribution 

f x dt

: indirect marginal contribution Thus, adjoint equation implies  MC = MR

Example 2.1

Consider the problem subject to the state equation and the control constraint Note that

T

=1,

F=-x

,

S

=0, and

f =u

. Because

F=-x

, we can interpret the problem as one of minimizing the (signed) area under the curve

x

(

t

) for 0 

t

 1.

Solution

.First we form the Hamiltonian and note that, because the Hamiltonian is linear in

u

, the form of the optimal control, i.e., the one that would maximize the Hamiltonian, is or referring to the notation in Section 1.4,

To find  , we write the adjoint equation Because this equation does not involve

x

and

u

, we can easily solve it as It follows that  (

t

) =

t-

1  0 for all

t

 [0,1], and since we can set

u

*(1)=-1, which defines u at the single point

t

=1, we have the optimal control

Substituting this into the state equation(2.34) we have whose solution is The graphs of the optimal state and adjoint trajectories appear in Figure 2.2. Note that the optimal value of the objective function is

J

*= -1/2.

Figure 2.2 Optimal State and Adjoint Trajectories for Example 2.1

Example 2.2

objective is to Let us solve the same problem as in example 2.1 over the interval [0,2] so that the The dynamics and constraints are (2.33) and (2.34), respectively, as before. Here we want to minimize the signed area between the horizontal axis and the trajectory of

x

(t) for 0 

t

 2.

Solution

. As before the Hamiltonian is defined by (2.36) and the optimal control is as in (2.38). The adjoint equation is the same as (2.39) except that now

T

=2 instead of

T

=1. The solution of (2.44) is easily found to be Hence the state equation (2.41) and its solution (2.42) are exactly the same. The graphs of the optimal state and adjoint trajectories appear in Figure 2.3. Note that the optimal value of the objective function here is

J

*=0.

Figure 2.3 Optimal State and Adjoint Trajectories for Example 2.2

Example 2.3 The next example is:

subject to the same constraints as in Example 2.1, namely, Here

F

= - (1/2)

x

2 so that the interpretation of the objective function (2.46) is that we are trying to find the trajectory

x

(

t

) in order that the area under the curve (1/2)

x

2 is minimized.

Solution.

The Hamiltonian is which is linear in

u

so that the optimal policy is The adjoint equation is Here the adjoint equation involves

x

so that we cannot solve it directly. Because the state equation (2.47) involves

u

, which depends on  , we also cannot integrate it independently without knowing  .

The way out of this dilemma is to use some intuition. Since we want to minimize the area under (1/2)

x

2 and since

x

(0)=1, it is clear that we want

x

to decrease as quickly as possible. Let us therefore temporaily assume that  is nonpositive in the interval [0,1] so that from (2.49) we have

u

=-1 throughout the interval. (In Exercise 2.5, you will be asked to show that this assumption is correct.) With this assumption, we can solve (2.47) as

Substituting this into (2.50) gives Integrating both sides of this equation from

t

to1 gives or which, using  (1)=0, yields

The reader may now verify that  (t) is nonpositive in the interval [0,1], verifying our original assumption. Hence (2.51) and (2.52) satisfy the necessary conditions. In Exercise 2.6, you will be asked to show that they satisfy sufficient conditions derived in Section 2.4 as well, so that they are indeed optimal. Figure 2.4 shows the graphs of the optimal trajectories.  It would be clearly optimal if we could keep x*(t)=0, t ≥1.

 This is possible by setting

Figure 2.4 Optimal Trajectories for Example 2.3 and Example 2.4

Example 2.4

Let us rework Example 2.3 with

T

=2, i.e, Subject to the constraints: It would be clearly optimal if we could keep

x

*(

t

)=0,

t

≥1. This is possible by setting: 1

t

≤1

u

*(

t

)= -1

t

≥1.

Note

u

*(

t

)=0,

t

≥1 is singular control.

Example 2.5

The problem is:

Solution

:

Figure 2.5 Optimal control for Example 2.5

2.4 Sufficient Conditions

Since either or or both.

Theorem 2.1 (Sufficiency Conditions). Let u*(t), and

the corresponding x*(t) and

(t) satisfy the maximum principle necessary condition (2.32) for all t

[0,T]. Then, u* is an optimal control if H 0 (x,

(t),t) is concave in x for each t and S(x,T) is concave in x.

Example 2.6:

Examples 2.1 and 2.2 satisfy the sufficient conditions.

Fixed-end-point

problem: Transversality condition is:

2.5 Solving a TPBVP by Using Spreadsheet Software

Example 2.7

Consider the problem:

Let:  t=0.01, initial value  (0)=-0.2, x(0)=5.