The Maximum Principle: Continuous Time - PowerPoint PPT Presentation

1 / 57
About This Presentation
Title:

The Maximum Principle: Continuous Time

Description:

Necessary conditions for optimization of dynamic systems --General derivation by Pontryagin in 1956-60 ... of time, and S[x,T] is the salvage value, in dollars. ... – PowerPoint PPT presentation

Number of Views:145
Avg rating:3.0/5.0
Slides: 58
Provided by: varghes
Category:

less

Transcript and Presenter's Notes

Title: The Maximum Principle: Continuous Time


1
The Maximum Principle Continuous Time
  • Main purpose to introduce the maximum principle
    as a necessary condition that must be satisfied
    by any optimal control.
  • --Necessary conditions for optimization of
    dynamic systems
  • --General derivation by Pontryagin in 1956-60
  • --A simple (but not completely rigorous) proof
    using dynamic programming
  • -- Examples
  • -- Statement of sufficiency conditions
  • -- Computational method

2
2.1 Statement of the problem
  • Optimal control theory deals with the problem of
  • optimization of dynamic systems
  • 2.1.1 The Mathematical Model
  • State equation
  • where state variables, x(t)? En ,control
    variables,
  • u(t) ? Em , and f En x Em x E1 ?En .
  • f is assumed to be continuously differentiable.
    The
  • path x(t), t ? 0,T is called a state trajectory
    and u(t),
  • t ?0,T is called a control trajectory.

3
2.1.2 Constraints
  • Admissible control u(t), t ?0,T is piecewise
    continuous
  • and satisfies, in addition,
  • 2.1.3 The Objective Function
  • Objective function is defined as follows
  • Where F En x Em x E1 ? E1 and S En x E1 ? E1
    . F and
  • S are assumed to be continuously differentiable.

4
2.1.4 The Optimal Control Problem
  • The problem is to find an admissible control u ,
  • which maximizes the objective function (2.3)
    subject
  • to (2.1) and (2.2). We restate the optimal
    control
  • problem as
  • Control u is called optimal control, x is
    called the optimal trajectory under u, J(u) or
    J will denote optimal value of J.

5
  • Case 1. The optimal control problem (2.4) is said
    to be
  • in Bolza form.
  • Case 2. When S?0, it is said to be in Lagrange
    form.
  • Case 3. When F ?0, it is said to be in Mayer
    form, If
  • F ?0 and S is linear, it is in linear Mayer form,
    i.e.,
  • where c(c1,c2,,cn) ? En.

6
  • The Bolza form can be reduced to the linear Mayer
  • form,we define a new state vector
    y(y1,y2,,yn1),
  • having n1 components defined as followsyixi
    for
  • i1,2,n and
  • so that the objective function is
    Jcy(T)yn1(T). If
  • we now integrate (2.6) from 0 to T, we have
  • which is the same as the objective function J in
    (2.4)

7
Statement of Bellmans Optimality Principle
  • An optimal policy has the property that
    whatever the initial state and initial decision
    are, the remaining decisions must constitute an
    optimal policy with regard to the state resulting
    from the first decision.

8
PRINCIPLE OF OPTIMALITY
  • a
  • c
  • Assertion If abe is the optimal path from a to
    e, then be is
  • the optimal path from b to e.
  • Proof Suppose it is not. Then there is
    another path
  • (note that existence is assumed
    here) bce which
  • is optimal from b to e, i.e.
  • Jbce gt Jbe
  • but then
  • Jabe JabJbelt Jab Jbce
    Jabce
  • This contradicts the
    hypothesis that abe is the
  • optimal path from a to e.

Jbe
Jab
b
Jbce
e
9
A Dynamic Programming Example
  • Stagecoach Problem

Costs
10
Solution
  • Let 1-u1-u2-u3-10 be the optimal path.
  • Let fn(s,un) be the minimal cost path given that
    current
  • state is s and the decision taken is un .
  • fn(s) min f (s,un) min cost(s, un)
    fn1(un)
  • un un
  • This is the Recursion Equation of D.P. It can be
  • solved by a backward procedure which starts at
    the
  • terminal stage and stops at the initial statge.

11
  • Note 1-2-6-9-10 with cost13 is a greedy path
    that
  • minimizes cost at each stage. This may not be
    minimal cost
  • solution, however, E.g. 1-4-6 is cheaper overall
    than 1-2-6.

12
(No Transcript)
13
2.2 Dynamic Programming and the Maximum Principle
  • 2.2.1 The Hamilton Jacobi- Bellman Equation
  • where for s ? t ,
  • Principle of Optimality
  • An optimal policy has the property that, whatever
    the
  • initial state and initial decision are, the
    remaining
  • decision must constitute an optimal policy with
    regard
  • to the outcome resulting from the first decision.

14
Figure 2.1 An Optimal Path in the State-Time
Space
15
The change in the objective function consists of
two parts
  • 1. The incremental changes in J from t to t?t,
    which is
  • given by the integral of F(x,u,t) from t to
    t?t,
  • 2. The value function V(x?x, t?t) at time t?t.
  • In equation form we have
  • Since F is continuous, the integral in (2.9) is
  • approximately F(x,u,t) ?t so that we rewrite
    (2.9) as

16
Assume that V is a continuously differentiable.
Use Taylor series expansion of V with respect to
?t and obtain
  • Substituting for x from (2.1)
  • canceling V(x,t) on both sides and then dividing
    by ?t
  • we get

17
Let ?t ? 0
  • for which the boundary condition is
  • Vx(x,t) can be interpreted as the marginal
    contribution
  • of the state variable x to the maximized
    objective function. Denote it by ?(t) ? En called
    the adjoint (row)
  • vector i.e.,

18
We introduce the so-called Hamiltonian
  • or
  • The (2.14) can rewritten as the following
  • (2.19) will be called the Hamilton-Jacobi-Bellman
  • equation or, HJB equation.

19
From (2.19) we can get Hamiltonian maximizing
condition of the maximum principle
  • canceling the term Vt on both sides, we obtain
  • for all u ? ?(t).
  • Remark H decouples the problem over time by
  • means of ?(t), which is analogous to dual
    variables or
  • shadow prices in Linear Programming.

20
2.2.2 Derivation of the Adjoint Equation
  • Let
  • where for a small
    positive ?.
  • Fix t and use H-J-B equation (2.19)
  • LHS0 from (2.19) since u maximizes HVt. RHS
  • will be zero if u(t) also maximizes HVt with
    x(t) as
  • the state. In general x(t) ? x(t), thus RHS ? 0,
    But then
  • RHSx(t)x(t) 0 ?RHS is maximized at x(t)x(t).

21
Since x(t) is unconstrained, we have?RHS/ ?x
x(t)x(t) 0
  • or,
  • By definition of H
  • Note (2.25) assumes V to be twice continuously
  • differentiable.

22
By definition (2.16) of ?(t)
23
Using (2.25) and (2.26) we have
  • Using (2.16), we have
  • From (2.18), we have

24
Terminal boundary condition or transversality
condition
  • (2.28) and (2.29) can determine the adjoint
    variables.
  • From (2.28), we can rewrite the state equation as
  • From (2.28), (2.29), (2.30) and (2.1), we get
  • (2.31) is called a canonical system of equation
    or
  • canonical adjoints.

25
Free and Fixed End Point Problems for fixed
Terminal time T
  • (1) x(T) is free ? ?(T) Sx
  • (2) x(T) is fixed ? ? (T) some constant
  • (given)
  • (3) x(T) is lying on Mx(T),T0
  • ? ?(T) Sx MxT ? and Mx(T),T0,
  • where ? is the Lagrange vector.

26
Free and Fixed End Point Problems for free
Terminal time T
(1) x(T) is free ? ?(T) Sx HSt0
at tT. (2) x(T) is fixed ? HSt0 at
tT. (3) x(T) is lying on Mx(T),T0 ?
?(T) Sx MxT ? , HSt0 at
tT, and Mx(T),T0, where
? is the Lagrange vector.
27
2.2.3 The Maximum Principle
  • The necessary conditions for u to be an optimal
  • control are

28
2.2.4 Economic Interpretation of the Maximum
Principles
  • where F is considered to be the instantaneous
    profit rate measured in dollars per unit of
    time, and Sx,T is the salvage value, in
    dollars.
  • Multiplying (2.18) formally by dt and from (2.1),
    we
  • have
  • F(x,u,t)dt direct contribution to J in from t
    to tdt.
  • ?dx indirect contribution to J in dollars.
  • Hdt total contribution to J from time t to tdt
    when
  • x(t)x and u(t) u in the interval t,tdt.

29
By (2.28) and (2.29) we have
  • Rewriting the first equation as
  • -d? marginal cost of holding capital from t to
    tdt
  • Hxdt marginal revenue of investing the capital
  • Fxdt direct marginal contribution
  • ?fxdt indirect marginal contribution
  • Thus, adjoint equation implies ? MC MR

30
Example 2.1 Consider the problem
  • subject to the state equation
  • and the control constraint
  • Note that T1, F-x, S0, and f u. Because F-x,
    we
  • can interpret the problem as one of minimizing
    the
  • (signed) area under the curve x(t) for 0?t ? 1.

31
Solution.First we form the Hamiltonian
  • and note that, because the Hamiltonian is linear
    in u,
  • the form of the optimal control, i.e., the one
    that would
  • maximize the Hamiltonian, is
  • or referring to the notation in Section 1.4,

32
To find ?, we write the adjoint equation
  • Because this equation does not involve x and u,
    we
  • can easily solve it as
  • It follows that ?(t) t-1 ? 0 for all t?0,1,
    and since we
  • can set u(1)-1, which defines u at the single
    point
  • t1, we have the optimal control

33
Substituting this into the state equation(2.34)
we havewhose solution is
  • The graphs of the optimal state and adjoint
    trajectories
  • appear in Figure 2.2. Note that the optimal value
    of the
  • objective function is J -1/2.

34
Figure 2.2 Optimal State and Adjoint Trajectories
for Example 2.1
35
Example 2.2 Let us solve the same problem as in
example 2.1 over the interval 0,2 so that the
objective is to
  • The dynamics and constraints are (2.33) and
    (2.34),
  • respectively, as before. Here we want to minimize
    the
  • signed area between the horizontal axis and the
  • trajectory of x(t) for 0?t?2.

36
Solution. As before the Hamiltonian is defined by
(2.36) and the optimal control is as in (2.38).
The adjoint equation
  • is the same as (2.39) except that now T2 instead
    of
  • T1. The solution of (2.44) is easily found to be
  • Hence the state equation (2.41) and its solution
    (2.42)
  • are exactly the same. The graphs of the optimal
    state
  • and adjoint trajectories appear in Figure 2.3.
    Note
  • that the optimal value of the objective function
    here is
  • J0.

37
Figure 2.3 Optimal State and Adjoint Trajectories
for Example 2.2
38
Example 2.3 The next example is
  • subject to the same constraints as in Example
    2.1,
  • namely,
  • Here F - (1/2)x2 so that the interpretation of
    the
  • objective function (2.46) is that we are trying
    to find
  • the trajectory x(t) in order that the area under
    the
  • curve (1/2)x2 is minimized.

39
Solution. The Hamiltonian is
  • which is linear in u so that the optimal policy
    is
  • The adjoint equation is
  • Here the adjoint equation involves x so that we
  • cannot solve it directly. Because the state
    equation
  • (2.47) involves u, which depends on ?, we also
  • cannot integrate it independently without knowing
    ?.

40
  • The way out of this dilemma is to use some
    intuition.
  • Since we want to minimize the area under (1/2)x2
    and
  • since x(0)1, it is clear that we want x to
    decrease as
  • quickly as possible. Let us therefore temporaily
  • assume that ? is nonpositive in the interval
    0,1 so
  • that from (2.49) we have u-1 throughout the
    interval. (In Exercise 2.5, you will be asked to
    show that this
  • assumption is correct.) With this assumption, we
    can
  • solve (2.47) as

41
  • Substituting this into (2.50) gives
  • Integrating both sides of this equation from t
    to1 gives
  • or
  • which, using ?(1)0, yields

42
  • The reader may now verify that ?(t) is
    nonpositive in
  • the interval 0,1, verifying our original
    assumption.
  • Hence (2.51) and (2.52) satisfy the necessary
  • conditions. In Exercise 2.6, you will be asked to
    show
  • that they satisfy sufficient conditions derived
    in Section
  • 2.4 as well, so that they are indeed optimal.
    Figure 2.4
  • shows the graphs of the optimal trajectories.
  • It would be clearly optimal if we could keep
    x(t)0, t 1.
  • This is possible by setting

43
Figure 2.4 Optimal Trajectories for Example 2.3
and Example 2.4
44
Example 2.4 Let us rework Example 2.3 with T2,
i.e,
  • Subject to the constraints

It would be clearly optimal if we could keep
x(t)0, t 1. This is possible by setting
1 t 1 u(t) -1 t
1. Note u(t)0, t 1 is singular control.
45
Example 2.5 The problem is
46
Solution
47
Figure 2.5 Optimal control for Example 2.5
48
2.4 Sufficient Conditions
  • Since either or or both.

49
  • Theorem 2.1 (Sufficiency Conditions). Let u(t),
    and
  • the corresponding x(t) and ?(t) satisfy the
    maximum
  • principle necessary condition (2.32) for all
    t?0,T.
  • Then, u is an optimal control if H0(x,?(t),t) is
    concave
  • in x for each t and S(x,T) is concave in x.

50
Example 2.6 Examples 2.1 and 2.2 satisfy the
sufficient conditions. Fixed-end-point problem
Transversality condition is
51
2.5 Solving a TPBVP by Using Spreadsheet Software
  • Example 2.7 Consider the problem

52
  • Let ?t0.01, initial value ?(0)-0.2, x(0)5.

53
(No Transcript)
54
(No Transcript)
55
(No Transcript)
56
(No Transcript)
57
(No Transcript)
Write a Comment
User Comments (0)
About PowerShow.com