Optimal Control of Systems
Given a system: π₯ = π(π₯, π’, π);
;
; x(0) = x0
Find the control u(t) for t e [0,T] such that
Instantaneous
(Stage) Cost
Terminal Cost
π
π’ = arg lim πΏ π₯, π’ ππ‘ + π π₯ π , π’ π
π’πβ¦
0
π½(π₯, π’)
β’ Can include constraints on control u and state x
β (along trajectory or at final time):
β’ Final time T may or may not be free (weβll first derive fixed T case)
1
Function Optimization
β’ Necessary condition for optimality is that gradient is zero
β Characterizes local extrema; need to check sign of second derivative
β Need convexity to guarantee global optimum
Keep this analogy in mind:
at an extremum
lim π π₯ β + βπ₯ β π(π₯ β ) = 0
βπ₯β0
Gradient of f(x)
orthogonal to level sets
lim π
βπ₯β0
π₯β
ππ(π₯ β )
+
βπ₯ + β¦ β π(π₯ β ) = 0
ππ₯
2
Constrained Function optimization
β’ At the optimal solution, gradient of
F(x) must be parallel to gradient of
G(x):
Consider:
Constraint G(x)=0
Constraint G(x)=0
x*
Level sets of F(x)
Level sets of F(x)
Constrained Function optimization
β’ At the optimal solution, gradient of
F(x) must be parallel to gradient of
G(x):
β’ More generally, define:
Constraint G(x)=0
Lagrange
multiplier
β’ Then a necessary condition is:
β’ The Lagrange multipliers ο¬ are the
sensitivity of the cost to a change in G
x*
Level sets of F(x)
Solution approach
β’ Add Lagrange multiplier ο¬(t) for dynamic constraint
β And additional multipliers for terminal constraints or state constraints
β’ Form augmented cost functional:
β’ where the Hamiltonian is:
β’ Necessary condition for optimality:
vanishes for any
perturbation (variation) in x, u, or ο¬ about optimum:
x(t) = x*(t) + dx(t);
u(t) = u*(t) + du(t);
ο¬(t) = ο¬*(t) + dο¬ (t);
Variations must satisfy path end conditions!
βvariationsβ
variation
dc(tβ)
dc(t)={dx(t), du(t), dο¬(t)}
xT=c(T)
c(tβ)
={x(t),u(t),ο¬(t)}
x0=c(0)
6
Derivationβ¦
β’ Note that (integration by parts):
β’ So:
We want this to be stationary for all variations
Pontryaginβs Maximum Principle
β’ Optimal (x*,u*) satisfy:
Optimal control
is solution to
O.D.E.
Unbounded
controls
β’
β’ Can be more general and include terminal constraints
β’ Follows directly from:
0
0
0
0
Interpretation of ο¬
β’
β’
β’
β’
Two-point boundary value problem: ο¬ is solved backwards in time
ο¬ is the βco-stateβ (or βadjointβ variable)
Recall that H = L(x,u) + ο¬Tf(x,u)
If L=0, ο¬(t) is the sensitivity of the cost to a perturbation in state x(t)
β In the integral as
β Recall dJ = β¦ +ο¬(0)dx(0)
Terminal Constraints
Assume q terminal constraints of the form: οΉ(x(T))=0
β’ Then
ο¬(T) =
ππ
ππ₯
x T
+
ποΉ
ππ₯
(x(T)) π£
β’ Where n is a set of undetermined Lagrange Multipliers
β’ Under some conditions, n is free, and therefore ο¬(T) is free as well
When the final time T is free (i.e., it is not predetermined), then the cost function
J must be stationary with respect to perturbations T: T* + dT. In this case:
H(T) = 0
11
General Remarks
12
Example: Bang-Bang Control
βminimum time controlβ
H
-1
ο¬T B>0
+1
u
Since H is linear w.r.t. u,
minimization occurs at boundary
13
Linear system, Quadratic cost
β’ Apply PMP:
BT selects the part of the state
that is influenced by u, so BTο¬ is
sensitivity of aug. state cost to u
β’ Guess that ο¬(t)=P(t)x(t):
β’ xTPx has an interpretation as the βcost to goβ
β’ Often see the infinite-horizon solution where dP/dt=0
© Copyright 2026 Paperzz