1 Introduction
Reference e-book: An Introduction to Mathematical Optimal Control Theory by Lawrence C. Evans in 2024
1.1 basic problem
DYNAMICS: Consider an ODE:
: initial point : dynamic function : unknown; dynamic evolution of the state of some "system"
CONTROLLED DYNAMICS: Generalize a bit,
Change the value
Then the dynamic equation becomes:
We call a function
Notation
Introduce:
To denote the collection of all admissible controls
would be more precise
Payoffs
Overall task will be to determine what is the "best" control for our system. For this we need to specify a specific payoff (or reward) criterion. Let us define the payoff functional
where:
solves ODE for the control : Given; runnig payoff : Given; terminal payoff; : Given; terminal time
The basic problem
Aim: find a control
for all controls
This task presents us with these mathematical issues:
- Does an optimal control exist?
- How can we characterize an optimal control mathematically?
- How can we construct an optimal control?
These turn out to be sometimes subtle problems, as the following collection of examples illustrates.
1.2 Examples
Example 1: Control of Production and Consumption
Suppose we own, say, a factory whose output we can control. Let us begin to construct a mathematical model by setting
We suppose that we consume some fraction of our output at each time, and likewise can reinvest the remaining fraction. Let us denote:
This will be our control, and s.t. the obvious constraint that:
The corresponding dynamic equation is:
The constant
Take as a payoff functional:
That means we want to maximize our total consumption of the output, our consumption at a given time t being
In 4.4.2, we will see that the optimal control is
In other words, we should reinvest all the output (and therefore consume nothing) up until time
Example 2: Reproduction Strategies in Social Insects
In this example, we consider a population of social insects, a population if bees. Write
number of workers at time number of queens fraction of colony effort devoted to increasing work force
Constraint of
Introduce the dynamic for the numbers of workers and the number of queens:
- workers:
:is the death rate of workers; given constant : the known rate at which each worker contributes to the bee economy
- queens:
constant
Goal: maximize the queens at time
We have
answer will again turn out to be a bang–bang control
Example 3: A Pendulum
A hanging pendlum:
If no external force:
The solution will be a damped oscillation, provided
Let
Our dynamics now become
Define
We introduce as well
for
Maximize
The terminal time isn't fixed, but rather depends upon the control. This's a fixed endpoint, free time problem.
Example 4: A Moon Lander
This model asks us to bring a spacecraft to a soft landing on the lunar surface, using the least amount of fuel.
Introduce the notation:
Notation
: height at time : velocity : mass of spacecraft at time (changing as fuel is burned) : thrust at time t, assumed that
For Newton's law:
Modelled by ODE:
We want to minimize the amount of fuel used up, that is, to maximize the amount remaining once we have landed. Thus
where
We have also the extra constraints
Example 5: Rocket Railroad Car
Imagine a railroad car powered by rocket engines on each side. We introduce the variables:
: position at time : velocity at time : thrust from rockets at time , assumed that
We want to figure out how to fire the rockets, so as to arrive at the origin 0 with zero velocity in a minimum amount of time. Assuming the car has mass
Rewrie by setting
Take
for
1.3 A geometric solution
Introduce some ad hoc calculus and geometry methods for the rocke car problem.
First of all, let us guess that to find an optimal solution we will need only to consider the cases
CASE 1:
Then
And so
Let
belong to the time interval where and interate from to : Then
In other words, so long as the control is set for
, the trajectory stays on the curve for some constant . CASE 2:
Then
Let
As long as the control is set for
Geometric interpretation
Now we can design an optimal control