Skip to content

2 Controllablity, bang-bang control

2.1 Definitions

Controllability question

Given the initial point x0 and a “target” set SRn, does there exist a control steering the system to S in finite time?

For the time being we will therefore not introduce any payoff criterion that would characterize an “optimal” control, but instead will focus on the question as to whether or not there exist controls that steer the system to a given goal. In this chapter we will mostly consider the problem of driving the system to the origin S={0}.

Definition

Definition: reachable set

  • reachable set for time t: C(t)= set of initial points x0 for which there exists a control such that x(t)=0

  • reachable set: C= set of initial points x0 for which there exists a control such that x(t)=0 for some finite time t;

    C=t0C(t)

Let Mn×m denote the set of all n×m matrices. Assume that this and next chapter, the ODE is linear in both the state x() and the control α(), and he ODE hasthe form

{x˙(t)=Mx(t)+Nα(t)x(0)=x0

where MMn×n and NMn×m. Assume the set A of conrol parameters is a cube in Rm:

A=[1,1]m={aRm||ai|1,i=1,,m}

2.2 Quick review of linear ODE

Definition: fundamental solution

Let X():RMn×n be the unique solution of the matrix ODE:

{X˙(t)=MX(t)(tR)X(0)=I.

We call X(t) the fundamental solution and sometimes write

X(t)=etM:=t=0tkMkk!

Last formula being the definition of the exponential etM and observe that

X1(t)=X(t)

Theorem 2.1: Solving linear systems of ODE

  1. The unique solution of the homogeneous system of ODE

    {x˙(t)=Mx(t)x(0)=x0.

is

x(t)=X(t)x0=etMx0
  1. The unique solution of the nonhomogeneous system
{x˙(t)=Mx(t)+f(t)x(0)=x0.

is

x(t)=X(t)x0+X(t)0tX1(s)f(s)ds

This expression is the variation of parameters formula.

2.3 Controllability of linear equations

According to the variation of parameters formula, the solution of (linear ODE) for a given control α(·) is

x(t)=X(t)x0+X(t)0tX1(s)Nα(s)dsx0C(t)there exists a control α()A s.t. x(t)=00=X(t)x0+X(t)0tX1(s)Nα(s)ds for some control α()Ax0=0tX1(s)Nα(s)ds for some control α()A

Theorem 2.2: Structure of reachable set

  1. The reachable set C is symmetric and convex.
  2. Also, if x0C(t¯), then x0C(t) for all times tt¯

Definition

Definition: symmetric & convex

  1. A set S is symmetric if xS implies xS
  2. The set S is convex if x,x^S and 0λ1 imply λx+(1λ)x^S

Proof of theorem 2.2:

  1. (Symmetric) Let t0 and x0C(t). Then x0=0tX1(s)Nα(s)ds for some admissible control αA.

Therefore x0=0tX1(s)N(α(s))ds and α(s)A since set A is symmetric

Therefore x0C(t), and so each set C(t) symmetric. It follows that C is symmetric

  1. (Convexity) Take x0,x^0C so that x0C,x^0C(t^) for appropriate time t,t^0. Assume tt^. Then
x0=0tX1(s)Nα(s)ds for some control α()Ax^0=0t^X1(s)Nα^(s)ds for some control α^()A

Define a new control

α~(s):={α(s)if 0st0if s>t

Then

x0=0t^X1(s)Nα~(s)ds

and hence x0C(t^). Now let 0λ1, and observe

λx0+(1λ)x^0=0t^X1(s)N(λα~(s)+(1λ)α^(s))ds

Therefore λx0+(1λ)x^0C(t^)C

  1. Assertion (ii) follows from the foregoing if we take t¯=t^.

A simple example

Let n=2 and m=1,A=[1,1], and write x(t)=(x1(t),x2(t))T. Suppose

{x˙1=0x˙2=α(t).

This is a system of the form x˙=Mx+Nα, for

M=(0000),N=(01)

Clearly C={(x1,x2)x1=0}, the x2-axis.

We next wish to establish some general algebraic conditions ensuring that C contains a neighborhood of the origin

Controllability

Definition: controllability matrix

The controllability matrix is

G=G(M,N):=[NMNM2NMn1N]n×(mn) matrix

Theorem 2.3: Controllability matrix

rankG=n0C

Notation

  • C: the interior of the set C, with its own and neighbor fields in the set
  • rank of G = number of linearly independent rows / columns of G; rankGn

Proof:

  1. Suppose 1st that rankG<n. This means that the linear span of the columns of G has dimension less than or equal to n1. Thus there exists a vector bRn,b0, orthogonal to each column of G. This implies bG=0. So
bN=bMN==bMn1N=0
  1. In fact, bMkN=0,kR+ To confirm this, recall that
p(λ):=det(λIM)

is the characteristic polynomial of M. The Cayley–Hamilton Theorem states that p(M)=0 So if we write

p(λ)=λn+βn1λn1++β1λ1+β0

then

p(M)=Mn+βn1Mn1++β1M+β0I=0

Therefore

Mn=βn1Mn1βn2Mn2β1Mβ0I

and so

bMnN=b(βn1Mn1)N=0

Similarly, bMn+1N=0, etc.

Now notice that

bX1(s)N=besMN=bTk=0(s)kMkNk!=k=0(s)kk!bTMkN=0
  1. Assume next that x0C(t). This is equivalent to having
x0=0tX1(s)Nα(s)ds for some control α()A

Then

bx0=0tbX1(s)Nα(s)ds=0

This says that b is orthogonal x0. In other words, C must lie in the hyperplane orthogonal to b0. Consequently C=ϕ.

(How to understand: in the hyperplane, there is no hypersphere in the set)

  1. Conversely, assume that 0C. Thus 0C(t),t>0. Since C(t) is convex, there exits a support hyperplane to C(t) through 0(This hyperplane put the set into just one side, and 0 is not in te interior, so can do this). This means that b0, s.t. bx00,x0C(t)

(An equation for hyperplane that crosses thre origin is bx=0)

Choose any x0C(t). Then

x0=0tX1(s)Nα(s)ds

for some control α, and therefore

0bx0=0tbX1(s)Nα(s)ds

Thus

0tbX1(s)Nα(s)ds0 for all controls α()

We assert that therefore

bX1(s)N0

a proof of which follows as a lemma below. We rewrite it as

besMN0

Let s=0 to see that bN=0. Next differentiate it with respect to s, to find that

b(M)esMN0

For s=0 this says

bMN=0

We repeatedly differentiate, to deduce

bMkN=0,=0,1,

and so bG=0. This implies rankG<n, since b0.

Lemma 2.4: Integral inequalities

Assume that

0tbX1(s)Nα(s)ds0

for all controls α(). Then

bX1(s)N0

Proof: Replacing α with α, we see that

0tbX1(s)Nα(s)ds=0

for all controls α().

Define

v(s):=bX1(s)N

If v0, then v(s0)0 for some s0. Then there exists an interval I s.t. s0I and v(s)0 on I. Now define α()A this way:

{α(s)=0,(sI)α(s)=v(s)|v(s)|1n,(sI)

Then

0=0tv(s)α(s)ds=Iv(s)nv(s)|v(s)|ds=1nI|v(s)|ds

This implies the contradiction that v0 in I.

Definition: controllable

We say the linear system (ODE) is controllable if C=Rn

Theorem 2.5: Criterion for controllability

Let A be the cube [1,1]n in Rn. Suppose as well that rankG=n, and Reλ<0 for each eigenvalue λ of the matrix M. Then the system(ODE) is controllable

Proof: Since rankG=n, Theroem 2.3 tells us that C contains some ball B centered at 0. Now take any x0Rn and consider the evolution

{x˙(t)=Mx(t)x(0)=x0

in other words, take the control α()0. Since Reλ<0 for each eigenvalue λ of the matrix M, then the origin is asymptotically stable. So ther exists a time T s.t. x(t)B. Thus x(T)BC; and hence there exists a control α()A steering x(t) into 0 in finite time.

Example We once again consider the rocket railroad car, from §1.2, for which n=2,m=1,A=[1,1], and

x˙=[0100]x+[01]α

Then

G=[N,MN]=[0110]

Therefore rankG=2=n

Also, the characteristic polynomial of the matrix M is

p(λ)=det(λIM)=det(λ10λ)=λ2

Since the eigenvalues are both 0, we fail to satisfy the hypotheses of Theorem 2.5.

This example motivates the following extension of the previous theorem:

Theorem 2.6: Improved criterion for controllability

Assume rankG=n and Reλ0 for each eigenvalue λ of M. Then the system(ODE) is controllable.

Proof:

  1. If CRn, then the convexity of C implies that there exists a vector b0 and a real number μ s.t.
bx0μ,x0C

(Must contain a support hyperplane if C doesn't contain the whole space)

Indeed, in the picture we see that b(x0z0)0; and this implies that μ:=bz0.

plane

We will derive a contradiction.

  1. Given b0,μR, our intention is to find x0C s.t. bx0μ fails. Recall x0C iff t>0 and a control α()A s.t.
x0=0tX1(s)Nα(s)ds

Then

bx0=0tbX1(s)Nα(s)ds

Define

v(s):=bX1(s)N
  1. We assert that v0

To see this, suppose instead that v0. Then k times differentiate the expression bX1(s)N w.r.t. s and set s=0, to discover

bMkN=0,k=0,1,2,

This implies b is orthogonal to the columns of G, and so rankG<n. This is a contradiction to our hypothesis, and therefore v0 holds.

  1. Next, define α() this ay:
α(s):={v(s)|v(s)|,if v(s)0,0,if v(s)=0.

Then

bx0=0tv(s)α(s)ds=0t|v(s)|ds

We want to find a time t>0 s.t. 0t|v(s)|ds>μ. In fact, we assert that

0|v(s)|ds=+

To begin the proof above introduce the function

ϕ(t):=tv(s)ds

We will find an ODE ϕ satisfies. Take p() to be the characteristic polynomial of M. Then

p(ddt)v(t)=p(ddt)[betMN]=b(p(ddt)etM)N=b(p(M)etM)N0

Since p(M)=0, according to the Cayley–Hamilton Theorem. But since p(ddt)v(t)0, it follows that

ddtp(ddt)ϕ(t)=p(ddt)(ddtϕ)=p(ddt)v(t)=0

Hence ϕ solves the (n+1)th order ODE

ddtp(ddt)ϕ(t)=0

We also know that ϕ()0. Let μ1,,μn+1 be the solutions of μp(μ)=0. According to ODE theory, we can write

ϕ(t)=sum of the terms of the form pi(t)eμit

for appropriate polynomials pi()

Furthermore, we see that μn+1=0 and μk=λk, where λ1,,λn are the eigenvalues of M. By assumption Reμk0,k=0,1,,n. If 0|v(s)|ds<, then

|ϕ(t)|0|v(s)|ds0,as t

that is, ϕ(t)0 as t. This's a contradiction to the representation formula of ϕ(t)=pi(t)eμit, with Reμi0. Assertaion is proved.

  1. Consequently given any μ,t>0 s.t.
bx0=0t|v(s)|ds>μ

a contradiction to (2.8). Therefore C=Rn.

2.4 Observability

Consider the linear system of ODE

{x˙(t)=Mx(t)x(0)=x0.

where MMn×n.

In this section we address the observability problem, modeled as follows. We suppose that we can observe

y(t):=Nx(t)(t0)

for a given matrix NMm×n. Consequently, y(t)Rm. The interesting situation is when m<<n and we interpret y() as low-dimensional “observations” or “measurements” of the high-dimensional dynamics x()

Observability question: Given the observation y(), can we in principle reconstruct x()? In particular, do observations of y() provide enough information for us to deduce the initial value x0 for (ODE)?

Definition: observable

The pair (ODE, Observation) called observable if the knowledge of y() on any time interval [0,t] allows us to compute x0.

More precisely, (ODE, Observation) is observable if for all solutions x1(),x2(),Nx1()Nx2() on a time interval [0,t] implies x1(0)=x2(0).

2 simple examples

  1. If N0, then clearly the system is not observable.
  2. On the other hand, if m=n and N is invertible, then clearly x(t)=N1y(t) is observable.

The interesting cases lie between these extremes.

Theorem 2.7: Observability and controllability The system 1

{x˙(t)=Mx(t)y(t)=Nx(t)

is observable iff the system 2

z˙(t)=Mz(t)+Nα(t),A=Rm

is controllable, meaning that C=Rn

INTERPRETATION. This theorem asserts that somehow “observability and controllability are dual concepts” for linear systems.

Proof:

  1. () Suppose the system 1 is not observable. Then x1x2Rn, s.t.

    {x˙1(t)=Mx1(t),x1(0)=x1x˙2(t)=Mx2(t),x2(0)=x2

but y(t):=Nx1(t)Nx2(t),t0. Let

x(t):=x1(t)x2(t),x0:=x1x2

Then

x˙(t)=Mx(t),x(0)=x00

but

Nx(t)=0(t0)

Now

x(t)=X(t)x0=etMx0

Thus

NetMx0=0(t0)

Let t=0, to find Nx0=0. Then differentiate this expression k times in t and let t=0, to discover as well that

NMkx0=0

for k=0,1,2 Hence (x0)(Mk)N=0 and hence (x0)(M)kN=0. This implies

(x0)[N,MN,,(M)n1N]=0

Since x00,rank[N,,(M)n1N]<n. Thus system 2 is not controllable. Consequently, system 2 controllable implies system 1 is observable.

  1. ()Assume now system 2 is not controllable. Then rank[N,,(M)n1N]<n, and consequently according to Theorem 2.3, x00, s.t.
(x0)[N,,(M)n1N]=0

That is, NMkx0=0,k=0,1,2,,n1

We want to show that y(t)=Nx(t)0, where

{x˙(t)=Mx(t)x(0)=x0.

According to the Cayley–Hamilton Theorem, we can write

Mn=βn1Mn1β0I

for appropriate constants. Consequently NMnx0=0. Likewise

Mn+1=M(βn1Mn1β0I)=βn1Mnβ0M

and so NMn+1x0=0. Similarly, NMkx0=0,k.

Now

x(t)=X(t)x0=eMtx0=k=0tkMkk!x0

and therefore Nx(t)=Nk=0tkMkk!x0=0

We have shown that if system 2 is not controllable, then system 1 is not observable.

2.5 bang-bang control

Again take A to be the cube [1,1]mRm.

Defnition: bang-bang

A control α()A is called bang-bang if t0 and for each index i=1,,m, we have |αi(t)|=1, where

α(t)=[α1(t)αm(t)]

Theorem 2.8: bang-bang principle

Let t>0 and suppose x0C(t), for the system

x˙(t)=Mx(t)+Nα(t).

Then there exists a bang-bang control α() which steers x0 to 0 at time t.

To prove the theorem we need some tools from functional analysis, among them the Krein–Milman Theorem, expressing the geometric fact that every bounded convex set has an extreme point.

2.5.1 Some functional analysis

We will study the “geometry” of certain infinite dimensional spaces of functions.

Notation

L=L(0,t;Rm)={α():(0,t)Rm|sup0st|α(s)|<}.αL=sup0st|α(s)|.

Definition: converge in the weak* sense

Let αL for n=1, and αL. We say αn converges to α in the weak* sense, written

αnα.

provided

0tαn(s)v(s)ds0tα(s)v(s)ds

as n, for all v():[0,t]Rm satisfying 0t|v(s)|ds<.

We will the following useful weak* compactness theorem for L.

Alaoglu's Theorem

Let αnA,n=1,. Then there exists a subsequence αnk and αA s.t.

αnkα.

Definition: convex; extreme point

  • The set K is convex if x,x^K and all real numbers 0λ1,

    λx+(1λ)x^K.
  • A point zK called extreme provided there do not exist points x,x^K and 0<λ<1 s.t.

    z=λx+(1λ)x^.

Krein-Milman Theorem

Let K be a convex, nonempty subset of L, which is compact in the weak* topology.

Then K has at least one extreme point.

2.5.2 Application to bang-bang control

The foregoing abstract theory will be useful for us in the following setting. We will take K to be the set of controls which steer x0 to 0 at time t, prove it satisfies the hypotheses of Krein–Milman Theorem and finally show that an extreme point is a bang-bang control.

So consider again the linear dynamics

{x˙(t)=Mx(t)+Nα(t)x(0)=x0.

take x0C(t) and write

K={α()A|α() steers x0 to 0 in time t}.

Lemma 2.9: Geometry of set of controls

The collection K of admissible controls satisfies the hypotheses of the Krein-Milman Theorem.

Proof: Since x0C(t), we see that Kϕ.

Next we show that K is convex. For this, recall that α()K iff

x0=0tX1(s)Nα(s)ds.

Now take also α^K and 0λ1. Then

x0=0tX1(s)Nα^(s)ds.

and so

x0=0tX1(s)N(α(s)+(1λ)α^(s))ds.

Hence λα+(1λ)α^K.

Lastly, we confirm the compactness. Let αnK for n=1,. According to Alaoglu’s Theorem \existssnk and αA s.t. αnkα. We need to show that αK.

Now αnkK implies

x0=0tX1(s)Nαnk(s)ds0tX1(s)Nα(s)ds.

by definition of weak-* convergence. Hence αK.

We can now apply the Krein–Milman Theorem to deduce that there exists an extreme point αK. What is interesting is that such an extreme point corresponds to a bang-bang control.

Theorem 2.10: Extremality and bang-bang principle

The control α() is bang-bang.

Proof:

  1. We must show that for almost all times 0st and for each i=1,,m, we have
|αi(s)|=1.

Suppose not. Then there exists an index i{1,,m} and a subset E[0,t] of positive measure s.t. |αi(s)|<1 for sE. In fact, \existssε>0 and a subset FE s.t.

|F|>0 and |αi(s)|1ε for sF.

Define

IF(β()):=FX1(s)Nβ(s)ds.

for β=(0,,β(),,0)., the function β in the ith slot. Choose any real-valued function β()0, s.t.

IF(β())=0

and |β()|1. Define

α1():=α()+εβ()α2():=α()εβ(),

where we redefine β to be zero off the set F

  1. We claim that α1(),α2()K.

To see this, observe that

0tX1(s)Nα1(s)ds=0tX1(s)Nα(s)dsε0tX1(s)Nβ(s)ds=x0εFX1(s)Nβ(s)dsIF(β())=0=x0

Note also α1()A. Indeed,

{α1(s)=α(s)(sF)α1(s)=α(s)+εβ(s)(sF).

But on the set F, we have |αi(s)|1ε, and therefore

|α1(s)||α(s)|+ε|β(s)|1ε+ε=1

Similar considerations apply for α2. Hence α1,α2K, as claimed above.

  1. Finally, observe that
α1()=α+εβ,α1αα2()=αεβ,α2α

But

12α1+12α2=α.

and this is a contradiction, since α is an extreme point of K.

Released under the MIT License.