Chap1 PDF
Chap1 PDF
Chap1 PDF
This course focuses on the state space approach to the analysis and design of control systems. The idea
of state of a system dates back to classical physics. Roughly speaking, the state of a system is that
quantity which, together with knowledge of future inputs to the system, determine the future behaviour
of the system. For example, in mechanics, knowledge of the current position and velocity (equivalently,
momentum) of a particle, and the future forces acting it, determines the future trajectory of the particle.
Thus, the position and the velocity together qualify as the state of the particle.
In the case of a continuous-time linear system, its state space representation corresponds to a system
of first order differential equations describing its behaviour. The fact that this description fits the idea of
state will become evident after we discuss the solution of linear systems of differential equation.
To illustrate how we can obtain state equations from simple differential equation models, consider the
second order linear differential equation
ÿ(t) = u(t) (1.1)
F
This is basically Newton’s law F = ma for a particle moving on a line, where we have used u to denote m.
It also corresponds to a transfer function from u to y to be s12 . Let
x1 (t) = y(t)
x2 (t) = ẏ(t)
Set xT (t) = [x1 (t) x2 (t)]. Then we can write
dx(t) 0 1 0
= x(t) + u(t) (1.2)
dt 0 0 1
which is a system of first order linear differential equations.
The above construction can be readily extended to higher order linear differential equations of the form
y (n) (t) + a1 y (n−1) (t) + · · · + an−1 ẏ(t) + an y(t) = u(t)
h n−1
iT
dy
Define x(t) = y dt · · · ddtn−1y . It is straightforward to check that x(t) satisfies the differential
equation
0 1 0 0 ··· 0
0
0 0 1 0 ··· 0
..
dx(t) .. .. .. .. ..
.
= x(t) + u(t)
. . ··· . . .
dt 0
0 0 0 1
1
−an −an−1 ··· −a2 −a1
1
2 CHAPTER 1. ANALYSIS OF LINEAR SYSTEMS IN STATE SPACE FORM
On the other hand, it is not quite obvious how we should define the state x for a differential equation of
the form
ÿ(t) + 3ẏ(t) + 2y(t) = u̇(t) + 4u(t) (1.3)
The general problem of connecting state space representations with transfer function representations will
be discussed later in this chapter. We shall only consider differential equations with constant coefficients
in this course, although many results can be extended to differential equations whose coefficients vary with
time.
Once we have obtained a state equation, we need to solve it to determine the state. For example, in the
case of Newton’s law (1.1), solving the corresponding state equation (1.2) gives the position and velocity
of the particle. We now turn our attention, in the next few sections, to the solution of state equations.
ẋ(t) = Ax(t)
x(t) ∈ Rn (1.4)
x(0) = x0
Very often, for simplicity, we shall suppress the time dependence in the notation. We shall now develop a
systematic method for solving (1.4) using a matrix-valued function called the transition matrix.
It is known from the theory of ordinary differential equations that a linear systems of differential
equations of the form
has a unique solution passing through x0 at t = 0. To obtain the explicit solution for x, recall that the
scalar linear differential equation
ẏ = ay
y(0) = y0
and satisfies
d at
e = aeat
dt
By analogy, let us define the matrix exponential
∞
X An
eA =
n=0
n!
1.2. PROPERTIES OF THE TRANSITION MATRIX AND SOLUTION OF LINEAR ODES 3
where A0 is defined to be I, the identity matrix (analogous to a0 = 1). We can then define a function eAt ,
which we call the transition matrix of (1.4), by
∞
X Ak tk
eAt = (1.7)
k!
k=0
The infinite sum can be shown to converge always, and its derivative can be evaluated by differentiating
the infinite sum term by term. The analogy with the solution of the scalar equation suggests that eAt x0 is
the natural candidate for the the solution of (1.4).
(Property 2) d At
e = AeAt (1.9)
dt
To prove this, differentiate the infinite series in (1.7) term by term to get
∞
d At d X Ak tk
e =
dt dt k!
k=0
∞
X Ak tk−1
=
(k − 1)!
k=1
∞
X Ak tk
= A
k!
k=0
At
= Ae
Properties (1) and (2) together can be considered the defining equations of the transition matrix. In
other words, eAt is the unique solution of the linear matrix differential equation
dF
= AF, with F (0) = I (1.10)
dt
To see this, note that (1.10) is a linear differential equation. Hence there exists a unique solution to the
initial value problem. Properties (1) and (2) show that eAt is a solution to (1.10). By uniqueness, eAt is
the only solution. This shows (1.10) can be considered the defining equation for eAt .
If A and B commutes, eAt B = BeAt so that the R.H.S. of (1.11) becomes (A + B)eAt eBt . Furthermore,
at t = 0, eAt eBt = I. Hence eAt eBt satisfies the same differential equation as well as initial condition as
e(A+B)t . By uniqueness, they must be equal.
Setting t = 1 gives
eA+B = eA eB whenever AB = BA
In particular, since At commutes with As for all t, s, we have
Proof: We have
e−At eAt = eA(t−t) = I
Since eAt is a square matrix, this shows that (Property 4) is true. Thus, eAt is invertible for all t.
Using properties (1) and (2), we immediately verify that
The λi ’s are not necessarily distinct and they may be complex, provided we interpret (1.13) as a differential
equation in C n . Since A is diagonal,
k
λ1 0 0
0 λk
k 2
A =
. .
. 0
0 0 λkn
1.3. COMPUTING E AT : LINEAR ALGEBRAIC METHODS 5
If we directly evaluate the sum of the infinite series in the definition of eAt in (1.7), we find for example
that the (1, 1) entry of eAt is given by
∞
X λk1 tk
= eλ1 t
k!
k=0
Hence, in the diagonal case,
eλ1 t
0 0
0 eλ2 t
eAt = (1.15)
..
. 0
0 0 eλn t
We can also interpret this result by examining the structure of the state equation. In this case (1.13) is
completely decoupled into n differential equations
so that
xi (t) = eλi t x0i (1.17)
where x0i is the ith component of x0 . It follows that
λt
e 1 0 0
0 eλ2 t
x(t) = x0 (1.18)
..
. 0
0 0 eλn t
Case I: A Diagonalizable
By definition, A diagonalizable means there exists a nonsingular matrix P (in general complex) such
that P −1 AP is a diagonal matrix, say
λ1 0 0
0 λ2
P −1 AP = Λ = (1.19)
..
. 0
0 0 λn
eAt = eP ΛP
−1 t
(1.20)
(P BP −1 )2 = P BP −1 P BP −1 = P B 2 P −1
so that in general
(P BP −1 )k = P B k P −1 (1.21)
We then have the following
eP BP t
= P eBt P −1
−1
for any n × n matrix B (1.22)
6 CHAPTER 1. ANALYSIS OF LINEAR SYSTEMS IN STATE SPACE FORM
Proof:
∞
X (P BP −1 )k tk
eP BP
−1 t
=
k!
k=0
∞
X B k tk −1
= P P
k!
k=0
= P eBt P −1
eλ1 t
0 0
0 eλ2 t
eAt
−1
= P P (1.23)
..
. 0
0 0 eλn t
From linear algebra, we know that a matrix A is diagonalizable if and only if it has n linearly indepen-
dent eigenvectors. Each of the following two conditions is sufficient, though not necessary, to guarantee
diagonalizability:
In these two cases, A has a complete set of n independent eigenvectors v1 , v2 , · · · , vn , with corresponding
eigenvalues λ1 , λ2 , · · · , λn . Write
P = v1 v2 · · · vn (1.24)
Then
λ1 0 0
0 λ2
P −1 AP = Λ = (1.25)
..
. 0
0 0 λn
The equations (1.24), (1.25), and (1.23) together provide a linear algebraic method for omputing eAt .
From (1.23), we also see that the dynamic behaviour of eAt is completely characterized by the behaviour
of eλi t , i = 1, 2, · · · , n.
Example 1:
1 2
A =
−1 4
The characteristic equation is given by
det(sI − A) = s2 − 5s + 6 = 0
Example 2:
1 0 0
A = 1 2 0
1 0 −1
Owing to the form of A, we see right away that the eigenvalues are 1, 2 and -1 so that
1 0 0
Λ = 0 2 0
0 0 −1
T T
The eigenvector corresponding to the eigenvalue 2 is given by 0 1 0 , while that of -1 is 0 0 1 .
To find the eigenvector for the eigenvalue 1, we have
1 0 0 v1 v1
1 2 0 v2 = v2
1 0 −1 v3 v3
2
so that −2 is an eigenvector. The diagonalizing matrix P is then
1
2 0 0
P = −2 1 0
1 0 1
1
2 0 0
P −1 = 1 1 0
− 21 0 1
8 CHAPTER 1. ANALYSIS OF LINEAR SYSTEMS IN STATE SPACE FORM
t 1
et
2 0 0 e 0 0 2 0 0 0 0
eAt = −2 1 0 0 e2t 0 1 1 0 = −et + e2t e2t 0
1 0 1 0 0 e−t − 21 0 1 1 t
2 (e − e )
−t 0 e−t
Model Decomposition:
We can give a dynamical interpretation to the above results. Suppose A has distinct eigenvalues
λ1 , ..., λn so that it has a set of linearly independent eigenvectors v1 , ..., vn . If x0 = vj , then
This means that if we start along en eigenvector of A, the solution x will stay in the direction of the eigen-
vector, with length being stretched or shrunk by eλj t . In general, because the vi′ s are linearly independent,
an arbitrary initial condition x0 can be expressed as
n
X
x0 = ξj vj = P ξ (1.26)
j=1
where
P = v1 v2 · · · vn
ξ1
..
and ξ = . Using this representation, we have
ξn
n
X
x(t) = eAt x0 = eAt ξj vj
j=1
n
X Xn
= ξj eAt vj = ξj eλj t vj (1.27)
j=1 j=1
P −1 AP = Λ
and
the same result as in (1.27). The representation (1.27) of the solution x(t) in terms of the eigenvectors of
A is often called the modal decomposition of x(t).
Write A = λI + N where
0 1 0
N = 0 0 1 (1.29)
0 0 0
Direct calculation shows that
0 0 1
N 2 = 0 0 0
0 0 0
and
N3 = 0
A matrix B satisfying B k = 0 for some k ≥ 1 is called nilpotent. The smallest integer k such that B k = 0
is called the index of nilpotence. Thus the matrix N in (1.29) is nilpotent with index 3. But then the
transition matrix eN t is easily evaluated to be
t2
X2 j
N t j 1 t 2!
eN t = = 0 1 t (1.30)
j!
j=0 0 0 1
Once we understand the 3 × 3 case, it is not difficult to check that an l × l matrix of the form
0 1 0 ··· 0
..
.
N = ..
..
. 0
.
1
0 ··· 0
i.e., 1′ s along the superdiagonal, and 0 everywhere else, is nilpotent and satisfies N l = 0
10 CHAPTER 1. ANALYSIS OF LINEAR SYSTEMS IN STATE SPACE FORM
To compute eAt when A is of the form (1.28), recall Property 4 of matrix exponentials:
If A and B commute, i.e. AB = BA, then e(A+B)t = eAt eBt .
If A is of the form (1.28), then since λI commutes with N , we can write, using Property 3 of transition
matrices,
using (1.30).
We are now in a position to sketch the general structure of the matrix exponential. More details are
given in the Appendix to this chapter. The idea is to transform a general A using a nonsingular matrix P
into
J1 0
J2
P −1 AP = J =
. .
.
0 Jk
where each Ji = λi I + Ni , Ni nilpotent. J is called the Jordan form of A. With the block diagonal
structure, we can verify m
J1 0
J2m
Jm =
..
.
0 m
Jk
By applying the definition of the transition matrix, we obtain
Jt
e 1 0
eJ2 t
eJt = (1.32)
..
.
0 eJk t
Each eJi t can be evaluated by the method described above. Finally,
The above procedure in principle enables us to evaluate eAt , the only problem being to find the matrix
P which transforms A either to diagonal or Jordan form. In general, this can be a very tedious task (See
Appendix to Chapter 1 for an explanation on how it can be done). The above formula is most useful as
a means for studying the qualitative dependence of eAt on t, and will be particularly important in our
discussion of stability.
We remark that the above results hold regardless of whether the λi ’s are real or complex. In the latter
case, we shall take the underlying vector space to be complex and the results then go through without
modification.
If we let G(t) = eAt , then the Laplace transform of G(t), denoted by Ĝ(s), satisfies
sĜ(s) = AĜ(s) + I
or
Ĝ(s) = (sI − A)−1 (1.34)
If we denote the inverse Laplace transform operation by L−1
(a) Find (sI − A)−1 . The entries of (sI − A)−1 will be strictly proper rational functions.
(b) Determine the inverse Laplace transform for each entry of (sI − A)−1 . This is usually accomplished
1
by doing partial fractions expansion and using the fact that L−1 [ s−λ ] = eλt (or looking up a mathe-
matical table!).
Determining (sI − A)−1 using the cofactor expansion method from linear algebra can be quite tedious.
Here we give an alternative method. Let det(sI − A) = sn + p1 sn−1 + ... + pn = p(s). Write
B1 = I
Bk+1 = ABk + pk I 1 ≤ k ≤ n − 1 (1.37)
To illustrate the Laplace transform method for evaluating eAt , again consider
1 0 0
A = 1 2 0
1 0 −1
s−1 0 0
p(s) = det −1 s − 2 0 = (s − 1)(s − 2)(s + 1)
−1 0 s+1
= s3 − 2s2 − s + 2
The matrix polynomial B(s) can then be determined using the recursive procedure (1.37).
−1 0 0
B2 = A − 2I = 1 0 0
1 0 −3
−2 0 0
B3 = AB2 − I = 1 −1 0
−2 0 2
12 CHAPTER 1. ANALYSIS OF LINEAR SYSTEMS IN STATE SPACE FORM
giving
s2 − s − 2
0 0
B(s) = s + 1 s2 − 1 0
s−2 0 2
s − 3s + 2
where u is a piecewise continuous Rm -valued function. Define a function z(t) = e−At x(t). Then z(t)
satisfies
Since the above equation does not depend on z(t) on the right hand side, it can be directly integrated to
give Z t Z t
z(t) = z(0) + e−As Bu(s)ds = x(0) + e−As Bu(s)ds
0 0
Hence
x(t) = eAt z(t)
Rt (1.39)
= eAt x0 + 0 eA(t−s) Bu(s)ds
More generally, we have
Z t
x(t) = eA(t−t0 ) x(t0 ) + eA(t−s) Bu(s)ds (1.40)
t0
where the output y is a piecewise continuous Rp -valued function. Using (1.39) in (1.42) give the solutions
immediately
Z t
y(t) = CeAt x0 + CeA(t−s) Bu(s)ds + Du(t) (1.43)
0
L R
x2
+ +
C x1
u -
-
The choice of x1 to denote the voltage across the capacitor and x2 the current through the inductor
gives a state description for the circuit. The equations are
C ẋ1 = x2 (1.46)
Lẋ2 + Rx2 + x1 = u (1.47)
We have
−1
s −2
(sI − A)−1 =
1 s+3
s+3 2
−1 s
= 2
s + 3s + 2
2
− 1 2 2
s+1 − s+2
= s+1 s+2
1 1 1 2
− s+1 + s+2 − s+1 + s+2
The impluse response from the input voltage source to the output voltage is given by
If we have a constant input u = 1, the output voltage using (1.39), assuming initial rest (x0 = 0), is given
by
Z t
y(t) = [2e−(t−s) − 2e−2(t−s) ]ds
0
= 2(1 − e−t ) − (1 − e−2t )
Armed with (1.39) and (1.43), we can see why (1.38) deserves the name of state equation. The solutions
for x(t) and y(t) from (1.39) and (1.43) show that indeed knowledge of the current state (x0 ) and future
inputs u(t), t ≥ 0 determines the future states x(t), t ≥ 0 and outputs y(t), t ≥ 0.
1.6 Stability
Let us consider the unforced state equation first:
ẋ = Ax, x(0) = x0 .
The vector x(t) → 0 as t → ∞ for every x0 ⇔ all the eigenvalues of A lie in the open left half-plane.
The entries in the matrix eJt are of the form eλi t p(t) where p(t) is a polynomial in t, and λi is an eigenvalue
of A. These entries converge to 0 if and only if Reλi < 0. Thus
x(t) → 0 ∀x0
ẋ = Ax + Bu
y = Cx + Du .
Hence, if all the eigenvalues of A are in the open left half-plane, then H(s) is a stable transfer matrix. The
converse is not necessarily true.
16 CHAPTER 1. ANALYSIS OF LINEAR SYSTEMS IN STATE SPACE FORM
Example:
−1 0 1
A= , B=
0 2 0
C = [1 0], D=0
1
H(s) = .
s+1
So the transfer function has stable poles, but the system is obviously unstable. Any initial condition
x0 with a nonzero second component will cause x(t) to grow without bound.
A more subtle example is the following
Example:
0 1 1 −1 0
A=
−2 −2 ,
0 B=
2
1
2 1 −1 −1 −1
C = [2 2 1], D = [0 0]
1
H(s) = C · adj(sI − A) · B
det(sI − A)
s2 + 3s + 2 s+2 s+2 −1 0
1
= · [2 2 1] ·
−2s − 2 s2 + s − 2 −2 2 1
s2 + 3s2 + 2s
2s + 2 s+2 s2 + 2s + 2 −1 −1
−1 0
1
[2s2 + 4s + 2 2s2 + 5s + 2 s2 + 4s + 2]
= 2 2 2 1
s + 3s + 2s
−1 −1
1
= [s2 + 2s s2 + s]
s3
+ 3s 2 + 2s
s+2 s+1
=
s2 + 3s + 2 s2 + 3s + 2
Thus {poles of H(s)} = {-1,-2}. Hence the eigenvalue of A at λ = 0 does not appear as a pole of H(s). Note
that there is a cancellation of a factor of s in the numerator and denominator of H(s). This corresponds
to a pole-zero cancellation in the determination of H(s), and indicates something internal in the system
cannot be seen from the input-output behaviour.
harder problem of determining a state space description of a linear time-invariant system from its transfer
function. This is also called the realization problem in control theory. We shall only solve the single-
input single-output case, as the general multivariable problem is much harder and beyond the scope of this
course.
Suppose we are given the scalar-valued transfer function H(s) of a single-input single-output linear
time invariant system. We assume H(s) is a proper rational function; otherwise we cannot find a state
space representation of the form (1.41), (1.42) for H(s). We also assume that there are no common factors
in the numerator and denominator of H(s). After long division, if necessary, we can write
q(s)
H(s) = d +
p(s)
where degree of q(s) < degree of p(s). The constant d corresponds to the direct transmission term from u
to y, so the realization problem reduces to finding a triple (cT , A, b) such that
q(s)
cT (sI − A)−1 b =
p(s)
Note that we have used the notation cT for the 1 × n matrix C, in keeping with our normal convention
that lower case letters denote column vectors.
We now give 2 solutions to the realization problem. Let
q(s)
Ĥ(s) =
p(s)
with
p(s) = sn + p1 sn−1 + · · · + pn
q(s) = q1 sn−1 + q2 sn−2 + · · · + qn
q(s)
Then the following (cT , A, b) triple realizes p(s)
0 ··· 0 −pn
..
1 0 ··· .
A= .. .. .. ..
.
. . .
··· ··· 1 −p1
qn
..
b= .
q1
cT = [0 · · · 0 1]
To see this, let
ξ T (s) = [ξ1 (s) ξ2 (s) ... ξn (s)]
18 CHAPTER 1. ANALYSIS OF LINEAR SYSTEMS IN STATE SPACE FORM
= cT (sI − A)−1
Then
ξ T (s)(sI − A) = cT (1.51)
and
n
X
n
s ξ1 (s) + pi sn−i ξ1 (s) = 1 (1.52)
i=1
so that
1
ξ1 (s) =
p(s)
Hence
1
ξ T (s) = [1 s s2 · · · sn−1 ] (1.53)
p(s)
and
q(s)
cT (sI − A)−1 b = ξ T (s)b =
p(s)
The resulting system is often said to be in observable canonical form, We shall discuss the meaning of
the word “observable” in control theory later in the course.
As an example, we can now write down the state space representation for the equation given in (1.3).
The transfer function of (1.3) is
s+4
H(s) = 2
s + 3s + 2
The observable representation is therefore given by
0 −2 4
ẋ = x + u
1 −3 1
h i
y = 0 1 x
1.8. LINEARIZATION 19
Now note that because H(s) is scalar-valued, H(s) = H T (s). We can immediately conclude that the
q(s)
following (cT1 , A1 , b1 ) triple also realizes p(s)
0 1 0 0 ··· 0
0 0 1 0 ··· 0
.. .. .. .. ..
A1 = .
. . ··· . .
0 0 0 1
−pn −pn−1 ··· −p2 −p1
0
..
.
b1 =
0
1
1.8 Linearization
So far, we have only considered systems described by linear constant coefficient differential equations.
Many systems are nonlinear, with dynamics described by nonlinear differential equations of the form
ẋ = f (x, u) (1.54)
x = xp + δx
u = up + δu
If we drop higher order terms (h.o.t.) involving δxj and δuk , we obtain a linear differential equation for δx
d
δx = Aδx + Bδu (1.55)
dt
∂fj ∂fj
where [A]jk = ∂xk (xp , up ), [B]jk = ∂uk (xp , up ). A is called the Jacobian matrix of f with respect to x,
∂f
sometimes, simply written as Similarly, B = ∂f
∂x (xp , up ). ∂u (xp , up ).
The resulting system (1.55) is called the linearized system of the nonlinear system (1.54) about the
operating point (xp , up ). The process is called linearization of the nonlinear system about (xp , up ). Keeping
the original system near the operating point can often be achieved by keeping the linearized system near
0.
Example:
Consider a pendulum subjected to an applied torque. The normalized equation of motion is
g
θ̈ + sin θ = u (1.56)
l
h iT
where g is the gravitational constant, and l is the length of the pendulum. Let x = θ θ̇ . We have
the following nonlinear differential equation
ẋ1 = x2
g
ẋ2 = − sin x1 + u (1.57)
l
We recognize
x2
f (x, u) =
− gl sin x1 + u
The system matrix A0 has purely imaginary eigenvalues, corresponding to harmonic motion. Note,
however, that (x1 , x2 , u) = (π, 0, 0) is also an equilibrium point, corresponding to the pendulum in the
inverted vertical position. The linearized system is then given by
0 1 0
ẋ = x + u = Aπ x + Bu
g
l 0 1
The system matrix Aπ has an unstable eigenvalue, and the behaviour of the linearized system about
the 2 equilibrium points are quite different.
APPENDIX TO CHAPTER I 21
This appendix provides more details on the Jordan form of a matrix and its use in computing the transition
matrix. The results are deep and complicated, and we cannot give a complete exposition here. Standard
textbooks on linear algebra should be consulted for a more thorough treatment.
When a matrix A has repeated eigenvalues, it is usually not diagonalizable. However, there always
exists a nonsingular matrix P such that
J1 0
J
−1
2
P AP = J =
..
(A.1)
.
0 Jk
The λ′i s need not be all distinct, but ki=1 mi = n. The special form on the right hand side of (A.1) is
P
called the Jordan form of A, and Ji is the Jordan block associated with eigenvalue λi . Note that if mi = 1
for all i, i.e., each Jordan block is 1 × 1 and k = n, the Jordan form specializes to the diagonal form.
There are immediately some natural questions which arise in connection with the Jordan form: for an
eigenvalue λi with multiplicity ni , how many Jordan blocks can there be associated with λi , and how do we
determine the size of the various blocks? Also, how do we determine the nonsingular P which transforms
A into Jordan form?
First recall that the nullspace of A, denoted by N (A), is the set of vectors x such that Ax = 0. N (A)
is a subspace and has a dimension. If λ is an eigenvalue of A, the dimension of N (A − λI) is called the
geometric multiplicity of λ. The geometric multiplicity of λ corresponds to the number of independent
eigenvectors associated with λ.
Fact: Suppose λ is an eigenvalue of A. The number of Jordan blocks associated with λ is equal to the
geometric multiplicity of λ.
The number of times λ is repeated as a root of the characteristic equation det(sI − A) = 0 is called
its algebraic multiplicity (commonly simplified to just multiplicity). The geometric multiplicity of λ is
always less than or equal to its algebraic multiplicity. If the geometric multiplicity is strictly less than the
algebraic multiplicity, λ has generalized eigenvectors. A generalized eigenvector is a nonzero vector v
such that for some positive integer k, (A − λI)k v = 0, but that (A − λI)k−1 v 6= 0. Such a vector v defines
a chain of generalized eigenvectors {v1 , v2 , · · · , vk } through the equations vk = v, vk−1 = (A − λI)vk ,
vk−2 = (A − λI)vk−1 , · · · , v1 = (A − λI)v2 . Note that (A − λI)v1 = (A − λI)k v = 0 so that v1 is an
22 CHAPTER 1. ANALYSIS OF LINEAR SYSTEMS IN STATE SPACE FORM
Av2 = λv2 + v1
Av3 = λv3 + v2
..
.
Avk = λvk + vk−1
It can be shown that the generalized eigenvectors associated with an eigenvalue λ are linearly indepen-
dent. The nonsingular matrix P which transforms A into its Jordan form is constructed from the set of
eigenvectors and generalized eigenvectors.
The complete procedure for determining all the generalized eigenvectors, though too complex to describe
precisely here, can be well-illustrated by an example.
Example:
Consider the matrix A given by
0 0 1 0
0 0 0 1
A=
0 0 −1 1
0 0 2 −2
det(sI − A) = s3 (s + 3)
so that the eigenvalues are −3, and 0, with algebraic multiplicity 3. Let us first determine the geometric
multiplicity of the eigenvalue 0. The equation
0 0 1 0 w1
0 0 0 1 w2
Aw = = 0 (A.3)
0 0 −1 1 w3
0 0 2 −2 w4
gives w3 = 0, w4 = 0, and w1 , w2 arbitrary. Thus the geometric multiplicity of 0 is 2, i.e., there are 2 Jordan
blocks associated with 0. Since 0 has algebraic multiplicity 3, this means there are 2 eigenvectors and 1
generalized eigenvector. To determine the generalized eigenvector, we solve the homogeneous equation
A2 v = 0, for solutions v such that Av 6= 0. Now
0 0 1 0 0 0 1 0
0 0 0 1 0 0 0 1
A2 =
0 0 −1 1 0 0 −1 1
0 0 2 −2 0 0 2 −2
0 0 −1 1
0 0 2 −2
=
0 0 3 −3
0 0 −6 6
APPENDIX TO CHAPTER I 23
solves A2 v = 0 and
1
1
Av =
0
0
so that we have the chain of generalized eigenvectors
0 1
0 1
v2 = , v1 =
1 0
1 0
h iT
with v1 an eigenvector. From (A.3), a second independent eigenvector is w = 1 0 0 0 . This
completes the determination of the eigenvectors and generalized eigenvectors associated with 0. Finally
the eigenvector for the eigenvalue −3 is given by the solution of
3 0 1 0
0 3 0 1
(A + 3I)z = z = 0
0 0 2 1
0 0 2 1
h iT
yielding z = 1 −2 −3 6 The nonsingular matrix bringing A to Jordan form is then given by
1 1 0 1
h i 0 1 0 −2
P = w v1 v2 z =
0 0 1 −3
0 0 1 6
Note that the chain of generalized eigenvectors v1 , v2 go together. One can verify that
1 −1 13 − 31
0 1 − 29 − 92
P −1 =
0 2 1
0 3 3
0 0 − 19 1
9
24 CHAPTER 1. ANALYSIS OF LINEAR SYSTEMS IN STATE SPACE FORM
and
0 0 0 0
0 0 1 0
P −1 AP =
0
(A.4)
0 0 0
0 0 0 −3
0 1 0 0
0 0 0 0
M −1 AM =
0
(A.5)
0 0 0
0 0 0 −3
although it is common to choose the blocks ordered in decreasing size, i.e., that given in (A.5).
Once we have transformed A into its Jordan form (A.1), we can readily determine the transition matrix
eAt . As before, we have
Am = (P JP −1 )m = P J m P −1
By the block diagonal nature of J, we readily see
m
J1 0
m
J2m
J =
..
.
0 Jkm
Thus,
eJ1 t
0
eJ2 t
eAt = P
−1
P
..
.
0 eJk t
Since the mi × mi block Ji = λi I + Ni , using the methods of Section 1.3, we can immediately write down
tmi −1
1 t · · · (m i −1)!
..
Ji t λi t
.
e = e
..
. t
0 1
We have now a complete linear algebraic procedure for determining eAt when A is not diagonalizable.
For
0 0 1 0
0 0 0 1
A= 0 0 −1 1
0 0 2 −2
APPENDIX TO CHAPTER I 25