Skip to content

5 Eigenvalues and Eigenvectors

5.1 Eigenvalues and Eigenvectors

Definition and Theorem

This chapter is concerned with the so-called diagonalization problem. For a given linear operator T on a finite-dimensional vector space V, we seek answers to the following questions.

  1. Does there exist an ordered basis β for V such that [T]β is a diagonal matrix?
  2. If such a basis exists, how can it be found?

Definitions: diagonalizable

A linear operator T on a finite-dimensional vector space V is called diagonalizable if there is an ordered basis β for V such that [T]β is a diagonal matrix. A square matrix A is called diagonalizable if LA is diagonalizable.

Definitions: eigenvalue & eigenvector

Let T be a linear operator on a vector space V. A nonzero vector vV is called an eigenvector of T if there exists a scalar λ such that T(v)=λv. The scalar λ is called the eigenvalue corresponding to the eigenvector V.

Let A be in Mn×n(F). A nonzero vector vFn is called an eigenvector of A if v is an eigenvector of LA; that is, if Av=λv for some scalar λ. The scalar λ is called the eigenvalue of A corresponding to the eigenvector V.

Theorem 5.1

A linear operator T on a finite-dimensional vector space V is diagonalizable if and only if there exists an ordered basis β for V consisting of eigenvectors of T.

Furthermore, if T is diagonalizable, β={v1,v2,...,vn} is an ordered basis of eigenvectors of T, and D=[T]β, then D is a diagonal matrix and Djj is the eigenvalue corresponding to vj for 1jn.

Corollary

A matrix AMn×n(F) is diagonalizable if and only if there exists an ordered basis for Fn consisting of eigenvectors of A. Furthermore, if {v1,v2,...,vn} is an ordered basis for Fn consisting of eigenvectors of A and Q is the n×n matrix whose jth column is vj for j=1,2,...,n, then D=Q1AQ is a diagonal matrix such that Djj is the eigenvalue of A corresponding to vj. Hence A is diagonalizable if and only if it is similar to a diagonal matrix.

[T]β=[λ1000λ2000λn]

Example 5.1.1

Let

A=[1342],v1=[11],v2=[34].

Since

LA(v1)=Av1=[1342][11]=[22]=2[11]=2v1

v1 is an eigenvector of LA, and hence of A. Here λ1=2 is the eigenvalue corresponding to v1. Furthermore,

LA(v2)=Av2=[1342][34]=[1520]=5[34]=5v2,

and so v2 is an eigenvector of LA, and hence of A, with the corresponding eigenvalue λ2=5. Note that β={v1,v2} is an ordered basis for R2 consisting of eigenvectors of both A and LA, and therefore A and LA are diagonalizable. Moreover, by Theorem 5.1 and its corollary, if

Q=[2314]

then

Q1AQ=[LA]β=[2005]

Example 5.1.2

Let T be the linear operator on R2 that rotates each vector in the plane through an angle of π/2.

It is clear geometrically that for any nonzero vector v, the vector T(v) does not lie on the line through 0 determined by v (geometrical meaning of eigenvector, but for this T); hence T(v) is not a multiple of v. Therefore T has no eigenvectors and, consequently, no eigenvalues. Thus there exist operators (and matrices) with no eigenvalues or eigenvectors. Of course, such operators and matrices are not diagonalizable.

While in the field of complex number C2, it exists:

A=[cosθsinθsinθcosθ]pA(t)=det(AtI)=t22(cosθ)t+1e±iθ=cosθ±isinθ,λ1=eiθ,λ2=eiθ

For basis β={x1,x2},

[A]β=[eiθ00eiθ]

Let

x=c1x1+c2x2,[x]β=(c1,c2)

Then

[Ax]β=[A]β[x]β=[c1eiθ,c2eiθ].

Example 5.1.3

Let C(R) denote the set of all functions f:RR having derivatives of all orders(including the polynominl functions, the sine an cosine functions, the exponential functions, etc.). Clearly, C(R) is a subspace of the vector space F(R,R) of all functions from R to R as defined in Section 1.2.

Let T:C(R)C(R) be the function defined by T(f)=f, the derivative of f. It's easily verified that T is a linear operator on C(R). We determine the eigenvalues and eigenvectors of T.

Suppose f is an eigenvector of T with eigenvalue λ. Then f=λf. This's a 1st-order differential equation whose solutions are of the form f(t)=ceλt for some constant c. Consequently, every real number λ is an eigenvalue of T, and corresponds to eigenvectors of the form ceλt for c0. Note that for λ=0, the eigevectors are the nonzero constant functions, i.e. c.

(Here, the function f is the eigenvector v in Av=λv)

Theorem 5.2

Let AMn×n(F). A scalar λ is an eigenvalue of A if and only if

det(AλIn)=0

Definition: characteristic polynomial (For matrix)

The polynomial f(t)=det(AtIn) is called the characteristic polynomial of A.

Definition: characteristic polynomial (For linear transformation)

Let T be a linear operator on an n-dimensional vector space V with ordered basis β. We define the characteristic polynomial f(t) of T to be the characteristic polynomial of A=[T]β. That is,

f(t)=det(AtIn)

Example 5.1.4

To find the eigenvalues of

A=[1141]M2×2(R),

we compute its characteristic polynomial:

det(AtI2)=det[1t141t]=(1t)24=t22t3=(t3)(t+1).

Hence, the only eigenvalues of A are 3 and -1.

Example 5.1.5

Let T be the linear operator on P2(R) defined by

T(f(x))=f(x)+(x+1)f(x),

let β be the standard ordered basis for P2(R), and let A=[T]β. Then

A=[110022003]

(Explanation:

f(x)=a+bx+cx2f(x)=b+2cxT(f(x))=(a+bx+cx2)+(x+1)(b+2cx)=a+bx+cx2+bx+b+2cx2+2cx=(a+b)+(2b+2c)x+3cx2

)

The characteristic polynomial of T is

det(AtI3)=|1t1002t2003t|=(1t)(2t)(3t)=(t1)(t2)(t3).

Hence λ is an eigenvalue of T if and only if λ=1,2, or 3.

Theorem 5.3

Let AMn×n(F).

  • The characteristic polynomial of A is a polynomial of degree n with leading coefficient (1)n.
  • A has at most n distinct eigenvalues.

Theorem 5.4

Let T be a linear operator on a vector space V, and let λ be an eigenvalue of T. A vector vV is an eigenvector of T corresponding to λ if and only if v0 and vN(TλI).

Example 5.1.6 [core]

To find all eigenvectors of the matrix

A=[1141],

recall A has two eigenvalues λ1=3 and λ2=1. We begin with eigenvectors corresponding to λ1=3. Let

B1=A3I=[131413]=[2142].

A vector x=(x1,x2)TR2 is an eigenvector corresponding to λ1=3 iff x0 and xN(LB1), i.e.,

B1x=[2142][x1x2]=[2x1+x24x12x2]=0.

Clearly the set of all solutions to this equation is

{t[12]:tR}.

Now, suppose x is an eigenvector corresponding to λ2=1. Let

B2=A(1)I=A+I=[1+1141+1]=[2142].

xN(LB2) iff

B2x=0{2x1+x2=04x1+2x2=0

Hence

N(LB2)={t[12]:tR}.

Thus x is an eigenvector corresponding to λ2=1 iff

x=t(1,2)for some t0.

Observe that

{(1,2),(1,2)}

is a basis for R2 consisting of eigenvectors of A. Thus LA, and hence A, is diagonalizable.

Suppose that β is a basis for Fn consisting of eigenvectors of A. The corollary to Theorem 2.23 assures that if Q is the n×n matrix whose columns are the vectors in β, then Q1AQ is a diagonal matrix. For example, in Example 5.1.6, if

Q=[1122],D=Q1AQ=[3001].

The diagonal entries of this matrix are the eigenvalues of A that correspond to the respective columns of Q.

To find the eigenvectors of a linear operator T on an n-dimensional vector space, select an ordered basis β for V and let A=[T]β. Then for vV, ϕβ(v)=[v]β, the coordinate vector of v relative to β. We show that vV is an eigenvector of T corresponding to λ if and only if ϕβ(v) is an eigenvector of A corresponding to λ. Suppose v is an eigenvector with eigenvalue λ, then T(v)=λv. Hence

Aϕβ(v)=LAϕβT(v)=ϕβ(T(v))=ϕβ(λv)=λϕβ(v),

Now ϕβ(v)0, since ϕβ is an isomorphism; hence ϕβ(v) is an eigenvector of A. This argument is reversible, and so we can establish that if ϕβ(v) is an eigenvector of A corresponding to λ, then v is an eigenvector of T corresponding to λ.

An equivalent formulation of the result discussed in the preceding paragraph is that for an eigenvalue λ of A (and hence of T ), a vector yFn is an eigenvector of A corresponding to λ if and only if ϕβ1(y) is an eigenvector of T corresponding to λ.

We can choose to solve the problem using whichever path is easier.

Example 5.1.7

Let T be the linear operator on P2(R) defined in Example 5.1.5, i.e., T(f(x))=f(x)+(x+1)f(x), and let β be the standard ordered basis for P2(R). Recall that T has eigenvalues 1,2,3 and that

A=[T]β=(110022003)

We consider each eigenvalue separately. Let λ1=1, and define

B1=Aλ1I=(010012002)

Then

x=(x1x2x3)R3

is an eigenvector corresponding to λ1=1 if and only if x0 and xN(LB1); that is, x is a nonzero solution to the system

x2=0x2+2x3=02x3=0

Notice that this system has three unknowns, x1,x2, and x3, but one of these, x1, does not actually appear in the system. Since the values of x1 do not affect the system, we assign x1 a parametric value, say x1=a, and solve the system for x2 and x3. Clearly, x2=x3=0, and so the eigenvectors of A corresponding to λ1=1 are of the form

a(100)=ae1

for a0. Consequently, the eigenvectors of T corresponding to λ1=1 are of the form

ϕβ1(ae1)=aϕβ1(e1)=a1=a

for any a0. Hence the nonzero constant polynomials are the eigenvectors of T corresponding to λ1=1.

Next let λ2=2, and define

B2=Aλ2I=(110002001)

It is easily verified that

N(LB2)={a(110):aR},

and hence the eigenvectors of T corresponding to λ2=2 are of the form

ϕβ1(a(110))=aϕβ1(e1+e2)=a(1+x)

for a0. Finally, consider λ3=3 and

B3=Aλ3I=(210012000)

Since

N(LB3)={a(121):aR}

the eigenvectors of T corresponding to λ3=3 are of the form

ϕβ1(a(121))=aϕβ1(e1+2e2+e3)=a(1+2x+x2)

for a0.

For each eigenvalue, select the corresponding eigenvector with a=1 in the preceding descriptions to obtain γ={1,1+x,1+2x+x2}, which is an ordered basis for P2(R) consisting of eigenvectors of T . Thus T is diagonalizable, and

[T]γ=(100020003)

5.2 Diagonalizability

Theorem 5.5 and Corollary

What is still needed is a simple test to determine whether an operator or a matrix can be diagonalized, as well as a method for actually finding a basis of eigenvectors.

Theorem 5.5

Let T be a linear operator on a vector space, and let λ1,λ2,,λk be distinct eigenvalues of T. For each i=1,2,,k, let Si(i=1,2,,k) be a finite set of eigenvectors of T corresponding to λi. If each Si is linearly independent, then S1S2Sk is also linearly independent.

Corollary

Let T be a linear operator on an n-dimensional vector space V. If T has n distinct eigenvalues, then T is diagonalizable.

Example 5.2.1

Let

A=[1111]M2×2(R).

The characteristic polynomial of A ( and hence of LA) is t(t2), giving two distinct eigenvalues λ1=0 and λ2=2. By the corollary to Theorem 5.5, A is diagonalizable for LA is a linear operator on the 2-dimensional vector space R2.

split

Definition: split over

A polynomial f(t) in P(F) splits over F if there are scalars c,a1,...,an (not necessarily distinct) in F such that

f(t)=c(ta1)(ta2)(tan).

For example, t21=(t+1)(t1) splits over R, but (t2+1)(t2) does not split over R because t2+1 cannot be factored into a product of linear factors. However, (t2+1)(t2) does split over C because it factors into the product (t+i)(ti)(t2).

If f(t) is the characteristic polynomial of a linear operator or a matrix over a field F, then the statement that f(t) splits is understood to mean that it splits over F.

Theorem 5.6

The characteristic polynomial of any diagonalizable linear operator splits.

From this theorem, it is clear that if T is a diagonalizable linear operator on an n-dimensional vector space that fails to have distinct eigenvalues, then the characteristic polynomial of T must have repeated zeros.

The converse of Theorem 5.6 is false; that is, the characteristic polynomial of T may split, but T need not be diagonalizable. (See Example 5.2.3, which follows.) The following concept helps us determine when an operator whose characteristic polynomial splits is diagonalizable.

Definition: multiplicity

Let λ be an eigenvalue of a linear operator or matrix with characteristic polynomial f(t). The multiplicity (sometimes called the algebraic multiplicity) of λ is the largest positive integer k for which (tλ)k is a factor of f(t).

Example 5.2.2

Let

A=[310034004],

which has characteristic polynomial f(t)=(t3)2(t4). Hence λ=3 is an eigenvalue of A with multiplicity 2, and λ=4 is an eigenvalue of A with multiplicity 1.

Remark

Let A be a square matrix. A (complex) number λ is an eigenvalue of A if and only if λ is a root of the characteristic equation of A.

The multiplicity of λ being a root of the characteristic equation is called the algebraic multiplicity of λ, denoted ma.

The dimension of the eigenspace associated to eigenvalue λ, that is, the maximum number of linearly independent eigenvectors associated with eigenvalue λ, is called the geometric multiplicity of λ, denoted mg. It can be shown that for each eigenvalue λ, we have

1mgma,

where mg and ma are the geometric and algebraic multiplicities of λ respectively.

Therefore we have A is diagonalizable if and only if the algebraic multiplicity and the geometric multiplicity are equal for all eigenvalues of A.

The eigenvectors of T corresponding to the eigenvalue λ are the nonzero vectors in the null space of TλI, we are led naturally to the study of this set.

eigenspace

Definition: eigenspace

Let T be a linear operator on a vector space V , and let λ be an eigenvalue of T . Define Eλ={xV:T(x)=λx}=N(TλIV). The set Eλ is called the eigenspace of T corresponding to the eigenvalue λ. Analogously, we define the eigenspace of a square matrix A corresponding to the eigenvalue λ to be the eigenspace of LA corresponding to λ.

Theorem 5.7

Let T be a linear operator on a finite-dimensional vector space V, and let λ be an eigenvalue of T having multiplicity m. Then

1dim(Eλ)m,

where Eλ is the eigenspace corresponding to λ.

(1mgma)

Example 5.2.3

Let T be the linear operator on P2(R) defined by T(f(x))=f(x). The matrix representation of T with respect to the standard ordered basis β for P2(R) is

[T]β=[010002000].

Consequently, the characteristic polynomial of T is

det([T]βtI)=det[t100t200t]=t3.

Thus T has only one eigenvalue (λ=0) with multiplicity 3. Solving T(f(x))=f(x)=0 shows that Eλ=N(TλI)=N(T) is the subspace of P2(R) consisting of the constant polynomials. So {1} is a basis for Eλ, and therefore dim(Eλ)=1. Consequently, there is no basis for P2(R) consisting of eigenvectors of T, and therefore T is not diagonalizable.

A matrix is diagonalizable if and only if the algebraic multiplicity and geometric multiplicity are equal for all eigenvalues.

Example 5.2.4

Let T be the linear operator on R3 defined by

T[a1a2a3]=[4a1+a32a1+3a2+2a3a1+4a3].

We determine the eigenspace of T corresponding to each eigenvalue. Let β be the standard ordered basis for R3. Then

[T]β=[401232104],

and hence the characteristic polynomial of T is

det([T]βtI)=det[4t0123t214a3t]=(t5)(t3)2.

So the eigenvalues of T are λ1=5 and λ2=3 with multiplicities 1 and 2, respectively.

Since

Eλ1=N(Tλ1I)={xR3:(T5I)x=0},

Eλ1 is the solution space of the system of linear equations

{x1+x3=02x1+2x2+2x3=0x1+x3=0,

which means Eλ1 is spanned by the basis {(1,2,1)}. Hence, dim(Eλ1)=1.

Similarly, Eλ2=N(T3I) is the solution space of the system

{x1+x3=02x1+2x3=0x1+x3=0.

Since the unknown x2 does not appear in the first two equations, we assign it an arbitrary parameter s, and solve the system for x1 and x3, introducing another parameter t. The general solution to the system is

x=s[010]+t[101],for some x,tR.

So a basis for Eλ2 is {(0,1,0),(1,0,1)} and dim(Eλ2)=2.

In this case, the multiplicity of each eigenvalue λi is equal to the dimension of the corresponding eigenspace Eλi. The union of the two bases just derived, namely,

{(1,2,1),(0,1,0),(1,0,1)},

is linearly independent and hence is a basis for R3 consisting of eigenvectors of T. Consequently, T is diagonalizable.

diagram

Lemma

Let T be a linear operator, and let λ1,λ2,,λk be distinct eigenvalues of T. For each i=1,2,,k, let viEλi, the eigenspace corresponding to λi. If

v1+v2++vk=0,

then vi=0 for all i.

Theorem 5.5 says that eigenvectors which corresponds to different eigenvalues are linearly independent, so the only possibility is that they are all zeros.

Theorem 5.8

Let T be a linear operator on a vector space V, and let λ1,λ2,,λk be distinct eigenvalues of T. For each i=1,,k, let Si be a finite linearly independent subset of the eigenspace Eλi. Then S=S1S2Sk is a linearly independent subset of V.

Theorem 5.9

Let T be a linear operator on an n-dimensional vector space V such that the characteristic polynomial of T splits. Let λ1,λ2,,λk be the distinct eigenvalues of T. Then

  • T is diagonalizable if and only if the algebraic multiplicity of λi equals dim(Eλi) for all i (i.e. algebraic multiplicity = geometric multiplicity).
  • If T is diagonalizable and βi is an ordered basis for Eλi for each i, then
β=β1β2βk

is an ordered basis for V consisting of eigenvectors of T.

Test for Diagonalizability

Let T be a linear operator on an n-dimensional vector space V. Then T is diagonalizable if and only if

  1. The characteristic polynomial of T splits.
  2. For each eigenvalue λ of T, the multiplicity of λ equals the corresponding geometrical multiplicity:
multiplicity(λ)=nullity(TλI)=nrank(TλI).

Example 5.2.5

We test the matrix

A=[310030004]M3×3(R)

for diagonalizability.

The characteristic polynomial of A is

det(AtI)=(t4)(t3)2,

which splits and so condition 1 of the test for diagonalization is satisfied. The eigenvalues are λ1=4 and λ2=3 with multiplicities 1 and 2, respectively.

Since λ1 has multiplicity 1, condition 2 is satisfied for λ1. For λ2:

Aλ2I=[010000001]

the matrix A3I has rank 2, so the nullity is 32=1, which is not the multiplicity of λ2. Thus, condition 2 fails for λ2 and A is not diagonalizable.

Example 5.2.6 [core]

Let T be the linear operator on P2(R) defined by

T(f(x))=f(1)+f(0)x+(f(0)+f(0))x2.

With α={1,x,x2} as the standard ordered basis for P2(R), let B=[T]α. Then

B=[111010012].

The characteristic polynomial of B, and hence of T, is

(t1)2(t2),

which splits. Hence, condition 1 of the test for diagonalizability is satisfied.

Also, B has eigenvalues λ1=1 and λ2=2 with multiplicities 2 and 1, respectively.

Condition 2 is satisfied for λ2 because it has multiplicity 1.

Checking condition 2 for λ1=1:

rank(Bλ1I)=rank(011000011)=1,

the rank of BI is 1, so nullity is 31=2, which equals the multiplicity of λ1. Therefore T is diagonalizable.

We now find an ordered basis γ for R3 of eigenvectors of B. For each eigenvalue:

  • The eigenspace corresponding to λ1=1 is
Eλ1={xR3:(BI)x=0}

which is the solution space for the system

x2+x3=0,

and has

γ1={[100],[011]}.

as a basis.

  • The eigenspace corresponding to λ2=2 is
Eλ2={xR3:(B2I)x=0}

which is the solution space for the system

x1+x2+x3=0x2=0,

and has

γ1={[101]}.

as a basis.

Let

γ=γ1γ2={[100],[011],[101]}

Then γ is an ordered basis for R3 consisting of eigenvectors ofB.

Finally, observe that the vector in γ are the coordinatevectors relative to α of the vectors in the set

β={1,x+x2,1+x2},

which is an ordered basis for P2(R) consisting of eigenvectors of T. Thus

[T]β=[100010002].

Example 5.2.7

Let

A=[0213].

We show A is diagonalizable and find a 2×2 matrix Q such that Q1AQ is diagonal. We then show how to use this result to compute An for any positive integar n.

First observe the characteristic polynomial of A is (t1)(t2), so A has two distinct eigenvalues, λ1=1 and λ2=2.

By applying the corollary to Theorem 5.5 to the operator LA, A is diagonalizable. Moreover,

γ1={(2,1)},γ2={(1,1)}.

are bases for the eigenspaces Eλ1 and Eλ2, respectively. Therefore,

γ=γ1γ2={(2,1),(1,1)}

is an ordered basis for R2 consisting of eigenvectors of R2. Let

Q=[2111]

be the matrix whose columns are the vectors in γ. Then, by the corollary to Theorem 2.23,

D=Q1AQ=[1002].

To compute An for any positive integer n,

An=(QDQ1)n=(QDQ1)(QDQ1)(QDQ1)=QDnQ1=Q[1n002n]Q1=[2111][1002n][1112]=[22n22n+11+2n1+2n+1].

Systems of Differential Equations

Exercise 5.2.8

Consider the system of differential equations

{x1=3x1+x2+x3x2=2x1+4x2+2x3x3=x1x2+x3,

where each xi=xi(t) is a differentiable real-valued function of the real value t. Clearly, this system has a solution, namel, the solution in which each xi(t) is the zero function. We determine all of the solutions to this system.

Let x:RR3 be

x(t)=[x1(t)x2(t)x3(t)],

and x(t) is the derivative of x.

Let

A=[311242111].

be the coefficient omatrix of the given system, so that we can rewrite the system as matrix equation x=Ax.

It can be verified that for

Q=[101012111]andD=[200020004],

we have Q1AQ=D. Substitute A=QDQ1 into x=Ax to obtain

x=QDQ1x,

or equivalently,

Q1x=DQ1x.

Define y(t)=Q1x(t) yielding y=Q1x. Hence,

y=Dy,

which is a system of decoupled equations

{y1=2y1y2=2y2y3=4y3.

Solutions for each yi are

y1(t)=c1e2t,y2(t)=c2e2t,y3(t)=c3e4t,

with arbitrary constants c1,c2,c3.

Back-substituting,

x(t)=Qy(t)=[101012111][c1e2tc2e2tc3e4t]=[c1e2t+c3e4tc2e2tc3e4tc1e2t+c2e2t+c3e4t].

Thus the general solution is

x(t)=e2tz1+e4tz2=e2t[c1[101]+c2[011]]=e4t[c3[121]].

where z1Eλ1 and z2Eλ2, with λ1=2 and λ2=4.

This concludes the main part of the example.

Here is more generalized conclusion.

Application for ODE

Let n×n matrix A=(aij) be the coefficient matrix of a system of differential equations

xi=j=1naijxj,i=1,,n.

Suppose A is diagonalizable and the distinct eigenvalues of A are λ1,λ2,,λk.

Prove that a differentiable function x:RRn is a solution to the system if and only if

x(t)=eλ1tz1+eλ2tz2++eλktzk,

where ziEλi for i=1,,k. Use this result to prove that the set of solutions form an n-dimensional real vector space.

Following the step of the previous exercise, we may pick a matrix Q whose column vectors consist of eigenvectors and Q is invertible. Let D be the diagonal matrix Q1AQ. And we also know that finally we'll have the solution x=Qu for some vector u whose i-th entry is cieλ if the i-th column of Q is an eigenvector corresponding to λ. By denoting D¯ to be the diagonal matrix with D¯ii=eDii, we may write x=QD¯y. where the i-th entry of y is ci. So the solution must be of the form described in the exercise.

For the second statement, we should know first that the set

{eλ1t,eλ2t,eλkt}

are linearly independent in the space of real functions. Since Q invertible, we know that the solution set

{QD¯y:yRn}

is an n-dimensional real vector space.

Released under the MIT License.