Skip to content

6 Inner Products Spaces

6.1 Inner Products and Norms

Definition

Definition: Inner product [core]

Let V be a vector space over field F. An inner product on V is a function assigning each ordered pair of vectors x,y in V×V a scalar x,yF,satisfying:

  • x+z,y=x,y+z,y.
  • cx,y=cx,y for cF.
  • x,y=y,x where the bar denotes complex conjugation.
  • x,x>0 if x0

Note that:

  • The 3rd one reduces to x,y=y,x if F=R
  • The 1st and 2nd parts simply require that the inner product be linear in the first component, (and conjugate linear in the second component).
  • It's easily shown that if a1,a2,,anF and y,v1,v2,,vnV, theni=1naivi,y=i=1naivi,y.

The idea of distance or length is missing. Therefore we need a richer structure, the so-called inner product space structure, by adding a new inner product function.

Example 6.1.1

Example 1 For x=(a1,a2,,an) and y=(b1,b2,,bn) in Fn, define

x,y=i=1naibi

The verification that , satisfies conditions the 1st through (4th) is easy. For example, if z=(c1,c2,,cn), we have for (a)

x+z,y=i=1n(ai+ci)bi=i=1naibi+i=1ncibi=x,y+z,y

Thus, for x=(1+i,4) and y=(23i,4+5i) in C2,

x,y=(1+i)(2+3i)+4(45i)=1515i.

The inner product in Example 1 is called the standard inner product on Fn. When F=R the conjugations are not needed, and in early courses this standard inner product is usually called the dot product and is denoted by xy instead of x,y.

Example 6.1.2

If x,y is any inner product on a vector space V and r>0, define another inner product by x,y=rx,y. If r0, then the 4th one of definition for inner product would not hold.

Example 6.1.3

Let V=C([0,1]), the vector space of real-valued continuous functions on [0,1]. For f,gV, define

f,g=01f(t)g(t)dt.
  • Since the proceding integral is linear in f, the 1st and the 2nd parts are immediate.
  • the 3rd one is trival (real-value).
  • If f0, then f2 is bounded away from zero on some subinterval of [0,1] (continuity is used here), and hence f,f=01[f(t)]2dt>0.

conjugate transpose

Definition: conjugate transpose

Let AMm×n(F). Define the conjugate transpose or adjoint of A as the n×m matrix A such that (A)ij=Aji for all i,j.

Example 6.1.4

If

A=[i1+2i23+4i],

then

A=[i212i34i].

Notice that:

  • if x,y are viewed as column vector in Fn, then x,y=yx.
  • The conjugate transpose of a matrix plays a very important role in the remainder of this chapter. In the case that A has only real entries, A is simply the transpose of A.

Example 6.1.5

Let V=Mn×n(F), and define A,B=tr(BA) for A,BV. (Recall that the trace of a matrix A is defined by tr(A)=i=1nAii.) We verify that the 1st and 4th of the definition of inner product hold and leave (b) and (c) to the reader. For this purpose, let A,B,CV. Then

A+B,C=tr(C(A+B))=tr(CA+CB)=tr(CA)+tr(CB)=A,C+B,C.

Also

A,A=tr(AA)=(AA)ii=i=1nk=1n(A)ikAki=i=1nk=1nAkiAki=i=1nk=1n|Aki|2.

Now if A0, then Aki0 for some k and i. So A,A>0.

inner product space

The inner product on Mn×n(F) in Example 6.1.5 is called the Frobenius inner product.

A vector space V over F endowed with a specific inner product is called an inner product space. If F=C, we call V a complex inner product space, whereas if F=R, we call V a real inner product space.

For the remainder of this chapter, Fn denotes the inner product space with the standard inner product as defined in Example 6.1.1. Likewise, Mn×n(F) denotes the inner product space with the Frobenius inner product as defined in Example 6.1.5. The reader is cautioned that two distinct inner products on a given vector space yield two distinct inner product spaces. For instance, it can be shown that both

f(x),g(x)1=11f(t)g(t)dtandf(x),g(x)2=01f(t)g(t)dt

are inner products on the vector space P(R). Even though the underlying vector space is the same, however, these two inner products yield two different inner product spaces. For example, the polynomials f(x)=x and g(x)=x2 are orthogonal in the second inner product space, but not in the first. About orthogonal, introduce it later.

A very important inner product space that resembles C([0,1]) is the space H of continuous complex-valued functions defined on the interval [0,2π] with the inner product

f,g=12π02πf(t)g(t)dt.

Show that the vector space H with , defined above is an inner product space.

Check the condition one by one.

f+h,g=12π02π(f(t)+h(t))g(t)dt=12π02πf(t)g(t)dt+12π02πh(t)g(t)dt=f,g+h,g.cf,g=12π02π(cf(t))g(t)dt=c12π02πf(t)g(t)dt=cf,g.f,g=g,f.f,f=12π02π|f(t)|2dt>0if f0.

At this point, we mention a few facts about integration of complex-valued functions.

  1. the imaginary number i can be treated as a constant under the integration sign.
  2. every complex-valued function f may be written as f=f1+if2, where f1 and f2 are real-valued functions. Thus we have
f=f1+if2andf=f¯.

Theorem 6.1

Let V be an inner product space. Then for x,y,zV and cF, the following statements are true.

  • x,y+z=x,y+x,z.
  • x,cy=cx,y.
  • x,0=0,x=0.
  • x,x=0 if and only if x=0.
  • If x,y=x,z for all xV, then y=z.

The 1st and 2nd of Theorem 6.1 show that the inner product is conjugate linear in the second component.

Norms

Definition: Norm / Length

Let V be an inner product space. For xV, we define the norm or length of x by x=x,x.

Example 6.1.6

Let V=Fn. If x=(a1,a2,,an), then

x=(a1,a2,,an)=i=1n|ai|2

is the Euclidean definition of length. Note that if n=1, we have a=|a|.

Theorem 6.2

Theorem 6.2

Let V be an inner product space over F. Then for all x,yV and cF, the following statements are true.

  • cx=|c|x.
  • x=0 if and only if x=0. In any case, x0.
  • (Cauchy-Schwarz Inequality) |x,y|xy.
  • (Triangle Inequality) x+yx+y.

Example 6.1.7

For Fn, we may apply 3rd and 4th of Theorem 6.2 to the standard inner product to obtain the following well-known inequalities:

|i=1naib¯i|[i=1n|ai|2]1/2[i=1n|bi|2]1/2

and

[i=1n|ai+bi|2]1/2[i=1n|ai|2]1/2+[i=1n|bi|2]1/2.

Orthogonal [core]

Definitions: orthogonal & orthonormal

Let V be an inner product space. Vectors x and y in V are orthogonal (perpendicular) if x,y=0.

A subset S of V is orthogonal if any two distinct vectors in S are orthogonal. A vector x in V is a unit vector if x=1. Finally, a subset S of V is orthonormal if S is orthogonal and consists entirely of unit vectors.

Note that if S={v1,v2,} (infinite dimension), then S is orthonormal if and only if vi,vj=δij, where δij denotes the Kronecker delta. Also, observe that multiplying vectors by nonzero scalars does not affect their orthogonality and that if x is any nonzero vector, then (1/x)x is a unit vector. The process of multiplying a nonzero vector by the reciprocal of its length is called normalizing.

Example 6.1.8

In F3, {(1,1,0),(1,1,1),(1,1,2)} is an orthogonal set of nonzero vectors, but it is not orthonormal; however, if we normalize the vectors in the set, we obtain the orthonormal set

{12(1,1,0),13(1,1,1),16(1,1,2)}.

Example 6.1.9

Recall the inner product space H (defined above). We introduce an important orthonormal subset S of H.

For what follows, i is the imaginary number such that i2=1. For any integer n, let fn(t)=eint, where 0t2π. (Recall that eint=cosnt+isinnt.)

Now define S={fn:n is an integer}. Clearly S is a subset of H. Using the property that eit=eit for every real number t, we have, for mn,

fm,fn=12π02πeimteintdt=12π02πei(mn)tdt=0.

Also,

fn,fn=12π02π1dt=1.

In other words, fm,fn=δmn, the Kronecker delta.


6.2 The Gram-Schmidt Orthogonalization Process and Orthogonal Complements

Orthonormal basis

Definition: orthonormal basis

Let V be an inner product space. A subset of V is an orthonormal basis for V if it is an ordered basis that is orthonormal.

Example 6.2.1

The standard ordered basis for Fn is an orthonormal basis for Fn.

Example 6.2.2

The set {(1/5,2/5),(2/5,1/5)} is an orthonormal basis for R2.

Theorem 6.3

Theorem 6.3

Let V be an inner product space and S={v1,v2,,vk} be an orthogonal subset of V consisting of nonzero vectors. If yspan(S), then

y=i=1kaivi,

where

ai=y,vivi2for i=1,2,,k.

Corollary of Theorem 6.3

Corollary

  1. If, in addition to the hypotheses of Theorem 6.3, S is orthonormal and yspan(S), then
y=i=1ky,vivi.
  1. Let V be an inner product space, and let S be an orthogonal subset of V consisting of nonzero vectors. Then S is linearly independent.

For Corollary 1: If V possesses a finite orthonormal basis, then Corollary 1 allows us to compute the coefficients in a linear combination very easily.

Example 6.2.3

By Corollary 2, the orthonormal set

{12(1,1,0),13(1,1,1),16(1,1,2)}

obtained in Example 6.1.8 is an orthonormal basis for R3. Let x=(2,1,3). The coefficients given by Corollary 1 to Theorem 6.3 that express x as a linear combination of the basis vectors are

a1=x,v1v12=2+12,a2=21+33,a3=2+1+66.

As a check, we have

(2,1,3)=a1(1,1,0)+a2(1,1,1)+a3(1,1,2).

Gram-Schmidt process

Before stating this theorem, let us consider a simple case. Suppose that {w1,w2} is a linearly independent subset of an inner product space (and hence a basis for some two-dimensional subspace). We want to construct an orthogonal set from {w1,w2} that spans the same subspace.

gram-schmidt

Figure 6.1 suggests that the set {v1,v2}, where v1=w1 and v2=w2cw1, has this property if c is chosen so that v2 is orthogonal to w1. To find c, we solve

0=v2,w1=w2cw1,w1=w2,w1cw1,w1.

So

c=(w2,w1)w12.

Thus

v2=w2w2,w1w12w1

Theorem 6.4: Gram-Schmidt process [core]

Let V be an inner product space and S={w1,w2,,wn} be a linearly independent subset of V. Define S={v1,v2,,vn}, where v1=w1, and for 2kn,

vk=wkj=1k1wk,vjvj2vj.

Then S is an orthogonal set of nonzero vectors such that span(S)=span(S).

This construction of {v1,v2,,vn} by the use of Theorem 6.4 is called the Gram-Schmidt process.

Example 6.2.4 [core]

In R4, let w1=(1,0,1,0), w2=(1,1,1,1), and w3=(0,1,2,1). Then {w1,w2,w3} is linearly independent. We use the Gram-Schmidt process to compute orthogonal vectors v1,v2,v3, then normalize them to obtain an orthonormal set.

Take v1=w1=(1,0,1,0). Then

v2=w2w2,v1v12v1=(1,1,1,1)22(1,0,1,0)=(0,1,0,1).

Finally,

v3=w3w3,v1v12v1w3,v2v22v2=(0,1,2,1)22(1,0,1,0)22(0,1,0,1)=(1,0,1,0).

Normalization yields the orthonormal basis {u1,u2,u3} where

u1=v1v1=12(1,0,1,0),u2=v2v2=12(0,1,0,1),u3=v3v3=(1,0,1,0).

Example 6.2.5

Let V=P(R) with the inner product

f(x),g(x)=11f(t)g(t)dt,

and consider the subspace P2(R) with the standard ordered basis β={1,x,x2}. We use the Gram-Schmidt process to replace β with an orthogonal basis {v1,v2,v3} for P2(R), then obtain an orthonormal basis.

Take v1=1. Then v12=1112dt=2, and x,v1=11t1dt=0. Thus

v2=xv1,xv12=x02=x.

Furthermore,

x2,v1=11t21dt=23 and x2,v2=11t2tdt=0

Therefore

v3=x2x2,v1v12v1x2,v2v22v2=x21310x=x213.

We conclude that {1,x,x213} is an orthogonal basis for P2(R).

To obtain an orthonormal basis, we normalize v1,v2, and v3 to obtain

u1=11112dt=12u2=x11t2dt=32x

and similarly,

u3=v3v3=58(3x21).

Thus {u1,u2,u3} is the desired orthonormal basis for P2(R).

If we continue applying the Gram-Schmidt orthogonalization process to the basis {1,x,x2,} for P(R), we obtain an orthogonal basis whose elements are called the Legendre polynomials. The orthogonal polynomials v1,v2, and v3 in Example 6.2.5 are the first three Legendre polynomials.

Theorem 6.5

Theorem 6.5

Let V be a nonzero finite-dimensional inner product space. Then V has an orthonormal basis β. Furthermore, if β={v1,v2,,vn} and xV, then

x=i=1nx,vivi.

Example 6.2.6

We use Theorem 6.5 to represent the polynomial f(x)=1+2x+3x2 as a linear combination of the vectors in the orthonormal basis {u1,u2,u3} for P2(R) obtained in Example 6.2.5. Observe that:

f(x),u1=1112(1+2t+3t2)u1(t)dt=22,f(x),u2=1132t(1+2t+3t2)u1(t)dt=263,f(x),u3=11(1+2t+3t2)u1(t)dt=2105.

Therefore,

f(x)=22u1+263u2+2105u3.

Corollary of Theorem 6.5

Corollary

Let V be a finite-dimensional inner product space with an orthonormal basis β={v1,v2,,vn}. Let T be a linear operator on V, and let A=[T]β. Then for any i and j,

Aij=T(vj),vi.

Fourier coefficients [core]

Definition: Fourier coefficients

Let β be an orthonormal subset (possibly infinite) of an inner product space V, and let xV. We define the Fourier coefficients of x relative to β to be the scalars x,y where yβ.

Example 6.2.7

Let S={eint:n is an integer}. In Example 6.1.9, S was shown to be an orthonormal set in H. We compute the Fourier coefficients of f(t)=t relative to S. Using integration by parts, for n0,

f,fn=12π02πteintdt=in,

and for n=0,

f,1=12π02πtdt=π.

Orthogonal complement

Definition: orthogonal complement

Let S be a nonempty subset of an inner product space V. We define S (read "S perp") to be the set of all vectors in V that are orthogonal to every vector in S; that is,

S={xV:x,y=0 for all yS}.

The set S is called the orthogonal complement of S. It is easily seen that S is a subspace of V for any subset S of V.

Example 6.2.8

The reader should verify that {0}=V and V={0} for any inner product space V.

Example 6.2.9

If V=R3 and S={e3}, then S equals the xy-plane.

Let S0={x0}, where x0 is a nonzero vector in R3. Describe S geometrically. Now suppose S={x1,x2} is a linearly independent subset of R3. Describe S geometrically.

We may think of S as the plane orthogonal to x0, and if S spans a plane, then S is the line orthogonal to that plane.

Consider the problem in R3 of finding the distance from a point P to a plane W. (See Figure 6.2.) If we let y be the vector determined by 0 and P, we may restate the problem as follows:

Determine the vector u in W that is "closest" to y. The desired distance is clearly given by yu. Notice from the figure that the vector z=yu is orthogonal to every vector in W, so zW.

perp_plane

Orthogonal projection

Theorem 6.6

Let W be a finite-dimensional subspace of an inner product space V, and let yV. Then there exist unique vectors uW and zW such that y=u+z. Furthermore, if {v1,v2,,vk} is an orthonormal basis for W, then

u=i=1ky,vivi.

P.S. From Corollary 1 of Theorem 6.3:

u=i=1ku,vivi.

Corollary

In the notation of Theorem 6.6, the vector u is the unique vector in W that is "closest" to y; that is, for any xW, yxyu, and this inequality is an equality if and only if x=u.

The vector u is called the orthogonal projection of y onto W.

Example 6.2.10

Let V=P3(R) with inner product

f(x),g(x)=11f(t)g(t)dtfor all f,gV.

We compute the orthogonal projection f1(x) of f(x)=x3 on P2(R).

By Example 6.2.5, {u1,u2,u3} is an orthonormal basis for P2(R) with

{12,32x,58(3x21)}

Computing inner products:

f,u1=0,f,u2=11t332tdt=65,f,u3=0.

Hence

f1(x)=i=13f(x),uiui=35x.

Theorem 6.7

Theorem 6.7

Suppose that S={v1,v2,,vk} is an orthonormal set in an n-dimensional inner product space V. Then

  • S can be extended to an orthonormal basis {v1,v2,,vk,vk+1,,vn} for V.

  • If W=span(S), then

S1={vk+1,vk+2,,vn}

is an orthonormal basis for W.

  • If W is any subspace of V, then
dim(V)=dim(W)+dim(W).

Example 6.2.11

Let W=span({e1,e2}) in F3. Then x=(a,b,c)W if and only if

x,e1=0,x,e2=0.

So x=(0,0,c), and therefore W=span({e3}). Thus e3W and from 3rd claim of Theorem 6.7, that dim(W)=32=1.

Released under the MIT License.