Skip to content

David. C. Lay: Linear Algebra and its Applications", Addison-Wesley.

5.3 Theorem 5

The Diagonalization Theorem: An n×n matrix A is diagonalizable if and only if A has n linearly independent eigenvectors.

First, prove that

An n×n matrix A is diagonalizable A has n linearly independent eigenvectors.

A=PDP1D=[λ1000λ2000λn]P=[v1v2vn]PD=[v1v2vn][λ1000λ2000λn]=[λ1v1λ2v2λnvn]A=A×I=A(PP1)=(AP)P1=APP1APP1=PDP1APP1P=PDP1PAP(P1P)=PD(P1P)AP×I=PD×IAP=PDAP=A[v1v2vn]=[Av1Av2Avn][Av1Av2Avn]=[λ1v1λ2v2λnvn]Avi=λivi

vi is the ith eigenvactor of A and λi is the corresponding eigenvalue.

Because P is invertible, [v1v2vn] is linearly independent.

Then, prove that

A has n linearly independent eigenvectors An n×n matrix A is diagonalizable.

Let A=PBP1.

A is similar to B

Let B=[λ1000λ2000λn]

Where λi is the ith eigenvalue of A.

Avi=λivi, vi is the corresponding eigenvector.

P is some matirx. But such P may not exit, resulting A=PBP1 is invalid.

So we just need to find a P.

Let P=[p1p2pn], where pi is the ithe column vector.

PB=[p1p2pn][λ1000λ2000λn]=[λ1p1λ2p2λnpn]

Construct the same euqation again.

A=A×I=A(PP1)=(AP)P1=APP1APP1=PBP1APP1P=PBP1PAP(P1P)=PB(P1P)AP×I=PB×IAP=PB

Let A=[a1a2an], where ai is the ithe raw vector.

AP=[a1a2an][p1p2pn]=[a1p1a1p2a1pna2p1a2p2a2pnanp1anp2anpn]AP=PB[a1p1a1p2a1pna2p1a2p2a2pnanp1anp2anpn]=[λ1p1λ2p2λnpn][a1pia2pianpi]=λipi

Let ai=[ai1ai2ain].

Let pi=[p1ip2ipni].

ajpi=[aj1aj2ajn][p1ip2ipni]

By the definition of matrix multiplication,

[a1pia2pianpi]=[a11a12a1na21a22a2nan1an2ann][p1ip2ipni]=[a1a2an]Pi=ApiApi=λipi

Thus, each p is an eigenvector of A.

Because n eigenvectors of A are linearly independet, P is invertible.

A=PBP1 is indeed valid.

6.1 Theorem 3

Let A be an m×n matrix. The orthogonal complement of the row space of A is the null space of A, and the orthogonal complement of the column space of A is the null space of AT : (RowA)=NulA and (ColA)=NulAT

Let A=[a1a2an], and xNulA.

By the definition of Null Space,

Ax=0[a1a2an]x=[000]aix=0

Because ai is the row vector and x is the column vector, aix is the the inner product of the row vector and any vector from Null Space, resulting the linear combinations of row vector are also perpendicular to Null Space.

Thus, RowA=(NulA), namely (RowA)=NulA.

And we can get (ColA)=NulAT, by transposing both sides.

6.2 Theorem 7

Let U be an m×n matrix with orthonormal columns, and let x and y be in Rn. Then

a. Ux=x

Ux=UxUx=(Ux)TUx=xTUTUx=xT(UTU)x=xTIx=xTxx=xx=xTx

b. (Ux)(Uy)=xy

(Ux)(Uy)=(Ux)T(Uy)=xTUTUy=xT(UTU)y=xTIy=xTyxy=xTy

c. (Ux)(Uy)=0xy=0

This is just a special case of b.

6.3 Theorem 8

The Orthogonal Decomposition Theorem: Let W be a subspace of Rn. Then each y in Rn can be written uniquely in the form y=y^+z (1) where y^ is in W and z is in W. In fact, if {u1,,up} is any orthogonal basis of W , then y^=yu1u1u1u1++yupupupup (2) and z=yy^.

The textbook provides the proof. Therefore, we only show how to derive formula (2) here.

(yy^)y^=0yy^y^y^=0yy^=y^y^

Let y^=c1u1++cpup.

Namely,

y^=[u1u2up][c1c2cp]yy^=yTy^=yT[u1u2up][c1c2cp]y^y^=(c1u1++cpup)(c1u1++cpup)=(c1)2u1u1++(cp)2upup=[c1(u1u1)cp(upup)][c1c2cp]

Apply yy^=y^y^:

yT[u1u2up][c1c2cp]=[c1(u1u1)cp(upup)][c1c2cp]yT[u1u2up]=[c1(u1u1)cp(upup)]

By the definition of matrix multiplication,

ci(uiui)=yTui=yuici=yuiuiui

i ranges from 1 to p.

6.5 Theorem 15

Given an m×n matrix A with linearly independent columns, let A=QR be a QR factorization of A as in Theorem 12. Then, for each b in Rm, the equation Ax=b has a unique least-squares solution, given by x^=R1QTb^

The textbook provides the proof. We only show how to derive formula (6) here.

Ax^=b^(QR)x^=b^(QR)1(QR)x^=(QR)1b^Ix^=R1Q1b^x^=R1QTb^

Because Q has orthonormal columns,

QTQ=I=Q1QQT=Q1