Characteristic Polynomial and Eigenvalues
Characteristic Polynomial and Eigenvalues
Table of Contents
CHAPTER ONE ........................................................................................................................................... 1
THE CHARACTERISTIC EQUATION OF A MATRIX ........................................................................... 1
1.1 Eigenvalues and Eigenvectors ............................................................................................................ 1
1.2. The Characteristic Polynomial ........................................................................................................... 5
1.3. Similarity of Matrices and Characteristic Polynomial ..................................................................... 10
1.4. The Spectral radius of a matrix ........................................................................................................ 12
1.5. Diagonalization ................................................................................................................................ 13
1.6. Decomposable Matrices ................................................................................................................... 15
1.7. Minimal Polynomial and Cayley-Hamilton Theorem...................................................................... 17
CHAPTER TWO ........................................................................................................................................ 22
ORTHOGONALITY .................................................................................................................................. 22
2.1. The Inner Product ....................................................................................................................... 22
2.2. Inner Product space ..................................................................................................................... 22
2.3. Orthonormal Sets ........................................................................................................................ 25
2.4. The Gram-Schmidt Orthogonalization process ........................................................................... 26
2.5. Cauchy-Schwarz and Triangular Inequalities. ............................................................................ 28
2.6. The Dual Space ........................................................................................................................... 30
2.7. The Adjoint of Linear Operator .................................................................................................. 32
2.8. Self- adjoint Linear Operators..................................................................................................... 33
2.9. Isometry ...................................................................................................................................... 37
2.10. Normal Operators and the Spectral Theorem.......................................................................... 38
2.11. Factorization of a matrix / LU, Cholesky, QR/ ....................................................................... 39
2.11.1 LU- factorization ................................................................................................................. 40
2.11.2 Cholesky factorization ............................................................................................................ 42
2.11.2 QR- factorization................................................................................................................. 44
2.11.4 Singular Value Decomposition ........................................................................................... 48
CHAPTER THREE .................................................................................................................................... 51
CANONICAL FORMS............................................................................................................................... 51
3.1. Elementary row and column operation on Matrices ................................................................... 51
I
Linear Algebra II
II
Linear Algebra II
CHAPTER ONE
A linear mapping F:V → V’ is a mapping which satisfies the following two properties.
F(u+v)= F(u)+F(v)
F(cv) =cF(v).
If we wish to specify the field K, we also say that F is K-linear. Since we usually deal with a
fixed field K, we omit the prefix K, and say simply Linear.
Example-1:
Let V be any vector space. The mapping which associates to any element u of V this element
itself is obviously a linear mapping, which is called the identity mapping .We denote it by id or
simple I. Thus id(u)=u.
Let V, W be vector spaces over K, and let F:V → W be a linear map. The set of elements
Example-2:
P(x,y,z)=(x,y).
Then the Kernel of P is a linear map whose first two coordinates are equal to 0, i.e. all vectors
(0,0,z) with arbitrary component z.
1
Linear Algebra II
Example-3:
3 0
a) Consider M= [ ] Mat2*2 (R) . We have :
0 2
3 0 1 1
[ ] [ ] = 3 [ ] and
0 2 0 0
3 0 0 0
[ ] [ ] = 2 [ ].
0 2 1 1
3 1
b) Consider M= [ ] Mat2*2 (R) . We have :
0 2
3 1 a a
[ ] [ ] = 3 [ ] for all a R.
0 2 0 0
3 1 a 3a + b a
[ ][ ] = [ ] = 2 [ ] a=-[Link] for all a R. We have
0 2 b 2b b
3 1 a a
[ ] [ ] = 2 [ ].
0 2 −a −a
5 1
c) Consider M= [ ] Mat2*2 (R) . We have :
−4 10
5 1 1 1
[ ] [ ] = 9 [ ] and
−4 10 4 4
5 1 1 1
[ ] [ ] = 6 [ ].
−4 10 1 1
0 1
d) Consider M= [ ] Mat2*2 (R) . We look at
−1 0
0 1 a b a
[ ] [ ] = [ ]. This vector is equal to [ ] iff b= a and a =- b. This gives
−1 0 b −a b
a 0
a= - a. Thus there is no R with this property if [ ] ≠ [ ].
2
b 0
We will study these phenomena in general .Given a linear operator T:V → V, it is important in
many problems in science and mathematics to determine those scalars for which T(v)= v
has non-zero solutions.
Definition 1.1:
Let T:V → V be a linear operator on a vector space V over a field K. Then an element v V is
called an eigenvector of T if there exists K such that T(v)= v. is called an eigenvalue of T
belonging to the eigenvector v. we also say that v is an eigenvector with eigen value .
2
Linear Algebra II
Examples:
1. Let V=R2 and consider f(x,y) =(y,x). To determine the eigenvectors and eidenvalues by the
above definition
f(x,y) =(y,x)
= (x,y)=( x, y)
y = x and x = y
y- x =0.
Y- ( y) =0.
y- 2 y =0.
y(1- 2 )=0.
1- 2 =0 or y=0.
=1,-1.
Then 1 and -1 are eigenvalues of f, and
E1(f) = {(x,x):x R},
E-1(f) = {(x,-x):x R}.The eigenvectors (1,1) and (1,-1) form a basis of V.
2. Let V be the vectorspace over R generated by all infinitely differentiable functions. Let
t
R. Then the function f such that f(t) = e is an eigenvector of the derivative d/dt
because df/dt = e .And its eigenvalue .
t
Note :
Example:
3
Linear Algebra II
a11 ⋯ 0
A = [ ⋮ ⋱ ⋮ ] is a diagonal matrix in K. Then every unit vector (i=1,…,n) is an
0 ⋯ ann
eigenvector of A. In fact, we have: Aei = aiei.
Note that:
1. One of the meanings of the word “eigen“ in German is “proper.” Thus eigenvalues are also
called proper values or characteristics values or latent roots.
2. If v≠0, then is uniquely determined, because:
1v= 2v.
1= 2.
3. If v and w are eigenvectors with eigenvalue ,then:
i) V+w is also an eigenvector with eigenvalue , because :
T(v+w) =T(v) + T(w) =λv + λw = λ(v+w).
ii) Each scalar multiple kv is also an eigenvector with eigenvalue λ, because:
T(kv) =kT(v)=k(λv) = (kv).
4. Let T:V → V be a linear operator with ker T≠{0}.Let v be any non-zero vector in ker T. Then
T(v) =0=0.v.
So, v is an eigenvector of T with eigenvalue 0.
Examples:
4
Linear Algebra II
Theorem -1.1:
Vλ is a subspace of V.
Proof:
AX = λ X.
Note:
5
Linear Algebra II
(A - λ I)X= 0. …………..(*)
a11 − λ ⋯ a1n x1 0
[ ⋮ ⋱ ⋮ ] [ ⋮ ]=[ ⋮ ].
an1 ⋯ ann − λ xn 0
X=0 is the trivial solution of (*).
Further solutions will exist iff |A − λIn |=0. Hence, solving the equation |A − λIn |=0
gives the eigenvalue of A.
Note:
1 −1 3
A= [−2 1 1 ].
0 1 −1
Solution:
λ−1 1 −3
PA (λ)= | 2 λ − 1 −1 |=|λIn − A| which we can expand according to the first
0 −1 λ + 1
column, to find
PA (λ)= λ3 - λ2 - 4 λ +6.
1 6
Example-2: A= [ ] . Find the eigenvectors and eigenvalues of A.
5 2
Solution:
6
Linear Algebra II
Or |A − λIn | = 0
1 6 1 0
|( ) − λ( )|= 0.
5 2 0 1
1−λ 6
| |= 0.
5 2−λ
λ2 - 3 λ -28=0.
λ= 7 or λ=-4.
a) Let λ=7.
Then (A − 7I2 )X =0.
1 6 1 0 x 0
[( ) − 7( )] [ ] = [ ].
5 2 0 1 y 0
−6 6 x 0
( ) [ ] = [ ].
5 −5 y 0
y=x.
Hence, any vector of the type (1,1), where isa non-zero scalar, is an eigenvector
corresponding to the eigenvalue 7.
b) Let λ= - 4.
Then (A + 4I2 )X =0.
1 6 1 0 x 0
[( ) + 4( )] [y] = [ ].
5 2 0 1 0
5 6 x 0
( ) [ ] = [ ].
5 6 y 0
5x + 6y= 0.
−5
y= 6
x.
−5
Hence, any vector of the type (1, ), where is a non-zero scalar, is an eigenvector
6
corresponding to the eigenvalue - 4.
Note:
Justification
7
Linear Algebra II
3 −2 0
Example-3: Let A= [−2 3 0].Find bases for eigenspace of A.
0 0 5
Solution:
Characteristic equation of A:
3 − λ −2 0
| −2 3 − λ 0 |=0.
0 0 5−λ
(3 − λ) (3 − λ) (5 − λ) - 4(5 − λ) =0.
[(3 − λ)2 – 4] (5−λ)=0.
(λ2- 6 λ +5)( λ-5) = 0.
( λ − 1)( λ - 5) 2= 0.
So the eigenvalue of A:
λ = 1 and λ= 5.
x
From the definition, X =[y]is an eigenvector corresponding to λ iff X is non-trivial solution of
z
(A − λI3 )X =[Link] is,
3 − λ −2 0 x 0
[ −2 3 − λ y
0 ] [ ] = [0]…………………….(*)
0 0 5−λ z 0
a) Let λ=1. Then (*) becomes
3 − 1 −2 0 x 0
[ −2 3 − 1 0 ] [y] = [0]
0 0 5−1 z 0
2 −2 0 x 0
[−2 2 0] [y] = [0].
0 0 4 z 0
2x-2y = 0 and 4z = 0.
x= y and z=0.
Thus, the eigenvectors corresponding to eigenvalue 1 are non-zero vectors of the form:
x 1
X =[ x]. Therefore, {[1]} is a base for V1 = The eigenspace of A corresponding to
0 0
eigenvalue 1.
b) Let λ=5. Then (*) becomes
8
Linear Algebra II
3 − 5 −2 0 x 0
[ −2 3 − 5 0 ] [y ] = [ 0]
0 0 5−5 z 0
−2 −2 0 x 0
[−2 −2 0] [y] = [0].
0 0 0 z 0
-2x-2y = 0 .
x= - y .
Thus, the eigenvectors corresponding to eigenvalue 5 are non-zero vectors of the form:
x x 0 1 0
−x −x
X =[ ]= [ ] + [0]= x [−1] + z [0].
z 0 z 0 1
1 0
Therefore, {[−1] , [0]} is a base for V5 = The eigenspace of A corresponding to eigenvalue 5.
0 1
Note:
If v1, v2 are eigenvectors of a linear operator T:V → V, with eigenvalues λ1≠ λ2 , then v1 + v2
isnot an eigenvector of T.
Theorem 1.1.3: Let v1, v2,…, vn be eigenvectors of a linear operator T:V → V, with eigenvalues
λ1, λ2,…, λn respectively. If λi≠ λj , then { v1, v2,…, vn } is Linearly independent.
Proof: By induction on n.
If n=1, an element v1 ∈V, v1 ≠ 0, is linearly independent.
Assume n >1.
Suppose α1 v1+α2 v2+…,+αn vn =0, with αi ∈ K. ----------------------- (1)
WTS: αi =0, for all i.
Multiplying equation (1) by λ1 , we obtain:
i. λ1 α1 v1+ α2 v2 λ1 +…,+αn vn λ1 =0
We apply T to equation (1) , we obtain:
ii. λ1 α1 v1+ α2 v2 λ2 +…+αn vn λn =0.
Subtracting (i) from (ii), we obtain:
α2 v2 (λ2- λ1) + α3 v3 (λ3- λ1) +…,+αn vn( λn- λ1) =0.
Since λj- λ1 ≠ for j = 2,3,…,n. We conclude by induction that
α2 = α3 =…=αn = 0.
Going back to the original relation, we see that α1 v1 =0. Hence α1 =0, and the theorem is
proved.
Exercise - 1.1:
9
Linear Algebra II
1. Find the eigenvalues and associated eigenspaces of each of the following matrices.
0 1 0 1 1 1
1 1
a) A=[ ] b) A = [0 0 1] c) A = [0 2 1]
−2 3
0 0 0 0 0 1
2. Let A be an n*n matrix. Prove that A is singular if and only if λ=0 is an eigenvalue of A.
1
3. Let A be a non-singular matrix and let λ be nonzero eigenvalue of A. Show that is an
λ
eigenvalue of A−1.
4. Let A be n*n matrix. Show that A and AT have the same eigenvalues. Do they also share
eigenvectors?
5. Show that the eigenvalues of a triangular matrix are the diagonal elements of the matrix.
Let A and B be n*n matrices over a field K. We say that B is similar to A over K if there is an
invertible n*n matrix P over K such that B =P-1AP./PB=AP/
Notation: B∼A.
1 2 1 0
Example: Let A= [ ] and B = [ ], show that B∼A.
0 −1 −2 −1
1 2 1 −1 3 1 1 −1 1 0
AP=PB → [ ][ ]=[ ]=[ ][ ].
0 −1 1 1 −1 −1 1 1 −2 −1
Properties of similar matrices
1. A is similar to itself.
2. If B∼A, then A∼ B.
→ C∼A.
10
Linear Algebra II
Proof: / of the first two only/. Suppose that A and B are similar,i.e. that B= P-1AP for some
matrix P .
a) Det(B) = det(P-1AP)
= det(P-1)det(A)det(P).
= [det(P)]-1det(A)det(P) = det(A).
So the matrices have the same determinant, and one is invertible if the other is.
b) Det(B-λI)=det(P-1AP- λP-1P)
= det[P-1(A- λI)P]
= det(P-1)det(A- λI)det(P).
= det(A- λI).
So the matrices have the same characteristic polynomial and hence the same eigenvalues.
1 2 2 1
i) Let A= [ ] and B = [ ] are not similar since det(A)=-3 but det(B)=3.
2 1 1 2
11
Linear Algebra II
1 3 1 1
ii) Let A= [ ] and B = [ ] are not similar since the characteristic polynomial of
2 2 3 −1
A is λ2-3λ-4 while B has λ2-4.
Activity
1 0 1 1
Consider A= [ ] and B = [ ] both have determinant 1 and rank 2, are invertible and have
0 1 0 1
characteristic polynomial (1- λ)2 and eigenvalues λ1= λ2=[Link] these two matrices similar?
Theorem 1.3.1:
An eigenvalue of A that is larger in absolute value than any other eigenvalue is called a
dominant eigenvalue.
9 −1 2
Example: Consider the matrix A = [−2 8 4] whose eigenvalues are 5,10,10.
1 1 8
σ(A)= {5,10}.
ρ(A)= max{5,10}.
Exercise-1.3:
5 1 2 1 0 1
i) A=[ ] ii) A=[ ] iii) A=[ ]
−4 10 0 2 −1 0
12
Linear Algebra II
1.5. Diagonalization
The best we can hope for when we are given a matrix is when it is similar to a diagonal matrix.
In fact there is a close relationship between when a matrix is diaonalizable and the eigenvalues
and eigenvectors of a matrix.
1 3 1 3
Example: The matrix A =[ ] is diagonalizable since the matrix P= [ ] and
2 2 1 −2
4 0
D=[ ] produce the identity P-1AP=D or equivalently AP=PD.
0 −1
This is wonderful, but we have no idea where P and D arose from .To answer this question we
note that the diagonal entries 4 and -1 of D are the eigenvalues of A . How P is found is a more
interesting question, as in the case of D, the entries of P are related to the eigenvectors of A.
Theorem: Let A be an n*n matrix, then A is diagonalizable if and only if A has n linearly
independent eigenvectors.
To be precise , there exists an invertible matrix P with a diagonal matrix D such that P-1AP=D if
and only if the columns of P are linearly independent eigenvectors of A and the diagonal entries
of D are the eigenvalues of A corresponding to the eigenvectors in P in the same order.
Corrollary:
Let A be 2*2 matrix having distinct eigenvalues λ1, λ2 and corresponding eigenvectors X1 and
X2 .Let P be the matrix whose columns are X1 and X2 respectively. Then P is non-singular and P-
λ 0
1
AP = [ 1 ].
0 λ2
2 1 −1 1
Example-1: Let A= [ ]. Then X1 = ( ) and X2 = ( ) are eigenvectors assoiciated to the
1 2 1 1
−1 1 1 0
eigenvalues 1 and 3 respectively. Hence if P=[ ],we have P-1AP = [ ].
1 1 0 3
0 1 0
Example-2: Determine whether a matrix P exists to diagonalizable A = [0 0 1].
2 −5 4
Solution: We have seen that is matrix has eigenvalues λ1 = λ2 =1 and λ3 =2 with the
corresponding eigenspaces
1 1
E1 = span{[1]}, E2 = span{[2]}.
1 4
13
Linear Algebra II
As all other eigenvectors are just multiples of one of these two basis vectors there cannot be
three linearly dependent eigenvectors. Thus it cannot be diagonalized.
−1 0 1
A = [ 3 0 −3].
1 0 −1
Solution: The eigenvalues λ1 = λ2 =0 and λ3 = -2 with the basis for the eigenspaces.
0 1 −1
E0 = span {[1] , [0]}, E-2 = span{[ 3 ]}.
0 1 1
It is easy to check that the three vectors are linearly independent, and so we from P.
0 1 −1
P = [1 0 3 ] this matrix will be invertible and furthermore
0 1 1
0 0 0
P-1AP = [0 0 0 ]=D.
0 0 −2
Properties of Diagonalize a matrix
Proof: Let v1, v2,…, vn be eigenvectors corresponding to the n distinct eigenvalues of A. These
vectors linearly independent by previous theorem. A is diagonalizable.
2 −3 7
Example: The matrix A = [0 5 1 ] has eigenvalues λ1 = 2, λ2 =5 and λ3 = -1, as these
0 0 −1
eigenvalues are distinct for the 3*3 matrix A, A is diagonalizable.
2. To calculate An ./A2*2 /
λ 0 λ 0 -1
If P-1AP = [ 1 ], then A = P [ 1 ] P and
0 λ2 0 λ2
n λ1 0 -1 n λ1 0 n -1 λ1 n 0
A = [P [ ]P ] = P[ ] P =P[ ] P-1
0 λ2 0 λ2 0 λ2 n
0 1
Example: Compute A10 if A= [ ].
2 1
Solution: Here
14
Linear Algebra II
−1
λ1 = -1 → v1 = ( )
1
1
λ2 =2 → v2 = ( ).
2
1 1 −1 0
→P = [ ] and D = [ ].
−1 2 0 2
2 −1
10
λ 0 -1 10 −1 0 10 -1 1 1 (−1) 0
A = [P [ 1
10
]P ] = P[ ] P =[ ] [ ] [3 3
].
0 λ2 0 2 −1 2 0 (2)10 1 1
3 3
342 341
=[ ].
682 683
Let A be a matrix over a field K. Partition A in to parts means of a system of horizontal and
vertical lines.
1 2 3 4 1 2 3 4 1 2 3 4
Example: A= [5 6 7 8] = [5 6 7 8 ] = [5 6 7 8].
9 0 1 3 9 0 1 3 9 0 1 3
The parts are smaller matrices and are called blocks of the matrix A.
Given a square matrix A, it is often necessary to partition in to blocks so that the diagonal blocks
are square matrices.
A1 0 … 0
0 A2 … 0
A block matrix of the form: A = [ ]. Where A1 , A2 ,…, Ar are square matrices
⋮ ⋮ ⋱ ⋮
0 0 ⋯ Ar
and each zero is a zero matrix of proper dimensions, is called a diagonal block matrix.
Note:
1. We say that :
i. A is decomposed into blocks A1 , A2 ,…, Ar .
15
Linear Algebra II
Or
Then
2 −3 7 2 −3 7
Example-1: Let A = [0 5 1 ] , B = [0 5 1 ]. Then it is not difficult to see that
0 0 −1 0 0 −1
1 2 6 7 0 22 25 0
{[ ][ ]} [ ]
AB= diag {[ 3 4 8 9 0 ]} = [ 50 57 0].
0 0 5∗1 0 0 5
1 3 0 0
Example-2: Find |A|. Let A= [0 2 0 0 ].
0 0 4 3
0 0 −1 1
1 3 4 3
Solution: A1 = | |=2. A2 = | |=4+3=7.
0 2 −1 1
|A|= |A1 ||A2 | = 2*7=14.
1 3 0
Example-3: Find A-1 . Let A = [2 2 0] .
0 0 1
1 3
Solution : A1 = [ ] ; A2 = [1].
2 2
−1 3
−1 2 4
A1 = [ 1 −1]; A2 −1 = [1].
2 4
−1 3
0
2 4
-1
A =[ 1 −1
0].
2 4
0 0 1
16
Linear Algebra II
Proof: Let A = diag(A1 , A2 ,…, Ar ), where Ai is a square matrix of order ni for i= 1,2,…,r.
Then, A − λIn = diag (A1 -λIn1 , A2 -λIn2 , …, Ar − λInr ), where Ini is an identity matrix of order
ni .
1 3 0
Example-4: Determine the characteristics polynomial of A = [2 2 0].
0 0 1
1−λ 3
Solution: 𝒳A1 (λ) = | |=(1 − λ) (2 − λ)-6= λ 2-3λ-4.
2 2−λ
𝒳A2 (λ) =|1 − λ|= (1 − λ)
Exercise-1.4:
A1 B
1. Show that , suppose M= [ ] , where A1 and A2 are square matrices. Then
0 A2
A1 B … C
0 A2 … D
2. Suppose M = [ ], Where A1 , A2 ,…, Ar are square matrices.
⋮ ⋮ ⋱ ⋮
0 0 ⋯ Ar
Then, show that:
𝒳M (λ) = 𝒳A1 (λ) 𝒳A2 (λ)… 𝒳Ar (λ).
17
Linear Algebra II
Proof: All n*n matrices form a vectorspace of dimension n2. Hence, the matrices
2
I,A,A2 , …,An are linearly independent. Therefore, there exists scalars c0 , c1 , c2 ,…, cn2 , not all
2
zero, such that : c0 I+c1 A+c2 A2 + ⋯+cn2 An =0.
2
Let f(x) = c0 +c1 x+c2 x 2 + ⋯+cn2 x n .
Among all the non-zero polynomials annihilating A, consider those of least degree. By
multiplying by suitable non-zero constants, we can ensure that they are monic.
Let g(A)= f1 (A)- f2(A)=0.→← / This contradict the definition of f1 ,f2 as non-zero polynomials of
least degree annihilating A./
Thus, f1 = f2.
Definition 1.7.1:
Let A be an n*n matrix over a field K. The unique monic polynomial of least degree, which
annihilates A is called the minimial polynomial of A and is denoted by mA (x).
Theorem 1.7.1: Let A be an n*n matrix over a field K. A is non-singular if and only if the
constant term of mA (x) is non-zero.
Therefore, a0 ≠ 0.
18
Linear Algebra II
(←): Suppose a0 ≠ 0.
Hence, A is non-singular.
𝒳A (A)=0.
Where the Pij (x) are polynomials in x. B can be written in polynomial in x having matrix
coefficients. Let B= Bk x k +Bk−1 x k−1 + ….+ B1 x + B0 = f(x).
B(xI-A) = 𝒳A (x)I.
→(Bk x k +Bk−1 x k−1 + ….+ B1 x + B0 )(xI-A) = (an x n + an−1 x n−1 + ….+ a1 x + a0 )I.
→f(x) (xI-A) = q(x), where q(x)= (an I)x n + (an−1 I) x n−1 + ….+ (a1 I) x + a0 I.
→q(A)= 0,
=0.
Theorem 1.7.3:
19
Linear Algebra II
Proof: Let f(x) ∈ K[x] such that f(A) =0. By division Algorithm, there exist q(x), r(x) ∈ K[x]
such that : f(x) = q(x) mA (x) + r(x), where either r(x) = 0or else deg r(x) < deg mA (x) .
Example-1: Here are key examples to understand the difference between minimal and
characteristic polynomial. The following three matrices have the same characteristic
polynomials,(x-1)2 :
1 0 1 1 1 169
M1= [ ], M2 = [ ], M3 = [ ].
0 1 0 1 0 1
0 1 0 169
The minimal polynomial of M1 is [Link] M2 − I = [ ] ≠ 02 and M3 − I = [ ] ≠ 02 ,
0 0 0 0
the minimal polynomial has to be them.
1 4
Example-2: Verified Cayley Hamilton Theorem for A= [ ]. Find A−1.
2 3
Solution: We know that |A − λI2 | =0.
1−λ 4
| |= 0.
2 3−λ
λ2 -4λ -5= 0.
1 42 1 4 1 0
So, [ ] − 4[ ]-5[ ]
2 3 2 3 0 1
9 16 −4 −16 −5 0 0 0
[ ]+[ ]+[ ]= [ ].
8 17 −8 −12 0 −5 0 0
Cayley- Hamilton theorem is verified.
20
Linear Algebra II
A-4I-5A−1=0.
1
A−1 = 5 [A-4I].
1 1 4 1 0
= 5 [[ ]-4 [ ]].
2 3 0 1
1 −3 4
=5[ ].
2 −1
Exercise-1.5:
21
Linear Algebra II
CHAPTER TWO
ORTHOGONALITY
2.1. The Inner Product
Let V be a vector space over K where K=R or C.
Definition -2.1.1:
Note:
Let u= (u1 , u2 , … , un ) ∈ Rn .
v= (v1 , v2 , … , vn ) ∈ Rn .
Solution
i) u, u = ∑ni=1 ui ui = u1 u1 + u2 u2 + …. + un un .
22
Linear Algebra II
= u1 2 + u2 2 + … + un 2 ≥ 0.
and u, u = 0 iff u1 2 + u2 2 + … + un 2 = 0.
↔ u1 2 = u2 2 = … = un 2 = 0.
↔ ui =0,for all i=1,2,….
↔u = 0.
ii) u, v = ∑ni=1 ui vi = ∑ni=1 vi ui = v, u = v, u .
iii) Let w= (w1 , w2 , … , wn ) ∈ Rn .
v, u + w = 〈(v1 , v2 , … , vn ), (u1 + w1 , u2 + w2 , … , un + wn )〉.
= v1 (u1 + w1 ) + v2 (u2 + w2 ) + …+ vn (un + wn ).
= ( v1 u1 ++ v2 u2 + …+ vn wn ) + ( v1 w1 ++ v2 w2 + …+ vn wn )
= v, u + u, v .
iv) u, v = 〈(αu1 , αu2 , … , αun ), (v1 , v2 , … , vn )〉
= αu1 v1 + αu2 v2 + …+ αun vn .
=α[u1 v1 + u2 v2 +…+ un vn ].
= α u, v .
Example-2: Let f,g ∈ C[a,b] →C./ The space of continuous complex valued functions on [a,b].
b
̅̅̅̅̅ dt. Show that f , g inner product.
Define f , g = ∫ f(t)g(t)
a
b
i) ̅̅̅̅̅ =∫b|f|2dt ≥ 0.
f , f =∫a f(t)f(t) a
b
and ∫a |f|2 dt = 0 iff f(t)= 0.
b
ii) f ,g ̅̅̅̅̅ dt
=∫a f(t)g(t)
b
g, f =∫a g(t)f(t)̅̅̅̅̅ dt
b ̅̅̅̅̅̅̅̅̅̅
g, f =∫a g(t)f(t) ̅̅̅̅̅ dt.
b
̅̅̅̅̅ f(t) dt.
=∫a g(t)
b
̅̅̅̅̅ dt.
=∫a f(t)g(t)
Therefore, f , g = g, f
.
iii) Let α be a scalar
b
̅̅̅̅̅ dt =α ∫b f(t)g(t)
f , g =∫a αf(t)g(t) ̅̅̅̅̅ dt = α f , g
a .
iv) Let h ∈ C[a,b].
b ̅̅̅̅̅̅̅̅̅̅̅̅̅̅
f , g + h =∫a f(t)g(t) + h(t) dt
23
Linear Algebra II
b ̅̅̅̅̅ } dt
̅̅̅̅̅ + f(t)h(t)
=∫a {f(t)g(t)
b
̅̅̅̅̅ dt + ∫b f(t)h(t)
=∫a f(t)g(t) ̅̅̅̅̅ dt
a
= f , g + f ,h
.
Exercise-2.1:
Definition 2.2.2
Let V be an inner product space. Let v, w ∈ V. We say v and w are orthogonal, and denote v⊥w,
if v, w = 0.
Definition 2.2.3
Theorem 2.2.1
S ⊥ is a subspace of V.
Proof:
24
Linear Algebra II
→ v, w = 0 for all w ∈ S.
→ v, w = λ v, w = λ×0=0 for all w ∈ S.
→ λv ∈ S ⊥ .
𝐃𝐞𝐟𝐢𝐧𝐢𝐭𝐢𝐨𝐧 𝟐.2.4
v, w
Let V be a Euclidean space. Let v, w ∈ V, w ≠0. The number is called the Fourier
w, w
coefficient of v with respect to w.
Theorem 2.2.2
v, w
Let V be a Euclidean space. Let v, w ∈ V, w ≠0. Let λ = be the Fourier coefficient of v
w, w
with respect to w. Then, (v- λw ) ⊥ w.
= v, w - λ w, w
v, w
= v, w - w, w
w, w
=0.
Definition 2.2.5
v, w
Let V be a Euclidean space. Let v, w ∈ V, w ≠0. Let λ = be the Fourier coefficient of v
w, w
with respect to w. The vector λw is called the projection of v along w and is denoted by Projvw .
‖v‖ = √ v, v .
Note:
25
Linear Algebra II
Definition 2.3.2
If in addition, each vi is a unit vector, the basis {v1 , v2 , … , vn } is called an orthonormal basis.
Theorem 2.3.1
Let V be an inner product space. Let {v1 , v2 , … , vn } is called an orthogonal basis of V. Suppose
{x1 , x2 , … , xn } is the coordinate vector of v ∈ V with respect to the given orthogonal basis.
〈v,v 〉
Then, xi = 〈v ,vi 〉 = The Fourier coefficient of v with respect to vi .
i i
Proof:
V = ∑nj=1 xj vj .
= ∑nj=1 xj 〈vj , vi 〉
=xi ∑nj=1〈vj , vi 〉.
〈v,v 〉
Therefore, xi = 〈v ,vi 〉 .
i i
Then
(i) {u1 , u2 , … , ur } is linearly independent.
(ii) For any v ∈ V , the vector
w = v- 〈v, u1 〉u1 -〈v, u2 〉u2 - …-〈v, ur 〉ur is orthogonal to each of the ui .
Proof:
i) Suppose α1 u1 + α2 u2 , … + αr ur = 0.
Taking the inner product of both sides with respect to ui :
26
Linear Algebra II
Proof:
v
i) Set u1 = ‖v1‖ . Then { ui } is orthonormal.
1
w2
ii) Next, set w2 = v2 - 〈v2 , u1 〉u1 and u2 = ‖w2 ‖
.By Lemma , w2 /and hence u2 / is
orthogonal to u1 . Then {u1 , u2 } is orthonormal.
w
iii) Next, set w3 = v3 - 〈v3 , u1 〉u1 −〈v3 , u2 〉u2 and u3 = ‖w3‖ .By Lemma , w3 /and hence
3
u3 / is orthogonal to u1 and u2 .Then {u1 , u2 , u3 } is orthonormal.
iv) In general , after obtaining {u1 , u2 , … , ui }, set
wi+1 = vi+1 - 〈vi+1 , u1 〉u1 −〈vi+1 , u2 〉u2 - …- 〈vi+1 , ui 〉ui ; and
w
ui+1 = ‖wi+1 ‖. Then, {u1 , u2 , … , ui , ui+1 } is [Link] induction, we obtain an
i+1
orthonormal set {u1 , u2 , … , un }which is linearly independent and hence a basis for V.
Note:
The proof of the above theorem is also a method for finding an orthogonal basis given an
arbitrary basis. It is called the Gram-Schmidt orthonogonalization process.
Example-1
Consider R3 with the inner product is the dot product. Consider the basis {v1 , v2 , v3 }. Where
Solution
v (1,1,1) (1,1,1) 1 1 1
i) Set u1 = ‖v1‖ = ‖(1,1,1)‖ = =( , , ).
1 √3 √3 √3 √3
ii) Next, set w2 = v2 - 〈v2 , u1 〉u1
1 1 1 1 1 1
= (0,1,1) - 〈(0,1,1), ( , , )〉 ( , , ).
√3 √3 √3 √3 √3 √3
27
Linear Algebra II
2 1 1 1
= (0,1,1) – ( , , )
√3 √3 √3 √3
2 2 2 −2 1 1
=(0,1,1) – (3 , 3 , 3) = ( 3 , 3 , 3).
−2 1 1 −2 1 1
w (, , ) ( , , ) −2 1 1
and u2 = ‖w2‖ = 3 33
−2 1 1 = 3 33
=( , , ).
2 ‖( , , )‖ √
2 √6 √6 √6
3 33
3
−1 1 −1 1
w (0,
, ) (0, , ) −1 1
and u3 = ‖w3‖ = 2 2
−1 1 = 2 2
= (0, , ).
3 ‖(0, , )‖ √
1 √2 √2
2 2
2
Exercise-2.2:
‖λv‖ = |λ|‖v‖.
Proof: ‖v + w‖2 = 〈v + w, v + w〉
28
Linear Algebra II
=‖v‖2 + ‖w‖2 .
Proof:
Case 1: w=0.
Case 2: ‖w‖=1.
v, w
Let λ = = 〈v, w〉.
w, w
Therefore, λ2 ≤ ‖v‖2
→ |λ| ≤ ‖v‖.
Case 3: w ≠0.
1
Let u= ‖w‖ w.
1
→ ‖w‖ |〈v, w〉| ≤ ‖v‖.
29
Linear Algebra II
Proof: ‖v + w‖2 = 〈v + w, v + w〉
= (‖v‖ + ‖w‖)2 .
→ ‖v + w‖ ≤ ‖v‖ + ‖w‖.
Definition 2.6.1:
Notation: Let V* and v [Link] shall use the notation 〈 , v〉= (v).
Note:
1) 〈 + , v〉=〈 , v〉 + 〈 , v〉.
1 2 1 2
2) 〈 , v1 + v2 〉=〈 , v1 〉 + 〈 , v2 〉.
3) 〈λ , v〉=λ〈 , v〉.
4) 〈 , λv〉=λ〈 , v〉.
Here, in the symbole 〈 , v〉 the two components donot belong to the same space.
Example-1:
Lemma 1:
30
Linear Algebra II
Assume that V and W are vector spaces over a field K. Let {v1 , v2 , … , vn } be a basis of V. Let
w1 , w2 , … , wn be arbitrary elements of W. Then , there exists a unique linear map T:V→W such
that T(vi ) = wi for i=1,2,…,n.
Proof:
i) Existence
Let v ∈ V. Define T(v) = ∑ni=1 xi wi , where v = ∑ni=1 xi vi . Clearly, T is a linear map. T(vi ) = wi
for i=1,2,…,n.
ii) Uniqueness
Suppose T’: V→W is also a linear map such that T’(vi ) = wi for i=1,2,…,n.
Therefore, T’=T.
Theorem 2.6.1
Let V be finite dimensional vector space over a field K. Then dim V* = dim V.
Proof
Let {v1 , v2 , … , vn }be a basis of V. By Lemma 1, for each i=1,2,…,n, there exists a linear
functional vi ∗ , such that
1, ifi = j
〈vi ∗ , vj 〉 = / is called the Kronecker delta function/
0, i j
=αi 〈vi ∗ , vi 〉= αi =〈 , vi 〉.
31
Linear Algebra II
Definition 2.6.2
{ v1 ∗ , v2 ∗ ,…, vn ∗ } is called the dual basis of {v1 , v2 , … , vn }.Let V be an inner product space.
Then to each v ∈ V, we can associate linear functional Lv ∈ V* given by:
To check that
𝐓𝐡𝐞𝐨𝐫𝐞𝐦 𝟐. 𝟔. 𝟐
Let V be a finite dimensional inner product space. Let ∈ V*. Then there exists a unique v ∈V
such that = Lv .
Proof:
i) Existence
Let w ∈ V. Define :V→K by (v)= 〈T(v), w〉. Clearly is a linear functional on
[Link] theorem 2.6.2, there exist a unique v’∈ V such that = Lv ’.
Define T* : V→V by T*(w)= v’.
〈T(v), w〉= (v)= Lv ‘(v) = 〈v, v ′ 〉 = 〈v, T ∗ (w)〉.
Now, we need only show that T* is linear.
∗
a) 〈v, T (w1 + w2 )〉 = 〈T(v), w1 + w2 〉.
= 〈T(v), w1 〉 + 〈T(v), w2 〉
= 〈v, T ∗ (w1 )〉 + 〈v, T ∗ (w2 )〉.
=〈v, T ∗ (w1 ) + T ∗ (w2 ) 〉
Therefore, T ∗ (w1 + w2 ) = T ∗ (w1 ) + T ∗ (w2 ).
b) 〈v, T ∗ (λw)〉 = 〈T(v), λw)〉 =λ̅ 〈v, T ∗ (w)〉 = 〈v, λT ∗ (w)〉.
32
Linear Algebra II
Theorem 2.7.2:
Let V be a finite-dimensional inner product space. Let λϵK and let T,S be linear operators on V.
Then :
1. (T + S)∗ = T ∗ + S ∗ .
2. (λT)∗ = λ̅ T ∗ .
3. (T S)∗ = s ∗ T ∗ .
4. (T ∗ )∗ = T.
5. I ∗ = I.
6. 0∗ =0.
7. If T is invertible, then (T −1 )∗ = (T ∗ )−1 .
33
Linear Algebra II
i) A̿ = A.
ii) ̅λA
̅̅̅ = λ̅ A
̅.
iii) ̅̅̅̅̅̅̅
A+B=A ̅+B ̅/ provided A and B have the same sizes/
iv) ̅̅̅̅
AB = A ̅B̅
v) (A̅)t = A̅t ./ provided A is a square matrix /
𝐍𝐨𝐭𝐚𝐭𝐢𝐨𝐧: A* = A̅t .
i)
ii) Skew- Hermitian if A*=-A.
Note:
Theorem 2.8.3: Suppose A, B are unitary matrices of the same size. Then, At ,A-1 and AB are
also unitary.
Note:
34
Linear Algebra II
T = a linear operator on V.
→ T(v)=0,for all v ∈ V.
→ T=0.
T = a linear operator on V.
Proof:
Then,
35
Linear Algebra II
Proof:
Then, [T ∗ ]β = A*.
T(x,y,z)= (x+2y,3x-4z,y).
Find T*(x,y,z).
Solution:
T(1,0,0)=(1,3,0)
T(0,1,0)=(2,0,1)
T(0,0,1)=(0,-4,0).
1 2 0
Therefore, A=[T]= [3 0 −4].
0 1 0
36
Linear Algebra II
1 3 0
Therefore, A*=[T*]= [2 0 1].
0 −4 0
2.9. Isometry
Theorem 2.9.1: Let V= a finite dimensional inner product space.
i) T*=T-1./T*T=I/
ii) T preserves inner products, i.e., 〈T(v), T(w)〉 = 〈v, w〉 ;for all v,w ∈ V.
iii) T preserves length,i.e.,‖T(v)‖ = ‖v‖ ;for all v ∈ V.
Proof:
(iii)→(i):
Claim: T is invertible.
T(v)=0. → ‖T(v)‖=0.
→ ‖v‖=0.
→ v=0.
→ 〈(T ∗ T − I)(v), v〉 = 0.
→ T ∗ T − I = 0.
→ T ∗ T = I.
→ T ∗ = T −1.
𝐃𝐞𝐟𝐢𝐧𝐢𝐭𝐢𝐨𝐧 2.9.1: Let V be a finite dimensional inner product space. Let T be a linear operator
on V. T is called an isometry if it satisfies any one / and hence all/ of the three equivalent
conditions stated in theorem 2.9.1.
Note:
37
Linear Algebra II
↔ [ T*] =[T-1].
↔ [ T*] =[T]-1.
↔ A*=A-1.
38
Linear Algebra II
↔ T*(v) = λ̅v.
Proof:
𝒳T (λ)= (x-λ1 ) (x-λ2 )… (x-λn ) over C. Since T is self-adjoint each eigenvalues of T is real.
Therefore, the λi ’s are all real.
Definition: Matrix factorization is the process of factorize matrix as a product of lower and
upper triangular matrices and upper triangular and transpose factor of matrix A by different
methods.
39
Linear Algebra II
Note: Matrix factorization is not unique but it has many factors in different methods of
factorization.
That is : A=LU.
Definition: Let A be square matrix if there is a lower triangular matrix with all diagonal
elements are equal to 1/lii =1/ and upper triangular matrix such that A=LU then we say this is
LU- decomposition is Doolitteles Method.
Definition: Let A be square matrix if there is an upper triangular matrix with all diagonal
elements are equal to 1/uii =1/ and lower triangular matrix such that A=LU then we say this is
LU- decomposition is Crout’s Method.
1 2 4
A=(3 8 14) where L=lower triangular matrix U33 and U=upper triangular matrix.
2 6 13
40
Linear Algebra II
1 2 4 1 0 0 1 2 4
Therefore A=LU=(3 8 14)=(3 1 0) (0 2 2) is the result of LU-decomposition.
2 6 13 2 1 1 0 0 3
Definition:-LU-decomposition is used when the matrix is invertible and all its leading sub
matrices have non-zero determinant. Otherwise it does not work.
1 2 4
Example-2: A=(3 8 14)
2 6 13
1 2 4
1 2
Solution: A1 = 1, A2 = | |,A |3 8 14| and the matrix is invertible anda11= |A1 | = 1
3 8 3=
2 6 13
,|A2 | =(1×2) − (2 × 3) = 2
8 14 3 14 3 8
And |A3 |=| | -2| | +4| | =20-(2×11) + (4×2) =6
6 13 2 13 2 6
Therefore all determinants are non-zero then a matrix A has LU- decomposition.
41
Linear Algebra II
1 2 3
A=(2 4 5)
1 3 4
Solution:- The leading sub matrices are
1 2 3
1 2
A1 =1, A2 =( ) , A3 =(2 4 5)
2 4
1 3 4
|A1 |=1,|A2 |=(1×4) - (2×2)=0
2 4
|A3 |=|4 5 2
|-2|
5
|+3| |=1+6+6=13
3 4 1 4 1 3
Therefore the matrix A has no LU-decomposition because determinant of one sub matrices A2 is
0 by the fact we have above it has no LU- decomposition.
Note: By reordering row of invertible sub matrices we can make it has non-zero determinant and
also has LU-decomposition.
Or
Is called Cholesky decomposition. Where L= lower triangular matrix and LT =transpose of matrix
L and U=upper triangular matrix and U T =transpose of matrix U.
Note: A matrix A must be positive definite to be has cholesky decomposition only symmetric is
not enough.
1 2 3
A=(2 8 22)
3 22 82
42
Linear Algebra II
i) l11 = 1,
l21 l11 = 2 → l21 =2
l31 l11 = 2 → l31 =3.
ii) l21 2 + l22 2 = 8.
→ l22 = √8 − l21 2 = √8 − 22 = √8 − 4 = √4 = 2.
l21 l31 + l22 l32 =22.
2 ∗ 3 + 2*l32 = 22.
22−6
→ l32 = = 8.
2
iii) l31 2 + l32 2 + l33 2 = 82.
1 0 0 1 2 3
T
Therefore, L= (2 2 0) and L = (0 2 8).
3 8 3 0 0 3
25 15 −5
A=( 15 18 0 )
−5 0 11
• First column of R
25 15 −5 5 0 0 5 3 −1
A = RRT = ( 15 18 0 ) = ( 3 R 22 0 ) (0 R 22 R 32 )
−5 0 11 −1 R 32 R 33 0 0 R 33
• Second column of R
18 0 3 R 0 R R 32
=( ) − ( ) (3 −1) = ( 22 ) ( 22 )
0 11 −1 R 32 R 33 0 R 33
9 3 3 0 3 1
=( )=( )( )
3 10 1 R 33 0 R 33
• Third column of R
43
Linear Algebra II
1+ R 33 2 =10
R 33 2 = 10 − 1=9.
R 33 = 3.
25 15 −5 5 0 0 5 3 −1
Then the conclusion is A = RRT ( 15 18 0 ) = ( 3 3 0 ) (0 3 1 ).
−5 0 11 −1 1 3 0 0 3
A = QR,
There are several methods for actually computing the QR decomposition. One of such method is
the Gram-Schmidt process.
Note: Decompose matrix by QR- decomposition method with Gram- Schmidt process has
difficult if matrix is linearly dependent columns has.
Processes:
To determine Q:
To determine R:
Coordinates of xi :
44
Linear Algebra II
α1i
α11 α12
⋮
0 α22
αii
x1 = 0 , x2 = 0 ,…, xi = .
⋮ ⋮ 0
[ 0 ] [ 0 ] ⋮
[0]
Q= [u1 u2 … un ].
1 0 0
A = [1 1 0].
1 0 1
Solution: A= [ x1 x2 x3 ].This vectors are linear independent. And apply Gram-Schmidt
orthogonalization process:
x (1,1,1) (1,1,1) 1 1 1
i) Set u1 = ‖v1 ‖ = ‖(1,1,1)‖ = =( , , ).
1 √3 √3 √3 √3
ii) Next, set w2 = x2 - 〈x2 , u1 〉u1
1 1 1 1 1 1
= (0,1,1) - 〈(0,1,1), ( , , )〉 ( , , ).
√3 √3 √3 √3 √3 √3
2 1 1 1
= (0,1,1) – ( , , )
√3 √3 √3 √3
2 2 2 −2 1 1
=(0,1,1) – (3 , 3 , 3) = ( 3 , 3 , 3).
−2 1 1 −2 1 1
w (, , ) ( , , ) −2 1 1
and u2 = ‖w2‖ = 3 33
−2 1 1 = 3 33
=( , , ).
2 ‖( , , )‖ √
2 √6 √6 √6
3 33
3
−1 1 −1 1
w3 (0,
, ) (0, , ) −1 1
2 2 2 2
and u3 = ‖w3 ‖
= −1 1 = = (0, , ).
‖(0, , )‖ √
1 √2 √2
2 2
2
45
Linear Algebra II
1 −2
0
√3 √6 α11 α12 α13
1 1 −1
Q=[u1 u2 u3 ] = √3 √6 √2
. And R= [ 0 α22 α23 ].
1 1 1 0 0 α33
[√3 √6 √2]
From x1 = α11 u1
1 1 1
→(1,1,1) =α11 ( , , ).
√3 √3 √3
→ α11 = √3 .
x2 = α12 u1 + α22 u2 .
1 1 1 −2 1 1
→(0,1,1)= α12 ( , , ) + α22 ( , , ).
√3 √3 √3 √6 √6 √6
2√3 √6
α12 = and α22 = .
3 3
And
√3 √6 √2
α13 = , α23 = and α33 = .
3 6 2
2√3 √3
√3 3 3
√6 √6
Therefore, R = 0 .
3 6
√2
[ 0 0 2]
→A= QR.
−1 3
A=( )
1 5
Solution: Let as use Gram-Schmidt process with columns
46
Linear Algebra II
u1= a1=[−1]
1
u1= u1 1 −1
= [ ]
‖u1 ‖ √2 1
3 1 1 −1 3 −1 4
=[ ] − ( (−3 + 5)) [ ] = [ ] − [ ] = [ ]
5 √2 √2 1 5 1 4
u2 = u2
=
1 4 1 1
[ ]= [ ]
‖u2 ‖ 4√2 4 √2 1
−1 1
1 −1 1
Therefore the matrix Q is Q=[√2
1
√2
1 ]=√2 [ 1 ] this produce by Gram schmdit process then
1
√2 √2
multiplies by elementary matrices on right as
−1 3
A=[ ]
1 5
−1
1
0 3
A[√2 ] = [√2
1 ]
0 1 5
√2
−1
1
0 1 4
A[√2 ][ −√2] = [√2 ]
1
0 1 0 1
√2
4
−1 1
1 1 0
0 1 −√2] [
A[√2 ][ 0
1 ] = [√2
1
√2
1] =Q
0 1 0 1 4√2
√2 √2
1 −1
1 0 0
A=Q [0 4 ] [1 √2] [√2 ]
√2 0 1 0 1
1 0
=Q[0 ] [1 √2] [√2 0]
4
√2 0 1 0 1
1 1
=Q[√2 √2 ]=QR where R =[√2 √2 ]=2[ ].
0 4√2 0 4√2 0 4
−1 3 1 −1 1 1 1
Now QR-factorization is : A= [ ]= [ ] √2 [ ].
5 5 √2 1 1 0 4
47
Linear Algebra II
which is sometimes called the singular value factorization of A. The diagonal elements of σ are
called the singular values of A. The corresponding columns of U and V are called left and right
singular vectors of A.
ATA = VΣ2VT.
Note:
. Σ is an m*n matrix where the (i,j)th entry σ i ,0,…,i= 1… min (m,n) and other entries are zero.
1 0
A= [ ].
0 4
Solution:
A = U Σ VT, where the columns of V are the eigenvectors of ATA and the columns of U
are the eigenvectors of AAT.
1 0
First we form: A TA = [ ], and identify the eigenvalues as 𝛌𝐢 =16,[Link]
0 16
corresponding normalized eigenvectors are the columns of V:
0 1
V=[ ].
1 0
4 0
Thus, Σ=[ ].
0 1
48
Linear Algebra II
1 0
Next we form: AAT = [ ], and as this matrix identical to ATA, we can construct U=V
0 16
without much work at all.
0 1 4 0 0 1
A = U Σ VT = [ ][ ][ ].
1 0 0 1 1 0
3 2 2
Example-2: Find the SVD of A, UΣV T , where A = [ ] .
2 3 −2
17 8
AAT = [ ]
8 17
The characteristic polynomial is det (AAT − λI) = λ2 − 34λ + 225 = (λ − 25)(λ − 9), so the
singular values are σ1 = √25 = 5 and σ2 = √9 = 3. Now we find the right singular vectors (the
columns of V) by finding an orthonormal set of eigenvectors of ATA. It is also possible to
proceed by finding the left singular vectors (columns of U) instead. The eigenvalues of are 25, 9,
and 0, and since AT A is symmetric we know that the eigenvectors will be orthogonal.
−12 12 2
T
For λ = 25, we have: A A−25I= [ 12 −12 −2 ]
2 −2 −17
1 −1 0
Which row -reduces to [0 0 1] . A unit-length vector in the kernel of that matrix is
0 0 0
1/√2
u1 = (1/√2).
0
1
4 12 2 1 0 4
For λ = 9 we have AT A−9I = [12 4 −2] which row reduces to [1 0 4
1 ].
2 −2 −1
0 0 0
1⁄√18
A unit-length vector in the kernel is u2 = (−1⁄√18).
4⁄√18
For the last eigen vector, we could compute the kernel of AT A or find a unit vector perpendicular
a
to u1 and u2 . To be perpendicular to u1 = ( b ) we need −a = b.
c
49
Linear Algebra II
a 2⁄3
u3 = ( −a ) and for it to be unit-length we need a = 2⁄3 so u3 = (−2⁄3).
−a⁄2 −1⁄3
1⁄√2 1⁄√2
U=( ).
1⁄√2 1⁄√2
50
Linear Algebra II
CHAPTER THREE
CANONICAL FORMS
We wish to study matrices over F[x]. Consider elementary operations in such matrices, that is
Definition 3.1.1: A matrix obtained from the identity matrix I by applying an F[x]-elementary
operation is called F[x]-elementary matrix.
Note:
Every F[x]- elementary matrix is non-singular and its inverse is also F[x]- elementary matrix.
1 0
Example: A matrix A= [ 2 ] is an F[x]- elementary matrix. Then A is non-singular.
x 1
1 0
i.e.,|A|=1≠0 and its inverse is A−1 = [ 2 ] and also A−1 is an F[x]- elementary matrix. We
−x 1
can obtain by multiply identity
Notation: A F[x] B.
Note: A F[x] B if and only if B= PAQ, where P, Q are products of F[x]- elementary matrices.
x x+1 1 0
a) A[x]= [ ] and B[x] = [ ].
x2 − x x2 − 1 0 0
51
Linear Algebra II
2
b) A[x] = [ x x + 1] , B[x] = [1 0
].
x−1 x 2 0 x − x2 + 1
4
0 1 x 1 0 0
c) A[x]= [ x x 1 ] ,B[x]= [0 1 0]
x2 − x x2 − 1 x2 − 1 0 0 0
Solution:
a)
i) Add –(x-1) times the first row to the second row to obtain :
1 0 x x+1
[ ]A[x] = [ ].
1−x 1 0 0
ii) Add -1 times the first column to the second column to obtain :
x x + 1 1 −1 x 1
[ ][ ]= [ ].
0 0 0 1 0 0
iii) Add –x times the second column to the first column to obtain :
x 1 1 0 0 1
[ ][ ]=[ ].
0 0 −x 1 0 0
iv) Interchange columns to obtain :
0 1 0 1 1 0
[ ][ ]=[ ] =B[x].
0 0 1 0 0 0
1 0
Thus, P(x) = [ ].
1−x 1
1 −1 1 0 0 1 −1 x + 1
Q(x)= [ ][ ][ ]=[ ].
0 1 −x 1 1 0 1 −x
(b) and (c) are exercises.
There are (m
k
)×(nk) k-order minors of A.
1 2x 3
Example-2: Let A= [0 4 5x]. Find all the 2nd- order minors of A.
x 2 1
Solution:
52
Linear Algebra II
|1|,|2x|,|3|,|0|,|4|,|5x|,|x|,|2|,|1| = 1,2x,3,0,4,5x,x,2,1.
Second-order minor
Rows Columns
1,2 1,3 2,3
1 2x 1 3 2x 3
| |=4 | | = 5x | | = 10x 2 − 12
1,2 0 4 0 5x 4 5x
1 2x 1 3 2x 3
| | = 2 − 2x 2 | | = 1 − 3x | | = 2x − 6
1,3 x 2 x 1 2 1
0 4 0 5x 4 5x
| |= -4x | | = −5x 2 | |= 4-10x
2,3 x 2 x 1 2 1
Third-order minor
1 2x 3
4 5x 2x 3
|0 4 5x|=1 | |+x| |
2 1 4 5x
x 2 1
= 4-10x + x[10x 2 − 12]= 10x 3 -22x +4.
And dk(x):
d1(x)= gcd{1,2x,3,4,5x,x,2,1}=1.
Rank=3.
Definition 3.2.3: Let A be a non-zero m*n matrix over F[x]. A is said to be of rank k if k is the
largest integer such that not all kth- order minors of A are identically zero.
Lemma-1: Suppose B=PA, where P is a product of F[x]- elementary matrices. Then, every Kth –
order minor of B is a linear combination, over F[x], of Kth –order minor of A.
1 2x 3
A= [0 4 5x] from the above example-2.
x 2 1
53
Linear Algebra II
1 0 0 1 2x 3
P= A= [0 0 1] and PA = A= [ x 2 1 ]=B.
0 1 0 0 4 5x
All the second order minors of B are
Columns
Rows 1,2 1,3 2,3
1 2x 1 3 2x 3
| |= 2 − 2x 2 | | = 1 − 3x | | = 2x − 6
1,2 x 2 x 1 2 1
1 2x 1 3 2x 3
| |=4 | | = 5x | | = 10x 2 − 12
1,3 0 4 0 5x 4 5x
x 2 x 1 2 1
| |= 4x | | = 5x 2 | |= 10x-4
2,3 0 4 0 5x 4 5x
Thus, the Kth –order minor of A=N= the Kth –order minor of B=M. Or N=-M or N= another
second order minor except for sign.
1 0 0 1 2x 3
P= [0 α 0] and PA = [0 4α 5αx] = B.
0 0 1 x 2 1
All the second minors of B are:
Columns
Rows 1,2 1,3 2,3
1 2x 1 3 2x 3
| |=4 α | | = 5αx | | = 10αx 2 − 12α
1,2 0 4α 0 5xα 4α 5αx
1 2x 1 3 2x 3
| | = 2 − 2x 2 | | = 1 − 3x | | = 2x − 6
1,3 x 2 x 1 2 1
0 4α 0 5αx 4α 5αx
| |= -4 α x | | = −5αx 2 | |= 4 α -10 α x
2,3 x 2 x 1 2 1
Thus, N=M or N= α M
1 0 0 1 2x 3
P= [0 1 0] and PA = [ 0 4 5x ] = B.
x 0 1 2x 2x 2 + 2 3x + 1
54
Linear Algebra II
Columns
Rows 1,2 1,3 2,3
1 2x 1 3 2x3
| |=4 | | = 5x | | = 10x 2 − 12
1,2 0 4 0 5x 45x
1 2x 1 3 2x 3
| | = 2 − 2x 2 | | = 1 − 3x | | = 2x − 6
1,3 x 2 x 1 2 1
0 4 0 5x 4 5x
| 2 |= -8x | | = −10x 2 | 2 |= -10x 3 + 2x + 4
2,3 2x 2x + 2 2x 3x + 1 2x + 2 3x + 1
Thus, N=M or N=M-XL where L=(1,2),(2,3).
A F[x] B → Every Kth –order minor of B is a linear combination, over F[x], of Kth –order minor
of A.
Corollary -1: Let A,B be m*n matrices over F[x]. Then A F[x] B→ rank A=rank B.
→ rank B ≤ rank A.
Corollary -2: Let A F[x] B. Let dk (x) be the gcd of all kth –order minors of A. Then, dk (x) is
Proof: Let d′ k (x) be the gcd of all kth –order minors of B. Theorem 3.3.1, dk (x) ∖ d′ k (x).
Since also, B F[x] A , again by Theorem 3.3.1, d′ k (x) ∖ dk (x) .So, both dk (x) and d′ k (x) are
Lemma-2: Let A be a non-zero m*n matrix over F[x].Then, A is F[x]-equivalent to the m*n
matrix, over F[x], of the form:
55
Linear Algebra II
f (x) 0
[1 ] where f1 (x) is monic polynomial of minimal degree among all non-zero elements of
0 A1
all matrices F[x]- equivalent to A.
Theorem-3.3.2: Let A= a non-zero m*n matrix of rank r over F[x].Then, A is F[x]- equivalent to
a matrix B of one of the following types:
D D 0
D, [D 0],[ ],[ ] ,where
0 0 0
i) D= diag[ f1 (x), f2 (x),…, fr (x)].
ii) fi (x)’s are monic polynomials such that fi (x)/ fi+1 (x) for i= 1,2,…,r-1.
Theorem 3.3.4: Let A, B be as in theorem 3.3.2. The polynomials f1 (x), f2 (x),…, fr (x) are
uniquely determine by the matrix A.
Proof: Suppose g1 (x), g 2 (x),…, g r (x) also satisfy the conditions of theorem 3.3.2.
By Theorem 3.3.3, dk (x) =f1 (x) f2 (x)… fk (x)= g1 (x) g 2 (x)… g k (x). In particular ,
1. The polynomials f1 (x), f2 (x),…, fr (x) are called the invariant factors of A.
2. The matrix B is called the Smith Canonical form of A.
Corollary-3: Let A1 , A2 be m*n matrices over F[x]. A1 F[x] A2 , if and only if A1 , A2 have
56
Linear Algebra II
(←): Suppose A1 , A2 have the same invariant [Link] they are F[x]-equivalent to the same
x 1 0
Example-1: Let A= [0 x 1 ]. Find the Smith canonical form of A.
2 3 x−1
Solution:
d1(x)= gcd{x,1,2,3,x-1}
d2(x)=gcd{
x 1 x 0 1 0 x 1 x 0 1 0 0 x 0 1 x 1
| |,| |,| |,| |,| |,| |,| |,| |,| |}
0 x 0 1 x 1 2 3 2 x−1 3 x − 1 2 3 2 x−1 3 x − 1
=gcd{x2,x,1,3x-2,x2-x,x-1,-2x,-2,x2-x-3}=1.
x 1 1 0
d3(x)= x| | + 2| | =x(x2-x-3) +2= x3-x2-3x +2.
3 x−1 x 1
f1(x)= d1(x)=1.
d2(x) 1
f2(x)= d1(x) = 1 =1.
d3(x) x3 −x2 −3x +2
f3(x)= d2(x) = =x3-x2-3x +2.
1
1 0 0
Therefore, the SCF of A=A=[0 1 0 ].
0 0 x 3 − x 2 − 3x + 2
0 1 x
Example-2: Let A= [ x x 1 ] . Find the Smith canonical form of A.
x2 − x x2 − 1 x2 − 1
Solution:
d2(x)=gcd{
x 1 x 1 1 0 x x 1 x
| 2 2 |,| 2 2 |,| |,| 2 2 |,| 2 2 |,
x −1 x −1 x −x x −1 x 1 x −x x −1 x −1 x −1
0 x 0 1 0 x 0 1 1 x
| 2 |,| |,| |,| |,| |}
x − x x2 − 1 x2 − x x2 − 1 x 1 x x x 1
57
Linear Algebra II
x 1 x x
d3(x)= -1| 2 | + x| 2 |
x −x x2 − 1 x −x x2 − 1
= -1(x3-x- x2 +x ) +x(x+x2)
= -1(x3-x2 ) + x(x2-x)
= x2- x3+x3-x2=0.
f1(x)= d1(x)=1.
d2(x) 1
f2(x)= d1(x) = 1 =1.
d3(x) 0
f3(x)= d2(x) = 1 = 0.
1 0 0
Therefore, the SCF of A=A=[0 1 0].
0 0 0
Exercise-3.1: Find the Smith canonical form of A if:
x x x2 ] x 2 ].
a) A[x] = [ 2 ] b) B[x] = [ x3 c)A[x]= [ x3
x +x x x x5 x x4
A is a product of F[x]- elementary matrices if and only if A is square and det(A) is a non-zero
constant.
Observe:
58
Linear Algebra II
p= det(P).
q=det(Q).
Therefore, B=I.
A has an inverse over F[x] if and only if A is a product of F[x]- elementary matrices.
AA−1 = I.
[det(A)][det(A−1)]=1.
(←): Suppose Ais a product of elementary matrices. By Theorem 3.4.1, A is a square matrix and
det(A) is a non-zero constant. Hence, A−1 exists in F[x].
Examples: Determine whether or not the following matrices over R[x] are non-singular over
R[x].
x x+1 1 x 1 x
a) A[x] = [ ] b) B[x] = [ 2 ] c) C [x]= [ ].
x+2 x+3 x x +1 x+1 x+3
Solution:
59
Linear Algebra II
Proof: Since det(xI-A) = 𝒳A (x) ≠ 0, rank of (xI-A) is n. Hence,(xI-A) has n invariant factors
f1(x), f2(x),…, fn(x).
Let P,Q be products of elementary matrices over F[x] such that P(xI-A)Q is the Smith canonical
form of xI-A.
→ pq𝒳A (x)= f1(x) f2(x)…fn(x), where p= |P|,q=|Q|. Since 𝒳A (x) and all fi(x) are monic, pq=1.
Let mA (x) = The minimal polynomial of A. Then mA (x) = fn(x), where fn(x) is similarity
invariant of A of highest degree.
0 1 0 … 0 0
0 0 1 … 0 0
C(f) = ⋮ ⋮ ⋮ ⋱ ⋮ ⋮ .
0 0 0 … 0 1
[−a0 −a1 − a2 … −an−2 −an−1 ]
Note:
60
Linear Algebra II
0 1 0 0
C4 = ( 0 0 1 0 ).
0 0 0 1
−2 5 3 −2
Theorem 3.5.1: Let f(x) = x n + an−1 x n−1 + an−2 x n−2 + …+ a1 x + a0 .Then, the characteristic
and minimal polynomials of C(f) are both equal to f(x).
Theorem 3.5.2: Let f(x) = x n + an−1 x n−1 + an−2 x n−2 + …+ a1 x + a0 be monic polynomial of
degree n. Then
1 0 0 … 0 0
x 1 0 … 0 0
x 2 x 1 … 0 0
Let Q(x)= .
⋮ ⋮ ⋮ ⋱ ⋮ ⋮
x n−2 x n−3 x n−4 … 1 0
[x n−1 x n−2 x n−3 … x 1]
0 −In−1
(xI-C)Q(x) = [ ], where B= 1*n-1 matrix over F[x].
f(x) B
I 0n−1∗1 0 1
Let P(x) = [ n−1 ], R[x]= [ 1∗n−1 ].
B 1 −In−1 0n−1∗1
In−1 01∗n−1
Then, P(x)(xI-C)Q(x)R(x)= [ ].
01∗n−1 f(x)
Theorem 3.5.3: Let f1(x), f2(x),…,fr(x) be non-constant monic polynomials over F such that fi(x)
divides fi+1(x) for i= 1,2,…,r-1.
61
Linear Algebra II
Then, the matrix B= diag[C1 , C2 ,…, Cr ] has f1(x), f2(x),…,fr(x) as its non-trivial similarity
invariants.
= f1(x) f2(x)…fr(x).
Theorem 3.5.4: Every square matrix A over a field F is similar to a diagonal block matrix,
where each diagonal block is the companion matrix of one of the non-trivial similarity invariants
of A.
Proof: Let A be an n*n matrix over a field F. Let f1(x), f2(x),…,fr(x) be non-trivial similarity
invariants of A.
B is an n*n matrix, since 𝒳B (x)= f1(x) f2(x)…fr(x) which is of degree n. By Theorem 3.5.3, the
non-trivial similarity invariants of B are also f1(x) ,f2(x),…,fr(x).Hence, A is similar to B.
Definition 3.5.1: A diagonal block matrix diag[C1 , C2 ,…, Cr ] of theorem 3.5.4 is called the
Rational Canonical form for the matrices similar to A.
6 2 −2
Example -1: Let A = [−2 2 2 ]. Determine the Rational Canonical form of matrices
2 2 2
similar to A.
Solution:
62
Linear Algebra II
x − 6 −2 2
xI- A =[ 2 x−2 −2 ].
−2 −2 x−2
d1(x)=gcd{x-6,-2,2,x-2}=1.
x − 6 −2 x − 6 2 −2 2
d2(x)=gcd{ | |,| |, | |,
2 x−2 2 −2 x−2 −2
2 x−2 2 −2 x − 2 −2 x − 6 −2 x − 6 2 −2 2
| |,| |,| |,| |,| |,| |}
−2 −2 −2 x − 2 −2 x − 2 −2 −2 −2 x − 2 −2 x − 2
=gcd{(x − 6)(x − 2) + 4,-2(x-6)-4,4-2(x-2),-4+2(x-2),-4+2(x-2),(x − 2)2-4,
-2 (x − 6) − 4, (x − 6)(x − 2) + 4,-2(x-2)+4}.
x − 2 −2 2 −2 2 x−2
d3(x)=(x-6)| |+2| |+2 | |.
−2 x − 2 −2 x − 2 −2 −2
=(x-6) (x 2 − 4x)+ 2(2x-8)+2(2x-8).
=(x-4)( x 2 -6x+8).
f1(x)=d1(x)=1.
d2(x)
f2(x)=d1(x) =x-4. →C(f2(x))=(4).
d3(x) 0 1
f3(x)= =(x-4)(x-2)=x 2 − 6x + 8. →C(f3(x))=( )
d2(x) −8 6
c) Characteristic polynomial of A:
d) Minimal polynomial of A:
63
Linear Algebra II
4 0 0
[0 0 1].
0 −8 6
Suppose XA (x) = p1 e1 p2 e2 …pi ei , where p1 , p2 , …,pi are distinct monic polynomials which
are irreducible over F, and each ei is a positive integer.
Notice, eij may be 0. But if eij is positive, then ei+1,j is also positive.
Definition 3.6.1: The polynomials pj eij which appear in the similarity invariants of A with non-
zero exponents eij are called the elementary divisors of A over F.
Example-2: Suppose x,x, x 2 , (x + 1), (x + 1)2 , (x-1), (x − 1)3 are the list of elementary
divisors of A. Then the non-trivial similarity invariants of A are:
f3 = x 2 (x + 1)2 (x − 1)3
f2 = x (x+1)(x-1)
f1 = x.
64
Linear Algebra II
Theorem 3.6.1: Let A, B= n*n matrices over a field F. Then, A∼B if and only if A and B have
the same elementary divisors.
Then A and B have the same similarity invariants and hence the same elementary divisors.
(←): Suppose A and B have the same elementary divisors. Then A and B have the same
similarity invariants.
Hence, A∼B.
We can write f(x) = p1 e1 p2 e2 …pi ei .Where the pi ‘s are distinct, monic, irreducible polynomials
over F; and the ei ‘s are positive integers.
Since the pi ’s are irreducible and distinct, the gcd of the (n-1)th order minors of the last matrix is
1.
Hence, A ∼C.
65
Linear Algebra II
Proof: Let f1(x), f2(x),…,ft(x) be non-trivial similarity invariants of A. Since we know that
6 2 2
Example-1: Let A = [−2 2 0]. Determine the Normal Canonical form of matrices similar to
0 0 2
A.
Solution:
x − 6 −2 −2
xI- A =[ 2 x−2 0 ].
0 0 x−2
d1(x)=gcd{x-6,-2,2,x-2}=1.
x − 6 −2 x − 6 −2 −2 −2
d2(x)=gcd{ | |,| |, | |,
2 x−2 2 0 x−2 0
2 x−2 2 0 x−2 0 x − 6 −2 x − 6 −2 −2 −2
| |,| |,| |,| |,| |,| |}
0 0 0 x−2 0 x−2 0 0 0 x−2 0 x−2
=gcd{(x − 6)(x − 2) + 4,4,2(x-2),0,2(x-2),(x − 2)2,0,(x-6)(x-2),-2(x-2)].
= 1.
x − 6 −2
d3(x)=(x-2)| |= (x-2)[(x-6)(x-2)+4]= (x-2) (x 2 -8x+12+4]
2 x−2
=(x-2)( x 2 -8x+16).
= (x − 4)2(x-2).
f1(x) = d1(x)=1.
d2(x)
f2(x) = d1(x) = 1.
66
Linear Algebra II
d3(x)
f3(x) = d2(x) =(x − 4)2 (x-2)
c) Characteristic polynomial of A:
d) Minimal polynomial of A:
2 0 0
A~ [0 0 1].
0 −16 8
We define the Jordan block corresponding to the elementary divisor (x − λj )eij to be the
λj 1 0 … 0
0 λj 1 … 0
Jj = ⋮ ⋮ ⋱ ⋮ ⋮ .
0 0 0 λj 1
[0 0 0 … λj ]
𝐓𝐡𝐞𝐨𝐫𝐞𝐦 3.7.2: Let F be algebraically closed field. If the n*n matrix A over F has r elementary
divisors with associated Jordan blocks J1 , J2 ,…, Jr then
𝐏𝐫𝐨𝐨𝐟: We need only show that the elementary divisors in Jj and C( pj eij ) coincide for i=1,2,…,r.
67
Linear Algebra II
Examination of xI- Jj shows that there are minors of orders 1,2,…, eij − 1 having values
±[Link],
It follows that pj eij is the only elementary divisors of Jj .Definition -3.7.2: Let A, J1 , J2 ,…, Jr be
as in Theorem 3.7.2. The matrix diag [J1 , J2 ,…, Jr ] is called the Jordan Canonical Form of the
matrices similar to A.
6 2 2
Example-2: Let A = [−2 2 0]. Determine the Jordan Canonical form of matrices similar to
0 0 2
A.
Solution:
x − 6 −2 −2
xI- A =[ 2 x−2 0 ].
0 0 x−2
d1(x)=gcd{x-6,-2,2,x-2}=1.
x − 6 −2 x − 6 −2 −2 −2
d2(x)=gcd{ | |,| |, | |,
2 x−2 2 0 x−2 0
2 x−2 2 0 x−2 0 x − 6 −2 x − 6 −2 −2 −2
| |,| |,| |,| |,| |,| |}
0 0 0 x−2 0 x−2 0 0 0 x−2 0 x−2
=gcd{(x − 6)(x − 2) + 4,4,2(x-2),0,2(x-2),(x − 2)2,0,(x-6)(x-2),-2(x-2)].
= 1.
x − 6 −2
d3(x)=(x-2)| |= (x-2)[(x-6)(x-2)+4]= (x-2) (x 2 -8x+12+4]
2 x−2
=(x-2)( x 2 -8x+16).
= (x − 4)2(x-2).
f1(x)=d1(x)=1.
d2(x)
f2(x)=d1(x) = 1.
68
Linear Algebra II
d3(x)
f3(x)=d2(x) =(x − 4)2 (x-2)
𝐄𝐱𝐞𝐫𝐜𝐢𝐬𝐞 − 𝟑. 𝟑:
x 1 0
1. Let A[x] = [0 x 1 ].
2 3 x−1
a) Find the invariant and smith canonical forms of A .
b) Determine whether or not A is non-singular over R[x].
2. Determine the Similarity invariants of A:
2 2 −1 0 −1 2
a) A= [−1 −1 1 ] b) A=[3 −4 6].
−1 −2 −2 2 −2 3
In each case, write the Rational Canonical form of A.
0 0 1
3. Let A= [1 0 −1] .Find the Normal and Jordan canonical Forms/over C/ of matrices
0 1 1
similar to A.
4. Suppose x+1,x(x+1),x(x+1)(x 2 + 2) are the non-trivial similarity invariants of a matrix A.
a) Find the elementary divisors of A over R.
b) Find the elementary divisors of A over C.
c) Write the Jordan Canonical form of matrices similar to A over C.
69
Linear Algebra II
CHAPTER FOUR
Note: Condition (i) states that φ is a linear map with respect to the first variable and condition
(ii) states that φ is a linear map with respect to the second variable.
Example 4.1.1:
1. Let V= a vector over a field K with an inner product ⟨ , ⟩.Then the inner product
⟨ , ⟩: V×V→ K is a bilinear map.
2. Let V be a ε vector space over the field K. Let V* be the dual space of V. Then the map
φ(v,w) = φ (∑m n
i=1 xi vi , ∑j=1 yj wj ).
= ∑m n
i=1[xi φ(vi ), ∑j=1 yj wj ].
= ∑m n
i=1 xi [ ∑j=1 yj φ(vi, wj )].
= ∑m n
i=1 ∑j=1 xi yj φ(vi, wj ).
Thus, the mn scalars φ(vi, wj ) completely determine the value of the map φ. Let aij = φ(vi, wj )
and consider the m*n matrix : A = (aij ).
70
Linear Algebra II
x1 y1
x2 y2
Then, φ(v,w) = X t A Y where X = [ ⋮ ] , Y= [ ⋮ ].
xm yn
Definition 4.1.2: A is called the matrix representation of the bilinear map : V×W → K with
respect to the basis β, β, of V and W respectively.
0 −1
Therefore, A= [ ].
−1 −2
71
Linear Algebra II
Example 4.1.2: V= an inner product space with inner product ⟨ , ⟩. The inner product
Note: Let B(V) = the set of all bilinear forms on V. If f,g ∈B(V) and λ ∈ K, we can define f+g
and λf by:
Theorem 4.1.1: Let V be a vector space of dimension n over K. Let {φ1 , φ2 , …,φn } be a basis
of the dual space V*.
Then {fij : i,j =1,2,…,n}, where fij is defined by fij (v,w) = φi (v)φj (w) , is a basis of B(V).
Thus, in particular, dim B(V) = n2 .
Proof: Let {e1 , e2 , …,en } be a basis of V dual to {φ1 , φ2 , …,φn }.Thus, φi (ej ) = δij .
72
Linear Algebra II
Theorem 4.1.2: Let [f] be a matrix representation of f ∈B(V) relative to a basis {e1 , e2, …,en }
of V . Then the mapping f→ [f] is an isomorphism of B(V) on to the vector space Mn (K) of n*n
matrices over K.
Proof: since f is completely determined by the scalars f(ei , ej ), the mapping f→ [f] is one to one
correspondence . It only remains to show that the mapping f→ [f] is a homomorphism.
But
A= y1 y2 y3
x1 5 −2 0
x2 [7 −5 6].
x3 1 4 1
Example-2: For the following matrix determine the bilinear form b(x,y)?
3 5 2 1 −3
a) A= [ 7 −5 1] b) A = [9 10 ].
−3 8 1 1 2
Solution:
73
Linear Algebra II
Note:
Thus, alternating and skew symmetric bilinear forms are equivalent when 1+1≠ 0 in K.
Example: Show that f(v,w) = 2xy ′ -2x ′ y , for v =(x,y) and w= (x ′ , y ′ ) is an alternating bilinear
form?
0 2
The matrix representation is skew symmetric: A= [ ].
−2 0
Theorem 4.3.1: An n*n matrix A over a field K represents a symmetric bilinear form if and only
if it is a symmetric matrix.
→ A = At .
74
Linear Algebra II
Theorem 4.3.2: Let f be a symmetric bilinear form on V over K /in which 1+1 ≠ 0/. Then V has
a basis {v1 , v2 , …,vn }in which f is represented by a diagonal matrix, i.e., f(vi , vj ) =0 for i ≠ j.
Example-1: For the following symmetric bilinear form determine the matrix A?
A= y1 y2 y3
x1 4 1 0
x2 [1 −2 −4].
x3 0 −4 7
Example-2: For the following symmetric matrix determine the symmetric bilinear form, b(x,y)?
3 5 2
A= [5 −5 1] .
2 1 1
Solution:
Exercise-4.1:
[Link] u=(x1 , x2 ) and u=(y1 , y2 ) .Determine which of the following are bilinear forms on R2 .
i. f(u,v)=x1 + y2 iv. f(u,v)= 1
ii. f(u,v)=2 x1 y1 v. f(u,v)=0
iii. f(u,v)=2x1 x2 + 3y1 y2 vi. f(u,v)= 3x1 y2 + x2 y1 .
2
2. Let f be the bilinear form on R defined by
a)F((x1 , x2 ), (y1 , y2 ))= 2x1 y1 +3x1 y2 + x2 y1 -2 x2 y2 .
b)F((x1 , x2 ), (y1 , y2 ))= 3x1 x2 + 2y1 y2 .
i. Find the matrix A of f in the basis {u1 = (1, 1), u2 = (2,1)}
ii. Find the matrix A of f in the basis {v1 = (1, 3), v2 = (1,1)}
2 5
3. Let V be the vector space of 2 2 matrices over R. Let M= [ ],and let f(A,B)=tr(At MB)
1 1
where A,B V and ‘’tr’’ denotes trace.
75
Linear Algebra II
1 0 0 1 0 0 0 0
ii. Find the matrix of f in the basis {[ ],[ ],[ ],[ ]}.
0 0 0 0 1 0 0 1
4.3.2. Quadratic Forms
Definition 4.2.1: Let V be a finite dimensional vector space over a field K.
A mapping q:V→ K is called a quaradratic form if q(v) = f(v,v) for some symmetric bilinear
form f on V.
Notes:
The formula (**) is called the quadratic polynomial corresponding to the symmetric matrix A.
Observe that if A is diagonal, then q(X) = a11 x1 2 + a22 x2 2 +…+ ann xn 2 is called canonical
form.
Example 4.2.2: Find the symmetric matrix which corresponds to the quadratic polynomial
q(x,y) = xy + y 2 .
Solution:
The symmetric matrix A= A = (aij ) representing q(x1 x2 … xn ) has the diagonal entry
n∗n
aii equal to the coefficient of xi 2 and has the entries aij each equal to half the coefficient of xi xj .
76
Linear Algebra II
1
0 2
Thus, A= [ 1 ].
1
2
Example 4.2.3: Let q(x,y) = 4x 2 − 6 xy -7 y 2 . Find the symmetric matrix which corresponds to
the quadratic polynomial
Solution:
The symmetric matrix A= A = (aij ) representing q(x1 x2 … xn ) has the diagonal entry
n∗n
aii equal to the coefficient of xi 2 and has the entries aij each equal to half the coefficient of xi xj .
4 −3
Thus, A= [ ].
−3 −7
Example 4.2.4: For the following symmetric matrices determine the quadratic form, q(X)?
2 6
a) A= [ ].
6 1
3 3 1
b) A= [3 5 2] .
1 2 1
Solution:
a) q(X) = q(x,y) = 2x 2 + 12 xy + y 2 .
b) q(X)= q(x1 , x2 , x3 ) = 3x1 2 + 5x2 2 + x3 2 +6 x1 x2 + 2x1 x3 + 4x2 x3 .
Exercise-4.2:
1. Find the symmetric matrix which corresponds to each of the following quadratic
polynomials.
a)q(x,y) = x 2 −xy + y 2 .
b)q(x,y) =x 2 +4yz + 2xz + 5y 2 .
c) q(x,y) =x 2 +2xy-y 2 + 6xz – 4yz + 3z 2 .
2. For each of the following matrices A, find a non-singular matrix P such that P t AP is
diagonal:
2 3
i) A= [ ].
3 4
1 −2 3
ii) A = [−2 6 −9].
3 −9 4
77
Linear Algebra II
3. Let q be the quadratic form associated with a symmetric bilinear form f. Verify the following
1
alternate polar form of f(u,v)=4 [q(u+v)- q(u-v)].
Proof:
We may assume without any loss of generality that in each matrix the positive entries appear
first. Since rank(f) = P +N= P ′ + N′ , it suffice to prove that P=P ′ .
Let U = {u1 , u2 , … , up }.
W={wp′ +1 , wp′ +2 , … , wn }.
Then f(v,v) > 0 for every non-zero v ∈U, and f(v,v) ≤ 0 for every non-zero v ∈W.
Hence, U W={0}.
dim(W)= n-P ′ .
= P + n-P ′ -0 = P -P ′ + n.
Hence, P -P ′ + n ≤ n or P ≤ P ′ . Similarly, P ′ ≤ P.
78
Linear Algebra II
Therefore, P= P ′ .
We call:
Rank= p + n.
p= The index of f.
p – n = The signature of f.
nature:
q(x,y) = x 2 − y 2 . The matrix representation of q with respect to the standard ordered basis of
R2 is:
1 0
A= [ ].
0 −1
A is also the matrix representation of the bilinear form f on R2 corresponding to q.
Signature of f = signature of q = 0.
Example-2: Reduce the following quadratic forms into canonical form. And determine rank,
signature, index and nature of the quadratic forms.
79
Linear Algebra II
Solution:
8 −6 2
a) First , A = [−6 7 −4].
2 −4 3
6−λ −6 2
The characteristic equation is |A − λI| = 0= | −6 7−λ −4 |=0= λ3 - D1 λ2 + λD2 - D3
2 −4 3−λ
D1 = 8+7+3=18.
7 −4 8 2 8 −6
D2 = | |+ | |+| | = 5 + 20 +20 = 45.
−4 3 2 3 −6 7
7 −4 −6 −4 −6 7
D3 =8| | + 6| | + 2| | = 40-60+20=0.
−4 3 2 3 2 −4
This implies: λ3 - 18λ2 + 45λ = 0.
→ λ (λ - 15) (λ - 3) =0.
→ λ =0,3,15.
0 0 0
Therefore , D = [0 3 0 ].
0 0 15
The canonical form is : 3y 2 + 15z 2 .
Exercise-4.3:
into canonical form. And determine rank, signature, index and nature of the quadratic form.
80
Linear Algebra II
Definition -4.4.2: A real symmetric n*n matrix A is said to be positive definite if X t AX > 0 for
all X≠ 0.
.Symbolized by A>0.
f(v,w) = v • w = x1 y1 + x2 y2 + …+ xn yn , where
Generally, the following methods are test for positive definite and positive semi-definite.
Example: Check that the following symmetric matrices are positive definite, positive semi-
definite or not.
4 6
a) A= [ ]
6 10
1 1 2
b) B = [1 2 3]
0 −1 −1
81
Linear Algebra II
Solution:
4 6
a) A = [ ]
6 10
i) By using pivot element method:
4 6 4 6
[ ]↔[ ]. Therefore, all the pivots are positive:1and 4.
6 10 0 1
ii) By taking left-upper matrices determinants:
|4| =4> 0.
4 6
| |=4*10- 6*6=40-36=4 >0.
6 10
→ A is positive definite.
1 1 2
b) B= [1 2 3 ].By taking left-upper matrices determinants:
0 −1 −1
|1| =5> 0.
1 1
| |=2- 1=1>0.
1 2
1 1 2
1 1 1 1
|1 2 3 |=1 | |- | | = (-2+3) – (-1+2)=1-1=0≥ 0.
1 2 1 2
0 −1 −1
→ A is positive semi-definite.
2 −1 −1
B = [−1 2 −1 ]
−1 −1 2 + c
Positive definite and positive semi-definite ?
82
Linear Algebra II
CHAPTER FIVE
Let W1 , W2 subspace of V.
v = w1 + w2 .
Let W1 , W2 subspace of V.
Proof: Let v V .
Then, v = w1 + w2 = w1 ′ + w2 ′ .
→ w1 - w1 ′ = w2 ′ - w2 ∈ W1 W2 .
→ w1 = w1 ′ and w2 ′ = w2 .
Definition 5.1.2:
83
Linear Algebra II
→ V = 𝑅 3 ≠ U W.
Solution:
i) (a, b, c) = (a , b, 0) + (0,0,c) → is sum.
ii) U W = (0, 0, 0) = {0}.→ is unique.
→ V = 𝑅 3 = U W.
3. V = 𝑅 3 , U = {(a, b, c): a= b=c ∈ R}; W = {(0,b,c): b,c ∈ R}.Is V a direct sum of U and W?
Solution:
i) 𝑅3 = U + W .
→ (a, b, c) = (a , a, a) + (0,b-a,c-a) → is sum.
→ a = b= c and a=0.
→b=c=0.
→ v = ( 0, 0, 0). → is unique.
V = 𝑅 3 = U W.
84
Linear Algebra II
V = W1 + W2 +…+ Wk .
v = w1 ′ + w2 ′ + …+ wk ′ .
→ wi = wi ′ for i= 1,2,…,k.
(ii)→ (iii):
→ v = 0.
(iii)→ (i):
→ wk = 0.
85
Linear Algebra II
→ w1 + w2 +…+ wk−1 = 0.
V = W1 W2 … Wk .
Proof:
d d d
1
→ ∑j=1 2
a1j v1j = ∑j=1 a2j v2j = …= ∑j=1
k
akj vkj = 0.
Note:
86
Linear Algebra II
T= a linear operator on V.
Tw (W) = T(w), w ∈ W.
R = Range of T.
N = Null space of T.
Then, V= R N.
Proof:
87
Linear Algebra II
(i) Let 𝑣 ∈ V.
Then, v = T(v) + [v – T(v)], T(v) ∈ R.
v – T(v) ∈ N, since T(v-T(v)) = T(v) - 𝑇 2 (v) = T(v)- T(v) = 0.
Therefore, 0= T(v) = 𝑇 2 (w) = T(w) = v.
Hence, R N= {0}.
Hence, V= R N.
Note:
Proof:
88
Linear Algebra II
Tj 2 = Tj -------------------------(a)
Null space of Tj = W1 + W2 +…+ Wj + Wj−1 + …+ Wk .
Furthermore, v = T1 (v)+ T2 (v) +…+ Tk (v).
Therefore, T1 + T2 +…+ Tk = I. --------------------(c)
If i≠j, then Ti Tj = 0, because the Range Tj is the subspace of Wj which is contained
in the Null space of Ti . -------------------(b)
(ii) Clearly, V = W1 + W2 + …+ Wk , for by (c) , we have
v = T1 (v)+ T2 (v) +…+ Tk (v), for all v∈V and Tj (v) ∈ Wj . We need only show that
this expression for v is unique.
Suppose
v = w1 + w2 +…+ wk with wj ∈ Wj . Since Wj =Range Tj , wj = Tj (vj ) for some
vj ∈ V.
Then, Tj (v) = Tj (w1 + w2 +…+ wk ).
= Tj (w1 )+ Tj (w2 )+…+ Tj (wk ).
= Tj T1(v1 )+ Tj T2 (v2 )+…+ Tj Tk (vk ).
= Tj 2 ( vj ).
== Tj ( vj ).
= wj .
P= 𝑝1 𝑟1 𝑝2 𝑟2 .... 𝑝𝑘 𝑟𝑘 ;
Where the 𝑝𝑖 are distinct irreducible monic polynomials over F and ri are positive integers.
(i) V = W1 W2 … Wk ;
(ii) Each Wi is invariant under T;
(iii) If Ti is the operator induced on Wi by T, then the minimal polynomial for Ti is pi ri .
89