Definition 1.1.1. A rectangular array of numbers is called a matrix.
In this book, we shall mostly be concerned with complex numbers. The horizontal arrays of a matrix
are called its rows and the vertical arrays are called its columns. Let A be a matrix having m rows
and n columns. Then, A is said to have order m × n or is called a matrix of size m × n and can be
represented in either of the following forms:
where aij is the entry at the intersection of the ith row and jth column. One writes A ∈ Mm,n(ℂ) to
say that A is an m × n matrix with complex entries, A ∈ Mm,n(ℝ) to say that A is an m × n matrix
with real entries and A = [aij], when the order of the matrix is understood from the context. We will
also use A[i,:] to denote the i-th row of A, A[:,j] to denote the j-th column of A and aij or (A)ij, for
the (i,j)-th entry of A.
For example, if A = then A[1,:] = [13 + i7], A[:,3] = and
a22 = 5. In general, in row vector commas are inserted to differentiate between entries. Thus,
A[1,:] = [1,3 + i,7]. A matrix having only one column is called a column vector and a matrix
with only one row is called a row vector. All our vectors will be column vectors and
will be represented by bold letters. Thus, A[1,:] is a row vector and A[:,3] is a column
vector.
Definition 1.1.3. Two matrices A = [aij],B = [bij] ∈ Mm,n(ℂ) are said to be equal if
aij = bij, for each i = 1,2,…,m and j = 1,2,…,n.
In other words, two matrices are said to be equal if they have the same order and their
corresponding entries are equal.
Definition 1.1.4.
-
1.
- A matrix in which each entry is zero is called a zero-matrix, denoted 0. For example,
-
2.
- A matrix that has the same number of rows as the number of columns, is called a square
matrix. A square matrix is said to have order n if it’s order is n × n and is denoted
either by writing A ∈ Mn(ℝ) or A ∈ Mn(ℂ), depending on whether the entries are real or
complex numbers, respectively.
DRAFT
-
3.
- Let A = [aij] ∈ Mm,n(ℂ).
-
(a)
- Then, the entries a11,a22,…,ann are called the diagonal entries and they constitute
the principal diagonal of A.
-
(b)
- Then, A is said to be a diagonal matrix if aij = 0 for i≠j, denoted diag(a11,…,ann).
For example, the zero matrix 0n and are two diagonal matrices.
-
(c)
- If A = diag(a11,…,ann) and aii = d for all i = 1,…,n then the diagonal matrix A
is called a scalar matrix.
-
(d)
- Then, A = diag(1,…,1) is called the identity matrix, denoted In, or in short I.
For example, I2 = and I3 = .
-
4.
- For 1 ≤ i ≤ n, define ei = In[:,i], a matrix of order n × 1. Then, the set {e1,…,en}, where
ei ∈ Mn,1(ℂ), for 1 ≤ i ≤ n, is called the standard basis of ℂn. Note that even though the
order of the column vectors ei’s depend on n, we don’t mention it as the size is understood from
the context. For example, if e1 ∈ ℂ2 then, e1T = [1,0]. If e1 ∈ ℂ3 then, e1T = [1,0,0] and so
on.
-
5.
- Let A = [aij] be a square matrix.
-
(a)
- Then, A is said to be an upper triangular matrix if aij = 0 for i > j.
-
(b)
- Then, A is said to be a lower triangular matrix if aij = 0 for i < j.
-
(c)
- Then, A is said to be triangular if it is an upper or a lower triangular matrix.
For example, is upper triangular, is lower triangular and
the matrices 0,I are upper as well as lower triangular matrices.
-
6.
- An m × n matrix A = [aij] is said to have an upper triangular form if aij = 0 for all i > j.
For example, the matrices , and have upper
DRAFT
triangular forms.
Definition 1.2.1. Let A = [aij] ∈ Mm,n(ℂ).
-
1.
- Then, the transpose of A, denoted AT = [bij] ∈ Mn,m(ℂ) and bij = aji, for all i,j.
-
2.
- Then, the conjugate transpose of A, denoted A* = [cij] ∈ Mn,m(ℂ) and cij = aji, for
all i,j, where for a ∈ ℂ, a denotes the complex-conjugate of a.
Thus, if x is a column vector then xT and x* are row vectors and vice-versa. For example, if
A = then A* = AT = , whereas if A = then AT = and
A* = . Note that A*≠AT .
Theorem 1.2.2. For any matrix A, (A*)* = A. Thus, (AT )T = A.
Proof. Let A = [aij],A* = [bij] and (A*)* = [cij]. Clearly, the order of A and (A*)* is the same. Also,
by definition cij = bji = aij = aij for all i,j and hence the result follows. _
Definition 1.2.3. Let A = [aij],B = [bij] ∈ Mm,n(ℂ). Then, the sum of A and B, denoted
A + B, is defined to be the matrix C = [cij] ∈ Mm,n(ℂ) with cij = aij + bij.
DRAFT
Definition 1.2.4. Let A = [aij] ∈ Mm,n(ℂ). Then, the product of k ∈ ℂ with A, denoted
kA, is defined as kA = [kaij] = [aijk] = Ak.
For example, if A = then 5A = and (2 + i)A = .
Theorem 1.2.5. Let A,B,C ∈ Mm,n(ℂ) and let k,ℓ ∈ ℂ. Then,
-
1.
- A + B = B + A (commutativity).
-
2.
- (A + B) + C = A + (B + C) (associativity).
-
3.
- k(ℓA) = (kℓ)A.
-
4.
- (k + ℓ)A = kA + ℓA.
Proof. Part 1.
Let A = [aij] and B = [bij]. Then, by definition
as
complex numbers commute. The reader is required to prove the other parts as all the results follow
from the properties of complex numbers. _
DRAFT
Definition 1.2.6. Let A ∈ Mm,n(ℂ).
-
1.
- Then, the matrix 0m×n is called the additive identity as A + 0 = 0 + A = A.
-
2.
- Then, there exists a matrix B with A + B = 0. This matrix B is called the additive
inverse of A, and is denoted by -A = (-1)A.
Definition 1.2.8. Let A = [aij] ∈ Mm,n(ℂ) and B = [bij] ∈ Mn,r(ℂ). Then, the product of
A and B, denoted AB, is a matrix C = [cij] ∈ Mm,r(ℂ) with
Thus, AB is defined if and only if number of columns of A = number of rows of
B.
For example, if A = and B = then
Note that the rows of the matrix AB can be written directly as
and similarly, the columns of the matrix AB can be written directly as
| (1.2.3) |
(AB)[:,2] = βA[:,1] + yA[:,2] + vA[:,3],,(AB)[:,4] = δA[:,1] + tA[:,2] + sA[:,3].
Remark 1.2.9. Observe the following:
DRAFT
-
1.
- In this example, while AB is defined, the product BA is not defined. However, for square
matrices A and B of the same order, both the product AB and BA are defined.
-
2.
- The product AB corresponds to operating (adding or subtracting multiples of different
rows) on the rows of the matrix B(see Equation (1.2.2)). This is row method for
calculating the matrix product.
-
3.
- The product AB also corresponds to operating (adding or subtracting multiples of different
columns) on the columns of the matrix A (see Equation (1.2.3)). This is column method
for calculating the matrix product.
-
4.
- Let A and B be two matrices such that the product AB is defined. Then, verify that
-
(a)
- Then, verify that (AB)[i,:] = A[i,:]B. That is, the i-th row of AB is obtained by
multiplying the i-th row of A with B.
-
(b)
- Then, verify that (AB)[:,j] = AB[:,j]. That is, the j-th column of AB is obtained
by multiplying A with the j-th column of B.
Hence,
| (1.2.4) |
DRAFT
DRAFT
Exercise 1.2.11.
-
1.
- For 1 ≤ i ≤ n, recall the basis elements ei ∈ Mn,1(ℂ) (see Definition 4). If A ∈ Mn(ℂ) then
-
(a)
- Ae1 = A[:,1],…,Aen = A[:,n].
-
(b)
- e1T A = A[1,:],…,enT A = A[n,:].
-
2.
- Let A ∈ Mn(ℂ) and D = diag(d1,d2,…,dn). Then
-
(a)
- (DA)[i,:] = diA[i,:], for 1 ≤ i ≤ n, and
-
(b)
- (AD)[:,j] = djA[:,j], for 1 ≤ j ≤ n.
In particular, if D = αI is a scalar matrix, for some α ∈ ℂ, then DA = αA = AD.
-
3.
- If x = ∈ Mn,1(ℂ) then x*x = ∑
i=1n|xi|2.
-
4.
- Let A be an upper triangular matrix. If A*A = AA* then prove that A is a diagonal matrix. The
same holds for lower triangular matrix.
-
5.
- Let x = ,y = ∈ Mn,1(ℂ). Then, prove that xy* = and
y*x = ∑
i=1nyixi.
Definition 1.2.12. Two square matrices A and B are said to commute if AB = BA.
Theorem 1.2.14. Suppose that the matrices A,B and C are so chosen that the matrix
multiplications are defined.
-
1.
- Then, (AB)C = A(BC). That is, the matrix multiplication is associative.
-
2.
- For any k ∈ ℝ,(kA)B = k(AB) = A(kB).
-
3.
- Then, A(B + C) = AB + AC. That is, multiplication distributes over addition.
-
4.
- If A ∈ Mn(ℂ) then AIn = InA = A.
Proof. Part 1. Let A = [aij] ∈ Mm,n(ℂ),B = [bij] ∈ Mn,p(ℂ) and C = [cij] ∈ Mp,q(ℂ).
Then,
Therefore,
Using a similar argument, the next part follows. The other parts are left for the reader. _
Exercise 1.2.15.
-
1.
- Let L1,L2 ∈ Mn(ℂ) be lower triangular matrices and U1,U2 ∈ Mn(ℂ) be upper
triangular matrices. Further, let D ∈ Mn(ℂ) be a diagonal matrix. Then, prove that
-
(a)
- L1L2 is a lower triangular matrix.
-
(b)
- U1U2 is an upper triangular matrix.
-
(c)
- DL1 and L1D are lower triangular matrices.
-
(d)
- DU1 and U1D are upper triangular matrices.
-
2.
- Let A ∈ Mm,n(ℂ). If Ax = 0 for all x ∈ Mn,1(ℂ) then prove that A = 0, the zero
matrix.
-
3.
- Let A,B ∈ Mm,n(ℂ). If Ax = Bx, for all x ∈ Mn,1(ℂ) then prove that A = B.
-
4.
- Let A and B be two matrices such that the matrix product AB is defined.
-
(a)
- Prove that (AB)* = B*A*.
-
(b)
- If A[1,:] = 0* then (AB)[1,:] = 0*.
DRAFT
-
(c)
- If B[:,1] = 0 then (AB)[:,1] = 0.
-
(d)
- If A[i,:] = A[j,:] for some i and j then (AB)[i,:] = (AB)[j,:].
-
(e)
- If B[:,i] = B[:,j] for some i and j then (AB)[:,i] = (AB)[:,j].
-
5.
- Construct matrices A and B, different from the one given earlier, that satisfy the following
statements.
-
(a)
- The product AB is defined but BA is not defined.
-
(b)
- The products AB and BA are defined but they have different orders.
-
(c)
- The products AB and BA are defined, they have the same order but AB≠BA.
-
(d)
- Let A = and B = . Guess a formula for An
and Bn and prove
it?
-
(e)
- Let A = , B = and C = . Is it true that A2
-2A+I =
0? What is B3 - 3B2 + 3B - I? Is C3 = 3C2?
__________________________________
-
6.
- Let A and B be two m × n matrices. Then, prove that (A + B)* = A* + B*.
-
7.
- Find a 2 × 2 nonzero matrix A satisfying A2 = 0.
-
8.
- Find a 2 × 2 nonzero matrix A satisfying A2 = A and A≠I2.
-
9.
- Find 2 × 2 nonzero matrices A,B and C satisfying AB = AC but B≠C. That is, the cancelation
law doesn’t hold.
-
10.
- Let S = and T = . Then, determine the smallest positive integers m,n such
that Sm = I and Tn = I.
DRAFT
-
11.
- Let A = . Compute A2
and A3. Is A3 = I? Determine aA3 + bA + cA2.
-
12.
- Let A = and B = . Compute
-
(a)
- A - A*,A + A*,(3AB)*- 4B*A and 3A - 2A*.
-
(b)
- (AB)[1,:],(AB)[3,:],(AB)[:,1] and (AB)[:,2].
-
(c)
- (B*A*)[:,1],(B*A*)[:,3],(B*A*)[1,:] and (B*A*)[2,:].
-
13.
- Let a,b and c be indeterminate. Then, can we find A with complex entries satisfying
A = ? What if A = ? Give reasons for your answer.
Definition 1.2.16. Let A ∈ Mn(ℂ).
-
1.
- Then, a square matrix B is said to be a left inverse of A, if BA = In.
-
2.
- Then, a square matrix C is called a right inverse of A, if AC = In.
-
3.
- Then, A is said to be invertible (or is said to have an inverse) if there exists a matrix
B such that AB = BA = In.
DRAFT
Lemma 1.2.17. Let A ∈ Mn(ℂ). If that there exist B,C ∈ Mn(ℂ) such that AB = In and
CA = In then B = C.
Proof. Note that C = CIn = C(AB) = (CA)B = InB = B. _
Remark 1.2.18. Lemma 1.2.17 implies that whenever A is invertible, the inverse is unique.
Thus, we denote the inverse of A by A-1. That is, AA-1 = A-1A = I.
Example 1.2.19.
-
1.
- Let A = .
-
(a)
- If ad - bc≠0. Then, verify that A-1 = .
-
(b)
- In particular, the inverse of equals .
-
(c)
- If ad - bc = 0 then prove that either A[1,:] = 0* or A[:,1] = 0 or A[2,:] = αA[1,:]
or A[:,2] = αA[:,1] for some α ∈ ℂ. Hence, prove that A is not invertible.
-
(d)
- Matrices , and do not have inverses. Justify your answer.
-
2.
- Let A = . Then, A-1
= .
-
3.
- Prove that the matrices A = and B = are not invertible.
Solution: Suppose there exists C such that CA = AC = I. Then, using matrix
product
DRAFT
But A[1,:] = A[2,:] and thus [1,0,0] = [0,1,0], a contradiction.
Similarly, if there exists D such that BD = DB = I then
But B[:,3] = B[:,1] + B[:,2] and hence I[:,3] = I[:,1] + I[:,2], a contradiction.
Theorem 1.2.20. Let A and B be two invertible matrices. Then,
-
1.
- (A-1)-1 = A.
-
2.
- (AB)-1 = B-1A-1.
-
3.
- (A*)-1 = (A-1)*.
Proof. Part 1. Let B = A-1 be the inverse of A. Then, AB = BA = I. Thus, by definition, B is
invertible and B-1 = A. Or equivalently, (A-1)-1 = A.
Part 2. By associativity (AB)(B-1A-1) = A(BB-1)A-1 = I = (B-1A-1)(AB).
Part 3. As AA-1 = A-1A = I, we get (AA-1)* = (A-1A)* = I*. Or equivalently,
(A-1)*A* = A*(A-1)* = I. Thus, by definition (A*)-1 = (A-1)*. _
DRAFT
We will again come back to the study of invertible matrices in Sections 2.2 and 2.3.1.
Exercise 1.2.21.
-
1.
- Let A be an invertible matrix. Then, prove that (A-1)r = A-r, for all integers r.
-
2.
- Find the inverse of and .
-
3.
- Let A1,…,Ar be invertible matrices. Then, prove that the matrix B = A1A2Ar is also
invertible.
-
4.
- Let A ∈ Mn(ℂ) be an invertible matrix. Then, prove that
-
(a)
- A[i,:]≠0T , for any i.
-
(b)
- A[:,j]≠0, for any j.
-
(c)
- A[i,:]≠A[j,:], for any i and j.
-
(d)
- A[:,i]≠A[:,j], for any i and j.
-
(e)
- A[3,:]≠αA[1,:] + βA[2,:], for any α,β ∈ ℂ, whenever n ≥ 3.
-
(f)
- A[:,3]≠αA[:,1] + βA[:,2], for any α,β ∈ ℂ, whenever n ≥ 3.
____________________
-
5.
- Let x* = [1 + i,2,3] and y* = [2,-1 + i,4]. Prove that y*x is invertible but yx* is not
invertible.
-
6.
- Determine A that satisfies (I + 3A)-1 = .
-
7.
- Determine A that satisfies (I - A)-1 = . [See Example 1.2.19.2].
-
8.
- Let A be a square matrix satisfying A3 + A - 2I = 0. Prove that A-1 = .
DRAFT
-
9.
- Let A = [aij] be an invertible matrix. If B = [pi-jaij], for some p ∈ ℂ, p≠0 then relate A-1 and
B-1.
Definition 1.3.1.
-
1.
- For 1 ≤ k ≤ m and 1 ≤ ℓ ≤ n, define a matrix kℓ ∈ Mm,n(ℂ) by (kℓ)ij =
Then, the matrices kℓ, for 1 ≤ k ≤ m and 1 ≤ ℓ ≤ n are called
the standard basis elements for Mm,n(ℂ).
So, if kℓ ∈ M2,3(ℂ) then 11 = = ,12 = =
and 22 = = .
-
2.
- Let A ∈ Mm,n(ℝ).
-
(a)
- Then, A is called symmetric if AT = A. For example, A = .
-
(b)
- Then, A is called skew-symmetric if AT = -A. For example, A = .
-
(c)
- Then, A is called orthogonal if AAT = AT A = I. For example, A = .
-
(d)
- Let A ∈ Mn(ℂ). Then, A is said to be a permutation matrix if A has exactly
one non-zero entry, namely 1, in each row and column. For example, In, for each
positive integer n, , , and are permutation
matrices. Verify that permutation matrices are Orthogonal matrices.
-
3.
- Let A ∈ Mm,n(ℂ).
DRAFT
-
(a)
- Then, A is called normal if A*A = AA*. For example, is a normal matrix.
-
(b)
- Then, A is called Hermitian if A* = A. For example, A = .
-
(c)
- Then, A is called skew-Hermitian if A* = -A. For example, A = .
-
(d)
- Then, A is called unitary if AA* = A*A = I. For example, A = .
Verify that Hermitian, skew-Hermitian and Unitary matrices are normal matrices.
-
4.
- Then, A is called idempotent if A2 = A. For example, A = is idempotent.
-
5.
- A vector u ∈ Mn,1(ℂ) such that u*u = 1 is called a unit vector.
-
6.
- A matrix that is symmetric and idempotent is called a projection matrix. For example, let
u ∈ Mn,1(ℝ) be a unit vector. Then, A = uuT is a symmetric and an idempotent matrix. Hence,
A is a projection matrix. In particular, let u = [1,2]T and A = uuT . Then, uT u = 1 and for
any vector x = [x1,x2]T ∈ M2,1(ℝ) note that
Thus, Ax is the foot of the perpendicular from the point x on the vector [12]T .
-
7.
- Fix a unit vector a ∈ Mn,1(ℝ) and let A = 2aaT - In. Then, verify that A ∈ Mn(ℝ) and
Ay = 2(aT y)a - y, for all y ∈ ℝn. This matrix is called the reflection matrix about the line
containing the points 0 and a.
-
8.
- Let A ∈ Mn(ℂ). Then, A is said to be nilpotent if there exists a positive integer n such that
An = 0. The least positive integer k for which Ak = 0 is called the order of nilpotency. For
example, if A = [aij] ∈ Mn(ℂ) with aij equal to 1 if i-j = 1 and 0, otherwise then An = 0 and
Aℓ≠0 for 1 ≤ ℓ ≤ n - 1.
DRAFT
Exercise 1.3.2.
-
1.
- Consider the matrices ij ∈ Mn(ℂ) for 1 ≤ i,j,≤ n. Is 1211 = 1112? What about
1222 and 2212?
-
2.
- Let {u1,u2,u3} be three vectors in ℝ3 such that ui*ui = 1, for 1 ≤ i ≤ 3, and ui*uj = 0
whenever i≠j. Then prove the following.
-
(a)
- Let U = [u1u2u3]. Then U*U = I.
-
(b)
- If A = uiui*, for 1 ≤ i ≤ 3 then A2 = A. Is A symmetric? Is A a projection matrix?
-
(c)
- If A = uiui* + ujuj*, for i≠j then A2 = A. Is A a projection matrix?
-
3.
- Let A,B ∈ Mn(ℂ) be two unitary matrices. Then, prove that AB is also a unitary
matrix.
-
4.
- Let A ∈ Mn(ℂ) be a Hermitian matrix (A* = A).
-
(a)
- Then, the diagonal entries of A are necessarily real numbers.
-
(b)
- For each B ∈ Mn(ℂ) prove that B*AB is a Hermitian matrix.
-
(c)
- Further assume that A2 = 0. Is it necessarily true that A = 0?
-
(d)
- Then, for any x ∈ Mn,1(ℂ), prove that x*Ax is a real number.
-
5.
- Let A ∈ Mn(ℂ). If x*Ax is a real number for every x ∈ Mn,1(ℂ) then prove that A is a
Hermitian matrix. [Hint: Use ej,ej + ek and ej + iek of Mn,1(ℂ) for x.]
-
6.
- Let A and B be Hermitian matrices. Then, prove that AB is Hermitian if and only if
AB = BA.
DRAFT
-
7.
- Let A ∈ Mn(ℂ) be a skew-Hermitian matrix. Then prove that
-
(a)
- the diagonal entries of A are either zero or purely imaginary.
-
(b)
- for each B ∈ Mn(ℂ) prove that B*AB is a skew-Hermitian matrix.
-
8.
- Let A be a complex square matrix. Then S1 = (A + A*) is Hermitian, S2 = (A - A*) is
skew-Hermitian, and A = S1 + S2.
-
9.
- Let A,B be skew-Hermitian matrices with AB = BA. Is the matrix AB Hermitian or
skew-Hermitian?
-
10.
- Let A be a nilpotent matrix. Prove that there exists a matrix B such that B(I + A) = I = (I + A)B.
[If Ak = 0 then look at I - A + A2 - + (-1)k-1Ak-1].
-
11.
- Are the matrices and orthogonal, for θ ∈ [-π,π)?
Definition 1.3.3. For a positive integer k, let us denote [k] = {1,…,k}. Also, let A ∈ Mm×n(ℂ).
-
1.
- Then, a matrix obtained by deleting some of the rows and/or columns of A is said to be
a submatrix of A.
-
2.
- If S ⊆ [m] and T ⊆ [n] then by A(S|T) , we denote the submatrix obtained from A by
deleting the rows with indices in S and columns with indices in T. By A[S,T], we mean
A(Sc|Tc), where Sc is the complement of S in [m] and Tc is the complement of T in [n].
Whenever, S or T consist of a single element, then we just write the element. If S = [m],
then we write A[S,T] = A[:,T] and if T = [n] then A[S,T] = A[S,:] which matches with
our notation in Definition 1.1.1.
DRAFT
-
3.
- If m = n, the submatrix A[S,S] is called a principal submatrix of A.
Example 1.3.4.
-
1.
- Let A = . Then, A[{1,2},{1,3}] = A[:,{1,3}] = , A[1,1] =
[1],A[2,3] = [2],A[{1,2},1] = A[:,1] = ,A[1,{1,3}] = [15] and A are a few
submatrices of A. But the matrices and are not submatrices of A.
-
2.
- Take A = , S = {1,3} and T = {2,3}. Then, A[S,S] = ,A[T,T] =
,A(S∣S) = and A(T∣T) = are principal submatrices of A.
Let A be an n × m matrix and B be an m × p matrix. Suppose r < m. Then, we can decompose the
matrices A and B as A = [PQ] and B = , where P has order n×r and H has order r ×p. That
is, the matrices P and Q are submatrices of A and P consists of the first r columns of A and Q
consists of the last m - r columns of A. Similarly, H and K are submatrices of B and H consists of
the first r rows of B and K consists of the last m - r rows of B. We now prove the following
important theorem.
Proof. The matrix products PH and QK are valid as the order of the matrices P,H,Q and K are
respectively, n×r,r ×p,n× (m-r) and (m-r) ×p. Also, the matrices PH and QK are of the same
order and hence their sum is justified. Now, let P = [Pij],Q = [Qij],H = [Hij], and K = [Kij]. Then,
for 1 ≤ i ≤ n and 1 ≤ j ≤ p, we have
Thus, the required result follows. _
Remark 1.3.6. Theorem 1.3.5 is very useful due to the following reasons:
-
1.
- The order of the matrices P,Q,H and K are smaller than that of A or B.
-
2.
- The matrices P,Q,H and K can be further partitioned so as to form blocks that are
either identity or zero or matrices that have nice forms. This partition may be quite useful
during different matrix operations.
-
3.
- If we want to prove results using induction then after proving the initial step, one assumes
the result for all r×r submatrices and then try to prove it for (r+1)×(r+1) submatrices.
For example, if A = and B = then AB = .
DRAFT
Suppose A =
and B = .
Then, the matrices P,Q,R,S and E,F,G,H, are called the blocks of the matrices A and B,
respectively. Note that even if A + B is defined, the orders of P and E need not be the
same. But, if the block sums are defined then A + B = . Similarly, if the
product AB is defined, the product PE may not be defined. Again, if the block products are
defined, one can verify that AB = . That is, once a partition of A
is fixed, the partition of B has to be properly chosen for purposes of block addition or
multiplication.
Exercise 1.3.7.
-
1.
- Complete the proofs of Theorems 1.2.5 and 1.2.14.
-
2.
- Let x = , y = , A = and B = .
-
(a)
- Then, prove that y = Ax gives the counter-clockwise rotation through an angle α.
-
(b)
- Then, prove that y = Bx gives the reflection about the line y = tan(θ)x.
-
(c)
- Let α = θ and compute y = (AB)x and y = (BA)x. Do they correspond to
reflection? If yes, then about which line(s)?
-
(d)
- Further, if y = Cx gives the counter-clockwise rotation through β and y = Dx gives the
reflections about the line y = tan(δ)x.
-
i.
- Then, prove that AC = CA and y = (AC)x gives the counter-clockwise rotation
through α + β.
-
ii.
- Then, prove that y = (BD)x and y = (DB)x give rotations. Which angles do
they represent?
DRAFT
-
3.
- Let A be an n × n matrix such that AB = BA for all n × n matrices B. Then,
prove that A is a scalar matrix. That is, A = αI for some α ∈ ℂ (use matrices in
Definition 1.3.1.1).
-
4.
- Consider the two coordinate transformations
x1 | = a11y1 + a12y2 |
x2 | = a21y1 + a22y2 |
and
y1 | = b11z1 + b12z2 |
y2 | = b21z1 + b22z2 |
.
-
(a)
- Compose the two transformations to express x1,x2 in terms of z1,z2.
-
(b)
- Does the composition of two transformations obtained in the previous part correspond
to multiplying two matrices? Give reasons for your answer.
-
5.
- For An×n = [aij], the trace of A, denoted tr(A), is defined by tr(A) = a11 + a22 + + ann.
-
(a)
- Compute tr(A) for A = and A = .
-
(b)
- Let A be a matrix with A = 2 and A = 3. If B =
then compute tr(AB).
-
(c)
- Let A and B be two square matrices of the same order. Then, prove that
-
i.
- tr(A + B) = tr(A) + tr(B).
-
ii.
- tr(AB) = tr(BA).
-
(d)
- Prove that there doesn’t exist matrices A,B ∈ Mn(ℂ) such that AB - BA = cI, for any
c≠0.
-
6.
- Let J ∈ Mn(ℝ) be a matrix having each entry 1.
-
(a)
- Then, verify that J2 = nJ.
-
(b)
- Also, for any α1,α2,β1,β2 ∈ ℝ, verify that there exist α3,β3 ∈ ℝ such that
DRAFT
-
(c)
- Let α,β ∈ ℝ such that α≠0 and α + nβ≠0. Now, define A = αIn + βJ. Then, use
the above to prove that A is invertible.
_________________________________________
-
7.
- Let A = .
-
(a)
- Find a matrix B such that AB = I2.
-
(b)
- What can you say about the number of such matrices? Give reasons for your answer.
-
(c)
- Does there exist a matrix C such that CA = I3? Give reasons for your answer.
-
8.
- Let A = and B = . Compute the matrix product AB using
the block matrix multiplication.
-
9.
- Let A = . If P,Q and R are Hermitian, is the matrix A Hermitian?
-
10.
- Let A = , where A11 is an n × n invertible matrix and c ∈ ℂ.
-
(a)
- If p = c - y*A11-1x is nonzero, then verify that
is the inverse of A.
-
(b)
- Use the above to find the inverse of and .
-
11.
- Let x ∈ Mn,1(ℝ) be a unit vector.
-
(a)
- Define A = In - 2xxT . Prove that A is symmetric and A2 = I. The matrix A is
commonly known as the Householder matrix.
-
(b)
- Let α≠1 be a real number and define A = In - αxxT . Prove that A is symmetric
and invertible. [The inverse is also of the form In + βxxT , for some β.]
-
12.
- Let A ∈ Mn(ℝ) be an invertible matrix and let x,y ∈ Mn,1(ℝ). Also, let β ∈ ℝ such that
α = 1 + βyT A-1x≠0. Then, verify the famous Shermon-Morrison formula
This formula gives the information about the inverse when an invertible matrix is modified by a
rank (see Definition 2.2.26) one matrix.
-
13.
- Suppose the matrices B and C are invertible and the involved partitioned products are defined,
then verify that that
-
14.
- Let A be an m × n matrix. Then, a matrix G of order n×m is called a generalized inverse
(for short, g-inverse) of A if AGA = A. For example, a generalized inverse of the matrix
A = [1,2] is a matrix G = , for all α ∈ ℝ. A generalized inverse G is called a
pseudo inverse or a Moore-Penrose inverse if GAG = G and the matrices AG and
GA are symmetric. Check that for α = the matrix G is a pseudo inverse of A.
Further, among all the g-inverses, the inverse with the least euclidean norm also has
α = .
In this chapter, we started with the definition of a matrix and came across lots of examples. We recall
these examples as they will be used in later chapters to relate different ideas:
-
1.
- The zero matrix of size m × n, denoted 0m×n or 0.
-
2.
- The identity matrix of size n × n, denoted In or I.
-
3.
- Triangular matrices.
-
4.
- Hermitian/Symmetric matrices.
-
5.
- Skew-Hermitian/skew-symmetric matrices.
-
6.
- Unitary/Orthogonal matrices.
-
7.
- Idempotent matrices.
-
8.
- nilpotent matrices.
DRAFT
We also learnt product of two matrices. Even though it seemed complicated, it basically tells that
multiplying by a matrix on the
-
1.
- left to a matrix A is same as operating on the rows of A.
-
2.
- right to a matrix A is same as operating on the columns of A.
DRAFT
DRAFT
DRAFT