Inner Product Spaces

Inner product spaces (IPS) are generalizations of the three dimensional Euclidean space, equipped with the notion of distance between points represented by vectors and angles between vectors, made possible through the concept of an inner product. These are broadly categorized as - real inner product spaces and complex inner product spaces and may be defined as follows. Let K commonly denote any of the fields R, or, C. Let V be a vector space over K. A map (. , .) : V ´ V ® K, satisfying: (a) (x, x) ³ 0, and (x, x) = 0, iff x = 0; (b) (x, y) = (y,x)* (star denoting complex conjugate); and, (c) (cx+y, z) = c(x, z) + (y, z), for all x, y, z Î V and c Î K, is called an inner-product in V.

Note that (a) asserts that (x, x) is a real number and is non-negative, even when K = C. A vector space V with an inner product on it is called an inner product space. If K = R, V is called a real inner-product space and if K = C, V is called a complex inner-product space. The scalar (x, y) is called the inner product of x and y. Note that in (b) the bar denotes complex conjugation, and so when K = R, (b) simply reads as (x,y) = (y,x). In the sequel, unless otherwise specified, an inner product refers to the field K which could be either of the fields R or C. Taking c = 1 in (c), (x+y, z) = (x,z) + (y,z). Putting x = 0 in it (0, z) = 0 and hence with y = 0 in (c), (cx,z) = c(x,z). The distributivity and conjugate linearity with respect to the second argument, namely (x, y+z) = (x, y) + (x, z), and (x, cz) = c(x, z), follow from the axiom (b) for an inner product.

Examples: In the following the verification of the inner product axioms is elementary and is left as an exercise.

(1) The Euclidean space Rn and the Unitary space Cn may be commonly defined as the inner product space Kn over the field K, with the inner product given by (x, y) = y* x = S 1£ i£ n xiyi* , (x, y Î K ). This inner product is known as the standard inner product in K . In the sequel, unless otherwise specified, K would stand for the above standard inner product space. Note that when K = R, the complex conjugation is redundant and (x, y) = y¢x = x¢y. (2) The sequence space l2(K) consisting of infinite square summable sequences x = (x1, x2, … , xn, … ), i.e., S i³ 1 |xn|2 < ¥ , with the inner product (x, y) = S i³ 1 xiyi* , (x, y Î l2(K)), is an infinite dimensional generalization of the inner product space K . (3) The space L2(X, m ), over K, of functions f that are square integrable on a domain X with respect to the measure m , i.e., ò X |f|2 dm < ¥ , with the inner product (f, g) = ò X fg* dm , is a further generalization of l2(K).

The following is the most improtant result about an inner product:

Cauchy's inequality. If x, y Î V, an inner product space, |(x, y)|2 £ (x, x)(y, y), with equality iff x and y are linearly dependent.

Proof: If x and y are linearly dependent, without loss of generality we may take y = l x, in which case both sides of the inequality equal |l |2(x, x) . If x and y are linearly independent, (by the axiom (a)) for any scalar l , 0 < (y-l x, y-l x) = (y, y) - 2Re [l (x,y)] + |l |2(x,x). Taking l = (y, x)/(x, x), this inequality reduces to 0 < (y,y) - |(x,y)|2/(x,x), which, since (x,x) > 0 (due to (a)), is equivalent to the Cauchy's inequality, with strict inequality, completing the proof. #

The Cauchy's inequality is also termed as Cauchy-Schwarz, or, Cauchy-Schwarz-Bunyakovsky inequality.

 

PROBLEMS

1. Verify the following for an inner product: (1) (0, y) = 0 = (x, 0); (2) (cx, y) = c(x, y), and (x, cy) = c* (x,y); (3) (cx, cx) = |c|2 (x, x); (4) (S cjvj , v) = S cj(vj, v), and, (v, S cjvj) = S cj* (v, vj ); (5) (S aiui, S bjvj) = S ai S bj* (ui, vj); (6) (x+y, x+y) = (x, x) + (y, y) + 2 Re(x,y); (7) if (x, y) = 0 for all x Î V, then y = 0; (8) if Re (x, y) ³ 0 for all x Î V, then y = 0; and, (9) if (x, y) = (x, z) for all x Î V, then y = z.

2. Show that the sum of two inner products on the same vector space V and a positive scalar multiple of an inner product on V is an inner product on V. Is the difference of two inner products an inner product?

3. Let (. , .) and (. , .) be two inner products on a vector space V. Show that there exist a scalar d such that (x, y) = (x, y) - c(x, y) is an inner product on V for all c < d, but for no c ³ d. How many such scalars d are there?

4. Describe explicitly all real inner products on R1 and on C1, and all complex inner products on C1.

5. Let (. , .) be an inner product on a vector space V over C. Show that (. , .) continues to remain an inner product if V is regarded as a vector space over R.

6. Verify that the standard inner product on Kn is an inner product and that it is characterized by (ei, ej) = d ij, where ei 's are the standard unit (basis) vectors.

7. Let {v1, v2, … , vn} be an ordered basis of an inner product space V. Show than given any n-scalars c1, … , cn, there exists a unique v = S 1£ i£ n xjvj Î V such that (v, vj) = cj, 1 £ j £ n. Hence derive a system of linear equations for determining xj 's.

8. Determine all linear operators T on Kn for which (Tv, v) = 0 for all v.

9. If A Î Kn´ n, prove that fA(x, y) = y* Ax is an inner product on the vector space Kn over K iff A is positive definite, i.e., x* Ax is positive for all non-zero x Î C n.

10. Verify for an inner product that (u+v, u+v) + (u-v, u-v) = 2(u, u) + 2(v, v). Interpret this as saying that the sum of squares of lengths of diagonals equals the sum of squares of lengths of all the four sides of a parallelogram. Deduce Pythagoras theorem from this.

11. The matrix M of an inner product with respect to an ordered basis b is defined by: (u, v) = [v]b * M[u]b . If K = C (R), prove that M is hermitian (real symmetric) and is unique.

12. Show that (x, y) = x1y1 -3x2y1 -3x1y2 +10x2y2, defines an inner product on R2. What is the matrix of this inner product relative to the standard basis in R2? What are all bases with respect to which the matrix of this inner product becomes the identity matrix?

13. For the real inner product (f, g) = ò [0,1] f(t)g(t)dt, on L2[0, 1], verify (S 0£ j£ n-1 ajxj, S 0£ k£ n-1 bkxk) = S 0£ j,k£ n (ajbk)/(j+k+1).

14. Prove that the Hilbert matrix (1/(i+j-1))1£ i,j£ n is positive definite.

15. Determine the inner product on R2, the quadratic form of which is given by P xP 2 = (2x1-3x2)2 + (3x1-2x2)2.

16. If (u, v) is an inner product in Kn such that |(u, v)| £ C|v*u|, for all u, v Î Kn, where C is a constant, prove that (u, v) = cv*u, for all u, v Î Kn, where c is a constant.

17. Which of (a) (x, y) = x* y, (b) (x, y) = y* x, (c) (x, y) = y¢x, and, (d) (x, y) = x¢y, is an inner product in Cn? Which of these are inner products in R n.

18. Let {ui}1£ i£ n be a basis of an inner product space V. Show that <S, T> = S 1£ i£ n (Sui, Tui) is an inner product in L(V).

19. Show that the standard basis {Eij} of Kn´ n= L(Kn), is orthonormal with respect to the inner product: <S, T> = S 1£ i£ n (Sei , Tei).

 

Norm Induced by an Inner Product

An inner product on a vector space V induces a natural norm in V by defining P xP 2 = (x, x) , x Î V. Of the norm axioms: (a) P xP ³ 0, and P xP = 0 iff x = 0; (b) P cxP = |c|P xP ; and, (c) P x+yP £ P xP + P yP , (a)-(b) trivially follow from the inner product axioms. For the triangle inequality (c), we have P x+yP 2 = P xP 2 + P yP 2 + 2 Re(x, y) £ P xP 2 + P yP 2 + 2|(x,y)| £ (P xP + P yP )2, as follows from the Cauchy-Schwarz inequality.

In the sequel, unless otherwise specified, the norm in an inner-product space context would mean the induced norm defined as above. A sequence of vectors {x(n)} in a normed (vector) space V (over K) is called Cauchy if given any e > 0, there exists an N such that P x(m) -x(n)P < e , for all m, n > N. A sequence {x(n)} is said to converge if there exists an x Î V such that P x(n) -xP ® 0, as n ® ¥ . A normed space V is said to be complete if every Cauchy sequence in it converges. A complete inner product space is called a Hilbert space. (In the literature, however, several authors exclude finite dimensional inner product spaces from being called Hilbert spaces). Bolzano-Weierstrass theorem and the equivalence of any two norms in a finite dimensional space imply that all finite dimensional vector spaces over K are complete. A subspace W of a normed space V is called closed if W ' w(n) ® w, implies that w Î W (i.e., a sequence in W can converge only to an element in W itself).

 

PROBLEMS

1. Let V be an inner product space. For what real values of q does the function d(u, v) = (u-v, u-v)q , (u, v Î V), defines a distance in V, i.e., there hold: (1) d(u, v) ³ 0; (2) d(u, v) = 0 iff u = v; (3) d(u, v) = d(v, u) (symmetry); and, (4) d(u, w) £ d(u, v) + d(v, w), u, v, w Î V (triangle inequality).

2. Let V be a complex inner product space. If u, v Î V are such that (u, v) is real, i.e., u has a zero axial-twist with respect to y, show that P u+ivP 2 = P uP 2 + P vP 2 and that u + iv is orthogonal to u - iv iff u is orthogonal to v and P uP = P vP .

3. Let V be an inner product space and G a finite dimensional subspace of V. Let v Î V\G and g Î G . Prove that P v-g0P £ P v-gP , for all g Î G , i.e., g0 is a best approximation of v from G, iff there exists an element u Î V such that (i) P uP = 1, (ii) (g, u) = 0, for all g Î G , and, (iii) (v, u) = P v-g0P .

4. Let V be an inner product space and v1, v2, … , vn Î V be linearly independent. Let G = < v1, v2, … , vn >, the span of the vectors v1, v2, v3, … , vn. Show thatinf P v - gP = det ([(vi, vj)]0£ i,j£ n)/det ([(vi, vj)]1£ i,j£ n] ).

5. Consider L2(0,1), the Hilbert space of square integrable real valued functions on the interval (0,1) with the inner product (f, g) = ò [0,1] f(x)g(x)dx. Let a i > -1/2, 0 £ i £ n, and en = inf {P xa 0 - S 1£ i£ n cixa iP : ci Î R}. Prove that en2 = D 0/D 1 = [1/(2a 0+1)]P 1£ i£ n {(a 0 -a i)2/(a 0+a i+1)}, where D k = P k£ i<j£ n {(a i -a j)2/P k£ i,j£ n (a i+a j+1)} = det (1/(a i+a j+1)) k£ i,j£ n. If 0 < a 0 < a 1 < … , deduce that any f Î L2(0,1) can be approximated arbitrarily closely by linear combinations of the functions {xa i, i ³ 1} iff the series S i³ 1 a i diverges.

6. Verify the parallelogram law for a norm induced by an innner product: P a+bP 2 + P a-bP 2 = 2P aP + 2P bP 2.

7. In a complex IPS verify the polarization identity: 4(x,y) = P x+yP 2 -P x-yP 2 +iP x+iyP 2 -iP x-iyP 2.

8. In a real inner product space verify the identity: 4(x,y) = P x+yP 2 -P x-yP 2 .

 

Orthogonality in Inner Product Spaces

Let V be an inner product space. Given two vectors x, y in V, x is said to be orthogonal to y, if (x, y) = 0. Symbolically one writes x ^ y (read x is perpendicular, or orthogonal, to y). It is clear that x ^ y Û y ^ x. If in addition to being orthogonal, the vectors x and y are also normalized, i.e., PxP = PyP = 1, we say that x and y are orthonormal. A set S is said to be orthogonal to a set S 2 (S1 ^ S 2) if s1 ^ s2 for all s1 Î S1 and s2 Î S 2.

A subset S of an inner product space is said to be orthogonal if u, v Î S and u ¹ v imply (u, v) = 0. A subset S is called orthonormal if in addition to being orthogonal, every element of it has unit norm. Every orhtonormal set is linearly independent, for if S civi = 0, taking inner product with vj we get cj = 0. It follows that an orthogonal set is linearly independent iff the zero vector does not belong to it.

A basis of an inner product space V which is orthonormal is called an orhonormal basis or a unitary basis of V. If a basis if just orthogonal it is called an orthogonal basis. A triangulization, or a diagonalization of a linear operator T on V by a unitary basis is referred to as a unitary triangulization, or unitary diagonalization.

 

PROBLEMS

1. Let T be a linear operator on a finite dimensional inner product space V. If Tu is orthogonal to u for all u Î V, show that T = 0.

2. Is it true that every inner product space V ¹ {0} has an othonormal basis?

3. Prove that any orthonormal set of vectors in an inner product space is linearly independent. Is the converse true?

 

The Angle Between Two Vectors

In view of the Cauchy-Schwarz inequality, for a real inner-product space, defining cos q = (x,y)/(||x||||y||), we have -1 £ cos q £ 1. Taking the principal value of q Î [0, p ], we get the notion of angle q between two vectors x and y, which coincides with the usual angle between two vectors in R 2 and R 3. For a complex inner-product space V the notion of an angle between vectors, though not very neat, may however be defined as follows:

First of all let us consider a one dimensional subspace <v> = {cv: c Î K } spanned by a vector v of V . How does it look like? Unlike when the field K is R in which case <v> could be imagined as a line passing through origin, in the case K = C , it requires something like an imaginary surface of a right circular cylinder whose axis lies along the direction of v passing through the origin, so that assuming its base to be in the plane perpendicular to the axis through the origin and taking v to be represented by an arbitrary point on the surface at a positive height ||v|| marking the angle zero on the circle, the vector cv = reiq v, (q Î [-p , p ]) could be represented as a point on the cylinder at signed height r and making angle q in the anticlockwise sense. The angle q could be imagined as an axial twist of cv over v. Then, interpreting the polar representation (x,y)/( P xP P yP ) = reif , (-1 £ r £ 1, -p /2 £ f £ p /2), we find that the vector x has an additional minimal magnitude axial-twist of f radians over y. Thus if we wish to do meaningful real geometric operations compatible with the inner product we should work with the pair of vectors e-if x, y, or, the vectors x, eif y, the cosine of the real angle q between which is given by cos q = r Î [-1,1]. The principal value of q Î [0, p ] then may be defined as the required angle between x and y. In the case of real inner-product spaces this notion is compatible with the standard one.

 

PROBLEMS

1. Show that given any basis in a vector space V over K, there exists an inner product on V with respect to which the basis members are orthonormal.

2. If w Î Rn, w¢ w = 1 and A = I - 2ww¢ , show that A reflects vectors in the hyperplane H = {x Î Rn: w¢ x = 0} and is an orthogonal matrix. Interpret the complex case w Î Cn.

 

Gram-Schmidt Orthonormalization Procedure

Given a sequence {v1, v2, … , vm} of linearly independent vectors in an inner product space, the Gram-Schmidt procedure constructs another sequence {u1, u2, … , um} of vectors such that span {v1, v2, … , vk} = span {u1, u2, … , uk}, k = 1, 2, … , m, while in addition the vectors ui 's are orthonormal: (ui, uj) = d ij, i, j = 1, 2, ... , m. The construction goes as follows: Put u1 = v1/P v1P and for i = 2, 3, ... , m, form the intermediate vectors wi = vi - S 1£ j<i-1 (vi, uj)uj and put ui = wi/P wiP .

From the above construction it is clear that wi, ui belong to the span of{v1, v2, … ,vi}, wi ¹ 0 and that P uiP = P wiP /P wiP , 1 £ i £ m. Further, since ui is a non-zero scalar multiple of wi, to prove (ui, uk) = 0, i ¹ k, it is sufficient to prove that (wi,uk) = 0, i > k. For this: (wi,uk) = (vi, uk) - S 1£ j£ i-1 (vi,uj)(uj, uk) = (vi, uk) - S 1£ j£ i-1 (vi,uj)(d jk)= (vi, uk) - (vi, uk) = 0, completing the proof. #

Note that if the vi 's are linearly dependent, for definiteness, say, if k is smallest such that v1, v2, … , vk are l.d. then wk becomes zero and the Gram-Schmidt process breaks down, and conversely. Thus, the G-S process could also be used to check the l.i. of vectors in an i.p.s.

Corollary. Every finite dimensional inner product space V (¹ {0}) has an orthonormal basis.

Proof: The G-S process applied to a basis gives an orthonormal basis. #

Unitary and orthogonal matrices have a very simple interconnection with a set of orthonormal vectors: An m´ n U is sub-unitary (sub-orthogonal) i.e., U* U = I, iff its column vectors uj 's are orthonormal in Cn (Rn). Note that an m´ n U being sub-unitary (sub-orthonormal) implies that m ³ n and that when m = n, U becomes unitary (orthonormal) and then also UU* = I.

 

Theorem. If V and W are two n-dimensional inner product spaces over K, there exists an isomorphism h : V ® W, such that (h (x), h (y)) = (x,y), x, y Î V.

Proof: Take {v1, v2, … , vn} and {w1, w2, … , wn} to be orthonormal bases of V and W, respectively, and for x = S 1£ i£ n aivi, defining h (x) = S 1£ i£ n aiwi, h is an isomorphism and with y = S 1£ i£ n bivi, (h (x), h (y)) = (h S 1£ i£ n aivi, h S 1£ j£ n bjvj) = (S 1£ i£ n aiwi, S 1£ j£ n bjwj) = S 1£ i£ n aibj* = (x, y). #

Note that since the map h in the above theorem preserves the inner products, it preserves the distances and so it is an isometry. In particular, Taking W as Kn, it follows that problems involving an n-dimensional inner product space can be isometrically transferred to the standard inner product space K n (i.e., R n, or C n).

The adjoint T* of a linear operator T : V ® W, where V and W are inner product spaces, is defined implicitly by the requirement (Tx, y) = (x, T*y), x Î V, y Î W, and is a linear operator from W to V. If V and W are finite dimensional inner product spaces, the existence, uniqueness and a construction of the adjoint is easy.

If T is a matrix operator A : K n ® K m, and as usual, A* for a complex matrix A denotes its adjoint matrix, i.e., the transpost of the complex conjugate of A, then for the standard inner products on K n and K m, (x, T* y) = (Tx, y) = (Ax, y) = (x, A* y) so that T* = A* , i.e., for a matrix operator over Euclidean or Unitary spaces with the standard inner products, the adjoint operator is nothing but the adjoint of the matirx.

Let V = {v1, v2, … , vn} and W = {w1, w2, … , wn} to be orthonormal bases of V and W, respectively. Let [x] = [x]V , [y] = [y]W , [T] = [T]V ,W . Then (Tx, y) = ([T][x], [y]) = ([x], [T]* [y]) = (x, T* y) implies that T* y = V [T]* [y] and that [T* ]W ,V = ([T]V ,W )* . Conversely, an operator T* from W to V defined by y ® V [T]* [y] satisfies (Tx, y) = (x, T*y), x Î V, y Î W is the same as the adjoint of T* .

Thus, the adjoint T* of T is the operator whose matrix with respect to a pair of orthonormal bases W , V is the adjoint of the matrix of T with respect to the pair V , W . T is called sub-unitary if T* T = I, the identity operator on V. T is said to be unitary if T* T = I and TT* = I (or equivalently that both T and T* are sub-unitary). A linear operator T on an inner product space V is called self adjoint if T* = T, unitary if T*T = I (in which case the accompanying relation TT* = I automatically holds!) and normal if T* T = TT* , i.e., T and T* commute. In the case of a finite dimensional inner product space V, the self-adjoint, unitary and normal operators are easily associated with the corresponding matrices as follows:

 

Theorem. If T is a linear operator on a finite dimensional inner product space V, b is an ordered unitary basis of V and A = [T]b , then T is self-adjoint, unitary or normal according as A is.

Proof: Let b = {w1, w2, … , wn}. For any u, v Î V, let x = [u]b , and y = [v]b . Then, (u, v) = (x,y), the standard inner product of x and y in Kn. Hence (x,[T* ]b y) = (u, T* v) = (Tu, v) = (Ax, y) = (x, A* y), implying that A* = [T* ]b , from which the results are clear. #

Theorem. Let T be a linear transformation from an IPS V to an IPS W . Then N (T) ^ R (T* ) and R (T) ^ N (T* ).

Proof: Let u Î N (T) and v Î R (T* ). Then v = T* w for some w Î W . Hence, (u, v) = (u, T* w) = (Tu, w) = (0, w) = 0. Similarly, if w Î R (T) and v Î N (T* ). Then w = Tu for some u Î V . Hence, (w, v) = (Tu, v) = (u, T* v) = (u, 0) = 0. #

 

PROBLEMS

1. Let T : V ® W, both being inner product spaces with b and g being two unitary bases of the respective spaces. If A = [T]b ,g , show that [T* ]g ,b = A* , [T* T]g = A* A and [TT* ]b = AA* . Deduce that T is sub-unitary or unitary according as A is.

2. Let u1, u2, … , un and v1, v2, … , vn be two bases in an inner product space. Let T and S be two linear operators such that (Tui, vj) = (ui, Svj), i, j = 1, ... , n. Prove that T = S* .

3. Let b be an orthonormal basis and A = [T]b , where T is a linear operator. Prove that A* = [T* ] . Does this result remain true if B is not necessarily orthonormal?

4. Prove that for a linear operator T on a finite dimensional inner product space V, the following statements are equivalent: (a) T is unitary; (b) (Tu,Tu) = (u, u), for all u Î V; (c) (Tu, Tv) = (u, v), for all u, v Î V.

5. Prove that a linear operator T on a finite dimensional inner product space V is normal iff ?Tu? = ?u?, for all u Î V.

 

Projections and Orthogonal Projections

A linear operator (or a matrix) P is called a projection (projection operator or a projection matrix) iff P2 = P, i.e., P is idempotent. Writing I = P + (I-P) = P Å (I-P) is a direct sum in the sense that R (P)Ç R (I-P) = {0} and that R (P) and R (I-P) are invariant, respectively under P and I-P. To see this, if x Î R (P), then x = Py for some y and so Px = P2y = Py = x Î R (P), i.e., P is an identity operator on R (P). As, (I-P)2 = I-2P+P2 = I-P, I-P itself is a projection and I-P acts as an identity on R (I-P) which therefore is invariant under I-P. Now let R (P)Ç R (I-P) ' x = Py = (I-P)z. Then x = Py = P2y = (P-P2)z = 0. Hence R (P)Ç R (I-P) = {0}. Note that P, respectively, I-P act as null operators on R (I-P) and R (P).

Also we have V = R (P)Å R (I-P) and x = Px + (I-P)x, is the unique decomposition of an x Î V into elements of R (P) and R (I-P). Pooling up the bases of R (P) and R (I-P), it follows that a projection is diagonalizable and its eigenvalues are only 0 and 1's, and conversely. A projection P is called an orthogonal projection if R (P) ^ N (P) = R (I-P).

Theorem. A projection P is an orthogonal projection iff P* = P (i.e., P is self-adjoint).

Proof: P is orthogonal projection Û P is projection and (Px, (I-P)y) = 0 for all x, y Û P is projection and P* = P*P Û P is a projection and P = P*P = P* Û P is a projection and P* = P. #

Corollary 1. P is an orthogonal projection iff it is unitarily diagonalizable with eigenvalues 0 and 1's.

Proof: Self-adjointness implies normalcy, which implies unitary diagonalization. Eigenvalues 0 and 1's only and diagonalization implies projection property. Projection and unitary diagonalization implies projection and self-adjointness which is equivalent to orthogonal projection property. #

Corollary 2. If an m´ n A is sub-unitary (A*A = I, m ³ n) then AA* is an orthogonal projection.

Proof: (AA*)2 = A(A*A)A* = AIA* = AA*. Hence AA* is a projection. But (AA*)* = (AA*). Hence AA* is also self-adjoint and hence an orthogonal projection. #

Theorem. If A is a projection matrix over F , tr(A) = r(A)1, where 1 is the unity in F .

Proof: Iff A is a projection, for some non-singular C, A = C-1 diag (d1, … , dn) C, where di = 0, or, 1. Using tr AB = tr BA, we have tr A = tr (C-1 diag (d , ... , d ) C) = tr (diag(d , ... , d ) CC-1) = tr (diag(d1, … , dn)) = r(A)1. #

 

Corollary. If the field is of characteristic zero and I = E1 + E2 + … + Ek, where Ei 's are projections, then EiEj = 0, i ¹ j.

Proof: I = E1 + … + Ek implies that V = E1V + … + EkV. Let W = EiV, 1 £ i £ k. As dim Wi = rank Ei and S 1£ i£ k tr Ei = tr I, the characteristic of the field being zero, n = dim W1 + ... + dim Wk . It follows that V = W1 ÅÅ Wk. Now, N (E1) = R (I - E1) = R(E2 + … + Ek) and therefore N (E1) Ì W2 ÅÅ Wk . But R (E1) = W1. Hence since W1 Ç (W2 ÅÅ Wk) = {0} and V = R (E1) Å N (E1), N (E1) = W1 ÅÅ Wk. It follows that E1Ej = 0, j ¹ 1 (as (E1Ej)v = E1(Ejv) Ì E1Wj ). By reordering, E1 could be taken to be any of the Ei 's. Hence EiEj = 0, i ¹ j. #

 

PROBLEMS

1. Let V be an (n+1)-dimensional vector space over a field F and T a linear operator on V. Let c0, c1, … , cn Î F be distinct. Let p(x) = P 0£ i£ n (x-ci), and lk(x), 0 £ k £ n denote the basic Lagrange interpolation polynomials given by lk(x) = P j¹ k [(x-cj)/(ck -cj )], where j runs from 0 to n. Let Ek = lk(T). Show that : (i) I = E1 + E2 + … + En; (ii) T = S 0£ k£ n ckEk ; (iii) if p(T) = 0, show that Ek 's are projections such that Ei Ej = 0, 0 £ i ¹ j £ n.

2. Find, using observations alone and without computations, an orthogonal basis of R4 one member of which is to be (1,1,1,1)¢. [Hint: Use 0, ± 1 as components.]

3. Let W be the subspace of R4 consisting of all vectors which are orthogonal to both u = (1, 2, 3, 4)¢ and v = (4, 3, 2, 1)¢ . What relations do the components of w Î W satisfy? Use these to find an orthonormal basis for W.

4. Use the Gram-Schmidt process on u = (1, 0, 0)¢, v = (1, 2, 0)¢, and w = (1, 2, 3)¢, to obtain an orthonormal basis for R3. Redo the problem with u = (1, 2, 3)¢, v = (1, 2, 0)¢, w = (1, 0, 0)¢. What do you conclude?

5. Find an orthonormal basis for the subspace of C3 spanned by u = (1, 0, i), and, v = (1, 0, -i), using observations only.

6. Let W be the subspace of R2 spanned by the vector (11, 12)¢. and P denote the orthogonal projection of R2 onto W. Find: (a) a formula for P(x1, x2)¢; (b) the matrix of P in the standard ordered basis; (c) W^ ; (d) an orthonormal basis in which P is represented by the matrix

.

Also, show that the matrix of P with respect to any orthonormal basis is of the obove type. Repeat (a)-(d) with another inner product on R the quadratic form of which is given by ?(x ,x )? = (2x1 -x2)2 + (x1 -2x2)2.

7. Find all inner product on R2 such that (e1, e2) = 2, where ei 's are the standard basis vectors.

8. Let V denote the vector space of real valued functions, square integrable with respect to a measure m on an interval I. Show that (f,g) = ò I fgdm is an inner product on V. If every monomial belongs to V, show that there exists a unique sequence of monic polynomials Pn, n ³ 0, of degree n such that (Pn, Pm) = 0, n ¹ m. Prove that for all n ³ 1: (a) there exist scalars an, and, bn such that Pn+1(x) = (x + bn)Pn(x) + cnPn-1(x), n ³ 1; (b) cn < 0; (c) all the n-zeros of Pn are real, simple and lie in the interior of the interval I; (d) between any two zeros of Pn+1 there lies a zero of Pn; (e) if Qn is a polynomial of degree n such that (Qn, Rn-1) is zero for all polynomials Rn-1 of degree £ n-1, then Qn = cPn for some scalar c; and (f) if P n denotes the subspace of polynomials of degree £ n, show that the orthogonal complement of P n in P m equals the span of Pn+1, Pn+2, … , Pm, (m > n).

9. With dm (x) = w(x)dx, for the intervals I and weight functions as follows, compute the first four monic orthogonal polynomials: (a) I = [-1,1], w(x) = 1 (Legendre polynomials, Pn); (b) I = [-1,1], w(x) = 1/Ö (1-x2) (Tchebycheff polynomials, Tn); (c) I = [0, ¥ ), w(x) = e-x (Laguerre polynomials, Ln); (d) I = (-¥ , ¥ ), w(x) = exp {-x2} (Hermite polynomials, Hn) and (e) I = [-1,1], w(x) = (1-x)a (1+x)b (Jacobi polynomials, Jn(a ,b )).

10. Let V be the vector space of all n´ n matrices over C, with the inner product (A, B) = tr (AB* ). Find the orthogonal complement of the subspace spanned by: (i) diagonal matrices, (ii) real symmetric matrices, (iii) hermitian matrices, (iv) skew hermitian matrices, (v) real skew symmetric matrices, (vi) unitary matrices, (vii) non-singular matrices, (viii) lower triangular matrices, and (ix) upper triangular matrices.

11. If V is an inner product space, and v1, … , vn Î V, show that for all u, v Î V there holds (u, v) = S 1£ k£ n (u, vk)(vk,v), iff v1, … , vn are orthonormal and dim V = n.

12. Let P be a projection of an inner product space V on a finite dimensional subspace W. Show that (Pu, v) = (u, Pv) for all u, v Î V iff P is an orthogoanl projection.

13. For any subset S of an inner product space V, show that (S ^ )^ contains <S >, the subspace spanned by S . When <S > is finite dimensional, prove the equality (S ^ )^ = <S>. Give a counter example when <S > is not finite dimensional

14. If b = {v1, … , vn} is an ordered orthonormal basis of an inner product space V, show that the matrix A = [T]b of a linear operator T on V could be calculated from aij = (Tvj,vi).

15. Suppose V = W1 Å W2 and that f1 and f2 are inner products on W1 and W2, respectively. Let P be the projection of V on W1 along W2. Show that f(u, v) = f1(Pu, Pv) + f2((I-P)u, (I-P)v) is the unique inner product on V such that: (a) W2 = W1^ ; (b) f(u, v) = fk(u, v), when u, v are in Wk , k = 1, 2; and, (c) P is the orhogonal projection of V on W1.

16. Show that in an inner product space V, of all projections of V on a subspace W, the orthogonal projection on W is characterized by: P PvP £ P vP for all v Î V.

17. Consider the vector space C[-1,1] with the inner product (f, g) = ò [-1,1] f(t)g(t)dt. Find the orthogonal complement of the subspace of even functions? What is the orhogonal complement of the subspace of odd functions? Repeat the exercise if the inner product is changed to (f, g) = ò [-1,1] w(x)f(t)g(t) dt, where w(x) > 0 is general.

18. If P and Q are commuting (orthogonal) projections, then PQ is also (orthogonal) projection.

19. A hermitian matrix has only real eigenvalues and as a matrix operator it is self adjoint with respect to the standard inner product. Is is true that the eigenvalues of any self adjoint operator on a finite dimensioal inner product space are real?

20. Given a diagonalizable operator T on Rn, does there always exist an inner product on Rn with respect to which the eigenvectors of T are orthonormal?

21. For x = (x1, x2)¢, find T* x if the linear operator T on C is defined through: (a) Te1 =(1, i)¢, Te2 = (i, -1)¢; (b) Te1 = (1+i, 1-i)¢, Te2 = (1-i, 1+i)¢; and, (c) Te1 = (1+i, i-1)¢, Te2 = (1-i, 1+i)¢. Also determine the matrices of the operators T, T* and TT* -T* T in the standard ordered basis and describe the ranges and the null spaces of these operators and check if they are related?

22. Let T denote the linear operator on C3 whose matrix A in the standard ordered basis is defined by: ajk = ij-k, (i = Ö (-1)). Find orthonormal bases for the null spaces of T and T* .

23. Let V be a finite-dimensional inner product space and T a linear operator on V. Prove that V = R (T* ) Å ^ N (T) = R(T) Å ^ N(T* ), where W1 Å ^ W2 means a direct sum in which the component subspaces are orthogonal to each other.

24. Let V be a finite-dimensional inner product space and T a linear operator on V. Prove that T is invertible iff T* is invertible and that there holds (T* )-1 = (T-1)* .

25. Let y = (y1, … , yn)¢ and z = (z1, … ,zn)¢ be two fixed vectors. If T is given by Tx = (x, y)z, x Î Cn, determine the matrix of T in the standard ordered basis and find the range and null spaces of T* .

26. Show that the product of two self-adjoint operators T and S is self-adjoint if and only if their commutator TS-ST is so.

27. If {uk}1£ k£ n is an orthonormal basis of an inner product space V, show that Tv = S 1£ kn (v, T* uk)uk, for every linear operator T on V. If V = Kn (K = R or C), deduce that: I = S 1£ k£ n ukuk* .

28. Consider the vector space of real polynomials of degree £ n with the inner product (P, Q) = ò [0,1] P(x)Q(x) dx. If D denotes the differentiation operator and A the left averaging operator given by AP(x) = x ò [0,x] P(t)dt, find D* and A* . Are they invertible?

29. In Cn´ n with the inner product (A, B) = tr (AB* ), determine T* if T is defined by: (a) TA = BA, (b) TA = AB, (c) TA = BAC, and, (d) TA = AB-CA+EAF.

30. Prove that a projection operator on a finite dimensional inner product space is self adjoint iff it is normal.

31. Show that a linear operator T on a finite-dimensional complex inner product space is self-adjoint if and only if (Tv, v) is real for every v Î V.

32. Find a unitary matrix which is not complex-orthogonal, and find a complex-orthogonal matrix which is not unitary.

33. Consider the vector space Cn´ n with the inner product (A, B) = tr (AB* ). For M, N Î Cn´ n, let TM,N be the linear operator on Cn´ n defined by TM,N (A) = MAN. Show that TM,N is unitary if and only if nM/tr(M) and nN/tr(N) are unitary, and |tr(M)tr(N)| = 1.

34. Let Cn(R) denote Cn regarded as a real vector space.

(a) Show that (x, y) = Re (y* x) defines an inner product on C n(R).

(b) Exhibit an isometric isomorphism of Cn(R) onto R 2n with the standard inner product.

(c) For each A Î C n´ n, let TA be the linear operator on C n(R) defined by TA(x) = Ax. Show that (TA)* =TA*

(d) For which complex A is TA self-adjoint?

(e) For which A is TA unitary?

(f) For which A is TA positive definite?

(g) What is det (TA)?

(h) Find the matrix of T in the ordered basis: {e1, e2, … , en, ie1, ie2, … , ien}.

(i) If T is a linear operator on C n(R), find necessary and sufficient conditions for T to be a TA.

(j) Find a unitary operator on C n(R) which is not a TA .

35. Considering R as the Euclidean plane, write down the transformation equations governing: (a) anti-clockwise rotation by an angle q about a point (x ,y )¢; (b) reflection in the line y = mx +c, where m = tan a.

36. Prove that the operation of rotating the plane about a point is a linear operation in R2 iff the point is the origin.

37. Show that the operation of reflecting points in the plane in a line is a linear operation iff the line passes through the origin.

38. From the unitary isomorphism (x, y)¢ ® x + iy = z = reiq , of R2 onto C(R), it follows that a linear rotation (anti-clockwise) of the plane R by an angle f increases q by f so that a point P = (x, y)¢ moves to the point P = (x , h )¢, with z = rei(q +f ) = x + ih , so that x = x cos f - y sin f and h = x sin f + y cos f . Hence, show that the matrix T of this transformation, with respect to the standard ordered basis, is given by

T = .

39. Show that the reflection of a point in the x-axis in the plane R2 corresponds to a linear operator whose matrix with respect to the standard ordered basis is given by

.

40. Prove that the operation of reflection about the line y = mx, (m = tan f ), is equivalent to a linear rotation by the angle -f , followed by a reflection in the x-axis and then another linear rotation by the angle f . Hence, or otherwise, show that this constitutes a linear operation in R whose matrix is given by

Rf = .

41. Prove that a linear rotation is never a reflection and a reflection is never a linear rotation, but that a product of linear rotations and linear reflections is either a linear rotation or a linear reflection.

42. Show that every unitary operator in R is either a linear rotation or a linear reflection.

43. Show that the adjoint of a linear rotation or a linear reflection is the inverse of the operator.

44. Prove that (a) a linear rotation followed by a linear rotation is a linear rotation, (b) a linear rotation followed by a linear reflection is a linear reflection, (c) a linear reflection followed by a linear rotation is a linear reflection, and that, (d) a linear reflection followed by a linear reflection is a linear rotation. Hence deduce that a product of linear rotations and linear reflections is a linear reflection if it has an odd number of linear reflections and that it is a linear rotation otherwise. Generalize for rotations about arbitrary points and reflections in arbitrary lines.

45. Let W be a subspace of a finite dimensional vector space V and let P denote the orthogonal projection on W. If U = 2P-I, show that U may be interpreted as a reflection of V in the subspace W and that U is both self-adjoint and unitary. If V = R and W is the span of the vector (1, 1, 1)¢ , find the matrix of U in the standard ordered basis and interpret U as a 180 -rotation about an appropriate axis.

46. If w Î Rn is a unit vector (w¢w = 1), the n´ n matrix operator H = (I-2ww¢) is called a Householder rotation. Prove that H is actually a reflection in the hyperplane w^ , whereas -H is a reflecion in the w-axis, which for n ³ 3 may be regarded as a rotation about the w-axis.

47. If T is any self-adjoint linear operator on a complex inner product space V, prove that P v+iTvP = P v-iTvP = (P vP 2 + P TvP 2)1/2 , v Î V and deduce that I ± iT is non-singular and hence that if V is finite dimensional, the related Cayley transform of T, defined by U = (I-iT)(I+iT)-1, is unitary. Relate the result with the fact that the conformal map f(z) = (1-iz)/(1+iz) takes the x-axis into the unit circle.

48. If q is a real number, prove that the matrices

A = , B = ,

are unitarily equivalent by finding a unitary U so that B = U* AU.

49. Find all matrices B unitarily equivalent to the matrix A in the previous problem.

50. Let V be a finite-dimensional inner product space and T a positive definite linear operator on V. Let [. , .] be another inner product on V defined by [u, v] = (Tu, v). Prove that U is unitary with respect to both the inner products iff U commutes with T.

51. Let u, w Î V, a finite-dimensional inner product space. If T denotes the linear operator on V defined by Tu,w(v) = (v, w)u, show that: (a) if u, w, g, h ¹ 0, Tu,w commutes with Tg,h iff the sets {u, g} and {w, h} are linearly independent; (b) Tu,w self adjoint iff one of u and w is a real scalar multiple of the other; (c) If u, w ¹ 0, Tu,w is diagonalizable iff trace (Tu,w) ¹ 0; (d) Tu,w is normal iff u and w are linearly dependent; (e) a linear operator T on V commutes with Tu,w iff for some scalar c, Tv = cv and T* v = c* v; and, (f) P Tu,wP = P uP P wP , where the norms are the ones induced by the inner product.

52. Let V be a finite dimensional inner product space and L(V) denote the vector space of linear operators on V. Prove that (T, S) = trace (S* T), T, S Î L(V), defines the unique inner product on L(V) for which P Tu,wP F = P uP P wP for all u, w Î V, where Tu,w(v) = (v, w)u, v Î V. (Frobenius norm).

53. Show that every self adjoint unitary operator on a finite-dimensional inner product space is a reflection in some subspace of V.

54 Let U Î L(V, W), where V and W are finite dimensional inner product spaces. Let U : V ® W be a linear operator. In each of the following cases find the minimal conditions on U so that: (a) the map T ® UTU* is an isomorphism of the algebra L(V,V) onto the algebra L(W,W); (b) trace (UTU* ) = trace (T), T Î L(V,V). (c) UTu,wU* = TUu,Uw , where Tu,w(v) = (v, w)u; (d) (UTU-1)* = UT* U-1, assuming that U is invertible; (e) the map T ® UTU* preserves inner products in L(V,V) and L(W,W ) given by (T, S) = tr (S* T).

55. For the following real symmetric matrices A, find orthogonal matrices P such that P¢AP is diagonal:

.

56. Is the square of every complex symmetric matrix positive definite? Is it self-adjoint? Is it normal?

57. Show that there exist only a finite number of orthogonal matrices P that diagonalize the matrix

A = .

Hence describe all the matrices C such that C-1AC is diagonal.

58. Show that the matrix A = is normal and find an orthonormal basis of C2 of its eigenvectors.

59. Give examples of 3´ 3 matrices A such that A3 is: (a) normal, (b) hermitian, and (c) positive define, without A being so.

60. If T is a normal operator on a finite-dimensional inner product space, show that T is, respectively, self-adjoint, positive definite, or, unitary iff all the eigenvalues of T are real, positive, or, of absolute value 1.

61. If T is a normal operator on a finite-dimensional inner product space, prove that it is both self-adjoint and unitary iff T = I. When is T both positive definite and unitary?

62. Use spectral theorem to show that a linear operator T on an i.p.s. V is normal iff T = H1+iH2, where H1, H2 are commuting self-adjoint operators.

63. Show that T is self-adjoint iff it can be written as a difference of two positive definite operators: T = P1-P2.

64. If T is self-adjoint show that there exist unique p.s.d. P1, P2 such that T = P1-P2 and P TP 2 = P P1P 2+P P2P 2.

65. If m = 1, 2, … , prove that a real symmetric matrix A has a real symmetric (2m+1)-th root B (i.e., A = B2m+1), which is also real symmetric and that it has a real symmetric 2m-th root iff it is postitive semi-definite.

66. Prove that every positive (semi-) definite matrix is the square of a positive (semi-) definite matrix.

67. Prove that a nilpotent operator is normal iff it is the zero operator.

68. Prove that an n´ n complex matrix A is normal iff A is diagonalizable and its eigenvectors vectors associated with distinct eigenvalues are orthogonal.

69. Is the sum of two commuting normal operators normal?

70. Is the product of two commuting normal operators normal?

71. Let ai Î Cn, i = 1, 2, ..., n, be linearly independent, P aiP = 1, Ri = I-2aiai* , A = R1R2 … Rn. Show that A-I is invertible. If H = (A+I )(A-I) , prove that ai* Hai = 0, 1 £ i £ n. [Hint: If Ax = x, x ¹ 0, R1-1 = R1* = R1 Þ R2R3…Rnx = R1x Þ x-2a1a1* x = x+S 2£ j£ n a jaj, a j Î C, say. By l.i. of aj 's, a1* x = 0. Þ R3…Rnx = R2x. Þ a2* x = 0. Continuing in this fashion, we get ai* x = 0, 1 £ i £ n. Þ x = 0. Hence A-I is invertible. Next, let 1 £ i £ n, and (A-In)x = ai. Put R = Ri+1 … Rn, and L = R1 … Ri-1. Þ Rx-2aiai* Rx = RiRx = L-1Ax = Ri-1…R1(ai+x), or, x+S i+1£ j£ n b jaj -2aiai* Rx = ai+x+S 1£ j£ i-1 b jaj , b j Î C. By l.i. of aj 's, Rx = x (b i 's are zero), -2ai* Rx = 1 and -2ai* x = 1. Þ ai* Hai = ai* (A+In)(A-In)-1ai = ai* (A+In)x = ai* (ai+2x) = (P aiP 2)2 + 2ai* x = 0. #]

72. If v1, v2, … , vn are orthonormal, prove that Tu = S 1£ i£ n (u, vi)vi, is the orthogonal projection of u on the subspace G = <v1, v2, … , vn> and, moreover, that Tu is the unique vector from G nearest to u.

 

THE KACZMARZ METHOD

Let C n and C m have some prescibed inner products. The Kaczmarz method is an iterative procedure for providing a solution of a consistent equation Ax = b deviating least from the starting initial vector. The method, nevertheless, converges even for an inconsistent system.

Orthogonal Projection on a Hyper Plane

Consider a hyper plane given by Rx = b , where R is a row vector and b is a scalar. The unit normal to the hyper plane is given by R* /P R* P and the orthogonal projection of a vector x on the hyper plane is obtained by removing the vector joining the foot of the perpendicular from x to its orthogonal projection on the hyperplane Rx = b . Here R* is the adjoint of the linear functional R, i.e., (Rx, y) = (x, R* y). Hence the projection has the form

z = x - q R* /P R* P ,

where q is obtained by the condition Rz = b , so that Rx - q RR* /P R* P = b , i.e., q = (Rx - b )/P R* P . Thus

z = x - (Rx - b )R* /P R* P 2 = [I - R* R/P R* P 2]x + b R* /P R* P 2 = [I - PR*]x + b R* /P R* P 2,

where PR* denotes the orthogonal projection on the span of R* .

 

The Kaczmarz method

The Kaczmarz method for Ax = b, where A is an m´ n complex matrix begins with an initial vector x(0) and generates a sequence of iterations x(k), k ³ 1 as follows. The vector x(k) is orthogonally projected on the i-th hyperplane Rix = bi, for 1 £ i £ m, to arrive at x(k+1), the expression for which is given by

,

where Qi = [I - PRi*], the orthogonal projection on the othogonal complement of the subspace spanned by Ri* . It is clear that

= Hx(k) + c,

say, where c Î R (A* ), A* being the adjoint of the matrix operator A : R n ® R m. Since, R (A* ) = N (A)^ ,

Qi = [I - PRi*] : R (A* ) ® R (A* ), 1 £ i £ m,

and

Qix = x, x Î N (A), 1 £ i £ m,

i.e., Qi is an identity on N (A). Hence Qix = Qi[PR (A*)x + PN (A)x] = Qi[PR (A*)x] + PN (A)x, x Î C n.

It follows that Hx = H[PR (A*)x] + PN (A)x. If P H[PR (A*)x]P = P PR (A*)xP , then writing z = PR (A*)x, Qiz = z, 1 £ i £ m, so that Riz = 0, 1 £ i £ m, i.e., Az = 0, or that z Î N (A). But z Î R (A* ), so that z = 0. Hence the restriction of H to R (A* ) has norm less than 1, so that

limk® ¥ Hkx(0) = limk® ¥ Hk PR (A*)x(0) + PN (A)x(0) = PN (A)x(0).

Moreover, h = limk® ¥ (I + H + H2 + … + Hk-1)c exists and belongs to R (A* ). Therefore the Kaczmarz method converges for arbitrary A, b and x(0) and

limk® ¥ x(k) = PN (A)x(0). + h, provides a description of the limiting vector provided by the method. Noting that h does not depend on x(0), and that both c and h belong to R (A* ), h = Hh + c = [I - H|R (A*)]-1c.

Now, let us consider the case when Ax = b is consistent and x is a solution of it. Then, x lies on the intersection of all the m-hyperplanes so that x = Hx + c. Taking x as x(0) we get x(k) = x , for all k, and so x = PN (A)x + h . Since h Î R (A* ) and PN (A)x Î N (A), it follows that h is the minimum norm solution of Ax = b, and it is the limiting vector obtained by the Kaczmarz method after choosing x(0) = 0. Finally, the solution of Ax = b nearest to a given vector z is obtained by minimizing P z - (y+h )P with respect to y Î N (A), it being the orthogonal projection of z on the solution space, and hence is given by PN (A)[z - h ] + h = PN (A)z + h , as h Î R (A*). Since the vector PN (A)z + h is the limiting vector provided by the Kaczmarz method using x(0) = z , it follows that the Kaczmarz method for a consistent system converges to the solution of the system nearest to the starting vector x(0). Summarizing, we have:

Theorem. The iterations {x(k)}k³ 0 in the Kaczmarz method for Ax = b converge to the limiting vector x(¥ ) = limk® ¥ x(k) = PN (A)x(0). + h , where h does not depend on x(0). If Ax = b is consistent h is the mininimum norm solution of it, and, moreover, x(¥ ) is the sloution of the system nearest to the initial vector x(0).

 

The Method of Residual Projections

The analysis in the sequel shows that the sequence {x(k)}k³ 0 generated by the residual projection method converges to a vector x minimizing the residual b-Ax in the inner product space Cm with (y, z) = z* y and the norm P yP = (y* y)1/2, note that y* here and in the sequel denotes the adjoint of y with respect to the inner product under consideration. Thus, if M deotes the positive definite matrix associated with an inner product on K m, z* y = (y, z) = (z¢ )-My, so that z* º (z¢ )-M. [Similarly for a linear functional Rx where R is a row vector, if N is the matrix associated with an inner product on K n, (x, R* ) = Rx = RN-1Nx = ((N-1(R')-)¢ )-Nx, so that R* = (N-1(R')-)].

Let A be an m´ n complex matrix with each column containing at least one nonzero entry, and b an m´ 1 vector. The j-th column vector of A is denoted by Cj. Let pj = (Cj* Cj)-1 and Dj = pjCj. With x(0) as the initial vector we put

r(0) = r(0,0) = b-Ax(0).

The residual projection algorithm generates a sequence {x(k)}k³ 0 of iterates in which the j-th component x(k+1)j of x(k+1) is computed through

d j(k) = Dj* r(k,j-1), xj(k+1) = xj(k) + d j(k), j = 1, 2, … , n,

and

r(k,j) = r(k,j-1) - d j(k)Cj, j = 1, 2, … , n,

where

r(k+1) = r(k+1,0) = r(k,n), k = 0, 1, 2, … .

Let Pj = pjCjCj* , j = 1, 2, … , n, denote the orthogonal projection on the subspace <Cj> generated by the column vector Cj and put Qj = I -Pj. Since

r(k,j) = r(k,j-1) - d j(k)Cj = (I-(Cj* Cj)-1CjCj* )r(k,j-1) = (I-Pj)r(k,j-1) = Qjr(k,j-1) = QjQj-1… Q1 r(k,0) = QjQj-1… Q1 r(k),

r(k+1) = QnQn-1 Q2Q1r(k) = Hr(k), k ³ 0,

where

H = QnQn-1 Q2Q1,

and Qj denotes the orthogonal projection on <Cj>^ , the orthogonal complement of <Cj>.

Note that

xj(k+1) = xj(k) + pjCj* r(k,j-1) = xj(k) + pjCj* Qj-1… Q1 r(k) = xj(k) + pjCj* Qj-1… Q1 [b - Ax(k)].

Hence

x(k+1) = [I - RA]x(k) + Rb = Sx(k) + Rb,

say, where R denotes the matrix with j-th row vector equalling pjCj* Qj-1… Q1. Since K m = R (A) Å ^ N (A* ), for x Î K m, using x = PAx + PN (A*)x = y + z, where y Î R (A) and z Î N (A* ), we have Qjz = (I - (Cj* Cj) -1CjCj* )z = z, and Qjy = (I - (Cj* Cj) -1 CjCj* )y Î R (A). Thus, R (A) and N (A* ) are invariant under each Qj which is an identity operator on N (A* ). It follows that R (A) and N (A* ) are invariant under H, and that H is an identity operator on N (A* ).

We show that the restriction H|R (A) of H to R (A) has norm less than 1. For this, since H = QnQn-1 … Q2Q1 and each Qj is an orthogonal projection P HP £ 1 and so P H|R (A)P £ 1.

Let y Î R (A) have P HyP = P yP . Then, Qjy = y, 1£ j £ n, so that Cj* y = 0 for each j, implying that A* y = 0, i.e., y Î N (A* ). Hence y = 0. Hence for every 0 ¹ y Î R (A), P HyP < P yP , so that sup0¹ yÎ R (A) P HyP /P yP being attained on the surface of the unit sphere in R (A), we have proved that

r º P H|R(A)P < 1.

It follows that

r(k) = Hkr(0) = (H|R (A))kPR (A)r(0) + PN (A*)r(0) ® PN (A*)r(0) = PN (A*)b, k ® ¥ .

Note that P r(k,j)P = P Qjr(k,j-1)P £ P r(k,j-1)P , 1 £ j £ n, so that P r(k+1)P = P r(k,n)P £ P r(k,0)P = P r(k)P , so that the norms of the residuals constitute a monotone non-increasing sequence. Also,

d (k)j Cj = [Dj* r(k,j-1)]Cj = [pjCj* r(k,j-1)]Cj = [pjCjCj* r(k,j-1)] = Pj r(k,j-1) = PjQj-1 r(k,j-2) = …

= PjQj-1 Qj-2 … Q1 r(k,0) = PjQj-1 Qj-2 … Q1 r(k ),

so that

d (k)jCj = PjQj-1Qj-2 … Q1 r(k) = PjQj-1Qj-2 … Q1 [(H|R (A))kPR (A)r(0) + PN (A*)r(0)] = PjPN (A*)r(0) + PjQj-1Qj-2 … Q1 [(H|R (A))kPR (A)r(0)] = PjQj-1Qj-2 … Q1 [(H|R (A))kPR (A)r(0)].

It follows that P d (k)jCjP £ s kP r(0)P . Hence P d (k)jP £ r kP r(0)P /P CjP , so that

x(k)j = x(k-1)j + d (k-1)j = x(0)j + d (0)j + d (1)j + … + d (k-1)j ® x(¥ )j = x(0)j + S k³ 0 d (k)j, 1 £ j £ n,

the last series converging, it being dominated by the geometric series S k³ 0 r kP r(0)P /P CjP , in norm. Hence the method of residual projections converges. Since, PAb - Ax(¥ ) = PA[b - Ax(¥ )] = PA[PN (A*)b] = 0, i.e., Ax(¥ ) = PAb, it follows that x(¥ ) is a least squares solution of Ax = b.