Chapter 12

STABILITY OF A MATRIX

 

An matrix A Î C n´ n is called stable if the initial value problem (IVP): dx/dt = Ax, x(0) = x0, has a solution x(t) ® 0, as t ® ¥ , for any choice of the initial vector x0, whatsoever. Note that the solution of the IVP is given by x(t) = etA x0, where eAt º etA = I+tA+(t2/2!)A2+...+(tn/n!)An+... . Hence, a necessary and sufficient condition for the stability is eAt ® 0, as t ® ¥ . This is equivalent to r (eA) < 1, where r (B) denotes the spectral radius of the matrix B and is the largest magnitude of the eigenvalues of B. The condition is equivalent to eRl (l ) < 1 Û Rl (l ) < 0, for each eigenvalue l = l (A) of A. Hence A is stable iff all the eigenvalues of A lie in the open left half plane {z : Rl z < 0} or using the notion of inertia that In (A) = (0,n,0).

 

PROBLEMS

1. Which of the following matrices are stable:

.

2. Show that A is stable iff A* is stable iff A¢ is stable.

3. Is it true that there exist matrices A such that Ak is stable for all even positive integers k?

4. Is it true that there exist matrices A such that Ak is stable for all odd positive integers k?

5. Prove that an n´ n complex matrix A is stable if it satisfies Rl aii > S j¹ i |aij | > 0, 1 £ i £ n.

6. Prove that a real 2´ 2 symmetric matrix

is stable iff a < 0 and ac > b2.

7. Can a partitioned matrix of type B = be stable?

8. If In A = (p ,n ,d ), show that for the matrix

B = ,

In B = (p ,n ,p +n +2d ). Deduce that B-cI is stable for all complex c with Re c > 0 iff A is stable.

 

Lyapunov’s Stability Theorem

As such the notion of stability of a matrix involves a knowledge of the eigenvalues of the matrix. We know that a determination of the eigenvalues of a general matrix is a transcendental process in the sense that it involves an infinity of arithmetic operations. Thus, first determining the eigenvalues (even if by an approximate numerical method, and then checking if their real parts are all negative is not the right way of establishing the stability of a matrix. In this context, the celebrated Lyapunov’s stability theorem enables a theoretical determination of the stability of a matrix by using only a finite number of arithmetic operations in solving a linear system and then determining if a certain matrix is positive definite or not.

Lyapunov’s Stability Theorem. If A is stable and C is a positive definite matrix there exists an X p.d. such that AX+XA* = -C. Conversely, if X, C are p.d. and the above equation is satisfied, then A is stable.

Proof: If the equation is satisfied with X, C p.d. let (l , y) be an e.p. (eigen pair) of A*, i.e., y ¹ 0 and Ay = l y. Pre-and post-multiplying the equation by y* and y, respectively, we have –y*Cy = (A*y)*Xy+y*XA*y = (l *+l )y*Xy. Since y*Cy, y*Xy > 0 Þ l *+l < 0, i.e., Rl l < 0, showing that A is stable.

Conversely, let A be stable. Then s (A)Ç s (-A*) = f and so the system AX+XA* = -C has a unique solution X for every C. If C is positive definite, *-ing the equation (i.e., taking the adjoint or the conjugate transpose on both sides) we have -C = X*A*+AX* = AX*+X*A . It follows from the uniqueness of the solution that X = X*, i.e., X is hermitian.

Since A is stable so is A* and so both eAt and eA* t ® 0, as t ® ¥ . Hence eAtC eA* t ® 0, t ® ¥ . But (d/dt)( eAtC eA* t) = AeAtC eA*t+eAtC eA* t A*, i.e., (d/dt)Z(t) = AZ(t)+Z(t)A*, where Z(t) = eAtCeA*t.Integrating w.r.t. t from 0 to ¥ , we therefore have

-C = Z(¥ ) – Z(0) = = AX+XA*, X = .

Note that the existence of the integral is guaranteed by the fact that A is stable. If s = max{Rl l : l an e.v. of A} < 0, r (eA ) = es . Consequently, there is a matrix norm ||.||1 s.t. ||eA||1 < es +d < 1, for a d sufficiently small so that s +d < 0. Similarly, ||eA*||2 < es +d < 1. Hence ||eAtC eA* t|| £ ||eAt|| ||C|| ||eA* t|| £ C1 ||eAt ||1 ||C||C2||eA* t||2. But for any matrix norm ||× ||, ||eBt || £ ||eB[t]|| ||eB(t-[t]) || £ ||eB||[t] e||B||(t-[t]) £ ||eB||t e||B||/ ||eB||, if ||eB|| < 1. Put e-a = max {||eA||1, ||eA*||2} < 1, i.e., a > 0. Then ||eAtC eA*t|| £ C1C2CQe-2a t, where

Choosing the matrix norm ||× || to be ||X|| = n´ max{|Xij| : 1 £ i, j £ n}, the integral converges by the Lebesgue’s dominated convergence theorem.

Finally, to show that X is p.d., let z ¹ 0. Then z*Xz = > 0, the integrand being positive for all t > 0, as C is positive definite. Hence, z*Xz > 0 Þ X is p.d., completing the proof of the Lyapunov’s stability theorem. #

 

PROBLEMS

1. Is it true that AX+XA* is necessarily negative define if A is stable and X is positive definite? If so give a proof? If not, produce a counter example.

2. The left hand side of AX+XA* = -C, is a linear vector function in the n2-variables xij‘s. Produce a matrix B and a vectror c such that Bx = c, where x = vec (X) is defined by xi+(j-1)n = (X)ij for all 1 £ i, j £ n.

3. Check the stability of the following matrices:

.

4. Let the hermitian part H of a complex matrix A be defined by H = (A+A*)/2 and the skew hermitian part S by S = (A-A*)/2. If the hermitian part H of A is negative definite prove that A is stable. Is the converse also true?

 

Equivalent Formulations of Sylvester’s Inertia Theorem

Recall that the inertia of a matrix A is defined as the triple In(A) = (p (A), n (A), d (A)), where the entries respectively denote the number of eigenvalues with positive, negative and zero real parts. Thus an n´ n matrix A being stable is equivalent to In(A) = (0, n, 0). The following equivalents of Sylvester’s inertia theorem are sometimes useful in determining stability of certain matrices:

Theorem. Let Hn denote the class of all n´ n hermitian matrices and B > 0 mean that B is p.d. The following statements are equivalent: (a) If H Î Hn and S is non-singular, In(S*HS) = In(H) (Sylvester’s law of inertia). (b) If P > 0 and H Î Hn, In(PH) = In(H). (c) If AH > 0 and H Î Hn Þ In A = In H.

Proof: Note that In (S*HS) = In (S-1((SS*)H)S) = In ((SS*)H) and that In ((AH)-1H) = In (H-1A-1H) = In (A-1) = In A. Hence, taking P = SS*, (a) Ü (b). Taking S to be such that SS* = P (e.g., LL*-decomposition, or S = Ö P), (b) Ü (a). Hence (a) Û (b). Taking P = (AH)-1, (b) Þ (c). Finally, assume (c). If H is non-singular, choosing A = P-1H-1, so that P = (AH)-1, In (PH) = In (HP) = In (HP)-1 = In A = In H. Thus (c) Þ (b) for H non-singular, which by the above proof of (b) Þ (a) shows that (a) holds for non-singular H. Hence (c) Þ (a) in the non-singular case and to complete the proof it is sufficient to show that (a) in the non-singular case implies (a) in the singular case. This we do below:

If H is singular, let H(d ) denote the matrix obtained from the unitary diagonalization of H by replacing the zero diagonal entries by d . Let d > 0 and sufficiently small. Then H(d ) and S*H(d )S have the same number of negative eigen values as H. Passing to the limit as d ® 0, we find that H and S*HS have the same number of negative eigen values (since a positive eigen value could at best tend to zero). Replacing H by -H it follows that they have the same number of positive eigen values and therefore also the same naumber of zero eigen values. This completes the proof. #

 

PROBLEMS

1. Prove that for a complex matrix A there holds A = PH, where P is positive definite and H is hermitian iff pA(x) = P 1£ j£ k (x-cj), where cj‘s are real and distinct.

2. Is it true that a hermitian H is stable iff -H is p.d.? [Give two reasonings, one of which uses Lyapunov’s theorem].

3. If H is hermitian an non-singular, H(-H)+(-H)H = -2H2 and 2H2 is positive definite. What do you conclude?

 

Some Equivlaent Criteria of Stability

The following list states several equivalent criteria for the stability of a matrix:

Theorem. For an n´ n complex matrix A, the following statements are equivalent:

(1) A is stable (i.e., s (A) Ì {z: Rl z < 0}.

(2) All solutions of the system dx/dt = Ax approach zero as t ® ¥ .

(3) limt® ¥ eAt = 0.

(4) For a > 0, A-a I is non-singular and the spectral radius of B = (A-a I)-1(A+a I) is less than unity.

(5) For a > 0, A-a I is non-singular and H = (A-a I)-1(A+a I) is convergent (i.e., Hk ® 0 as k ® ¥ .

(6) For a > 0, A-a I is non-singular and the sequence {xk} defined by (A-a I)xk+1 = -(A+a I)xk+2b, converges to the solution of the linear system Ax = b, for any initial guess x0.

(7) For C n.d. there exists H p.d. such that AH+HA* = C.

(8) There exists H p.d. such that AH+HA* is n.d.

(9) There exists H p.d. such that Rl (x*AHx) < 0, for all non-zero x Î C n.

(10) There exists H p.d. such that the numerical range of AH is contained in the open left half plane. [The numerical range, or, the field of values of A is defined as: W(A) = {x*Ax : x*x = 1}].

(11) A is similar to a matrix which can be expressed as the sum of a skew-Hermitian matrix and a negative definite hermitian matrix.

(12) There exist H p.d., Q n.d and a skew-hermitian S such that A = (S+Q)H.

(13) For C n.d., the matrix equation A*H+HA = C has a solution and 0 ¹ AY+YA* p.s.d. implies tr (Y) < 0.

(14) For a > 0, A-a I is non-singular and for every negative definite C there exists H p.d. such that BHB*-H = C, where B = (A-a I)-1(A+a I).

(15) For a > 0, A-a I is non-singular and there exists H p.d. such that BHB*-H is n.d., where B = (A-a I)-1 (A+a I).

Proof: The implications (1) Û (3) Û (2) have already been proved above. Next, A-a I is non-singular and |l ((A-a I)-1 (A+a I))|2 = |(l (A)+a )/(l (A)-a )|2 = {(Rl l (A)+a )2+(Im l (A))2}/{(Rl l (A)-a )2+(Im l (A))2} < 1, iff Rl l (A) < 0. Hence (1) Û (4) Û (5) Û (6).(1) Þ (7) has been proved before and clearly (7) Þ (8). Also (8) Þ (1) is proved before. Thus (1) Þ (7) Þ (8) Þ (1).

With H as in (8), if x Î C n is non-zero then 0 > x*(AH+HA*)x = 2 Rl (x*AHx). Hence (8) Þ (9). Converlely, if there is H p.d. such that Rl (x*AHx) < 0 for x Î C n not zero, it follows that AH+HA is n.d. Hence (9) Þ (8). Consequently (9) Û (8).

Noting that W(AH) = {x*AHx : x*x = 1}, W(AH) Ì {z : Rl z < 0} Û Rl x*AHx < 0, x ¹ 0. Hence (9) Û (10).

To prove that (11) Þ (1), we have to show that (via similarity) that the sum of a skew-hermitian K and a negative definite Q is stable. For this, if x ¹ 0, (x*Kx)* = x*K*x = - x*Kx Þ Rl x*Kx = 0. Hence, as I is p.d., for every x ¹ 0, Rl (x*(K+Q)Ix) = Rl (x*Qx) = x*Qx < 0. By (9) Þ (1), K+Q is stable. Thus (11) Þ (1).

Conversely, if (1) holds and Em = (1/m)Jm , m = 1, 2, ... , where Jm is the Jordan canonical form of mA, Em is simialr to A. If m is suficiently large, -(1/2)(Em+Em*) = Q is a tridiagonal matrix which is real, symmetric, strictly diagonally dominant with positive diagonal entries (negatives of real parts of e.v.’s of A) and hence is p.d. (|-l (Q)+Rl l (A)| < 2/m Þ -Rl l (A) -2/m £ -l (Q) £ -Rl l (A)+2/m Þ l (Q) > 0 : Gershgorin circle theorem). Hence Em = (Em+Em* )/2+(Em-Em*)/2 = Q+K, where Q is n.d. and K is skew-hermitian. Hence by the similarity of A and E we have (1) Þ (11).

Thus (1) Û (11).

Assuming (12), i.e., A = (S+Q1) H, H p.d., Q1 n.d. and S skew-hermitian, we have A = (Ö H)-1(Ö H(S+Q1)Ö H)Ö H = (Ö H)-1(K+Q)Ö H, where K is skew-hermitian and Q is n.d. Thus (12) Þ (11).

Conversely, if A = C-1(K+Q)C, K skew hermitian, Q n.d., we have A = C-1(K+Q)(C-1)*(C*C) = (S+Q1)H, where S = C-1K(C-1)* skew-hermitian, Q1 = C-1Q(C-1)* n.d. and H = C*C p.d. Hence (11) Þ (12) and so (11) Û (12).

To prove (1) Þ (13), by (1) Þ (7), for n.d. C, A* being stable as A is, A*H+HA = C has a p.d. solution H. Also, if 0 ¹ AY+YA* = G is p.s.d., since Y is unique (s (A)Ç s (-A ) = f ), for each xÎ C n,

Consequently Y is n.s.d. and as Y ¹ 0, tr Y < 0. Hence (1) Þ (13).

Conversely, let (13) hold. Let (l , y) be an e.p. of A. Let Y1 = (y | y | ... | y). Then with Y = Y1Y1*, if Rl l > 0, AY+YA* = A Y1Y1*+Y1(AY1)* = (l +l *)Y1Y1* ¹ 0 is p.s.d. but tr Y = tr (Y1Y1*) > 0, #, a contradiction. Hence Rl l < 0. Then, since for n.d. C, A*H+HA = C has a solution H and as also A*H*+H*A = C, w.l.g., H can be assumed to be hermitian (else take (H+H*)/2 as solution). Then 0 > y*Cy = y*A*Hy+y*HAy = (l *+l )y*Hy Þ l *+l = 2 Rl l ¹ 0. Hence Rl l < 0 for every e.v. l of A. Consequently A is stable. Thus (13) Þ (1). Hence (1) Û (13).

Next, assume (4). Then r (B) < 1. Let C = -E be n.d. If X-BXB* = 0, we have X = BmX(Bm)* ® 0, m ® ¥ . Hence X = 0, and so the homogeneous system X-BXB = 0, has only trivial solution. Consequently, the system H - BHB = E has a solution H (unique all the more) for every E and so for the present p.d. E. Define Hm = E + BHm-1B , m = 1, 2, ... . Then, for an arbitrary H0 , since r (B) < 1, Hm-H = B(Hm-1 -H)B* = Bm(H0 - H)(B*) ® 0, m ® ¥ . Hence H = limm® ¥ Hm. With H0 = I, Hm = E + BEB* + ... + Bm-1 E(Bm-1)*+Bm(Bm)* and so x*Hmx ³ x*Ex > 0, for all x ¹ 0. Taking limit as m ® ¥ , x*Hx ³ x*Ex > 0, for all x ¹ 0. Hence H is p.d. Thus (4) Þ (14). Taking H as in (14) we have (14) Þ (15).

If (15) holds, H is p.d. and H –BHB* is also p.d. Let (l , u) be an e.p. of B*. Then 0 < u*Hu –u*BHB*u = (1 -|l |2)u*Hu. Hence 1 -|l |2 > 0, i.e., |l | < 1. Hence r (B) = r (B*) < 1. Thus (15) Þ (4). Hence (4) Þ (14) Þ (15) Þ (4).

Summarizing, we have proved the implications:

 from which the equivalence of (1)-(16) follows. #