Images

Canonical Forms

10.1 Introduction

Let T be a linear operator on a vector space of finite dimension. As seen in Chapter 6, T may not have a diagonal matrix representation. However, it is still possible to “simplify” the matrix representation of T in a number of ways. This is the main topic of this chapter. In particular, we obtain the primary decomposition theorem, and the triangular, Jordan, and rational canonical forms.

We comment that the triangular and Jordan canonical forms exist for T if and only if the characteristic polynomial Δ(t) of T has all its roots in the base field K. This is always true if K is the complex field C but may not be true if K is the real field R.

We also introduce the idea of a quotient space. This is a very powerful tool, and it will be used in the proof of the existence of the triangular and rational canonical forms.

10.2 Triangular Form

Let T be a linear operator on an n-dimensional vector space V. Suppose T can be represented by the triangular matrix

Images

Then the characteristic polynomial Δ(t) of T is a product of linear factors; that is,

Images

The converse is also true and is an important theorem (proved in Problem 10.28).

THEOREM 10.1:   Let T:VV be a linear operator whose characteristic polynomial factors into linear polynomials. Then there exists a basis of V in which T is represented by a triangular matrix.

THEOREM 10.1:   (Alternative Form) Let A be a square matrix whose characteristic polynomial factors into linear polynomials. Then A is similar to a triangular matrix—that is, there exists an invertible matrix P such that P−1AP is triangular.

We say that an operator T can be brought into triangular form if it can be represented by a triangular matrix. Note that in this case, the eigenvalues of T are precisely those entries appearing on the main diagonal. We give an application of this remark.

EXAMPLE 10.1   Let A be a square matrix over the complex field C. Suppose λ is an eigenvalue of A2. Show that Images or Images is an eigenvalue of A.

By Theorem 10.1, A and A2 are similar, respectively, to triangular matrices of the form

Images

Because similar matrices have the same eigenvalues, Images for some i. Hence, Images is an eigenvalue of A.

10.3 Invariance

Let T:VV be linear. A subspace W of V is said to be invariant under T or T-invariant if T maps W into itself—that is, if υW implies T(υ) ∈ W. In this case, T restricted to W defines a linear operator on W; that is, T induces a linear operator Images defined by Images for every wW.

EXAMPLE 10.2

(a) Let T: R3R3 be the following linear operator, which rotates each vector υ about the z-axis by an angle θ (shown in Fig. 10-1):

Images

Images

Figure 10-1

Observe that each vector w = (a, b, 0) in the xy-plane W remains in W under the mapping T; hence, W is T-invariant. Observe also that the z-axis U is invariant under T. Furthermore, the restriction of T to W rotates each vector about the origin O, and the restriction of T to U is the identity mapping of U.

(b) Nonzero eigenvectors of a linear operator T:VV may be characterized as generators of T-invariant one-dimensional subspaces. Suppose T(υ) = λυ, υ ≠ 0. Then W = {kυ, kK}, the one-dimensional subspace generated by υ, is invariant under T because

Images

Conversely, suppose dim U = 1 and u ≠ 0 spans U, and U is invariant under T. Then T(u) ∈ U and so T(u) is a multiple of u—that is, T(u) = μu. Hence, u is an eigenvector of T.

The next theorem (proved in Problem 10.3) gives us an important class of invariant subspaces.

THEOREM 10.2:   Let T:VV be any linear operator, and let f(t) be any polynomial. Then the kernel of f(t) is invariant under T.

The notion of invariance is related to matrix representations (Problem 10.5) as follows.

THEOREM 10.3:   Suppose W isan invariant subspace of T:VV.Then T hasa blockmatrix representation Images, where A is a matrix representation of the restriction Images of T to W.

10.4 Invariant Direct-Sum Decompositions

A vector space V is termed the direct sum of subspaces W1, ..., Wr, written

Images

if every vector υV can be written uniquely in the form

Images

The following theorem (proved in Problem 10.7) holds.

THEOREM 10.4:   Suppose W1, W2, ..., Wr are subspaces of V, and suppose

Images

are bases of W1, W2, ..., Wr, respectively. Then V is the direct sum of the Wi if and only if the union B = B1 ∪ ... ∪ Br is a basis of V.

Now suppose T:VV is linear and V is the direct sum of (nonzero) T-invariant subspaces W1, W2, ..., Wr; that is,

Images

Let Ti denote the restriction of T to Wi. Then T is said to be decomposable into the operators Ti or T is said to be the direct sum of the Ti, written T = T1 ⊕ ... ⊕ Tr. Also, the subspaces W1, ..., Wr are said to reduce T or to form a T-invariant direct-sum decomposition of V.

Consider the special case where two subspaces U and W reduce an operator T:VV; say dim U = 2 and dim W = 3, and suppose {u1, u2} and {w1, w2, w3} are bases of U and W, respectively. If T1 and T2 denote the restrictions of T to U and W, respectively, then

Images

Accordingly, the following matrices A, B, M are the matrix representations of T1, T2, T, respectively,

Images

The block diagonal matrix M results from the fact that {u1, u2, w1, w2, w3} is a basis of V (Theorem 10.4), and that T(ui) = T1(ui) and T(wj) = T2(wj).

A generalization of the above argument gives us the following theorem.

THEOREM 10.5:   Suppose T:VV is linear and suppose V is the direct sum of T-invariant subspaces, say, W1, ..., Wr. If Ai is a matrix representation of the restriction of T to Wi, then T can be represented by the block diagonal matrix:

Images

10.5 Primary Decomposition

The following theorem shows that any operator T:VV is decomposable into operators whose minimum polynomials are powers of irreducible polynomials. This is the first step in obtaining a canonical form for T.

THEOREM 10.6:   (Primary Decomposition Theorem) Let T:VV be a linear operator with minimal polynomial

Images

where the fi(t) are distinct monic irreducible polynomials. Then V is the direct sum of T-invariant subspaces W1, ..., Wr, where Wi is the kernel of fi(T)ni. Moreover, fi(t)ni is the minimal polynomial of the restriction of T to Wi.

The above polynomials fi(t)ni are relatively prime. Therefore, the above fundamental theorem follows (Problem 10.11) from the next two theorems (proved in Problems 10.9 and 10.10, respectively).

THEOREM 10.7:   Suppose T:VV is linear, and suppose f(t) = g(t)h(t) are polynomials such that f(t) = 0 and g(t) and h(t) are relatively prime. Then V is the direct sum of the T-invariant subspace U and W, where U = Ker g(T) and W = Ker h(T).

THEOREM 10.8:   In Theorem 10.7, if f(t) is the minimal polynomial of T [and g(t) and h(t) are monic], then g(t) and h(t) are the minimal polynomials of the restrictions of T to U and W, respectively.

We will also use the primary decomposition theorem to prove the following useful characterization of diagonalizable operators (see Problem 10.12 for the proof).

THEOREM 10.9:   A linear operator T:VV is diagonalizable if and only if its minimal polynomial m(t) is a product of distinct linear polynomials.

THEOREM 10.9:   (Alternative Form) A matrix A is similar to a diagonal matrix if and only if its minimal polynomial is a product of distinct linear polynomials.

EXAMPLE 10.3   Suppose AI is a square matrix for which A3 = I. Determine whether or not A is similar to a diagonal matrix if A is a matrix over: (i) the real field R, (ii) the complex field C.

Because A3 = I, A is a zero of the polynomial f(t) = t3 − 1 = (t − 1)(t2 + t + 1). The minimal polynomial m(t) of A cannot be t − 1, because AI. Hence,

Images

Because neither polynomial is a product of linear polynomials over R, A is not diagonalizable over R. On the other hand, each of the polynomials is a product of distinct linear polynomials over C. Hence, A is diagonalizable over C.

10.6 Nilpotent Operators

A linear operator T:VV is termed nilpotent if Tn = 0 for some positive integer n; we call k the index of nilpotency of T if Tk = 0 but Tk − 10. Analogously, a square matrix A is termed nilpotent if An = 0 for some positive integer n, and of index k if Ak = 0 but Ak − 1 ≠ 0. Clearly the minimum polynomial of a nilpotent operator (matrix) of index k is m(t) = tk; hence, 0 is its only eigenvalue.

EXAMPLE 10.4   The following two r-square matrices will be used throughout the chapter:

Images

The first matrix N, called a Jordan nilpotent block, consists of 1’s above the diagonal (called the super-diagonal), and 0’s elsewhere. It is a nilpotent matrix of index r. (The matrix N of order 1 is just the 1 × 1 zero matrix [0].)

The second matrix J(λ), called a Jordan block belonging to the eigenvalue λ, consists of λ’s on the diagonal, 1’s on the superdiagonal, and 0’s elsewhere. Observe that

Images

In fact, we will prove that any linear operator T can be decomposed into operators, each of which is the sum of a scalar operator and a nilpotent operator.

The following (proved in Problem 10.16) is a fundamental result on nilpotent operators.

THEOREM 10.10:   Let T:VV be a nilpotent operator of index k. Then T has a block diagonal matrix representation in which each diagonal entry is a Jordan nilpotent block N. There is at least one N of order k, and all other N are of orders ≤k. The number of N of each possible order is uniquely determined by T. The total number of N of all orders is equal to the nullity of T.

The proof of Theorem 10.10 shows that the number of N of order i is equal to 2mimi + 1mi − 1, where mi is the nullity of Ti.

10.7 Jordan Canonical Form

An operator T can be put into Jordan canonical form if its characteristic and minimal polynomials factor into linear polynomials. This is always true if K is the complex field C. In any case, we can always extend the base field K to a field in which the characteristic and minimal polynomials do factor into linear factors; thus, in a broad sense, every operator has a Jordan canonical form. Analogously, every matrix is similar to a matrix in Jordan canonical form.

The following theorem (proved in Problem 10.18) describes the Jordan canonical form J of a linear operator T.

THEOREM 10.11:   Let T:VV be a linear operator whose characteristic and minimal polynomials are, respectively,

Images

where the λi are distinct scalars. Then T has a block diagonal matrix representation J in which each diagonal entry is a Jordan block Jij = J(λi). For each λij, the corresponding Jij have the following properties:

(i) There is at least one Jij of order mi; all other Jij are of order ≤mi.

(ii) The sum of the orders of the Jij is ni.

(iii) The number of Jij equals the geometric multiplicity of λi.

(iv) The number of Jij of each possible order is uniquely determined by T.

EXAMPLE 10.5   Suppose the characteristic and minimal polynomials of an operator T are, respectively,

Images

Then the Jordan canonical form of T is one of the following block diagonal matrices:

Images

The first matrix occurs if T has two independent eigenvectors belonging to the eigenvalue 2; and the second matrix occurs if T has three independent eigenvectors belonging to the eigenvalue 2.

10.8 Cyclic Subspaces

Let T be a linear operator on a vector space V of finite dimension over K. Suppose υV and υ ≠ 0. The set of all vectors of the form f (T)(υ), where f(t) ranges over all polynomials over K, is a T-invariant subspace of V called the T-cyclic subspace of V generated by υ ; we denote it by Z(υ, T) and denote the restriction of T to Z(υ, T) by Tυ. By Problem 10.56, we could equivalently define Z(υ, T) as the intersection of all T-invariant subspaces of V containing υ.

Now consider the sequence

Images

of powers of T acting on υ. Let k be the least integer such that Tk(υ) is a linear combination of those vectors that precede it in the sequence, say,

Images

Then

Images

is the unique monic polynomial of lowest degree for which mυ(T)(υ) = 0. We call mυ(t) the T-annihilator of υ and Z(υ, T).

The following theorem (proved in Problem 10.29) holds.

THEOREM 10.12:   Let Z(υ, T), Tυ, mυ(t) be defined as above. Then

(i) The set {υ, T(υ), ..., Tk − 1(υ)} is a basis of Z(υ, T); hence, dim Z(υ, T) = k.

(ii) The minimal polynomial of Tυ is mυ(t).

(iii) The matrix representation of Tυ in the above basis is just the companion matrix C(mυ) of mυ(t); that is,

Images

10.9 Rational Canonical Form

In this section, we present the rational canonical form for a linear operator T:VV. We emphasize that this form exists even when the minimal polynomial cannot be factored into linear polynomials. (Recall that this is not the case for the Jordan canonical form.)

LEMMA 10.13:   Let T:VV be a linear operator whose minimal polynomial is f(t)n, where f(t) is a monic irreducible polynomial. Then V is the direct sum

Images

of T-cyclic subspaces Z(υi, T) with corresponding T-annihilators

Images

Any other decomposition of V into T-cyclic subspaces has the same number of components and the same set of T-annihilators.

We emphasize that the above lemma (proved in Problem 10.31) does not say that the vectors υ i or other T-cyclic subspaces Z(υi, T) are uniquely determined by T, but it does say that the set of T-annihilators is uniquely determined by T. Thus, T has a unique block diagonal matrix representation:

Images

where the Ci are companion matrices. In fact, the Ci are the companion matrices of the polynomials f(t)ni.

Using the Primary Decomposition Theorem and Lemma 10.13, we obtain the following result.

THEOREM 10.14:   Let T:VV be a linear operator with minimal polynomial

Images

where the fi(t) are distinct monic irreducible polynomials. Then T has a unique block diagonal matrix representation:

Images

where the Cij are companion matrices. In particular, the Cij are the companion matrices of the polynomials fi(t)nij, where

Images

The above matrix representation of T is called its rational canonical form. The polynomials fi(t)nij are called the elementary divisors of T.

EXAMPLE 10.6   Let V be a vector space of dimension 8 over the rational field Q, and let T be a linear operator on V whose minimal polynomial is

Images

Thus, because dim V = 8, the characteristic polynomial Δ(t) = f1(t) f2(t)4. Also, the rational canonical form M of T must have one block the companion matrix of f1(t) and one block the companion matrix of f2(t)2. There are two possibilities:

(a) diag[C(t4 − 4t3 + 6t2 − 4t − 7), C((t − 3)2), C((t − 3)2)]

(b) diag[C(t4 − 4t3 + 6t2 − 4t − 7), C((t − 3)2), C(t − 3), C(t − 3)]

That is,

Images

10.10 Quotient Spaces

Let V be a vector space over a field K and let W be a subspace of V. If υ is any vector in V, we write υ + W for the set of sums υ + w with wW; that is,

Images

These sets are called the cosets of W in V. We show (Problem 10.22) that these cosets partition V into mutually disjoint subsets.

EXAMPLE 10.7   Let W be the subspace of R2 defined by

Images

Figure 10-2

Images

that is, W is the line given by the equation xy = 0. We can view υ + W as a translation of the line obtained by adding the vector υ to each point in W. As shown in Fig. 10-2, the coset υ + W is also a line, and it is parallel to W. Thus, the cosets of W in R2 are precisely all the lines parallel to W.

In the following theorem, we use the cosets of a subspace W of a vector space V to define a new vector space; it is called the quotient space of V by W and is denoted by V/W.

THEOREM 10.15:   Let W be a subspace of a vector space over a field K. Then the cosets of W in V form a vector space over K with the following operations of addition and scalar multiplication:

Images

We note that, in the proof of Theorem 10.15 (Problem 10.24), it is first necessary to show that the operations are well defined; that is, whenever u + W = u′ + W and υ + W = υ′ + W, then

Images

In the case of an invariant subspace, we have the following useful result (proved in Problem 10.27).

THEOREM 10.16:   Suppose W is a subspace invariant under a linear operator T:VV. Then T induces a linear operator Images on V/W defined by Images. Moreover, if T is a zero of any polynomial, then so is Images. Thus, the minimal polynomial of Images divides the minimal polynomial of T.

SOLVED PROBLEMS

Invariant Subspaces

10.1.   Suppose T:VV is linear. Show that each of the following is invariant under T:

(a) {0}, (b) V, (c) kernel of T, (d) image of T.

(a) We have T(0) = 0 ∈ {0}; hence, {0} is invariant under T.

(b) For every υV, T(υ) ∈ V; hence, V is invariant under T.

(c) Let u ∈ Ker T. Then T(u) = 0 ∈ Ker T because the kernel of T is a subspace of V. Thus, Ker T is invariant under T.

(d) Because T(υ) ∈ Im T for every υV, it is certainly true when υ ∈ Im T. Hence, the image of T is invariant under T.

10.2.   Suppose {Wi} is a collection of T-invariant subspaces of a vector space V. Show that the intersection W = ∩i Wi is also T-invariant.

Suppose υW; then υWi for every i. Because Wi is T-invariant, T(υ) ∈ Wi for every i. Thus, T(υ) ∈ W and so W is T-invariant.

10.3.   Prove Theorem 10.2: Let T:VV be linear. For any polynomial f(t), the kernel of f(T) is invariant under T.

Suppose υ ∈ Ker f(t)—that is, f(T)(υ) = 0. We need to show that T(υ) also belongs to the kernel of f(T)—that is, f(T)(T(υ)) = (f(T) ∘ T)(υ) = 0. Because f(t)t = tf(t), we have f(T) ∘ T = Tf(t). Thus, as required,

Images

10.4.   Find all invariant subspaces of Images viewed as an operator on R2.

By Problem 10.1, R2 and {0} are invariant under A. Now if A has any other invariant subspace, it must be one-dimensional. However, the characteristic polynomial of A is

Images

Hence, A has no eigenvalues (in R) and so A has no eigenvectors. But the one-dimensional invariant subspaces correspond to the eigenvectors; thus, R2 and {0} are the only subspaces invariant under A.

10.5.   Prove Theorem 10.3: Suppose W is T-invariant. Then T has a triangular block representation Images, where A is the matrix representation of the restriction Images of T to W.

We choose a basis {w1, ..., wr} of W and extend it to a basis {w1, ..., wr, υ1, ..., υs} of V. We have

Images

But the matrix of T in this basis is the transpose of the matrix of coefficients in the above system of equations (Section 6.2). Therefore, it has the form Images, where A is the transpose of the matrix of coefficients for the obvious subsystem. By the same argument, A is the matrix of Images relative to the basis {wi} of W.

10.6.   Let Images denote the restriction of an operator T to an invariant subspace W. Prove

(a) For any polynomial f(t), f(Images)(w) = f (T)(w).

(b) The minimal polynomial of Images divides the minimal polynomial of T.

(a) If f(t) = 0 or if f(t) is a constant (i.e., of degree 1), then the result clearly holds. Assume deg f = n > 1 and that the result holds for polynomials of degree less than n. Suppose that

Images

Then

Images

(b) Let m(t) denote the minimal polynomial of T. Then by (a), m(Images)(w) = m(T)(w) = 0(w) = 0 for every wW; that is, Images is a zero of the polynomial m(t). Hence, the minimal polynomial of Imagesdivides m(t).

Invariant Direct-Sum Decompositions

10.7.   Prove Theorem 10.4: Suppose W1, W2, ..., Wr are subspaces of V with respective bases

Images

Then V is the direct sum of the Wi if and only if the union B = ∪i Bi is a basis of V.

Suppose B is a basis of V. Then, for any υV,

Images

where wi = ai1wi1 + ⋯ + ainiwiniWi. We next show that such a sum is unique. Suppose

Images

Because {wi1, ..., wini} is a basis of Images, and so

Images

Because B is a basis of V, aij = bij, for each i and each j. Hence, Images, and so the sum for υ is unique. Accordingly, V is the direct sum of the Wi.

Conversely, suppose V is the direct sum of the Wi. Then for any υV, υ = w1 + ⋯ + wr, where wiWi. Because {wiji} is a basis of Wi, each wi is a linear combination of the wiji, and so υ is a linear combination of the elements of B. Thus, B spans V. We now show that B is linearly independent. Suppose

Images

Note that ai1wi1 + ⋯ + ainiwiniWi. We also have that 0 = 0 + 0 ⋯ 0 ∈ Wi. Because such a sum for 0 is unique,

Images

The independence of the bases {wiji} implies that all the a’s are 0. Thus, B is linearly independent and is a basis of V.

10.8.   Suppose T:VV is linear and suppose T = T1T2 with respect to a T-invariant direct-sum decomposition V = UW. Show that

(a) m(t) is the least common multiple of m1(t) and m2(t), where m(t), m1(t), m2(t) are the minimum polynomials of T, T1, T2, respectively.

(b) Δ(t) = Δ1(t2(t), where Δ(t), Δ1(t), Δ2(t) are the characteristic polynomials of T, T1, T2, respectively.

(a) By Problem 10.6, each of m1(t) and m2(t) divides m(t). Now suppose f(t) is a multiple of both m1(t) and m2(t), then f(T1)(U) = 0 and f(T2)(W) = 0. Let υV, then υ = u + w with uU and wW. Now

Images

That is, T is a zero of f(t). Hence, m(t) divides f(t), and so m(t) is the least common multiple of m1(t) and m2(t).

(b) By Theorem 10.5, T has a matrix representation Images, where A and B are matrix representations of T1 and T2, respectively. Then, as required,

Images

10.9.   Prove Theorem 10.7: Suppose T:VV is linear, and suppose f(t) = g(t)h(t) are polynomials such that f(t) = 0 and g(t) and h(t) are relatively prime. Then V is the direct sum of the T-invariant subspaces U and W where U = Ker g(T) and W = Ker h(T).

Note first that U and W are T-invariant by Theorem 10.2. Now, because g(t) and h(t) are relatively prime, there exist polynomials r(t) and s(t) such that

Images

Hence, for the operator T,

Images

Let υV; then, by (*),

Images

But the first term in this sum belongs to W = Ker h(T), because

Images

Similarly, the second term belongs to U. Hence, V is the sum of U and W.

To prove that V = UW, we must show that a sum υ = u + w with uU, wW, is uniquely determined by υ. Applying the operator r(T)g(T) to υ = u + w and using g(T)u = 0, we obtain

Images

Also, applying (*) to w alone and using h(T)w = 0, we obtain

Images

Both of the above formulas give us w = r(T)g(T)υ, and so w is uniquely determined by υ. Similarly u is uniquely determined by υ. Hence, V = UW, as required.

10.10.  Prove Theorem 10.8: In Theorem 10.7 (Problem 10.9), if f(t) is the minimal polynomial of T (and g(t) and h(t) are monic), then g(t) is the minimal polynomial of the restriction T1 of T to U and h(t) is the minimal polynomial of the restriction T2 of T to W.

Let m1(t) and m2(t) be the minimal polynomials of T1 and T2, respectively. Note that g(T1) = 0 and h(T2) = 0 because U = Ker g(T) and W = Ker h(T). Thus,

Images

By Problem 10.9, f(t) is the least common multiple of m1(t) and m2(t). But m1(t) and m2(t) are relatively prime because g(t) and h(t) are relatively prime. Accordingly, f(t) = m1(t)m2(t). We also have that. These two equations together with (1) and the fact that all the polynomials are monic imply that g(t) = m1(t) and h(t) = m2(t), as required.

10.11.  Prove the Primary Decomposition Theorem 10.6: Let T:VV be a linear operator with minimal polynomial

Images

where the fi(t) are distinct monic irreducible polynomials. Then V is the direct sum of T-invariant subspaces W1, ..., Wr where Wi is the kernel of fi(T)ni. Moreover, fi(t)ni is the minimal polynomial of the restriction of T to Wi.

The proof is by induction on r. The case r = 1 is trivial. Suppose that the theorem has been proved for r − 1. By Theorem 10.7, we can write V as the direct sum of T-invariant subspaces W1 and V1, where W1 is the kernel of f1(T)n1 and where V1 is the kernel of f2(T)n2fr(T)nr. By Theorem 10.8, the minimal polynomials of the restrictions of T to W1 and V1 are f1(t)n1 and f2(t)n2fr(t)nr, respectively.

Denote the restriction of T to V1 by Images. By the inductive hypothesis, V1 is the direct sum of subspaces W2, ..., Wr such that Wi is the kernel of fi(T1)ni and such that fi(t)ni is the minimal polynomial for the restriction of Images to Wi. But the kernel of fi(T)ni, for i = 2, ..., r is necessarily contained in V1, because fi(t)ni divides f2(t)n2fr(t)nr. Thus, the kernel of fi(T)ni is the same as the kernel of fi(T1)ni, which is Wi. Also, the restriction of T to Wi is the same as the restriction of Images to Wi (for i = 2, ..., r); hence, fi(t)ni is also the minimal polynomial for the restriction of T to Wi. Thus, V = W1W2Wr is the desired decomposition of T.

10.12.  Prove Theorem 10.9: A linear operator T:VV has a diagonal matrix representation if and only if its minimal polynomal m(t) is a product of distinct linear polynomials.

Suppose m(t) is a product of distinct linear polynomials, say,

Images

where the λi are distinct scalars. By the Primary Decomposition Theorem, V is the direct sum of subspaces W1, ..., Wr, where Wi = Ker(TλiI). Thus, if υWi, then (TλiI)(υ) = 0 or T(υ) = λiυ. In other words, every vector in Wi is an eigenvector belonging to the eigenvalue λi. By Theorem 10.4, the union of bases for W1, ..., Wr is a basis of V. This basis consists of eigenvectors, and so T is diagonalizable.

Conversely, suppose T is diagonalizable (i.e., V has a basis consisting of eigenvectors of T ). Let λ1, ..., λs be the distinct eigenvalues of T. Then the operator

Images

maps each basis vector into 0. Thus, f(t) = 0, and hence, the minimal polynomial m(t) of T divides the polynomial

Images

Accordingly, m(t) is a product of distinct linear polynomials.

Nilpotent Operators, Jordan Canonical Form

10.13.  Let T:V be linear. Suppose, for υV, Tk(υ) = 0 but Tk − 1(υ) ≠ 0. Prove

(a) The set S = {υ, T(υ), ..., Tk − 1(υ)} is linearly independent.

(b) The subspace W generated by S is T-invariant.

(c) The restriction Images of T to W is nilpotent of index k.

(d) Relative to the basis {Tk − 1(υ), ..., T(υ), υ} of W, the matrix of T is the k-square Jordan nilpotent block Nk of index k (see Example 10.5).

(a) Suppose

Images

Applying Tk − 1 to (*) and using Tk(υ) = 0, we obtain aTk − 1(υ) = 0; because Tk − 1(υ) ≠ 0, a = 0. Now applying Tk − 2 to (*) and using Tk(υ) = 0 and a = 0, we fiind a1Tk − 1(υ) = 0; hence, a1 = 0. Next applying Tk − 3 to (*) and using Tk(υ) = 0 and a = a1 = 0, we obtain a2Tk − 1(υ) = 0; hence, a2 = 0. Continuing this process, we find that all the a’s are 0; hence, S is independent.

(b) Let υW. Then

Images

Using Tk(υ) = 0, we have

Images

Thus, W is T-invariant.

(c) By hypothesis, Tk(υ) = 0. Hence, for i = 0, ..., k − 1,

Images

That is, applying Images to each generator of W, we obtain 0; hence, Images and so Images is nilpotent of index at most k. On the other hand, Images; hence, T is nilpotent of index exactly k.

(d) For the basis {Tk − 1(υ), Tk − 2(υ), ..., T(υ), υ} of W,

Images

Hence, as required, the matrix of T in this basis is the k-square Jordan nilpotent block Nk.

10.14.  Let T:VV be linear. Let U = Ker Ti and W = Ker Ti + 1. Show that

(a) UW, (b) T(W) ⊆ U.

(a) Suppose uU = Ker Ti. Then Ti(u) = 0 and so Ti+ 1u) = T(Ti(u)) = T(0) = 0. Thus, u ∈ Ker Ti + 1 = W. But this is true for every uU; hence, UW.

(b) Similarly, if wW = Ker Ti + 1, then Ti + 1(w) = 0. Thus, Ti + 1(w) = Ti(T(w)) = Ti(0) = 0 and so T(W) ⊆ U.

10.15. Let T:V be linear. Let X = Ker Ti − 2, Y = Ker Ti − 1, Z = Ker Ti. Therefore (Problem 10.14),

Images

are bases of X, Y, Z, respectively. Show that

Images

is contained in Y and is linearly independent.

By Problem 10.14, T(Z) ⊆ Y, and hence SY. Now suppose S is linearly dependent. Then there exists a relation

Images

where at least one coefficient is not zero. Furthermore, because {ui} is independent, at least one of the bk must be nonzero. Transposing, we find

Images

Hence,

Images

Thus,

Images

Because {ui, υj} generates Y, we obtain a relation among the ui, υj, wk where one of the coefficients (i.e., one of the bk) is not zero. This contradicts the fact that {ui, υj, wk} is independent. Hence, S must also be independent.

10.16.  Prove Theorem 10.10: Let T:VV be a nilpotent operator of index k. Then T has a unique block diagonal matrix representation consisting of Jordan nilpotent blocks N. There is at least one N of order k, and all other N are of orders ≤k. The total number of N of all orders is equal to the nullity of T.

Suppose dim V = n. Let W1 = Ker T, W2 = Ker T2, ..., Wk = Ker Tk. Let us set mi = dim Wi, for i = 1, ..., k. Because T is of index k, Wk = V and Wk − 1V and so mk − 1 < mk = n. By Problem 10.14,

Images

Thus, by induction, we can choose a basis {u1, ..., un} of V such that {u1, ..., umi} is a basis of Wi.

We now choose a new basis for V with respect to which T has the desired form. It will be convenient to label the members of this new basis by pairs of indices. We begin by setting

Images

and setting

Images

By the preceding problem,

Images

is a linearly independent subset of Wk − 1. We extend S1 to a basis of Wk − 1 by adjoining new elements (if necessary), which we denote by

Images

Next we set

Images

Again by the preceding problem,

Images

is a linearly independent subset of Wk − 2, which we can extend to a basis of Wk − 2 by adjoining elements

Images

Continuing in this manner, we get a new basis for V, which for convenient reference we arrange as follows:

Images

The bottom row forms a basis of W1, the bottom two rows form a basis of W2, and so forth. But what is important for us is that T maps each vector into the vector immediately below it in the table or into 0 if the vector is in the bottom row. That is,

Images

Now it is clear [see Problem 10.13(d)] that T will have the desired form if the υ(i, j) are ordered lexicographically: beginning with υ (1, 1) and moving up the first column to υ(1, k), then jumping to υ(2, 1) and moving up the second column as far as possible.

Moreover, there will be exactly mkmk − 1 diagonal entries of order k3. Also, there will be

Images

as can be read off directly from the table. In particular, because the numbers m1, … ,mk are uniquely determined by T, the number of diagonal entries of each order is uniquely determined by T. Finally, the identity

Images

shows that the nullity m1 of T is the total number of diagonal entries of T.

10.17.  Let Images. The reader can verify that A and B are both nilpotent of index 3; that is, A3 = 0 but A2 ≠ 0, and B3 = 0 but B2 ≠ 0. Find the nilpotent matrices MA and MB in canonical form that are similar to A and B, respectively.

Because A and B are nilpotent of index 3, MA and MB must each contain a Jordan nilpotent block of order 3, and none greater then 3. Note that rank(A) = 2 and rank(B) = 3, so nullity(A) = 5 − 2 = 3 and nullity(B) = 5 − 3 = 2. Thus, MA must contain three diagonal blocks, which must be one of order 3 and two of order 1; and MB must contain two diagonal blocks, which must be one of order 3 and one of order 2. Namely,

Images

10.18.  Prove Theorem 10.11 on the Jordan canonical form for an operator T.

By the primary decomposition theorem, T is decomposable into operators T1, ..., Tr; that is, T = T1 ⊕ ⋯ ⊕ Tr, where (tλi)mi is the minimal polynomial of Ti. Thus, in particular,

Images

Set Ni = TiλiI. Then, for i = 1, ..., r,

Images

That is, Ti is the sum of the scalar operator λiI and a nilpotent operator Ni, which is of index mi because Images is the minimal polynomial of Ti.

Now, by Theorem 10.10 on nilpotent operators, we can choose a basis so that Ni is in canonical form. In this basis, Ti = Ni + λiI is represented by a block diagonal matrix Mi whose diagonal entries are the matrices Jij. The direct sum J of the matrices Mi is in Jordan canonical form and, by Theorem 10.5, is a matrix representation of T.

Last, we must show that the blocks Jij satisfy the required properties. Property (i) follows from the fact that Ni is of index mi. Property (ii) is true because T and J have the same characteristic polynomial. Property (iii) is true because the nullity of Ni = TiλiI is equal to the geometric multiplicity of the eigenvalue λi. Property (iv) follows from the fact that the Ti and hence the Ni are uniquely determined by T.

10.19.  Determine all possible Jordan canonical forms J for a linear operator T:VV whose characteristic polynomial Δ(t) = (t − 2)5 and whose minimal polynomial m(t) = (t − 2)2.

J must be a 5 × 5 matrix, because Δ(t) has degree 5, and all diagonal elements must be 2, because 2 is the only eigenvalue. Moreover, because the exponent of t − 2 in m(t) is 2, J must have one Jordan block of order 2, and the others must be of order 2 or 1. Thus, there are only two possibilities:

Images

10.20.  Determine all possible Jordan canonical forms for a linear operator T:VV whose characteristic polynomial Δ(t) = (t − 2)3(t − 5)2. In each case, find the minimal polynomial m(t).

Because t − 2 has exponent 3 in Δ(t), 2 must appear three times on the diagonal. Similarly, 5 must appear twice. Thus, there are six possibilities:

Images

The exponent in the minimal polynomial m(t) is equal to the size of the largest block. Thus,

(a) m(t) = (t − 2)3(t − 5)2, (b) m(t) = (t − 2)3(t − 5), (c) m(t) = (t − 2)2(t − 5)2,

(d) m(t) = (t − 2)2(t − 5), (e) m(t) = (t − 2)(t − 5)2, (f) m(t) = (t − 2)(t − 5)

Quotient Space and Triangular Form

10.21.  Let W be a subspace of a vector space V. Show that the following are equivalent:

(i) uυ + W, (ii) uυW, (iii) υu + W.

Suppose uυ + W. Then there exists w0W such that u = υ + w0. Hence, u υ = w0W. Conversely, suppose uυW. Then uυ = w0 where w0W. Hence, u = υ + w0υ + W. Thus, (i) and (ii) are equivalent.

We also have uυW iff − (uυ ) = υuW iff υu + W. Thus, (ii) and (iii) are also equivalent.

10.22.  Prove the following: The cosets of W in V partition V into mutually disjoint sets. That is,

(a) Any two cosets u + W and υ + W are either identical or disjoint.

(b) Each υV belongs to a coset; in fact, υυ + W.

Furthermore, u + W = υ + W if and only if uυW, and so (υ + w) + W = υ + W for any wW.

Let υV. Because 0 ∈ W, we have υ = υ + 0 ∈ υ + W, which proves (b).

Now suppose the cosets u + W and υ + W are not disjoint; say, the vector x belongs to both u + W and υ + W. Then uxW and xυW. The proof of (a) is complete if we show that u + W = υ + W. Let u + w0 be any element in the coset u + W. Because ux, xυ, w0 belongs to W,

Images

Thus, u + w0υ + W, and hence the cost u + W is contained in the coset υ + W. Similarly, υ + W is contained in u + W, and so u + W = υ + W.

The last statement follows from the fact that u + W = υ + W if and only if uυ + W, and, by Problem 10.21, this is equivalent to uυW.

10.23.  Let W be the solution space of the homogeneous equation 2x + 3y + 4z = 0. Describe the cosets of W in R3.

W is a plane through the origin O = (0, 0, 0), and the cosets of W are the planes parallel to W. Equivalently, the cosets of W are the solution sets of the family of equations

Images

In fact, the coset υ + W, where υ = (a, b, c), is the solution set of the linear equation

Images

10.24.  Suppose W is a subspace of a vector space V. Show that the operations in Theorem 10.15 are well defined; namely, show that if u + W = u′ + W and υ + W = υ ′ + W, then

Images

(a) Because u + W = u′ + W and υ + W = υ′ + W, both uu′ and υυ ′ belong to W. But then (u + υ) (u′ + υ′) = (uu′) + (υυ′) ∈ W. Hence, (u + υ) + W = (u′ + υ′) + W.

(b) Also, because uu′ ∈ W implies k(uu′) ∈ W, then kuku′ = k(uu′) ∈ W; accordingly, ku + W = ku′ + W.

10.25.  Let V be a vector space and W a subspace of V. Show that the natural map η: VV/W, defined by η(υ) = υ + W, is linear.

For any u, υV and any kK, we have

Images

and

Images

Accordingly, η is linear.

10.26.  Let W be a subspace of a vector space V. Suppose {w1, ..., wr} is a basis of W and the set of cosets Images, where Images, is a basis of the quotient space. Show that the set of vectors B = {υ1, ..., υs, w1, ..., wr} is a basis of V. Thus, dim V = dim W + dim(V/W).

Suppose uV. Because Images is a basis of V/W,

Images

Hence, u = a1υ1 + ⋯ + asυs + w, where wW. Since {wi} is a basis of W,

Images

Accordingly, B spans V.

We now show that B is linearly independent. Suppose

Images

Then

Images

Because Images is independent, the c’s are all 0. Substituting into (1), we find d1w1 + ⋯ + drwr = 0. Because {wi} is independent, the d’s are all 0. Thus, B is linearly independent and therefore a basis of V.

10.27.  Prove Theorem 10.16: Suppose W is a subspace invariant under a linear operator T:VV. Then T induces a linear operator Images on V/W defined by Images. Moreover, if T is a zero of any polynomial, then so is Images. Thus, the minimal polynomial of Images divides the minimal polynomial of T.

We first show that Images is well defined; that is, if u + W = υ + W, then Images. If u + W = υ + W, then uυW, and, as W is T-invariant, T(uυ ) = T(u) − T(υ) ∈ W. Accordingly,

Images

as required.

We next show that Images is linear. We have

Images

Furthermore,

Images

Thus, Images is linear.

Now, for any coset u + W in V/W,

Images

Hence, Images. Similarly, Images for any n. Thus, for any polynomial

Images

and so Images. Accordingly, if T is a root of f(t) then Images; that is, Images is also a root of f(t). The theorem is proved.

10.28.  Prove Theorem 10.1: Let T:VV be a linear operator whose characteristic polynomial factors into linear polynomials. Then V has a basis in which T is represented by a triangular matrix.

The proof is by induction on the dimension of V. If dim V = 1, then every matrix representation of T is a 1 × 1 matrix, which is triangular.

Now suppose dim V = n > 1 and that the theorem holds for spaces of dimension less than n. Because the characteristic polynomial of T factors into linear polynomials, T has at least one eigenvalue and so at least one nonzero eigenvector υ, say T(υ) = a11υ. Let W be the one-dimensional subspace spanned by υ. Set Images = V/W. Then (Problem 10.26) dim Images = dim V − dim W = n − 1. Note also that W is invariant under T. By Theorem 10.16, T induces a linear operator Images on Images whose minimal polynomial divides the minimal polynomial of T. Because the characteristic polynomial of T is a product of linear polynomials, so is its minimal polynomial, and hence, so are the minimal and characteristic polynomials of Images. Thus, Images and Images satisfy the hypothesis of the theorem. Hence, by induction, there exists a basis Images of Images such that

Images

Now let υ2, ..., υn be elements of V that belong to the cosets υ2, ..., υn, respectively. Then {υ, υ2, ..., υn} is a basis of V (Problem 10.26). Because Images, we have

Images

But W is spanned by υ ; hence, T(υ2) − a22υ2 is a multiple of υ, say,

Images

Similarly, for i = 3, ..., n

Images

Thus,

Images

and hence the matrix of T in this basis is triangular.

Cyclic Subspaces, Rational Canonical Form

10.29.  Prove Theorem 10.12: Let Z(υ, T) be a T-cyclic subspace, Tυ the restriction of T to Z(υ, T), and mυ(t) = tk + ak − 1tk − 1 + ⋯ + a0 the T-annihilator of υ. Then,

(i) The set {υ, T(υ), ..., Tk − 1(υ)} is a basis of Z(υ, T); hence, dim Z(υ, T) = k.

(ii) The minimal polynomial of Tυ is mυ(t).

(iii) The matrix of Tυ in the above basis is the companion matrix C = C(mυ) of mυ(t) [which has 1’s below the diagonal, the negative of the coefficients a0, a1, ..., ak − 1 of mυ(t) in the last column, and 0’s elsewhere].

(i) By definition of mυ(t), Tk(υ) is the first vector in the sequence υ, T(υ), T2(υ), ... that, is a linear combination of those vectors that precede it in the sequence; hence, the set B = {υ, T(υ), ..., Tk − 1(υ)} is linearly independent. We now only have to show that Z(υ, T) = L(B), the linear span of B. By the above, Tk(υ) ∈ L(B). We prove by induction that Tn(υ) ∈ L(B) for every n. Suppose n > k and Tn − 1(υ) ∈ L(B)—that is, Tn − 1(υ) is a linear combination of υ, ..., Tk − 1(υ). Then Tn(υ) = T(Tn − 1(υ)) is a linear combination of T(υ), ..., Tk(υ). But Tk(υ) ∈ L(B); hence, Tn(υ) ∈ L(B) for every n. Consequently, f(T)(υ) ∈ L(B) for any polynomial f(t). Thus, Z(υ, T) = L(B), and so B is a basis, as claimed.

(ii) Suppose m(t)= ts + bs − 1ts − 1 + ⋯ + b0 is the minimal polynomial of Tυ. Then, because υZ(υ, T),

Images

Thus, Ts(υ) is a linear combination of υ, T(υ), ..., Ts − 1(υ), and therefore ks. However, mυ(T) = 0 and so mυ(Tυ) = 0. Then m(t) divides mυ(t), and so sk. Accordingly, k = s and hence mυ(t) = m(t).

Images

By definition, the matrix of Tυ in this basis is the tranpose of the matrix of coefficients of the above system of equations; hence, it is C, as required.

10.30.  Let T:VV be linear. Let W be a T-invariant subspace of V and Images the induced operator on V/W. Prove

(a) The T-annihilator of υV divides the minimal polynomial of T.

(b) The Images-annihilator of Images divides the minimal polynomial of T.

(a) The T-annihilator of υV is the minimal polynomial of the restriction of T to Z(υ, T); therefore, by Problem 10.6, it divides the minimal polynomial of T.

(b) The Images-annihilator of Images divides the minimal polynomial of Images, which divides the minimal polynomial of T by Theorem 10.16.

Remark: In the case where the minimum polynomial of T is f(t)n, where f(t) is a monic irreducible polynomial, then the T-annihilator of υV and the Images-annihilator of Images are of the form f(t)m, where mn.

10.31.  Prove Lemma 10.13: Let T:VV be a linear operator whose minimal polynomial is f(t)n, where f(t) is a monic irreducible polynomial. Then V is the direct sum of T-cyclic subspaces Zi = Z(υi, T), i = 1, ..., r, with corresponding T-annihilators

Images

Any other decomposition of V into the direct sum of T-cyclic subspaces has the same number of components and the same set of T-annihilators.

The proof is by induction on the dimension of V. If dim V = 1, then V is T-cyclic and the lemma holds. Now suppose dim V > 1 and that the lemma holds for those vector spaces of dimension less than that of V.

Because the minimal polynomial of T is f(t)n, there exists υ1V such that f(T)n − 1(υ1) ≠ 0; hence, the T-annihilator of υ1 is f(t)n. Let Z1 = Z(υ1, T) and recall that Z1 is T-invariant. Let Images and let Images be the linear operator on Images induced by T. By Theorem 10.16, the minimal polynomial of Images divides f(t)n; hence, the hypothesis holds for Images and Images. Consequently, by induction, Images is the direct sum of Images-cyclic subspaces; say,

Images

where the corresponding Images-annihilators are f(t)n2, ..., f(t)nr, nn2nr.

We claim that there is a vector υ2 in the coset Images whose T-annihilator is f(t)n2, the Images-annihilator of Images. Let w be any vector in Images. Then f(t)n2(w) ∈ Z1. Hence, there exists a polynomial g(t) for which

Images

Because f(t)n is the minimal polynomial of T, we have, by (1),

Images

But f(t)n is the T-annihilator of υ1; hence, f(t)n divides f(t)nn2g(t), and so g(t) = f(t)n2 h(t) for some polynomial h(t). We set

Images

Because wυ2 = h(T)(υ1) ∈ Z1, υ2 also belongs to the coset Images. Thus, the T-annihilator of υ2 is a multiple of the Images-annihilator of Images. On the other hand, by (1),

Images

Consequently, the T-annihilator of υ2 is f(t)n2, as claimed.

Similarly, there exist vectors υ3, ..., υrV such that Images and that the T-annihilator of υi is f(t)ni, the Images-annihilator of Images. We set

Images

Let d denote the degree of f(t), so that f(t)ni has degree dni. Then, because f(t)ni is both the T-annihilator of υi and the Images-annihilator of Images, we know that

Images

are bases for Z(υi, T) and Images, respectively, for i = 2, ..., r. But Images hence,

Images

is a basis for Images. Therefore, by Problem 10.26 and the relation Images (see Problem 10.27),

Images

is a basis for V. Thus, by Theorem 10.4, V = Z(υ1, T) ⊕ ⋯ ⊕ Z(υr, T), as required.

It remains to show that the exponents n1, ..., nr are uniquely determined by T. Because d = degree of f(t),

Images

Also, if s is any positive integer, then (Problem 10.59) f(T)s(Zi) is a cyclic subspace generated by f(T)s(υi), and it has dimension d(nis) if ni > s and dimension 0 if nis.

Now any vector υV can be written uniquely in the form υ = w1 + ⋯ + wr, where wiZi. Hence, any vector in f(T)s(V) can be written uniquely in the form

Images

where f(t)s(wi) ∈ f(T)s(Zi). Let t be the integer, dependent on s, for which

Images

Then

Images

and so

Images

The numbers on the left of (2) are uniquely determined by T. Set s = n − 1, and (2) determines the number of ni equal to n. Next set s = n − 2, and (2) determines the number of ni (if any) equal to n − 1. We repeat the process until we set s = 0 and determine the number of ni equal to 1. Thus, the ni are uniquely determined by T and V, and the lemma is proved.

10.32.  Let V be a seven-dimensional vector space over R, and let T:VV be a linear operator with minimal polynomial m(t) = (t2 − 2t + 5)(t − 3)3. Find all possible rational canonical forms M of T.

Because dim V = 7, there are only two possible characteristic polynomials, Δ1(t)= (t2 − 2t + 5)2(t − 3)3 or Δ1(t)= (t2 − 2t + 5)(t − 3)5. Moreover, the sum of the orders of the companion matrices must add up to 7. Also, one companion matrix must be C(t2 − 2t + 5) and one must be C((t − 3)3) = C(t3 − 9t2 + 27t − 27). Thus, M must be one of the following block diagonal matrices:

Images

Projections

10.33.  Suppose V = W1 ⊕ ⋯ ⊕ Wr. The projection of V into its subspace Wk is the mapping E: VV defined by E(υ) = wk, where υ = w1 + ⋯ + wr, wiWi. Show that (a) E is linear, (b) E2 = E.

(a) Because the sum υ = w1 + ⋯ + wr, wiW is uniquely determined by υ, the mapping E is well defined. Suppose, for uV, Images. Then

Images

are the unique sums corresponding to υ + u and . Hence,

Images

and therefore E is linear.

(b) We have that

Images

is the unique sum corresponding to wkWk; hence, E(wk) = wk. Then, for any υV,

Images

Thus, E2 = E, as required.

10.34.  Suppose E:VV is linear and E2 = E. Show that (a) E(u) = u for any u ∈ Im E (i.e., the restriction of E to its image is the identity mapping); (b) V is the direct sum of the image and kernel of E:V = Im E ⊕ Ker E; (c) E is the projection of V into Im E, its image. Thus, by the preceding problem, a linear mapping T:VV is a projection if and only if T2 = T; this characterization of a projection is frequently used as its definition.

(a) If u ∈ Im E, then there exists υV for which E(υ) = u; hence, as required,

Images

(b) Let υV. We can write υ in the form υ = E(υ) + υE(υ). Now E(υ) ∈ Im E and, because

Images

υE(υ) ∈ Ker E. Accordingly, V = Im E + Ker E.

Now suppose w ∈ Im E ∩ Ker E. By (i), E(w) = w because w ∈ Im E. On the other hand, E(w) = 0 because w ∈ Ker E. Thus, w = 0, and so Im E ∩ Ker E = {0}. These two conditions imply that V is the direct sum of the image and kernel of E.

(c) Let υV and suppose υ = u + w, where u ∈ Im E and w ∈ Ker E. Note that E(u) = u by (i), and E(w) = 0 because w ∈ Ker E. Hence,

Images

That is, E is the projection of V into its image.

10.35.  Suppose V = UW and suppose T:VV is linear. Show that U and W are both T-invariant if and only if TE = ET, where E is the projection of V into U.

Observe that E(υ) ∈ U for every υV, and that (i) E(υ) = υ iff υU, (ii) E(υ) = 0 iff υW. Suppose ET = TE. Let uU. Because E(u) = u,

Images

Hence, U is T-invariant. Now let wW. Because E(w) = 0,

Images

Hence, W is also T-invariant.

Conversely, suppose U and W are both T-invariant. Let υV and suppose υ = u + w, where uT and wW. Then T(u) ∈ U and T(w) ∈ W; hence, E(T(u)) = T(u) and E(T(w)) = 0. Thus,

Images

and

That is, (ET)(υ) = (TE)(υ) for every υV; therefore, ET = TE, as required.

SUPPLEMENTARY PROBLEMS

Invariant Subspaces

10.36.  Suppose W is invariant under T:VV. Show that W is invariant under f(t) for any polynomial f(t).

10.37.  Show that every subspace of V is invariant under I and 0, the identity and zero operators.

10.38.  Let W be invariant under T1: VV and T2: VV. Prove W is also invariant under T1 + T2 and T1T2.

10.39.  Let T:VV be linear. Prove that any eigenspace, EΛ is T-invariant.

10.40.  Let V be a vector space of odd dimension (greater than 1) over the real field R. Show that any linear operator on V has an invariant subspace other than V or {0}.

10.41.  Determine the invariant subspace of Images viewed as a linear operator on (a) R2, (b) C2.

10.42.  Suppose dim V = n. Show that T:VV has a triangular matrix representation if and only if there exist T-invariant subspaces W1W2 ⊂ ⋯ ⊂ Wn = V for which dim Wk = k, k = 1, ..., n.

Invariant Direct Sums

10.43.  The subspaces W1, ..., Wr are said to be independent if w1 + ⋯ + wr = 0, wiWi, implies that each wi = 0. Show that span(Wi) = W1 ⊕ ⋯ ⊕ Wr if and only if the Wi are independent. [Here span(Wi) denotes the linear span of the Wi.]

10.44.  Show that V = W1 ⊕ ⋯ ⊕ Wr if and only if (i) V = span(Wi) and (ii) for k = 1, 2, ..., r, Wk ∩ span(W1, ..., Wk − 1, Wk + 1, ..., Wr) = {0}.

10.45.  Show that span(Wi) = W1 ⊕ ⋯ ⊕ Wr if and only if dim [span(Wi) = dim W1 + ⋯ + dim Wr.

10.46.  Suppose the characteristic polynomial of T:VV is Δ(t) = f1(t)n1f2(t)n2fr(t)nr, where the fi(t) are distinct monic irreducible polynomials. Let V = W1 ⊕ ⋯ ⊕ Wr be the primary decomposition of V into T-invariant subspaces. Show that fi(t)ni is the characteristic polynomial of the restriction of T to Wi.

Nilpotent Operators

10.47.  Suppose T1 and T2 are nilpotent operators that commute (i.e., T1T2 = T2T1). Show that T1 + T2 and T1T2 are also nilpotent.

10.48.  Suppose A is a supertriangular matrix (i.e., all entries on and below the main diagonal are 0). Show that A is nilpotent.

10.49.  Let V be the vector space of polynomials of degree ≤n. Show that the derivative operator on V is nilpotent of index n + 1.

10.50.  Show that any Jordan nilpotent block matrix N is similar to its transpose NT (the matrix with 1’s below the diagonal and 0’s elsewhere).

10.51.  Show that two nilpotent matrices of order 3 are similar if and only if they have the same index of nilpotency. Show by example that the statement is not true for nilpotent matrices of order 4.

Jordan Canonical Form

10.52.  Find all possible Jordan canonical forms for those matrices whose characteristic polynomial Δ(t) and minimal polynomial m(t) are as follows:

(a) Δ(t) = (t − 2)4(t − 3)2, m(t) = (t − 2)2(t − 3)2,

(b) Δ(t) = (t − 7)5, m(t) = (t − 7)2, (c) Δ(t) = (t − 2)7, m(t) = (t − 2)3

10.53.  Show that every complex matrix is similar to its transpose. (Hint: Use its Jordan canonical form.)

10.54.  Show that all n × n complex matrices A for which An = I but AkI for k < n are similar.

10.55.  Suppose A is a complex matrix with only real eigenvalues. Show that A is similar to a matrix with only real entries.

Cyclic Subspaces

10.56.  Suppose T:VV is linear. Prove that Z(υ, T) is the intersection of all T-invariant subspaces containing υ.

10.57.  Let f(t) and g(t) be the T-annihilators of u and υ, respectively. Show that if f(t) and g(t) are relatively prime, then f(t)g(t) is the T-annihilator of u + υ.

10.58.  Prove that Z(u, T) = Z(υ, T) if and only if g(T)(u) = υ where g(t) is relatively prime to the T-annihilator of u.

10.59.  Let W = Z(υ, T), and suppose the T-annihilator of υ is f(t)n, where f(t) is a monic irreducible polynomial of degree d. Show that f(t)s(W) is a cyclic subspace generated by f(T)s(υ) and that it has dimension d(ns) if n > s and dimension 0 if ns.

Rational Canonical Form

10.60.  Find all possible rational forms for a 6 × 6 matrix over R with minimal polynomial:

(a) m(t) = (t2 − 2t + 3)(t + 1)2, (b) m(t) = (t − 2)3.

10.61.  Let A be a 4 × 4 matrix with minimal polynomial m(t) = (t2 + 1)(t2 − 3). Find the rational canonical form for A if A is a matrix over (a) the rational field Q, (b) the real field R, (c) the complex field C.

10.62.  Find the rational canonical form for the four-square Jordan block with λ’s on the diagonal.

10.63.  Prove that the characteristic polynomial of an operator T:VV is a product of its elementary divisors.

10.64.  Prove that two 3 × 3 matrices with the same minimal and characteristic polynomials are similar.

10.65.  Let C( f(t)) denote the companion matrix to an arbitrary polynomial f(t). Show that f(t) is the characteristic polynomial of C( f(t)).

Projections

10.66.  Suppose V = W1 ⊕ ⋯ ⊕ Wr. Let Ei denote the projection of V into Wi. Prove (i) EiEj = 0, ij; (ii) I = E1 + ⋯ + Er.

10.67.  Let E1, ..., Er be linear operators on V such that

(i) Images (i.e., the Ei are projections); (ii) EiEj = 0, ij; (iii) I = E1 + ⋯ + Er

Prove that V = Im E1 ⊕ ⋯ ⊕ Im Er.

10.68.  Suppose E: VV is a projection (i.e., E2 = E). Prove that E has a matrix representation of the form Images, where r is the rank of E and Ir is the r-square identity matrix.

10.69.  Prove that any two projections of the same rank are similar. (Hint: Use the result of Problem 10.68.)

10.70.  Suppose E: VV is a projection. Prove

(i) IE is a projection and V = Im E ⊕ Im (IE), (ii) I + E is invertible (if 1 + 1 ≠ 0).

Quotient Spaces

10.71.  Let W be a subspace of V. Suppose the set of cosets {υ1 + W, υ2 + W, ..., υn + W} in V/W is linearly independent. Show that the set of vectors {υ1, υ2, ..., υn} in V is also linearly independent.

10.72.  Let W be a substance of V. Suppose the set of vectors {u1, u2, ..., un} in V is linearly independent, and that L(ui) ∩ W = {0}. Show that the set of cosets {u1 + W, ..., un + W} in V/W is also linearly independent.

10.73.  Suppose V = UW and that {u1, ..., un} is a basis of U. Show that {u1 + W, ..., un + W} is a basis of the quotient spaces V/W. (Observe that no condition is placed on the dimensionality of V or W.)

10.74.  Let W be the solution space of the linear equation

Images

and let υ = (b1, b2, ..., bn) ∈ Kn. Prove that the coset υ + W of W in Kn is the solution set of the linear equation

Images

10.75.  Let V be the vector space of polynomials over R and let W be the subspace of polynomials divisible by t4 (i.e., of the form a0t4 + a1t5 + ⋯ + an − 4tn). Show that the quotient space V/W has dimension 4.

10.76.  Let U and W be subspaces of V such that WUV. Note that any coset u + W of W in U may also be viewed as a coset of W in V, because uU implies uV; hence, U/W is a subset of V/W. Prove that (i) U/W is a subspace of V/W, (ii) dim(V/W) − dim(U/W) = dim(V/U).

10.77.  Let U and W be subspaces of V. Show that the cosets of UW in V can be obtained by intersecting each of the cosets of U in V by each of the cosets of W in V:

Images

Images

10.78.  Let T:VV′ be linear with kernel W and image U. Show that the quotient space V/W is isomorphic to U under the mapping θ:V/WU defined by θ(υ + W) = T(υ). Furthermore, show that T = iθη, where η:VV/W is the natural mapping of V into V/W (i.e., η(υ) = υ + W), and Images is the inclusion mapping (i.e., i(u) = u). (See diagram.)

ANSWERS TO SUPPLEMENTARY PROBLEMS

10.41. (a) R2 and {0}, (b) C2, {0}, W1 = span(2, 1 − 2i), W2 = span(2, 1 + 2i)

10.52.

Images

Images

(c) Let Mk denote a Jordan block with λ = 2 and order k. Then diag(M3, M3, M1), diag(M3, M2, M2), diag(M3, M2, M1, M1), diag(M3, M1, M1, M1, M1)

10.60.  Let Images.

(a) diag(A, A, B), diag(A, B, B), diag(A, B, –1, –1); (b) diag(C, C), diag(C, D, 2), diag(C, 2, 2, 2)

10.61.  Let Images.

(a) diag(A, B), (b) Images, (c) Images

10.62.  Companion matrix with the last column [λ4, 4λ3, –6λ2, 4λ]T