Consider a basis S = {u1, u2,..., un} of a vector space V over a field K. For any vector υ 2 V, suppose υ = a1 u1 + a2 u2 + … + an un
Then the coordinate vector of υ relative to the basis S, which we assume to be a column vector (unless otherwise stated or implied), is denoted and defined by
Recall (Section 4.11) that the mapping υ 7![υ S, determined by the basis S, is an isomorphism between V and Kn.
This chapter shows that there is also an isomorphism, determined by the basis S, between the algebra A (V)of linear operators on V and the algebra M of n-square matrices over K. Thus, every linear mapping F: V → V will correspond to an n-square matrix [F]S determined by the basis S. We will also show how our matrix representation changes when we choose another basis.
Let T be a linear operator (transformation) from a vector space V into itself, and suppose S = {u1, u2,..., un} is a basis of V. Now T (u1), T (u2),…, T (un) are vectors in V, and so each is a linear combination of the vectors in the basis S; say,
The following definition applies.
DEFINITION: The transpose of the above matrix of coefficients, denoted by mS (T) or [T S, is called the matrix representation of T relative to the basis S, or simply the matrix of T in the basis S. (The subscript S may be omitted if the basis S is understood.)
Using the coordinate (column) vector notation, the matrix representation of T may be written in the form
That is, the columns of m (T) are the coordinate vectors of T (u1), T (u2),…, T (un), respectively.
EXAMPLE 6.1 Let F: R 2 → R 2 be the linear operator defined by F (x, y) = (2x + 3 y, 4 x 5y).
(a) Find the matrix representation of F relative to the basis S = {u1, u2} = {(1, 2), (2, 5)}.
(1) First find F (u1), and then write it as a linear combination of the basis vectors u1 and u2. (For notational convenience, we use column vectors.) We have
Solve the system to obtain x = 52, y = 22. Hence, F(u1) = 52 u1 − 22u2.
(2) Next find F(u2), and then write it as a linear combination of u1 and u2:
Solve the system to get x = 129, y = −55. Thus, F (u2) = 129 u1 55 u2.
Now write the coordinates of F (u1) and F (u2) as columns to obtain the matrix
(b) Find the matrix representation of F relative to the (usual) basis E = {e1, e2} = {(1, 0), (0, 1)}.
Find F (e1) and write it as a linear combination of the usual basis vectors e1 and e2, and then find F (e2) and write it as a linear combination of e1 and e2. We have
Note that the coordinates of F (e1) and F (e2) form the columns, not the rows, of [F E. Also, note that the arithmetic is much simpler using the usual basis of R 2.
EXAMPLE 6.2 Let V be the vector space of functions with basis S = {sin t, cos t, e3t}, and let D : V → V be the differential operator defined by D (f (t)) = d(f (t))/dt. We compute the matrix representing D in the basis S:
and so
Note that the coordinates of D (sin t), D (cos t), D (e3t) form the columns, not the rows, of [D].
Consider the following matrix A, which may be viewed as a linear operator on R2, and basis S of R2:
(We write vectors as columns, because our map is a matrix.) We find the matrix representation of A relative to the basis S.
(1) First we write A (u1) as a linear combination of u1 and u2. We have
Solving the system yields x = 7, y = 4. Thus, A(u1) = 7 u1 4u2.
(2) Next we write A (u2) as a linear combination of u1 and u2. We have
Solving the system yields x = 14, y = −9. Thus, A (u2) = 14 u1 9 u2. Writing the coordinates of A (u1)and A (u2) as columns gives us the following matrix representation of A:
Remark: Suppose we want to find the matrix representation of A relative to the usual basis E = {e1, e2} = {[1, 0]T, [0, 1T} of R2: We have
Note that [A]E is the original matrix A. This result is true in general:
Next follows an algorithm for finding matrix representations. The first Step 0 is optional. It may be useful to use it in Step 1 (b), which is repeated for each basis vector.
ALGORITHM 6.1: The input is a linear operator T on a vector space V and a basis S = {u1, u2,..., un} of V. The output is the matrix representation [T]S.
Step 0. Find a formula for the coordinates of an arbitrary vector υ relative to the basis S.
Step 1. Repeat for each basis vector uk in S:
(a) Find T (uk).
(b) Write T (uk) as a linear combination of the basis vectors u1, u2,..., un.
Step 2. Form the matrix [T]S whose columns are the coordinate vectors in Step 1 (b).
EXAMPLE 6.3 Let F: R 2 → R2 be defined by F (x, y) = (2x + 3 y, 4x − 5 y). Find the matrix representation [F]S of F relative to the basis S = {u1, u2} = {(1, 2), (2, 5)}.
(Step 0) First find the coordinates of (a, b) ∈ R2 relative to the basis S. We have
Solving for x and y in terms of a and b yields x = 5a + 2b, y = 2 a b. Thus,
(Step 1) Now we find F(u1) and write it as a linear combination of u1 and u2 using the above formula for (a; b), and then we repeat the process for F(u2). We have
(Step 2) Finally, we write the coordinates of F(u1) and F (u2) as columns to obtain the required matrix:
This subsection gives the main properties of the matrix representations of linear operators T on a vector space V. We emphasize that we are always given a particular basis S of V.
Our first theorem, proved in Problem 6.9, tells us that the “action” of a linear operator T on a vector υ is preserved by its matrix representation.
THEOREM 6.1: Let T: V → V be a linear operator, and let S be a (finite) basis of V. Then, for any vector υ in V, [T]S[υ]S = [T(υ)]S.
EXAMPLE 6.4 Consider the linear operator F on R2 and the basis S of Example 6.3; that is,
Let
Using the formula from Example 6.3, we get
We verify Theorem 6.1 for this vector υ (where [F] is obtained from Example 6.3):
Given a basis S of a vector space V, we have associated a matrix [T to each linear operator T in the algebra A (V) of linear operators on V. Theorem 6.1 tells us that the “action” of an individual linear operator T is preserved by this representation. The next two theorems (proved in Problems 6.10 and 6.11) tell us that the three basic operations in A (V) with these operators—namely (i) addition, (ii) scalar multiplication, and (iii) composition—are also preserved.
THEOREM 6.2: Let V be an n-dimensional vector space over K, let S be a basis of V, and let M be the algebra of n n matrices over K. Then the mapping
is a vector space isomorphism. That is, for any F, G ∈ A (V) and any k ∈ K,
(i) m (F + G) = m (F)+ m (G) or [F + G = [F +[G
(ii) m (kF) = km (F) or [kF = k[F
(iii) m is bijective (one-to-one and onto).
THEOREM 6.3: For any linear operators F, G ∈ A (V),
(Here G ° F denotes the composition of the maps G and F.)
Let V be an n-dimensional vector space over a field K. We have shown that once we have selected a basis S of V, every vector υ ∈ V can be represented by means of an n-tuple [υ]S in Kn, and every linear operator T in A (V) can be represented by an n n matrix over K. We ask the following natural question:
In order to answer this question, we first need a definition.
DEFINITION: Let S = {u1, u2,..., un} be a basis of a vector space V, and let S′ = fυ1, υ 2,..., υ ng be another basis. (For reference, we will call S the “old” basis and S′ the “new” basis.) Because S is a basis, each vector in the “new” basis S′ can be written uniquely as a linear combination of the vectors in S; say,
Let P be the transpose of the above matrix of coefficients; that is, let P = [pij],where pij = aji. Then P is called the change-of-basis matrix (or transition matrix) from the “old” basis S to the “new” basis S′.
The following remarks are in order.
Remark 1: The above change-of-basis matrix P may also be viewed as the matrix whose columns are, respectively, the coordinate column vectors of the “new” basis vectors υi relative to the “old” basis S; namely,
Remark 2: Analogously, there is a change-of-basis matrix Q from the “new” basis S′ to the “old” basis S. Similarly, Q may be viewed as the matrix whose columns are, respectively, the coordinate column vectors of the “old” basis vectors ui relative to the “new” basis S′; namely,
Remark 3: Because the vectors υ 1, υ 2,..., υ n in the new basis S′ are linearly independent, the matrix P is invertible (Problem 6.18). Similarly, Q is invertible. In fact, we have the following proposition (proved in Problem 6.18).
PROPOSITION 6.4: Let P and Q be the above change-of-basis matrices. Then Q = P-1.
Now suppose S = {u1, u2,..., un} is a basis of a vector space V, and suppose P = [pij is any nonsingular matrix. Then the n vectors
corresponding to the columns of P, are linearly independent [Problem 6.21 (a)]. Thus, they form another basis S′ of V. Moreover, P will be the change-of-basis matrix from S to the new basis S′.
EXAMPLE 6.5 Consider the following two bases of R2:
(a) Find the change-of-basis matrix P from S to the “new” basis S′.
Write each of the new basis vectors of S′ as a linear combination of the original basis vectors u1 and u2 of S. We have
Thus,
Note that the coordinates of υ 1 and υ 2 are the columns, not rows, of the change-of-basis matrix P.
(b) Find the change-of-basis matrix Q from the “new” basis S′ back to the “old” basis S.
Here we write each of the “old” basis vectors u1 and u2 of S′ as a linear combination of the “new” basis vectors υ 1 and υ 2 of S′. This yields
As expected from Proposition 6.4, Q = P−1. (In fact, we could have obtained Q by simply finding P−1.)
EXAMPLE 6.6 Consider the following two bases of R3:
and
(a) Find the change-of-basis matrix P from the basis E to the basis S.
Because E is the usual basis, we can immediately write each basis element of S as a linear combination of the basis elements of E. Specifically,
Again, the coordinates of u1, u2, u3 appear as the columns in P. Observe that P is simply the matrix whose columns are the basis vectors of S. This is true only because the original basis was the usual basis E.
(b) Find the change-of-basis matrix Q from the basis S to the basis E.
The definition of the change-of-basis matrix Q tells us to write each of the (usual) basis vectors in E as a linear combination of the basis elements of S. This yields
We emphasize that to find Q, we need to solve three 3 × 3 systems of linear equations—one 3 × 3 system for each of e1, e2, e3.
Alternatively, we can find Q = P1 by forming the matrix M = [P, I] and row reducing M to row canonical form:
thus,
(Here we have used the fact that Q is the inverse of P.)
The result in Example 6.6 (a) is true in general. We state this result formally, because it occurs often.
PROPOSITION 6.5: The change-of-basis matrix from the usual basis E of Kn to any basis S of Kn is the matrix P whose columns are, respectively, the basis vectors of S.
First we show how a change of basis affects the coordinates of a vector in a vector space V. The following theorem is proved in Problem 6.22.
THEOREM 6.6: Let P be the change-of-basis matrix from a basis S to a basis S′ in a vector space V. Then, for any vector υ 2 V, we have
Namely, if we multiply the coordinates of υ in the original basis S by P-1, we get the coordinates of υ in the new basis S′.
Remark 1: Although P is called the change-of-basis matrix from the old basis S to the new basis S′, we emphasize that P-1 transforms the coordinates of υ in the original basis S into the coordinates of υ in the new basis S′.
Remark 2: Because of the above theorem, many texts call Q = P-1, not P, the transition matrix from the old basis S to the new basis S′. Some texts also refer to Q as the change-of-coordinates matrix.
We now give the proof of the above theorem for the special case that dim V = 3. Suppose P is the change-of-basis matrix from the basis S = {u1, u2, u3g to the basis S′ = fυ1, υ 2, υ 3g; say,
Now suppose υ ∈ V and, say, υ = k1υ1 + k2υ2 + k3υ3. Then, substituting for υ 1, υ 2, υ 3 from above, we obtain
Thus,
Accordingly,
Finally, multiplying the equation [υ S = P[υ S, by P-1, we get
The next theorem (proved in Problem 6.26) shows how a change of basis affects the matrix representation of a linear operator.
THEOREM 6.7: Let P be the change-of-basis matrix from a basis S to a basis S′ in a vector space V. Then, for any linear operator T on V,
That is, if A and B are the matrix representations of T relative, respectively, to S and S′, then
EXAMPLE 6.7 Consider the following two bases of R3:
and
The change-of-basis matrix P from E to S and its inverse P-1 were obtained in Example 6.6.
(a) Write υ = (1, 3, 5) as a linear combination of u1, u2, u3, or, equivalently, find [υ S.
One way to do this is to directly solve the vector equation υ = xu1 + yu2 + zu3; that is,
The solution is x = 7, y = 5, z = 4, so υ = 7 u1 5 u2 + 4 u3.
On the other hand, we know that [υE = [1, 3, 5]T, because E is the usual basis, and we already know P-1.
Therefore, by Theorem 6.6,
Thus, again, υ = 7 u1 5 u2 + 4 u3.
(b) Let , which may be viewed as a linear operator on R3. Find the matrix B that represents A relative to the basis S.
The definition of the matrix representation of A relative to the basis S tells us to write each of A (u1), A (u2), A (u3)as a linear combination of the basis vectors u1, u2, u3 of S. This yields
We emphasize that to find B, we need to solve three 3 × 3 systems of linear equations—one 3 × 3 system for each of A (u1), A (u2), A (u3).
On the other hand, because we know P and P-1, we can use Theorem 6.7. That is,
This, as expected, gives the same result.
Suppose A and B are square matrices for which there exists an invertible matrix P such that B = P-1 AP; then B is said to be similar to A,or B is said to be obtained from A by a similarity transformation. We show (Problem 6.29) that similarity of matrices is an equivalence relation.
By Theorem 6.7 and the above remark, we have the following basic result.
THEOREM 6.8: Two matrices represent the same linear operator if and only if the matrices are similar.
That is, all the matrix representations of a linear operator T form an equivalence class of similar matrices.
A linear operator T is said to be diagonalizable if there exists a basis S of V such that T is represented by a diagonal matrix; the basis S is then said to diagonalize T. The preceding theorem gives us the following result.
THEOREM 6.9: Let A be the matrix representation of a linear operator T. Then T is diagonalizable if and only if there exists an invertible matrix P such that P-1 AP is a diagonal matrix.
That is, T is diagonalizable if and only if its matrix representation can be diagonalized by a similarity transformation.
We emphasize that not every operator is diagonalizable. However, we will show (Chapter 10) that every linear operator can be represented by certain “standard” matrices called its normal or canonical forms. Such a discussion will require some theory of fields, polynomials, and determinants.
Suppose f is a function on square matrices that assigns the same value to similar matrices; that is, f (A) = f (B)whenever A is similar to B. Then f induces a function, also denoted by f, on linear operators T in the following natural way. We define
where S is any basis. By Theorem 6.8, the function is well defined.
The determinant (Chapter 8) is perhaps the most important example of such a function. The trace (Section 2.7) is another important example of such a function.
EXAMPLE 6.8 Consider the following linear operator F and bases E and S of R 2:
By Example 6.1, the matrix representations of F relative to the bases E and S are, respectively,
Using matrix A, we have
(i) Determinant of F = det (A) = 10 12 = 22;
(ii) Trace of F = tr (A) = 2 − 5 = −3.
On the other hand, using matrix B, we have
(i) Determinant of F = det (B) = 2860 + 2838 = 22;
(ii) Trace of F = tr (B) = 52 55 = 3.
As expected, both matrices yield the same result.
Last, we consider the general case of linear mappings from one vector space into another. Suppose V and U are vector spaces over the same field K and, say, dim V = m and dim U = n. Furthermore, suppose
are arbitrary but fixed bases, respectively, of V and U.
Suppose F: V → U is a linear mapping. Then the vectors F (υ1), F (υ2),..., F (υm) belong to U, and so each is a linear combination of the basis vectors in S′; say,
DEFINITION: The transpose of the above matrix of coefficients, denoted by mS;S0 (F) or [F S;S0, is called the matrix representation of F relative to the bases S and S′. [We will use the simple notation m (F) and [F when the bases are understood.]
The following theorem is analogous to Theorem 6.1 for linear operators (Problem 6.67).
THEOREM 6.10: For any vector υ 2 V, [F S;S[υ]S = [F (υ)S0.
That is, multiplying the coordinates of υ in the basis S of V by [F, we obtain the coordinates of F (υ) in the basis S′ of U.
Recall that for any vector spaces V and U, the collection of all linear mappings from V into U is a vector space and is denoted by Hom (V, U). The following theorem is analogous to Theorem 6.2 for linear operators, where now we let M = M m;n denote the vector space of all m × n matrices (Problem 6.67).
THEOREM 6.11: The mapping m: Hom (V, U) → M defined by m (F)= [F is a vector space isomorphism. That is, for any F, G ∈ Hom (V, U) and any scalar k,
(i)m (F + G) = m (F)+ m (G) or [F + G = [F +[G
(ii) m (kF) = km (F) or [kF = k[F
(iii) m is bijective (one-to-one and onto).
Our next theorem is analogous to Theorem 6.3 for linear operators (Problem 6.67).
THEOREM 6.12: Let S, S′, S" be bases of vector spaces V, U, W, respectively. Let F: V → U and G ∘ U → W be linear mappings. Then
That is, relative to the appropriate bases, the matrix representation of the composition of two mappings is the matrix product of the matrix representations of the individual mappings.
Next we show how the matrix representation of a linear mapping F: V → U is affected when new bases are selected (Problem 6.67).
THEOREM 6.13: Let P be the change-of-basis matrix from a basis e to a basis e′ in V, and let Q be the change-of-basis matrix from a basis f to a basis f0 in U. Then, for any linear map F: V → U,
In other words, if A is the matrix representation of a linear mapping F relative to the bases e and f, and B is the matrix representation of F relative to the bases e′ and f', then
Our last theorem, proved in Problem 6.36, shows that any linear mapping from one vector space V into another vector space U can be represented by a very simple matrix. We note that this theorem is analogous to Theorem 3.18 for m × n matrices.
THEOREM 6.14: Let F: V → U be linear and, say, rank (F) = r. Then there exist bases of V and U such that the matrix representation of F has the form
where Ir is the r-square identity matrix.
The above matrix A is called the normal or canonical form of the linear map F.
SOLVED PROBLEMS
6.1. Consider the linear mapping F: R 2 → R 2 defined by F (x, y) = (3x + 4 y, 2x 5y) and the following bases of R2:
(a) Find the matrix A representing F relative to the basis E.
(b) Find the matrix B representing F relative to the basis S.
(a) Because E is the usual basis, the rows of A are simply the coefficients in the components of F (x, y); that is, using (a, b) = ae1 + be2, we have
Note that the coefficients of the basis vectors are written as columns in the matrix representation.
(b) First find F (u1) and write it as a linear combination of the basis vectors u1 and u2. We have
Solve the system to obtain x = −49, y = 30. Therefore,
Next find F (u2) and write it as a linear combination of the basis vectors u1 and u2. We have
Solve for x and y to obtain x = −76, y = 47. Hence,
Write the coefficients of u1 and u2 as columns to obtain
(b′) Alternatively, one can first find the coordinates of an arbitrary vector (a, b) in R2 relative to the basis S.
We have
Solve for x and y in terms of a and b to get x = −3 a + 2 b, y = 2a − b. Thus,
Then use the formula for (a; b) to find the coordinates of F(u1) and F(u2) relative to S:
6.2. Consider the following linear operator G on R 2 and basis S:
(a) Find the matrix representation [G S of G relative to S.
(b) Verify [G S[υ S = [G (υ)S for the vector υ = (4, 3) in R2.
First find the coordinates of an arbitrary vector υ = (a, b) in R2 relative to the basis S.
We have
Solve for x and y in terms of a and b to get x = −5a + 2b, y = 3a − b. Thus,
(a) Using the formula for (a, b) and G (x, y) = (2 x 7 y, 4 x + 3 y), we have
(We emphasize that the coefficients of u1 and u2 are written as columns, not rows, in the matrix representation.)
(b) Use the formula (a, b) = (−5a + 2b)u1 +(3a − b) u2 to get
Then
Accordingly,
(This is expected from Theorem 6.1.)
6.3. Consider the following 2 × 2 matrix A and basis S of R2:
The matrix A defines a linear operator on R2. Find the matrix B that represents the mapping A relative to the basis S.
First find the coordinates of an arbitrary vector (a, b)T with respect to the basis S. We have
Solve for x and y in terms of a and b to obtain x = 7 a + 3 b, y = 2 a b. Thus,
Then use the formula for (a, b)T to find the coordinates of Au1 and Au2 relative to the basis S:
Writing the coordinates as columns yields
6.4. Find the matrix representation of each of the following linear operators F on R3 relative to the usual basis E = {e1, e2, e3} of R3; that is, find [F] = [F]E:
(a) F defined by F (x, y, z) = (x + 2y −3 z, 4x − 5y 6z, 7x + 8y + 9z).
(b) F defined by the 3 × 3 matrix
(c) F defined by F (e1) = (1, 3, 5), F (e2) = (2, 4, 6), F (e3) = (7, 7, 7). (Theorem 5.2 states that a linear map is completely defined by its action on the vectors in a basis.)
(a) Because E is the usual basis, simply write the coefficients of the components of F (x, y, z) as rows:
(b) Because E is the usual basis, [F] = A, the matrix A itself.
(c) Here
That is, the columns of [F] are the images of the usual basis vectors.
6.5. Let G be the linear operator on R3 defined by G (x, y, z) = (2 y + z, x 4 y, 3 x).
(a) Find the matrix representation of G relative to the basis
(b) Verify that [G [υ = [G (υ) for any vector υ in R3.
First find the coordinates of an arbitrary vector (a, b, c) ∈ R3 with respect to the basis S. Write (a, b, c) as a linear combination of w1, w2, w3 using unknown scalars x, y, and z:
Set corresponding components equal to each other to obtain the system of equations
Solve the system for x, y, z in terms of a, b, c to find x = c, y = b c, z = a b. Thus,
(a) Because G (x, y, z) = (2y + z, x 4y, 3 x),
Write the coordinates G(w1), G (w2), G (w3) as columns to get
(b) Write G (υ) as a linear combination of w1, w2, w3, where υ = (a, b, c) is an arbitrary vector in R3,
or equivalently,
Accordingly,
6.6. Consider the following 3 × 3 matrix A and basis S of R 3:
The matrix A defines a linear operator on R 3. Find the matrix B that represents the mapping A relative to the basis S. (Recall that A represents itself relative to the usual basis of R 3.)
First find the coordinates of an arbitrary vector (a, b, c) in R3 with respect to the basis S. We have
Solve for x, y, z in terms of a, b, c to get
thus,
Then use the formula for (a, b, c)T to find the coordinates of Au1, Au2, Au3 relative to the basis S:
6.7. For each of the following linear transformations (operators) L on R 2, find the matrix A that represents L (relative to the usual basis of R 2):
(a) L is defined by L (1, 0) = (2, 4) and L (0, 1) = (5, 8).
(b) L is the rotation in R 2 counterclockwise by 90°.
(c) L is the reflection in R 2 about the line y = x.
(a) Because {(1, 0), (0, 1)} is the usual basis of R2, write their images under L as columns to get
(b) Under the rotation L, we have L (1, 0) = (0, 1) and L (0, 1) = (1, 0). Thus,
(c) Under the reflection L, we have L (1, 0) = (0, −1) and L (0, 1) = (−1, 0). Thus,
6.8. The set S = {e3t, te3t, t2 e3t} is a basis of a vector space V of functions f : R → R. Let D be the differential operator on V; that is, D (f) = df = dt. Find the matrix representation of D relative to the basis S.
Find the image of each basis function:
6.9. Prove Theorem 6.1: Let T: V ! V be a linear operator, and let S be a (finite) basis of V. Then, for any vector υ in V, [T S[υ S = [T (υ)S.
Suppose S = {u1, u2,..., un}, and suppose, for i = 1,..., n,
Then [T]S is the n-square matrix whose jth row is
Now suppose
Writing a column vector as the transpose of a row vector, we have
Furthermore, using the linearity of T,
Thus, [T (υ)S is the column vector whose jth entry is
On the other hand, the jth entry of [T]S[υ] S is obtained by multiplying the jth row of [T S by [υ S—that is (1) by (2). But the product of (1) and (2) is (3). Hence, [T S[υ S and [T (υ)S have the same entries. Thus, [T S[υ] S = [T (υ)S .
6.10. Prove Theorem 6.2: Let S = {u1, u2,..., ung be a basis for V over K, and let M be the algebra of n-square matrices over K. Then the mapping m: A (V) → M defined by m (T) = [T]S is a vector space isomorphism. That is, for any F, G ∈ A (V) and any k 2 K, we have
(i) [F + G = [F +[G],
(ii) [kF = k[F],
(iii) m is one-to-one and onto.
(i) Suppose, for i = 1,..., n,
Consider the matrices A = [aij] and B = [bij]. Then [F] = AT and [G] = BT. We have, for i = 1,..., n,
Because A + B is the matrix (aij + bij), we have
(ii) Also, for i = 1,..., n,
Because kA is the matrix (kaij), we have
(iii) Finally, m is one-to-one, because a linear mapping is completely determined by its values on a basis. Also, m is onto, because matrix A = [aij in M is the image of the linear operator,
Thus, the theorem is proved.
6.11. Prove Theorem 6.3: For any linear operators G, F ∈ A (V), [G F = [G [F].
Using the notation in Problem 6.10, we have
Recall that AB is the matrix AB = [cik, where . Accordingly,
The theorem is proved.
6.12. Let A be the matrix representation of a linear operator T. Prove that, for any polynomial f (t), we have that f (A) is the matrix representation of f (T). [Thus, f (T) = 0 if and only if f (A) = 0.]
Let φ be the mapping that sends an operator T into its matrix representation A. We need to prove that φ (f (T)) = f (A). Suppose f (t) = an tn + + a1 t + a0. The proof is by induction on n, the degree of f (t). Suppose n = 0. Recall that φ (I′) = I, where I′ is the identity mapping and I is the identity matrix. Thus,
and so the theorem holds for n = 0.
Now assume the theorem holds for polynomials of degree less than n. Then, because φ is an algebra isomorphism,
and the theorem is proved.
The coordinate vector [υ S] in this section will always denote a column vector; that is,
6.13. Consider the following bases of R2:
(a) Find the change-of-basis matrix P from the usual basis E to S.
(b) Find the change-of-basis matrix Q from S back to E.
(c) Find the coordinate vector [υ of υ = (5, 3) relative to S.
(a) Because E is the usual basis, simply write the basis vectors in S as columns:
(b) Method 1. Use the definition of the change-of-basis matrix. That is, express each vector in E as a linear combination of the vectors in S. We do this by first finding the coordinates of an arbitrary vector υ = (a, b)relative to S. We have
Solve for x and y to obtain x = 4a b, y = 3 a + b. Thus,
Using the above formula for [υ]S and writing the coordinates of the ei as columns yields
Method 2. Because Q = P-1, find P-1, say by using the formula for the inverse of a 2 × 2 matrix.
Thus,
(c) Method 1. Write υ as a linear combination of the vectors in S, say by using the above formula for υ = (a, b). We have υ = (5, 3) = 23u1 − 18 u2, and so [υ]S = [23, −18]T.
Method 2. Use, from Theorem 6.6, the fact that [υ]S = P -1[v] E and the fact that [υ]E = [5, −3]T:
6.14. The vectors u1 = (1, 2, 0), u2 = (1, 3, 2), u3 = (0, 1, 3) form a basis S of R 3. Find
(a) The change-of-basis matrix P from the usual basis E = {e1, e2, e3} to S.
(b) The change-of-basis matrix Q from S back to E.
(a) Because E is the usual basis, simply write the basis vectors of S as columns:
(b) Method 1. Express each basis vector of E as a linear combination of the basis vectors of S by first finding the coordinates of an arbitrary vector υ = (a, b, c) relative to the basis S. We have
Solve for x, y, z to get x = 7a − 3b + c, y = 6a + 3b c, z = 4a −2 b + c. Thus,
or
Using the above formula for [υ]S and then writing the coordinates of the ei as columns yields
Method 2. Find P-1 by row reducing M = [P, I] to the form [I, P-1:
Thus,
6.15. Suppose the x-axis and y-axis in the plane R 2 are rotated counterclockwise 45° so that the new x0-axis and y0-axis are along the line y = x and the line y = x, respectively.
(a) Find the change-of-basis matrix P.
(b) Find the coordinates of the point A (5, 6) under the given rotation.
(a) The unit vectors in the direction of the new x0- and y0-axes are
(The unit vectors in the direction of the original x and y axes are the usual basis of R2.) Thus, write the coordinates of u1 and u2 as columns to obtain
(b) Multiply the coordinates of the point by P-1:
(Because P is orthogonal, P-1 is simply the transpose of P.)
6.16. The vectors u1 = (1, 1, 0), u2 = (0, 1, 1), u3 = (1, 2, 2) form a basis S of R 3. Find the coordinates of an arbitrary vector υ = (a, b, c) relative to the basis S.
Method 1. Express υ as a linear combination of u1, u2, u3 using unknowns x, y, z. We have
this yields the system
Solving by back-substitution yields x = b c, y = 2 a + 2 b c, z = a b + c. Thus,
Method 2. Find P-1 by row reducing M = [P, I to the form [I, P-1, where P is the change-of-basis matrix from the usual basis E to S or, in other words, the matrix whose columns are the basis vectors of S.
We have
Thus,
6.17. Consider the following bases of R 2:
(a) Find the coordinates of υ = (a, b) relative to the basis S.
(b) Find the change-of-basis matrix P from S to S′.
(c) Find the coordinates of υ = (a, b) relative to the basis S′.
(d) Find the change-of-basis matrix Q from S′ back to S.
(e) Verify Q = P-1
(f) Show that, for any vector υ = (a, b) in R2, P-1[υ S = [υ S0. (See Theorem 6.6.)
(a) Let υ = xu1 + yu2 for unknowns x and y; that is,
Solve for x and y in terms of a and b to get x = 2 a 32 b and y = a + 12 b. Thus,
(b) Use part (a) to write each of the basis vectors υ 1 and υ 2 of S′ as a linear combination of the basis vectors u1 and u2 of S; that is,
Then P is the matrix whose columns are the coordinates of υ 1 and υ 2 relative to the basis S; that is,
(c) Let υ = xυ1 + yυ2 for unknown scalars x and y:
Solve for x and y to get x = 8 a + 3 b and y = 3 a b. Thus,
(d) Use part (c) to express each of the basis vectors u1 and u2 of S as a linear combination of the basis vectors υ1 and υ 2 of S′:
Write the coordinates of u1 and u2 relative to S′ as columns to obtain
(f) Use parts (a), (c), and (d) to obtain
6.18. Suppose P is the change-of-basis matrix from a basis {uig to a basis {wig, and suppose Q is the change-of-basis matrix from the basis {wig back to {uig. Prove that P is invertible and that Q = P-1
Suppose, for i = 1, 2,..., n, that
and, for j = 1, 2,..., n,
Let A = [aij and B = [bjk. Then P = AT and Q = BT. Substituting (2) into (1) yields
Because {wi} is a basis, Σ aij bjk = δik, where δik is the Kronecker delta; that is, δik = 1if i = k but δik = 0if i ≠ k. Suppose AB = [cik. Then cik = δik. Accordingly, AB = I, and so QP = BT AT = (AB)T = IT = I
Thus, Q = P-1
6.19. Consider a finite sequence of vectors S = {u1, u2,..., un}. Let S′ be the sequence of vectors obtained from S by one of the following “elementary operations”:
(1) Interchange two vectors.
(2) Multiply a vector by a nonzero scalar.
(3) Add a multiple of one vector to another vector.
Show that S and S′ span the same subspace W. Also, show that S′ is linearly independent if and only if S is linearly independent.
Observe that, for each operation, the vectors S′ are linear combinations of vectors in S. Also, because each operation has an inverse of the same type, each vector in S is a linear combination of vectors in S′. Thus, S and S′ span the same subspace W. Moreover, S′ is linearly independent if and only if dim W = n, and this is true if and only if S is linearly independent.
6.20. Let A = [aij] and B = [bij] be row equivalent m × n matrices over a field K, and let υ1, υ2,..., υn be any vectors in a vector space V over K. For i = 1, 2,..., m, let ui and wi be defined by
Show that {ui} and {wi} span the same subspace of V.
Applying an “elementary operation” of Problem 6.19 to {ui} is equivalent to applying an elementary row operation to the matrix A. Because A and B are row equivalent, B can be obtained from A by a sequence of elementary row operations. Hence, {wi} can be obtained from {ui} by the corresponding sequence of operations. Accordingly, {ui} and {wi} span the same space.
6.21. Suppose u1, u2,..., un belong to a vector space V over a field K, and suppose P = [aij] is an n-square matrix over K. For i = 1, 2,..., n, let υi = ai1 u1 + ai2 u2 + … + ain un.
(a) Suppose P is invertible. Show that {ui} and {υi} span the same subspace of V. Hence, {ui} is linearly independent if and only if {υi} is linearly independent.
(b) Suppose P is singular (not invertible). Show that {υi} is linearly dependent.
(c) Suppose {υi} is linearly independent. Show that P is invertible.
(a) Because P is invertible, it is row equivalent to the identity matrix I. Hence, by Problem 6.19, {υi} and {ui} span the same subspace of V. Thus, one is linearly independent if and only if the other is linearly independent.
(b) Because P is not invertible, it is row equivalent to a matrix with a zero row. This means {υi} spans a subspace that has a spanning set with less than n elements. Thus, {υi} is linearly dependent.
(c) This is the contrapositive of the statement of part (b), and so it follows from part (b).
6.22. Prove Theorem 6.6: Let P be the change-of-basis matrix from a basis S to a basis S′ in a vector space V. Then, for any vector υ ∈ V, we have P[υS′ = [υS, and hence, P-1[vS = [υS′.
Suppose S = {u1,..., un} and S′ = {w1,..., wn}, and suppose, for i = 1,..., n,
Then P is the n-square matrix whose jth row is
Also suppose . Then
Substituting for wi in the equation for υ, we obtain
Accordingly, [υ]S is the column vector whose jth entry is
On the other hand, the jth entry of P[υ]S′ is obtained by multiplying the jth row of P by [υ]S′—that is, (1) by (2). However, the product of (1) and (2) is (3). Hence, P[υS′ and [υ]S have the same entries. Thus, P[υ]S′ = [υ]S, as claimed.
Furthermore, multiplying the above by P–1 gives .
6.23. Consider the linear transformation F on R2 defined by F (x, y) = (5x – y, 2x + y) and the following bases of R2:
(a) Find the change-of-basis matrix P from E to S and the change-of-basis matrix Q from S back to E.
(b) Find the matrix A that represents F in the basis E.
(c) Find the matrix B that represents F in the basis S.
(a) Because E is the usual basis, simply write the vectors in S as columns to obtain the change-of-basis matrix P. Recall, also, that Q = P–-. Thus,
(b) Write the coefficients of x and y in F(x, y) = (5x – y, 2x + y) as rows to get
(c) Method 1. Find the coordinates of F (u1) and F(u2) relative to the basis S. This may be done by first finding the coordinates of an arbitrary vector (a, b) in R2 relative to the basis S. We have
Solve for x and y in terms of a and b to get x = –7 a + 2 b, y = 4 a – b. Then
Now use the formula for (a, b) to obtain
Method 2. By Theorem 6.7, B = P–1 AP. Thus,
6.24. Let . Find the matrix B that represents the linear operator A relative to the basis
. [Recall A defines a linear operator A: R2 → R2 relative to the usual basis E of R2].
Method 1. Find the coordinates of A (u1) and A (u2) relative to the basis S by first finding the coordinates of an arbitrary vector [a, b T in R 2 relative to the basis S. By Problem 6.2,
Using the formula for [a, b]T, we obtain
and
Thus,
Method 2. Use B = P–1 AP, where P is the change-of-basis matrix from the usual basis E to S. Thus, simply write the vectors in S (as columns) to obtain the change-of-basis matrix P and then use the formula
for P–1. This gives
Then
6.25. Let . Find the matrix B that represents the linear operator A relative to the basis
[Recall A that defines a linear operator A: R3 → R3 relative to the usual basis E of R3.]
Method 1. Find the coordinates of A(u1), A(u2), A(u3) relative to the basis S by first finding the coordinates of an arbitrary vector υ = (a, b, c) in R3 relative to the basis S. By Problem 6.16,
Using this formula for [a, b, c]T, we obtain
Writing the coefficients of u1, u2, u3 as columns yields
Method 2. Use B = P–1 AP, where P is the change-of-basis matrix from the usual basis E to S. The matrix P (whose columns are simply the vectors in S) and P–1 appear in Problem 6.16. Thus,
6.26. Prove Theorem 6.7: Let P be the change-of-basis matrix from a basis S to a basis S′ in a vector space V. Then, for any linear operator T on V, [T S′ = P–1 [T]SP.
Let υ be a vector in V. Then, by Theorem 6.6, P[υS′ = [υ]S. Therefore,
But [TS′ [υS′ = [T (υ)\S′. Hence,
Because the mapping . Thus, P–1 [T]SP = [T]S′, as claimed.
(a) First find P–1 using the formula for the inverse of a 2 × 2 matrix. We have
Then
(b) tr (A) = 4 + 6 = 10 and tr (B) = 25 15 = 10. Hence, tr (B) = tr (A).
(c) det (A) = 24 + 6 = 30 and det (B) = 375 + 405 = 30. Hence, det (B) = det (A).
6.28. Find the trace of each of the linear transformations F on R3 in Problem 6.4.
Find the trace (sum of the diagonal elements) of any matrix representation of F such as the matrix representation [F] = [F]E of F relative to the usual basis E given in Problem 6.4.
(a) tr (F) = tr([F) = 1 – 5 + 9 = 5.
(b) tr (F) = tr([F) = 1 + 3 + 5 = 9.
(c) tr (F) = tr([F) = 1 + 4 + 7 = 12.
6.29. Write A ≈ B if A is similar to B—that is, if there exists an invertible matrix P such that A = P–1 BP. Prove that ≈ is an equivalence relation (on square matrices); that is,
(a) A ≈ A, for every A.
(b) If A ≈ B, then B ≈ A.
(c) If A ≈ B and B ≈ C, then A ≈ C.
(a) The identity matrix I is invertible, and I–1 = I. Because A = I–1 AI, we have A ≈ A.
(b) Because A ≈ B, there exists an invertible matrix P such that A = P–1 BP. Hence, and P–1 is also invertible. Thus, B ≈ A.
(c) Because A ≈ B, there exists an invertible matrix P such that A = P–1 BP, and as B ≈ C, there exists an invertible matrix Q such that B = Q–1 CQ. Thus,
and QP is also invertible. Thus, A ≈ C.
6.30. Suppose B is similar to A, say B = P–1 AP. Prove
(a) Bn = P–1 An P, and so Bn is similar to An.
(b) f (B) = P–1 AnP, for any polynomial f(x), and so f(B) is similar to f (A).
(c) B is a root of a polynomial g(x) if and only if A is a root of g(x).
(a) The proof is by induction on n. The result holds for n = 1 by hypothesis. Suppose n > 1 and the result holds for n – 1. Then
(b) Suppose . Using the left and right distributive laws and part (a), we have
(c) By part (b), g(B) = 0 if and only if P–1 g(A) P = 0 if and only if g(A) = P0 P–1 = 0.
6.31. Let F: R3 → R2 be the linear map defined by F (x, y, z) = (3x + 2y – 4z, x 5y + 3z).
(a) Find the matrix of F in the following bases of R3 and R2:
(b) Verify Theorem 6.10: The action of F is preserved by its matrix representation; that is, for any v in R3, we have .
(a) From Problem 6.2, . Thus,
Write the coordinates of F(w1), F(w2), F(w3) as columns to get
(b) If υ = (x, y, z), then, by Problem 6.5, υ = zw1 + (y – z) w2 + (x – y) w3 . Also,
6.32. Let F: Rn ! Rm be the linear mapping defined as follows:
(a) Show that the rows of the matrix [F representing F relative to the usual bases of Rn and Rm are the coefficients of the xi in the components of F(x1,..., xn).
(b) Find the matrix representation of each of the following linear mappings relative to the usual basis of Rn:
(a) We have
(b) By part (a), we need only look at the coefficients of the unknown x, y,... in F(x, y,...). Thus,
6.33. Let . Recall that A determines a mapping F: R3 → R2 defined by F(v) = Av, where vectors are written as columns. Find the matrix [F] that represents the mapping relative to the following bases of R3 and R2:
(a) The usual bases of R3 and of R2.
(b) S = {w1, w2, w3} = {(1, 1, 1), (1, 1, 0), (1, 0, 0)} and S′ = {u1, u2} = f (1, 3), (2, 5)}.
(a) Relative to the usual bases, [F is the matrix A.
(b) From Problem 6.2, (a, b) = (–5a + 2b) u1 + (3a – b) u2. Thus,
Writing the coefficients of F(w1), F(w2), F(w3) as columns yields .
6.34. Consider the linear transformation T on R2 defined by T(x, y) = (2x – 3y, x + 4y) and the following bases of R2:
(a) Find the matrix A representing T relative to the bases E and S.
(b) Find the matrix B representing T relative to the bases S and E.
(We can view T as a linear mapping from one space into another, each having its own basis.)
(a) From Problem 6.2, (a, b) = (–5a + 2b) u1 + (3a – b)u2. Hence,
(b) We have
6.35. How are the matrices A and B in Problem 6.34 related?
By Theorem 6.12, the matrices A and B are equivalent to each other; that is, there exist nonsingular matrices P and Q such that B = Q–1 AP, where P is the change-of-basis matrix from S to E, and Q is the change-of-basis matrix from E to S. Thus,
and
6.36. Prove Theorem 6.14: Let F: V → U be linear and, say, rank (F) = r. Then there exist bases V and of U such that the matrix representation of F has the following form, where Ir is the r-square identity matrix:
Suppose dim V = m and dim U = n. Let W be the kernel of F and U′ the image of F. We are given that rank (F) = r. Hence, the dimension of the kernel of F is m – r. Let {w1,..., wm–r} be a basis of the kernel of F and extend this to a basis of V:
Set
Then {u1,..., ur} is a basis of U′, the image of F. Extend this to a basis of U, say
Observe that
Thus, the matrix of F in the above bases has the required form.
SUPPLEMENTARY PROBLEMS
6.37. Let F: R2 → R2 be defined by F(x, y) = (4x + 5y, 2x y).
(a) Find the matrix A representing F in the usual basis E.
(b) Find the matrix B representing F in the basis S = {u1, u2} = {(1, 4), (2, 9)}.
(c) Find P such that B = P–1 AP.
(d) For υ = (a, b), find [υ]S and [F (υ)]S. Verify that [F]S[υ]S = [F(υ)S.
6.38. Let A: R2 → R2 be defined by the matrix
(a) Find the matrix B representing A relative to the basis S = {u1, u2} = f(1, 3), (2, 8)}. (Recall that A represents the mapping A relative to the usual basis E.)
(b) For υ = (a, b), find [υ]S and [A(υ)S.
6.39. For each linear transformation L on R2, find the matrix A representing L (relative to the usual basis of R2):
(a) L is the rotation in R2 counterclockwise by 45°.
(b) L is the reflection in R2 about the line y = x.
(c) L is defined by L (1, 0) = (3, 5) and L (0, 1) = (7, – 2).
(d) L is defined by L (1, 1) = (3, 7) and L (1, 2) = (5, –4).
6.40. Find the matrix representing each linear transformation T on R3 relative to the usual basis of R3:
(a) T (x, y, z) = (x, y, 0).
(b) T (x, y, z) = (z, y + z, x + y + z).
(c) T (x, y, z) = (2x – 7y 4z, 3x + y + 4z, 6x – 8y + z).
6.41. Repeat Problem 6.40 using the basis S = {u1, u2, u3} = {(1, 1, 0), (1, 2, 3), (1, 3, 5)}.
6.42. Let L be the linear transformation on R3 defined by
(a) Find the matrix A representing L relative to the usual basis of R3.
(b) Find the matrix B representing L relative to the basis S in Problem 6.41.
6.43. Let D denote the differential operator; that is, D(f(t)) = df/dt. Each of the following sets is a basis of a vector space V of functions. Find the matrix representing D in each basis:
6.44. Let D denote the differential operator on the vector space V of functions with basis S = {sin θ, cos θ}.
(a) Find the matrix A = [D]S.
(b) Use A to show that D is a zero of f(t) = t2 + 1.
6.45. Let V be the vector space of 2 ′ 2 matrices. Consider the following matrix M and usual basis E of V:
Find the matrix representing each of the following linear operators T on V relative to E:
(a) T (A) = MA.
(b) T (A) = AM.
(c) T (A) = MA – AM.
6.46. Let 1V and 0V denote the identity and zero operators, respectively, on a vector space V. Show that, for any basis S of V, (a) [1V]S = I, the identity matrix. (b) [0V]S = 0, the zero matrix.
6.47. Find the change-of-basis matrix P from the usual basis E of R2 to a basis S, the change-of-basis matrix Q from S back to E, and the coordinates of υ = (a, b) relative to S, for the following bases S:
(a) S = {(1, 2), (3, 5)}.
(b) S = {(1, – 3), (3, – 8)}.
(c) S = {(2, 5), (3, 7)}.
(d) S = {(2, 3), (4, 5)}.
6.48. Consider the bases S = {(1, 2), (2, 3)g and S′ = {(1, 3), (1, 4)g of R2. Find the change-of-basis matrix:
(a) P from S to S′.
(b) Q from S′ back to S.
6.49. Suppose that the x-axis and y-axis in the plane R2 are rotated counterclockwise 30° to yield new x′-axis and y 0-axis for the plane. Find
(a) The unit vectors in the direction of the new x′-axis and y′-axis.
(b) The change-of-basis matrix P for the new coordinate system.
(c) The new coordinates of the points A (1, 3), B (2, 5), C (a, b).
6.50. Find the change-of-basis matrix P from the usual basis E of R3 to a basis S, the change-of-basis matrix Q from S back to E, and the coordinates of υ = (a, b, c) relative to S, where S consists of the vectors:
(a) u1 = (1, 1, 0), u2 = (0, 1, 2), u3 = (0, 1, 1).
(b) u1 = (1, 0, 1), u2 = (1, 1, 2), u3 = (1, 2, 4).
(c) u1 = (1, 2, 1), u2 = (1, 3, 4), u3 = (2, 5, 6).
6.51. Suppose S1, S2, S3 are bases of V. Let P and Q be the change-of-basis matrices, respectively, from S1 to S2 and from S2 to S3. Prove that PQ is the change-of-basis matrix from S1 to S3.
6.52. Consider the linear operator F on R2 defined by F(x, y) = (5x + y, 3x – 2y) and the following bases of R2:
(a) Find the matrix A representing F relative to the basis S.
(b) Find the matrix B representing F relative to the basis S′.
(c) Find the change-of-basis matrix P from S to S′.
(d) How are A and B related?
6.53. Let A: R2 → R2 be defined by the matrix . Find the matrix B that represents the linear operator
A relative to each of the following bases: (a) S = {(1, 3)T, (2, 5)T}. (b) S = {(1, 3)T, (2, 4)T}.
6.54. Let F: R2 → R2 be defined by F(x, y) = (x − 3y, 2x − 4y). Find the matrix A that represents F relative to each of the following bases: (a) S = {(2, 5), (3, 7)}. (b) S = {(2, 3), (4, 5)}.
6.55. Let A: R3 → R3 be defined by the matrix . Find the matrix B that represents the linear operator A relative to the basis S = {(1, 1, 1)T, (0, 1, 1)T, (1, 2, 3)T}.
(a) Find B = P−1AP.
(b) Verify that tr(B) = tr(A).
(c) Verify that det(B) = det(A).
6.57. Find the trace and determinant of each of the following linear maps on R2:
(a) F(x, y) = (2x − 3y, 5x + 4y).
(b) G(x, y) = (ax + by, cx + dy).
6.58. Find the trace and determinant of each of the following linear maps on R3:
(a) F(x, y, z) = (x + 3y, 3x − 2z, x − 4y − 3z).
(b) G(x, y, z) = (y + 3z, 2x − 4z, 5x + 7y).
6.59. Suppose S = {u1, u2} is a basis of V, and T: V → V is defined by T(u1) = 3u1 − 2u2 and T(u2) = u1 + 4u2. Suppose S′ = {w1, w2} is a basis of V for which w1 = u1 + u2 and w2 = 2u1 + 3u2.
(a) Find the matrices A and B representing T relative to the bases S and S′, respectively.
(b) Find the matrix P such that B = P−1AP.
6.60. Let A be a 2 × 2 matrix such that only A is similar to itself. Show that A is a scalar matrix, that is, that .
6.61. Show that all matrices similar to an invertible matrix are invertible. More generally, show that similar matrices have the same rank.
6.62. Find the matrix representation of each of the following linear maps relative to the usual basis for Rn:
(a) F: R3 → R2 defined by F(x, y, z) = (2x − 4y + 9z, 5x + 3y − 2z).
(b) F: R2 → R4 defined by F(x, y) = (3x + 4y, 5x − 2y, x + 7y, 4x).
(c) F: R4 → R defined by F(x1, x2, x3, x4) = 2x1 + x2 − 7x3 − x4.
6.63. Let G: R3 → R2 be defined by G(x, y, z) = (2x + 3y z, 4x y + 2z).
(a) Find the matrix A representing G relative to the bases
(b) For any υ = (a, b, c) in R3, find [υ]S and [G(υ)]S′.
(c) Verify that A[υ]S = [G(υ)]S′.
6.64. Let H: R2 → R2 be defined by H(x, y) = (2x + 7y, x − 3y) and consider the following bases of R2:
(a) Find the matrix A representing H relative to the bases S and S′.
(b) Find the matrix B representing H relative to the bases S′ and S.
6.65. Let F: R3 → R2 be defined by F(x, y, z) = (2x + y z, 3x − 2y + 4z).
(a) Find the matrix A representing F relative to the bases
(b) Verify that, for any υ = (a, b, c) in R3, A[υ]S = [F(υ)]S′.
6.66. Let S and S′ be bases of V,and let 1V be the identity mapping on V. Show that the matrix A representing 1V relative to the bases S and S′ is the inverse of the change-of-basis matrix P from S to S′; that is, A = P−1.
6.67. Prove (a) Theorem 6.10, (b) Theorem 6.11, (c) Theorem 6.12, (d) Theorem 6.13. [Hint: See the proofs of the analogous Theorems 6.1 (Problem 6.9), 6.2 (Problem 6.10), 6.3 (Problem 6.11), and 6.7 (Problem 6.26).]
6.68. Suppose F: V → V is linear. A subspace W of V is said to be invariant under F if F(W) ⊆ W. Suppose W is invariant under F and dim W = r. Show that F has a block triangular matrix representation where A is an r × r submatrix.
6.69. Suppose V = U + W, and suppose U and V are each invariant under a linear operator F: V → V. Also, suppose dim U = r and dim W = S. Show that F has a block diagonal matrix representation where A and B are r × r and s × s submatrices.
6.70. Two linear operators F and G on V are said to be similar if there exists an invertible linear operator T on V such that G = T−1 ∘ F ∘ T. Prove
(a) F and G are similar if and only if, for any basis S of V, [F]S and [G]S are similar matrices.
(b) If F is diagonalizable (similar to a diagonal matrix), then any similar matrix G is also diagonalizable.
ANSWERS TO SUPPLEMENTARY PROBLEMS
Notation: M = [R1; R2; …] represents a matrix M with rows R1, R2, ….
6.37. (a) A = [4, 5; 2, −1; (b) B = [220, 487; −98, −217]; (c) P = [1, 2; 4, 9]; (d) [υ]S = [9a − 2b, −4a + b]T and [F(υ)]S = [32a + 47b, −14a − 21b]T
6.40. (a) [1, 0, 0; 0, 1, 0; 0, 0, 0]; (b) [0, 0, 1; 0, 1, 1; 1, 1, 1];
(c) [2, −7, −4; 3, 1, 4; 6, −8, 1]
6.41. (a) [1, 3, 5; 0, −5, −10; 0, 3, 6]; (b) [0, 1, 2; 1, 2, 3; 1, 0, 0];
(c) [15, 65, 104; 49, −219, −351; 29, 130, 208]
6.42. (a) [1, 1, 2; 1, 3, 2; 1, 5, 2]; (b) [0; 2, 14, 22; 0, −5, −8]
6.43. (a) [1, 0, 0; 0, 2, 1; 0, 0, 2]; (b) [0, 1, 0, 0; 0; 0, 0, 0, −3; 0, 0, 3, 0]; (c) [5, 1, 0; 0, 5, 2; 0, 0, 5]
6.44. (a) A = [0, −1; 1, 0;] (b) A2 + I = 0
6.45. (a) [a, 0, b, 0; 0, a, 0, b; c, 0, d, 0; 0, c, 0, d];
(b) [a, c, 0, 0; b, d, 0, 0; 0, 0, a, c; 0, 0, b, d];
(c) [0, −c, b, 0; −b, a − d, 0, b; c, 0, d − a, −c; 0, c, −b, 0]
6.48. (a) P = [3, 5; 1, −2 ×]; (b) Q = [2, 5; 1, −3]
6.50. P is the matrix whose columns are u1, u2, u3, Q = P−1, [υ] = Q[a, b, cT:
6.52. (a) [23, −39; 15, 26]; (b) [35, 41; 27, −32]; (c) [3, 5; 1, −2]; (d) B = P−1AP
6.55. [10, 8, 20; 13, 11, 28; 5, −4, −10]
6.56. (a) [34, 57; 19, 32]; (b) tr(B) = tr(A) = −2; (c) det(B) = det(A) = −5
6.57. (a) tr(F) = 6, det(F) = 23; (b) tr(G) = a + d, det(G) = ad − bc
6.58. (a) tr(F) = −2, det(F) = 13; (b) tr(G) = 0, det(G) = 22
6.59. (a) A = [3, 1; −2, 4], B = [8, 11; 2, −1]; (b) P = [1, 2; 1, 3]
6.62. (a) [2, −4, 9; 5, 3, −2 ×]; (b) [3, 5, 1, 4; 4, −2, 7, 0]; (c) [2, 1, −7, −1]
6.63. (a) [9, 1, 4; 7, 2, 1]; (b) [υ]S = [−a + 2b − c, 5a − 5b + 2c, −3a + 3b − c]T, and [G(υ)]S′ = [2a − 11b + 7c, 7b − 4c]T