Skip to main content

Section 5.4 Direct Sums and Invariant Subspaces

This section continues the discussion of direct sums (from Section 1.8) and invariant subspaces (from Section 4.1), to better understand the structure of linear operators.

Subsection 5.4.1 Invariant subspaces

Definition 5.4.1.

Given an operator T:Vโ†’V, we say that a subspace UโІV is T-invariant if T(u)โˆˆU for all uโˆˆU.
Given a basis B={u1,u2,โ€ฆ,uk} of U, note that U is T-invariant if and only if T(ui)โˆˆU for each i=1,2,โ€ฆ,k.
For any operator T:Vโ†’V, there are four subspaces that are always T-invariant:
{0},V,kerโกT, and imT.
Of course, some of these subspaces might be the same; for example, if T is invertible, then kerโกT={0} and imT=V.

Exercise 5.4.2.

Show that for any linear operator T, the subspaces kerโกT and imT are T-invariant.
Hint.
In each case, choose an element v of the subspace. What does the definition of the space tell you about that element? (For example, if vโˆˆkerโกT, what is the value of T(v)?) Then show that T(v) also fits the defintion of that space.
A subspace U is T-invariant if T does not map any vectors in U outside of U. Notice that if we shrink the domain of T to U, then we get an operator from U to U, since the image T(U) is contained in U.

Definition 5.4.3.

Let T:Vโ†’V be a linear operator, and let U be a T-invariant subspace. The restriction of T to U, denoted T|U, is the operator T|U:Uโ†’U defined by T|U(u)=T(u) for all uโˆˆU.

Exercise 5.4.4.

    True or false: the restriction T|U is the same function as the operator T.
  • True.

  • The definition of a function includes its domain and codomain. Since the domain of T|U is different from that of T, they are not the same function.
  • False.

  • The definition of a function includes its domain and codomain. Since the domain of T|U is different from that of T, they are not the same function.
A lot can be learned by studying the restrictions of an operator to invariant subspaces. Indeed, the textbook by Axler does almost everything from this point of view. One reason to study invariant subspaces is that they allow us to put the matrix of T into simpler forms.
Reducing a matrix to block triangular form is useful, because it simplifies computations such as determinants and eigenvalues (and determinants and eigenvalues are computationally expensive). In particular, if a matrix A has the block form
A=[A11A12โ‹ฏA1n0A22โ‹ฏA2nโ‹ฎโ‹ฎโ‹ฑโ‹ฎ00โ‹ฏAnn],
where the diagonal blocks are square matrices, then det(A)=det(A11)det(A22)โ‹ฏdet(Ann) and cA(x)=cA11(x)cA22(x)โ‹ฏcAnn(x).

Subsection 5.4.2 Eigenspaces

An important source of invariant subspaces is eigenspaces. Recall that for any real number ฮป, and any operator T:Vโ†’V, we define
Eฮป(T)=kerโก(Tโˆ’ฮป1V)={vโˆˆV|T(v)=ฮปv}.
For most values of ฮป, weโ€™ll have Eฮป(T)={0}. The values of ฮป for which Eฮป(T) is non-trivial are precisely the eigenvalues of T. Note that since similar matrices have the same characteristic polynomial, any matrix representation MB(T) will have the same eigenvalues. They do not generally have the same eigenspaces, but we do have the following.
In other words, the two eigenspaces are isomorphic, although the isomorphism depends on a choice of basis.

Subsection 5.4.3 Direct Sums

Recall that for any subspaces U,W of a vector space V, the sets
U+W={u+w|uโˆˆU and wโˆˆW}UโˆฉW={vโˆˆV|vโˆˆU and vโˆˆW}
are subspaces of V. Saying that vโˆˆU+W means that v can be written as a sum of a vector in U and a vector in W. However, this sum may not be unique. If vโˆˆUโˆฉW, uโˆˆU and wโˆˆW, then we can write (u+v)+w=u+(v+w), giving two different representations of a vector as an element of U+W.
We proved in Theorem 1.8.9 in Section 1.8 that for any vโˆˆU+W, there exist unique vectors uโˆˆU and wโˆˆW such that v=u+w, if and only if UโˆฉW={0}.
In Definition 1.8.8, we said that a sum U+W where UโˆฉW={0} is called a direct sum, written as UโŠ•W.
Typically we are interested in the case that the two subspaces sum to V. Recall from Definition 1.8.11 that if V=UโŠ•W, we say that W is a complement of U. We also say that UโŠ•W is a direct sum decomposition of V. Of course, the orthogonal complement UโŠฅ of a subspace U is a complement in this sense, if V is equipped with an inner product. (Without an inner product we have no concept of โ€œorthogonalโ€.) But even if we donโ€™t have an inner product, finding a complement is not too difficult, as the next example shows.

Example 5.4.7. Finding a complement by extending a basis.

The easiest way to determine a direct sum decomposition (or equivalently, a complement) is through the use of a basis. Suppose U is a subspace of V with basis {e1,e2,โ€ฆ,ek}, and extend this to a basis
B={e1,โ€ฆ,ek,ek+1,โ€ฆ,en}
of V. Let W=span{ek+1,โ€ฆ,en}. Then U+W=V, since the first k vectors in B belong to U, and the remaining vectors in B belong to W. And UโˆฉW={0}, since if vโˆˆUโˆฉW, then vโˆˆU and vโˆˆW, so we have
v=a1e1+โ‹ฏ+akek=b1ek+1+โ‹ฏ+bnโˆ’ken,
which gives
a1e1+โ‹ฏ+akekโˆ’b1ek+1โˆ’โ‹ฏโˆ’bnโˆ’ken=0,
so a1=โ‹ฏbnโˆ’k=0 by the linear independence of B, showing that v=0.
Conversely, if V=UโŠ•W, and we have bases {u1,u2,โ€ฆ,uk} of U and {v1,v2,โ€ฆ,vl} of W, then
B={u1,โ€ฆ,uk,w1,โ€ฆ,wl}
is a basis for V. Indeed, B spans V, since every element of V can be written as v=u+w with uโˆˆU,wโˆˆW. Independence follows by reversing the argument above: if
a1u1+โ‹ฏ+akuk+b1w1+โ‹ฏblwl=0
then a1u1+โ‹ฏ+akuk=โˆ’b1w1โˆ’โ‹ฏโˆ’blwl, and equality is only possible if both sides belong to UโˆฉW={0}. Since {u1,u2,โ€ฆ,uk} is independent, the ai have to be zero, and since {w1,w2,โ€ฆ,wl} is independent, the bj have to be zero.
The argument given in the second part of Example 5.4.7 has an immediate, but important consequence.

Example 5.4.9.

Suppose V=UโŠ•W, where U and W are T-invariant subspaces for some operator T:Vโ†’V. Let BU={u1,u2,โ€ฆ,um} and let BW={w1,w2,โ€ฆ,wn} be bases for U and W, respectively. Determine the matrix of T with respect to the basis B=BUโˆชBW of V.
Solution.
Since we donโ€™t know the map T or anything about the bases BU,BW, weโ€™re looking for a fairly general statement here. Since U is T-invariant, we must have T(ui)โˆˆU for each i=1,โ€ฆ,m. Similarly, T(wj)โˆˆW for each j=1,โ€ฆ,n. This means that we have
T(u1)=a11u1+โ‹ฏ+am1um+0w1+โ‹ฏ+0wnโ‹ฎT(um)=a1mu1+โ‹ฏ+ammum+0w1+โ‹ฏ+0wnT(w1)=0u1+โ‹ฏ+0um+b11w1+โ‹ฏ+bn1wnโ‹ฎT(wn)=0u1+โ‹ฏ+0um+b1nw1+โ‹ฏ+bnnwn
for some scalars aij,bij. If we set A=[aij]mร—m and B=[bij]nร—n, then we have
MB(T)=[A00B].
Moreover, we can also see that A=MBU(T|U), and B=MBW(T|W).
You have attempted 1 of 3 activities on this page.