6 minutes read

Just as vectors are essential in understanding the movement and relationships within a space, linear transformations provide a framework to explore how one vector space can be transformed into another.

You're already familiar with the basics of linear transformations, but now you'll learn to manipulate them just as fluently as vectors. Also, you'll start using an effective operation that only functions have, and that behaves like the product of matrices.

The linear transformations are vectors, too!

Vectors may be abstract objects, don't you think? But so far, you've been able to work with them fluidly and get a lot of useful results thanks to their structure: addition and scalar product have given you all the flexibility you need. What's more, we often don't think about vectors individually but rather exploit these operations. It doesn't matter if you are adding vectors in the plane or matrices, in the end, the operation behaves in the same way, and you get similar results.

Well, it turns out that among the linear transformations, you can also define a sum and a product by a scalar. But remember, the sum of two vectors and their product by a scalar must also be another vector, so the result of combining the transformations must be another transformation.

Let's take two vector spaces VV and WW. From now on, denote the set of all linear transformations from VV to WW as L(V,W)L(V, W). If S,TL(V,W)S,T \in L(V,W), then define their sum, S+TS+T, as the transformation that assigns to each vector vVv \in V the vector T(v)+S(v)WT(v) + S(v) \in W. Similarly, for any scalar λ\lambda, you can define a new transformation λT\lambda T whose value in vv is simply λT(v)W\lambda T(v) \in W. In summary:

The sum and product by a scalar in L(V,W)L(V, W) are:

T+S:VW,(T+S)(v)=T(v)+S(v)T+S:\, V \rightarrow W, \quad (T+S)(v) = T(v) + S(v)λT:VW,(λT)(v)=λT(v)\lambda T:\, V \rightarrow W, \quad (\lambda T)(v) = \lambda T(v)

But here comes the litmus test: does this sum and this scalar product make L(V,W)L(V, W) a vector space? The answer is yes! And the proof is simple, the only fine point is that the set needs to have a 00. Well, this role is fulfilled by the constant function O\mathscr{O} given by O(v)=0\mathscr{O}(v) = 0 for all vVv \in V.

In particular, the set of linear operators on VV, denoted by L(V,V)L(V, V), is simply written as L(V)L(V).

Composing transformations

Functions are like processes that are applied to objects in a set, so they can be chained together to form more complex processes that connect different sets. Since linear transformations are functions, of course, you can compose them, you only need one more space, say UU. If SL(U,V)S \in L(U,V) and TL(V,W)T \in L(V,W), then their composition is:

TS:UW,(TS)(u)=T(S(u))T \circ S:U \rightarrow W, \quad (T \circ S)(u) = T(S(u))That is, first SS transforms uUu \in U into S(u)VS(u) \in V and then TT transforms the latter into T(S(u))WT(S(u)) \in W.

Composition of linear transformations

But what really matters is whether the composition preserves linearity, that is, whether TST \circ S is linear. Fortunately, the answer is also affirmative and will have invaluable consequences later. So TSL(U,W)T \circ S \in L(U,W).

Proof, TST \circ S is linear

(TS)(u1+u2)=T(S(u1+u2))=T(S(u1)+S(u2))=T(S(u1))+T(S(u2))=(TS)(u1)+(TS)(u2)=(TS)(u1+u2)\begin{align*} (T\circ S)(u_1 + u_2) &= T(S(u_1 + u_2)) \\ &= T(S(u_1) +S( u_2)) \\ &= T(S(u_1)) + T(S( u_2)) \\ &= (T \circ S)(u_1) + (T \circ S)( u_2) \\ &= (T \circ S)(u_1 + u_2) \end{align*}

(TS)(λu)=T(S(λu))=T(λS(u))=λT(S(u))=λ(TS)(u)\begin{align*} (T\circ S)(\lambda u) &= T(S(\lambda u)) \\ &= T( \lambda S(u)) \\ &= \lambda T( S(u)) \\ &= \lambda (T \circ S)(u) \end{align*}

This new operation is quite peculiar and unique, it is an independent quality of the vector space structure. In reality, it is similar to the product of matrices. Actually, you'll find out later that this relationship is much more intimate than it may seem, and incidentally it has similar properties:

  • (TS)R=T(SR)(T \circ S) \circ R = T \circ ( S \circ R)
  • IT=T=TII \circ T = T = T \circ I
  • (T1+T2)S=(T1S)+(T2S)(T_1 + T_2) \circ S = (T_1 \circ S) +( T_2 \circ S)
  • T(S1+S2)=(TS1)+(TS2)T \circ (S_1 + S_2) = (T \circ S_1 ) + (T \circ S_2)
  • T(λS)=(λT)S=λ(TS)T \circ (\lambda S ) = (\lambda T) \circ S = \lambda (T \circ S )

Looking carefully at these properties, it is impossible not to notice their resemblance to multiplication. This is why they are often also called a product. But this is only an appearance, because the composition is not commutative, since the expression ST=TSS \circ T = T \circ S doesn't even make sense. Even when T,SL(V)T,S \in L(V) it is rare for commutativity to hold, another similarity with the product of matrices.

Taking powers

Let's go further with the product. As with the multiplication of numbers, you can define the powers of an operator TL(V)T \in L(V) through composition. For example, the square of TT is the product of TT with itself, that is, the composition of TT with TT:

T2=TTT^2 = T \circ T

Larger powers are defined in a similar way, but in general, you can write them compactly as Tn=TTn1T^n = T \circ T^ {n-1}. Again in analogy with numbers, you can define the powers 00 and 11 as T0=IT^ 0 = I (the identity function) and T1=TT^1 = T. Thinking of the composition as the successive application of the same transformation, it is evident that the power complies with the elemental properties:

  • TmTn=Tm+nT^m T^n = T^{m + n}
  • (Tm)n=Tmn(T^m)^n = T^{mn}

But the functions are not numbers, and the composition is not a product, therefore it has other properties. Sometimes after applying an operator several times, it stops having any effect. These types of operators are called idempotent, and the simplest example is the projection on a line.

Take a vector uu in the plane, after projecting it with a transformation PL(R2)P \in L(\mathbb{R}^2) on a line, no matter how many more times you project it again, the result doesn't change.

Projection on a line

The geometry of transformations

Since linear transformations are vectors, let's start treating them as such with some examples. Let's go to the plane and take a vector u=(x,y)u=(x,y), you are going to investigate the geometry of the projections in the coordinate axes P,QL(R2)P, Q \in L(\mathbb{R}^2) given by:

P(x,y)=(x,0)Q(x,y)=(0,y)P(x,y)=(x,0) \qquad Q(x,y)=(0,y)

The projections on the axes

Let's start combining the projections. Just by adding them, you have that (P+Q)(u)=P(u)+Q(u)=(x,0)+(0,y)=(x,y)=u=I(u)(P+Q) (u) = P(u) +Q(u) = (x,0) + (0,y) = (x,y) = u = I(u) and so P+Q=IP+Q =I.

The sum of the projections is the identity

If you now scale PP by a scalar, say 55, then you get that (5P)(u)=5P(u)=5(x,0)=(5x,0)(5P)(u) = 5P(u) = 5(x,0)=(5x,0), which means that the result is simply a stretch of the projection.

A curious combination is PQP-Q. Note that (PQ)(u)=P(u)Q(u)=(x,0)(0,y)=(x,y)(P-Q)(u) = P(u)-Q(u) = (x,0)-(0,y)=(x,-y) and therefore PQP-Q is the transformation that reflects each vector about the horizontal axis. What geometric meaning do you imagine that QPQ-P have?

A reflexion respect to the horizontal axis

Since PP and PP are projections, they are idempotent. For example, P2(u)=P(P(u))=P(x,0)=(x,0)P^2(u) = P(P(u)) = P(x,0)=(x,0), so any power of PP is equal to PP, you can verify that the same thing happens with QQ.

It's time to compose the transformations! On one hand, (PQ)(u)=P(Q(u))=P(0,y)=(0,0)(P\circ Q)(u) = P(Q(u)) = P(0,y)= (0,0) and on the other hand(QP)(u)=Q(P(u))=Q(x,0)=(0,0)(Q \circ P)(u) = Q(P(u)) = Q(x,0)=(0,0). This means that PQ=QP=OP\circ Q = Q \circ P = \mathscr{O},

So, the projections commute with respect to composition. But remember, in general, this rarely happens!

Even more surprising, you can easily calculate their powers since (PQ)2(u)=O2(u)=O(u)(P\circ Q)^2 (u)=\mathscr{O}^2(u) = \mathscr{O}(u).

Conclusion

Let VV and WW be two vector spaces.

  • L(V,W)L(V, W) is the set of linear transformations from VV into WW
  • L(V)L(V) is the set of linear operators in VV
  • L(V,W)L(V, W) is a vector space with the operations:
  • T+S:VW,(T+S)(v)=T(v)+S(v)T+S:\, V \rightarrow W, \quad (T+S)(v) = T(v) + S(v)
  • λT:VW,(λT)(v)=λT(v)\lambda T:\, V \rightarrow W, \quad (\lambda T)(v) = \lambda T(v)
  • If SL(U,V)S \in L(U,V) and TL(V,W)T \in L(V,W), then their composition is in L(U,W)L(U, W) and is given by:
  • TS:UW,(TS)(u)=T(S(u))T \circ S:U \rightarrow W, \quad (T \circ S)(u) = T(S(u))
  • Usually TSSTT \circ S \neq S \circ T
  • If TL(V)T \in L(V) then their powers are:

T0=IT1=TT2=TTTn=TTn1\begin{align*} T^0 &= I \\ T^1 &= T \\ T^2 &= T \circ T \\ & \vdots \\ T^n &= T \circ T^{n-1} \end{align*}

  • If TL(V)T \in L(V) then it's idempotent if T2=TT^2 = T
3 learners liked this piece of theory. 1 didn't like it. What about you?
Report a typo