Now that you know what linear transformations are, you're ready to delve deeper into their properties. Now, we're going to introduce two new sets intimately related to them: the null space and the image. Both arise naturally in the context of systems of equations.
Thanks to them, you will be able to reinterpret these systems, discover several relevant facts about them, and even analyze them geometrically. Here, you'll work with a linear transformation between two vector spaces and .
Null space
Let's look at an homogeneous system of linear equations. You already know that it can be represented as a matrix equation:
But if you remember the linear transformation associated with , which is given by , then you can write the equation:
So, the solutions to are the vectors that maps to . Well, it turns out that for any linear transformation , this kind of vectors make up an important set, which is called the null space of (or kernel of ):
The null space is full of surprises that you will discover now and in future topics. The first one is that it's not a simple set, but a subspace. In fact, it's quite straightforward to check this out.
Proof that is a subspace
If , then . It follows that , and by definition . Similarly, .
Going back to the systems of linear equations, you've just shown that the solutions of the homogeneous system are the null space of ! Now, let's say that:
You can find the solutions to the system by computing a row-echelon form of . Then you'll find out that:This means that and are a basis, thus the null space of is a plane.
Null spaces properties
Transformations that send different vectors to different vectors are called injective. This means that the unique vector that you can map to is indeed . In consequence . So, injective transformations have the smallest possible null space.
It seems intuitive that the more vectors maps to , the less injective it is. And this suggests that when the null space is as small as possible, is, in fact, injective. Well, the intuition is correct, and this is the second surprise of the null space:
is injective if and only if
Proof
We have already shown that if is injective its null space only contains .
Conversely, if , then , from which we immediately obtain that . By definition this means that is in the null space. But by hypothesis the unique vector that is in the null space is , so that . Clearly this implies that .
This result means that, in order to prove that is injective, it's enough to show that implies that .
Do you remember that linear transformations preserve the structure of vector spaces? Well, the third surprise is that transformations whose null space is as small as possible preserve linear independence. More precisely, if are linearly independent and is injective, then are linearly independent..
Proof
If are linearly independent, then let's consider . This means that , but as is injective, from the previous result we get that . But as these vectors are linearly independent, we know that . So are linearly independent.
Range
Let's go back to systems of equations. When they are not homogeneous they look like this:
But that's exactly . This means that it must be a vector whose image under is . In general, if you put together the image under of all the vectors of , then the result is a set called the range of (or image of ):
The range of has a close relationship to its kernel. Both possess similar properties and actually complement each other.
To begin with, while the null space contains vectors of , the range is a subset of . And like the null space, it is a subspace.
Proof
If , then . Clearly , and as , we get that . Similarly, as it's immediate that .
If, for example, is a transformation from into , then this means that it collapses all into a point, a line, a plane or even all .
Now, the system has a solution if which now means that . Therefore, the fact that the system of equations has a solution is the same that is in ! Let's continue with the matrix from the first example.
Then generates the range of , and so it is a line:
Range properties
Transformations that occupy the entire W space are called surjective. This means that . In other words, surjective transformations have the biggest possible range.
Linear transformations "preserve" spanning sets:
If then
Proof
If , then and by hypothesis . Applying we get that so
You can use this result to calculate the rank of a matrix. Since is the ith column of , it follows immediately that the range of is the space spanned by the columns of . The dimension of is known as the rank of .
In summary, injective transformations have the smallest null space and preserve linear independence. On the other hand, surjective transformations have the largest range and preserve the spanning sets.
Putting it all together, bijective transformations (injective and surjective simultaneously) preserve linearly independent spanning sets, in other words, they convert bases of V to bases of W! They are so important that they have their own name, isomorphisms.
Equilibrium
Up to this point, it seems pretty clear that the kernel and image tell us several things about the behavior of T and have similar properties.
However, their connection is so intimate that they balance each other: the smaller one is, the bigger the other. More precisely, as is a subspace of , it's clear that . Well, it turns out that is the number that complements in the sense that it is the exact amount that is missing to reach , in the sense that .
This is precisely the meaning of the last theorem, which is both a theoretical and a practical tool that you will use frequently in the future. It receives a name that is not at all modest: the dimension theorem (also known as the fundamental theorem of linear transformations).
Proof
Let be a basis for . We can extend it to a basis of by adding more vectors . Then and . What we're going to proof is that is a basis for .
First, it's a spanning set because if , then there are such that:
By applying we get that:
Finally, we will verify that are linearly independent. If , then. This means that and as is a basis for we get that:But this implies that . And as is a basis, we get that .
As a first sample of the power of this theorem, you can show really easily that when it is equivalent that is injective, surjective and bijective!
Here are a couple of applications to systems of linear equations. It is usually quite easy to know the null space of , so from the Dimension Theorem, you can quickly calculate the rank of and then determine if it is a line, a plane, or something else.
On the other hand, when the matrix is squared of , you can interpret its rank as a measure of how likely it is to find unique solutions. The larger the rank, the closer is to , and when we get . From the last theorem, you know that this means that is bijective, so is satisfied by a unique . This simplifies to the fact that the system has a solution and that it is also unique.
Conclusion
- The null space of is .
- The solutions of a homogeneous system of linear equations are the null space of the matrix of the system.
- is injective if and only if .
- The range of is .
- An inhomogeneous system of linear equations has a solution if and only if .
-
is surjective if and only if .
-
The dimension theorem states that