The polymath Henri Poincaré once said "Mathematics is the art of giving the same name to different things". Now you will learn the connection between invertible matrices, bases and operators as well as their geometry. They are all different ways of talking about the same concept.
These relationships will help you better understand invertible matrices, understand deeper topics, and save you a lot of unnecessary calculations. In the following you'll work with a square matrix of size .
Basic properties
The following properties show the behavior of the inverse with other common matrix operations. They come in handy when you need to compute the inverse of some matrix in terms of a simpler one.
If is invertible then:
- is invertible and .
- is invertible and for any non-zero real number .
- is invertible and .
- is invertible and for every positive integer .
Proof
- This is immediate from the definition of inverse matrix. For the next statements, we only need to check the definition of matrix inverse
- For the second point:
- The third claim is just:
- For the last statement:
Suppose you want to find the inverse of the following matrix . Since all its inputs are even, you can factor the 2 so that
. Now, the new matrix is easier to invert. You can easily compute , so thanks to the second property:
Systems of linear equations
So far you know that when a matrix is invertible, any system of equations it appears in will have a unique solution. Now, you'll look at other equivalent conditions that are easier to check to find out if the matrix is invertible.
-
is invertible.
-
The linear system has a solution for all .
-
The linear system has a unique solution for all -
-
The linear system has a unique solution (the trivial one ).
Proof
Let's denote the reduced row echelon form of by . We're going to proof that 1 3 4 1 and 1 2 1. Thus, we can arrive at any statement starting from any of them. And this implies their equivalence.
- implies and . This has been showed in the first topic about matrix inverse
- implies . This is obvious by taking .
- imples . First of all, represent the system in augmented matrix form . By applying row operations in order to find the reduced row echelon form of , the augmented matrix becomes . Since is in reduced echelon form and is square, if it weren't the identity matrix, then its last row would only contain zeros. But then in the augmented matrix you could delete the last equation and then you would have an homogeneous system with more variables than equations. Thus the system would have infinitely many solutions, which is a contradiction. Thus and this means that is invertible.
- implies . Consider the system with augmented matrix . You can apply some row operations in this augmented matrix to convert into . These operations turn into some vector . Then the augmented matrix becomes . Again, since is in reduced echelon form and is square, if it weren't the identity matrix, then its last row would only contain zeros. But this implies that the has no solution. In consequence the system has no solution and this is a contradiction. Thus and this signifies that is invertible.
This means that it is enough to guarantee that the homogeneous system has a unique solution (the trivial one) in order to solve every system for each possible . But let's see a geometric interpretation of this result.
The geometry of invertible matrices
The above characterization of invertibility can be paraphrased in terms of the linear transformation associated with . In particular the third point means that for every there's a unique such that . Well, this signifies, by definition, that is bijective. But this in turn is equivalent to being an invertible function. With this you've just discovered one of the most profound relationships between matrices and linear transformations:
is invertible if and only if is invertible
From a geometric point of view, this means that invertible matrices have the best possible behavior in the sense that does not agglomerate vectors (injective) and covers all (surjective). For this reason, you can think of an invertible matrix as a transformation that reassigns the position of all points in space: each point changes position but no pair of points is sent to the same location. You can interpret this as the invertible operators "shaking" the space.
For example, a matrix of size 3 that is not invertible has an associated operator in that collapses space into a point, a line or a plane.
But the operator of an invertible matrix doesn't collapse the space, it only changes the positions of its points.
A quick test of invertibility
Time to analyze the relationship between invertibility and the columns of the matrix. The key is that the linear system can be rewritten in terms of the columns of as:
Recall that the vectors are linearly independent if and only if . Now, notice that this is the same as being the only possible solution of . And you've already seen that this means that is invertible. You just proved the next result:
The following conditions are equivalent:
- is invertible.
- The columns of form a linearly independent set.
- The columns of form a basis for .
Linearly dependent vectors are those that contain redundant information (in the sense that one of them can be obtained through the others). So invertible matrices are those whose columns have the minimum information needed to reconstruct the whole space .
Up to this point you're familiar with a lot of properties of invertible matrices. You also know how to calculate them and their geometric interpretation. But first of all, how can you tell if a matrix is invertible? Although the simplest way is to check if the columns are linearly independent, there is a much more direct technique that requires less calculations and is the most widely used:
is invertible if and only if .
Proof
The determinant detects linearly independence. That is, the columns of are linearly independent if and only if . But you've just seen that the first condition is equivalent to the fact that is invertible.
Let's see this technique in action with the following matrices:
The determinants of the first two matrices are and , respectively. Therefore both are invertible. But the determinant of the third is , so it is not invertible. Quick and straightforward.
Conclusion
You just developed a lot of synonyms for invertibility, so let's go over them. The square matrix is invertible if and only if:
-
The linear system has a unique solution for all .
-
is invertible
-
The columns of form a basis for .
-
Some useful properties of the inverse are:
- for any non-zero real number .
- .
- for every positive integer .