Skip to main content

Linear Algebra

Module 12 Inverse Functions & Inverse Matrices

In this module you will learn
  • The definition of an inverse function and an inverse matrix.
  • How to decompose a matrix into the product of elementary matrices and how to use elementary matrices to compute inverses.
  • How the order of matrix multiplication matters.
  • How row-reduction and matrix inverses relate.
We should think of transformations or functions as machines that perform some manipulation of their input and then give an output. This perspective allows us to divide functions into two natural categories: those that can be undone and those that cannot. The official term for a function that can be undone is an invertible function.

Section 12.1 Invertible Functions

The simplest function is the identity function.

Definition 12.1.1. Identity Function.

Let \(X\) be a set. The identity function with domain and codomain \(X\text{,}\) notated \(\Ident:X\to X\), is the function satisfying
\begin{equation*} \Ident(x)=x \end{equation*}
for all \(x\in X\text{.}\)
The identity function is the function that does nothing to its input
 1 
Technically, for every set there exists a unique identity function with that set as the domain/codomain, but we won’t belabor this point.
. When doing precise mathematics, we often prove a function or composition of functions does nothing to its input by showing it is equal to the identity function
 2 
This is similar to saying that we know \(\vec x=\vec y\) if and only if \(\vec x-\vec y=\vec 0\text{.}\)
.
In plain terms, a function is invertible if it can be undone. More precisely a function is invertible if there exists an inverse function that when composed with the original function produces the identity function and vice versa.

Definition 12.1.2. Inverse Function.

Let \(f:X\to Y\) be a function. We say \(f\) is invertible if there exists a function \(g:Y\to X\) so that \(f\circ g=\Ident\) and \(g\circ f=\Ident\text{.}\) In this case, we call \(g\) an inverse of \(f\) and write
\begin{equation*} f^{-1}=g. \end{equation*}
Let’s consider an example. You have some money in your pockets. Let \(l:\Set{\text{nickels in left pocket}}\to \N\) be the function that adds up the value of all the nickels in your left pocket. Let \(r:\Set{\text{nickels in either pocket}}\to \N\) be the function that adds up the value of all the nickels in both of your pockets. In this case, \(l\) would be invertible—if you know that \(l(\text{# nickels})=25\text{,}\) you must have had \(5\) nickels in your left pocket. We can write down a formula for \(l^{-1}\) as
\begin{equation*} l^{-1}(n)=\frac{n}{5}. \end{equation*}
However, \(r\) is not invertible. If \(r(\text{# nickels})=25\text{,}\) you might have had \(5\) nickels in your left pocket, but you might have \(3\) nickels in your left pocket and \(2\) in your right. We just don’t know, so no inverse to \(r\) can exist.
What we’ve just learned is that for a function to be invertible, it must be one-to-one.

Definition 12.1.3. One-to-one.

Let \(f:X\to Y\) be a function. We say \(f\) is one-to-one (or injective) if distinct inputs to \(f\) produce distinct outputs. That is \(f(x)=f(y)\) implies \(x=y\text{.}\)
Whenever a function \(f\) is one-to-one, there exists a function \(g\) so that \(g\circ f=\Ident\text{.}\) However, this is not enough to declare that \(f\) is invertible
 3 
In this situation, we say that \(f\) is left-invertible.
because we also need \(f\circ g=\Ident\text{.}\) To ensure this, we need \(f\) to be onto.

Definition 12.1.4. Onto.

Let \(f:X\to Y\) be a function. We say \(f\) is onto (or surjective) if every point in the codomain of \(f\) gets mapped to. That is \(\Range(f)=Y\text{.}\)
Every invertible function is both one-to-one and onto, and every one-to-one and onto function is invertible. And, as we will learn, this has implications for the rank and nullity of linear transformations.

Section 12.2 Invertibility and Linear Transformations

Let’s now focus on linear transformations. We know that a linear transformation \(\mathcal{T}:\R^{n}\to\R^{m}\) is invertible if and only if it is one-to-one and onto.
If \(\mathcal{T}\) is one-to-one, that means that distinct inputs to \(\mathcal{T}\) yield distinct outputs. In other words, the solution to \(\mathcal{T}(\vec x)=\vec b\) is always unique. But, the set of all solutions to \(\mathcal{T}(\vec x)=\vec b\) can be expressed as
\begin{equation*} \Null(\mathcal{T})+\Set{\vec p}. \end{equation*}
Therefore, \(\mathcal{T}\) is one-to-one if and only if \(\Nullity(\mathcal{T})=0\text{.}\) If \(\mathcal{T}\) is onto, then \(\Range(\mathcal{T})=\R^{m}\) and so \(\Rank(\mathcal{T})=m\text{.}\)
Now, suppose \(\mathcal{T}\) is one-to-one and onto. By the rank-nullity theorem,
\begin{equation*} \Rank(\mathcal{T})+\Nullity(\mathcal{T}) = 0+m=m=n=\Dim(\text{domain of $\mathcal{T}$}), \end{equation*}
and so \(\mathcal{T}\) has the same domain and codomain (at least a domain and codomain of the same dimension).
Using the rank-nullity theorem, we can start developing a list of properties that are equivalent to invertibility of a linear transformation.
  • \(\mathcal{T}:\R^{n}\to\R^{m}\) is invertible if and only if \(\Nullity(\mathcal{T})=0\) and \(\Rank(\mathcal{T})=m\text{.}\)
  • \(\mathcal{T}:\R^{n}\to\R^{m}\) is invertible if and only if \(m=n\) and \(\Nullity(\mathcal{T})=0\text{.}\)
  • \(\mathcal{T}:\R^{n}\to\R^{m}\) is invertible if and only if \(m=n\) and \(\Rank(\mathcal{T})=m\text{.}\)

Example 12.2.1.

Let \(\mathcal{P}:\R^{2}\to\R^{2}\) be projection onto the \(x\)-axis and let \(\mathcal{R}:\R^{2}\to\R^{2}\) be rotation counter-clockwise by \(15^{\circ}\text{.}\) Classify each of \(\mathcal{P}\) and \(\mathcal{R}\) as invertible or not.

Solution.

Notice that \(\mathcal{P}(\yhat)=\mathcal{P}(2\yhat)=\vec 0\text{,}\) therefore \(\mathcal{P}\) is not one-to-one and so is not invertible.
Let \(\mathcal{Q}: \R^{2}\to\R^{2}\) be rotation clockwise by \(15^{\circ}\text{.}\) \(\mathcal{R}\) and \(\mathcal{Q}\) will undo each other. Phrased mathematically,
\begin{equation*} \mathcal{R}\circ \mathcal{Q}=\Ident\qquad\text{and}\qquad \mathcal{Q}\circ \mathcal{R}=\Ident. \end{equation*}
Therefore, \(\mathcal{Q}\) is an inverse of \(\mathcal{R}\text{,}\) and so \(\mathcal{R}\) is invertible.
One important fact about linear transformations is that if a linear transformation is invertible, its inverse is also a linear transformation.

Proof.

Let \(\mathcal{T}\) be an invertible linear transformation and let \(\mathcal{T}^{-1}\) be its inverse. We need to show that (i) \(\mathcal{T}^{-1}\) distributes over addition and (ii) \(\mathcal{T}^{-1}\) distributes over scalar multiplication.
(i)
First observe that since \(\mathcal{T}\circ \mathcal{T}^{-1}=\Ident\) and because \(\mathcal{T}\) is linear, we have
\begin{equation*} \vec a+\vec b = \mathcal{T}\circ \mathcal{T}^{-1}\vec a+\mathcal{T}\circ \mathcal{T}^{-1}\vec b= \mathcal{T}(\mathcal{T}^{-1}\vec a+ \mathcal{T}^{-1}\vec b). \end{equation*}
Since \(\mathcal{T}^{-1}\circ \mathcal{T}=\Ident\text{,}\) by using the fact that \(\vec a+\vec b=\mathcal{T}(\mathcal{T}^{-1}\vec a+ \mathcal{T}^{-1}\vec b)\) we know
\begin{equation*} \mathcal{T}^{-1}(\vec a+\vec b) = \mathcal{T}^{-1}\Big(\mathcal{T}(\mathcal{T}^{-1}\vec a+ \mathcal{T}^{-1}\vec b)\Big)= \mathcal{T}^{-1}\vec a+ \mathcal{T}^{-1}\vec b. \end{equation*}
(ii)
Similar to the proof of (i), we see
\begin{equation*} \mathcal{T}^{-1}(\alpha\vec a) = \mathcal{T}^{-1}\Big(\alpha \big(\mathcal{T}\circ \mathcal{T}^{-1}\vec a\big)\Big) =\mathcal{T}^{-1}\circ \mathcal{T}(\alpha \mathcal{T}^{-1}\vec a) = \alpha\mathcal{T}^{-1}\vec a. \end{equation*}

Section 12.3 Invertibility and Matrices

In the world of matrices, the identity matrix takes the place of the identity function.

Definition 12.3.1. Identity Matrix.

An identity matrix is a square matrix with ones on the diagonal and zeros everywhere else. The \(n\times n\) identity matrix is denoted \(I_{n\times n}\text{,}\) or just \(I\) when its size is implied.
We can now define what it means for a matrix to be invertible
 1 
This should look very similar to what it means for a function to be invertible.
.

Definition 12.3.2. Matrix Inverse.

The inverse of a matrix \(A\) is a matrix \(B\) such that \(AB=I\) and \(BA=I\text{.}\) In this case, \(B\) is called the inverse of \(A\) and is notated \(A^{-1}\text{.}\)

Example 12.3.3.

Determine whether the matrices \(A=\mat{2&5\\-3&-7}\) and \(B=\mat{-7&-5\\3&2}\) are inverses of each other.

Solution.

\begin{align*} AB&=\mat{2&5\\-3&-7}\mat{-7&-5\\3&2}=\mat{1&0\\0&1}= I\\ BA&=\mat{-7&-5\\3&2}\mat{2&5\\-3&-7}=\mat{1&0\\0&1}= I . \end{align*}
Therefore, \(A\) and \(B\) are inverses of each other.

Example 12.3.4.

Determine whether the matrices \(A=\mat{2&5&0\\-3&-7&0}\) and \(B=\mat{-7&-5\\3&2\\1&1}\) are inverses of each other.

Solution.

\begin{equation*} AB =\mat{2&5&0\\-3&-7&0}\mat{-7&-5\\3&2\\1&1}=\mat{1&0\\0&1}=I \end{equation*}
but
\begin{equation*} BA =\mat{-7&-5\\3&2\\1&1}\mat{2&5&0\\-3&-7&0}=\mat{1&0&0\\0&1&0\\-1&-2&0}\neq I. \end{equation*}
Therefore, \(A\) and \(B\) are not inverses of each other.
Since every matrix induces a linear transformation, we can use the facts we know about invertible linear transformations to produce facts about invertible matrices. In particular:
  • An \(n\times m\) matrix \(A\) is invertible if and only if \(\Nullity(A)=0\) and \(\Rank(A)=n\text{.}\)
  • An \(n\times n\) matrix \(A\) is invertible if and only if \(\Nullity(A)=0\text{.}\)
  • An \(n\times n\) matrix \(A\) is invertible if and only if \(\Rank(A)=n\text{.}\)

Section 12.4 Matrix Algebra

The linear equation \(ax=b\) has solution \(x=\frac{b}{a}\) whenever \(a\neq 0\text{.}\) We arrive at this solution by dividing both sides of the equation by \(a\text{.}\) Does a similar process exist for solving the matrix equation \(A\vec x=\vec b\text{?}\) It sure does!
Unfortunately, we cannot divide by a matrix, but to solve \(A\vec x=\vec b\text{,}\) we don’t need to “divide” by a matrix, we just need to eliminate \(A\) from the left side. This could be accomplished by using an inverse.
Suppose \(A\) is invertible, then
\begin{equation*} A\vec x=\vec b\qquad\implies\qquad A^{-1}A\vec x=A^{-1}\vec b\qquad\implies\qquad \vec x=A^{-1}\vec b. \end{equation*}
Thus, if we have the inverse of a matrix handy, we can use it to solve a system of equations.

Example 12.4.1.

Use the fact that \(\mat{2&5\\-3&-7}^{-1}=\mat{-7&-5\\3&2}\) to solve the system \(\left\{\begin{array}{crcrl}&2x&+&5y&=2\\-&3x&-&7y&=1\end{array}\right.\text{.}\)

Solution.

The system can be rewritten as
\begin{equation*} \mat{2&5\\-3&-7}\mat{x\\y}=\mat{2\\1}. \end{equation*}
Multiplying both sides by \(\mat{2&5\\-3&-7}^{-1}\) gives
\begin{equation*} \mat{x\\y}=\mat{2&5\\-3&-7}^{-1}\mat{2\\1}=\mat{-7&-5\\3&2}\mat{2\\1}=\mat{-19\\8}. \end{equation*}
It’s important to note that, unlike in the case with regular scalars, the order of matrix multiplication matters. So, whereas with scalars you could get away with something like
\begin{equation*} ax=b \qquad\implies \qquad \tfrac{1}{a}ax=b\tfrac{1}{a}\qquad\implies \qquad x=\tfrac{b}{a}, \end{equation*}
with matrices \(A\vec x=\vec b\) does not imply \(A^{-1}A\vec x=\vec b A^{-1}\text{.}\) In fact, if \(\vec b\) is a column vector, the expression \(\vec bA^{-1}\) is almost always undefined!

Section 12.5 Finding a Matrix Inverse

Whereas before we only knew how to solve a matrix equation \(A\vec x=\vec b\) using row reduction, we now know how to use \(A^{-1}\) to solve the same system. In fact, \(A^{-1}\) is the exact matrix so that \(\vec x=A^{-1}\vec b\) is the solution to \(A\vec x=\vec b\text{.}\) Therefore, by picking different \(\vec b\)’s and solving for \(\vec x\text{,}\) we can find \(A^{-1}\text{.}\)

Example 12.5.1.

Let \(A=\mat{2&5\\-3&-7}\text{.}\) Find \(A^{-1}\text{.}\)

Solution.

We know \(A^{-1}=\mat{a&b\\c&d}\) will be a \(2\times 2\) matrix, and we know \(\vec x=A^{-1}\vec b\) will always be the unique solution to \(A\vec x=\vec b\text{.}\) Therefore, we can find \(A^{-1}\) by finding \(\vec x,\vec b\) pairs that satisfy \(A\vec x=\vec b\text{.}\)
Using row reduction, we see
\begin{equation*} A\vec x=\mat{1\\0}\quad\text{has solution}\quad \vec x=\mat{-7\\3}, \qquad\text{and}\qquad A\vec x=\mat{0\\1}\quad\text{has solution}\quad \vec x=\mat{-5\\2}. \end{equation*}
Therefore
\begin{equation*} A^{-1}\mat{1\\0}=\mat{a\\c}=\mat{-7\\3}\qquad \text{and}\qquad A^{-1}\mat{0\\1}=\mat{b\\d}=\mat{-5\\2}, \end{equation*}
and so
\begin{equation*} A^{-1}=\mat{-7&-5\\3&2}. \end{equation*}

Section 12.6 Elementary Matrices

Finding the inverse of a matrix can be a lot of work. However if you already know how to undo what the matrix does, finding the inverse might not be so hard. For example, if \(R_{30}\) is the matrix that rotates vectors in \(\R^{2}\) counter-clockwise by \(30^{\circ}\text{,}\) its inverse must be \(R_{-30}\text{,}\) the matrix that rotates vectors in \(\R^{2}\) clockwise by \(30^{\circ}\text{.}\)
Like before when we analyzed linear transformations by breaking them up into compositions of simpler linear transformations, another strategy for finding an inverse matrix is to break a matrix into simpler ones whose inverses we can just write down.
Some of the simplest matrices around are the elementary matrices.

Definition 12.6.1. Elementary Matrix.

A matrix is called an elementary matrix if it is an identity matrix with a single elementary row operation applied.
Examples of elementary matrices include
\begin{equation*} \mat{1&0&0\\0&1&0\\0&0&-5}\qquad \mat{1&0&7\\0&1&0\\0&0&1}\qquad \mat{0&1&0\\1&0&0\\0&0&1}. \end{equation*}
These matrices are obtained from the row operations “multiply the last row by \(-5\)”, “add \(7\) times the last row to the first”, and “swap the first two rows”.
Elementary matrices are useful because multiplying by an elementary matrix performs the corresponding elementary row operation! See for yourself:
\begin{align*} \mat{1&0&0\\0&1&0\\0&0&-5}\mat{a&b&c\\d&e&f\\g&h&i}&= \mat{a&b&c\\d&e&f\\-5g&-5h&-5i}\\ \mat{1&0&7\\0&1&0\\0&0&1}\mat{a&b&c\\d&e&f\\g&h&i}&= \matc{a+7g&b+7h&c+7i\\d&e&f\\g&h&i}\\ \mat{0&1&0\\1&0&0\\0&0&1}\mat{a&b&c\\d&e&f\\g&h&i}&= \mat{d&e&f\\a&b&c\\g&h&i} \end{align*}
As a refresher, the elementary row operations are:
  • multiply a row by a non-zero constant;
  • add a multiple of one row to another; and
  • swap two rows.
Each one of these operations can be undone, and so every elementary matrix is invertible. What’s more, the inverse is another elementary matrix that is easy to write down.

Example 12.6.2.

Find the inverse of \(E= \mat{1&0&7\\0&1&0\\0&0&1}\text{.}\)

Solution.

Since \(E\) corresponds to the row operation “add \(7\) times the last row to the first”, \(E^{-1}\) must correspond to the row operation “subtract \(7\) times the last row from the first”. Therefore,
\begin{equation*} E^{-1}=\mat{1&0&-7\\0&1&0\\0&0&1}. \end{equation*}

Section 12.7 Elementary Matrices and Inverses

For a matrix \(M\) to be invertible, we know that \(M\) must be square and \(\Nullity(M)=0\text{.}\) That means, \(M\) is invertible if and only if \(\Rref(M)=I\text{.}\) In other words, \(M\) is invertible if there is a sequence of elementary row operations that turn \(M\) into \(I\text{.}\) Each one of these row operations can be represented by an elementary matrix, which gives us the following theorem.
Now, suppose \(M\) is invertible and let \(E_{1},\ldots,E_{k}\) be elementary matrices so that \(E_{k}\cdots E_{2}E_{1}M=I\text{.}\) We now know
\begin{equation*} E_{k}\cdots E_{2}E_{1}M=\underbrace{(E_k\cdots E_2E_1)}_{Q}M=QM=I. \end{equation*}
If we can argue that \(MQ=I\text{,}\) then \(Q\) will be the inverse of \(M\text{!}\)

Proof.

Suppose \(A\) is a square matrix and that \(AB=I\text{.}\) Since \(AB=I\text{,}\) \(B\) must also be square. Since \(\Null(B)\subseteq\Null(AB)\text{,}\) we know \(\Nullity(B)\leq \Nullity(AB)=\Nullity(I)=0\text{,}\) and so \(B\) is invertible (since it’s a square matrix whose nullity is \(0\)). Let \(B^{-1}\) be the inverse of \(B\text{.}\) Observe now that
\begin{equation*} A=AI=A(BB^{-1})=(AB)B^{-1}=IB^{-1}=B^{-1}, \end{equation*}
and so \(A=B^{-1}\text{.}\) Finally, substituting \(B^{-1}\) for \(A\) shows
\begin{equation*} BA=BB^{-1}=I. \end{equation*}
In light of this theorem, we now have a new algorithm for finding the inverse of a matrix—find elementary matrices that turn the matrix into the identity matrix and multiply those elementary matrices together to find the inverse.

Example 12.7.3.

Let \(A=\mat{1&2&0\\0&4&0\\0&-1&1}\text{.}\) Find \(A^{-1}\) using elementary matrices.

Solution.

We can row-reduce \(A\) with the following steps:
\begin{equation*} \mat{1&2&0\\0&4&0\\0&-1&1}\to \mat{1&2&0\\0&1&0\\0&-1&1}\to \mat{1&0&0\\0&1&0\\0&-1&1}\to \mat{1&0&0\\0&1&0\\0&0&1} \end{equation*}
The elementary matrices corresponding to these steps are
\begin{equation*} E_{1}=\mat{1&0&0\\0&\frac{1}{4}&0\\0&0&1}\qquad E_{2}=\mat{1&-2&0\\0&1&0\\0&0&1}\qquad E_{3}=\mat{1&0&0\\0&1&0\\0&1&1}. \end{equation*}
We now have
\begin{equation*} E_{3} E_{2} E_{1} A = I, \end{equation*}
and so
\begin{equation*} A^{-1}=E_{3} E_{2} E_{1} = \mat{1&0&0\\0&1&0\\0&1&1}\mat{1&-2&0\\0&1&0\\0&0&1}\mat{1&0&0\\0&\frac{1}{4}&0\\0&0&1}= \mat{1&-\frac{1}{2}&0\\ 0&\frac{1}{4}&0\\ 0&\frac{1}{4}&1}. \end{equation*}

Section 12.8 Decomposition into Elementary Matrices

If \(A\) is an invertible matrix, then the double-inverse of \(A\) (i.e., \((A^{-1})^{-1}\)) is \(A\) itself
 1 
Formally we say that the operation of taking a matrix inverse is an involution.
. This is easily proved. By definition, \((A^{-1})^{-1}\) is a matrix \(B\) so that \(BA^{-1}=I\) and \(A^{-1}B=I\text{.}\) But \(B=A\) satisfies this condition!
Now, suppose \(M\) is an invertible matrix. Then, there exists a sequence of elementary matrices \(E_{1},\ldots, E_{k}\) so that \(E_{k}\cdots E_{2}E_{1}M=I\) and
\begin{equation*} M^{-1}=E_{k}\cdots E_{2}E_{1}. \end{equation*}
Therefore
\begin{equation*} M=(M^{-1})^{-1}= (E_{k}\cdots E_{2}E_{1})^{-1}. \end{equation*}
Thinking carefully about what \((E_{k}\cdots E_{2}E_{1})^{-1}\) should be, we see that
\begin{equation*} (E_{1}^{-1}E_{2}^{-1}\cdots E_{k}^{-1})E_{k}\cdots E_{2}E_{1}=I\qquad\text{and}\qquad E_{k}\cdots E_{2}E_{1}(E_{1}^{-1}E_{2}^{-1}\cdots E_{k}^{-1})=I, \end{equation*}
and so
\begin{equation*} M=(E_{k}\cdots E_{2}E_{1})^{-1}=E_{1}^{-1}E_{2}^{-1}\cdots E_{k}^{-1}. \end{equation*}
(Notice the order of matrix multiplication reversed!) Each \(E_{i}^{-1}\) is also an elementary matrix, and so we have just shown that every invertible matrix can be written as the product of elementary matrices. This is actually a double-sided implication (if and only if).

Proof.

Suppose \(M\) is invertible. Then, there exists a sequence of elementary matrices \(E_{1},\ldots,E_{k}\) so that \(E_{k}\cdots E_{1}M=I\text{.}\) It follows that
\begin{equation*} M=E_{1}^{-1}E_{2}^{-1}\cdots E_{k}^{-1} \end{equation*}
is the product of elementary matrices. Conversely, since the product of invertible matrices is invertible and every elementary matrix is invertible, the product of elementary matrices must be invertible. Therefore, if \(M\) is not invertible, it cannot be written as the product of elementary matrices.

Exercises 12.9 Exercises

1.

Determine whether the following linear transformations are one-to-one, onto, or both. As well, determine whether or not they are invertible. Justify your answers.
  1. \(\mathcal{S}:\R^{2}\to\R^{2}\text{,}\) where \(\mathcal{S}\) is the linear transformation that doubles every vector.
  2. \(\mathcal{R}:\R^{2}\to\R^{2}\text{,}\) where \(\mathcal{R}\) the linear transformation that rotates every vector clockwise by \(72^{\circ}\text{.}\)
  3. \(\mathcal{P}:\R^{2}\to\R^{2}\text{,}\) where \(\mathcal{P}\) the linear transformation that projects every vector onto the \(y\)-axis.
  4. \(\mathcal{F}:\R^{2}\to\R^{2}\text{,}\) where \(\mathcal{F}\) is the linear transformation that reflects every vector over the line \(y=x\text{.}\)
  5. \(\mathcal{T}:\R^{3}\to\R^{3}\text{,}\) where \(\mathcal{T}\) is the linear transformation induced by the matrix \(M_{\mathcal{T}}= \mat{1&2&3\\4&5&6\\7&8&9}\text{.}\)
  6. \(\mathcal{U}:\R^{3}\to\R^{2}\text{,}\) where \(\mathcal{U}\) is the linear transformation induced by the matrix \(M_{\mathcal{U}}= \mat{1&2&3\\3&4&5}\text{.}\)
Solution.
  1. \(\mathcal{S}\) is both one-to-one and onto. No two distinct vectors can be doubled to become the same vector, thus the transformation is one-to-one. Every vector is double of the vector that is half of itself, thus the transformation is onto. \(\mathcal{S}\) is invertible, its inverse is the transformation that halves every vector.
  2. \(\mathcal{R}\) is both one-to-one and onto. No two distinct vectors can be rotated by \(72^{\circ}\) clockwise to become the same vector, thus the transformation is one-to-one. Every vector is the \(72^{\circ}\) clockwise rotation of the vector that is the \(72^{\circ}\) counter-clockwise rotation of itself, thus the transformation is onto. \(\mathcal{R}\) is invertible, its inverse is the transformation that rotates every vector counter-clockwise by \(72^{\circ}\text{.}\)
  3. \(\mathcal{P}\) is neither one-to-one nor onto. Every vector on the x-axis is sent to the origin so there are distinct inputs that do not map to distinct outputs, thus the transformation is not one-to-one. Vectors that do not lie on the y-axis are not in the range of the transformation, thus the transformation is not onto. \(\mathcal{P}\) is not invertible.
  4. \(\mathcal{F}\) is both one-to-one and onto. No two distinct vectors can be reflected to become the same vector, thus the transformation is one-to-one. Every vector is the reflection of the vector that is the reflection of itself, thus the transformation is onto. \(\mathcal{F}\) is invertible, its inverse is itself—the reflection about the line \(y=x\text{.}\)
  5. \(\mathcal{T}\) is neither one-to-one nor onto. The RREF of \(M_{\mathcal{T}}\) is \(\mat{1&0&-1\\0&1&0\\0&0&0}\text{.}\) \(\Rank(\mathcal{T})\) = \(\Rank(M_{\mathcal{T}})\) = 2. by the rank-nullity theorem the nullity is 1. An entire line of vectors is getting mapped to \(\vec{0}\) so there are distinct inputs that do not map to distinct outputs, thus the transformation is not one-to-one. The range of \(\mathcal{T}\) = the column space of \(M_{\mathcal{T}}\) which is not all of \(\R^{3}\text{,}\) Thus the transformation is not onto. \(\mathcal{T}\) is not invertible.
  6. \(\mathcal{U}\) is onto but is not one-to-one. Because \(\mathcal{U}\) maps from \(\R^{3}\) to \(\R^{2}\text{,}\) the rank is at most 2, so by the Rank-Nullity Theorem, the nullity must be at least 1. More than one vector is getting mapped to \(\vec{0}\) so there are distinct inputs that do not map to distinct outputs, thus the transformation is not one-to-one. Since \(\Rref(M_{\mathcal{U}})\) is \(\mat{1&0&-1\\0&1&2}\text{.}\) \(\Rank(\mathcal{U})=\Rank(M_{\mathcal{U}})=2\text{,}\) which is the dimension of the codomain, thus the transformation is onto. \(\mathcal{U}\) is not invertible.

2.

Invert the following matrices or explain why they are not invertible.
  1. \(\displaystyle M_{1}= \mat{2&3\\1&-1}\)
  2. \(\displaystyle M_{2}= \mat{1&0\\1&0}\)
  3. \(\displaystyle M_{3}= \mat{0&2&1\\1&0&1\\-2&3&0}\)
  4. \(\displaystyle M_{4}= \mat{2&0&1&8\\1&-5&2&2\\3&-1&0&7}\)
  5. \(\displaystyle M_{5}= \mat{0&-3&1&2\\1&0&-1&1\\2&-1&0&0\\0&0&1&3}\)
Solution.
  1. To find the inverse we row reduce \(M_{1}\text{,}\) applying each of the row operations to the identity matrix.
    \(\left[\begin{array}{cc|cc}2&3&1&0\\ 1&-1&0&1\\\end{array}\right]\)
    \(\left[\begin{array}{cc|cc}1&\frac{3}{2}&\frac{1}{2}&0\\ 1&-1&0&1\\\end{array}\right]\)
    \(\left[\begin{array}{cc|cc}1&\frac{3}{2}&\frac{1}{2}&0\\ 0&-\frac{5}{2}&-\frac{1}{2}&1\\\end{array}\right]\)
    \(\left[\begin{array}{cc|cc}1&\frac{3}{2}&\frac{1}{2}&0\\ 0&1&\frac{1}{5}&-\frac{2}{5}\\\end{array}\right]\)
    \(\left[\begin{array}{cc|cc}1&0&\frac{1}{5}&\frac{3}{5}\\ 0&1&\frac{1}{5}&-\frac{2}{5}\\\end{array}\right]\)
    So \(M_{1}^{-1}=\mat{\frac{1}{5}&\frac{3}{5}\\ \frac{1}{5}&-\frac{2}{5}\\}\text{.}\)
  2. \(M_{2}\) does not row reduce to the identity matrix so it is not invertible.
  3. \(\displaystyle M_{3}^{-1}=\mat{3&-3&-2\\2&-2&-1\\-3&4&2}\)
  4. \(M_{4}\) is not invertible because it is not a square matrix.
  5. \(\displaystyle M_{5}^{-1}=\frac{1}{23}\mat{-4&-1&12&3\\-8&-2&1&6\\-3&-18&9&8\\1&6&-3&5}\)

3.

Solve the following systems in two ways: (i) by using row reduction, and (ii) by using inverse matrices.
  1. \(\displaystyle \left\{\begin{array}{rcrl}2x&+&y&=5\\3x&+&7y&=3\end{array}\right.\)
  2. \(\displaystyle \left\{\begin{array}{rcrcrl}2x&+&2y&+&3z&=4\\2x&+&2y&+&z&=0\\4x&+&5y&+&6z&=2\end{array}\right.\)

4.

Let \(A=\mat{1&3&5\\0&2&0\\-1&0&2}\text{.}\)
  1. Express \(A^{-1}\) as the product of elementary matrices.
  2. Express \(A\) as the product of elementary matrices.

5.

For each statement below, determine whether it is true or false. Justify your answer.
  1. For an arbitrary linear transformation \(T:\R^{m}\to\R^{n}\text{,}\) if \(n \neq m\text{,}\) then the linear transformation is not invertible.
  2. The matrix \(M = \mat{1&0&0\\1&1&0}\) is an elementary matrix.
  3. Every elementary matrix is invertible.
  4. The product of elementary matrices is sometimes an elementary matrix.
  5. The product of elementary matrices is always an elementary matrix.
  6. A matrix that induces an invertible linear transformation is necessarily invertible.
  7. A transformation that is one-to-one and onto is always invertible.
  8. For two matricies \(A\) and \(B\text{,}\) if \(AB=I\text{,}\) then \(A\) and \(B\) are invertible.
Solution.
  1. True. If \(m<n\) then \(T\) cannot be onto. If \(m>n\) then \(T\) cannot be one-to-one. In both cases, \(T\) cannot be invertible. We may conclude that the dimension of the domain and codomain must be equal in order for a transformation to be invertible.
  2. False. Elementary matricies are one row operation away from the identity matrix. The identity matrix of any size is always square. There is no identity matrix such that performing one row operation on it yields \(M\text{.}\) We may conclude that elementary matricies are always square.
  3. True. The inverse of an elementary matrix \(E\) is another elementary matrix \(E^{-1}\) which corresponds to the row operation that turns \(E\) into the identity matrix. \(E^{-1}\) is the “opposite” row operation.
  4. True. Let \(E_{1}\) correspond to multiplying row 1 by 4. Let \(E_{2}\) correspond to multiplying row 1 by 2. The product \(E_{2}E_{1}\) corresponds to multiplying row 1 by 8, which is also a single row operation and thus has a corresponding elementary matrix.
  5. False. Let \(E_{1}\) correspond to multiplying row 1 by 4. Let \(E_{2}\) correspond to multiplying row 2 by 2. The product \(E_{2}E_{1}\) corresponds to multiplying row 1 by 4 and row 2 by 2, which is not a single row operation and thus does not have a corresponding elementary matrix.
  6. True.
  7. True.
  8. False. Let \(A = \mat{1&0&0\\0&1&0}\text{.}\) Let \(B = \mat{1&0\\0&1\\0&0}\text{.}\) \(AB = I\text{.}\) \(BA = \mat{1&0&0\\0&1&0\\0&0&0}\neq I\text{.}\) \(A\) and \(B\) are not invertible. It is required that \(AB = BA = I\) in order for matrices \(A\) and \(B\) to be invertible.