Skip to main content

Section 4.4 Matrix inverses

In Theorem 4.1.19 we saw that the system of linear equations \([A|\vec{b}]\) can be equivalently expressed as a matrix equation \(A\vec{x} = \vec{b}\text{.}\) If this equation was about numbers instead of matrices and vectors then we would easily solve the equation \(ax=b\) by dividing to get \(x = \frac{b}{a}\) (as long as \(a \neq 0\)). Dividing by \(a\) is really the same as multiplying by \(a^{-1}\text{,}\) so we could write our solution as \(x=a^{-1}b\text{.}\) Since we know how to multiply matrices, this brings up a natural question: Is there a matrix \(A^{-1}\) so that the equations \(A\vec{x}=\vec{b}\) and \(\vec{x} = A^{-1}\vec{b}\) are equivalent? Solving this problem is the goal of this section.

Subsection 4.4.1 The inverse of a square matrix

Definition 4.4.1.

Let \(A\) be an \(n \times n\) matrix. An inverse of \(A\) is an \(n \times n\) matrix \(B\) such that \(AB=I_n\) and \(BA = I_n\text{.}\)

If there is an inverse of \(A\) then we say that \(A\) is invertible.

Notice that in the definition we require that multiplication in both orders gives us the identity matrix. Remember that, in general, \(AB \neq BA\text{!}\) Nevertheless, later (Theorem 4.4.14) we will see that it turns out that if \(A\) and \(B\) are square matrices and \(AB=I_n\) then \(BA=I_n\) happens automatically.

Observe that if \(B\) is an inverse for \(A\) then we can accomplish the goal described in the introduction to this section: Multiplying both sides of \(A\vec{x}=\vec{b}\) by \(B\) on the left will give \(BA\vec{x} = B\vec{b}\text{,}\) and \(BA=I_n\text{,}\) so we get \(I_n\vec{x} = B\vec{b}\text{,}\) which is the same as \(\vec{x} = B\vec{b}\)

Let \(A = \begin{bmatrix}1 \amp 1 \\ 1 \amp 0\end{bmatrix}\) and \(B = \begin{bmatrix}0 \amp 1 \\ 1 \amp -1\end{bmatrix}\text{.}\) By direct calculation you can find that \(AB = BA = I_2\text{,}\) so \(B\) is an inverse of \(A\text{,}\) and the matrix \(A\) is invertible.

In Subsection 4.4.4 we will see the techniques needed to find this \(B\) when given \(A\text{.}\)

You will probably not be surprised to learn that the zero matrix is not invertible. Indeed, for any \(n \times n\) matrix \(B\) we have \(0_{n \times n}B = 0_{n\times n} \neq I_n\text{,}\) so \(B\) cannot be an inverse for \(0_{n \times n}\text{.}\)

Let \(A = \begin{bmatrix}2 \amp 1 \\ 6 \amp 3\end{bmatrix}\text{.}\) We will show that \(A\) is not invertible - this is more surprising than the previous example, because not only is \(A\) not the zero matrix, it doesn't even have any zero entries! Suppose that we did have an inverse, say \(B = \begin{bmatrix}a \amp b \\ c \amp d\end{bmatrix}\text{.}\) Then we would have \(AB = I_2\text{.}\) When we carry out the multiplication on the left side of this equation it becomes

\begin{equation*} \begin{bmatrix}2a+c \amp 2b+d \\ 6a+3c \amp 6b+3d\end{bmatrix} = \begin{bmatrix}1 \amp 0 \\ 0 \amp 1\end{bmatrix}\text{.} \end{equation*}

Setting corresponding entries equal, we obtain a system of four linear equations in four variables.

\begin{gather*} 2a+c=1\\ 2b+d=0\\ 6a+3c=0\\ 6b+3d=1 \end{gather*}

If we attempt to solve this system we find that there are no solutions - you could do this by setting up an augmented matrix, or by noticing that the third equation is \(3(2a+c)=0\text{,}\) which contradicts the first equation. In any case, there are no \(a, b, c, d\) satisfying the requirements we have found, so there is no matrix \(B\) that is an inverse for \(A\text{.}\) The matrix \(A\) is not invertible.

The last example gives us a method for trying to check if a matrix is invertible: Multiply it by an arbitrary matrix, set the result equal to the identity matrix, and try to solve. Unfortunately, this method is horrendously inefficient. Even for a \(2 \times 2\) matrix we ended up with a system of \(4\) equations in \(4\) variables. For a \(3 \times 3\) matrix the system would have had \(9\) equations and \(9\) variables! Fortunately, after we develop a bit more machinery, in Subsection 4.4.4 we will be able to describe a much more efficient method for testing if a matrix is invertible (and if so, calculating the inverse).

So far we have been careful to speak of an inverse for a matrix \(A\text{,}\) because as far as we know it is possible that a single matrix could have many different inverses. In fact, that isn't true, as we now prove.

Suppose that \(A\) is an \(n \times n\) matrix and that both \(B_1\) and \(B_2\) are inverses for \(A\text{.}\) We will prove that \(B_1 = B_2\text{.}\) By definition of being inverses for \(A\) we have \(AB_1 = B_1A = AB_2 = B_2A = I_n\text{.}\) We calculate as follows:

\begin{align*} B_1 \amp = B_1I_n \\ \amp = B_1(AB_2) \\ \amp = (B_1A)B_2 \\ \amp = I_nB_2 \\ \amp = B_2 \end{align*}

Therefore any two inverses of \(A\) are actually equal, so \(A\) has at most one inverse.

Since we have shown that any given square matrix has at most one inverse, when \(A\) is invertible we will speak of the inverse of \(A\text{,}\) and we will name it \(A^{-1}\text{.}\)

Here are some elementary properties of inverses. We will use these frequently, and usually without explicitly referring back to this theorem.

We prove only the second claim, leaving the first and third as exercises. We just need to calculate:

\begin{equation*} (AB)(B^{-1}A^{-1}) = A(BB^{-1})A^{-1} = AI_nA^{-1} = AA^{-1} = I_n\text{,} \end{equation*}

and

\begin{equation*} (B^{-1}A^{-1})(AB) = B^{-1}(A^{-1}A)B = B^{-1}I_nB = B^{-1}B = I_n\text{.} \end{equation*}

Therefore \(B^{-1}A^{-1} = (AB)^{-1}\text{.}\)

Just like with numbers, knowing that \(A\) and \(B\) are invertible tells you nothing about whether or not \(A+B\) is invertible.

Subsection 4.4.2 Elementary matrices

Our next major goal is to find an efficient way of determining whether or not a matrix is invertible, and if so, finding the inverse. Both of those goals will be accomplished in Subsection 4.4.4, but in order to do so we need some preliminary material that gives us a way of connecting row operations to matrix multiplication.

Definition 4.4.7.

An \(n\times n\) elementary matrix is any matrix that can be obtained from \(I_n\) by performing exactly one row operation.

Here are some elementary matrices (see if you can work out which row operation was performed on \(I_2\) to get each of these!):

  • \(\displaystyle \begin{bmatrix}0 \amp 1 \\ 1 \amp 0\end{bmatrix}\)

  • \(\displaystyle \begin{bmatrix}-2 \amp 0 \\ 0 \amp 1\end{bmatrix}\)

  • \(\displaystyle \begin{bmatrix}1 \amp 0 \\ 5 \amp 1\end{bmatrix}\)

By contrast, the matrix \(A = \begin{bmatrix}1 \amp 1 \\ 2 \amp 1\end{bmatrix}\) is not an elementary matrix, because there is no single row operation that takes \(I_2\) to \(A\) (we could do several row operations to get from \(I_2\) to \(A\text{,}\) but the definition of elementary matrices requires that we only use one operation).

Let \(A = \begin{bmatrix}1 \amp 2 \\ 3 \amp 4\end{bmatrix}\text{.}\) If we perform the row operation \(R_1 - 2R_2\) on \(A\) then we get \(B = \begin{bmatrix}-5 \amp -6 \\ 2 \amp 3\end{bmatrix}\text{.}\) If we do the same row operation to \(I_2\) then we get the elementary matrix \(E = \begin{bmatrix}1 \amp -2 \\ 0 \amp 1\end{bmatrix}\text{.}\) If we calculate the product \(EA\text{,}\) we get:

\begin{equation*} EA = \begin{bmatrix}1 \amp -2 \\ 0 \amp 1\end{bmatrix}\begin{bmatrix}1 \amp 2 \\ 3 \amp 4\end{bmatrix} = \begin{bmatrix}-5 \amp -6 \\ 3 \amp 4 \end{bmatrix} = B\text{,} \end{equation*}

as predicted by the theorem.

The purpose of elementary matrices is that they allow us to transform questions about row operations into questions about matrix multiplication, which allows us to use the tools of matrix algebra that we have been developing.

Suppose that \(E\) is an elementary matrix, so \(E\) was obtained from \(I_n\) by a single row operation. We know that every row operation can be reversed, that is, there is some row operation that takes \(E\) back to \(I_n\text{.}\) Let \(F\) be the elementary matrix corresponding to this "reversing" row operation. By Theorem 4.4.9 the matrix \(FE\) is the matrix obtained from \(E\) by the row operation that created \(F\text{;}\) by our choice of \(F\) this means that \(FE = I_n\text{.}\)

On the other hand, the row operation that created \(E\) is also the "reverse" of the operation that created \(F\text{,}\) so by a very similar argument we also have that \(EF = I_n\text{.}\) Thus \(F = E^{-1}\text{.}\)

Subsection 4.4.3 The fundamental theorem

Recall Theorem 3.3.14, which gave us several equivalences relating to solving systems of linear equations. We are now prepared to add some very important items to that list of equivalences.

In Theorem 3.3.14 we proved the equivalences between (1), (3), (5), (7), (8), and (9). The equivalences of (3) with (4) and (5) with (6) both follow immediately from Theorem 4.1.19. To complete the proof we will prove that (1) implies (10), (10) implies (2), and (2) implies (4).

Figure 4.4.13. A diagram of the implications to be proved.

\(1 \implies 10\text{:}\) Suppose that \(\RREF(A) = I_n\text{.}\) Then there is a sequence of row operations that takes \(A\) to \(I_n\text{.}\) Let \(E_1\) be the elementary matrix corresponding to the first row operation used, \(E_2\) the elementary matrix for the second row operation used, and so on, up to \(E_k\) for the last row operation. Then by Theorem 4.4.9 we have \(E_k\cdots E_2E_1A = I_n\text{.}\) By Theorem 4.4.11 each \(E_j\) is invertible, and each \(E_j^{-1}\) is also an elementary matrix. By Theorem 4.4.6 \((E_k\cdots E_2E_1)^{-1} = E_1^{-1}E_2^{-1} \cdots E_k^{-1}\text{,}\) so multiplying both sides of \(E_k\cdots E_2E_1A = I_n\) on the left by this expression we obtain

\begin{equation*} A = E_1^{-1}E_2^{-1}\cdots E_k^{-1}\text{,} \end{equation*}

which is a product of elementary matrices.

\(10 \implies 2\text{:}\) If \(A\) can be written as a product of elementary matrices then since each elementary matrix is invertible (Theorem 4.4.11) and products of invertible matrices are invertible (Theorem 4.4.6) we conclude that \(A\) is invertible.

\(2 \implies 4\text{:}\) Suppose that \(A\) is invertible. Then the equation \(A\vec{x} = \vec{0}\) is equivalent to the equation \(\vec{x} = A^{-1}\vec{0} = \vec{0}\text{,}\) meaning that the unique solution to \(A\vec{x} = \vec{0}\) is \(\vec{x}=\vec{0}\text{.}\)

Subsection 4.4.4 Calculating matrix inverses

It might not be obvious at first glance, but the Fundamental Theorem can be used to give us a method for finding the inverse of a matrix. We will need a preliminary result, which is helpful in its own right.

Consider the equation \(A\vec{x} = \vec{0}\text{.}\) Multiplying both sides on the left by \(B\) we obtain \(BA\vec{x} = B\vec{0} = \vec{0}\text{,}\) and since \(BA = I_n\) this gives us \(\vec{x} = \vec{0}\text{.}\) That is, the equation \(A\vec{x} = \vec{0}\) has a unique solution. By Theorem 4.4.12 the matrix \(A\) is invertible. Now, to show that \(B = A^{-1}\text{,}\) we calculate:

\begin{equation*} B = BI_n = B(AA^{-1}) = (BA)A^{-1} = I_nA^{-1} = A^{-1}\text{.} \end{equation*}

Suppose that \(E_1, \ldots, E_k\) are the elementary matrices corresponding to the sequence of row operations taking \(A\) to \(I_n\text{,}\) so we have

\begin{equation*} E_k \cdots E_1 A = I_n\text{.} \end{equation*}

If we let \(B = E_k \cdots E_1\) then this equation says \(BA=I_n\text{,}\) which by Theorem 4.4.14 is enough to let us conclude that \(B = A^{-1}\text{.}\) That is,

\begin{equation*} A^{-1} = E_k \cdots E_1 = E_k \cdots E_1 I_n\text{,} \end{equation*}

which means that performing the sequence of row operations on \(I_n\) gives us \(A^{-1}\text{.}\)

Combining Theorem 4.4.12 with Theorem 4.4.15 yields an algorithm for both checking the invertibility of a matrix and finding its inverse: Set up the large augmented matrix \([A|I_n]\) and row reduce, aiming to get the left side into reduced row echelon form. If \(\RREF(A) = I_n\) then (since we are performing the same row operations on \(I_n\)), we will have \([A|I_n] \to [I_n|A^{-1}]\text{.}\) On the other hand, if \(\RREF(A) \neq I_n\) then \(A\) is not invertible.

Let \(A = \begin{bmatrix}3 \amp 3 \amp 1 \\ 0 \amp 0 \amp 1 \\ 2 \amp 2 \amp 1\end{bmatrix}\text{.}\) Determine whether or not \(A\) is invertible, and if it is, find \(A^{-1}\text{.}\)

Solution.

We set up the augmented matrix \([A|I_3]\) and row reduce.

\begin{align*} [A|I_3] \amp = \matr{ccc|ccc}{3 \amp 3 \amp 1 \amp 1 \amp 0 \amp 0 \\ 0 \amp 0 \amp 1 \amp 0 \amp 1 \amp 0 \\ 2 \amp 2 \amp 1 \amp 0 \amp 0 \amp 1}\\ \amp \to_{R_1 - R_3} \matr{ccc|ccc}{1 \amp 1 \amp 0 \amp 1 \amp 0 \amp -1 \\ 0 \amp 0 \amp 1 \amp 0 \amp 1 \amp 0 \\ 2 \amp 2 \amp 1 \amp 0 \amp 0 \amp 1}\\ \amp \to_{R_3 - 2R_1} \matr{ccc|ccc}{1 \amp 1 \amp 0 \amp 1 \amp 0 \amp -1 \\ 0 \amp 0 \amp 1 \amp 0 \amp 1 \amp 0 \\ 0 \amp 0 \amp 1 \amp -2 \amp 0 \amp 3}\\ \amp \to_{R_3-R_2}\matr{ccc|ccc}{1 \amp 1 \amp 0 \amp 1 \amp 0 \amp -1 \\ 0 \amp 0 \amp 1 \amp 0 \amp 1 \amp 0 \\ 0 \amp 0 \amp 0 \amp -2 \amp -1 \amp -3} \end{align*}

We see that \(\RREF(A) \neq I_3\text{,}\) so \(A\) is not invertible. The matrix appearing on the right side of the augmentation line has no particular meaning for us in this case.

Let \(A = \begin{bmatrix}1 \amp 1 \amp 1 \\ 1 \amp 2 \amp 1 \\ 0 \amp 0 \amp 1\end{bmatrix}\text{.}\) Determine whether or not \(A\) is invertible, and if it is, find \(A^{-1}\text{.}\)

Solution.

We set up the augmented matrix \([A|I_3]\) and row-reduce.

\begin{align*} [A|I_3] \amp = \matr{ccc|ccc}{1 \amp 1 \amp 1 \amp 1 \amp 0 \amp 0 \\ 1 \amp 2 \amp 1 \amp 0 \amp 1 \amp 0 \\ 0 \amp 0 \amp 1 \amp 0 \amp 0 \amp 1}\\ \amp \to_{R_2 - R_1} \matr{ccc|ccc}{1 \amp 1 \amp 1 \amp 1 \amp 0 \amp 0 \\ 0 \amp 1 \amp 0 \amp -1 \amp 1 \amp 0 \\ 0 \amp 0 \amp 1 \amp 0 \amp 0 \amp 1}\\ \amp \to_{R_1-R_2} \matr{ccc|ccc}{1 \amp 0 \amp 1 \amp 2 \amp -1 \amp 0 \\ 0 \amp 1 \amp 0 \amp -1 \amp 1 \amp 0 \\ 0 \amp 0 \amp 1 \amp 0 \amp 0 \amp 1}\\ \amp \to_{R_1-R_3}\matr{ccc|ccc}{1 \amp 0 \amp 0 \amp 2 \amp -1 \amp -1 \\ 0 \amp 1 \amp 0 \amp -1 \amp 1 \amp 0 \\ 0 \amp 0 \amp 1 \amp 0 \amp 0 \amp 1} \end{align*}

This calculation shows that \(\RREF(A) = I_3\text{,}\) so \(A\) is invertible. It also shows that \(A^{-1} = \begin{bmatrix}2 \amp -1 \amp -1 \\ -1 \amp 1 \amp 0 \\ 0 \amp 0 \amp 1\end{bmatrix}\text{.}\)

Subsection 4.4.5 The inverse of a linear transformation

Definition 4.4.18.

Let \(T : \mathbb{R}^n \to \mathbb{R}^n\) be a linear transformation. The inverse for \(T\text{,}\) if it exists, is a function \(T^{-1} : \mathbb{R}^n \to \mathbb{R}^n\) such that for every \(\vec{v}\) in \(\mathbb{R}^n\text{,}\)

\begin{equation*} (T \circ T^{-1})(\vec{v}) = \vec{v} = (T^{-1} \circ T)(\vec{v})\text{.} \end{equation*}

This is the same definition of "inverse function" that you have likely encountered in other mathematics courses. If you have believed the slogan that linear transformations and matrices are fundamentally the same thing, then the next result is probably not surprising.

Suppose first that \([T]\) is invertible, and let \(S : \mathbb{R}^n \to \mathbb{R}^n\) be defined by \(S(\vec{v}) = [T]^{-1}\vec{v}\text{.}\) By Theorem 4.1.18 \(S\) is a linear transformation, and \([S] = [T]^{-1}\text{.}\) Using Theorem 4.1.13 we have, for any \(\vec{v}\text{,}\)

\begin{equation*} (S \circ T)(\vec{v}) = S(T(\vec{v})) = S([T]\vec{v}) = [S][T]\vec{v} = [T]^{-1}[T]\vec{v} = \vec{v}\text{,} \end{equation*}

and likewise

\begin{equation*} (T \circ S)(\vec{v}) = [T][S]\vec{v} = [T][T]^{-1}\vec{v} = \vec{v}\text{.} \end{equation*}

Thus \(T\) is invertible.

Now suppose that \(T\) is invertible, with inverse function \(T^{-1}\text{.}\) To show that \(T^{-1}\) is a linear transformation, suppose that \(\vec{v}\) and \(\vec{w}\) are vectors in \(\mathbb{R}^n\text{,}\) and \(c\) is a scalar. Then:

\begin{equation*} \vec{v}+\vec{w} = T(T^{-1}(\vec{v})) + T(T^{-1}(\vec{w})) = T(T^{-1}(\vec{v}) + T^{-1}(\vec{w}))\text{,} \end{equation*}

and applying \(T^{-1}\) on both sides then gives us

\begin{equation*} T^{-1}(\vec{v}+\vec{w}) = T^{-1}(\vec{v}) + T^{-1}(\vec{w})\text{.} \end{equation*}

Similarly,

\begin{equation*} c\vec{v} = cT(T^{-1}(\vec{v})) = T(cT^{-1}(\vec{v}))\text{,} \end{equation*}

and applying \(T^{-1}\) to both sides gives

\begin{equation*} T^{-1}(c\vec{v}) = cT^{-1}(\vec{v})\text{.} \end{equation*}

Now since we have shown that \(T^{-1}\) is a linear transformation it has a matrix \([T^{-1}]\text{.}\) By Theorem 4.1.13, for any \(\vec{v}\text{,}\)

\begin{equation*} [T^{-1}][T]\vec{v} = [T^{-1}]T(\vec{v}) = T^{-1}(T(\vec{v})) = \vec{v}\text{,} \end{equation*}

from which it follows that \([T^{-1}][T] = I_n\text{.}\) This is enough to prove that \([T]\) is invertible, and also that \([T^{-1}] = [T]^{-1}\text{,}\) by Theorem 4.4.14.

Let \(T : \mathbb{R}^3 \to \mathbb{R}^3\) be given by \(T\left(\begin{bmatrix}x\\y\\z\end{bmatrix}\right) = \begin{bmatrix}x+y+z\\x-y\\x\end{bmatrix}\text{.}\) Show that \(T\) is invertible, and find a formula for \(T^{-1}\left(\begin{bmatrix}x\\y\\z\end{bmatrix}\right)\text{.}\)

Solution.

The matrix of \(T\) is \([T] = \begin{bmatrix}1 \amp 1 \amp 1 \\ 1 \amp -1 \amp 0 \\ 1 \amp 0 \amp 0\end{bmatrix}\text{.}\) Using our method for finding the inverse,

\begin{equation*} \matr{ccc|ccc}{1 \amp 1 \amp 1 \amp 1 \amp 0 \amp 0 \\ 1 \amp -1 \amp 0 \amp 0 \amp 1 \amp 0 \\ 1 \amp 0 \amp 0 \amp 0 \amp 0 \amp 1} \to \matr{ccc|ccc}{1 \amp 0 \amp 0 \amp 0 \amp 0 \amp 1 \\ 0 \amp 1 \amp 0 \amp 0 \amp -1 \amp 1 \\ 0 \amp 0 \amp 1 \amp 1 \amp 1 \amp -2}\text{.} \end{equation*}

Thus \([T]^{-1} = \begin{bmatrix}0 \amp 0 \amp 1 \\ 0 \amp -1 \amp 1 \\ 1 \amp 1 \amp -2\end{bmatrix}\text{.}\) By Theorem 4.4.19 we see that \(T\) is invertible, and moreover that \([T^{-1}] = [T]^{-1}\text{.}\) Thus, using Theorem 4.1.13 we have that for any \(\begin{bmatrix}x\\y\\z\end{bmatrix}\) in \(\mathbb{R}^3\text{,}\)

\begin{equation*} T^{-1}\left(\begin{bmatrix}x\\y\\z\end{bmatrix}\right) = [T^{-1}]\begin{bmatrix}x\\y\\z\end{bmatrix} = [T]^{-1}\begin{bmatrix}x\\y\\z\end{bmatrix} = \begin{bmatrix}0 \amp 0 \amp 1 \\ 0 \amp -1 \amp 1 \\ 1 \amp 1 \amp -2\end{bmatrix}\begin{bmatrix}x\\y\\z\end{bmatrix} = \begin{bmatrix}z \\ -y+z \\ x+y-2z\end{bmatrix}\text{.} \end{equation*}

Exercises 4.4.6 Exercises

1.

For each of the following matrices, find the inverse if possible. If it doesn't exist, explain why.
  1. \(\displaystyle \begin{bmatrix} 2 \amp 1 \\ -1 \amp 3 \end{bmatrix} \)

  2. \(\displaystyle \begin{bmatrix} 0 \amp 1 \\ 5 \amp 3 \end{bmatrix} \)

  3. \(\displaystyle \begin{bmatrix} 2 \amp 1 \\ 3 \amp 0 \end{bmatrix} \)

  4. \(\displaystyle \begin{bmatrix} 2 \amp 1 \\ 4 \amp 2 \end{bmatrix} \)

  5. \(\displaystyle \begin{bmatrix} 0 \amp 1 \amp 2 \\ 1 \amp 2 \amp 5 \end{bmatrix} \)

Hint.
To find the inverse of a matrix \(A\) of size \(n\times n\text{,}\) one writes down the augmented matrix
\begin{equation*} \matr{c|c} {A\amp I_{n}} \end{equation*}
and then row-reduces. If the RREF of \(A\) is \(I_{n}\) (the identity matrix of size \(n\times n\)), so that the above augmented matrix becomes
\begin{equation*} \matr{c|c} {A\amp I_{n}} \overset{\text{row reduction}}{\longrightarrow} \matr{c|c} {I_{n}\amp B} \end{equation*}
for some matrix \(B\text{,}\) then \(B\) is the inverse \(A^{-1}\) of \(A\text{.}\) If the RREF of \(A\) is not \(I_{n}\text{,}\) then \(A\) is not invertible.
Answer.

  1. Answer.

    \(\begin{bmatrix} 2 \amp 1 \\ -1 \amp 3 \end{bmatrix}^{-1} = \frac{1}{7}\begin{bmatrix} 3 \amp -1 \\ 1 \amp 2 \end{bmatrix}\)

  2. Answer.

    \(\begin{bmatrix} 0 \amp 1 \\ 5 \amp 3 \end{bmatrix}^{-1}=\frac{1}{5}\begin{bmatrix} -3 \amp 1 \\ 5 \amp 0 \end{bmatrix}\)

  3. Answer.

    \(\begin{bmatrix} 2 \amp 1 \\ 3 \amp 0 \end{bmatrix}^{-1} = \frac{1}{3}\begin{bmatrix} 0 \amp 1 \\ 3 \amp -2 \end{bmatrix}\)

  4. Answer.

    The matrix \(\begin{bmatrix} 2 \amp 1 \\ 4 \amp 2 \end{bmatrix}\) does not have an inverse.

  5. Answer.

    The matrix \(\begin{bmatrix} 0 \amp 1 \amp 2 \\ 1 \amp 2 \amp 5 \end{bmatrix} \) does not have an inverse.

Solution.

  1. Solution.

    We will use the approach explained in Hint 4.4.6.1.1.
    \begin{align*} \matr{cc|cc} { 2 \amp 1 \amp 1 \amp 0 \\ -1 \amp 3 \amp 0 \amp 1 } \overset{R_{1}+2R_{2}}{\longrightarrow} \amp \matr{cc|cc} { 2 \amp 1 \amp 1 \amp 0 \\ 0 \amp 7 \amp 1 \amp 2 }\\ \overset{-R_{2}+7R_{1}}{\longrightarrow} \amp \matr{cc|cc} { 14 \amp 0 \amp 6 \amp -2 \\ 0 \amp 7 \amp 1 \amp 2 }\\ \overset{\frac{1}{14}R_{1}}{\underset{\frac{1}{7}R_{2}}{\longrightarrow}} \amp \matr{cc|cc} { 1 \amp 0 \amp \frac{3}{7} \amp -\frac{1}{7} \\ 0 \amp 1 \amp \frac{1}{7} \amp \frac{2}{7} }. \end{align*}
    We conclude that \(\begin{bmatrix} 2 \amp 1 \\ -1 \amp 3 \end{bmatrix}^{-1} = \frac{1}{7}\begin{bmatrix} 3 \amp -1 \\ 1 \amp 2 \end{bmatrix}.\)

  2. Solution.

    \begin{align*} \matr{cc|cc} { 0 \amp 1 \amp 1 \amp 0 \\ 5 \amp 3 \amp 0 \amp 1 } \overset{R_{1}\leftrightarrow R_{2}}{\longrightarrow} \amp \matr{cc|cc} { 5 \amp 3 \amp 0 \amp 1 \\ 0 \amp 1 \amp 1 \amp 0 }\\ \overset{-3R_{2}+R_{1}}{\longrightarrow} \amp \matr{cc|cc} { 5 \amp 0 \amp -3 \amp 1 \\ 0 \amp 1 \amp 1 \amp 0 }\\ \overset{\frac{1}{5}R_{1}}{\longrightarrow} \amp \matr{cc|cc} { 1 \amp 0 \amp \frac{-3}{5} \amp \frac{1}{5} \\ 0 \amp 1 \amp 1 \amp 0 } \end{align*}
    We conclude that \(\begin{bmatrix} 0 \amp 1 \\ 5 \amp 3 \end{bmatrix}^{-1}=\frac{1}{5}\begin{bmatrix} -3 \amp 1 \\ 5 \amp 0 \end{bmatrix}\text{.}\)

  3. Solution.

    \begin{align*} \matr{cc|cc} { 2 \amp 1 \amp 1 \amp 0 \\ 3 \amp 0 \amp 0 \amp 1 } \overset{R_{1}\leftrightarrow R_{2}}{\longrightarrow} \amp \matr{cc|cc} { 3 \amp 0 \amp 0 \amp 1 \\ 2 \amp 1 \amp 1 \amp 0 }\\ \overset{-2R_{1}+3R_{2}}{\longrightarrow} \amp \matr{cc|cc} { 3 \amp 0 \amp 0 \amp 1 \\ 0 \amp 3 \amp 3 \amp -2 }\\ \overset{\frac{1}{3}R_{1}}{\underset{\frac{1}{3}R_{2}}{\longrightarrow}} \amp \matr{cc|cc} { 1 \amp 0 \amp 0 \amp \frac{1}{3} \\ 0 \amp 1 \amp 1 \amp -\frac{2}{3} } \end{align*}
    We conclude that \(\begin{bmatrix} 2 \amp 1 \\ 3 \amp 0 \end{bmatrix}^{-1} = \frac{1}{3}\begin{bmatrix} 0 \amp 1 \\ 3 \amp -2 \end{bmatrix}\text{.}\)

  4. Solution.

    \begin{align*} \matr{cc|cc} { 2 \amp 1 \amp 1 \amp 0 \\ 4 \amp 2 \amp 0 \amp 1 } \overset{-2R_{1}+R_{2}}{\longrightarrow} \amp \matr{cc|cc} { 2 \amp 1 \amp 1 \amp 0 \\ 0 \amp 0 \amp -2 \amp 1 } \end{align*}
    We see that the left-hand side will not have the identity matrix as its RREF. Therefore, the matrix \(\begin{bmatrix} 2 \amp 1 \\ 4 \amp 2 \end{bmatrix}\) does not have an inverse.

  5. Solution.

    The matrix \(\begin{bmatrix} 0 \amp 1 \amp 2 \\ 1 \amp 2 \amp 5 \end{bmatrix} \) is not a square matrix, so it cannot be invertible.

2.

Let \(A \) be a \(2 \times 2 \) invertible matrix, with \(A = \begin{bmatrix} a \amp b \\ c \amp d \end{bmatrix}. \) Find a formula for \(A^{-1} \) in terms of \(a,b,c,d. \)
Hint.
You can use Hint 4.4.6.1.1. Be careful not to divide by zero (you need to make a case distinction).
Answer.
If \(A = \begin{bmatrix} a \amp b \\ c \amp d \end{bmatrix}\) is invertible, its inverse is given by
\begin{equation*} \begin{bmatrix} a \amp b \\ c \amp d \end{bmatrix}^{-1} = \frac{1}{ad-bc} \begin{bmatrix} d \amp -b \\ -c \amp a \end{bmatrix} . \end{equation*}
Solution.

We use the approach from Hint 4.4.6.1.1, i.e. we want to row-reduce the following augmented matrix:
\begin{align*} \matr{cc|cc} { a \amp b \amp 1 \amp 0 \\ c \amp d \amp 0 \amp 1 } \end{align*}
But we need to be careful not to divide by zero in the process. By assumption, the matrix \(A\) is invertible. In particular, according to Hint 4.4.6.1.1, the reduced row-echelon form of \(A\) is the identity matrix. Therefore, if \(a=0\text{,}\) then neither \(c\) nor \(b\) can be zero (because else the first column resp. row of \(A\) would be zero). This means that we can make a case distinction: Assume first that \(a\neq 0\text{,}\) so that multiplying a row by \(a\) is a valid row operation:
\begin{align*} \matr{cc|cc} { a \amp b \amp 1 \amp 0 \\ c \amp d \amp 0 \amp 1 } \overset{aR_{2}}{\longrightarrow} \amp \matr{cc|cc} { a \amp b \amp 1 \amp 0 \\ ac \amp ad \amp 0 \amp a }\\ \overset{-cR_{1} + R_{2}}{\longrightarrow} \amp\matr{cc|cc} { a \amp b \amp 1 \amp 0 \\ 0 \amp ad-bc \amp -c \amp a } . \end{align*}
Since the RREF of \(A\) is the identity (by the assumption that \(A\) is invertible), we conclude that \(\det(A) = ad-bc \neq 0\text{.}\) We proceed:
\begin{align*} \matr{cc|cc}{ a \amp b \amp 1 \amp 0 \\ 0 \amp ad-bc \amp -c \amp a } \overset{\frac{1}{\det(A)}R_{2}}{\longrightarrow} \amp \matr{cc|cc} { a \amp b \amp 1 \amp 0 \\ 0 \amp 1 \amp \frac{-c}{\det(A)} \amp \frac{a}{\det(A)} }\\ \overset{-bR_{2}+R_{1}}{\longrightarrow} \amp \matr{cc|cc} { a \amp 0 \amp 1 + \frac{bc}{\det(A)} \amp \frac{-ab}{\det(A)} \\ 0 \amp 1 \amp \frac{-c}{\det(A)} \amp \frac{a}{\det(A)} }\\ =\amp \matr{cc|cc} { a \amp 0 \amp \frac{ad}{\det(A)} \amp \frac{-ab}{\det(A)} \\ 0 \amp 1 \amp \frac{-c}{\det(A)} \amp \frac{a}{\det(A)} }\\ \overset{\frac{1}{a}R_{1}}{\longrightarrow} \amp \matr{cc|cc} { 1 \amp 0 \amp \frac{d}{\det(A)} \amp \frac{-b}{\det(A)} \\ 0 \amp 1 \amp \frac{-c}{\det(A)} \amp \frac{a}{\det(A)} } , \end{align*}
where the last operation is valid since \(a\neq 0\text{.}\) We conclude: If \(A= \begin{bmatrix} a \amp b \\ c \amp d \end{bmatrix}\) is invertible and \(a\neq 0\text{,}\) then \(\det(A) \neq 0\text{,}\) and \(A^{-1} =\frac{1}{\det(A)} \begin{bmatrix} d \amp -b \\ -c \amp a \end{bmatrix}.\) The case where \(a=0\) works similarly: As explained above, this implies that \(c\neq 0\) and \(b\neq 0\text{.}\) This allows us to do the following row operations:
\begin{align*} \matr{cc|cc} { a \amp b \amp 1 \amp 0 \\ c \amp d \amp 0 \amp 1 } = \amp \matr{cc|cc} { 0 \amp b \amp 1 \amp 0 \\ c \amp d \amp 0 \amp 1 }\\ \overset{ \frac{1}{c}R_{2}}{\underset{\frac{1}{b}R_{1}}{\longrightarrow}} \amp \matr{cc|cc} { 0 \amp 1 \amp \frac{1}{b} \amp 0 \\ 1 \amp \frac{d}{c} \amp 0 \amp \frac{1}{c} }\\ \overset{R_{1}\leftrightarrow R_{2}}{\longrightarrow} \amp \matr{cc|cc} { 1 \amp \frac{d}{c} \amp 0 \amp \frac{1}{c} \\ 0 \amp 1 \amp \frac{1}{b} \amp 0 }\\ \overset{-\frac{d}{c}R_{1}+R_{2}}{\longrightarrow} \amp \matr{cc|cc} { 1 \amp 0 \amp -\frac{d}{c}\cdot \frac{1}{b} \amp \frac{1}{c} \\ 0 \amp 1 \amp \frac{1}{b} \amp 0 } . \end{align*}
In the case where \(a=0\text{,}\) we have found the inverse of \(A\) to be
\begin{equation*} A^{-1} = \begin{bmatrix} -\frac{d}{c}\cdot \frac{1}{b} \amp \frac{1}{c} \\ \frac{1}{b} \amp 0 \end{bmatrix} = \frac{1}{-bc} \begin{bmatrix} d \amp -b \\ -c \amp a \end{bmatrix} = \frac{1}{\det (A)} \begin{bmatrix} d \amp -b \\ -c \amp a \end{bmatrix}. \end{equation*}
All in all, we have shown: If \(A= \begin{bmatrix} a \amp b \\ c \amp d \end{bmatrix}\) is invertible (where \(a\) is any value), then \(\det(A) \neq 0\text{,}\) and \(A^{-1} =\frac{1}{\det(A)} \begin{bmatrix} d \amp -b \\ -c \amp a \end{bmatrix}.\)

3.

Using the inverse of the matrix, find the solution to the systems:
  1. \begin{equation*} \begin{bmatrix} 2 \amp 4 \\ 1 \amp 1 \end{bmatrix} \begin{bmatrix}x \\ y \end{bmatrix} = \begin{bmatrix} 1 \\ 2 \end{bmatrix} \end{equation*}

  2. \begin{equation*} \begin{bmatrix} 2 \amp 4 \\ 1 \amp 1 \end{bmatrix} \begin{bmatrix}x \\ y \end{bmatrix} = \begin{bmatrix} 2 \\ 0 \end{bmatrix} \end{equation*}

  3. \begin{equation*} \begin{bmatrix} 2 \amp 4 \\ 1 \amp 1 \end{bmatrix} \begin{bmatrix}x \\ y \end{bmatrix} = \begin{bmatrix} a \\ b \end{bmatrix} \end{equation*}
    (give the solution in terms of \(a \) and \(b \))

Hint 1.
It is easiest to use Answer 4.4.6.2.1 to determine the inverse matrix.
Hint 2.
Note that, given an invertible matrix \(A\text{,}\) we can multiply the vector equation
\begin{equation*} A \vec{x} = \vec{y} \end{equation*}
on the left by \(A^{-1}\text{,}\) which yields the equivalent equation:
\begin{equation*} \vec{x} = A^{-1}\vec{y} . \end{equation*}
Answer.

  1. Answer.

    \(\begin{bmatrix} x \\ y \end{bmatrix}=\begin{bmatrix} \frac{7}{2} \\ \frac{-3}{2} \end{bmatrix}\)

  2. Answer.

    \(\begin{bmatrix} x \\ y \end{bmatrix}=\begin{bmatrix} -1 \\ 1 \end{bmatrix}\)

  3. Answer.

    \(\begin{bmatrix} x \\ y \end{bmatrix}=\begin{bmatrix} \frac{-a}{2} + 2 b \\ \frac{a}{2} - b \end{bmatrix}\)

Solution.

First, we determine the inverse of \(\begin{bmatrix} 2 \amp 4 \\ 1 \amp 1 \end{bmatrix} \) using Answer 4.4.6.2.1:
\begin{equation*} \begin{bmatrix} 2 \amp 4 \\ 1 \amp 1 \end{bmatrix}^{-1} = \frac{1}{2\cdot 1 - 4\cdot 1} \begin{bmatrix} 1 \amp -4 \\ -1 \amp 2 \end{bmatrix} = \frac{1}{2} \begin{bmatrix} -1 \amp 4 \\ 1 \amp -2 \end{bmatrix} . \end{equation*}
  1. Solution.

    Note that
    \begin{equation*} \begin{bmatrix} 2 \amp 4 \\ 1 \amp 1 \end{bmatrix} \begin{bmatrix}x \\ y \end{bmatrix} = \begin{bmatrix} 1 \\ 2 \end{bmatrix} \iff \begin{bmatrix}x \\ y \end{bmatrix} = \begin{bmatrix} 2 \amp 4 \\ 1 \amp 1 \end{bmatrix}^{-1} \begin{bmatrix} 1 \\ 2 \end{bmatrix}. \end{equation*}
    We compute:
    \begin{equation*} \begin{bmatrix} 2 \amp 4 \\ 1 \amp 1 \end{bmatrix}^{-1} \begin{bmatrix} 1 \\ 2 \end{bmatrix} = \frac{1}{2} \begin{bmatrix} -1 \amp 4 \\ 1 \amp -2 \end{bmatrix} \begin{bmatrix} 1 \\ 2 \end{bmatrix} = \frac{1}{2} \begin{bmatrix} -1 + 4\cdot 2 \\ 1 - 2 \cdot 2 \end{bmatrix} = \begin{bmatrix} \frac{7}{2} \\ \frac{-3}{2} \end{bmatrix} . \end{equation*}
    The solution is therefore given by \(\begin{bmatrix} x \\ y \end{bmatrix}=\begin{bmatrix} \frac{7}{2} \\ \frac{-3}{2} \end{bmatrix}.\)

  2. Solution.

    Note that
    \begin{equation*} \begin{bmatrix} 2 \amp 4 \\ 1 \amp 1 \end{bmatrix} \begin{bmatrix}x \\ y \end{bmatrix} = \begin{bmatrix} 2 \\ 0 \end{bmatrix} \iff \begin{bmatrix}x \\ y \end{bmatrix} = \begin{bmatrix} 2 \amp 4 \\ 1 \amp 1 \end{bmatrix}^{-1}\begin{bmatrix} 2 \\ 0 \end{bmatrix} . \end{equation*}
    We compute:
    \begin{equation*} \begin{bmatrix} 2 \amp 4 \\ 1 \amp 1 \end{bmatrix}^{-1}\begin{bmatrix} 2 \\ 0 \end{bmatrix} = \frac{1}{2} \begin{bmatrix} -1 \amp 4 \\ 1 \amp -2 \end{bmatrix} \begin{bmatrix} 2 \\ 0 \end{bmatrix} = \frac{1}{2} \begin{bmatrix} -2 + 4\cdot 0 \\ 2 -2\cdot 0 \end{bmatrix} = \begin{bmatrix} -1 \\ 1 \end{bmatrix} . \end{equation*}
    The solution is therefore given by \(\begin{bmatrix} x \\ y \end{bmatrix}=\begin{bmatrix} -1 \\ 1 \end{bmatrix}\text{.}\)

  3. Solution.

    Just as before, we see that
    \begin{equation*} \begin{bmatrix} 2 \amp 4 \\ 1 \amp 1 \end{bmatrix} \begin{bmatrix}x \\ y \end{bmatrix} = \begin{bmatrix} a \\ b \end{bmatrix} \iff \begin{bmatrix}x \\ y \end{bmatrix} = \begin{bmatrix} 2 \amp 4 \\ 1 \amp 1 \end{bmatrix}^{-1}\begin{bmatrix} a \\ b \end{bmatrix}. \end{equation*}
    We compute:
    \begin{equation*} \begin{bmatrix} 2 \amp 4 \\ 1 \amp 1 \end{bmatrix}^{-1} \begin{bmatrix} a \\ b \end{bmatrix} = \frac{1}{2} \begin{bmatrix} -1 \amp 4 \\ 1 \amp -2 \end{bmatrix} \begin{bmatrix} a \\ b \end{bmatrix} = \frac{1}{2} \begin{bmatrix} -a + 4 b \\ a - 2b \end{bmatrix}. \end{equation*}
    The solution is therefore given by \(\begin{bmatrix} x \\ y \end{bmatrix}=\begin{bmatrix} \frac{-a}{2} + 2 b \\ \frac{a}{2} - b \end{bmatrix}.\)

4.

For each of the following matrices, determine whether \(B \) is an inverse of \(A. \)
  1. \begin{equation*} A = \begin{bmatrix} 2 \amp 4 \\ 1 \amp 4 \end{bmatrix}, \ \ B = \frac{1}{2}\begin{bmatrix} 3 \amp -4 \\ -1 \amp 2 \end{bmatrix}. \end{equation*}

  2. \begin{equation*} A = \begin{bmatrix} 1 \amp -2 \\ 4 \amp -7 \end{bmatrix}, \ \ B = \begin{bmatrix} -1 \amp 2 \\ -4 \amp 7 \end{bmatrix}. \end{equation*}

  3. \begin{equation*} A = \begin{bmatrix} 4 \amp 1 \amp 3 \\ 2 \amp 1 \amp 2 \\ 1 \amp 0 \amp 1 \end{bmatrix}, \ \ B = \begin{bmatrix} 1 \amp -1 \amp -1 \\ 0 \amp 1 \amp -2 \\ -1 \amp 1 \amp 2 \end{bmatrix}. \end{equation*}

Hint.
You can use Answer 4.4.6.2.1 to find the inverse of a \(2\times 2\) matrix.
Answer.

  1. Answer.

    No.

  2. Answer.

    No.

  3. Answer.

    Yes.

Solution.

In the situation of \(2\times 2\) matrices, we can approach the question in two different ways: one option is using Answer 4.4.6.2.1 to compute the inverse \(A^{-1}\) of \(A\) and then check whether it coincides with the given matrix \(B\text{.}\) The second option is to check whether \(AB=I\) and \(BA=I\text{,}\) which is exactly the case if \(B=A^{-1}\text{.}\) This latter approach also works for matrices of larger size.
  1. Solution.

    For this example, we use Answer 4.4.6.2.1 to compute \(A^{-1}\text{:}\)
    \begin{align*} A^{-1} =\amp \begin{bmatrix} 2 \amp 4 \\ 1 \amp 4 \end{bmatrix}^{-1} = \frac{1}{2\cdot 4 - 1\cdot 4} \begin{bmatrix} 4 \amp -4 \\ -1 \amp 2 \end{bmatrix}\\ =\amp \frac{1}{4} \begin{bmatrix} 4 \amp -4 \\ -1 \amp 2 \end{bmatrix} \neq \frac{1}{2}\begin{bmatrix} 3 \amp -4 \\ -1 \amp 2 \end{bmatrix} = B. \end{align*}
    We conclude that, no, \(B\) is not the inverse of \(A\text{.}\)

  2. Solution.

    For this example, we will actually check whether \(B\) behaves like the inverse of \(A\text{,}\) so we compute:
    \begin{align*} AB \amp= \begin{bmatrix} 1 \amp -2 \\ 4 \amp -7 \end{bmatrix} \begin{bmatrix} -1 \amp 2 \\ -4 \amp 7 \end{bmatrix}\\ \amp= \begin{bmatrix} 1\cdot (-1) - 2 \cdot (-4) \amp 1 \cdot 2 - 2 \cdot 7 \\ 4\cdot (-1) - 7 \cdot (-4) \amp 4 \cdot 2 - 7 \cdot 7 \end{bmatrix} = \begin{bmatrix} 7 \amp -12 \\ 24 \amp -41 \end{bmatrix}. \end{align*}
    Since \(AB \neq I\text{,}\) we conclude that, no, \(B\) is not the inverse of \(A\text{.}\) (We note that it was not necessary to compute all entries of \(AB\text{;}\) it suffices to see that its \((1,1)\)-entry is not equal to \(1\text{.}\))

  3. Solution.

    We compute:
    \begin{align*} AB \amp= \begin{bmatrix} 4 \amp 1 \amp 3 \\ 2 \amp 1 \amp 2 \\ 1 \amp 0 \amp 1 \end{bmatrix} \begin{bmatrix} 1 \amp -1 \amp -1 \\ 0 \amp 1 \amp -2 \\ -1 \amp 1 \amp 2 \end{bmatrix}\\ \amp= \begin{bmatrix} 4 +0 -3 \amp -4+1+3 \amp -4-2+6 \\ 2+0-2 \amp -2+1+2 \amp -2-2+4 \\ 1+0-1 \amp -1+0+1 \amp -1+0+2 \end{bmatrix} = \begin{bmatrix} 1 \amp 0 \amp 0 \\ 0 \amp 1 \amp 0 \\ 0 \amp 0 \amp 1 \end{bmatrix}. \end{align*}
    At this point, there is hope that \(B\) is the inverse of \(A\text{,}\) but we still need to check the other equation, so we compute:
    \begin{align*} BA \amp= \begin{bmatrix} 1 \amp -1 \amp -1 \\ 0 \amp 1 \amp -2 \\ -1 \amp 1 \amp 2 \end{bmatrix} \begin{bmatrix} 4 \amp 1 \amp 3 \\ 2 \amp 1 \amp 2 \\ 1 \amp 0 \amp 1 \end{bmatrix}\\ \amp= \begin{bmatrix} 4-2-1 \amp 1-1+0 \amp 3-2-1 \\ 0+2-2 \amp 0+1+0 \amp 0+2-2 \\ -4+2+2 \amp -1+1+0 \amp -3+2+2 \end{bmatrix} = \begin{bmatrix} 1 \amp 0 \amp 0 \\ 0 \amp 1 \amp 0 \\ 0 \amp 0 \amp 1 \end{bmatrix}. \end{align*}
    We conclude that, indeed, \(B\) is the inverse of \(A\text{.}\)

5.

Suppose \(A, B\) are two invertible matrices of the same size. Show that \((AB)^{-1} = B^{-1}A^{-1} \) by verifying that
\begin{equation*} A B (B^{-1}A^{-1}) =I \quad\text{and}\quad B^{-1}A^{-1} (AB) =I. \end{equation*}
Solution.

Recall that \(A^{-1}\) is the unique matrix that satisfies \(AA^{-1}=I=A^{-1}A\text{,}\) and similarly for \(B^{-1}\text{.}\) We use this in the following computation:
\begin{align*} A B (B^{-1}A^{-1}) \amp= A (B B^{-1})A^{-1} \amp \text{by associativity of matrix multiplication}\\ \amp= A (I)A^{-1} \amp \text{since } B B^{-1}=I\\ \amp= A A^{-1} \amp \text{by the properties of } I\\ \amp= I \amp\text{since } AA^{-1} = I. \end{align*}
For the other equality, we proceed similarly:
\begin{align*} B^{-1}A^{-1} (AB) \amp= B^{-1}(A^{-1} A)B \amp \text{by associativity of matrix multiplication}\\ \amp= B^{-1}(I)B \amp \text{since } A^{-1}A=I\\ \amp= B^{-1}B \amp \text{by the properties of } I\\ \amp= I \amp\text{since } B^{-1}B=I. \end{align*}

6.

Let \(A \) and \(B \) denote \(n \times n \) invertible matrices.
  1. Show that \(A^{-1} + B^{-1} = A^{-1}(A + B) B^{-1}. \)

    Hint.

    Recall that, for a matrix \(A\text{,}\) its inverse \(A^{-1}\) is uniquely determined by satisfying the property \(AA^{-1}=I\) and \(A^{-1}A=I\text{.}\) This means: If a matrix \(C\) satisfies \(AC=I\) and \(CA=I\text{,}\) then \(C=A^{-1}\text{.}\)

    Solution.

    By substracting \(B^{-1}\) from each side of the equation, we see that the claim is equivalent to showing that \(A^{-1} = A^{-1}(A + B) B^{-1} - B^{-1}. \) Recall that \(A^{-1}\) is the unique matrix that satisfies both \(AA^{-1}=I\) and \(A^{-1}A=I\text{,}\) so it suffices to show that \(A^{-1}(A + B) B^{-1} - B^{-1}\) has this property. We compute (using Theorem 2.1.1 in Section 2.1 of [4]):
    \begin{align*} \amp A \bigl( A^{-1}(A + B) B^{-1} - B^{-1} \bigr) \amp\\ \quad\amp= AA^{-1}(A + B) B^{-1} - AB^{-1} \amp\text{by distributivity}\\ \quad\amp= I(A + B) B^{-1} - AB^{-1} \amp\text{by the properties of } A^{-1}\\ \amp = I(AB^{-1} + BB^{-1}) - AB^{-1} \amp\text{by distributivity}\\ \quad\amp= AB^{-1} + BB^{-1} - AB^{-1} \amp\text{by the properties of } I\\ \amp = AB^{-1} + I - AB^{-1} \amp\text{by the properties of } B^{-1}\\ \quad\amp= I. \end{align*}
    Similarly,
    \begin{align*} \amp\bigl( A^{-1}(A + B) B^{-1} - B^{-1} \bigr) A\\ \quad\amp= A^{-1}(A + B) B^{-1}A - B^{-1}A \amp\text{by distributivity}\\ \quad\amp= (A^{-1}A + A^{-1}B) B^{-1}A - B^{-1}A \amp\text{by distributivity}\\ \amp = (I + A^{-1}B) B^{-1}A - B^{-1}A \amp\text{by the properties of } A^{-1}\\ \quad\amp= IB^{-1}A + A^{-1}BB^{-1}A - B^{-1}A \amp\text{by distributivity}\\ \quad\amp= B^{-1}A + A^{-1}IA - B^{-1}A \amp\text{by the properties of } I \text{ and } B^{-1}\\ \quad\amp= A^{-1}A = I \amp\text{by the properties of } A^{-1}. \end{align*}
    We have checked that the matrix \(A^{-1}(A + B) B^{-1} - B^{-1} \) indeed satisfies the defining property of \(A^{-1}\text{.}\)

  2. If \(A + B \) is also invertible, show that \(A^{-1} + B^{-1} \) is invertible and find a formula for \((A^{-1} + B^{-1})^{-1}. \)

    Hint.

    Use the first part of the exercise. Then use Exercise 4.4.6.5.

    Solution.

    By assumption, \(C:= A + B \) is invertible. By the first part of the exercise, we know that \(A^{-1} + B^{-1} = A^{-1}(A + B) B^{-1},\) i.e. \(A^{-1} + B^{-1}= A^{-1}CB^{-1}.\) The matrix \(A^{-1}CB^{-1}\) is invertible as product of invertible matrices by Exercise 4.4.6.5, which shows that \(A^{-1} + B^{-1}\) is invertible. Exercise 4.4.6.5 also gives a formula for the inverse of a product, which we use for the equalities \((*)\) in the following computation:
    \begin{align*} (A^{-1}CB^{-1})^{-1} \amp=(A^{-1}[CB^{-1}])^{-1} \overset{(*)}{=}[CB^{-1}]^{-1} (A^{-1})^{-1}\\ \amp\overset{(*)}{=}(B^{-1})^{-1} C^{-1} A =B C^{-1} A. \end{align*}
    We conclude: The inverse of \(A^{-1} + B^{-1} \) is given by \(B C^{-1} A = B (A + B)^{-1} A.\)

    Answer.

    The inverse of \(A^{-1} + B^{-1} \) is given by \(B (A + B)^{-1} A.\)

7.

In each case, find the elementary matrix \(E \) such that \(B = EA. \)
  1. \begin{equation*} A =\begin{bmatrix} 2 \amp 1 \\ 3 \amp -1 \end{bmatrix}, \ \ B = \begin{bmatrix} 2 \amp 1 \\ 1 \amp -2 \end{bmatrix} \end{equation*}
    Solution.

    We see that \(A \overset{-R_{1} + R_{2}}{\longrightarrow}B\text{,}\) so the row operation is of type III. This means:
    \begin{equation*} B = EA \ \text{ where } \ E = \begin{bmatrix} 1 \amp 0 \\ -1 \amp 1 \end{bmatrix}. \end{equation*}
    Answer.

    \begin{equation*} E = \begin{bmatrix} 1 \amp 0 \\ -1 \amp 1 \end{bmatrix}. \end{equation*}

  2. \begin{equation*} A =\begin{bmatrix} -1 \amp 2 \\ 0 \amp 1 \end{bmatrix}, \ \ B = \begin{bmatrix} 1 \amp -2 \\ 0 \amp 1 \end{bmatrix} \end{equation*}
    Solution.

    We see that \(A \overset{(-1)R_{1}}{\longrightarrow}B\text{,}\) so the row operation is of type I. This means:
    \begin{equation*} B = EA \ \text{ where } \ E = \begin{bmatrix} -1 \amp 0 \\ 0 \amp 1 \end{bmatrix}. \end{equation*}
    Answer.

    \begin{equation*} E = \begin{bmatrix} -1 \amp 0 \\ 0 \amp 1 \end{bmatrix}. \end{equation*}

  3. \begin{equation*} A = \begin{bmatrix} 1 \amp 1 \\ -1 \amp 2 \end{bmatrix}, \ \ B = \begin{bmatrix} -1 \amp 2 \\ 1 \amp 1 \end{bmatrix} \end{equation*}
    Solution.

    We see that \(A \overset{R_{1}\leftrightarrow R_{2}}{\longrightarrow}B\text{,}\) so the row operation is of type II. This means:
    \begin{equation*} B = EA \ \text{ where } \ E = \begin{bmatrix} 0 \amp 1 \\ 1 \amp 0 \end{bmatrix}. \end{equation*}
    Answer.

    \begin{equation*} E = \begin{bmatrix} 0 \amp 1 \\ 1 \amp 0 \end{bmatrix}. \end{equation*}

  4. \begin{equation*} A =\begin{bmatrix} 4 \amp 1 \\ 3 \amp 2 \end{bmatrix}, \ \ B = \begin{bmatrix} 1 \amp -1 \\ 3 \amp 2 \end{bmatrix} \end{equation*}
    Solution.

    We see that \(A \overset{-R_{2}+R_{1}}{\longrightarrow}B\text{,}\) so the row operation is of type III. This means:
    \begin{equation*} B = EA \ \text{ where } \ E = \begin{bmatrix} 1 \amp -1 \\ 0 \amp 1 \end{bmatrix}. \end{equation*}
    Answer.

    \begin{equation*} E = \begin{bmatrix} 1 \amp -1 \\ 0 \amp 1 \end{bmatrix}. \end{equation*}

  5. \begin{equation*} A =\begin{bmatrix} -1 \amp 1 \\ 1 \amp -1 \end{bmatrix}, \ \ B = \begin{bmatrix} -1 \amp 1 \\ -1 \amp 1 \end{bmatrix} \end{equation*}
    Solution.

    We see that \(A \overset{(-1)R_{2}}{\longrightarrow}B\text{,}\) so the row operation is of type I. This means:
    \begin{equation*} B = EA \ \text{ where } \ E = \begin{bmatrix} 1 \amp 0 \\ 0 \amp -1 \end{bmatrix}. \end{equation*}
    Answer.

    \begin{equation*} E = \begin{bmatrix} 1 \amp 0 \\ 0 \amp -1 \end{bmatrix}. \end{equation*}

  6. \begin{equation*} A =\begin{bmatrix} 2 \amp 1 \\ -1 \amp 3 \end{bmatrix}, \ \ B = \begin{bmatrix} -1 \amp 3 \\ 2 \amp 1 \end{bmatrix} \end{equation*}
    Solution.

    We see that \(A \overset{R_{1}\leftrightarrow R_{2}}{\longrightarrow}B\text{,}\) so the row operation is of type II. This means:
    \begin{equation*} B = EA \ \text{ where } \ E = \begin{bmatrix} 0 \amp 1 \\ 1 \amp 0 \end{bmatrix}. \end{equation*}
    Answer.

    \begin{equation*} E = \begin{bmatrix} 0 \amp 1 \\ 1 \amp 0 \end{bmatrix}. \end{equation*}

Hint.
Recall the three different types of elementary row operations from Hint 1.2.4.8.1. Each of them corresponds to left-multiplication by an elementary matrix \(E\text{:}\) Hint 1.
The row operation \(R_{k}\leftrightarrow R_{l}\) corresponds to multiplication on the left with the matrix \([e_{i,j}]\text{,}\) where
\begin{align*} e_{k,l} \amp= 1 = e_{l,k},\\ e_{i,i} \amp= \left\{ \begin{array}{2} 1 \amp\amp \text{if } i\neq k \text{ and } i \neq l, \\ 0 \amp\amp \text{if } i= k \text{ or } i = l, \end{array} \right.\\ e_{i,j} \amp= 0 \text{ for all remaining } i,j . \end{align*}
We define elementary matrices of this type to be of type I.
Hint 2.
The row operation \(\alpha R_{k}\) where \(\alpha\) is a non-zero number corresponds to multiplication on the left with the matrix \([e_{i,j}]\text{,}\) where
\begin{align*} e_{i,j} \amp= \left\{ \begin{array}{2} 1 \amp\amp \text{if } i=j\neq k, \\ \alpha \amp\amp \text{if } i=j= k, \text{ and} \\ 0 \amp\amp \text{if } i\neq j. \end{array} \right. \end{align*}
We define elementary matrices of this type to be of type II.
Hint 3.
The row operation \(\alpha R_{k} + R_{l}\) corresponds to multiplication on the left with the matrix \([e_{i,j}]\text{,}\) where
\begin{align*} e_{i,j} \amp= \left\{ \begin{array}{2} 1 \amp\amp \text{if } i=j, \\ \alpha \amp\amp \text{if } i=l \text{ and } j=k, \text{ and} \\ 0 \amp\amp \text{for all remaining } i, j. \end{array} \right. \end{align*}
We define elementary matrices of this type to be of type III.

8.

Let \(A = \begin{bmatrix} 1 \amp 2 \amp 1 \\ 0 \amp 5 \amp 1 \\ 2 \amp -1 \amp 4 \end{bmatrix}. \) and \(B = \begin{bmatrix} 1 \amp 2 \amp 1 \\ 0 \amp 5 \amp 1 \\ 1 \amp -\frac{1}{2} \amp 2 \end{bmatrix}.\)
  1. Find the elementary matrix \(E \) such that \(EA = B. \)

  2. Find the inverse of \(E, E^{-1},\) such that \(E^{-1}B = A. \)

Hint.
Look into Hint 4.4.6.7.1 for the different types of elementary matrices.
Answer.

  1. \(\displaystyle E=\begin{bmatrix} 1 \amp 0 \amp 0 \\ 0 \amp 1 \amp 0 \\ 0 \amp 0 \amp \frac{1}{2} \end{bmatrix} \)

  2. \(\displaystyle E^{-1}=\begin{bmatrix} 1 \amp 0 \amp 0 \\ 0 \amp 1 \amp 0 \\ 0 \amp 0 \amp 2 \end{bmatrix}\)

Solution.

We see that \(A\overset{\frac{1}{2}R_{3}}{\longrightarrow}B\text{,}\) so the row operation is of type I. This means that
\begin{equation*} B = EA \ \text{ where } \ E = \begin{bmatrix} 1 \amp 0 \amp 0 \\ 0 \amp 1 \amp 0 \\ 0 \amp 0 \amp \frac{1}{2} \end{bmatrix}. \end{equation*}
In particular, \(B\overset{2R_{3}}{\longrightarrow}A\) undoes this operation, so that
\begin{equation*} E^{-1} = \begin{bmatrix} 1 \amp 0 \amp 0 \\ 0 \amp 1 \amp 0 \\ 0 \amp 0 \amp 2 \end{bmatrix}. \end{equation*}

9.

Suppose \(AB = AC \) and \(A \) is an invertible \(n \times n \) matrix. Does it follow that \(B = C? \) Explain why or why not.
Answer.

Yes, \(B = C. \)
Solution.

By assumption, \(A^{-1}\) exists. In particular, we can multiply the equation \(AB = AC \) on the left by \(A^{-1}\) and we get \(A^{-1}AB = A^{-1}AC \text{.}\) Recall that \(A^{-1}\) has the property that \(A^{-1}A=I\text{,}\) so our equation can be rewritten as \(IB=IC\text{.}\) Since \(I\) has the property \(B=IB\) and \(C=IC\text{,}\) we conclude that, indeed, \(B = C. \)

10.

Suppose \(AB = AC \) and \(A \) is a non-invertible \(n \times n \) matrix. Does it follow that \(B = C? \) Explain why or why not.
Answer.

No, \(B\) is not necessarily equal to \(C\text{.}\)
Solution.

Suppose \(A=0\text{,}\) the matrix all of whose entries are \(0\text{.}\) Then for any matrix \(B\text{,}\) we have \(AB=0\text{.}\) In particular, \(AB=AC\) for all matrices \(B, C\text{.}\) This shows that \(AB = AC \) does not necessarily imply \(B = C \text{.}\)

11.

Construct an example to demonstrate that \((A + B)^{-1} = A^{-1} + B^{-1} \) is not true for all invertible square matrices \(A \) and \(B \) of the same size.
Hint.
It is possible to choose \(A=B\) to be a very nice matrix.
Answer.

The equality is not true for \(A=B=I\text{.}\)
Solution.

We choose \(A=B=I\text{,}\) so that
\begin{equation*} (A + B)^{-1} = (2I)^{-1} = \frac{1}{2}I^{-1} = \frac{1}{2}I, \end{equation*}
while
\begin{equation*} A^{-1} + B^{-1} = 2 I^{-1} = 2 I. \end{equation*}
This shows that \((A + B)^{-1}\) does not equal \(A^{-1} + B^{-1}\text{.}\)

12.

Let
\begin{equation*} A = \begin{bmatrix} 1 \amp 2 \amp 3 \\ 2 \amp 1 \amp 4 \\ 1 \amp 0 \amp 4 \end{bmatrix}. \end{equation*}
Find \(A^{-1} \) if possible. If \(A^{-1} \) does not exist, explain why.
Hint.
See Hint 4.4.6.1.1 for an algorithm to find \(A^{-1}\text{,}\) if it exists.
Answer.

\(A^{-1}= \frac{1}{7} \begin{bmatrix} -4 \amp 8 \amp -5 \\ 4 \amp -1 \amp -2 \\ 1 \amp -2 \amp 3 \end{bmatrix}\text{.}\)
Solution.

We compute
\begin{align*} \matr{ccc|ccc} { 1 \amp 2 \amp 3 \amp 1 \amp 0 \amp 0 \\ 2 \amp 1 \amp 4 \amp 0 \amp 1 \amp 0 \\ 1 \amp 0 \amp 4 \amp 0 \amp 0 \amp 1 } \overset{-2R_{1}+R_{2}}{\underset{-R_{1}+R_{3}}{\longrightarrow}} \amp\matr{ccc|ccc} { 1 \amp 2 \amp 3 \amp 1 \amp 0 \amp 0 \\ 0 \amp -3 \amp -2 \amp -2 \amp 1 \amp 0 \\ 0 \amp -2 \amp 1 \amp -1 \amp 0 \amp 1 }\\ \overset{-2R_{2}+3R_{3}}{\longrightarrow} \amp\matr{ccc|ccc} { 1 \amp 2 \amp 3 \amp 1 \amp 0 \amp 0 \\ 0 \amp -3 \amp -2 \amp -2 \amp 1 \amp 0 \\ 0 \amp 0 \amp 7 \amp 1 \amp -2 \amp 3 } \end{align*}
From this, we can already see that the matrix is invertible. We continue:
\begin{align*} \matr{ccc|ccc} { 1 \amp 2 \amp 3 \amp 1 \amp 0 \amp 0 \\ 0 \amp -3 \amp -2 \amp -2 \amp 1 \amp 0 \\ 0 \amp 0 \amp 7 \amp 1 \amp -2 \amp 3 } \overset{\frac{1}{7}R_{3}}{\underset{2R_{3}'+R_{2}}{\longrightarrow}} \amp\matr{ccc|ccc} { 1 \amp 2 \amp 3 \amp 1 \amp 0 \amp 0 \\ 0 \amp -3 \amp 0 \amp \frac{-12}{7} \amp \frac{3}{7} \amp \frac{6}{7} \\ 0 \amp 0 \amp 1 \amp \frac{1}{7} \amp \frac{-2}{7} \amp \frac{3}{7} }\\ \overset{-\frac{1}{3}R_{2}}{\underset{-3R_{3}+R_{1}}{\longrightarrow}} \amp\matr{ccc|ccc} { 1 \amp 2 \amp 0 \amp \frac{4}{7} \amp \frac{6}{7} \amp \frac{-9}{7} \\ 0 \amp 1 \amp 0 \amp \frac{4}{7} \amp \frac{-1}{7} \amp \frac{-2}{7} \\ 0 \amp 0 \amp 1 \amp \frac{1}{7} \amp \frac{-2}{7} \amp \frac{3}{7} }\\ \overset{-2R_{2}+R_{1}}{\longrightarrow} \amp\matr{ccc|ccc} { 1 \amp 0 \amp 0 \amp \frac{-4}{7} \amp \frac{8}{7} \amp \frac{-5}{7} \\ 0 \amp 1 \amp 0 \amp \frac{4}{7} \amp \frac{-1}{7} \amp \frac{-2}{7} \\ 0 \amp 0 \amp 1 \amp \frac{1}{7} \amp \frac{-2}{7} \amp \frac{3}{7} }. \end{align*}
We conclude that \(A^{-1}= \frac{1}{7} \begin{bmatrix} -4 \amp 8 \amp -5 \\ 4 \amp -1 \amp -2 \\ 1 \amp -2 \amp 3 \end{bmatrix}\)

13.

If \(c \ne 0, \) find the inverse of \(A=\begin{bmatrix} 1 \amp -1 \amp 1 \\ 2 \amp -1 \amp 2 \\ 0 \amp 2 \amp c \end{bmatrix}\) in terms of \(c. \)
Hint.
See Hint 4.4.6.1.1 for an algorithm to find \(A^{-1}\text{,}\) if it exists.
Answer.

\begin{equation*} A^{-1} = \frac{1}{c} \begin{bmatrix} -c-4 \amp c+2 \amp -1 \\ -2c \amp c \amp 0 \\ 4 \amp -2 \amp 1 \end{bmatrix} . \end{equation*}
Solution.

We use Hint 4.4.6.1.1, i.e. we will row-reduce the augmented matrix \(\matr{c|c}{A\amp I_{3}}\text{:}\)
\begin{align*} \matr{ccc|ccc} { 1 \amp -1 \amp 1 \amp 1 \amp 0 \amp 0 \\ 2 \amp -1 \amp 2 \amp 0 \amp 1 \amp 0 \\ 0 \amp 2 \amp c \amp 0 \amp 0 \amp 1 } \overset{-2R_{1}+R_{2}}{\longrightarrow} \amp\matr{ccc|ccc} { 1 \amp -1 \amp 1 \amp 1 \amp 0 \amp 0 \\ 0 \amp 1 \amp 0 \amp -2 \amp 1 \amp 0 \\ 0 \amp 2 \amp c \amp 0 \amp 0 \amp 1 }\\ \overset{-2R_{2}+R_{3}}{\underset{R_{2}+R_{1}}{\longrightarrow}} \amp\matr{ccc|ccc} { 1 \amp 0 \amp 1 \amp -1 \amp 1 \amp 0 \\ 0 \amp 1 \amp 0 \amp -2 \amp 1 \amp 0 \\ 0 \amp 0 \amp c \amp 4 \amp -2 \amp 1 }. \end{align*}
Since \(c\neq 0\text{,}\) dividing by \(c\) is a valid row-operation, so that
\begin{align*} \matr{ccc|ccc} { 1 \amp 0 \amp 1 \amp -1 \amp 1 \amp 0 \\ 0 \amp 1 \amp 0 \amp -2 \amp 1 \amp 0 \\ 0 \amp 0 \amp c \amp 4 \amp -2 \amp 1 } \overset{\frac{1}{c}R_{3}}{\longrightarrow} \amp\matr{ccc|ccc} { 1 \amp 0 \amp 1 \amp -1 \amp 1 \amp 0 \\ 0 \amp 1 \amp 0 \amp -2 \amp 1 \amp 0 \\ 0 \amp 0 \amp 1 \amp \frac{4}{c} \amp \frac{-2}{c} \amp \frac{1}{c} }\\ \overset{-R_{3}+R_{1}}{\longrightarrow} \amp\matr{ccc|ccc} { 1 \amp 0 \amp 0 \amp -1-\frac{4}{c} \amp 1+\frac{2}{c} \amp \frac{-1}{c} \\ 0 \amp 1 \amp 0 \amp -2 \amp 1 \amp 0 \\ 0 \amp 0 \amp 1 \amp \frac{4}{c} \amp \frac{-2}{c} \amp \frac{1}{c} }. \end{align*}
We conclude that
\begin{equation*} A^{-1} = \begin{bmatrix} -1-\frac{4}{c} \amp 1+\frac{2}{c} \amp \frac{-1}{c} \\ -2 \amp 1 \amp 0 \\ \frac{4}{c} \amp \frac{-2}{c} \amp \frac{1}{c} \end{bmatrix} = \frac{1}{c} \begin{bmatrix} -c-4 \amp c+2 \amp -1 \\ -2c \amp c \amp 0 \\ 4 \amp -2 \amp 1 \end{bmatrix} . \end{equation*}