What is called an inverse matrix. Finding the inverse matrix

For any non-singular matrix A there is a unique matrix A -1 such that

A*A -1 =A -1 *A = E,

where E is the identity matrix of the same orders as A. The matrix A -1 is called the inverse of matrix A.

In case someone forgot, in the identity matrix, except for the diagonal filled with ones, all other positions are filled with zeros, an example of an identity matrix:

Finding the inverse matrix using the adjoint matrix method

The inverse matrix is ​​defined by the formula:

where A ij - elements a ij.

Those. To calculate the inverse matrix, you need to calculate the determinant of this matrix. Then find the algebraic complements for all its elements and compose a new matrix from them. Next you need to transport this matrix. And divide each element of the new matrix by the determinant of the original matrix.

Let's look at a few examples.

Find A -1 for matrix

Solution. Let's find A -1 using the adjoint matrix method. We have det A = 2. Let us find the algebraic complements of the elements of matrix A. In this case, the algebraic complements of the matrix elements will be the corresponding elements of the matrix itself, taken with a sign in accordance with the formula

We have A 11 = 3, A 12 = -4, A 21 = -1, A 22 = 2. We form the adjoint matrix

We transport the matrix A*:

We find the inverse matrix using the formula:

We get:

Using the adjoint matrix method, find A -1 if

Solution. First of all, we calculate the definition of this matrix to verify the existence of the inverse matrix. We have

Here we added to the elements of the second row the elements of the third row, previously multiplied by (-1), and then expanded the determinant for the second row. Since the definition of this matrix is ​​nonzero, its inverse matrix exists. To construct the adjoint matrix, we find the algebraic complements of the elements of this matrix. We have

According to the formula

we transport the matrix A*:

Then according to the formula

Finding the inverse matrix using the method of elementary transformations

In addition to the method of finding the inverse matrix, which follows from the formula (the adjoint matrix method), there is a method for finding the inverse matrix, called the method of elementary transformations.

Elementary matrix transformations

The following transformations are called elementary matrix transformations:

1) rearrangement of rows (columns);

2) multiplying a row (column) by a number other than zero;

3) adding to the elements of a row (column) the corresponding elements of another row (column), previously multiplied by a certain number.

To find the matrix A -1, we construct a rectangular matrix B \u003d (A | E) of orders (n; 2n), assigning to the matrix A on the right the identity matrix E through the dividing line:

Let's look at an example.

Using the method of elementary transformations, find A -1 if

Solution. We form matrix B:

Let us denote the rows of matrix B by α 1, α 2, α 3. Let us perform the following transformations on the rows of matrix B.

The matrix $A^(-1)$ is called the inverse of the square matrix $A$ if the condition $A^(-1)\cdot A=A\cdot A^(-1)=E$ is satisfied, where $E $ is the identity matrix, the order of which is equal to the order of the matrix $A$.

A non-singular matrix is ​​a matrix whose determinant is not equal to zero. Accordingly, a singular matrix is ​​one whose determinant is equal to zero.

The inverse matrix $A^(-1)$ exists if and only if the matrix $A$ is non-singular. If the inverse matrix $A^(-1)$ exists, then it is unique.

There are several ways to find the inverse of a matrix, and we will look at two of them. This page will cover the adjoint matrix method, which is considered standard in most courses. higher mathematics. The second way to find the inverse matrix (method of elementary transformations), which involves the use of the Gauss method or the Gauss-Jordan method, is considered in the second part.

Adjoint matrix method

Let the matrix $A_(n\times n)$ be given. In order to find the inverse matrix $A^(-1)$, three steps are required:

  1. Find the determinant of the matrix $A$ and make sure that $\Delta A\neq 0$, i.e. that matrix A is non-singular.
  2. Compose algebraic complements $A_(ij)$ of each element of the matrix $A$ and write down the matrix $A_(n\times n)^(*)=\left(A_(ij) \right)$ from the found algebraic complements.
  3. Write the inverse matrix taking into account the formula $A^(-1)=\frac(1)(\Delta A)\cdot (A^(*))^T$.

The matrix $(A^(*))^T$ is often called adjoint (reciprocal, allied) to the matrix $A$.

If the decision is made manually, then the first method is good only for matrices of relatively small orders: second (), third (), fourth (). To find the inverse matrix for a higher order matrix, other methods are used. For example, the Gaussian method, which is discussed in the second part.

Example No. 1

Find matrix inverse to matrix $A=\left(\begin(array) (cccc) 5 & -4 &1 & 0 \\ 12 &-11 &4 & 0 \\ -5 & 58 &4 & 0 \\ 3 & - 1 & -9 & 0 \end(array) \right)$.

Since all elements of the fourth column are equal to zero, then $\Delta A=0$ (i.e. the matrix $A$ is degenerate). Since $\Delta A=0$, there is no inverse matrix to matrix $A$.

Answer: matrix $A^(-1)$ does not exist.

Example No. 2

Find the inverse of matrix $A=\left(\begin(array) (cc) -5 & 7 \\ 9 & 8 \end(array)\right)$. Perform check.

We use the adjoint matrix method. First, let's find the determinant of the given matrix $A$:

$$ \Delta A=\left| \begin(array) (cc) -5 & 7\\ 9 & 8 \end(array)\right|=-5\cdot 8-7\cdot 9=-103. $$

Since $\Delta A \neq 0$, then the inverse matrix exists, therefore we will continue the solution. Finding algebraic complements

\begin(aligned) & A_(11)=(-1)^2\cdot 8=8; \; A_(12)=(-1)^3\cdot 9=-9;\\ & A_(21)=(-1)^3\cdot 7=-7; \; A_(22)=(-1)^4\cdot (-5)=-5.\\ \end(aligned)

We compose a matrix of algebraic additions: $A^(*)=\left(\begin(array) (cc) 8 & -9\\ -7 & -5 \end(array)\right)$.

We transpose the resulting matrix: $(A^(*))^T=\left(\begin(array) (cc) 8 & -7\\ -9 & -5 \end(array)\right)$ (the resulting matrix is ​​often is called the adjoint or allied matrix to the matrix $A$). Using the formula $A^(-1)=\frac(1)(\Delta A)\cdot (A^(*))^T$, we have:

$$ A^(-1)=\frac(1)(-103)\cdot \left(\begin(array) (cc) 8 & -7\\ -9 & -5 \end(array)\right) =\left(\begin(array) (cc) -8/103 & 7/103\\ 9/103 & 5/103 \end(array)\right) $$

So, the inverse matrix is ​​found: $A^(-1)=\left(\begin(array) (cc) -8/103 & 7/103\\ 9/103 & 5/103 \end(array)\right) $. To check the truth of the result, it is enough to check the truth of one of the equalities: $A^(-1)\cdot A=E$ or $A\cdot A^(-1)=E$. Let's check the equality $A^(-1)\cdot A=E$. In order to work less with fractions, we will substitute the matrix $A^(-1)$ not in the form $\left(\begin(array) (cc) -8/103 & 7/103\\ 9/103 & 5/103 \ end(array)\right)$, and in the form $-\frac(1)(103)\cdot \left(\begin(array) (cc) 8 & -7\\ -9 & -5 \end(array )\right)$:

$$ A^(-1)\cdot(A) =-\frac(1)(103)\cdot \left(\begin(array) (cc) 8 & -7\\ -9 & -5 \end( array)\right)\cdot\left(\begin(array) (cc) -5 & 7 \\ 9 & 8 \end(array)\right) =-\frac(1)(103)\cdot\left( \begin(array) (cc) -103 & 0 \\ 0 & -103 \end(array)\right) =\left(\begin(array) (cc) 1 & 0 \\ 0 & 1 \end(array )\right) =E $$

Answer: $A^(-1)=\left(\begin(array) (cc) -8/103 & 7/103\\ 9/103 & 5/103 \end(array)\right)$.

Example No. 3

Find the inverse matrix for the matrix $A=\left(\begin(array) (ccc) 1 & 7 & 3 \\ -4 & 9 & 4 \\ 0 & 3 & 2\end(array) \right)$. Perform check.

Let's start by calculating the determinant of the matrix $A$. So, the determinant of the matrix $A$ is:

$$ \Delta A=\left| \begin(array) (ccc) 1 & 7 & 3 \\ -4 & 9 & 4 \\ 0 & 3 & 2\end(array) \right| = 18-36+56-12=26. $$

Since $\Delta A\neq 0$, then the inverse matrix exists, therefore we will continue the solution. We find the algebraic complements of each element of a given matrix:

$$ \begin(aligned) & A_(11)=(-1)^(2)\cdot\left|\begin(array)(cc) 9 & 4\\ 3 & 2\end(array)\right| =6;\; A_(12)=(-1)^(3)\cdot\left|\begin(array)(cc) -4 &4 \\ 0 & 2\end(array)\right|=8;\; A_(13)=(-1)^(4)\cdot\left|\begin(array)(cc) -4 & 9\\ 0 & 3\end(array)\right|=-12;\\ & A_(21)=(-1)^(3)\cdot\left|\begin(array)(cc) 7 & 3\\ 3 & 2\end(array)\right|=-5;\; A_(22)=(-1)^(4)\cdot\left|\begin(array)(cc) 1 & 3\\ 0 & 2\end(array)\right|=2;\; A_(23)=(-1)^(5)\cdot\left|\begin(array)(cc) 1 & 7\\ 0 & 3\end(array)\right|=-3;\\ & A_ (31)=(-1)^(4)\cdot\left|\begin(array)(cc) 7 & 3\\ 9 & 4\end(array)\right|=1;\; A_(32)=(-1)^(5)\cdot\left|\begin(array)(cc) 1 & 3\\ -4 & 4\end(array)\right|=-16;\; A_(33)=(-1)^(6)\cdot\left|\begin(array)(cc) 1 & 7\\ -4 & 9\end(array)\right|=37. \end(aligned) $$

We compose a matrix of algebraic additions and transpose it:

$$ A^*=\left(\begin(array) (ccc) 6 & 8 & -12 \\ -5 & 2 & -3 \\ 1 & -16 & 37\end(array) \right); \; (A^*)^T=\left(\begin(array) (ccc) 6 & -5 & 1 \\ 8 & 2 & -16 \\ -12 & -3 & 37\end(array) \right) . $$

Using the formula $A^(-1)=\frac(1)(\Delta A)\cdot (A^(*))^T$, we get:

$$ A^(-1)=\frac(1)(26)\cdot \left(\begin(array) (ccc) 6 & -5 & 1 \\ 8 & 2 & -16 \\ -12 & - 3 & 37\end(array) \right)= \left(\begin(array) (ccc) 3/13 & -5/26 & 1/26 \\ 4/13 & 1/13 & -8/13 \ \ -6/13 & -3/26 & 37/26 \end(array) \right) $$

So $A^(-1)=\left(\begin(array) (ccc) 3/13 & -5/26 & 1/26 \\ 4/13 & 1/13 & -8/13 \\ - 6/13 & -3/26 & 37/26 \end(array) \right)$. To check the truth of the result, it is enough to check the truth of one of the equalities: $A^(-1)\cdot A=E$ or $A\cdot A^(-1)=E$. Let's check the equality $A\cdot A^(-1)=E$. In order to work less with fractions, we will substitute the matrix $A^(-1)$ not in the form $\left(\begin(array) (ccc) 3/13 & -5/26 & 1/26 \\ 4/13 & 1/13 & -8/13 \\ -6/13 & -3/26 & 37/26 \end(array) \right)$, and in the form $\frac(1)(26)\cdot \left( \begin(array) (ccc) 6 & -5 & 1 \\ 8 & 2 & -16 \\ -12 & -3 & 37\end(array) \right)$:

$$ A\cdot(A^(-1)) =\left(\begin(array)(ccc) 1 & 7 & 3 \\ -4 & 9 & 4\\ 0 & 3 & 2\end(array) \right)\cdot \frac(1)(26)\cdot \left(\begin(array) (ccc) 6 & -5 & 1 \\ 8 & 2 & -16 \\ -12 & -3 & 37\ end(array) \right) =\frac(1)(26)\cdot\left(\begin(array) (ccc) 26 & 0 & 0 \\ 0 & 26 & 0 \\ 0 & 0 & 26\end (array) \right) =\left(\begin(array) (ccc) 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1\end(array) \right) =E $$

The check was successful, the inverse matrix $A^(-1)$ was found correctly.

Answer: $A^(-1)=\left(\begin(array) (ccc) 3/13 & -5/26 & 1/26 \\ 4/13 & 1/13 & -8/13 \\ -6 /13 & -3/26 & 37/26 \end(array) \right)$.

Example No. 4

Find the matrix inverse of matrix $A=\left(\begin(array) (cccc) 6 & -5 & 8 & 4\\ 9 & 7 & 5 & 2 \\ 7 & 5 & 3 & 7\\ -4 & 8 & -8 & -3 \end(array) \right)$.

For a fourth-order matrix, finding the inverse matrix using algebraic additions is somewhat difficult. However, such examples in tests meet.

To find the inverse of a matrix, you first need to calculate the determinant of the matrix $A$. The best way to do this in this situation is by decomposing the determinant along a row (column). We select any row or column and find the algebraic complements of each element of the selected row or column.

For example, for the first line we get:

$$ A_(11)=\left|\begin(array)(ccc) 7 & 5 & 2\\ 5 & 3 & 7\\ 8 & -8 & -3 \end(array)\right|=556; \; A_(12)=-\left|\begin(array)(ccc) 9 & 5 & 2\\ 7 & 3 & 7 \\ -4 & -8 & -3 \end(array)\right|=-300 ; $$ $$ A_(13)=\left|\begin(array)(ccc) 9 & 7 & 2\\ 7 & 5 & 7\\ -4 & 8 & -3 \end(array)\right|= -536;\; A_(14)=-\left|\begin(array)(ccc) 9 & 7 & 5\\ 7 & 5 & 3\\ -4 & 8 & -8 \end(array)\right|=-112. $$

The determinant of the matrix $A$ is calculated using the following formula:

$$ \Delta(A)=a_(11)\cdot A_(11)+a_(12)\cdot A_(12)+a_(13)\cdot A_(13)+a_(14)\cdot A_(14 )=6\cdot 556+(-5)\cdot(-300)+8\cdot(-536)+4\cdot(-112)=100. $$

$$ \begin(aligned) & A_(21)=-77;\;A_(22)=50;\;A_(23)=87;\;A_(24)=4;\\ & A_(31) =-93;\;A_(32)=50;\;A_(33)=83;\;A_(34)=36;\\ & A_(41)=473;\;A_(42)=-250 ;\;A_(43)=-463;\;A_(44)=-96. \end(aligned) $$

Matrix of algebraic complements: $A^*=\left(\begin(array)(cccc) 556 & -300 & -536 & -112\\ -77 & 50 & 87 & 4 \\ -93 & 50 & 83 & 36\\ 473 & -250 & -463 & -96\end(array)\right)$.

Adjoint matrix: $(A^*)^T=\left(\begin(array) (cccc) 556 & -77 & -93 & 473\\ -300 & 50 & 50 & -250 \\ -536 & 87 & 83 & -463\\ -112 & 4 & 36 & -96\end(array)\right)$.

Inverse matrix:

$$ A^(-1)=\frac(1)(100)\cdot \left(\begin(array) (cccc) 556 & -77 & -93 & 473\\ -300 & 50 & 50 & -250 \\ -536 & 87 & 83 & -463\\ -112 & 4 & 36 & -96 \end(array) \right)= \left(\begin(array) (cccc) 139/25 & -77/100 & -93/100 & 473/100 \\ -3 & 1/2 & 1/2 & -5/2 \\ -134/25 & 87/100 & 83/100 & -463/100 \\ -28/ 25 & 1/25 & 9/25 & -24/25 \end(array) \right) $$

The check, if desired, can be done in the same way as in the previous examples.

Answer: $A^(-1)=\left(\begin(array) (cccc) 139/25 & -77/100 & -93/100 & 473/100 \\ -3 & 1/2 & 1/2 & -5/2 \\ -134/25 & 87/100 & 83/100 & -463/100 \\ -28/25 & 1/25 & 9/25 & -24/25 \end(array) \right) $.

In the second part, we will consider another way to find the inverse matrix, which involves the use of transformations of the Gaussian method or the Gauss-Jordan method.

Matrix algebra - Inverse matrix

inverse matrix

Inverse matrix is a matrix that, when multiplied both on the right and on the left by a given matrix, gives the identity matrix.
Let us denote the inverse matrix of the matrix A through , then according to definition we get:

Where E– identity matrix.
Square matrix called non-special (non-degenerate) if its determinant is not zero. Otherwise it is called special (degenerate) or singular.

The theorem holds: Every non-singular matrix has an inverse matrix.

The operation of finding the inverse matrix is ​​called appeal matrices. Let's consider the matrix inversion algorithm. Let a non-singular matrix be given n-th order:

where Δ = det A ≠ 0.

Algebraic addition of an element matrices n-th order A is called the determinant of a matrix taken with a certain sign ( n–1)th order obtained by deleting i-th line and j th matrix column A:

Let's create the so-called attached matrix:

where are the algebraic complements of the corresponding elements of the matrix A.
Note that algebraic additions of matrix row elements A are placed in the corresponding columns of the matrix à , that is, the matrix is ​​transposed at the same time.
By dividing all the elements of the matrix à by Δ – the value of the matrix determinant A, we get the inverse matrix as a result:

Let us note a number of special properties of the inverse matrix:
1) for a given matrix A its inverse matrix is the only one;
2) if there is an inverse matrix, then right reverse And left reverse the matrices coincide with it;
3) a singular (singular) square matrix does not have an inverse matrix.

Basic properties of an inverse matrix:
1) the determinant of the inverse matrix and the determinant of the original matrix are reciprocals;
2) the inverse matrix of the product of square matrices is equal to the product of the inverse matrices of factors, taken in reverse order:

3) the transposed inverse matrix is ​​equal to the inverse matrix of the given transposed matrix:

EXAMPLE Calculate the inverse of the given matrix.

inverse matrix is a matrix A−1, when multiplied by which the given initial matrix A results in the identity matrix E:

AA −1 = A −1 A =E.

Inverse matrix method.

Inverse matrix method- this is one of the most common methods for solving matrices and is used to solve systems of linear algebraic equations (SLAEs) in cases where the number of unknowns corresponds to the number of equations.

Let there be a system n linear equations With n unknown:

Such a system can be written as a matrix equation A* X = B,

Where
- system matrix,

- column of unknowns,

- column of free coefficients.

From the derived matrix equation, we express X by multiplying both sides of the matrix equation on the left by A-1, resulting in:

A -1 * A * X = A -1 * B

Knowing that A -1 * A = E, Then E * X = A -1 * B or X = A -1 * B.

The next step is to determine the inverse matrix A-1 and multiplied by the column of free terms B.

Inverse Matrix to Matrix A exists only when det A≠ 0 . In view of this, when solving SLAE by the inverse matrix method, the first step is to find det A. If det A≠ 0 , then the system has only one solution, which can be obtained using the inverse matrix method, but if det A = 0, then such a system inverse matrix method can't be solved.

Solving the inverse matrix.

Sequence of actions for inverse matrix solutions:

  1. Get the matrix determinant A. If the determinant is greater than zero, we solve the inverse of the matrix further; if it is equal to zero, then we cannot find the inverse matrix here.
  2. Finding the transposed matrix AT.
  3. We look for algebraic complements, after which we replace all elements of the matrix with their algebraic complements.
  4. We assemble the inverse matrix from algebraic additions: we divide all the elements of the resulting matrix by the determinant of the initially given matrix. The final matrix will be the desired inverse matrix with respect to the original one.

Below algorithm inverse matrix solutions essentially the same as the one above, the difference is only in a few steps: first of all we define the algebraic complements, and after that we calculate the allied matrix C.

  1. Determine whether a given matrix is ​​square. If the answer is negative, it becomes clear that there cannot be an inverse matrix for it.
  2. Determine whether a given matrix is ​​square. If the answer is negative, it becomes clear that there cannot be an inverse matrix for it.
  3. We calculate algebraic complements.
  4. We compose a union (mutual, adjoint) matrix C.
  5. We compose the inverse matrix from algebraic additions: all elements of the adjoint matrix C divide by the determinant of the initial matrix. The final matrix will be the required inverse matrix relative to the given one.
  6. We check the work done: multiply the initial and resulting matrices, the result should be an identity matrix.

This is best done using an attached matrix.

Theorem: If to a square matrix with right side assign an identity matrix of the same order and, using elementary transformations over the rows, transform the initial matrix on the left into the identity matrix, then the one obtained on the right side will be inverse to the initial one.

An example of finding an inverse matrix.

Exercise. For matrix find the inverse using the adjoint matrix method.

Solution. Add to the given matrix A on the right is a 2nd order identity matrix:

From the 1st line we subtract the 2nd:

From the second line we subtract the first 2:

This topic is one of the most hated among students. Worse, probably, only determinants.

The trick is that the very concept of an inverse element (and I’m not just talking about matrices) refers us to the operation of multiplication. Even in the school curriculum, multiplication is considered a complex operation, and multiplication of matrices is generally a separate topic, to which I have an entire paragraph and video lesson dedicated.

Today we will not go into the details of matrix calculations. Let’s just remember: how matrices are designated, how they are multiplied, and what follows from this.

Review: Matrix Multiplication

First of all, let's agree on notation. A matrix $A$ of size $\left[ m\times n \right]$ is simply a table of numbers with exactly $m$ rows and $n$ columns:

\=\underbrace(\left[ \begin(matrix) ((a)_(11)) & ((a)_(12)) & ... & ((a)_(1n)) \\ (( a)_(21)) & ((a)_(22)) & ... & ((a)_(2n)) \\ ... & ... & ... & ... \\ ((a)_(m1)) & ((a)_(m2)) & ... & ((a)_(mn)) \\\end(matrix) \right])_(n)\]

To avoid accidentally mixing up rows and columns (believe me, in an exam you can confuse a one with a two, let alone some rows), just look at the picture:

Determining indices for matrix cells

What's happening? If you place standard system coordinates $OXY$ in the upper left corner and direct the axes so that they cover the entire matrix, then each cell of this matrix can be uniquely associated with coordinates $\left(x;y \right)$ - this will be the row number and column number.

Why is the coordinate system placed in the upper left corner? Yes, because it is from there that we begin to read any texts. It's very easy to remember.

Why is the $x$ axis directed downwards and not to the right? Again, it's simple: take a standard coordinate system (the $x$ axis goes to the right, the $y$ axis goes up) and rotate it so that it covers the matrix. This is a 90 degree clockwise rotation - we see the result in the picture.

In general, we have figured out how to determine the indices of matrix elements. Now let's look at multiplication.

Definition. Matrices $A=\left[ m\times n \right]$ and $B=\left[ n\times k \right]$, when the number of columns in the first coincides with the number of rows in the second, are called consistent.

Exactly in that order. One can be confused and say that the matrices $A$ and $B$ form an ordered pair $\left(A;B \right)$: if they are consistent in this order, then it is not at all necessary that $B$ and $A$ those. the pair $\left(B;A \right)$ is also consistent.

Only matched matrices can be multiplied.

Definition. The product of matched matrices $A=\left[ m\times n \right]$ and $B=\left[ n\times k \right]$ is the new matrix $C=\left[ m\times k \right]$ , the elements of which $((c)_(ij))$ are calculated according to the formula:

\[((c)_(ij))=\sum\limits_(k=1)^(n)(((a)_(ik)))\cdot ((b)_(kj))\]

In other words: to get the element $((c)_(ij))$ of the matrix $C=A\cdot B$, you need to take the $i$-row of the first matrix, the $j$-th column of the second matrix, and then multiply in pairs elements from this row and column. Add up the results.

Yes, that’s such a harsh definition. Several facts immediately follow from it:

  1. Matrix multiplication, generally speaking, is non-commutative: $A\cdot B\ne B\cdot A$;
  2. However, multiplication is associative: $\left(A\cdot B \right)\cdot C=A\cdot \left(B\cdot C \right)$;
  3. And even distributively: $\left(A+B \right)\cdot C=A\cdot C+B\cdot C$;
  4. And once again distributively: $A\cdot \left(B+C \right)=A\cdot B+A\cdot C$.

The distributivity of multiplication had to be described separately for the left and right multiplier-sum just because of the non-commutativity of the multiplication operation.

If, nevertheless, it turns out that $A\cdot B=B\cdot A$, such matrices are called permutable.

Among all the matrices that are multiplied by something there, there are special ones - those that, when multiplied by any matrix $A$, again give $A$:

Definition. A matrix $E$ is called identity if $A\cdot E=A$ or $E\cdot A=A$. In the case of a square matrix $A$ we can write:

The identity matrix is ​​a frequent guest when solving matrix equations. And in general, a frequent guest in the world of matrices. :)

And because of this $E$, someone came up with all the nonsense that will be written next.

What is an inverse matrix

Since matrix multiplication is a very labor-intensive operation (you have to multiply a bunch of rows and columns), the concept of an inverse matrix also turns out to be not the most trivial. And requiring some explanation.

Key Definition

Well, it's time to know the truth.

Definition. A matrix $B$ is called the inverse of a matrix $A$ if

The inverse matrix is ​​denoted by $((A)^(-1))$ (not to be confused with the degree!), so the definition can be rewritten as follows:

It would seem that everything is extremely simple and clear. But when analyzing this definition, several questions immediately arise:

  1. Does an inverse matrix always exist? And if not always, then how to determine: when it exists and when it does not?
  2. And who said that there is exactly one such matrix? What if for some initial matrix $A$ there is a whole crowd of inverses?
  3. What do all these “reverses” look like? And how, exactly, should we count them?

As for calculation algorithms, we will talk about this a little later. But we will answer the remaining questions right now. Let us formulate them in the form of separate statements-lemmas.

Basic properties

Let's start with how the matrix $A$ should, in principle, look in order for $((A)^(-1))$ to exist for it. Now we will make sure that both of these matrices must be square, and of the same size: $\left[ n\times n \right]$.

Lemma 1. Given a matrix $A$ and its inverse $((A)^(-1))$. Then both of these matrices are square, and of the same order $n$.

Proof. It's simple. Let the matrix $A=\left[ m\times n \right]$, $((A)^(-1))=\left[ a\times b \right]$. Since the product $A\cdot ((A)^(-1))=E$ exists by definition, the matrices $A$ and $((A)^(-1))$ are consistent in the order shown:

\[\begin(align) & \left[ m\times n \right]\cdot \left[ a\times b \right]=\left[ m\times b \right] \\ & n=a \end( align)\]

This is a direct consequence of the matrix multiplication algorithm: the coefficients $n$ and $a$ are “transit” and must be equal.

At the same time, the inverse multiplication is also defined: $((A)^(-1))\cdot A=E$, therefore the matrices $((A)^(-1))$ and $A$ are also consistent in the specified order:

\[\begin(align) & \left[ a\times b \right]\cdot \left[ m\times n \right]=\left[ a\times n \right] \\ & b=m \end( align)\]

Thus, without loss of generality, we can assume that $A=\left[ m\times n \right]$, $((A)^(-1))=\left[ n\times m \right]$. However, according to the definition of $A\cdot ((A)^(-1))=((A)^(-1))\cdot A$, therefore the sizes of the matrices strictly coincide:

\[\begin(align) & \left[ m\times n \right]=\left[ n\times m \right] \\ & m=n \end(align)\]

So it turns out that all three matrices - $A$, $((A)^(-1))$ and $E$ - are square matrices of size $\left[ n\times n \right]$. The lemma is proven.

Well, that's already good. We see that only square matrices are invertible. Now let's make sure that the inverse matrix is ​​always the same.

Lemma 2. Given a matrix $A$ and its inverse $((A)^(-1))$. Then this inverse matrix is ​​the only one.

Proof. Let's go by contradiction: let the matrix $A$ have at least two inverses - $B$ and $C$. Then, according to definition, the following equalities are true:

\[\begin(align) & A\cdot B=B\cdot A=E; \\ & A\cdot C=C\cdot A=E. \\ \end(align)\]

From Lemma 1 we conclude that all four matrices - $A$, $B$, $C$ and $E$ - are squares of the same order: $\left[ n\times n \right]$. Therefore, the product is defined:

Since matrix multiplication is associative (but not commutative!), we can write:

\[\begin(align) & B\cdot A\cdot C=\left(B\cdot A \right)\cdot C=E\cdot C=C; \\ & B\cdot A\cdot C=B\cdot \left(A\cdot C \right)=B\cdot E=B; \\ & B\cdot A\cdot C=C=B\Rightarrow B=C. \\ \end(align)\]

Received only possible variant: two instances of the inverse matrix are equal. The lemma is proven.

The above arguments repeat almost verbatim the proof of the uniqueness of the inverse element for all real numbers $b\ne 0$. The only significant addition is taking into account the dimension of matrices.

However, we still do not know anything about whether every square matrix is ​​invertible. Here the determinant comes to our aid - this is a key characteristic for all square matrices.

Lemma 3. Given a matrix $A$. If its inverse matrix $((A)^(-1))$ exists, then the determinant of the original matrix is ​​nonzero:

\[\left| A\right|\ne 0\]

Proof. We already know that $A$ and $((A)^(-1))$ are square matrices of size $\left[ n\times n \right]$. Therefore, for each of them we can calculate the determinant: $\left| A\right|$ and $\left| ((A)^(-1)) \right|$. However, the determinant of a product is equal to the product of the determinants:

\[\left| A\cdot B \right|=\left| A \right|\cdot \left| B \right|\Rightarrow \left| A\cdot ((A)^(-1)) \right|=\left| A \right|\cdot \left| ((A)^(-1)) \right|\]

But according to the definition, $A\cdot ((A)^(-1))=E$, and the determinant of $E$ is always equal to 1, so

\[\begin(align) & A\cdot ((A)^(-1))=E; \\ & \left| A\cdot ((A)^(-1)) \right|=\left| E\right|; \\ & \left| A \right|\cdot \left| ((A)^(-1)) \right|=1. \\ \end(align)\]

The product of two numbers is equal to one only if each of these numbers is non-zero:

\[\left| A \right|\ne 0;\quad \left| ((A)^(-1)) \right|\ne 0.\]

So it turns out that $\left| A \right|\ne 0$. The lemma is proven.

In fact, this requirement is quite logical. Now we will analyze the algorithm for finding the inverse matrix - and it will become completely clear why, with a zero determinant, no inverse matrix in principle can exist.

But first, let’s formulate an “auxiliary” definition:

Definition. A singular matrix is ​​a square matrix of size $\left[ n\times n \right]$ whose determinant is zero.

Thus, we can claim that every invertible matrix is ​​non-singular.

How to find the inverse of a matrix

Now we will consider a universal algorithm for finding inverse matrices. In general, there are two generally accepted algorithms, and we will also consider the second one today.

The one that will be discussed now is very effective for matrices of size $\left[ 2\times 2 \right]$ and - partially - size $\left[ 3\times 3 \right]$. But starting from the size $\left[ 4\times 4 \right]$ it is better not to use it. Why - now you will understand everything yourself.

Algebraic additions

Get ready. Now there will be pain. No, don’t worry: a beautiful nurse in a skirt, stockings with lace will not come to you and give you an injection in the buttock. Everything is much more prosaic: algebraic additions and Her Majesty the “Union Matrix” come to you.

Let's start with the main thing. Let there be a square matrix of size $A=\left[ n\times n \right]$, whose elements are called $((a)_(ij))$. Then for each such element we can define an algebraic complement:

Definition. Algebraic complement $((A)_(ij))$ to the element $((a)_(ij))$ located in the $i$th row and $j$th column of the matrix $A=\left[ n \times n \right]$ is a construction of the form

\[((A)_(ij))=((\left(-1 \right))^(i+j))\cdot M_(ij)^(*)\]

Where $M_(ij)^(*)$ is the determinant of the matrix obtained from the original $A$ by deleting the same $i$th row and $j$th column.

Again. The algebraic complement to a matrix element with coordinates $\left(i;j \right)$ is denoted as $((A)_(ij))$ and is calculated according to the scheme:

  1. First, we delete the $i$-row and $j$-th column from the original matrix. We obtain a new square matrix, and we denote its determinant as $M_(ij)^(*)$.
  2. Then we multiply this determinant by $((\left(-1 \right))^(i+j))$ - at first this expression may seem mind-blowing, but in essence we are simply figuring out the sign in front of $M_(ij)^(*) $.
  3. We count and get a specific number. Those. the algebraic addition is precisely a number, and not some new matrix, etc.

The matrix $M_(ij)^(*)$ itself is called an additional minor to the element $((a)_(ij))$. And in this sense, the above definition of an algebraic complement is a special case of a more complex definition- what we considered in the lesson about the determinant.

Important note. Actually, in “adult” mathematics, algebraic additions are defined as follows:

  1. We take $k$ rows and $k$ columns in a square matrix. At their intersection we get a matrix of size $\left[ k\times k \right]$ - its determinant is called a minor of order $k$ and is denoted $((M)_(k))$.
  2. Then we cross out these “selected” $k$ rows and $k$ columns. Once again you get a square matrix - its determinant is called an additional minor and is denoted $M_(k)^(*)$.
  3. Multiply $M_(k)^(*)$ by $((\left(-1 \right))^(t))$, where $t$ is (attention now!) the sum of the numbers of all selected rows and columns . This will be the algebraic addition.

Look at the third step: there is actually a sum of $2k$ terms! Another thing is that for $k=1$ we will get only 2 terms - these will be the same $i+j$ - the “coordinates” of the element $((a)_(ij))$ for which we are looking for an algebraic complement.

So today we're using a slightly simplified definition. But as we will see later, it will be more than enough. The following thing is much more important:

Definition. The allied matrix $S$ to the square matrix $A=\left[ n\times n \right]$ is a new matrix of size $\left[ n\times n \right]$, which is obtained from $A$ by replacing $(( a)_(ij))$ by algebraic additions $((A)_(ij))$:

\\Rightarrow S=\left[ \begin(matrix) ((A)_(11)) & ((A)_(12)) & ... & ((A)_(1n)) \\ (( A)_(21)) & ((A)_(22)) & ... & ((A)_(2n)) \\ ... & ... & ... & ... \\ ((A)_(n1)) & ((A)_(n2)) & ... & ((A)_(nn)) \\\end(matrix) \right]\]

The first thought that arises at the moment of realizing this definition is “how much will have to be counted!” Relax: you will have to count, but not that much. :)

Well, all this is very nice, but why is it necessary? But why.

Main theorem

Let's go back a little. Remember, in Lemma 3 it was stated that the invertible matrix $A$ is always non-singular (that is, its determinant is non-zero: $\left| A \right|\ne 0$).

So, the opposite is also true: if the matrix $A$ is not singular, then it is always invertible. And there is even a search scheme for $((A)^(-1))$. Check it out:

Inverse matrix theorem. Let a square matrix $A=\left[ n\times n \right]$ be given, and its determinant is nonzero: $\left| A \right|\ne 0$. Then the inverse matrix $((A)^(-1))$ exists and is calculated by the formula:

\[((A)^(-1))=\frac(1)(\left| A \right|)\cdot ((S)^(T))\]

And now - everything is the same, but in legible handwriting. To find the inverse matrix, you need:

  1. Calculate the determinant $\left| A \right|$ and make sure it is non-zero.
  2. Construct the union matrix $S$, i.e. count 100500 algebraic additions $((A)_(ij))$ and place them in place $((a)_(ij))$.
  3. Transpose this matrix $S$, and then multiply it by some number $q=(1)/(\left| A \right|)\;$.

That's all! The inverse matrix $((A)^(-1))$ has been found. Let's look at examples:

\[\left[ \begin(matrix) 3 & 1 \\ 5 & 2 \\\end(matrix) \right]\]

Solution. Let's check the reversibility. Let's calculate the determinant:

\[\left| A\right|=\left| \begin(matrix) 3 & 1 \\ 5 & 2 \\\end(matrix) \right|=3\cdot 2-1\cdot 5=6-5=1\]

The determinant is different from zero. This means the matrix is ​​invertible. Let's create a union matrix:

Let's calculate the algebraic additions:

\[\begin(align) & ((A)_(11))=((\left(-1 \right))^(1+1))\cdot \left| 2 \right|=2; \\ & ((A)_(12))=((\left(-1 \right))^(1+2))\cdot \left| 5 \right|=-5; \\ & ((A)_(21))=((\left(-1 \right))^(2+1))\cdot \left| 1 \right|=-1; \\ & ((A)_(22))=((\left(-1 \right))^(2+2))\cdot \left| 3\right|=3. \\ \end(align)\]

Please note: the determinants |2|, |5|, |1| and |3| are determinants of matrices of size $\left[ 1\times 1 \right]$, and not modules. Those. If there were negative numbers in the determinants, there is no need to remove the “minus”.

In total, our union matrix looks like this:

\[((A)^(-1))=\frac(1)(\left| A \right|)\cdot ((S)^(T))=\frac(1)(1)\cdot ( (\left[ \begin(array)(*(35)(r)) 2 & -5 \\ -1 & 3 \\\end(array) \right])^(T))=\left[ \begin (array)(*(35)(r)) 2 & -1 \\ -5 & 3 \\\end(array) \right]\]

OK it's all over Now. The problem is solved.

Answer. $\left[ \begin(array)(*(35)(r)) 2 & -1 \\ -5 & 3 \\\end(array) \right]$

Task. Find the inverse matrix:

\[\left[ \begin(array)(*(35)(r)) 1 & -1 & 2 \\ 0 & 2 & -1 \\ 1 & 0 & 1 \\\end(array) \right] \]

Solution. We calculate the determinant again:

\[\begin(align) & \left| \begin(array)(*(35)(r)) 1 & -1 & 2 \\ 0 & 2 & -1 \\ 1 & 0 & 1 \\\end(array) \right|=\begin(matrix ) \left(1\cdot 2\cdot 1+\left(-1 \right)\cdot \left(-1 \right)\cdot 1+2\cdot 0\cdot 0 \right)- \\ -\left (2\cdot 2\cdot 1+\left(-1 \right)\cdot 0\cdot 1+1\cdot \left(-1 \right)\cdot 0 \right) \\\end(matrix)= \ \ & =\left(2+1+0 \right)-\left(4+0+0 \right)=-1\ne 0. \\ \end(align)\]

The determinant is nonzero—the matrix is ​​invertible. But now it’s going to be really tough: we need to count as many as 9 (nine, motherfucker!) algebraic additions. And each of them will contain the determinant $\left[ 2\times 2 \right]$. Flew:

\[\begin(matrix) ((A)_(11))=((\left(-1 \right))^(1+1))\cdot \left| \begin(matrix) 2 & -1 \\ 0 & 1 \\\end(matrix) \right|=2; \\ ((A)_(12))=((\left(-1 \right))^(1+2))\cdot \left| \begin(matrix) 0 & -1 \\ 1 & 1 \\\end(matrix) \right|=-1; \\ ((A)_(13))=((\left(-1 \right))^(1+3))\cdot \left| \begin(matrix) 0 & 2 \\ 1 & 0 \\\end(matrix) \right|=-2; \\ ... \\ ((A)_(33))=((\left(-1 \right))^(3+3))\cdot \left| \begin(matrix) 1 & -1 \\ 0 & 2 \\\end(matrix) \right|=2; \\ \end(matrix)\]

In short, the union matrix will look like this:

Therefore, the inverse matrix will be:

\[((A)^(-1))=\frac(1)(-1)\cdot \left[ \begin(matrix) 2 & -1 & -2 \\ 1 & -1 & -1 \\ -3 & 1 & 2 \\\end(matrix) \right]=\left[ \begin(array)(*(35)(r))-2 & -1 & 3 \\ 1 & 1 & -1 \ \2 & 1 & -2 \\\end(array) \right]\]

That's it. Here is the answer.

Answer. $\left[ \begin(array)(*(35)(r)) -2 & -1 & 3 \\ 1 & 1 & -1 \\ 2 & 1 & -2 \\\end(array) \right ]$

As you can see, at the end of each example we performed a check. In this regard, an important note:

Don't be lazy to check. Multiply the original matrix by the found inverse matrix - you should get $E$.

It is much easier and faster to perform this check than to look for an error in further calculations, when, for example, you solve a matrix equation.

Alternative way

As I said, the inverse matrix theorem works great for sizes $\left[ 2\times 2 \right]$ and $\left[ 3\times 3 \right]$ (in the latter case, it’s not so “great” "), but for larger matrices the sadness begins.

But don't worry: there is an alternative algorithm that can be used to calmly find the inverse even for the $\left[ 10\times 10 \right]$ matrix. But, as is often the case, to consider this algorithm, we need a little theoretical background.

Elementary transformations

Among the various transformations of the matrix, there are several special ones - they are called elementary. There are exactly three such transformations:

  1. Multiplication. You can take the $i$th row (column) and multiply it by any number $k\ne 0$;
  2. Addition. Add to the $i$-th row (column) any other $j$-th row (column), multiplied by any number $k\ne 0$ (you can, of course, do $k=0$, but what's the point? ? Nothing will change).
  3. Rearrangement. Take the $i$th and $j$th rows (columns) and swap places.

Why these transformations are called elementary (for large matrices they do not look so elementary) and why there are only three of them - these questions are beyond the scope of today's lesson. Therefore, we will not go into details.

Another thing is important: we have to perform all these perversions on the adjoint matrix. Yes, yes: you heard right. Now there will be one more definition - the last one in today's lesson.

Adjoint matrix

Surely at school you solved systems of equations using the addition method. Well, there, subtract another from one line, multiply some line by a number - that’s all.

So: now everything will be the same, but in an “adult” way. Ready?

Definition. Let a matrix $A=\left[ n\times n \right]$ and an identity matrix $E$ of the same size $n$ be given. Then the adjoint matrix $\left[ A\left| E\right. \right]$ is a new matrix of size $\left[ n\times 2n \right]$ that looks like this:

\[\left[ A\left| E\right. \right]=\left[ \begin(array)(rrrr|rrrr)((a)_(11)) & ((a)_(12)) & ... & ((a)_(1n)) & 1 & 0 & ... & 0 \\((a)_(21)) & ((a)_(22)) & ... & ((a)_(2n)) & 0 & 1 & ... & 0 \\... & ... & ... & ... & ... & ... & ... & ... \\((a)_(n1)) & ((a)_(n2)) & ... & ((a)_(nn)) & 0 & 0 & ... & 1 \\\end(array) \right]\]

In short, we take the matrix $A$, on the right we assign to it the identity matrix $E$ of the required size, we separate them with a vertical bar for beauty - here you have the adjoint. :)

What's the catch? And here's what:

Theorem. Let the matrix $A$ be invertible. Consider the adjoint matrix $\left[ A\left| E\right. \right]$. If using elementary string conversions bring it to the form $\left[ E\left| B\right. \right]$, i.e. by multiplying, subtracting and rearranging rows to obtain from $A$ the matrix $E$ on the right, then the matrix $B$ obtained on the left is the inverse of $A$:

\[\left[ A\left| E\right. \right]\to \left[ E\left| B\right. \right]\Rightarrow B=((A)^(-1))\]

It's that simple! In short, the algorithm for finding the inverse matrix looks like this:

  1. Write the adjoint matrix $\left[ A\left| E\right. \right]$;
  2. Perform elementary string conversions until $E$ appears instead of $A$;
  3. Of course, something will also appear on the left - a certain matrix $B$. This will be the opposite;
  4. PROFIT!:)

Of course, this is much easier said than done. So let's look at a couple of examples: for sizes $\left[ 3\times 3 \right]$ and $\left[ 4\times 4 \right]$.

Task. Find the inverse matrix:

\[\left[ \begin(array)(*(35)(r)) 1 & 5 & 1 \\ 3 & 2 & 1 \\ 6 & -2 & 1 \\\end(array) \right]\ ]

Solution. We create the adjoint matrix:

\[\left[ \begin(array)(rrr|rrr) 1 & 5 & 1 & 1 & 0 & 0 \\ 3 & 2 & 1 & 0 & 1 & 0 \\ 6 & -2 & 1 & 0 & 0 & 1 \\\end(array) \right]\]

Since the last column of the original matrix is ​​filled with ones, subtract the first row from the rest:

\[\begin(align) & \left[ \begin(array)(rrr|rrr) 1 & 5 & 1 & 1 & 0 & 0 \\ 3 & 2 & 1 & 0 & 1 & 0 \\ 6 & - 2 & 1 & 0 & 0 & 1 \\\end(array) \right]\begin(matrix) \downarrow \\ -1 \\ -1 \\\end(matrix)\to \\ & \to \left [ \begin(array)(rrr|rrr) 1 & 5 & 1 & 1 & 0 & 0 \\ 2 & -3 & 0 & -1 & 1 & 0 \\ 5 & -7 & 0 & -1 & 0 & 1 \\\end(array) \right] \\ \end(align)\]

There are no more units, except for the first line. But we don’t touch it, otherwise the newly removed units will begin to “multiply” in the third column.

But we can subtract the second line twice from the last - we get one in the lower left corner:

\[\begin(align) & \left[ \begin(array)(rrr|rrr) 1 & 5 & 1 & 1 & 0 & 0 \\ 2 & -3 & 0 & -1 & 1 & 0 \\ 5 & -7 & 0 & -1 & 0 & 1 \\\end(array) \right]\begin(matrix) \ \\ \downarrow \\ -2 \\\end(matrix)\to \\ & \left [ \begin(array)(rrr|rrr) 1 & 5 & 1 & 1 & 0 & 0 \\ 2 & -3 & 0 & -1 & 1 & 0 \\ 1 & -1 & 0 & 1 & -2 & 1 \\\end(array) \right] \\ \end(align)\]

Now we can subtract the last row from the first and twice from the second - this way we “zero” the first column:

\[\begin(align) & \left[ \begin(array)(rrr|rrr) 1 & 5 & 1 & 1 & 0 & 0 \\ 2 & -3 & 0 & -1 & 1 & 0 \\ 1 & -1 & 0 & 1 & -2 & 1 \\\end(array) \right]\begin(matrix) -1 \\ -2 \\ \uparrow \\\end(matrix)\to \\ & \ to \left[ \begin(array)(rrr|rrr) 0 & 6 & 1 & 0 & 2 & -1 \\ 0 & -1 & 0 & -3 & 5 & -2 \\ 1 & -1 & 0 & 1 & -2 & 1 \\\end(array) \right] \\ \end(align)\]

Multiply the second row by −1 and then subtract it 6 times from the first and add 1 time to the last:

\[\begin(align) & \left[ \begin(array)(rrr|rrr) 0 & 6 & 1 & 0 & 2 & -1 \\ 0 & -1 & 0 & -3 & 5 & -2 \ \ 1 & -1 & 0 & 1 & -2 & 1 \\\end(array) \right]\begin(matrix) \ \\ \left| \cdot \left(-1 \right) \right. \\ \ \\\end(matrix)\to \\ & \to \left[ \begin(array)(rrr|rrr) 0 & 6 & 1 & 0 & 2 & -1 \\ 0 & 1 & 0 & 3 & -5 & 2 \\ 1 & -1 & 0 & 1 & -2 & 1 \\\end(array) \right]\begin(matrix) -6 \\ \updownarrow \\ +1 \\\end (matrix)\to \\ & \to \left[ \begin(array)(rrr|rrr) 0 & 0 & 1 & -18 & 32 & -13 \\ 0 & 1 & 0 & 3 & -5 & 2 \\ 1 & 0 & 0 & 4 & -7 & 3 \\\end(array) \right] \\ \end(align)\]

All that remains is to swap lines 1 and 3:

\[\left[ \begin(array)(rrr|rrr) 1 & 0 & 0 & 4 & -7 & 3 \\ 0 & 1 & 0 & 3 & -5 & 2 \\ 0 & 0 & 1 & - 18 & 32 & -13 \\\end(array) \right]\]

Ready! On the right is the required inverse matrix.

Answer. $\left[ \begin(array)(*(35)(r))4 & -7 & 3 \\ 3 & -5 & 2 \\ -18 & 32 & -13 \\\end(array) \right ]$

Task. Find the inverse matrix:

\[\left[ \begin(matrix) 1 & 4 & 2 & 3 \\ 1 & -2 & 1 & -2 \\ 1 & -1 & 1 & 1 \\ 0 & -10 & -2 & -5 \\\end(matrix) \right]\]

Solution. We compose the adjoint again:

\[\left[ \begin(array)(rrrr|rrrr) 1 & 4 & 2 & 3 & 1 & 0 & 0 & 0 \\ 1 & -2 & 1 & -2 & 0 & 1 & 0 & 0 \ \ 1 & -1 & 1 & 1 & 0 & 0 & 1 & 0 \\ 0 & -10 & -2 & -5 & 0 & 0 & 0 & 1 \\\end(array) \right]\]

Let's cry a little, be sad about how much we have to count now... and start counting. To begin with, we “zero out” the first column by subtracting row 1 from rows 2 and 3:

\[\begin(align) & \left[ \begin(array)(rrrr|rrrr) 1 & 4 & 2 & 3 & 1 & 0 & 0 & 0 \\ 1 & -2 & 1 & -2 & 0 & 1 & 0 & 0 \\ 1 & -1 & 1 & 1 & 0 & 0 & 1 & 0 \\ 0 & -10 & -2 & -5 & 0 & 0 & 0 & 1 \\\end(array) \right]\begin(matrix) \downarrow \\ -1 \\ -1 \\ \ \\\end(matrix)\to \\ & \to \left[ \begin(array)(rrrr|rrrr) 1 & 4 & 2 & 3 & 1 & 0 & 0 & 0 \\ 0 & -6 & -1 & -5 & -1 & 1 & 0 & 0 \\ 0 & -5 & -1 & -2 & -1 & 0 & 1 & 0 \\ 0 & -10 & -2 & -5 & 0 & 0 & 0 & 1 \\\end(array) \right] \\ \end(align)\]

We see too many “cons” in lines 2-4. Multiply all three rows by −1, and then burn out the third column by subtracting row 3 from the rest:

\[\begin(align) & \left[ \begin(array)(rrrr|rrrr) 1 & 4 & 2 & 3 & 1 & 0 & 0 & 0 \\ 0 & -6 & -1 & -5 & - 1 & 1 & 0 & 0 \\ 0 & -5 & -1 & -2 & -1 & 0 & 1 & 0 \\ 0 & -10 & -2 & -5 & 0 & 0 & 0 & 1 \\ \end(array) \right]\begin(matrix) \ \\ \left| \cdot \left(-1 \right) \right. \\ \left| \cdot \left(-1 \right) \right. \\ \left| \cdot \left(-1 \right) \right. \\\end(matrix)\to \\ & \to \left[ \begin(array)(rrrr|rrrr) 1 & 4 & 2 & 3 & 1 & 0 & 0 & 0 \\ 0 & 6 & 1 & 5 & ​​1 & -1 & 0 & 0 \\ 0 & 5 & 1 & 2 & 1 & 0 & -1 & 0 \\ 0 & 10 & 2 & 5 & 0 & 0 & 0 & -1 \\\end (array) \right]\begin(matrix) -2 \\ -1 \\ \updownarrow \\ -2 \\\end(matrix)\to \\ & \to \left[ \begin(array)(rrrr| rrrr) 1 & -6 & 0 & -1 & -1 & 0 & 2 & 0 \\ 0 & 1 & 0 & 3 & 0 & -1 & 1 & 0 \\ 0 & 5 & 1 & 2 & 1 & 0 & -1 & 0 \\ 0 & 0 & 0 & 1 & -2 & 0 & 2 & -1 \\\end(array) \right] \\ \end(align)\]

Now is the time to “fry” the last column of the original matrix: subtract row 4 from the rest:

\[\begin(align) & \left[ \begin(array)(rrrr|rrrr) 1 & -6 & 0 & -1 & -1 & 0 & 2 & 0 \\ 0 & 1 & 0 & 3 & 0 & -1 & 1 & 0 \\ 0 & 5 & 1 & 2 & 1 & 0 & -1 & 0 \\ 0 & 0 & 0 & 1 & -2 & 0 & 2 & -1 \\\end(array ) \right]\begin(matrix) +1 \\ -3 \\ -2 \\ \uparrow \\\end(matrix)\to \\ & \to \left[ \begin(array)(rrrr|rrrr) 1 & -6 & 0 & 0 & -3 & 0 & 4 & -1 \\ 0 & 1 & 0 & 0 & 6 & -1 & -5 & 3 \\ 0 & 5 & 1 & 0 & 5 & 0 & -5 & 2 \\ 0 & 0 & 0 & 1 & -2 & 0 & 2 & -1 \\\end(array) \right] \\ \end(align)\]

Final throw: “burn out” the second column by subtracting line 2 from lines 1 and 3:

\[\begin(align) & \left[ \begin(array)(rrrr|rrrr) 1 & -6 & 0 & 0 & -3 & 0 & 4 & -1 \\ 0 & 1 & 0 & 0 & 6 & -1 & -5 & 3 \\ 0 & 5 & 1 & 0 & 5 & 0 & -5 & 2 \\ 0 & 0 & 0 & 1 & -2 & 0 & 2 & -1 \\\end( array) \right]\begin(matrix) 6 \\ \updownarrow \\ -5 \\ \ \\\end(matrix)\to \\ & \to \left[ \begin(array)(rrrr|rrrr) 1 & 0 & 0 & 0 & 33 & -6 & -26 & -17 \\ 0 & 1 & 0 & 0 & 6 & -1 & -5 & 3 \\ 0 & 0 & 1 & 0 & -25 & 5 & 20 & -13 \\ 0 & 0 & 0 & 1 & -2 & 0 & 2 & -1 \\\end(array) \right] \\ \end(align)\]

And again the identity matrix is ​​on the left, which means the inverse is on the right. :)

Answer. $\left[ \begin(matrix) 33 & -6 & -26 & 17 \\ 6 & -1 & -5 & 3 \\ -25 & 5 & 20 & -13 \\ -2 & 0 & 2 & - 1 \\\end(matrix) \right]$

OK it's all over Now. Do the check yourself - I'm screwed. :)

Loading...
Top