Finding Inverse of Matrix by using Gaussian-Jordan Elimination

Click For Summary

Discussion Overview

The discussion revolves around understanding the process of finding the inverse of a matrix using Gaussian-Jordan elimination. Participants explore the theoretical underpinnings of why the method works, particularly the relationship between row operations and elementary matrices, as well as the implications for solving systems of equations.

Discussion Character

  • Exploratory
  • Technical explanation
  • Conceptual clarification
  • Debate/contested

Main Points Raised

  • Some participants explain that each row operation corresponds to an elementary matrix, which transforms the identity matrix into the inverse of the original matrix through a series of operations.
  • One participant questions the reasoning behind the transformation from AI to IA' during row operations, indicating a lack of understanding of the underlying principles.
  • Another participant provides a more intuitive explanation by relating the process to solving systems of equations, suggesting that the same principles apply when finding the inverse of a matrix.
  • There is a discussion about the general method of solving linear equations using Gaussian elimination, with emphasis on the multiplication of matrices and how it leads to the identity matrix.
  • Some participants clarify that applying row operations to the identity matrix yields elementary matrices that can be used to derive the inverse of the original matrix.

Areas of Agreement / Disagreement

Participants express varying levels of understanding regarding the theoretical justification for the method. While some explanations are provided, there is no consensus on the clarity of the reasoning behind the transformation of matrices or the application of row operations.

Contextual Notes

Some participants express uncertainty about specific steps in the Gaussian-Jordan elimination process and the relationship between row operations and elementary matrices. There are also references to solving systems of equations that may not be fully resolved within the discussion.

shuiki0306
Messages
2
Reaction score
0
Hello. Nice to meet you. I have just enrolled. :)
I knew how to solve and to find out inverse Matrix by using Gaussian elimination.
However, I was wondering why AI -> IA' is satisfactory.
In my university, I was just taught how to use but wasn't taught why it is satisfactory.
Thank you for answer.
(I am non-native, so my English is a little bad. Sorry giving challenge to you)
 
Physics news on Phys.org
That works because each row operation corresponds to an "elementary matrix", the matrix we get by applying that row operation to the identity matrix. That is the row operation "add 3 times the second row to the first row" corresponds to the "elementary matrix"
\begin{pmatrix}1 & 3 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1\end{pmatrix}

Multiplying that matrix by any matrix, A, will add three times the second row of A to its first row:
\begin{pmatrix}1 & 3 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1\end{pmatrix}\begin{pmatrix}a_{11} & a_{12} & a_{13} \\ a_{21} & a_{22} & a_{23} \\ a_{31} & a_{32} & a_{33}\end{pmatrix}= \begin{pmatrix}a_{11}+ 3a_{21} & a_{12}+ 3a_{22} & a_{13}+ 3a_{23} \\ a_{21} & a_{22} & a_{23} \\ a_{31} & a_{32} & a_{33}\end{pmatrix}


So deciding what row operations will reduce A to the identity matrix means finding elementary matrices, e1, e2, e3, ..., en, each corresponding to a row operation, such that (en...e3e2e1)A= I. Of course, the product of those elementary matrices, en...e3e2d1, is the inverse matrix to A. Applying those same row operations to the identity matrix is just the same as multiplying them: en...e3e2e1I= en...e3e2e1= A^-1.
 
ImageUploadedByPhysics Forums1409759840.771257.jpg

I was wondering why this is correct.
Actually, I solved problem in my exam (this style).
But, I still cannot understand why A I change to I A' with row operation.
 
What part of my previous response did you not understand?
 
A slightly longer but (I hope) more intuitive explanation for why the method works.

Do you understand why solving a system of equations with Gauss Jordan works? That is, do you understand why solving the systems
<br /> \begin{align*}<br /> 5x + 7y = 1 &amp; \\<br /> 8x - 3y = 0 &amp; <br /> \end{align*}<br />

by reducing this augmented matrix

<br /> \begin{bmatrix} 5 &amp; 7 &amp; 1 \\ 8 &amp; -3 &amp; 0 \end{bmatrix}<br />

gives the solution?

If so: If you want to find the inverse of the matrix from that system you want another matrix that satisfies
<br /> \begin{bmatrix} 5 &amp; 7 \\ 8 &amp; -3 \end{bmatrix} \begin{bmatrix} x &amp; s \\ y &amp; t \end{bmatrix} = \begin{bmatrix} 1 &amp; 0 \\ 0 &amp; 1 \end{bmatrix}<br />

The solution to the system above gives the required elements x and y of the second matrix: the elements s and t would be the solutions to
this system

\begin{align*}
5s + 7t = 0 & \\
8s - 3t = 1 &
\end{align*}

and could be found by applying Gauss Jordan to this augmented matrix

<br /> \begin{bmatrix} 5 &amp; 7 &amp; 0 \\ 8 &amp; -3 &amp; 1 \end{bmatrix}<br />

Solve those two systems and you have the entries for the inverse of A. The method you ask about for finding inverses simply says this: instead of going through two bouts of row operations and augmented matrices, simply but both columns of the identity matrix in a single augmented matrix and do the row operations once - that is, set up
\begin{bmatrix} 5 & 7 & 1 & 0 \\ 8 & -3 & 0 & 1 \end{bmatrix}

and reduce it. The third and fourth columns of the result are the required entries for the inverse.
 
Here is why that algorithm works. I will express it more generally in terms of solving a system of linear equations: A.x = B where B is a column vector or a matrix. Solution by Gaussian elimination consists of multiplying both sides on the left by some matrices, one for each step of elimination, pivoting, and backsubstitution.. Matrices T1, T2, ..., Tn:
Tn...T2.T1.A.x = Tn...T2.T1.B
Or for short,
T.A.x = T.B

When we have reached a solution, T.A = I, the identity matrix, making T = A-1. If B = I, then T.B = A-1, which is why this algorithm works.
 
The matrices lpetrich is talking about are called, in some texts, "elementary" matrices. If you apply a "row operation" to the identity matrix, you get a matrix, R, that has the nice property that the multiplication, RA, gives the same result as applying that row operation to matrix A. For example, in three dimensions, applying the row operation "add twice the first row to the third row" to the identity matrix gives \begin{bmatrix} 1 &amp; 0 &amp; 0 \\ 0 &amp; 1 &amp; 0 \\ 2 &amp; 0 &amp; 1\end{bmatrix}.

And multiplying that by matrix A gives
\begin{bmatrix} 1 &amp; 0 &amp; 0 \\ 0 &amp; 1 &amp; 0 \\ 2 &amp; 0 &amp; 1\end{bmatrix}\begin{bmatrix}a_{11} &amp; a_{12} &amp; a_{13} \\ a_{21} &amp; a_{22} &amp; a_{23} \\ a_{31} &amp; a_{32} &amp; a_{33}\end{bmatrix}= \begin{bmatrix}a_{11} &amp; a_{12} &amp; a_{13} \\ a_{21} &amp; a_{22} &amp; a_{23} \\ 2a_{11}+ a_{31} &amp; 2a_{12}+ a_{32} &amp; 2a_{13}+ a_{33}\end{bmatrix}, exactly the same as adding twice the first row of A to its third row.

So if I can find a sequence of row operations that reduce matrix A to the identity matrix then I have found a sequence of elementary matrices (even if I never write them explicitly) whose product is the inverse matrix to A. Applying those operations to the identity matrix is the same as multiplying all of the corresponding elementary matrices so gives the inverse matrix to A. Applying those operations to a vector is the same as multiplying that vector by all those elementary matrices which is the same as A^{-1}v.
 

Similar threads

  • · Replies 34 ·
2
Replies
34
Views
3K
  • · Replies 14 ·
Replies
14
Views
4K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 8 ·
Replies
8
Views
1K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 7 ·
Replies
7
Views
4K
  • · Replies 4 ·
Replies
4
Views
6K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 7 ·
Replies
7
Views
3K