Why does Gaussian elimination work?

Click For Summary

Discussion Overview

The discussion revolves around the workings of Gaussian elimination, specifically addressing its validity and the meaning behind obtaining an identity matrix on one side of an augmented matrix. Participants explore the theoretical underpinnings, practical applications, and historical context of the method as it relates to solving systems of linear equations.

Discussion Character

  • Exploratory
  • Technical explanation
  • Conceptual clarification
  • Debate/contested
  • Mathematical reasoning

Main Points Raised

  • Some participants express curiosity about the fundamental reasons why Gaussian elimination works, particularly in relation to the identity matrix and solutions.
  • Others explain that the operations performed in Gaussian elimination correspond to valid manipulations of the original system of equations, preserving the solution set.
  • A participant highlights that Gaussian elimination predates the formal introduction of matrices, suggesting that the method can be understood without them, focusing instead on the algebraic manipulations involved.
  • One participant describes the process of eliminating variables step-by-step, demonstrating how to simplify a system of equations by expressing one variable in terms of others.
  • Another participant emphasizes that the three operations used in Gaussian elimination are those that maintain the solution set, suggesting a deeper exploration of why these operations are valid.
  • There is a discussion about the equivalence of operations performed on both sides of the equations, reinforcing the idea that the solutions remain unchanged throughout the process.

Areas of Agreement / Disagreement

Participants generally agree on the validity of the operations used in Gaussian elimination and their role in preserving solutions. However, there is no consensus on the necessity of matrices in understanding the method, with some arguing for their historical context and others focusing on the algebraic principles.

Contextual Notes

Some participants mention the importance of performing legal algebraic operations and avoiding illegal steps, such as dividing by zero, but do not resolve the implications of these limitations in the context of Gaussian elimination.

Mathematicsss
Post moved by moderator from Homework section
Hello, I was curious as to why Gaussian elimination works. I know that if we have two ( or more) systems of two (or more) linear equations, we can write then in terms of a matrix. However, what does it mean when I get the identity on the left hand side and the solution on the right?
 
Last edited by a moderator:
Physics news on Phys.org
Mathematicsss said:
Post moved by moderator from Homework section
Hello, I was curious as to why Gaussian elimination works. I know that if we have two ( or more) systems of two (or more) linear equations, we can write then in terms of a matrix. However, what does it mean when I get the identity on the left hand side and the solution on the right?
How would you solve a system like
$$
2x+3y = 4 \\
x-3y = 0
$$
or
$$
-x+ 4y = -3 \\
y= 2
$$
Gauß elimination is basically the same process in general terms, so it fits all possible linear equation systems.
 
Mathematicsss said:
Post moved by moderator from Homework section
Hello, I was curious as to why Gaussian elimination works. I know that if we have two ( or more) systems of two (or more) linear equations, we can write then in terms of a matrix. However, what does it mean when I get the identity on the left hand side and the solution on the right?
You don't have "two (or more) systems" of linear equations -- you have a system of one or more linear equations, which you can write in a shorthand form as an augmented matrix.

The operations you do on the matrix are the same that you can do with the system of equations: exchange two equations, add a multiple of one equation to another equation, replace one equation by a nonzero multiple of itself. In each of these operations, there is no change in the solution.

Suppose you have completely reduced a matrix equation so that the augmented matrix looks like this:
##\begin{bmatrix} 1 & 0 & 0 & | & 3 \\ 0 & 1 & 0 & | & 1 \\ 0 & 0 & 1 & | & -3\end{bmatrix}##
The corresponding system of equations would be
1x + 0y + 0z = 3
0x + 1y + 0z = 1
0x + 0y + 1z = -1

Or more simply,
x = 3
y = 1
z = -1
 
Mathematicsss said:
Post moved by moderator from Homework section
Hello, I was curious as to why Gaussian elimination works. I know that if we have two ( or more) systems of two (or more) linear equations, we can write then in terms of a matrix. However, what does it mean when I get the identity on the left hand side and the solution on the right?

Gaussian elimination has been around hundreds of years before the invention of matrices, so matrices are not needed at all (but they make the work cleaner and easier). The method just solves equations successively, each time eliminating a new independent variable.

For example, if we have the system
$$\begin{array}{crcc}
(1):& x + 2y - 3z &=& 10\\
(2):& 2x - y + 4z &=& 20 \\
(3):& 2x - z &=& 8
\end{array}
$$
We can use eq (1) to express ##x## in terms of ##y## and ##z##:
$$(1a): \; x = 10 - 2y + 3z. $$
Substitute (1a) into (2) and (3), to get
$$\begin{array}{crrcc}
(2a):& 20-5y+10z = 20 & \Rightarrow& 5y - 10 z = 0 \\
(3a): & 20-4y+5z = 8 & \Rightarrow& 4y - 5z = 12
\end{array}
$$
So, we have a new system of 2 equations in the two unknowns ##y,z##. We have "eliminated" ##x## and made the problem smaller and easier to deal with. We can continue, for example by using eq. (2a) to express ##y## in terms of ##z##. That will leave us with a single equation (3b) that contains ##y## alone.

The reason it "works" is that each step involves a valid re-writing of the equation or equations, using only standard algebraic rules, and never, ever, performing an illegal step such as dividing by 0. For example, we obtained eq. (2a) by re-writing (2) as ##2(10 - 2y + 3z) -y + 4z = 20## and simplifying.
 
  • Like
Likes   Reactions: mfb
To rephrase, the three operations used in Gaussian elimination are precisely those that preserve the solution set. It is a nice exercise to show those three are precisely the ones that preserve solutions (EDIT: Trivial for all but one; adding linear combinations of one row to another)..
 
You start with a set of equations, with formulas on the left side and constants on the right. You are looking for variable values that give equality. As long as you do the same thing on both sides, they remain equal and valid for those variable values. On one side you can perform operations in terms of the formulas and on the other side you can perform the same operations on the constants -- always the same operation, just in a different form. The variable solution values never change because the two sides always remain equal given those values. Finally, you end up with a set of simple equations like x1=c1; x2=c2; x3=c3; ... They give the solution values of the original equations and of every step in between.
 

Similar threads

  • · Replies 14 ·
Replies
14
Views
2K
  • · Replies 26 ·
Replies
26
Views
5K
  • · Replies 1 ·
Replies
1
Views
5K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 6 ·
Replies
6
Views
5K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 4 ·
Replies
4
Views
4K
Replies
1
Views
4K
  • · Replies 4 ·
Replies
4
Views
6K
  • · Replies 7 ·
Replies
7
Views
2K