Can You Reduce Simultaneous Equations by Eliminating Variables?

  • Context: Undergrad 
  • Thread starter Thread starter BobbyBear
  • Start date Start date
  • Tags Tags
    Variables
Click For Summary

Discussion Overview

The discussion revolves around the concept of eliminating variables from simultaneous equations, particularly in the context of reducing the number of equations and variables. Participants explore whether it is possible to systematically reduce a set of equations while maintaining the relationships between the variables, considering both linear and non-linear cases.

Discussion Character

  • Exploratory
  • Technical explanation
  • Conceptual clarification
  • Debate/contested
  • Mathematical reasoning

Main Points Raised

  • One participant questions whether it is possible to eliminate variables from a set of simultaneous equations, suggesting that if there are more variables than equations, one could reduce the system to fewer variables.
  • Another participant agrees with the general principle that m equations with n variables can be reduced to m-1 equations with n-1 variables, eventually leading to a single equation with n-(m-1) variables.
  • Examples are provided to illustrate the process of eliminating variables, including finding a basis for a subspace defined by two equations in three variables.
  • A participant discusses the transition from implicit to explicit functions, suggesting that eliminating variables can lead to explicit equations from implicit ones, and vice versa.
  • Concerns are raised about the independence of equations, with participants noting that dependent equations may yield the same results when attempting to eliminate variables.
  • One participant proposes that the same dependency issues may arise in non-linear equations, seeking clarification on whether this holds true and how it might be proven.

Areas of Agreement / Disagreement

Participants express a general agreement on the principles of variable elimination, but there is contention regarding the independence of equations and the implications of dependency on the elimination process. The discussion remains unresolved on the broader applicability of these principles to non-linear equations.

Contextual Notes

Participants acknowledge the need for assumptions about the independence of equations and the potential limitations of their examples. There is also an exploration of the implications of variable elimination in both linear and non-linear contexts, but no definitive conclusions are reached.

BobbyBear
Messages
162
Reaction score
1
I'm a bit stuck with the idea of eliminating variables from a set of simultaneous equations . . . for example, suppose you have two equations (with more than 3 variables), could you, in principle, reduce it to one equation with one variable less?

And if you had three equations with let's say 4 variables, could you, in principle, reduce them to 2 equations with 3 variables, and them in turn to one equation with 2 variables?

So in general, if you have m equations and n variables, with n>m, then can you, in principle, eliminate m-1 variables, so that you end up with a single equation with n-(m-1) variables? Is this correct? (I say in principle because I suppose you need to assume all the equations are independent and stuff like that:P).
 
Physics news on Phys.org
Yes, in general, if you have m equations with n variables, n> m, you can reduce to m-1 equations with n-1 variables and, repeating, eventually arrive at one equation with n-(m-1)= n-m+1 variable. You could then write one of those variables as a function of the other n-m variables and, working back, eventually write all variables as functions of those n-m "free variables".

Here is an example. Find a basis for the subspace of all (x, y, z) in R3 satisfying x+ y- z= 0 and x- y+ 2z= 0 (two equations in 3 variables). Adding those equations together I get 2x+ z= 0 (one equation in two variabes) so z= -2x. Putting that into the first equation, x+ y-(-2x)= 3x+ y= 0 so y= -3x. Any such vector is of the form (x, -3x, -2x)= x(1, -3, -2). The single vector (1, -3, -2) forms a basis for that (one dimensional) subspace.
 
HallsofIvy said:
Yes, in general, if you have m equations with n variables, n> m, you can reduce to m-1 equations with n-1 variables and, repeating, eventually arrive at one equation with n-(m-1)= n-m+1 variable. You could then write one of those variables as a function of the other n-m variables and, working back, eventually write all variables as functions of those n-m "free variables".

Here is an example. Find a basis for the subspace of all (x, y, z) in R3 satisfying x+ y- z= 0 and x- y+ 2z= 0 (two equations in 3 variables). Adding those equations together I get 2x+ z= 0 (one equation in two variabes) so z= -2x. Putting that into the first equation, x+ y-(-2x)= 3x+ y= 0 so y= -3x. Any such vector is of the form (x, -3x, -2x)= x(1, -3, -2). The single vector (1, -3, -2) forms a basis for that (one dimensional) subspace.

Thank you Ivy:)

so basically what you have done in that example is write a set of two implicit functions as explicit functions of x, that is, you have solved the system for y and z (can I put it like that?). It's like, you've considered the 'free' variable to be x, ie., as if it were a parameter instead of a variable, and you've solved the system of two equations with two unknown variables y and z.

So eliminating variables is the first step to actually solving an implicit system of equations to end up with explicit equations, or parametric equations, yes? But it's also obtaining implicit equations from explicit ones, no?

Say for example, I'm given a curve in space by its parametric equations:

x=f(s)
y=g(s)
z=h(s)

I could eliminate s to obtain two implicit functions by doing the following:

x=f(s) \rightarrow s=f^{-1}(x)
y=g(s) \rightarrow s=g^{-1}(y)
z=h(s) \rightarrow s=h^{-1}(z)

from which:

f^{-1}(x)=g^{-1}(y)
g^{-1}(y)=h^{-1}(z)

ie.
F(x,y)=0
G(y,z)=0

which would be the two surfaces whose intersection is the curve.



But now suppose I'm given the parametric equations of a surface:

x=f(u,v)
y=g(u,v)
z=h(u,v)

In theory, I could eliminate u and v and I'd be left with a single equation in the variables x,y and z, which would be the implicit equation of the surface. But, say if from the first two equations I solve for u and v, and I do the same using the second pair of equations, I'd have:

u=f_1(x,y)
v=f_2(x,y)
and
u=g_1(y,z)
v=g_2(y,z)

and eliminating u and v from these four equations, I have:

f_1(x,y)=g_1(y,z)
f_2(x,y)=g_2(y,z)

ie. I end up with two equations in the variables x,y and z instead of just one!

f_1(x,y)=g_1(y,z) \rightarrow F(x,y,z)=0
f_2(x,y)=g_2(y,z) \rightarrow G(x,y,z)=0

which define a curve, not a surface. Can you tell me where I am going wrong?
 
The two equations are not independent. For example, suppose your equations are x= u+v, y= u- v, z= 2u+ v (describing a plane). From the first two equations you get u= (x+y)/2 and then v= u- y= (x- y)/2. From the last two equations, u= (y+z)/3 and then v= u- y= (z- 2y)/3. Now combine those: u= (x+y)/2= (y+z)/3 give 3x+ 3y= 2y+ 2z or 3x+ y- 2z= 0. v= (x- y)/2= (z- 2y)/3 or 3x- 3y= 2z- 4y: 3x+ y- 2z= 0 again!
 
HallsofIvy said:
The two equations are not independent.

Ivy, I love you!

Okay but, I assume this will always be the case . . . but how can one see this if it's not through a particular example?
 
Okay wait, in the case of linear equations it's easy to show in general that the equations are dependent:

x=f(u,v)=Au+Bv
y=g(u,v)=Cu+Dv
z=h(u,v)=Eu+Fv

Applying Cramer's rule, from the frist two we get:

\begin{array}{cc}<br /> x=Au+Bv\\<br /> y=Cu+Dv<br /> \end{array} \right\} \rightarrow}<br />

u=\frac{ det \left(<br /> \begin{array}{cc}<br /> x &amp; B\\<br /> y &amp; D<br /> \end{array}<br /> \right)}{det \left(<br /> \begin{array}{cc}<br /> A &amp; B\\<br /> C &amp; D<br /> \end{array}<br /> \right)} = \frac{xD-yB}{AD-CB}; (1)

v=\frac{ det \left(<br /> \begin{array}{cc}<br /> A &amp; x\\<br /> C &amp; y<br /> \end{array}<br /> \right)}{det \left(<br /> \begin{array}{cc}<br /> A &amp; B\\<br /> C &amp; D<br /> \end{array}<br /> \right)} = \frac{yA-xC}{AD-CB}; (2)


and from the second two:

\begin{array}{cc}<br /> y=Cu+Dv\\<br /> z=Eu+Fv<br /> \end{array} \right\} \rightarrow}<br />

u=\frac{ det \left(<br /> \begin{array}{cc}<br /> y &amp; D\\<br /> z &amp; F<br /> \end{array}<br /> \right)}{det \left(<br /> \begin{array}{cc}<br /> C &amp; D\\<br /> E &amp; F<br /> \end{array}<br /> \right)} = \frac{yF-zD}{CF-ED}; (3)

v=\frac{ det \left(<br /> \begin{array}{cc}<br /> C &amp; y\\<br /> E &amp; z<br /> \end{array}<br /> \right)}{det \left(<br /> \begin{array}{cc}<br /> C &amp; D\\<br /> E &amp; F<br /> \end{array}<br /> \right)} = \frac{zC-yE}{CF-ED}; (4)

From (1) and (3) we obtain, after simplifying, the equation of the plane:

(CF-ED)Dx-(FA-BE)Dy-(AD-CB)Dz=0

and it's easy to see that from (2) and (4) we get the same equation of the plane.



But the same would happen if the equations were non-linear? And is there a way to prove it?
 

Similar threads

  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 3 ·
Replies
3
Views
2K
Replies
2
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 40 ·
2
Replies
40
Views
5K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 7 ·
Replies
7
Views
3K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K