Can You Reduce Simultaneous Equations by Eliminating Variables?

  • Thread starter Thread starter BobbyBear
  • Start date Start date
  • Tags Tags
    Variables
AI Thread Summary
Eliminating variables from simultaneous equations allows for the reduction of the number of equations and variables, enabling a clearer understanding of the relationships between them. In general, if there are m equations with n variables (where n > m), it is possible to reduce the system to m-1 equations with n-1 variables, ultimately leading to a single equation with n-m+1 variables. This process can be used to express one or more variables as functions of the remaining "free" variables. However, when attempting to eliminate variables from parametric equations of surfaces, one may end up with multiple equations instead of a single implicit equation, indicating that the equations may not be independent. The discussion highlights the importance of understanding the dependency of equations, whether linear or nonlinear, when performing variable elimination.
BobbyBear
Messages
162
Reaction score
1
I'm a bit stuck with the idea of eliminating variables from a set of simultaneous equations . . . for example, suppose you have two equations (with more than 3 variables), could you, in principle, reduce it to one equation with one variable less?

And if you had three equations with let's say 4 variables, could you, in principle, reduce them to 2 equations with 3 variables, and them in turn to one equation with 2 variables?

So in general, if you have m equations and n variables, with n>m, then can you, in principle, eliminate m-1 variables, so that you end up with a single equation with n-(m-1) variables? Is this correct? (I say in principle because I suppose you need to assume all the equations are independent and stuff like that:P).
 
Mathematics news on Phys.org
Yes, in general, if you have m equations with n variables, n> m, you can reduce to m-1 equations with n-1 variables and, repeating, eventually arrive at one equation with n-(m-1)= n-m+1 variable. You could then write one of those variables as a function of the other n-m variables and, working back, eventually write all variables as functions of those n-m "free variables".

Here is an example. Find a basis for the subspace of all (x, y, z) in R3 satisfying x+ y- z= 0 and x- y+ 2z= 0 (two equations in 3 variables). Adding those equations together I get 2x+ z= 0 (one equation in two variabes) so z= -2x. Putting that into the first equation, x+ y-(-2x)= 3x+ y= 0 so y= -3x. Any such vector is of the form (x, -3x, -2x)= x(1, -3, -2). The single vector (1, -3, -2) forms a basis for that (one dimensional) subspace.
 
HallsofIvy said:
Yes, in general, if you have m equations with n variables, n> m, you can reduce to m-1 equations with n-1 variables and, repeating, eventually arrive at one equation with n-(m-1)= n-m+1 variable. You could then write one of those variables as a function of the other n-m variables and, working back, eventually write all variables as functions of those n-m "free variables".

Here is an example. Find a basis for the subspace of all (x, y, z) in R3 satisfying x+ y- z= 0 and x- y+ 2z= 0 (two equations in 3 variables). Adding those equations together I get 2x+ z= 0 (one equation in two variabes) so z= -2x. Putting that into the first equation, x+ y-(-2x)= 3x+ y= 0 so y= -3x. Any such vector is of the form (x, -3x, -2x)= x(1, -3, -2). The single vector (1, -3, -2) forms a basis for that (one dimensional) subspace.

Thank you Ivy:)

so basically what you have done in that example is write a set of two implicit functions as explicit functions of x, that is, you have solved the system for y and z (can I put it like that?). It's like, you've considered the 'free' variable to be x, ie., as if it were a parameter instead of a variable, and you've solved the system of two equations with two unknown variables y and z.

So eliminating variables is the first step to actually solving an implicit system of equations to end up with explicit equations, or parametric equations, yes? But it's also obtaining implicit equations from explicit ones, no?

Say for example, I'm given a curve in space by its parametric equations:

x=f(s)
y=g(s)
z=h(s)

I could eliminate s to obtain two implicit functions by doing the following:

x=f(s) \rightarrow s=f^{-1}(x)
y=g(s) \rightarrow s=g^{-1}(y)
z=h(s) \rightarrow s=h^{-1}(z)

from which:

f^{-1}(x)=g^{-1}(y)
g^{-1}(y)=h^{-1}(z)

ie.
F(x,y)=0
G(y,z)=0

which would be the two surfaces whose intersection is the curve.



But now suppose I'm given the parametric equations of a surface:

x=f(u,v)
y=g(u,v)
z=h(u,v)

In theory, I could eliminate u and v and I'd be left with a single equation in the variables x,y and z, which would be the implicit equation of the surface. But, say if from the first two equations I solve for u and v, and I do the same using the second pair of equations, I'd have:

u=f_1(x,y)
v=f_2(x,y)
and
u=g_1(y,z)
v=g_2(y,z)

and eliminating u and v from these four equations, I have:

f_1(x,y)=g_1(y,z)
f_2(x,y)=g_2(y,z)

ie. I end up with two equations in the variables x,y and z instead of just one!

f_1(x,y)=g_1(y,z) \rightarrow F(x,y,z)=0
f_2(x,y)=g_2(y,z) \rightarrow G(x,y,z)=0

which define a curve, not a surface. Can you tell me where I am going wrong?
 
The two equations are not independent. For example, suppose your equations are x= u+v, y= u- v, z= 2u+ v (describing a plane). From the first two equations you get u= (x+y)/2 and then v= u- y= (x- y)/2. From the last two equations, u= (y+z)/3 and then v= u- y= (z- 2y)/3. Now combine those: u= (x+y)/2= (y+z)/3 give 3x+ 3y= 2y+ 2z or 3x+ y- 2z= 0. v= (x- y)/2= (z- 2y)/3 or 3x- 3y= 2z- 4y: 3x+ y- 2z= 0 again!
 
HallsofIvy said:
The two equations are not independent.

Ivy, I love you!

Okay but, I assume this will always be the case . . . but how can one see this if it's not through a particular example?
 
Okay wait, in the case of linear equations it's easy to show in general that the equations are dependent:

x=f(u,v)=Au+Bv
y=g(u,v)=Cu+Dv
z=h(u,v)=Eu+Fv

Applying Cramer's rule, from the frist two we get:

\begin{array}{cc}<br /> x=Au+Bv\\<br /> y=Cu+Dv<br /> \end{array} \right\} \rightarrow}<br />

u=\frac{ det \left(<br /> \begin{array}{cc}<br /> x &amp; B\\<br /> y &amp; D<br /> \end{array}<br /> \right)}{det \left(<br /> \begin{array}{cc}<br /> A &amp; B\\<br /> C &amp; D<br /> \end{array}<br /> \right)} = \frac{xD-yB}{AD-CB}; (1)

v=\frac{ det \left(<br /> \begin{array}{cc}<br /> A &amp; x\\<br /> C &amp; y<br /> \end{array}<br /> \right)}{det \left(<br /> \begin{array}{cc}<br /> A &amp; B\\<br /> C &amp; D<br /> \end{array}<br /> \right)} = \frac{yA-xC}{AD-CB}; (2)


and from the second two:

\begin{array}{cc}<br /> y=Cu+Dv\\<br /> z=Eu+Fv<br /> \end{array} \right\} \rightarrow}<br />

u=\frac{ det \left(<br /> \begin{array}{cc}<br /> y &amp; D\\<br /> z &amp; F<br /> \end{array}<br /> \right)}{det \left(<br /> \begin{array}{cc}<br /> C &amp; D\\<br /> E &amp; F<br /> \end{array}<br /> \right)} = \frac{yF-zD}{CF-ED}; (3)

v=\frac{ det \left(<br /> \begin{array}{cc}<br /> C &amp; y\\<br /> E &amp; z<br /> \end{array}<br /> \right)}{det \left(<br /> \begin{array}{cc}<br /> C &amp; D\\<br /> E &amp; F<br /> \end{array}<br /> \right)} = \frac{zC-yE}{CF-ED}; (4)

From (1) and (3) we obtain, after simplifying, the equation of the plane:

(CF-ED)Dx-(FA-BE)Dy-(AD-CB)Dz=0

and it's easy to see that from (2) and (4) we get the same equation of the plane.



But the same would happen if the equations were non-linear? And is there a way to prove it?
 
Thread 'Video on imaginary numbers and some queries'
Hi, I was watching the following video. I found some points confusing. Could you please help me to understand the gaps? Thanks, in advance! Question 1: Around 4:22, the video says the following. So for those mathematicians, negative numbers didn't exist. You could subtract, that is find the difference between two positive quantities, but you couldn't have a negative answer or negative coefficients. Mathematicians were so averse to negative numbers that there was no single quadratic...
Insights auto threads is broken atm, so I'm manually creating these for new Insight articles. In Dirac’s Principles of Quantum Mechanics published in 1930 he introduced a “convenient notation” he referred to as a “delta function” which he treated as a continuum analog to the discrete Kronecker delta. The Kronecker delta is simply the indexed components of the identity operator in matrix algebra Source: https://www.physicsforums.com/insights/what-exactly-is-diracs-delta-function/ by...
Thread 'Unit Circle Double Angle Derivations'
Here I made a terrible mistake of assuming this to be an equilateral triangle and set 2sinx=1 => x=pi/6. Although this did derive the double angle formulas it also led into a terrible mess trying to find all the combinations of sides. I must have been tired and just assumed 6x=180 and 2sinx=1. By that time, I was so mindset that I nearly scolded a person for even saying 90-x. I wonder if this is a case of biased observation that seeks to dis credit me like Jesus of Nazareth since in reality...
Back
Top