Multiplying by dx to Solve ODEs

  • Thread starter flatmaster
  • Start date
  • Tags
    Dx
In summary: But even in that case, you would only be able to get the derivative if you knew the function's entire domain. And even then, the function might not have a derivative at all.
  • #1
flatmaster
501
2
When solving an ODE, physicsts sometimes multiply the differential dx.

Example

df/dx = x/2

df = (x/2)dx

f = x^2 + c

I've always been told that this mathematically unsound, but I've never been told why. Is it because the necessary constant c may not necessarily be additive? Are there any specific counter-examples where this doesn't work?
 
Physics news on Phys.org
  • #2
It works fine. The objection is that, in Leibnitz's notation, df/dx is not the quotient of something df by something dx. The calculation is conveniently done that way, and eaisiy checked at the end, so even if someone "questions" the method, you still get the right answer.
 
  • #3
flatmaster said:
When solving an ODE, physicsts sometimes multiply the differential dx.

Example

df/dx = x/2

df = (x/2)dx

f = x^2 + c

I've always been told that this mathematically unsound, but I've never been told why. Is it because the necessary constant c may not necessarily be additive? Are there any specific counter-examples where this doesn't work?



1. It is mathematically unsound because "dx" is not a number, so you can't "multiply" with it.

2. Doing so "works" in the very limited case of an ordinary first derivative, because it can then be coupled to the rigorous mathematical theory concerning differential forms.

It does NOT work, in general, for higher order derivatives, or partial derivatives, not even first-order ones.
 
  • #4
Wonderful, two opposite answers. I think I know who's the physicist and who's the mathematician.
 
  • #5
How about a third answer?

The differential (a.k.a. covector field) df is, in fact, the scalar product of the differential dx by the function df/dx.

But that fact has nothing to do with the related property of real numbers. (at least, not directly)

(okay, I lied -- arildno already said this)
 
  • #6
flatmaster said:
Wonderful, two opposite answers. I think I know who's the physicist and who's the mathematician.

How are the two "opposite"??

All g_edgar did was to pre-gurgitate my point 2.

(Is pregurgitate an English word, by the way?)
 
  • #7
Hey, English is flexible. If it wasn't, it is now!

My look at it:
[tex]\frac{dy}{dx}= \lim_{h\rightarrow 0}\frac{f(x+h)- f(x)}{h}[/tex]

Writing "dy/dx= f(x) so dy= f(x)dx" by "multiplying both sides by dx" is "mathematically unsound" because dy/dx is the limit of a fraction, not a fraction.

But you can use that equation: "dy= f(x)dx" (though not the justification "multiplying both sides by dx") because it is the limit of a fraction!

Basically, you can justify treating dy/dx as if it were a fraction (note the use of subjunctive mode?) by going back before the limit, using the fact that (f(x+h)- f(x))/h is a fraction, then taking the limit again. Take any Calculus text, and look at how the "chain rule" (which could be symbolized by
[tex]\frac{dy}{dt}= \frac{dy}{dx}\frac{dx}{dt}[/tex]
and note that they don't "cancel" the "dx"s.


In order to make use that, we define the "differentials": "dx" is an "undefined ("primitive") term and then we define "dy" by "dy= f'(x)dx". (I intentionally used the notation "y' " rather than "dy/dx" to avoid anyone thinking I was "cancelling"!)

Although used a lot in Calculus and Differential Equations, differentials really come into there own in Differential Geometry.
 
  • #8
"Writing "dy/dx= f(x) so dy= f(x)dx" by "multiplying both sides by dx" is "mathematically unsound" because dy/dx is the limit of a fraction, not a fraction."

This was, as it happens, also Newton's view:

He made the analogy that if we take the fraction 2x/x and let x pass into infinity, the expression still equals 2, but that nobody would say that in the limit, we had a fraction between two numbers.

Ps:
Thanks, Halls, I thought pregurgitate was a nice word myself! :smile:
 
  • #9
Yes, it was Newton's view, but he was using "infinitesmals" and it has taken us almost 500 years to define those properly! Anyway, in "non-standard" analysis, I would have had to tell the OP that his first question made no sense!
 
  • #10
Leibniz notation is unsound, but miraculously produces the correct answer most of the time anyway (at least for first-year calculus).

As it's been said before, the derivative and integral are defined by limits. You can't break the limit into two halves. In other words, it is blatantly false that [tex]\lim \frac{a}{b} = \frac{\lim a}{\lim b}[/tex] or variation like that. You simply can't split a limit. The "dx" can't exist on its own like that. It's not a real number. It's not an infinitesimal. It's nonsense.

But it's nonsense that produces the correct answer (in most cases). The reason for this probably has to do with a generalized notion of a limit. If you had a limit operator that acted on an entire equation (as opposed to an expression -- ie: on both sides of the = sign, not just on the left or right), then you can do a few more tricks. In that case, the definition of a derivative would be

[tex]\lim_{h->0} (f'(x) = \frac{f(x+h) - f(x)}{h})[/tex]

The parentheses are there to show that the limit applies to the whole equation. The important difference is that the variable h is allowed to live anywhere inside the equation (as opposed to just on the left or right side). That means, we can rearrange the equation like this:

[tex]\lim_{h->0} (f'(x) h = f(x+h) - f(x))[/tex]

If someone learned calculus with a concentration in physics, they will recognize this form immediately. The h plays the role of the dx. The limit suggests it plays the role of an infinitesimal. If you use the first equation, you can reduce the limit equation to a regular one, and in the process, you learn the derivative. In the second equation, (the one right above here), the equation reduces to 0 = 0, which (while useless) is still true.

So these bastard "dx"'s can be used to do simple calculus. If you can solve problems using this method, that's cool. But you should understand that you're really working with a more powerful framework. The dx's are NOT real numbers. They are NOT functions. They are just the shadows of more powerful tools.

These methods tend to break down when you get to more difficult problems. When there are more than one function or you are working with a function of more than one parameter, it would not be difficult to run into problems. Especially in multivariable calculus, when learning to do multiple integrals, surface, or line integrals, the equation is more of a summary of the problem than the whole problem.
 
  • #11
I know people have already tried to give an explanation, but I would recommend you to browser Thomas's Calculus book and look for section 3.8.

I am glad I looked it up, I had the same question for ages. Its no fun solving problems when you know you don't understand the first principles.
 
  • #12
In nonstandard analysis the derivative is a perfectly good fraction and things like df = (x/2)dx describe numbers and make sense.
 
  • #13
Phrak said:
In nonstandard analysis the derivative is a perfectly good fraction and things like df = (x/2)dx describe numbers and make sense.

Sure, but then again, neither division or multiplication in non-standard analysis are the same operations as within "standard" maths.
 
  • #14
Phrak said:
In nonstandard analysis the derivative is a perfectly good fraction and things like df = (x/2)dx describe numbers and make sense.

This is perhaps misleading. In nonstandard analysis, the quotient of infinitesimals is not equal to the derivative, it is just infinitely close to the derivative.
 
  • #15
dy = a dx is just a different (and, especially in physics, often more natural) way of writing dy/dx = a.
 
  • #16
Preno said:
dy = a dx is just a different (and, especially in physics, often more natural) way of writing dy/dx = a.

Nope, it is not.
 
  • #17
arildno said:
Sure, but then again, neither division or multiplication in non-standard analysis are the same operations as within "standard" maths.

I'm not sure what you mean. There are things that have no counterpart in real numbers, like

[tex]\frac{y}{dx} .[/tex]

Can you supply an example?
 
  • #18
Phrak said:
In nonstandard analysis the derivative is a perfectly good fraction and things like df = (x/2)dx describe numbers and make sense.

Not quite! In non-standard analysis the derivative of a function is not defined as the quotient of two infinitesimal numbers, but rather their "standard part." Hence, we arrive at the definition:

[tex]\frac{dy}{dx} = \mathrm{st}\left (\frac{f(x + \epsilon) - f(x)}{\epsilon} \right )[/tex]

Note: If [tex]f[/tex] is differentiable, then [tex]f(x + \epsilon) - f(x)[/tex] is infinitesimal if [tex]\epsilon[/tex] is infinitesimal. Also, if [tex]\Delta{y}[/tex] and [tex]\Delta{x}[/tex] are infinitesimal, then their quotient is infinitely close to the function's derivative.
 
  • #19
jgens said:
Not quite! In non-standard analysis the derivative of a function is not defined as the quotient of two infinitesimal numbers, but rather their "standard part." Hence, we arrive at the definition:

[tex]\frac{dy}{dx} = \mathrm{st}\left (\frac{f(x + \epsilon) - f(x)}{\epsilon} \right )[/tex]

Note: If [tex]f[/tex] is differentiable, then [tex]f(x + \epsilon) - f(x)[/tex] is infinitesimal if [tex]\epsilon[/tex] is infinitesimal. Also, if [tex]\Delta{y}[/tex] and [tex]\Delta{x}[/tex] are infinitesimal, then their quotient is infinitely close to the function's derivative.

To be honest, it was an intentionally provocative claim, and I do appreciate all the criticism, but it really depends on how one defines a fraction--a quotient of reals or a quotient that may include infinitesimals in quotient or denominator, doesn't it?
 
Last edited:
  • #20
it really depends on how one defines a fraction
The phrase "the derivative is a fraction" is most commonly to mean the assertion
[tex]f'(x) = \frac{f(x+\epsilon) - \f(x)}{\epsilon}[/itex]
for... well... people usually aren't clear on what they really think [itex]\epsilon[/itex] means here.

Of course, one can find all sorts of equations that assert the derivative is equal to a fraction. For example, if
[tex]f(x) = \frac{1}{1 + x^2}[/tex]​
then
[tex]f'(x) = \frac{-2x}{(1 + x^2)^2}[/tex]​
Voila! The derivative is a fraction!


Even classicaly, I might be inspired to define, for any univariate function f, a binary function (which I will boldly call df) defined by
[tex]d_\epsilon f(x) = \epsilon f'(x)[/tex]​
and whose domain is nonzero [itex]\epsilon[/itex] and those x for which f is differentiable

Then, we would have the equation
[tex]\frac{d_\epsilon f(x)}{d_\epsilon x} = f'(x)[/tex]​

but this is very different from what the phrase "the derivative is a fraction" is most commonly used to mean.


(Incidentally, the function df I write above is very closely related both to differential forms and to what one might find in a nonstandard analysis textbook such as Keisler's online book)
 
  • #21
I've read one section by Keisler's calc text, as you recommended. The sections relevant to nonstandard analysis seem to be confined to chapter 1, sections 1.4, 1.5 and 1.6. It's an elementary text, as claimed, so the chapter on vector calculus, unfortunately, doesn't address differential forms.
 
  • #22
There are plenty of online reference to differential forms but I haven't found any developed, rigorously upon infinitesimals.

Like, what are these things, really?

[tex]\frac{d}{d \lambda } = \frac{ dx^i }{ d \lambda } \frac{ \partial }{\partial x^i} [/tex]

[tex] \nabla = \frac{ \partial }{ \partial x^i } dx^i[/tex]
 
  • #23
The impression I got from a non-standard analyst is "I have infinitessimals -- why would I bother with such things?"

But I did sit down with him and worked out enough of the foundation that it seemed clear to me.

If you fix a "length scale" -- i.e. some nonzero infinitessimal number e, then you can define two relations on points of *Rn:
x ~ y if and only if d(x,y) is a finite multiple of e
x # y if and only if d(x,y) is an infinitessimal multiple of e
The relevant quality of differentiable functions is
if x~y, then f(x)~f(y)
if x#y, then f(x)#f(y)​
The tangent space to any standard point P is simply the set of all points such that P~x, modulo the equivalence relation #. (i.e. two points x and y name the same tangent vector if and only if x#y)

It's not difficult to show this is a vector space -- e.g. that the sum of two tangent vectors is the same (modulo #) no matter what differentiable coordinate chart you use to compute it.
 
  • #24
I'm sorry I can't follow that. What are the symbols # and ~ ?
 
  • #25
Phrak said:
I'm sorry I can't follow that. What are the symbols # and ~ ?
Binary relations whose definition is given in the first indented section of my post.
 
  • #26
You're way ahead of me. I'd need a short course in differential forms presented mathematically, rather than for relativity, to fully appreciate your post.
 

1. What does it mean to "multiply by dx" in solving ODEs?

Multiplying by dx in solving ODEs refers to the process of using the Euler method to approximate the solution of a differential equation. This involves breaking down the original equation into small intervals and using the derivative at each interval to find the value of the solution at the next interval.

2. How does multiplying by dx help to solve ODEs?

Multiplying by dx helps to solve ODEs by allowing us to approximate the solution of the differential equation at different points. This method is especially useful for nonlinear and complicated equations that cannot be solved analytically.

3. Are there any limitations to using the Euler method to solve ODEs?

Yes, there are some limitations to using the Euler method to solve ODEs. This method can produce inaccurate results for highly oscillatory or stiff equations. Additionally, the step size (dx) must be small enough to ensure accurate results, which can be computationally expensive.

4. What are some other methods for solving ODEs?

There are several other methods for solving ODEs, including the Runge-Kutta method, the Taylor series method, and the Adams-Bashforth method. These methods may offer better accuracy and efficiency compared to the Euler method, but they also have their own limitations.

5. How can I determine which method is best for solving a specific ODE?

The best method for solving a specific ODE depends on various factors, such as the type of equation, the desired level of accuracy, and the computational resources available. It is recommended to consult with a mathematician or use numerical analysis techniques to determine the most suitable method for a particular ODE.

Similar threads

Replies
3
Views
781
Replies
2
Views
2K
  • Differential Equations
Replies
1
Views
5K
  • Differential Equations
Replies
1
Views
2K
  • Differential Equations
2
Replies
52
Views
775
  • Differential Equations
Replies
5
Views
1K
Replies
3
Views
2K
Replies
3
Views
1K
Replies
8
Views
143
  • Differential Equations
Replies
3
Views
2K
Back
Top