How Does the Lagrangian Transform in Polar Coordinates for Orbital Motion?

  • Thread starter Thread starter cepheid
  • Start date Start date
  • Tags Tags
    Field Motion
AI Thread Summary
The discussion focuses on the transformation of the Lagrangian in polar coordinates for orbital motion, specifically addressing the kinetic energy expression. The Lagrangian is initially presented as a function of the radial and angular velocities, leading to confusion about the validity of assuming constant radius. Participants clarify that the correct expression for kinetic energy in polar coordinates involves both radial and tangential components, derived from the time derivative of the position vector. The conversation also touches on the mathematical treatment of differentials and infinitesimals, debating their rigorous use in calculus versus intuitive approaches. Ultimately, the discussion highlights the importance of understanding vector relationships and the derivation of the Lagrangian in polar coordinates.
cepheid
Staff Emeritus
Science Advisor
Gold Member
Messages
5,197
Reaction score
38
In describing the motion (orbit) of a particle of mass m moving in a plane in a potential V(r), my prof wrote the Lagrangian as:

\mathcal{L} = \frac{1}{2}m|\dot{\vec{r}}|^2 - V(r)

Without further ado (or explanation of any kind), he wrote this as:

\mathcal{L} = \frac{1}{2}m(\dot{r}^2 + r^2\dot{\phi}^2) - V(r)

Where (r,\phi) are the coordinates in the orbital plane.

I understand that he is expressing the magnitude of r'(t) in terms of its components, and that \dot{r} is the velocity in the radial direction, whereas the second term should represent tangential velocity. But that part bothers me. After all...how did he arrive at this second term? It seems to me that he assumed r was constant, otherwise the second term is invalid. Yet if r were constant, then \dot{r} would be zero!

Shouldn't it be done as follows?

|\dot{\vec{r}}|^2 = \left(\frac{ds}{dt}\right)^2 = \left(\frac{d(r \phi)}{dt}\right)^2 = (\dot{r}\phi + r\dot{\phi})^2

What am I doing wrong, and how did he get his formula?
 
Physics news on Phys.org
cepheid said:
Shouldn't it be done as follows?

|\dot{\vec{r}}|^2 = \left(\frac{ds}{dt}\right)^2 = \left(\frac{d(r \phi)}{dt}\right)^2 = (\dot{r}\phi + r\dot{\phi})^2

What am I doing wrong, and how did he get his formula?

s is not equal to phi*r unless it is a circular motion (constant r). In polar coordinates,

ds=\sqrt{(dr)^2+(rd\phi)^2}. See attached picture. The yellow thing is a rectangular triangle when the displacement is very small.

ehild
 
Last edited:
He's computing .5*m*v^2 to get the kinetic energy for the first term.

In cartesian coordinates, it's .5*m*(vx^2 + vy^2).

To duplicate his result, you can either

1) note that the vectors \frac{d \hat r}{dt} and \frac {d \phi}{dt} are orthogonal, so you just add the squares of the two velocites, or

2) you can make the coordinate transformation

x = r(t)*cos(phi(t))
y = r(t)*sin(phi(t))

use the chain rule to compute dx/dt and dy/dt

square the results and add them together

but the first way is much easier. (The second is not too bad if you have a symbolic algebra program).
 
Hmm...I think I've got it. However, it is somewhat confusing when two people give slightly different answers. However, am I right in saying that pervect's method number 1 is the same as ehild's method? The only difference is that pervect assumed that I would be able to get the two expressions for radial and tangential velocity components just like that, whereas ehild showed me how to do it on a diagram. Another thing that bothers me: to get the desired answer from what ehild showed on the diagram, you have to do this:


ds=\sqrt{(dr)^2+(rd\phi)^2}

\frac{ds}{dt} = \frac{\sqrt{(dr)^2+(rd\phi)^2}}{dt}

= \sqrt{\frac{(dr)^2+(rd\phi)^2}{(dt)^2}}

= \sqrt{\frac{(dr)^2}{(dt)^2}+ \frac{(rd\phi)^2}{(dt)^2}}

= \sqrt{\left(\frac{dr}{dt}\right)^2 + r^2\left(\frac{d\phi}{dt}\right)^2}

= \sqrt{\dot{r}^2 + r^2\dot{\phi}^2}

So, we get the right answer, but I'm wondering about the mathematical validity of the manipulations I performed. I understand that there must be some validity to algebraic manipulations of differentials (i.e. infinitesimals), because it is done all the time in integral calculus. But it seems to me that physicists are a lot more free and easy with differntials than mathematicians would say is permissible.

My question is, are we allowed to interpret them the way we do, as actual infinitesimal quantities, that can be squared, grouped together etc. Can we say that the derivative is a ratio of infinitesimal changes in (i.e. "elements" of or "little bits" of) two variables? This is the method of presenting calculus in Silvanus P. Thompson's Calculus Made Easy, and it is totally intuitive to me. He even gives reference to a mathematician (I can't remember who) who wrote a paper asserting (with some proof) that infinitesimals are legitimate quantities in mathematics. However, I have seen people on this site freaking out about that. My calculus prof would also have none of it. "You could interpret the derivative (etc.) that way," he said, "but you would merely be restating the same thing a different way. A limit is a well defined concept, and the definition of a derivative is given in terms of one."

In fact, isn't the definition of a differential a limit as well? I haven't looked it up, but I think it goes something like, for a differentiable function y = f(x), on an interval \Delta x, we can make the following definition:

dy = \lim_{\substack{\Delta x \rightarrow 0}}f'(x) \Delta x

I guess I could have derived the equation above more rigorously with that in mind. Still, clarification on this very confusing issue would be appreciated.
 
This sort of thing is a real quagmire, because a great many number of concepts all get similar notation because they all "act like infintiessimals".


Infinitessimals can, in fact, be used rigorously, but you have to learn non-standard analysis. For example, the derivative is defined by:

<br /> \frac{d}{dx} f(x) := \mathrm{Std} \frac{f^*(x^* + \Delta x) - f^*(x^*)}{\Delta x}<br /> &lt;br /&gt; &lt;br /&gt; Where \Delta x is any nonzero &lt;i&gt;infinitessimal&lt;/i&gt;, Std means to round to the nearest standard number, and the superscript * means to use the nonstandard version of that symbol. (It&amp;#039;s sort of like distinguishing between 1 the integer and 1 the rational number)&lt;br /&gt; &lt;br /&gt; &lt;br /&gt; However, you&amp;#039;re probably better off translating your intuition to the standard way of doing things than trying to do nonstandard stuff intuitively.&lt;br /&gt; &lt;br /&gt; &lt;br /&gt; &lt;br /&gt; Differentials, like f(x) dx, are simply functions on paths. That is, given a differential, f(x) dx, you can apply it to a path, \gamma, to get a number: \int_{\gamma} f(x) \, dx.&lt;br /&gt; &lt;br /&gt; &lt;br /&gt; However, I don&amp;#039;t know if the &amp;quot;arithmetic&amp;quot; used in computing ds is an actual arithmetic, or just a formal symbol that means &amp;quot;do the right thing&amp;quot;.
 
cepheid said:
Hmm...I think I've got it. However, it is somewhat confusing when two people give slightly different answers. However, am I right in saying that pervect's method number 1 is the same as ehild's method? The only difference is that pervect assumed that I would be able to get the two expressions for radial and tangential velocity components just like that, whereas ehild showed me how to do it on a diagram.

Yes, it is. In a polar coordinate system the base vectors (angular and azimuthal) can be written in terms of Cartesian coordinates as

\vec e_{r} = \cos\phi \vec i + \sin \phi \vec j \mbox { and }\vec e_\phi= -\sin \phi \vec i + \cos \phi \vec j.

The position of a point in vector form is

\vec r = r \vec e_r

The velocity is the time derivative of the position vector.

\vec v = \dot \vec r = \dot r \vec e _r +r \dot \vec e_r

We have to find the time derivative of the radial unit vector.

First, as

\vec e \cdot \vec e = 1\rightarrow \vec e \cdot \dot \vec e =0

the time derivative of an unit vector is normal to the original vector, which means it is parallel to \vec e _\phi.

The magnitude of an unit vector \dot \vec e can be found from a geometrical approach. In the limit when

\Delta\phi \rightarrow 0

\frac{\Delta |\vec e|}{\Delta t}=\frac{\Delta \phi}{\Delta t} \rightarrow \dot \phi

If you do not like this method, you can derive the radial unit vector written in Cartesian components

\dot \vec e_{r} = -\sin\phi *\dot\phi *\vec i + \cos \phi *\dot\phi*\vec j =\dot \phi * \vec e_\phi

We get the velocity now in polar coordinates :

\vec v = \dot r \vec e _r +r\dot \phi \vec e_\phi.

Take the square :

(\vec v) ^2 = (\dot r )^2 + (r \dot \phi)^2

As for your problem with ds, how did you learn the line integral? How do you calculate the length of a line y=f(x) from A to B?

\int _A^B ds

You replace ds with

\sqrt{dx^2+dy^2}=\sqrt{ 1+(f&#039;(x))^2}dx,

don't you? This is the same method with infinitesimal quantities I used to get ds from dr and d(phi).

ehild
 
How did we get this step? Sorry, probably something really easy I'm missing.

ehild said:
\vec e \cdot \vec e = 1\rightarrow \vec e \cdot \dot \vec e =0
 
cepheid said:
How did we get this step? Sorry, probably something really easy I'm missing.

It is differentiation of a product.
If you do not like vectors,

( \vec e )^2= \vec e \cdot \vec e = = e_x^2+e_y^2=1
\frac{d(\vec e)^2}{dt}= 2 (e_x*\dot e_x+e_y*\dot e_y) = 0
Notice that
(e_x*\dot e_x+e_y*\dot e_y)=\vec e \cdot \dot \vec e

dot product of he unit vector and its derivative, and this is zero, which means that they are orthogonal.


ehild
 
Back
Top