What's the deal on infinitesimal operators?

Click For Summary
The discussion centers on the quest for a rigorous treatment of infinitesimal operators from an epsilon-delta perspective, particularly in the context of Lie groups and transformations. Participants express frustration that most literature on infinitesimals lacks a formal approach, often relying on abstract concepts rather than concrete calculus methods. There is a call for resources that explain infinitesimal transformations without invoking infinitesimals, focusing instead on limits and Taylor expansions. Some mention that existing texts, while rigorous in abstract terms, do not adequately address practical applications or the foundational calculus behind these concepts. The conversation highlights a gap in the literature for a mathematically rigorous yet accessible treatment of infinitesimal operators.
  • #31
Sorry, I don't think you're on topic. So please, make a new thread.
 
Physics news on Phys.org
  • #32
Stephen Tashi said:
Every book that takes a concrete approach to Lie Groups proves a result that says

f(x_1,y_1) = f(x,y) + U f\ \alpha + \frac{1}{2!}U^2 f \ \alpha^2 + \frac{1}{3!}U^3 f \ \alpha^3 + ..

by using Taylor series.

However, the function they are expanding is (to me) unclear.

If I try to expand f(x_1,y_1) = f(T_x(x,y,\alpha),T_y(x,y,\alpha)) in Taylor series, only the first two terms of that result work.

I would like to denote the series as

<br /> (e^{\alpha U}f)(x,y) = f(x,y) + \alpha (Uf)(x,y) + \frac{1}{2!}\alpha^2 (U^2f)(x,y) +\cdots<br />

Are we speaking about the same thing?

If I expand f(x_1,y_1) = f( x + \alpha \ u_x(x,y), y + \alpha \ u_y(x,y) ) then I get the desired result.

This looks like a mistake to me. The operator U^2 behaves like this:

<br /> U^2f = u_x\frac{\partial}{\partial x}(Uf) + u_y\frac{\partial}{\partial y}(Uf)<br />
<br /> = u_x\Big(\big(\frac{\partial u_x}{\partial x}\frac{\partial f}{\partial x}<br /> + u_x\frac{\partial^2 f}{\partial x^2}<br /> + \frac{\partial u_y}{\partial x}\frac{\partial f}{\partial y}<br /> + u_y\frac{\partial^2 f}{\partial x\partial y}\Big)<br />
<br /> + u_y\Big(\frac{\partial u_x}{\partial u_y}\frac{\partial f}{\partial x}<br /> + u_y\frac{\partial^2 f}{\partial x \partial y}<br /> + \frac{\partial u_y}{\partial y} \frac{\partial f}{\partial y}<br /> + u_y\frac{\partial^2 f}{\partial y^2}\Big)<br />

If you compute derivatives of f(x+\alpha u_x,y+\alpha u_y) with respect to \alpha, the partial derivatives of u_x and u_y will never appear.
 
  • #33
Stephen Tashi said:
and the property that allows the homomorphism you mentioned:

T_x(x,y,\alpha + \beta)= T_x(T_x(x,y,\alpha),\beta), T_y(T_y(x,y,\alpha),\beta)
T_y(x,y,\alpha + \beta)= T_x(T_x(x,y,\alpha),\beta), T_y(T_y(x,y,\alpha),\beta)

This could be a minor thing, but these formulas probably should have been

<br /> T_x(x,y,\alpha + \beta) = T_x(T_x(x,y,\alpha),T_y(x,y,\alpha),\beta)<br />
<br /> T_y(x,y,\alpha + \beta) = T_y(T_x(x,y,\alpha),T_y(x,y,\alpha),\beta)<br />

?
 
  • #34
I don't remember precise conditions for uniqueness of the solutions of some DEs, but anyway I think the seeked result comes when some uniqueness result is used.

We want to prove

<br /> (e^{\alpha U}f)(x,y) = f(T_x(x,y,\alpha),T_y(x,y,\alpha))<br />

The left side satisfies the DE

<br /> D_{\alpha} (e^{\alpha U}f)(x,y) = (U(e^{\alpha U}f)(x,y)<br />

so the proof should be reasonably close if we succeed in proving that the right side satisfies the same DE, that means

<br /> D_{\alpha} f(T_x(x,y,\alpha),T_y(x,y,\alpha)) = (Uf)(T_x(x,y,\alpha),T_y(x,y,\alpha))<br />

(update: I think I made a mistake here. A comment in #36.)

Direct computation gives

<br /> D_{\alpha} f(T_x(x,y,\alpha),T_y(x,y,\alpha))<br />
<br /> = (D_{\alpha}T_x(x,y,\alpha))\frac{\partial f}{\partial x}(T_x(x,y,\alpha),T_y(x,y,\alpha))<br />
<br /> + (D_{\alpha}T_y(x,y,\alpha))\frac{\partial f}{\partial y}(T_x(x,y,\alpha),T_y(x,y,\alpha)) = \cdots<br />

We assume that the transformation satisfies

jostpuur said:
<br /> T_x(x,y,\alpha + \beta) = T_x(T_x(x,y,\alpha),T_y(x,y,\alpha),\beta)<br />
<br /> T_y(x,y,\alpha + \beta) = T_y(T_x(x,y,\alpha),T_y(x,y,\alpha),\beta)<br />

Hence it also satisfies

<br /> D_{\alpha}T_x(x,y,\alpha) = \frac{\partial}{\partial\alpha}T_x(T_x(x,y,\alpha),T_y(x,y,\alpha),0)<br />
<br /> D_{\alpha}T_y(x,y,\alpha) = \frac{\partial}{\partial\alpha}T_y(T_x(x,y,\alpha),T_y(x,y,\alpha),0)<br />

So we get

<br /> \cdots<br /> = \frac{\partial}{\partial\alpha}T_x(T_x(x,y,\alpha),T_y(x,y,\alpha),0)<br /> \frac{\partial f}{\partial x}(T_x(x,y,\alpha),T_y(x,y,\alpha))<br />
<br /> + \frac{\partial}{\partial\alpha}T_y(T_x(x,y,\alpha),T_y(x,y,\alpha),0)<br /> \frac{\partial f}{\partial y}(T_x(x,y,\alpha),T_y(x,y,\alpha))<br />
<br /> = (Uf)(T_x(x,y,\alpha),T_y(x,y,\alpha))<br />
 
Last edited:
  • #35
After struggling with physicists' infinitesimal stuff myself during my studies, I eventually arrived at the enlightened conclusion, that the infinitesimal arguments can often be transformed into differential equations. I think this settles lot of problems, but as I pointed out in my off topic posts, there are also some infinitesimal arguments which I have been unable to transform into rigor form.

When I have attempted to discuss those problems, the physicists are usually unable to understand my complaints, since they have already been indoctrinated into believing anything that has the magic word "infinitesimal" as a part. Then usually the physicists attempt to convince me to abandon the differential equations or other rigor concepts, and accept the infinitesimal arguments for sake of making life easier.

So, to answer the question in the topic, I would say that the "situation with infinitesimal operators" is serious.
 
  • #36
Oh dear, it could be there is something wrong in my "answer".

I think I was supposed to prove

<br /> D_{\alpha} (f\circ T(\alpha))(x,y) = (U(f\circ T(\alpha)))(x,y)<br />

but the result that I proved was

<br /> D_{\alpha} (f\circ T(\alpha))(x,y) = ((Uf)\circ T(\alpha))(x,y)<br />

I'm feeling little confused now...
 
  • #37
If a function has been written in polar coordinates like f(r,\theta), then

<br /> e^{\theta\frac{\partial}{\partial \theta}} f(r,\theta_0)<br />

is the Taylor series with respect to the angle variable. Hence the operator should be seen as a rotation operator. On the other hand

<br /> \frac{\partial}{\partial\theta} = -y\frac{\partial}{\partial x} + x\frac{\partial}{\partial y}<br />

holds too, so we should be able to write the rotation operator in the Cartesian coordinates. Let's define

<br /> \phi(x,y,\theta) = e^{\theta(-y\frac{\partial}{\partial x} + x\frac{\partial}{\partial y})}f(x,y)<br />

and

<br /> \psi(x,y,\theta) = f(x\cos\theta - y\sin\theta,\; x\sin\theta + y\cos\theta)<br />

Now

<br /> \phi(x,y,\theta) = \psi(x,y,\theta)<br />

is supposed to hold, right?

<br /> \frac{\partial}{\partial\theta}\phi(x,y,\theta) = \Big(-y\frac{\partial}{\partial x}+ x\frac{\partial}{\partial y}\Big)\phi(x,y,\theta)<br />

<br /> \frac{\partial}{\partial\theta}\psi(x,y,\theta)<br /> = (-x\sin\theta - y\cos\theta)\frac{\partial f}{\partial x}<br /> + (x\cos\theta - y\sin\theta)\frac{\partial f}{\partial y}<br />
<br /> = -y\Big(\cos\theta\frac{\partial f}{\partial x} +\sin\theta \frac{\partial f}{\partial y}\Big)<br /> + x\Big(\cos\theta\frac{\partial f}{\partial y} - \sin\theta \frac{\partial f}{\partial x}\Big)<br />
<br /> = \Big(-y\frac{\partial}{\partial x} + x \frac{\partial}{\partial y}\Big)\psi(x,y,\theta)<br />

Here equations

<br /> \frac{\partial \psi}{\partial x} = \cos\theta \frac{\partial f}{\partial x} +\sin\theta \frac{\partial f}{\partial y}<br />
<br /> \frac{\partial \psi}{\partial y} = -\sin\theta \frac{\partial f}{\partial x} +\cos\theta\frac{\partial f}{\partial y}<br />

were used.

Ok so \phi and \psi satisfy the same PDE and therefore it should be possible to prove that they are the same thing rigorously.

I'm sure a similar manipulation should be possible in more general case, but I did something wrong in my previous attempt, and all pieces didn't fit together for some reason. It is still little mysterious, but I'm getting optimistic on this now.
 
  • #38
jostpuur said:
This looks like a mistake to me. The operator U^2 behaves like this:

<br /> U^2f = u_x\frac{\partial}{\partial x}(Uf) + u_y\frac{\partial}{\partial y}(Uf)<br />
<br /> = u_x\Big(\big(\frac{\partial u_x}{\partial x}\frac{\partial f}{\partial x}<br /> + u_x\frac{\partial^2 f}{\partial x^2}<br /> + \frac{\partial u_y}{\partial x}\frac{\partial f}{\partial y}<br /> + u_y\frac{\partial^2 f}{\partial x\partial y}\Big)<br />
<br /> + u_y\Big(\frac{\partial u_x}{\partial u_y}\frac{\partial f}{\partial x}<br /> + u_y\frac{\partial^2 f}{\partial x \partial y}<br /> + \frac{\partial u_y}{\partial y} \frac{\partial f}{\partial y}<br /> + u_y\frac{\partial^2 f}{\partial y^2}\Big)<br />


If you compute derivatives of f(x+\alpha u_x,y+\alpha u_y) with respect to \alpha, the partial derivatives of u_x and u_y will never appear.

I see your point. I way trying to interpret U^n f as an operation that is applied only to the original function f and not to the function that is the result of previous operations.
 
  • #39
Just in case somebody feels unsure about how to interpret the U^n, I would recommend studying the example

<br /> \frac{\partial}{\partial\theta} = -y\frac{\partial}{\partial x} + x\frac{\partial}{\partial y}<br />

I didn't check this now myself, but I'm sure that with some example, it can be verified that

<br /> \frac{\partial^2}{\partial\theta^2},\frac{\partial^3}{\partial\theta^3},\ldots<br />

will not turn out correctly, if it is assumed that \frac{\partial}{\partial x},\frac{\partial}{\partial y} would commute with x,y.

I thought that I recognized the Stephen Tashi's problem as something that I had thought through earlier, but now it became clear that I hadn't thought it through completely after all. This seems to be coming a good thread, although it also contains lot of mistakes now, and some from me. :blushing: I'll try to take a break now, and return later to see what's happened.
 
  • #40
With the correction and also adding the evaluation \alpha = 0 to the left hand side,

<br /> D_{\alpha} f(T_x(x,y,\alpha),T_y(x,y,\alpha))_{|\ \alpha = 0} = U (f(T_x(x,y,\alpha),T_y(x,y,\alpha)))<br />

is my interpretation of what old books on Lie groups prove and then they immediately assert
something that (I think) amounts to

D^n_{\alpha} f(T_x(x,y,\alpha),T_y(x,y,\alpha))_{|\ \alpha = 0} = U^n (f(T_x(x,y,\alpha),T_y(x,y,\alpha)))

as if the result is just straightforward calculus. (Perhaps with your correction to my interpretation of U^2, it is.)

In those books, the right hand side is denoted as U( f(x_1,y_1) ).

In previous pages, notation amounting to x_1 = T_x(x,y,\alpha),\ y_1 = T_y(x,y,\alpha) is employed, and in other places the same symbols are also defined implicityly using "infinitesimals" as x_1 + \delta x_1 = T_x(x,y,\alpha + \delta \alpha),\ y1 +\delta y_1 = T_y(x,y,\alpha + \delta \alpha). So I think the interpretation in the first equation is correct.

For matrix groups, lovina's proof works. (Perhaps even my wrong interpretation of U^2 works.)
 
  • #41
Many people react to questions about the expansion in U by brining up differential equations. Intuitively, I understand why. We trace out the curve followed by the point (T_x(x,y.\alpha),T_y(x,y,\alpha) and the value of function f(T_x,T_y) as it varies along that curve, thinking of \alpha as "time". The derivative of f(x,y) with respect to \alpha depends on the gradient of f and the directional derivative of the curve at (x,y).

Since the curve is a group of transformations, the directional derivative at (x,y) is determined by what the transformation does to the point (x,y) by values of \alpha close to zero.

I'd think that some advanced calculus book somewhere would treat expanding a function f(x,y) in Taylor series "along a curve". Then the problem for adapting this result to 1-parameter Lie groups would be to show how that result becomes simpler when the curve is generated by a tranformation group.
 
  • #42
One new notation:

<br /> T(\alpha)(x,y) = \big(T_x(x,y,\alpha),T_y(x,y,\alpha)\big)<br />

so that T(\alpha) is a function with two variables.

Stephen Tashi said:
D^n_{\alpha} f(T_x(x,y,\alpha),T_y(x,y,\alpha))_{|\ \alpha = 0} = U^n (f(T_x(x,y,\alpha),T_y(x,y,\alpha))) \quad\quad\quad (1)<br />

Expressions like this are dangerous, since there could be ambiguity about whether we mean

<br /> \big(U(f\circ T(\alpha)\big)(x,y)\quad\quad\quad (2)<br />

or

<br /> \big((Uf)\circ T(\alpha)\big)(x,y)\quad\quad\quad (3)<br />

The difference between these is that (2) involves weights

<br /> u_x(x,y),\quad u_y(x,y)<br />

while (3) involves weights

<br /> u_x\big(T_x(x,y,\alpha),T_y(x,y,\alpha)\big),\quad u_u\big(T_x(x,y,\alpha),T_y(x,y,\alpha)\big)<br />

I did some mistakes with these issues in my previous attempt to deal with the infinitesimal problem through PDEs.

Being fully strict with the notation, a f(x_0,y_0) is only a real number once x_0,y_0 have been fixed. An operation by U from left on a real number is not defined, hence U(f(x_0,y_0)) is not defined. I wouldn't get too strict merely for the sake of being strict, but critical ambiguity should be avoided.
 
Last edited:
  • #43
Stephen Tashi said:
Many people react to questions about the expansion in U by brining up differential equations. Intuitively, I understand why.

IMO the PDE approach is extremely critical. The reason is that sometimes the exponential series can diverge, but the transformation itself may still make sense. Even then, the transformation could be seen as generated by some PDE. The exponential operator could then be seen as a formal notation for an operator that generates the transformation.

For example, the translation operator is well defined even if Taylor series diverge at some points. Also, a transport equation (PDE) will generate the translation very well for differentiable functions whose Taylor series diverges at some points.

Stephen Tashi said:
Is there a treatment of "infinitesimal operators" that is rigorous from the epsilon-delta point of view?

In looking for material on the infinitesimal transformations of Lie groups, I find many things online about infinitesimal operators. Most seem to be by people who take the idea of infinitesimals seriously and I don't think they are talking about the rigorous approach to infinitesimals [a la Abraham Robinson and "nonstandard analysis".

I have never seen any such rigour material. My own replies were combination of knowledge on PDEs, groups, and some own thinking.

Everytime I have read about Lie groups, it has been about actions on some finite dimensional manifolds. Now these differential operators are acting on infinite dimensional vector spaces. Are these infinite dimensional Lie groups? I guess not. The groups themselves are still finite dimensional?

Returning to the opening post... it would be interesting to learn about some pedagogical material on this. It seems replies containing links to such did not appear yet.
 
Last edited:
  • #44
By comparing my posts #34 and #37, it is clear that #34 contained some fatal mistakes. I would like to outline the goal more clearly now.

We define two functions \phi,\psi by formulas

<br /> \phi(x,y,\alpha) = (e^{\alpha U}f)(x,y)<br />

<br /> \psi(x,y,\alpha) = f\big(T_x(x,y,\alpha),T_y(x,y,\alpha)\big)<br />

Now \phi(x,y,0)=\psi(x,y,0) holds obviously, and

<br /> \frac{\partial\phi(x,y,\alpha)}{\partial\alpha} = U\phi(x,y,\alpha)<br />

holds almost by the definition. The goal should be to prove that \psi satisfies the same PDE, which would then imply \phi=\psi.

<br /> \frac{\partial\psi(x,y,\alpha)}{\partial\alpha}<br /> = \frac{\partial T_x(x,y,\alpha)}{\partial\alpha} \frac{\partial f}{\partial x}<br /> + \frac{\partial T_y(x,y,\alpha)}{\partial\alpha} \frac{\partial f}{\partial y}<br />

<br /> \frac{\partial\psi}{\partial x} =\frac{\partial f}{\partial x}\frac{\partial T_x}{\partial x}<br /> + \frac{\partial f}{\partial y}\frac{\partial T_y}{\partial x}<br />

<br /> \frac{\partial\psi}{\partial y} =\frac{\partial f}{\partial x}\frac{\partial T_x}{\partial y}<br /> + \frac{\partial f}{\partial y}\frac{\partial T_y}{\partial y}<br />

In the end we want U\psi to appear, so we need to get the \frac{\partial\psi}{\partial\alpha} written in such form that \frac{\partial\psi}{\partial x} and \frac{\partial\psi}{\partial y} would be present, and not the derivatives of f.

It's not immediately obvious how to accomplish that. The properties of the T_x,T_y shoul be studied in more detail.
 
  • #45
jostpuur said:
One new notation:

<br /> T(\alpha)(x,y) = \big(T_x(x,y,\alpha),T_y(x,y,\alpha)\big)<br />

so that T(\alpha) is a function with two variables.



Expressions like this are dangerous, since there could be ambiguity about whether we mean

<br /> \big(U(f\circ T(\alpha)\big)(x,y)\quad\quad\quad (2)<br />

or

<br /> \big((Uf)\circ T(\alpha)\big)(x,y)\quad\quad\quad (3)<br />

What about
<br /> U( f\circ (T(\alpha)(x,y) )) \quad\quad\quad (4)<br />

which is how I'd translate

D_{\alpha} f(T_x(x,y,\alpha),T_y(x,y,\alpha))_{|\ \alpha = 0} = U (f(T_x(x,y,\alpha),T_y(x,y,\alpha)))

into your notation.

Of course, I'm just guessing at what the old books mean. The result I'm trying to translate into modern calculus is shown in An Introduction to the Lie Theory of one-parameter groups" by Abraham Cohen (1911) http://archive.org/details/introlietheory00coherich page 30 of the PDF, page 14 of the book.

The expansion is also asserted in the modern book Solution of Ordinary Differential Equations by Continuous Groups by George Emmanuel (2001) page 13,
 
  • #46
Continuing the work from post #44

When you don't know how to arrive at the goal, the basic trick is to see what happens if you work backwards from the goal. It turns out this:

<br /> U\psi(x,y,\alpha) = \frac{\partial T_x(x,y,0)}{\partial\alpha}<br /> \frac{\partial\psi(x,y,\alpha)}{\partial x}<br /> + \frac{\partial T_y(x,y,0)}{\partial\alpha}\frac{\partial\psi(x,y,\alpha)}{\partial y}<br />
<br /> =\Big(\frac{\partial T_x(x,y,0)}{\partial\alpha}\frac{\partial T_x(x,y,\alpha)}{\partial x}<br /> + \frac{\partial T_y(x,y,0)}{\partial\alpha}\frac{\partial T_x(x,y,\alpha)}{\partial y}\Big)<br /> \frac{\partial f(T_x(x,y,\alpha),T_y(x,y,\alpha))}{\partial x}<br />
<br /> +\Big(\frac{\partial T_x(x,y,0)}{\partial\alpha}\frac{\partial T_y(x,y,\alpha)}{\partial x}<br /> +\frac{\partial T_y(x,y,0)}{\partial\alpha}\frac{\partial T_y(x,y,\alpha)}{\partial y}\Big)<br /> \frac{\partial f(T_x(x,y,\alpha),T_y(x,y,\alpha))}{\partial y}<br />

Comparing this with equation of post #44 we see that \psi will satisfy the desired PDE if the equations

<br /> \frac{\partial T_x(x,y,\alpha)}{\partial\alpha}<br /> = \frac{\partial T_x(x,y,0)}{\partial\alpha}\frac{\partial T_x(x,y,\alpha)}{\partial x}<br /> + \frac{\partial T_y(x,y,0)}{\partial\alpha}\frac{\partial T_x(x,y,\alpha)}{\partial y}<br />
<br /> \frac{\partial T_y(x,y,\alpha)}{\partial\alpha}<br /> = \frac{\partial T_x(x,y,0)}{\partial\alpha}\frac{\partial T_y(x,y,\alpha)}{\partial x}<br /> +\frac{\partial T_y(x,y,0)}{\partial\alpha}\frac{\partial T_y(x,y,\alpha)}{\partial y}<br />

hold.

Do these follow from some axioms about the transformation?
 
  • #47
Stephen Tashi said:
What about
<br /> U( f\circ (T(\alpha)(x,y) )) \quad\quad\quad (4)<br />

This contains two mistakes. You probably meant

<br /> U((f\circ T(\alpha))(x,y))<br />

?

The problem with this is that (f\circ T(\alpha))(x,y) is a real number, and we cannot operate with U from left on real numbers. The equation (4) also halts because T(\alpha)(x,y) is a point on plane, and f\circ (x&#039;,y&#039;) is not defined for any point (x&#039;,y&#039;), but only f(x&#039;,y&#039;) is.

Notice that the difference between equations (2) and (3) was that the weights u_x,u_y were evaluated at different points. There are only two obvious ways to evaluate u_x,u_y, so the ambiguous expressions will consequently have at most two obvious interpretations.
 
  • #48
jostpuur said:
This contains two mistakes. You probably meant

<br /> U((f\circ T(\alpha))(x,y))<br />

I won't say I mean that since I don't know what it means yet.

Is the convention you're using that any function G mapping 3 variables x,y,\alpha) to a point on the plane (G_x(x,y,\alpha), G_y(x,y,\alpha) ) will be denoted as G(\alpha) (x,y)? Or is this honor reserved for T(\alpha)(x,y) alone ?

I don't see f \circ T as a function that maps 3 variables to a point on the plane and I don't mean that.


The problem with this is that (f\circ T(\alpha))(x,y) is a real number, and we cannot operate with U from left on real numbers.

I can't interpret (f\circ T(\alpha))(x,y). The two arguments (x,y) sitting outside the parentheses confuse me.

Returning to the notation T(x,y,\alpha) = (T_x(x,y,\alpha), T_y(x,y,\alpha))
I consider f\circ T(x,y,\alpha) as a real valued function of 2 variables where each of those variables is replaced by a real valued function of 3 variables. .So f is a only a specific real number when x, y, \alpha are specific real numbers.


The equation (4) also halts because T(\alpha)(x,y) is a point on plane, and f\circ (x&#039;,y&#039;) is not defined for any point (x&#039;,y&#039;), but only f(x&#039;,y&#039;) is.

I don't understand what the primes mean, but I think that argument is a further objection to something that I don't mean anyway.


Notice that the difference between equations (2) and (3) was that the weights u_x,u_y were evaluated at different points. There are only two obvious ways to evaluate u_x,u_y, so the ambiguous expressions will consequently have at most two obvious interpretations.

I heartily agree that where the weights are to be evaluated needs precise notation. Denoting a real valued function g as g(A,B) to avoid the confusing proliferation of x's and y's, I think that U is the operator

U g(A,B) = u_x(A,B)\frac{\partial}{\partial A} + u_y(A,B) \frac{\partial}{\partial B}.

So in the case that A = T_x(x,y,\alpha),\ B = T_x(x,y,\alpha) those are the
values where the derivatives and weights are evaluated. An \alpha is thus introduced into this evaluation. When we set \alpha = 0 to get the Taylor coefficient we have T_x(x,y,0) = x, \ T_y(x,y,0) = y so the evaluation in the Taylor coefficient is done at (A,B) = (x,y).

Of course, I'm not yet certain how the old books are defining U.
 
  • #49
jostpuur said:
Comparing this with equation of post #44 we see that \psi will satisfy the desired PDE if the equations

<br /> \frac{\partial T_x(x,y,\alpha)}{\partial\alpha}<br /> = \frac{\partial T_x(x,y,0)}{\partial\alpha}\frac{\partial T_x(x,y,\alpha)}{\partial x}<br /> + \frac{\partial T_y(x,y,0)}{\partial\alpha}\frac{\partial T_x(x,y,\alpha)}{\partial y}<br />
<br /> \frac{\partial T_y(x,y,\alpha)}{\partial\alpha}<br /> = \frac{\partial T_x(x,y,0)}{\partial\alpha}\frac{\partial T_y(x,y,\alpha)}{\partial x}<br /> +\frac{\partial T_y(x,y,0)}{\partial\alpha}\frac{\partial T_y(x,y,\alpha)}{\partial y}<br />

hold.

Do these follow from some axioms about the transformation?

I think they do.

I don't follow the entire argument yet, but I at least understand the above question.

We assume the coordinate functions obey the composition law:

T_x( T_x(x,y,\beta), T_y(x,y,\beta),\alpha) = T_x( x,y,\alpha + \beta) [1]
T_y( T_x(x,y,\beta), T_y(x,y,\beta),\alpha) = T_y( x,y,\alpha + \beta) [2]

I copy the technique that lovinia showed for matrix groups.

Consider T_x to be T_x(A,B,C) to avoid confusion. Differentiate both sides of [1] with respect to \beta :

[3]:
\frac{\partial T_x}{\partial A} \frac{\partial Tx}{\partial C} + \frac{\partial T_x}{\partial B} \frac{\partial T_y}{\partial C} = \frac{\partial T_x}{\partial C}

The evaluations of these functions are:

Left hand side:

\frac{\partial T_x}{\partial A} at A = T_x(x,y,\beta),\ B = T_y(x,y,\beta),\ C = \alpha
\frac{\partial T_x}{\partial B} at A = T_x(x,y,\beta),\ B = T_y(x,y,\beta),\ C = \alpha
\frac{\partial T_x}{\partial C} at A = x,\ B = y ,\ C = \beta)
\frac{\partial T_y}{\partial C} at A = x,\ B = y ,\ C = \beta)
Right hand side:
\frac{\partial T_x}{\partial C} at A = T_x(x,y,\beta+\alpha),\ B = T_y(x,y,\beta+\alpha),\ C = \alpha + \beta




Set \beta = 0 in [3]. This sets T_x(x,y,\beta)= x,\ T_y(x,y,\beta) = y so we have the same functions with the evaluations:

\frac{\partial T_x}{\partial A} at A = x,\ B = y ,\ C = \alpha
\frac{\partial T_x}{\partial B} at A = x,\ B = y, \ C = \alpha
\frac{\partial T_x}{\partial C} at A = x,\ B = y,\ C = 0
\frac{\partial T_y}{\partial C} at A = x,\ B = y,\ C = 0

Right hand side:
\frac{\partial T_x}{\partial C} at A =x,\ B = y,\ C = \alpha

I interpret the above right hand side as the same thing as \frac{\partial T_x}{\partial \alpha}

Using your notation ( \frac{\partial T_x}{\partial x} instead of \frac{\partial T_x}{\partial A} etc.), I interpret the above left hand side to be the same thing as

\frac{\partial T_x}{\partial x}(x,y,\alpha) \frac{\partial T_x}{\partial \alpha}(x,y,0) + \frac{\partial T_x}{\partial y}(x,y,\alpha) \frac{\partial T_y}{\partial \alpha}(x,y,0)

so I get the equation you needed for the derivatives of T_x.
 
  • #50
Stephen Tashi said:
We assume the coordinate functions obey the composition law:

T_x( T_x(x,y,\beta), T_y(x,y,\beta),\alpha) = T_x( x,y,\alpha + \beta) [1]
T_y( T_x(x,y,\beta), T_y(x,y,\beta),\alpha) = T_y( x,y,\alpha + \beta) [2]

I copy the technique that lovinia showed for matrix groups.

Consider T_x to be T_x(A,B,C) to avoid confusion. Differentiate both sides of [1] with respect to \beta :

I had only considered derivatives with respect to \alpha from this equation. The derivative with respect to \beta was the trick.

I don't see much problems with this anymore. IMO this is settled now reasonably.

The two dimensional plane was chosen for simplicity of study in the beginning. It could be interesting to check the same thing in higher dimension, and get some guarantees for the uniqueness of solutions of PDEs.
 
  • #52
Sorry for such a late response, here we go:

Stephen Tashi said:
An example that I've given in another thread (on using Lie Groups to solve differential equations) is the treatment of the "infinitesimal operator" of a 1-parameter Lie group.

Avoiding the traditional zoo of Greek letters, I prefer the terminology:

Let T(x.y,\alpha) = ( \ T_x(x,y,\alpha), T_y(x,y,\alpha )\ ) denote an element of a Lie group of 1-parameter transformations of the xy-plane onto itself.

Let f(x,y) be a real valued function whose domain is the xy-plane.

Let u_x(x,y) = D_\alpha \ T_x(x,y,\alpha)_{\ |\alpha=0}
Let u_y(x,y) = D_\alpha \ T_y(x,y,\alpha)_{\ |\alpha =0}

(u_x and u_y are the " infinitesimal elements ".)

Let U be the differential operator defined by the operation on the function g(x,y) by:

U g(x,y) = u_x(x,y) \frac{\partial}{\partial x} g(x,y) + u_y(x,y)\frac{\partial}{\partial y} g(x,y)

(The operator U is "the symbol of the infinitesimal transformation" .)

I think the notation is the source of the problem here, it gets a bit confusing not specifying which point your starting from etc... I'll just derive it the way I would do it, substituting your notation into what I've done at one or two points so we can compare the results.

If I set x_0 = \phi(x_0,y_0,\alpha_0), \ y_0 = \psi(x_0,y_0,\alpha_0) & x_1 = \phi(x_0,y_0,\alpha_0 + \delta \alpha), \ y_1 = \psi(x_0,y_0,\alpha_0 + \delta \alpha) then

\Delta x = x_1 - x_0 = \phi(x_0,y_0,\alpha_0 + \delta \alpha) - \phi(x_0,y_0,\alpha_0) = \delta x + \varepsilon_{error}(\delta x) \approx \delta x = \frac{\partial \phi}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0} \delta \alpha

Similarly \delta y = \frac{\partial \psi}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0} \delta \alpha

Therefore (x_1,y_1) = T(x_0,y_0,\alpha) = (\phi(x_0,y_0,\alpha),\psi(x_0,y_0,\alpha) w/ T(x_0,y_0,\alpha_0) = (x_0,y_0) &

(x_1,y_1) = T(x_0,y_0,\alpha) = (\phi(x_0,y_0,\alpha),\psi(x_0,y_0,\alpha)) \\<br /> \ \ \ \ \ \ \ \ \ \ \ \ \ = (\phi(x_0,y_0,\alpha_0) + \frac{\partial \phi}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0} \delta \alpha,\psi(x_0,y_0,\alpha_0) + \frac{\partial \psi}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0}\delta \alpha)

Using your notation I have

T_x(x_0,y_0,\alpha)= \phi(x_0,y_0,\alpha) = \phi(x_0,y_0,\alpha_0) + \frac{\partial \phi}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0} \delta \alpha \\<br /> \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ = T_x(x_0,y_0,\alpha_0) + \frac{\partial T_x}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0}\delta \alpha = T_x(x_0,y_0,\alpha_0) + u_x(x_0,y_0)\delta \alpha

Now we want to deal with f(x_1,y_1) = f(\phi(x_0,y_0,\alpha),\psi(x_0,y_0,\alpha)).

Since the Taylor series expansion of f is

f(x_1,y_1) = f(x_0,y_0) + \frac{\partial f}{\partial x}(x_0,y_0)\delta x + \frac{\partial f}{\partial y}(x_0,y_0)\delta y + (\delta x \frac{\partial}{\partial x} + \delta y \frac{\partial}{\partial y})^2f|_{(x,y) = (x_0,y_0)} + ...

written more succinctly as

f(x_1,y_1) = \sum_{i=0}^\infty \frac{1}{n!}(\delta x \frac{\partial}{\partial x} + \delta y \frac{\partial}{\partial y})^nf|_{(x,y) = (x_0,y_0)}

we see that simply by substituting in our \delta y = \frac{\partial \psi}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0} \delta \alpha terms we get:

f(x_1,y_1) = \sum_{i=0}^\infty \frac{1}{n!}(\delta x \frac{\partial}{\partial x} + \delta y \frac{\partial}{\partial y})^nf|_{(x,y) = (x_0,y_0)} \\<br /> <br /> \ \ \ \ \ \ \ \ \ \ \ \ \ \ = \sum_{i=0}^\infty \frac{1}{n!} (\frac{\partial \phi}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0} \delta \alpha \frac{\partial}{\partial x} + \frac{\partial \psi}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0} \delta \alpha \frac{\partial}{\partial y})^nf|_{(x,y) = (x_0,y_0)} \\<br /> <br /> \ \ \ \ \ \ \ \ \ \ \ \ \ \ = \sum_{i=0}^\infty \frac{(\delta \alpha)^n}{n!}(\frac{\partial \phi}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0} \frac{\partial}{\partial x} + \frac{\partial \psi}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0} \frac{\partial}{\partial y})^nf|_{(x,y) = (x_0,y_0)} \\<br /> <br /> \ \ \ \ \ \ \ \ \ \ \ \ \ \ = \sum_{i=0}^\infty \frac{(\delta \alpha)^n}{n!}(\varepsilon(x_0,y_0) \frac{\partial}{\partial x} + \eta(x_0,y_0) \frac{\partial}{\partial y})^nf|_{(x,y) = (x_0,y_0)} \\

Stephen Tashi said:
Every book that takes a concrete approach to Lie Groups proves a result that says

f(x_1,y_1) = f(x,y) + U f\ \alpha + \frac{1}{2!}U^2 f \ \alpha^2 + \frac{1}{3!}U^3 f \ \alpha^3 + ..

by using Taylor series.

However, the function they are expanding is (to me) unclear.

If I try to expand f(x_1,y_1) = f(T_x(x,y,\alpha),T_y(x,y,\alpha)) in Taylor series, only the first two terms of that result work.

Just substitute T_x(x_0,y_0,\alpha)= \phi(x_0,y_0,\alpha) into what I derived above:

\sum_{i=0}^\infty \frac{(\delta \alpha)^n}{n!}(\frac{\partial \phi}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0} \frac{\partial}{\partial x} + \frac{\partial \psi}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0} \frac{\partial}{\partial y})^nf|_{(x,y) = (x_0,y_0)} \\<br /> <br /> \sum_{i=0}^\infty \frac{(\delta \alpha)^n}{n!}(\frac{\partial T_x}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0} \frac{\partial}{\partial x} + \frac{\partial T_y}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0} \frac{\partial}{\partial y})^nf|_{(x,y) = (x_0,y_0)}\\<br /> <br /> \sum_{i=0}^\infty \frac{(\delta \alpha)^n}{n!}(u_x \frac{\partial}{\partial x} + u_y \frac{\partial}{\partial y})^nf|_{(x,y) = (x_0,y_0)}

(I don't like the u_x, \varepsilon or \eta terms, a) because the subscript looks like a partial derivative & b) because defining u_x or \eta etc... makes me forget what we're doing, I think including the partial derivatives explicitly is least confusing, though that may be due to inexperience)

Stephen Tashi said:
If I expand f(x_1,y_1) = f( x + \alpha \ u_x(x,y), y + \alpha \ u_y(x,y) ) then I get the desired result. So I think this is equivalent to what the books do because they do not give an elaborate proof of the result. They present it as being "just calculus" and expanding f(x_1,y_1) = f( x + \alpha \ u_x(x,y), y + \alpha \ u_y(x,y) ) is indeed just calculus.

This is exactly equivalent to what I wrote above, only instead of \alpha u_x(x,y) I used \frac{\partial \phi}{\partial \alpha}\delta \alpha, the fact we can write both things basically amounts to whether we use \alpha - 0 or \alpha - \alpha_0 in our notation (where \alpha_0 & \alpha = 0 do the same thing!). If there's any ambiguity here just let me know.

Stephen Tashi said:
The books then proceed to give examples where the above result is applied to expand f(x_1,y_1) = f(T_x(x,y,\alpha),T_y(x,y,\alpha)) I haven't found any source that justifies this expansion except by using the concept of infinitesimals.

Hopefully what I wrote is enough, it all relies on nothing more than Taylor series thus you could just modify a more rigorous proof of Taylor's theorem I'd imagine. This is all a mix of what's in Cohen & Emanuel so hopefully it's alright!

Stephen Tashi said:
Many people react to questions about the expansion in U by brining up differential equations. Intuitively, I understand why. We trace out the curve followed by the point (T_x(x,y.\alpha),T_y(x,y,\alpha) and the value of function f(T_x,T_y) as it varies along that curve, thinking of \alpha as "time". The derivative of f(x,y) with respect to \alpha depends on the gradient of f and the directional derivative of the curve at (x,y).

Since the curve is a group of transformations, the directional derivative at (x,y) is determined by what the transformation does to the point (x,y) by values of \alpha close to zero.

I'd think that some advanced calculus book somewhere would treat expanding a function f(x,y) in Taylor series "along a curve". Then the problem for adapting this result to 1-parameter Lie groups would be to show how that result becomes simpler when the curve is generated by a tranformation group.

The proof of Taylor's theorem in two variables follows exactly the same approach we're following, almost word for word as far as I can see. You could substitute the t in their proof with \alpha for us & get the same result, just modifying a tiny bit of notation, thus you are indeed incorporating the idea of a directional derivative in a sense. Their proof just assumes you've got a line in your domain between the two points you're deriving the taylor series with, the transformation group idea is analogous to saying we can move our second point along any point on this line, there isn't really any difference between them as far as I can see (on an intuitive level, modulo the formalism of establishing what transformation groups are to allow for curved paths between those points etc...). Furthermore references to differential equations & matrix exponentials are implicitly encoding our basic ideas as far as I can see. Doing it any other way is merely a formalistic way of justifying our intuition in the context of a formalism we've built up to get the answer we already know we're looking for.
 
Last edited:
  • #53
bolbteppa said:
f(x_1,y_1) = \sum_{i=0}^\infty \frac{1}{n!}(\delta x \frac{\partial}{\partial x} + \delta y \frac{\partial}{\partial y})^nf|_{(x,y) = (x_0,y_0)}

Am I understanding correctly, that you have written f in (x_1,y_1) as a Taylor series with respect to the point (x_0,y_0), under the assumption that these points are infinitesimally close to each other? Doesn't look very useful.

Stephen Tashi said:
If I expand f(x_1,y_1) = f( x + \alpha \ u_x(x,y), y + \alpha \ u_y(x,y) ) then I get the desired result.
This is exactly equivalent to what I wrote above, only instead of \alpha u_x(x,y) I used \frac{\partial \phi}{\partial \alpha}\delta \alpha,...

What Tashi wrote was incorrect, so this isn't looking good. Explanation here #32
 
  • #54
bolbteppa said:
We see that simply by substituting in our \delta y = \frac{\partial \psi}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0} \delta \alpha terms we get:

How is such a substitution justifed by standard calculus? You're putting an approximation into an infinite series of terms.

It looks to me like you are reproducing what the old books do using an argument based on infinitesimals. You do supply more details than the old books. However, if we use standard calculus, we obtain the 3rd term of the Taylor series that I got in post #24 of https://www.physicsforums.com/showthread.php?t=697036&page=2. I think I can get the third term to work out without any infinitesimal reasoning. I'll try to accomplish that in my next post in that thread.

I think the old books "have something going for them" and the way they reason would be useful to know. However, I want it translated to (or at least clearly related to) the modern viewpoint of calculus.
 
  • #55
jostpuur said:
What Tashi wrote was incorrect, so this isn't looking good. Explanation here #32

My goal to expand that function was my intended goal. My method was incorrect. It would be interesting to see if the correct interpretation of U^n makes any difference in the final result for the simple arguments that I used in f.
 
  • #56
jostpuur said:
Am I understanding correctly, that you have written f in (x_1,y_1) as a Taylor series with respect to the point (x_0,y_0), under the assumption that these points are infinitesimally close to each other? Doesn't look very useful.

Our goal in the other thread referred to is to use Lie groups to solve ordinary differential equations. There is a theorem, which I hope we'll prove, is that the transformations T of the form T(x_0,y_0,\alpha) = (\phi(x_0,y_0,\alpha),\psi(x_0,y_0,\alpha)) of a one-parameter group can be put in one-one correspondence with transformations of a group generated by the infinitesimal generator

Uf \ = \ \varepsilon(x_0,y_0) \frac{\partial f}{\partial x} + \eta(x_0,y_0)\frac{\partial f}{\partial y} = \frac{\partial \phi}{\partial \alpha}(x_0,y_0,\alpha)|_{\alpha = \alpha_0}\frac{\partial f}{\partial x} + \frac{\partial \psi}{\partial \alpha}|_{\alpha = \alpha_0}\frac{\partial f}{\partial y}

The idea is that those coefficients are basically the coefficients of the tangent vector to the curve (thus need to be evaluated at a point!) & we'll use this local idea to generate our entire (global) curve & also solve some differential equations... What you would say doesn't look very useful is something someone like Lie would say has the potential to revolutionize an entire subject :wink:

jostpuur said:
What Tashi wrote was incorrect, so this isn't looking good. Explanation here #32

First of all you actually made the mistake when you wrote \frac{\partial u_x}{\partial x}, the u_x should be written as u_x = u_x(x_0,y_0,\alpha), in the context of what we're doing it is only a function of \alpha being evaluated at x = x_0, y = y_0, thus the last sentence of your post is also a misinterpretation of what we're actually doing as those partials are never meant to show up. Second if what Tashi wrote was incorrect, & I said what I'd written was equivalent to what he'd written, then it should have been easy to spot the flaw in what I'd written also. The reason you're allowed to interpret it as an operator acting on f is because you end up evaluating the derivatives at x = x_0, y = y_0 thus it is nothing but a shortcut to an assumption you're making in the derivation of the Taylor series, in the Taylor series you have to evaluate all partials at the point (x_0,y_0). Thus what he wrote is fine, & this is the method used in those old scary books where they went out of their way to prove the theorem I referred to above. What you've basically done is to take Taylor's theorem:

f(x_0 + \delta x,y + \delta y) = f(x_0,y_0) + \frac{\partial f}{\partial x}|_{(x_0,y_0)} \delta x + \frac{\partial f}{\partial y}|_{(x_0,y_0)} \delta y + \frac{1}{2!}(\frac{\partial f^2}{\partial x^2}|_{(x_0,y_0)} \delta x^2 + 2\frac{\partial ^2 f}{\partial x \partial y}\delta x \delta y + \frac{\partial ^2 f}{\partial y^2}|_{(x_0,y_0)} \delta y ^2) + ... \\<br /> \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ = \sum_{n = 0}^\infty \frac{1}{n!}(\delta x \frac{\partial}{\partial x}|_{(x_0,y_0)} + \delta y \frac{\partial}{\partial y}|_{(x_0,y_0)})^n f \\<br /> \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ = \sum_{n = 0}^\infty \frac{1}{n!}U^n f<br />

which I'm sure you know how to use & then, because of sloppy notation said that, for instance, in the second term above:

U^2 f = U(\delta x \frac{\partial f}{\partial x} + \delta y \frac{\partial f}{\partial y})\\<br /> \ \ \ \ \ \ \ = (U \delta x) \frac{\partial f}{\partial x} + \delta x (U \frac{\partial f}{\partial x}) + ...

which is obviously wrong here, this does not equal the Taylor series I wrote above when you expand it out (if you can make sense of it), yet that's exactly what you did, albeit with the 'dx' & 'dy' replaced by something equivalent to them (re-read my derivation).

Stephen Tashi said:
How is such a substitution justifed by standard calculus? You're putting an approximation into an infinite series of terms.

What I'm doing is replacing one real number by another slightly smaller real number, where the numbers were so small to begin with that the difference between them doesn't matter. The infinite series x_1 = x_0 + Ux|_{x_0} \delta \alpha + ... is assumed to converge, i.e. you are supposed to choose \alpha so small that the above series converges, i.e. so small that the difference between x_1 & x_0 is negligible - where these are two real numbers - not variables (sloppy notation again). Again it is by doing this that we'll end up, hopefully, proving the theorem I mentioned above (or at least implicitly using it). The reason the books are alright in using x & not x_0 in the notation is that the point x_0 is arbitrary & it's a standard abuse of notation for authors not to be so pedantic all the time, but it can lead you into trouble (as this thread clearly shows) so use this as a lesson to become familiar with what we're doing. For me I had trouble with these things when learning how to translate multivariable calculus to manifolds & found a lot of flawed assumptions I was making was based on this exact abbreviation in notation so don't feel bad :-p

Stephen Tashi said:
It looks to me like you are reproducing what the old books do using an argument based on infinitesimals. You do supply more details than the old books.

I'm just expanding a Taylor series as is done in Emanuel & Cohen, the only difference is that I'm being more careful than they are in specifying my initial points (x_0,y_0) with subscripts.

Stephen Tashi said:
However, if we use standard calculus, we obtain the 3rd term of the Taylor series that I got in post #24 of https://www.physicsforums.com/showthread.php?t=697036&page=2. I think I can get the third term to work out without any infinitesimal reasoning. I'll try to accomplish that in my next post in that thread.

I'll be honest, your notation is very confusing, it's already managed to lead jostpuur into doing something equivalent to differentiating the "dx" & "dy" terms in \frac{\partial f}{\partial x}dx + \frac{\partial f}{\partial y}dy & some of what you're doing is not consistent with standard practice. When you refer to your "corrected definition" you are actually doing something completely incorrect, I've explained above why what you're doing is wrong. What you had originally was correct, so if you are now managing to get the third term without infinitesimal reasoning it's a complete fluke (though I don't see anything you wrote in your last post in that thread as being correct). I honestly would be more careful with my notation, with so many S's, T's & W's thrown around it's very easy to get confused, there are very standard conventions for doing everything you're doing, like replacing an x in a partial with an x' or X when using the chain rule. If you want to use tons of letters that's fine, but it'll probably lead to more problems unless you're really careful so keep an eye on it.

Stephen Tashi said:
I think the old books "have something going for them" and the way they reason would be useful to know. However, I want it translated to (or at least clearly related to) the modern viewpoint of calculus.

What we're doing is on the level of modern calculus. If you want to do this from an analysis standpoint you'll have to be comfortable differential forms, be willing to start constructing tangent & cotangent spaces & phrase everything you're doing in terms of the definition of a one-parameter group I gave in my post in the other thread, something encoding about 4 layers of pedanticism not present in anything we're seeing, & we'd probably have to invoke jets etc... just to do what we're doing. It's far too much for us, far too roundabout a way of getting the exact same answer we're striving for based off pure intuition, which is why I'm not trying to learn the manifold perspective of this theory until I know what I'm doing because all intuition is lost. Even in the general relativity lectures by Susskind on youtube he talks about having to go back to basic tensor analysis when doing his own work, not being able to invoke the formalism naturally because it's tough to do... Turning this into a modern perspective will become a matter of notation once we're familiar with it.
 
  • #57
bolbteppa said:
First of all you actually made the mistake when you wrote \frac{\partial u_x}{\partial x}, the u_x should be written as u_x = u_x(x_0,y_0,\alpha), in the context of what we're doing it is only a function of \alpha being evaluated at x = x_0, y = y_0

This is as wrong as 0=1.

If \mathcal{C}^{\infty}(\mathbb{R}^2) is the set of smooth functions f:\mathbb{R}^2\to\mathbb{R}, the operator U can be seen as a mapping U:\mathcal{C}^{\infty}(\mathbb{R}^2)\to\mathcal{C}^{\infty}(\mathbb{R}^2). That means that if f is a smooth function on the plane, also Uf is a smooth function on the plane. In other words

<br /> f\in\mathcal{C}^{\infty}(\mathbb{R}^2)\quad\implies\quad<br /> Uf\in\mathcal{C}^{\infty}(\mathbb{R}^2)<br />

And the smooth function Uf is defined by the formula

<br /> (Uf)(x,y) = u_x(x,y)\frac{\partial f(x,y)}{\partial x} + u_y(x,y)\frac{\partial f(x,y)}{\partial y}<br />

In other words there is no special point (x_0,y_0) which would be substituted into u_x,u_y always indepedent of the input of Uf.

If you want to compute the partial derivative of Uf with respect to x, the definition is this:

<br /> \frac{\partial (Uf)(x,y)}{\partial x} = \lim_{\Delta x\to 0} \frac{(Uf)(x+\Delta x, y) - (Uf)(x,y)}{\Delta x}<br />

and the u_x,u_y will not be constants.

Finally, when you want to study U^2f = U(Uf), you will need the partial derivatives of Uf.
 
  • #58
After all, U(Uf) is a smooth function defined by the formula

<br /> (U(Uf))(x,y) = u_x(x,y)\frac{\partial (Uf)(x,y)}{\partial x} + u_y(x,y)\frac{\partial (Uf)(x,y)}{\partial y}<br />

...

Wait a minute, what was alpha doing in u_x(x,y,\alpha)? The u_x is defined as

<br /> u_x(x,y) = \frac{\partial T_x(x,y,0)}{\partial \alpha}<br />

where the parameters of T_x are T_x(x,y,\alpha) by convention.
 
  • #59
If you are going to invoke the language akin to that of the language of manifolds to analyze this then you should really see how what you've written makes no sense in our context because you are conflating terms from the cotangent space basis with vectors in the tangent space, picking them out of their respective spaces & just mushing them together in some magical operator U - do you not see that? That alone should convince you what you're writing is flawed, but forget about the U operator, go back to where I wrote out Taylor's theorem to second order (before I ever even mentioned U) & replace the \delta x & \delta y terms with \frac{\partial \phi}{\partial \alpha}|_{\alpha = \alpha_0}\delta \alpha & \frac{\partial \psi}{\partial \alpha}|_{\alpha = \alpha_0}\delta \alpha, then compare your two answers to have another reason why what you've written is wrong. You will now unequivocally see you are misinterpreting the meaning of U, because we are using what U is actually derived from. Another way you should see this is wrong is when you try to explain to me why what you did makes perfect sense yet what I did (when I wrote (U\delta x)f + \delta x (Uf) + ...), i.e. follow your own logic, is illogical (or is it?).

Yes there is no special point, as I made clear in my post, but in the derivation of Taylor's theorem you need to specify the point you're working from in order to derive the series expansion, again as I made clear in my post. Only then can you invoke the arbitrariness of the point we're deriving things from, which implies the operator U does not act on the coefficients (this should be obvious because you are getting the wrong answer based on the little test I've given you above). Furthermore you've ignored my illustration of the flaw in what you've done - why is it in your case it makes perfect sense yet when I use your own logic on a basic Taylor series it becomes nonsensical? You have not answered this.

jostpuur said:
Wait a minute, what was alpha doing in u_x(x,y,\alpha)?

Sorry that was a mistake, I only deleted one of the (x_0,y_0,\alpha)|_{\alpha = \alpha_0} it seems, corrected now.
 
  • #60
I don't understand most of the issues in the last few posts, so I proclaim my neutrality.


bolbteppa said:
What I'm doing is replacing one real number by another slightly smaller real number, where the numbers were so small to begin with that the difference between them doesn't matter. The infinite series x_1 = x_0 + Ux|_{x_0} \delta \alpha + ... is assumed to converge, i.e. you are supposed to choose \alpha so small that the above series converges, i.e. so small that the difference between x_1 & x_0 is negligible - where these are two real numbers - not variables (sloppy notation again).

It seems to me that if I try put this in a modern context, that I should avoid the idea of "a real number so close to another real number that it doesn't matter". Is there any way to look at it as an iterated limit? - along the lines of:

Limit as something ( perhaps \delta x, \delta y?)) approaches 0 of ( Limit as n -> infinity of a series of terms that are a function of alpha and the something).
Swap the limits.
= limit as n->infinity of ( limit of each term as the something ->0)
The limit of each term is an appropriate term for a power series in alpha
= Limit as n->infinity of a Taylor series in alpha.

Emmanuel remarks that he isn't concerned with the radius of convergence of the Taylor series. I don't know why. My intuition confused by the fact that the Taylor series expansion in alpha makes no (overt) assumption that alpha is small. Yet deriving seems to depend delicately on making (x1,y1) close to (x0,y0). I don't understand why(x0,y0) is needed at all. From some of what you just wrote, I have the thought that the Taylor series is first approached as a series in \delta x , \delta y and that the \alpha gets in there by a substitution.

I think my derivation begun the other thread will work. I don't think it will be a fluke if I show it utilizes the result of post #49 of this thread.
 

Similar threads

Replies
24
Views
6K
Replies
81
Views
10K
  • · Replies 2 ·
Replies
2
Views
353
  • · Replies 2 ·
Replies
2
Views
1K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 37 ·
2
Replies
37
Views
9K
Replies
7
Views
2K
  • · Replies 87 ·
3
Replies
87
Views
8K
  • · Replies 7 ·
Replies
7
Views
3K
  • · Replies 10 ·
Replies
10
Views
3K