What's the deal on infinitesimal operators?

Stephen Tashi
Science Advisor
Homework Helper
Education Advisor
Messages
7,864
Reaction score
1,602
Is there a treatment of "infinitesimal operators" that is rigorous from the epsilon-delta point of view?

In looking for material on the infinitesimal transformations of Lie groups, I find many things online about infinitesimal operators. Most seem to be by people who take the idea of infinitesimals seriously and I don't think they are talking about the rigorous approach to infinitesimals [a la Abraham Robinson and "nonstandard analysis".

I suppose people who work on mainfolds and various morphisms also can deal with infinitesimal operators via some abstraction. However, I'd like to know if there is an approach to infinitesimal operators ( in general, not simply Lie group operators) that is essentially from the "advanced calculus" point of view.

(If not, I suppose I'll have to think about them like a physicist.)
 
Physics news on Phys.org
http://en.wikipedia.org/wiki/Infinitesimal_transformation

Any book on differential geometry or Lie algebras should be rigorous. You can just go through and add epsilons and deltas.

Maybe you are talking about things like

$$e^{t \, D} \mathop{f}(x)=\mathop{f}(x+t) \\
e^{t \, D} \, \sim \, 1+t \, D $$

That is entirely rigorous, it is just that the equivalence is to first order.
 
lurflurf said:

That's the correct subject, but it isn't the correct approach. (I find it amusing that calculus students are discouraged from proving things in terms of infinitesimals, but when we get to some advanced topics, reasoning by way of infinitesimals suddenly becomes respectable.)

Any book on differential geometry or Lie algebras should be rigorous. You can just go through and add epsilons and deltas.

I have yet to see such a book that was compatible with an rigorous advanced calculus point of view. The abstract approaches are rigorous, but that's not my interest. The down to Earth books speak in terms of infinitesimals. Do you know of a book with a practical orientation that treats infinitesimal transformations without using infinitesimals?

Maybe you are talking about things like
<br /> e^{t \, D} \mathop{f}(x)=\mathop{f}(x+t) \\<br />
(I first saw that result in a volume of Oliver Heaviside's Electrical Papers.)


<br /> e^{t \, D} \, \sim \, 1+t \, D <br />
That is entirely rigorous, it is just that the equivalence is to first order.

As far as I can interpret the materials I've seen on applications of Lie groups, the arguments that begin with showing things are true "to the first order" proceed to claim things that are true with "=" meaing equal, not "equal to the first order". Perhaps there is a rigorous way to phrase these arguments, but I have not seen it done.





.
 
I'm not sure what you mean by "rigorous from the epsilon-delta point of view"- the whole point of "epsilon delta limit arguments" is to avoid "infinitesmals", which, in order to be introduced rigorously require an extended number system.
 
Stephen Tashi said:
As far as I can interpret the materials I've seen on applications of Lie groups, the arguments that begin with showing things are true "to the first order" proceed to claim things that are true with "=" meaing equal, not "equal to the first order". Perhaps there is a rigorous way to phrase these arguments, but I have not seen it done.

I am positively surprised to find that I'm not the only one who's annoyed by this.

Eventually I have not evolved to become an expert on this stuff, so I cannot answer the original question. But I am under impression that for example "Lie Groups Beyond an Introduction" by Anthony W. Knapp would be rigorous like real mathematics. What I do not know is how much it actually answers to those who are puzzled by theoretical physics.
 
HallsofIvy said:
I'm not sure what you mean by "rigorous from the epsilon-delta point of view"- the whole point of "epsilon delta limit arguments" is to avoid "infinitesmals", which, in order to be introduced rigorously require an extended number system.

I do mean that I would like to see a treatment of "infinitesimal operators" that did not make reference to "infinitesimals" but rather used the more familiar (to me) concepts of limits (in the epsilon-delta sense) of functions, sequences of functions etc. My web search on "infinitesimal operators" didn't even turn up a source that defined "infinitesimal operators" except in terms of infinitesimals.

Perhaps I'm using terminology that only turns up things written by physicists. Or perhaps "infinitesimal operators" are one of these areas where physicists have gotten ahead of mathematicians and no non-infinitesimal treatment of them has been written.

An example that I've given in another thread (on using Lie Groups to solve differential equations) is the treatment of the "infinitesimal operator" of a 1-parameter Lie group.

Avoiding the traditional zoo of Greek letters, I prefer the terminology:

Let T(x.y,\alpha) = ( \ T_x(x,y,\alpha), T_y(x,y,\alpha )\ ) denote an element of a Lie group of 1-parameter transformations of the xy-plane onto itself.

Let f(x,y) be a real valued function whose domain is the xy-plane.

Let u_x(x,y) = D_\alpha \ T_x(x,y,\alpha)_{\ |\alpha=0}
Let u_y(x,y) = D_\alpha \ T_y(x,y,\alpha)_{\ |\alpha =0}

(u_x and u_y are the " infinitesimal elements ".)

Let U be the differential operator defined by the operation on the function g(x,y) by:

U g(x,y) = u_x(x,y) \frac{\partial}{\partial x} g(x,y) + u_y(x,y)\frac{\partial}{\partial y} g(x,y)

(The operator U is "the symbol of the infinitesimal transformation" .)

Every book that takes a concrete approach to Lie Groups proves a result that says

f(x_1,y_1) = f(x,y) + U f\ \alpha + \frac{1}{2!}U^2 f \ \alpha^2 + \frac{1}{3!}U^3 f \ \alpha^3 + ..

by using Taylor series.

However, the function they are expanding is (to me) unclear.

If I try to expand f(x_1,y_1) = f(T_x(x,y,\alpha),T_y(x,y,\alpha)) in Taylor series, only the first two terms of that result work.

If I expand f(x_1,y_1) = f( x + \alpha \ u_x(x,y), y + \alpha \ u_y(x,y) ) then I get the desired result. So I think this is equivalent to what the books do because they do not give an elaborate proof of the result. They present it as being "just calculus" and expanding f(x_1,y_1) = f( x + \alpha \ u_x(x,y), y + \alpha \ u_y(x,y) ) is indeed just calculus.

The books then proceed to give examples where the above result is applied to expand f(x_1,y_1) = f(T_x(x,y,\alpha),T_y(x,y,\alpha)) I haven't found any source that justifies this expansion except by using the concept of infinitesimals.
 
Last edited:
I am not sure I understand the question. But it seems to me that epsilon delta proof are avoided for simplicity of exposition.
 
lavinia said:
I am not sure I understand the question.

I'd be happy to have answers to either of the following questions.

1. If you are famililar with books about Lie groups that do a Taylor expansion in terms of the operator U what is the definition of the function [f(x_1,y_1) that they are expanding?

2. What justifies using that expansion for the function f(T_x(x,y,\alpha),T_y(x,y,\alpha) )?
But it
seems to me that epsilon delta proof are avoided for simplicity of exposition.
I suppose lack of clarity is a type of simplicity.
 
The only context in which I've ever seen the phrases "infinitesimal operators" and "infinitesimal transformations" are in QM and GR texts respectively with the latter being related to one-parameter Lie groups (specifically one-parameter diffeomorpism groups of space-times). I've never seen a math book that uses the terminology. In the case of "infinitesimal transformations", one starts with a one-parameter group of diffeomorphisms ##\varphi_t## on a smooth manifold ##M## generated by a vector field ##v^{a}## and defines the Lie derivative of a tensor field ##T^{a_1...a_k}{}{}_{b_1...b_l}## on ##M## as \mathcal{L}_{v}T^{a_1...a_k}{}{}_{b_1...b_l} = \lim_{t\rightarrow 0}\{\frac{\varphi^{*}_{-t}T^{a_1...a_k}{}{}_{b_1...b_l} - T^{a_1...a_k}{}{}_{b_1...b_l}}{t}\} I guess in that sense the "infinitesimal transformations" are codified by regular ##\epsilon-\delta## limits.
 
Last edited:
  • #10
WannabeNewton said:
In the case of "infinitesimal transformations", one starts with a one-parameter group of diffeomorphisms ##\varphi_t## on a smooth manifold ##M## generated by a vector field ##v^{a}## and defines the Lie derivative of a tensor field ##T^{a_1...a_k}{}{}_{b_1...b_l}## on ##M## as \mathcal{L}_{v}T^{a_1...a_k}{}{}_{b_1...b_l} = \lim_{t\rightarrow 0}\{\frac{\varphi^{*}_{-t}T^{a_1...a_k}{}{}_{b_1...b_l} - T^{a_1...a_k}{}{}_{b_1...b_l}}{t}\} I guess in that sense the "infinitesimal transformations" are codified by regular ##\epsilon-\delta## limits.

My simplistic view of the above is that the vector field in my example is (u_x,u_y) and one can define a directional derivative of a function at each point (x,y) that is taken with respect to the field vector at that point. Those concepts are defined using the ordinary notion of limit. However, I don't understand how Taylor expansions of functions in terms of the operator U are proven without resorting to arguments using infinitesimals..
 
  • #11
Ah yes. The book "Modern Quantum Mechanics"-Sakurai does that over and over and the justifications are by mathematical standards abysmal. All he does is ignore the terms of second order and higher in the infinitesimals. To give an example: http://postimg.org/image/fxhr3sgmf/

This is one of the reasons I absolutely hate that book. Apparently the author had some kind of grudge against mathematical rigor. I don't get how anyone can settle for such hand-waviness. The book by Ballentine is supposed to be more rigorous mathematically but: http://postimg.org/image/h29z9wlzb/

You might be interested in Stone's theorem: http://en.wikipedia.org/wiki/Stone's_theorem_on_one-parameter_unitary_groups and this thread: http://physics.stackexchange.com/questions/62876/translator-operator

I'm sure there is a QM for mathematicians book out there that treats these "infinitesimal operators" with more care. I agree with you that the notion of infinitesimal generators in the context of vector fields is much simpler to make mathematically rigorous than in the context of operators.
 
Last edited by a moderator:
  • #12
Stephen Tashi said:
I'd be happy to have answers to either of the following questions.

1. If you are famililar with books about Lie groups that do a Taylor expansion in terms of the operator U what is the definition of the function [f(x_1,y_1) that they are expanding?

I suppose lack of clarity is a type of simplicity.

Since I don't have your book let's start with an example to see if this is the right track.

SO(2) parameterized as

G(x) =

cos(x) -sin(x)
sin(x) cos(x)G can be expressed as a Taylor series around the identity matrix. This series can be computed directly.

On the other hand since G(x) is a homomorphism from the real line with addition, to the group of rotations under matrix multiplication, one can derive the equation

dG/dx = VG where VG is the matrix product of V and G and V is the matrix,

0 -1
1 0

V is the derivative of G at zero and the iterated derivatives of G at zero are just the iterated powers of V.

So in this case G(x) = exp(xV).

V is called the infinitesimal generator of G because every element of G is the exponential of V times the parameter,X. This simplifies G since now it can be derived from a single matrix.

Another way to look at infinitesimal generators of SO(2) is to look at the infinitesimal effect of a rotation of the plane upon a differentiable function.

The function, F(x,y) under rotation becomes a function of θ. That is F(x,y) = F(x(θ),y(θ))

where x(θ) = cosθx_{0} - sinθy_{0} and y'(θ) = sinθx_{0} + cosθy_{0} and (x_{0},y_{0}) is the point at which F is being differentiated and θ is the angle of rotation.

so dF/dθ = (∂F/∂x)(∂x/∂θ) + (∂F/∂y)(∂y/∂θ) = (∂F/∂x)(-x_{0}sinθ - y_{0}cosθ) + (∂F/∂y)(x_{0}cosθ -y_{0}sinθ)

evaluated at θ = 0. This is -y_{0}∂F/∂x + x_{0}∂F/∂y.

F(x,y) = F(x_{0},y_{0}) + (-y_{0}∂F/∂x + x_{0}∂F/∂y)dθ up to first order.

The infinitesimal generator of the rotation is defined as the operator

-y∂/∂x + x∂/∂y
 
  • #13
lavinia said:
V is the derivative of G at zero and the iterated derivatives of G at zero are just the iterated powers of V.

The question that I've asked (I think) is why, in general, does such a derivative turn out to be iterated powers of a single operator? Matrix groups are special case and ordinary calculus may suffice to show this. Perhaps that gives insight into the general situation.
Another way to look at infinitesimal generators of SO(2) is to look at the infinitesimal effect of a rotation of the plane upon a differentiable function.

The original post asks how to look at it without using arguments that assume existence of "infinitesimals".
 
  • #14
Stephen Tashi said:
The question that I've asked (I think) is why, in general, does such a derivative turn out to be iterated powers of a single operator? Matrix groups are special case and ordinary calculus may suffice to show this. Perhaps that gives insight into the general situation.

The generalization is the 1 parameter subgroup whose derivative at the identity is a particular element of the Lie algebra. In the case of SO(2), the rotation matrix of sines and cosines is a 1 parameter group.

All of the properties of the exponential follow because it is a homomorphism of the real numbers under addition into the Lie group.



The original post asks how to look at it without using arguments that assume existence of "infinitesimals".

Not sure about your problem here. You can think of dθ as the differential of the function,θ. For this first order approximation it just means that for small enough increments in theta the approximation is close and higher order terms become insignificant. This is just your epsilon delta proof. But the important observation is that (the Higher Order Terms/Δθ) all go to zero. So evetually the first order approximation is arbitrarily accurate.
 
Last edited:
  • #15
lavinia said:
The generalization is the 1 parameter subgroup whose derivative at the identity is a particular element of the Lie algebra. In the case of SO(2), the rotation matrix of sines and cosines is a 1 parameter group.

I think that's the problem I have described, if we assume T_x , T_y are infinitely differntiable functions.


All of the properties of the exponential follow because it is a homomorphism of the real numbers under addition into the Lie group.

To justify the exponential in the first place, you need establish that the expansion can be done using powers of the same operator, which, as I said, is what I want to see demonstrated.



Not sure about your problem here. You can think of dθ as the differential of the function,θ.

I don't have a problem with differentials provided a "differential" can be defined without resorting to the terminology of "infinitesimals". The idea that that a first order approximation can be defined is not a problem. The problem is how that implies that the Taylor expansion can be written with an "=" (instead of a \approx) using operators that are a first order approximations in the variable \alpha that is used in the expansions.
 
  • #16
Stephen Tashi said:
To justify the exponential in the first place, you need establish that the expansion can be done using powers of the same operator, which, as I said, is what I want to see demonstrated

For matrix groups this is just a Taylor series expansion. You are solving the differential equation dH/dt = XH or X = (dH/dt)H^{-1} which shows that the tangent vector to H is right invariant.

On an abstract Lie group the differential equation is

dR_{g_{-1}}dH/dt = X where X is an element of the tangent space at the identity.
 
Last edited:
  • #17
lavinia said:
For matrix groups this is just a Taylor series expansion.

You are solving the differential equation dH/dt = XH or X = (dH/dt)H^{-1} which shows that the tangent vector to H is right invariant.

On an abstract Lie group the differential equation is

dR_{g_{-1}}dH/dt = X where X is an element of the tangent space at the identity.

In the books that take a concrete approach to "abstract" Lie groups (i.e. don't restrict themselves to matrix groups) the differential equation approach is mentioned. However, as an independent demonstration of the operator expansion, they purport to do the Taylor expansion directly. When I try this the 3rd term of the expansion

f(T_x(x,y,\alpha),T_y(x,y,\alpha)) = f(x,y) + Uf + \frac{1}{2!} U^2 f\ \alpha + ...

does not come out to be \frac{1}{2!} U^2f\ \alpha^2.

So a particular instance of my original question is whether there is a way to argue that the 3rd term is \frac{1}{2!} U^2f\ \alpha^2. using the special properties of the transformation.

Those special properties being

T_x(x,y,0) = x,\ \ T_y(x,y,0) = y

and the property that allows the homomorphism you mentioned:

T_x(x,y,\alpha + \beta)= T_x(T_x(x,y,\alpha),\beta), T_y(T_y(x,y,\alpha),\beta)
T_y(x,y,\alpha + \beta)= T_x(T_x(x,y,\alpha),\beta), T_y(T_y(x,y,\alpha),\beta)

Or is the claim that the result can be established by doing Taylor expansion directly false?
 
  • #18
Wow, make sure you link this thread to our one in your next post, good stuff!

Stephen Tashi said:
When I try this the 3rd term of the expansion

f(T_x(x,y,\alpha),T_y(x,y,\alpha)) = f(x,y) + Uf + \frac{1}{2!} U^2 f\ \alpha + ...

does not come out to be \frac{1}{2!} U^2f\ \alpha^2.

So a particular instance of my original question is whether there is a way to argue that the 3rd term is \frac{1}{2!} U^2f\ \alpha^2. using the special properties of the transformation.

Emanuel derives it on page 13, explicitly showing how you get the third term, is there something wrong with what he does?

WannabeNewton said:
Ah yes. The book "Modern Quantum Mechanics"-Sakurai does that over and over and the justifications are by mathematical standards abysmal. All he does is ignore the terms of second order and higher in the infinitesimals.

I was reading Gelfand's proof of Noether's theorem yesterday (Page 168) & level to which terms of second order & higher are abused in that proof are stunning. I love the book but in that proof there's just too much of it going on, & not coincidentally it's intimately related to the topic of infinitesimal transformations.

WannabeNewton said:
The only context in which I've ever seen the phrases "infinitesimal operators" and "infinitesimal transformations" are in QM and GR texts respectively

Another place I've seen infinitesimal transformations come up is in the derivation of the Euler angles in classical mechanics, you could now have three subjects you';; have seen this stuff in :-p
 
  • #19
bolbteppa said:
Wow, make sure you link this thread to our one in your next post, good stuff!

only if this thread enlightens me.

Emanuel derives it on page 13, explicitly showing how you get the third term, is there something wrong with what he does?

He just claims its true. And he's correct if f_1(x,y) is defined to be f(x + \alpha\ u_x(x,y), y+ \alpha\ u_y(x,y)), which isn't the function I want to expand.

Try to derrive it using f_1(x,y) = f(T_x(x,y,\alpha), T_y(x,y,\alpha)) and using ordinary calculus - the chain rule, the product rule. It doesn't work because \frac{d^2 f_1}{d\alpha^2} has terms involving factors like \frac{d^2T_x(x,y,\alpha)}{d\alpha^2} When you set \alpha = 0 that factor becomes \frac{d^2 T_x(x,y,\alpha)}{d\alpha^2}_{|\ \alpha = 0}, not (u_x(x,y))^2 = \big( \frac{dT_x(x,y,\alpha)}{d\alpha}_{|\ \alpha = 0} \big)^2. So the derrivation (if there is one) requires something besides straightforward differentiation.
 
  • #20
From the differential equation

dH/dt = XH

one gets d^{2}H/dt^{2} = XdH/dt = X^{2}H

Inductively, d^{n}H/dt^{n} = X^{n}H

H(0) = Identity matrix

so the Taylor series at the identity follows.
The differential equation derives from the homomorphism equation.

H(x+y) = H(x)H(y)

for instance,

dH(x+y)/dx = dH/dxH(y) At x = 0 the right hand side is just XH where X is the derivative of H at 0.

The left hand side can be rewritten as

[dH(x + y)/x+y]d(x+y)/dx using the Chain Rule. This is just the derivative of H.

In sum dH/dt = XH
 
Last edited:
  • Like
Likes 1 person
  • #21
lavinia said:
From the differential equation

dH/dt = XH

one gets d^{2}H/dt^{2} = XdH/dt = X^{2}H

Inductively, d^{n}H/dt^{n} = X^{n}H

H(0) = Identity matrix

so the Taylor series at the identity follows.


The differential equation derives from the homomorphism equation.

H(x+y) = H(x)H(y)

for instance,

dH(x+y)/dx = dH/dxH(y) At x = 0 the right hand side is just XH where X is the derivative of H at 0.

The left hand side can be rewritten as

[dH(x + y)/x+y]d(x+y)/dx using the Chain Rule. This is just the derivative of H.

In sum dH/dt = XH


Thank you very much for that explanation. The part that justifies the differential equation works for a matrix group because composition of the functions in the group becomes multiplication of matrices.
 
  • #22
Most seem to be by people who take the idea of infinitesimals seriously and I don't think they are talking about the rigorous approach to infinitesimals [a la Abraham Robinson and "nonstandard analysis".
Actually, you can define Lie algebras rigorously using nonstandard analysis; see Abraham Robinson's original book.
 
  • #23
A long time ago I was wondering how to prove that the total charge is invariant in Lorentz boosts, under the assumption the the total charge is obtained by a spatial integral over a charge density which is the first component of a four current, also assuming that the four current satisfies a conservation equation.

https://www.physicsforums.com/showthread.php?t=180779

The thing ended in a situation where I was trying to prove that one mapping Q:[0,1]\to\mathbb{R} would be constant, by showing it derivative to be zero, but was unable to accomplish it (trying this: Q&#039;(x)=0\;\forall x.). Then samalkhaiat (now a science advisor) tried to convince me that I can prove Q to be constant by proving that it derivative is zero at one point (like this Q&#039;(0)=0 would be sufficient).

The original problem was left as a mystery to me.
 
  • #24
jostpuur, I'm not seeing the relevance of that to the question of the thread at hand to be honest. As an aside, Lorentz invariance of charge is a consequence of Stokes' theorem and the conserved Noetherian current coming from ##\partial_{a}j^{a} = 0##, that's all there is to it. See e.g. Franklin section 14.10.2
 
Last edited:
  • #25
Are you sure that the result is really proven in that book? Or could it be, that it only proves that the charge is conserved in infinitesimal boosts, then assuming that it implies the full result?
 
  • #26
Charge is conserved is different from charge is invariant but yes it proves that charge is a Lorentz invariant in the usual special relativistic sense under Lorentz boosts. The proof is very simple (it is identical to the one samalkhaiat gave in the thread you linked).
 
  • #27
ok my choice of words "conserved in boosts" was a mistake, but anyway people understand what I meant.

Your answer seems contradictory to me because samalkhait gave the proof only for infinitesimal boosts!

Besides I myself proved the result for infinitesimal boosts too. I was patiently trying to explain that that's how far I got on my own, and I was trying find a way to complete the proof, and then I just got bombarded with the repetition on the infinitesimal boost proof.
 
  • #28
I was speaking of his second method in post #24. This result holds for global Lorentz boosts in Minkowski spacetime. As I said, it is simply a consequence of Stokes' theorem and the fact that ##\partial_{a}j^{a} = 0##, for ##j^{a}## that are compactly supported in a worldtube.
 
  • #29
Can we please keep this thread on topic? Jostpuur, your question is very interesting, but can you ask it in a different thread?
 
  • #30
I am on the topic IMO. The topic is rather broad, because physics is filled with this infinitesimal stuff.

I'm going to show you a magic trick. I'll prove that a function

<br /> f:[0,1]\to [0,1],\quad\quad f(x)=x<br />

is constant! :wink:

The proof happens so that I prove that its derivative is zero everywhere. (In other words, given arbitrary point x\in [0,1], the function is constant in its infinitesimal environment.)

Let x\in [0,1] be fixed. Let's define

<br /> \phi_x(u) = x + (u- x)^3.<br />

Now we prove that the derivative of f\circ\phi_x is zero at x.

<br /> \big(D_u (f\circ\phi_x )\big)(u) = f&#039;(\phi_x(u)) \;3 (u - x)^2<br />

It is zero when you substitute u=x!

So the lesson is this: If you want to prove that f is constant, it is fully ok to carry out some transformation, and prove that f\circ \phi is constant, if see it convenient for some reason. However, it is not ok to first fix some point x, then carry out a transformation that depends on that point, and then conclude that the derivative looks zero there.

That's the problem with the infinitesimal Lorentz boost thing. If you have parametrized some set of coordinates so that they depend on a parameter \alpha, and then have solved a formula for a total charge Q(\alpha), you must actually prove D_{\alpha}Q(\alpha)=0 for all \alpha with this particular representation of this function Q. It is not ok to move into a local representation for each fixed \alpha, and then investigate infinitesimal boosts there.
 
  • #31
Sorry, I don't think you're on topic. So please, make a new thread.
 
  • #32
Stephen Tashi said:
Every book that takes a concrete approach to Lie Groups proves a result that says

f(x_1,y_1) = f(x,y) + U f\ \alpha + \frac{1}{2!}U^2 f \ \alpha^2 + \frac{1}{3!}U^3 f \ \alpha^3 + ..

by using Taylor series.

However, the function they are expanding is (to me) unclear.

If I try to expand f(x_1,y_1) = f(T_x(x,y,\alpha),T_y(x,y,\alpha)) in Taylor series, only the first two terms of that result work.

I would like to denote the series as

<br /> (e^{\alpha U}f)(x,y) = f(x,y) + \alpha (Uf)(x,y) + \frac{1}{2!}\alpha^2 (U^2f)(x,y) +\cdots<br />

Are we speaking about the same thing?

If I expand f(x_1,y_1) = f( x + \alpha \ u_x(x,y), y + \alpha \ u_y(x,y) ) then I get the desired result.

This looks like a mistake to me. The operator U^2 behaves like this:

<br /> U^2f = u_x\frac{\partial}{\partial x}(Uf) + u_y\frac{\partial}{\partial y}(Uf)<br />
<br /> = u_x\Big(\big(\frac{\partial u_x}{\partial x}\frac{\partial f}{\partial x}<br /> + u_x\frac{\partial^2 f}{\partial x^2}<br /> + \frac{\partial u_y}{\partial x}\frac{\partial f}{\partial y}<br /> + u_y\frac{\partial^2 f}{\partial x\partial y}\Big)<br />
<br /> + u_y\Big(\frac{\partial u_x}{\partial u_y}\frac{\partial f}{\partial x}<br /> + u_y\frac{\partial^2 f}{\partial x \partial y}<br /> + \frac{\partial u_y}{\partial y} \frac{\partial f}{\partial y}<br /> + u_y\frac{\partial^2 f}{\partial y^2}\Big)<br />

If you compute derivatives of f(x+\alpha u_x,y+\alpha u_y) with respect to \alpha, the partial derivatives of u_x and u_y will never appear.
 
  • #33
Stephen Tashi said:
and the property that allows the homomorphism you mentioned:

T_x(x,y,\alpha + \beta)= T_x(T_x(x,y,\alpha),\beta), T_y(T_y(x,y,\alpha),\beta)
T_y(x,y,\alpha + \beta)= T_x(T_x(x,y,\alpha),\beta), T_y(T_y(x,y,\alpha),\beta)

This could be a minor thing, but these formulas probably should have been

<br /> T_x(x,y,\alpha + \beta) = T_x(T_x(x,y,\alpha),T_y(x,y,\alpha),\beta)<br />
<br /> T_y(x,y,\alpha + \beta) = T_y(T_x(x,y,\alpha),T_y(x,y,\alpha),\beta)<br />

?
 
  • #34
I don't remember precise conditions for uniqueness of the solutions of some DEs, but anyway I think the seeked result comes when some uniqueness result is used.

We want to prove

<br /> (e^{\alpha U}f)(x,y) = f(T_x(x,y,\alpha),T_y(x,y,\alpha))<br />

The left side satisfies the DE

<br /> D_{\alpha} (e^{\alpha U}f)(x,y) = (U(e^{\alpha U}f)(x,y)<br />

so the proof should be reasonably close if we succeed in proving that the right side satisfies the same DE, that means

<br /> D_{\alpha} f(T_x(x,y,\alpha),T_y(x,y,\alpha)) = (Uf)(T_x(x,y,\alpha),T_y(x,y,\alpha))<br />

(update: I think I made a mistake here. A comment in #36.)

Direct computation gives

<br /> D_{\alpha} f(T_x(x,y,\alpha),T_y(x,y,\alpha))<br />
<br /> = (D_{\alpha}T_x(x,y,\alpha))\frac{\partial f}{\partial x}(T_x(x,y,\alpha),T_y(x,y,\alpha))<br />
<br /> + (D_{\alpha}T_y(x,y,\alpha))\frac{\partial f}{\partial y}(T_x(x,y,\alpha),T_y(x,y,\alpha)) = \cdots<br />

We assume that the transformation satisfies

jostpuur said:
<br /> T_x(x,y,\alpha + \beta) = T_x(T_x(x,y,\alpha),T_y(x,y,\alpha),\beta)<br />
<br /> T_y(x,y,\alpha + \beta) = T_y(T_x(x,y,\alpha),T_y(x,y,\alpha),\beta)<br />

Hence it also satisfies

<br /> D_{\alpha}T_x(x,y,\alpha) = \frac{\partial}{\partial\alpha}T_x(T_x(x,y,\alpha),T_y(x,y,\alpha),0)<br />
<br /> D_{\alpha}T_y(x,y,\alpha) = \frac{\partial}{\partial\alpha}T_y(T_x(x,y,\alpha),T_y(x,y,\alpha),0)<br />

So we get

<br /> \cdots<br /> = \frac{\partial}{\partial\alpha}T_x(T_x(x,y,\alpha),T_y(x,y,\alpha),0)<br /> \frac{\partial f}{\partial x}(T_x(x,y,\alpha),T_y(x,y,\alpha))<br />
<br /> + \frac{\partial}{\partial\alpha}T_y(T_x(x,y,\alpha),T_y(x,y,\alpha),0)<br /> \frac{\partial f}{\partial y}(T_x(x,y,\alpha),T_y(x,y,\alpha))<br />
<br /> = (Uf)(T_x(x,y,\alpha),T_y(x,y,\alpha))<br />
 
Last edited:
  • #35
After struggling with physicists' infinitesimal stuff myself during my studies, I eventually arrived at the enlightened conclusion, that the infinitesimal arguments can often be transformed into differential equations. I think this settles lot of problems, but as I pointed out in my off topic posts, there are also some infinitesimal arguments which I have been unable to transform into rigor form.

When I have attempted to discuss those problems, the physicists are usually unable to understand my complaints, since they have already been indoctrinated into believing anything that has the magic word "infinitesimal" as a part. Then usually the physicists attempt to convince me to abandon the differential equations or other rigor concepts, and accept the infinitesimal arguments for sake of making life easier.

So, to answer the question in the topic, I would say that the "situation with infinitesimal operators" is serious.
 
  • #36
Oh dear, it could be there is something wrong in my "answer".

I think I was supposed to prove

<br /> D_{\alpha} (f\circ T(\alpha))(x,y) = (U(f\circ T(\alpha)))(x,y)<br />

but the result that I proved was

<br /> D_{\alpha} (f\circ T(\alpha))(x,y) = ((Uf)\circ T(\alpha))(x,y)<br />

I'm feeling little confused now...
 
  • #37
If a function has been written in polar coordinates like f(r,\theta), then

<br /> e^{\theta\frac{\partial}{\partial \theta}} f(r,\theta_0)<br />

is the Taylor series with respect to the angle variable. Hence the operator should be seen as a rotation operator. On the other hand

<br /> \frac{\partial}{\partial\theta} = -y\frac{\partial}{\partial x} + x\frac{\partial}{\partial y}<br />

holds too, so we should be able to write the rotation operator in the Cartesian coordinates. Let's define

<br /> \phi(x,y,\theta) = e^{\theta(-y\frac{\partial}{\partial x} + x\frac{\partial}{\partial y})}f(x,y)<br />

and

<br /> \psi(x,y,\theta) = f(x\cos\theta - y\sin\theta,\; x\sin\theta + y\cos\theta)<br />

Now

<br /> \phi(x,y,\theta) = \psi(x,y,\theta)<br />

is supposed to hold, right?

<br /> \frac{\partial}{\partial\theta}\phi(x,y,\theta) = \Big(-y\frac{\partial}{\partial x}+ x\frac{\partial}{\partial y}\Big)\phi(x,y,\theta)<br />

<br /> \frac{\partial}{\partial\theta}\psi(x,y,\theta)<br /> = (-x\sin\theta - y\cos\theta)\frac{\partial f}{\partial x}<br /> + (x\cos\theta - y\sin\theta)\frac{\partial f}{\partial y}<br />
<br /> = -y\Big(\cos\theta\frac{\partial f}{\partial x} +\sin\theta \frac{\partial f}{\partial y}\Big)<br /> + x\Big(\cos\theta\frac{\partial f}{\partial y} - \sin\theta \frac{\partial f}{\partial x}\Big)<br />
<br /> = \Big(-y\frac{\partial}{\partial x} + x \frac{\partial}{\partial y}\Big)\psi(x,y,\theta)<br />

Here equations

<br /> \frac{\partial \psi}{\partial x} = \cos\theta \frac{\partial f}{\partial x} +\sin\theta \frac{\partial f}{\partial y}<br />
<br /> \frac{\partial \psi}{\partial y} = -\sin\theta \frac{\partial f}{\partial x} +\cos\theta\frac{\partial f}{\partial y}<br />

were used.

Ok so \phi and \psi satisfy the same PDE and therefore it should be possible to prove that they are the same thing rigorously.

I'm sure a similar manipulation should be possible in more general case, but I did something wrong in my previous attempt, and all pieces didn't fit together for some reason. It is still little mysterious, but I'm getting optimistic on this now.
 
  • #38
jostpuur said:
This looks like a mistake to me. The operator U^2 behaves like this:

<br /> U^2f = u_x\frac{\partial}{\partial x}(Uf) + u_y\frac{\partial}{\partial y}(Uf)<br />
<br /> = u_x\Big(\big(\frac{\partial u_x}{\partial x}\frac{\partial f}{\partial x}<br /> + u_x\frac{\partial^2 f}{\partial x^2}<br /> + \frac{\partial u_y}{\partial x}\frac{\partial f}{\partial y}<br /> + u_y\frac{\partial^2 f}{\partial x\partial y}\Big)<br />
<br /> + u_y\Big(\frac{\partial u_x}{\partial u_y}\frac{\partial f}{\partial x}<br /> + u_y\frac{\partial^2 f}{\partial x \partial y}<br /> + \frac{\partial u_y}{\partial y} \frac{\partial f}{\partial y}<br /> + u_y\frac{\partial^2 f}{\partial y^2}\Big)<br />


If you compute derivatives of f(x+\alpha u_x,y+\alpha u_y) with respect to \alpha, the partial derivatives of u_x and u_y will never appear.

I see your point. I way trying to interpret U^n f as an operation that is applied only to the original function f and not to the function that is the result of previous operations.
 
  • #39
Just in case somebody feels unsure about how to interpret the U^n, I would recommend studying the example

<br /> \frac{\partial}{\partial\theta} = -y\frac{\partial}{\partial x} + x\frac{\partial}{\partial y}<br />

I didn't check this now myself, but I'm sure that with some example, it can be verified that

<br /> \frac{\partial^2}{\partial\theta^2},\frac{\partial^3}{\partial\theta^3},\ldots<br />

will not turn out correctly, if it is assumed that \frac{\partial}{\partial x},\frac{\partial}{\partial y} would commute with x,y.

I thought that I recognized the Stephen Tashi's problem as something that I had thought through earlier, but now it became clear that I hadn't thought it through completely after all. This seems to be coming a good thread, although it also contains lot of mistakes now, and some from me. :blushing: I'll try to take a break now, and return later to see what's happened.
 
  • #40
With the correction and also adding the evaluation \alpha = 0 to the left hand side,

<br /> D_{\alpha} f(T_x(x,y,\alpha),T_y(x,y,\alpha))_{|\ \alpha = 0} = U (f(T_x(x,y,\alpha),T_y(x,y,\alpha)))<br />

is my interpretation of what old books on Lie groups prove and then they immediately assert
something that (I think) amounts to

D^n_{\alpha} f(T_x(x,y,\alpha),T_y(x,y,\alpha))_{|\ \alpha = 0} = U^n (f(T_x(x,y,\alpha),T_y(x,y,\alpha)))

as if the result is just straightforward calculus. (Perhaps with your correction to my interpretation of U^2, it is.)

In those books, the right hand side is denoted as U( f(x_1,y_1) ).

In previous pages, notation amounting to x_1 = T_x(x,y,\alpha),\ y_1 = T_y(x,y,\alpha) is employed, and in other places the same symbols are also defined implicityly using "infinitesimals" as x_1 + \delta x_1 = T_x(x,y,\alpha + \delta \alpha),\ y1 +\delta y_1 = T_y(x,y,\alpha + \delta \alpha). So I think the interpretation in the first equation is correct.

For matrix groups, lovina's proof works. (Perhaps even my wrong interpretation of U^2 works.)
 
  • #41
Many people react to questions about the expansion in U by brining up differential equations. Intuitively, I understand why. We trace out the curve followed by the point (T_x(x,y.\alpha),T_y(x,y,\alpha) and the value of function f(T_x,T_y) as it varies along that curve, thinking of \alpha as "time". The derivative of f(x,y) with respect to \alpha depends on the gradient of f and the directional derivative of the curve at (x,y).

Since the curve is a group of transformations, the directional derivative at (x,y) is determined by what the transformation does to the point (x,y) by values of \alpha close to zero.

I'd think that some advanced calculus book somewhere would treat expanding a function f(x,y) in Taylor series "along a curve". Then the problem for adapting this result to 1-parameter Lie groups would be to show how that result becomes simpler when the curve is generated by a tranformation group.
 
  • #42
One new notation:

<br /> T(\alpha)(x,y) = \big(T_x(x,y,\alpha),T_y(x,y,\alpha)\big)<br />

so that T(\alpha) is a function with two variables.

Stephen Tashi said:
D^n_{\alpha} f(T_x(x,y,\alpha),T_y(x,y,\alpha))_{|\ \alpha = 0} = U^n (f(T_x(x,y,\alpha),T_y(x,y,\alpha))) \quad\quad\quad (1)<br />

Expressions like this are dangerous, since there could be ambiguity about whether we mean

<br /> \big(U(f\circ T(\alpha)\big)(x,y)\quad\quad\quad (2)<br />

or

<br /> \big((Uf)\circ T(\alpha)\big)(x,y)\quad\quad\quad (3)<br />

The difference between these is that (2) involves weights

<br /> u_x(x,y),\quad u_y(x,y)<br />

while (3) involves weights

<br /> u_x\big(T_x(x,y,\alpha),T_y(x,y,\alpha)\big),\quad u_u\big(T_x(x,y,\alpha),T_y(x,y,\alpha)\big)<br />

I did some mistakes with these issues in my previous attempt to deal with the infinitesimal problem through PDEs.

Being fully strict with the notation, a f(x_0,y_0) is only a real number once x_0,y_0 have been fixed. An operation by U from left on a real number is not defined, hence U(f(x_0,y_0)) is not defined. I wouldn't get too strict merely for the sake of being strict, but critical ambiguity should be avoided.
 
Last edited:
  • #43
Stephen Tashi said:
Many people react to questions about the expansion in U by brining up differential equations. Intuitively, I understand why.

IMO the PDE approach is extremely critical. The reason is that sometimes the exponential series can diverge, but the transformation itself may still make sense. Even then, the transformation could be seen as generated by some PDE. The exponential operator could then be seen as a formal notation for an operator that generates the transformation.

For example, the translation operator is well defined even if Taylor series diverge at some points. Also, a transport equation (PDE) will generate the translation very well for differentiable functions whose Taylor series diverges at some points.

Stephen Tashi said:
Is there a treatment of "infinitesimal operators" that is rigorous from the epsilon-delta point of view?

In looking for material on the infinitesimal transformations of Lie groups, I find many things online about infinitesimal operators. Most seem to be by people who take the idea of infinitesimals seriously and I don't think they are talking about the rigorous approach to infinitesimals [a la Abraham Robinson and "nonstandard analysis".

I have never seen any such rigour material. My own replies were combination of knowledge on PDEs, groups, and some own thinking.

Everytime I have read about Lie groups, it has been about actions on some finite dimensional manifolds. Now these differential operators are acting on infinite dimensional vector spaces. Are these infinite dimensional Lie groups? I guess not. The groups themselves are still finite dimensional?

Returning to the opening post... it would be interesting to learn about some pedagogical material on this. It seems replies containing links to such did not appear yet.
 
Last edited:
  • #44
By comparing my posts #34 and #37, it is clear that #34 contained some fatal mistakes. I would like to outline the goal more clearly now.

We define two functions \phi,\psi by formulas

<br /> \phi(x,y,\alpha) = (e^{\alpha U}f)(x,y)<br />

<br /> \psi(x,y,\alpha) = f\big(T_x(x,y,\alpha),T_y(x,y,\alpha)\big)<br />

Now \phi(x,y,0)=\psi(x,y,0) holds obviously, and

<br /> \frac{\partial\phi(x,y,\alpha)}{\partial\alpha} = U\phi(x,y,\alpha)<br />

holds almost by the definition. The goal should be to prove that \psi satisfies the same PDE, which would then imply \phi=\psi.

<br /> \frac{\partial\psi(x,y,\alpha)}{\partial\alpha}<br /> = \frac{\partial T_x(x,y,\alpha)}{\partial\alpha} \frac{\partial f}{\partial x}<br /> + \frac{\partial T_y(x,y,\alpha)}{\partial\alpha} \frac{\partial f}{\partial y}<br />

<br /> \frac{\partial\psi}{\partial x} =\frac{\partial f}{\partial x}\frac{\partial T_x}{\partial x}<br /> + \frac{\partial f}{\partial y}\frac{\partial T_y}{\partial x}<br />

<br /> \frac{\partial\psi}{\partial y} =\frac{\partial f}{\partial x}\frac{\partial T_x}{\partial y}<br /> + \frac{\partial f}{\partial y}\frac{\partial T_y}{\partial y}<br />

In the end we want U\psi to appear, so we need to get the \frac{\partial\psi}{\partial\alpha} written in such form that \frac{\partial\psi}{\partial x} and \frac{\partial\psi}{\partial y} would be present, and not the derivatives of f.

It's not immediately obvious how to accomplish that. The properties of the T_x,T_y shoul be studied in more detail.
 
  • #45
jostpuur said:
One new notation:

<br /> T(\alpha)(x,y) = \big(T_x(x,y,\alpha),T_y(x,y,\alpha)\big)<br />

so that T(\alpha) is a function with two variables.



Expressions like this are dangerous, since there could be ambiguity about whether we mean

<br /> \big(U(f\circ T(\alpha)\big)(x,y)\quad\quad\quad (2)<br />

or

<br /> \big((Uf)\circ T(\alpha)\big)(x,y)\quad\quad\quad (3)<br />

What about
<br /> U( f\circ (T(\alpha)(x,y) )) \quad\quad\quad (4)<br />

which is how I'd translate

D_{\alpha} f(T_x(x,y,\alpha),T_y(x,y,\alpha))_{|\ \alpha = 0} = U (f(T_x(x,y,\alpha),T_y(x,y,\alpha)))

into your notation.

Of course, I'm just guessing at what the old books mean. The result I'm trying to translate into modern calculus is shown in An Introduction to the Lie Theory of one-parameter groups" by Abraham Cohen (1911) http://archive.org/details/introlietheory00coherich page 30 of the PDF, page 14 of the book.

The expansion is also asserted in the modern book Solution of Ordinary Differential Equations by Continuous Groups by George Emmanuel (2001) page 13,
 
  • #46
Continuing the work from post #44

When you don't know how to arrive at the goal, the basic trick is to see what happens if you work backwards from the goal. It turns out this:

<br /> U\psi(x,y,\alpha) = \frac{\partial T_x(x,y,0)}{\partial\alpha}<br /> \frac{\partial\psi(x,y,\alpha)}{\partial x}<br /> + \frac{\partial T_y(x,y,0)}{\partial\alpha}\frac{\partial\psi(x,y,\alpha)}{\partial y}<br />
<br /> =\Big(\frac{\partial T_x(x,y,0)}{\partial\alpha}\frac{\partial T_x(x,y,\alpha)}{\partial x}<br /> + \frac{\partial T_y(x,y,0)}{\partial\alpha}\frac{\partial T_x(x,y,\alpha)}{\partial y}\Big)<br /> \frac{\partial f(T_x(x,y,\alpha),T_y(x,y,\alpha))}{\partial x}<br />
<br /> +\Big(\frac{\partial T_x(x,y,0)}{\partial\alpha}\frac{\partial T_y(x,y,\alpha)}{\partial x}<br /> +\frac{\partial T_y(x,y,0)}{\partial\alpha}\frac{\partial T_y(x,y,\alpha)}{\partial y}\Big)<br /> \frac{\partial f(T_x(x,y,\alpha),T_y(x,y,\alpha))}{\partial y}<br />

Comparing this with equation of post #44 we see that \psi will satisfy the desired PDE if the equations

<br /> \frac{\partial T_x(x,y,\alpha)}{\partial\alpha}<br /> = \frac{\partial T_x(x,y,0)}{\partial\alpha}\frac{\partial T_x(x,y,\alpha)}{\partial x}<br /> + \frac{\partial T_y(x,y,0)}{\partial\alpha}\frac{\partial T_x(x,y,\alpha)}{\partial y}<br />
<br /> \frac{\partial T_y(x,y,\alpha)}{\partial\alpha}<br /> = \frac{\partial T_x(x,y,0)}{\partial\alpha}\frac{\partial T_y(x,y,\alpha)}{\partial x}<br /> +\frac{\partial T_y(x,y,0)}{\partial\alpha}\frac{\partial T_y(x,y,\alpha)}{\partial y}<br />

hold.

Do these follow from some axioms about the transformation?
 
  • #47
Stephen Tashi said:
What about
<br /> U( f\circ (T(\alpha)(x,y) )) \quad\quad\quad (4)<br />

This contains two mistakes. You probably meant

<br /> U((f\circ T(\alpha))(x,y))<br />

?

The problem with this is that (f\circ T(\alpha))(x,y) is a real number, and we cannot operate with U from left on real numbers. The equation (4) also halts because T(\alpha)(x,y) is a point on plane, and f\circ (x&#039;,y&#039;) is not defined for any point (x&#039;,y&#039;), but only f(x&#039;,y&#039;) is.

Notice that the difference between equations (2) and (3) was that the weights u_x,u_y were evaluated at different points. There are only two obvious ways to evaluate u_x,u_y, so the ambiguous expressions will consequently have at most two obvious interpretations.
 
  • #48
jostpuur said:
This contains two mistakes. You probably meant

<br /> U((f\circ T(\alpha))(x,y))<br />

I won't say I mean that since I don't know what it means yet.

Is the convention you're using that any function G mapping 3 variables x,y,\alpha) to a point on the plane (G_x(x,y,\alpha), G_y(x,y,\alpha) ) will be denoted as G(\alpha) (x,y)? Or is this honor reserved for T(\alpha)(x,y) alone ?

I don't see f \circ T as a function that maps 3 variables to a point on the plane and I don't mean that.


The problem with this is that (f\circ T(\alpha))(x,y) is a real number, and we cannot operate with U from left on real numbers.

I can't interpret (f\circ T(\alpha))(x,y). The two arguments (x,y) sitting outside the parentheses confuse me.

Returning to the notation T(x,y,\alpha) = (T_x(x,y,\alpha), T_y(x,y,\alpha))
I consider f\circ T(x,y,\alpha) as a real valued function of 2 variables where each of those variables is replaced by a real valued function of 3 variables. .So f is a only a specific real number when x, y, \alpha are specific real numbers.


The equation (4) also halts because T(\alpha)(x,y) is a point on plane, and f\circ (x&#039;,y&#039;) is not defined for any point (x&#039;,y&#039;), but only f(x&#039;,y&#039;) is.

I don't understand what the primes mean, but I think that argument is a further objection to something that I don't mean anyway.


Notice that the difference between equations (2) and (3) was that the weights u_x,u_y were evaluated at different points. There are only two obvious ways to evaluate u_x,u_y, so the ambiguous expressions will consequently have at most two obvious interpretations.

I heartily agree that where the weights are to be evaluated needs precise notation. Denoting a real valued function g as g(A,B) to avoid the confusing proliferation of x's and y's, I think that U is the operator

U g(A,B) = u_x(A,B)\frac{\partial}{\partial A} + u_y(A,B) \frac{\partial}{\partial B}.

So in the case that A = T_x(x,y,\alpha),\ B = T_x(x,y,\alpha) those are the
values where the derivatives and weights are evaluated. An \alpha is thus introduced into this evaluation. When we set \alpha = 0 to get the Taylor coefficient we have T_x(x,y,0) = x, \ T_y(x,y,0) = y so the evaluation in the Taylor coefficient is done at (A,B) = (x,y).

Of course, I'm not yet certain how the old books are defining U.
 
  • #49
jostpuur said:
Comparing this with equation of post #44 we see that \psi will satisfy the desired PDE if the equations

<br /> \frac{\partial T_x(x,y,\alpha)}{\partial\alpha}<br /> = \frac{\partial T_x(x,y,0)}{\partial\alpha}\frac{\partial T_x(x,y,\alpha)}{\partial x}<br /> + \frac{\partial T_y(x,y,0)}{\partial\alpha}\frac{\partial T_x(x,y,\alpha)}{\partial y}<br />
<br /> \frac{\partial T_y(x,y,\alpha)}{\partial\alpha}<br /> = \frac{\partial T_x(x,y,0)}{\partial\alpha}\frac{\partial T_y(x,y,\alpha)}{\partial x}<br /> +\frac{\partial T_y(x,y,0)}{\partial\alpha}\frac{\partial T_y(x,y,\alpha)}{\partial y}<br />

hold.

Do these follow from some axioms about the transformation?

I think they do.

I don't follow the entire argument yet, but I at least understand the above question.

We assume the coordinate functions obey the composition law:

T_x( T_x(x,y,\beta), T_y(x,y,\beta),\alpha) = T_x( x,y,\alpha + \beta) [1]
T_y( T_x(x,y,\beta), T_y(x,y,\beta),\alpha) = T_y( x,y,\alpha + \beta) [2]

I copy the technique that lovinia showed for matrix groups.

Consider T_x to be T_x(A,B,C) to avoid confusion. Differentiate both sides of [1] with respect to \beta :

[3]:
\frac{\partial T_x}{\partial A} \frac{\partial Tx}{\partial C} + \frac{\partial T_x}{\partial B} \frac{\partial T_y}{\partial C} = \frac{\partial T_x}{\partial C}

The evaluations of these functions are:

Left hand side:

\frac{\partial T_x}{\partial A} at A = T_x(x,y,\beta),\ B = T_y(x,y,\beta),\ C = \alpha
\frac{\partial T_x}{\partial B} at A = T_x(x,y,\beta),\ B = T_y(x,y,\beta),\ C = \alpha
\frac{\partial T_x}{\partial C} at A = x,\ B = y ,\ C = \beta)
\frac{\partial T_y}{\partial C} at A = x,\ B = y ,\ C = \beta)
Right hand side:
\frac{\partial T_x}{\partial C} at A = T_x(x,y,\beta+\alpha),\ B = T_y(x,y,\beta+\alpha),\ C = \alpha + \beta




Set \beta = 0 in [3]. This sets T_x(x,y,\beta)= x,\ T_y(x,y,\beta) = y so we have the same functions with the evaluations:

\frac{\partial T_x}{\partial A} at A = x,\ B = y ,\ C = \alpha
\frac{\partial T_x}{\partial B} at A = x,\ B = y, \ C = \alpha
\frac{\partial T_x}{\partial C} at A = x,\ B = y,\ C = 0
\frac{\partial T_y}{\partial C} at A = x,\ B = y,\ C = 0

Right hand side:
\frac{\partial T_x}{\partial C} at A =x,\ B = y,\ C = \alpha

I interpret the above right hand side as the same thing as \frac{\partial T_x}{\partial \alpha}

Using your notation ( \frac{\partial T_x}{\partial x} instead of \frac{\partial T_x}{\partial A} etc.), I interpret the above left hand side to be the same thing as

\frac{\partial T_x}{\partial x}(x,y,\alpha) \frac{\partial T_x}{\partial \alpha}(x,y,0) + \frac{\partial T_x}{\partial y}(x,y,\alpha) \frac{\partial T_y}{\partial \alpha}(x,y,0)

so I get the equation you needed for the derivatives of T_x.
 
  • #50
Stephen Tashi said:
We assume the coordinate functions obey the composition law:

T_x( T_x(x,y,\beta), T_y(x,y,\beta),\alpha) = T_x( x,y,\alpha + \beta) [1]
T_y( T_x(x,y,\beta), T_y(x,y,\beta),\alpha) = T_y( x,y,\alpha + \beta) [2]

I copy the technique that lovinia showed for matrix groups.

Consider T_x to be T_x(A,B,C) to avoid confusion. Differentiate both sides of [1] with respect to \beta :

I had only considered derivatives with respect to \alpha from this equation. The derivative with respect to \beta was the trick.

I don't see much problems with this anymore. IMO this is settled now reasonably.

The two dimensional plane was chosen for simplicity of study in the beginning. It could be interesting to check the same thing in higher dimension, and get some guarantees for the uniqueness of solutions of PDEs.
 
Back
Top