Possible mistake in an article (rotations and boosts).

  • Thread starter Thread starter Fredrik
  • Start date Start date
  • Tags Tags
    article Mistake
  • #51
Thank you for those comments. I appreciate them a lot.

1. Fixed.

2. I meant the angle that corresponds to the velocity c, i.e. the ##\theta## such that ##c\tan\theta=c##. You're right that this is ##\pi/4##, not ##\pi/2##.

3. Yes, I've been changing my mind over and over about when to use the words "theorem", "lemma", "corollary", so I ended up with lots of mistakes like this. I had already fixed most of them when I uploaded the pdf, but apparently not all. I found this specific mistake in four places. I will make another sweep for similar mistakes.

4. The velocity addition formula is
$$V(\Lambda\Lambda')=\frac{\rho'v+v'}{1+Kvv'\rho'}.$$ When K<0 and ##\Lambda'=\Lambda##, this turns into
$$V(\Lambda^2)=\frac{\rho v+v}{1-|K|v^2\rho}.$$ This blows up at v=c (and v=-c) only for proper transformations. But that's actually all I need, so I should definitely use this.

My strategy to rule out K<0 is to prove the following:

  • There is NO proper transformation with velocity c.
  • If ε>0 is such that (-ε,ε) is in the range of the velocity function V (such an ε exists because of my assumption 1b), then for each v in that interval, there's a proper transformation with velocity v.
  • For each proper ##\Lambda##, the angle corresponding to ##\Lambda^n## is n times the angle corresponding to ##\Lambda##.
  • Find the angle ##\theta_c## corresponding to the velocity c (which is forbidden for proper transformations), then chose an integer n such that ##\theta_c/n<\varepsilon##. Let ##\Lambda## be a proper transformation with velocity ##\arctan(\theta_c/n)##. Then ##\Lambda^n## is proper and the angle corresponding to ##\Lambda^n## is ##n\theta_c/n=\theta_c##, so the velocity of ##\Lambda^n## is ##c\tan\theta_c##. Since ##\theta_c=\pi/4## and ##\tan(\pi/4)=1##, this means that the velocity of ##\Lambda^n## is c, and we have a contradiction.
I see that I need to make some rewrites to make this clearer. I will upload a new version when I've fixed these problems. I will add a comment about it to this post when I'm done.

5. Good idea. I will do this.

Edit: The inequality ##\theta_c/n<\varepsilon## above doesn't make much sense. The right-hand side should be the angle that corresponds to the velocity ##\varepsilon##. See the lemma titled "The relativity is non-negative" in the latest version of the document. (Scroll down). In version 2, this is lemma 17.
 
Last edited:
Physics news on Phys.org
  • #52
Fredrik said:
Thank you for those comments. I appreciate them a lot.
Now if only one of the SAs or mentors would even try to answer my questions on the extremely rare occasions when I ask a question... (sigh).
1. There is NO proper transformation with velocity c.
Ah, but we do often analyze physical situations in a limit as ##v/c\to 1##. So this opens a new can of worms for you: how to deal more satisfactorily with these limiting situations? Currently, it seems you have no way of handling these usefully. :-)
 
  • #53
strangerep said:
Now if only one of the SAs or mentors would even try to answer my questions on the extremely rare occasions when I ask a question... (sigh).
I've been very lucky with that sort of thing. Most of my questions are math-oriented these days, and micromass have been answering pretty much all of them within an hour of me asking the question. That guy is awesome. :smile:

I don't open the relativity and QM forums to look for new posts as often as I used to, so I'm likely to miss most questions that are being asked. If you ever ask something that you think I might be able to answer, don't hesitate to send me a PM with a link to the thread.

strangerep said:
Ah, but we do often analyze physical situations in a limit as ##v/c\to 1##. So this opens a new can of worms for you: how to deal more satisfactorily with these limiting situations? Currently, it seems you have no way of handling these usefully. :-)
I don't see the problem. Do you have an example in mind?

I'm attaching version 2 of the document to this post. I will remove the old version above. The biggest change is to the lemma that rules out K<0. I also split the velocity addition corollary into two very similar corollaries just for clarity, and I made some minor changes here and there.

Edit: I have found some mistakes myself. Just before Lemma 27 (in version 2), I said "we ruled out the possibility K>0". It should of course be K<0. And in Lemma 27 (e) one of the uparrows should be a downarrow. In the unnumbered formula after (52), I have set the velocity to 0 without explaining why. I will have to do something about that.
 
Last edited:
  • #54
Fredrik said:
I don't see the problem. Do you have an example in mind?
Oh, never mind for now. If it's a problem that exists anywhere outside my vague imagination, it will re-emerge later. :-)
I'm attaching version 2 of the document to this post. I will remove the old version above. The biggest change is to the lemma that rules out K<0.
I think that lemma (17) needs a bit more work. Since you're only using a single velocity ##v##, I think you've only proven that rapidities in a certain discrete set ##\{\theta(c)/n\}## are excluded from the allowable group parameter value set. Of course, I'm sure this hole can be plugged by exploiting your original assumption that rational parameter values are dense in an open neighbourhood of 0.
 
  • #55
strangerep said:
I think that lemma (17) needs a bit more work. Since you're only using a single velocity ##v##, I think you've only proven that rapidities in a certain discrete set ##\{\theta(c)/n\}## are excluded from the allowable group parameter value set. Of course, I'm sure this hole can be plugged by exploiting your original assumption that rational parameter values are dense in an open neighbourhood of 0.

I agree that what I'm doing in lemmas 16-17 is shows that those rapidities are excluded (for proper transformations). There are no members of G that have determinant 1 and a rapidity in the set ##\{\theta(c)/n|n\in\mathbb Z^+\}##.

However, assumption 1b says that there's an ε>0 such that the interval (-ε,ε) is a subset of the range of the velocity function V. This means that for each v in that interval, there's a member of G with velocity v. Lemma 15 uses that to show that for each v in that interval, there's a member of G that has velocity v and determinant 1. This implies that for all ##\varphi\in(-\operatorname{arctan}(\varepsilon/c),\operatorname{arctan}(\varepsilon/c))##, there's a member of G that has rapidity ##\varphi## and determinant 1.

These results contradict each other, since for large enough n, we have ##\theta(c)/n\in(-\operatorname{arctan}(\varepsilon/c),\operatorname{arctan}(\varepsilon/c))##. That contradiction is what rules out K<0.

So I disagree that there's a hole in the proof, but I still consider this very valuable input, because I think this means that I need to explain the overall plan for lemmas 15-17 somewhere. I think this is what I'll do: Right after the second version of the velocity addition rule (corollary 12 in version 2), I add a comment about how it looks like we may have a division by 0 problem when K<0. (When K=0, there's clearly no problem, and we have already ruled out velocities v such that |v|≥c for the case K>0). Then I explain that I'm going to use this observation to rule out K<0, and describe the strategy for lemmas 15-17.
 
  • #56
I found a serious mistake as I was thinking about what to say after the velocity addition rule. Lemma 9 (The range of ##\Lambda_K## is closed under matrix multiplication) is wrong. When K<0, it's simply not true that the range is closed under matrix multiplication. I'm sure that the problem is fixable, but it requires a substantial rewrite.
 
Last edited:
  • #57
Fredrik said:
I found a serious mistake as I was thinking about what to say after the velocity addition rule. Lemma 9 (The range of ##\Lambda_K## is closed under matrix multiplication) is wrong. When K<0, it's simply not true that the range is closed under matrix multiplication.
Yeah, I had wondered about something similar, but I hadn't gotten around to thinking about it carefully...
 
  • #58
Unfortunately, I've not the time to dulge into this interesting thread. What I liked most as a "derivation" of the Lorentz transform is the following paper. Perhaps, you find it interesting too:

V. Berzi, V. Gorini, Reciprocity Principle and the Lorentz Transformations, Jour. Math. Phys. 10, 1518 (1969)
http://dx.doi.org/10.1063/1.1665000
 
  • #59
vanhees71 said:
Unfortunately, I've not the time to dulge into this interesting thread. What I liked most as a "derivation" of the Lorentz transform is the following paper. Perhaps, you find it interesting too:

V. Berzi, V. Gorini, Reciprocity Principle and the Lorentz Transformations, Jour. Math. Phys. 10, 1518 (1969)
http://dx.doi.org/10.1063/1.1665000
Thanks for the tip. I haven't been able to access that paper (I searched for it a few weeks ago), but the paper by Giulini that I linked to in the OP claims to be doing essentially the same thing as Berzi and Gorini. There are a few things I don't like about that approach. In particular, I think it's a bit ugly to assume that the domain of the function that takes velocities to boosts is an open ball of radius c, where c is a non-negative real number or +∞. I want the possibility of a "speed limit" to be a derived result, not one of the assumptions. Giulini also assumes that this velocity function is continuous, and uses that to make a fairly sophisticated argument based on analysis in one step. He also claims that Berzi & Gorini made an additional assumption of continuity that he didn't need to make.

I think I can avoid all of that by starting with a set of assumptions that makes better use of the principle of relativity. You could say my mathematical assumptions that are based on "principles" are stronger, and as a result, (I think) I can avoid technical assumptions and arguments based on analysis. But there are still mistakes in my pdf, so I guess I can't say that for sure yet. I'm trying to fix them now.

My pdf is about the 1+1-dimensional case, but I think that once I've gotten that right, the step to 3+1 dimensions will be much easier than the full proof of that 1+1-dimensional case. I have a pretty good idea about how to do it.

Another issue I have is with Giulini's approach is that he doesn't rigorously prove that Euclidean rotations of spacetime can be ruled out as an option. Instead of showing that they contradict his assumptions, he argues that they contradict physical common sense. To make his version of that part of the proof rigorous, we would have to make another assumption that makes that common sense precise. I think I can do this part much better.

Also, the first step in Giulini's article is incorrect. This is what we discussed on page 1. I don't know if he inherited that mistake from Berzi & Gorini or if it's one of the things he did differently.
 
Last edited:
  • #61
I'm still working on the rewrite of my pdf. That mistake I made has caused an avalanche of changes. It's super annoying. It will probably take another day or two.

In the mean time, I want to mention that I have some concerns about my assumption 2 (which says that ##\Lambda## and ##\Lambda^{-1}## have the same diagonal elements). The concern is that it may not make sense to interpret it as a mathematically precise statement of an aspect of the principle of relativity alone. In that case, it's probably a precise statement of an aspect of the combination of the principle of relativity and the idea of reflection invariance. The problem with that is that I'm defining
$$P=\begin{pmatrix}1 & 0\\ 0 & -1\end{pmatrix}$$ and want to interpret the statements ##P\in G## and ##P\notin G## respectively as "space is reflection invariant" and "space is not reflection invariant". This won't make sense if we have already made a mathematical assumption inspired by the principle of reflection invariance.

I got concerned about this when I read a comment in Berzi & Gorini (I have obtained a copy of the article) that I had already read in Giulini, but not given enough thought. What they say is this: If v is the velocity of S' in S, and v' is the velocity of S in S', then the principle of relativity doesn't justify the assumption ##v'=-v##. If the function that takes v to v' is denoted by ##\varphi##, the principle of relativity does however justify the assumptions ##\varphi(v)=v'## and ##\varphi(v')=v##, which imply that ##\varphi\circ\varphi## is the identity map. But that's it. So now they have to make some continuity assumption and use analysis to prove that the continuity assumption and the result ##\phi\circ\phi=\operatorname{id}## together imply that ##\phi(v)=-v## for all v.

I tried to think of a physical argument for why we should have v'=-v, but they all start with something like "consider two identical guns pointing in opposite directions, both fired at the same event, while moving such that the bullet fired from gun A will end up comoving with gun B".

This is definitely something I will have to think about some more. If my assumption 2 has the same problem as the assumption v'=-v (it probably does), then maybe I can still avoid reflection invariance by stating the assumptions in the context of 3+1 dimensions and using rotation invariance.
 
  • #62
Fredrik said:
So now they have to make some continuity assumption and use analysis to prove that the continuity assumption and the result ##\phi\circ\phi=\operatorname{id}## together imply that ##\phi(v)=-v## for all v.
[...]
If my assumption 2 has the same problem as the assumption v'=-v (it probably does), then maybe I can still avoid reflection invariance by stating the assumptions in the context of 3+1 dimensions and using rotation invariance.
In my 1+3D derivation (i.e., my rework of Manida's derivation), I started out with such a reciprocity assumption, just like Manida. But then I found I was able to use spatial isotropy (i.e., invariance of the transformation equations under rotation around the boost axis) to derive the desired condition. I.e., that the parameter for the inverse transformation corresponds to ##-v##.

Levy-Leblond does a similar trick (a bit less obviously) in the paper I cited earlier.

In your 1+1D derivation, I don't think you have any choice but to rely on parity invariance. But when you graduate up to 1+3D, that part of the proof can indeed be changed to use rotational invariance. I wouldn't waste too much time worrying about it in the 1+1D case.
 
  • #63
strangerep said:
In my 1+3D derivation (i.e., my rework of Manida's derivation), I started out with such a reciprocity assumption, just like Manida. But then I found I was able to use spatial isotropy (i.e., invariance of the transformation equations under rotation around the boost axis) to derive the desired condition. I.e., that the parameter for the inverse transformation corresponds to ##-v##.

Levy-Leblond does a similar trick (a bit less obviously) in the paper I cited earlier.

In your 1+1D derivation, I don't think you have any choice but to rely on parity invariance. But when you graduate up to 1+3D, that part of the proof can indeed be changed to use rotational invariance. I wouldn't waste too much time worrying about it in the 1+1D case.
That sounds good. Makes me a bit less worried.

For anyone who's interested, here's version 3 of the pdf that proves the theorem that was posted (incorrectly) in post #42 and (correctly) in post #48.

If this post doesn't have an attachment, look for a newer version in my posts below.
 

Attachments

  • #64
strangerep said:
For those who have trouble accessing behind the paywall, some related material is here:

http://books.google.com.au/books?id...8T1KCDS&dq=gorini+"rotational+invariance"&lr=
Hey, this is a great link. Thanks for finding it and posting it. I can't see all the pages, but I can see the statement of the theorem, and he makes exactly the kind of assumptions that I'm OK with. There are no weird technical assumptions about continuity, about the group being a connected Lie group, or anything like that. There's no assumption about some function that takes velocities to boosts, or anything like that. He just sets out to find all groups ##G\subset\operatorname{GL}(\mathbb R^4)## such that the subgroup ##\{\Lambda\in G|V(\Lambda^{-1})=0\}## is the set of all matrices
$$\begin{pmatrix}1 & 0 & 0 & 0\\ 0 & & &\\ 0 & & R &\\ 0 & & &\end{pmatrix},$$ with ##R\in\operatorname{SO}(3)##. His notation and statement of the theorem is kind of ugly, but that's a ******* beautiful theorem. It's a far more awesome theorem than I thought would exist, after I had read Giulini. I'm going to have to get a copy of that book somehow.
 
Last edited:
  • #65
Gorini's theorem looks so awesome that it really frustrates me that the library isn't open today. He's really making the absolute minimum of assumptions.
 
  • #66
Fredrik said:
Gorini's theorem looks so awesome that it really frustrates me that the library isn't open today. He's really making the absolute minimum of assumptions.
If, when you visit the library, you're then able to access behind paywalls, or hard copies of old journals, try typing "gorini reciprocity" into Google Scholar. It turns up some other potentially-relevant papers, including one where Gorini tries to get a better handle on what "isotropy of space" means.

[Edit: I just found out that the textbook by Sexl & Urbantke does a "nothing but relativity" derivation. I was surprised, but pleased, to find this sort of thing in a textbook.]
 
Last edited:
  • #67
strangerep said:
If, when you visit the library, you're then able to access behind paywalls, or hard copies of old journals, try typing "gorini reciprocity" into Google Scholar. It turns up some other potentially-relevant papers, including one where Gorini tries to get a better handle on what "isotropy of space" means.

[Edit: I just found out that the textbook by Sexl & Urbantke does a "nothing but relativity" derivation. I was surprised, but pleased, to find this sort of thing in a textbook.]
I went to a university library and borrowed the book. I will post some comments when I've studied the proof some more. I had read your post before I went there, but when I was there, I completely forgot to check for other articles.

I have a digital copy of Sexl & Urbantke that I haven't read. I had a quick look at their proof. It looks OK, but I didn't try to understand the details. It looked less awesome than Gorini's theorem. (It was more like the Berzi & Gorini article with "reciprocity" in the title).
 
  • #68
Some of my early thoughts on the proof, after studying only the first two lemmas in Gorini's chapter of the book...

I will use lowercase letters for numbers and 3×1 matrices, and uppercase letters for square matrices (2×2 or bigger). (See e.g. my notation for an arbitrary ##\Lambda## below). I'm still numbering my rows and columns from 0 to 3.

Let G be a subgroup of GL(ℝ4) such that
$$\big\{\Lambda\in G\,|\, \Lambda_{10}=\Lambda_{20} =\Lambda_{30}=0\big\} =\left\{\begin{pmatrix}1 & 0^T\\ 0 & R\end{pmatrix}\bigg|R\in\operatorname{SO}(3)\right\}.$$ The goal is to show, without any other assumptions, that G is the restricted Lorentz group, the group of Galilean rotations and boosts, or SO(4).

Here's the gist of the first two lemmas. Let ##\Lambda\in G## be arbitrary. I will write it as
$$\Lambda=\begin{pmatrix}a & b^T\\ c & D\end{pmatrix}.$$ Let U, U' be such that
$$U=\begin{pmatrix}1 & 0^T\\ 0 & R\end{pmatrix},\quad U'=\begin{pmatrix}1 & 0^T\\ 0 & R'\end{pmatrix},$$ where ##R,R'\in SO(3)##. Choose R such that ##Rc## is parallel to the standard basis vector ##e_1##. Let s be the real number such that ##Rc=se_1##. Choose ##R'## such that the first column of R' is parallel to the first row of RD. (This makes the other two columns of R' orthogonal to the first row of RD). Let ##\Lambda'=U\Lambda U'##, ##D'=RDR'## and ##b'=b^TR'##. We have
$$\Lambda' =U\Lambda U'=\begin{pmatrix}a & b^TR'\\ Rc & RDR'\end{pmatrix} =\begin{pmatrix}a & b_1' & b_2' & b_3 '\\ s & D'_{11} & D'_{12} & D'_{13}\\ 0 & 0 & D'_{22} & D'_{23}\\ 0 & 0 & D'_{32} & D'_{33}\end{pmatrix}.$$ So now we know that there's a member of G that has only zeros in the lower left quarter. It's easy to see that
$$0\neq \det\Lambda'=\begin{vmatrix}a & b_1'\\ s & D'_{11}\end{vmatrix}\begin{vmatrix}D'_{22} & D'_{23}\\ D'_{32} & D'_{33}\end{vmatrix}.$$
Now we want to prove that ##a\neq 0## and ##D'_{11}\neq 0##. I don't understand what Gorini is doing there. It looks wrong to me. But I think I see another way to obtain a contradiction from the assumption that one of these two variables is 0. So hopefully I have either just misunderstood something simple, or I have a way around the problem.

This is why I think what he's doing is wrong. Define
$$P=\begin{pmatrix}1 & 0 & 0 & 0\\ 0 & -1 & 0 & 0\\ 0 & 0 & -1 & 0\\ 0 & 0 & 0 & 1\end{pmatrix}.$$ Note that ##P\Lambda'^{-1}## has the same components as ##\Lambda'^{-1}##, except that the middle two rows have the opposite sign. This implies that ##\Lambda' P\Lambda'^{-1}## can differ from ##\Lambda'\Lambda'^{-1}## only in the middle two columns. (We can make a similar case for why they can only differ in the middle two rows). So the 0 column of ##\Lambda' P\Lambda'^{-1}## is the same as the 0 column of ##\Lambda'\Lambda'^{-1}=I##. In particular, ##(\Lambda'P\Lambda'^{-1})_{00}=1##. But my translation of what Gorini is saying into my notation, is that ##D'_{11}=0## implies that ##(\Lambda'P\Lambda'^{-1})_{00}=-1##.

I'm still not sure about this, but I think that one way or another, it is possible to prove that those two variables are non-zero. And I think that's very cool. When I proved my theorem for 1+1 dimensions, I had to assume that the 00 component is non-zero. (This is part of my assumption 1a). Here we seem to have the weakest possible assumptions, and we are already recovering my most basic assumption.
 
Last edited:
  • #69
Fredrik said:
In particular, ##(\Lambda'P\Lambda'^{-1})_{00}=1##. But my translation of what Gorini is saying into my notation, is that ##D'_{11}=0## implies that ##(\Lambda'P\Lambda'^{-1})_{00}=-1##.
I didn't make it clear why this bothered me. The contradiction isn't a problem, since we want to obtain a contradiction. I was thinking that my argument proves that an explicit calculation of ##(\Lambda'P\Lambda'^{-1})_{00}## can't possibly have any other result than 1. But I just did the calculation with ##D'_{11}=0## and got -1. I'm still a bit confused about what's going on here, but it will probably clear up when I work through this stuff one more time. Edit: It did. My argument about how ##\Lambda'P\Lambda'^{-1}## can differ from ##\Lambda'\Lambda'^{-1}## only in the middle is (very) wrong.
 
Last edited:
  • #70
strangerep said:
If, when you visit the library, you're then able to access behind paywalls, or hard copies of old journals, try typing "gorini reciprocity" into Google Scholar. It turns up some other potentially-relevant papers, including one where Gorini tries to get a better handle on what "isotropy of space" means.

Fredrik said:
I went to a university library and borrowed the book. I will post some comments when I've studied the proof some more. I had read your post before I went there, but when I was there, I completely forgot to check for other articles.
...and now I see why it would have been a good idea to get that article too, because a key part of lemma 3 is not proved in the book, because he wants people to read that article on isotropy.

Compared to lemmas 1-2, it was much harder to understand what lemma 3 was about. I'll write down some of my thoughts here. (This is mainly to get things straight in my own head). Consider the subgroup of G that consists of matrices of the form
$$\begin{pmatrix}A & B\\ 0 & C\end{pmatrix},$$ where A,B,C are 2×2 matrices, and det A>0. Let X be an arbitrary member of that subgroup, and write it as
$$X=\begin{pmatrix}A & B\\ 0 & C\end{pmatrix},$$ The inverse of X is
$$X^{-1}=\begin{pmatrix}A^{-1} & A^{-1}BC^{-1}\\ 0 & C^{-1}\end{pmatrix}.$$ Lemmas 1-2 tell us that the 00 components of X are both non-zero. These results simplify the formula for the velocity of X.
$$V(X)=\begin{pmatrix}X^{-1}{}_{10}/X^{-1}{}_{00}\\ X^{-1}{}_{20}/X^{-1}{}_{00}\\ X^{-1}{}_{30}/X^{-1}{}_{00}\end{pmatrix} =\begin{pmatrix}-X_{10}/X_{11}\\ 0\\ 0\end{pmatrix}.$$ So if
$$Y=\begin{pmatrix}D & E\\ 0 & F\end{pmatrix}$$ is another member of that same subgroup, and and V(X)=V(Y), we have
$$XY^{-1}=\begin{pmatrix}A & B\\ 0 & C\end{pmatrix} \begin{pmatrix}D^{-1} & D^{-1}EF^{-1}\\ 0 & F^{-1}\end{pmatrix} =\begin{pmatrix}AD^{-1} & AD^{-1}EF^{-1}\\ 0 & CF^{-1}\end{pmatrix}.$$ \begin{align}(XY^{-1})_{10} &=(AD^{-1})_{10} =\begin{pmatrix}A_{10} & A_{11}\end{pmatrix} \frac{1}{\det D}\begin{pmatrix}D_{11} \\ -D_{10}\end{pmatrix} =\frac{1}{\det D}\big(A_{10}D_{11}-A_{11}D_{10}\big)\\ &\frac{A_{11}D_{11}}{\det D}\left( \frac{A_{10}}{A_{11}} -\frac{D_{10}}{D_{11}}\right) =\frac{1}{\det D}\big(V(X)-V(Y)\big)=0\\
(XY^{-1})_{20} &= 0\\
(XY^{-1})_{30} &=0.\end{align} The theorem's main assumption is that all the transformations with the i0 components =0 are rotations. So this means that for some ##R\in\operatorname{SO}(3)##,
$$XY^{-1}=\begin{pmatrix}1 & 0^T\\ 0 & R\end{pmatrix}.$$ Denote the right-hand side by U. We have X=UY. Since ##U_{1i}=0## and ##Y_{21}=Y_{31}=0##, this implies that
\begin{align}0 &=X_{21}=U_{2\mu}Y_{\mu 1} =U_{2i}Y_{i 1} =R_{21}Y_{11}\\ 0&=X_{31}=U_{3\mu}Y_{\mu 1} =U_{3i}Y_{i 1} =R_{31}Y_{11}.\end{align} Since ##Y_{11}\neq 0##, this implies that ##R_{21}=R_{31}=0##. This implies that ##R_{11}=\pm 1##. The negative sign can be ruled out (it has something to do with determinants that's not clear in my head right now). So U is actually of the form
$$\begin{pmatrix}I & 0\\ 0 & R'\end{pmatrix}$$ where ##R'\in\operatorname{SO}(2)##. This implies that
$$X=YU=\begin{pmatrix}D & E\\ 0 & R'F\end{pmatrix}.$$ This is a pretty significant result. It implies that A=D, B=E and ##C=R'F##. So transformations of this "block upper diagonal" form are almost completely determined by the velocity. The upper left and upper right are completely determined, and the lower right is determined up to multiplication by a member of SO(2). This implies that for all
$$U(R)=\begin{pmatrix}I & 0\\ 0 & R\end{pmatrix}$$ with R in SO(2), ##V(U(R)X)=V(X)##. This implies that there's an R' in SO(2) such that ##U(R)X=XU(R')##. This is where he refers to "reference 12" for the proof that this implies that B=0, that C is diagonal, and that there are some additional constraints on A and C. Edit: I'm thinking about how to do this now, and it looks like this might be easy. The argument is similar to the things I said about Giulini's article on page 1.
 
Last edited:
  • #71
If X and Y are members of the subgroup that consists of all the matrices of the form
$$\begin{pmatrix}A & B\\ 0 & C\end{pmatrix}$$ with det A>0, then there's an R in SO(3) such that ##XY^{-1}=U(R)##, where the right-hand side is defined by
$$U(R)=\begin{pmatrix}I & 0\\ 0 & R\end{pmatrix}.$$ This implies that if
$$X=\begin{pmatrix}A & B\\ 0 & C\end{pmatrix},\quad Y=\begin{pmatrix}D & E\\ 0 & F\end{pmatrix},\quad V(X)=V(Y),$$ we have
$$X=U(R)Y=\begin{pmatrix}D & E\\ 0 & RF\end{pmatrix}.$$ So two members of this subgroup with the same velocity differ only in the lower right, and there they differ only by multiplication of a member of SO(2).

Since for all R, ##V(U(R)X)=V(X)=V(XU(R))##, this implies that the following statements are true:
  1. For all R in SO(2), we have BR=B.
  2. For all R in SO(2), there's an R' in SO(2) such that ##RC=CR'##.
  3. For all R' in SO(2), there's an R in SO(2) such that ##RC=CR'##.
The first one implies that B=0. (Choose R to be a rotation by π/2, and the rest is obvious). The results 2 and 3 imply that C is a number times a member of SO(2). I don't see a way to prove that C is diagonal, so I think Gorini has made essentially the same mistake as Giulini. The proof that C is a number times a member of SO(2) is a bit trickier than the corresponding proof in the OP, since now only one of the rotation matrices (R or R') is arbitrary. It's very convenient to choose the arbitrary SO(2) matrix to be a rotation by π/2. To see if the other SO(2) matrix exists at all, we start with the following observations. An SO(2) matrix acting on a 2×2 matrix from the left doesn't change the norm of the columns (viewed as members of ℝ2). An SO(2) matrix acting on a 2×2 matrix from the right doesn't change the norm of the rows. These observations and the results 2-3 above imply that ##a^2+c^2=b^2+d^2## and ##a^2+b^2=c^2+d^2##. These results imply that ##a^2=d^2## and ##b^2=c^2##. So C is of the form
$$\begin{pmatrix}a & \pm b\\ b & \pm a\end{pmatrix}$$ and it turns out that three of the four possible sign combinations can be ruled out by the observation that an SO(2) matrix acting from the left on a 2×2 matrix doesn't change the inner product of the columns. So the final result is that there exist numbers a,b such that
$$C=\begin{pmatrix}a & -b\\ b & a\end{pmatrix}.$$ The columns (and the rows) are orthogonal and have the same norm. So if we define ##k=\sqrt{a^2+b^2}##, there's an R in SO(2) such that C=kR. This means that the X that we started with is of the form
$$\begin{pmatrix}A & 0\\ 0 & kR\end{pmatrix}$$ where k is a real number and R is a member of SO(2). I don't see a way to prove that k=0 right now, but I have only just started to think about it.

It seems impossible to me to prove that the group contains a member with velocity v for each v with |v|<c. If I'm right, it's a pretty big flaw, and the theorem would have be repaired by adding an assumption like my "0 is an interior point of V(G)". Then we would have to go through the same sort of stuff I did in my pdf for the 1+1-dimensional case.
 
Last edited:
  • #72
It's a bit difficult for me to follow along properly, since I don't have complete copies of all the papers, and won't be on-campus again for a while. So you might need to include a bit more context in your posts.

As an aside (or maybe a large tangent?) I'll just make the general remark that I think part of the difficulty is that you're still approaching all this as a "geometrical" problem, instead of a "dynamical symmetry" problem. (I had no great difficulty reaching a equivalent point to what these authors reach.)

In the (simplest case of) a "dynamical symmetry" approach, one assumes an independent variable ##t## (time), and a dependent variable ##q = q(t)## (position), and the equation of motion ##\Delta \equiv \ddot q = 0## . In the general theory of symmetries of (systems of) differential equations, one works in a larger space in a larger space, being a Cartesian product of spaces of all variables and the partial derivatives of the dependent variables. The condition ##\Delta=0## then specifies a solution variety within that space.

One then considers a Lie group ##G## acting on both the dependent and independent variables, and the so-called higher prolongation(s) [*ref 1] of ##G## acting in a space thus augmented by Cartesian product with spaces of various derivatives of the dependent variable(s). The idea is to find the most general transformation of the larger space such that the variety ##\Delta=0## is mapped into itself. There are reasonably straightforward formulas for the 1st and 2nd prolongations of the Lie algebra of ##G##, and the symmetries can thus be found in a couple of pages. (I now know that this is actually easier than all the previous ways we've discussed about obtaining the straight-line-preserving maps.)

My point here is that velocity, i.e., ##\dot q##, is an integral part of this whole approach, rather than an afterthought, and one can also apply the 1st prolongation to find out how velocity is involved in the transformations. Since the basic variables are continuous and differentiable, so is the velocity, at least piece-wise.

Anyway, maybe this was indeed just me running off on a tangent.

*Ref 1: P. J. Olver, "Applications of Lie Groups to Differential Equations", 2nd Ed.
 
  • #73
I don't know... The approach you're describing sounds a lot less appealing to me. One of the reasons is that it involves so many technical terms that even if I would find it easier to learn, I still wouldn't be able to explain it to a physics student who has studied linear algebra and special relativity, or even to some physics PhD's. But I'm still curious enough that I might take a look at that approach when I'm done with this one. (If nothing else, I'll at least find out what equations of motion have to do with this).

If it really is possible to do what Gorini claims to be able to do (I will know when I've worked through the rest of the proof), then it seems to be the ultimate theorem of this type. Gorini's theorem says that if G is a subgroup of GL(ℝ4) such that the subgroup of G that takes the 0 axis to itself is equal to
$$\left\{\begin{pmatrix}1 & 0^T\\ 0 & R\end{pmatrix}\bigg| \,R\in\operatorname{SO}(3)\right\}$$ (where 0 is a 3×1 matrix), then G is either the group of Galilean boosts, the group of Lorentz boosts with some invariant speed ##c\in(0,\infty)##, or the group of rotations that I just mentioned. (Unfortunately he says this in a rather complicated way).

It's understandable that you're having difficulties following my notes on lemmas 1-3 above. They're not as well thought out or as detailed as the stuff in my pdf. They are more like a "version 0.1" of a new document. I need to make notes of what I find somewhere, that I can later develop into a "version 1", and I figured I might as well make them here. If you find them useful, that's a bonus. If not, I'm not offended or anything. :smile: This also applies to what I'm saying in this post. If you too are interested in understanding Gorini's proof, I'll be happy to answer questions on the parts of it I understand so far.

Here's a summary of my thoughts on lemmas 1-3. First a comment about velocities. For all ##\Lambda\in G## such that ##(\Lambda^{-1})_{00}\neq 0##, I will call the 3×1 matrix with components ##(\Lambda^{-1})_{i0}/(\Lambda^{-1})_{00}## the velocity of ##\Lambda##. (If ##\Lambda## changes coordinates from a system S to a system S', then this is the velocity of S' in S). For all ##\Lambda\in G## such that ##\Lambda_{00}\neq 0##, I will call the 3×1 matrix with components ##\Lambda_{i0}/\Lambda_{00}## the reciprocal velocity of ##\Lambda##. (This is the velocity of S in S').

Let ##\Lambda\in G## be arbitrary and consider a transformation
$$\Lambda\mapsto U(R)\Lambda U(R')$$ where ##U:\operatorname{SO}(3)\to \operatorname{GL}(\mathbb R^4)## is defined by
$$U(R)=\begin{pmatrix}1 & 0^T\\ 0 & R\end{pmatrix}$$ for all R in SO(3). (This 0 is a 3×1 matrix). Denote ##U(R)\Lambda U(R')## by ##\Lambda'##. It turns out that there's a clever choice of R and R' that ensures that ##\Lambda'## is of the form
$$\begin{pmatrix}A & B\\ 0 & C\end{pmatrix},$$ where A,B,C are 2×2 matrices such that ##\det A>0##. When we prove this, we also see that ##\Lambda_{00}\neq 0## and ##\Lambda'_{11}\neq 0##. Since the inverse of the matrix above is
$$\begin{pmatrix}A^{-1} & A^{-1}B C^{-1}\\ 0 & C^{-1}\end{pmatrix}$$
and we have
$$A^{-1}=\frac{1}{\det A}\begin{pmatrix}A_{11} & -A_{01}\\ -A_{10} & A_{00}\end{pmatrix},$$
the velocity of ##\Lambda'## is
$$\begin{pmatrix}(\Lambda'^{-1})_{10}/(\Lambda'^{-1})_{00}\\ (\Lambda'^{-1})_{20}/(\Lambda'^{-1})_{00}\\ (\Lambda'^{-1})_{30}/(\Lambda'^{-1})_{00}\end{pmatrix} =\begin{pmatrix}-\Lambda'_{10}/\Lambda'_{11}\\ 0 \\ 0\end{pmatrix}.$$
So every member of G has a well-defined reciprocal velocity, and every member of G that has the special form above has a well-defined velocity, and it's given by a simple formula: ##V(\Lambda)=-\Lambda_{10}/\Lambda_{11}.##.

Now let X,Y be two arbitrary members of G that have the special form above, and also have the same velocity. Then we can prove that ##XY^{-1}## has velocity 0, and must therefore (by the only assumption we made about G) be equal to U(R) for some R in SO(3). Further, we can show that ##(XY^{-1})_{11}=1##. This implies that the R is a rotation around the 1 axis. So if we define ##T:SO(2)\to\operatorname{GL}(\mathbb R^4)## by
$$T(R)=\begin{pmatrix}I & 0\\ 0 & R\end{pmatrix}$$ for all R in SO(2), then we have ##XY^{-1}=T(R)## for some R in SO(2). This implies that X=T(R)Y, and this tells us that X and Y can only differ in the lower right 2×2 corner. In that corner they differ at most by a factor of R. This means that the top two rows of ##\Lambda'## are completely determined by the velocity ##v=-\Lambda_{10}/\Lambda_{11}##.

Since ##T(R)\Lambda'## and ##\Lambda' T(R)## have the same velocity as ##\Lambda'## for all R in SO(2), these three matrices can only differ in the lower right. We can use this to show that the upper right 2×2 corner of ##\Lambda'## is 0.

Now Gorini claims that these results imply that there's a R in SO(2) such that ##\Lambda' T(R)## is of the form
$$\begin{pmatrix}d(v) & c(v) & 0 & 0\\ -va(v) & a(v) & 0 & 0\\ 0 & 0 & e(v) & 0\\ 0 & 0 & 0 & f(v)\end{pmatrix}$$ where e(v) is positive and f(v)=±e(v). He doesn't prove this in the book. He just claims that reference 12 (the article about isotropy) proves it.

I actually got the result that e(v)=f(v). I will have to check my calculations to see if this is a mistake. Edit: I did the calculation again. This time my result agrees with Gorini's. So in spite of the suspicions I've had, I now think that there are no mistakes in lemmas 1-3.

Now at this point, it would be incredibly awesome if we could somehow prove a result like that a transformation of this form has the same diagonal elements as its inverse. Then we can proceed as I did in the 1+1-dimensional case. But Gorini doesn't do anything like that. Instead he starts a long and tedious-looking argument that I still haven't studied. I'm sort of hoping that it can be avoided, but even if it can, I probably won't see how to avoid it until I've studied the argument. So I will have to do that. But maybe not today.
 
Last edited:
  • #74
I started looking at the main body of the proof today. It's really beautiful and really ugly at the same time. The beauty is in the following idea. Denote the last 4×4 matrix in my previous post by N(v). Its velocity is in the 1 direction. There's no way to get a transformation with a velocity that has a magnitude different from |v| just by applying rotation operators to N(v), but if Q is a rotation around the 3 axis, then ##N(v)^{-1}QN(v)## is a transformation with a velocity that's not in the 1 direction. Now we can find two rotations U and U' such that the lower left corner of ##UN(v)^{-1}QN(v)U'## are all zeroes. Then the upper right corner is automatically all zeroes. Finally, we can multiply this from the right with a rotation around the 1 axis, to get a result that we can denote by N(w), where N is the same function as before, but w is a velocity in the 1 direction that may be different from v.

The really ugly part is that this is a crazy exercise in matrix multiplication. Having to calculate ##N(v)^{-1}QN(v)## is annoying enough (and the result is ugly), but then we have to find the appropriate U and U' (both ugly), and compute ##UN(v)^{-1}QN(v)U'##.

This way Gorini finds the result ##N(w)^{-1}=N(-w)##, as well as a lot of other details about the components of N(v). I'm still hoping that I can find some sort of shortcut, but I'm perhaps being naive.
 
Last edited:
  • #75
I'm still working on this. I've been trying to find ways to simplify the proof, but I have so far failed miserably at that. I can make some statements clearer, but that's it.

We want to find all groups ##G\subset\mathrm{GL}(\mathbb R^4)## such that the set of all ##\Lambda\in G## that take points on the 0 axis to points on the 0 axis is the rotation subgroup
$$\left\{\left.\begin{pmatrix}1 & 0^T\\ 0 & R\end{pmatrix}\right|R\in\mathrm{SO}(3)\right\},$$ where 0 denotes the 3×1 matrix with all components zero.

It's useful to define the following notations:
$$\begin{align}U(R) &=\begin{pmatrix}1 & 0^T\\ 0 & R\end{pmatrix}\\
T(Q)&=\begin{pmatrix}I & 0\\ 0 & Q\end{pmatrix}\\
F(t)&=\begin{pmatrix}1 & 0 & 0 & 0\\ 0 & \cos t & -\sin t & 0\\ 0 & \sin t & \cos t & 0\\ 0 & 0 & 0 & 1\end{pmatrix},\end{align}$$ where I and Q (and the zeroes next to them) are 2×2 matrices. I denotes the identity matrix.

Let's write an arbitrary ##\Lambda\in G## as
$$\Lambda=\begin{pmatrix}a & b^T\\ c & D\end{pmatrix},$$ where b,c are 3×1 matrices and D is a 3×3 matrix. Let ##R,R'\in\mathrm{SO}(3)##. Define ##\Lambda'=U(R)\Lambda U(R')##. We have
$$\Lambda'=\begin{pmatrix}a & b^T R'\\ Rc & RDR'\end{pmatrix}.$$ If we choose the first row of R parallel to c, and the first column of R' orthogonal to the second and third rows of RD, then the 20,30,21,31 components of ##\Lambda'## will all be zero. This result is lemma 1.

The other lemmas show that if ##\Lambda## is such that the 20,30,21,31 components (the lower left) are all zero, then so are the 02,03,12,13 components (the upper right). Further, the upper left is fully determined by the velocity, and the lower right is determined up to multiplication by an O(2) matrix. We can use this to show that there's a unique k>0, a unique 2×2 matrix A, and a unique O(2) matrix Q, all determined by the velocity of ##\Lambda##, such that
$$\Lambda=\begin{pmatrix}A & 0\\ 0 & kQ\end{pmatrix}.$$ Now, if we multiply this from the right by T(R), where R is either equal to Q (if det Q=1) or equal to Q with the sign of the second row flipped (if det Q=-1), then we get a matrix with an even simpler form, like the N I will mention below.

The strategy in the main body of the proof is this: If there's a ##\Lambda\in G## that isn't a rotation, then there's also an ##N\in G## of the form
$$N=\begin{pmatrix}d & c & 0 & 0\\ -va & a & 0 & 0\\ 0 & 0 & e & 0\\ 0 & 0 & 0 & f\end{pmatrix},$$ where v≠0, e>0, f=±e, and a,d,c,e,f are all fully determined by v. (The point of the lemmas is to show this. I'm sure I understand this part well enough).

Now we note that for all t, the matrix ##N^{-1}F(t)N## can be brought to this pretty form by a transformation ##X\mapsto U(R)XU(R')##. So let's write ##N'=U(R)N^{-1}F(t)NU(R')##, where ##R,R'\in\mathrm{SO}(3)## are chosen only to ensure that we get the simple form above. Somewhere in this calculation of N', a miracle occurs, and we end up with an even simpler form.
$$N'=\begin{pmatrix}a' & c' & 0 & 0\\ -wa' & a' & 0 & 0\\ 0 & 0 & 1 & 0\\ 0 & 0 & 0 & 1\end{pmatrix}.$$ The value of w depends on the angle t, and Gorini claims that the possible values of w form a closed interval. (This is why he doesn't need an assumption like my 1b). Now he starts deriving a bunch of results about these especially simple matrices. v may not be in that closed interval, but there's an ##N'## with velocity w in that interval such that ##(N')^n=N## for some positive integer n. At least I think that's what he's saying. I'm still not clear on all the details at the end.

I've been thinking that if I can understand why that "miracle" occurs (i.e. why N' has an even simpler form than N), then maybe I can use that insight to simplify the proof considerably. But I still don't see what's really going on there.

I have typed up my version of the lemmas. I guess I'll start typing up the main proof as well.
 
Last edited:
  • #76
Any thoughts on what "rotational" invariance really means in this approach? The title of Gorini's article is "Derivation of the Lorentz and Galilei groups from rotational invariance". But the assumption in his theorem is that the zero-velocity subgroup is equal the rotation subgroup. That's hardly the condition that best matches our intuition about what rotation invariance of space means. It may however be the best match to our intuition about "rotation invariance, and no kind of reflection invariance".

I'm thinking that a statement that we choose to think of as a mathematically precise statement of rotation invariance should imply this:

(1) The set
$$\left\{\left.\begin{pmatrix}1 & 0^T\\
0 & R\end{pmatrix}\right|R\in\mathrm{SO}(3)\right\},$$ where 0 denotes a 3×1 matrix whose components are all zeroes, is a subgroup of G.

This is a weaker statement than Gorini's. I think it's too weak to imply anything like his result. But I also think that there's more to rotation invariance than this. The velocity of a boost singles out a direction in space, but a zero-velocity transformation doesn't do anything like that. So a mathematically precise statement of rotation invariance should also imply things like this:

(2) For all ##\Lambda\in G## with zero velocity, the projection of ##\Lambda x## onto the 0 axis should have the same value for all ##x\in\mathbb R^4## such that ##x_0=0## and ##x_1^2+x_2^2+x_3^2=1##.

If we assume (2), then all transformations of the form
\begin{pmatrix}* & * & * & *\\ 0 & * & * & *\\ 0 & * & * & *\\ 0 & * & * & *\end{pmatrix} are actually of the form \begin{pmatrix}* & 0 & 0 & 0\\ 0 & * & * & *\\ 0 & * & * & *\\ 0 & * & * & *\end{pmatrix} This takes us a long way toward Gorini's assumption. I think Gorini took things a bit too far, but I think that we should be able to conclude that transformations of the form above are actually of the form
\begin{pmatrix}a & 0^T\\ 0 & Q\end{pmatrix} where ##a\neq 0## and ##Q\in\mathrm{O}(3)##. Unfortunately I don't see how to get there with an assumption like (2), which everyone would agree is an aspect of rotation invariance.


Edit: It makes sense to require that an arbitrary ##\Lambda\in G## with zero velocity takes the unit sphere to a sphere. (If it takes the unit sphere to some other shape, then some directions in space are different from others). But why does it have to have the same radius? I don't see why the inclusion of a transformation of the form
\begin{pmatrix}a & 0^T\\ 0 & bQ\end{pmatrix} in the group should be thought of as inconsistent with rotation invariance. ##a,b\neq 0, Q\in\mathrm{O}(3)##.

Edit 2: Nevermind. I figured out the answers. The statements (1), (2), and the requirement that zero-velocity transformations take spheres to spheres, can together be thought of as "rotation invariance". Transformations like
\begin{pmatrix}a & 0^T\\ 0 & bQ\end{pmatrix} with Q in O(2) and a,b not both equal to 1, aren't ruled out by rotation invariance. They are ruled out by "dilation non-invariance", something that's even more intuitive than rotation invariance; it's obvious that if a transformation e.g. changes the length of a meter stick, length measurements will not have the same results as before.

Gorini doesn't derive these groups "from rotation invariance". He derives them from rotation invariance, dilation non-invariance, spatial reflection non-invariance and time reversal non-invariance. I think the proof would work with only minor modifications if we drop the last two assumptions, but we're going to have to keep the dilation non-invariance.
 
Last edited:
  • #77
Dunno whether you've already got this earlier Gorini paper (attached hopefully). Still can't find the "Isotropy of Space" paper though.
 

Attachments

  • #78
Thank you. I already had it though. I found a pdf version of it at this URL a couple of weeks ago: http://physics.sharif.ir/~sperel/paper1.pdf. I haven't had a chance to look at the isotropy paper yet. I didn't find a pdf of that one.
 
Last edited:
  • #79
I've been trying to generalize Gorini's theorem, and I'm stuck on a silly-looking detail, so figured I might as well ask if you see something I don't.

Gorini's assumption is that the zero-velocity subgroup is equal to the rotation subgroup. Since this only gives us the proper and orthochronous groups, I want to weaken the assumption. I'm not 100% sure what the appropriate weaker assumption should be, but I suspect that it's this one: The zero-velocity subgroup has the rotations as a subgroup, and is itself a subgroup of the group
$$\left\{\begin{pmatrix}\sigma & 0^T\\ 0 & Q\end{pmatrix}\,\bigg|\,\sigma\in\{-1,1\}, Q\in\mathrm{O}(3)\right\}$$ (where 0 denotes a 3×1 matrix).

Because of this weakening, I quickly run into a problem. Let ##\Lambda## be an arbitrary member of the group, with only zeroes in the lower left corner (i.e. the 20,30,21,31 components). Gorini proved that the 00 and 11 components of such a ##\Lambda## must be non-zero, so this is what I would like to do. Let F be a rotation by ##\pi## around the 3 axis:
$$F=\begin{pmatrix}1 & 0 & 0 & 0\\ 0 & -1 & 0 & 0\\ 0 & 0 & -1 & 0\\ 0 & 0 & 0 & 1\end{pmatrix}.$$ Now I can show e.g. that if ##\Lambda_{11}=0##, then ##\Lambda F\Lambda^{-1}## is a zero-velocity transformation with its 00 component equal to -1. With Gorini's assumption, this is clearly a contradiction that proves that ##\Lambda_{11}\neq 0## (because every zero-velcocity transformation is a rotation and the 00 component of a rotation is 1). But with my weaker assumption, I don't see how ##(\Lambda F\Lambda^{-1})_{00}=-1## contradicts anything.

I may have to modify the assumption to ensure that this results contradicts it. But I want the assumptions to be mathematical statements that can be thought of as making aspects of the concept of "rotation invariance" mathematically precise. So the question is, what aspect of rotation invariance is violated by this result?

Note that ##\Lambda## having all zeroes in the lower left corner means that it's going to turn out to be a boost in the 1-direction. So in physical terms, what this result is saying is that if I stand on a treadmill, speed it up so I have to run, then turn around and run backwards for a while, and then stop the treadmill, my clock is know running backwards! But to make this argument, I have to anticipate the result of the theorem. At the point where I'm making the calculation, I'm not sure we know enough about the group to even define what an orthochronous transformation is.
 
  • #80
Have you been able to obtain a copy of Gorini's "Isotropy of Space" paper? I.e., J. Math. Phys. 11, 2226 (1970) ?
 
  • #81
No, I haven't. I guess I should make another trip to the library to get that. I didn't need it to understand the part of the proof that he said could be found in the isotropy article, so I wasn't very motivated to go back to the library just for that. Now that want to generalize the theorem, I seem to need a deeper understanding of what rotation invariance entails than I currently have. Maybe I can get that from the isotropy article, maybe not.
 
  • #82
Yeah -- I just figured it would be nice to see what Gorini actually says. (I haven't been on-campus for a while either.)

As for what rotation invariance (or spatial isotropy) means, I think of it as follows.

Suppose you have an expression ##F(t,x)## (which may involve derivatives). Then the equation $$F(t,x) ~=~ 0$$ (in general, on some domain) is said to be rotationally invariant if $$F(t, Rx) ~=~ 0$$ on the same domain (where ##R## is an arbitrary spatial rotation).
 
  • #83
Maybe you should just try to contact him direct:
http://www.uninsubria.eu/research/physmath/cv_Gorini.htm

Perhaps he'll send you a copy... :-)
 
  • #84
The thing is, I need to translate "rotation invariance" into a set of conditions on the group G that we're trying to find. An obvious condition is that the group of spatial rotations is a subgroup of G. This corresponds to saying that no matter which way we rotate the laboratory, there's an inertial coordinate system with its spatial axes aligned with the laboratory walls.

A less obvious condition is that zero-velocity transformations must preserve simultaneity. The reason is, a linear transformation either preserves simultaneity or "tilts" the simultaneity hyperplanes. Such a "tilt" would make one direction in space "special". (It's easy to visualize this in a 2+1-dimensional spacetime diagram). This is OK when we're dealing with a transformation with a non-zero velocity in that special direction. But when we're dealing with a zero-velocity transformation, the transformation doesn't single out a direction, so if space itself doesn't have a preferred direction, the simultaneity plane can't be tilted. (In what direction would it tilt?)

The consequence of this is that a zero-velocity transformation (which by definition has three zeroes in column 0) must have three zeroes in row 0. This is easy to see. Just note that the "t=0" hyperplane is preserved by ##\Lambda## only if the 0 component of
$$\Lambda\begin{pmatrix}0\\ x\\ y\\ z\end{pmatrix}$$ is 0 for all ##x,y,z\in\mathbb R##. This implies that ##\Lambda_{01}=\Lambda_{02}=\Lambda_{03}=0##. Because of this, we choose the theorem's assumptions such that every ##\Lambda\in G## with ##\Lambda_{10}=\Lambda_{20}=\Lambda_{30}=0## also satisfies ##\Lambda_{01}=\Lambda_{02}=\Lambda_{03}=0##.

I suspect that I will just have to think of more arguments of this sort, until I find one that takes care of the problem with time reversal that I described.
 
  • #85
Fredrik said:
A less obvious condition is that zero-velocity transformations must preserve simultaneity.
But how do you know that a priori if you're starting from the relativity principle alone, and trying to derive the relativity group(s)? You can't appeal to geometric intuitions from Minkowski spacetime since the latter is really only an afterthought -- a homogeneous space constructed from a given relativity group.

(BTW, your term "zero-velocity transformation" is a bit misleading. I think "velocity-preserving transformation" is clearer.)
 
  • #86
strangerep said:
But how do you know that a priori if you're starting from the relativity principle alone, and trying to derive the relativity group(s)? You can't appeal to geometric intuitions from Minkowski spacetime since the latter is really only an afterthought -- a homogeneous space constructed from a given relativity group.
I think I explained how I know that, but feel free to ask about the details if I need to clarify something. Note that I didn't use the Minkowski metric. I just used what I know about linear transformations. There isn't a whole lot of things that a linear transformation can do to a simultaneity hyperplane. It can preserve it, or it can tilt it. If a transformation tilts a simultaneity hyperplane, that always favors a direction in space: the direction of the tilt.

A linear transformation can also stretch or rotate a simultaneity hyperplane, but the preservation of simultaneity depends only on whether the hyperplane gets tilted or not.

strangerep said:
(BTW, your term "zero-velocity transformation" is a bit misleading. I think "velocity-preserving transformation" is clearer.)
I define the velocity of a transformation ##\Lambda## as the vector with components ##(\Lambda^{-1})_{i0}/(\Lambda^{-1})_{00}##. I think this terminology is appropriate. The velocity of the transformation is the velocity of the second observer in the coordinates of the first.
 
Back
Top