Series expansion of the Lorentz Transformation

In summary: The last line is just$$(1_{2\times2}+i \sigma_2\epsilon)^{\mu}_{ \ \ \ \nu}$$Thus we have$$\eta_{\sigma \mu}R(\epsilon^{\mu}_{\ \ \nu})=(1_{2\times2}+i \sigma_2\epsilon)_{\sigma \
  • #1
JD_PM
1,131
158
Homework Statement
Proper Lorentz Transformations' expansion is given to be

$$\Lambda^{\mu}_{ \ \ \nu}(\epsilon)=(\exp(\epsilon))^{\mu}_{ \ \ \nu} = \delta^{\mu}_{\nu} + \epsilon^{\mu}_{ \ \ \nu} +\frac{1}{2!} \epsilon^{\mu}_{ \ \ \rho}\epsilon^{\rho}_{ \ \ \nu}+ \ ..., \tag1$$

Where

$$\epsilon_{\mu \nu}=-\epsilon_{\nu \mu} \tag2$$

a) Show that the Lorentz Transformation condition ##\eta_{\mu\nu}\Lambda^{\mu}{}_{\rho} (\epsilon)\Lambda^{\nu}{}_{\sigma}(\epsilon) = \eta_{\rho \sigma}## holds for ##(1)##

b) Show (by working out the matrix exponential explicitly) that only when ##\epsilon_{12}## does not vanish we get a rotation around the 3-axis over an angle ##\epsilon_{12}##

c) Show that only when ##\epsilon_{03}## does not vanish we get a boost along the 3-axis with rapidity ##\epsilon_{03}##
Relevant Equations
$$\Lambda^{\mu}_{ \ \ \nu}(\epsilon)=(\exp(\epsilon))^{\mu}_{ \ \ \nu} = \delta^{\mu}_{\nu} + \epsilon^{\mu}_{ \ \ \nu} +\frac{1}{2!} \epsilon^{\mu}_{ \ \ \rho}\epsilon^{\rho}_{ \ \ \nu}+ \ ..., \tag1$$
a) I think I got this one (I have to thank samalkhaiat and PeroK for helping me with the training in LTs :) )

$$\eta_{\mu\nu}\Big(\delta^{\mu}_{\rho} + \epsilon^{\mu}_{ \ \ \rho} +\frac{1}{2!} \epsilon^{\mu}_{ \ \ \lambda}\epsilon^{\lambda}_{ \ \ \rho}+ \ ...\Big)\Big(\delta^{\nu}_{\sigma} + \epsilon^{\nu}_{ \ \ \sigma} +\frac{1}{2!} \epsilon^{\nu}_{ \ \ \alpha}\epsilon^{\alpha}_{ \ \ \sigma}+ \ ...\Big)=$$

$$=\eta_{\mu \nu} \delta^{\mu}_{\rho}\delta^{\nu}_{\sigma}+\eta_{\mu\nu} \epsilon^{\mu}_{ \ \ \rho}\epsilon^{\nu}_{ \ \ \sigma}+\frac 1 4 \eta_{\mu \nu}\epsilon^{\mu}_{ \ \ \lambda}\epsilon^{\lambda}_{ \ \ \rho} \epsilon^{\nu}_{ \ \ \alpha}\epsilon^{\alpha}_{ \ \ \sigma} +\eta_{\mu \nu}\delta^{\mu}_{\rho}\epsilon^{\nu}_{ \ \ \sigma}+\frac{1}{2!}\eta_{\mu\nu}\delta^{\mu}_{\rho}\epsilon^{\nu}_{ \ \ \alpha}\epsilon^{\alpha}_{ \ \ \sigma}+\eta_{\mu\nu}\epsilon^{\mu}_{ \ \ \rho}\delta^{\nu}_{\sigma}$$

$$+\frac{1}{2!}\eta_{\mu\nu}\epsilon^{\mu}_{ \ \ \rho}\epsilon^{\nu}_{ \ \ \alpha}\epsilon^{\alpha}_{ \ \ \sigma} + \frac{1}{2!} \eta_{\mu\nu} \epsilon^{\mu}_{ \ \ \lambda}\epsilon^{\lambda}_{ \ \ \rho}\delta^{\nu}_{\sigma}+\frac{1}{2!} \eta_{\mu\nu} \epsilon^{\mu}_{ \ \ \lambda}\epsilon^{\lambda}_{ \ \ \rho}\epsilon^{\nu}_{ \ \ \sigma} + \ ... =$$

$$ = \eta_{\mu \nu} \delta^{\mu}_{\rho}\delta^{\nu}_{\sigma}+\eta_{\mu\nu} \epsilon^{\mu}_{ \ \ \rho}\epsilon^{\nu}_{ \ \ \sigma}+\eta_{\mu \nu}\delta^{\mu}_{\rho}\epsilon^{\nu}_{ \ \ \sigma}+\frac{1}{2!}\eta_{\mu\nu}\delta^{\mu}_{\rho}\epsilon^{\nu}_{ \ \ \alpha}\epsilon^{\alpha}_{ \ \ \sigma}+\eta_{\mu\nu}\epsilon^{\mu}_{ \ \ \rho}\delta^{\nu}_{\sigma}+\frac{1}{2!} \eta_{\mu\nu} \epsilon^{\mu}_{ \ \ \lambda}\epsilon^{\lambda}_{ \ \ \rho}\delta^{\nu}_{\sigma}+\mathcal O(\epsilon^3)=$$ $$=\eta_{\rho \sigma} + \epsilon_{\nu \rho}\epsilon^{\nu}_{ \ \sigma}+ \epsilon_{\rho \sigma}+\frac{1}{2!}\epsilon_{\rho \alpha}\epsilon^{\alpha}_{ \ \sigma}+\epsilon_{\sigma \rho}+\frac{1}{2!}\epsilon_{\sigma \lambda}\epsilon^{\lambda}_{ \ \rho} + \mathcal O(\epsilon^3)$$

We now massage one of the terms to get

$$\epsilon_{\sigma \lambda}\epsilon^{\lambda}_{ \ \rho}=\epsilon^{\lambda}_{ \ \rho}\epsilon_{\sigma \lambda}=-\epsilon^{\lambda}_{ \ \rho}\epsilon_{\lambda \sigma}=\epsilon_{\rho \lambda}\epsilon^{\lambda}_{ \ \sigma}$$

Neglecting order three and higher we get

$$\eta_{\mu\nu}\Lambda^{\mu}{}_{\rho} (\epsilon)\Lambda^{\nu}{}_{\sigma}(\epsilon)=\eta_{\rho \sigma} - \epsilon_{\rho \beta}\epsilon^{\beta}_{ \ \sigma}+ \epsilon_{\rho \sigma}+\frac{1}{2!}\epsilon_{\rho \beta}\epsilon^{\beta}_{ \ \sigma}-\epsilon_{\rho \sigma}+\frac{1}{2!}\epsilon_{\rho \beta}\epsilon^{\beta}_{ \ \sigma}=\eta_{\rho \sigma}$$

QED.

For b) and c) I am afraid I need hints. What do they mean by saying 'a rotation around the 3-axis over an angle ##\epsilon_{12}##' and 'a boost along the 3-axis with rapidity ##\epsilon_{03}##'?

Thanks :smile:
 
Physics news on Phys.org
  • #2
JD_PM said:
For b) and c) I am afraid I need hints. What do they mean by saying 'a rotation around the 3-axis over an angle ##\epsilon_{12}##' and 'a boost along the 3-axis with rapidity ##\epsilon_{03}##'?
Think about "rotation around the z-axis by an angle ##\theta##". I.e., a rotation in the x-y plane by an angle ##\theta##. What is the general (matrix) expression for such a 2D rotation of the x,y coordinates? Try to express that matrix in terms of an exponential of a generator matrix. (Hint: think about the Taylor series for sin and cos.)

A similar approach applies to part (c), except that there you work with a boost transformation involving the ##z,t## coordinates instead of a rotation of the ##x,y## coordinates.
 
  • Like
Likes hutchphd and JD_PM
  • #3
strangerep said:
Think about "rotation around the z-axis by an angle ##\theta##". I.e., a rotation in the x-y plane by an angle ##\theta##. What is the general (matrix) expression for such a 2D rotation of the x,y coordinates? Try to express that matrix in terms of an exponential of a generator matrix. (Hint: think about the Taylor series for sin and cos.)
b) Alright so the Rotation matrix in 2D

$$R(\theta)=
\begin{pmatrix}
\cos(\theta) & -\sin(\theta) \\
\sin(\theta) & \cos(\theta) \\
\end{pmatrix}
$$

Can also be written as follows (I checked out the rotation case here)

$$R(\theta)=e^{G \theta}= 1_{2\times2} + G\sin(\theta) + G^2(1-\cos\theta)$$

Where

$$G=ba^T -ab^T, \ \ \ \ a^T b=0, \ \ \ \ G^2=-aa^T-bb^T$$

In our problem we have

$$R(\epsilon^{\mu}_{\ \ \nu})=e^{G \epsilon^{\mu}_{\ \ \nu}}= 1_{2\times2} + G\sin(\epsilon^{\mu}_{\ \ \nu}) + G^2(1-\cos\epsilon^{\mu}_{\ \ \nu})$$

Applying Taylor series for ##\cos## and ##\sin## up to second order we get

$$R(\epsilon^{\mu}_{\ \ \nu}) = 1_{2\times2} + G \epsilon^{\mu}_{\ \ \nu}+G^2\frac{\epsilon^{\mu}_{\ \ \lambda}\epsilon^{\lambda}_{\ \ \nu}}{2!} \tag{*}$$

We now multiply both sides of ##(*)## by ##\eta_{\sigma \mu}## to get

$$\eta_{\sigma \mu}R(\epsilon^{\mu}_{\ \ \nu}) = \eta_{\sigma \mu}1_{2\times2} + G \epsilon_{\sigma\nu}+G^2\frac{\epsilon_{\sigma\lambda}\epsilon^{\lambda}_{\ \ \nu}}{2!} \tag{**}$$

OK so far.

However, I do not see why Eq ##(**)## only holds when ##\epsilon_{12} \neq 0##.
 
  • #4
Actually, I was hoping you would see the correct matrix ##G## almost by inspection.

Hint: expand ## e^{i \sigma_2\, \theta}##, where ##\sigma_2## is the 2nd Pauli matrix.
 
  • Informative
Likes JD_PM
  • #5
strangerep said:
Hint: expand ## e^{i \sigma_2\, \theta}##, where ##\sigma_2## is the 2nd Pauli matrix.

$$e^{i \sigma_2\, \theta}=1+i\sigma_2\theta-\frac{(\sigma_2\theta)^2}{2!}+ \ ...$$

strangerep said:
Actually, I was hoping you would see the correct matrix ##G## almost by inspection.

I see that if we pick ##a=
\begin{pmatrix}
1 \\
0 \\
\end{pmatrix}
## and ##b=
\begin{pmatrix}
0 \\
1 \\
\end{pmatrix}
## then we get ##G=
\begin{pmatrix}
0 & -1 \\
1 & 0 \\
\end{pmatrix}
## and ##G^2=
\begin{pmatrix}
-1 & 0 \\
0 & -1 \\
\end{pmatrix}
##

Thus we see that

$$\sigma_2=iG$$

Thus ##(**)## can be rewritten as follows (note that in my previous message I forgot to explicitly state that ##(1_{2\times2})^{\mu}_{ \ \ \ \nu}##)

$$\eta_{\sigma \mu}R(\epsilon^{\mu}_{\ \ \nu}) = \eta_{\sigma \mu}(1_{2\times2})^{\mu}_{ \ \ \ \nu} -i \sigma_2 \epsilon_{\sigma\nu}-\sigma_2^2 \frac{\epsilon_{\sigma\lambda}\epsilon^{\lambda}_{\ \ \nu}}{2!} \tag{***}$$
 
  • #6
I was a bit stuck once I got ##(***)## but then I recalled I came across rotation block matrices ##S^{ij}## containing Pauli matrices while studying Tong's notes (page 85). They are given to be

$$S^{ij} = \frac 1 2 \begin{pmatrix}
0 & \sigma^i \\
-\sigma^i & 0 \\
\end{pmatrix} \begin{pmatrix}
0 & \sigma^j \\
-\sigma^j & 0 \\
\end{pmatrix}=-\frac{i}{2} \mathscr{E}^{ijk}\begin{pmatrix}
\sigma^k & 0 \\
0 & \sigma^k \\
\end{pmatrix} \tag{4.25}$$

Where ##\mathscr{E}^{ijk}## is the Levi-Civita symbol .

We note that ##(4.25)## only holds when ##i \neq j##. Thus we could define a rotation parameter ##\Omega_{ij} = -\mathscr{E}_{ijk} \varphi^k##. Then Tong asserts that the rotation matrix ##S^{ij}## becomes

$$S[\Lambda]=\exp(\frac 1 2 \Omega_{\rho \sigma}S^{\rho \sigma})=\begin{pmatrix}
e^{\frac 1 2 i \vec \varphi \cdot \vec \sigma} & 0 \\
0 & e^{\frac 1 2 i \vec \varphi \cdot \vec \sigma} \\
\end{pmatrix} \tag{4.26}$$

OK.

Coming back to the original problem b)

I was hoping to be able to use the above to solve the b) but I noticed that in ##(4.25)## we need to be dealing with ##i \neq j## if we want to get ##S^{ij} \neq 0## and in ##(***)## we're dealing all the time with ##\sigma_2##; so either ##(***)## is wrong or the rotation matrices provided by Tong are not useful to solve this particular problem.

PS: I may be overcomplicating things though.
 
  • #7
Usually it's tough to do literally the infinite matrix-exponential calculation. It's easy for spin-1/2 with Pauli matrices (which also can be used for the Lorentz transformations for the Weyl-spinor representations of the Lorentz group, or rather the covering group of the original Lorentz group, ##\mathrm{SL}(2,\mathbb{C})##).

Often it's easier to define the vector
$$\vec{x}(t)=\exp(t \hat{\epsilon})$$
and then solve for
$$\dot{\vec{x}}(t)=\hat{\epsilon} \vec{x}.$$
Here ##\vec{x}## is of course a vector appropriate for the problem, i.e., ##\mathbb{R}^n## or ##\mathbb{C}^n## when you have a real or complex ##(n \times n)## matrix to exponentiate.
 
  • #8
JD_PM said:
I was hoping to be able to use the above to solve the b) but I noticed that in ##(4.25)## we need to be dealing with ##i \neq j## if we want to get ##S^{ij} \neq 0## and in ##(***)## we're dealing all the time with ##\sigma_2##; so either ##(***)## is wrong or the rotation matrices provided by Tong are not useful to solve this particular problem.
Take another look at eq(4.9) in Tong's notes, and the following paragraph.
A finite Lorentz transformation takes the form of Tong's eq(4.10). Which elements in Tong's ##\omega^\mu_{~\nu}## must be nonzero to get a rotation in the ##(x^1,x^2)##-plane?

So yes, you were indeed overcomplicating it again. The answer is (almost) staring you in the face in Tong's eqns (4.9) and (4.10). Take the square of the matrix ##M^{12}##. Which terms remain zero? Do the same with the cube of ##M^{12}##. You'll find simply that the Taylor series I hinted at involving just the 2D matrix ##\sigma_2## remains nestled inside the 4D matrix.
 
  • Like
Likes JD_PM, vanhees71 and PeroK
  • #9
OK I see we can construct a basis for the antisymmetric matrix ##\omega^\mu_{~\nu}##. Such a basis ##(\mathcal{M}^{\rho \sigma})^{\mu}_{ \ \ \nu}## contains 6 independent elements (3 boosts and 3 rotations).

Rotations basis elements are ##(\mathcal{M}^{1 3})^{\mu}_{ \ \ \nu}##, ##(\mathcal{M}^{2 3})^{\mu}_{ \ \ \nu}## and ##(\mathcal{M}^{1 2})^{\mu}_{ \ \ \nu}##

I see that ##(\mathcal{M}^{1 2})^{\mu}_{ \ \ \nu}## generates rotations around the ##x^3## axis, which are the ones they are asking at b). Then by Tong's ##(4.10)##

$$\omega^{\mu}_{ \ \ \nu}= \frac 1 2 \Omega_{12} (\mathcal{M}^{1 2})^{\mu}_{ \ \ \nu}$$

And the rotation parameter is indeed ##\Omega_{12}##

I'd rather use the series expansion method to solve the problem though.

strangerep said:
So yes, you were indeed overcomplicating it again. The answer is (almost) staring you in the face in Tong's eqns (4.9) and (4.10). Take the square of the matrix ##M^{12}##. Which terms remain zero? Do the same with the cube of ##M^{12}##. You'll find simply that the Taylor series I hinted at involving just the 2D matrix ##\sigma_2## remains nestled inside the 4D matrix.

Mmm thanks to your comment I've noticed that I gave ##G## (which is ##(\mathcal{M}^{12})^{\mu}_{ \ \ \nu}## indeed) and ##G^2## incorrectly; they are ##4\times4## matrices instead. Let's use ##(\mathcal{M}^{12})^{\mu}_{ \ \ \nu}## notation from now on.$$(\mathcal{M}^{12})^{\mu}_{ \ \ \nu}=
\begin{pmatrix}
0 & 0 & 0 & 0\\
0 & 0 & -1 & 0 \\
0 & 1 & 0 & 0\\
0 & 0 & 0 & 0\\
\end{pmatrix}, \ \ \ \ \ \
\Big((\mathcal{M}^{12})^{\mu}_{ \ \ \nu}\Big)^2=
\begin{pmatrix}
0 & 0 & 0 & 0\\
0 & -1& 0& 0 \\
0 & 0 & -1 & 0\\
0 & 0 & 0 & 0\\
\end{pmatrix}
$$

Now I am trying to set up the correct series expansion (including the right generator); I am thinking of something of the form

$$R(\epsilon^{\mu}_{\ \ \nu}) = (1_{4\times4})^{\mu}_{ \ \ \nu} + (\mathcal{M}^{12})\epsilon^{\mu}_{\ \ \nu}+\Big(\mathcal{M}^{12}\Big)^2\frac{\epsilon^{\mu}_{\ \ \lambda}\epsilon^{\lambda}_{\ \ \nu}}{2!} \tag{****}$$

But I'm a bit stuck...
 
  • #10
JD_PM said:
I'd rather use the series expansion method to solve the problem though.
[...]
Mmm thanks to your comment I've noticed that I gave ##G## (which is ##(\mathcal{M}^{12})^{\mu}_{ \ \ \nu}## indeed) and ##G^2## incorrectly; they are ##4\times4## matrices instead. Let's use ##(\mathcal{M}^{12})^{\mu}_{ \ \ \nu}## notation from now on.
Do you still not get the point that you can concentrate just on ##\sigma_2## (which is nestled inside the 4x4 ##M^{12}## matrix)? ##(M^{12})^n## (for any power ##n##) always has zeroes around the outside.

Go back and do what I suggested in my post #4 (you have not yet computed the matrices properly). Do the expansion up to 4th order and then compare with the Taylor series for ##\sin\theta## and ##\cos\theta##. Relate this back to the expression for a finite rotation in the x-y place by angle ##\theta##. (You still haven't actually done this properly.)
 
  • Like
Likes JD_PM
  • #11
I'd solve the 1st-order system of ODE's. It's much simpler. You've done the calculation when you solved for the relativistic motion of particles in homogeneous electric fields ("boost") or magnetic fields ("rotation").
 
  • #12
vanhees71 said:
I'd solve the 1st-order system of ODE's. It's much simpler.

Hi vanhees71.

That’s the approach you suggested at #7, right? I have to acknowledge I don’t really understand it. Could you please give an explicit worked-out example? (Maybe I’m asking too much).

Solving the problem using both ways sound great to me.
 
  • #13
Well, let's take the example of a rotation around a fixed axis ##\vec{n}## in ##\mathbb{R}^3##. Writing
$$\vec{x}(\phi)=\hat{D}_{\text{n}}(\phi) \vec{x}_0,$$
you know that
$$\vec{x}'(\phi)=\vec{n} \times \vec{x}(\phi).$$
It's most easy to choose a Cartesian reference frame with ##\vec{n}=\vec{e}_3##. Then you have
$$\begin{pmatrix}x' \\ y' \\ z' \end{pmatrix} = \begin{pmatrix} -y \\ x \\0 \end{pmatrix}.$$
This set of ODE's you can solve with the initial condition ##\vec{x}(0)=\vec{x}_0##.

Taking the derivative of the first component of the equation leads to
$$x''=-y'=-x \; \Rightarrow \; x=A \cos \varphi + B \sin \varphi.$$
From this
$$y=-x'=-A \sin \varphi + B \cos \varphi.$$
The 3rd equation is trivial:
$$z'=0 \; \Rightarrow \; z=z_0=\text{const}.$$
The initial condition for ##x## and ##y## leads to
$$x(0)=A=x_0, \quad y(0)=B=y_0.$$
From that you have
$$\vec{x}(\varphi)=\begin{pmatrix} \cos \varphi & \sin \varphi & 0 \\
-\sin \varphi & \cos \varphi & 0\\
0 & 0 & 1 \end{pmatrix} \vec{x}_0.$$
The matrix is ##\hat{D}_{\vec{e}_3}(\varphi)##. Now to express this with invariant vectors just take
$$\vec{n} \times \vec{x}_0=\vec{e}_3 \times \vec{x}_0=\begin{pmatrix}-y_0 \\ x_0 \\ 0 \end{pmatrix}$$
and
$$\vec{n} \times (\vec{n} \times \vec{x}_0) = - \begin{pmatrix} x_0 \\ y_0 \\ 0 \end{pmatrix}.$$
This leads to
$$\hat{D}_{\varphi} \vec{x}_0=\vec{n} (\vec{n} \cdot \vec{x}_0) - \sin \varphi \vec{n} \times \vec{x}_0 - \cos \varphi \vec{n} \times (\vec{n} \times \vec{x}_0).$$
 
  • Like
Likes JD_PM
  • #14
strangerep said:
Do you still not get the point that you can concentrate just on ##\sigma_2## (which is nestled inside the 4x4 ##M^{12}## matrix)? ##(M^{12})^n## (for any power ##n##) always has zeroes around the outside.

Yes, I see it.

strangerep said:
Go back and do what I suggested in my post #4 (you have not yet computed the matrices properly). Do the expansion up to 4th order and then compare with the Taylor series for ##\sin\theta## and ##\cos\theta##. Relate this back to the expression for a finite rotation in the x-y place by angle ##\theta##. (You still haven't actually done this properly.)
Let's go step by step.

strangerep said:
Go back and do what I suggested in my post #4 (you have not yet computed the matrices properly).

You suggested to write the Taylor expansion of ##e^{i \sigma_2\, \theta}##. Let's now do it up to fourth order

$$e^{i \sigma_2\, \theta}=1+i\sigma_2\theta+i^2 \frac{(\sigma_2\theta)^2}{2!}+i^3 \frac{(\sigma_2\theta)^3}{3!}+i^4 \frac{(\sigma_2\theta)^4}{4!} \ +...$$

Where

$$\sigma_2 =

\begin{pmatrix}

0 & -i \\

i & 0 \\

\end{pmatrix}$$

$$\sigma_2^2 =

\begin{pmatrix}

1 & 0 \\

0 & 1 \\

\end{pmatrix}=1_{2\times2}$$

$$\sigma_2^3 =

\begin{pmatrix}

0 & -i \\

i & 0 \\

\end{pmatrix}=\sigma_2$$

$$\sigma_2^4 =

\begin{pmatrix}

1 & 0 \\

0 & 1 \\

\end{pmatrix}=1_{2\times2}$$

strangerep said:
then compare with the Taylor series for ##\sin\theta## and ##\cos\theta##. Relate this back to the expression for a finite rotation in the x-y place by angle ##\theta##. (You still haven't actually done this properly.)

I see how to relate the expansion of the exponential and ##\sin\theta## and ##\cos\theta## through Euler's formula ##e^{\pm i \sigma_2\, \theta}=\cos(\sigma_2\, \theta) \pm i\sin(\sigma_2\, \theta)##. However I do not see how to relate this to the expression for a finite rotation in the x-y plane by angle ##\theta##.

I know once I see it I'll say oh, it was really easy but unfortunately I did not 'click' yet. I was thinking that an example could help me out to see your point.

Thanks for your patience.
 
  • #15
vanhees71 said:
Well, let's take the example of a rotation around a fixed axis ##\vec{n}## in ##\mathbb{R}^3##. Writing
$$\vec{x}(\phi)=\hat{D}_{\text{n}}(\phi) \vec{x}_0,$$
you know that
$$\vec{x}'(\phi)=\vec{n} \times \vec{x}(\phi).$$
It's most easy to choose a Cartesian reference frame with ##\vec{n}=\vec{e}_3##. Then you have
$$\begin{pmatrix}x' \\ y' \\ z' \end{pmatrix} = \begin{pmatrix} -y \\ x \\0 \end{pmatrix}.$$
This set of ODE's you can solve with the initial condition ##\vec{x}(0)=\vec{x}_0##.

Taking the derivative of the first component of the equation leads to
$$x''=-y'=-x \; \Rightarrow \; x=A \cos \varphi + B \sin \varphi.$$
From this
$$y=-x'=-A \sin \varphi + B \cos \varphi.$$
The 3rd equation is trivial:
$$z'=0 \; \Rightarrow \; z=z_0=\text{const}.$$
The initial condition for ##x## and ##y## leads to
$$x(0)=A=x_0, \quad y(0)=B=y_0.$$
From that you have
$$\vec{x}(\varphi)=\begin{pmatrix} \cos \varphi & \sin \varphi & 0 \\
-\sin \varphi & \cos \varphi & 0\\
0 & 0 & 1 \end{pmatrix} \vec{x}_0.$$
The matrix is ##\hat{D}_{\vec{e}_3}(\varphi)##. Now to express this with invariant vectors just take
$$\vec{n} \times \vec{x}_0=\vec{e}_3 \times \vec{x}_0=\begin{pmatrix}-y_0 \\ x_0 \\ 0 \end{pmatrix}$$
and
$$\vec{n} \times (\vec{n} \times \vec{x}_0) = - \begin{pmatrix} x_0 \\ y_0 \\ 0 \end{pmatrix}.$$
This leads to
$$\hat{D}_{\varphi} \vec{x}_0=\vec{n} (\vec{n} \cdot \vec{x}_0) - \sin \varphi \vec{n} \times \vec{x}_0 - \cos \varphi \vec{n} \times (\vec{n} \times \vec{x}_0).$$

Thanks vanhees71.
 
  • #16
JD_PM said:
$$e^{i \sigma_2\, \theta}=1+i\sigma_2\theta+i^2 \frac{(\sigma_2\theta)^2}{2!}+i^3 \frac{(\sigma_2\theta)^3}{3!}+i^4 \frac{(\sigma_2\theta)^4}{4!} \ +...$$
Continue this a bit more. E.g., replace ##i\sigma_2## by the corresponding matrix with real entries only (I'll call it ``"##\alpha##"). Rewrite your 4th order expansion in terms of ##\alpha## and ##1_{2\times 2}##. I.e., get rid of all powers on your matrices by substituting what they evaluate to, e.g., ##\alpha^3 = -\alpha##.

Then rearrange the expansion so that even and odd powers of ##\theta## are grouped together separately (say, by putting those respective terms on 2 separate lines).

Then WRITE OUT the Taylor series for ##\sin\theta## and ##\cos\theta## (each on a separate line) and compare with the 2-line expansion). Notice anything?

I know once I see it I'll say oh, it was really easy.
Yep.

Late Edit: I've decided to make it easier for you. Prove that $$e^{-i \sigma_2\, \theta} ~=~ \begin{pmatrix} \cos\theta & -\sin\theta \\ \sin\theta & \cos\theta \end{pmatrix} ~.$$
 
Last edited:
  • Love
Likes JD_PM
  • #17
Yes, love the 'I got it' feeling! 😍 (thanks for your patience 🤝)

strangerep said:
Continue this a bit more. E.g., replace ##i\sigma_2## by the corresponding matrix with real entries only (I'll call it ``"##\alpha##"). Rewrite your 4th order expansion in terms of ##\alpha## and ##1_{2\times 2}##. I.e., get rid of all powers on your matrices by substituting what they evaluate to, e.g., ##\alpha^3 = -\alpha##.

$$e^{i \sigma_2\, \theta}=\begin{pmatrix}
1 & 0 \\
0 & 1 \\
\end{pmatrix}\theta^0 +\begin{pmatrix}
0 & 1 \\
-1 & 0 \\
\end{pmatrix}\theta
+
\frac{1}{2!} \begin{pmatrix}
-1 & 0 \\
0 & -1 \\
\end{pmatrix}\theta^2
+
\frac{1}{3!} \begin{pmatrix}
0 & -1 \\
1 & 0 \\
\end{pmatrix}\theta^3
+
\frac{1}{4!} \begin{pmatrix}
1 & 0 \\
0 & 1 \\
\end{pmatrix}\theta^4
\ +...$$

strangerep said:
Then rearrange the expansion so that even and odd powers of ##\theta## are grouped together separately (say, by putting those respective terms on 2 separate lines).

$$e^{i \sigma_2\, \theta}=\begin{pmatrix}
1 & 0 \\
0 & 1 \\
\end{pmatrix}\theta^0 +
\frac{1}{2!} \begin{pmatrix}
-1 & 0 \\
0 & -1 \\
\end{pmatrix}\theta^2 +
\frac{1}{4!} \begin{pmatrix}
1 & 0 \\
0 & 1 \\
\end{pmatrix}\theta^4$$ $$+
\begin{pmatrix}
0 & 1 \\
-1 & 0 \\
\end{pmatrix}\theta
+
\frac{1}{3!} \begin{pmatrix}
0 & -1 \\
1 & 0 \\
\end{pmatrix}\theta^3 \ + ...$$

strangerep said:
Then WRITE OUT the Taylor series for ##\sin\theta## and ##\cos\theta## (each on a separate line) and compare with the 2-line expansion).

$$\cos \theta = 1 - \frac{\theta^2}{2!} + \frac{\theta^4}{4!} \ - ...$$

$$\sin \theta = \theta - \frac{\theta^3}{3!} \ + ...$$

strangerep said:
Notice anything?

Ahhh so the key was to focus on matrix entries! For instance if we focus on ##x_{11}## position we indeed get ##\cos \theta##

$$\sum_{i=0}^n x_{i 11} = \theta^0 - \frac{\theta^2}{2!} + \frac{\theta^4}{4!} \ - ... = \cos \theta$$

For the other three we get

$$\sum_{i=0}^n x_{i 12} = \theta - \frac{\theta^3}{3!} \ + ... = \sin \theta$$

$$\sum_{i=0}^n x_{i 21} = -\theta + \frac{\theta^3}{3!} \ - ... = -\sin \theta$$

$$\sum_{i=0}^n x_{i 22} = \theta^0 - \frac{\theta^2}{2!} + \frac{\theta^4}{4!} \ - ... = \cos \theta$$

Where ##i=0## refers to the matrix ##\begin{pmatrix}
1 & 0 \\
0 & 1 \\
\end{pmatrix}\theta^0## and so on

Thus we have just proven that

$$e^{i \sigma_2\, \theta} ~=~ \begin{pmatrix} \cos\theta & \sin\theta \\ -\sin\theta & \cos\theta \end{pmatrix} ~.$$

By analogous work we can prove

$$e^{-i \sigma_2\, \theta} ~=~ \begin{pmatrix} \cos\theta & -\sin\theta \\ \sin\theta & \cos\theta \end{pmatrix} ~.$$
 
  • #18
OK. I see how to relate the exponential to the 2D rotation matrix.

But once we've worked out the matrix exponential, how can we see that only when ##\epsilon_{12}## does not vanish we get a rotation around the 3-axis over an angle ##\epsilon_{12}##?
 
  • #19
JD_PM said:
By analogous work we can prove
$$e^{-i \sigma_2\, \theta} ~=~ \begin{pmatrix} \cos\theta & -\sin\theta \\ \sin\theta & \cos\theta \end{pmatrix} ~.$$
It doesn't to be "analogous work". Just substitute ##\theta \to -\theta##.
 
  • Like
Likes JD_PM
  • #20
JD_PM said:
OK. I see how to relate the exponential to the 2D rotation matrix.

But once we've worked out the matrix exponential, how can we see that only when ##\epsilon_{12}## does not vanish we get a rotation around the 3-axis over an angle ##\epsilon_{12}##?
You've got the general 2D rotation transformation ##x \to x'(\theta) = x \cos\theta - y\sin\theta; ~~~ y \to y'(\theta) = x \sin\theta + y\cos\theta##.

So now write down the general 3D rotation transformation (in terms of coordinates ##x,y,z##) that rotate ##x,y## around the ##z## axis, i.e., which leave ##z## unchanged. This will be 3 equations ##x' = ?;~## ##y' = ?;~## ##z' = ?\;##. Then combine these 3 equations into a single 3D matrix equation.

What are the corresponding 3D matrix equations for ##x - z## rotations (which preserve ##y##) and for ##y - z## rotations (which preserve ##x##) ?
 
  • #21
strangerep said:
So now write down the general 3D rotation transformation (in terms of coordinates ##x,y,z##) that rotate ##x,y## around the ##z## axis, i.e., which leave ##z## unchanged. This will be 3 equations ##x' = ?;~## ##y' = ?;~## ##z' = ?\;##. Then combine these 3 equations into a single 3D matrix equation.

The rotation transformations around the ##z## axis

$$x \to x'(\theta) = x \cos\theta + y\sin\theta; ~~~ y \to y'(\theta) = -x \sin\theta + y\cos\theta; ~~~ z \to z' = z$$

Thus the z-rotation matrix is

$$R_z(\theta)=
\begin{pmatrix}
\cos\theta & \sin\theta & 0 \\
-\sin\theta & \cos\theta & 0 \\
0 & 0 & 1 \\
\end{pmatrix}
$$

strangerep said:
What are the corresponding 3D matrix equations for ##x - z## rotations (which preserve ##y##) [...] ?

The rotation transformations around the ##y## axis

$$x \to x'(\theta) = x \cos\theta - z\sin\theta; ~~~ y \to y' = y; ~~~ z \to z'(\theta) = x \sin\theta + z\cos\theta$$

Thus the y-rotation matrix is

$$R_y(\theta)=
\begin{pmatrix}
\cos\theta & 0 & -\sin\theta \\
0 & 1 & 0 \\
\sin\theta & 0 & \cos\theta \\
\end{pmatrix}
$$

strangerep said:
[...] for ##y - z## rotations (which preserve ##x##) ?

The rotation transformations around the ##x## axis

$$x \to x' = x; ~~~ y \to y'(\theta) = y \cos\theta + z\sin\theta; ~~~ z \to z'(\theta) = -y \sin\theta + z\cos\theta$$

Thus the x-rotation matrix is

$$R_x(\theta)=
\begin{pmatrix}
1 & 0 & 0 \\
0 & \cos\theta & \sin\theta \\
0 & -\sin\theta & \cos\theta \\
\end{pmatrix}
$$
 
  • #22
OK I now understand the statement 'only when ##\epsilon_{12}## does not vanish we get a rotation around the 3-axis over an angle ##\epsilon_{12}##'; they mean that only when the z-coordinate remains unchanged we get a rotation around the z-axis.

However, they want us to show it explicitly in terms of ##\epsilon## and matrix notation. Does that mean we have to introduce the Levi-Civita symbol and proceed like below:

Captura de pantalla (1105).png


?
 
  • #23
JD_PM said:
However, they want us to show it explicitly in terms of ##\epsilon## and matrix notation. Does that mean we have to introduce the Levi-Civita symbol and proceed like below:

View attachment 266675
?
Similar, but not exactly like that. The text you've reproduced looks like how one would represent 3D rotations that act on a Dirac 4-spinor. But in previous posts you've been working rotations acting on ordinary 3-space vectors. To understand this distinction better, I suggest you take your ##R_i(\theta)## matrices, (where ##i=x,y,z##) and compute their corresponding generators.

If you don't know what a "generator" is, it's the thing you use in the exponential version of a rotation matrix. So to recover a generator from the corresponding finite transformation, just differentiate wrt ##\theta## and then set ##\theta=0##.

So for each of your ##R_i(\theta)## matrices, you'll get a "##J_i##" generator matrix.

Further exercises:
1) Show that the set of the 3 ##J_i## matrices form a Lie algebra. (If you don't know what a Lie algebra is, then consider it part of this exercise to look that up.)

2) Show that the set of the 3 ##J_i## matrices form a vector space, in the sense that any linear combination of them is in a fact a generator for some 3D rotation.

3) Show that ANY finite 3D rotation can be obtained by exponentiating a linear combination of the ##J_i##'s, (meaning that exponentiating any such linear combination yields a finite 3D rotation matrix).

[Btw, if you haven't yet acquired a copy of Greiner's "Quantum Mechanics -- Symmetries", I can assure you it would be money well spent. I learned a vast amount from that book when I was getting started into serious study of QM. Greiner's textbooks are good for a beginner student because they give lots of worked examples, and he doesn't skip steps in the detailed computations.]
 
  • Like
Likes JD_PM

1. What is the series expansion of the Lorentz Transformation?

The series expansion of the Lorentz Transformation is a mathematical tool used to approximate the transformation equations between two reference frames in special relativity. It is based on the Taylor series expansion and allows for a more precise calculation of the transformation at high velocities.

2. How is the series expansion of the Lorentz Transformation derived?

The series expansion of the Lorentz Transformation is derived by using the Taylor series expansion on the transformation equations, which are based on the principles of special relativity. This allows for a more accurate calculation of the transformation at high velocities.

3. What is the significance of the series expansion of the Lorentz Transformation?

The series expansion of the Lorentz Transformation is significant because it allows for a more precise calculation of the transformation at high velocities, which is important in understanding the effects of special relativity. It also helps to explain the observed phenomena such as time dilation and length contraction.

4. How accurate is the series expansion of the Lorentz Transformation?

The series expansion of the Lorentz Transformation is accurate to a certain degree, depending on the number of terms used in the expansion. The more terms that are included, the more accurate the calculation will be. However, at extremely high velocities, the series expansion may not be as accurate as other methods of calculation.

5. How is the series expansion of the Lorentz Transformation used in practical applications?

The series expansion of the Lorentz Transformation is used in a variety of practical applications, such as in particle physics, astrophysics, and in the design of high-speed technologies. It is also used in GPS systems, which rely on the principles of special relativity to accurately calculate positions and time.

Similar threads

  • Advanced Physics Homework Help
Replies
2
Views
470
  • Advanced Physics Homework Help
Replies
3
Views
1K
  • Advanced Physics Homework Help
Replies
1
Views
1K
  • Advanced Physics Homework Help
Replies
1
Views
2K
  • Advanced Physics Homework Help
Replies
11
Views
2K
  • Advanced Physics Homework Help
Replies
10
Views
2K
  • Advanced Physics Homework Help
Replies
3
Views
2K
  • Advanced Physics Homework Help
Replies
30
Views
5K
  • Advanced Physics Homework Help
Replies
5
Views
2K
  • Advanced Physics Homework Help
Replies
0
Views
293
Back
Top