Index Notation for Lorentz Transformation

Click For Summary

Discussion Overview

The discussion centers on the application of index notation to the Lorentz transformation, particularly focusing on the correct representation of the transpose of the transformation matrix and the implications for raising and lowering indices. Participants explore the mathematical relationships and definitions involved in these transformations, including the use of the metric tensor.

Discussion Character

  • Technical explanation
  • Debate/contested
  • Mathematical reasoning

Main Points Raised

  • Some participants express confusion regarding the definition of the transpose for Lorentz transformation matrices and how it relates to index notation.
  • There is a proposal that the transpose of the Lorentz transformation matrix should be written as (ΛT)μ ν = Λν μ, but this raises questions about consistency with the definition of transpose.
  • Others argue that the distinction between raised and lowered indices complicates the understanding of transposition and suggest using only raised indices for clarity.
  • One participant emphasizes that the condition Λνμ = (Λ-1)μν does not imply that ΛΛT = I, highlighting the need for careful treatment of indices.
  • Another participant points out that when taking the transpose, one must also swap the indices from up to down, which is crucial for maintaining the correct relationships.
  • There is a discussion about the implications of using matrix notation versus tensor notation, with some participants advocating for a more intuitive understanding of Lorentz transformations as derivatives of coordinates.
  • Participants reference specific examples and equations to illustrate their points, including the matrix forms of Lorentz transformations and their inverses.

Areas of Agreement / Disagreement

Participants do not reach consensus on the correct treatment of indices in the context of Lorentz transformations. There are competing views on the interpretation of transposition and the implications for matrix versus tensor notation.

Contextual Notes

Limitations include unresolved questions about the correct application of raising and lowering indices, the implications of matrix notation, and the definitions of transpose and inverse in the context of Lorentz transformations.

fayled
Messages
176
Reaction score
0
The Lorentz transformation matrix may be written in index form as Λμ ν. The transpose may be written (ΛT)μ νν μ.

I want to apply this to convert the defining relation for a Lorentz transformation η=ΛTηΛ into index form. We have
ηρσ=(ΛT)ρ μημνΛν σ

The next step to obtain the correct result seems to be to write (ΛT)ρ μμ ρ

However I don't see how this agrees with the way I have defined the transpose above. Why can we not use (ΛT)ρ μμ ρ, which seems to follow from the definition of the transpose above?

Thanks for any help!
 
Physics news on Phys.org
fayled said:
The Lorentz transformation matrix may be written in index form as Λμ ν. The transpose may be written (ΛT)μ νν μ.

I want to apply this to convert the defining relation for a Lorentz transformation η=ΛTηΛ into index form. We have
ηρσ=(ΛT)ρ μημνΛν σ

The next step to obtain the correct result seems to be to write (ΛT)ρ μμ ρ

However I don't see how this agrees with the way I have defined the transpose above. Why can we not use (ΛT)ρ μμ ρ, which seems to follow from the definition of the transpose above?

Thanks for any help!

The word "transpose" is appropriate for matrices, not tensors. For matrices, the distinction between raised indices and lowered indices is more confusing than helpful. So let's write it all in terms of raised indices. Then the Lorentz transform is:

\Lambda^{\alpha \beta} x^\beta = (x')^\alpha

(\Lambda^T)^{\alpha \beta} = \Lambda^{\beta \alpha}

So the defining relation for \eta is:

(\Lambda^T) \eta \Lambda = \eta

or in terms of components:

\Lambda^{\beta \alpha} \eta^{\beta \mu} \Lambda^{\mu \nu} = \eta^{\alpha \nu}
 
stevendaryl said:
The word "transpose" is appropriate for matrices, not tensors. For matrices, the distinction between raised indices and lowered indices is more confusing than helpful. So let's write it all in terms of raised indices. Then the Lorentz transform is:

\Lambda^{\alpha \beta} x^\beta = (x')^\alpha

(\Lambda^T)^{\alpha \beta} = \Lambda^{\beta \alpha}

So the defining relation for \eta is:

(\Lambda^T) \eta \Lambda = \eta

or in terms of components:

\Lambda^{\beta \alpha} \eta^{\beta \mu} \Lambda^{\mu \nu} = \eta^{\alpha \nu}

Thanks for your reply.

Unfortunately I think this just sidetracks the problem. I really need to be able to understand how to deal with the raising and lowering of indices.

For example, I want to be able to understand why the condition Λνμ=(Λ-1)μν doesn't imply that ΛΛT=I. If I just use all indices as raised, then that doesn't seem to help.
 
To be honest, things such as ##(\Lambda^T)^\alpha{}_{\beta}## do not really make sense - the indices are in the wrong place in relation to the transpose and the parenthesis. You can define a matrix
$$
\Lambda = (\Lambda^\alpha{}_{\beta})
$$
which is the matrix whose entry in row ##\alpha## and column ##\beta## is ##\Lambda^\alpha{}_\beta##. By definition of matrix transposition, the transpose ##\Lambda^T## would then be the matrix
$$
(\Lambda^\alpha{}_\beta)^T
$$
with ##\Lambda^\alpha{}_\beta## being the entry in column ##\alpha## and row ##\beta##. If you anyway insist on writing things such as ##(\Lambda^T)_\beta{}^\alpha##, the only meaningful relation involving indices up and down would be ##(\Lambda^T)_\beta{}^\alpha = \Lambda^\alpha{}_\beta##. You cannot change where the indices are located (i.e., up/down).

Edit: That being said, I strongly discourage thinking of the Lorentz transformation as a matrix. You can represent it by a matrix, but really the most intuitive way of thinking about it is as ##\Lambda^\alpha{}_\beta = \partial x'^\alpha/\partial x^\beta## (this will also help in making the transition to general coordinate transforms).
 
Orodruin said:
To be honest, things such as ##(\Lambda^T)^\alpha{}_{\beta}## do not really make sense - the indices are in the wrong place in relation to the transpose and the parenthesis. You can define a matrix
$$
\Lambda = (\Lambda^\alpha{}_{\beta})
$$
which is the matrix whose entry in row ##\alpha## and column ##\beta## is ##\Lambda^\alpha{}_\beta##. By definition of matrix transposition, the transpose ##\Lambda^T## would then be the matrix
$$
(\Lambda^\alpha{}_\beta)^T
$$
with ##\Lambda^\alpha{}_\beta## being the entry in column ##\alpha## and row ##\beta##. If you anyway insist on writing things such as ##(\Lambda^T)_\beta{}^\alpha##, the only meaningful relation involving indices up and down would be ##(\Lambda^T)_\beta{}^\alpha = \Lambda^\alpha{}_\beta##. You cannot change where the indices are located (i.e., up/down).

Ok, well in some places I seem to be reading that transposing involves swapping the indices left-right, whereas in other places it seems to imply the indices are completely swapped left-right and top-bottom.

So, assuming we swap left-right only, would you be able to help identify why the following is incorrect (which has led me to all this confusion)?

We know that
Λνμ=(Λ-1)μν
So, why can't I just swap the indices left-right on the left hand side term, to give
T)μν=(Λ-1)μν
which would then imply in matrix form
ΛT-1
This is obviously not correct, because the Lorentz group actually satisfies η=ΛTηΛ...
 
fayled said:
We know that
Λνμ=(Λ-1)μν.
This is incorrect. Here you need to swap the indices from up to down - you do that with the metric tensor. Another reason to avoid matrix notation completely.
 
Orodruin said:
This is incorrect. Here you need to swap the indices from up to down - you do that with the metric tensor. Another reason to avoid matrix notation completely.

See for example Equation 1.34 here:
http://epx.phys.tohoku.ac.jp/~yhitoshi/particleweb/ptest-1.pdf
 
Last edited by a moderator:
fayled said:
Unfortunately I think this just sidetracks the problem. I really need to be able to understand how to deal with the raising and lowering of indices.

fayled said:
Thanks for your reply.

Unfortunately I think this just sidetracks the problem. I really need to be able to understand how to deal with the raising and lowering of indices.

For example, I want to be able to understand why the condition Λνμ=(Λ-1)μν doesn't imply that ΛΛT=I. If I just use all indices as raised, then that doesn't seem to help.

Well, it isn't true that (\Lambda^{-1})^\mu_\nu = \Lambda^\nu_\mu. Let's just look at 2-D spacetime. Then

\Lambda = \left( \begin{array} \\ \gamma & - \gamma v/c\\ -\gamma v/c & \gamma \end{array} \right)

where \gamma = \frac{1}{\sqrt{1 - \frac{v^2}{c^2}}}

\Lambda^{-1} = \left( \begin{array} \\ \gamma & + \gamma v/c \\ +\gamma v/c & \gamma \end{array} \right)

\Lambda^T = \left( \begin{array} \\ \gamma & - \gamma v/c \\ -\gamma v/c & \gamma \end{array} \right)

\eta = \left( \begin{array} \\ 1 & 0 \\ -1 & 0 \end{array} \right)

So it's not true that \Lambda \Lambda^T = I.
 
@fayled :

1) Have a look at this recent thread in which I learned the error of my ways regarding transposition and index placement. See especially post #14 in that thread.

2) tl:dr : In taking a transpose, one must indeed flip upstairs##\leftrightarrow##downstairs as well as swapping the index symbols. See the end of post #14 in the above thread.

Re-summarizing, if ##V,W## are vector spaces, and ##L## is a linear map: $$L : V \to W $$then the transpose ##L^T## is defined as $$L^T : W^* \to V^* ~,$$whereas the inverse is $$L^{-1} : W \to V ~.$$ That's why it's correct to do an upstairs##\leftrightarrow##downstairs flip when taking the transpose -- because you're now working with the dual spaces.

(##V^*## denotes the dual space of ##V##.)
 
  • #10
By definition a Lorentz transformation matrix obeys
$$\eta_{\mu \nu} {\Lambda^{\mu}}_{\rho} {\Lambda^{\nu}}_{\sigma}=\eta_{\rho \sigma},$$
where (in the "west-coast convention") ##(\eta_{\mu \nu})=\text{diag}(1,-1,-1,-1)##. Note that it is crucial to keep the indices in clear order, both horizontally and vertically.

Sometimes the matrix-vector calculus is shortening some calculations, but there's big confusion in the literature and of course also here in this forum. We had a recent thread about it somewhere here in the relativity subforum. The upshot is that by definition
$${(\Lambda^T)^{\mu}}_{\nu}={\Lambda^{\nu}}_{\mu}.$$
With this notation, putting ##(\eta_{\mu \nu})=\hat{\eta}## and ##({\Lambda^{\mu}}_{\nu})=\hat{\Lambda}## (which btw. makes the shortcoming of the matrix-vector notation apparent, because the "automatic syntax check" that co- and contravariant indices are not longer clearly distinguished as in the Ricci calculus, where the former are lower and the latter are upper indices) you can write the above "pseudo-orthogonality condition" as
$$\hat{\Lambda}^T \hat{\eta} \hat{\Lambda}=\hat{\eta},$$
which, because of ##\hat{\eta}^2=\hat{1}##, implies
$$\hat{\Lambda}^{-1}=\hat{\eta} \hat{\Lambda}^T \hat{\eta}.$$

This is easily translated back into index notation
$${(\Lambda^{-1})^{\mu}}_{\nu} = \eta^{\mu \rho} \eta_{\nu \sigma} {\Lambda^{\sigma}}_{\rho}={\Lambda_{\nu}}^{\mu},$$
i.e., the inverse LT is not simply given by the transverse matrix but the indices must be dragged by the pseudo-metric matrix too.
 
  • #11
fayled said:
The Lorentz transformation matrix may be written in index form as Λμ ν. The transpose may be written (ΛT)μ νν μ.

I want to apply this to convert the defining relation for a Lorentz transformation η=ΛTηΛ into index form. We have
ηρσ=(ΛT)ρ μημνΛν σ

The next step to obtain the correct result seems to be to write (ΛT)ρ μμ ρ

However I don't see how this agrees with the way I have defined the transpose above. Why can we not use (ΛT)ρ μμ ρ, which seems to follow from the definition of the transpose above?

Thanks for any help!

The usual confusion again!

1) The non-covariant expression (\Lambda^{T})^{\mu}{}_{\nu} = \Lambda^{\nu}{}_{\mu}, can only be used for the concrete elements of \Lambda, for example: \Lambda^{1}{}_{2} = (\Lambda^{T})^{2}{}_{1}.

2) When \Lambda^{T} comes multiplied by other matrices, don’t use any up-down index structure for \Lambda^{T}. See (3).

3) To obtain the index structure of the matrix equation \eta = \Lambda^{T} \ \eta \ \Lambda, do the following: Since \eta^{T} = \eta, i.e., \eta_{\alpha\beta} = \eta_{\beta\alpha}, then \eta_{\alpha\beta} = \left( ( \eta \ \Lambda)^{T} \ \Lambda \right)_{\beta\alpha} . Now, if we choose the matrix element of \Lambda to be of the form \Lambda^{\rho}{}_{\sigma} with upper index represents the rows and the lower one for columns, then the above equation becomes \eta_{\alpha\beta} = ( \eta \ \Lambda)^{T}_{\beta\mu} \ \Lambda^{\mu}{}_{\alpha} . \ \ \ \ \ \ \ \ (3) For any matrix M, the transpose M^{T} is defined by M^{T}_{ab} = M_{ba}. Thus, (\eta \ \Lambda)^{T}_{\beta\mu} = (\eta \ \Lambda)_{\mu\beta}, and Eq(3) becomes the one we know \eta_{\alpha\beta} = (\eta \ \Lambda)_{\mu\beta} \ \Lambda^{\mu}{}_{\alpha} = \eta_{\mu\nu} \ \Lambda^{\nu}{}_{\beta} \ \Lambda^{\mu}{}_{\alpha} .

4) The expression (\Lambda^{-1})^{\sigma}{}_{\rho} = \Lambda_{\rho}{}^{\sigma} , \ \ \ \ \ \ \ \ \ \ \ (4’) is an allowed covariant expression because you can lower and raise the indices on \Lambda by \eta and \eta^{-1} respectively: \Lambda_{\rho}{}^{\sigma} = \eta^{\sigma \nu} \ \eta_{\rho\mu} \ \Lambda^{\mu}{}_{\nu} . \ \ \ \ \ \ \ \ \ (4) Substituting (4’) in the LHS of (4), we find (\Lambda^{-1})^{\sigma}{}_{\rho} = \eta^{\sigma \nu} \ \left( \eta_{\rho\mu} \ \Lambda^{\mu}{}_{\nu} \right) = \eta^{\sigma \nu} \ \left( \eta \ \Lambda \right)_{\rho \nu} . \ \ \ \ (5) This can be rewritten as (\Lambda^{-1})^{\sigma}{}_{\rho} = \eta^{\sigma\nu} \ \left(\eta \ \Lambda \right)^{T}_{\nu \rho} = \left( \eta \ ( \eta \ \Lambda)^{T} \ \right)^{\sigma}{}_{\rho} .

And, therefore, we have the correct matrix equations \Lambda^{-1} = \eta \ \Lambda^{T} \ \eta = \left( \eta \ \Lambda \ \eta \right)^{T} , \ \ \ \ \ \ (6) and \Lambda^{T} = \eta \ \Lambda^{-1} \ \eta = \left( \eta \ \Lambda \ \eta \right)^{-1} . \ \ \ \ \ \ (7)

5) The matrix \Lambda^{-1} can be constructed either from its elements as given by Eq(5), or directly by simple matrix multiplication as given by Eq(6). To check that I did not make a mistake, let us do them both. We write (5) as (\Lambda^{-1})^{\sigma}{}_{\rho} = \eta_{\mu\rho} \ \Lambda^{\mu}{}_{\nu} \ \eta^{\nu\sigma} , and evaluate the relevant matrix elements using the fact that \eta = \eta^{-1} = \mbox{diag} (1, -1, -1, -1):

(\Lambda^{-1})^{0}{}_{k} = (-1) \ \Lambda^{k}{}_{0} \ (+1) = - \Lambda^{k}{}_{0} . In other words, we have the following equalities (\Lambda^{-1})^{0}{}_{k} = \Lambda_{k}{}^{0} = - \Lambda^{k}{}_{0} = - (\Lambda^{T})^{0}{}_{k} .

Next, we calculate the space-space matrix elements (\Lambda^{-1})^{i}{}_{k} = \eta_{\mu k} \ \Lambda^{\mu}{}_{\nu} \ \eta^{\nu i} = (-1) \ \Lambda^{k}{}_{i} \ (-1) . Thus (\Lambda^{-1})^{i}{}_{k} = \Lambda_{k}{}^{i} = \Lambda^{k}{}_{i} = (\Lambda^{T})^{i}{}_{k} . And similarly, we find (\Lambda^{-1})^{k}{}_{0} = \Lambda_{0}{}^{k} = - \Lambda^{0}{}_{k} = - (\Lambda^{T})^{k}{}_{0} .

From these matrix elements, we can write

\Lambda^{-1} = \begin{pmatrix}<br /> <br /> \Lambda^{0}{}_{0} &amp; -\Lambda^{1}{}_{0} &amp; -\Lambda^{2}{}_{0} &amp; -\Lambda^{3}{}_{0} \\<br /> <br /> -\Lambda^{0}{}_{1} &amp; \Lambda^{1}{}_{1} &amp; \Lambda^{2}{}_{1} &amp; \Lambda^{3}{}_{1} \\<br /> <br /> -\Lambda^{0}{}_{2} &amp; \Lambda^{1}{}_{2} &amp; \Lambda^{2}{}_{2} &amp; \Lambda^{3}{}_{2} \\<br /> <br /> -\Lambda^{0}{}_{3} &amp; \Lambda^{1}{}_{3} &amp; \Lambda^{2}{}_{3} &amp; \Lambda^{3}{}_{3}<br /> <br /> \end{pmatrix} . \ \ \ (8)

Indeed, the same matrix can be obtained from Eq(6) as follow: From

\Lambda = \begin{pmatrix}<br /> \Lambda^{0}{}_{0} &amp; \Lambda^{0}{}_{1} &amp; \Lambda^{0}{}_{2} &amp; \Lambda^{0}{}_{3} \\<br /> \Lambda^{1}{}_{0} &amp; \Lambda^{1}{}_{1} &amp; \Lambda^{1}{}_{2} &amp; \Lambda^{1}{}_{3} \\<br /> \Lambda^{2}{}_{0} &amp; \Lambda^{2}{}_{1} &amp; \Lambda^{2}{}_{2} &amp; \Lambda^{2}{}_{3} \\<br /> \Lambda^{3}{}_{0} &amp; \Lambda^{3}{}_{1} &amp; \Lambda^{3}{}_{2} &amp; \Lambda^{3}{}_{3} \end{pmatrix} ,

and
\eta = \begin{pmatrix}<br /> 1 &amp; 0 &amp; 0 &amp; 0 \\<br /> 0 &amp; -1 &amp; 0 &amp; 0 \\<br /> 0 &amp; 0 &amp; -1&amp; 0 \\<br /> 0 &amp; 0 &amp; 0 &amp; -1 \end{pmatrix} ,

We fined

\eta \ \Lambda \ \eta = \begin{pmatrix}<br /> \Lambda^{0}{}_{0} &amp; -\Lambda^{0}{}_{1} &amp; -\Lambda^{0}{}_{2} &amp; -\Lambda^{0}{}_{3} \\<br /> -\Lambda^{1}{}_{0} &amp; \Lambda^{1}{}_{1} &amp; \Lambda^{1}{}_{2} &amp; \Lambda^{1}{}_{3} \\<br /> -\Lambda^{2}{}_{0} &amp; \Lambda^{2}{}_{1} &amp; \Lambda^{2}{}_{2} &amp; \Lambda^{2}{}_{3} \\<br /> -\Lambda^{3}{}_{0} &amp; \Lambda^{3}{}_{1} &amp; \Lambda^{3}{}_{2} &amp; \Lambda^{3}{}_{3}<br /> \end{pmatrix} . \ \ \ (9)

Clearly, you obtain Eq(8) by taking the transpose of Eq(9).

6) Here is an exercise for you (probably the most important thing you should do). Use the matrix \eta \ \Lambda \ \eta, as given in Eq(9), to show that (\eta \ \Lambda \ \eta) \Lambda^{T} = I_{4 \times 4} .
 
  • Like
Likes   Reactions: vanhees71
  • #12
fayled said:
I really need to be able to understand how to deal with the raising and lowering of indices.
If this is your goal then I would recommend not trying to mix tensor and matrix notation. Focus entirely on the raising and lowering tensor indices, and drop the matrix and matrix transposition stuff entirely.

IMO, the best approach will be to take 2D tensors and expand them out explicitly writing the sums.
 
Last edited:
  • Like
Likes   Reactions: vanhees71 and Orodruin

Similar threads

  • · Replies 9 ·
Replies
9
Views
3K
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 7 ·
Replies
7
Views
1K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 5 ·
Replies
5
Views
1K
  • · Replies 30 ·
2
Replies
30
Views
7K
  • · Replies 9 ·
Replies
9
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K