Confused by proof of Lorentz properties from invariance of interval

Click For Summary

Discussion Overview

The discussion revolves around the proof of Lorentz properties derived from the invariance of the spacetime interval. Participants explore the implications of transformations that preserve the interval and the mathematical justification behind certain algebraic steps in these proofs, particularly focusing on the equality of matrices involved in the transformation.

Discussion Character

  • Technical explanation
  • Debate/contested
  • Mathematical reasoning

Main Points Raised

  • One participant expresses confusion about the justification for concluding that two matrices are equal based on their relation via the dot product for any vector x, questioning the validity of this step.
  • Another participant points out the issue of canceling factors of zero in equations, highlighting the potential for contradictions.
  • Several participants discuss the properties of rotation matrices and their implications for the invariance of vector lengths under transformations, noting that not all matrices behave similarly.
  • A participant provides an example from a textbook to illustrate the invariance of the interval and discusses the subtleties involved in proving the result in more detail.
  • There is a discussion about the necessity of using properties of the metric g on both sides of the equation to establish the equality of matrices, with one participant detailing a method to show that the entries of the matrices must be equal.
  • Another participant reflects on the intuitive nature of the result while acknowledging the complexity of the argument used to arrive at it, expressing a desire for clearer mathematical understanding.

Areas of Agreement / Disagreement

Participants generally agree on the importance of the properties of the metric and the implications of transformations preserving the interval, but there is no consensus on the clarity and validity of the algebraic steps involved in proving the equality of matrices.

Contextual Notes

Some participants note that the arguments may not hold in lower dimensions, such as 2D, and emphasize the need for careful consideration of the properties of the matrices involved.

Who May Find This Useful

This discussion may be useful for those studying the mathematical foundations of relativity, particularly in understanding the subtleties of transformations and the invariance of the spacetime interval.

superbro
Messages
5
Reaction score
0
I've seen a few short proofs that if that some transformation \Lambda preserves the spacetime interval, then

\Lambda^\top g \Lambda = g

where g is the spacetime metric.

They have all relied on an argument using some simple algebra to show that

(\Lambda^\top g \Lambda) x \cdot x = g x \cdot x

and since this is true for *any* x, it must be true that

\Lambda^\top g \Lambda = g

This confuses me. I don't see how this "since it's true for any x" step is justified.

For example,

\left( \begin{array}{ccc} 0 & -1 \\ 1 & 0 \end{array} \right) x \cdot x = \left( \begin{array}{ccc} 0 & 0 \\ 0 & 0 \end{array} \right) x \cdot x

for any x, and by the same argument, since it is true for *any* x, it must be true that the arrays are equal and -1 = 0 = 1. Did I just break math?

Intuitively, I don't think I could do the same trick to produce a counterexample in 3+ dimensions, but this seems like kind of a subtle point to sweep under the rug in a proof.

Does anyone have a more detailed version of this argument that would (hopefully) make more sense to me?
 
Physics news on Phys.org
You cannot cancel a factor of zero from an equation because it can give contradictions

0 x = 0 y

therefore

x = y

for all x, y.
 
If it's the zero matrix that is making you think of division by zero, it's also true for a 90 degree rotation in the opposite direction.

\left( \begin{array}{ccc} 0 & -1 \\ 1 & 0 \end{array} \right) x \cdot x = \left( \begin{array}{ccc} 0 & 1 \\ -1 & 0 \end{array} \right) x \cdot x

Those matrices aren't equal, either.

To be clear, my confusion is that in several proofs the claim is made that two matrices are equal because of their relation via the dot product of any vector. It's clearly not true for all matrices as the above counterexample shows, so what properties of these particular matrices are exploited to make it true for them?
 
superbro said:
If it's the zero matrix that is making you think of division by zero, it's also true for a 90 degree rotation in the opposite direction.

\left( \begin{array}{ccc} 0 & -1 \\ 1 & 0 \end{array} \right) x \cdot x = \left( \begin{array}{ccc} 0 & 1 \\ -1 & 0 \end{array} \right) x \cdot x

Those matrices aren't equal, either.

The equation you've written is an example of the length of vectors being invariant under spatial rotations, which form a group. The rotation matrices are unimodular with orthogonal rows and columns.

To be clear, my confusion is that in several proofs the claim is made that two matrices are equal because of their relation via the dot product of any vector.
That is confusing.
 
Here's an example of this from "Problem Book in Quantum Field Theory" by Voda Radovanovic...

The square of the length of a four-vector, x is x^2 = g_{\mu \nu} x^{\mu} x^{\nu}. By substituting x'^{\mu} = \Lambda^{\mu}_{\rho}x^{\rho} into the condition x'^2 = x^2 one obtains:

g_{\mu\nu}\Lambda^{\mu}_{\rho}\Lambda^{\nu} _{\sigma} x^{\rho}x^{\sigma} = g_{\rho\sigma}x^{\rho}x^{\sigma}

Since (1.1) is valid for any vector x \in M_4, we get \Lambda^{\mu}_{\rho}g_{\mu\nu}\Lambda^{\nu} _{\sigma} = g_{\rho\sigma}.

He does specifically mention x is in M4, which rules out my counterexamples, since they only work in 2D. I presume this is a well-known and obvious argument in order to be the solution to the very first problem of the book, but it seems pretty subtle to me if I want to prove the result in more detail...
 
That seems logical. If \Lambda is length preserving then g, which defines the interval should be invariant under it.
 
Mentz114 said:
The equation you've written is an example of the length of vectors being invariant under spatial rotations, which form a group. The rotation matrices are unimodular with orthogonal rows and columns.

Actually, thinking about this a bit, I think it's different than a statement that vector length or dot product is preserved.

For example, for any orthogonal matrices A and B, (Ax) \cdot (Ax) = x \cdot x = (Bx) \cdot (Bx), which is directly related to the length being preserved.

But (Ax) \cdot x = (Bx) \cdot x fails for most orthogonal matrices. (As an example, check the identity versus a 90 degree rotation where you end up with x * x = 0 which clearly can't be true for all x).

In the 2D rotational case, for example, A and B would have to rotate x by the same absolute value of theta for their projections onto the original x to be equal. I can't think of any examples of nonequal rotations A and B from higher dimensions where this would hold, since any 3D+ rotation A has an eigenvector on its axis (with A x * x = x * x) which is not an eigenvector of any other rotation B (so B x * x < x * x). I think this must ultimately be related to why this argument works, but there's a fair amount of linear algebra behind it.

Mentz114 said:
That seems logical. If \Lambda is length preserving then g, which defines the interval should be invariant under it.

I agree it's an intuitive result, and I have no doubt it's true. I just think the argument used to get there (If A x * x = B x * x for all x in M4, then A = B) is not so obvious, and in fact it's not even true in M2.

Maybe it's just part of me getting used to physics to accept the more hand-wavy "it seems reasonable" kind of argument even if I don't fully understand the mathematics behind it. It tends to leave me terribly confused, though.
 
If Ax=Bx then (Ax).x = (Bx).x. I don't know if it's a sufficient condition though.
 
Well, I managed to figure this out after a bit more pondering. It's necessary to use the properties of g on both sides of the equation.

(My thought that maybe it worked because of some nutty property of rotations in dimensions greater than 2 was crap, since e.g. \left( \begin{array}{ccc} 0 &amp; -1 &amp; 0 &amp; 0 \\ 1 &amp; 0 &amp; 0 &amp; 0 \\ 0 &amp; 0 &amp; 0 &amp; -1 \\ 0 &amp; 0 &amp; 1 &amp; 0 \end{array} \right) x \cdot x = 0 for all x. [Stupid even dimensions.])

I did it in more detail this way...

For arbitrary matrices A and B, where (Ax) \cdot x = (Bx) \cdot x for all x, you can show that their entries on the main diagonal must be equal by picking each unit vector in the basis and substituting it for x.

For example, picking x = \left( \begin{matrix} 1 \\ 0 \\ 0 \\ 0 \end{matrix} \right) in A_{ij} x_i x_j = B_{ij} x_i x_j yields A_{00} = B_{00}

Now that we know the diagonals are equal, you can show by picking each vector in the basis with two 1's in it (e.g. (0, 1, 1, 0)) that A_{ij} + A_{ji} = B_{ij} + B_{ji}.

For example, picking x = \left( \begin{matrix} 1 \\ 0 \\ 1 \\ 0 \end{matrix} \right) in A_{ij} x_i x_j = B_{ij} x_i x_j yields A_{00} + A_{02} + A_{20} + A_{22} = B_{00} + B_{02} + B_{20} + B_{22} which implies A_{02} + A_{20} = B_{02} + B_{20} since A_{nn} = B_{nn} for all indices n.

Now going from the general case to the special case where B = g, and i ≠ j, then A_{ij} + A_{ji} = g_{ij} + g_{ji} = 0, so the entries off diagonal are antisymmetric.

Finally, we know that \Lambda^{\top} g \Lambda is symmetric, since g is symmetric and (\Lambda^{\top} g \Lambda)^{\top} = \Lambda^{\top} g^{\top} \Lambda = \Lambda^{\top} g \Lambda.

The only way that the off-diagonal entries can be both symmetric and antisymmetric is if they are zero, and the on-diagonal entries must be equal, so \Lambda^{\top} g \Lambda = g as expected.

The same proof seems like it would work for any diagonal matrix substituted for g, so there's probably some slightly more sophisticated linear algebra that would make short work of this through some decomposition or something.
 

Similar threads

Replies
3
Views
2K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 9 ·
Replies
9
Views
2K
  • · Replies 101 ·
4
Replies
101
Views
7K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 146 ·
5
Replies
146
Views
12K
  • · Replies 9 ·
Replies
9
Views
2K
  • · Replies 123 ·
5
Replies
123
Views
8K