Graduate Killing vectors corresponding to the Lorentz transformations

wafelosek
Messages
2
Reaction score
0
Hi everyone! I have a problem with one thing.

Let's consider the Lorentz group and the vicinity of the unit matrix. For each ##\hat{L}##
from such vicinity one can prove that there exists only one matrix ##\hat{\epsilon}## such that ##\hat{L}=exp[\hat{\epsilon}]##. If we take ##\epsilon^{μν}##, μ<ν, as the parameters on the Lorentz group in a vicinity of the unit matrix, then we can compute the corresponding Killing vectors as ##\xi ^μ_{αβ}=\frac{∂x′^μ}{∂ϵ^{αβ}}(\hat{ϵ}=0)## where ##x′^μ=L^μ_νx^ν##. Here is my problem: during the computations there is one line that I do not get, namely: ##\cfrac{∂L^{μν}}{∂ϵ^{αβ}}(\hat{ϵ}=0)=δ^μ_αδ^ν_β−δ^μ_βδ^ν_α##. The ## (\epsilon^{μν}) ## matrix is antisymmetric, so that is why we get the difference of the Kronecer delta?

Thank you in advance!

MW
 
Physics news on Phys.org
Yes. That partial derivative is defined such that the (a)symmetries of alpha and beta are the same on the left and right hand side. You can also check explicitly by taking e.g. alpha=0, beta=1.
 
Because that page treats a general matrix where the components are all independent, this is not the case for the Lorentz group generators.
 
So, to extend a little bit on Orodruin's remark: in this case, the antisymmetry in alpha and beta indicates that not all the components are independent. So if you then define the functional derivative of a tensor with respect to that epsilon, you get an expression containing again indices alpha and beta. By definition, these components are then still considered to be dependent. This means the (a)symmetry carries over in the functional derivative.

Or, rephrased: the (a)symmetry has to be respected on the left and right hand side in the definition of the functional derivative.
 
To expand a little on the little expansion, just with a hands-on example, consider the Lorentz transformations in 1+1D Minkowski space, where you would have
$$
\hat \epsilon = \begin{pmatrix}0 & \epsilon_{01} \\ - \epsilon_{01} & 0 \end{pmatrix}.
$$
Differentiating ##\hat L = \exp(\hat\epsilon)## with respect to ##\epsilon_{01}## would give you
$$
\frac{\partial \hat L}{\partial \epsilon_{01}} = \frac{\partial \hat \epsilon}{\partial \epsilon_{01}} \hat L,
$$
which evaluated in ##\epsilon = 0## leads to
$$
\left. \frac{\partial \hat L}{\partial \epsilon_{01}}\right|_{\epsilon = 0} = \begin{pmatrix}0 & 1 \\ -1 & 0\end{pmatrix}.
$$
You will find that this satisfies your relation.
 

Similar threads

  • · Replies 11 ·
Replies
11
Views
9K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 20 ·
Replies
20
Views
5K
  • · Replies 30 ·
2
Replies
30
Views
3K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 2 ·
Replies
2
Views
3K
  • · Replies 10 ·
Replies
10
Views
3K
  • · Replies 6 ·
Replies
6
Views
2K