How can the dual tensors derivation be achieved using rotation matrices?

Click For Summary
SUMMARY

The discussion focuses on deriving the relationship between dual tensors and rotation matrices, specifically within the context of group theory as presented in A Zee's "Group Theory in a Nutshell for Physicists." The key equation discussed is $$\epsilon^{ijk\cdots n}R^{ip}R^{jq}=\epsilon^{pqr\cdots s}R^{kr}\cdots R^{ns}$$, where ##R \in SO(N)## represents an N-dimensional rotation. Participants elaborate on the determinant relation and the application of group relations, ultimately leading to a clearer understanding of the derivation process for both 3-dimensional and N-dimensional cases.

PREREQUISITES
  • Understanding of dual tensors and their properties
  • Familiarity with rotation matrices in the context of special orthogonal groups (SO(N))
  • Knowledge of the antisymmetric symbol and its applications in tensor calculus
  • Basic grasp of matrix determinants and their significance in linear algebra
NEXT STEPS
  • Study the properties of the antisymmetric symbol in tensor analysis
  • Learn about the structure and applications of special orthogonal groups, specifically SO(3) and SO(N)
  • Explore the derivation of determinant properties in relation to rotation matrices
  • Investigate the implications of group relations in tensor equations
USEFUL FOR

This discussion is beneficial for physicists, mathematicians, and students studying advanced topics in group theory, particularly those interested in the applications of dual tensors and rotation matrices in theoretical physics.

euphoricrhino
Messages
23
Reaction score
14
Hello,
I'm reading Group Theory in a nutshell for physicist by A Zee. When he introduced Dual tensors (pp 192), he made a claim with a light hint, and I have had great trouble deriving this claim, any help would be appreciated -

Let ##R \in SO(N)## be an ##N##-dimensional rotation, then the following is true
$$
\epsilon^{ijk\cdots n}R^{ip}R^{jq}=\epsilon^{pqr\cdots s}R^{kr}\cdots R^{ns}
$$
(where ##\epsilon## is the antisymmetric symbol and the above uses repeated index summing convention).
The hint was to use the ##N\times N## matrix determinant
$$
\epsilon^{ijk\cdots n}R^{ip}R^{jq}R^{kr}\cdots R^{ns}=\epsilon^{pqr\cdots s} \mbox{det}R=\epsilon^{pqr\cdots s} \quad(\mbox{since }R\mbox{ is special})
$$
and multiply it "by a bunch of ##R^T##s carrying appropriate indices".

I have tried to understand the claim with ##N=3## which I think is the cross product relation, but I couldn't see how that could be obtained by involving ##R^T##, and how it could be extended to ##N## dimensions.

Thanks for the help!
 
  • Like
Likes   Reactions: yanpengju
Physics news on Phys.org
I don't have Zee's book but does he talk about,

##\epsilon^{ijk\cdots n}R^{ip}R^{jq}\cdots R^{ns} = \det(R)\epsilon^{pqr\cdots s}##

which I think this follows from the definition of the determinate. Now, for ##SO(N)## one has ##\det(R) = 1##. From this one applies the group relation ##R^{ij}R^{ik} = \delta^{jk}## to both sides twice and you have it. Hope this helps.
 
Thanks for the reply!

However I must be missing something really obvious, I don't see how to "apply the group relation ##R^{ij}R^{ik}=\delta^{jk}## to both sides" of the determinant equality.

The LHS of the determinant relation is a sum of ##N!## terms, each of which is a product whose factors don't share any index. Can you kindly elaborate for the ##N=3## case here?

From determinant equality
$$
\epsilon^{ijk}R^{ip}R^{jq}R^{kr}=\epsilon^{pqr}
$$
where ##(pqr)## is a given permutation
how to derive (for any given ##p,q,k##)
$$
\epsilon^{ijk}R^{ip}R^{jq}=\epsilon^{pqr}R^{kr}
$$

Thank you very much!
 
In matrix form

##R^TR = RR^T = I##

so, one also has the relation,

##R^{ip}R^{kp} = \delta^{ik}##
 
  • Like
Likes   Reactions: euphoricrhino
I finally figured it out, it's actually quite simple, but all the symbols there have been distracting.

The determinant relation
$$
\epsilon^{ijk\cdots n}R^{ip}R^{jq}R^{kr}\cdots R^{ns}=\epsilon^{pqr\cdots s}
$$
can be viewed as an inner product relation
$$
v^nR^{ns}=\epsilon^{pqr\cdots s}
$$
where ##v^n## is defined by
$$
v^n=\epsilon^{ijk\cdots n}R^{ip}R^{jq}R^{kr}\cdots
$$
Since columns ##R^{\cdot s}## form an orthonormal basis of the ##N##-dimensional space, the inner product relation above actually gives the decomposition of vector ##v## into this basis, i.e.
$$
v=\epsilon^{pqr\cdots s}R^{\cdot s}
$$
Taking the ##n##-th component of this yields
$$
\epsilon^{ijk\cdots n}R^{ip}R^{jq}R^{kr}\cdots=v^n=\epsilon^{pqr\cdots s}R^{ns}
$$

Now we just need to repeat the same argument on all the other ##R##s on the left until only ##R^{ip}## and ##R^{jq}## were left
 
I have been struggling with this question for days and came across here. I don't understand euphoricrhino's final solution, but instead got my own derivations here. Hope it helps anyone who might Google here in the future.

The key obstacle here is that Zee didn't explicitly specify the indices to be summed on. That's Einstein's fault anyway, for his invention of summation convention messes all it up. :smile:

What we have and want to prove are:

$$\begin{aligned}
\epsilon^{ijk\cdots n}R^{ip}R^{jq}R^{kr}\cdots R^{ns} &= \epsilon^{pqr\cdots s} \\
\epsilon^{ijk\cdots n}R^{ip}R^{jq} &= \epsilon^{pqr\cdots s}R^{kr}\cdots R^{ns}
\end{aligned}$$

Let's make the summation operation explicit:

$$\begin{aligned}
\sum_{ijk\cdots n}\epsilon^{ijk\cdots n}R^{ip}R^{jq}R^{kr}\cdots R^{ns} &= \epsilon^{pqr\cdots s} \\
\sum_{ij}\epsilon^{ijk\cdots n}R^{ip}R^{jq} &= \sum_{r\cdots s}\epsilon^{pqr\cdots s}R^{kr}\cdots R^{ns}
\end{aligned}$$

(You may stop here and continue on your own derivation if you like.)

We introduce some fixed indices ##k^{\prime},\cdots,n^{\prime}## and multiply the following same terms on both sides like this:

$$\sum_{ijk\cdots n}\epsilon^{ijk\cdots n}R^{ip}R^{jq}R^{kr}\cdots R^{ns}\cdot R^{k^{\prime}r}\cdots R^{n^{\prime}s}=\epsilon^{pqr\cdots s}\cdot R^{k^{\prime}r}\cdots R^{n^{\prime}s}$$

Summing both sides over indices ##r\cdots s##, then the left side becomes:

$$\begin{aligned}
& \sum_{r\cdots s}\sum_{ijk\cdots n}\epsilon^{ijk\cdots n}R^{ip}R^{jq}R^{kr}\cdots R^{ns}\cdot R^{k^{\prime}r}\cdots R^{n^{\prime}s} \\
=& \sum_{ij}R^{ip}R^{jq}\sum_{k\cdots}\sum_{r\cdots}R^{kr}\cdots R^{k^{\prime}r}\cdots\sum_{n}\epsilon^{ijk\cdots n}\sum_{s}R^{ns}R^{n^{\prime}s} \\
=& \sum_{ij}R^{ip}R^{jq}\sum_{k\cdots}\sum_{r\cdots}R^{kr}\cdots R^{k^{\prime}r}\cdots\sum_{n}\delta^{nn^{\prime}}\epsilon^{ijk\cdots n} \\
=& \sum_{ij}R^{ip}R^{jq}\sum_{k\cdots}\sum_{r\cdots}\epsilon^{ijk\cdots n^{\prime}}R^{kr}\cdots R^{k^{\prime}r}\cdots \\
=& \cdots \\
=& \sum_{ij}\epsilon^{ijk^{\prime}\cdots n^{\prime}}R^{ip}R^{jq}
\end{aligned}$$

Rename ##k^{\prime},\cdots,n^{\prime}## to ##k,\cdots,n## and remove the explicit summation, we get the equation finally:

$$\begin{aligned}
\sum_{ij}\epsilon^{ijk^{\prime}\cdots n^{\prime}}R^{ip}R^{jq} &= \sum_{r\cdots s}\epsilon^{pqr\cdots s}R^{k^{\prime}r}\cdots R^{n^{\prime}s} \\
\sum_{ij}\epsilon^{ijk\cdots n}R^{ip}R^{jq} &= \sum_{r\cdots s}\epsilon^{pqr\cdots s}R^{kr}\cdots R^{ns} \\
\epsilon^{ijk\cdots n}R^{ip}R^{jq} &= \epsilon^{pqr\cdots s}R^{kr}\cdots R^{ns}
\end{aligned}$$
 

Similar threads

  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 1 ·
Replies
1
Views
1K
Replies
14
Views
4K
  • · Replies 6 ·
Replies
6
Views
3K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 1 ·
Replies
1
Views
5K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 4 ·
Replies
4
Views
4K
  • · Replies 1 ·
Replies
1
Views
2K