Imaginary eigenvalues of gamma matrices

In summary, the conversation discusses a statement in David Tong's notes on QFT where it is claimed that the square of a matrix being positive and negative respectively means real and imaginary eigenvalues. Further discussion involves trying to obtain a specific equation involving exponential expansion and factorizing out gammas, with some difficulties arising due to incorrect expansions. Eventually, the issue is resolved and it is determined that the incorrect expansion was the source of the problem.
  • #1
center o bass
560
2
Hi! I'm reading David Tong's notes on QFT and I'm now reading on the chapter on the dirac equation

http://www.damtp.cam.ac.uk/user/tong/qft/four.pdf

and I stumbled across a statement where he claims that

[tex] (\gamma^0)^2 = 1 \ \ \Rightarrow \text{real eigenvalues}[/tex]

while

[tex] (\gamma^i)^2 = -1 \ \ \Rightarrow \text{imaginary eigenvalues}.[/tex]

I'm a bit rusty on my linear algebra and just wondered why this is necessarily true. Why does the square of a matrix being positive and negative respectively mean real and imaginary eigenvalues?
 
Last edited:
Physics news on Phys.org
  • #2
Suppose x is an eigenvector of the matrix A with eigenvalue a, so that Ax = ax. Then A^2 x = a^2 x.

Suppose A^2 x = x for all x (including the eigenvectors of A). Then in view of the above, all eigenvalues of A satisfy a^2 = 1. So the possible eigenvalues are +1 and -1.

Now suppose A^2 x = -x for all x. Then all eigenvalues of A satisfy a^2 = -1. So the possible eigenvalues are +i and -i.
 
  • #3
Ah thanks a lot! :)
 
  • #4
The_Duck said:
Suppose x is an eigenvector of the matrix A with eigenvalue a, so that Ax = ax. Then A^2 x = a^2 x.

Suppose A^2 x = x for all x (including the eigenvectors of A). Then in view of the above, all eigenvalues of A satisfy a^2 = 1. So the possible eigenvalues are +1 and -1.

Now suppose A^2 x = -x for all x. Then all eigenvalues of A satisfy a^2 = -1. So the possible eigenvalues are +i and -i.
Maybe you could help me with the next thing I'm stuck on? :)

After obtaining that

[tex] (S^{\mu \nu})^\dagger = \frac{1}4 [\gamma^\mu, \gamma^\nu]^\dagger = - \gamma^0 S^{\mu \nu} \gamma^0[/tex]

D. Tong claims that

[tex] \exp ( \frac{1}{2} \Omega_{\mu \nu} (S^{\mu \nu})^\dagger ) = \gamma^0 \exp( -\frac{1}{2} \Omega_{\mu \nu} S^{\mu \nu} ) \gamma^0.[/tex]

I tried obtaining this by doing an exponential expansion and factorizing out the gammas, plus the property that
[tex](\gamma^0)^2 = 1[/tex]

but I get problems with the even terms in the expansion.
 
Last edited:
  • #5
center o bass said:
I tried obtaining this by doing an exponential expansion and factorizing out the gammas, plus the property that
[tex](\gamma^0)^2 = 1[/tex]

but I get problems with the even terms in the expansion.

What problems? Could you write out what you think the expansions of the left and right-hand sides of the equations are?
 
  • #6
The_Duck said:
What problems? Could you write out what you think the expansions of the left and right-hand sides of the equations are?

Ofcourse.

[tex] \exp(\gamma^0 (-\frac{1}2 \Omega_{\mu \nu} S^{\mu \nu}) \gamma^0)
= 1 + \gamma^0 A \gamma^0 + \frac{1}{2!} (\gamma^0)^2 A^2 (\gamma^0)^2 + \frac{1}{3!}(\gamma^0)^3 A^3 (\gamma^0)^3 + \ldots = \gamma^0 ( 1 + A + \frac{1}{2!} \gamma^0 A \gamma^0 + \frac{1}{3!} A^3 + \ldots )\gamma^0[/tex]

where i have defined

[tex]A = -\frac{1}2 \Omega_{\mu \nu} S^{\mu \nu}[/tex]
 
  • #7
OK, two things:

1) Your expansion of the LHS isn't right because gamma-0 and A don't necessarily commute. If A and B are not-necessarily commuting matrices or operators, then

exp(AB) = 1 + AB + (1/2!)ABAB + (1/3!)ABABAB + ...

(recall that, for instance, (AB)^2 = (AB)*(AB)). But you seem to be assuming the incorrect form

exp(AB) = 1 + AB + (1/2!)AABB + (1/3!)AAABBB + ...

which is only equivalent to the correct form if A and B commute (because (AB)(AB) != AABB unless A and B commute). (Of course, you need to apply the equivalent statement for the exponential of the product of three matrices, not just two).2) Why are there some gamma-0's inside the parentheses in your expansion of the RHS?
 
  • #8
The_Duck said:
OK, two things:

1) Your expansion of the LHS isn't right because gamma-0 and A don't necessarily commute. If A and B are not-necessarily commuting matrices or operators, then

exp(AB) = 1 + AB + (1/2!)ABAB + (1/3!)ABABAB + ...

(recall that, for instance, (AB)^2 = (AB)*(AB)). But you seem to be assuming the incorrect form

exp(AB) = 1 + AB + (1/2!)AABB + (1/3!)AAABBB + ...

which is only equivalent to the correct form if A and B commute (because (AB)(AB) != AABB unless A and B commute). (Of course, you need to apply the equivalent statement for the exponential of the product of three matrices, not just two).


2) Why are there some gamma-0's inside the parentheses in your expansion of the RHS?

Ah, thank you again. The problem lied in the wrong expansion :) The gamma-0's were there due to this wrong expansion. When i tried to factor out gamma-0's to the left and right of the sum gamma-0's remained at the even terms.
 

What are imaginary eigenvalues of gamma matrices?

Imaginary eigenvalues of gamma matrices are complex numbers that are solutions to the characteristic equation of a gamma matrix. They can be found by solving the equation det(A-λI) = 0, where A is the gamma matrix and λ is the eigenvalue.

What is the significance of imaginary eigenvalues of gamma matrices?

Imaginary eigenvalues of gamma matrices play a crucial role in theoretical physics, specifically in the study of quantum field theory. They are associated with the spinor representations of the Lorentz group and are essential for understanding the behavior of fermions in relativistic systems.

Can a gamma matrix have only imaginary eigenvalues?

No, a gamma matrix can have a mix of real and imaginary eigenvalues. This is due to the fact that gamma matrices are Hermitian, meaning they are equal to their conjugate transpose. Therefore, their eigenvalues must be either real or occur in complex conjugate pairs.

How are imaginary eigenvalues of gamma matrices related to the Dirac equation?

The Dirac equation, which describes the behavior of fermions in a relativistic system, uses gamma matrices to represent the spin of the particles. The imaginary eigenvalues of these gamma matrices correspond to the energy levels of the fermions, and are crucial in solving the Dirac equation and understanding the behavior of these particles.

Are there any practical applications of studying imaginary eigenvalues of gamma matrices?

Yes, the study of imaginary eigenvalues of gamma matrices has practical applications in various fields such as particle physics, condensed matter physics, and quantum computing. They are essential for understanding the behavior of fermions and can help in predicting and analyzing their properties in different systems.

Similar threads

Replies
4
Views
3K
Replies
5
Views
2K
Replies
1
Views
627
  • Quantum Physics
Replies
5
Views
2K
Replies
4
Views
999
Replies
1
Views
2K
Replies
9
Views
1K
  • High Energy, Nuclear, Particle Physics
Replies
6
Views
2K
  • Quantum Physics
Replies
6
Views
2K
Replies
4
Views
3K
Back
Top