Ballentine-sections 3.2 and 3.3

  • Thread starter Thread starter WannabeNewton
  • Start date Start date
WannabeNewton
Science Advisor
Gold Member
Messages
5,848
Reaction score
552
Hi! In section 3.2 of Ballentine, the author considers the most general effect of a Galilean transformation on space-time coordinates ##\{\mathbf{x}, t \}## (3.5) and denotes such a transformation by ##\tau = \tau (R,\mathbf{a},\mathbf{v}, s)## where ##R## is a rotation matrix, ##\mathbf{a}## is a space displacement, ##\mathbf{v}## is a 3-velocity, and ##s## is a time displacement. Based on his later notation directly below, ##\tau \{\mathbf{x}, t \} = \{\mathbf{x'}, t' \}## it seems like ##\tau## is a map ##\tau : \mathbb{R}^{3}\times \mathbb{R} \rightarrow \mathbb{R}^{3}\times \mathbb{R}##.

He then says that ##\tau## must have corresponding to it a unitary transformation ##U(\tau)## with ##|\psi \rangle \rightarrow U|\psi \rangle ## and ##A\rightarrow UAU^{-1}## as usual (top of p.67). Furthermore, the Galilei group has ten parameters, ##s_{\mu}## (the 3 rotation angles, 3 velocity components, 3 spatial translations, and 1 temporal translation), and associated with each ##s_{\mu}## is a Hermitian generator ##K_{\mu}## such that ##U(s_{\mu}) = e^{i s_{\mu} K_{\mu}}##. At the inception of section 3.3 Ballentine notes that a unitary transformation ##U(\tau)## corresponding to a space-time transformation ##\tau## will have the most general form ##U(\tau) = \prod_{\mu=1}^{10}U(s_{\mu}) = \prod_{\mu=1}^{10}e^{is_{\mu}K_{\mu}}## (3.9).

What's confusing me is the functional form of ##U(\tau)## when compared to the functional forms of the ##U(s_{\mu})##. Each ##U(s_{\mu})## corresponds to a one-parameter family of unitary operators but ##\tau## seems to be an endomorphism of Galilean space-time so what does ##U(\tau)## correspond to? At first I thought it corresponded to a ten-parameter family of unitary operators ##U(s_1,...,s_{10})## since ##\tau = \tau (R,\mathbf{a},\mathbf{v}, s) = \tau(s_1,...,s_{10})## but the fact that he subsequently writes ##\tau \{\mathbf{x}, t \} = \{\mathbf{x'}, t' \}## has me confused as to what ##\tau## is as a mathematical object, what explicit functional dependence it has on the ##s_{\mu}##, and if I can actually call ##U(\tau)## a ten-parameter family of unitary operators.

Thanks in advance for clearing this up!
 
Physics news on Phys.org
I took the curly brackets to mean doing the transform on the something with the coordinates of {x,t} gives you something with the coordinates of {x',t'}. But perhaps I am not reading enough into it.
 
I don't have my Ballentine home with me, but it seems like

WannabeNewton said:
Based on his later notation directly below, ##\tau \{\mathbf{x}, t \} = \{\mathbf{x'}, t' \}## it seems like ##\tau## is a map ##\tau : \mathbb{R}^{3}\times \mathbb{R} \rightarrow \mathbb{R}^{3}\times \mathbb{R}##.

Yes.

Think of each real number ##s_\mu## as corresponding to a specific ##\tau##. For example: ##s_7## is the specific ##\tau## that translates by ##s_7## along ##x##, while leaving all other coordinates untouched; ##s_{10}## is the specific ##\tau## that translates by ##s_{10}## temporally, while leaving all other coordinates untouched; etc.

Also, there is no sum on ##s_\mu K_\mu##.
 
George Jones said:
Yes.

Its been a while since I went through that chapter but as far as I recall that's it.

The thing I recall is it took me a little while to fully understand - its a bit dense.

The other thing I recall is there is a bit of thought required about exactly why a free particle forms an irreducible set.

The other thing is the exact key assumption being made in quantitization - it took me a while to nut that one out.

Thanks
Bill
 
George Jones said:
I don't have my Ballentine home with me, but it seems like



Yes.

Think of each real number ##s_\mu## as corresponding to a specific ##\tau##. For example: ##s_7## is the specific ##\tau## that translates by ##s_7## along ##x##, while leaving all other coordinates untouched; ##s_{10}## is the specific ##\tau## that translates by ##s_{10}## temporally, while leaving all other coordinates untouched; etc.

Also, there is no sum on ##s_\mu K_\mu##.

There is in 3.10
U = I + iƩsK summing over μ from 1 to 10
 
Expanding 3.9 gives

$$\begin{align}
U &=e^{is_{1}K_{1}}e^{is_{2}K_{2}}\cdots e^{is_{10}K_{10}}\\
&=\left( 1+is_{1}K_{1}+\ldots \right) \left( 1+is_{2}K_{2}+\ldots \right) \cdots \left( 1+is_{10}K_{10}+\ldots \right) \\
&=\left(1+is_{1}K_{1}+is_{2}K_{2}+is_{3}K_{3}+is_{4}K_{4}+is_{5}K_{5}+is_{6}K_{6}+is_{7}K_{7}+is_{8}K_{8}+is_{9}K_{9}+is_{10}K_{10}+\ldots \right),
\end{align}$$

where ##\ldots## in the last line refers to terms of second order and higher.
 
Last edited:
Oops, I originally left out a bunch of factors of ##i## in my previous post, which I have now edited and corrected. I inadvertently used standard math notation instead of standard physics notation. I will atone for my sins by using below the physicists' fast and loose version of functional analysis.

I now am in my office, and I have Ballentine at hand. There is a bit of a clash of notation between the material near equations 3.2 and 3.3, and the material near equations 3.9 and 3.10.

George Jones said:
Think of each real number ##s_\mu## as corresponding to a specific ##\tau##. For example: ##s_7## is the specific ##\tau## that translates by ##s_7## along ##x##, while leaving all other coordinates untouched; ##s_{10}## is the specific ##\tau## that translates by ##s_{10}## temporally, while leaving all other coordinates untouched; etc.

This is what the chart after equation 3.13 indicates.

Let's do the ##s_7## example. Take ##s_7## to be a real number and ##K_7 = -id/dx## (suppressing all other coodinates). Then

$$\begin{align}
U\left(s_7 \right) \psi \left(x\right) &= e^{is_7 K_7} \psi \left(x\right)\\
&= \left( 1 + s_7 \frac{d}{dx} +\frac{s_7^2}{2!} \frac{d^2}{dx^2} + \dots \right) \psi \left(x\right) \\
&= \psi \left(x\right) + s_7 \frac{d\psi}{dx}\left(x\right) +\frac{s_7^2}{2!} \frac{d^2 \psi}{dx^2}\left(x\right) + \dots \\
&= \psi \left(x+s_7 \right)\\
&= U \left( \tau \right) \psi \left(x\right)
\end{align}$$

The last equality, with ##\tau## a spatial x-translation by amount ##s_7##, follows from equation 3.8. I hope I got the signs right.
 
  • Like
Likes 1 person
Oh wow, that definitely clears it up quite elegantly. Thank you so much George!

If I may ask another question: starting from (3.12) on p.69 and assuming ##\omega \sim s_{\mu} \sim \epsilon##, we have ##e^{i\omega}U = (1 + i\omega + O(\epsilon^2))(1 + i\sum_{\mu=1}^{10}s_{\mu}K_{\mu} + O(\epsilon^2)) = (1 + i\omega)I + i\sum_{\mu=1}^{10}s_{\mu}K_{\mu} + O(\epsilon^2)## and separately ##\epsilon^{i\epsilon K_{\mu}}\epsilon^{i\epsilon K_{\nu}}\epsilon^{-i\epsilon K_{\mu}}\epsilon^{-i\epsilon K_{\nu}} = I - \epsilon^{2}[K_{\mu},K_{\nu}] + O(\epsilon^3)##.

Now I understand the general argument for why the left hand sides of these two expressions are supposed to be equal (as elucidated by Ballentine in the paragraph directly under (3.11)) but I don't get how the right hand sides equate since one is written to 2nd order in ##\epsilon## whereas the other is written to 1st order in ##\epsilon##. Furthermore, even if we ignore the error terms and just equate the two (as Ballentine says to do directly under (3.12)) we seem to get ##- \epsilon^{2}[K_{\mu},K_{\nu}] = i\omega I + i\sum_{\mu=1}^{10}s_{\mu}K_{\mu} ## and I don't immediately see how this leads to ##[K_{\mu},K_{\nu}] = i\sum_{\lambda}c^{\lambda}_{\mu\nu}K_{\lambda} + ib_{\mu\nu}I## (3.13); the indices don't even match up.

Thanks again for the help.
 
George Jones said:
Oops, I originally left out a bunch of factors of ##i## in my previous post, which I have now edited and corrected. I inadvertently used standard math notation instead of standard physics notation. I will atone for my sins by using below the physicists' fast and loose version of functional analysis.

I now am in my office, and I have Ballentine at hand. There is a bit of a clash of notation between the material near equations 3.2 and 3.3, and the material near equations 3.9 and 3.10.



This is what the chart after equation 3.13 indicates.

Let's do the ##s_7## example. Take ##s_7## to be a real number and ##K_7 = -id/dx## (suppressing all other coodinates). Then

$$\begin{align}
U\left(s_7 \right) \psi \left(x\right) &= e^{is_7 K_7} \psi \left(x\right)\\
&= \left( 1 + s_7 \frac{d}{dx} +\frac{s_7^2}{2!} \frac{d^2}{dx^2} + \dots \right) \psi \left(x\right) \\
&= \psi \left(x\right) + s_7 \frac{d\psi}{dx}\left(x\right) +\frac{s_7^2}{2!} \frac{d^2 \psi}{dx^2}\left(x\right) + \dots \\
&= \psi \left(x+s_7 \right)\\
&= U \left( \tau \right) \psi \left(x\right)
\end{align}$$

The last equality, with ##\tau## a spatial x-translation by amount ##s_7##, follows from equation 3.8. I hope I got the signs right.

Thanks, that really cool to see it that way round. I am less clear though why the unitary condition requires dU/ds + dU/ds dagger needs to be zero? I'm not getting the line above equation 3.3
 
  • #10
Jilang said:
Thanks, that really cool to see it that way round. I am less clear though why the unitary condition requires dU/ds + dU/ds dagger needs to be zero? I'm not getting the line above equation 3.3

##I = UU^{\dagger}\\ = (I + s\frac{\mathrm{d} U}{\mathrm{d} s}|_{s=0} + O(s^2))(I + s\frac{\mathrm{d} U^{\dagger}}{\mathrm{d} s}|_{s=0} + O(s^2)) \\= I + s(\frac{\mathrm{d} U}{\mathrm{d} s}|_{s=0} +\frac{\mathrm{d} U^{\dagger}}{\mathrm{d} s}|_{s=0} ) + O(s^2)##

so ##\frac{\mathrm{d} U}{\mathrm{d} s}|_{s=0} +\frac{\mathrm{d} U^{\dagger}}{\mathrm{d} s}|_{s=0} = 0## (ignoring 2nd order and higher terms in ##s##).
 
  • Like
Likes 1 person
  • #11
WannabeNewton said:
##I = UU^{\dagger}\\ = (I + s\frac{\mathrm{d} U}{\mathrm{d} s}|_{s=0} + O(s^2))(I + s\frac{\mathrm{d} U^{\dagger}}{\mathrm{d} s}|_{s=0} + O(s^2)) \\= I + s(\frac{\mathrm{d} U}{\mathrm{d} s}|_{s=0} +\frac{\mathrm{d} U^{\dagger}}{\mathrm{d} s}|_{s=0} ) + O(s^2)##

so ##\frac{\mathrm{d} U}{\mathrm{d} s}|_{s=0} +\frac{\mathrm{d} U^{\dagger}}{\mathrm{d} s}|_{s=0} = 0## (ignoring 2nd order and higher terms in ##s##).

That's brilliant. Many thanks. I now understand mathematically why unitary transformations look like they do. That's some accomplishment on both our parts especially on a Friday!
 
  • #12
I just had a horrible thought though... What happens to equation 3.3 if K does not equal K dagger? What requires it to be Hermitian?
 
  • #13
WannabeNewton said:
Now I understand the general argument for why the left hand sides of these two expressions are supposed to be equal (as elucidated by Ballentine in the paragraph directly under (3.11)) but I don't get how the right hand sides equate since one is written to 2nd order in ##\epsilon## whereas the other is written to 1st order in ##\epsilon##.

Do you buy Ballentine's statement "It is clear that all 11 parameters ##\left\{ \omega, s_\lambda \right\}## must be infinitesimal and of order ##\epsilon^2##"? If you don't, pretend that you do for the rest of this post, and we can talk about it later.

WannabeNewton said:
the indices don't even match up.

As relativists everywhere know:biggrin:, Ballentine's (3.12) is an example really horrible notation, since ##\mu## is used there as a dummy summation index on the right. and a free index on the left. Better to use ##\lambda## as the summation index on the right in (3.12). If the left sides of (3.11) and (3.12) are to be equated, we must have (with slight abuse of notation) ##s_\lambda = s_\lambda \left(\mu,\nu\right)## and ##\omega = \omega\left(\mu,\nu\right)##, because (3.11) depends on ##\mu## and ##\nu##.

Now I am going to wave my hands so furiously that folks everywhere should be able to feel the breeze.

Putting everything together, write ##s_\lambda \left(\mu,\nu\right) = \epsilon^2 c^\lambda_{\mu\nu}##, where ##\epsilon## is "small"', bu the ##c^\lambda_{\mu\nu}## aren't, and put this in the sum (now over ##\lambda##) on the right of 3.12.
 
  • #14
Jilang said:
I just had a horrible thought though... What happens to equation 3.3 if K does not equal K dagger? What requires it to be Hermitian?

Take ##\frac{\mathrm{d} U}{\mathrm{d} s}|_{s=0} = iK##, and, initially, don't assume anything about ##K## What do you then get for

$$\frac{\mathrm{d} U}{\mathrm{d} s}|_{s=0} +\frac{\mathrm{d} U^{\dagger}}{\mathrm{d} s}|_{s=0}?$$
 
  • #15
WannabeNewton said:
Oh wow, that definitely clears it up quite elegantly. Thank you so much George!

If I may ask another question: starting from (3.12) on p.69 and assuming ##\omega \sim s_{\mu} \sim \epsilon##, we have ##e^{i\omega}U = (1 + i\omega + O(\epsilon^2))(1 + i\sum_{\mu=1}^{10}s_{\mu}K_{\mu} + O(\epsilon^2)) = (1 + i\omega)I + i\sum_{\mu=1}^{10}s_{\mu}K_{\mu} + O(\epsilon^2)## and separately ##\epsilon^{i\epsilon K_{\mu}}\epsilon^{i\epsilon K_{\nu}}\epsilon^{-i\epsilon K_{\mu}}\epsilon^{-i\epsilon K_{\nu}} = I - \epsilon^{2}[K_{\mu},K_{\nu}] + O(\epsilon^3)##.

Now I understand the general argument for why the left hand sides of these two expressions are supposed to be equal (as elucidated by Ballentine in the paragraph directly under (3.11)) but I don't get how the right hand sides equate since one is written to 2nd order in ##\epsilon## whereas the other is written to 1st order in ##\epsilon##. Furthermore, even if we ignore the error terms and just equate the two (as Ballentine says to do directly under (3.12)) we seem to get ##- \epsilon^{2}[K_{\mu},K_{\nu}] = i\omega I + i\sum_{\mu=1}^{10}s_{\mu}K_{\mu} ## and I don't immediately see how this leads to ##[K_{\mu},K_{\nu}] = i\sum_{\lambda}c^{\lambda}_{\mu\nu}K_{\lambda} + ib_{\mu\nu}I## (3.13); the indices don't even match up.

Thanks again for the help.
I think your multiplication may be wrong in the first equation on the RHS. The later terms are missing a factor w.
 
  • #16
George Jones said:
Take ##\frac{\mathrm{d} U}{\mathrm{d} s}|_{s=0} = iK##, and, initially, don't assume anything about ##K## What do you then get for

$$\frac{\mathrm{d} U}{\mathrm{d} s}|_{s=0} +\frac{\mathrm{d} U^{\dagger}}{\mathrm{d} s}|_{s=0}?$$

Thanks Mr Jones, I can see that it works if K is Hermitian, I am just wondering why it has to be. It seems like the introduction of imaginary numbers is necessary because K is Hermitian, but I am unclear as to why K needs to be just Hermitan, though I understand why U is. Ballentine says the text is "that we may write". But it's not saying it has to be. Sorry if I'm being a bit dense, but it's been a long week.
I think you get iK + iK dagger
 
Last edited:
  • #17
Jilang said:
Thanks Mr Jones, I can see that it works if K is Hermitian, I am just wondering why it has to be. It seems like the introduction of imaginary numbers is necessary because K is Hermitian, but I am unclear as to why K needs to be just Hermitan, though I understand why U is. Ballentine says the text is "that we may write". But it's not saying it has to be. Sorry if I'm being a bit dense, but it's been a long week.
I think you get iK + iK dagger

If ##\frac{\mathrm{d} U}{\mathrm{d} s} = -\frac{\mathrm{d} U^{\dagger}}{\mathrm{d} s}## then ##\frac{\mathrm{d} U}{\mathrm{d} s} ## is "purely imaginary" in the sense that ##\frac{\mathrm{d} U}{\mathrm{d} s} = iK## for some ##K = K^{\dagger}##. It's analogous to how if ##a## is a complex number and ##a = -a^{*}## then ##a = ib## for some real number ##b##.
 
  • Like
Likes 1 person
  • #18
George Jones said:
Take ##\frac{\mathrm{d} U}{\mathrm{d} s}|_{s=0} = iK##, and, initially, don't assume anything about ##K## What do you then get for

$$\frac{\mathrm{d} U}{\mathrm{d} s}|_{s=0} +\frac{\mathrm{d} U^{\dagger}}{\mathrm{d} s}|_{s=0}?$$

Then,

$$\frac{\mathrm{d} U}{\mathrm{d} s}|_{s=0} +\frac{\mathrm{d} U^{\dagger}}{\mathrm{d} s}|_{s=0} = iK -iK^\dagger = i\left( K - K^\dagger \right)?$$

In order for this to be zero, we must have ##K = K^\dagger##.

[edit]Scooped by WannabeNewton while typing[/edit].
 
  • #19
George Jones said:
Do you buy Ballentine's statement "It is clear that all 11 parameters ##\left\{ \omega, s_\lambda \right\}## must be infinitesimal and of order ##\epsilon^2##"? If you don't, pretend that you do for the rest of this post, and we can talk about it later.

Thanks, as a matter of fact I'm actually unsure as to why Ballentine's statement is true.

George Jones said:
As relativists everywhere know:biggrin:, Ballentine's (3.12) is an example really horrible notation, since ##\mu## is used there as a dummy summation index on the right. and a free index on the left. Better to use ##\lambda## as the summation index on the right in (3.12). If the left sides of (3.11) and (3.12) are to be equated, we must have (with slight abuse of notation) ##s_\lambda = s_\lambda \left(\mu,\nu\right)## and ##\omega = \omega\left(\mu,\nu\right)##, because (3.11) depends on ##\mu## and ##\nu##.

Now I am going to wave my hands so furiously that folks everywhere should be able to feel the breeze.

Putting everything together, write ##s_\lambda \left(\mu,\nu\right) = \epsilon^2 c^\lambda_{\mu\nu}##, where ##\epsilon## is "small"', bu the ##c^\lambda_{\mu\nu}## aren't, and put this in the sum (now over ##\lambda##) on the right of 3.12.

Thanks again George, that makes perfect sense. The abuse of indices in (3.12) confused the living hell out of me.
 
  • #20
George Jones said:
Then,

$$\frac{\mathrm{d} U}{\mathrm{d} s}|_{s=0} +\frac{\mathrm{d} U^{\dagger}}{\mathrm{d} s}|_{s=0} = iK -iK^\dagger = i\left( K - K^\dagger \right)?$$

In order for this to be zero, we must have ##K = K^\dagger##.

[edit]Scooped by WannabeNewton while typing[/edit].

Thanks Mr Jones, it's obvious when you do it. I'll try not to bother you for a few chapters!
 
  • #21
WannabeNewton said:
Thanks, as a matter of fact I'm actually unsure as to why Ballentine's statement is true.

Let's rearrange the presentation a bit. Using (3.7) a few time gives ##e^{i\omega}U = e^{i\epsilon K_\mu} e^{i\epsilon K_\nu} e^{-i\epsilon K_\mu} e^{-i\epsilon K_\mu}##. But (3.11) shows ##e^{i\omega}U = e^{i\epsilon K_\mu} e^{i\epsilon K_\nu} e^{-i\epsilon K_\mu} e^{-i\epsilon K_\mu} = I - \epsilon^2 \left[ K_\mu , K_\nu] \right] + O\left( \epsilon^3 \right)##. Hence, ##e^{i\omega}U = I - \epsilon^2 \left[ K_\mu , K_\nu] \right] + O\left( \epsilon^3 \right)##. Since the right side of the last equation has no first-order ##\epsilon## terms, the left side can't have any either.

Equivalently, write ##s_\lambda \left(\mu,\nu\right) =\epsilon a^\lambda_{\mu\nu} + \epsilon^2 c^\lambda_{\mu\nu}##. Since the right side has epsilon times zero all the ##a^\lambda_{\mu\nu}## must be zero.
 
  • #22
George Jones said:
Hence, ##e^{i\omega}U = I - \epsilon^2 \left[ K_\mu , K_\nu] \right] + O\left( \epsilon^3 \right)##. Since the right side of the last equation has no first-order ##\epsilon## terms, the left side can't have any either.

Equivalently, write ##s_\lambda \left(\mu,\nu\right) =\epsilon a^\lambda_{\mu\nu} + \epsilon^2 c^\lambda_{\mu\nu}##. Since the right side has epsilon times zero all the ##a^\lambda_{\mu\nu}## must be zero.

Gotcha, thanks! Just to make sure I'm on the same page overall, consider the sequence of space-time transformations given by a boost ##-\epsilon## along the ##x##-axis, a time translation by ##-\epsilon##, and the respective inverses.

We have on the one hand ##e^{i\epsilon H}e^{i\epsilon G_{x}}e^{-i\epsilon H}e^{-i\epsilon G_{x}} = I + \epsilon^{2}[G_{x},H]+O(\epsilon^{3})## where ##H## is the generator of time translations and ##G_{x}## the generator of boosts in the ##x## direction. On the other hand the sequence of space-time transformations give us ##(t,x,y,z)\rightarrow (t,x - \epsilon^2,y,z)## which is a spatial translation along ##x## by ##-\epsilon^2## so the associated unitary transformation is ##U = I + i \epsilon^2 P_{x} + O(\epsilon^3)## (using Ballentine's sign convention for the generator of spatial translations ##P_{\alpha}##) where ##s_{x}(t,x) = \epsilon^2## and ##s_{\mu}(t,x) = 0## for all ##\mu \neq x##.

The two different expressions we have must be equivalent up to a phase factor ##\omega(t,x) = \epsilon^2 b_{tx}## so ##[G_x,H]=i P_{x} +i b_{tx}I ##. To relate back to (3.13) would it then be fair to say that ##\sum c^{\lambda}_{tx}K_{\lambda} = P_{x}\Rightarrow c^{x}_{tx} = 1,c^{\mu}_{tx} = 0(\forall \mu\neq x)##? Also just to be careful, is Ballentine's "(?)" in front of the identity in (3.21) simply the factor ##ib_{tx}##?

Thanks!
 
  • #23
WannabeNewton said:
Gotcha, thanks! Just to make sure I'm on the same page overall, consider the sequence of space-time transformations given by a boost ##-\epsilon## along the ##x##-axis, a time translation by ##-\epsilon##, and the respective inverses.

We have on the one hand ##e^{i\epsilon H}e^{i\epsilon G_{x}}e^{-i\epsilon H}e^{-i\epsilon G_{x}} = I + \epsilon^{2}[G_{x},H]+O(\epsilon^{3})## where ##H## is the generator of time translations and ##G_{x}## the generator of boosts in the ##x## direction. On the other hand the sequence of space-time transformations give us ##(t,x,y,z)\rightarrow (t,x - \epsilon^2,y,z)## which is a spatial translation along ##x## by ##-\epsilon^2## so the associated unitary transformation is ##U = I + i \epsilon^2 P_{x} + O(\epsilon^3)## (using Ballentine's sign convention for the generator of spatial translations ##P_{\alpha}##) where ##s_{x}(t,x) = \epsilon^2## and ##s_{\mu}(t,x) = 0## for all ##\mu \neq x##.

The two different expressions we have must be equivalent up to a phase factor ##\omega(t,x) = \epsilon^2 b_{tx}## so ##[G_x,H]=i P_{x} +i b_{tx}I ##. To relate back to (3.13) would it then be fair to say that ##\sum c^{\lambda}_{tx}K_{\lambda} = P_{x}\Rightarrow c^{x}_{tx} = 1,c^{\mu}_{tx} = 0(\forall \mu\neq x)##? Also just to be careful, is Ballentine's "(?)" in front of the identity in (3.21) simply the factor ##ib_{tx}##?

Thanks!

I'm agreeing with everything you just said. Just hit another query at 3.22 where ε seems to mean something else entirely before it switches back to being infinitesimal in the next equation. Am I being picky or just missing something obvious?
 
  • #24
I didn't bring Ballentine home with me, as my knapsack only has a finite volume. I might stop by and pick it up, but I'm not sure.
 
Last edited:
  • #25
George Jones said:
I didn't bring Ballentine home with me, as my knapsack only has a finite volume. I might stop by and pick it up, but I'm sure.
Please don't go to any trouble as I am not unduly troubled by it. It just seems strange that when there are all those Greek letters to choose from you use the same one to mean different things on the same page. There might be a good reason for it though that is alluding me!
 
  • #26
Jilang said:
Am I being picky or just missing something obvious?

##\epsilon_{\alpha_1...\alpha_n}## is the standard notation for the n-dimensional Levi-Civita symbol. ##\epsilon## by itself is also standard notation for expansion parameters when working with perturbed quantities; it isn't really a problem since the former has indices to indicate that it's the Levi-Civita symbol.

George Jones said:
I didn't bring Ballentine home with me, as my knapsack only has a finite volume. I might stop by and pick it up, but I'm sure.

Thanks!
 
  • #27
WannabeNewton said:
##\epsilon_{\alpha_1...\alpha_n}## is the standard notation for the n-dimensional Levi-Civita symbol. ##\epsilon## by itself is also standard notation for expansion parameters when working with perturbed quantities; it isn't really a problem since the former has indices to indicate that it's the Levi-Civita symbol.
Thanks! I wondered if there might be a good reason. Just 2 pages of maths left to go now before the Physics kicks in! :smile:
 
  • #28
Since George hasn't answered yet...

WannabeNewton said:
[...] To relate back to (3.13) would it then be fair to say that ##\sum c^{\lambda}_{tx}K_{\lambda} = P_{x}\Rightarrow c^{x}_{tx} = 1,c^{\mu}_{tx} = 0(\forall \mu\neq x)##?
That looks right to me.

BTW, the generators of a Lie algebra are also basis elements in an abstract vector space. So in cases like this you could think of the 2 sides as vectors, and make some rapid deductions by relying on linear independence of the basis elements. (I hope that makes sense.)

Also just to be careful, is Ballentine's "(?)" in front of the identity in (3.21) simply the factor ##ib_{tx}##?
More likely it's ##ib_{t\alpha}##. :biggrin:

But yes, it denotes the coefficient of the central element ##I## in the Lie algebra.
 
  • #29
strangerep said:
BTW, the generators of a Lie algebra are also basis elements in an abstract vector space. So in cases like this you could think of the 2 sides as vectors, and make some rapid deductions by relying on linear independence of the basis elements. (I hope that makes sense.)

That makes perfect sense, thank you!

strangerep said:
More likely it's ##ib_{t\alpha}##. :biggrin:

But yes, it denotes the coefficient of the central element ##I## in the Lie algebra.

Cool. Sorry for the onslaught of questions but it's almost winter break and I figured I should finally bring myself around to working through the book in a more serious manner :biggrin:
 
  • #30
WannabeNewton said:
Sorry for the onslaught of questions
No need to be sorry.

but it's almost winter break and I figured I should finally bring myself around to working through the book in a more serious manner :biggrin:
Great idea. Maybe that will eventually force me to figure out whether certain little things later in the book are typos or just my misunderstandings. (I couldn't find an errata list back when I was going through Ballentine closely.)

Seems like Jilang is also working through it?
 
  • #31
This isn't a question about chapter 3 of Ballentine itself but is related to the notation of the chapter. In chapter 2 of Srednicki's QFT text, Srednicki goes about deriving the commutator relations for the generators of the proper orthochronous Lorentz group and consequently the structure coefficients of the Lorentz algebra similarly to Ballentine's derivation of these quantities for the Galilei group. However the notations differ slightly and I'm having trouble parsing Srednicki's notation having become comfortable with Ballentine's notation.

He starts with an infinitesimal Lorentz transformation ##\Lambda^{\mu}{}{}_{\nu} = \delta^{\mu}{}{}_{\nu} + \delta \omega^{\mu}{}{}_{\nu}## where ##\delta \omega_{\mu\nu} = -\delta \omega_{\nu\mu}##. In expression 2.12 he writes, for a unitary transformation ##U(\Lambda)## of an infinitesimal Lorentz transformation, ##U(1 + \delta \omega) = I + \frac{i}{2\hbar}\delta \omega_{\mu\nu}M^{\mu\nu}##.

I assume by ##1 + \delta \omega## he means the identity operator ##1## and the matrix ##\delta \omega^{\mu}{}{}_{\nu}## now construed as an operator ##\delta \omega##, with both operators acting on Minkowski space-time. Now he's using abstract indices to represent the 2-form ##\delta \omega_{\mu\nu}## and then contracting it with the indices on ##M^{\mu\nu}## and based on his definition of the angular momentum operator ##\mathbf{J}## component-wise by ##J_i = \frac{1}{2}\epsilon_{ijk}M^{jk}## and the Lorentz boost operator ##\mathbf{K}## also component-wise by ##K_{i} = M^{0i}## it seems that the indices on ##M^{\mu\nu}## aren't abstract indices but rather just labels for Hermitian generators of the proper orthochronous Lorentz group along the different spatial axes in which case I can neither make sense of how he gets ##U(1 + \delta \omega) = I + \frac{i}{2\hbar}\delta \omega_{\mu\nu}M^{\mu\nu}## nor make sense of how to even parse it.

Thanks.
 
Last edited:
  • #32
Each each ##\Lambda## is Lorentz transformation matrix; consequently, each ##\Lambda^{\mu}{}{}_{\nu}## is a component of a matrix, i.e., a number; so, each ##\delta \omega_{\mu \nu}## is a number (infintesimal?); aach ##M^{\mu \nu}## is an operator (or matrix, if the space is finite-dimensional and a basis has been chosen); ##\delta \omega_{\mu \nu} M^{\nu \mu}## is a linear combination of operators.
 
  • #33
Adding a little to what George said...

WannabeNewton said:
I assume by ##1 + \delta \omega## he means the identity operator ##1## and the matrix ##\delta \omega^{\mu}{}{}_{\nu}## now construed as an operator ##\delta \omega##, with both operators acting on Minkowski space-time.
An element of the abstract Lorentz group can be represented as 4x4 matrices acting as transformations on Minkowski spacetime, or it can be represented as a unitary operator on Hilbert space. Srednicki is obviously mapping between these 2 representations. In more general situations, however, you may need to think of ##U## as a mapping from abstract group elements to their concrete representations as unitary operators. But of course, when reading Srednicki's textbook, you got to keep these two distinctions straight, since he's doing a rep-to-rep mapping.

Hence: when he writes ##U(1+\delta\omega)##, the argument is obviously a 4x4 matrix, denoting a transformation applicable on Minkowski spacetime. But the rhs of (2.12) is an operator on Hilbert space, so he uses ##I## instead of ##1##.

(And now I wonder whether I've only confused the issue.)
 
  • #34
George Jones said:
Each each ##\Lambda## is Lorentz transformation matrix; consequently, each ##\Lambda^{\mu}{}{}_{\nu}## is a component of a matrix, i.e., a number; so, each ##\delta \omega_{\mu \nu}## is a number (infintesimal?); aach ##M^{\mu \nu}## is an operator (or matrix, if the space is finite-dimensional and a basis has been chosen); ##\delta \omega_{\mu \nu} M^{\nu \mu}## is a linear combination of operators.

Ok good that's what I figured the notation was representing. So just to tie it back to Ballentine's notation, if we take the definitions ##J_{i} = \epsilon_{ijk}M^{jk}## and ##K_{i} = M^{i0}## and note that ##M^{ij} = \epsilon^{ijk}J_{k}## as a result of the first definition, we have ##\frac{1}{2}\delta \omega_{\mu\nu}M^{\mu\nu} = \delta \omega_{i0}M^{i0}+\frac{1}{2}\epsilon^{ijk}\omega_{ij}J_{k} = \delta \omega_{0x}K_{x}+\delta \omega_{0y}K_{y}+\delta \omega_{0z}K_{z} +\omega_{xy}J_{z} + \omega_{yz}J_{x} + \omega_{zx}J_{y} ##
so if I wrote ##U(1 + \delta \omega) = I + i\sum_{\mu = 1}^{6}s_{\mu}K_{\mu}## as in Ballentine (units of ##\hbar = 1##) and made the identifications ##s_1 = \delta \omega_{0x}##, ##s_{4} = \delta \omega_{xy}## etc. and ##K_{1} = K_{x}##, ##K_{4} = J_{z}## etc. would it be fair to say it's equivalent to what Srednicki writes?

Also, consider a killing field ##\xi^{\mu}## in Minkowski space-time and a constant 4-velocity field ##u^{\mu}##. Letting ##x^{\mu}## be the position vector field relative to ##u^{\mu}##, we can decompose ##\xi^{\mu}## as ##\xi^{\mu} = 2E^{[\nu}u^{\mu]}x_{\nu} + \epsilon^{\mu\nu\alpha\beta}x_{\nu}u_{\alpha}B_{\beta}## where ##E^{\mu},B^{\mu}## are space-like vectors. Pure Lorentz boosts correspond to the "electric" part ##2E^{[\nu}u^{\mu]}x_{\nu}## and pure rotations correspond to the "magnetic part" ##\epsilon^{\mu\nu\alpha\beta}x_{\nu}u_{\alpha}B_{\beta}## where the rotation axis is given by ##B^{\mu}##.

If we go to the rest frame of ##u^{\mu}## and consider an elemental pure rotation given by ##B^{\mu} = (\partial_z)^{\mu}## for example then ##\xi = x\partial_{y}-y\partial_{x} ## which is in obvious 1-1 correspondence with the generator of rotations about the ##z## axis ## J_z = X P_{y} -Y P_{x}##. For an elemental pure Lorentz boost in the ##x## direction, for which ##E^{\mu} = (\partial_{x})^{\mu}##, we get ##\xi = x\partial_t+t\partial_x ##. Can an analogous 1-1 correspondence between this and ##K_z##, the generator of Lorentz boosts in the ##x## direction, be made in the same obvious manner? This is just so I can connect the Hermitian generators of the proper orthochronous Lorentz group back to the familiar notion of Minkowski space-time killing fields as the generators of isometric flows in Minkowski space-time.

strangerep said:
Adding a little to what George said...
In more general situations, however, you may need to think of ##U## as a mapping from abstract group elements to their concrete representations as unitary operators.

Could you elucidate this point? I understand that when we write ##U(1 + \delta \omega)##, we basically just saying that to each representation of the infinitesimal Lorentz transformation as an operator ##1 + \delta \omega; \delta^{\mu}{}{}_{\nu} + \delta \omega^{\mu}{}{}_{\nu}## on Minkowski space-time there is associated a representation of the same infinitesimal Lorentz transformation as a unitary operator ##I + \frac{i}{2\hbar}\delta \omega_{\mu\nu} M^{\mu\nu}## on Hilbert space, a rep to rep mapping as you put it, but I'm afraid I don't quite understand the quoted statement above.

Thanks guys!
 
  • #35
WannabeNewton said:
Could you elucidate this point? I understand that when we write ##U(1 + \delta \omega)##, we basically just saying that to each representation of the infinitesimal Lorentz transformation as an operator ##1 + \delta \omega; \delta^{\mu}{}{}_{\nu} + \delta \omega^{\mu}{}{}_{\nu}## on Minkowski space-time there is associated a representation of the same infinitesimal Lorentz transformation as a unitary operator ##I + \frac{i}{2\hbar}\delta \omega_{\mu\nu} M^{\mu\nu}## on Hilbert space, a rep to rep mapping as you put it, but I'm afraid I don't quite understand the quoted statement above.

The correct mathematical setting to talk about this stuff is the theory of unitary representations of Lie groups. In general, you are given a Lie group ##G##, that is a group with an additional manifold structure, such that the group multiplication and the inversion are smooth maps. In your case, this group is the Poincare group (or rather it's connected component of the identity). A unitary representation of a group is a pair ##(\mathcal H, \pi)## with ##\mathcal H## being a Hilbert space and ##\pi:G\rightarrow U(\mathcal H)## is a (strongly continuous) map from the group into the unitary operators on ##\mathcal H## such that ##\pi(e) = \mathrm{id}_\mathcal{H}## and ##\pi(g)\pi(h)=\pi(gh)##. The goal of representation theory is to classify all these representations (up to equivalence).

For the Poincare group, it turns out that you can classify all the (irreducible, roughly speaking "minimal") (projective) representations by two numbers, which we call mass and spin. This is due to Wigner ("On unitary representations of the inhomogeneous Lorentz group").

There is much more to be said about this. For example the role of the infinitesimal transformations is played by Lie algebras and their representations and there is a way to relate them to the representations of the Lie group (that's why we actually do it) and one has to deal with some complications. But I'm not sure how much you want to learn about it, so I stop here for now.
 
  • #36
WannabeNewton said:
For an elemental pure Lorentz boost in the ##x## direction, for which ##E^{\mu} = (\partial_{x})^{\mu}##, we get ##\xi = x\partial_t+t\partial_x ##. Can an analogous 1-1 correspondence between this and ##K_z##, the generator of Lorentz boosts in the ##x## direction, be made in the same obvious manner? This is just so I can connect the Hermitian generators of the proper orthochronous Lorentz group back to the familiar notion of Minkowski space-time killing fields as the generators of isometric flows in Minkowski space-time.
I guess you meant ##K_x## ?

It's interesting to me to see how you're approaching this in terms of more general techniques usually associated with GR and DG. When I learned this stuff, I never thought about Killing fields, etc, at all.

But anyway, the answer is essentially "yes". A more low-brow way to get the explicit generator is to consider an ordinary 1-parameter transformation of the form ##x^\mu \to x'^{\,\mu}(x^\lambda, p)##, where ##p## is the parameter. The associated generator ##P## can be found via this formula:
$$
\def\Pdrv#1#2{\frac{\partial #1}{\partial #2}}
\def\pdrv#1{\frac{\partial}{\partial#1}}
P ~:=~ \left. \Pdrv{x'^{\,\mu}}{p} \right|_{p = p_0} \pdrv{x^\mu} ~,
$$
where ##p_0## denotes the value of ##p## corresponding to the identity transformation. I.e., the first derivative above is to be evaluated at the identity.

In the case of a coordinate transformation corresponding to a boost in velocity, one gets essentially what you wrote.


Could you elucidate this point? I understand that when we write ##U(1 + \delta \omega)##, we basically just saying that to each representation of the infinitesimal Lorentz transformation as an operator ##1 + \delta \omega; \delta^{\mu}{}{}_{\nu} + \delta \omega^{\mu}{}{}_{\nu}## on Minkowski space-time there is associated a representation of the same infinitesimal Lorentz transformation as a unitary operator ##I + \frac{i}{2\hbar}\delta \omega_{\mu\nu} M^{\mu\nu}## on Hilbert space, a rep to rep mapping as you put it, but I'm afraid I don't quite understand the quoted statement above.
Urk, I knew I was just confusing matters by jumping ahead like that.

Rubi has already given a good answer, and I'd probably just add to the confusion by trying to embelish it. But feel free to ask further questions if you can't keep your curiosity under control. :biggrin:
 
  • #37
strangerep said:
I guess you meant ##K_x## ?

Ah yeah sorry typo.

strangerep said:
In the case of a coordinate transformation corresponding to a boost in velocity, one gets essentially what you wrote.
Thanks! So it seems then, after rewriting the pure boost killing fields along the ##x##-axis as ##t\partial_{x} + x\partial_{t}\rightarrow x^{0}\partial^{1} - x^{1}\partial^{0}## (where the negative sign comes from raising time indices using the Minkowski metric) and similarly for pure boost killing fields along the other axes, as well as rewriting the rotational killing fields about the ##z##-axis as ##x\partial_{y} - y\partial_{x}\rightarrow x^{1}\partial^{2} - x^{2}\partial^{1}## and similarly for the pure rotational killing fields about the other axes, that ##M^{\mu\nu} \propto (x^{\mu}\partial^{\nu} - x^{\nu}\partial^{\mu})## with the proportionality being given by some appropriate quantization factor. If so, why didn't Srednicki just write down an explicit form for ##M^{\mu\nu}## in the above manner and how is ##M^{\mu\nu}## different from the ##\mathcal{L}^{\mu\nu} = \frac{\hbar}{i}(x^{\mu}\partial^{\nu} - x^{\nu}\partial^{\mu})## that Srednicki defines in problem 2.8? For example ##\mathcal{L}^{01} = \frac{\hbar}{i}(x^{1}\partial^{0} - x^{0}\partial^{1}) = i\hbar(x^{0}\partial^{1}-x^{1}\partial^{0})## and this seems to correspond to the boost operator ##K_{x}## based on the form of the boost killing field in the ##x## direction.
 
  • #38
rubi said:
There is much more to be said about this. For example the role of the infinitesimal transformations is played by Lie algebras and their representations and there is a way to relate them to the representations of the Lie group (that's why we actually do it) and one has to deal with some complications. But I'm not sure how much you want to learn about it, so I stop here for now.

Thanks for the awesome primer rubi! I'm familiar with this from classical mechanics as well as general smooth manifold theory but I haven't personally had a chance to find a QM/QM for mathematicians book that developed all of the above in a detailed manner in the context of QM (Hall contains it but the text does it very briefly in a single chapter so I found it hard to absorb anything effectively from it). Do you have any particular reference(s) in mind?
 
  • #39
This is beyond me but I think the book Peter Woit is preparing (current version available here http://www.math.columbia.edu/~woit/qmbook.pdf) might interest you. From the preface:

These are the course notes prepared for a class taught at Columbia during the
2012-13 academic year. The intent was to cover the basics of quantum mechanics,
from a point of view emphasizing the role of unitary representations of
Lie groups in the foundations of the subject. The approach to this material is
simultaneously rather advanced, using crucially some fundamental mathematical
structures normally only discussed in graduate mathematics courses, while
at the same time trying to do this in as elementary terms as possible.
 
  • Like
Likes 1 person
  • #40
WannabeNewton said:
[...] So it seems then, after rewriting the pure boost killing fields along the ##x##-axis as ##t\partial_{x} + x\partial_{t}\rightarrow x^{0}\partial^{1} - x^{1}\partial^{0}## (where the negative sign comes from raising time indices using the Minkowski metric)
Yes.

why didn't Srednicki just write down an explicit form for ##M^{\mu\nu}## in the above manner and [...]
I learned my basic QFT from Peskin & Schroeder. Later, I bought a copy of Srednicki when several other people said nice things about it, but I was a bit disappointed, tbh.

But rest assured, you'll find plenty more (bigger) issues to complain about as you wend your way through multiple QFT texts at various levels. :rolleyes:


BTW, regarding Peter Woit's draft textbook,... keep in mind that it's aimed at mathematicians. Therefore, it inevitably omits many of the important physical insights that are contained in Ballentine (yet are hard to find elsewhere). This is not to criticize Peter's book -- but only as a warning that a mathematical emphasis tends to gloss over important physical insights (just as a physics emphasis tends to gloss over some of the subtler maths). One needs to study the subject well from both perspectives.

You might also want to take a look at this book:

Michele Maggiore, "A Modern Introduction to QFT",
Oxford Univ. Press, 2005, ISBN 0-19-852074-3

He places less emphasis than other QFT books on practical stuff like calculating cross-sections, etc, in particle physics, and more emphasis on the underlying methods and ideas. I found that side of the subject easier to absorb from Maggiore than from Weinberg's lengthy tomes. If you look through ch2 of Maggiore, you should be able to decide whether it's suitable for you at this time.
 
  • Like
Likes 1 person
  • #41
WannabeNewton said:
Ah yeah sorry typo. Thanks! So it seems then, after rewriting the pure boost killing fields along the ##x##-axis as ##t\partial_{x} + x\partial_{t}\rightarrow x^{0}\partial^{1} - x^{1}\partial^{0}## (where the negative sign comes from raising time indices using the Minkowski metric) and similarly for pure boost killing fields along the other axes, as well as rewriting the rotational killing fields about the ##z##-axis as ##x\partial_{y} - y\partial_{x}\rightarrow x^{1}\partial^{2} - x^{2}\partial^{1}## and similarly for the pure rotational killing fields about the other axes, that ##M^{\mu\nu} \propto (x^{\mu}\partial^{\nu} - x^{\nu}\partial^{\mu})## with the proportionality being given by some appropriate quantization factor. If so, why didn't Srednicki just write down an explicit form for ##M^{\mu\nu}## in the above manner and how is ##M^{\mu\nu}## different from the ##\mathcal{L}^{\mu\nu} = \frac{\hbar}{i}(x^{\mu}\partial^{\nu} - x^{\nu}\partial^{\mu})## that Srednicki defines in problem 2.8? For example ##\mathcal{L}^{01} = \frac{\hbar}{i}(x^{1}\partial^{0} - x^{0}\partial^{1}) = i\hbar(x^{0}\partial^{1}-x^{1}\partial^{0})## and this seems to correspond to the boost operator ##K_{x}## based on the form of the boost killing field in the ##x## direction.

Coming from a thorough study of differential geometry into quantum mechanics/field theory, I assume you wish to stick as closely as possible to the differential geometric concepts proved useful to you in learning relativity. There is only one source that I know which attacks QFT coming from the diff. geometry used in relativity: the notes of R. Geroch at the Univ. of Chicago which are now being published as a book and offered for sale. You should venture yourself in reading those.
 
  • Like
Likes 1 person
  • #42
I'd rather recommend Weinberg's marvelous books on QFT. They are not very much based on differential geometry, but have a very clear and complete exposition of the symmetry of the Poincare group, which you do not find so easily in other textbooks.

Some aspects you also find in the very nice book

Sexl, Roman U., Urbandtke, Helmuth K.: Relativity, Groups, Particles, Springer, 2001

but this deals only with the representation theory not with the further development of QFT and S-Matrix theory.

Peskin and Schroeder is a standard text, but has to read with some caveats. There are a lot of topics not very carefully treated. The most shocking moment was when I found logarithms with dimensionful arguments in the chapter about the renormalization group (sic!). It's not so much the many typos in this book which I find disturbing but such conceptional weaknesses which are the more serious in view of the fact that relativistic interacting QFT is anyway mathematically not strictly formulated and maybe can never be. The more importance should be given to the parts which can be made right.

Srednicky is not so bad, but I don't understand, why he deals with \phi^3 theory, which is a priori unstable. Formally you can demonstrate some formal issues in perturbative renormalization, but otherwise it has not much value. On the other hand, it gives a careful exposition of the important and not so trivial topics like the LSZ reduction formalism and S-matrix theory.

To begin with, I'd recommend also Ryder. A more path-integral based approach you find in the book by Bailin and Love, Gauge Field Theories, which I liked very much too when I learned QFT. The best choice, in my opinion, still is Weinberg's Quantum Theory of Fields (3 vols., with vol. 3 about SUSY QFT), as mentioned above. Another somewhat more advanced book is Boehm, Dittmaier, Joos, Gauge theories of the strong and electroweak interaction (Teubner).

If you are interested in the special topic of QFT in curved spacetime, then have a look at deWitts famous review:

B. deWitt, Quantum Field Theory in Curved Space Time, Phys. Rept. 19, 295 (1975)
http://dx.doi.org/10.1016/0370-1573(75)90051-4
 
Last edited by a moderator:
  • #43
bhobba said:
The thing I recall is it took me a little while to fully understand - its a bit dense.

The other thing I recall is there is a bit of thought required about exactly why a free particle forms an irreducible set.

The other thing is the exact key assumption being made in quantitization - it took me a while to nut that one out.
I'm having problems with those two , any help would be appreciated.
 
  • #44
andresB said:
I'm having problems with those two , any help would be appreciated.

This thread is over 18 months old.

I think it would be wise to start a new one detailing exactly what your problems are.

It been a long while since I have gone through it myself.

Thanks
Bill
 
  • #45
andresB said:
I'm having problems with those two , any help would be appreciated.
I agree with Bhobba -- start a new thread. Also, about irreducibility, be sure to read the relevant appendices in Ballentine's textbook which cover this in a bit more detail.
 
Back
Top