Solving Tensor Confusion: Electromagnetic Field Tensor

  • Thread starter Thread starter AxiomOfChoice
  • Start date Start date
  • Tags Tags
    Confusion Tensor
AxiomOfChoice
Messages
531
Reaction score
1
I don't think this should be a very difficult question for people who are used to working with tensors, but I'm new to it, so I'm confused. The Wikipedia article on the electromagnetic field tensor F^{\mu \nu} asserts that

<br /> F_{\mu \nu} F^{\mu \nu} = 2 \left( B^2 - \frac{E^2}{c^2} \right).<br />

But if you look at the way they've written out F_{\mu \nu} and F^{\mu \nu} in matrix form, there is no way you get that when you just simply multiply the two matrices together. I mean, how can one obtain a constant from multiplying two square matrices (that aren't one-by-one, obviously) together? What am I missing here?

(The relevant Wikipedia article is http://en.wikipedia.org/wiki/Electromagnetic_tensor" .)
 
Last edited by a moderator:
Physics news on Phys.org
Because this is not matrix multiplication! The fact that mu and nu are used on both tensors means they are summed over, so this is an inner product of two tensors, not the product of two matrices!
 
Of course you're missing. Taking the minus trace on the resulting matrix. In tensor notation

F_{\mu\nu}F^{\mu\nu} = -\delta_{\mu}^{\lambda} F_{\lambda\nu}F^{\nu\mu}
 
R u guys taking antifrction into account? I think ur overcomplicating things
 
AxiomOfChoice said:
What am I missing here?
I answered this question quite recently (for arbitrary matrices), so I'll just quote myself.

Fredrik said:
If you want to treat this expression as something involving a product of matrices, this is what you need:

Definition of matrix multiplication: (XY)ij=XikYkj
Definition of transpose: (XT)ij=Xji
Definition of trace: Tr X=Xii

These definitions tell us that

A_{ij}S_{ij}=(A^T)_{ji}S_{ij}=(A^TS)_{jj}=\operatorname{Tr}(A^TS)

Alternatively,

A_{ij}S_{ij}=A_{ij}(S^T)_{ji}=(AS^T)_{ii}=\operatorname{Tr}(AS^T)

You can also reverse the order of the matrices in the trace, because Tr(XY)=Tr(YX) for all matrices X and Y. (This is very easy to prove using the definitions above).
 
Thanks a lot, guys. I think this has cleared up a lot. Can I argue as follows?

<br /> F_{\mu \nu} F^{\mu \nu} = (\partial_\mu A_\nu - \partial_\nu A_\mu)(\partial^\mu A^\nu - \partial^\nu A^\mu),<br />

from which one obtains

<br /> F_{\mu \nu} F^{\mu \nu} = 2 (\partial_\mu \partial^\mu A_\nu A^\nu - \partial_\nu A^\nu \partial_\mu A^\mu)<br />
 
Last edited:
AxiomOfChoice said:
Thanks a lot, guys. I think this has cleared up a lot. Can I argue as follows?

<br /> F_{\mu \nu} F^{\mu \nu} = (\partial_\mu A_\nu - \partial_\nu A_\mu)(\partial^\mu A^\nu - \partial^\nu A^\mu),<br />

from which one obtains

<br /> F_{\mu \nu} F^{\mu \nu} = 2 (\partial_\mu \partial^\mu A_\nu A^\nu - \partial_\nu A^\nu \partial_\mu A^\mu)<br /> [/tex
<br /> <br /> You maye want to correct the sloppy notation and LaTex formatting. And what you wrote, assumingly correct, has nothing to do with the initial question to which answers were already formulated.
 
bigubau said:
You maye want to correct the sloppy notation and LaTex formatting.

Done. Sorry; I temporarily lost my wireless connection while trying to correct.

bigubau said:
And what you wrote, assumingly correct, has nothing to do with the initial question to which answers were already formulated.

...I guess I don't see why. The key ingredient I was missing was the fact that repeated indices implied a sum. I put this together with the identity F_{\mu \nu} = \partial_\mu A_\nu - \partial_\nu A_\mu to produce the above.
 
The LaTeX needs further fixing. Well, your question was about the matrices of the e-m tensor. The elements of these matrices are made up of E and B. There's no help you can get on your question by bringing potentials into the discussion, because the E and B are enough (your lagrangian is a quadratic function of E and B, after all) to formulate the simplest possible correct answer you could get (post 3).
 
  • #10
bigubau said:
The LaTeX needs further fixing.

The only thing I can see that you might be talking about is the fact that I have written \partial_\mu A_\nu \partial^\nu A^\mu as \partial_\nu A^\nu \partial_\mu A^\mu. But what is the difference between \partial^\nu A_\nu and \partial_\mu A^\mu?
 
  • #11
AxiomOfChoice said:
Thanks a lot, guys. I think this has cleared up a lot. Can I argue as follows?

<br /> F_{\mu \nu} F^{\mu \nu} = (\partial_\mu A_\nu - \partial_\nu A_\mu)(\partial^\mu A^\nu - \partial^\nu A^\mu),<br />

from which one obtains

<br /> F_{\mu \nu} F^{\mu \nu} = 2 (\partial_\mu \partial^\mu A_\nu A^\nu - \partial_\nu A^\nu \partial_\mu A^\mu)<br />
I'm also wondering what this has to do with your original question. What you overlooked was, as you said yourself, that repeated indices are summed over. Now that you know that, you should see that it's simply a matter of using the definitions of matrix multiplication, transpose and trace. The trace of a matrix is a number. There's no need to use any knowledge of what the components of F are. I'll go one step further and say that it doesn't help at all to know what the components are.

AxiomOfChoice said:
But what is the difference between \partial^\nu A_\nu and \partial_\mu A^\mu?

\partial^\nu A_\nu=\eta^{\nu\rho}\partial_\rho\eta_{\nu\sigma}A^\sigma=\delta^\rho_\sigma\partial_\rho A^\sigma=\partial_\mu A^\mu
 
  • #12
AxiomOfChoice said:
[...]<br /> F_{\mu \nu} F^{\mu \nu} = 2 (\partial_\mu \partial^\mu A_\nu A^\nu - \partial_\nu A^\nu \partial_\mu A^\mu)<br />

You wrote that, which is wrong.

\partial_\mu \partial^\mu A_\nu A^\nu \neq \partial_{\mu} A^{\nu}\partial_{\nu} A^{\mu}
 
  • #13
Fredrik said:
I'm also wondering what this has to do with your original question. What you overlooked was, as you said yourself, that repeated indices are summed over. Now that you know that, you should see that it's simply a matter of using the definitions of matrix multiplication, transpose and trace. The trace of a matrix is a number. There's no need to use any knowledge of what the components of F are. I'll go one step further and say that it doesn't help at all to know what the components are.
\partial^\nu A_\nu=\eta^{\nu\rho}\partial_\rho\eta_{\nu\sigma}A^\sigma=\delta^\rho_\sigma\partial_\rho A^\sigma=\partial_\mu A^\mu

Did you just use eta to denote the metric tensor? Can I formally protest because I'm much more used to seeing it as g? >_>
 
  • #14
Matterwave said:
Did you just use eta to denote the metric tensor? Can I formally protest because I'm much more used to seeing it as g? >_>

Well presumably the OP is referring to electromagnetism in flat, aka minkowski, spacetime, rather than a general spacetime... In which case we usually write eta instead of g.
 
  • #15
Oh, even in minkowski spacetime, I usually see g=diag(1,-1,-1,-1). This is the convention I am familiar with haha.
 
  • #16
I use g for the metric tensor and also for its matrix of components in a coordinate system, but η is the specific matrix you mentioned (sometimes defined with the opposite sign), so Minkowski spacetime is defined by the choice "g=η, in an inertial coordinate system". In a coordinate-independent expression, I would always use g for the metric tensor, even if we're dealing with Minkowski spacetime.
 
  • #17
bigubau said:
You wrote that, which is wrong.

\partial_\mu \partial^\mu A_\nu A^\nu \neq \partial_{\mu} A^{\nu}\partial_{\nu} A^{\mu}

Sorry...I don't see what the issue is. Obviously,

<br /> (\partial_\mu A_\nu - \partial_\nu A_\mu)(\partial^\mu A^\nu - \partial^\nu A^\mu) = \partial_\mu A_\nu \partial^\mu A^\nu - \partial_\mu A_\nu \partial^\nu A^\mu - \partial_\nu A_\mu \partial^\mu A^\nu + \partial_\nu A_\mu \partial^\nu A^\mu.<br />

I don't see how one gets the \partial_{\mu} A^{\nu}\partial_{\nu} A^{\mu} you mention from that. As best I can tell, the mistake I have made is commuting some of the derivatives; i.e., that \partial_\mu \partial^\mu A_\nu A^\nu \neq \partial_\mu A_\nu \partial^\mu A^\nu. But that raises another issue: What can I commute in this business, and what can't I?

Thanks for your help, and for being patient. I'm in a Intro GR class where the professor assumes familiarity with this notation. Which I lack.
 
  • #18
Fredrik said:
There's no need to use any knowledge of what the components of F are. I'll go one step further and say that it doesn't help at all to know what the components are.

Really? I don't see how I'm supposed to eventually get E and B into this business without referring to what the components of F are.
 
  • #19
AxiomOfChoice said:
Really? I don't see how I'm supposed to eventually get E and B into this business without referring to what the components of F are.
I thought the original question was "how can this be a number and not a matrix?", but if you also want to know why the right-hand side is that specific combination of E and B, you obviously have to use the equalities that describe how A is related to B,E and F.

AxiomOfChoice said:
As best I can tell, the mistake I have made is commuting some of the derivatives; i.e., that \partial_\mu \partial^\mu A_\nu A^\nu \neq \partial_\mu A_\nu \partial^\mu A^\nu. But that raises another issue: What can I commute in this business, and what can't I?
This isn't just a matter of commuting derivative operators. You need to keep track of when they act on one of the factors and when they act on a product of two of them. Use the product rule when necessary.
 
  • #20
Fredrik said:
This isn't just a matter of commuting derivative operators. You need to keep track of when they act on one of the factors and when they act on a product of two of them. Use the product rule when necessary.

Ok. I'm trying to work this out. I'm guessing that the expression

<br /> \partial_\mu A_\nu \partial^\mu A^\nu,<br />

corresponds to a sum that contains 16 summands, and that the (0,0) summand looks like

<br /> \underbrace{\frac{1}{c} \frac{\partial}{\partial t} \frac{\phi}{c}}_{\partial_0 A_0} \underbrace{\frac{1}{c} \frac{\partial}{\partial t} \frac{\phi}{c}}_{\partial^0 A^0}.<br />

Are you saying that this is actually

<br /> \frac{1}{c^4} \frac{\partial}{\partial t} \left( \phi \frac{\partial \phi}{\partial t} \right)<br />
 
  • #21
No, what I'm saying is that if you want to move the derivative operators around in that middle expression, you have to do it with the product rule in mind: Df Dg = D(Df g)-D2f g.
 
  • #22
Fredrik said:
No, what I'm saying is that if you want to move the derivative operators around in that middle expression, you have to do it with the product rule in mind: Df Dg = D(Df g)-D2f g.

Well, that's what I thought I was saying...so is it wrong to say

<br /> \partial_0 A_0 \partial^0 A^0 = \frac{1}{c^4} \left( \left( \frac{\partial \phi}{\partial t} \right)^2 + \phi \frac{\partial^2 \phi}{\partial t^2}\right)<br />

?
 
  • #23
Yes it's wrong, because the left-hand side is equal to the first term (give or take minus signs and factors of c...I didn't check those details).
 
  • #24
Fredrik said:
Yes it's wrong, because the left-hand side is equal to the first term (give or take minus signs and factors of c...I didn't check those details).

Wow... :frown: ... I'm starting to feel like a complete idiot...so how do I know that in the expression

<br /> \underbrace{\frac{1}{c} \frac{\partial}{\partial t} \frac{\phi}{c}}_{\partial_0 A_0} \underbrace{\frac{1}{c} \frac{\partial}{\partial t} \frac{\phi}{c}}_{\partial^0 A^0},<br />

that first partial derivative with respect to t only acts on the first \phi and not the second one?
 
  • #25
Because the product of the two real numbers f'(x) and g'(x) is f'(x)g'(x), not D(Df·g)(x). It really is as simple as that. :smile:
 
  • #26
Fredrik said:
Because the product of the two real numbers f'(x) and g'(x) is f'(x)g'(x), not D(Df·g)(x). It really is as simple as that. :smile:

Ok. Well, first of all, thanks for all of your help. Second of all, are there any good online resources you could recommend for learning this stuff systematically, as opposed to bugging you guys incessantly?
 
  • #27
I don't know any, but it might help you to look at some older threads about expressions with lots of indices. One thing that will definitely help is to simply decide to make sure at every step that you're using the definitions and not throwing in assumptions of your own. I think you're finding this difficult just because you expect it to be much more difficult than it really is. These expression don't involve anything more complicated than sums and products of real numbers, and partial derivatives.
 
  • #28
Fredrik said:
I don't know any, but it might help you to look at some older threads about expressions with lots of indices. One thing that will definitely help is to simply decide to make sure at every step that you're using the definitions and not throwing in assumptions of your own. I think you're finding this difficult just because you expect it to be much more difficult than it really is. These expression don't involve anything more complicated than sums and products of real numbers, and partial derivatives.

You're probably right. But I'm trying not to throw in any assumptions. See, for example, the \partial_0 A_0 \partial^0 A^0 confusion with the first partial derivative and what it acts on...there is some convention that says the first partial acts ONLY on the first \phi and not on the second. What is this convention? Because, if one were to just write out

<br /> \partial_0 A_0 \partial^0 A^0 = \frac{1}{c} \frac{\partial}{\partial t} \frac{\phi}{c} \frac{1}{c} \frac{\partial}{\partial t} \frac{\phi}{c},<br />

it seems to require an assumption not to let that first \partial_t act through to the second occurrence of \phi... :confused:
 
  • #29
What you're saying about the notation is true. It's not 100% obvious what it means, if you only look at the notation. But in this case, you can also look at the expression you obtained it from. You started with

F_{\mu\nu}=\partial_\mu A_\nu-\partial_\nu A_\mu

F_{\mu\nu}F^{\mu\nu}=(\partial_\mu A_\nu - \partial_\nu A_\mu)(\partial^\mu A^\nu - \partial^\nu A^\mu)

Let's look at the expression \partial_\mu A_\nu. This is just a function from \mathbb R^4 into \mathbb R. Alternatively, you can interpret it as meaning \partial_\mu A_\nu(x) with the x suppressed (i.e. not typed because you're trying to save time). This is just a real number. Let's go with the latter interpretation. Then the expression above is of the form (A+B)(C+D) where A,B,C,D are real numbers. You know that this is equal to AC+AD+BC+BD, and somehow you still managed to convince yourself that the product of A and C isn't AC. :smile:

If you find the notation confusing, just insert parentheses where you think they're needed. Most authors seem to use the convention that a derivative operator only acts on the first thing that's immediately to its right, and if you can remember that, you can avoid having to type a lot of parentheses.
 
Last edited:
Back
Top