Do Tensor Fields Commute When Multiplied?

  • Thread starter Thread starter Kreizhn
  • Start date Start date
  • Tags Tags
    Field Tensor
Kreizhn
Messages
714
Reaction score
1

Homework Statement


The SR&GR guy aren't being very helpful, so maybe I can get this quickly resolved here.

I want to know if the product of two tensor fields is generally non-commutative. That is, if I have two tensor fields A_{ij}, B_k^\ell do these representations commute?

The Attempt at a Solution



I feel generally quite conflicted about this subject, and I think it's because I don't fully understand what the representations mean. On one hand, I want to say that for a fixed i,j these simply represent scalar elements and so certainly commute. However, taken as general tensors (for example matrices), they would not commute. That is, if A and B were matrices, then AB \neq BA in general, but given the representations (A)_{ij} = a_{ij}, (B)_{ij} = b_{ij} then certainly a_{ij}b_{k\ell} = b_{k\ell} a_{ij} - the only "non-commutativity" comes in the ordering of the indices. Can anybody shed some light on this situation?
 
Physics news on Phys.org
If A and B are two (0, 2) tensors with components a_{ij} and b_{ij} respectively, then the (0, 4) tensor with components c_{ijkl} = a_{ij} b_{kl} is the tensor product A \otimes B. You are correct to observe that this tensor differs from B \otimes A, which has components c'_{ijkl} = b_{ij} a_{kl}, only in the ordering of the indices -- but since the order of the indices for tensors is very important, you can't think of \otimes as a commutative product.

When you mention matrices, you are talking about a different product, which is noncommutative in a more fundamental way. In their role as linear transformations, you can think of matrices as (1, 1) tensors. If A and B are matrices expressed this way, with components a^i_j and b^i_j, then the matrix product AB (composition of linear transformations) has components c^i_j = \sum_k a^i_k b^k_j, or simply a^i_k b^k_j in the Einstein summation convention. In tensor language, the matrix product (composition) is actually reflected as a contraction (which is also how the product of two (1, 1) tensors can be another (1, 1) tensor and not a (2, 2) tensor, as the tensor product would be).
 
ystael said:
If A and B are two (0, 2) tensors with components a_{ij} and b_{ij} respectively, then the (0, 4) tensor with components c_{ijkl} = a_{ij} b_{kl} is the tensor product A \otimes B. You are correct to observe that this tensor differs from B \otimes A, which has components c'_{ijkl} = b_{ij} a_{kl}, only in the ordering of the indices -- but since the order of the indices for tensors is very important, you can't think of \otimes as a commutative product.

When you mention matrices, you are talking about a different product, which is noncommutative in a more fundamental way. In their role as linear transformations, you can think of matrices as (1, 1) tensors. If A and B are matrices expressed this way, with components a^i_j and b^i_j, then the matrix product AB (composition of linear transformations) has components c^i_j = \sum_k a^i_k b^k_j, or simply a^i_k b^k_j in the Einstein summation convention. In tensor language, the matrix product (composition) is actually reflected as a contraction (which is also how the product of two (1, 1) tensors can be another (1, 1) tensor and not a (2, 2) tensor, as the tensor product would be).

Some prefer to use (1,1) tensors as matrices and some say that (0,2) and (2,0) tensors are to be called the second-rank matrices which of course sounds so correct. The reason is that 4-by-4 matrices (on a 4d spacetime) are mixed tensors that are not that much identified among physicists and in their language you can find a lot of things like a metric tensor is a second-rank square matrix and if this is the case, then claiming mixed tensors as being of the nature of the same matrices does seem absurd. Besides, if we represent v^i (i=0,...,3) as a 1-by-4 matrix (i.e. a row-vector) and a mixed tensor as a 4-by-4 matrix, then from the transformation formula

v^i=\frac{\partial x^i}{\partial \bar{x}^j}\bar{v}^j

one would expect to have a (4\times 4)(1\times 4) = (4\times 4) vector which is absurd whereas if the transformation formula was written as

v^i=\bar{v}^j\frac{\partial x^i}{\partial \bar{x}^j},

everything would be okay. The same situation happens to exist when one wants to take an upper index down or vice versa using the metric matrix g_{ij}, i.e.

v_i=g_{ij}v^j,

then taking the preceding path, a 4-by-4 matrix is to be assigned to a vector vi!

So to simply answer our OP's question about why non-commutativity does not hold for componential representation of matrices\tensors, I got to say that you can readily change the position of two numbers under the usual operation of multiplication, while you can't do the same stuff to an arrangement of numbers with a different law of multiplication which is, deep down, not commutative. So we you deal with tensors (actually with rank 2) as matrices, or a complex of matrices and tensors together, like Ag_{ab}C where A and C are 4-by-4 matrices and g_{ab} is the second-rank metric tensor, then non-commutativity is to be strongly considered in our calculations.

And I assume that you know for second-rank tensors, tensor rank and matrix rank are the same. So these things won't work if we are given something like a (0,3) tensor.

AB
 
Last edited:
Prove $$\int\limits_0^{\sqrt2/4}\frac{1}{\sqrt{x-x^2}}\arcsin\sqrt{\frac{(x-1)\left(x-1+x\sqrt{9-16x}\right)}{1-2x}} \, \mathrm dx = \frac{\pi^2}{8}.$$ Let $$I = \int\limits_0^{\sqrt 2 / 4}\frac{1}{\sqrt{x-x^2}}\arcsin\sqrt{\frac{(x-1)\left(x-1+x\sqrt{9-16x}\right)}{1-2x}} \, \mathrm dx. \tag{1}$$ The representation integral of ##\arcsin## is $$\arcsin u = \int\limits_{0}^{1} \frac{\mathrm dt}{\sqrt{1-t^2}}, \qquad 0 \leqslant u \leqslant 1.$$ Plugging identity above into ##(1)## with ##u...
Back
Top