TimeRip496 said:
I am reading a notes about tensor when I came across this which the notes did not elaborate more on it. As a result I don't quite understand why.
Here it is : " Note that we mark the covariant basis vectors with an upper index and the contravariant basis vectors with a lower index. This may sounds counter-intuitive ('did we not decide to use upper indices for contravariant vectors?') but this is precisely what we mean with the 'different meaning of the indices' here: this time they label the vectors and do not denote their components. "
I can follow except the last sentence and I do not know why. Can anyone enlighten me?
If you've taken vector calculus, you probably have seen a 2-D vector \vec{A} written in the form A^x \hat{x} + A^y \hat{y}. In that notation, \hat{x} means a "unit vector" in the x-direction, while the coefficient A^x means the component of \vec{A} in that direction. When you get to relativity, the notion of a "unit vector" becomes not well-defined, so the more general notion is a "basis vector". You would write an arbitrary vector \vec{A} in the form \sum_\mu A^\mu e_\mu, where the sum ranges over all basis vectors (there are 4 in SR--3 spatial directions and one time direction). By convention, people leave off the \sum_\mu and it's assumed that if an index appears in both lowered and raised forms, then it means that it is summed over. So people would just write a vector as A^\mu e_\mu
Now, although the components A^\mu are different in different coordinate systems, so people say that the vector "transforms" when you change coordinates, the combination A^\mu e_\mu is actually coordinate-independent. The vector has the same value, as a vector, in every coordinate system. What that means is that if you change coordinates from x^\mu to some new coordinates x^\alpha, the value of \vec{A} doesn't change:
A^\mu e_\mu = A^\alpha e_\alpha
The components A^\mu change, and the basis vectors e_\mu change, but the combination remains the same.
We can relate the old and new components through a matrix L^\alpha_\mu:
A^\alpha = L^\alpha_\mu A^\mu
If we use this matrix to rewrite A^\alpha in our equation relating the two vectors, we see:
A^\mu e_\mu = L^\alpha_\mu A^\mu e_\alpha = A^\mu (L^\alpha_\mu e_\alpha)
Note that since this equation holds for any vector \vec{A}, it must mean that
e_\mu = L^\alpha_\mu e_\alpha
or if we let (L^{-1})^\mu_\alpha be the inverse matrix, we can apply it to both sides to get:
(L^{-1})^\mu_\alpha e_\mu = e_\alpha
So we have the pair of transformation equations:
- A^\alpha = L^\alpha_\mu A^\mu
- e_\alpha = (L^{-1})^\mu_\alpha e_\mu
The basis vectors e_\mu transform in the opposite way from the components A^\mu, so that the combination A^\mu e_\mu has the same value in every coordinate system.