Aero51 said:
I really wish I had a better background in higher level mathematics beyond calculus and linear algebra so these concepts could be understood more rigorously. When you wrote the gradient in index notation, were you using contra-variant components instead of covariant components (usually written as a lower subscript in the texts I've seen). Too tell you the truth, I'm not even sure what the difference is between co-variant and contra-variant components of a tensor are. The bad thing about engineering is that nobody cares about mathematics..
:(
Really you didn't have to have more advanced math courses to run into this stuff (in a perfect world, I realize this subject is highly neglected in our curriculum imo), it's really only the application of things you already know from vector calculus/linear algebra applied to the case of a generalized curvilinear coordinate system. Often this stuff gets introduced (by a good professor) when talking about generalized curvilinear coordinates at the beginning of an upper level undergraduate physics course like mechanics or E&M. However, the formal math course you would find all of this stuff in is differential geometry.
Aero51 said:
Also, why do author's stick with using co-variant notation when multiplying vector components by their basis (in cartesian coordinates) even though this is not the "true" mathematics that should be carried out. It is very misleading! For instance, many authors define the gradient using index notation as:
grad(x) = \sum e_i \partial_i
Panton doesn't even bother to include the basis vector.
Really this is because using contravariant/covariant notation when students are just doing problems in cartesian coordinates (where there's no difference between contravariant/covariant vectors) in basic vector calculus can raise more questions than it answers. I agree though, it's inconsistent, but it's for pedagogical purposes I'm sure.
I recently wrote a pedagogical guide to tensors for undergraduate physics majors this summer, and so I would like to copy and paste the treatment I gave to basis vectors and dual basis vectors as it's probably the most basic and straightforward way to introduce the subject:
Every vector can be represented by either specifying the contravarient components along with the usual basis vectors, "\vec{v} = v^{1}\vec{e}_{1} + v^{2}\vec{e}_{2} + v^{3}\vec{e}_{3}", or the very same vector can be represented with covarient components and "dual basis vectors", "\vec{v} = v_{1}\vec{e}^{1} + v_{2}\vec{e}^{2} + v_{3}\vec{e}^{3}". Ordinarily, in orthogonal coordinate systems (\hat{x} \perp \hat{y} \perp \hat{z}), such distinctions aren't noticed because both the contravariant and covariant components of the vector are equal, as well as the fact that the basis vectors are equal to the dual basis vectors. But let's see how this changes with non-orthogonal coordinate systems and why these contravariant and covariant concepts should be introduced.
Consider vector '
A' in
fig 1. Consider how one would find the components for vector
A if the 'y' axis were rotated as shown in the figure (the 'y' axis is rotated to form the new axis 'y', where as the new 'x' axis is not changed and simply overlays the old 'x' axis.) This new primed coordinate system is known as a non-orthogonal coordinate system, and in such systems the same vector can be represented multiple ways. One way one could represent
A' would be to run lines parallel to the primed axises and see where they intercept 'x' and 'y' and make these our components for
A. In doing so we come up with the contravariant vector components for
A, A^{x} and A^{y}. But what if we found the components of
A by running perpendicularly to the primed coordinate system to find our components instead? In doing that, we come up with the covariant components A_{x} and A_{y}. But how can these both be components of the same vector? Surely A_{x'} \vec{e}_{x'} + A_{y'} \vec{e}_{y'} \neq A^{x'} \vec{e}_{x'} + A^{y'} \vec{e}_{y'}, well this would be a correct assessment, they're NOT equal, but which one of the component schemes is the culprit?
<
Insert Figure 1 which I've added as a thumbnail attachment at the bottom of this post..>
Notice that only one of these schemes produces the true vector
A upon vector addition, and that is the vector "A^{x'} \vec{e}_{x'} + A^{y'} \vec{e}_{y'}". This can clearly be seen by applying the parallelogram rule for vector addition to each vector geometrically. So does this make the covariant components worthless? Not really, you just have to choose different basis vectors to maintain the legitimate representation of the vector
A. These new basis vectors, often called "dual basis vectors" or "1 forms" will allow you to represent
A in the usual way, notice that when writing the dual basis vectors I simply raise the index as if I were working with the components: \vec{A} = A_{x'} \vec{e}^{x'} + A_{y'} \vec{e}^{y'}. The dual basis vectors are found by rules for their dot products with all original basis vectors. For original basis vectors with different indices than the basis vector in question, the dot product must be zero, and for the basis vector of the same index, the dot product must be 1.
\vec{e}^{i} \circ \vec{e}_{j} = {\delta}_{ij}
Note that there is no need for the dual basis vectors to have the same magnitude as their original basis vector counterparts. From this equation (or rather to meet the needs of this equation), the relationship between the dual basis vectors and the original basis vectors can be found in three dimensions, and the 3 equations for each dual basis vector are as follows:
\vec{e}^{1} = \frac{{\vec{e}_{2}} \times {\vec{e}_{3}}}{{\vec{e}_{1}} \circ {\vec{e}_{2}} \times {\vec{e}_{3}}}
\vec{e}^{2} = \frac{{\vec{e}_{3}} \times {\vec{e}_{1}}}{{\vec{e}_{1}} \circ {\vec{e}_{2}} \times {\vec{e}_{3}}}
\vec{e}^{3} = \frac{{\vec{e}_{1}} \times {\vec{e}_{2}}}{{\vec{e}_{1}} \circ {\vec{e}_{2}} \times {\vec{e}_{3}}}
These equations allow the dual basis vectors to be calculated if the original basis vectors are known. The equations were simply found by creating an object that would abide by our dot product rules stated previously using some ingenuity. You may have already seen these equations in physics if you've studied physics of the solid state, these were the equations relating the reciprocal lattice vectors to the primitive translational vectors, but now you know what these vectors are on a much deeper level!
Thus, we have now seen that any vector in both orthogonal and non-orthogonal coordinate systems can be represented two different ways:
\vec{A} = A^{1}\vec{e}_{1} + A^{2}\vec{e}_{2} + A^{3}\vec{e}_{3} = A_{1}\vec{e}^{1} + A_{2}\vec{e}^{2} + A_{3}\vec{e}^{3}
Now that was a nice little geometric approach to the contravariant vs. covariant distinction with vectors, but I would like to stress another major point about contravariant and covariant components. And that is that the real definition for contravariant and covariant components comes from transformation laws, the behavior of these components under a change of coordinate system, the same exact transformation laws that are necessarily abided by tensors. What are these transformation laws? Well, there's a different transformation law for contravariant components versus covariant components. Anything that transforms like A^{i'} = \frac{{\partial}{x^{i'}}}{{\partial} x^{j}} A^{j} transforms "contravariantly", and anything that transforms like A_{i'} = \frac{{\partial}{x^{j}}}{{\partial} x^{i'}} A_{j} does so "covariantly". Why do the coefficients of the transformations look the way that they do (partial derivatives of the variables)? I make the claim that covariant vector components transform the same way that regular basis vectors do from one coordinate system to another. This is the origin of the term "covariant", "co-" being a prefix that means "with" or "together". Likewise, "contra-" means "against" or "opposite". Contravariant vector components transform in the opposite way as basis vectors (the transformation factors are "flipped"), instead transforming in the way that dual basis vectors transform from one coordinate system to another. Thus we need to study how standard basis vectors transform. We'll consider the transformation of a basis vector to a generalized curvilinear coordinate axis in one dimension (the process can then be expanded to take into account 3 or more dimensions).
The situation can be seen in
fig. 2. In curvilinear coordinates, the basis vectors change depending on where in the space you are located, so we define a set of axes (here called the 'b_{1}' axis) that are tangent to the curvilinear axes at a specific point. It is along these axes that the curvilinear basis vectors point. Now, we want to find the projection of the vector '\vec{b}_{1}' onto the 'x' axis ('\vec{b}_{1}' is simply a vector along the 'b_{1}' axis. It is
not, in general, normalized.):
comp_{x}^{b} = \frac{\vec{b}_{1} \circ \vec{e}_{1}}{|\vec{e}_{1}|} = \frac{|\vec{b}_{1}||\vec{e}_{1}|}{|\vec{e}_{1}|} \cos{\theta}
We can simplify this expression by noticing the triangle formed by the vectors \vec{e}_{1} and \vec{b}_{1}. Further, we notice that this same ratio can be described by differential length elements along each axis (you'll see in a second why we want to notice both relations to the "\cos{\theta}" term):
\cos{\theta} = \frac{|\vec{e}_{1}|}{|\vec{b}_{1}|} = \frac{d x_{1}}{d x '_{1}}
Plugging back into the projection equation:
comp_{x}^{b} = |\vec{b}_{1}| \frac{d x_{1}}{d x '_{1}}
This equation can be generalized into the 3 dimensions, projecting the curvilinear axis b_{1} onto each of the 3 cartesian axes to obtain a complete representation of \vec{b}_{1} in terms of the cartesian basis vectors:
\begin{eqnarray*}
\vec{b}_{1} &=& comp_{x}^{b} \vec{e}_{1} + comp_{y}^{b} \vec{e}_{2} + comp_{z}^{b} \vec{e}_{3} \\
&=& |\vec{b}_{1}| \frac{{\partial} {x_{1}}}{{\partial}{x '_{1}}} \vec{e}_{1} + |\vec{b}_{1}| \frac{{\partial}{x_{2}}}{{\partial}{x '_{1}}} \vec{e}_{2} + |\vec{b}_{1}| \frac{{\partial}{x_{3}}}{{\partial}{x '_{1}}} \vec{e}_{3}
\end{eqnarray*}
But we don't just want '
b' vectors that span the space, we want a basis for the curvilinear coordinate space, thus:
\frac{\vec{b}_{1}}{|\vec{b}_{1}|} = \vec{e'}_{1} = \frac{{\partial}{x_{i}}}{{\partial}{x '_{1}}} \vec{e}_{i}
Or in general for all curvilinear basis vectors the transformation between cartesian basis vectors and curvilinear basis vectors can be written as follows:
\vec{e'}_{j} = \frac{{\partial}{x_{i}}}{{\partial}{x '_{j}}} \vec{e}_{i}
And thus basis vectors transform covariantly.
<
Insert Figure 2 from the thumbnails..>
The reader might also recognize the transformation factors as components of the Jacobian transformation matrix:
J = \frac{\partial (x_{1}, x_{2}, x_{3})}{\partial ({x'}_{1}, {x'}_{2}, {x'}_{3})} = \begin{bmatrix}<br />
\frac{{\partial}{x_{1}}}{{\partial}{{x'}_{1}}} & \frac{{\partial}{x_{1}}}{{\partial}{{x'}_{2}}} & \frac{{\partial}{x_{1}}}{{\partial}{{x'}_{3}}} \\<br />
\frac{{\partial}{x_{2}}}{{\partial}{{x'}_{1}}} & \frac{{\partial}{x_{2}}}{{\partial}{{x'}_{2}}} & \frac{{\partial}{x_{2}}}{{\partial}{{x'}_{3}}} \\<br />
\frac{{\partial}{x_{3}}}{{\partial}{{x'}_{1}}} & \frac{{\partial}{x_{3}}}{{\partial}{{x'}_{2}}} & \frac{{\partial}{x_{3}}}{{\partial}{{x'}_{3}}}<br />
\end{bmatrix}
One now sees why vectors are invariant. If a vector is constructed from components and basis vectors that each use these different, but related, transformations, the result will be the canceling out of the transformation factors and nothing will happen to the actual vector.
\vec{A} = A^{i} \vec{e}_{i}
\frac{{\partial}{x^{i'}}}{{\partial}{x^{i}}} \frac{{\partial}{x^{i}}}{{\partial}{x^{i'}}} \vec{A} = \frac{{\partial}{x^{i'}}}{{\partial}{x^{i}}} A^{i} \frac{{\partial}{x^{i}}}{{\partial}{x^{i'}}} \vec{e}_{i}
\vec{A} = A^{i} \vec{e}_{i}
Finally, notice that the inner product between a contravariant and covariant vector is an invariant scalar quantity:
s = \vec{A} \circ \vec{A} = A_{i} A^{i}
This is easily proved upon, again, multiplication of both sides of the equation by the transformation factors and noticing their cancelation:
\frac{{\partial}{x^{i}}}{{\partial}{x^{i'}}} \frac{{\partial}{x^{i'}}}{{\partial}{x^{i}}} s = A_{i} \frac{{\partial}{x^{i}}}{{\partial}{x^{i'}}} A^{i} \frac{{\partial}{x^{i'}}}{{\partial}{x^{i}}}
s = A_{i} A^{i}
And so the equation (and the quantity 's') is unaffected by a coordinate transformation.
This is NOT true (unless it's in cartesian coordinate systems where covariant and contravariant components are indistinguishable form each other) in the case of the inner product between between two covariant vectors or two contravariant vectors together.
I never had this article reviewed but I'm sure it's all correct..
Also, I agree with/like Murphid's posts on the subject.