Why Does d/dx Not Equal d/dx' Even When x Equals x'?

bigbalpha
Messages
4
Reaction score
0
Summary: I'm stuck on this simple excersize, to show that in this coord transform, despite x = x', d/dx != d/dx'

From "Intro to Smooth Manifolds" (this is a calculus excersize), The Problem I have is with showing d/dx != d/dx'
246034


When I write out the Jacobian matrix, I get exactly d/dx = d/dx' and not sure what I'm doing wrong, here is my work:
246035

[Moderator's note: Moved from a technical forum and thus no template.]
 
Last edited by a moderator:
Physics news on Phys.org
You have the transpose of the Jacobian in place of the Jacobian.
 
Alright, always get index notation very confused, for example, earlier in the book, it says this. This is contradictory?
246054
 
That notation looks very confused so I would not say it is strange that you are confused. Just from the image I am not sure what the author intends.

Anyway, your problem does not really need the Jacobian - just the chain rule for partial derivatives.
 
Orodruin said:
That notation looks very confused so I would not say it is strange that you are confused. Just from the image I am not sure what the author intends.

Which image, the one in post #1, or the one in post #3 ? And in what way is the author's intention unclear?
 
Post #3. By unclear I mean that it is not clear to me what the different quantities represent. It would be good to have the author’s definitions. If you have a better feeling for the notation, please feel free to answer the OP.
 
bigbalpha said:
Alright, always get index notation very confused, for example, earlier in the book, it says this. This is contradictory?
View attachment 246054

Matrix multiplication in terms of suffix notation is written a_i = M_{ij}b_j: the second index of the matrix is summed over. The first index labels the rows, and the second the columns. (This assumes we regard vectors as column vectors.)

What (3.9) is saying is that \frac{\partial F^j}{\partial x^i}(p) is the jth component of the image of the ith basis vector. Thus if you have a tangent vector X^i \left.\frac{\partial}{\partial x^i}\right|_p, then dF_p will send it to Y^j \left.\frac{\partial}{\partial y^j}\right|_{F(p)} where <br /> Y^j = \frac{\partial F^j}{\partial x^i}(p) X^i. Note that the index i is being summed over, so must be the second index (labelling columns) and j is the first index (labelling rows). This is of course the opposite of the standard convention, but consistent with the matrix following (3.9).

On the other hand, one naturally writes the chain rule, and thus the jacobian, as <br /> \frac{\partial}{\partial x&#039;^i} = \frac{\partial x^j}{\partial x&#039;^i} \frac{\partial}{\partial x^j}. Here, in accordance with the usual convention, the first index (labelling rows) is i and the second index (labelling columns) is j. Also, in this instance the partial derivative operators are the components, not the basis vectors.

TLDR: Pay attention to which index is being summed over. This index will label the columns.
 
  • Like
Likes bigbalpha
Getting back to the concrete exercise in the original post, @bigbalpha has, after matrix multiplication,
$$\begin{bmatrix} \frac{\partial}{\partial x} \\
\frac{\partial}{\partial y} \end{bmatrix}
=
\begin{bmatrix} \frac{\partial \tilde{x}}{\partial x} & \frac{\partial \tilde{x}}{\partial y} \\
\frac{\partial \tilde{y}}{\partial x} & \frac{\partial \tilde{y}}{\partial y} \end{bmatrix}
\begin{bmatrix} \frac{\partial}{\partial \tilde{x}} \\
\frac{\partial}{\partial \tilde{y}} \end{bmatrix}
=
\begin{bmatrix} \frac{\partial \tilde{x}}{\partial x} \frac{\partial}{\partial \tilde{x}} + \frac{\partial \tilde{x}}{\partial y} \frac{\partial}{\partial \tilde{y}} \\
\frac{\partial \tilde{y}}{\partial x} \frac{\partial}{\partial \tilde{x}} + \frac{\partial \tilde{y}}{\partial y} \frac{\partial}{\partial \tilde{y}} \end{bmatrix} .
$$

This gives, for example,
$$\frac{\partial}{\partial x} = \frac{\partial \tilde{x}}{\partial x} \frac{\partial}{\partial \tilde{x}} + \frac{\partial \tilde{x}}{\partial y} \frac{\partial}{\partial \tilde{y} },$$
which is an incorrect application of the chain rule.

Used correctly, the chain rule gives
$$\frac{\partial}{\partial x} = \frac{\partial \tilde{x}}{\partial x} \frac{\partial}{\partial \tilde{x}} + \frac{\partial \tilde{y}}{\partial x} \frac{\partial}{\partial \tilde{y} }.$$

The exercise is meant to illustrate an error (that I have seen in at least two general relativity books) that is sometimes called "Woodhouse's Second Fundamental Confusion of Calculus".
 
  • Like
Likes bigbalpha
To see how this (that ##\partial/\partial x \neq \partial/\partial \tilde x## even when ##x = \tilde x##) arises, we need to consider what partial derivatives actually means. A partial derivative is the change in a function when you change only one of its parameters while keeping the other parameters constant. The partial derivative ##\partial f/\partial x## means the change in ##f## with ##x## while keeping ##y## constant whereas ##\partial f/\partial \tilde x## is the change in ##f## with ##\tilde x## while keeping ##\tilde y## constant. In the first case, we would compare the value of ##f(x,y_0)## with the value of ##f(x+\epsilon,y_0)##. Consider now the case ##\tilde x = x## and ##\tilde y = x+y##. The partial derivative ##\partial f/\partial \tilde x## now means comparing ##f(x,\tilde y_0 + x)## with ##f(x + \epsilon,\tilde y_0 + x +\epsilon)##, since we must keep ##\tilde y## constant. This is not the same as when taking the partial derivative with respect to ##x##.
 
  • #10
pasmith said:
Matrix multiplication in terms of suffix notation is written a_i = M_{ij}b_j: the second index of the matrix is summed over. The first index labels the rows, and the second the columns. (This assumes we regard vectors as column vectors.)

What (3.9) is saying is that \frac{\partial F^j}{\partial x^i}(p) is the jth component of the image of the ith basis vector. Thus if you have a tangent vector X^i \left.\frac{\partial}{\partial x^i}\right|_p, then dF_p will send it to Y^j \left.\frac{\partial}{\partial y^j}\right|_{F(p)} where <br /> Y^j = \frac{\partial F^j}{\partial x^i}(p) X^i. Note that the index i is being summed over, so must be the second index (labelling columns) and j is the first index (labelling rows). This is of course the opposite of the standard convention, but consistent with the matrix following (3.9).

On the other hand, one naturally writes the chain rule, and thus the jacobian, as <br /> \frac{\partial}{\partial x&#039;^i} = \frac{\partial x^j}{\partial x&#039;^i} \frac{\partial}{\partial x^j}. Here, in accordance with the usual convention, the first index (labelling rows) is i and the second index (labelling columns) is j. Also, in this instance the partial derivative operators are the components, not the basis vectors.

TLDR: Pay attention to which index is being summed over. This index will label the columns.

Thanks a lot for that explanation, that really helped!
 
Back
Top