What is derivative of a vector respect to another vector?

  • Thread starter Thread starter yungman
  • Start date Start date
  • Tags Tags
    Derivative Vector
yungman
Messages
5,741
Reaction score
294
I am confused. I never seen derivative of a vector respect to another vector. When I go on the web, the article just show divergence, curl, gradient etc. But not derivative of a vector respect to another vector?

For example what is

\frac{d(\vec{x}-\vec{x_0})^2}{d \vec{x}} ?

For \vec{x_0} is a constant vector.


The book seems to imply:

\frac{d[(\vec{x}-\vec{x_0})^2]}{d \vec{x}} = 2(\vec{x}-\vec{x_0}) \frac{d \vec{x}}{d \vec{x}} = 2(\vec{x}-\vec{x_0})

I guess I don't know how to do a derivative like this. Can anyone help? I have looked through the multiple variable book and nothing like this. The variable is always scalar. The closest I seen is:

\int_C \vec{F} \cdot d\vec{r} \;=\; \int_C \vec{F} \cdot \hat{r}dr

But this is not exactly what the book discribed.

The only one that is remotely close is Directional Derivative which I don't think so.
 
Last edited:
Physics news on Phys.org
Wow, I agree that is really confusing notation.
Probably, they mean
(\vec x - \vec x_0)^2 = (\vec x - \vec x_0) \cdot (\vec x - \vec x_0)
so the square is actually a scalar.

Then in components, you could write
<br /> \left( \frac{\mathrm d [(\vec x - \vec x_0)^2] }{ \mathrm d\vec x } \right)_j = <br /> \frac{\mathrm d [(\vec x - \vec x_0)^2] }{ \mathrm d\vec x_j } =<br /> 2(\vec x - \vec x_0)_j = 2(\vec x_j - (\vec{x_0})_j) )<br />

If you doubt this, you can write out
(\vec x - \vec x_0) \cdot (\vec x - \vec x_0) = \left( \sum_{i = 1}^n (\vec x_i)^2 \right) + 2 \left( \sum_{i = 1}^n (\vec x_i) (\vec x_0)_i \right) + \left( \sum_{i = 1}^n ((\vec{x_0})_i)^2 \right)
and use that
\frac{\mathrm d}{\mathrm d \vec x_j} \left( \sum_{i = 1}^n (\vec x_i)^2 \right) = 2 \vec x_j
etc
 
yungman said:
The book seems to imply ... But this is not exactly what the book discribed.

What book?
 
The derivative of a vector function of a vector \vec f(\vec x) with respect to a vector \vec x is a 1-1 tensor, with the i,j element being

\frac{\partial f_i(\vec x)}{x_j}}

However, (\vec x - \vec x_0)^2 is not a vector function. It is a scalar. You are just calculating the gradient:

\nabla f(\vec x) = \sum_j \frac{\partial f(\vec x)}{x_j}}\hat x_j

Note that the gradient looks a lot like a vector. It is better thought of as being a covector.

So what about the gradient of f(\vec x) = (\vec x - \vec x_0)^2? Expanding this, we get

f(\vec x) = (\vec x - \vec x_0)\cdot (\vec x - \vec x_0) = \sum_i (x_i - x_{0,i})^2

Taking the gradient, the jth of the gradient is

\left(\nabla f(\vec x)\right)_j = \sum_i 2 (x_i - x_{0,i}) \frac{\partial x_i}{\partial x_j}<br /> = \sum_i 2 (x_i - x_{0,i})\delta_{ij} = 2(x_j - x_{0,j})
 
The book is PDE by Strauss p194 to p195. It is part of the derivation of the Green's Function for sphere. The part is about normal derivative of G. It talked about derivation respect to \vec{x} and some very funcky statement I still don't understand. But the later part just went back to the ordinary definition of normal derivative:

\frac{\partial G}{\partial n} = \nabla G \cdot \hat{n}

and derive the equation accordinary as if nothing happened! So it is a non question at this point. Strauss is not a good book in any stretch. I just cannot find any PDE book that cover the Green's Function and the EM book that I ordered is still in shipment!

Thanks

Alan
 
amazon.com's search function let's me look at some, but not all, of the pages in the book. Do you mean the statement
Let's not forget that \bold{x}_0 is considered to be fixed, and the derivatives are with respect \bold{x}.

on page 185?

Notice that equation (10) on page 185 gives G as a function of both \bold{x} and \bold{x}_0, so the quoted statement just means that normal partial derivatives, gradients, divergences, etc., are with respect to the coordinates of \bold{x} and not with respect to the coordinates of \bold{x}_0. The quoted statement does not actually mean "take the derivative with respect to a vector."
 
George Jones said:
amazon.com's search function let's me look at some, but not all, of the pages in the book. Do you mean the statement


on page 185?

Notice that equation (10) on page 185 gives G as a function of both \bold{x} and \bold{x}_0, so the quoted statement just means that normal partial derivatives, gradients, divergences, etc., are with respect to the coordinates of \bold{x} and not with respect to the coordinates of \bold{x}_0. The quoted statement does not actually mean "take the derivative with respect to a vector."

Yes, that is the sentence I refer to. I just took it literally that derivative with respect to \vec{x}. I have absolutely no issue on the normal derivative. That is the reason I reposted that I have no question on the complete derivation.

I have not gone into the exercise yet. I still have one more question regarding to a zero vector in my other post that I got stuck! If you can help, that would be really appreciated.

Thanks.
 
Back
Top