## What is derivative of a vector respect to another vector?

I am confused. I never seen derivative of a vector respect to another vector. When I go on the web, the article just show divergence, curl, gradient etc. But not derivative of a vector respect to another vector?

For example what is

$$\frac{d(\vec{x}-\vec{x_0})^2}{d \vec{x}} ?$$

For $$\vec{x_0}$$ is a constant vector.

The book seems to imply:

$$\frac{d[(\vec{x}-\vec{x_0})^2]}{d \vec{x}} = 2(\vec{x}-\vec{x_0}) \frac{d \vec{x}}{d \vec{x}} = 2(\vec{x}-\vec{x_0})$$

I guess I don't know how to do a derivative like this. Can anyone help? I have looked through the multiple variable book and nothing like this. The variable is always scalar. The closest I seen is:

$$\int_C \vec{F} \cdot d\vec{r} \;=\; \int_C \vec{F} \cdot \hat{r}dr$$

But this is not exactly what the book discribed.

The only one that is remotely close is Directional Derivative which I don't think so.

 PhysOrg.com science news on PhysOrg.com >> Heat-related deaths in Manhattan projected to rise>> Dire outlook despite global warming 'pause': study>> Sea level influenced tropical climate during the last ice age
 Blog Entries: 5 Recognitions: Homework Help Science Advisor Wow, I agree that is really confusing notation. Probably, they mean $$(\vec x - \vec x_0)^2 = (\vec x - \vec x_0) \cdot (\vec x - \vec x_0)$$ so the square is actually a scalar. Then in components, you could write $$\left( \frac{\mathrm d [(\vec x - \vec x_0)^2] }{ \mathrm d\vec x } \right)_j = \frac{\mathrm d [(\vec x - \vec x_0)^2] }{ \mathrm d\vec x_j } = 2(\vec x - \vec x_0)_j = 2(\vec x_j - (\vec{x_0})_j) )$$ If you doubt this, you can write out $$(\vec x - \vec x_0) \cdot (\vec x - \vec x_0) = \left( \sum_{i = 1}^n (\vec x_i)^2 \right) + 2 \left( \sum_{i = 1}^n (\vec x_i) (\vec x_0)_i \right) + \left( \sum_{i = 1}^n ((\vec{x_0})_i)^2 \right)$$ and use that $$\frac{\mathrm d}{\mathrm d \vec x_j} \left( \sum_{i = 1}^n (\vec x_i)^2 \right) = 2 \vec x_j$$ etc

Mentor
 Quote by yungman The book seems to imply ... But this is not exactly what the book discribed.
What book?

Mentor

## What is derivative of a vector respect to another vector?

The derivative of a vector function of a vector $$\vec f(\vec x)$$ with respect to a vector $$\vec x$$ is a 1-1 tensor, with the i,j element being

$$\frac{\partial f_i(\vec x)}{x_j}}$$

However, $(\vec x - \vec x_0)^2$ is not a vector function. It is a scalar. You are just calculating the gradient:

$$\nabla f(\vec x) = \sum_j \frac{\partial f(\vec x)}{x_j}}\hat x_j$$

Note that the gradient looks a lot like a vector. It is better thought of as being a covector.

So what about the gradient of $f(\vec x) = (\vec x - \vec x_0)^2$? Expanding this, we get

$$f(\vec x) = (\vec x - \vec x_0)\cdot (\vec x - \vec x_0) = \sum_i (x_i - x_{0,i})^2$$

$$\left(\nabla f(\vec x)\right)_j = \sum_i 2 (x_i - x_{0,i}) \frac{\partial x_i}{\partial x_j} = \sum_i 2 (x_i - x_{0,i})\delta_{ij} = 2(x_j - x_{0,j})$$

 The book is PDE by Strauss p194 to p195. It is part of the derivation of the Green's Function for sphere. The part is about normal derivative of G. It talked about derivation respect to $\vec{x}$ and some very funcky statement I still don't understand. But the later part just went back to the ordinary definition of normal derivative: $$\frac{\partial G}{\partial n} = \nabla G \cdot \hat{n}$$ and derive the equation accordinary as if nothing happened!!!! So it is a non question at this point. Strauss is not a good book in any stretch. I just cannot find any PDE book that cover the Green's Function and the EM book that I ordered is still in shipment!!! Thanks Alan

Mentor
amazon.com's search function lets me look at some, but not all, of the pages in the book. Do you mean the statement
 Let's not forget that $\bold{x}_0$ is considered to be fixed, and the derivatives are with respect $\bold{x}$.
on page 185?

Notice that equation (10) on page 185 gives $G$ as a function of both $\bold{x}$ and $\bold{x}_0$, so the quoted statement just means that normal partial derivatives, gradients, divergences, etc., are with respect to the coordinates of $\bold{x}$ and not with respect to the coordinates of $\bold{x}_0$. The quoted statement does not actually mean "take the derivative with respect to a vector."

 Quote by George Jones amazon.com's search function lets me look at some, but not all, of the pages in the book. Do you mean the statement on page 185? Notice that equation (10) on page 185 gives $G$ as a function of both $\bold{x}$ and $\bold{x}_0$, so the quoted statement just means that normal partial derivatives, gradients, divergences, etc., are with respect to the coordinates of $\bold{x}$ and not with respect to the coordinates of $\bold{x}_0$. The quoted statement does not actually mean "take the derivative with respect to a vector."
Yes, that is the sentence I refer to. I just took it literally that derivative with respect to $\vec{x}$. I have absolutely no issue on the normal derivative. That is the reason I reposted that I have no question on the complete derivation.

I have not gone into the exercise yet. I still have one more question regarding to a zero vector in my other post that I got stuck!!! If you can help, that would be really appreciated.

Thanks.