The Gradient of a Vector: Understanding Second Order Derivatives

  • Thread starter Thread starter Weather Freak
  • Start date Start date
  • Tags Tags
    Gradient Vector
Weather Freak
Messages
40
Reaction score
0
First off, this is not a homework problem, but rather is an issue that I've had for a while not and haven't quite been able to reason out to my satisfaction on my own.

u-vector = ui + vj + wk
What is grad(u-vector)?

I know what the gradient of a function is, but this is the gradient of a vector. I know what the answer is, because we did it a kazillion times in class, and I know how to get it by memorizing, but what is the technique at work here? There must be a method to the madness somewhere. I've tried looking up the gradient of a vector, gradient of a tensor (thinking there might be a general formula for gradient of a tensor that would reduce to gradient of a vector), but it has all led to nothing but confusion.

Could someone open my eyes a bit?

Thanks!

Kyle
 
Physics news on Phys.org
granpa said:
Neither of these is the gradient of a vector field. The divergence of a vector, \boldsymol{\nabla}\cdot \mathbf F, is a scalar while the curl of a vector field, \boldsymol{\nabla}\times \mathbf F, is a vector. The gradient of a vector field is a second order tensor:

(\boldsymol{\nabla}\mathbf F)_{ij} = \frac{\partial F_i(\boldsymbol x)}{\partial x_j}[/itex]<br /> <br /> One way to look at this: The <i>i<sup>th</sup></i> row of the gradient of a vector field \mathbf F(\mathbf x) is the plain old vanilla gradient of the scalar function F_i(\mathbf x).<br /> <br /> One place where the concept is useful is in forming a Taylor expansion of a scalar function. To first order,<br /> <br /> f(\mathbf x_0 + \Delta \mathbf x) \approx f(\mathbf x_0) +&lt;br /&gt; \boldsymol{\nabla}\mathbf f(\mathbf x)|_{\mathbf x=\mathbf x_0}\cdot \Delta \mathbf x<br /> <br /> Higher order expansions require higher order derivatives. The second order expansion requires taking the gradient of the gradient (i.e., taking the gradient of a vector).<br /> <br /> f(\mathbf x_0 + \Delta \mathbf x) \approx&lt;br /&gt; f(\mathbf x_0) +&lt;br /&gt; \sum_i&lt;br /&gt; (\boldsymol{\nabla}(\mathbf f))(\mathbf x_0)_i&lt;br /&gt; \Delta x_i +&lt;br /&gt; \sum_{i,j}&lt;br /&gt; \Delta x_i&lt;br /&gt; (\boldsymol{\nabla}(\boldsymol{\nabla}(\mathbf f)))(\mathbf x_0)_{i,j}&lt;br /&gt; \Delta x_j&lt;br /&gt;<br /> <br /> One application of this is computing the gravity gradient torque induced on a vehicle.
 
Last edited:
I would think that the curl of a vector field would only be a vector (technically a pseudovector, which is really a tensor) in 3 dimensions. in more than 3 its a tensor.

perhaps the tensor you are talking about is simply the true value of the curl. otherwise I have no idea what you are talking about.

I didnt mention grad because he asked for the gradient of a vector not a scalar.
 
I am not talking about curl, which is a pseudovector in three dimensions and generalizes to a N(N-1)/2\times N(N-1)/2 skew-symmetric tensor in N dimensions. I am talking about about the N\times N tensor

(\boldsymol{\nabla}\mathbf F)_{ij} = \frac{\partial F_i(\boldsymbol x)}{\partial x_j}[/itex]<br /> <br /> which I goofed up in my first post (now corrected).<img src="https://www.physicsforums.com/styles/physicsforums/xenforo/smilies/oldschool/redface.gif" class="smilie" loading="lazy" alt=":redface:" title="Red Face :redface:" data-shortname=":redface:" /><br /> <br /> If f(\mathbf x)is a scalar function, then the gradient \boldsymol{\nabla}f = \partial f(\mathbf x)/\partial x_i is a vector field. The &quot;gradient&quot; of this vector is what I was talking about in the second part of my post.<br /> <br /> Aside: Is there a name for the second-order spatial derivative \partial^2 f(\mathbf x)/\partial x_i \partial x_j?
 
D H said:
I am not talking about curl, which is a pseudovector in three dimensions and generalizes to a N(N-1)/2\times N(N-1)/2 skew-symmetric tensor in N dimensions. I am talking about about the N\times N tensor

(\boldsymol{\nabla}\mathbf F)_{ij} = \frac{\partial F_i(\boldsymbol x)}{\partial x_j}[/itex]<br /> <br /> which I goofed up in my first post (now corrected).<img src="https://www.physicsforums.com/styles/physicsforums/xenforo/smilies/oldschool/redface.gif" class="smilie" loading="lazy" alt=":redface:" title="Red Face :redface:" data-shortname=":redface:" /><br /> <br /> If f(\mathbf x)is a scalar function, then the gradient \boldsymol{\nabla}f = \partial f(\mathbf x)/\partial x_i is a vector field. The &quot;gradient&quot; of this vector is what I was talking about in the second part of my post.<br /> <br /> Aside: Is there a name for the second-order spatial derivative \partial^2 f(\mathbf x)/\partial x_i \partial x_j?
<br /> <br /> Thanks a lot, that definitely answers the question! The trick of each row being the gradient of Fi really makes it easy to remember as well.
 
I have a follow-up question to this thread. The gradient of a dot product is given

<br /> \nabla ( A \cdot B) = \underbrace{(B\cdot \nabla)A + (A\cdot \nabla)B}_{\mbox{gradients of vectors?}} + B \times (\nabla\times A) + A \times (\nabla \times B) <br />

All the terms in the equation should be vectors, not second order tensors, which is what gradients of vectors were explained to be earlier in this thread. How then to interpret the first two terms of the right hand side?

Also, the hint I've seen for deriving the identity is to use the "BAC-CAB" identity

<br /> A \times (B \times C) = B (A \cdot C) - C (A \cdot B),<br />

which can be rewritten

<br /> B(A \cdot C) = A \times (B \times C) + (A \cdot B) C.<br />

But using this to expand the gradient of the dot product of two vectors (letting B = \nabla, A = A, \mbox{ and } C = B) appears to yield

<br /> \nabla (A \cdot B) = A \times (\nabla \times B) + (A \cdot \nabla) B,<br />

which is not consistent with the expansion given in textbooks unless

(B \cdot \nabla) A + A \times (\nabla \times B) = 0,

and by symmetry, I don't think that can be true (wouldn't the whole right hand side of the textbook grad of dot product expansion then be zero?).
Please help.

Thanks, Genya
 
musemonkey said:
<br /> B(A \cdot C) = A \times (B \times C) + (A \cdot B) C.<br />

But using this to expand the gradient of the dot product of two vectors (letting B = \nabla, A = A, \mbox{ and } C = B) appears to yield

<br /> \nabla (A \cdot B) = A \times (\nabla \times B) + (A \cdot \nabla) B,<br />


You can't do that. \nabla is not a vector. It's an differential operator that in some cases we can treat like a vector but this is not one of those cases because in doing so you didn't take account of the product rule.

For example, for ordinary vectors \mathbf{A} and \mathbf{B} and a scalar function \psi, the following identity holds:

\mathbf{A} \cdot (\psi \mathbf{B}) = \psi \mathbf{A}\cdot\mathbf{B}

If you were to just plug in \mathbf{A} &quot;=&quot; \nabla now, you would arrive at

\nabla \cdot (\psi \mathbf{B}) = \psi \nabla \cdot \mathbf{B}

which is simply not correct. The correct expression is

\nabla \cdot (\psi \mathbf{B}) = \psi \nabla \cdot \mathbf{B} + \mathbf{B}\cdot \nabla \psi
 
I thought it was OK to substitute \nabla into the BAC-CAB identity because Feynman Lectures vol. II sec. 2-7 contain the following:

A \times (B \times C) = B ( A \cdot C) - (A \cdot B) C
\nabla \times (\nabla \times h) = \nabla (\nabla \cdot h) - (\nabla \cdot \nabla) h

Thank you Mute for the speedy response, but I'm not sure what to make of it. It's still unclear what the terms of form {(B\cdot \nabla)A mean and how the derivation of the gradient of a dot product formula is supposed to be done using the BAC-CAB identity.
 
  • #10
musemonkey said:
<br /> \nabla ( A \cdot B) = \underbrace{(B\cdot \nabla)A + (A\cdot \nabla)B}_{\mbox{gradients of vectors?}} + B \times (\nabla\times A) + A \times (\nabla \times B) <br />

All the terms in the equation should be vectors, not second order tensors, which is what gradients of vectors were explained to be earlier in this thread. How then to interpret the first two terms of the right hand side?

The first term in the above equation in cartesian coordinates is

(B\cdot \nabla)A = \bmatrix<br /> b_x\frac{\partial a_x}{\partial x} +<br /> b_y\frac{\partial a_x}{\partial y} +<br /> b_z\frac{\partial a_x}{\partial z} \\<br /> b_x\frac{\partial a_y}{\partial x} +<br /> b_y\frac{\partial a_y}{\partial y} +<br /> b_z\frac{\partial a_y}{\partial z} \\<br /> b_x\frac{\partial a_z}{\partial x} +<br /> b_y\frac{\partial a_z}{\partial y} +<br /> b_z\frac{\partial a_z}{\partial z}<br /> \endbmatrix

One way to think of B\cdot\nabla is as defining a new operator:

B\cdot\nabla \equiv \sum_j b_j\frac{\partial}{\partial x_j}


Also, the hint I've seen for deriving the identity is to use the "BAC-CAB" identity ...

The path you took is, as Mute noted, invalid. The "BAC-CAB" identity can be used if one uses Feynman's notation:

\nabla(A\cdot B) = \nabla_A(A\cdot B) + \nabla_B(A\cdot B)<br /> = \nabla_A(B\cdot A) + \nabla_B(A\cdot B)

where \nabla_A only operates on A and \nabla_B only operates on B. Then one can "safely" use the BAC-CAB identity as you did:

\aligned<br /> \nabla_A(B\cdot A) &amp;= B \times (\nabla_A \times A) + (B \cdot \nabla_A) A \\<br /> &amp;= B \times (\nabla \times A) + (B \cdot \nabla) A \\<br /> \nabla_B(A\cdot B) &amp;= A \times (\nabla_B \times B) + (A \cdot \nabla_B) B \\<br /> &amp;= A \times (\nabla \times B) + (A \cdot \nabla) B \\<br /> \nabla(A\cdot B) &amp;= \nabla_A(B\cdot A) + \nabla_B(A\cdot B) \\<br /> &amp;= (B \cdot \nabla) A + (A \cdot \nabla) B + B \times (\nabla \times A) + A \times (\nabla \times B)<br /> \endaligned<br />

This, however, is too much sleight-of-hand for me. It happens to work. Your approach happened not to work.
 
  • #11
Thank you DH and Mute! You answered everything, and I appreciate your taking the time to write everything out.
 
  • #12
Hello,

About the second order spatial derivative \partial^2 f(\mathbf x)/\partial x_i \partial x_j which DH wrote above, in tensor notation can it be written as

\nabla(\nabla{f}(\mathbf x))?
 
Back
Top