The Gradient of a Vector: Understanding Second Order Derivatives

  • Thread starter Thread starter Weather Freak
  • Start date Start date
  • Tags Tags
    Gradient Vector
Click For Summary
SUMMARY

The discussion centers on the gradient of a vector field, specifically the mathematical expression for the gradient as a second-order tensor, defined as (\boldsymol{\nabla}\mathbf F)_{ij} = \frac{\partial F_i(\boldsymbol x)}{\partial x_j}. Participants clarify that the gradient of a vector field differs from the gradient of a scalar function, emphasizing its utility in Taylor expansions and applications such as gravity gradient torque. The conversation also addresses the complexities of deriving the gradient of a dot product and the proper interpretation of terms involving differential operators.

PREREQUISITES
  • Understanding of vector calculus, specifically gradients and tensors.
  • Familiarity with Taylor series expansions in multivariable calculus.
  • Knowledge of differential operators and their applications in physics.
  • Basic concepts of vector fields and their properties.
NEXT STEPS
  • Study the properties of second-order tensors in vector calculus.
  • Learn about Taylor expansions involving vector fields and their applications.
  • Explore the derivation of the gradient of a dot product using the BAC-CAB identity.
  • Investigate the implications of the gradient of a vector field in physical applications, such as fluid dynamics.
USEFUL FOR

Mathematicians, physicists, and engineers who require a deeper understanding of vector calculus, particularly those working with gradients and tensor analysis in applied contexts.

Weather Freak
Messages
40
Reaction score
0
First off, this is not a homework problem, but rather is an issue that I've had for a while not and haven't quite been able to reason out to my satisfaction on my own.

u-vector = ui + vj + wk
What is grad(u-vector)?

I know what the gradient of a function is, but this is the gradient of a vector. I know what the answer is, because we did it a kazillion times in class, and I know how to get it by memorizing, but what is the technique at work here? There must be a method to the madness somewhere. I've tried looking up the gradient of a vector, gradient of a tensor (thinking there might be a general formula for gradient of a tensor that would reduce to gradient of a vector), but it has all led to nothing but confusion.

Could someone open my eyes a bit?

Thanks!

Kyle
 
Physics news on Phys.org
granpa said:
Neither of these is the gradient of a vector field. The divergence of a vector, \boldsymol{\nabla}\cdot \mathbf F, is a scalar while the curl of a vector field, \boldsymol{\nabla}\times \mathbf F, is a vector. The gradient of a vector field is a second order tensor:

(\boldsymol{\nabla}\mathbf F)_{ij} = \frac{\partial F_i(\boldsymbol x)}{\partial x_j}[/itex]<br /> <br /> One way to look at this: The <i>i<sup>th</sup></i> row of the gradient of a vector field \mathbf F(\mathbf x) is the plain old vanilla gradient of the scalar function F_i(\mathbf x).<br /> <br /> One place where the concept is useful is in forming a Taylor expansion of a scalar function. To first order,<br /> <br /> f(\mathbf x_0 + \Delta \mathbf x) \approx f(\mathbf x_0) +&lt;br /&gt; \boldsymol{\nabla}\mathbf f(\mathbf x)|_{\mathbf x=\mathbf x_0}\cdot \Delta \mathbf x<br /> <br /> Higher order expansions require higher order derivatives. The second order expansion requires taking the gradient of the gradient (i.e., taking the gradient of a vector).<br /> <br /> f(\mathbf x_0 + \Delta \mathbf x) \approx&lt;br /&gt; f(\mathbf x_0) +&lt;br /&gt; \sum_i&lt;br /&gt; (\boldsymol{\nabla}(\mathbf f))(\mathbf x_0)_i&lt;br /&gt; \Delta x_i +&lt;br /&gt; \sum_{i,j}&lt;br /&gt; \Delta x_i&lt;br /&gt; (\boldsymol{\nabla}(\boldsymol{\nabla}(\mathbf f)))(\mathbf x_0)_{i,j}&lt;br /&gt; \Delta x_j&lt;br /&gt;<br /> <br /> One application of this is computing the gravity gradient torque induced on a vehicle.
 
Last edited:
I would think that the curl of a vector field would only be a vector (technically a pseudovector, which is really a tensor) in 3 dimensions. in more than 3 its a tensor.

perhaps the tensor you are talking about is simply the true value of the curl. otherwise I have no idea what you are talking about.

I didnt mention grad because he asked for the gradient of a vector not a scalar.
 
I am not talking about curl, which is a pseudovector in three dimensions and generalizes to a N(N-1)/2\times N(N-1)/2 skew-symmetric tensor in N dimensions. I am talking about about the N\times N tensor

(\boldsymol{\nabla}\mathbf F)_{ij} = \frac{\partial F_i(\boldsymbol x)}{\partial x_j}[/itex]<br /> <br /> which I goofed up in my first post (now corrected).<img src="https://www.physicsforums.com/styles/physicsforums/xenforo/smilies/oldschool/redface.gif" class="smilie" loading="lazy" alt=":redface:" title="Red Face :redface:" data-shortname=":redface:" /><br /> <br /> If f(\mathbf x)is a scalar function, then the gradient \boldsymol{\nabla}f = \partial f(\mathbf x)/\partial x_i is a vector field. The &quot;gradient&quot; of this vector is what I was talking about in the second part of my post.<br /> <br /> Aside: Is there a name for the second-order spatial derivative \partial^2 f(\mathbf x)/\partial x_i \partial x_j?
 
D H said:
I am not talking about curl, which is a pseudovector in three dimensions and generalizes to a N(N-1)/2\times N(N-1)/2 skew-symmetric tensor in N dimensions. I am talking about about the N\times N tensor

(\boldsymol{\nabla}\mathbf F)_{ij} = \frac{\partial F_i(\boldsymbol x)}{\partial x_j}[/itex]<br /> <br /> which I goofed up in my first post (now corrected).<img src="https://www.physicsforums.com/styles/physicsforums/xenforo/smilies/oldschool/redface.gif" class="smilie" loading="lazy" alt=":redface:" title="Red Face :redface:" data-shortname=":redface:" /><br /> <br /> If f(\mathbf x)is a scalar function, then the gradient \boldsymol{\nabla}f = \partial f(\mathbf x)/\partial x_i is a vector field. The &quot;gradient&quot; of this vector is what I was talking about in the second part of my post.<br /> <br /> Aside: Is there a name for the second-order spatial derivative \partial^2 f(\mathbf x)/\partial x_i \partial x_j?
<br /> <br /> Thanks a lot, that definitely answers the question! The trick of each row being the gradient of Fi really makes it easy to remember as well.
 
I have a follow-up question to this thread. The gradient of a dot product is given

<br /> \nabla ( A \cdot B) = \underbrace{(B\cdot \nabla)A + (A\cdot \nabla)B}_{\mbox{gradients of vectors?}} + B \times (\nabla\times A) + A \times (\nabla \times B) <br />

All the terms in the equation should be vectors, not second order tensors, which is what gradients of vectors were explained to be earlier in this thread. How then to interpret the first two terms of the right hand side?

Also, the hint I've seen for deriving the identity is to use the "BAC-CAB" identity

<br /> A \times (B \times C) = B (A \cdot C) - C (A \cdot B),<br />

which can be rewritten

<br /> B(A \cdot C) = A \times (B \times C) + (A \cdot B) C.<br />

But using this to expand the gradient of the dot product of two vectors (letting B = \nabla, A = A, \mbox{ and } C = B) appears to yield

<br /> \nabla (A \cdot B) = A \times (\nabla \times B) + (A \cdot \nabla) B,<br />

which is not consistent with the expansion given in textbooks unless

(B \cdot \nabla) A + A \times (\nabla \times B) = 0,

and by symmetry, I don't think that can be true (wouldn't the whole right hand side of the textbook grad of dot product expansion then be zero?).
Please help.

Thanks, Genya
 
musemonkey said:
<br /> B(A \cdot C) = A \times (B \times C) + (A \cdot B) C.<br />

But using this to expand the gradient of the dot product of two vectors (letting B = \nabla, A = A, \mbox{ and } C = B) appears to yield

<br /> \nabla (A \cdot B) = A \times (\nabla \times B) + (A \cdot \nabla) B,<br />


You can't do that. \nabla is not a vector. It's an differential operator that in some cases we can treat like a vector but this is not one of those cases because in doing so you didn't take account of the product rule.

For example, for ordinary vectors \mathbf{A} and \mathbf{B} and a scalar function \psi, the following identity holds:

\mathbf{A} \cdot (\psi \mathbf{B}) = \psi \mathbf{A}\cdot\mathbf{B}

If you were to just plug in \mathbf{A} &quot;=&quot; \nabla now, you would arrive at

\nabla \cdot (\psi \mathbf{B}) = \psi \nabla \cdot \mathbf{B}

which is simply not correct. The correct expression is

\nabla \cdot (\psi \mathbf{B}) = \psi \nabla \cdot \mathbf{B} + \mathbf{B}\cdot \nabla \psi
 
I thought it was OK to substitute \nabla into the BAC-CAB identity because Feynman Lectures vol. II sec. 2-7 contain the following:

A \times (B \times C) = B ( A \cdot C) - (A \cdot B) C
\nabla \times (\nabla \times h) = \nabla (\nabla \cdot h) - (\nabla \cdot \nabla) h

Thank you Mute for the speedy response, but I'm not sure what to make of it. It's still unclear what the terms of form {(B\cdot \nabla)A mean and how the derivation of the gradient of a dot product formula is supposed to be done using the BAC-CAB identity.
 
  • #10
musemonkey said:
<br /> \nabla ( A \cdot B) = \underbrace{(B\cdot \nabla)A + (A\cdot \nabla)B}_{\mbox{gradients of vectors?}} + B \times (\nabla\times A) + A \times (\nabla \times B) <br />

All the terms in the equation should be vectors, not second order tensors, which is what gradients of vectors were explained to be earlier in this thread. How then to interpret the first two terms of the right hand side?

The first term in the above equation in cartesian coordinates is

(B\cdot \nabla)A = \bmatrix<br /> b_x\frac{\partial a_x}{\partial x} +<br /> b_y\frac{\partial a_x}{\partial y} +<br /> b_z\frac{\partial a_x}{\partial z} \\<br /> b_x\frac{\partial a_y}{\partial x} +<br /> b_y\frac{\partial a_y}{\partial y} +<br /> b_z\frac{\partial a_y}{\partial z} \\<br /> b_x\frac{\partial a_z}{\partial x} +<br /> b_y\frac{\partial a_z}{\partial y} +<br /> b_z\frac{\partial a_z}{\partial z}<br /> \endbmatrix

One way to think of B\cdot\nabla is as defining a new operator:

B\cdot\nabla \equiv \sum_j b_j\frac{\partial}{\partial x_j}


Also, the hint I've seen for deriving the identity is to use the "BAC-CAB" identity ...

The path you took is, as Mute noted, invalid. The "BAC-CAB" identity can be used if one uses Feynman's notation:

\nabla(A\cdot B) = \nabla_A(A\cdot B) + \nabla_B(A\cdot B)<br /> = \nabla_A(B\cdot A) + \nabla_B(A\cdot B)

where \nabla_A only operates on A and \nabla_B only operates on B. Then one can "safely" use the BAC-CAB identity as you did:

\aligned<br /> \nabla_A(B\cdot A) &amp;= B \times (\nabla_A \times A) + (B \cdot \nabla_A) A \\<br /> &amp;= B \times (\nabla \times A) + (B \cdot \nabla) A \\<br /> \nabla_B(A\cdot B) &amp;= A \times (\nabla_B \times B) + (A \cdot \nabla_B) B \\<br /> &amp;= A \times (\nabla \times B) + (A \cdot \nabla) B \\<br /> \nabla(A\cdot B) &amp;= \nabla_A(B\cdot A) + \nabla_B(A\cdot B) \\<br /> &amp;= (B \cdot \nabla) A + (A \cdot \nabla) B + B \times (\nabla \times A) + A \times (\nabla \times B)<br /> \endaligned<br />

This, however, is too much sleight-of-hand for me. It happens to work. Your approach happened not to work.
 
  • #11
Thank you DH and Mute! You answered everything, and I appreciate your taking the time to write everything out.
 
  • #12
Hello,

About the second order spatial derivative \partial^2 f(\mathbf x)/\partial x_i \partial x_j which DH wrote above, in tensor notation can it be written as

\nabla(\nabla{f}(\mathbf x))?
 

Similar threads

  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 8 ·
Replies
8
Views
5K
Replies
5
Views
4K
Replies
4
Views
2K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 5 ·
Replies
5
Views
28K