Undergrad How does the Vector Laplacian come about?

  • Thread starter Thread starter yosimba2000
  • Start date Start date
  • Tags Tags
    Laplacian Vector
Click For Summary
SUMMARY

The discussion centers on the derivation and understanding of the Vector Laplacian, defined as the divergence of the gradient of a vector field. The general definition of the Laplacian is given by the equation $$\nabla^{2}\mathbf{T}=\nabla\cdot\nabla\mathbf{T}$$, where ##\mathbf{T}## can be a scalar, vector, or tensor. The participants clarify that the Laplacian of a vector field results in a vector, with specific calculations provided for Cartesian coordinates. The conversation emphasizes the importance of understanding the gradient and divergence in the context of differential forms and tensor calculus.

PREREQUISITES
  • Understanding of vector calculus, specifically divergence and gradient operations.
  • Familiarity with differential forms and tensor calculus.
  • Knowledge of Cartesian coordinates and their application in vector analysis.
  • Basic concepts of Hodge duality and codifferentials in differential geometry.
NEXT STEPS
  • Study the derivation of the Laplacian in different coordinate systems, including polar and spherical coordinates.
  • Explore the application of the Hodge dual in the context of differential forms.
  • Learn about the properties and applications of the divergence theorem in vector calculus.
  • Investigate the relationship between the Laplacian and physical phenomena, such as heat conduction and wave propagation.
USEFUL FOR

Mathematicians, physicists, and engineers interested in advanced vector calculus, particularly those working with differential equations and fields in physics.

yosimba2000
Messages
206
Reaction score
9
So the Laplacian of a scalar is divergence of the gradient of a scalar field, and it comes out to the double derivative of the field in X, Y, and Z.

My book says the Laplacian of a vector field is the double derivative of the X component of the field with respect to X, the double derivative of the Y component with respect to Y, and same with Z.

I'm not sure how this comes about from the definition of the Scalar Laplacian.
 
Physics news on Phys.org
The most general definition of the Laplacian is
$$\nabla^{2}\mathbf{T}=\nabla\cdot\nabla\mathbf{T}$$
where ##\mathbf{T}## is some tensor, which could be a scalar, vector, matrix, etc. The gradient of ##\mathbf{T}## in Cartesian coordinates is
$$\nabla\mathbf{T}=\left[\nabla T_{x}\;\;\nabla T_{y}\;\;\nabla T_{z}\right]$$
where
$$\nabla T_{x}=\begin{bmatrix}\frac{\partial}{\partial x}T_{x} \\ \frac{\partial}{\partial y}T_{x} \\ \frac{\partial}{\partial z}T_{x} \end{bmatrix}$$
and so on for ##T_{y}## and ##T_{z}##. The dot product ##\nabla\cdot\nabla\mathbf{T}## is just
$$\nabla\cdot\nabla\mathbf{T}=\left[\nabla\cdot\nabla T_{x}\;\;\nabla\cdot\nabla T_{y}\;\;\nabla\cdot\nabla T_{z}\right]$$
Thus if ##\mathbf{T}## is a scalar then the Laplacian returns a scalar, if ##\mathbf{T}## is a vector then the Laplacian returns a vector, and so on.
 
How would you do ∇T if T is a vector? I thought gradient only acts on a scalar field.

Just examining derivatives wrt X and components in X-hat direction to make it simpler.
So gradient of scalar takes the derivative of field with respect to X, then tacks on the X-hat direction.

So say vector T is only made of Tx in X direction.

How would you do the gradient of Tx in X-hat direction? It seems like you would do derivative of Tx in the X-hat direction wrt to X, then tack on another X-hat direction, but that doesn't make any sense.
 
I think you should view it within the calculus of differential forms. The Laplacian for any form (and hence also for vector fields, which are in one-to-one correspondence to 1-forms, and for functions, which are ##0##-forms) is per definition given by
$$ \Delta = d \circ \delta + \delta \circ d \, : \Omega^p(M) \rightarrow \Omega^p(M)$$, where ##\delta## is the codifferential given by ##\delta = (-1)^a* d *## where I don't know the exact expression for ##a## at the moment, but it depends on the dimension of the underlying space and the signature of the metric. ##*## is the hodge dual, which maps ##p##-forms to ##(n-p)##-forms, where ##n## is the dimension of the underlying space.
Applied to a function, ##\Delta## gives you the usual scalar Laplacian (if ##M=\mathbb{R}^n##), applied to one-forms and identifying one-forms and vectors gives you the vector-Laplacian (also if ##M = \mathbb{R}^n##).

PS: This is in fact just the same NFuller wrote, but in a different representation, where you do not have to apply some nabla to a tensor. The advantage of the form-approach is, that it is valid for arbitrary (regular) coordinate systems and also for curved spacetimes.
 
Last edited by a moderator:
  • Like
Likes NFuller
The vector Laplacian has to be handled with some care. In Cartesian components you simply have
$$\Delta A^j=\partial_k \partial^k A^j=\delta^{ik} \partial_i \partial_k A^j.$$
It's not as easy in general coordinates (even not in general orthogonal curvilinear coordinates). To define it unambiguously in a covariant way we note that in Cartesian coordinates
$$\vec{\nabla} \times (\vec{\nabla} \times \vec{A})=\vec{\nabla} (\vec{\nabla} \cdot \vec{A})-\Delta \vec{A},$$
and thus it's safe to define generally
$$\Delta \vec{A}=\vec{\nabla} (\vec{\nabla} \cdot \vec{A}) - \vec{\nabla} \times (\vec{\nabla} \times \vec{A}),$$
where the right-hand side has a well-defined covariant meaning.
 
  • Like
Likes Metmann and NFuller
Sorry Metmann, that is way beyond me.
 
yosimba2000 said:
How would you do ∇T if T is a vector? I thought gradient only acts on a scalar field.
Let's assume ##\mathbf{T}## has the form ##\mathbf{T}=T_{x}\hat{x}+T_{y}\hat{y}+T_{z}\hat{z}##. The gradient is
$$\nabla\mathbf{T}=\begin{bmatrix}\frac{\partial}{\partial x}T_{x} & \frac{\partial}{\partial x}T_{y} & \frac{\partial}{\partial x}T_{z} \\ \frac{\partial}{\partial y}T_{x} & \frac{\partial}{\partial y}T_{y} & \frac{\partial}{\partial y}T_{z} \\ \frac{\partial}{\partial z}T_{x} & \frac{\partial}{\partial z}T_{y} & \frac{\partial}{\partial z}T_{z} \end{bmatrix}$$
So just as the gradient of a scalar produces a vector, the gradient of a vector produces a matrix. Then the Laplacian is
$$\nabla\cdot\nabla\mathbf{T}=\begin{bmatrix}\frac{\partial^{2}}{\partial x^{2}}T_{x} + \frac{\partial^{2}}{\partial y^{2}}T_{x} + \frac{\partial^{2}}{\partial z^{2}}T_{x} \\ \frac{\partial^{2}}{\partial x^{2}}T_{y} + \frac{\partial^{2}}{\partial y^{2}}T_{y} + \frac{\partial^{2}}{\partial z^{2}}T_{y} \\ \frac{\partial^{2}}{\partial x^{2}}T_{z} + \frac{\partial^{2}}{\partial y^{2}}T_{z} + \frac{\partial^{2}}{\partial z^{2}}T_{z} \end{bmatrix}$$
Which is just a vector, as expected. As stated by Metmann, this is only valid in Cartesian coordinates and there are more general ways to describe this. However, this is probably the simplest to understand.
 
  • Like
Likes Metmann
Thanks. How do we know this, though? Is it just definition?
 
That's just how it's defined. Although this is technically an abuse of notation, I like to think of the gradient as a vector
$$\nabla=\begin{bmatrix}\frac{\partial}{\partial x} \\ \frac{\partial}{\partial y} \\ \frac{\partial}{\partial z} \end{bmatrix}$$
such that the gradient of a vector is the outer product
$$\nabla(\mathbf{T}^{T})=\begin{bmatrix}\frac{\partial}{\partial x} \\ \frac{\partial}{\partial y} \\ \frac{\partial}{\partial z} \end{bmatrix}\begin{bmatrix}T_{x} & T_{y} & T_{z}\end{bmatrix}=\begin{bmatrix}\frac{\partial}{\partial x}T_{x} & \frac{\partial}{\partial x}T_{y} & \frac{\partial}{\partial x}T_{z} \\ \frac{\partial}{\partial y}T_{x} & \frac{\partial}{\partial y}T_{y} & \frac{\partial}{\partial y}T_{z} \\ \frac{\partial}{\partial z}T_{x} & \frac{\partial}{\partial z}T_{y} & \frac{\partial}{\partial z}T_{z} \end{bmatrix}$$
 
  • #10
I see, so the definition of the Laplacian of a Scalar is different than the definition of the Laplacian of a Vector, but they use the same notation?
 
  • #11
yosimba2000 said:
I see, so the definition of the Laplacian of a Scalar is different than the definition of the Laplacian of a Vector, but they use the same notation?
The definition also works with a scalar variable such as ##\mathbf{T}=\phi##
$$\nabla\phi=\begin{bmatrix}\frac{\partial}{\partial x} \\ \frac{\partial}{\partial y} \\ \frac{\partial}{\partial z} \end{bmatrix}\begin{bmatrix}\phi\end{bmatrix}=\begin{bmatrix}\frac{\partial\phi}{\partial x} \\ \frac{\partial\phi}{\partial y} \\ \frac{\partial\phi}{\partial z} \end{bmatrix}$$
This produces a vector where the components are just the spatial derivatives of the scalar function ##\phi##, just as one would expect.
 
  • #12
yosimba2000 said:
I see, so the definition of the Laplacian of a Scalar is different than the definition of the Laplacian of a Vector, but they use the same notation?

The definition is in fact the same, the notation of usual vector analysis is just not good enough to display this fact ;)

I admit that my post might have been a bit over top, but consider the following (all valid only for ##\mathbb{R}^n##): You may know the differential of a function: ##df = \sum_i \frac{\partial f}{\partial x^i}dx^i##. This is nothing but the gradient on scalar fields, when identifiying ## a_i dx^i \leftrightarrow (a^0,\ldots,a^n)^T ## with ##a_i = a^i##.

Now, if you have a vectorfield ##(X^0,\ldots,X^n) ## and identify it with ## X= \sum_i X_i dx^i ## with ##X^i=X_i##, than ## dX = \sum_{i,j} \frac{\partial X^i}{\partial x^j}dx^j\wedge dx^i ##, which is nothing but a generalization of the differential on functions. The "inverse" of ##d## is ##\delta## given for vectors as ##\delta (\sum_i X_i dx^i) = \sum_i \frac{\partial X_i}{\partial x^i}##, which is nothing but the divergence of ##\vec{X}##. Then ## \Delta f = (d + \delta)^2f =\delta(df) + d(\delta f) = \sum_i \frac{\partial^2 f}{\partial (x^i)^2} ##, where ##\delta f = 0##. There is also a formula for the codifferential for objects of the type ##a_{ij}dx^j \wedge dx^i## (derived from the general definition given in my previous post). Then you can define the Laplacian for vectors exactly the same way as for functions, giving you ##\Delta X = \sum_{ij} \frac{\partial^2 X_i}{\partial (x^j)^2}dx^i ## which is nothing but the Vector-Laplacian, i.e. scalar Laplacian in each component.

In fact you get: ##d(\delta(X)) = \nabla(\nabla \cdot \vec{X}) = grad(div(\vec{x}))## and ##\delta (d(X)) = -\nabla \times (\nabla \times \vec{X}) =- rot(rot(\vec{X}))##, which is what vanhees71 gave.

The same way (with the same definition) you get a Laplacian for all tensors.
 
Last edited by a moderator:
  • #13
NFuller said:
That's just how it's defined. Although this is technically an abuse of notation, I like to think of the gradient as a vector
$$\nabla=\begin{bmatrix}\frac{\partial}{\partial x} \\ \frac{\partial}{\partial y} \\ \frac{\partial}{\partial z} \end{bmatrix}$$
such that the gradient of a vector is the outer product
$$\nabla(\mathbf{T}^{T})=\begin{bmatrix}\frac{\partial}{\partial x} \\ \frac{\partial}{\partial y} \\ \frac{\partial}{\partial z} \end{bmatrix}\begin{bmatrix}T_{x} & T_{y} & T_{z}\end{bmatrix}=\begin{bmatrix}\frac{\partial}{\partial x}T_{x} & \frac{\partial}{\partial x}T_{y} & \frac{\partial}{\partial x}T_{z} \\ \frac{\partial}{\partial y}T_{x} & \frac{\partial}{\partial y}T_{y} & \frac{\partial}{\partial y}T_{z} \\ \frac{\partial}{\partial z}T_{x} & \frac{\partial}{\partial z}T_{y} & \frac{\partial}{\partial z}T_{z} \end{bmatrix}$$
I always prefer the index calculus (aka Ricci calculus), from which it becomes clear how the various components "naturally" transform under rotations (for 3D "classical" vector analysis in Euclidean space). The gradient is a co-vector, i.e., it has lower indices ##\nabla_j=\partial_j## wrt. to a Cartesian (co-)basis. Only because in Euclidean space wrt. a orthonormal (Cartesian) basis the metric reads ##g_{ij}=\delta_{ij}## the gradient looks like vector components (and this holds true of course only wrt. Cartesian bases).
 
  • Like
Likes NFuller

Similar threads

  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 6 ·
Replies
6
Views
3K
  • · Replies 8 ·
Replies
8
Views
3K
  • · Replies 9 ·
Replies
9
Views
4K
Replies
1
Views
3K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 20 ·
Replies
20
Views
4K