I Why are the equations for dt/du and Dt[a]/Du equal in the geodesic equation?

Apashanka Das
In the geodesic equation why is dt/du=λ(u)t ,where t is the tangent vector along the curve
and why Dt[a]/Du=λ(u)dx[a]/du equated same,as given in hobson
 
Physics news on Phys.org
There is no relevant coordinate independent meaning with dt/du unless what you mean is ##Dt/Du = \nabla_{t}t##. Also, the relevant interpretation of ##dx/du## is ##dx/du = t## so the equations are the same.
 
Mr. Orodruin can you please explain the first elaborately ,its very much needed
 
Last edited by a moderator:
Apashanka Das said:
Mr. Orodruin can you please explain the first elaborately ,its very much needed

Please note that you originally marked this thread as "A", which indicates a graduate level knowledge of the subject matter. I have changed it to "I", which indicates undergraduate level, because your question can't really be answered at the "B" level. But even at the undergraduate level, the expressions @Orodruin is using should be fairly common knowledge; and in any case, explaining them "elaborately" as you request would go far beyond the scope of a PF discussion. What background do you have in this area?
 
Ok please give some precise explanation
 
Apashanka Das said:
please give some precise explanation

We need to know your background in this area. What textbooks have you studied?
 
Nothing to tell just a post graduate student at st xaviers college ,the question strikes me that's why I m asking ,I have read Hobson
 
yes it is this book,
pg no-76,section-3.17,equation -3.44,3.45,3.46,3.47
 
  • #10
Apashanka Das said:
Nothing to tell just a post graduate student at st xaviers college ,the question strikes me that's why I m asking ,I have read Hobson
I am sorry, but it is impossible to decipher what you are actually asking about. Can you reformulate the question in such a way that it becomes clear what you want to know? Please explain what it is in the equations you listed that gives you problems. If you are reading these things, I assume that you are at least relatively familiar with calculus on manifolds?

Also, nobody is going to "explain something elaborately" as in writing you a textbook segment. It is up to you to write a question where it is clear what precisely you have problems with. If you do not, you are very unlikely to get a satisfactory answer.
 
  • #11
I can see something confusing in the notation used in that book. It appears little d is used to specify the abstract notion of absolute derivative along a curve, while big D is used for the coordinate representation. Also little d is used for ordinary derivative on coordinate basis. Thus, the first equation after 3.41 is just a translation from abstract to coordinate basis. Then, 3.45 follows trivially. The rest is restricting consideration to affine parameters.

[edit: above, substitute 3.44 for 3.41]
 
Last edited:
  • #12
strangerep said:
Actually, Hobson uses big D for the intrinsic (or "absolute") derivative along a curve. [See eq(3.37) in section 3.14.]

@Apashanka Das : I guess the 1st part of your question is about why the notion of "geodesic" is defined by the requirement that the intrinsic derivative of the tangent vector ##{\mathbf t}## along the curve should be a multiple of itself, i.e., why
$$\frac{d{\mathbf t}}{du} ~=~ \lambda(u) {\mathbf t} ~,$$where ##u## is an arbitrary parameter. Then, since this ordinary derivative is not covariant, Hobson switches to the intrinsic derivative ##D/Du## which involves the connection ##\Gamma##, ending up at eq(3.45).
Since I don't have the book, and some pages (including the one you reference) are not accessible to me via the link, I have to intuit the meaning. I agree with you and state as much for the the interpretation of D. However, several usages of d applied to a vector without a coordinate basis are all consistent (IMO) with this being an abstract absolute derivative, as distinct from d applied to components which is always an ordinary derivative. The discussion from 3.39 to 3.41 seems consistent with my interpretation, but I cannot see pages where certain quantities are formally defined. In any case, I find the notation and ambiguity about what derivative is meant pointlessly confusing. Note also the discussion of covariant derivatives on page 69. To me, this also supports my interpretation that Hobson uses ordinary derivative symbols applied to an abstract vector as equivalent to the covariant (or absolute) derivative. See especially just before 3.32 on this page.
 
  • Like
Likes strangerep
  • #13
Ohh,that means here Hobson has defined 'geodesic' along which the the parallel transport of the tangent vector is takes place considering direction ,and 'affinely parametrised curve' along which parallel transport of the tangent vector considering direction and magnitude both,I may be wrong,need correction
 
  • #14
PAllen said:
Since I don't have the book, and some pages (including the one you reference) are not accessible to me via the link, I have to intuit the meaning. I agree with you and state as much for the the interpretation of D. However, several usages of d applied to a vector without a coordinate basis are all consistent (IMO) with this being an abstract absolute derivative, as distinct from d applied to components which is always an ordinary derivative.
Aaargh! You are right -- this is clear from p72. (I tried to insert corrections in my earlier post but it was too messy, so I deleted the entire post to avoid confusing anyone else.)

When Hobson uses ##d## for a vector it's the absolute derivative. But when switches to components of the vector, he must use ##D## for the absolute derivative. I agree this is a confusing notation -- since Hobson was successful in confusing me. :sorry:
 
Last edited:
  • #15
##\newcommand{\bvec}[1]{\boldsymbol{#1}}##
Well, the point is that you have to distinguish between vectors and vector components. All physicists (including myself) are very sloppy not doing so all the time, and unfortunately we don't even take the time to explain it to the students in Physics 101, when we should.

Let's write ##\bvec{V}## for a vector (bold italics), which is a function of some scalar parameter (e.g., a world-line parameter for the motion of a particle in spacetime). Then let ##\bvec{b}_{\mu}## be an arbitrary basis. Then of course
$$\bvec{W}(\lambda)=\frac{\mathrm{d}}{\mathrm{d} \lambda} \bvec{V}(\lambda)$$
is still a vector, and no reference has been made to the basis and no components occurred yet. So let's see, how the components of ##\bvec{W}## with respect to the given basis are related to the components of ##\bvec{V}## wrt. the same basis:
$$\bvec{W}=\frac{\mathrm{d}}{\mathrm{d} \lambda} (V^{\mu} \bvec{b}_{\mu})=\bvec{b}_{\mu} \frac{\mathrm{d} V^{\mu}}{\mathrm{d} \lambda} + V^{\mu} \frac{\mathrm{d} \bvec{b}_{\mu}}{\mathrm{d} \lambda}=\dot{V}^{\mu} \bvec{b}_{\mu} + V^{\mu} \dot{\bvec{b}}_{\mu}.$$
Since now ##\dot{\bvec{b}}_{\mu}## is again a vector, I can write it in terms of the basis, defining the connection coefficients
$$\dot{\bvec{b}}_{\mu}={\gamma^{\nu}}_{\mu} \bvec{b}_{\nu},$$
and finally we get
$$\bvec{W}=\dot{\bvec{V}}=\bvec{b}_{\nu} (\dot{V}^{\nu} + {\gamma^{\nu}}_{\mu} V^{\mu}).$$
For convenience, when calculating with components in the Ricci calculus one defines some kind of derivative operator,
$$W^{\nu} = \mathrm{D}_{\lambda} V^{\nu} = \dot{V}^{\nu} + {\gamma^{\nu}}_{\mu} V^{\mu}.$$
Of course you can generalize this principle in many ways, e.g., to introduce the Christoffel symbols in affine bases, as usually used in introductory GR.

Then you have space-time coordinates ##q^{\mu}## (no vector components!), and you can define four basis vectors as tangents to the coordinate lines, i.e., curves defined by only varying ##q^{\mu}## and keeping the other three space-time coordinates fixed. The corresponding basis vectors are intuitively written as
$$\bvec{b}_{\mu}=\partial_{\mu}$$
Now if you have a vector field ##\bvec{V}=V^{\mu} \bvec{b}_{\mu}##, you can define a tensor field in the usual way by
$$\bvec{W}=\mathrm{d} q^{\mu} \otimes \partial_{\mu} \bvec{V}.$$
Here ##\mathrm{d} q^{\mu}## defines the dual basis of dual space to the holonomous basis ##\partial_{\mu}##.

This leads you to define a "covariant derivative" as components of this tensor:
$$\bvec{W}=\mathrm{d} q^{\mu} \otimes \partial_{\mu} (V^{\nu} \bvec{b}_{\nu}).$$
Then you define the affine connection
$$\mathrm{d} q^{\mu} \otimes \partial_{\mu} \bvec{b}_{\nu}=\mathrm{d} q^{\mu} \otimes {\Gamma^{\rho}}_{\mu \nu} \bvec{b}_{\rho}={\Gamma^{\rho}}_{\mu \nu} \partial_{\rho}\mathrm{d} q^{\mu} \otimes \partial_{\rho} $$
to get
$$\bvec{W}=:(\nabla_{\mu} V^{\nu}) \mathrm{d} q^{\mu} \otimes \partial_{\nu} = (\partial_{\mu} V^{\nu} + {\Gamma^{\nu}}_{\mu \rho} V^{\rho})\mathrm{d} q^{\mu} \otimes \partial_{\nu}.$$
This defines the "covariant derivative" of vector components, leading to a 2nd-rank mixed tensor components
$${W_{\mu}}^{\nu} = \nabla_{\mu} V^{\nu}=\partial_{\mu} V^{\nu} + {\Gamma^{\nu}}_{\mu \rho} V^{\rho}.$$
The "connection terms" always occur due to taking derivatives of the basis vectors and writing them again in terms of the basis.
 
  • #16
Ok, @vanhees71, that justifies his notation, however, it still seems confusing and uncommon to me.
 
  • #17
Hm, but why is the (very common) notation of covariant derivatives for components confusing? It's used for decades in differential geometry, general relativity, and gauge-field theory...
 
  • #18
vanhees71 said:
Hm, but why is the (very common) notation of covariant derivatives for components confusing? It's used for decades in differential geometry, general relativity, and gauge-field theory...
No, what is confusing is the use of ordinary derivative symbols for covariant or absolute derivatives when components are not used. I have literally never seen this before. Your derivation provides some motivation, however all books I've seen use some symbol other than ordinary derivative symbols for abstract coordinate free equations.
 
  • Like
Likes strangerep
  • #19
I don't see the motivation in vanhees71's post. For me it is equally unusual to see the ordinary derivative symbol there in ##\frac{d}{d\lambda}V(\lambda)##.
 
  • #20
But the point is that if ##\boldsymbol{V}(\lambda)## is a vector and ##\lambda## is a scalar, then ##\mathrm{d} \boldsymbol{V}/\mathrm{d} \lambda## is a vector. So why should you introduce another symbol for such a derivative than the usual one? My motivation was to make clear, why you don't need a special symbol for the vectors themselves and why you need one for the components.
 
  • Like
Likes strangerep
  • #21
vanhees71 said:
But the point is that if ##\boldsymbol{V}(\lambda)## is a vector and ##\lambda## is a scalar, then ##\mathrm{d} \boldsymbol{V}/\mathrm{d} \lambda## is a vector. So why should you introduce another symbol for such a derivative than the usual one? My motivation was to make clear, why you don't need a special symbol for the vectors themselves and why you need one for the components.
But to define ##\mathrm{d} \boldsymbol{V}/\mathrm{d} \lambda## you need additional structure. Just the differentiable manifold is not enough you need a connection. That's why it is more common to denote it ##\mathrm{D} \boldsymbol{V}/\mathrm{d} \lambda##.
 
  • #22
vanhees71 said:
But the point is that if ##\boldsymbol{V}(\lambda)## is a vector and ##\lambda## is a scalar, then ##\mathrm{d} \boldsymbol{V}/\mathrm{d} \lambda## is a vector. So why should you introduce another symbol for such a derivative than the usual one? My motivation was to make clear, why you don't need a special symbol for the vectors themselves and why you need one for the components.
And now, with more thought, I question your whole justification. Independent of any coordinate basis, there is no notion of vector derivative without a connection. Otherwise there is no way to compare or subract vectors at different points, even along a given curve, because the vectors are in different tangent spaces - and the connection is what maps between tangent spaces of nearby points. Given this need for connection to even define a vector derivative, it is confusing to use an ordinary derivative symbol for this.

[edit: I see I cross posted with @martinbn ]
 
  • #23
PAllen said:
[...] Independent of any coordinate basis, there is no notion of vector derivative without a connection. Otherwise there is no way to compare or subtract vectors at different points, even along a given curve, because the vectors are in different tangent spaces - and the connection is what maps between tangent spaces of nearby points. [...]
That makes me wonder: why then are we allowed to subtract scalars at different points?

The vector tangent space is (isomorphic to) ##{\mathbb R}^4## whereas the "scalar tangent space" is just ##{\mathbb R}##.

I'm guessing the answer has something to do with whether a "natural" mapping between tangent spaces exists. But I don't see a "natural" mapping between 2 copies of ##{\mathbb R}## since arbitrary translations and scale changes are possible.

:oldconfused:
 
  • #24
strangerep said:
why then are we allowed to subtract scalars at different points?

Are we? Can you give an example?
 
  • #25
PeterDonis said:
Can you give an example?
Well, I was thinking of the case of a scalar-valued function ##f(\lambda)## defined on a curve ##\gamma^i(\lambda)##. Calculating its derivative involves ##f(\lambda+\epsilon) - f(\lambda)##.
 
  • #26
The scalars have values real numbers. We can subtract real numbers. The vectors belong to different vector spaces. They may be isomorphic, but there are many different isomorphisms. How do you pick an isomorphism? How do you subtract ##v\in T_pM## and ##w\in T_qM##? On the other hand ##\alpha\in\mathbb R## and ##\beta\in\mathbb R## can be subtracted.
 
  • #27
martinbn said:
The scalars have values real numbers. We can subtract real numbers. The vectors belong to different vector spaces. They may be isomorphic, but there are many different isomorphisms. How do you pick an isomorphism? How do you subtract ##v\in T_pM## and ##w\in T_qM##? On the other hand ##\alpha\in\mathbb R## and ##\beta\in\mathbb R## can be subtracted.
Yes, that's similar to what I was thinking. But... there's also more than 1 possible isomorphism between different copies of ##\mathbb R##, unless you invoke some extra criteria. E.g., 0 matches 0, and the mapping is isometric (hence no scale changes allowed). :oldconfused:
 
  • #28
Scalars are defined as maps from the manifold to the real numbers i.e. ##\alpha, \beta : M\rightarrow \mathbb R##. The target set is the same, so we can subtract their values. Vectors are not maps to ##\mathbb R^n##. If they were we could subtract them. But they are derivations of functions at the point (for example). They form a vector space, which is isomorphic to ##\mathbb R^n##.
 
  • #29
PeterDonis said:
Are we? Can you give an example?

Well, one way of defining tangent vectors is via differences of scalar fields at different locations:

If you have a parametrized path \gamma(s), the corresponding tangent vector \mathbf{V}(s) is that operator which given a scalar field \Phi(\mathcal{P}) returns \frac{d}{ds} \Phi(\gamma(s)).
 
  • Like
Likes vanhees71
  • #30
PAllen said:
And now, with more thought, I question your whole justification. Independent of any coordinate basis, there is no notion of vector derivative without a connection. Otherwise there is no way to compare or subract vectors at different points, even along a given curve, because the vectors are in different tangent spaces - and the connection is what maps between tangent spaces of nearby points. Given this need for connection to even define a vector derivative, it is confusing to use an ordinary derivative symbol for this.

[edit: I see I cross posted with @martinbn ]

Okay, but in the context of talking about a smooth manifold with connection (which is the case for GR), the notation \frac{d\mathbf{V(s)}}{ds} is unambiguous.
 
  • #31
stevendaryl said:
Okay, but in the context of talking about a smooth manifold with connection (which is the case for GR), the notation \frac{d\mathbf{V(s)}}{ds} is unambiguous.
Unambiguous, yes , if defined by the author. Confusing and nonstandard, also yes. Also, my argument was in response to an argument that connections were only relevant in a coordinate basis - which is not correct.
 
  • #32
The entire vector analysis on differentiable manifolds do not need additional structures a priori. As long as you restrict yourself to alternating differential forms you don't need covariant derivatives.
PAllen said:
And now, with more thought, I question your whole justification. Independent of any coordinate basis, there is no notion of vector derivative without a connection. Otherwise there is no way to compare or subract vectors at different points, even along a given curve, because the vectors are in different tangent spaces - and the connection is what maps between tangent spaces of nearby points. Given this need for connection to even define a vector derivative, it is confusing to use an ordinary derivative symbol for this.

[edit: I see I cross posted with @martinbn ]
Why then don't you need a connection when dealing with plain differentiable manifolds without any additional structures? You can go pretty far, introducing tangent and cotangent spaces and introducing differential forms. For this I don't need a connection, let alone a fundamental form (i.e., metric of pseudometric).

Also see the remark on Wikipedia:

https://en.wikipedia.org/wiki/Covariant_derivative#Remarks
 
Last edited:
  • #33
If you want to differentiate objects other than functions in the direction of a vector you need a connection.
 
  • #34
Then, why don't I need a connection to define (alternating) differential forms?
 
  • #35
vanhees71 said:
Then, why don't I need a connection to define (alternating) differential forms?
Why would you?
 
  • #36
strangerep said:
I was thinking of the case of a scalar-valued function ##f(\lambda)## defined on a curve ##\gamma^i(\lambda)##. Calculating its derivative involves ##f(\lambda+\epsilon) - f(\lambda)##.

But the curve plays no role in this whatsoever; ##\lambda## is just the independent variable for the function ##f(\lambda)## and the derivative makes no use of the curve ##\gamma^i(\lambda)## at all. So you're not subtracting scalars at "different points" in some manifold, or in different tangent spaces; you're just subtracting values of a function at different values of its independent variable. Or, to put it another way, there is no need to define a "connection" (along a curve or by any other means) in order to take the ordinary derivative of a function, which is all you are doing here. But that's not the same thing as trying to compare scalars at different events in spacetime.
 
  • #37
PeterDonis said:
But the curve plays no role in this whatsoever; ##\lambda## is just the independent variable for the function ##f(\lambda)## and the derivative makes no use of the curve ##\gamma^i(\lambda)## at all. So you're not subtracting scalars at "different points" in some manifold, or in different tangent spaces; you're just subtracting values of a function at different values of its independent variable. Or, to put it another way, there is no need to define a "connection" (along a curve or by any other means) in order to take the ordinary derivative of a function, which is all you are doing here. But that's not the same thing as trying to compare scalars at different events in spacetime.
Suppose the scalar is function on the manifold. Then you can take its derivative along along a curve, and this does depend on the curve, and it subtracts scalar values at different points of the manifold (along the curve). However, none of this need a connection for the reason @martinbn pointed out several posts ago: the scalar is a mapping from the manifold to reals; there is no notion of a different set of reals for each manifold point.
 
  • #38
PeterDonis said:
[...] So you're not subtracting scalars at "different points" in some manifold, or in different tangent spaces; you're just subtracting values of a function at different values of its independent variable. [...]
My response would have been similar to PAllen's in post #37.

PAllen said:
there is no notion of a different set of reals for each manifold point.
But isn't that merely because we have tacitly identified the different copies of ##\mathbb R## at each manifold point, and thus treat them all as one and the same?

BTW, @Apashanka Das : I'm assuming that your original question has been answered sufficiently, (hence a bunch of SAs debating like this is not an impolite hijack of your thread). If your original question remains insufficiently answered, please say so. :oldbiggrin:
 
Last edited:
  • #39
PAllen said:
Suppose the scalar is function on the manifold. Then you can take its derivative along along a curve, and this does depend on the curve, and it subtracts scalar values at different points of the manifold (along the curve).

But if you say that the derivative involves ##f(\lambda + \epsilon) - f(\lambda)## (with all the appropriate language about limits, etc.), you aren't doing that. You're just taking the ordinary derivative of a function with one variable.

To do what you're talking about, you have to start with the parameterization of the curve, and use it in defining the derivative. And if you do that, since, as you say, the result will depend on the curve you choose, it is the curve that is providing the connection between scalars at different points on the manifold.
 
  • #40
PeterDonis said:
To do what you're talking about, you have to start with the parameterization of the curve, and use it in defining the derivative. And if you do that, since, as you say, the result will depend on the curve you choose, it is the curve that is providing the connection between scalars at different points on the manifold.
This is what I am saying. The point is exactly that different curves through a point lead to different scalar derivatives at that point, and in all cases, the limit defining the derivative involves subtracting scalar function values at different points in the manifold. Again, this is needs no extra structure beyond a differential manifold because the scalar is a mapping between the manifold and R.
 
  • #41
strangerep said:
But isn't that merely because we have tacitly identified the different copies of ##\mathbb R## at each manifold point, and thus treat them all as one and the same?

BTW, @Apashanka Das : I'm assuming that your original question has been answered sufficiently, (hence a bunch of SAs debating like this is not an impolite hijack of your thread). If your original question remains insufficiently answered, please say so. :oldbiggrin:
No , it is because a scalar is defined as mapping between the manifold and R.

You could, for example, define a mapping between the manifold and some fixed vector space isomorphic to Rn. Then, derivatives of this vector valued function along curves, or in different directions on the manifold (e.g. divergences) would not need a connection. But this vector function would be invariant, not contravariant; it would be totally unaffected by coordinate transforms on the manifold. I am not aware of any physical theory making use of such a construct, but it is perfectly well defined mathematically.

However, in GR, we normally want to say that a vector field, or vectors along a curve, are vectors from the tangent space at each point. Since this is a different vector space at each point, this is what makes it impossible to subtract vectors for different points without a connection, and thus a connection is required for derivative of such a vector function.
 
  • #42
##\newcommand{\dd}{\mathrm{d}}## ##\newcommand{\vv}[2]{\begin{pmatrix} #1 \\ #2 \end{pmatrix}}## ##\newcommand{\vvv}[3]{\begin{pmatrix}
#1 \\ #2 \\3 \end{pmatrix}}## ##\newcommand{\bvec}[1]{\boldsymbol{#1}}##Well after reading to last few postings, I think, @martinbn is right after all. The point indeed is that we don't just consider some vector-valued function of a scalar parmater ##\lambda## as I did in one of my previous postings but a vector field ##\bvec{V}:M \rightarrow \mathbb{R}## (or ##\mathbb{C}## if needed), and we want a "canonical" (i.e., basis independent) derivative of ##\bvec{V}## along a curve or in direction of the tangent vectors along the curve. For this you have various options like the Lie derivative, for which you don't need an affine connection but also the covariant derivative, where you need some rule to infinitesimally move the vector from one tangent space at point ##P## to some point ##P'## along the given direction, and how you do this defines the connection.

In my derivation for the derivative of the vector field along a curve that's hidden in the fact that I've not only used the dependence of ##\bvec{V}## on the parameter ##\lambda##, which parametrizes the curve, but the dependence of the complete basis ##\bvec{b}_{\mu}(\lambda)## along the curve, and I can choose this dependence quite arbitrarily, defining different affine connections along the curve.

Of course, as I said before, you can do a lot of vector analysis on a plain differentiable manifold without additional structures like an affine connection (making the manifold to a affine manifold) or fundamental bilinear form (making the manifold to a Riemann or pseudo-Riemann space when choosing the torsion free affine connection compatible with this (pseudo-)metric). Then you are however more or less restricted to alternating forms and their derivatives.
 
Back
Top