MUST the variation of the action be zero?

  • Thread starter Thread starter friend
  • Start date Start date
  • Tags Tags
    Variation Zero
friend
Messages
1,448
Reaction score
9
Feynman's path integral is:

\[<br /> \int {Dx\,e^{{\textstyle{i \over \hbar }}\int {L(x,\dot x,t)dt} } } <br /> \]

where the Action is:

\[<br /> \int {L(x,\dot x,t)dt} <br /> \]

and the Lagrangian is: \[<br /> {L(x,\dot x,t)}<br /> \]<br />

Now we are told that as we functionally integrate the path integral in the region of far flung paths, contributions of the integral cancel out with other parts of other far flung paths, leaving only the classical path that contributes most to the final result. This classical part is also obtained by setting the functional derivative of the Action to zero and getting the Euler-Lagrange equations of motion.

My question is if the functional derivative of the Action integral does not have a zero for any path, no classical path, then is it still possible to evaluate the path integral at all? Or will everything cancel out? Thanks.
 
Physics news on Phys.org
if you don't have a classical path..then it means you're still in the quantum domain. hence apply QM or QFT tools..
 
friend said:
My question is if the functional derivative of the Action integral does not have a zero for any path, no classical path, then is it still possible to evaluate the path integral at all? Or will everything cancel out? Thanks.

The path integral is equated to <x1|x2> which is elsewhere also equated to the dirac delta function D(x1-x2). So the path integral itself equates to a Dirac delta function:

\[<br /> &lt; x|x_0 &gt; \, = \,\,\delta (x - x_0 )\, = \int {{\rm{Dx}}\,{\rm{e}}^{\frac{{\rm{i}}}{\hbar }\int {{\rm{L(x,\dot x,t)dt}}} } } <br /> \]

Now it occurs to me that the variation of the action in the exponential might be required to be zero if the functional derivative of the whole path integral were required to be zero. Taking the functional derivative of the whole path integral may result in a factor that is the functional derivative of the action integral in the exponential. So if the functional derivative of the whole thing were required to be zero, then this would require the action integral to have a variation of zero. So the question for me is what is the functional derivative of a Dirac delta function. Or what is,

\[<br /> \frac{\delta }{{\delta (x)}}\,\delta (x - x_0 )<br /> \]

My first instinct would be that it is 1. But when this is put in QM terms:

\[<br /> \frac{\delta }{{\delta (x)}}\, &lt; x|x_0 &gt; <br /> \]

Then perhaps it is zero if the states are eigenstates with discrete spectrum, for then there is no small variation of quantum states. Anyone have any idea what the meaning of this variation of an inner product would mean, physically or mathematically? Does this sound like a reasonable physical requirement? Thanks.
 
Last edited:
friend said:
The path integral is equated to <x1|x2> which is elsewhere also equated to the dirac delta function D(x1-x2). So the path integral itself equates to a Dirac delta function:

\[<br /> &lt; x|x_0 &gt; \, = \,\,\delta (x - x_0 )\, = \int {{\rm{Dx}}\,{\rm{e}}^{\frac{{\rm{i}}}{\hbar }\int {{\rm{L(x,\dot x,t)dt}}} } } <br /> \]

Now it occurs to me that the variation of the action in the exponential might be required to be zero if the functional derivative of the whole path integral were required to be zero. Taking the functional derivative of the whole path integral may result in a factor that is the functional derivative of the action integral in the exponential. So if the functional derivative of the whole thing were required to be zero, then this would require the action integral to have a variation of zero.

OK, how does this sound? If we integrate the path integral one more time, it should equal 1 since it's a Dirac delta function. Then the functional derivative of that should be zero, since the functional derivative of a constant is zero. This much is necessarily True. But since the path integral is already an infinite number of integrations, I'm not sure that integrating it one more time would really affect its formulation. And then taking the functional derivative of that would probably create a factor consisting of the variation of the action. This would be required to be zero since nothing else could be set to zero.

The path integral written out more fully is:

\[<br /> \mathop {\lim }\limits_{n \to \infty } \,\,\int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {...\int_{ - \infty }^{ + \infty } {\prod\nolimits_{j = 1}^n {dx_j \,\left( {{\textstyle{m \over {2\pi i\hbar \varepsilon }}}} \right)^{{\raise0.5ex\hbox{$\scriptstyle 1$}<br /> \kern-0.1em/\kern-0.15em<br /> \lower0.25ex\hbox{$\scriptstyle 2$}}} \,\exp \left( {{\textstyle{i \over \hbar }}\,S[x\left( t \right)]} \right)} } } } <br /> \]<br />

So if we integrate one more time the product would go from \[<br /> \prod\nolimits_{j = 1}^n {} <br /> \]<br /> to \[<br /> \prod\nolimits_{j = 1}^{n + 1} {} <br /> \]<br />

But what real difference would this make since they both go to infinity anyway?

In any event, I have seen where the functional derivative commutes with integration. So the functional derivative of the path integral will commute with all the integrals (all infinity of them) and pass inside to the exponential. So functional differentiation commutes with functional integration, right?

Then I assume:

\[<br /> \delta \left( {\exp \left( {\frac{{\rm{i}}}{\hbar }{\rm{S[x}}\left( {\rm{t}} \right){\rm{]}}} \right)} \right){\rm{ = }}\exp \left( {\frac{{\rm{i}}}{\hbar }{\rm{S}}\left[ {{\rm{x}}\left( {\rm{t}} \right)} \right]} \right){\rm{\cdot}}\left( {\frac{{\rm{i}}}{\hbar }\delta {\rm{S}}\left[ {{\rm{x}}\left( {\rm{t}} \right)} \right]} \right)<br /> \]<br />

And also that the only way this could be identically zero is if the variation of the action is zero. This might be where the principle of least action comes from.

Some of this functional calculus doesn't seem to be very well developed in the literature I've seen. I'd appreciate those with more experience could check the accuracy of these assumptions. Thank you.
 
Last edited:
friend said:
OK, how does this sound? If we integrate the path integral one more time, it should equal 1 since it's a Dirac delta function. Then the functional derivative of that should be zero, since the functional derivative of a constant is zero. This much is necessarily True.
Yes, this much is true, but is it actually useful? Does it apply to any functional whatsoever, or just those that equal the Dirac delta function?

I do find it interesting that here we have the possibility of discovering the mathematical origin of the least action principle.

It's curious to note that the path integral can be developed from principle alone from the identiy operator:

\[<br /> 1 = \int_{ - \infty }^{ + \infty } {\left| {x_1 } \right\rangle \left\langle {x_1 } \right|dx_1 } <br /> \]<br />

And when this is applied an infinite number of times we get:

\[<br /> 1 = \int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {...\int_{ - \infty }^{ + \infty } {\left| {x_1 } \right\rangle \left\langle {{x_1 }}<br /> \mathrel{\left | {\vphantom {{x_1 } {x_2 }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_2 }} \right\rangle \left\langle {{x_2 }}<br /> \mathrel{\left | {\vphantom {{x_2 } {x_3 }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_3 }} \right\rangle \left\langle {x_3 } \right|...\left| {x_n } \right\rangle \left\langle {x_n } \right|dx_1 dx_2 dx_3 ...dx_n } } } } <br /> \]

where we can recognize \[<br /> \left\langle {{x_i }}<br /> \mathrel{\left | {\vphantom {{x_i } {x_j }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_j }} \right\rangle <br /> \]<br /> as the inner product between orthonormal vectors. And when this identity is applied to a general inner product we get:

\[<br /> \left\langle {{x_F }}<br /> \mathrel{\left | {\vphantom {{x_F } {x_0 }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_0 }} \right\rangle = \int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {...\int_{ - \infty }^{ + \infty } {\left\langle {{x_F }}<br /> \mathrel{\left | {\vphantom {{x_F } {x_1 }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_1 }} \right\rangle \left\langle {{x_1 }}<br /> \mathrel{\left | {\vphantom {{x_1 } {x_2 }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_2 }} \right\rangle \left\langle {{x_2 }}<br /> \mathrel{\left | {\vphantom {{x_2 } {x_3 }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_3 }} \right\rangle \left\langle {x_3 } \right|...\left| {x_n } \right\rangle \left\langle {{x_n }}<br /> \mathrel{\left | {\vphantom {{x_n } {x_0 }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_0 }} \right\rangle dx_1 dx_2 dx_3 ...dx_n } } } } <br /> \]<br />

If the vectors have finite components, then the \[<br /> \left\langle {{x_i }}<br /> \mathrel{\left | {\vphantom {{x_i } {x_j }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_j }} \right\rangle <br /> \]<br /> is the Kronecker delta, \[<br /> \delta _j^i <br /> \]<br />. But if the vectors have an infinite number of components, then the inner product is the Dirac delta function \[<br /> \delta \left( {x_i - x_j } \right)<br /> \]<br />.

Then if we use the gaussian from of the delta function, \[<br /> \delta \left( {x - x_0 } \right) = \mathop {\lim }\limits_{\Delta \to 0} {\textstyle{1 \over {\left( {\pi \Delta ^2 } \right)^{{\raise0.5ex\hbox{$\scriptstyle 1$}<br /> \kern-0.1em/\kern-0.15em<br /> \lower0.25ex\hbox{$\scriptstyle 2$}}} }}}e^{ - \left( {x - x_0 } \right)^2 /\Delta ^2 } <br /> \]<br />, then set \[<br /> \Delta = i^{{\raise0.5ex\hbox{$\scriptstyle 1$}<br /> \kern-0.1em/\kern-0.15em<br /> \lower0.25ex\hbox{$\scriptstyle 2$}}} \Delta t<br /> \], the exponential becomes \[<br /> i\left( {\left( {x - x_0 } \right)/\Delta t} \right)^2 <br /> \], which can be recognized as the square of the velocity when t is considered the time variable. With this substitution for each of the \[<br /> \left\langle {{x_i }}<br /> \mathrel{\left | {\vphantom {{x_i } {x_j }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_j }} \right\rangle <br /> \]<br />, the above integral becomes:

\[<br /> \int_{ - \infty }^{ + \infty } {...\int_{ - \infty }^{ + \infty } {\prod\nolimits_{1 = 1}^n {dx_i } } } \,e^{ - i\int_0^t {\dot x^2 \,\,dt} } <br /> \]

where each of the inner products contributes infinitesimally to the integral in the exponent. And apart from the constant m/2 in the exponent, this is the lagrangian for the kinetic energy of a free particle. And this is very similar to the way I've seen the path integral derived is some textbooks.

If we think more generally, we can substitute x or even x' in the exponent with some other function or field. The requirement would be that these functions be square-integrable so that the the exponent exists. This is all part of the definition of a Hilbert space of functions that seem to be the only functions applicable. And then we are talking about quantum field theory when functions replace x, etc. (not necessarily physical fields). Wouldn't it be interesting if ALL of physics could be derived from nothing more than the Identity? Perhaps there are some interesting philosophical implications that can be gained from that.

So it seems the path integral can be derived apart from physical concerns, and it may be that physics is just a part of these results.

So one question in this context is if the least action principle can be required on principle alone as well. Or is this purely a requirement of physics?

friend said:
In any event, I have seen where the functional derivative commutes with integration. So the functional derivative of the path integral will commute with all the integrals (all infinity of them) and pass inside to the exponential. So functional differentiation commutes with functional integration, right?

Then I assume:

\[<br /> \delta \left( {\exp \left( {\frac{{\rm{i}}}{\hbar }{\rm{S[x}}\left( {\rm{t}} \right){\rm{]}}} \right)} \right){\rm{ = }}\exp \left( {\frac{{\rm{i}}}{\hbar }{\rm{S}}\left[ {{\rm{x}}\left( {\rm{t}} \right)} \right]} \right){\rm{\cdot}}\left( {\frac{{\rm{i}}}{\hbar }\delta {\rm{S}}\left[ {{\rm{x}}\left( {\rm{t}} \right)} \right]} \right)<br /> \]<br />

And also that the only way this could be identically zero is if the variation of the action is zero. This might be where the principle of least action comes from.

I've seen in the book, Quantum Field Theory, by Lowell S. Brown, page 13 that:

\[<br /> {\textstyle{1 \over i}}{\textstyle{\delta \over {\delta f\left( t \right)}}}\left\langle {q&#039;,t_2 |q&#039;&#039;,t_1 } \right\rangle ^f = \int {[dq]\,e^{i\int_{t1}^{t2} {dt&#039;L_0 } } \,{\textstyle{1 \over i}}{\textstyle{\delta \over {\delta f\left( t \right)}}}\,e^{i\int_{t1}^{t2} {dt&#039;\,f\left( {t&#039;} \right)q\left( {t&#039;} \right)} } \, = \,} \int {[dq]\,q\left( t \right)} \,e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } <br /> \]<br />

But it's just given as is, and they don't say how they derive this, or how the functional derivate commutes with integration and passes onto the exponential and then to the action. But it is as if to say,

\[<br /> {\textstyle{1 \over i}}{\textstyle{\delta \over {\delta f\left( t \right)}}}\int {[dq]} \,e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } \, = \,\int {[dq]} \,{\textstyle{1 \over i}}{\textstyle{\delta \over {\delta f\left( t \right)}}}\left( {e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } } \right)<br /> \]<br />

\[<br /> = \int {[dq]} \,e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } {\textstyle{1 \over i}}{\textstyle{\delta \over {\delta f\left( t \right)}}}\left( {i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } \right)<br /> \]<br />

\[<br /> = \int {[dq]} \,e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } {\textstyle{\partial \over {\partial f\left( t \right)}}}\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]\, = \,\int {[dq]} \,e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } q\left( {t&#039;} \right)<br /> \]<br />

as I suspect is the case. So perhaps my suspicions are correct. Does anyone want to comment?

It might be that physicists do not concern themselves too much with the details of functional calculus. They seem to only use it in passing to get the least action principle and move on from there. I feel that the question I raise here should be of interest to physicists. But perhaps this is more of a technical math question and should be moved to the appropriate forum. I'm not sure which forum would be most interested in functional derivatives and integrals.
 
Last edited:
Here are some random ramblings.

friend said:
So one question in this context is if the least action principle can be required on principle alone as well. Or is this purely a requirement of physics?

It seems you partly search for a deeper understanding of things, but you seek to seek a deduction of some of the laws of physics from deductive logic, is that right?

I can related to your thinking, I seek something similar but in a different way where I replace the logical system with a more general one. I seek an inductive type of inference of the laws of physics from the a more general (not generally deductive) logic.

As I see it, it's a sort of parallell to your vision, but we use different logic. You seem to seek a closed fixed logical system, mine is an open evolving one.

In my think, the "least action principle" is treated on the same basis as "the second law". So Max ent rules and least action rules are two special cases of a more general idea. Conceptually I think this belongs in evolving inductive logic, rather than deductive.

The connection is that the action take by a system, relates to it's expectations of the future. And the least action then suggest that each possible future has a "weight" according to it's plausability. The relation between least action and entropy is that least action is loosely a form of "state relative entropy". This unifies the two principles.

The weighted possibilities, is quite analogous to a risk analysis, where the action takes the form of minimum speculation.

A hint of this is described here
http://en.wikipedia.org/wiki/Kullback–Leibler_divergence

Although that is still a simplification, the full evolving idea is not incorporated there.

So if we only can described the expected encoding information microstructure of a system, I think, it's expected actions should follow from such a intrinsic principle of minimum information divergence - close to - from "principle alone".

minimum information divergence is in effect a principle of minimum speculation or least action.

/Fredrik
 
Fra said:
It seems you partly search for a deeper understanding of things, but you seek to seek a deduction of some of the laws of physics from deductive logic, is that right?
This seems like the reasonable thing to do.

Fra said:
I can related to your thinking, I seek something similar but in a different way where I replace the logical system with a more general one. I seek an inductive type of inference of the laws of physics from the a more general (not generally deductive) logic.

As I see it, it's a sort of parallell to your vision, but we use different logic. You seem to seek a closed fixed logical system, mine is an open evolving one.

Inductive logic comes from deductive logic. Deductive logic is concerned with whether a proposition is true or false. Inductive logic counts the relative number of samples in various sets. But you have to be able to acknowledge whether it is true or false that a particular sample is included in your set before you can count how many there are.



Fra said:
In my think, the "least action principle" is treated on the same basis as "the second law". So Max ent rules and least action rules are two special cases of a more general idea. Conceptually I think this belongs in evolving inductive logic, rather than deductive.
If you don't know how your system of reasoning is evolving, then you can't say you have a valid system that will give you reliable answers. You might be using the "old" system which no longer applies. But if you do know how your system of reasoning is evolving, then this is because you know the rules by which it evolves. But then these rules by which it evolves are now more basic and fixed and serve as a rigid logic to be used.

Fra said:
The connection is that the action take by a system, relates to it's expectations of the future. And the least action then suggest that each possible future has a "weight" according to it's plausability. The relation between least action and entropy is that least action is loosely a form of "state relative entropy". This unifies the two principles.

The weighted possibilities, is quite analogous to a risk analysis, where the action takes the form of minimum speculation.

/Fredrik

It's interesting to note that the Dirac delta function is also considered a probability distributions. It's a gaussian distribution which represents completely random processes. In the development of the path integral (post 5) the inner product was a dirac delta function. The inner product also describes a transition from one state to the next. This is a kind of implication where the previous state necessarily results in the next state. This also may be a hint that deductive logic can be used to derive the path integral.

So perhaps this structure of physics ( the path integral) from completely random processes (the Dirac delta gaussian distribution) is the evolving law that you're looking for.
 
Last edited:
friend said:
I've seen in the book, Quantum Field Theory, by Lowell S. Brown, page 13 that:

\[<br /> {\textstyle{1 \over i}}{\textstyle{\delta \over {\delta f\left( t \right)}}}\left\langle {q&#039;,t_2 |q&#039;&#039;,t_1 } \right\rangle ^f = \int {[dq]\,e^{i\int_{t1}^{t2} {dt&#039;L_0 } } \,{\textstyle{1 \over i}}{\textstyle{\delta \over {\delta f\left( t \right)}}}\,e^{i\int_{t1}^{t2} {dt&#039;\,f\left( {t&#039;} \right)q\left( {t&#039;} \right)} } \, = \,} \int {[dq]\,q\left( t \right)} \,e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } <br /> \]<br />

But it's just given as is, and they don't say how they derive this, or how the functional derivate commutes with integration and passes onto the exponential and then to the action. But it is as if to say,

\[<br /> {\textstyle{1 \over i}}{\textstyle{\delta \over {\delta f\left( t \right)}}}\int {[dq]} \,e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } \, = \,\int {[dq]} \,{\textstyle{1 \over i}}{\textstyle{\delta \over {\delta f\left( t \right)}}}\left( {e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } } \right)<br /> \]<br />

\[<br /> = \int {[dq]} \,e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } {\textstyle{1 \over i}}{\textstyle{\delta \over {\delta f\left( t \right)}}}\left( {i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } \right)<br /> \]<br />

\[<br /> = \int {[dq]} \,e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } {\textstyle{\partial \over {\partial f\left( t \right)}}}\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]\, = \,\int {[dq]} \,e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } q\left( {t&#039;} \right)<br /> \]<br />

This somewhat confuses me because I thought that any variation of the action integral was considered to be identically zero. And this seems to be what he's doing. He then goes on to interate this process in order to form the definition of time ordered products. How can he do this?

Is the general variation, \[<br /> \delta S<br /> \]<br />, somehow different than a specific variation with respect to a function, \[<br /> \frac{\delta }{{\delta f\left( t \right)}}S<br /> \]<br /> such that the first is zero when the second is not zero?
 
Last edited:
friend said:
This somewhat confuses me because I thought that any variation of the action integral was considered to be identically zero. And this seems to be what he's doing. He then goes on to interate this process in order to form the definition of time ordered products. How can he do this?

Is the general variation, \[<br /> \delta S<br /> \]<br />, somehow different than a specific variation with respect to a function, \[<br /> \frac{\delta }{{\delta f\left( t \right)}}S<br /> \]<br /> such that the first is zero when the second is not zero?

Yes, this seems to be the case because what I found is that if

\[<br /> S\left( {x,y\left( x \right),z\left( x \right)} \right) = \int_a^b {L\left( {x,y\left( x \right),z\left( x \right)} \right)} \,dx<br /> \]<br />

then

\[<br /> \delta S = \frac{{\partial L}}{{\partial y}}\delta y + \frac{{\partial L}}{{\partial z}}\delta z<br /> \]<br />

but

\[<br /> \frac{{\delta S}}{{\delta z}} = \frac{{\partial L}}{{\partial z}}<br /> \]<br />

So there are extra terms in \[<br /> \delta S<br /> \]<br /> that are not in \[<br /> \frac{{\delta S}}{{\delta z}}<br /> \]<br /> which can be used to make first zero when the second is not.

This makes me wonder if when we start adding fields in the Lagrangian to account for other kinds of particles and interactions, is there any care taken to make sure that the whole variation of the action remains zero? From what I've seen in my limited experience, they add terms to the lagrangian in a somewhat ad-hoc way to account for extra fields and self-interactions. I don't remember seeing any procedure where they take the time to check. Perhaps the coupling constants or the exponent of the fields are restricted in such a procedure. As just a wild quess, maybe this is where the cosmological constant problem comes in.
 
Last edited:
  • #10
friend said:
OK, how does this sound? If we integrate the path integral one more time, it should equal 1 since it's a Dirac delta function. Then the functional derivative of that should be zero, since the functional derivative of a constant is zero. This much is necessarily True... And then taking the functional derivative of that would probably create a factor consisting of the variation of the action. This would be required to be zero since nothing else could be set to zero.

As a brief introduction for purposes of context, functional derivatives are developed as follows:

Let \[<br /> S\left[ {x,y\left( x \right),z\left( x \right)} \right]<br /> \]<br /> be a functional of the functions y(x) and z(x).

Then this functional, S, can be expanded in a taylor series

\[<br /> S\left[ {x ,y_0(x) + h_y ,z_0(x) + h_z } \right] = \,S\left[ {x,y_0 ,z_0 } \right] + \frac{{\partial S}}{{\partial y}}|_{y_0 ,z_0 } h_y + \frac{{\partial S}}{{\partial z}}|_{y_0 ,z_0 } h_z + \frac{1}{{2!}}\left( {\frac{{\partial ^2 S}}{{\partial y^2 }}|_{y_0 ,z_0 } h_y ^2 + 2\frac{{\partial S}}{{\partial y}}\frac{{\partial S}}{{\partial z}}|_{y_0 ,z_0 } h_y h_z + \frac{{\partial ^2 S}}{{\partial z^2 }}|_{y_0 ,z_0 } h_z ^2 } \right) + ...<br /> \]<br />

Or the finite difference is

\[<br /> \Delta S = S\left[ {x,y + h_y ,z + h_z } \right] - S\left[ {x,y,z} \right] = \frac{{\partial S}}{{\partial y}}h_y + \frac{{\partial S}}{{\partial z}}h_z + \frac{1}{{2!}}\left( {\frac{{\partial ^2 S}}{{\partial y^2 }}h_y ^2 + 2\frac{{\partial S}}{{\partial y}}\frac{{\partial S}}{{\partial z}}h_y h_z + \frac{{\partial ^2 S}}{{\partial z^2 }}h_z ^2 } \right) + ...<br /> \]<br />

where it is understood that we evaluate S using the functions \[<br /> ,y_0(x) ,z_0(x) <br /> \]<br />

We can change notation so that \[<br /> h_y = \delta y<br /> \]<br /> and \[<br /> h_z = \delta z<br /> \]<br /> and get

\[<br /> \begin{array}{l}\\ <br /> \Delta S = S\left[ {x,y + \delta y,z + \delta z } \right] - S\left[ {x,y,z} \right] = \frac{{\partial S}}{{\partial y}}\delta y + \frac{{\partial S}}{{\partial z}}\delta z + \frac{1}{{2!}}\left( {\frac{{\partial ^2 S}}{{\partial y^2 }}\left( {\delta y} \right)^2 + 2\frac{{\partial S}}{{\partial y}}\frac{{\partial S}}{{\partial z}}\delta y\delta z + \frac{{\partial ^2 S}}{{\partial z^2 }}\left( {\delta z} \right)^2 } \right) + ... \\ <br /> \end{array}<br /> \]<br />

The first variation of S, labeled \[<br /> \delta S<br /> \]<br /> , which is called the first functional derivative of S, is the linear part of the Taylor expansion, or

\[<br /> \delta S = \frac{{\partial S}}{{\partial y}}\delta y + \frac{{\partial S}}{{\partial z}}\delta z<br /> \]<br />

and the term with second power in \[<br /> \delta y<br /> \]<br /> and \[<br /> \delta z<br /> \]<br /> is called the second variation or second functional derivative, or

\[<br /> \delta ^2 S = \frac{{\partial ^2 S}}{{\partial y^2 }}\left( {\delta y} \right)^2 + 2\frac{{\partial S}}{{\partial y}}\frac{{\partial S}}{{\partial z}}\delta y\delta z + \frac{{\partial ^2 S}}{{\partial z^2 }}\left( {\delta z} \right)^2 <br /> \]<br />

and similarly for higher order variations, \[<br /> \delta ^3 S,\,\delta ^4 S,\,...<br /> \]<br />

And with this notation,

\[<br /> \Delta S = \delta S + \frac{1}{{2!}}\delta ^2 S + \frac{1}{{3!}}\delta ^3 S + ...<br /> \]<br />


What is \[<br /> \delta \left( {\delta S} \right)<br /> \]<br /> ?

\[<br /> \delta \left( {\delta S} \right) = \delta \left( {\frac{{\partial S}}{{\partial y}}\delta y + \frac{{\partial S}}{{\partial z}}\delta z} \right) = \frac{\partial }{{\partial y}}\left( {\frac{{\partial S}}{{\partial y}}\delta y + \frac{{\partial S}}{{\partial z}}\delta z} \right)\delta y + \frac{\partial }{{\partial z}}\left( {\frac{{\partial S}}{{\partial y}}\delta y + \frac{{\partial S}}{{\partial z}}\delta z} \right)\delta z<br /> \]<br />

\[<br /> = \left( {\frac{{\partial ^2 S}}{{\partial y^2 }}\delta y + \frac{{\partial S}}{{\partial y}}\frac{\partial }{{\partial y}}\left( {\delta y} \right) + \frac{{\partial ^2 S}}{{\partial y\partial z}}\delta z + \frac{{\partial S}}{{\partial z}}\frac{\partial }{{\partial y}}\left( {\delta z} \right)} \right)\delta y + \left( {\frac{{\partial ^2 S}}{{\partial z\partial y}}\delta y + \frac{{\partial S}}{{\partial y}}\frac{\partial }{{\partial z}}\left( {\delta y} \right) + \frac{{\partial ^2 S}}{{\partial z^2 }}\delta z + \frac{{\partial S}}{{\partial z}}\frac{\partial }{{\partial z}}\left( {\delta z} \right)} \right)\delta z<br /> \]<br />

But y(x) and z(x) do not depend on one another, so terms like \[<br /> {\frac{\partial }{{\partial y}}\left( {\delta z} \right)}<br /> \]<br /> are obviously zero.

And with terms like \[<br /> {\frac{\partial }{{\partial y}}\left( {\delta y} \right)}<br /> \]<br /> we have that \[<br /> {\delta y}<br /> \]<br /> is a function of x and is independent of y. So they go to zero as well. What is left is

\[<br /> \delta \left( {\delta S} \right) = \frac{{\partial ^2 S}}{{\partial y^2 }}\left( {\delta y} \right)^2 + \frac{{\partial ^2 S}}{{\partial y\partial z}}\delta y\delta z + \frac{{\partial ^2 S}}{{\partial z\partial y}}\delta y\delta z + \frac{{\partial ^2 S}}{{\partial z^2 }}\left( {\delta z} \right)^2 = \frac{{\partial ^2 S}}{{\partial y^2 }}\left( {\delta y} \right)^2 + 2\frac{{\partial ^2 S}}{{\partial y\partial z}}\delta y\delta z + \frac{{\partial ^2 S}}{{\partial z^2 }}\left( {\delta z} \right)^2 <br /> \]<br />

But this last part was exactly how \[<br /> \delta ^2 S<br /> \]<br /> was defined. So we have

\[<br /> \delta \left( {\delta S} \right) = \delta ^2 S<br /> \]<br />

Now if,

\[<br /> S\left[ {x,y\left( x \right),z\left( x \right)} \right] = dF\left[ {x,y\left( x \right),z\left( x \right)} \right]<br /> \]<br />

then,

\[<br /> \Delta S\left[ {x,y\left( x \right),z\left( x \right)} \right] = S\left[ {x,y + \delta y,z + \delta z} \right] - S\left[ {x,y,z} \right] = \Delta dF\left[ {x,y\left( x \right),z\left( x \right)} \right]<br /> \]<br />

\[<br /> = dF\left[ {x,y + \delta y,z + \delta z} \right] - dF\left[ {x,y\left( x \right),z\left( x \right)} \right] = d\left( {F\left[ {x,y + \delta y,z + \delta z} \right] - F\left[ {x,y\left( x \right),z\left( x \right)} \right]} \right) = d\Delta F\left[ {x,y\left( x \right),z\left( x \right)} \right]<br /> \]<br />

So that,

\[<br /> \Delta d = d\Delta <br /> \]<br />

\[<br /> \Delta <br /> \]<br /> commutes with \[<br /> d<br /> \]<br />

Then,

\[<br /> \Delta dF = \delta \left( {dF} \right) + \frac{1}{{2!}}\delta ^2 \left( {dF} \right) + \frac{1}{{3!}}\delta ^3 \left( {dF} \right) + ... = d\left( {\delta F + \frac{1}{{2!}}\delta ^2 F + \frac{1}{{3!}}\delta ^3 F + ...} \right) = d\delta F + \frac{1}{{2!}}d\delta ^2 F + ...<br /> \]<br />

And keeping only the first approximation linear terms on both sides of the equation,

\[<br /> \delta d = d\delta <br /> \]<br />

\[<br /> \delta <br /> \]<br /> commutes with \[<br /> d<br /> \]<br />

And if

\[<br /> S\left[ {x,y\left( x \right),z\left( x \right)} \right] = \int_a^b {L\left[ {x,y\left( x \right),z\left( x \right)} \right]} \,dx<br /> \]<br />

Then

\[<br /> \Delta S = S\left[ {x,y + \delta y,z + \delta z} \right] - S\left[ {x,y,z} \right] = \int_a^b {L\left[ {x,y + \delta y,z + \delta z} \right]} \,dx - \int_a^b {L\left[ {x,y,z} \right]} \,dx = \Delta \int_a^b {L\left[ {x,y,z} \right]} \,dx<br /> \]<br />

\[<br /> = \int_a^b {L\left[ {x,y + \delta y,z + \delta z} \right] - L\left[ {x,y,z} \right]} \,dx = \int_a^b {\Delta L\left[ {x,y,z} \right]} \,dx<br /> \]<br />

So we see here that \[<br /> \Delta <br /> \]<br /> commutes with \[<br /> \int {} <br /> \]<br />

But also,

\[<br /> \Delta S = \delta S + \frac{1}{{2!}}\delta ^2 S + ... = \delta \left( {\int_a^b L \,dx} \right) + \frac{1}{{2!}}\delta ^2 \left( {\int_a^b L \,dx} \right) + ...<br /> \]<br />

\[<br /> = \int_a^b {\Delta L} \,dx = \int_a^b {\delta L + \frac{1}{{2!}}\delta ^2 L + ...} \,dx = \int_a^b {\delta L\,dx + \frac{1}{{2!}}\int_a^b {\delta ^2 L\,dx + ...} } <br /> \]<br />

Then keeping first approximation linear terms on each side of the equation we see

\[<br /> \delta \int_a^b L \,dx = \int_a^b {\delta L\,dx} <br /> \]<br />

Or, \[<br /> \delta \,\,\,{\rm{commutes}}\,\,{\rm{with}}\,\,\,\int {} <br /> \]<br />

Many text use the integral definition of a functional, \[<br /> S = \int {L\,dx} <br /> \]<br /> in its development. I suppose they do this because it makes it easier to justify taking only the linear terms inside the integral since differentials approach zero more naturally in the process of integration.

It's easy to see that variation commutes with any number of integral signs since a difference outside the integral is translated to a difference inside the integral signs. Or,

\[<br /> \Delta \int {\int {...\int {L\left[ {x,y,z} \right]} } } dx_1 dx_2 ...dx_n = \Delta \int {d[x]} L\left[ {x,y,z} \right]<br /> \]<br />

\[<br /> = \int {d[x]} L\left[ {x,y + \delta y,z + \delta z} \right] - \int {d[x]} L\left[ {x,y,z} \right] = \int {d[x]\left( {L\left[ {x,y + \delta y,z + \delta z} \right] - L\left[ {x,y,z} \right]} \right)} = \int {d[x]\Delta L\left[ {x,y,z} \right]} <br /> \]<br />

So following a similar procedure of keeping only linear terms,

\[<br /> \delta \int {d[x]\,L = \int {d[x]\delta L} } <br /> \]<br />

This means that the variation of the integration of the path integral gets passed inside all the infinite number of integrations to taking the variation of the exponent of the action.
 
Last edited:
  • #11
friend said:
OK, how does this sound? If we integrate the path integral one more time, it should equal 1 since it's a Dirac delta function. Then the functional derivative of that should be zero, since the functional derivative of a constant is zero. This much is necessarily True.

So what does it mean if this is not true? If the first variation of the integration of the path integral is not zero, then what does this mean? Feel free to offer your thoughts, but I think it means that the path integral is not a Dirac delta function, not a wavefunction or a transition function <x|x_0>. And if the path integral as a whole is not a delta function, then I think this means that one or more of the infinite number of <xi|xj> that are inside the path integral is not a delta function. Or at least it means that the infinite multiplication of the <xi|xj> is not an infinite multiplication of delta functions. And since the delta functions, <xi|xj> , are expressed in terms of an exponential whose exponents can be added up to equal the action integral of one exponential, at least this means the action type integral is not adding up right. So I can see how this would place some restrictions on the action integral, but I'm not sure what.
 
  • #12
friend said:
Yes, this much is true, but is it actually useful? Does it apply to any functional whatsoever, or just those that equal the Dirac delta function?

I do find it interesting that here we have the possibility of discovering the mathematical origin of the least action principle.

It's curious to note that the path integral can be developed from principle alone from the identiy operator:

\[<br /> 1 = \int_{ - \infty }^{ + \infty } {\left| {x_1 } \right\rangle \left\langle {x_1 } \right|dx_1 } <br /> \]<br />

And when this is applied an infinite number of times we get:

\[<br /> 1 = \int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {...\int_{ - \infty }^{ + \infty } {\left| {x_1 } \right\rangle \left\langle {{x_1 }}<br /> \mathrel{\left | {\vphantom {{x_1 } {x_2 }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_2 }} \right\rangle \left\langle {{x_2 }}<br /> \mathrel{\left | {\vphantom {{x_2 } {x_3 }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_3 }} \right\rangle \left\langle {x_3 } \right|...\left| {x_n } \right\rangle \left\langle {x_n } \right|dx_1 dx_2 dx_3 ...dx_n } } } } <br /> \]

where we can recognize \[<br /> \left\langle {{x_i }}<br /> \mathrel{\left | {\vphantom {{x_i } {x_j }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_j }} \right\rangle <br /> \]<br /> as the inner product between orthonormal vectors. And when this identity is applied to a general inner product we get:

\[<br /> \left\langle {{x_F }}<br /> \mathrel{\left | {\vphantom {{x_F } {x_0 }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_0 }} \right\rangle = \int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {...\int_{ - \infty }^{ + \infty } {\left\langle {{x_F }}<br /> \mathrel{\left | {\vphantom {{x_F } {x_1 }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_1 }} \right\rangle \left\langle {{x_1 }}<br /> \mathrel{\left | {\vphantom {{x_1 } {x_2 }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_2 }} \right\rangle \left\langle {{x_2 }}<br /> \mathrel{\left | {\vphantom {{x_2 } {x_3 }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_3 }} \right\rangle \left\langle {x_3 } \right|...\left| {x_n } \right\rangle \left\langle {{x_n }}<br /> \mathrel{\left | {\vphantom {{x_n } {x_0 }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_0 }} \right\rangle dx_1 dx_2 dx_3 ...dx_n } } } } <br /> \]<br />

If the vectors have finite components, then the \[<br /> \left\langle {{x_i }}<br /> \mathrel{\left | {\vphantom {{x_i } {x_j }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_j }} \right\rangle <br /> \]<br /> is the Kronecker delta, \[<br /> \delta _j^i <br /> \]<br />. But if the vectors have an infinite number of components, then the inner product is the Dirac delta function \[<br /> \delta \left( {x_i - x_j } \right)<br /> \]<br />.

Then if we use the gaussian from of the delta function, \[<br /> \delta \left( {x - x_0 } \right) = \mathop {\lim }\limits_{\Delta \to 0} {\textstyle{1 \over {\left( {\pi \Delta ^2 } \right)^{{\raise0.5ex\hbox{$\scriptstyle 1$}<br /> \kern-0.1em/\kern-0.15em<br /> \lower0.25ex\hbox{$\scriptstyle 2$}}} }}}e^{ - \left( {x - x_0 } \right)^2 /\Delta ^2 } <br /> \]<br />, then set \[<br /> \Delta = i^{{\raise0.5ex\hbox{$\scriptstyle 1$}<br /> \kern-0.1em/\kern-0.15em<br /> \lower0.25ex\hbox{$\scriptstyle 2$}}} \Delta t<br /> \], the exponential becomes \[<br /> i\left( {\left( {x - x_0 } \right)/\Delta t} \right)^2 <br /> \], which can be recognized as the square of the velocity when t is considered the time variable. With this substitution for each of the \[<br /> \left\langle {{x_i }}<br /> \mathrel{\left | {\vphantom {{x_i } {x_j }}}<br /> \right. \kern-\nulldelimiterspace}<br /> {{x_j }} \right\rangle <br /> \]<br />, the above integral becomes:

\[<br /> \int_{ - \infty }^{ + \infty } {...\int_{ - \infty }^{ + \infty } {\prod\nolimits_{1 = 1}^n {dx_i } } } \,e^{ - i\int_0^t {\dot x^2 \,\,dt} } <br /> \]

where each of the inner products contributes infinitesimally to the integral in the exponent. And apart from the constant m/2 in the exponent, this is the lagrangian for the kinetic energy of a free particle. And this is very similar to the way I've seen the path integral derived is some textbooks.

If we think more generally, we can substitute x or even x' in the exponent with some other function or field. The requirement would be that these functions be square-integrable so that the the exponent exists. This is all part of the definition of a Hilbert space of functions that seem to be the only functions applicable. And then we are talking about quantum field theory when functions replace x, etc. (not necessarily physical fields). Wouldn't it be interesting if ALL of physics could be derived from nothing more than the Identity? Perhaps there are some interesting philosophical implications that can be gained from that.

So it seems the path integral can be derived apart from physical concerns, and it may be that physics is just a part of these results.

So one question in this context is if the least action principle can be required on principle alone as well. Or is this purely a requirement of physics?



I've seen in the book, Quantum Field Theory, by Lowell S. Brown, page 13 that:

\[<br /> {\textstyle{1 \over i}}{\textstyle{\delta \over {\delta f\left( t \right)}}}\left\langle {q&#039;,t_2 |q&#039;&#039;,t_1 } \right\rangle ^f = \int {[dq]\,e^{i\int_{t1}^{t2} {dt&#039;L_0 } } \,{\textstyle{1 \over i}}{\textstyle{\delta \over {\delta f\left( t \right)}}}\,e^{i\int_{t1}^{t2} {dt&#039;\,f\left( {t&#039;} \right)q\left( {t&#039;} \right)} } \, = \,} \int {[dq]\,q\left( t \right)} \,e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } <br /> \]<br />

But it's just given as is, and they don't say how they derive this, or how the functional derivate commutes with integration and passes onto the exponential and then to the action. But it is as if to say,

\[<br /> {\textstyle{1 \over i}}{\textstyle{\delta \over {\delta f\left( t \right)}}}\int {[dq]} \,e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } \, = \,\int {[dq]} \,{\textstyle{1 \over i}}{\textstyle{\delta \over {\delta f\left( t \right)}}}\left( {e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } } \right)<br /> \]<br />

\[<br /> = \int {[dq]} \,e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } {\textstyle{1 \over i}}{\textstyle{\delta \over {\delta f\left( t \right)}}}\left( {i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } \right)<br /> \]<br />

\[<br /> = \int {[dq]} \,e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } {\textstyle{\partial \over {\partial f\left( t \right)}}}\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]\, = \,\int {[dq]} \,e^{i\int_{t1}^{t2} {dt&#039;\,\left[ {L_0 + f\left( {t&#039;} \right)q\left( {t&#039;} \right)} \right]} } q\left( {t&#039;} \right)<br /> \]<br />

as I suspect is the case. So perhaps my suspicions are correct. Does anyone want to comment?

It might be that physicists do not concern themselves too much with the details of functional calculus. They seem to only use it in passing to get the least action principle and move on from there. I feel that the question I raise here should be of interest to physicists. But perhaps this is more of a technical math question and should be moved to the appropriate forum. I'm not sure which forum would be most interested in functional derivatives and integrals.

Is there a reason why you're answering to your own posts?
Just curious, that's all...
 
  • #13
Friend, I didn't respond because I felt I would just be repeating arguments I've made before and that doesn't convey anything.

Since you like play around with formulas I'll just throw in random comments/critique instead regarding your elaborations.

1. Your connection of dirac delta ang gaussian probability is a bit ambigous. Although the delta distribution can be defined by a construction involving the limit of a gaussian distribution, this choice of construction is not unique. Ie. you can find other families of functions that are not gaussan that are ALSO equally valid for defining the dirac distribution.

So if we are really talking about a proper distribution, the dirac delta has IMO not much specifically to do with gaussian distributions.

2. OTOH, if you think the gaussan really is the key, then the actual limiting process must contain physics. Which means we can not take the limit, the running of the n is something that has physical significance, do you agree? (If so, this is something that I am trying to do as well)

3. Also the choice of decomposition or partitioning when working with an identity operator is also ambigous. Similarly, I argue that this choice and the size or the partitioning index has physical significance. How about that?

4. I won't repeat any old argumetns but you think of induction in terms of counting sets: OK, if we take this view, then what I suggest that prevents a deductive scheme is that these sets are not given, they are changing. In particular can the SIZE of hte sets change. The size of the set represents the span of the counting index.

friend said:
If you don't know how your system of reasoning is evolving, then you can't say you have a valid system that will give you reliable answers. You might be using the "old" system which no longer applies. But if you do know how your system of reasoning is evolving, then this is because you know the rules by which it evolves. But then these rules by which it evolves are now more basic and fixed and serve as a rigid logic to be used.

I don't think we can not prove that our system is valid in a meaningful way. My point is that we don't need to do this. Nature is as far as I can see not an axiomatic construction.

The old system is updated in the light of new input, as per a rationality rule. But this is not predictable in advance simply because the new input hasn't arrived yet. The feedback drives the selection and evolution.

Anyway I think it would be interesting if you would consider the PHYSICS of the limiting procedure. And motivate it from there, rather than choosing at will a possible (but not unique) way of defining the dirac delta.

/Fredrik
 
  • #14
Neo_Anderson said:
Is there a reason why you're answering to your own posts?
Just curious, that's all...
It is a work in progress for me to understand these things. So I post what I think I understand as I come to know it so that someone can correct me if I'm wrong. That would be very much appreciated.
 
  • #15
Fra said:
Friend, I didn't respond because I felt I would just be repeating arguments I've made before and that doesn't convey anything.

Since you like play around with formulas I'll just throw in random comments/critique instead regarding your elaborations.

1. Your connection of dirac delta ang gaussian probability is a bit ambigous. Although the delta distribution can be defined by a construction involving the limit of a gaussian distribution, this choice of construction is not unique. Ie. you can find other families of functions that are not gaussan that are ALSO equally valid for defining the dirac distribution.

So if we are really talking about a proper distribution, the dirac delta has IMO not much specifically to do with gaussian distributions.

Right, there are non-gaussian ways of describing the Dirac delta function. So why a gaussian, you ask? The gaussian distribution represents a totally random process. It's also called a bell curve or normal distribution. People in quality control use it to prove that there is nothing within their control to better maintain a tighter tolerance in their manufacturing processes. So the question is what in the development of the path integral would indicate the use of the gaussian form of the dirac delta function.

The dirac delta function is used to describe the inner product of two states. But the path integral is constructed in such a way that the inner product of every two states of an infinite number of state is used. And it is constructed so that it comes out the same no matter how you label the states or in what order you take the products. So every inner product must have the same weight as every other, so it doesn't matter how you chose them. In other words, there is a completely random process involved with the inner product which is also described by the delta function. I believe that means we have to chose the gaussian form of the dirac delta function.

But you may say that a random process of chosing an inner product does not mean that there's a random process in the product itself. Perhaps, but then again using the gaussian form of the inner product indicates that there is no further structure within the process of taking the inner product. If the best we can do in the development is to say that there is an inner product, but we inherently cannot possibly find anymore structural information than that, then I think that dictates the use of the gaussian form of the delta function. The gaussian form should be the default choice unless further structure can be discerned. But then again this added mathematical structure would have some spacetime dependence, being the only variables at that level. And this would probably create different weights for different inner products and ruin the covariant nature of the formulation. So I'm seeing here a glimps of a connection between covariance, no hidden variables, and the gaussian form of the dirac delta function for the inner product.

Fra said:
2. OTOH, if you think the gaussan really is the key, then the actual limiting process must contain physics. Which means we can not take the limit, the running of the n is something that has physical significance, do you agree? (If so, this is something that I am trying to do as well)

I think the limiting process only indicates that a continum is involved.

Fra said:
3. Also the choice of decomposition or partitioning when working with an identity operator is also ambigous. Similarly, I argue that this choice and the size or the partitioning index has physical significance. How about that?

Yes, that's why I perfer to define the path integral in terms of the recursive relation of the dirac delta function and not the recursive relation of the identity operator. Because then, as you suggest, one has to justify the use of a decomposition into orthonormal vectors. Although, maybe it's just as obvious to some that the dirac delta is also described by the inner product of orthonormal vectors. But in my mind that takes a little more work.

Fra said:
4. I won't repeat any old argumetns but you think of induction in terms of counting sets: OK, if we take this view, then what I suggest that prevents a deductive scheme is that these sets are not given, they are changing. In particular can the SIZE of hte sets change. The size of the set represents the span of the counting index.

I don't know, you may have to take into account every sample (or state) in the space equally so that it may not matter which set it was in.

Fra said:
I don't think we can not prove that our system is valid in a meaningful way. My point is that we don't need to do this. Nature is as far as I can see not an axiomatic construction.

The old system is updated in the light of new input, as per a rationality rule. But this is not predictable in advance simply because the new input hasn't arrived yet. The feedback drives the selection and evolution.

Then what you are taking about is not deriving physics on principle alone, be it deductive or inductive, or whatever. Now what you are talking about is just using a trial and error method to find the best curve that fits that data, a curve fitting exercise. You're trying to discern fundamental physics by analyzing the underlying structure of the curves that fit the data. But that may change with better data. Obviously what is underlying any curve or equation is logic and counting.
 
Last edited:
  • #16
I'll pass going into argumenting details of my own thinking since it's clear enough that we have different strategies and we don't seem to connect but some convergence points might exists, which is why perhaps feedback from a different direction might still be useful fwiw?

friend said:
Right, there are non-gaussian ways of describing the Dirac delta function. So why a gaussian, you ask? The gaussian distribution represents a totally random process. It's also called a bell curve or normal distribution. People in quality control use it to prove that there is nothing within their control to better maintain a tighter tolerance in their manufacturing processes. So the question is what in the development of the path integral would indicate the use of the gaussian form of the dirac delta function.

Maybe I was unclear, my point was that your line of reasoning is towards a decuctive system is not deductive (and I think you want to make deductive progress?). I see this as an inconsistency. I mean if your aim is a logical axiomatic system - the road to that can hardly be a deductive process, right? Maybe this doesn't bother you, but it looks weird from my perspective:

Going from distributions, to a logicallt possible - but not unique choice - of familys of functions that can be used to defined the distribution as a limit, make further arguemtns based on that choice, and then take the limit again to get back at distributions seems like am ambigous argument to me. That was my main point.

My point is not that the connection to statistics is uninteresting, but I see not deductive path to such a connection.

Aldo, since not all random variables don't have gaussian distributions, but rather than the distribution of sampled means, converges to one as the sample size -> inf which is a difference, this I think calls for clarifications too. It seems to suggest the event space itself is emergent from an average process?

friend said:
The dirac delta function is used to describe the inner product of two states. But the path integral is constructed in such a way that the inner product of every two states of an infinite number of state is used. And it is constructed so that it comes out the same no matter how you label the states or in what order you take the products. So every inner product must have the same weight as every other, so it doesn't matter how you chose them. In other words, there is a completely random process involved with the inner product which is also described by the delta function. I believe that means we have to chose the gaussian form of the dirac delta function.

But you may say that a random process of chosing an inner product does not mean that there's a random process in the product itself. Perhaps, but then again using the gaussian form of the inner product indicates that there is no further structure within the process of taking the inner product. If the best we can do in the development is to say that there is an inner product, but we inherently cannot possibly find anymore structural information than that, then I think that dictates the use of the gaussian form of the delta function.

This argument is not clear to me. In your deductive view, how you introduce the notion of inner product in the first place and what it's physical significance is? And how does this follow from logic?

This thing with inner products and transition probabilities are usually part of what comes with the QM axioms or postulates such as projection and born postulates. But I thought you were going to find an alternative deduction.

If we take a step back, what's the starting point in your quest? Are you taking some structure of QM for granted, if so which postulates or axioms are you trying to explain form the others?

/Fredrik
 
  • #17
A short comment.

friend said:
Then what you are taking about is not deriving physics on principle alone, be it deductive or inductive, or whatever. Now what you are talking about is just using a trial and error method to find the best curve that fits that data, a curve fitting exercise. You're trying to discern fundamental physics by analyzing the underlying structure of the curves that fit the data. But that may change with better data. Obviously what is underlying any curve or equation is logic and counting.

There is a substantial difference between simple curve fitting, and adaptive learning such as biological evolution and human science is a good example of. You simply can not characterize in any sensible way, biological evolution as curve fitting. It's your characterisation of what I tried to say as basic curve fitting that most clearly illustrates that I totaly fail to convey much.

/Fredrik
 
  • #18
Fra said:
This argument is not clear to me. In your deductive view, how you introduce the notion of inner product in the first place and what it's physical significance is? And how does this follow from logic?

This thing with inner products and transition probabilities are usually part of what comes with the QM axioms or postulates such as projection and born postulates. But I thought you were going to find an alternative deduction.

If we take a step back, what's the starting point in your quest? Are you taking some structure of QM for granted, if so which postulates or axioms are you trying to explain form the others?

/Fredrik

You might remember from a previous post that using the identity an infinite number of times one can get,

\[\left\langle {{x_F }} \mathrel{\left | {\vphantom {{x_F } {x_0 }}} \right. \kern-\nulldelimiterspace} {{x_0 }} \right\rangle = \int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {...\int_{ - \infty }^{ + \infty } {\left\langle {{x_F }} \mathrel{\left | {\vphantom {{x_F } {x_1 }}} \right. \kern-\nulldelimiterspace} {{x_1 }} \right\rangle \left\langle {{x_1 }} \mathrel{\left | {\vphantom {{x_1 } {x_2 }}} \right. \kern-\nulldelimiterspace} {{x_2 }} \right\rangle \left\langle {{x_2 }} \mathrel{\left | {\vphantom {{x_2 } {x_3 }}} \right. \kern-\nulldelimiterspace} {{x_3 }} \right\rangle \left\langle {x_3 } \right|...\left| {x_n } \right\rangle \left\langle {{x_n }} \mathrel{\left | {\vphantom {{x_n } {x_0 }}} \right. \kern-\nulldelimiterspace} {{x_0 }} \right\rangle dx_1 dx_2 dx_3 ...dx_n } } } } \]

and then that \[ &lt; x|x_0 &gt; \, = \,\,\delta (x - x_0 )\ \] This much you may recognize from textbooks.

But it's easy to show from the recursive property of the dirac delta function that you can get the same result.

You might remember that,

\[<br /> \int_{ - \infty }^{ + \infty } {{\rm{f(x}}_1 {\rm{)\delta (x}}_1 {\rm{ - x}}_0 ){\rm{dx}}_1 } = {\rm{f(x}}_0 )<br /> \]

So that if \[<br /> {\rm{f(x}}_1 ) = {\rm{\delta (x - x}}_1 )<br /> \]<br />, then

\[<br /> \int_{ - \infty }^{ + \infty } {{\rm{\delta (x - x}}_1 {\rm{)\delta (x}}_1 {\rm{ - x}}_0 ){\rm{dx}}_1 } = {\rm{\delta (x - x}}_0 )<br /> \]<br />

So that this process could be iterated an infinite number of times to get,

\[<br /> \int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {...\int_{ - \infty }^{ + \infty } {{\rm{\delta (x - x}}_n {\rm{)\delta (x}}_n {\rm{ - x}}_{n - 1} {\rm{)}}...{\rm{\delta (x}}_1 {\rm{ - x}}_0 ){\rm{dx}}_n {\rm{dx}}_{n - 1} ...{\rm{dx}}_1 } } } = {\rm{\delta (x - x}}_0 )<br /> \]<br />

This is my starting point. But I wanted to show a similarity with the traditional path integral construction. You might remember how Ariel Caticha used the inner product \[ &lt; x|x_0 &gt; \,\] to represent implication. But I've seen more directly how implication is represented by the dirac delta function.

In any event, the next step is to replace the dirac delta with a function. Most QM books go through a derivation using the schroedinger equation. I make the replacement directly, which is what you seem to be having trouble with. Wouldn't it be advantageous to derive the path integral from first principles and then derive schroedinger's equation from that instead of only the other way around? And don't you find it the least bit curious that replacing the delta with a gaussian gives you something that looks like QM?

Yes, maybe I need to work on justifying the choice of a gaussian for the delta a bit more. Does it make sense to say that chosing a particular inner product (or dirac delta) must be a completely random process, that this means that the inner product must itself represent a random process (gaussian delta function)? The inner product is a combination of two things. So this is like asking if taking a combination of any two is a completely random process, then that combination represents a completely random process.

But the main question of this thread is that given a path integral, does the least action principle follow from that just by taking the first variation of the integration of the path integral. I was trying to justify lagrangian mechanic from first principles. This is where the cannonical version of QM comes from, right?
 
Last edited:
  • #19
friend said:
You might remember that,

\[<br /> \int_{ - \infty }^{ + \infty } {{\rm{f(x}}_1 {\rm{)\delta (x}}_1 {\rm{ - x}}_0 ){\rm{dx}}_1 } = {\rm{f(x}}_0 )<br /> \]
Yes this is from the definition of delta distribution.
friend said:
So that if \[<br /> {\rm{f(x}}_1 ) = {\rm{\delta (x - x}}_1 )<br /> \]<br />, then

\[<br /> \int_{ - \infty }^{ + \infty } {{\rm{\delta (x - x}}_1 {\rm{)\delta (x}}_1 {\rm{ - x}}_0 ){\rm{dx}}_1 } = {\rm{\delta (x - x}}_0 )<br /> \]<br />

This would need clarification I think. The diract delta is technically not a function, it's a distribution(http://en.wikipedia.org/wiki/Distribution_(mathematics )). And it operates on test functions, and now you use a distribution as a test function, which needs clarification to be more well defined.

More later, when I get more time to ttype.

/Fredrik
 
Last edited by a moderator:
  • #20
friend said:
\[<br /> \int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {...\int_{ - \infty }^{ + \infty } {{\rm{\delta (x - x}}_n {\rm{)\delta (x}}_n {\rm{ - x}}_{n - 1} {\rm{)}}...{\rm{\delta (x}}_1 {\rm{ - x}}_0 ){\rm{dx}}_n {\rm{dx}}_{n - 1} ...{\rm{dx}}_1 } } } = {\rm{\delta (x - x}}_0 )<br /> \]<br />

This is my starting point.
Ok, Set aside some technical details, which is not the biggest issues here anyway
- I don't see the physics in this starting point.

Suppose we accept what you write this as kind of identity of your choice, then what predictive power does it yield?
friend said:
But I wanted to show a similarity with the traditional path integral construction. You might remember how Ariel Caticha used the inner product \[ &lt; x|x_0 &gt; \,\] to represent implication.
I read Ariels arxiv papers on his attempt to "derive" the born postulate from a number of premises and I think it's what yuo mean. I like some of Ariels things, but overall I have a similar critique to him as to you.

His premises certainly does not follow deductively from an obvious starting point. They are possible but not unique starting points, and by no means plausible. If I remember corrently one key choice he makes is to introduce a complex number.

To me, his choices are motivated by what he is trying to prove. So it's a possible reconstruction, but that IMO is no more plausible than the normal axiomatic introduction of QM.

friend said:
In any event, the next step is to replace the dirac delta with a function. Most QM books go through a derivation using the schroedinger equation. I make the replacement directly, which is what you seem to be having trouble with. Wouldn't it be advantageous to derive the path integral from first principles and then derive schroedinger's equation from that instead of only the other way around? And don't you find it the least bit curious that replacing the delta with a gaussian gives you something that looks like QM?

I think I need to understand your overall strategy before I can comment on details, to be constructive.

(a) The QM formalism has several parts, first it's the general structure - hilbert spaces, operators, hamiltonians, projection postulates etc and how that supposedly connects to the notion of experimental probability.

(b) The other part is the choice of hamiltonian for a given system.

Which part are you trying to addressing and "replace from logic"?

friend said:
Wouldn't it be advantageous to derive the path integral from first principles and then derive schroedinger's equation from that instead of only the other way around?

If you can infer the hamiltonian for a system, from some constraining principles of the formalism, yes that would definitely be interesting. But then I think the constraining principles must be unique. Consistency isn't enough if you claim that you are deducing this. Othrewise it's like a reconstruction, designed to prove something we already "know".

I can appreciate your reflections but my personal feedback is that so far I find the reasoning to be weak.

... more later...

/Fredrik
 
  • #21
friend said:
Does it make sense to say that chosing a particular inner product
(or dirac delta) must be a completely random process,
that this means that the inner product must itself represent a random process
(gaussian delta function)? The inner product is a combination of two things.
So this is like asking if taking a combination of any two is a completely
random process, then that combination represents a completely random process.

At least I can say that the reasoning is fuzzy to me. In particular the physics connection. What is the physical correspondence to "choosing an inner product" you imagine? but maybe because I don't completely understand your overall strategy and ambition. When I understand that perhaps I could provide a more useful feedback on this part.

.. more later

/Fredrik
 
  • #22
It seems you are looking for a motivation for the variational method from deeper principles than what we usually see in standard textbooks? If so, I share a motivation for that overall quest.


friend said:
But the main question of this thread is that given a path integral, does the least action principle follow from that just by taking the first variation of the integration of the path integral.

With "given a path integral", do you actually mean, given a solution to the dynamical equations? or what do you mean?

If so, IMO the "variational problem" is not unique in the sense that it can be uniquely deduced as "the only problem that has that solution". The premise (the solution) is however consistent with a choice of variational problem that exists, but this is the standard motivation I think.

I think if we are to look for a first principle exposition of the least action principle, we need to give some physical motivation for all the baggage in that abstraction. State spaces, actions etc.

I see that you make partially interesting reflections with formulas but I do not quite understand where your _physical_ first principles are? And I have a strong feeling that your first principles may not be constructed by a deductive reasoning. Thus they are somewhat ambigous.

friend said:
I was trying to justify lagrangian mechanic from first principles. This is where the cannonical version of QM comes from, right?

The classical model (hamiltonian or lagrangian) is part of this, but another part is the general framework. State spaces etc.

I personally try to see the "information physics" in each mathematical entitiy:
What is the physical correspondence to mathematical representation?
What is the physical correspondence to "calculating a probability"?
What is the physical correspondence to "action"?

For these reasons I have large difficulty to just accept infinities and continuums as if it's physical significance was clear. In these details lies also the root of my scepsis about your arbitrary flipping between limits, non-limits and back again. I think you are ignoring some physics in that.

I'm not saying there isn't anything to your ideas, I just find the explanations you make unclear. In particular since your ambition is to prove this from logic. A deduction should not be ambigous or uncertain. To construct a consistent logical system, is itself not necessarily a deductive process. This is where the non-predictive evolution comes in, in my perspecive. I'm trying to understand how you can insist on your physics from logic, regarding this "holes" I see.

You seem to hint that a proper understanding on the transition amplitude calculation, could also put constraints on what langrangians that are "possible" or consistent with the framework. This is paradoxally a similar vision I have, by my idea of arriving at it, and make progress is very different. you seem to want to deduce the structure of the lagrangian from logic? I want to infer the structure for interaction histories. Ie. the structure of the lagrangian is emergent, in an evolutionary process. In this way I can not get information from nothing, but it might allow me as a human to a much deeper understanding of the formalism, and make me ask questions that makes my own search converge faster.

/Fredrik
 
  • #23
It seems you are looking for a motivation for the variational method from deeper principles than what we usually see in standard textbooks? If so, I share a motivation for that overall quest.


friend said:
But the main question of this thread is that given a path integral, does the least action principle follow from that just by taking the first variation of the integration of the path integral.

With "given a path integral", do you actually mean, given a solution to the dynamical equations? or what do you mean?

If so, IMO the "variational problem" is not unique in the sense that it can be uniquely deduced as "the only problem that has that solution". The premise (the solution) is however consistent with a choice of variational problem that exists, but this is the standard motivation I think.

I think if we are to look for a first principle exposition of the least action principle, we need to give some physical motivation for all the baggage in that abstraction. State spaces, actions etc.

I see that you make partially interesting reflections with formulas but I do not quite understand where your _physical_ first principles are? And I have a strong feeling that your first principles may not be constructed by a deductive reasoning. Thus they are somewhat ambigous.

friend said:
I was trying to justify lagrangian mechanic from first principles. This is where the cannonical version of QM comes from, right?

The classical model (hamiltonian or lagrangian) is part of this, but another part is the general framework. State spaces etc.

I personally try to see the "information physics" in each mathematical entitiy:
What is the physical correspondence to mathematical representation?
What is the physical correspondence to "calculating a probability"?
What is the physical correspondence to "action"?

For these reasons I have large difficulty to just accept infinities and continuums as if it's physical significance was clear. In these details lies also the root of my scepsis about your arbitrary flipping between limits, non-limits and back again. I think you are ignoring some physics in that.

I'm not saying there isn't anything to your ideas, I just find the explanations you make unclear. In particular since your ambition is to prove this from logic. A deduction should not be ambigous or uncertain. To construct a consistent logical system, is itself not necessarily a deductive process. This is where the non-predictive evolution comes in, in my perspecive. I'm trying to understand how you can insist on your physics from logic, regarding this "holes" I see.

You seem to hint that a proper understanding on the transition amplitude calculation, could also put constraints on what langrangians that are "possible" or consistent with the framework. This is paradoxally a similar vision I have, by my idea of arriving at it, and make progress is very different. you seem to want to deduce the structure of the lagrangian from logic? I want to infer the structure for interaction histories. Ie. the structure of the lagrangian is emergent, in an evolutionary process. In this way I can not get information from nothing, but it might allow me as a human to a much deeper understanding of the formalism, and make me ask questions that makes my own search converge faster.

/Fredrik
 
  • #24
Fra said:
At least I can say that the reasoning is fuzzy to me. In particular the physics connection. What is the physical correspondence to "choosing an inner product" you imagine?

Yes, it seems a little fuzzy to me too at this point. But what I think I meant was if the formalism includes each product with equal weight, then singling out anyone for special consideration must have been an arbitrary and thus totally random choice. Does this mean that the inner product <xi|xj>, which is a distribution, must be represented by a random process itself, by a gaussian distribution? What else could <xi|xj> mean if not the probability of obtaining it? That's how we are accustomed to thinking about <xi|xj> in QM. But more generally, is that how we should consider and inner product, as a amplitude or probability of obtaining it? You're all welcome to correct me if I'm wrong, but it seems to me that <xi|xj> has no meaning apart from the context of all the other |xn> that |xi> and |xj> are a part of. In fact it seems necessary to consider all the rest of the |xn> in order to even calculate <xi|xj>. For we have that <xi|xj> = 0 if i not equal j and <xi|xj>=1 if i=j. So it seems that the evaluation of <xi|xj> depends on the context of choicing it from all the other possible <xi|xj>. Any further thoughts out there?
 
Last edited:
  • #25
Fra said:
I think I need to understand your overall strategy before I can comment on details, to be constructive.

(a) The QM formalism has several parts, first it's the general structure - hilbert spaces, operators, hamiltonians, projection postulates etc and how that supposedly connects to the notion of experimental probability.

(b) The other part is the choice of hamiltonian for a given system.

Which part are you trying to addressing and "replace from logic"?

Isn't the goal of physics to provide a complete understanding of the laws of nature? Will physics be complete if we find even smaller particles? What if we prove that we have found the smallest possible particle? It would still need to be explained where particles or fields came from to begin with, or where spacetime came from to begin with. I don't think the questions will stop until we start explaining physics in terms of general principles of reason, until we derive physics from logic. Then the only thing to question will be reason itself. Yes, it is considered somewhat of a mystery how the universe can be comprehensible. And we will continue to probe until we are questioning comprehension itself.

So I have a philosophical interest in physics. It seems vain to me to try to explain anything with reason if reason really has nothing to do with what is physical. So my goal would be to derive physics from logic. And I think I've made some progress.

I've already shown how Feynman's path integral can be derived from the identity operator, or equivalently from the dirac delta function (See post 5 and 18 in this thread). And having obtained a path integral on principle alone, I'm now trying to see how other physical principles can be derived. To that end I ask the question of this thread, can the least action princple be derived on principle, by taking the first variation of the integration of the path integral which must of necessity be zero.

I don't want to get into an argument in this thread about how exactly physics can be derived from logic because that may be considered too speculative at this point for these forums. For the purposes of this thread let's assume the path integral can be derived as a purely mathematical construct apart from physics and see if it implied the least action principle. Thank you.
 
  • #26
hmm i see a problem with \langle x&#039; \mid x \rangle = \delta (x&#039;-x). that's not the case in Heisenberg picture. And i think the pathintegral equates to the Heisenberg picture, that means:
\langle x&#039;,t&#039; \mid x,t \rangle_{H}=G(x&#039;,t&#039;,x,t)=\int e^{i/h*S[x(t)]}D x(t). with G the propagator or greenfunction. and why should one make a functional deriviative of a propagator (but generating functional)?
 
Last edited:
  • #27
friend said:
For the purposes of this thread let's assume the path integral can be derived as a purely mathematical construct apart from physics and see if it implied the least action principle. Thank you.

Fair enough. I didn't want to change focus here, just wanted to give you some feedback.

So if I just play along with what I think you are doing:

If we simply take as a natural assumption that the transition amplitude is maximizes with respect to the "path", then your "inner product" is varied with respect to the path of choice. Then given your expression that should imply for a non-classical path, not that the variation of the action of a specific path is zero, but that the expectation value of the variation is zero. That seems to follow roughly straightforward.

ie. \delta P(final|initial) = 0 which seems reasonable enough implies as per your path integral \langle final|\delta S|initial \rangle = 0

So the way I see it, the least action principle follows if you look for the "most probable path" given the constraints at hand. One might even argue that this follows from the "definition" of probability. Is this good enough or were you having something else in mind?

But I think the path ingeral already contains the critical parts?

Edit: I was too fast, it's not a proper implication anway. Just a possibility.There are probably a wider class of paths that are also consistent with the premise.

But in either case, to instead consider the more fundamental principle to be variation of the transition probability to be zero is as I see it more fundamental than the original least ACTION principle - in this sense, the action and feynmann stuff is one possible way to compute this. But there might be other ways to do it too, that leads to the same transition probability.

(It's how I personally treat it in my view: I don't defined a regular action, I just manipulate transition probabilities directly).

Well this probably didn't help much but anways :)

/Fredrik
 
Last edited:
  • #28
OK, so... back to the problem at hand, how to take the first variation of the integration of the path integral.

I get a little bit confused because you can write the path integral as,

\[<br /> \delta \left( {x - x_0 } \right) = \int {d\left[ x \right]} \cdot e^{ - {\textstyle{i \over \hbar }}\int_0^t {L\left[ {x,\dot x,t} \right]\,\,dt} } <br /> \]<br />

in which case the x(t) and x'(t) are functions, not variables. But to integrate the path integral one more time, you must integrate with respect to a variable, the variable of integration. So it's not obvious how to integrate the path integral in this form. The x in the delta on the left is a variable, but the x in the path integral on the right is a function.

But when the path integral above was developed from the dirac delta function,

\[<br /> {\rm{\delta (x - x}}_0 ) = \int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {...\int_{ - \infty }^{ + \infty } {{\rm{\delta (x - x}}_n {\rm{)\delta (x}}_n {\rm{ - x}}_{n - 1} {\rm{)}}...{\rm{\delta (x}}_1 {\rm{ - x}}_0 ){\rm{dx}}_n {\rm{dx}}_{n - 1} ...{\rm{dx}}_1 } } } <br /> \]<br />

and substitute the gaussian form of the delta,

\[<br /> {\rm{\delta (x - x}}_0 ) = \mathop {\lim }\limits_{\Delta \to 0} \frac{1}{{(\pi \Delta ^2 )^{1/2} }}e^{ - (x - x_0 )^2 /\Delta ^2 } <br /> \]<br />

with,

\[<br /> \Delta = [\frac{{2i\hbar (t - t_0 )}}{m}]^{1/2} <br /> \]<br />

the path integral becomes,

\[<br /> \delta \left( {x - x_0 } \right) = \int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {...\int_{ - \infty }^{ + \infty } {({\textstyle{m \over {2\pi i\hbar \Delta t}}})^{n/2} {\mathop{\rm e}\nolimits} ^{(im(\frac{{x - x_n }}{{t - t_n }})^2 \Delta t/2\hbar )} \cdot {\mathop{\rm e}\nolimits} ^{(im(\frac{{x_n - x_{n - 1} }}{{t_n - t_{n - 1} }})^2 \Delta t/2\hbar )} \cdot ...{\rm{ }} \cdot {\mathop{\rm e}\nolimits} ^{(im(\frac{{x_1 - x_0 }}{{t_1 - t_0 }})^2 \Delta t/2\hbar )} {\rm{dx}}_n {\rm{dx}}_{n - 1} ...{\rm{dx}}_1 } } } <br /> \]<br /> <br />

Notice that the x variable of the delta on the left is in the exponent of the first exponential on the right. I suppose this makes for an interesting integration process.

But once the integration wrt x is done, in order to get 1 since that's integrating a delta function, then the question becomes how to take the first variation. Variations are done with respect to functions, although you can take the variation of a variable since \[<br /> \delta x = dx<br /> \]<br />.

Do we functionally differentiate

\[<br /> \delta \int_{ - \infty }^{ + \infty } {\left( {\int {d\left[ x \right]} \cdot e^{ - {\textstyle{i \over \hbar }}\int_0^t {{\textstyle{1 \over 2}}m\left( {\dot x} \right)^2 \,\,dt} } } \right)dx} <br /> \]<br /> <br />

with respect to x or x'? Or is it easier to functionally differentiate,

\[<br /> \delta \int_{ - \infty }^{ + \infty } {\left( {\int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {...\int_{ - \infty }^{ + \infty } {({\textstyle{m \over {2\pi i\hbar \Delta t}}})^{n/2} {\mathop{\rm e}\nolimits} ^{(im(\frac{{x - x_n }}{{t - t_n }})^2 \Delta t/2\hbar )} \cdot {\mathop{\rm e}\nolimits} ^{(im(\frac{{x_n - x_{n - 1} }}{{t_n - t_{n - 1} }})^2 \Delta t/2\hbar )} \cdot ...{\rm{ }} \cdot {\mathop{\rm e}\nolimits} ^{(im(\frac{{x_1 - x_0 }}{{t_1 - t_0 }})^2 \Delta t/2\hbar )} {\rm{dx}}_n {\rm{dx}}_{n - 1} ...{\rm{dx}}_1 } } } } \right)dx} <br /> \]<br />

with respect to each of the xi, which is more easily written,

\[<br /> \delta \int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {\int_{ - \infty }^{ + \infty } {...\int_{ - \infty }^{ + \infty } {({\textstyle{m \over {2\pi i\hbar \Delta t}}})^{n/2} {\mathop{\rm e}\nolimits} ^{\sum\nolimits_{i = 0}^{n} {(im(\frac{{x_{i + 1} - x_i }}{{t_{i + 1} - t_i }})^2 \Delta t/2\hbar )} } {\rm{dx}}_{n + 1} {\rm{dx}}_n ...{\rm{dx}}_1 } } } } <br /> \]<br />

It may not matter since the variation commutes with all the integration signs and pass to the exponential in both cases.

So what is the variation of a composite functional? If \[<br /> F\left[ {y\left( x \right),z\left( x \right)} \right] = F\left[ {S\left[ {y\left( x \right),z\left( x \right)} \right]} \right]<br /> \]<br />, then what is \[<br /> \delta F<br /> \]<br />?

Remember from post 10,

\[<br /> \delta F = \frac{{\partial F}}{{\partial y}}\delta y + \frac{{\partial F}}{{\partial z}}\delta z<br /> \]<br />

But we know from calculus that,

\[<br /> \frac{{\partial F}}{{\partial y}} = \frac{{dF}}{{dS}} \cdot \frac{{\partial S}}{{\partial y}}<br /> \]<br /> and \[<br /> \frac{{\partial F}}{{\partial z}} = \frac{{dF}}{{dS}} \cdot \frac{{\partial S}}{{\partial z}}<br /> \]<br />

So that,

\[<br /> \delta F = \frac{{dF}}{{dS}} \cdot \frac{{\partial S}}{{\partial y}}\delta y + \frac{{dF}}{{dS}} \cdot \frac{{\partial S}}{{\partial z}}\delta z<br /> \]<br />

\[<br /> \delta F = \frac{{dF}}{{dS}}\left( {\frac{{\partial S}}{{\partial y}}\delta y + \frac{{\partial S}}{{\partial z}}\delta z} \right) = \frac{{dF}}{{dS}} \cdot \delta S<br /> \]<br />

Or,

\[<br /> \delta F = \frac{{dF}}{{dS}} \cdot \delta S<br /> \]<br />

So since the variation passes into the exponential, we are looking at deriving

\[<br /> \delta e^{S\left[ {y,z} \right]} <br /> \]<br />

But from the above, this is just,

\[<br /> \delta e^S = e^S \cdot \delta S<br /> \]<br />

whether S is the integral or the discrete summation. In the integral version you'd have terms like, \[<br /> \delta x\,\,\,and\,\,\,\delta \dot x<br /> \]<br />. But in the discrete form you'd get terms like, \[<br /> \delta x_1 ,\,\,\delta x_2 ,\,\,\,\delta x_3 <br /> \]<br />, ... I think the difference can be explained by the understanding that \[<br /> x = \left( {x_1 ,x_2 ,x_3 ,...,x_n } \right)<br /> \]<br />. And I think this means that \[<br /> \delta x = \left( {\delta x_1 ,\delta x_2 ,\delta x_3 ,...,\delta x_n } \right)<br /> \]<br />. Does this sound right? Or maybe you can differentiate with respect to \[<br /> x_i <br /> \]<br /> in general and get terms like \[<br /> \delta x_i <br /> \]<br />

Either way the action integral must have a variation of zero in order for the first functional derivative of the integral of the path integral to be zero as it must. I think this is where the least action principle comes from.
 
Last edited:
Back
Top