Calculus of variations and integrands containing second derivatives

In summary: Euler-Lagrange equations for the metric connection.In summary, it is possible to solve the calculus of variations problem when L is a function of y'' but you will have to make the additional constraint that the variation at the end points of y'' also vanishes. Additionally, you will end up with a 3rd order differential equation instead of a second order ODE, which requires usually more initial data to solve than we have.
  • #1
ShayanJ
Insights Author
Gold Member
2,810
604
You know that the problem of calculus of variations is finding a y(x) for which [itex] \int_a^b L(x,y,y') dx [/itex] is stationary. I want to know is it possible to solve this problem when L is a function of also y'' ?
It happens e.g. in the variational method in quantum mechanics where we say that choosing any arbitrary wave function [itex] \psi(x) [/itex], the energy of the ground state of the system is always smaller than [itex] \int \psi^* H \psi dx [/itex]. But if we can find a [itex] \psi [/itex] for which this integral is stationary, then we have the wave function and also the energy of the ground state. But the problem is, for ordinary particle systems in non-relativistic quantum mechanics, H contains a second derivative.
 
Physics news on Phys.org
  • #2
You can certainly do it with a function of y''. However, you will have to integrate by parts twice, so you need to make the additional constraint that the variation at the end points of y'' also vanishes.

In addition, you will end up with probably a 3rd order differential equation instead of a second order ODE, which requires usually more initial data to solve than we have. We usually just have 2 initial data points in classical physics, the initial position and momentum. A 3rd order ODE would require an additional point.
 
  • #3
You mean variations of y' vanish at the boundary coming about from integration by parts? So that if

[itex]\delta y = \delta y' = 0[/itex]

at the boundaries the Euler-Lagrange equations then become:

[itex]{\partial L \over \partial y} - {d \over dx} \Big( {\partial L \over \partial y_x} \Big) +
{d^2 \over dx^2} \Big( {\partial L \over \partial y_{xx}} \Big) = 0.[/itex]

The Einstein-Hilbert action for general relativity is an example of an action principle that involves 2nd order derivatives of the basic variables i.e. the metric tensor. Under variations of the metric which vanish and who's normal derivatives also vanish on the boundary, the boundary of a compact region, this action is stationary if and only if the metric satisfies Einstein's equations - the Euler Lagrange equations are

[itex]{\partial L \over \partial g_{ab}} - {\partial \over \partial x^c} \Big( {\partial L \over
\partial g_{ab,c}} \Big) + {\partial^2 \over \partial x^c \partial x^d} \Big( {\partial^2 L \over
\partial g_{ab,cd}} \Big) = 0[/itex]

which gives

[itex]G_{ab} = \kappa T_{ab}[/itex]

(which turn out to be second order differential equations). If the normal derivatives don't vanish you have to add a boundary term to the action principle to get out the right equations of motion in the bulk (this term is important in quantum gravity).
 
Last edited:
  • #4
It is possible I recalled some of the discussion on higher derivative terms incorrectly. I apologize if I did. I do recall that a higher derivative term will require more initial conditions to specify the equations of motion than the regular term we are used to though.
 
  • #5
No problem. You will have an extra term

[itex]\int_{x_1}^{x_2} dx {\partial L \over \partial y_{xx}} {d^2 \delta y \over dx^2}[/itex]

which upon integration by parts gives

[itex]\Big[ {d \delta y \over dx} {\partial L \over \partial y_{xx}} \Big]_{x_1}^{x_2} - \int_{x_1}^{x_2} dx {d \over dx} \Big( {\partial L \over \partial y_{xx}} \Big) {d \delta y \over dx}[/itex]
= [itex]\Big[ {d \delta y \over dx} {\partial L \over \partial y_{xx}} \Big]_{x_1}^{x_2} -
\Big[ \delta y {d \over dx} \Big( {\partial L \over \partial y_{xx}} \Big) \Big]_{x_1}^{x_2}
+ \int_{x_1}^{x_2} dx {d^2 \over dx^2} \Big( {\partial L \over \partial y_{xx}} \Big) \delta y (x)[/itex]

so if [itex]\delta y = \delta y' = 0[/itex] at the boundaries we are left with an extra term to the Euler-Lagrange equations

[itex]+{d^2 \over dx^2} \Big( {\partial L \over \partial y_{xx}} \Big)[/itex]

In non-relativistic QM (as with any hyperbolic differential equations) it is the order of the time-derivative that dictates what initial conditions need to be specified. Shrodinger's equation is first order in time derivatives so you need only specify the state of the system a moment. It is in relativistic QMs, in particular the Klein-Gordon equation, that you encounter a differential equation of second order in time derivatives. This leads to the requirement of specifying [itex]\phi(t_0)[/itex] and [itex]{\partial \over \partial t} \phi(t_0)[/itex], which in turn leads to the well known problem of negative probability density! This, initially alarming problem, led Dirac to formulate a first order in time relativistic equation (and first order in spatial derivatives - so they are on an equal footing).

In GR it happens to turn out to be a second order in derivatives problem, higher order derivatives don't appear. The calculation using the Euler-Lagrange equations is horrendous, apparently. There is what is called a first order formulism where the metric and connection are taken to be independent variables and the connection appears only with first order derivatives making it much easier. The Euler-Lagrange equations for the connection give the connection as the metric connection. And the Euler-Lagrange equations for the (densitized) contravariant metric trivially give Einstein's equations.
 
Last edited:
  • #6
OK, thanks.
But now I want to find a [itex] \psi [/itex] for which [itex] \int \psi^*H\psi dx [/itex] is stationary.
For one particle in potential [itex] V(x) [/itex], [itex] H=-\frac{\hbar^2}{2m}\frac{d^2}{dx^2}+V(x)=-a \partial^2+V [/itex], so we have [itex] L=-a\psi^*\partial^2\psi+\psi^* V \psi [/itex] and using the Euler-Lagrange-(julian:biggrin:) equation:

[itex]

V\psi^*-a\psi^*=0 \Rightarrow \psi^*=0 \Rightarrow \psi=0

[/itex]!

I think its not that easy to say [itex] \delta y'=0 [/itex]!
 
  • #7
It isn't as simple as setting the variation equal to zero...there is the constraint that [itex]\int |\psi (x)|^2 dx = 1[/itex] which means you have to introduce Lagrangian multipliers...
 
  • #8
You then have [itex]{\delta \over \delta \psi^*} \int \psi^* H \psi dx = \lambda \psi[/itex] where [itex]\lambda[/itex] is a Lagrange multiplier. Which is basically

[itex]H \psi = E \psi[/itex]
 
Last edited:
  • #9
Shyan said:
OK, thanks.
But now I want to find a [itex] \psi [/itex] for which [itex] \int \psi^*H\psi dx [/itex] is stationary.
For one particle in potential [itex] V(x) [/itex], [itex] H=-\frac{\hbar^2}{2m}\frac{d^2}{dx^2}+V(x)=-a \partial^2+V [/itex], so we have [itex] L=-a\psi^*\partial^2\psi+\psi^* V \psi [/itex] and using the Euler-Lagrange-(julian:biggrin:) equation:

[itex]

V\psi^*-a\psi^*=0 \Rightarrow \psi^*=0 \Rightarrow \psi=0

[/itex]!

I think its not that easy to say [itex] \delta y'=0 [/itex]!

Hang on a moment...shouldn't you get a term:

[itex]{d^2 \over dx^2} \Big( {\partial \over \partial \psi_{xx}} (-a \psi^* \psi_{xx}) \Big) = -a {d^2 \over dx^2} \psi^*[/itex]?

And oh I should have been careful about the difference between initial conditions and boundary conditions on the wavefunction at spatial locations. It is the latter that is important here.
 
  • #10
julian said:
Hang on a moment...shouldn't you get a term:

[itex]{d^2 \over dx^2} \Big( {\partial \over \partial \psi_{xx}} (-a \psi^* \psi_{xx}) \Big) = -a {d^2 \over dx^2} \psi^*[/itex]?

And oh I should have been careful about the difference between initial conditions and boundary conditions on the wavefunction at spatial locations. It is the latter that is important here.

Yeah...that's right. I don't know why I did that mistake. Actually when I did the calculations, I wondered where is the [itex] E \psi [/itex] but I posted that when I was busy a little and so the statement of my problem changed to that wrong thing! But that problem is solved now. Thanks

And about [itex] \delta y'=0 [/itex]. I still have some doubt about it. Here we're actually excluding variations for which [itex] \delta y' \neq 0 [/itex]. It means we know the value of wavefunction's derivative at the boundaries. We also assume [itex] \delta y=0 [/itex] which means we know the value of wavefunction itself at the boundaries. So we're assuming that we have four boundary conditions where we need only two!
Its strange that its working. What's the meaning of this?
 
  • #11
A vaguely remember from undergrad that you solved Schrodinger's equation in separate regions and then had to match up solutions by both equating wavefunction at the boundary as well sometimes matching the derivative of the wavefunction at the boundary.

There is a more subtle issue, which as far as I know has only been brought up in what is called background-independent scattering amplitudes of loop quantum gravity, where we wish QM's equations to hold in a finite region but where we specify boundary data as coming from the experimental set-up. These can be motivated by: "finite closed boundaries represent the way real experiments are set up more directly than constant-time surfaces. A realistic experiment is confined to a finite region of spacetime." Anyway, this is called the "general boundary" approach. This seems exclusively targeted at QFT. Would be interesting to see how this can work for non-relativistic QM.
 

1. What is the purpose of the calculus of variations?

The calculus of variations is a mathematical theory that allows us to find the optimal solution to a problem by minimizing or maximizing a certain functional. It is particularly useful in physics and engineering, where we want to find the path or function that will minimize the energy or cost of a system.

2. What is an integrand containing second derivatives?

An integrand containing second derivatives is a function that involves the second derivative of a variable with respect to another variable. It can be written in the form of ∫F(x,y,y')dx, where y' is the first derivative of y with respect to x and y'' is the second derivative.

3. How is the Euler-Lagrange equation used in the calculus of variations?

The Euler-Lagrange equation is a necessary condition for finding the extremum of a functional in the calculus of variations. It is used to derive the differential equation that the optimal function must satisfy, and thus helps us find the solution to the problem.

4. Can the calculus of variations be applied to real-life problems?

Yes, the calculus of variations has a wide range of applications in various fields such as physics, engineering, economics, and biology. It can be used to solve problems related to minimizing energy, finding optimal control strategies, and optimizing resource allocation, among others.

5. Are there any limitations to the calculus of variations?

One limitation of the calculus of variations is that it assumes the existence of an extremum for the functional being optimized. In some cases, there may not be a unique solution, or the solution may not exist at all. Additionally, the calculus of variations can be quite complex and difficult to apply in certain situations, requiring advanced mathematical techniques and computer algorithms.

Similar threads

Replies
12
Views
1K
Replies
1
Views
936
Replies
9
Views
1K
Replies
3
Views
1K
Replies
2
Views
1K
Replies
22
Views
454
Replies
6
Views
1K
Replies
20
Views
2K
Replies
4
Views
347
  • Calculus
Replies
14
Views
1K
Back
Top