A Constrained variational problem

  • A
  • Thread starter Thread starter member 428835
  • Start date Start date
member 428835
Hi PF!

I'm trying to show that the eigenvalue problem $$L u = \lambda M u$$ is equivalent to solving $$\min_\phi (L\phi,\phi) : (M\phi,\phi) = 1$$
where ##\phi## is a real function of ##x## and ##L,M## are Hermitian operators and ##\lambda## is the Lagrange multiplier constant.

Applying Lagrange multipliers to the constrained problem yields a functional
$$ J = (L\phi,\phi) - \lambda\left[ (M\phi,\phi) - 1\right] \implies\\
\delta J = \delta (L\phi,\phi) - \delta\left[\lambda ((M\phi,\phi) - 1)\right]\\
= 2(L\phi,\delta\phi) - \delta\lambda ((M\phi,\phi) - 1) - 2\lambda(M\phi,\delta\phi)\\
=2(L\phi-\lambda M\phi,\delta\phi) - \delta\lambda ((M\phi,\phi) - 1) = 0.
$$
Now I know ##\delta\lambda ((M\phi,\phi) - 1) = 0## implies ##L u = \lambda M u##, but why is ##\delta\lambda ((M\phi,\phi) - 1) = 0##? Am I missing something? I think ##\delta \lambda = 0## (since ##\lambda## is a constant). What do you think?

Also, how is the operator ##\delta## above defined? I've been treating it as a derivative, but what's the formal definition? I've read several different websites now but can't find a direct definition.
 
Physics news on Phys.org
That second variation term recovers your original constraint.
\delta\lambda((M\phi,\phi)-1)=0 \implies (M\phi,\phi)-1 = 0
since \delta\lambda is arbitrary.
 
  • Like
Likes member 428835
As to your second question: (prepare for long exposition):

The variation operator \delta is a differential operator just like d in the sense that df(x)=f'(x)dx. However it is a differential on a functional which is itself a function with domain in a function space. As such there are two levels of differentiation and you have to distinguish them.

Consider a scalar function f and its "graph" i.e. it's use to relate two coordinate variables: y=f(x). The differential of a function emerges as a local linearization of this relation:
dy = d f(x) \equiv f'(x)dx
View this in terms of x and y being fixed and you're referencing a second point in the xy plane at (x+dx,y+dy). This is the "modern" (non-infinitesimal) interpretation of differential variables.

Now consider the same for a vector function, allow x and y to be vectors in their own spaces. The same relationship holds excepting that dx and dy are now vectors and the derivative is now an Operator Valued function of the original variables. I will often write the relation in this form:
d\mathbf{y} = F'(\mathbf{x})[d\mathbf{x}]
with the brackets indicating the operation of a linear operator. If you express your vectors as column matrices then the general derivative F' will take the form of a matrix of partial derivatives with respect to the components. For the specific case where F is a coordiant transformation its derivative is the Jacobian matrix... the matrix whose determinant is the Jacobian.

Ok this is conventional differentials. Now for the next stage. Let F be a functional, a function from a space of functions to \mathbb{R}. Now I'll use curly brackets to indicate functional evaluation of a function and let \phi be my archtypical function. For:
\psi = F\{\phi\}
the differential relation (expressing a local linear approximation) is:
\delta F\{\phi\} = F'\{\phi\}[\delta\phi]
Note that just as dx and dy were new independent variables expressing deviations from the original variables x,y so too now are \delta\psi and \delta\phi new indepent variables (in function space) expressing deviations from the original (function valued) variables.

But those variables as function valued variables are subject to the original differential operator d. So d: \phi(t) \mapsto d\phi(t,dt) = \phi'(t)dt is a distinct and still present mathematical operation and must be distinguished from \delta\phi.

I hope this makes some sense of it. Look up Gâteaux derivative which generalizes nicely the concept of differential and derivative to mutivariable and functional spaces.
In short define the differential before the derivative and then the derivative as the relationship between differential variables:
d F(X) \equiv \lim_{h\to 0} \frac{1}{h}\left[ F(X+hdX) - F(X)\right] \equiv F'(X)[dX]
Its relatively simple to proved the limit of the generalized difference quotient is a linear function(al) of the differential variable dX. All you need otherwise is that the domain is a linear space be it scalars, vectors, functions or something more.

This is something I've only come to understand fully in recent years after many years teaching mutivar calculus.
 
  • Like
Likes member 428835 and jim mcnamara
One more point. As this functional differential is just a conventional differential on a function space it obeys the same rules. It commutes with linear operators:
\delta L[\phi] = L[\delta\phi] and more generally obeys the Leibniz rule for multilinear forms:
\delta F[\phi_1,\phi_2,\phi_3,\ldots] = F[\delta \phi_1,\phi_2,\phi_3,\ldots] + F[\phi_1,\delta \phi_2,\phi_3,\ldots] + \cdots

This is how you evaluated the differential of your inner product. Note that since an definite integral is a linear functional on its argument you get likewise:
\delta \int_{\Omega}Fdx = \int_{\Omega} \delta F dx

Oh and since d is linear \delta and d commute:
\delta dF = d\delta F
That's a crucial step in deriving Euler-Lagrange equations as you may recall.
 
  • Like
Likes member 428835
And yet one final point (really, this time. I promise!).

There's a slight difficulty with your method in that, as I said \delta is a differential of functions on function space. Now you're also applying it to \lambda. If you are treating \lambda as a constant or as an independent variable which your functions in your function space may depend upon then \delta \lambda = 0. But if instead (and this is I think the proper case) your Lagrange multiplier is a scalar dependent variable, with dependence in the form of some (unknown) functional of your function then \delta \lambda is properly an independent variable which allows you to recover your constraint.

This is one of those nit-picking details that doesn't really affect calculations but can be the Achilles heel of an attempted proof allowing you to miss a pathological counter example.
 
  • Like
Likes member 428835
Back
Top