Proof of Lagrange multipliers method in calculus of variations.

Click For Summary

Discussion Overview

The discussion revolves around the application of Lagrange multipliers in the context of calculus of variations, particularly focusing on integral constraints. Participants explore the theoretical underpinnings and practical implications of using Lagrange multipliers when dealing with constraints in optimization problems.

Discussion Character

  • Exploratory
  • Technical explanation
  • Conceptual clarification
  • Debate/contested

Main Points Raised

  • One participant expresses confusion about the introduction of Lagrange multipliers in calculus of variations, questioning why the Euler-Lagrange equations must be multiples of each other rather than both being zero.
  • Another participant explains the geometric interpretation of gradients and constraints, using the example of minimizing distance to a point while constrained to a curve, emphasizing the relationship between the gradients of the distance function and the constraint function.
  • A different participant raises a question about applying the concept when the constraint is a line segment of a specific length, seeking clarity on how to maximize an integral under such conditions.
  • One participant shares a link to a proof and questions the necessity of a nonzero determinant and the conditions under which points satisfying the constraint can yield better integral values.

Areas of Agreement / Disagreement

Participants do not reach a consensus, as there are multiple competing views and ongoing questions regarding the application of Lagrange multipliers in various contexts within calculus of variations.

Contextual Notes

Participants highlight limitations in understanding the conditions under which Lagrange multipliers are applicable, particularly regarding the nature of constraints and the implications of determinant conditions in proofs.

chingel
Messages
307
Reaction score
23
I have been reading a little about calculus of variations. I understand the basic method and it's proof. I also understand Lagrange multipliers with regular functions, ie since you are moving orthogonal to one gradient due to the constraint, unless you are also moving orthogonal to the other gradient you have some component of movement along the gradient and it will change the value of the function, so the gradients have to be multiples of each other.

But when reading about the use of Lagrange multipliers in calculus of variations with an integral constraint, I haven't found an explanation which I could understand. If anyone has a good link or an explanation I would be glad.

For example, when I read this one:

https://netfiles.uiuc.edu/liberzon/www/teaching/cvoc/node38.html

Why are the Lagrange multipliers introduced? Both of the integrals have to be zero, when n is such a function that preserves the constraint, but why does the Euler-Lagrange equation of one have to be a multiple of the other? Wouldn't they have to be both zero, or what is going on?
 
Last edited by a moderator:
Physics news on Phys.org
This depends upon the fact that if g(x,y)= constant is a curve in the xy-plane (or g(x,y,z)= constant) then \nabla g is perpendicular to the curve (or plane).

Let's take a simple example: Find the point on the curve y= f(x) closest to (x_0, y_0). The function to be minimized is the distance function D= \sqrt{(x- x_0)^2+ (y- y_0)^2}. (Yes, it would be simpler to minimize "distance squared" and give the same answer but I want to be as straight forward as possible.)

It's not difficult to see that the "gradient" vector of the distance function just points directly from (x, y) to (x_0, y_0). With no constraint, we could just move in the direction of the gradient vector, eventually getting to (x_0, y_0), the point that obviously makes the distance minimum (in fact, 0).

But if we are restricted to stay on curve y= f(x), we cannot move directly toward the point. What we can do is move along the curve- and an obvious strategy is to look at the projection of the gradient onto the curve. If the gradient is "to the right", that is, if its projection onto the curve points to the right, move to the right. If the gradient is "to the left", move to the left. We can keep doing that, and getting closer to (x_0, y_0), as long as the gradient has such a projection- as long as the gradient is not perpendicular to the constraint curve. And we can write y= f(x) as g(x,y)= y- f(x)= 0: a "level curve". Finally, we know that the gradient of such a function is perpendicular to any level curve: that is, \nabla g is perpendicular to g(x,y)= 0 at every point on the curve.

Putting those together, \nabla D must be perpendicular to g(x,y)= y- f(x)= 0 and \nabla g is always perpendicular to that curve. That means the two vectors \nabla D and \nabla g must be parallel: \nabla D= \lambda\nabla g where \lambda is some real number.

For maximizing functions, h, other than just "distance", remember that \nabla h always points in the direction of fastest increase of h so the same applies to that situation. If \nabla h were not perpendicular to the curve g(x,y)= y- f(x)= 0, we could move "left" or "right" on the curve, depending on the direction of the projection of \nabla h on the curve. We cannot do that when there is no projection- when \nabla h is perpendicular to the curve which is the same as saying it is parallel to \nabla g: \nabla h= \lambda\nabla f.
 
That makes perfect sense, but how does the idea work when for example I have a constraint that a line segment has to be of a particular length and then I have to find a line segment of that given length that maximizes a particular integral? Then I don't have a line of the constraint, the constraint could be any line, as long as it is of given length. How does the idea work with functionals in calculus of variations?
 
I found some proof online, on page 32 and a small part of the lemma is on page 31:

http://la.epfl.ch/files/content/sites/la/files/shared/import/migration/IC_32/ic-32_lectures-9-16.pdf

However, I do not understand why does the determinant have to be nonzero and if an inverse function can be found, how can one always find points that satisfy the constraint yet are better for the integral?
 

Similar threads

  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 25 ·
Replies
25
Views
3K
  • · Replies 3 ·
Replies
3
Views
3K