Lagrange multipliers and variation of functions

Click For Summary

Discussion Overview

The discussion revolves around the application of Lagrange multipliers in the context of variational calculus, particularly in mechanical systems with constraints. Participants explore the implications of eliminating variables through constraints and how this affects the formulation of extremum problems in both theoretical and practical scenarios.

Discussion Character

  • Exploratory
  • Technical explanation
  • Conceptual clarification
  • Debate/contested
  • Mathematical reasoning

Main Points Raised

  • Some participants propose that eliminating variables using constraint equations incorporates information about the system, affecting how variations are treated during differentiation.
  • Others discuss the mathematical insight from literature, suggesting that reducing the problem to fewer variables can be seen as a "trick" that simplifies finding extrema under constraints.
  • A physical example involving a 2D pendulum is presented, illustrating how constraints can change the formulation of the Lagrangian and the resulting equations of motion.
  • Some participants express confusion about the mechanics of how Lagrange multipliers reduce the number of differential equations, questioning the relationship between ordinary calculus and variational calculus.
  • One participant clarifies that the method of Lagrange multipliers does not actually reduce the number of variables or equations but rather incorporates constraints into the equations, leading to a parallel gradient condition at stationary points.
  • Another participant acknowledges a misunderstanding regarding the reduction of equations and notes that the process results in simultaneous equations that include both the original and constraint equations.

Areas of Agreement / Disagreement

Participants exhibit a mix of agreement and disagreement regarding the implications of using Lagrange multipliers. While some acknowledge the method's utility in incorporating constraints, others contest the notion that it reduces the number of equations or variables, leading to an unresolved discussion on this point.

Contextual Notes

Some limitations in understanding arise from the complexity of the mathematical mechanics involved in variational calculus and the application of Lagrange multipliers, particularly in how they relate to ordinary calculus principles.

jackiefrost
Messages
140
Reaction score
1
Let F and f be functions of the same n variables where F describes a mechanical system and f defines a constraint. When considering the variation of these functions why does eliminating the nth term (for example using the Lagrange multiplier method) result in a free variation problem where it wasn't necessarily so before the elimination?
 
Physics news on Phys.org
I didn't state my question very well. Anyway, I think I may be getting some insight from a Dover booklet - "Calculus of Variations", Gelfand and Fomim. The portion that is bearing some fruit in my mind says:
"As is familiar from elementary analysis, in finding an extremum of a function of n variables subject to k constraints (k<n), we can use the constraints to express k variables in terms of the other n - k variables. In this way, the problem is reduced to that of finding an unconstrained extremum of a function of n - k variables; i.e. an extremum subject to no subsidiary conditions."​
I need to work out some simple examples to get the feel for why this works but it seems to me that this is a very amazing little mathematical "trick". Lagrange was pretty smart, huh? :wink:
 
I guess the gist of it is that when you eliminate variables using constraint equations, you are incorporating information about your system into the problem. When extremizing the result, the information is utilized, consequently, during differentiation. i.e. You no longer regard certain variables as having independent variations with respect to the varying parameter, rather, you acknowledge the constraint, substitute in for the variable in question and observe the differential changes in the substituted expression.

A physical example of this would be to take the Lagrangian of a 2d pendulum system, written in Cartesian (not polar!) coordinates and substitute length = x^2 + y^2 for x in the Lagrangian. We know that this system has 1 degree of freedom and, now, the Lagrangian is expressed using one coordinate, y. When we look at variations of L with respect to some parameter, we will also be looking at variations in the expression sqrt(length - y^2), wherever x had appeared. Thus we are asserting the truth of both the constraint equation and L when we extremize.

The method of Lagrange multipliers is a little different, procedurally and conceptually, yet with similar results. Returning to the 2D pendulum, this method suggests that we set the variational derivative of L - \lambda * (x^2 + y^2) to 0. The reason for this is often explained geometrically, where, at a stationary point, the gradient of the two terms should be parallel (hence the lambda and the 0).

A good way to get a sense for these equations is to break them, in a controlled way. So, for example, what happens when you don't include any constraint information into the 2D Lagrangian? You get a constant momentum in the x-direction and the point mass either hits the V=0 level with a thud (if we asserted there is a ground) or oscillates about it in this odd mathematical construction. It is as if we snipped the string connecting the point mass to the pendulum pivot point. The first point to make is that we were successful in extremizing the Lagrangian without the constraint equation. We, in return, were given the EOMs (because these are the extremums) of a system that had no constraints. The second point is that this result is not consistent with the pendulum system we had in mind when we began, as it should be, for there was no way our Lagrangian could acquire the knowledge that "there is a string."
 
derek e said:
I guess the gist of it is that when you eliminate variables using constraint equations, you are incorporating information about your system into the problem. When extremizing the result, the information is utilized, consequently, during differentiation. i.e. You no longer regard certain variables as having independent variations with respect to the varying parameter, rather, you acknowledge the constraint, substitute in for the variable in question and observe the differential changes in the substituted expression.
Yes - and what mystifies me is the actual "mechanics" (in the mathematical sense) of how the lambda method accomplishes a reduction in the number of resulting differential equations from n to n-m if when we have n variables for the functional:
F(x1,...,xn),​
and and m constraints,
Gj(x1,...,xn), {j=1,2,...,m},m<n.​
Actually, the variation of F (or the variation of the definite integral of F) under the constraining condition(s) given by G. The lambda (Lagrange multiplier) method seems to accomplish a result similar to what it does when using ordinary functions in multi-variable calculus, as you mentioned. In ordinary calculus, the lamda method works because the constraining function is tangent to the function under examination so that their respective gradients are parallel (or anti-parallel) and therefore are a real number multiple of each other. For some reason I'm having a mental block in seeing the similarity when dealing with the variation of functionals. I understand that in principle it should be the same. It's the mechanics of how variation(F+lamda G)=0 reduces the number of "admissible curves" that F can take as input from m to m-n. (If that's even an allowable way to say it?)

derek e said:
A physical example of this would be to take the Lagrangian of a 2d pendulum system...
Hey! I see what you mean here! I'm going to play with this today. Thanks a lot.
[I'm kind of playing in my head with the geometrical notion of how constraints (as functionals) accomplish dimensional reduction and therefore path limitation in configuration space and your example is a perfect playground]
 
Last edited:
It's probably best to split our peas and carrots. The method of Lagrange multipliers does not reduce the number of variables or the number of equations. What it does is incorporate the constraint information by adding a term to the constrained equation, then it observes a parallel gradient condition for singular points. In the 2D pendulum problem, one usually extremizes both the x and y coordinates, separately. On the other hand, solving the constraint for x and substituting this into the Lagrangian gives an expression free of x that can be normally extremized. The resulting y-EOM then gives the x-EOM without any further extremization. This sounds messy and I haven't tried it. The method of Lagrange multipliers is accessible in this problem, however.

There is a simple example of this problem in the mechanics book, by Hand and Finch, that I had used in school and I just picked it up and read that the authors suggest using the constraint length = sqrt(x^2 + y^2), rather than the length squared, because this gives a more interesting interpretation of the Lagrange multiplier.

https://www.amazon.com/dp/3540219250/?tag=pfamazon01-20 has a good section on the geometry of mechanics, as well. I also remember some Cal Tech notes that were floating around their physics site that handled this particular topic nicely. They were supplements (more like a step-by-step walk-through, lol) of Hand and Finch. Unfortunately, I do not remember the professor's name (I went to Berkeley).
 
Last edited by a moderator:
derek e said:
It's probably best to split our peas and carrots. The method of Lagrange multipliers does not reduce the number of variables or the number of equations. What it does is incorporate the constraint information by adding a term to the constrained equation, then it observes a parallel gradient condition for singular points.
Yes. I was wrong about the reduction. Not sure why I thought that. In fact, in going back and re-studying what I have on this, now it seems that when finding the stationary value of a functional with n degrees of freedom, F(u1,...,un), with m constraints, gi(u1,...,un)=0 {i=1,...,m}, we actually end up with n+m simultaneous equations. The n equations are from the modified functions (modified to incorporate the constraints into F by using separate Lagrange multipliers for each constraint g), F+lambda1g1+...+lambdamgm. The m additional equations would be the m constraining functions gi=0. So this would be n+m equations in m+n unknowns. The unknowns are u1,..., un, lambda1,..., lambdam.
At least, that's what I think today... :rolleyes:

https://www.amazon.com/dp/3540219250/?tag=pfamazon01-20 has a good section on the geometry of mechanics, as well.
That looks like a great book! I wish I could afford it :cry:
(I went to Berkeley).
Communist :smile:
 
Last edited by a moderator:

Similar threads

  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 10 ·
Replies
10
Views
2K
  • · Replies 3 ·
Replies
3
Views
1K
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 4 ·
Replies
4
Views
2K
Replies
3
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K