A question about the derivation of the Euler-Lagrange equation

Click For Summary

Discussion Overview

The discussion revolves around the derivation of the Euler-Lagrange equation, specifically focusing on the minimization of the functional I=\int_{x_1}^{x_2}{f(x,y,y')dx}. Participants explore the role of the parameter α in this context and its implications for finding stationary points of the functional.

Discussion Character

  • Technical explanation
  • Conceptual clarification
  • Debate/contested

Main Points Raised

  • One participant describes the derivation process involving families of functions Y(x) and Y'(x), questioning the introduction of α and its role in minimizing I.
  • Another participant clarifies that I is not a function of x, but rather a functional that changes with variations in α and η.
  • A different participant speculates on why α=0 is chosen to produce an extreme, suggesting that any value of α could potentially make the first derivative of I vanish, thus questioning the significance of the choice.
  • Another participant emphasizes that y(x) represents the function that minimizes the integral, while αη(x) represents an arbitrary deviation from that solution.

Areas of Agreement / Disagreement

Participants express differing views on the significance of α in the minimization process, with some agreeing on its role while others raise questions about its implications. The discussion remains unresolved regarding the necessity of choosing α=0 over other values.

Contextual Notes

There are assumptions regarding the nature of the functions involved and the conditions under which the functional is minimized. The discussion does not resolve the implications of these assumptions on the derivation process.

planck42
Messages
82
Reaction score
0
In the book Mathematical Methods for Engineers and Scientists 3, the derivation of the Euler-Lagrange equation starts roughly along the lines of this:

In order to minimize the functional I=\int_{x_1}^{x_2}{f(x,y,y')dx}, one should define two families of functions Y(x) and Y'(x), where Y(x) is y(x)+{\alpha}{\eta}(x) and Y'(x) is the x-derivative of y(x), \alpha is an arbitrary constant coefficient of {\eta}(x), which is an arbitrary function that must be zero at x_1 and x_2.

These steps aren't difficult to understand; it allows for any possible function that could render the functional stationary while maintaining the boundary conditions.

The minimum value of I should occur when \frac{dI}{d\alpha} is zero at {\alpha}=0

Whoa! Why are we dealing in \alpha's all of a sudden? How does this minimize I and not the derivative with respect to x? Does working it out with x always produce a trivial solution? I need help with this one step!
 
Last edited:
Physics news on Phys.org
I isn't a function of x (x is a dummy variable inside the integral, not a variable that you put into I).

On the other hand I DOES change as alpha changes (and as eta changes, but we're not going to worry about that)
 
Makes sense. However, I still have one question that I'll attempt to answer: why does {\alpha}=0 produce an extreme, and not, for instance, {\alpha}=1? My guess is that the value of \alpha is irrelevant; any value of \alpha can make the first derivative of I with respect to \alpha vanish. Hence the simplest choice, {\alpha}=0.
 
The point is that y(x) is a function that minimizes (or extremizes) the integral.
So in y(x)+{\alpha}{\eta}(x)[/itex], the y(x) is the sought after solution and {\alpha}{\eta}(x) is an arbitrary deviation from that solution.<br /> Therefore, the minimum occurs at {\alpha}=0
 

Similar threads

  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 12 ·
Replies
12
Views
3K
  • · Replies 16 ·
Replies
16
Views
3K
  • · Replies 6 ·
Replies
6
Views
5K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 12 ·
Replies
12
Views
3K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 8 ·
Replies
8
Views
3K
  • · Replies 1 ·
Replies
1
Views
3K