Variation sign and integral sign

In summary, the conversation discusses the feasibility of exchanging the order of a variation sign and an integral sign, with a proof given in a book. A step in the proof is questioned, specifically how certain variables can be moved into the integral sign. The respondent explains that there may be some sloppiness in the notation, as the same names are used for both variables and functions. They then walk through the derivation and make assumptions about the context. The derivation is properly denoted, but there are shortcuts that can be used, although they may be slightly erroneous. The respondent also discusses the use of the Gateaux differential and derivative in functional analysis and the relationship between integration and linear functionals. They conclude by mentioning that the derivative of a linear operator
  • #1
thaiqi
160
8
TL;DR Summary
wondering about a proof of exchange between variation sign and integral sign
Hello, everyone.
I know that it is feasible to exchange the order of one variation sign and one integral sign. But there gives a proof of this in one book. I wonder about a step in it. As below marked in the red
rectangle:
variation_question2.png

How can ##\delta y## and ##\delta y^\prime## be moved into the integral sign? Aren't they functions of ## x ## ?
 
Physics news on Phys.org
  • #2
You are correct in that the [itex]\delta y[/itex] and [itex]\delta y'[/itex] should not, as shown, appear outside the integrals. There is a bit of slopyness in the notation we use where we utilize the same names for variables and the functions such as [itex]y_{variable}= y_{function}(x)[/itex]. I'll walk through the above derivation in this context in a moment to show you what I mean.

Now you have not given the context of this derivation but I will make some assumptions and you can correct me if I err. Below I will distinguish functions from variables by using greek. I.e. if [itex]x,y[/itex] are respectively independent and dependent variables, that dependency is given by a function, say [itex] y = \varphi(x)[/itex]. (Toward the end I'll revert back to y as function.)

It looks to me like you are taking the variation of a functional, let's call it [itex]S[/itex] defined by:

[tex]S[\phi] = \int_{x_1}^{x_2} F(x,\phi(x),\phi'(x)) dx[/tex]
by analogy compare this to some scalar valued function of a vector [itex]\sigma(\mathbf{v})[/itex]. (Note that here I use brackets [] to indicate a (not necessarily linear) operator's action on a function as a whole. Rather than a function action on its value as a composition of functions.)

The variation [itex]\delta S[/itex] is analogous to the differential of the scalar valued function but since we don't have a nice indexible expansion it is harder to express in terms of components. It however has a given form in contrast to my arbitrary vector function analog. So let's break it down in terms of functional derivatives and limits of difference quotients:
[itex]\delta S[\varphi] [/itex] is the value of a linear functional on the variation of [itex]\varphi[/itex] defined by the Gateaux differential and derivative:
[tex] \delta S[\varphi] = S'[\varphi][\delta\varphi]= \lim_{h \to 0} \frac{1}{h}\left(S[\varphi+h\delta\varphi]-S[\varphi]\right)[/tex]
Here [itex]\delta\varphi[/itex] is an independent variation of the function in the same sense that say [itex]d\mathbf{v}[/itex] is an independent differential vector in the vector differential relation:
[tex]d\sigma(\mathbf{v})= \nabla\sigma(\mathbf{v})\bullet d\mathbf{v}[/tex]
In this vector example the linear functional can, via the Riesz Representation Theorem be expressed by dotting with a vector. The gradient is the dual of the vector derivative of [itex]\sigma[/itex].

So taking the form of [itex]S[/itex] and applying the limit of the difference quotient we get:
[tex]\delta S[\varphi][\delta\varphi] = \lim_{h\to 0} \frac{1}{h}\left( \int_{x_0}^{x_1} F(x,\varphi(x)+h\delta\varphi(x), \varphi'(x)+h\delta\varphi'(x)) - F(x,\varphi(x),\varphi'(x))dx\right)[/tex]
Remembering that the variation [itex]\varphi[/itex] is arbitrary and applying linearity to the integral we can internalize the limit of the difference quotient into the integral to yield an integral of partial derivatives:
[tex]\delta S=\int_{x_0}^{x_1} \left(F_2(x,\varphi(x),\varphi'(x)) \delta\varphi(x) + F_3(x,\varphi(x),\varphi'(x))\delta\varphi'(x) \right)dx = \int_{x_0}^{x_1} \delta F(x,\varphi(x),\varphi'(x))dx[/tex]

That is the derivation properly denoted. However as you can see the explicit details can be a bit overwhelming so one typically invokes some (somewhat sloppy) shortcuts. That slightly erroneous first step works because one is invoking a "functional gradient" in a sense when treating "[itex]\delta y[/itex]" as a differential vector (in the space of differentiable functions on the interval [itex]I=[x_0,x_1][/itex]).
[tex] \delta S = \frac{\delta}{\delta y}\left(\int_I F dx\right)[\delta y] = \nabla_y\left(\int_I F dx\right)\bullet \delta y[/tex]
where this "dot product" would be an integral contraction with another variable of integration, say [itex]\tilde{x}[/itex]. The fact that [itex]\delta y[/itex] as a functional differential is arbitrary doesn't quite imply that its value and derivative are independent. But one can invoke say arbitrary linear combinations of a delta function and the derivative of a delta function at separate points, e.g.
[itex]\delta y(x) = p\delta(x-a)+q\delta'(x-b)[/itex]
to show that the "sloppy" method will still lead to a valid result. I finally understood these variational derivations much better when I explored the implications of using the Gateaux differential and derivative in functional analysis. It is a little bulkier but one can get to the result without "bastardizing" the notation.

One final point I'd make. Integration of a function on a fixed interval is a linear functional. Consider in general that the derivative of a vector mapping is a linear operator (the Jacobi matrix represents this operator for vector mappings which are coordinate transformations.) If you look at the generalized derivative of a linear operator it will in fact be its own derivative.
For scalars: [itex]a: x \mapsto y=ax[/itex] means [itex]a(x)' = (ax)'=a[/itex]
For linear maps: [itex] A: \mathbf{x}\mapsto \mathbf{y}=A\mathbf{x}[/itex] means [itex]\frac{d\mathbf y}{\mathbf{x}} = A[/itex].
This makes more sense when we look at differentials (remembering derivatives are mappings of differentials to differentials)
[tex] y =ax\to dy = adx,\quad \mathbf{y}=A\cdot \mathbf{x}\to d\mathbf{y}=A\cdot \mathbf{x}[/tex]
This applies to function mappings since they are, after all, just another form of vector mapping:
[itex] I [\varphi] = \int_{[a,b]}\varphi(t) dt[/itex] is linear and so [itex]\delta (I[\varphi] = I\cdot\delta \varphi[/itex].

This is, I think the gist of what your author is deriving for the case at hand IMNSHO.
 
  • Like
Likes TheDS1337
  • #3
jambaugh said:
You are correct in that the [itex]\delta y[/itex] and [itex]\delta y'[/itex] should not, as shown, appear outside the integrals. ……
First thanks very much for your reply. I don't understand very well all parts of what you said. Do you agree it is a wrong step in the rectangle marked step at the third equal sign(though the result happens to be valid)? (Besides, I didn't say [itex]\delta y[/itex] and [itex]\delta y'[/itex] should not appear outside the integrals, but I said they cannot be moved into the integral. )
 
  • #4
To answer your direct question, it is really a matter of the reverse, moving the [itex]\delta y, \delta y'[/itex] outside the integral from the inside which would be a mistake. In point of fact, it's the very first step that has questionable validity.

Again it is a question of conflating the meaning of [itex]y[/itex], at one level as a variable with respect to which one can take a partial derivative when it appears inside a function an at another level as a function of [itex]x[/itex] which may occur within a definite integral.

As an exercise try evaluating each step in this derivation sequence for specific examples. Say let [itex]F(x,y,y')=x^2\cdot y^2\cdot y'[/itex] and let [itex] x_0 = 0, x_1=1[/itex]. For the initial functional and then the first derivation step one must evaluate the definite integral before considering variations or partial derivatives. To integrate you can't leave [itex]y[/itex] and [itex]y'[/itex] independent variables. They must be assigned some functional dependency on [itex]x[/itex]. Let's choose one, say [itex]y=x^3, y'=3x^2[/itex]. Then when you evaluate the definite integral you get a constant value. What does it mean to take the partial derivative of that with respect to [itex]y[/itex] or [itex]y'[/itex].

Alternatively if you treat [itex]y[/itex] and [itex]y'[/itex] as independent variables, then the integration would be:
[tex]\int_0^1 F(x,y,y')dx = \int_0^1 x^2 y^2 y' dx = \left[\frac{x^3}{3}\right]^1_0 y^2 y' = \frac{1}{3}y^2 y'[/tex]
That is most definitely not what is intended here.

Phrased properly the partial derivatives of the first step should be a single functional derivative of the integral as a functional on [itex]y[/itex]
[tex]= \frac{\delta }{\delta y}\left(\int_I F(x,y(x),y'(x))dx\right)[\delta y] = *...[/tex]
The functional derivative should yield a (linear) functional acting on the variation of [itex]y[/itex] the function.

But you can then, for the purposes of expanding in terms of partial derivatives, consider [itex]y[/itex] and [itex]y'[/itex] independent functions of [itex]x[/itex] (and here I'm going to go greek to emphasize these functions) and expand this in terms of partial functional derivatives:
[tex]*=\left. \frac{\delta}{\delta \phi}\left(\int_I F(x,\phi(x),\psi(x))dx\right)\right\rvert_{(\phi,\psi)=(y,y')}[\delta y] ...[/tex]
[tex] +\frac{\delta}{\delta \psi}\left(\int_I F(x,\phi(x),\psi(x))dx\right)\rvert_{(\phi,\psi)=(y,y')}[\delta y'][/tex]
The next step would then be to internalize the partial functional derivative which would then manifest as an integral of standard partial derivatives of the integrand function F.

It's bulky to express this correctly. What is happening is that each "partial functional derivative" of a functional is yielding a linear functional valued functional which is evaluated at a functional differential. e.g.
[tex]A[y]= \int_a^b F(x,y(x))dx,\quad \frac{\delta}{\delta y}A[y]=B[y], \quad (B[y])[z]=\int_a^b F_2(x,y(x))\cdot z(x) dx[/tex]
Here [itex]x[/itex] is a variable, and [itex]y,z[/itex] are functions (function valued variables to be precise). Also btw [itex]F_2(u,v)=\frac{\partial}{\partial v}F(u,v)[/itex]. I'm here again using [] to indicate evaluation of a functional in analogy to but distinguished from standard function notation's ().
 
Last edited:
  • Like
Likes jim mcnamara
  • #5
jambaugh said:
To answer your direct question, it is really a matter of the reverse, moving the [itex]\delta y, \delta y'[/itex] outside the integral from the inside which would be a mistake. In point of fact, it's the very first step that has questionable validity.
...
Thanks very much for your guidance again.
 

1. What is the difference between the variation sign and the integral sign?

The variation sign (∆) represents the change or difference between two values, while the integral sign (∫) represents the accumulation or total of a function over a certain interval.

2. How is the variation sign used in calculus?

In calculus, the variation sign (∆) is used to represent the change in a function's value with respect to a small change in its input variable. It is often used in the definition of the derivative and the fundamental theorem of calculus.

3. What is the meaning of the integral sign in mathematics?

The integral sign (∫) in mathematics represents the process of finding the total or accumulated value of a function over a given interval. It is commonly used in calculus to calculate areas, volumes, and other quantities.

4. Can the variation sign and integral sign be used together?

Yes, the variation sign and integral sign can be used together in certain applications. For example, in physics, the variation of a quantity with respect to time can be represented by the derivative (∆t) and the total change in that quantity over a period of time can be represented by the integral (∫t).

5. Are there other mathematical symbols that are similar to the variation sign and integral sign?

There are several other symbols that are similar to the variation sign and integral sign, such as the delta symbol (Δ) and the sigma symbol (∑). These symbols are used in different contexts and have slightly different meanings, but they all represent some form of accumulation or change.

Similar threads

Replies
1
Views
945
  • Calculus
Replies
29
Views
727
Replies
2
Views
299
Replies
24
Views
2K
Replies
7
Views
1K
Replies
20
Views
2K
Replies
3
Views
341
Replies
3
Views
585
Replies
2
Views
1K
Replies
6
Views
1K
Back
Top