Question about the def. of solving 2nd order ODEs through Var. of Parameters.

Click For Summary
SUMMARY

The method of variation of parameters is utilized to solve second-order ordinary differential equations (ODEs) of the form y'' + p(t)y' + q(t)y = g(t). The general solution is expressed as c_1y_1(t) + c_2y_2(t) + y_p, where y_p is the particular solution defined as y_p = u_1(t)y_1(t) + u_2(t)y_2(t). A crucial assumption in this method is that u'_1(t)y_1(t) + u'_2(t)y_2(t) = 0, which simplifies the derivation process by eliminating second derivatives of u's when substituted back into the original equation. This assumption is not derived from necessity but rather from convenience to facilitate finding u_1(t) and u_2(t).

PREREQUISITES
  • Understanding of second-order ordinary differential equations (ODEs)
  • Familiarity with the method of variation of parameters
  • Knowledge of complementary solutions in differential equations
  • Basic calculus, including differentiation and integration techniques
NEXT STEPS
  • Study the derivation of the Wronskian W(y_1, y_2) for two functions y_1(t) and y_2(t)
  • Explore the application of the method of variation of parameters in non-homogeneous linear ODEs
  • Learn about the implications of the assumption u'_1(t)y_1(t) + u'_2(t)y_2(t) = 0 in solving ODEs
  • Investigate alternative methods for solving second-order ODEs, such as the method of undetermined coefficients
USEFUL FOR

Mathematicians, engineering students, and anyone involved in solving second-order ordinary differential equations, particularly those interested in advanced techniques for handling non-homogeneous equations.

Je m'appelle
Messages
117
Reaction score
0
Ok, so I've been studying the method of variation of parameters in order to solve 2nd order ODEs, and I have a question regarding a supposition that is made in the definition of the method.

Say,

[tex]y'' + p(t)y' + q(t)y = g(t)[/tex]

Then the general solution to the above equation is

[tex]c_1y_1(t) + c_2y_2(t) + y_p[/tex], and now replacing [tex]c_1\ and\ c_2\ by\ u_1(t)\ and\ u_2(t)[/tex] on the complementary solution [tex]c_1y_1(t) + c_2y_2(t)[/tex] in order to find the particular solution [tex]y_p[/tex]

[tex]y_p = u_1(t)y_1(t) + u_2(t)y_2(t)[/tex]

[tex]y'_p = u'_1(t)y_1(t) + u_1(t)y'_1(t) + u'_2(t)y_2(t) + u_2(t)y'_2(t)[/tex]

And then the methods states that

[tex]u'_1(t)y_1(t) + u'_2(t)y_2(t) = 0[/tex]

And this is where I'm curious and confused at the same time, I would like to know why it can be considered true, is there some kind of a proof to back this up? I mean, the person who developed this method certainly had something in mind to state that, so what does it consist of?

Then we get to derive [tex]y'_p[/tex] once again and then we substitute the values for [tex]y''_p,\ y'_p\ y_p[/tex] on the original equation to get the following system

[tex]u'_1(t)y_1(t) + u'_2(t)y_2(t) = 0\ OBS:\ I\ wanna\ know\ why\ this\ is\ equal\ to\ zero.[/tex]

[tex]u'_1(t)y'_2(t) + u'_2(t)_2y'_2(t) = g(t)[/tex]

So we can solve for [tex]u_1(t)\ and\ u_2(t)[/tex] and find the particular solution for the ODE.

Leaving us with the general solution for the ODE

[tex]y = c_1y_1(t) + c_2y_2(t)\ -\ y_1(t)\int \frac{y_2(t)g(t)}{W(y_1,y_2)(t)}\ dt\ +\ y_2(t)\int \frac{y_1(t)g(t)}{W(y_1,y_2)(t)}\ dt[/tex]

Thanks.
 
Last edited:
Physics news on Phys.org
You have defined two functions [itex]u_i[/itex], and you only have one constraint on them, which is your differential equation. You are free to choose your second constraint however you wish, and the constraint you're asking about happens to be convenient.
 
Je m'appelle said:
Ok, so I've been studying the method of variation of parameters in order to solve 2nd order ODEs, and I have a question regarding a supposition that is made in the definition of the method.

Say,

[tex]y'' + p(t)y' + q(t)y = g(t)[/tex]

Then the general solution to the above equation is

[tex]c_1y_1(t) + c_2y_2(t) + y_p[/tex], and now replacing [tex]c_1\ and\ c_2\ by\ u_1(t)\ and\ u_2(t)[/tex] on the complementary solution [tex]c_1y_1(t) + c_2y_2(t)[/tex] in order to find the particular solution [tex]y_p[/tex]

[tex]y_p = u_1(t)y_1(t) + u_2(t)y_2(t)[/tex]

[tex]y'_p = u'_1(t)y_1(t) + u_1(t)y'_1(t) + u'_2(t)y_2(t) + u_2(t)y'_2(t)[/tex]

And then the methods states that

[tex]u'_1(t)y_1(t) + u'_2(t)y_2(t) = 0[/tex]

And this is where I'm curious and confused at the same time, I would like to know why it can be considered true, is there some kind of a proof to back this up? I mean, the person who developed this method certainly had something in mind to state that, so what does it consist of?

It's hard to guess what the original developer of the method was thinking. It might have been as easy as "since cty1+c2y2 doesn't have enough flexibility to work, maybe if I let the c's be variables, I can make it work". But keep in mind that the goal is to see if somehow a u1(t) and u2(t) can be found that will work. Given that these are two unknown functions that you hope to solve for you need solving for them not to be as difficult or more so than the original DE. The author cleverly noted that by setting that equation equal to zero, he can avoid getting any second derivatives of the u's when they are substituted into the second order DE. It isn't that he noted that the above equation must be zero. Rather, it was an assumption that was made to try to keep solving for the u's as simple as possible. So it turns out that the system of equations that come out for the u derivatives can be solved and, in principle, the u functions themselves and hence yp.

Surely he must have shouted "Yes!" and celebrated with a glass of beer!
 

Similar threads

Replies
2
Views
2K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 3 ·
Replies
3
Views
3K
Replies
9
Views
2K
  • · Replies 9 ·
Replies
9
Views
2K