Dismiss Notice
Join Physics Forums Today!
The friendliest, high quality science and math community on the planet! Everyone who loves science is here!

I General solution to linear homogeneous 2nd order ODEs

  1. Apr 27, 2016 #1
    Given a linear homogeneous 2nd order ODE of the form $$y''(x)+p(x)y'(x)+q(x)=0$$ the general solution is of the form $$y(x)=c_{1}y_{1}(x)+c_{2}y_{1}(x)$$ where ##c_{1},c_{2}## are arbitrary constants and ##y_{1}(x), y_{2}(x)## are linearly independent basis solutions.

    How does one prove that the general solution is given by the above?
    Last edited by a moderator: Apr 28, 2016
  2. jcsd
  3. Apr 27, 2016 #2

    Simon Bridge

    User Avatar
    Science Advisor
    Homework Helper

    substitute the proposed general solution into the DE and see.

    [edit] is there a typo in the general solution equation?
    Mod note: Now fixed in orig. post...
    Last edited by a moderator: Apr 28, 2016
  4. Apr 27, 2016 #3


    Staff: Mentor

    One explanation appeals to concepts from linear algebra. The solution space for a first-order, homogeneous differential equation has dimension one, so if a nontrivial solution (i.e., not identically zero) can be found, every solution will be a constant multiple of this solution.

    The solution space for a second-order, homogeneous differential equation has dimension two. If you have two linearly independent solutions, then the solution space is spanned by these two functions. That is, the general solution consists of all linear combinations of the two basis functions, exactly as you show in the second equation. If two initial conditions are given, then the constants ##c_1## and ##c_2## can be determined to give a unique solution.

    For a third-order, homogeneous differential equation, there need to be three linearly independent basis functions, and so on, for higher-order equations.
  5. Apr 28, 2016 #4
    Yes, sorry. It should read ##y(x)=c_{1}y_{1}(x)+c_{2}y_{2}(x)##.

    That's kind of how I intuitively see it, but I was wondering how one proves it (or is such a proof fiendishly hard)?
  6. Apr 28, 2016 #5


    Staff: Mentor

    I have several DE textbooks, but I don't remember any of them proving that a linear combination of basis functions is the general solution, but then, I haven't looked at these books for quite a while. However, a proof of this wouldn't be "fiendishly hard," I don't believe, and possibly would use a proof by contradiction. If I get a chance later today, I'll see what I can find.
  7. Apr 28, 2016 #6
    Ok great, I'd much appreciate that.
  8. Apr 28, 2016 #7
    If you can get hold of a copy of "An introduction to Linear Analysis", Kreider, Kuller, Ostberg, and Perkins, then you can see their partial proof of this in their Theorems 3.2 and 3.3.

    Theorem 3.2 states an existence and uniqueness theorem for the solutions of a normal n'th order linear DE, given a set of initial conditions but refers the reader elsewhere for the proof, and Theorem 3.3 relies on this to show that said DE has n l.i. solutions which span the solution space of the DE.

    They prove Theorem 3.3 by choosing a set of n initial conditions for which it is easy to show that the corresponding solutions (which exist and are unique by Theorem 3.2) are l.i.
  9. Apr 28, 2016 #8

    Simon Bridge

    User Avatar
    Science Advisor
    Homework Helper

    I see I misunderstood:

    Given DE: y'' + py' + q = 0

    knowing y1 and y2 are independent solutions, then y = Ay1 + By2 is also a solution, for any arbitrary A,B... prove by substitution.

    So what you want to prove is that any other solution can be written as a linear sum of two independent solutions, not just that any linear combination is a solution. i.e. is there a y, that is a solution, that cannot be written as a linear sum of y1 and y2?

    If y1 and y2 are orthogonal, and the solutions to the DE form a vector space of dimension 2, then doesn't it follow that any other solution must be a linear sum of y1 and y2?

    If we want to prove that it does follow - then look up the corresponding proof for a general vector space.
    What did I miss?
  10. Apr 29, 2016 #9
    Yes, this is basically what I want to prove.

    This follows if their corresponding Wronskian is non-zero, right? I can see how it must be the case if you consider ##y_{1}## and ##y_{2}## as a basis for a two dimensional vector space, but isn't there some sort of proof without using linear algebra?

    I think what I find hard to justify is why the solution space to an ##n##-th order differential equation must be ##n##-dimensional? I can kind of see that if this is the case, then if we can find ##n## linearly independent solutions ##y_{1},\cdots ,y_{n}## (whereby linear independence corresponds to ##W(y_{1},\cdots ,y_{n};x)\neq 0##), then every solution must be a linear combination of these "basis" solutions, but I'm unsure how to prove that this is true?!
    Last edited: Apr 29, 2016
Share this great discussion with others via Reddit, Google+, Twitter, or Facebook

Have something to add?
Draft saved Draft deleted