Proof Taylor series of (1-x)^(-1/2) converges to function

Click For Summary

Discussion Overview

The discussion revolves around proving the convergence of the Taylor series expansion of the function f(x) = 1/√(1-x) for the interval -1 < x < 1. Participants explore various methods to establish this convergence, including the use of the Lagrange error bound, the ratio test, and properties of composite functions.

Discussion Character

  • Exploratory
  • Technical explanation
  • Mathematical reasoning
  • Debate/contested

Main Points Raised

  • One participant proposes using the Lagrange error bound to show that the error tends to zero as n approaches infinity, but expresses difficulty in managing the limit due to the complexity of M.
  • Another participant suggests an alternative approach that involves proving the convergence of a series to a differentiable function and then relating it to the original function.
  • A different participant recommends using the ratio test to evaluate the limit for convergence.
  • One participant questions whether direct proof of convergence is necessary, suggesting that properties of composite functions could be leveraged instead.
  • Another participant expresses uncertainty about proving convergence for the full range -1 < x < 1, despite being able to prove it around x = 0.
  • A participant shares their results from the limit evaluation, noting discrepancies and questioning the implications for x > 0.5.
  • There is a discussion about the ratio test's requirements, clarifying that the limit should be less than 1 for convergence, not necessarily zero.
  • One participant reflects on the potential issues with the error bound affecting convergence and considers the implications of their findings.
  • Questions arise regarding the relationship between the ratio test results and proving that the series converges to a differentiable function.
  • Another participant advises looking for a theorem on power series to clarify the relationship between convergence and differentiability.

Areas of Agreement / Disagreement

Participants express various methods and approaches to proving convergence, but there is no consensus on the best method or the implications of their findings. Some participants agree on the utility of the ratio test, while others explore different avenues without reaching a definitive conclusion.

Contextual Notes

Participants note limitations in their approaches, including potential computational errors and the complexity of the error bounds. The discussion reflects a range of assumptions and conditions that may affect the conclusions drawn about convergence.

JD96
Messages
33
Reaction score
0
Hello,

I want to prove that the taylor expansion of f(x)={\frac{1}{\sqrt{1-x}}} converges to ƒ for -1<x<1. If I didn't make a mistake the maclaurin series should look like this:
Tf(x;0)=1+\sum_{n=1}^\infty{\frac{(2n)!}{(2^n n!)^2}}x^n
My attempt is to use the lagrange error bound, which is given by E(x)\leq M\frac{|x|^{n+1}}{(n+1)!}, where M is the upper bound for |f^{(n+1)}(x)| for the interval [0,x] and show it tends to zero for \lim_{n\rightarrow \infty}.
My problem is the following: For functions like ex or sin(x) M is independent of n and thus the limit is easy to compute, but here M is given by f^{(n+1)}(x)=\frac{(2n)!}{4^n n!}(1-x)^{-\frac{2n+1}{2}} (which I conclude from graphing the function). So I need to prove that the following holds true (for -1<x<1):
\lim_{n\rightarrow \infty}(\frac{(2n)!}{4^n n!}(1-x)^{-\frac{2n+1}{2}}\frac{|x|^{n+1}}{(n+1)!})=0
I have no clue how to solve such a limit and would appreciate any help! Of course I don't expect anyone to work it out, but maybe someone has an advice as how to tackle complicated limits like this one... .

Thanks in advance!
 
Last edited:
Physics news on Phys.org
Here's another way that does not involve the limit:

1) Prove the series ##\sum_{n=1}^{+\infty} \frac{(2n)!}{2^{2n} (n!)^2}x^n## converges to a differentiable function ##g(x)##.
2) Prove that ##\frac{d}{dx} g(x)\sqrt{1+x}## is ##0##, and hence that ##g(x) = \frac{c}{\sqrt{1+x}}##
3) Prove that ##c=1##.
 
If you do want to evaluate the limit, then you should consider the ratio test.
 
Is it a requirement to directly prove that the expansion converges? Instead, can you use the fact that it is the composition of functions whose radii of convergence are well known? (or easy to prove)
 
Thanks to both of you for your replies!
@micromass I will try to get my head around your method after I had some sleep. I just tried to work it out using the ratio test and got a wrong result (I have fairly little experience with computing limits), but maybe I will find the error tomorrow :-)
@FactChecker Not necessarily, however I not only want to prove that the expansion converges, but also that it converges to the function ƒ. So I am curious, is one able to prove the latter using the fact it is a composite function?
 
JD96 said:
So I am curious, is one able to prove the latter using the fact it is a composite function?
I am having second thoughts. It is possible to prove convergence around x=0, but it's not clear to me, without a lot of thinking, what it would take to prove it especially for the full range -1<x<1.
 
I decided to work out the limit and keep getting the same result, namely \displaystyle\lim_{n\rightarrow +\infty} {\frac{a_{n+1}}{a_n}}=\frac{|x|}{1-x}. But instead it should be zero or am I wrong here? Clearly the result I got is wrong, since that limit doesn't exist for x>0,5. So is it true that I should get \displaystyle\lim_{n\rightarrow +\infty} {\frac{|x|}{n(1-x)}}=0 if x\neq0 for the proof to succeed?
Also I want to mention that the limit for E(x) I wrote down in my original post is wrong since I mixed up f^{(n+1)}(x) and f^{(n)}(x).
 
Last edited:
The ratio tests states for a series ##\sum a_n## that if ##\lim_{n\rightarrow +\infty}\left|\frac{a_{n+1}}{a_n}\right|<1##, then the series converges (and hence the sequence of terms converges to ##0##).

So you don't want the limit to go to ##0##, you want it to be ##<1##. So indeed, it seems you have a problem for ##x>1/2##. This might be because of a computational error, or because the specific bound was to rough.
 
Mhh... I didn't consider the possibility that the error bound could cause problems and it really seems that it does, since I checked my result several times (and proving the taylor expansion converges for 1/2<x is not really satisfying). Thus I am considering your method, but I have a few questions regarding the individual steps: 1) I did the ratio test and the result is that the series converges for |x|<1. Is that already proving the series converges to a differentiable function in that domain? 2) How would I prove that without knowing how g(x) exactly looks like? 3) How would I determine the value of the integration constant?

Sorry if I'm asking too many questions at a time, I am very grateful for your input :-)
 
  • #10
JD96 said:
Mhh... I didn't consider the possibility that the error bound could cause problems and it really seems that it does, since I checked my result several times (and proving the taylor expansion converges for 1/2<x is not really satisfying). Thus I am considering your method, but I have a few questions regarding the individual steps: 1) I did the ratio test and the result is that the series converges for |x|<1. Is that already proving the series converges to a differentiable function in that domain?

Not in general, but for power series it does. You should look for a good theorem on power series which says exactly that. Any analysis book should have it.

How would I determine the value of the integration constant?

Evaluate both sides for a suitable concrete ##x## (for example ##x=0##).
 

Similar threads

  • · Replies 3 ·
Replies
3
Views
4K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 5 ·
Replies
5
Views
3K
  • · Replies 9 ·
Replies
9
Views
3K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 6 ·
Replies
6
Views
4K
  • · Replies 9 ·
Replies
9
Views
2K
  • · Replies 19 ·
Replies
19
Views
4K
  • · Replies 3 ·
Replies
3
Views
2K