# Does every continuous function has a power series expansion on a closed interval

1. ### kof9595995

679
By Weierstrass approximation theorem, it seems to be obvious that every continuous function has a power expansion on a closed interval, but I'm not 100% sure about this. Is this genuinely true or there're some counterexamples?

2. ### g_edgar

607
This is incorrect. The sum of a power series is an analytic function. So, for example, the function |x| on the interval [-1,1] is not the sum of a power series, since it is not even differentiable at x=0.

3. ### kof9595995

679
But by Weierstrass you can always construct a sequence of polynomials uniformly converging to it, isn't it? I thought differentiability is required only if you want to do something like Taylor series expansion.

4. ### kof9595995

679
Em, now I see what you mean. As long as a function can be represented by a power series, it can be differentiated as many times as you want, so |x| can not be represented by power series.
But by Weierstrass theorem you can always construct a sequence of polynomials uniformly converging to it, so in what sense is this different with saying that it can be written as a power series? I'm very confused

5. ### LCKurtz

8,338
A sequence of polynomials converging to a continuous function may be something other than a sequence of partial sums of a Taylor series. A sequence of Bernstein polynomials converging to f is a good example.

Also there exist infinitely smooth functions that aren't represented by their Taylor series. Consider
$$f(x) = e^{- \frac 1 {x^2}},\ x \neq 0$$
and f(0) = 0. This has derivatives of all orders equal to 0 when x = 0 so its Taylor series sums to 0, not to f(x) except at 0. Yet on any finite interval it can be uniformly approximated by polynomials.

6. ### kof9595995

679
Right, I know there needn't to be a Taylor series, but why there can be no infinite power series at all in some cases, like |x| around x=0.

7. ### LCKurtz

8,338
Didn't you read G_Edgar's explanation? Another way you might see it is that if f(x) is expressible as a series like:

$$f(x) = \sum_{n=0}^\infty a_nx^n$$

with a positive radius of convergence then f must be as differentiable as the right side and, moreover, differentiating both sides shows the only choice for the coefficients are the Taylor coefficients. So Taylor series is the only game in town for a power series.

8. ### kof9595995

679
Well, I thought any polynomials can be written as a power series; like Bernstein polynomials, I thought if you expand each term, and collect the terms with same power, then you get a power series. Is it wrong because we can't switch the order of terms in infinite series without proving absolute convergence, or is there some other reason?
And, why in the Bernstein case we don't need the differentiability condition?
Please bear with if the question is too naive; I'm not majoring in math so I'm really not good at these analysis stuff.
Thanks.

9. ### Office_Shredder

4,500
Staff Emeritus
The question is: given a sequence of Bernstein polynomials, do they actually converge to a power series? In general the coefficients will change from one polynomial to the next

10. ### kof9595995

679
So you mean if I try to rewrite Bernstein polynomials in terms of powers of x, then as the Bernstein polynomials approach the function, the coefficients of x^n won't converge, but probably oscillate or blows up instead, right?
But for those can be written in power series, the coefficients must converge to what we collect from Bernstein polynomials, is that right?

11. ### LCKurtz

8,338
I have never looked at those two specific questions, so I don't know for sure. I would expect that the answer to your first question is yes and the second is no.