Question about expanding a function to first order

Click For Summary
SUMMARY

The discussion centers on the validity of approximating the function f(x + Δx) using the first-order Taylor expansion, specifically f(x + Δx) ≈ f(x) + f'(x)Δx. It is established that this approximation is only valid when Δx is sufficiently small, regardless of its relative size to x. Examples such as f(x) = x² and f(x) = sin(x) illustrate that for functions with rapid curvature changes, the approximation may fail. Taylor's Theorem is referenced to explain the relationship between the approximation error and the second derivative of the function.

PREREQUISITES
  • Understanding of first-order Taylor expansion
  • Familiarity with derivatives and their notation
  • Knowledge of Taylor's Theorem
  • Basic concepts of function curvature and behavior
NEXT STEPS
  • Study Taylor's Theorem in depth, focusing on higher-order derivatives
  • Explore examples of first-order approximations with various functions
  • Analyze the impact of curvature on approximation accuracy
  • Investigate cases where linear functions provide exact approximations
USEFUL FOR

Mathematicians, physics students, and anyone interested in calculus and approximation methods will benefit from this discussion.

hgandh
Messages
26
Reaction score
2
If we have a function ##f(x+\Delta x)## where ##\Delta x << x##, is it valid to approximate this as:
$$f(x + \Delta x) \approx f(x) + f'(x)\Delta x$$
even if ##\Delta x## is not necessarily small? If not, what is the valid expansion to first order?
 
Physics news on Phys.org
If ##\Delta x## is big, then ##f'(x)## and ##\dfrac{f(x+\Delta x)-f(x)}{\Delta x}## can be very different, regardless how small ##\Delta x## versus ##x## is. E.g. ##x## could be positive and ##\Delta x## a large negative number, i.e. the function values can be quite different.
 
fresh_42 said:
If ##\Delta x## is big, then ##f'(x)## and ##\dfrac{f(x+\Delta x)-f(x)}{\Delta x}## can be very different, regardless how small ##\Delta x## versus ##x## is. E.g. ##x## could be positive and ##\Delta x## a large negative number, i.e. the function values can be quite different.
If I restrict it to positive values only, would this become valid then?
 
hgandh said:
If I restrict it to positive values only, would this become valid then?
No. To use a first order approximation, ##\Delta x ## must be "small" in some sense.

You can try some examples, like ##f(x)=x^2##.
 
PS another perhaps even more illustrative example would be ##\sin x##.
 
Exceptions would be linear, constant functions where expansion would be valid everywhere.
 
hgandh said:
If we have a function ##f(x+\Delta x)## where ##\Delta x << x##, is it valid to approximate this as:
$$f(x + \Delta x) \approx f(x) + f'(x)\Delta x$$
even if ##\Delta x## is not necessarily small? If not, what is the valid expansion to first order?

You can answer this yourself. Draw a curve, any curve. That's ##f(x)##. Draw the tangent line at any point ##x_0##. That's the line ##f(x_0) + f'(x_0) \Delta x##. How far is the line from the curve? Anywhere it's "close enough", whatever that means to you, is a place where the approximation is "good enough".

Intuitively you can see that if the curve bends rapidly away from the tangent line, then the error gets big quickly, whereas if the curve is relatively flat and stays close to the line, then the approximation is pretty good over a larger range.

Taylor's Theorem expresses that rigorously with the error term. If you stop at the first derivative, the amount of error depends on the second derivative, i.e. how fast the slope of the curve changes.
 
RPinPA said:
You can answer this yourself. Draw a curve, any curve. That's ##f(x)##. Draw the tangent line at any point ##x_0##. That's the line ##f(x_0) + f'(x_0) \Delta x##. How far is the line from the curve? Anywhere it's "close enough", whatever that means to you, is a place where the approximation is "good enough".

Intuitively you can see that if the curve bends rapidly away from the tangent line, then the error gets big quickly, whereas if the curve is relatively flat and stays close to the line, then the approximation is pretty good over a larger range.

Taylor's Theorem expresses that rigorously with the error term. If you stop at the first derivative, the amount of error depends on the second derivative, i.e. how fast the slope of the curve changes.
Interesting. This gives a formal proof that a linear function gives a perfect approximation, since f"(x)==0 for f(x)=mx+b.
 

Similar threads

  • · Replies 25 ·
Replies
25
Views
4K
  • · Replies 14 ·
Replies
14
Views
2K
  • · Replies 11 ·
Replies
11
Views
3K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 24 ·
Replies
24
Views
4K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 18 ·
Replies
18
Views
3K
  • · Replies 9 ·
Replies
9
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 13 ·
Replies
13
Views
2K