Error on using a curve of best fit for extrapolation

Click For Summary
The discussion centers on the challenges of calculating uncertainty when using an exponential curve of best fit for extrapolation. Participants suggest methods for determining uncertainty, including using the least-squares method to derive uncertainties in parameters and applying error propagation formulas. It is emphasized that the mean of the logarithm of y does not equal the logarithm of the mean, which can introduce bias in the fitting process. Additionally, the conversation highlights the importance of considering measurement uncertainties in the extrapolation process. Overall, the exchange provides insights into more accurate methods for incorporating uncertainty in data analysis.
Woland
Messages
18
Reaction score
0
Hello everyone,

I did a quick search but could not find this in the forums.

I have quite a basic situation. I have been gathering data points from an experiment and was able to fit an exponential curve of best fit to it. What I want to do is approximate some values between my data points using this curve. My approach is to just use the equation of the curve and plug in the desired number. I will then add the uncertainty to it.

This uncertainty is what I am drawing a blank with. Does anyone know what method I could use to calculate this uncertainty. I have a feeling this is really simple, but... I can't remember anything yet.

Any help would be appreciated.
 
Physics news on Phys.org
Woland said:
Hello everyone,

I did a quick search but could not find this in the forums.

I have quite a basic situation. I have been gathering data points from an experiment and was able to fit an exponential curve of best fit to it. What I want to do is approximate some values between my data points using this curve. My approach is to just use the equation of the curve and plug in the desired number. I will then add the uncertainty to it.

This uncertainty is what I am drawing a blank with. Does anyone know what method I could use to calculate this uncertainty. I have a feeling this is really simple, but... I can't remember anything yet.

Any help would be appreciated.


If your curve is:

y=exp(c*x)

Your extrapolation error will be roughly:

Y=exp(sigma_c (x2-x1))+sigma_yo

Where:
sigma_c is the standard deviation in your estimate of c.
x1 is the point you are extrapolating from
x2 is the point you are extrapolating to.
sigma_yo is your initial error.
 
which method are you using to get the error equation?

I was thinking of something along the lines of:
If I have y = A exp(Bx) as my best fit, I can take the ln of both sides:

ln y = B lnA + B ln x

I can now do a line of best fit and find sigma lny, sigma A and sigma B.
These are the uncertainties in the fit which I get from the least - squares method. I can now use the error propagation formula to get the uncertainty in y.

My question is, where do I incorporate the uncertainty in my measurments.
 
Woland said:
which method are you using to get the error equation?

I was thinking of something along the lines of:
If I have y = A exp(Bx) as my best fit, I can take the ln of both sides:

ln y = B lnA + B ln x

I can now do a line of best fit and find sigma lny, sigma A and sigma B.
These are the uncertainties in the fit which I get from the least - squares method. I can now use the error propagation formula to get the uncertainty in y.

My question is, where do I incorporate the uncertainty in my measurments.

Well, let's say y=<y(x)>+rv
Where:
<y(x)> is the mean value of y at x and rv is a random variable.

Lets assume for a sec that rv is a quasian distribution then:

P(y)=integral{-00,-00}1/(sigma*2pi)*exp(-(y-y_bar)^2/(2*sigma^2))dy

To get ln y do a change of variables on the above expression for P(y)

u=ln(y) <=> y=exp(u)

The term inside the integrand will be your new distribution function for ln y

See functions of random variables:
http://cnx.org/content/m11066/latest/

Keep in mind that
<ln y>
does not equal
ln <y>

Thus a least mean squared fit of ln y will be biased. Measurements where y is much smaller then the standard deviation of the random variable will be more biased. You can do the transformation before the fit but if your error is large compared to your measurement then you’ll need to use maximum likelihood. If you use least squares then weight large values of y exponentially more then small values of y.

Also note that:

if <y>=Aexp(BX)
Then:
<ln y> does not equal: BX + ln(A)

Rather it is equal to <ln (<y(x)> +rv)>

Which we can’t really separate neatly into the form you gave.
 
Ah thank you for clarifying. I knew something was too easy about my approach. I don't have time to try this now, but I am sure it will work. Thank you.
 
The standard _A " operator" maps a Null Hypothesis Ho into a decision set { Do not reject:=1 and reject :=0}. In this sense ( HA)_A , makes no sense. Since H0, HA aren't exhaustive, can we find an alternative operator, _A' , so that ( H_A)_A' makes sense? Isn't Pearson Neyman related to this? Hope I'm making sense. Edit: I was motivated by a superficial similarity of the idea with double transposition of matrices M, with ## (M^{T})^{T}=M##, and just wanted to see if it made sense to talk...

Similar threads

  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 28 ·
Replies
28
Views
3K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 16 ·
Replies
16
Views
3K
Replies
28
Views
4K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 30 ·
2
Replies
30
Views
4K
Replies
8
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K