Finding the Uncertainty of the Slope Parameter of a Liner Regression

Click For Summary
SUMMARY

This discussion focuses on calculating the uncertainty of the slope parameter (b) in linear regression when measurement errors are present in both x and y values. The common approach involves using the propagation of error formula and recognizing that ordinary least squares (OLS) regression does not account for errors in x measurements. The total least squares (TLS) regression method is suggested as an alternative that accommodates errors in both variables. The conversation highlights the complexity of estimating the standard deviation of b when only one sample value is available.

PREREQUISITES
  • Understanding of linear regression concepts, specifically ordinary least squares (OLS) and total least squares (TLS).
  • Familiarity with error propagation techniques and formulas.
  • Knowledge of statistical terms such as standard deviation and linear approximation.
  • Basic proficiency in mathematical notation and operations involving summations.
NEXT STEPS
  • Research the propagation of error formula and its application in linear regression contexts.
  • Study the differences between ordinary least squares (OLS) and total least squares (TLS) regression methods.
  • Learn about the estimation of standard deviation for slope parameters in the presence of measurement errors.
  • Explore advanced statistical techniques for handling errors in both independent and dependent variables.
USEFUL FOR

Statisticians, data analysts, and researchers involved in regression analysis, particularly those dealing with measurement errors in their datasets.

richardc
Messages
7
Reaction score
1
Finding the Uncertainty of the Slope Parameter of a Linear Regression

Suppose I have measurements x_i \pm \sigma_{xi} and y_i \pm \sigma_{yi} where \sigma is the uncertainty in the measurement. If I use a linear regression to estimate the value of b in y=a+bx, I'm struggling to find a straightforward way to compute the uncertainty of b that arises from the measurement uncertainties. This seems like it should be a very common problem, so I'm not sure why I can't find a simple algorithm or formula.

Thank you for any advice.
 
Last edited:
Physics news on Phys.org
Are you using "uncertainty" to mean "standard deviation"?

It's a common problem, but it's not simple. After all, your data gives only one value for b, so how can you estimate the standard deviation of b from a sample of size 1 ?

The common way to get an answer is to oversimplify matters and compute a "linearized asymptotic" estimate. The value of b is some function F of the (x_i,y_i). Let L be the linear approximation for the function F. Assume that near the observed values in the sample that this well approximates the random variable b as a linear combination of the x_i and y_i. When you have a random variable expressed as linear combination of other random variables, you can work on expressing its standard deviation in terms of the standard deviations of the other random variables.

That's the general picture. If it's what you want to do then we can try to look up the specifics. I don't know them from memory.
 
Thank you for clarifying the problem.

With N observation pairs I believe I can write b=\frac{N \sum x_i y_i - \sum x_i \sum y_i}{N \sum x_i^2 - (\sum x_i)^2}.

I suppose the propagation of error formula \sigma_f^2=\sum (\frac{\partial f}{\partial x_i} \sigma_{x_i} )^2 is then applied to a linear approximation of b?
 
You state a problem where there is an error in measurement for x_i as well as for y_i. In such a problem, people often use "total least squares" regression. I think the computation of the slope in "total least squares" regression is different than in ordinary least square regression, which assumes no error in the measurement of the x_i. I think the formula you gave for b is for ordinary least squares regression.

Of course, one may ask the question: If I fit a straight line to data using the estimator for slope used in ordinary least squares regression and my data also has errors in the x_i then what is the standard deviation of this estimator. If that's the question, you need terms involving \frac{\partial f}{\partial y_i} \sigma^2_{y_i} and \frac{\partial f}{\partial x_i} \sigma^2 x_i

I don't know if the estimator for slope in ordinary least squares regression is an unbiased estimator if there are errors in the x_i.
 
If there are an infinite number of natural numbers, and an infinite number of fractions in between any two natural numbers, and an infinite number of fractions in between any two of those fractions, and an infinite number of fractions in between any two of those fractions, and an infinite number of fractions in between any two of those fractions, and... then that must mean that there are not only infinite infinities, but an infinite number of those infinities. and an infinite number of those...

Similar threads

  • · Replies 11 ·
Replies
11
Views
5K
  • · Replies 23 ·
Replies
23
Views
4K
  • · Replies 19 ·
Replies
19
Views
3K
  • · Replies 64 ·
3
Replies
64
Views
5K
  • · Replies 13 ·
Replies
13
Views
2K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 2 ·
Replies
2
Views
1K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 21 ·
Replies
21
Views
162K
Replies
15
Views
3K