The linear in linear least squares regression

Click For Summary

Discussion Overview

The discussion revolves around the concept of linearity in linear least squares regression, particularly in relation to fitting various types of functions (e.g., quadratic, cubic) and the implications of normality and orthogonality in this context. Participants explore the conditions under which linear least squares can be applied and the nature of the coefficients involved.

Discussion Character

  • Debate/contested
  • Technical explanation
  • Conceptual clarification

Main Points Raised

  • Some participants propose that linearity in least squares refers specifically to the coefficients, allowing for fitting of non-linear functions through linear equations.
  • Others argue that normality is not a strict requirement for obtaining beta terms, suggesting that finite variance and a non-singular covariance matrix are sufficient.
  • A participant mentions that while normality is not essential for unbiased estimates, it is important for hypothesis testing in small samples.
  • One participant points out that the Gauss-Markov theorem provides good properties for least-squares estimates under certain conditions, indicating that normality is not critical for curve fitting.
  • There is a discussion about the distinction between linearity in parameters and the ability to fit non-linear models, with examples provided to illustrate this point.
  • Another participant expresses uncertainty about their understanding but emphasizes that the linearity of coefficients does not imply that error terms must be normally distributed.

Areas of Agreement / Disagreement

Participants do not reach a consensus on the necessity of normality for linear least squares regression or the implications of linearity in coefficients. Multiple competing views remain regarding the relationship between these concepts.

Contextual Notes

Some statements reflect a reliance on specific assumptions about error terms and the nature of the models being discussed, which may not be universally applicable. The discussion includes varying interpretations of the implications of linearity and normality in the context of least squares regression.

joshthekid
Messages
46
Reaction score
1
It is my understanding that you can use linear least squares to fit a plethora of different functions (quadratic, cubic, quartic etc). The requirement of linearity applies to the coefficients (i.e B in (y-Bx)^2). It seems to me that I can find a solution such that a coefficient b_i^2=c_i, in other words I can just express the squared b_i with a linear c_i. So can't I always find a linear solution?

My gut feeling is that linearity is required because normality is required for the orthoganality principle to hold? But I am not sure.

Thanks
 
Physics news on Phys.org
joshthekid said:
It is my understanding that you can use linear least squares to fit a plethora of different functions (quadratic, cubic, quartic etc). The requirement of linearity applies to the coefficients (i.e B in (y-Bx)^2). It seems to me that I can find a solution such that a coefficient b_i^2=c_i, in other words I can just express the squared b_i with a linear c_i. So can't I always find a linear solution?

My gut feeling is that linearity is required because normality is required for the orthoganality principle to hold? But I am not sure.

Thanks
I'm not sure what your question is here.

The least squares method for fitting a set of data to a general polynomial function (linear, quadratic, cubic, etc.) results in a set of linear equations which must be solved to determine the unknown coefficients of the polynomial. The number of equations to be solved is equal to the degree of the polynomial plus 1.
 
Normality is only required in OLS for accurate testing of error terms. As long as you have finite variance and a non-singular covariance matrix you can get the beta terms, nothing is required to be orthogonal.
 
It has nothing to with normality; linearity in the parameters just let's you derive a solution using linear algebra. It's entirely possible to fit non-linear functions by least squares, it just requires numerical methods.

The model
[tex]y = \alpha + \beta^2x + \epsilon[/tex]
is linear in the parameters [itex](\alpha, \beta^2)[/itex], so there's no problem. Conversely, the model
[tex]y = \alpha e^{\beta x} + \epsilon[/tex]
is clearly non-linear in [itex]\beta[/itex].

Normality is only required in OLS for accurate testing of error terms. As long as you have finite variance and a non-singular covariance matrix you can get the beta terms.

This is mostly true. Strictly speaking, the Gauss-Markov theorem guarantees that the least-squares estimates have good properties assuming only zero-mean, equal variance, and uncorrelated errors, so the normality assumption isn't so important if you just want to curve fit, but modelling is easier assuming normal errors. Under the usual normality assumptions, the least-squares estimates are the maximum likelihood estimates, which are better understood than the more general least-squares estimates. It also makes in easier test hypotheses about the coefficients.
 
Last edited:
I am not completely sure about my answers, but I try my best.

You do not need normally distributed error terms for an OLS estimator to be unbiased, nor is that the reason that you have linearity in coefficients. According to the central limit theorem your estimates will be normally distributed in the limit anyway. However, you want to have normally distributed error terms for hypothesis testing in small samples. Overall the assumption of normality has something to do with the assumption of the classical linear model in general an not with the linearity of its estimates (for an elaboration on the ols assumption see for example http://economictheoryblog.com/2015/04/01/ols_assumptions).

I think the answer to you question is far more technical, in the sense that what you get for your coefficient is a scalar which is constant per se. In that sense it cannot be a function of an non linear form. However, that is my own understanding of this subject and I would love to hear another answer to this question if there is one.

Cheers,
Stefan
 

Similar threads

  • · Replies 6 ·
Replies
6
Views
3K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 7 ·
Replies
7
Views
3K
Replies
3
Views
3K
  • · Replies 8 ·
Replies
8
Views
3K
  • · Replies 23 ·
Replies
23
Views
4K
  • · Replies 30 ·
2
Replies
30
Views
5K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 8 ·
Replies
8
Views
3K
  • · Replies 5 ·
Replies
5
Views
2K