Undergrad Why linear in linear regression?

Click For Summary
In linear regression, the parameters are estimated to be linear with respect to the dependent variable, but the functions themselves can be affine rather than strictly linear. The term "linear" is often used for convenience, despite the presence of translations in affine functions. When considering the tangent line at a point on a curve, it becomes a linear representation in the tangent space, highlighting the context-dependent nature of linearity. The distinction between linear and affine becomes significant in global coordinate systems and geometric contexts, but for practical purposes in regression, the focus is on linearity in parameters rather than the functional form itself. Thus, while the terminology may be misleading, it serves a functional role in statistical modeling.
schniefen
Messages
177
Reaction score
4
TL;DR
It appears to be common to say linear regression, but is this correct?
In linear regression, one estimates parameters that are supposed to be linear with respect to the dependent variable, for instance

##y=\theta_0 e^x+\epsilon \ ,##

or

##y=\theta_0+\theta_1 x_1+\theta_2x_2+...+\theta_n x_n+\epsilon \ . ##
Is it not true that neither ##y(\theta_0)## nor ##y(\theta_0,...,\theta_n)## are linear functions, but rather affine functions of the parameters?
 
Physics news on Phys.org
Affine linear is often abbreviated by calling it linear. It's for convenience, and it ##\epsilon=0## in your examples, then it is linear. However, it's not only convenience. E.g. if we consider the tangent line at ##x=2## at the curve ##y=x^2##, then it is an affine linear object on the surrounding ##x,y-##plane. But we also speak of the tangent space at ##x=2##. In that case we have implicitly identified the point ##(2,4)## with the origin of the tangent space and all of a sudden the affine linear line in the plane, became a linear line in the tangent space.

So, as long as not both terms are necessary in a certain context, affine linear is often just called linear.
 
Last edited:
  • Like
Likes DaveE, Dale and schniefen
Could you explain how the line ##x=2## is "an affine linear object on the surrounding ##x,y##-plane"? An affine transformation is a linear transformation plus a translation.

If ##\epsilon \neq 0##, are then both terms necessary?
 
schniefen said:
Could you explain how the line ##x=2## is "an affine linear object on the surrounding ##x,y##-plane"? An affine transformation is a linear transformation plus a translation.

If ##\epsilon \neq 0##, are then both terms necessary?
Both terms are necessary if you have a global coordinate system and you perform geometry. Then the distinction is reasonable. If you only want to say: not curved, then linear will do.
 
schniefen said:
TL;DR Summary: It appears to be common to say linear regression, but is this correct?

In linear regression, one estimates parameters that are supposed to be linear with respect to the dependent variable, for instance

##y=\theta_0 e^x+\epsilon \ ,##

or

##y=\theta_0+\theta_1 x_1+\theta_2x_2+...+\theta_n x_n+\epsilon \ . ##
Is it not true that neither ##y(\theta_0)## nor ##y(\theta_0,...,\theta_n)## are linear functions, but rather affine functions of the parameters?
In regression you make an assumption about the form of the functional relationship between your response and any predictor(s). LINEAR regression doesn't require the function to be itself linear, only that it be linear in the parameters to be estimated.

This expression would qualify as a functional form we'd lump into linear regression.

## y = \beta_0 + \beta_1 x_1^2 + \beta_2 \frac{x_2}{x_2^2 + 5} ##

This is not a linear function of the parameters.

## y = \beta_0 e^{\beta_1 x_1 + \beta_2 x_2} ##

since, for given values of the predictors it is a linear function of the betas. You don't need to worry about the affine stuff.
 
The standard _A " operator" maps a Null Hypothesis Ho into a decision set { Do not reject:=1 and reject :=0}. In this sense ( HA)_A , makes no sense. Since H0, HA aren't exhaustive, can we find an alternative operator, _A' , so that ( H_A)_A' makes sense? Isn't Pearson Neyman related to this? Hope I'm making sense. Edit: I was motivated by a superficial similarity of the idea with double transposition of matrices M, with ## (M^{T})^{T}=M##, and just wanted to see if it made sense to talk...

Similar threads

  • · Replies 13 ·
Replies
13
Views
4K
Replies
3
Views
3K
  • · Replies 6 ·
Replies
6
Views
3K
  • · Replies 23 ·
Replies
23
Views
4K
  • · Replies 64 ·
3
Replies
64
Views
5K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 7 ·
Replies
7
Views
3K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 9 ·
Replies
9
Views
2K