Can Alternative Least Squares Methods Be Used in Linear Regression?

Click For Summary

Discussion Overview

The discussion revolves around the application of alternative least squares methods in linear regression, specifically whether variants of least squares (LS) can be used to determine the best-fit line parameters instead of the traditional Ordinary Least Squares (OLS). The scope includes theoretical considerations and practical implications of different regression techniques.

Discussion Character

  • Exploratory
  • Technical explanation
  • Debate/contested

Main Points Raised

  • Some participants propose that OLS is based on specific assumptions, such as the random distribution of residuals, and question whether other LS methods can be applied to linear regression.
  • Others mention that the term "best-fit line" can have different meanings, particularly in relation to minimizing sum-squared errors versus minimizing perpendicular distances from data points.
  • A participant notes that robust regression methods exist to reduce the influence of outliers, referencing implementations in R and suggesting that these methods may provide alternatives to OLS.
  • Another viewpoint suggests that least squares is not the only method available, and one can minimize any penalty function, although this may involve more complex iterative algorithms.
  • Concerns are raised about the potential for local minima when using non-analytical methods for minimization.
  • Participants discuss the need for clarity around the definitions of "optimal" and "best fit," emphasizing that these terms require contextual understanding.

Areas of Agreement / Disagreement

Participants express a range of views on the applicability of alternative least squares methods, indicating that there is no consensus on whether these methods can effectively replace OLS in linear regression. The discussion remains unresolved regarding the definitions and implications of "optimal" and "best fit" in this context.

Contextual Notes

Limitations include the dependence on specific assumptions about residuals, the need for clarity in definitions of terms like "optimal," and the unresolved nature of the mathematical properties of alternative methods compared to OLS.

fog37
Messages
1,566
Reaction score
108
TL;DR
Understanding if linear regression can be done with other variants of least squares
Hello,

Simple linear regression aims at finding the slope and intercept of the best-fit line to for a pair of ##X## and ##Y## variables.
In general, the optimal intercept and slope are found using OLS. However, I learned that "O" means ordinary and there are other types of least square computations...

Question: is it possible to apply those variants of LS to the linear regression model, i.e., can we find the best-fit line parameters using something other than OLS (for example, I think there is "robust" LS, etc.)?

thank you!
 
Physics news on Phys.org
I have taken a course in regression and some courses that use regression techniques. From what I remember, the Ordinary in OLS refers to some assumptions we make, rather than the method
one assumption is: the residuals are randomly distributed.

I like this textbook (free download) https://www.statlearning.com/
 
  • Like
Likes   Reactions: BvU
fog37 said:
Simple linear regression aims at finding the slope and intercept of the best-fit line to for a pair of ##X## and ##Y## variables.
More precisely, it finds the line that uses the ##X## value to estimate the ##Y## values with the minimum sum-squared-errors for the ##Y## estimates. The phrase "best-fit line" can mean something different, referring to minimizing the sum-squared perpendicular distances from the data to the line.
fog37 said:
In general, the optimal intercept and slope are found using OLS. However, I learned that "O" means ordinary and there are other types of least square computations...

Question: is it possible to apply those variants of LS to the linear regression model, i.e., can we find the best-fit line parameters using something other than OLS (for example, I think there is "robust" LS, etc.)?
This is an interesting question. I am not an expert in this, but I see ( https://en.wikipedia.org/wiki/Robust_regression ) that there are attempts to decrease the influence of outliers. Some methods have been implemented in R (see https://stat.ethz.ch/R-manual/R-patched/library/MASS/html/rlm.html ). I don't know if that implementation is publicly available. It is applied in an example in https://stats.oarc.ucla.edu/r/dae/robust-regression/
 
  • Like
Likes   Reactions: fog37
Even least squares is not necessary. You can find a slope and intercept that minimize any penalty function you want.
 
  • Like
Likes   Reactions: fog37, scottdave and FactChecker
Office_Shredder said:
Even least squares is not necessary. You can find a slope and intercept that minimize any penalty function you want.
Good point, although most penalty functions would require non-analytical iterative minimization algorithms that are less intuitive. Also, I do not know what the risk of introducing local minimums would be.
 
  • Like
Likes   Reactions: fog37
fog37 said:
In general, the optimal intercept and slope are found using OLS.

When you read about this, what was the definition of "optimal"?

(A mathematical definition can be given in the context of statistical estimation and the properties of estimators.)
 
  • Like
Likes   Reactions: FactChecker
fog37 said:
TL;DR Summary: Understanding if linear regression can be done with other variants of least squares

Hello,

Simple linear regression aims at finding the slope and intercept of the best-fit line to for a pair of ##X## and ##Y## variables.
In general, the optimal intercept and slope are found using OLS. However, I learned that "O" means ordinary and there are other types of least square computations...

Question: is it possible to apply those variants of LS to the linear regression model, i.e., can we find the best-fit line parameters using something other than OLS (for example, I think there is "robust" LS, etc.)?

thank you!
The phrase "best fit line" is meaningless without some context.

In least squares regression the best fit line is the one that minimizes the sum of the squared residuals.

In robust regression based on the ideas that grew from Huber's M-estimate based ideas the residuals are not squared but are fed into some other function [designed to lessen the impact of certain kinds of outliers] and the best fit line is the one that minimizes the sum of the values of those function values.

In robust regression [at least the simplest methods, based on the initial work of Jana Jurečková, later expanded on by Joseph McKean and others] based on R-estimate [rank estimates] we view the squared residuals as (residual)(residual), replace one of the two factors by a weighted rank of the residual, and the best fit line is the one that minimizes the sum of those.
You should also look up the idea of Tukey's resistant line.

There are others, but the point is that phrases like "optimal" and "line of best fit" need to be placed into some context: optimal or best in what sense?
 
  • Like
Likes   Reactions: scottdave and FactChecker

Similar threads

  • · Replies 23 ·
Replies
23
Views
4K
Replies
3
Views
3K
  • · Replies 30 ·
2
Replies
30
Views
4K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 8 ·
Replies
8
Views
3K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 7 ·
Replies
7
Views
3K