Linear regression and bivariate normal, is there a relationship?

CantorSet
Messages
44
Reaction score
0
Hi everyone,

This is not a homework question. I just want to understand an aspect of linear regression better. The book "Applied Linear Models" by Kutchner et al, states that a linear regression model is of the form

Y_i = B_0 + B_1 X_i + \epsilon_i

where
Y_i is the value of the response variable in the ith trial
B_0, B_1 are parameters
X_i is a known constant
\epsilon_i is a random variable, normally distributed.
Therefore, Y_i is also a random variable, normally distributed but X_i is a constant.

This confused me a bit because I always associated linear regression with the bivariate normal distribution. That is, the underlying assumption of linear regression is the data \{(x_1,y_1), (x_2,y_2),...,(x_n,y_x) \} is sampled from a bivariate normal distribution. In which case, both X and Y are random variables. But in the formulation above, X is a known constant, while \epsilon and therefore Y are the random variables.

So in summary, what is the connection (if any) is between linear regression as formulated by Kutner and the bivariate normal.
 
Physics news on Phys.org
CantorSet said:
the underlying assumption of linear regression is the data \{(x_1,y_1), (x_2,y_2),...,(x_n,y_x) \} is sampled from a bivariate normal distribution. In which case, both X and Y are random variables.

I've never seen a treatment of regression that made that assumption. Are you confusing linear regession with some sort of "total least squares" regression?
http://en.wikipedia.org/wiki/Total_least_squares
 
Stephen Tashi said:
I've never seen a treatment of regression that made that assumption. Are you confusing linear regession with some sort of "total least squares" regression?
http://en.wikipedia.org/wiki/Total_least_squares

Thanks for responding, Stephen.

Yea, that was my own confusion for making that assumption. Thanks for clearing that up.

By the way, total least squares is just a generalization of linear regression in that the curve you're fitting the data points to can be polynomials with degrees higher than 1, right? Or is there more to total least squares?
 
Total least squares treats both X and Y as random variables.
 
Namaste & G'day Postulate: A strongly-knit team wins on average over a less knit one Fundamentals: - Two teams face off with 4 players each - A polo team consists of players that each have assigned to them a measure of their ability (called a "Handicap" - 10 is highest, -2 lowest) I attempted to measure close-knitness of a team in terms of standard deviation (SD) of handicaps of the players. Failure: It turns out that, more often than, a team with a higher SD wins. In my language, that...
Hi all, I've been a roulette player for more than 10 years (although I took time off here and there) and it's only now that I'm trying to understand the physics of the game. Basically my strategy in roulette is to divide the wheel roughly into two halves (let's call them A and B). My theory is that in roulette there will invariably be variance. In other words, if A comes up 5 times in a row, B will be due to come up soon. However I have been proven wrong many times, and I have seen some...
Back
Top