Least squares assumptions: finite and nonzero 4th moments

slakedlime
Messages
74
Reaction score
2
This isn't a homework problem - I'm just confused by something in a textbook that I'm reading (not for a class, either). I'd appreciate an intuitive clarification, or a link to a good explanation (can't seem to find anything useful on Google or in my textbook).

My book states that one of the least squares assumptions (e.g. for ordinary least squares, OLS, estimation) is that large outliers are unlikely.

That is, for the following equation:
Y_{i} = β_{0}+β_{1}X_i+u_{i}

It must be that (X_{i}, Y_{i}), i = 1, ..., n have nonzero finite fourth moments.

Why is this significant? What is the relationship between large outliers and nonzero finite fourth moments? I don't intuitively see the mathematical explanation. Any help and/or direction is much appreciated.
 
Physics news on Phys.org
The real importance of the fourth moment statement is that with it in place the underlying ideas needed for consistent estimation of variances are easily verified.
The argument that links the finite fourth moments to outliers can be intuitively stated as: if the fourth moments are finite, then the tails of the distribution are relatively short, so the PROBABILITY of unusually large observations occurring is small. In that regard it's an assumption made to try to account for the fact that least squares regression (least squares methods in general) are non-robust and results are very sensitive to the presence of outliers.
The better notion is: if you believe outliers could be an issue, use a method for robust regression.

The notion you ask about is discussed in this article.
http://www.aw-bc.com/info/stock_watson/Chapter4.pdf
 
Thank you so much!
 
thank you
 
statdad said:
The real importance of the fourth moment statement is that with it in place the underlying ideas needed for consistent estimation of variances are easily verified.
The argument that links the finite fourth moments to outliers can be intuitively stated as: if the fourth moments are finite, then the tails of the distribution are relatively short, so the PROBABILITY of unusually large observations occurring is small. In that regard it's an assumption made to try to account for the fact that least squares regression (least squares methods in general) are non-robust and results are very sensitive to the presence of outliers.
The better notion is: if you believe outliers could be an issue, use a method for robust regression.

The notion you ask about is discussed in this article.
http://www.aw-bc.com/info/stock_watson/Chapter4.pdf

Access is forbidden!
 
Prove $$\int\limits_0^{\sqrt2/4}\frac{1}{\sqrt{x-x^2}}\arcsin\sqrt{\frac{(x-1)\left(x-1+x\sqrt{9-16x}\right)}{1-2x}} \, \mathrm dx = \frac{\pi^2}{8}.$$ Let $$I = \int\limits_0^{\sqrt 2 / 4}\frac{1}{\sqrt{x-x^2}}\arcsin\sqrt{\frac{(x-1)\left(x-1+x\sqrt{9-16x}\right)}{1-2x}} \, \mathrm dx. \tag{1}$$ The representation integral of ##\arcsin## is $$\arcsin u = \int\limits_{0}^{1} \frac{\mathrm dt}{\sqrt{1-t^2}}, \qquad 0 \leqslant u \leqslant 1.$$ Plugging identity above into ##(1)## with ##u...

Similar threads

Back
Top