Having trouble understanding variance of OLS estimator

  • Context: Graduate 
  • Thread starter Thread starter chevrox
  • Start date Start date
  • Tags Tags
    Ols Variance
Click For Summary

Discussion Overview

The discussion centers around the variance of the OLS estimator (β-hat) in the context of linear regression models. Participants explore the implications of treating the matrix of independent variables (X) as non-stochastic and the resulting effects on the variance-covariance matrix of the estimator.

Discussion Character

  • Technical explanation
  • Conceptual clarification
  • Debate/contested

Main Points Raised

  • One participant expresses skepticism about the treatment of X as non-stochastic and questions how this affects the variance-covariance matrix of β-hat, particularly regarding its dependence on sample statistics.
  • Another participant clarifies that if X represents the independent variables, they are not considered stochastic when minimizing least squares error, suggesting a distinction between traditional regression and total least squares when errors are present in both X and Y.
  • A third participant agrees with the notion that for a consistent estimator, the variance should decrease with an increasing sample size, emphasizing the importance of this property for the utility of statistical estimation.
  • A later reply confirms understanding that X does not necessarily follow a stochastic process and discusses how the independent variables influence the dependent variable through observed values, maintaining the consistency of β-hat.

Areas of Agreement / Disagreement

Participants exhibit some agreement on the treatment of X as non-stochastic in traditional OLS regression, but there remains uncertainty regarding the implications of this assumption on the variance of the estimator and the interpretation of sample statistics.

Contextual Notes

Limitations include the dependence on the definition of stochastic processes and the implications of sample variability on the variance of the estimator, which are not fully resolved in the discussion.

chevrox
Messages
2
Reaction score
0
So in computing the variance-covariance matrix for β-hat in an OLS model, we arrive at

VarCov(β-hat)=(σ_ε)^2E{[X'X]^-1}

However, I'm incredulous as to how X is considered non-stochastic and how we can just eliminate the expectation sign and have

VarCov(β-hat)=(σ_ε)^2[X'X]^-1

I'm accepting this to be true (since it's so written in the text) but I'm taking a leap of faith here: if this is true, the elements in the VarCov matrix are expressed in terms of sample statistics and are therefore stochastic. I thought that the variance of an estimator of a parameter, if consistent, should be a deterministic parameter itself and should not depend on the sample observations (besides sample size, n), such as the ones we see in using Cramer-Rao lower bound to determine efficiency. Likely I'm understanding something wrong here, any pointers would be greatly appreciated!
 
Physics news on Phys.org
You haven't clearly stated a mathematical question. Is X supposed to be the vector of independent variables? If so, they aren't considered to be stochastic if you compute the regression so it minimizes the least square error in predicting the dependent variables, which are "Y", by tradition. If you have data of the form (X,Y) and there are "errors" in both X and Y, you should use a "total least squares" model.
 
Hey chevrox and welcome to the forums.

Like Stephen Tashi I am going wait for clarification of what your variables are but I did want to comment on one thing you said:

chevrox said:
I thought that the variance of an estimator of a parameter, if consistent, should be a deterministic parameter itself and should not depend on the sample observations (besides sample size, n), such as the ones we see in using Cramer-Rao lower bound to determine efficiency. Likely I'm understanding something wrong here, any pointers would be greatly appreciated!

That should definitely be the case for a consistent estimator and it should be the case that the variance 'shrinks' with a higher sample size. If the variance does not do this, then basically your estimate doesn't get 'better' with a higher sample size and it becomes rather pointless to do statistics with any kind of sample using that estimator.
 
Thanks for the replies! Yes, X is the nxk matrix of explanatory variables such that y=Xβ+ε. I think I understand it now. Variables in X do not necessarily follow a stochastic process, and even if they do, since all variability of y is explained by ε in the model, the independent variables affect the dependent variable solely through their observed values rather than a range of distribution where those values could fall, and it is therefore considered non-stochastic. And meanwhile β-hat does not lose its consistency since E(β-hat)=β (which is possible only if X is non-stochastic) and Var(β-hat)→0 even though Var(β-hat) varies with sample.
 

Similar threads

  • · Replies 7 ·
Replies
7
Views
3K
  • · Replies 23 ·
Replies
23
Views
4K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 1 ·
Replies
1
Views
4K
Replies
24
Views
6K
Replies
3
Views
1K
  • · Replies 1 ·
Replies
1
Views
3K
Replies
11
Views
6K