- #1
fog37
- 1,569
- 108
- TL;DR Summary
- estimators and their properties and relation to hypothesis testing
Hello,
An estimator is a random variable, i.e. a function that assigns a number to a random sample collected from a population with unknown parameters. More practically, an estimator is really a formula to calculate an estimated coefficient ##b## using the data from our single random sample...Some estimators are linear, if they are linear functions of the dependent variable ##Y##, some are nonlinear.
We want our estimator to be unbiased: collecting many many many samples, we calculate their many estimates and the arithmetic average of those estimate is equal to the true population parameter. That is really good even if our estimate is not exactly equal to the population parameter...Knowing that we used an unbiased estimator gives up more confidence on the "quality" of our estimate....
In the class of linear unbiased estimators, OLS estimators have minimum variance as long as the Gauss-Markov assumptions are met. That leads to a small standard error of the sampling distribution of the estimates and to tighter confidence interval, hence the least possible uncertainty. An unbiased estimator with the least variance is called an efficient estimator. Under the assumed conditions, OLS estimators are BLUE (best, linear, unbiased, estimators).
If, in addition to the GM assumptions we can require the errors to be normally distributed, the OLS estimators become themselves normally distributed and we can safely use hypothesis testing... With the added assumption of normality, the OLS estimators are best unbiased estimators (BUE) in the entire class of unbiased estimators among all linear and all nonlinear estimators...That is a big result because, without the normality assumptions, the OLS estimators are only best among the linear estimators.
That said, here my dilemmas:
Question: we only work with a single random sample of size ##n## and use fancy math to indirectly learn about the sampling distribution of the estimates without collecting millions of samples. If the estimators are BLUE, it essentially means that, even if our sample estimate is not really exactly equal to the population parameter, it is a "good quality" estimate. An estimator being BLUE gives us confidence that our estimation procedure is good and produces statistically reliable estimates, correct?
Question: what do we gain from hypothesis testing? Do we get to verify, based on our limited data, if our estimate is statistically sound? For example, in linear regression, we test if the slope is equal to zero against what we get from our single sample calculation.
For example, image we get an estimate from a BLUE estimator. That is good because our estimate ##b## is close to the real, true population value ##\beta## and on average, becomes equal to it! What if our hypothesis test, because of lack of error normality, gives us statistically not significant results? Does that make our good, unbiased, low variance estimate fall apart since the results we get from the tests tell us that our sample estimate don't really reflect what is going on in the population?
I am confused about what it means when getting an estimate from a BLUE estimator but the hypothesis test support ##H_0## that the estimated slope is zero....
Thank you!
An estimator is a random variable, i.e. a function that assigns a number to a random sample collected from a population with unknown parameters. More practically, an estimator is really a formula to calculate an estimated coefficient ##b## using the data from our single random sample...Some estimators are linear, if they are linear functions of the dependent variable ##Y##, some are nonlinear.
We want our estimator to be unbiased: collecting many many many samples, we calculate their many estimates and the arithmetic average of those estimate is equal to the true population parameter. That is really good even if our estimate is not exactly equal to the population parameter...Knowing that we used an unbiased estimator gives up more confidence on the "quality" of our estimate....
In the class of linear unbiased estimators, OLS estimators have minimum variance as long as the Gauss-Markov assumptions are met. That leads to a small standard error of the sampling distribution of the estimates and to tighter confidence interval, hence the least possible uncertainty. An unbiased estimator with the least variance is called an efficient estimator. Under the assumed conditions, OLS estimators are BLUE (best, linear, unbiased, estimators).
If, in addition to the GM assumptions we can require the errors to be normally distributed, the OLS estimators become themselves normally distributed and we can safely use hypothesis testing... With the added assumption of normality, the OLS estimators are best unbiased estimators (BUE) in the entire class of unbiased estimators among all linear and all nonlinear estimators...That is a big result because, without the normality assumptions, the OLS estimators are only best among the linear estimators.
That said, here my dilemmas:
Question: we only work with a single random sample of size ##n## and use fancy math to indirectly learn about the sampling distribution of the estimates without collecting millions of samples. If the estimators are BLUE, it essentially means that, even if our sample estimate is not really exactly equal to the population parameter, it is a "good quality" estimate. An estimator being BLUE gives us confidence that our estimation procedure is good and produces statistically reliable estimates, correct?
Question: what do we gain from hypothesis testing? Do we get to verify, based on our limited data, if our estimate is statistically sound? For example, in linear regression, we test if the slope is equal to zero against what we get from our single sample calculation.
For example, image we get an estimate from a BLUE estimator. That is good because our estimate ##b## is close to the real, true population value ##\beta## and on average, becomes equal to it! What if our hypothesis test, because of lack of error normality, gives us statistically not significant results? Does that make our good, unbiased, low variance estimate fall apart since the results we get from the tests tell us that our sample estimate don't really reflect what is going on in the population?
I am confused about what it means when getting an estimate from a BLUE estimator but the hypothesis test support ##H_0## that the estimated slope is zero....
Thank you!