Why is/was consistency of estimators desired?

  • Context: Graduate 
  • Thread starter Thread starter Stephen Tashi
  • Start date Start date
  • Tags Tags
    Estimators
Click For Summary

Discussion Overview

The discussion revolves around the concept of "consistency" in statistical estimators, exploring its historical context and modern interpretations. Participants examine the definitions and implications of consistency, particularly in relation to how estimators relate to population parameters, and the conditions under which they are deemed consistent.

Discussion Character

  • Exploratory
  • Technical explanation
  • Historical

Main Points Raised

  • One participant references Fisher's criteria for judging statistics, noting that consistency involves calculating the statistic in the same way for both sample and population.
  • Another participant questions whether the idea of consistency implies that the probability of the estimate being near the true value approaches 1.0 as sample size increases.
  • A different participant adds that consistency can be defined as a sequence of estimators converging in probability to the true value of the parameter, suggesting there may be equivalent definitions.
  • One participant discusses the importance of variance converging to 0 as sample size increases, and the relationship between the distributions of the population parameter and the estimated parameter.
  • Another participant seeks to clarify the historical definition of consistency, suggesting it requires the estimator to be computed in the same way as the parameter it estimates, and questions when the modern definition became dominant.
  • There is a mention that the unbiased estimator for variance in a Gaussian distribution may not be consistent under the old definition, as it is not computed in the same way as the population parameter.

Areas of Agreement / Disagreement

Participants express varying interpretations of the concept of consistency, with some focusing on historical definitions and others on modern interpretations. There is no consensus on the precise implications of these definitions or their historical evolution.

Contextual Notes

Participants highlight the potential ambiguity in the definitions of consistency, particularly regarding the conditions under which estimators are considered consistent. The discussion reflects a mix of modern and historical perspectives, with unresolved questions about the transition between these definitions.

Stephen Tashi
Science Advisor
Homework Helper
Education Advisor
Messages
7,864
Reaction score
1,605
Why is/was "consistency" of estimators desired?

In an article, I found while researching another thread ("Revisiting a 90-year-old debate: the advantages of the mean deviation", http://www.leeds.ac.uk/educol/documents/00003759.htm ), the author states this bit of statistics history:

Fisher had proposed that the quality of any statistic could be judged in terms of three characteristics. The statistic, and the population parameter that it represents, should be consistent (i.e. calculated in the same way for both sample and population). The statistic should be sufficient in the sense of summarising all of the relevant information to be gleaned from the sample about the population parameter. In addition, the statistic should be efficient in the sense of having the smallest probable error as an estimate of the population parameter. Both SD and MD meet the first two criteria (to the same extent). According to Fisher, it was in meeting the last criteria that SD proves superior.

I recognize the description of "sufficient" and "efficient" as modern criteria. But the description of "consistent" seems rather simple minded. Was the idea of "consistent" that if the estimator and the population parameter were calculated "in the same way" that the probability of the estimate being near true value of the parameter would approach 1.0 as the sample size approached infinity?
 
Physics news on Phys.org


Was the idea of "consistent" that if the estimator and the population parameter were calculated "in the same way" that the probability of the estimate being near true value of the parameter would approach 1.0 as the sample size approached infinity?

I glanced at my old mathematical statistics textbook, and it defines a sequence of estimators (in most cases, taken as sample size increases) as consistent if it converges in probability to the true value of the parameter, which is the only way I remember ever seeing it defined anywhere. I assume there are some other equivalent definitions.
 


My understanding is pretty much the same as Number Nine's with the exception that it is quantified in terms of the variance converging to 0 for some estimator as the number of samples reaches the size of the population: in something like a census, this is finite but for a theoretical distribution, it's infinite.

In terms of things being calculated "the same way", it would seem that there would be some similarity between the population parameter and the estimated parameter's distribution since they are both based on the same underlying PDF, but I'd be interested to here any further comments on this.

I guess the only other thing though that I see as important is the actual nature of the convergence as opposed to the condition that convergence simply exists.

Typically the way this is looked at is in terms of how the variance changes with an increasing sample size, but I would think that it's equally important to see how P(X = x) changes as n -> infinity rather than how just the variance changes.
 


I understand (or can understand if I read carefully) the modern definition of "consistency" for an estimator. My original post is mainly about the old fashioned definition of consistency that says the estimator must be computed "in the same way" as the parameter that it estimates.

(An interesting historical question is "When did the modern definition of consistency" supercede the old one?".)

I think the condition "in the same way" can be made precise by saying we compute a (old fashioned) consistent estimator for the parameter P by treating the sample as a population (i.e. as defining a distribution) and define the estimate by the same formula as we define the parameter P.

If that's what was meant in olden times, then technically the unbiased estimator for the variance of a Gaussian distribution was not consistent since it is not computed "in the same way" as the population parameter.
 

Similar threads

  • · Replies 7 ·
Replies
7
Views
3K
  • · Replies 23 ·
Replies
23
Views
4K
  • · Replies 3 ·
Replies
3
Views
4K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 21 ·
Replies
21
Views
4K
Replies
1
Views
2K
  • · Replies 7 ·
Replies
7
Views
3K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 15 ·
Replies
15
Views
3K