Why can t statistic deal with small numbers ?

In summary: The central limit theorem states that as the sample size increases, the sampling distribution of the sample mean approaches a normal distribution, regardless of the underlying population distribution. This means that as the sample size increases, the estimate of the population variance will also approach the true population variance. In summary, the t-test allows us to deal with smaller samples because it uses the t-distribution, which takes into account the variability of smaller sample sizes. The z-test, on the other hand, assumes a larger sample size and a normal distribution of the population. The central limit theorem also shows that as the sample size increases, the estimate of the population variance will approach the true population variance.
  • #1
thrillhouse86
80
0
Hi,

I've been trying to get my head around z and t statistics. and I almost have a matra in my head that "when the sample are small, use the t test, when the samples are big, use either the t or the z test".

Now As I understand it, the z test requires a large number of samples, because it assumes you have a normal distribution, and you need a certain number of samples before your samples will start to look like a normal distribution.

But Why does the t test allow us to deal with smaller samples ? what does it have (or what assumptions doesn't it have) which allow us to deal with smaller samples ?

Is it that in the z test the standard error of the mean distribution is determined from the KNOWN population variance, whereas the t test the standard error of the mean distribution is determined from the ESTIMATE of the population variance, and in the limit of large number of samples the ESTIMATE of the population variance will approach the TRUE population variance ?

If this is indeed the case, does the central limit theorem show us that in the limit of a large number of samples the estimate of the population variance will approach the true population variance ?

Thanks
 
Last edited:
Physics news on Phys.org
  • #2
""when the sample are small, use the t test, when the samples are big, use either the t or the z test"." is really old advice, originated before calculators and even computers brought cheap and easy calculations to us.

The origin of the idea: Gossett (the developer of the t-procedure) found that when sample sizes were small, estimates based on the normal distribution (the Z-test and Z-confidence interval) gave results that didn't match experimental observations. (Statistics using th sample mean and standard deviation were more variable for small samples than the normal distribution "expected".) Methods based on the "t"-distribution were developed empirically to circumvent this. By the time samples sizes were around 30, results from the two procedures were in general agreement, and this observation grew into the "small sample size" vs "large sample size" distinction. That was convenient: suppose you always create a 95% confidence interval. When you use the t-distribution intervals, you need a different critical value for each sample size, and many years ago this required tables. When you use the Z-distribution intervals, a single critical value works no matter what the sample size.

"Is it that in the z test the standard error of the mean distribution is determined from the KNOWN population variance"
It doesn't have to be - that really was the point of the distinction. If you are performing a hypothesis test, and all you have are the sample size, sample mean, and sample standard deviation, the form of the test statistic is

[tex]
\frac{\overline x - \mu_0}{\frac s {\sqrt{n}}}
[/tex]

For "small samples" you would compare this to critical values from the appropriate t-distribution: for "large samples" you compare it to a value from the standard normal distribution.

If you actually had the true population standard devation, and you were sure the underlying population were normal , then the test statistic would be

[tex]
\frac{\overline x - \mu_0}{\frac{\sigma}{\sqrt n}}
[/tex]

and you would compare it to a critical value from the normal distribution.

"If this is indeed the case, does the central limit theorem show us that in the limit of a large number of samples the estimate of the population variance will approach the true population variance ?"

Yes.
 

1. Why is it important to have a large sample size when using statistics?

Having a large sample size allows for more accurate and reliable results. With a larger sample size, the data is more representative of the entire population and reduces the chances of errors and biases affecting the results.

2. Can't statistical methods be applied to small numbers?

Yes, statistical methods can be applied to small numbers, but they may not be as effective. Small sample sizes can lead to unreliable and misleading results due to the lack of variability and representation of the population.

3. How does the power of a statistical test change with small numbers?

The power of a statistical test decreases with smaller sample sizes. This means that it is less likely to detect a true difference or relationship between variables. In order to have a higher power and increase the chances of finding significant results, a larger sample size is needed.

4. Why is the margin of error greater with small numbers?

The margin of error is greater with small numbers because there is less data to work with. This means that there is a larger range of possible values, leading to a wider margin of error. As the sample size increases, the margin of error decreases as the data becomes more precise.

5. Are there any alternatives to using statistics with small numbers?

Yes, there are alternatives to using statistics with small numbers. One option is to use non-parametric methods, which do not rely on assumptions about the data and can be more effective with smaller sample sizes. Another option is to use qualitative research methods, such as interviews or case studies, which can provide valuable insights and understanding without the need for statistical analysis.

Similar threads

  • Set Theory, Logic, Probability, Statistics
Replies
7
Views
442
  • Set Theory, Logic, Probability, Statistics
Replies
9
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
30
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
27
Views
3K
  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
4
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
5
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
7
Views
4K
  • Set Theory, Logic, Probability, Statistics
Replies
17
Views
2K
Back
Top