Why Do Degrees of Freedom Differ in Chi-Squared Distributions?

In summary, the degrees of freedom in the Chi-squared distribution can vary depending on whether the population parameters are known or unknown and whether the sample mean or the known mean is used. This difference in degrees of freedom is necessary to compensate for the inaccuracy in using the sample mean to estimate the population parameters.
  • #1
Usjes
9
0
Hi,

I am trying to understand the degrees of freedom parameter in the Chi_squared distribution and I have found two references from the same source that appear, to me, to contradict one-another. Can anyone explain what is going on ?
In https://onlinecourses.science.psu.edu/stat414/node/171 it states that:
Corollary. If X1, X2 , ... , Xn are independent normal random variables with mean=0 and variance=1, that is: Xi∼N(0,1) for i = 1, 2, ..., n. Then:

Sum_from_one_to_n(Xi)^2 ∼ χ2(n) ( I have simplified the formula in the source by setting all μi to 0 and σi to 1)

But https://onlinecourses.science.psu.edu/stat414/node/174 states that:
X1, X2, ... , Xn are observations of a random sample of size n from the normal distribution N(0,1) then:
Sum_from_one_to_n(Xi)^2 ∼ χ2(n-1) (Again setting all μi to 0 and σi to 1)

So it seems that a (slightly) different pdf is being given for the same R.V. , we have lost 1 degree of freedom. Can anyone explain this, or does the fact that the individual observations in the second case are described as a 'random sample' somehow impact on their independence ? If so, how exactly, is each element of the sample not an RV in its own right whose pdf is that of the population and => N(0,1) ?

Thanks,

Usjes
 
Last edited:
Physics news on Phys.org
  • #2
The sample mean of a random sample ([itex] \bar{X} [/itex] )is not the same as the mean (i.e. "population mean" [itex] \mu [/itex] ) of the random variable from which the sample is taken.
 
  • #3
"Can anyone explain this,..."

The explanation is in the proof shown below your second reference.
 
  • #4
In the first case the population parameters are somehow known. The second case is more usual, in which the population parameters are unknown and are estimated from a random sample. This inaccuracy is compensated for by loss of a degree of freedom.
 
  • #5
Your second statement is not what the reference says. Your statement uses the known mean, but the reference uses the sample mean. The sum of squares is the sum squares of differences from the known mean of 0. That has n degrees of freedom. It is only if the sample mean is used instead of the known mean that the degrees of freedom are reduced by 1. Notice that using the sample mean will reduce the sum of squares, so the different degrees of freedom is needed to compensate for that.
 
Last edited:

Related to Why Do Degrees of Freedom Differ in Chi-Squared Distributions?

What is the definition of "degrees of freedom" in science?

Degrees of freedom refer to the number of independent variables that can vary in a system without changing the outcome or result. In statistical analysis, it is the number of independent values or quantities that can vary in a data set.

Why is it important to know the degrees of freedom in an experiment?

Knowing the degrees of freedom in an experiment is important because it helps determine the appropriate statistical tests to use and the accuracy of the results. It also allows for the evaluation of the variability and reliability of the data.

How do you calculate the degrees of freedom in an experiment?

The degrees of freedom can be calculated by subtracting the number of constraints or fixed variables from the total number of variables in a system. In a statistical analysis, it is typically calculated as the difference between the sample size and the number of parameters being estimated.

Can the degrees of freedom ever be negative?

No, the degrees of freedom cannot be negative. It is a non-negative value that represents the number of independent variables in a system. If a calculation results in a negative value, it is likely an error or an incorrect method was used to calculate it.

How does the number of degrees of freedom affect the outcome of an experiment?

The number of degrees of freedom can affect the outcome of an experiment by influencing the accuracy and precision of the results. A higher number of degrees of freedom generally results in a more reliable and valid conclusion, while a lower number may lead to less accurate or biased results.

Similar threads

  • Set Theory, Logic, Probability, Statistics
Replies
5
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
4
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
9
Views
3K
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
850
  • Set Theory, Logic, Probability, Statistics
Replies
4
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
4
Views
3K
  • Set Theory, Logic, Probability, Statistics
Replies
10
Views
7K
  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
3K
Back
Top