When are 2 models comparable using an F Test?

In summary: The F statistic is the sum of the squares of the differences between the restricted and unrestricted models. In summary, the F statistic is a measure of the proportion of variation in the dependent variable that is explained by the model.
  • #1
David Laz
28
0
Say I'm given a bunch of models on the same set of data how does one determine whether a valid comparison can be made between any two of them using an F Test?

Is it that the estimation space of one has to be a subset of the other? Is there any easier, more practical way of determining this?

Thanks :redface:
 
Physics news on Phys.org
  • #2
Is it that the estimation space of one has to be a subset of the other?
No. For example, in a regression equation, you might want to test whether bi = bj. In this case, the relevant test statistic has an F distribution, although the hypothesis does not involve a subset relationship.

http://en.wikipedia.org/wiki/F_test
 
  • #3
I think I know what your getting at. But here's an example of the sort of question I need help with. Maybe you could explain/show me.

heres a question from an old exam:

The following table gives the yields from a field experiment on two varieties of wheat, Hard and Common, with four equally spaced levels of applied fertilizer. The plots were allocated at random to the various treatment combinations. Initially it was planned to have four replicates at each combination, but errors in applying the fertilizer reduced the final sample size.


http://img530.imageshack.us/img530/2558/statny0.jpg
(table of the data, probably not need to answer this question)

In what follows Var.f refers to variety treated as a factor with two levels. (1=hard, 2 = common) and Fert.f refers to fertilizer treated as a factor with 4 levels (1,2,3,4)

Since fertilizers are numeric it is possible to use the actual amount of fertilizer as a variable (denoted by x, taking values of 1,2,3,4)

http://img174.imageshack.us/img174/8470/stat2nc9.jpg
(table of different models and their associated deviance and df's)

Where a*b means a + b + a:b (interaction term)
Among all the models which cannot be validly compared using an F-Test?
 
Last edited by a moderator:
  • #4
To calculate the F statistic, you need a restricted model and an unrestricted model. The restricted model is a sub-type of the unrestricted model, in the sense that but for the restriction(s) being applied, it would have been identical to the unrestricted model. Put differently, starting from the unrestricted model, you should be able to arrive at the restricted model by imposing one or more linear restrictions on the set of parameters that the model is to estimate. Example: a Var.f + b Fert.f can be obtained from a Var.f + b Fert.f + c Var.f:Fert.f by imposing the linear restriction c = 0. (Each of a, b, c is an estimated parameter that would explain, say, "agricultural yield" as a function of the variable Var.f, Fert.f, Var.f:Fert.f in respective order.)
 
Last edited:

1. When should I use an F Test to compare two models?

The F Test is typically used to compare two models when one model is a nested, or a simplified version, of the other model. This means that the simpler model can be obtained by imposing constraints on the more complex model. In this case, the F Test can determine if the added complexity of the more complex model is justified compared to the simpler model.

2. What does the F statistic represent in an F Test?

The F statistic in an F Test is a ratio of two variances, specifically the variance explained by the model and the remaining unexplained variance. It measures the difference in fit between the two models being compared. A higher F statistic indicates a better fit for the more complex model compared to the simpler model.

3. How do I interpret the p-value from an F Test?

The p-value from an F Test represents the probability of obtaining a result as extreme as the observed F statistic, assuming the null hypothesis is true. A low p-value (usually less than 0.05) suggests that the observed difference in fit between the two models is unlikely to have occurred by chance and therefore the null hypothesis can be rejected. This means that the more complex model is a significantly better fit compared to the simpler model.

4. Can the F Test be used to compare models with different numbers of predictors?

Yes, the F Test can be used to compare models with different numbers of predictors. However, the models must be nested, meaning that the predictors in the simpler model are a subset of the predictors in the more complex model. This ensures that the models are directly comparable and the F Test can accurately assess the added complexity of the more complex model.

5. Are there any assumptions that need to be met for an F Test to be valid?

Yes, there are several assumptions that need to be met for an F Test to be valid. These include: the errors of the models are normally distributed, the variances of the errors are equal, and the observations are independent. Violations of these assumptions can lead to inaccurate results from the F Test and therefore, other methods of model comparison may be more appropriate.

Similar threads

  • Set Theory, Logic, Probability, Statistics
Replies
6
Views
818
  • Set Theory, Logic, Probability, Statistics
Replies
23
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
7
Views
472
  • Set Theory, Logic, Probability, Statistics
Replies
30
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
3
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
5
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
810
  • Set Theory, Logic, Probability, Statistics
Replies
5
Views
3K
  • Set Theory, Logic, Probability, Statistics
Replies
3
Views
820
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
1K
Back
Top