Testing multiple hypotheses

  • Context: Undergrad 
  • Thread starter Thread starter Agent Smith
  • Start date Start date
Click For Summary
SUMMARY

This discussion focuses on the complexities of hypothesis testing in statistics, particularly when considering multiple hypotheses (H0, H1, H2, ..., Hn) and their associated probabilities. It establishes that while the null hypothesis (H0) may be rejected at a significance level of 0.05, all alternative hypotheses (H1, H2, ..., Hn) can still have lower probabilities than H0, leading to a paradoxical situation where the sum of all probabilities equals 1. The conversation also contrasts frequentist and Bayesian statistics, emphasizing that frequentist methods do not assign probabilities to hypotheses, while Bayesian methods do, allowing for a different interpretation of evidence.

PREREQUISITES
  • Understanding of hypothesis testing and significance levels (e.g., P-value ≤ 0.05).
  • Familiarity with frequentist and Bayesian statistics concepts.
  • Knowledge of probability distributions, including normal distribution.
  • Experience with statistical tests, such as the F-test.
NEXT STEPS
  • Explore the implications of rejecting the null hypothesis in multiple hypothesis testing.
  • Study Bayesian statistics and its application in hypothesis testing.
  • Learn about the F-test and its role in comparing two hypotheses.
  • Investigate empirical probability and its relevance in actuarial sciences.
USEFUL FOR

Statisticians, data analysts, researchers in scientific fields, and anyone involved in hypothesis testing and statistical decision-making will benefit from this discussion.

Agent Smith
Messages
345
Reaction score
36
Say we're considering multiple hypotheses: ##H_0, H_1, H_2, ... H_ n##
##P(H_0) \leq 0.05## (our P-value).

Is it possible that ##P(H_1) < P(H_0), P(H_2) < P(H_0), ... P(H_n) < P(H_0)##

and yet ##P(H_0) + [P(H_1) + \dots + P(H_n)] = 1##?

We would reject the null hypothesis but all the alternative hypotheses are less probable than the null.
 
Physics news on Phys.org
Agent Smith said:
Say we're considering multiple hypotheses: ##H_0, H_1, H_2, ... H_ n##
##P(H_0) \leq 0.05## (our P-value).

Is it possible that ##P(H_1) < P(H_0), P(H_2) < P(H_0), ... P(H_n) < P(H_0)##

and yet ##P(H_0) + [P(H_1) + \dots + P(H_n)] = 1##?

We would reject the null hypothesis but all the alternative hypotheses are less probable than the null.
As an example ##\frac{1}{2} + \left[\frac{1}{4} + \frac{1}{8} + ...\right] = 1##
 
Agent Smith said:
Say we're considering multiple hypotheses: ##H_0, H_1, H_2, ... H_ n##
##P(H_0) \leq 0.05## (our P-value).

Is it possible that ##P(H_1) < P(H_0), P(H_2) < P(H_0), ... P(H_n) < P(H_0)##

and yet ##P(H_0) + [P(H_1) + \dots + P(H_n)] = 1##?

We would reject the null hypothesis but all the alternative hypotheses are less probable than the null.
This is an unusual statistical test setup. Usually, the probability parameters of the null hypothesis are initially assumed. Suppose that with the null hypothesis probability distribution, the sample is an unlikely event with a probability less than 0.05. Then the alternative hypotheses may be considered and their alternative probability parameters (mean, variance, etc.) are assumed. These are not the same probabilities that they had under the null hypothesis. With maximum likelihood estimated parameters the sample results are much more likely.

For instance, suppose that the null hypothesis is that the sample is from a normal distribution with a mean ##\mu_0=0## and variance ##\sigma_0 = 1##. Now suppose that the sample (N=101) has a sample results of ##\mu_s=0.4## and ##\sigma_s=1##. Then the null hypothesis mean of 0 is far (4 ##\sigma##) outside the 95% confidence interval for that sample (the 95% confidence interval is [0.2, 0.6]). So now we could consider the alternative hypothesis, ##H_1##, that the true population distribution has a mean of 0.4. That hypothesized mean would be much more compatible with the sample.
 
  • Wow
Likes   Reactions: Agent Smith
First, there exist tests that distinguish between exactly two hypotheses. (The F test is an example) You are not limited to tests of rejecting the null hypothesis. And yes, both hypotheses can be wrong.

Next, I would say if you calculate the p-value first and then set your threshold, you are not doing statistics. Not sure what I would call it, but this isn't statistics.

Finally, statistics are there to help you make a decision. I remember a test that caused us to reject the null hypothesis at 90 but not 95%. What we did was to make small interventions - less than we would do at 95% - but wouldn't be harmful if the null hypotheses were correct. Probably not a perfect outcome, but one supported by what data we had.
 
  • Wow
  • Like
Likes   Reactions: Agent Smith and FactChecker
Agent Smith said:
Say we're considering multiple hypotheses: ##H_0, H_1, H_2, ... H_ n##
##P(H_0) \leq 0.05## (our P-value).

Is it possible that ##P(H_1) < P(H_0), P(H_2) < P(H_0), ... P(H_n) < P(H_0)##

and yet ##P(H_0) + [P(H_1) + \dots + P(H_n)] = 1##?

We would reject the null hypothesis but all the alternative hypotheses are less probable than the null.
In frequentist statistics you never assign a probability to a hypothesis. You assign the probability to the data. So it would be $$P( data | H_0)$$ etc.

Bayesian statistics does assign probabilities to hypothesis. So you could indeed see that $$P(H_1|data)< P(H_0|data)<0.05$$ and consider the evidence to refute both hypotheses
 
  • Like
Likes   Reactions: Agent Smith, FactChecker and hutchphd
I think I'm talking about Bayesian statistics. The gist of what I want to say seems to be that every possible hypothesis is improbable at a threshold probability of 0.05.
 
Agent Smith said:
every possible hypothesis is improbable at a threshold probability of 0.05
Certainly not every possible hypothesis. But all of the hypotheses in some specified set of hypotheses.
 
  • Like
Likes   Reactions: Agent Smith
Dale said:
In frequentist statistics you never assign a probability to a hypothesis.
Apologies for the late reply, was engaged.

So I learned about empirical probability many suns ago which is required in some fields like actuarial sciences. So desired probabilities are computed from (say) birth and death records. Is this an example of frequentist statistics? Can Bayesian statistics be useful here?
 
Agent Smith said:
Apologies for the late reply, was engaged.

So I learned about empirical probability many suns ago which is required in some fields like actuarial sciences. So desired probabilities are computed from (say) birth and death records. Is this an example of frequentist statistics? Can Bayesian statistics be useful here?
An empirical probability is experimental evidence. It is the observed frequency in a given experiment. It can be useful in both frequentist statistics and in Bayesian statistics. It just depends if you want to determine ##P(hypothesis|evidence)## or ##P(evidence|hypothesis)##
 
  • Like
Likes   Reactions: Agent Smith
  • #10
@Dale

##\displaystyle \frac{2}{11} \sum_{k = 1} ^10 \frac{11 - k}{10} = 1##

Someone found for me a fraction viz. ##\frac{2}{11}## such that we take fractions < ##\frac{2}{11}## and add them all up we get ##1##. Each of these fractions could be the probability of a hypothesis. Our best hypothesis only has a probability of ##\frac{2}{11}##. I wonder what this would imply. 🤔
 
  • #11
It implies that none of your hypotheses are very likely.

If in addition you know that your hypotheses are mutually exclusive, then it also means that we are certain one is right but uncertain which one. Note that the mutually exclusive assumption is an additional one that is not generally satisfied by hypotheses. As a result, it is possible to have a set of hypotheses whose probabilities sum to greater or less than 1.
 

Similar threads

Replies
20
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 10 ·
Replies
10
Views
3K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 10 ·
Replies
10
Views
3K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 24 ·
Replies
24
Views
6K
  • · Replies 8 ·
Replies
8
Views
3K
  • · Replies 13 ·
Replies
13
Views
3K