Hypothesis Tests: Understanding p-value, alpha level & more

  • Context: Undergrad 
  • Thread starter Thread starter muktl
  • Start date Start date
Click For Summary

Discussion Overview

The discussion revolves around understanding hypothesis testing, specifically focusing on the concepts of p-value, alpha level, and the decision-making process regarding the rejection or acceptance of hypotheses. Participants explore both frequentist and Bayesian perspectives on hypothesis testing, as well as the implications of significance levels in practical scenarios.

Discussion Character

  • Exploratory
  • Technical explanation
  • Debate/contested
  • Conceptual clarification

Main Points Raised

  • One participant expresses confusion about when to reject or not reject a hypothesis and seeks clarification on the meaning and application of p-values and alpha levels.
  • Another participant critiques the traditional frequentist approach, arguing that it does not answer the question of the probability that a hypothesis is true given the data, instead focusing on the probability of the data given the hypothesis.
  • This participant suggests that the determination of what constitutes a "small" p-value is subjective and highlights the arbitrary nature of the rejection process in hypothesis testing.
  • A different viewpoint emphasizes the frequentist approach, explaining that under the null hypothesis, no significant difference is expected between treatment and control groups, and that a low alpha level is desired to minimize the risk of incorrectly rejecting the null hypothesis.
  • Another participant clarifies that a high significance level refers to the alpha error, which should be low, indicating a desire for confidence in rejecting the null hypothesis only when it is truly false.
  • Bayesian statistics is mentioned as an alternative approach that allows for the incorporation of prior probabilities, with a suggestion that mastering frequentist methods is beneficial before delving into Bayesian methods.

Areas of Agreement / Disagreement

Participants express differing views on the nature of hypothesis testing, particularly regarding the frequentist versus Bayesian approaches. There is no consensus on the best method or the interpretation of p-values and alpha levels, indicating an ongoing debate.

Contextual Notes

Participants note the limitations of the frequentist approach, particularly its inability to provide direct probabilities about hypotheses given the data. The discussion also highlights the subjective nature of significance levels and the potential consequences of varying p-value thresholds in practical applications.

muktl
Messages
6
Reaction score
0
Hi, i am just really confused about the hypothesis test part in my course.
all i know is i use the mean of the hypothesis and find the test point, but i don't understand when i need to reject/ not reject the hypothesis.
and what really is p-value ( significance level), how to use that to determine whether to reject or not.
and the alpha level ( % of making type I error)

Thanks.
 
Physics news on Phys.org
One way to look at the typical statistics course is that it doesn't make logical sense!

The situation is this:

The type of question that a practical person wants to know is "Given the data what is the probability that statement H is true? ".

There is not enough given information to answer this question in the problems presented in a typical statistics course. So, instead, a different question is answered, namely: "Given that statement H is true, what is the probability of the data?".

If the probability of the data is "small" then you are supposed to "reject" the statement H, but this is a completely subjective decision. There is no objective way to determine how small "small" should be. The process of "accepting or rejecting the null hypothesis" is simply an arbitrary procedure. There is no proof offered that it is the only correct one to follow.

You will find that the terms used in statistics ("level of significance", "rejection region", "confidence level", "confidence limits") have been cleverly chosen. They make the methods of statistics sound very objective. They confuse most people into thinking that they are getting the answer to "Given the Data, what is the probability that my idea is true?". However, underneath the hood, the methods are subjective and the numbers you compute don't answer this question.

If you want to study the type of statistics that does compute "The probability that statement H is true given the data", you'll have to study Bayesian statistics. If you want to study how to objectively set p-values, you should study a book like "Optimal Statistical Decisions" by Morris DeGroot.

In some situations, it may be possible to judge empirically how p-values are working. For example, if you publish a medical journal and you require that studies show a p-value is .05 then you'll have a smaller pile of papers to consider than if you set your p-value to .10. Or suppose you run a lab that screens thousands of substances to pick out ones that hold promise as drugs. Suppose you set a p-value of .10 and find hardly any substances to send-on for further testing. Other companies discover drugs based on substances that you have rejected. Your boss complains. One obvious course of action is increase your p-value to .15.
 
In a typical experimental situation we compare a treatment group with a control group in terms of some experimental parameter (such as the mean). Under the null hypothesis we assume the treatment will have no effect, so there should be no significant difference in the parameter value between the two groups. The standard (frequentist) approach is to assume some population based distribution of the parameter, usually a Gaussian (normal) distribution. This has been shown to work pretty well. After the experiment, we set a high significance level (low probability) that the null hypothesis is true, say a probability of 0.025 because we want to have confidence that a real effect exists. If the level is met or exceeded (a probability of less then 0.025) we reject the null hypothesis. This is called the alpha error because that is the probability that we wrongly rejected the null hypothesis in favor of the alternative hypothesis, which is that the treatment had an effect, and the findings were not just random variation.

The Bayesian approach is different in that it states a prior probability of an outcome before the experiment and then revises that probability based on the new data. It has a lot of applications, but I think one should master frequentist methods before studying Bayesian methods. They are not necessarily incompatible but the method I described above is what' s generally required if you are trying to publish experimental results or submitting data to a government agency.
 
Last edited:
To clarify something I said in the previous post. When I said that we want to set a high significance level, I was referring to the alpha error which is the probability of rejecting the null hypothesis when it is true. This probability should be low.
 

Similar threads

  • · Replies 43 ·
2
Replies
43
Views
6K
  • · Replies 24 ·
Replies
24
Views
7K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 5 ·
Replies
5
Views
3K
Replies
20
Views
3K
  • · Replies 2 ·
Replies
2
Views
3K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 8 ·
Replies
8
Views
3K
  • · Replies 20 ·
Replies
20
Views
4K
  • · Replies 5 ·
Replies
5
Views
4K