Relationship between a posterior distribution and the LLN

Click For Summary
SUMMARY

This discussion explores the relationship between posterior distributions in Bayesian statistics and the Law of Large Numbers (LLN). The user presents a scenario involving a biased coin with an unknown bias, Θ, and questions whether the expected percentage of heads from additional flips can be derived from the posterior distribution after initial flips. The consensus is that the expected value calculated from the posterior does not represent the expected outcome of repeated flips of the same coin, as it reflects a population of coins with varying biases rather than a single biased coin. The conversation highlights the distinction between the expected value from the posterior and the true mean of the specific biased coin.

PREREQUISITES
  • Understanding of Bayes' Theorem and its application in statistics.
  • Familiarity with the Law of Large Numbers (LLN) and its implications.
  • Knowledge of Central Limit Theorem (CLT) and its definitions.
  • Basic concepts of probability distributions, particularly unimodal distributions.
NEXT STEPS
  • Study the implications of Bayes' Theorem in practical scenarios involving unknown parameters.
  • Research the differences between the expected value from a posterior distribution and the true mean of a specific random variable.
  • Explore the Central Limit Theorem (CLT) and its technical definitions in the context of Bayesian statistics.
  • Examine case studies involving biased coins and their statistical analysis using Bayesian methods.
USEFUL FOR

Statisticians, data scientists, and researchers interested in Bayesian statistics, particularly those exploring the implications of posterior distributions and the Law of Large Numbers in practical applications.

scinoob
Messages
17
Reaction score
0
Hello everybody. This is my first post here and I hope I'm not asking a question that's been addressed already (I did try to use the search function, but couldn't find what I'm looking for).

Both the Bayes theorem and the law of large numbers are mathematical theorems derived from Kolmogorov's axioms. I've been thinking of a way to relate the two and I'm not sure how that can be done.

Let's say I have a coin with an unknown bias towards heads Θ. I start with a uniform prior distribution of Θ (between 0 and 1 obviously) and keep flipping the coin while updating the prior using Bayes' theorem. Let's say after 300 flips the posterior looks like a relatively peaked unimodal distribution centered at 0.3. Then I decide to flip the coin another, say, 10k times without updating the posterior anymore and somebody asks the question "approximately what percentage of heads do you expect to see after the 10k flips?" So, what I am likely going to do (at least according to the standard practice) is calculate the expected value of a single flip using the posterior distribution I obtained after the 300 flips. If I then go ahead and flip the coin 10k times, can I expect that the percentage will be close to the value I obtained in the analytic calculation? If I keep flipping it, can I expect that the % of heads will converge to the expected value by appealing to the LLN?

If the answer is 'no', then is there another way to relate Bayes' theorem to the LLN, say, using my coin example in particular?
 
Physics news on Phys.org
scinoob said:
Let's say after 300 flips the posterior looks like a relatively peaked unimodal distribution centered at 0.3.

Bayesians vary in how they interpret the bayesian approach to problems. One common way to think about the posterior distribution in your problem is that your particular coin was picked at from a population of coins with various probabilities of landing heads and that the posterior distribution gives the probabilities of picking such a coin. An expected value calculated using the posterior distribution is the expected number of heads over that entire population of coins, not the expected value of heads from tossing your particular coin repeatedly. So when you say that you continue to toss the coin 10k more times, you must be specific about whether you are tossing the same coin 10k times or whether you are imagining each of the 10k tosses of a coin to consist of two steps: a) Pick the coin at random from the population of all coins b) toss the coin you picked.
 
Stephen Tashi said:
Bayesians vary in how they interpret the bayesian approach to problems. One common way to think about the posterior distribution in your problem is that your particular coin was picked at from a population of coins with various probabilities of landing heads and that the posterior distribution gives the probabilities of picking such a coin. An expected value calculated using the posterior distribution is the expected number of heads over that entire population of coins, not the expected value of heads from tossing your particular coin repeatedly. So when you say that you continue to toss the coin 10k more times, you must be specific about whether you are tossing the same coin 10k times or whether you are imagining each of the 10k tosses of a coin to consist of two steps: a) Pick the coin at random from the population of all coins b) toss the coin you picked.

Good point. In this case I mean tossing the same coin 10k times. In other words, it's the same coin with some unknown bias θ. The posterior distribution represents the degree of belief in each value for θ.
 
If I keep flipping it, can I expect that the % of heads will converge to the expected value by appealing to the LLN?
No, you can't justify that by the CLT.
 
Stephen Tashi said:
No, you can't justify that by the CLT.

I'm afraid I have to disagree here. I think this is exactly what the CLT says. It will converge to the true mean of the biased coin. It will not converge to the mean of the 300 samples unless that happens to have 0 error.
 
But the true mean of the particular biased coin is not the mean calculated from the posterior distribution, which considers coins with a variety of biases in the calculation.

(Also "will converge" has a very technical definition - different than the definition of convergence used in calculus for the convergence of sequences or limits of functions.)
 
Last edited:
Stephen Tashi said:
But the true mean of the particular biased coin is not the mean calculated from the posterior distribution, which considers coins with a variety of biases in the calculation.

(Also "will converge" has a very technical definition - different than the definition of convergence used in calculus for the convergence of sequences or limits of functions.)

Oh. I missed that there were multiple coins with different biases. Then you are right.
 

Similar threads

  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 45 ·
2
Replies
45
Views
6K
Replies
3
Views
4K
  • · Replies 2 ·
Replies
2
Views
5K
Replies
2
Views
3K
Replies
2
Views
1K
  • · Replies 11 ·
Replies
11
Views
3K
  • · Replies 14 ·
Replies
14
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
Replies
2
Views
2K