Differences between binomial distribution and forced probability distribution

Click For Summary
SUMMARY

The discussion centers on the differences between binomial distribution and a concept termed "forced probability distribution." The original poster describes a personal example where the probability of attending the gym changes based on previous attendance, illustrating how perceived guilt alters probability. They created a C program demonstrating a model where a 36% winning chance increases by 20% after each failure, resulting in a distribution resembling a 50%-50% model. The conversation explores the feasibility of distinguishing between these two distributions using statistical methods.

PREREQUISITES
  • Understanding of binomial distribution principles
  • Basic knowledge of probability theory
  • Familiarity with stochastic processes
  • Proficiency in C programming for empirical modeling
NEXT STEPS
  • Explore advanced concepts in stochastic processes
  • Learn about statistical testing methods for discerning distributions
  • Investigate the implications of probability manipulation in real-world scenarios
  • Study the implementation of probability models in C programming
USEFUL FOR

Statisticians, data scientists, programmers interested in probability modeling, and anyone exploring the nuances of probability theory in practical applications.

Daaavde
Messages
29
Reaction score
0
Differences between binomial distribution and "forced" probability distribution

Hi everyone.

Yesterday I was thinking about probability and real life and about the fact that we always expect life's facts to behave according to probability theory.
If we flip a coin and we get 6 times heads in a row we get suspicious.

So what happen if we "force" probabilities when we temporarily see them under statistical fluctuations and we "make" them behave according to binomial distribution?

For example, I say that, in a month, I go to gym 50% of the days.
One day I go, the next two I don't, but the fourth day chance is not 50% anymore because I'm afraid that I'm skipping gym, so now is 60% (I feel guilty, I don't want to starting skip gym).

If the week starts and I don't go to gym on Monday, Tuesday, Wednesday and Thursday chance I'm going on Friday is now 70% and so on. But when I go to gym, probability reset itself at 50% again (I feel like I've done my duties).

It's just a silly example, but I hope you got the idea.

So this is the model of "forced probability"
If there's a fail -> Chance of winning increases by (let's say) 10%
If there's a winning -> Chance of winning goes back to 50%

My question is, it's possible to discern between the two cases just observing statistics?
What happen if, in physics, some phenomena seems to act like "coin flipping" while it behaves like "forced probability"?

I created a program in C and I empirically saw that a model in which there's 36% of winning but with increasing of 20% of winning everytime there's a failure shows a distribution very similar to the 50%-50% model. (Mean Value: 0.501758 / Standard Deviation: 0.00338905)
Probably a much more similar distribution can be obtained using float numbers (for simplicity of the program, I've set probability as an integer between 0 and 100, but using much more sophisticated approach would probably give better results).
If someone is interested I can send him the program.

I could have tried a mathematical approach, but honestly, I don't really feel into starting to fight with differential equations. That's why I'm posting this here.

So, do you think it's possible that by combining initial probability and increasing rate values the two distribution can be made not discirnible by statistics methods?

(I really apologize for my poor scholastic english, I'm not native speaker)
 
Physics news on Phys.org


You haven't described what data is available for statistical testing. For example, is the data one long series that begins at trial 1? Or do you have several series of trials, each series beginning at trial 1?

Statistics doesn't "discern" differences with certainty. Sometimes you can find a test that has a given probability of making the correct decision. In many cases, you don't even know that probability.

Suppose you have several series of trials of flipping a coin. You can pick out the trials that begin with pattern of results: H H H * , where '*' means a 'H' or a 'T" on the last flip. If the trials are independent flips of a fair coin then you might base a test on whether the fraction of trials that begins H H H T is about 1/2 of those trials. If the fraction that begins with H H H T is much more or less than 1/2, the test could decide that the flips were not independent. The exact details of how to do this depend on how you define "discerning" and what data you have.
 


Daaavde said:
Hi everyone.

Yesterday I was thinking about probability and real life and about the fact that we always expect life's facts to behave according to probability theory.
If we flip a coin and we get 6 times heads in a row we get suspicious.

So what happen if we "force" probabilities when we temporarily see them under statistical fluctuations and we "make" them behave according to binomial distribution?

For example, I say that, in a month, I go to gym 50% of the days.
One day I go, the next two I don't, but the fourth day chance is not 50% anymore because I'm afraid that I'm skipping gym, so now is 60% (I feel guilty, I don't want to starting skip gym).

If the week starts and I don't go to gym on Monday, Tuesday, Wednesday and Thursday chance I'm going on Friday is now 70% and so on. But when I go to gym, probability reset itself at 50% again (I feel like I've done my duties).

It's just a silly example, but I hope you got the idea.

So this is the model of "forced probability"
If there's a fail -> Chance of winning increases by (let's say) 10%
If there's a winning -> Chance of winning goes back to 50%

My question is, it's possible to discern between the two cases just observing statistics?
What happen if, in physics, some phenomena seems to act like "coin flipping" while it behaves like "forced probability"?

I created a program in C and I empirically saw that a model in which there's 36% of winning but with increasing of 20% of winning everytime there's a failure shows a distribution very similar to the 50%-50% model. (Mean Value: 0.501758 / Standard Deviation: 0.00338905)
Probably a much more similar distribution can be obtained using float numbers (for simplicity of the program, I've set probability as an integer between 0 and 100, but using much more sophisticated approach would probably give better results).
If someone is interested I can send him the program.

I could have tried a mathematical approach, but honestly, I don't really feel into starting to fight with differential equations. That's why I'm posting this here.

So, do you think it's possible that by combining initial probability and increasing rate values the two distribution can be made not discirnible by statistics methods?

(I really apologize for my poor scholastic english, I'm not native speaker)

Sure. This has to do with stochastic processes. It would be easy to tell. It is a bit much to teach here though.
 

Similar threads

  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 9 ·
Replies
9
Views
2K
  • · Replies 15 ·
Replies
15
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
Replies
1
Views
3K
  • · Replies 5 ·
Replies
5
Views
3K
  • · Replies 10 ·
Replies
10
Views
9K
  • · Replies 2 ·
Replies
2
Views
2K