Optimizing Data Sampling for Probabilities

  • Context: Graduate 
  • Thread starter Thread starter Loren Booda
  • Start date Start date
  • Tags Tags
    Data Statistical
Click For Summary

Discussion Overview

The discussion revolves around the mathematical methods for optimizing data sampling in the context of probabilities. Participants explore various scenarios that complicate the determination of probabilities, including physical and quantum factors, and question how to systematically approach the sampling process given these complexities.

Discussion Character

  • Exploratory
  • Technical explanation
  • Conceptual clarification
  • Debate/contested

Main Points Raised

  • Some participants propose that traditional probability models, such as coin flipping, may not account for all possible outcomes, including edge cases and quantum effects.
  • There is a suggestion that practical solutions, like rerolling dice until a "normal" outcome is achieved, are common in gaming contexts.
  • Questions are raised about how to mathematically determine the number of trials needed to achieve an average outcome, with concerns about the incorporation of errors in statistical measures like standard deviation.
  • One participant emphasizes the importance of defining the sample space and suggests that mathematics serves as a modeling tool rather than a definitive guide to real-life behavior.
  • Another participant inquires about online resources that compare various statistical tests and methods relevant to the discussion.

Areas of Agreement / Disagreement

Participants express differing views on the interpretation of "normal" outcomes and the role of mathematics in defining sample spaces. There is no consensus on the best approach to determine optimum sampling methods, and the discussion remains unresolved regarding the complexities introduced by physical variables.

Contextual Notes

Limitations include the dependence on definitions of "normal" outcomes and the unresolved nature of how various statistical methods apply to complex scenarios with multiple interacting variables.

Who May Find This Useful

This discussion may be of interest to those studying probability theory, statistical methods, and their applications in real-world scenarios, particularly in fields like gaming, physics, and data analysis.

Loren Booda
Messages
3,115
Reaction score
4
Is there a mathematical method to determine the optimum sampling of data for probabilities?

Flip a coin. Simplistically speaking from experience, it has a 1/2 chance of landing on either side. But what if it can land on its edge? What if it can fall through a crack? What if lava from a fissure invading the room can envelop and melt the coin? What if it can quantum mechanically flip itself after landing? Other examples of probability, like the nonlinear trajectory of a particle, have determinism not immediately apparent.

Even an electronic random number generator run by a quantum computer is susceptable to decoherence between the device and the observer. It seems that we must have extensive practical knowledge about the system under observation, then apply Occam's razor, if we are to determine the set of data required. But how may this be done systematically?
 
Physics news on Phys.org
Loren Booda said:
Flip a coin. Simplistically speaking from experience, it has a 1/2 chance of landing on either side. But what if it can land on its edge? What if it can fall through a crack? What if lava from a fissure invading the room can envelop and melt the coin? What if it can quantum mechanically flip itself after landing? Other examples of probability, like the nonlinear trajectory of a particle, have determinism not immediately apparent.
You could flip it until it behaves normally. :-p

In fact, that's a common practical solution in the gaming world -- one keeps rerolling a die until it doesn't fall off the table or lean against something or whatever.
 
By "normal" one might mean "average." How does one determine mathematically how many tries one needs to achieve average? Don't methods like standard deviation incorporate their own error, ad infinitum?

Overall, how and when can we be assured of precision's reproducibility?
 
But that's not what was meant when I said normal. I meant for "behaving normally" to be "lands heads up or lands tails up".
 
Duly noted.

Please allow me to repeat [with editing]:
How does one determine mathematically [with a huge number of interacting physical variables] how many tries one needs to achieve [a significant] average?

Don't [results] like standard deviation incorporate [errors of their own, each with their own statistical deviations] ad infinitum?

Overall, how and when can we be assured of precision's reproducibility?
 
You decide before hand what consitutes the sample space - either heads or tails. Any other outcome is deemed inadmissable. Why? Because that is what we want, and has nothing to do with mathematics. Mathematics is merely a tool for modelling, in this instance. Whether real life behaves sufficiently close to the model for the model to be valid is a different matter. There are plenty of tests to work out whether sample data is likely to have come from a population with assumed properties; they are taught to high school students such as confidence intervals; I'm surprised you've not met them. Then there is the strong law of large numbers, chi squared tests, t tests, ANOVA, et c.
 
Last edited:
Your examples are worthwhile studying. Do you know of an online tutorial that compares most of them?
 

Similar threads

  • · Replies 11 ·
Replies
11
Views
2K
  • · Replies 14 ·
Replies
14
Views
4K
  • · Replies 25 ·
Replies
25
Views
6K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 20 ·
Replies
20
Views
4K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 29 ·
Replies
29
Views
4K
  • · Replies 21 ·
Replies
21
Views
4K