Motivation behind random variables?

Click For Summary
SUMMARY

The discussion centers on the significance of random variables in probability theory, defined as integrable functions X:Ω→R within a probability space (Ω, μ). The utility of random variables is highlighted through examples, such as coin flipping, where they facilitate the analysis of events and their dependencies. The conversation also touches on the limitations of existing literature, particularly a paper by Mitchener, which inaccurately asserts that random variables with identical moments share the same probability laws, a claim countered by known distributions like the lognormal. Understanding random variables is essential for delving into non-commutative probability theory and its applications.

PREREQUISITES
  • Understanding of probability spaces and measures (Ω, μ)
  • Familiarity with random variables and their definitions
  • Basic knowledge of cumulative distribution functions (CDFs) and probability mass functions (PMFs)
  • Awareness of the moment problem and its implications in probability theory
NEXT STEPS
  • Study the role of random variables in defining moments of distributions
  • Explore non-commutative probability theory and its applications in quantum mechanics
  • Examine the implications of the moment problem, particularly with distributions like lognormal and Cauchy
  • Investigate software modeling techniques for probability theory applications
USEFUL FOR

Students and researchers in mathematics, statisticians, and anyone interested in the foundational aspects of probability theory and its applications in fields such as quantum mechanics and statistical modeling.

Tac-Tics
Messages
816
Reaction score
7
What is the motivation behind random variables in probability theory?

The definition is easy to understand. Given a probability space (Ω, μ), a random variable on that space is an integrable function X:Ω→R. So essentially, it allows you to work in the concrete representation R instead of the abstract Ω.

But why is that useful?

Take a simple example of flipping a pair of coins in order. Our event set is Ω={HH, HT, TH, TT}. Our probability function is μ({HH}) = μ({HT}) = μ({TH}) = μ({TT}) = 1/4.

The random variables on this set can be a simple numeric assignment: X(HH) = 1, X(HT) = 2, X(TH) = 3, X(TT) = 4. Or it can be in a different order. Or it can be non-injective, such as the constant random variable X(s) = 1, or the "outcome same" function, X(HH) = X(TT) = 1, X(HT) = X(TH) = 0.

How is that useful for analysis? Does it have to do with the random variable's role in defining the moments of a distribution? Or perhaps random variables simply aren't useful in the finite case?

The reason I'm trying to understand this is I am (trying) to learn a little about non-commutative probability theory (and ultimately, a little about quantum mechanics).

The source I'm working off of is a paper by Mitchener (http://www.uni-math.gwdg.de/mitch/free.pdf ) which has a very succinct introduction, but after chapter 2, becomes much more abstract than I care to deal with. I'm looking for simple applications that can be modeled in software, not the hardcore theory.

Ideally, I think I'd be satisfied if I could find a concrete non-commutative probability model for the game found in Sigfpe's blog on negative probabilities (http://blog.sigfpe.com/2008/04/negative-probabilities.html).

But the first step is to understand why random variables are so important in probability theory. since non-commutative probability is described exclusively in terms of them.

Any help would be greatly appreciated.
 
Last edited by a moderator:
Physics news on Phys.org
Tac-Tics said:
... The source I'm working off of is a paper by Mitchener (http://www.uni-math.gwdg.de/mitch/free.pdf ) which has a very succinct introduction, but after chapter 2, becomes much more abstract than I care to deal with. I'm looking for simple applications that can be modeled in software, not the hardcore theory.

...

But the first step is to understand why random variables are so important in probability theory. since non-commutative probability is described exclusively in terms of them.

One example is where you have several random variables representing various measurements of a single system; the event space formulation describes the dependence between those measurements in the most general possible way. For example with coin flipping, the measurements "number of heads" and "time of first head" relate in a nontrivial way but can be described in terms of their underlying events.

The event space formulation is most useful because it unifies probability and the theory of integration, one important consequence being that all random variables have a cdf regardless of whether they have a pmf or pdf or neither (e.g. Cantor), and other quantities such as moments (when they exist) can be expressed as a Stieltjes integral wrt the cdf.

Unfortunately the above paper's succint introduction puts the rest of it on shaky foundations. Theorem 1.9, which seems to be central to the arguments of the rest of the paper, states incorrectly and without proof that "random variables with the same moments have the same probability laws". A famous counterexample is the lognormal distribution, in fact there is an infinite family of distributions with all the same moments as lognormal (Heyde 1963; see also the Moment Problem). Fallacies like Theorem 1.9 tend to arise from misapplication of Taylor's theorem. Also there is no mention of random variables with infinite moments such as Cauchy and Pareto.
 
Last edited by a moderator:

Similar threads

  • · Replies 12 ·
Replies
12
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 7 ·
Replies
7
Views
4K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 3 ·
Replies
3
Views
4K
  • · Replies 14 ·
Replies
14
Views
3K
  • · Replies 21 ·
Replies
21
Views
7K
  • · Replies 13 ·
Replies
13
Views
4K
  • · Replies 7 ·
Replies
7
Views
9K