Summing Probabilities: Gaussian vs Uniform Distribution

  • Thread starter Thread starter SW VandeCarr
  • Start date Start date
  • Tags Tags
    Probabilities
Click For Summary
SUMMARY

This discussion centers on the comparison of probabilities under Gaussian and uniform distributions, particularly in the context of infinite sets. It establishes that for a Gaussian distribution, the sum of probabilities in an infinite set E(G) converges to one, while for a uniform distribution E(U), the sum of probabilities approaches zero when dealing with infinite events. The conversation highlights that a continuous uniform distribution cannot exist for countable infinite events, as it leads to contradictions regarding probability spaces. The implications for Bayesian inference and Freiling's axiom of symmetry are also examined.

PREREQUISITES
  • Understanding of Gaussian distribution and its properties
  • Familiarity with uniform distribution and its implications
  • Knowledge of probability spaces and integrals
  • Concept of countable vs uncountable sets in probability theory
NEXT STEPS
  • Explore the properties of Gaussian distributions in depth
  • Investigate the implications of uniform distributions in Bayesian inference
  • Study the concept of probability density functions and their integrals
  • Examine Freiling's axiom of symmetry and its applications in probability theory
USEFUL FOR

Mathematicians, statisticians, data scientists, and anyone involved in probability theory or Bayesian statistics will benefit from this discussion.

SW VandeCarr
Messages
2,193
Reaction score
77
Consider an infinite set E(G) where the elements are interpreted as independent events assigned probabilities under a Gaussian distribution. It can be shown that the probabilities of all events in E(G) will sum to one. Now consider an infinite set E(U) with the same interpretation under a uniform distribution (every event has equal probability). Given an infinite set, this implies that the probability of any randomly chosen event in E(U) is zero. Is it then true that for any continuous uniform probability distribution, the sum of probabilities will be zero?
 
Physics news on Phys.org
Do you mean integral of the probabilities (for the continuous distribution)?
 
If your infinite set of independent events (call it A) is countable, then there can be no uniform distribution. Proof: Suppose that each element of A had equal probability. If this is zero, then the sum of all probabilities is zero; if it is nonzero, then the sum of all probabilities is infinite. In either case, you don't have a probability space.

If your set is uncountable, then you can't sum the probabilities of all the events (unless all but countably many have zero probability).
 
rochfor1 said:
Do you mean integral of the probabilities (for the continuous distribution)?

I'm actually talking about a random variable that maps from an infinite set of elements termed "events" to a continuous probability space [0,1] according to a Gaussian distribution (G) vs a uniform (U). The events are countable. The sum of the probabilities in E(G) will be '1' at the limit. The open integral of G is '1'. My question was with E(U). It appears that in any mapping of countable events in E(U) under a (continuous)uniform distribution U, the sum of probabilities will be '1' for any finite number of events, but zero for infinite events and the open integral of U is '0'. That is, a continuous uniform distribution does not exist (which is essentially what I conclude from adriank's post).

I asked this because 1) The uniform distribution is often used as the prior assumption in Bayesian inference. 2) In demonstrating Freiling's axiom of symmetry (AX) (see 'maze' "dense geodesic" post 32 Nov 30 in topology), a random variable chooses a number on the real interval [0,1] and maps to an (infinite?) set of countable subsets (S)on the same interval. (S) would be an infinite set of countable elements under a uniform distribution given a complete mapping, which according to the above, would be impossible since the probabilities sum to zero under a distribution whose open integral is zero.
 
Last edited:
To clarify the previous post, the probability p(x), given on a continuous distribution G, is expressed as a probability density [0, p(x)] on the definite integral [0,1] of G. I used the term "open" integral to exclude events with probability zero or one.

With respect to the uniform distribution U the same holds for a finite set of possible events 'x'. However, the uniform distribution must be discrete since the probability of one event is fixed as the reciprocal of the total number of events. For an infinite number events, a (continuous) uniform distribution seems impossible unless the reciprocal of infinity can be defined.

The demonstration of Freiling's axiom of symmetry (in the Wiki) seems to contemplate a random variable 'choosing' a number from the real interval [0,1] which implies a continuous uniform distribution since there's no reason to consider anyone number has a different probability than any other to be chosen.
 
Last edited:
If there are an infinite number of natural numbers, and an infinite number of fractions in between any two natural numbers, and an infinite number of fractions in between any two of those fractions, and an infinite number of fractions in between any two of those fractions, and an infinite number of fractions in between any two of those fractions, and... then that must mean that there are not only infinite infinities, but an infinite number of those infinities. and an infinite number of those...

Similar threads

  • · Replies 6 ·
Replies
6
Views
3K
  • · Replies 13 ·
Replies
13
Views
3K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 11 ·
Replies
11
Views
3K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 11 ·
Replies
11
Views
3K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 8 ·
Replies
8
Views
3K
  • · Replies 9 ·
Replies
9
Views
5K
  • · Replies 2 ·
Replies
2
Views
2K