The Sum of Geometric Series from Probability Theory - Comments

Click For Summary

Discussion Overview

The discussion revolves around the sum of geometric series in the context of probability theory. Participants explore the implications of probability measures, the validity of certain identities, and the conditions under which these identities hold. The conversation includes both theoretical considerations and practical applications, with a focus on mathematical rigor and the use of probability theory as a tool for deriving results.

Discussion Character

  • Exploratory
  • Technical explanation
  • Debate/contested
  • Mathematical reasoning

Main Points Raised

  • Some participants express concerns about circular reasoning in the derivation of probabilities, emphasizing the need for a fixed probability measure.
  • Others argue that the derivation is a practical trick that simplifies finding mathematical identities, even if it lacks rigorous justification.
  • A participant introduces a generalized identity involving varying probabilities in a game setting, suggesting that it can be proven more easily using probability theory.
  • There is a discussion about the conditions under which the formula holds, specifically that the probabilities must not approach zero as trials increase.
  • Some participants highlight the importance of ensuring that the probabilities remain bounded away from zero to avoid convergence issues.
  • Concerns are raised about the implications of changing probabilities on the overall outcome, with some arguing that it could lead to a nonzero chance of losing even with infinite trials.
  • One participant suggests that stating probabilities as greater than zero rather than simply non-zero could clarify the conditions for eventual success.
  • Another participant challenges the assertion that a fixed probability guarantees eventual success, arguing that the nature of the probabilities involved is crucial.

Areas of Agreement / Disagreement

Participants do not reach a consensus on the validity of the derivation or the implications of varying probabilities. Multiple competing views remain regarding the conditions necessary for the identities discussed and the interpretation of the results.

Contextual Notes

Participants note that the discussion involves assumptions about the behavior of probabilities in the limit and the implications of those behaviors on the outcomes of the series. There are unresolved mathematical steps regarding the convergence of series and the conditions under which they sum to one.

Demystifier
Science Advisor
Insights Author
Messages
14,694
Reaction score
7,298
Greg Bernhardt submitted a new blog post

The Sum of Geometric Series from Probability Theory
probability_theory.png


Continue reading the Original Blog Post.
 

Attachments

  • probability_theory.png
    probability_theory.png
    11.2 KB · Views: 1,148
  • Like
Likes   Reactions: kith, stevendaryl and Greg Bernhardt
Physics news on Phys.org
Demystifier said:

I think there might be some circular reasoning here. Before you can calculate probabilities, there should be a fixed probability measure no? That is, you start with a given probability space ##(\Omega, \mathcal{F}, \mathbb{P}) ## where ##\mathbb{P}## is a probability measure. In particular, we must have ##\mathbb{P}(\Omega) =1##.

You just claim that ##\mathbb{P}(\Omega) =1##, but this must be verified before starting the entire reasoning and is the crux of the entire matter.

Hope it's clear what I'm trying to say.
 
Last edited by a moderator:
  • Like
Likes   Reactions: DrClaude
@Math_QED your objection makes sense from the point of mathematical rigor. But mathematical rigor is not the point of my "derivation". The point is a practical trick that allows you to find a shortcut in finding mathematical identities. This becomes particularly interesting if one generalizes my trick to something more complicated. For instance, suppose that ##p## is not constant but has a different value ##x_k## in the ##k##th play of the game. Then my trick gives an immediate identity
$$1=x_1+(1-x_1)x_2+(1-x_1)(1-x_2)x_3+\cdots$$
Try to prove this identity by more conventional methods that do not depend on probability theory. It's not so easy. The probability theory allows you to see that it is true immediately, without any hard work. I find it amazing.
 
Last edited:
  • Like
Likes   Reactions: PeroK
Demystifier said:
@Math_QED your objection makes sense from the point of mathematical rigor. But mathematical rigor is not the point of my "derivation". The point is a practical trick that allows you to find a shortcut in finding mathematical identities. This becomes particularly interesting if one generalizes my trick to something more complicated. For instance, suppose that ##p## is not constant but has a different value ##x_k## in the ##k##th play of the game. Then my trick gives an immediate identity
$$1=x_1+(1-x_1)x_2+(1-x_1)(1-x_2)x_3+\cdots$$
Try to prove this identity by more conventional methods that do not depend on probability theory. It's not so easy. The probability theory allows you to see that it is true immediately, without any hard work. I find it amazing.

It's true that for fixed ##p## you must eventually win, but if the probability of winning on the ##k##th play reduces quickly, then you may not eventually win and the sum may be less than ##1##. For example:

If we let ##x_1 = \frac14## and ##x_k = \frac{1}{4^k}(1-x_1)(1- x_2) \dots (1-x_{k-1})##, then:

##x_1 + (1-x_1)x_2 + (1-x_1)(1-x_2)x_3 \dots = \frac14 + \frac{1}{16} + \frac{1}{64} \dots = \frac13##
 
  • Like
Likes   Reactions: StoneTemplePython
PeroK said:
It's true that for fixed p you must eventually win, but if the probability of winning on the kth play reduces quickly
Of course, but the point is that the formula is valid if ##x_k## does not go to ##0## in the limit ##k\rightarrow\infty##. And even more important point is that it is not easy to see that without using probability theory.

If one wants to put it more precisely, the formula is valid if there is ##\epsilon>0## such that ##\epsilon<x_k\leq 1## for all ##k##.
 
Last edited:
Demystifier said:
Of course, but the point is that the formula is valid if ##x_k## does not go to ##0## in the limit ##k\rightarrow\infty##. And even more important point is that it is not easy to see that without using probability theory.

If one wants to put it more precisely, the formula is valid if there is ##\epsilon>0## such that ##\epsilon<x_k\leq 1## for all ##k##.

When I first saw the previous post, I assumed you meant ##x_k \rightarrow 0##. Certainly, if ##x_k## is bounded below, then it's a really clever result.
 
  • Like
Likes   Reactions: Demystifier
To put a finer point on this, you have a sequence of independent events (coin tosses) with a goal in mind that they keep occurring until the first success. They are inhomogenous. In slightly more general settings you can get in trouble with this line of thinking because stochastic processes can be defective (i.e. CDF doesn't tend to one, aka Complementary CDF doesn't tend to zero).
- - - -
For this particular problem with probability of success on trial ##i## given by ##p_i##, the complementary CDF after trial ##n## is given by

##\prod_{i=1}^n (1-p_i) = \prod_{i=1}^n q_i##

passing limits as ##n\to \infty## we have the standard analytic fact that for ##p_i \in (0,1)##
##0\lt \prod_{i=1}^\infty (1-p_i)##
iff ##\sum_{i=1}^\infty p_i \lt \infty##
- - - -
edit:
for a sketch of a proof of this result
the upper bound for the above is easy, just use
##1-p_i \leq e^{-p_i}##, so
##0\leq \prod_{i=1} (1-p_i) \leq e^{-\sum_i p_i}##
and if the series diverges then our product is squeezed between zero and zero

the lower bound takes a little more care, but one way to prove it is to take logs of the product and try to prove that
##-\infty \lt \sum_{i=1}^\infty \log(1-p_i)##

then exploiting geometric series and the power series for the logarithm, see the point-wise bound that holds for sufficiently small ##p_i## (in this case ##p_i\leq \frac{1}{2})##

##-2 p_i \leq \frac{-p_i}{1-p_i} = \sum_{k=1}^\infty -p_i^k \leq \sum_{k=1}^\infty -\frac{p_i^k}{k} = \log(1-p_i)##

then sum over the bound, for all but finitely many ##p_i##
- - - -
Demystifier said:
If one wants to put it more precisely, the formula is valid if there is ##\epsilon>0## such that ##\epsilon<x_k\leq 1## for all ##k##.
in this case since ## p_i## is bounded below by a positive constant, the series diverges which implies that the complementary CDF tends to zero and thus the probability distribution sums to one.
- - - -
For a slightly different look, in effect you're using the standard (countable state) Renewal Matrix for Age of a stochastic process

##
\left[\begin{matrix}
p_1 & q_1 & 0 & 0 & 0 & 0 & \dots\\
p_2 & 0 & q_2& 0 & 0 & 0 & \dots\\
p_3 & 0 & 0 & q_3 & 0 & 0 & \dots\\
p_4 & 0 & 0 & 0 & q_4 & 0 & \dots\\
p_5 & 0 & 0 & 0 & 0 & q_5 & \dots\\
p_6 & 0 & 0 & 0 & 0 & 0 & \dots\\
\vdots & \vdots & \vdots & \vdots & \vdots & \vdots & \ddots\end{matrix}\right]##

and trying to verify that state zero is not transient.
 
Last edited:
  • Like
Likes   Reactions: PeroK and Greg Bernhardt
That is a slick trick, which I think would be teaching someone how to figure the geometric series formula.

Should you state at the beginning that the probability ## p > 0 ##, rather than saying ##p\neq 0##? This goes along with ## q < 1 ## constraint at the end.

If you say ## p\neq 0 ##, it could suggest that p is less than zero. With a case of less than zero chance, you cannot accept that eventually you will win, anymore.
 
  • Like
Likes   Reactions: Demystifier
Demystifier said:
Of course, but the point is that the formula is valid if ##x_k## does not go to ##0## in the limit ##k\rightarrow\infty##. And even more important point is that it is not easy to see that without using probability theory.

If one wants to put it more precisely, the formula is valid if there is ##\epsilon>0## such that ##\epsilon<x_k\leq 1## for all ##k##.

One can do it with algebra instead of probability. For finite ##n## let
$$S = x_1 + (1-x_1)x_2 +(1-x_1)(1-x_2) x_3 + \cdots +(1-x_2)(1-x_2) \cdots (1-x_n) x_{n+1}.$$
We have
$$1-S = (1-x_1)(1-x_2) \cdots (1-x_{n+1}).$$
 
  • Like
Likes   Reactions: hutchphd, Demystifier and PeroK
  • #10
PeroK said:
It's true that for fixed ##p## you must eventually win, but if the probability of winning on the ##k##th play reduces quickly, then you may not eventually win and the sum may be less than ##1##.
Say what?

If the probability of winning one play is 1/4 then the probability of not winning after k plays is (1- 1/4)k and lim k→∞ (3/4)k = 0. I.e. the probability of NOT winning after ∞ plays is 0. You will eventually win regardless of p, assuming 0<p<1..

The fact that p = 1/4 or some other smaller but finite positive number < 1 is immaterial.

Good paper!
 
  • #11
rude man said:
Say what?

If the probability of winning one play is 1/4 then the probability of not winning after k plays is (1- 1/4)k and lim k→∞ (3/4)k = 0. I.e. the probability of NOT winning after ∞ plays is 0. You will eventually win regardless of p, assuming 0<p<1..

The fact that p = 1/4 or some other smaller but finite positive number < 1 is immaterial.

Good paper!

Not all series sum to 1.
 
  • #12
rude man said:
Say what?

If the probability of winning one play is 1/4 then the probability of not winning after k plays is (1- 1/4)k and lim k→∞ (3/4)k = 0. I.e. the probability of NOT winning after ∞ plays is 0. You will eventually win regardless of p, assuming 0<p<1..

The fact that p = 1/4 or some other smaller but finite positive number < 1 is immaterial.

Good paper!

I think the point is that if the odds are CHANGING, then you have nonzero chance of losing, even with infinitely many trials. For example, if the first time you play, the probability of winning is 1/4, and the next time it's 1/8, and the next time it's 1/16, etc, then your probability of winning for infinitely many trials is:

1/4 + 3/4 x 1/8 + 3/4 x 7/8 x 1/16 + ...

which is...I don't know what it is, but it adds up to something less than 1/2.
 
  • #13
rude man said:
Say what?

If the probability of winning one play is 1/4 then the probability of not winning after k plays is (1- 1/4)k and lim k→∞ (3/4)k = 0. I.e. the probability of NOT winning after ∞ plays is 0. You will eventually win regardless of p, assuming 0<p<1..

The fact that p = 1/4 or some other smaller but finite positive number < 1 is immaterial.

Good paper!

If the win-probabilities ##p_n## change with the round-number ##n## you may not be guaranteed an ultimate win. For instance, if ##p_n = 1/2^n,## the probability of losing the first ##N## games is
$$P(N) = \prod_{n=1}^N \left( 1 - \frac{1}{2^n} \right),$$ and the probability of no-win is ##P^* = \lim_{N \to \infty} P(N).## This limit is between 0.2887880492 and 0.2887881410, so the probability of a win is ##1-P^* \doteq 0.71121 ##
 
  • #14
Demystifier said:
The Sum of Geometric Series from Probability Theory
So you are showing that the mass function for a geometric distribution sums to 1? Not exactly a novel result.
 

Similar threads

  • · Replies 10 ·
Replies
10
Views
2K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 11 ·
Replies
11
Views
3K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 2 ·
Replies
2
Views
3K
  • Sticky
  • · Replies 44 ·
2
Replies
44
Views
14K
Replies
4
Views
2K
  • · Replies 21 ·
Replies
21
Views
363
  • · Replies 19 ·
Replies
19
Views
3K