Proof with regards to cumulative distribution function

Click For Summary

Discussion Overview

The discussion revolves around the properties and definitions of cumulative distribution functions (CDFs) in probability theory, specifically focusing on the relationship between probabilities involving random variables and their CDFs. Participants explore various formulations and implications of these relationships in both continuous and discrete cases.

Discussion Character

  • Technical explanation
  • Mathematical reasoning
  • Debate/contested

Main Points Raised

  • One participant seeks to prove the relationship P{x1 ≤ X ≤ x2} = F_X(x2) - F_X(x1^{-}) and expresses difficulty in rewriting this in terms of other related equations.
  • Another participant questions the definition of "cumulative distribution function" being used in the discussion.
  • A definition of the CDF is provided as F_X(x) = ∫_{-∞}^{x} f_X(μ) dμ, with a note on the importance of the limit from the left for both probability density functions (pdf) and probability mass functions (pmf).
  • It is noted that in the continuous case, the distinction between 'less than' and 'less than or equal to' does not affect the probability, while in the discrete case, it does.
  • One participant clarifies that for the discrete case, the probability must account for the lower boundary by considering the next lowest discrete point.
  • Another participant emphasizes that in the continuous case, the probability of a single data point is zero, leading to equivalence between two probability expressions.
  • Participants express appreciation for the clarifications provided, indicating that the discussion has been helpful.

Areas of Agreement / Disagreement

Participants generally agree on the definitions and implications of CDFs in continuous cases, while there is some contention regarding the treatment of boundaries in discrete cases. The discussion remains unresolved on certain nuances of these distinctions.

Contextual Notes

The discussion highlights the dependence on definitions of CDFs and the implications of continuity versus discreteness in probability calculations. There are unresolved aspects regarding the treatment of boundaries in discrete cases and the mathematical steps involved in the proofs.

Eidos
Messages
107
Reaction score
1
Hey guys

I'd like a steer in the right direction with this problem.
I would like to show that
[tex]P\{x_1\leq X \leq x_2\}=F_{X}(x_2)-F_{X}(x_1^{-})\quad(1)[/tex]

Where:
[tex]X[/tex] is a random variable.
[tex]F_{X}(x) \equiv P\{X \leq x \}[/tex] is its cumulative distribution function.

My notes only give an example (using dice) to show that this is true.

Generally
[tex]P\{x_1 < X \leq x_2\}=F_{X}(x_2)-F_{X}(x_1)\quad(2)[/tex]

and

[tex]P\{X = x_2\}=F_{X}(x_2)-F_{X}(x_2^{-})\quad (3)[/tex]
the latter of which is easy to prove.
I've been trying to rewrite (1) in terms of (2) & (3) but have had no success so far.
Any ideas would be most welcomed :smile:
 
Physics news on Phys.org
What definition of "cumulative distribution function" do you have?
 
[tex]F_{X}(x) = \int_{-\infty}^{x}f_{X}(\mu) d\mu[/tex]

That limit from the left in (1) is so that the same is true whether we have a pdf or a pmf. With the pmf we would have a sum, not an integral. It matters in the discrete case whether we have 'less than equals to' or just 'less than' for the lower bound in our probability, but in the continuous case (assuming of course that our cdf is differentiable everywhere) it doesn't matter since [tex]x_0^{-}=x_0[/tex].
 
Last edited:
Okay, so from that definition,
[tex]P(x_1< X\le x_2)= \int_{x_1}^{x_2}f_X(\mu)d\mu[/tex]
[tex]= \int_{-\infty}^{x_2}f_X(\mu)d\mu- \int_{-\infty}^{x_1} f_X(\mu)d\mu[/tex]
[tex]= F(x_2)- F(x_1)[/tex]
 
The only thing is though that we have not included the lower boundary x1 in our probability, but we have in the integral. How does that work, especially in the discrete case?

I know that the cdf is right continuous, and when we include the lower bound we take the next lowest discrete point than x1 which is x0.

That is [tex]P\{x_1 \leq X \leq x_2\}=F_{X}(x_2)-F_{X}(x_0)[/tex]

where: [tex]x_0=\lim_{x\rightarrow x_{1}^{-}}x[/tex]
 
In the continuous case, it doesn't matter: the probability of a single data point is always 0:
[tex]P(x_1< X\le x_2)= P(x1\le X\le x_2)[/itex]<br /> <br /> In the discrete case, there are two different probabilities:<br /> [tex]P(x_1< X\le x_2)= P(x1\le X\le x_2)- P(x_1)[/itex][/tex][/tex]
 
Cool thanks! :smile:

That last bit is exactly what I need.
 

Similar threads

  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 12 ·
Replies
12
Views
4K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 3 ·
Replies
3
Views
4K
  • · Replies 20 ·
Replies
20
Views
2K
  • · Replies 8 ·
Replies
8
Views
1K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 7 ·
Replies
7
Views
3K
  • · Replies 3 ·
Replies
3
Views
2K