Proof with regards to cumulative distribution function

AI Thread Summary
The discussion focuses on proving the relationship between probabilities and the cumulative distribution function (CDF) for a random variable X. The main equation under consideration is P{x1 ≤ X ≤ x2} = F_X(x2) - F_X(x1^-), highlighting the importance of the limit from the left in ensuring the validity for both probability density functions (pdf) and probability mass functions (pmf). The conversation clarifies that in continuous cases, the distinction between "less than" and "less than or equal to" does not affect the probability, while in discrete cases, it does, necessitating the use of the next lowest point. The participants discuss how to reconcile these differences in definitions and calculations, particularly regarding the treatment of boundaries in the probabilities. Overall, the thread emphasizes the nuances of applying the CDF in both continuous and discrete scenarios.
Eidos
Messages
107
Reaction score
1
Hey guys

I'd like a steer in the right direction with this problem.
I would like to show that
P\{x_1\leq X \leq x_2\}=F_{X}(x_2)-F_{X}(x_1^{-})\quad(1)

Where:
X is a random variable.
F_{X}(x) \equiv P\{X \leq x \} is its cumulative distribution function.

My notes only give an example (using dice) to show that this is true.

Generally
P\{x_1 < X \leq x_2\}=F_{X}(x_2)-F_{X}(x_1)\quad(2)

and

P\{X = x_2\}=F_{X}(x_2)-F_{X}(x_2^{-})\quad (3)
the latter of which is easy to prove.
I've been trying to rewrite (1) in terms of (2) & (3) but have had no success so far.
Any ideas would be most welcomed :smile:
 
Physics news on Phys.org
What definition of "cumulative distribution function" do you have?
 
F_{X}(x) = \int_{-\infty}^{x}f_{X}(\mu) d\mu

That limit from the left in (1) is so that the same is true whether we have a pdf or a pmf. With the pmf we would have a sum, not an integral. It matters in the discrete case whether we have 'less than equals to' or just 'less than' for the lower bound in our probability, but in the continuous case (assuming of course that our cdf is differentiable everywhere) it doesn't matter since x_0^{-}=x_0.
 
Last edited:
Okay, so from that definition,
P(x_1< X\le x_2)= \int_{x_1}^{x_2}f_X(\mu)d\mu
= \int_{-\infty}^{x_2}f_X(\mu)d\mu- \int_{-\infty}^{x_1} f_X(\mu)d\mu
= F(x_2)- F(x_1)
 
The only thing is though that we have not included the lower boundry x1 in our probability, but we have in the integral. How does that work, especially in the discrete case?

I know that the cdf is right continuous, and when we include the lower bound we take the next lowest discrete point than x1 which is x0.

That is P\{x_1 \leq X \leq x_2\}=F_{X}(x_2)-F_{X}(x_0)

where: x_0=\lim_{x\rightarrow x_{1}^{-}}x
 
In the continuous case, it doesn't matter: the probability of a single data point is always 0:
P(x_1&lt; X\le x_2)= P(x1\le X\le x_2)[/itex]<br /> <br /> In the discrete case, there are two different probabilities:<br /> P(x_1&amp;lt; X\le x_2)= P(x1\le X\le x_2)- P(x_1)[/itex]
 
Cool thanks! :smile:

That last bit is exactly what I need.
 
Back
Top