Calculating Conditional Probability of Y with Exponential RVs

  • Thread starter Thread starter EngWiPy
  • Start date Start date
  • Tags Tags
    Conditional Pdf
EngWiPy
Messages
1,361
Reaction score
61
Dear all,

How can I find the the following conditional probability: f_Y(y/\gamma_f\leq \gamma_0), where Y=\gamma_f+\text{min}(\gamma_h,\gamma_g), where all gammas with subscripts f, g, h are i.i.d exponential random variables.

Thanks in advance
 
Physics news on Phys.org
I would start out by working with cumulative distribution functions, and then take the derivative to get the density function. For instance, I would first find the cumulative distribution function for the random variable M = min(X_g, X_h), and then get the density from that. Then I would use the density for M together with the densities for X_f and X_0 to get the cumulative distribution function for Y given X_f <= X_0.

I'm guessing you are also doing this problem with the exponential distribution first in anticipation of doing it with more complicated distributions later. As for your other question about moment generating functions, I don't know if there is a general way to relate MGF of order statistics to the MGF of the i.i.d random variables. It sounds difficult.
 
techmologist said:
...

I'm guessing you are also doing this problem with the exponential distribution first in anticipation of doing it with more complicated distributions later.

...


Exactly. The problem is that I have worked it out and got a result that is differentf from the paper I am reading now. The authors divided the function into two intervals: \gamma\leq \gamma_0 and \gamma&gt;\gamma_0.
 
S_David said:
Exactly. The problem is that I have worked it out and got a result that is differentf from the paper I am reading now. The authors divided the function into two intervals: \gamma\leq \gamma_0 and \gamma&gt;\gamma_0.

Hmm. Did you work out an answer for the case of exponentially distributed RVs? For what it is worth, I got

f_Y(y|\gamma_f \leq \gamma_0) = \frac{4y}{\overline{\gamma}^2}e^{-\frac{2y}{\overline{\gamma}}}
 
techmologist said:
Hmm. Did you work out an answer for the case of exponentially distributed RVs? For what it is worth, I got

f_Y(y|\gamma_f \leq \gamma_0) = \frac{4y}{\overline{\gamma}^2}e^{-\frac{2y}{\overline{\gamma}}}

Actually, the answer in the paper is:

f_Y\left(y/\gamma_f\leq\gamma_0\right)=<br /> \begin{cases}<br /> \frac{\text{e}^{-y/\overline{\gamma}}-\text{e}^{-y/\overline{\gamma}_f}}{(\overline{\gamma}-\overline{\gamma}_f)\,\left(1-\text{e}^{-\gamma_0/\overline{\gamma}_f}\right)} &amp; \text{if}\,\,\, y\leq\gamma_0\\<br /> \frac{1-\text{e}^{-\gamma_0(1/\overline{\gamma}_f-1/\overline{\gamma})}}{(\overline{\gamma}-\overline{\gamma}_f)\,\left(1-\text{e}^{-\gamma_0/\overline{\gamma}_f}\right)}\,\text{e}^{-y/\overline{\gamma}} &amp; \text{if}\,\,\,y&gt;\gamma_0<br /> \end{cases}

where \overline{\gamma} is the mean of \text{min}(\gamma_h,\,\gamma_g)

Regards
 
***I just thought of this...does gamma_0 have the same distribution as the other gammas? I had assumed that in my answer above and in what follows below.*****

Is that answer supposed to be for the exponential distribution? I don't know how they got that, but I'm not saying it is incorrect. Also, if these are identically distributed random variables, then \overline{\gamma}_f = \overline{\gamma}_0 = 2\overline{\gamma}, which would simplify those expressions. I also don't see why it needs to be separated into two intervals of y. Do the authors explain any of this or do they present the answer out of thin air?

Here's how I got my answer. I am going to use X for the random variable so I don't have to keep typing out "gamma". Also, I will use the rate \lambda = 1/\overline{\gamma}, the inverse mean. So the density function for the X_i is


f_{X_i}(x) = \lambda e^{-\lambda x}

First of all, try to find the cumulative distribution function for Y given X_f <= X_0. This is

P\{Y \leq y | X_f \leq X_0\} = \frac{P\{Y \leq y , X_f \leq X_0\}}{P\{X_f \leq X_0\}}

I assumed that X_f and X_0 are identically distributed, in which case the probability in the denominator is just 1/2. But if that assumption was wrong, I don't think it is too hard to modify. So basically I'm interested in finding P\{Y \leq y , X_f \leq X_0\}. First I need to find the cumulative distribution function for the random variable M = min(X_h, X_g):

P\{M \leq x\} = 1 - P\{M &gt; x\} = 1 - P\{X_h&gt;x\}P\{X_g&gt;x\}
P\{M \leq x\} = 1 - e^{-2\lambda x} using i.i.d.

So the density for M is

f_M(x) = 2\lambda e^{-2\lambda x}, exponential with mean half that of the X's. Now to find

P\{X_f + M \leq y , X_f \leq X_0\}, I use the joint density for M, X_f, and X_0, which by independence is

f_{X_0, X_f, M}(r,s,t) = f_{X_0}(r)f_{X_f}(s) f_{M}(t) = \lambda e^{-\lambda r} \lambda e^{-\lambda s} 2\lambda e^{-2\lambda t}

Finally, compute the integral

P\{X_f + M \leq y , X_f \leq X_0\} = \int_{t=0}^y\int_{s=0}^{y-t}\int_{r=s}^{\infty}f_{X_0, X_f, M}(r,s,t)drdsdt = -\lambda y e^{-2\lambda y} + \frac{1}{2}(1-e^{-2\lambda y})

Differentiating this and multiplying by 2 (from my assumption that P{X_f <= X_0} = 1/2) gives the answer

f_Y(y| X_f \leq X_0) = 4\lambda ^2 y e^{-2 \lambda y}
 
Execuse me, but \gamma_0 is just a constant, and the random variables \gamma_f,\,\gamma_h,\,\text{and}\,\gamma_g are i.i.d, so \overline{\gamma}_f=\overline{\gamma}_h=\overline{\gamma}_g.

I am thinking in another way:

\text{Pr}\left[X_f+M\leq x,\,X_f\leq x_0\right]=\text{Pr}\left[2\,X_f+M\leq x+x_0\right]

given that x and x_0 must be >= 0.

I don't know if it is a valid approach or not.

Regards
 
S_David said:
Execuse me, but \gamma_0 is just a constant, and the random variables \gamma_f,\,\gamma_h,\,\text{and}\,\gamma_g are i.i.d, so \overline{\gamma}_f=\overline{\gamma}_h=\overline{\gamma}_g.

Ah, okay. I was confused about that. But this makes the problem easier, I think.

I am thinking in another way:

\text{Pr}\left[X_f+M\leq x,\,X_f\leq x_0\right]=\text{Pr}\left[2\,X_f+M\leq x+x_0\right]

given that x and x_0 must be >= 0.

I don't know if it is a valid approach or not.

That was actually my first thought, too. But it is incorrect for the following reason: although

X_f + M \leq y \text{ and } X_f \leq x_0 implies 2X_f + M \leq y+x_0 ,

the converse is not true. For example, you could have X_f + M = y + \epsilon, where epsilon is some positive number less than x_0. Then for values of X_f satisfying

X_f \leq x_0 - \epsilon ,

it would be true that 2X_f + M \leq y+x_0 .

Now that I know x_0 is a constant, I see why the answer is broken up into two intervals. You get the probability Pr\[ X_f + M \leq y , X_f \leq x_0 \] by integrating the joint density f_{X_f, M}(r,s) over a certain region in the r-s plane. The shape of that region depends on whether y is less than x_0 or greater than x_0. For the first case, the region is an equilateral right triangle. For the latter case, it is a quadrilateral (same triangle, but with one corner cut off).


EDIT: "equilateral right triangle" should be "isoscoles right triangle": two sides equal, not all three. haha :redface:
 
Last edited:
Were you able to get the author's answer?
 
  • #10
techmologist said:
Were you able to get the author's answer?

Not really. Have you?
 
  • #11
Yes. The joint density function for X_f and M is f_{X_f, M}(r,s) = \lambda_f e^{-\lambda_f r} \lambda_M e^{-\lambda_M s}, and

P\{X_f+M \leq y, X_f \leq x_0\} = \int_{r=0}^{y}\int_{s=0}^{y-r}f_{X_f, M}(r,s)dsdr

for y <= x_0, and

P\{X_f+M \leq y, X_f \leq x_0\} = \int_{r=0}^{x_0}\int_{s=0}^{y-r}f_{X_f, M}(r,s)dsdr

for y > x_0.

M is also an exponential random variable, independent of X_f. When I converted back to the notation of the authors: \lambda_f = 1/\overline{\gamma_f} and \lambda_M = 1/\overline{\gamma}, I got the same answers as you posted in post #5. EDIT: That is, of course, after taking the derivative w.r.t y and dividing by 1-e^{-\lambda_f x_0}

NOTE: I am having trouble previewing the latex images, so if what I posted looks totally strange, just ignore it. I'll fix it when I get to a better computer.
 
Last edited:
  • #12
It makes sense. That is great. Thanks for your help.

Regards
 
  • #13
You're welcome. Also, if you ever discover the general approach to that problem about moment generating functions, please post it. I'm curious now...
 
  • #14
techmologist said:
... if you ever discover the general approach to that problem about moment generating functions, please post it ...

Ok, I will. Thanks
 
Back
Top