I CDF of summation of random variables

AI Thread Summary
The discussion centers on finding the cumulative distribution function (CDF) of the random variable β, defined as the sum of independent and identically distributed (i.i.d.) random variables α_k, each with a specified CDF and PDF. The initial approach using the Moment Generating Function (MGF) leads to complications, particularly concerning the existence of moments and the validity of the MGF due to the infinite expected value of the random variables involved. Alternative methods, such as using characteristic functions or direct convolution, are suggested as potentially simpler approaches for computing the CDF. The conversation also touches on the implications of not having a finite expected value, which affects the applicability of various statistical theorems. The participants express a need for further exploration of these methods to find a more elegant solution.
EngWiPy
Messages
1,361
Reaction score
61
Hi,

I have this random variable ##\beta=\sum_{k=1}^K\alpha_k##, where ##\{\alpha_k\}_{k=1}^{K}## are i.i.d. random variables with CDF ##F_{\alpha}(\alpha)=1-\frac{1}{\alpha+1}## and PDF ##\frac{1}{(1+\alpha)^2}##. I want to find the CDF of the random variable ##\beta##. So, I used the Moment Generating Function (MGF) which is defined as:

\mathcal{M}_{\beta}(s)=E_{\beta}[e^{s\beta}]=\prod_{k=1}^K\mathcal{M}_{\alpha_k}(s)

The CDF then can be found using the inverse Laplace transform of ##\frac{\mathcal{M}_{\beta}(-s)}{s}##. After doing all of this I ended up with

\frac{\mathcal{M}_{\beta}(-s)}{s}=\sum_{l=0}^K{K\choose l}s^{l-1}\,e^{ls}\text{Ei}^l(-s)

where ##\text{Ei}(x)## is the exponential integral function. I have two questions:

  1. Is there another more simple way of computing the CDF?
  2. Can I evaluate the inverse Laplace transform of ##\frac{\mathcal{M}_{\beta}(-s)}{s}## other than numerically?
Thanks in advance
 
Physics news on Phys.org
There is an alternative method, using characteristic functions instead of moment generating functions, where the inverse transform is Fourier transform rather than Laplace. Whether this will be any better I couldn't say. If you are not familiar with ch. functions, it is like the moment generator with is in the exponent instead of s.
 
EngWiPy said:
Is there another more simple way of computing the CDF?

Yes, just do ##X_1 + X_2## (i.e. a convolution) directly. Then once you get comfortable with that, convolve that result with ##X_3##, and so on.
- - - -

There are a lot of problems in your writeup.

EngWiPy said:
Hi,

I have this random variable ##\beta=\sum_{k=1}^K\alpha_k##, where ##\{\alpha_k\}_{k=1}^{K}## are i.i.d. random variables with CDF ##F_{\alpha}(\alpha)=1-\frac{1}{\alpha+1}## and PDF ##\frac{1}{(1+\alpha)^2}##. I want to find the CDF of the random variable ##\beta##. So, I used the Moment Generating Function (MGF) which is defined as:

\mathcal{M}_{\beta}(s)=E_{\beta}[e^{s\beta}]=\prod_{k=1}^K\mathcal{M}_{\alpha_k}(s)

The convention is to use a capital letter like ##X## to denote the random variable, and specific values it takes on with lower case, as in ##X(\omega) = x## or in the common shorthand, for ##X = x##. You seem to have overloaded ##\alpha##.

You didn't state that domain that it takes values in, but from what I can tell, it is a real non-negative random variable. The PDF looks close to a cauchy distirbution, but you cannot have negative values here or else your CDF is illegal (i.e. CDF's are bounded in [0,1]).

The CDF only reaches one as ##x \to \infty##, but ##E[X] = \int_{\eta}^{\infty} (1-CDF)dx = \infty## hence your random variable doesn't even have a first moment.

for any real ##r \neq 0##

Your MGF transform = ##E\Big[\exp\big(rX\big)\Big] \geq \exp \Big( E \big[rX\big] \Big) = \exp \Big( r E \big[X\big] \Big) = \exp \Big( r \infty \Big) = \infty ##

by Jensen's Inequality,

hence the MGF is worthless, as it maps your random variable to 1, with r =0 and ##\infty## otherwise. There are many non-defective random variables that I can think of that don't have a first moment, that get the same results, hence your transform cannot be invertible.
- - - -
edit: technical nit:

it really should say something like

for any real ##r \gt 0##

Your MGF transform = ##E\Big[\exp\big(rX\big)\Big] \geq \exp \Big( E \big[rX\big] \Big) = \exp \Big( r E \big[X\big] \Big) = \exp \Big( r \infty \Big) = \infty ##
- - - - -
edit 2: the integral exists for any ##r \leq 0##. This transform doesn't generate moments anymore since the derivatives don't exist at r = 0, however there may be some useful things that can be done including a kind of Chernoff bound, again with restriction that ##r \lt 0##. That is of some interest.

Still, characteristic function approach seems maybe preferable.
 
Last edited:
mathman said:
There is an alternative method, using characteristic functions instead of moment generating functions, where the inverse transform is Fourier transform rather than Laplace. Whether this will be any better I couldn't say. If you are not familiar with ch. functions, it is like the moment generator with is in the exponent instead of s.

I am somewhat familiar with it, and as far as I know it follows the same procedure as the MGF, but it uses Fourier transform instead of Laplace transform. I need something that can simplify the result (if possible) such that I can write it in an elegant form.
 
Last edited:
StoneTemplePython said:
Yes, just do ##X_1 + X_2## (i.e. a convolution) directly. Then once you get comfortable with that, convolve that result with ##X_3##, and so on.
...

It doesn't sound like an efficient way for more than two random variables, unless there is a pattern in the convolution process.

I didn't understand what you mean by ##E[X]=\infty##! What is ##X##, ##\eta## and ##CDF## in your analysis?
 
EngWiPy said:
It doesn't sound like an efficient way for more than two random variables, unless there is a pattern in the convolution, which I suspect doesn't exist.

I didn't understand what you mean by ##E[X]=\infty##! What is ##X##, ##\eta## and ##CDF## in your analysis?

That's right -- sometimes, though, there is a pattern. One tell tale sign in convolving certain pathological distributions like Cauchy, is the nth convolution looks identical to the original distribution. Patiently convolving exponential distributions gives the pattern for the Erlang (which can be then proven with induction) and so on.

As for the expected value, you need a random variable with capital letters here. I went with ##X## by custom. You didn't specify the domain your random variable ##X## exists in, but based on your cdf formula ##X## can only be real non-negative. So for any ##\eta \geq 0## (i.e. where ##\eta ## is the lowest value that ##X## takes on) that integral to from ##\eta ## to ##\infty## gives you infinity. In general for real non-negative random variables, integrating 1 minus the CDF (i.e. integrating the complement of the CDF) will give you the expected value of the random variable. When your r.v. doesn't have a first moment, you end up quite limited in what you can do with it and what you can say about it.
 
StoneTemplePython said:
That's right -- sometimes, though, there is a pattern. One tell tale sign in convolving certain pathological distributions like Cauchy, is the nth convolution looks identical to the original distribution. Patiently convolving exponential distributions gives the pattern for the Erlang (which can be then proven with induction) and so on.

As for the expected value, you need a random variable with capital letters here. I went with ##X## by custom. You didn't specify the domain your random variable ##X## exists in, but based on your cdf formula ##X## can only be real non-negative. So for any ##\eta \geq 0## (i.e. where ##\eta ## is the lowest value that ##X## takes on) that integral to from ##\eta ## to ##\infty## gives you infinity. In general for real non-negative random variables, integrating 1 minus the CDF (i.e. integrating the complement of the CDF) will give you the expected value of the random variable. When your r.v. doesn't have a first moment, you end up quite limited in what you can do with it and what you can say about it.

Yes, the domain is ##[0,\infty)##. I will try the convolution for K=3 and see if I get a pattern.

I just viewed the Cauchy Distribution on Wikipedia and it says that:

\sum_{k=1}^Kw_k\frac{X_k}{Y_k} follows standard Cauchy distribution where ##X_k## and ##Y_k## are both Gaussian distributed with mean 0 and variance 1, and ##\sum_kw_k=1##. In my case, actually, ##\alpha_k=\frac{a_k}{b_k}##, but both ##a_k## and ##b_k## are exponential random variables with mean 1 (they are the magnitude squared of Gaussian RVs with mean 0 and variance 1). It is also said that Cauchy distribution has no mean and MGF, but it has a characteristic function. However, the pdf is different in both cases. For standar Cauchy distrution the PDF is

f_X(x)=\frac{1}{\pi}\frac{1}{1+x^2}

while in my case the PDF is given by

f_X(x)=\frac{1}{(1+x)^2}
 
Last edited:
I'd give it a shot. The idea of characteristic function is to use points on the unit circle to transform your problem, while not distorting length in the process -- hence your best shot of not having a blow up to infinity type of issue.

- - - - -
I actually think I've seen your PDF before, but I'm drawing a blank right now as to where. To the extent its (somewhat) well known, you could just look up the known results.

Just remember that when you don't have a finite expected value, all the various limit theorems fail to apply. This means there is no law of large numbers for your problem, and a lot of things I personally take for granted, won't apply here.
 
StoneTemplePython said:
I'd give it a shot. The idea of characteristic function is to use points on the unit circle to transform your problem, while not distorting length in the process -- hence your best shot of not having a blow up to infinity type of issue.

I think I will try the CF approach (as well as the convolution).

StoneTemplePython said:
I actually think I've seen your PDF before, but I'm drawing a blank right now as to where. To the extent its (somewhat) well known, you could just look up the known results.

"known results" like what and where?

StoneTemplePython said:
Just remember that when you don't have a finite expected value, all the various limit theorems fail to apply. This means there is no law of large numbers for your problem, and a lot of things I personally take for granted, won't apply here.

What are the implications of not having a finite expected value and not having the law of large numbers be applicable?
 
  • #10
EngWiPy said:
"known results" like what and where?
Basically like what you looked up for Cauchy -- you could look up the result of the convolution for ##n ## Cauchy r.v.'s -- people have already done this (and posted findings on the Internet). If I could only remember the name of the distribution -- it may or may not come to me.

EngWiPy said:
What are the implications of not having a finite expected value and not having the law of large numbers be applicable?

I'm not really sure what to say here. The Expected Value is, I'd suggest, the single most important calculation in all of probability theory. You don't have an expected value.

Most people have a notion that when you add up lots of observations of random variables, and average them, you tend to see 'long-run' behavior. But that is out -- this requires Weak (or Strong) Law of Large Numbers to justify but you don't have an expected value so no go. In general convolving lots of random variables is messy and difficult and its much easier to apply various useful bounds but they pretty much all require a first moment -- you can't even use Markov's Inequality here. I'm not aware of much work being done on random variables like this, but maybe there is literature on dealing with them.
- - -

EngWiPy said:
while in my case the PDF is given by

f_X(x)=\frac{1}{(1+x)^2}

Btw, I think your PDF integrates to ##\frac{\pi}{2} \neq 1##, hence you need to rescale the PDF so that it integrates to one, and thus becomes a valid probability density function.
 
  • #11
I, myself, suspect it must had be done, but not sure where to look.

As, for the PDF scaling, I derived it from the CDF which is ##1-\frac{1}{1+x}## which evaluates to 1 as ##x\to\infty##, which should be the same as integrating the PDF to infinity, right?
 
  • #12
EngWiPy said:
I, myself, suspect it must had be done, but not sure where to look.

As, for the PDF scaling, I derived it from the CDF which is ##1-\frac{1}{1+x}## which evaluates to 1 as ##x\to\infty##, which should be the same as integrating the PDF to infinity, right?

whoops. Ignore my PDF comment. You wrote ##f_X(x)=\frac{1}{(1+x)^2}## but for some reason I read it as ##
f_X(x)=\frac{1}{(1+x^2)}
## which is not the same thing, of course
 
  • #13
No worries.

I am trying to follow the CF approach, and got stuck at this integral:

\int_0^{\infty}\frac{e^{itx}}{(1+x)^2}\,dx

I searched some Fourier transform tables, and all of them include the expression ##1/(1+x^2)## (which is the Cauchy distribution) but not ##1/(1+x)^2##! I didn't find such an expression in the table of integrals either. I tried to do the integration by parts, but I got ##\infty##! Is this integral integrable?
 
  • #14
EngWiPy said:
No worries.

I am trying to follow the CF approach, and got stuck at this integral:

\int_0^{\infty}\frac{e^{itx}}{(1+x)^2}\,dx

I searched some Fourier transform tables, and all of them include the expression ##1/(1+x^2)## (which is the Cauchy distribution) but not ##1/(1+x)^2##! I didn't find such an expression in the table of integrals either. I tried to do the integration by parts, but I got ##\infty##! Is this integral integrable?
It is integrable, but I don't know if there is an explicit expression.
 
  • #15
Yes, I meant to write it as an expression in a compact form, not integration, because this is an intermediate step.
 
  • #16
Is it possible to approximate ##\sum_{k=1}^K\frac{x_i}{y_i}## as ##K\to\infty## in any simple way, where ##\{x_i\}_{i=1}^K## and ##\{y_i\}_{i=1}^K## are all i.i.d. random variables that follow exponential distribution with mean 1? If not, what about ##\sum_{k=1}^K\frac{1}{y_i}##?
 
  • #17
EngWiPy said:
\int_0^{\infty}\frac{e^{itx}}{(1+x)^2}\,dx

Could anyone comment on this integral if it can be evaluated in a compact form or I need to evaluate it numerically?

Thanks in advance
 
Back
Top