Moments of the Poisson distribution

Click For Summary

Discussion Overview

The discussion revolves around the calculation of the first moment of the Poisson distribution using its characteristic function and moment-generating function (MGF). Participants explore different methods for deriving the first moment and express confusion regarding the application of the characteristic function.

Discussion Character

  • Technical explanation
  • Mathematical reasoning
  • Debate/contested

Main Points Raised

  • One participant expresses difficulty in obtaining the first moment of the Poisson distribution using the characteristic function, leading to a result of zero instead of the expected value "a".
  • Another participant suggests using the moment-generating function (MGF) instead, stating that differentiating the MGF at t=0 is a more straightforward approach to find the moment.
  • A participant provides a derivation of the first moment using the summation definition, confirming that the result is indeed "a".
  • There is a discussion about the relationship between the MGF, characteristic function, and probability density function (PDF), with some participants emphasizing the importance of understanding these connections.
  • One participant points out a potential mistake in applying the chain rule when differentiating the characteristic function, suggesting that this could lead to the incorrect result of zero.
  • Another participant clarifies that the PDF for the Poisson distribution is not differentiable as it is a discrete distribution, leading to a discussion about the terminology used in relation to the PDF and probability mass function.
  • There is mention of using contour integration and the residue theorem to evaluate integrals related to the characteristic function, but the details remain unresolved.

Areas of Agreement / Disagreement

Participants do not reach a consensus on the best method to derive the first moment of the Poisson distribution. There are competing views on the use of the characteristic function versus the moment-generating function, and some participants express confusion about the calculations involved.

Contextual Notes

Some participants note that the characteristic function and MGF are related, but the exact nature of this relationship and its implications for calculating moments remains unclear. The discussion also highlights the distinction between the PDF and probability mass function for discrete distributions like the Poisson distribution.

fluidistic
Gold Member
Messages
3,932
Reaction score
283
I cannot seem to get the first moment of Poisson's distribution with parameter a: P(n_1)=\frac{a^{n_1}e^{-a}}{n_1!} when using the characteristic function \phi _X (k)=\exp [a(e^{ik}-1)].
The definition of the first moment involving the characteristic function is <n_1>=\frac{i}{n} \frac{d \phi _X (k)}{dk} \big | _{k=0}.
I get <n_1>=\frac{1}{i} a(e^{ik}-1)aie^{ik}e^{a(e^{ik}-1)} \big | _{k=0}=0 because of the factor e^{ik}-1\big | _{k=0}=1-1=0.
However I should reach <n_1>=a.
I really do not see what I did wrong. I do know that the characteristic function is OK, according to wikipedia and mathworld. I also do know I should reach that the mean or first moment is worth "a" but I'm getting 0.
I've applied twice the chain rule and do not see any mistake, but obviously I'm doing at least one somewhere. Any help is welcome!
 
Physics news on Phys.org
Hey fluidistic.

You are better off using a moment generating function and the differentiating once and setting the parameter t = 0 to get the moment.

The definition of the MGF is MGF_X(t) = E[e^tX].
 
First of all, thanks for helping me.
chiro said:
Hey fluidistic.

You are better off using a moment generating function and the differentiating once and setting the parameter t = 0 to get the moment.

The definition of the MGF is MGF_X(t) = E[e^tX].
Yeah I know another way (probably the one you mention?) to get the result, namely "a".
<n_1>=\sum _{n_1=0} ^ \infty n_1 \frac{a^{n_1}}{n_1!} e^{-a}=e^{-a} \sum _{n_1=0}^{ \infty } \frac{a^{n_1}}{(n_1-1)!}=e^{-a} \sum _{n_1=0}^{\infty } a \cdot \frac{a^{n-1}}{(n_1-1)!}=e^{-a}ae^a=a. I've no problem with this result since it is the right answer.
However what troubles me a lot is when I want to use the characteristic function. According to the book of Reichl's "A modern course in statistical physics", 2nd edition, page 189 it should be easy. Namely
Reichl said:
Furthermore, if we know the characteristic function we can obtain moments by differentiating: <x^n>= \lim _{k\to 0} (-i)^n \frac{d^n f_X (x)}{dk^n}
where I think he meant f_X (k) (as it is definied earlier and employed further.).
Using that given formula I could get the first moment of the binomial distribution. I'm losing my mind at understanding what's wrong with what I've done for the Poisson's distribution. It's not that I don't know how to apply the formula, it's that it gives me 0 no matter what instead of "a".
 
The MGF and PDF are directly related to the characteristic function.

You might want to look at this relation (basically an inverse Fourier transform type integral on the MGF over the real line gives back the PDF), and see the formula you have stated above relates to the actual MGF formula for getting moments.

I think this will clear up a lot of problems for since you can calculate the MGF easily and you can see analytically how the characteristic transformation and the MGF are both related to the PDF.
 
fluidistic said:
I cannot seem to get the first moment of Poisson's distribution with parameter a: P(n_1)=\frac{a^{n_1}e^{-a}}{n_1!} when using the characteristic function \phi _X (k)=\exp [a(e^{ik}-1)].
The definition of the first moment involving the characteristic function is <n_1>=\frac{i}{n} \frac{d \phi _X (k)}{dk} \big | _{k=0}.
I get <n_1>=\frac{1}{i} a(e^{ik}-1)aie^{ik}e^{a(e^{ik}-1)} \big | _{k=0}=0 because of the factor e^{ik}-1\big | _{k=0}=1-1=0.
However I should reach <n_1>=a.
I really do not see what I did wrong. I do know that the characteristic function is OK, according to wikipedia and mathworld. I also do know I should reach that the mean or first moment is worth "a" but I'm getting 0.
I've applied twice the chain rule and do not see any mistake, but obviously I'm doing at least one somewhere. Any help is welcome!

It looks like you made a mistake applying the chain rule.
$$\frac{d}{dk} \exp(f(k)) = \exp(f(k)) \frac{df(k)}{dk};$$
with ##f(k) = a(\exp(ik)-1)## you get ##df/dk = ai \exp(ik)##. The -1 drops out, so the moment doesn't vanish, and you get ##\langle n_1 \rangle = (1/i)(ai) = a##, as expected.

chiro said:
The MGF and PDF are directly related to the characteristic function.

You might want to look at this relation (basically an inverse Fourier transform type integral on the MGF over the real line gives back the PDF), and see the formula you have stated above relates to the actual MGF formula for getting moments.

The inverse Fourier transform over the characteristic function gives the probability density function. Maybe an analagous formula holds for getting the pdf from the moment generating function by manipulating the integral with contour methods (giving some sort of inverse Laplace transform, perhaps), but I haven't seen such a formula in use before.
 
Last edited:
Mute said:
The inverse Fourier transform over the characteristic function gives the probability density function. Maybe an analagous formula holds for getting the pdf from the moment generating function by manipulating the integral with contour methods (giving some sort of inverse Laplace transform, perhaps), but I haven't seen such a formula in use before.

It's a very simple relationship:

http://en.wikipedia.org/wiki/Moment-generating_function#Other_properties
 
chiro said:

Yes, I am aware of the relationship between the moment-generating function and the characteristic function.

What I was saying was the pdf is obtained directly by inverse Fourier transforming the characteristic function, not the moment-generating function. You can of course use the relation ##\varphi_X(t) = M_X(it)## to cosmetically rewrite the inverse formula in terms of the the moment-generating function of imaginary argument, but if you're going to stop there you might as well just keep things in terms of the characteristic function.

What I was getting at was that if you perform a rotation of the variables in the complex plane such that ##M_X(it) \rightarrow M_X(t)##, then the inverse formula is no longer an inverse Fourier transform, but rather an inverse Laplace transform, or something like it (there are perhaps some subtleties there, I haven't thought through it carefully, but naively it looks like you'd get an inverse Laplace transform). I was commenting that I haven't seen anyone write down the pdf in terms of an inverse Laplace transform over the moment-generating function, perhaps because it's easier to just inverse Fourier transform the characteristic function.

My goal here was to be precise so that the OP doesn't try to inverse Fourier transform ##M_X(t)## rather than ##\varphi_X(t) = M_X(it)## and then wonder why he's not getting the right answer.
 
Mute said:
I was commenting that I haven't seen anyone write down the pdf in terms of an inverse Laplace transform over the moment-generating function, perhaps because it's easier to just inverse Fourier transform the characteristic function.

My goal here was to be precise so that the OP doesn't try to inverse Fourier transform ##M_X(t)## rather than ##\varphi_X(t) = M_X(it)## and then wonder why he's not getting the right answer.

There is no pdf for the Poisson distribution. As a discrete distribution, it is not differentiable.
 
Last edited:
SW VandeCarr said:
There is no pdf for the Poisson distribution. As a discrete distribution, it is not differentiable.

Yes, in the spirit of being precise with terminology, there of course there is no pdf, but the probability mass function is still given by the inverse Fourier transform.

For example, inverse Fourier transforming the characteristic function of the Poisson distribution,

$$\int_{-\infty}^\infty \frac{dt}{2\pi} e^{\lambda(\exp(it)-1)} e^{-ikt},$$

let's change variables to the complex variable ##z = \exp(it)##. The integral can be written as the following contour integral in the complex plane, which we can then evaluate via the residue theorem, but we are forced to take ##k## to be an integer (as expected):

$$\frac{1}{2\pi i}\oint_\gamma \frac{dz}{z^{k+1}} e^{\lambda z} e^{-\lambda} = \frac{1}{k!}\frac{d^k}{dz^k}\left[e^{\lambda z}\right]_{z=0} e^{-\lambda} = \frac{\lambda^k}{k!}e^{-\lambda},$$
which is, of course, the probability mass function of the Poisson distribution. Here the contour ##\gamma## is a closed contour from ##-R## to ##R## along the real line, closed by a circular arc of radius ##R##, taking the ##R \rightarrow \infty## limit. The contribution from the arc will vanish in this limit, demonstrating equality between the contour integral and the original Fourier integral.

At any rate, the discussion of inverse transforms of characteristic functions to get pdf's (or pmf's) was tangential to the OP's original question, and we should probably not lead the discussion too far off topic.
 
  • #10
Mute said:
It looks like you made a mistake applying the chain rule.
$$\frac{d}{dk} \exp(f(k)) = \exp(f(k)) \frac{df(k)}{dk};$$
with ##f(k) = a(\exp(ik)-1)## you get ##df/dk = ai \exp(ik)##. The -1 drops out, so the moment doesn't vanish, and you get ##\langle n_1 \rangle = (1/i)(ai) = a##, as expected.
Thank you very, very much. I totally overlooked this even for the 10th times of rechecking.
I feel great now.
Thank you guys for all the help and insights.
 
  • #11
fluidistic said:
Thank you very, very much. I totally overlooked this even for the 10th times of rechecking.
I feel great now.
Thank you guys for all the help and insights.

Sometimes no matter how much you stare at something your brain just can't notice the mistake you made! It happens to the best of us!

Mute said:
Yes, in the spirit of being precise with terminology, there of course there is no pdf, but the probability mass function is still given by the inverse Fourier transform.

For example, inverse Fourier transforming the characteristic function of the Poisson distribution,

$$\int_{-\infty}^\infty \frac{dt}{2\pi} e^{\lambda(\exp(it)-1)} e^{-ikt},$$

let's change variables to the complex variable ##z = \exp(it)##. The integral can be written as the following contour integral in the complex plane, which we can then evaluate via the residue theorem, but we are forced to take ##k## to be an integer (as expected):

$$\frac{1}{2\pi i}\oint_\gamma \frac{dz}{z^{k+1}} e^{\lambda z} e^{-\lambda} = \frac{1}{k!}\frac{d^k}{dz^k}\left[e^{\lambda z}\right]_{z=0} e^{-\lambda} = \frac{\lambda^k}{k!}e^{-\lambda},$$
which is, of course, the probability mass function of the Poisson distribution. Here the contour ##\gamma## is a closed contour from ##-R## to ##R## along the real line, closed by a circular arc of radius ##R##, taking the ##R \rightarrow \infty## limit. The contribution from the arc will vanish in this limit, demonstrating equality between the contour integral and the original Fourier integral.

At any rate, the discussion of inverse transforms of characteristic functions to get pdf's (or pmf's) was tangential to the OP's original question, and we should probably not lead the discussion too far off topic.

I know I said we shouldn't focus on this too much, but I should be more careful than I was in that post, and correct some errors here:

For discrete probability distributions, the characteristic function, defined as the expectation of ##\exp(ikt)##, can be viewed as a Fourier series:

$$\varphi_x(t) = \mathbb{E}[e^{ikt}] = \sum_{k=-\infty}^\infty P(k)e^{ikt},$$
(where P(k) may be zero for some k).

This means that P(k) can be found via

$$P(k) = \int_{-\pi}^\pi \frac{dt}{2\pi} \varphi_X(t)e^{-ikt};$$
note the difference from my previous post. Here, the inverse is over a finite interval ##[-\pi,\pi)##, rather than the real line, due to the periodicity of the characteristic function in the discrete case (hence why it is a Fourier series of the probability mass function rather than a Fourier transform of the probability density function).

As such, the example I gave in my previous post should actually read

$$\int_{-\pi}^\pi \frac{dt}{2\pi}e^{\lambda(\exp(it)-1)} e^{-ikt} = \frac{1}{2\pi i}\oint_\gamma \frac{dz}{z^{k+1}} e^{\lambda z} e^{-\lambda} = \frac{1}{k!}\frac{d^k}{dz^k}\left[e^{\lambda z}\right]_{z=0} e^{-\lambda} = \frac{\lambda^k}{k!}e^{-\lambda},$$
where the contour ##\gamma## is the full circle ##|z| = 1##, not the infinite semi-circle in the incorrect derivation.
 

Similar threads

  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 14 ·
Replies
14
Views
2K
Replies
8
Views
3K
  • · Replies 12 ·
Replies
12
Views
3K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 1 ·
Replies
1
Views
8K
Replies
8
Views
3K