Kurtosis/4th central moment in terms of mean and variance

Click For Summary
SUMMARY

The discussion centers on expressing kurtosis (\kappa) and the 4th central moment (\mu_4) of a random variable X in terms of its mean (\mu) and variance (\sigma^2). The original poster, Chopin, explores the relationship between these moments and attempts to derive expressions applicable to any distribution, particularly focusing on a complex distribution resulting from the convolution of a generalized Pareto distribution and two uniform distributions. Key findings include the relation \mu_4(Y) = \mu_4(X) + \mu_4(C) + 6Var(X)Var(C), where C is the sum of the uniform distributions, and the realization that approximating \mu_4 with \sigma^4 is inaccurate for this case.

PREREQUISITES
  • Understanding of statistical moments, particularly the 4th central moment and kurtosis.
  • Familiarity with generalized Pareto distributions and their properties.
  • Knowledge of convolution of probability distributions.
  • Basic concepts of variance and its role in statistical analysis.
NEXT STEPS
  • Research the properties of generalized Pareto distributions and their types (Type I, II, III, IV).
  • Study the convolution theorem and its implications for moment generating functions.
  • Explore the delta method for approximating moments of complex distributions.
  • Investigate the relationship between moments and cumulants in statistical distributions.
USEFUL FOR

Statisticians, data scientists, and researchers working with complex probability distributions, particularly those involving kurtosis and higher-order moments.

fchopin
Messages
10
Reaction score
0
Hi All,

Is it possible to express the kurtosis \kappa, or the 4th central moment \mu_4, of a random variable X in terms of its mean \mu = E(X) and variance \sigma^2 = Var(X) only, without having to particularize to any distribution?

I mean, an expression like \kappa = f(\mu, \sigma^2) or \mu_4 = g(\mu, \sigma^2), valid for any distribution, where f(\mu, \sigma^2) and g(\mu, \sigma^2) are functions of the mean \mu and variance \sigma^2.

Thanks in advance!

Chopin

P.S.: Some comments on my attempts.

\kappa is related to \mu_4, and \mu_4 = E(X^4) - 4\mu E(X^3) + 6\mu^2 E(X^2) - 3\mu^4.

The term E(X^2) can be expressed as E(X^2) = \mu^2 + \sigma^2 but I didn't manage to find the way to express E(X^3) and E(X^4) in terms of \mu and \sigma^2.
 
Last edited:
Physics news on Phys.org
In general you can't express higher order moments in terms of mean and variance, although you can for a normal distribution.
 
  • Like
Likes   Reactions: 1 person
Thanks! Unfortunately the underlying distribution in my problem isn't Gaussian. But I think I could use approximations to the mean of a function of a random variable: E[g(X)] \approx g(\mu) + \frac{1}{2}\sigma^2g''(\mu)
 
fchopin said:
Thanks! Unfortunately the underlying distribution in my problem isn't Gaussian. But I think I could use approximations to the mean of a function of a random variable: E[g(X)] \approx g(\mu) + \frac{1}{2}\sigma^2g''(\mu)

Hey fchopin and welcome to the forums.

What distribution are you using? Is it a standard one?
 
No, the underlying distribution is much more complex. It is the result of the convolution of a generalized Pareto distribution with a uniform distribution, and with a uniform distribution again. The random variable is the sum of a generalized Pareto + uniform + uniform. I managed to obtain the expression of the pdf/cdf, but it is huge and computing \mu_4 by integration would definitely be a nightmare... I'm trying to find approximations based on the delta method. So far I have obtained \mu_4 \approx \sigma^4, but need to check its accuracy for my particular problem... Thank you all for you help and interest.
 
fchopin said:
No, the underlying distribution is much more complex. It is the result of the convolution of a generalized Pareto distribution with a uniform distribution, and with a uniform distribution again. The random variable is the sum of a generalized Pareto + uniform + uniform. I managed to obtain the expression of the pdf/cdf, but it is huge and computing \mu_4 by integration would definitely be a nightmare... I'm trying to find approximations based on the delta method. So far I have obtained \mu_4 \approx \sigma^4, but need to check its accuracy for my particular problem... Thank you all for you help and interest.

Can you calculate the Moment Generating Function either analytically as an exact expression or even as a good enough approximation?
 
fchopin said:
No, the underlying distribution is much more complex. It is the result of the convolution of a generalized Pareto distribution with a uniform distribution, and with a uniform distribution again. The random variable is the sum of a generalized Pareto + uniform + uniform.

The PDF of a sum of several random variables is a convolution only if the variables are independent.

On the other hand, by the convolution theorem, the moment generating function is a product of the corresponding moment generating functions. Taking the logarithm, the cumulant generating function is a sum of the corresponding cumulant generating function.

However, the 4th cumulant is not identical to the 4th central moment.
 
I'm afraid any attemp to obtain anything from the pdf/cdf analytically is a real nightmare as the resulting PDF is composed of four subequations, really huge... I spent one week computing these convolutions (yes, the involved variables are independent).

Basically, my random variable Y is Y = X + A + B, where X is generalized Pareto, and A,B are uniform (independent of each other the 3 random variables). I was interested in the 4th central moment of Y, \mu_4(Y), but I have finally found an alternative solution to my problem by expressing \mu_4(Y) as a function of \mu_4(X), \mu_4(C) and the variance of X and C, where C=A+B is a triangular distribution. This is enough for my problem.

By the way, the approximation \mu_4\approx\sigma^4 I was trying to use is really bad in this case since in practice \mu_4\gg\sigma^4.

Thank you all for your comments.
 
P.S.: The relation is

\mu_4(Y) = \mu_4(X) + \mu_4(C) + 6Var(X)Var(C)
 
  • #10
fchopin said:
P.S.: The relation is

\mu_4(Y) = \mu_4(X) + \mu_4(C) + 6Var(X)Var(C)

This is only true for statistically independent normally distributed random variables X and C.
 
  • #11
Dickfore said:
This is only true for statistically independent normally distributed random variables X and C.

Hi Dick, I'm both learning and refreshing my memory on many of these topics...

and I'm not certain what you mean; Don't normally distributed variables have an excess kurtosis of 0? / or kurtosis of 3?, so can his formula produce anything "but" 3, when applied to true normal distributions?

And second, I'm quite interested in the convolution approach for addition...
You mentioned cummulants, where the OP is asking about (central?) moments.
I thought that there is a bijection between cummulants and moments for a given distribution.

If that relationship holds for sums of distributions, such that the sum of the logs may be used for cummulants -- then it would seem to me that the OP might be able to generate a series of empirical random generator test cases for his distributions; and curve fit the bijection between moments and cummulants in the final distribution... ? Then he would only need to do interpolation on the curve fit as a very good approximation to the statistics he wishes to extract.

Thanks for your comments. :smile:

Oh, and if fchopin would comment, which generalized Pareto are you concerned with?

Vilfredo_Pareto.jpg


Type I, Type II, Lomax, Type III, or Type IV ?
http://en.wikipedia.org/wiki/Pareto_distribution
 
  • #12
4th moment and Kurtosis in convolution

With some hesitation (with mathematics nor physics being my core business) I would like to comment on the previous discussion:

The formula given by fchopin for the 4th moment of the convolution of two functions (y = x*h)

y4= x4 + h4 + 6 x2 h2

is valid for all regular functions, not only for a normal distribution.
While the first, second and third moments are nicely additive in convolution, the fourth (and higher?) is apparently not. I have no simple explanation for this.
With the Kurtosis defined as x4 / x2^2, a Guassian has K = 3.
You will find that the K-value of the convolution of 2 Gaussians, calculated with fchopins formula, is again 3, as it should be.
 
  • Like
Likes   Reactions: 1 person

Similar threads

Replies
5
Views
5K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 3 ·
Replies
3
Views
1K
  • · Replies 1 ·
Replies
1
Views
5K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 4 ·
Replies
4
Views
5K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
Replies
6
Views
2K
  • · Replies 9 ·
Replies
9
Views
2K