Interesting lognormal kurtosis discrepancy

In summary, lognormal distribution is a probability distribution where the logarithm of a random variable follows a normal distribution, while kurtosis is a measure of the degree of peakedness or flatness of a distribution's curve. The lognormal kurtosis discrepancy is calculated by taking the difference between the kurtosis of a lognormal distribution and the kurtosis of a normal distribution with the same mean and standard deviation. This discrepancy is interesting because it shows that the lognormal distribution has a different shape than the normal distribution, which has implications for data analysis and modeling in various fields. The asymmetry of the lognormal distribution is what causes the lognormal kurtosis discrepancy, and it is commonly used in research and analysis to better understand
  • #1
bpet
532
7
A colleague of mine noticed that samples from a lognormal distribution seem to have much smaller kurtosis than the theoretical value - for example with [tex]\sigma=2[/tex] the theoretical value is about 9,000,000 whereas for samples of size [tex]N=1000[/tex] we found that the sample kurtosis averages around 280 and is never larger than 1000. Even with [tex]N=10^6[/tex] the sample kurtosis is only around 50,000.

We used the standard formula
[tex]K=\frac{E[(X-E[X])^4]}{E[(X-E[X])^2]^2}=e^{4\sigma^2}+2e^{3\sigma^2}+3e^{2\sigma^2}-3 \approx e^{4\sigma^2}[/tex]
(for excess kurtosis subtract 3 again).

Several things could be contributing here though I think it's to do with slow convergence to the central limit; factors such as machine precision, random number generation method and use of the unbiased estimator don't seem to make a difference in the results. Anyone come across this before or know how to derive the convergence rate?
 
Physics news on Phys.org
  • #2


Thank you for bringing this up, colleague. The issue you have observed is actually quite common and can be explained by the slow convergence rate of the sample kurtosis to the theoretical value. This is due to the fact that the kurtosis is a higher order moment and is more sensitive to extreme values in the data.

To understand the convergence rate, we can look at the central limit theorem. This theorem states that as the sample size (N) increases, the sampling distribution of the sample mean approaches a normal distribution with mean equal to the population mean and standard deviation equal to the population standard deviation divided by the square root of the sample size. However, this theorem does not hold for higher order moments such as the kurtosis.

In fact, the convergence rate of the sample kurtosis is much slower than that of the sample mean. It has been shown that the convergence rate of the sample kurtosis is on the order of 1/\sqrt{N}, which means that as the sample size increases, the sample kurtosis will approach the theoretical value at a much slower rate compared to the sample mean.

This slow convergence rate is also affected by the shape of the underlying distribution. In the case of a lognormal distribution, the kurtosis is already high (around 9,000,000 for \sigma=2), so it will take a much larger sample size to see the sample kurtosis approach this value. This is why even with a sample size of 10^6, the sample kurtosis is only around 50,000.

In summary, the slow convergence rate of the sample kurtosis is a well-known phenomenon and is due to the higher order nature of this moment and the shape of the underlying distribution. It is important to keep this in mind when interpreting kurtosis values from a sample.
 

1. What is lognormal distribution and kurtosis?

Lognormal distribution is a probability distribution where the logarithm of a random variable follows a normal distribution. Kurtosis is a measure of the degree of peakedness or flatness of a distribution's curve.

2. How is lognormal kurtosis discrepancy calculated?

The lognormal kurtosis discrepancy is calculated by taking the difference between the kurtosis of a lognormal distribution and the kurtosis of a normal distribution with the same mean and standard deviation.

3. Why is the lognormal kurtosis discrepancy interesting?

The lognormal kurtosis discrepancy is interesting because it shows that the lognormal distribution has a different shape than the normal distribution, despite having the same mean and standard deviation. This has implications for data analysis and modeling in various fields, including finance and biology.

4. What causes the lognormal kurtosis discrepancy?

The lognormal kurtosis discrepancy is caused by the asymmetry of the lognormal distribution. Unlike the symmetric normal distribution, the lognormal distribution has a longer tail on one side, leading to a higher kurtosis value.

5. How is the lognormal kurtosis discrepancy used in research and analysis?

The lognormal kurtosis discrepancy is commonly used in research and analysis to better understand and model data with a lognormal distribution. It helps to identify when a lognormal distribution is a better fit for the data compared to a normal distribution. It is also used in financial risk management and in studying biological phenomena such as species abundance and body size distributions.

Similar threads

  • Set Theory, Logic, Probability, Statistics
Replies
8
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
841
  • Set Theory, Logic, Probability, Statistics
Replies
11
Views
10K
  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
3
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
9
Views
2K
  • MATLAB, Maple, Mathematica, LaTeX
Replies
2
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
24
Views
5K
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
8K
  • MATLAB, Maple, Mathematica, LaTeX
Replies
2
Views
736
Back
Top