Understanding the transformation of skewness formula

Click For Summary

Discussion Overview

The discussion revolves around the formula for calculating sample skewness, specifically addressing the correction factors involved in the formula and their implications for small and large sample sizes. Participants explore the mathematical reasoning behind these corrections and their relevance in statistical analysis.

Discussion Character

  • Technical explanation, Conceptual clarification, Debate/contested

Main Points Raised

  • One participant presents the sample skewness formula and questions the meaning and effect of the correction term n/[(n-1)(n-2)] for small sample sizes.
  • Another participant explains that for large n, the term n/[(n-1)(n-2)] approaches 1/n, providing a mathematical justification for this behavior.
  • There is a suggestion that understanding the correction for sample variance is essential before tackling sample skewness, as the logic is similar but applied to a higher moment.
  • Another participant asserts that the correction accounts for using the sample mean instead of the true mean, linking it to the mathematical expectation of the expression.
  • One participant notes that while corrections are significant for small samples, they become less critical as sample size increases.

Areas of Agreement / Disagreement

Participants express varying levels of understanding regarding the corrections in the skewness formula, with some agreeing on the importance of these corrections for small samples while others emphasize their diminishing relevance for larger samples. The discussion does not reach a consensus on the necessity of these corrections in practical applications.

Contextual Notes

Some participants mention the asymptotic behavior of the correction terms, indicating that the mathematical steps involved may not be straightforward and depend on the context of sample size.

Who May Find This Useful

This discussion may be useful for students and practitioners in statistics or data analysis who are interested in understanding the nuances of sample skewness and the implications of sample size on statistical corrections.

Vital
Messages
108
Reaction score
4
Hello.

Here is the formula that computes the sample skewness:

Sk = [ (n / [(n-1)(n-2)] ] x [ ∑ (Xi - X)3 / s3 ] ,
where n is the number of elements in the sample,
Xi is the specific i-th value of the sample, where i starts from 0 and ends at i=n,
X is the arithmetic mean, s - standard deviation

I have two questions about this formula:

1) It is said in the book that the term n/[(n − 1)(n − 2)] in the above equation corrects for a downward bias in small samples. What does it mean and how that correction happens? For example, if n = 5, then n/[(n − 1)(n − 2)] = 0.4167.

I see it as if by using this part of equation we are taking around only 42 percent of the second part of the formula [ ∑ (Xi - X)3 / s3 ]. How does that help to correct for downward bias?

2) Also in the book it is said that as n becomes large, the expression reduces to the mean cubed deviation: Sk ≈ [ (1 / n ] x [ SUM (Xi - X)3 / s3 ]
How does this happen mathematically? I don't see it. For example, n = 1000, then

1000 / ( 999 x 998) how does this turn into 1/n?

Thank you very much.
 
Physics news on Phys.org
This post really needs Latex. Consider using:

https://www.physicsforums.com/help/latexhelp/
and also

https://www.codecogs.com/latex/eqneditor.php
- - - -
Let's work backward:

Question 2:

are you familiar with the fact that
##\lim_{n \to \infty} \frac{n}{n-2} = 1##

or as people say
## \frac{n}{n-2} \approx 1##

for large enough ##n##.
- - - -
now consider

##\frac{1}{n-1} \approx \frac{1}{n}##

for large enough ##n## (why?)

putting these together

## \frac{n}{(n-1)(n-2)} = \big(\frac{n}{(n-2)}\big)\big(\frac{1}{n-1}\big) \approx \big(1\big)\big( \frac{1}{n}\big)= \frac{1}{n}##

for large enough ##n## -- i.e. you can consider the asymptotic behavior separately. (If you prefer: take advantage of positivity and consider the limiting effects while working in logspace and then exponentiate your way back at the end.)
- - - -
Outside the scope thought: the rate of convergence here is actually pretty good. In your example consider ##\frac{1}{1000} \text{vs} \frac{1000}{(999)(998)}## they are actually pretty close to each other -- i.e. the first is ##0.001000## and the second one is ##0.001003## --where I rounded to the 6th decimal place.

There are certain results that are asymptotically correct but require exponentially more data to be a valid approximation -- these are things to be suspicious of -- but they don't apply here because convergence is pretty good.
- - - -
Question 1:

Are you familiar with the ##(n-1)## correction used in calculating sample variance? I think you should know this inside and out before considering sample skewness -- (a) because it is simpler and (b) because variance is much more important than skew -- in particular for the Central Limit Theorem, but also because estimates get less and less reliable the further up the moment curve that you go when you have noisy data.

The corrections for skew are the same logic, just one moment higher -- i.e. if you look at the moments involved, it's 3rd moment, 1st moment and 2nd moment. But the first and second moment are not 'pure' -- they are each sample moments and hence there's a data point / degree of freedom being chewed up -- the ## \frac{n}{(n-1)(n-2)}## corrects for that. I'm sure you can dig around and find the exact derivation for these skew corrections, but I don't think it's going to be that insightful. And more to the point: as noted in Question 2, asymptotically these corrections don't matter. Put differently: if you are dealing with small amounts of data, you need to pay attention to this stuff. But if you are dealing with medium or big data, it really doesn't matter.
 
Last edited:
The (n-1)(n-2) term corrects for the fact that the expression uses the sample mean, rather than the true mean. Another way of looking at is that the mathematical expectation of the expression using (n-1)(n-2) is the theoretical value for the given distribution.
 
StoneTemplePython said:
...

Thank you very much for this truly helpful post. I am sorry I didn't reply earlier. Now I understand the concept, and yet, I wasn't familiar with corrections.
Thank you once again.
 

Similar threads

  • · Replies 42 ·
2
Replies
42
Views
6K
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 5 ·
Replies
5
Views
3K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 15 ·
Replies
15
Views
2K
  • · Replies 24 ·
Replies
24
Views
6K