# Random variables with no mean

Hello,

How do we interpret the fact that a random variable can have no mean? For example the Cauchy distribution, which arises from the ratio of two standard normal distributions.

I seek intuitive explanations or visualisations to understand math "facts" better.

Related Set Theory, Logic, Probability, Statistics News on Phys.org
Homework Helper
Hello,

How do we interpret the fact that a random variable can have no mean? For example the Cauchy distribution, which arises from the ratio of two standard normal distributions.

I seek intuitive explanations or visualisations to understand math "facts" better.
I'm not quite sure what you mean? The Cauchy distribution does not have a mean because its tails are "too heavy" - the ends of the distribution do not decrease quickly enough for the integral that defines the mean to converge.

Hello,

How do we interpret the fact that a random variable can have no mean? For example the Cauchy distribution, which arises from the ratio of two standard normal distributions.

I seek intuitive explanations or visualisations to understand math "facts" better.
Another example is the Pareto distribution for certain values of the tail parameter.

There's no requirement that a random variable has any finite moments. It just means that care needs to be taken when applying the usual theorems. For example, the mean of any finite sample is finite but since the law of large numbers no longer applies the mean does not converge to a finite value as the sample size increases. In fact, the average of 1000 Cauchy variables has the same distribution as a single Cauchy variable!

Homework Helper
"There's no requirement that a random variable has any finite moments. It just means that care needs to be taken when applying the usual theorems. For example, the mean of any finite sample is finite but since the law of large numbers no longer applies the mean does not converge to a finite value as the sample size increases. In fact, the average of 1000 Cauchy variables has the same distribution as a single Cauchy variable!"

There is a slight risk that, if this is read too quickly, a confusion of terms and ideas will result.

1) If you take a sample of any size from a Cauchy distribution, you can calculate $\bar x$, so a sample mean is always defined

2) The convergence issue means this: since a Cauchy distribution does not have a mean it does not have any higher-order moments: in particular, no variance. The lack of a population means that

a) The common statement that $\bar X \to \mu$ in probability does not apply

b) The CLT result, that the asymptotic distribution of the sample mean is a particular normal distribution, does not apply

c) It can be shown that, mathematically, if $X_1, X_n, \dots, X_n$ are i.i.d Cauchy, the distribution of $\bar X$ is still Cauchy - not , as we are used to seeing, approximately normal

bpet's initial comment that "there is no requirement that a random variable has any finite moments; it just means that care needs to be taken when applying the usual theorems" is spot on.

As another odd example, consider a $t$ distribution with $n$ degrees of freedom. For any particular $n > 2$ the distribution has moments of order $k < n$ but no others.

I am not too familiar with convergence and limits theoretically to understand how "the ends of the distribution do not decrease quickly enough for the integral that defines the mean to converge" though.

Is it equivalent to "The integral that defines the mean cannot be computed" or "The integral that defines the mean would result in an arbitarily large/small number"?

Homework Helper
Perhaps simply remembering that while probability does a good job of describing the world around us, it has mathematics as its underpinning and there are some instances where our intuition will cause us to expect things that don't occur. Happens to everyone.
I am not too familiar with convergence and limits theoretically to understand how "the ends of the distribution do not decrease quickly enough for the integral that defines the mean to converge" though.

Is it equivalent to "The integral that defines the mean cannot be computed" or "The integral that defines the mean would result in an arbitarily large/small number"?
For continuous distributions the mean and variance are, by definition, these integrals.

$$\mu = \int_{-\infty}^\infty x f(x) \, dx, \quad \sigma^2 = \int_{-\infty}^\infty (x-\mu)^2 f(x) \, dx$$

For either of these to exist the corresponding integral must exist. If the integral doesn't converge, the integrand (either $x f(x)$ or $(x-\mu)^2 f(x)$)
doesn't go to zero fast enough as $x \to \pm \infty$. That's all I meant.