Statistics: E(X) = Integral(0 to infinity) of (1-F(x))dx

  • Context: Graduate 
  • Thread starter Thread starter kingwinner
  • Start date Start date
  • Tags Tags
    Infinity Statistics
Click For Summary

Discussion Overview

The discussion centers around the expression for the expected value of a non-negative random variable X, specifically the formula E(X) = Integral(0 to infinity) of (1-F(x))dx, where F(x) is the cumulative distribution function. Participants explore the applicability of this formula to both continuous and discrete random variables, as well as the mathematical justification and proofs related to this expression.

Discussion Character

  • Exploratory
  • Technical explanation
  • Debate/contested
  • Mathematical reasoning

Main Points Raised

  • Some participants question whether the formula E(X) = Integral(0 to infinity) of (1-F(x))dx is valid for both continuous and discrete random variables.
  • Others assert that the formula applies to discrete, continuous, and mixed random variables, but note that the derivative aspect is relevant only for continuous cases.
  • There is a discussion about the appropriateness of using integration for discrete random variables, with some suggesting that sums should be used instead.
  • One participant proposes using integration by parts to derive the expected value from the definition involving the density function.
  • Another participant raises concerns about the limit involved in the integration by parts approach, specifically the indeterminate form "infinity times 0" and the need for L'Hôpital's Rule to resolve it.
  • Some participants mention alternative methods to derive the expected value, including changing the order of integration in double integrals.
  • There is mention of the Riemann-Stieltjes integral as a way to connect sums and integrals, although some argue that measure theory is not necessary for this discussion.
  • One participant emphasizes that the formula holds true for any type of random variable and does not require advanced calculus concepts beyond freshman-level understanding.

Areas of Agreement / Disagreement

Participants express differing views on the applicability of the formula to discrete random variables, with some advocating for the use of sums instead of integrals. There is no consensus on the necessity of measure theory or the validity of certain mathematical steps, indicating ongoing debate and exploration of the topic.

Contextual Notes

Limitations include the lack of rigorous proofs provided for the claims made, as well as unresolved questions regarding the handling of limits and the definitions of expectation in different contexts.

kingwinner
Messages
1,266
Reaction score
0
"If X is non-negative, then E(X) = Integral(0 to infinity) of (1-F(x))dx, where F(x) is the cumulative distribution function of X."

============================

First of all, does X have to be a continuous random variable here? Or will the above result hold for both continuous and discrete random variable X?

Secondly, the source that states this result gives no proof of it. I searched the internet but was unable to find a proof of it. I know that by definition, since X is non-negative, we have E(X) = Integral(0 to infinity) of x f(x)dx where f(x) is the density function of X. What's next?

Thanks for any help!
 
Physics news on Phys.org
kingwinner said:
"If X is non-negative, then E(X) = Integral(0 to infinity) of (1-F(x))dx, where F(x) is the cumulative distribution function of X."
...
E(X) = Integral(0 to infinity) of x f(x)dx where f(x) is the density function of X. What's next?
Well, the thing you know has an x in it, but the thing you're trying to get to doesn't... and the thing you're trying to get to has an F in it, but the thing you know has the derivative of F in it...
 
This works for discrete, cont., and mixed. Though the derivative statement applies for cont. only. Use integration by parts for cont. case.
 
But for discrete random variable X, would it still make sense to talk about "integration"? (i.e.INTEGRAL(0 to infinity) of (1-F(x))dx) Or should it be replaced by a (sigma) sum?

Do you mean using integration by parts for the expression of the definition of E(X)? What should I let u and dv be?

Thanks!
 
No, in the discrete case you would be using sums instead of integrals since expectation is defined in terms of a sum not integrals. Well if u = 1 - F(x) then du = -f'(x) dx and dv = dx then v = x. So now you have x*S(x) (evalulated between your limits) + integral(x*f(x) dx). Obviously the first part of your sum vanishes since at "infinity" S(x) -> 0 and at 0, x*S(x) = 0. And so now you are left with what your usual definition of E(X).

Note: this is a very handwavy proof as you would really want to be rigorous when talking about the limits that make the first term vanish.
 
kingwinner said:
But for discrete random variable X, would it still make sense to talk about "integration"? (i.e.INTEGRAL(0 to infinity) of (1-F(x))dx) Or should it be replaced by a (sigma) sum?
It does when you learn measure theory. Until then, just replace it with a sum without thinking about it.
 
kingwinner said:
What should I let u and dv be?
Did you think about that question at all? I practically told you what u and dv should be in post #2...
 
NoMoreExams said:
No, in the discrete case you would be using sums instead of integrals since expectation is defined in terms of a sum not integrals. Well if u = 1 - F(x) then du = -f'(x) dx and dv = dx then v = x. So now you have x*S(x) (evalulated between your limits) + integral(x*f(x) dx). Obviously the first part of your sum vanishes since at "infinity" S(x) -> 0 and at 0, x*S(x) = 0. And so now you are left with what your usual definition of E(X).

Note: this is a very handwavy proof as you would really want to be rigorous when talking about the limits that make the first term vanish.
Just one point I am having troubles with: (in red)

lim x(1-F(x))
x->inf
This actually gives "infinity times 0" which is an indeterminate form and requires L'Hopital's Rule. I tried many different ways but was still unable to figure out what the limit is going to be...how can we prove that the limit is equal to 0?

Thanks!
 
Hurkyl, you don't "need" measure theory to write a sum as an integral. The Riemann-Stieljes integral will do that.
 
  • #10
Another way to do it is to write the expected value as

E[X]=\int_{0}^{\infty}sf(s)ds = \int_{s=0}^{\infty}\int_{x=0}^{s}f(s)dxds

and then change the order of the integrals to get your formula. To see what the new bounds on the integrals would be, draw a picture of the region of integration. You can use this same approach to find that

E[X^2] = \int_{0}^{\infty}s^2f(s)ds = \int_{s=0}^{\infty}\int_{x=0}^{s}2xf(s)dxds=<br /> \int_{0}^{\infty}2x(1-F(x))dx<br />

which is also valid for X nonnegative.
 
  • #11
NoMoreExams said:
No, in the discrete case you would be using sums instead of integrals since expectation is defined in terms of a sum not integrals.
Or, equivalently, write the probability distribution as a sum of delta functions.
 
  • #12
Integration by parts:

m_X=\int^{\infty}_{0} x f_X(x) dx = -\int^{\infty}_{0} x (-f_X(x) dx) eq(1)

Let u=x and dv = -f_X(x) dx

Thus du=dx and v = 1-F_X(x)

Chech that dv/dx = d/dx (1-F_X(x)) = d/dx(-F_X(x)) = -f_X(x) o.k.

Then substitute in (1)

m_X=-[uv|^{\infty}_{0}-\int^{\infty}_{0}vdu]

m_X=-[x[1-F_X(x)]|^{\infty}_{0}]+\int^{\infty}_{0}[1-F_X(x)]dx

The first term is zero at x = 0. As x\rightarrow\infty, 1-F_X(x) tends to zero faster than the increase of x and thus x[1-F_X(x)]\rightarrow0

Therefore

m_X=\int^{\infty}_{0}[1-F_X(x)]dx

QED

Enjoy!
 
  • #13
<br /> \begin{align*}<br /> E[X] &amp;= E\bigg[\int_0^X 1\,dx\bigg]\\<br /> &amp;= E\bigg[\int_0^\infty 1_{\{X&gt;x\}}\,dx\bigg]\\<br /> &amp;= \int_0^\infty E[1_{\{X&gt;x\}}]\,dx\\<br /> &amp;= \int_0^\infty P(X &gt; x)\,dx\\<br /> &amp;= \int_0^\infty (1 - F(x))\,dx<br /> \end{align*}<br />

By the way, this formula is true no matter what kind of random variable X is, and we do not need anything more than freshman calculus to understand the integral on the right-hand side. (We need neither measure theory nor Stieltjes integrals.) Even when X is discrete, the function 1 - F(x) is still at least piecewise continuous, so the integral makes perfectly good sense, even when understood as a good old-fashioned Riemann integral.
 
Last edited:

Similar threads

  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K