Dismiss Notice
Join Physics Forums Today!
The friendliest, high quality science and math community on the planet! Everyone who loves science is here!

Var(x) = E[ x^2] - (E[X])^

  1. May 22, 2005 #1

    Can someone please please show me why Var(x) = E[ x^2] - (E[X])^2.

    I just dont get it. THanks in advance. :smile:
  2. jcsd
  3. May 22, 2005 #2
  4. May 22, 2005 #3
    yeah i got that part, similar to distributive property, but where does variance come from? Like how did they get E[x^2]-(E[X])^2 ??? How did they get E[x^2] and (E[X])^2 ??
  5. May 22, 2005 #4


    User Avatar
    Staff Emeritus
    Science Advisor
    Gold Member




    This just follows from the properties of expectation values, which follow from the properties of integrals:

    [tex]<x>=\frac{\int xf(x)dx}{\int f(x)dx}=\int xP(x)dx[/tex]
  6. May 23, 2005 #5
    So <x> = E(x). The denominator there, [tex]\int f(x)dx[/tex], always equals 1 because f is a probability density function. The definition of an expected value is just the numerator of that fraction.
  7. May 23, 2005 #6


    User Avatar
    Staff Emeritus
    Science Advisor
    Gold Member

    I'm not defining f(x) to be the probability density, just some distribution function. P(x) is the probability density. Sorry for not making that clear. But yes, <x>=E(x).
  8. May 23, 2005 #7
    So then what is <x>? Are you just defining it here or does it mean something else?

    What's the difference between a distribution function and a density function? In my course they were used synonymously, except a distribution function could also be a probability mass function.
  9. May 23, 2005 #8


    User Avatar
    Staff Emeritus
    Science Advisor
    Gold Member

    You were right, it's the expectation value.

    I suppose I was using the wrong terminology. What my advisors sometimes call simply a "distribution function" is sometimes actually a "frequency distribution". This is what I meant by f(x). The idea is that the integral over its domain is not equal to one, but is instead equal to the number of objects in the sample (for example). The "distribution function" is actually something entirely different; that is, the cumulative probability of the value being less than x. Check Mathworld for the definitions of these things if you want more precision. If you don't want to bother (I wouldn't blame you), then disregard the middle part of my last equation (with f(x)) and just consider the part with P(x), the probability density.
  10. May 23, 2005 #9
    Yes, I know what a distribution function is.

    It's just a different notation. In the course I took, P(...) means the probability of the stuff in the parentheses (which would be a logical formula). So you might say P(X=x). Also, distribution functions were denoted by capital letters and density/mass functions were denoted by the corresponding lowercase letters, so even if P didn't mean "the probability of," it would be a distribution function, not a density function.
  11. Jun 14, 2005 #10

    For a random variable X, the variance is defined as: Var(X) = E[(X-E[X])^2].
    Thus, Var(X) = E[(X^2 - 2XE[X] - (E[X])^2]. Remember that the expected value of a constant, say a, is this constant: E(a) = a, where a is a constant. And also that: E[ E[X] ] = E[X].
    Then, we have that:
    Var(X) = E[X^2] - E[2XE[X]] - E[(E[X])^2] = E[X^2] - 2E[X]E[X] - (E[X])^2
    Var(X) = E[X^2] - 2(E[X])^2 - (E[X])^2 = E[X^2] - (E[X])^2

    Var(X) = E[X^2] - (E[X])^2, which is what you are trying to prove.
Share this great discussion with others via Reddit, Google+, Twitter, or Facebook