Simple Expectation Value Question

Click For Summary
SUMMARY

The discussion centers on the calculation of expected values in probability theory, specifically the formula for the expected value of a function of a random variable. The key formula presented is = Ʃ f(xi) P(xi) = ∫ f(x) P(f(x)) dx, which generalizes the expected value calculation. A point of contention arises regarding the notation P(f(xi)) = P(xi), which some participants find confusing and potentially misleading. The "law of the unconscious statistician" is referenced as a foundational theorem that justifies this relationship, emphasizing the need for clarity in notation and understanding.

PREREQUISITES
  • Understanding of probability distributions and density functions
  • Familiarity with the concept of expected value in statistics
  • Knowledge of weighted averages and their application in statistics
  • Basic calculus, particularly integration techniques
NEXT STEPS
  • Study the "law of the unconscious statistician" in detail
  • Learn about the implications of different probability distributions on expected values
  • Explore advanced topics in statistical notation and its importance in mathematical rigor
  • Investigate practical applications of expected value in real-world scenarios
USEFUL FOR

Students of statistics, mathematicians, and anyone involved in data analysis or probability theory who seeks to deepen their understanding of expected values and their calculations.

Harrisonized
Messages
206
Reaction score
0
I was told that given a probability distribution p(x) dx, the expected value for x is given by:

<x> = Ʃ xi P(xi) = ∫ x P(x) dx

This part makes sense to me. It was justified to me through the use of weighted averages. However, my teacher then made a hand-wavy move to generalize the above formula. I quote:

This way of calculating the average can be easily generalized, since it depends neither on the numbers of different events nor on the total number of events, it only depends on the probabilities of all different possibilities. So we can consider an experiment where we are measuring some quantity x, and all the possible outcomes are x1, x2, ... , xn. If we denote the probability of the outcome xi to be P(xi) then we can write the average of x as

<x>= Ʃ xi P(xi) = ∫ x P(x) dx (17)

We may also be interested in calculating the average of some given function of x, call it f(x). The different possible values of f(x) are f(x1), f(x2), ... , f(xn), and the probability P(f(xi)) of the value f(xi) is, of course, the same as for x to have the value xi, i.e. P(xi)

P(f(xi)) = P(xi)

We can now use the rule (17) to find the average of f(x)

<f>= Ʃ f(xi) p(f(xi)) = ∫ f(x) P(f(x)) dx
<f>= Ʃ f(xi) p(xi) ∫ f(x) P(x) dx

-end of quote-

It's this last part I don't understand:

P(f(xi)) = P(xi)

I don't see how this can be true for anything other than f(xi)=xi or P(xi)=(a constant). Can someone please justify this to me?

-----

Edit: after extensive searching, I finally came across this:

http://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician

I've been searching for a long time now, and I still haven't found justification for why this works.
 
Last edited:
Physics news on Phys.org
Yeah that's really stupid way of putting it.

The expectation value of a function is just
<f> = Ʃ f(xi) P(xi) though,
right? That is, it's the average of the function weighted by the probability density of the individual events.

Now your teacher is using horrible notation P(f(x_i)) for the weight of each f(x_i) in the sum; in a rigorous sense it's complete nonsense, but his underlying point is correct.
 
Harrisonized said:
I was told that given a probability distribution p(x) dx, the expected value for x is given by:

<x> = Ʃ xi P(xi) = ∫ x P(x) dx

This part makes sense to me. It was justified to me through the use of weighted averages. However, my teacher then made a hand-wavy move to generalize the above formula. I quote:

This way of calculating the average can be easily generalized, since it depends neither on the numbers of different events nor on the total number of events, it only depends on the probabilities of all different possibilities. So we can consider an experiment where we are measuring some quantity x, and all the possible outcomes are x1, x2, ... , xn. If we denote the probability of the outcome xi to be P(xi) then we can write the average of x as

<x>= Ʃ xi P(xi) = ∫ x P(x) dx (17)

We may also be interested in calculating the average of some given function of x, call it f(x). The different possible values of f(x) are f(x1), f(x2), ... , f(xn), and the probability P(f(xi)) of the value f(xi) is, of course, the same as for x to have the value xi, i.e. P(xi)

P(f(xi)) = P(xi)

We can now use the rule (17) to find the average of f(x)

<f>= Ʃ f(xi) p(f(xi)) = ∫ f(x) P(f(x)) dx
<f>= Ʃ f(xi) p(xi) ∫ f(x) P(x) dx

-end of quote-

It's this last part I don't understand:

P(f(xi)) = P(xi)

I don't see how this can be true for anything other than f(xi)=xi or P(xi)=(a constant). Can someone please justify this to me?

-----

Edit: after extensive searching, I finally came across this:

http://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician

I've been searching for a long time now, and I still haven't found justification for why this works.

As you have already realized, the result [itex]E f(X) = \sum P(X=x) f(x)[/itex] or [itex]E f(X) = \int f(x) dP(X=x)[/itex] is true, and is know as the "law of the unconscious statistician", because it was reputed to be used indiscriminately by statisticians without knowing that it is a *theorem* rather than a definition. It can be found in Section 7 of
http://www.math.harvard.edu/~lauren/154/Outline8-DRV-Expect2009.pdf .
There the discussion is pretty elementary but also quite detailed. After reading it you will understand why your teacher's explanation is only partly correct. This link is much more informative than the one you gave above.

RGV
 
Last edited by a moderator:

Similar threads

  • · Replies 16 ·
Replies
16
Views
2K
Replies
4
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 10 ·
Replies
10
Views
3K
Replies
1
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K