Is the formula for conditional expectation valid for multiple random variables?

Click For Summary
SUMMARY

The formula for conditional expectation, E[Xg(Y)|Y] = g(Y)E[X|Y], is valid for discrete random variables X and Y, where g(Y) is a function of Y. The proof relies on the definition of conditional expectation, demonstrating that E[Xg(Y)|Y] simplifies to g(Y)E[X|Y]. However, this formula does not extend to three variables in the form E[Xg(Z)|Y] without additional conditions, as the independence of X and Z must be established for the equation to hold true. If Z is a function of Y, such as Z = h(Y), the formula can be applied correctly.

PREREQUISITES
  • Understanding of discrete random variables
  • Familiarity with the concept of conditional expectation
  • Knowledge of independence in probability theory
  • Basic understanding of functions of random variables
NEXT STEPS
  • Study the properties of conditional expectation in probability theory
  • Learn about the independence of random variables and its implications
  • Explore the concept of functions of random variables in depth
  • Investigate conditional distributions and their applications
USEFUL FOR

Statisticians, data scientists, and anyone involved in probability theory or statistical modeling will benefit from this discussion, particularly those working with conditional expectations and random variables.

jimmy1
Messages
60
Reaction score
0
[SOLVED] Conditional Expectation

I'm trying to understand the following proof I saw in a book. It says that:
E[Xg(Y)|Y] = g(Y)E[X|Y] where X and Y are discrete random variables and g(Y) is a function of the random variable Y.

Now they give the following proof:

E[Xg(Y)|Y] = \sum_{x}x g(Y) f_{x|y}(x|y)
= g(Y)\sum_{x}x f_{x|y}(x|y)
= g(Y)E[X|Y]

Now, the proof is very simple as they are just using the definition of conditional expectation (ie. E[X|Y]= \sum_{x}x f_{x|y}(x|y)).

But, would this formula also work for say 3 variables? That is, E[Xg(Z)|Y] = g(Z)E[X|Y], where Z is another discrete random variable.

It probably isn't right, but from the above proof I can't immediatley see what's wrong with it, as I'll be just switching the g(Y) with a g(Z), and as the summation in the proof is over x, I can take g(Z) out of the summation and similarly get the result E[Xg(Z)|Y] = g(Z)E[X|Y] ??
 
Last edited:
Physics news on Phys.org
No, the proof works because the event {Xg(Y)|Y=y} = {Xg(y)|y} = g(y){X|y}; but the event {Xg(Z)|y} cannot be written similarly, in general.

In general, V = Xg(Z) is a new random variable and has its own distribution and conditional distribution; so the sum is over v.

If X and Z are independent then E[Xg(Z)|y] = E[X|y]E[g(Z)|y].

If Z = h(Y) then E[Xg(h(y))|y] = g(h(y))E[X|y].
 
Ah yes, that clears things up. Cheers, you've been great help once again!
 

Similar threads

  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 2 ·
Replies
2
Views
1K
  • · Replies 3 ·
Replies
3
Views
3K