Why does conditional probability used in mean square error equal zero?

In summary, the conditional expectation of a function is the sum of the expectations of the functions corresponding to the different conditions that the function can take on. The notation E f(X,y) is ambiguous because it can either be referring to the expectation with respect to the distribution of X, or the distribution of Y conditional on X being equal to y0. It is not clear which distribution is being used.
  • #1
EdMel
13
0
Hi guys,

I am having trouble showing that [itex]\mathbb{E}\left[(Y-\mathbb{E}[Y|X])^{2}\right]=0[/itex].

I understand the proof of why E[Y|X] minimizes the mean square error, but I cannot understand why it is then equal to zero.

I tried multiplying out the square to get [itex]\mathbb{E}\left[Y^{2}\right]-2\mathbb{E}\left[Y\mathbb{E}[Y|X]\right]+\mathbb{E}\left[\mathbb{E}[Y|X]\mathbb{E}[Y|X]\right][/itex]
but have not been able to justify [itex]\mathbb{E}\left[Y\mathbb{E}[Y|X]\right]=\mathbb{E}\left[Y^{2}\right]
[/itex] or [itex]\mathbb{E}\left[\mathbb{E}[Y|X]\mathbb{E}[Y|X]\right]=\mathbb{E}\left[Y^{2}\right][/itex].

Thanks in advance.
 
Physics news on Phys.org
  • #2
Can you tell us your definition of the conditional expectation and what properties you are allowed to use?
 
  • #3
It would also be helpful to improve the notation. If [itex] X [/itex] and [itex] Y [/itex] are random variables and [itex] f(X,Y) [/itex] is a function of them then the notation [itex] E f(X,y) [/itex] is ambiguous. It is not clear whether the expectation is being computed with respect to the distribution of [itex] X [/itex] or the distribution of [itex] Y [/itex] - or perhaps with respect to the joint distribution for [itex] (X,Y) [/itex].

You can use a subscript to denote which distribution is used to compute the expectation. For example, if [itex] Y [/itex] is not a function of [itex] X [/itex] then [itex] E_X ( E_Y ( 3Y + 1) ) [/itex] is the expectation of with respect to the distribution of [itex] X [/itex] of the constant value [itex] E_Y(3Y + 1) [/itex] Hence [itex] E_X (E_Y (3Y+1)) = E_Y (3Y+ 1) [/itex].
 
  • #4
It's hard to prove because it is not true unless (Y-E(Y|X))==0. Are you sure that (Y-E(Y|X)) is supposed to be squared?
 
  • #5
FactChecker said:
It's hard to prove because it is not true unless (Y-E(Y|X))==0.

And before (Y - E(Y|X)) is equal or not equal to zero, it would have to mean something. How do we interpret Y - E(Y|X) ? Is it a random variable? To realize it , do we realize a value Y = y0 from the distribution of Y and then take the expected value of the constant y0 with respect to the distribution of X ?
 

1. What is conditional probability and how is it used in mean square error?

Conditional probability is a measure of the likelihood of an event occurring given that another event has already occurred. In the context of mean square error, conditional probability is used to determine the probability that a certain outcome will occur given a set of conditions or variables. This helps in calculating the expected error in a prediction or estimation.

2. Why is conditional probability used in mean square error instead of other probability measures?

Mean square error is a commonly used measure of the accuracy of a prediction or estimation. It takes into account both the magnitude and direction of the errors, making it a more comprehensive measure than other probability measures. Using conditional probability in mean square error allows for a more precise estimation of the expected error.

3. Can you provide an example of how conditional probability is used in mean square error?

Suppose we want to predict the height of a child based on their parents' heights. We can use mean square error to measure the accuracy of our prediction. Conditional probability would be used to calculate the probability of a certain height given the parents' heights. This helps in estimating the expected error in our prediction.

4. How does conditional probability affect the value of mean square error?

The value of conditional probability affects the value of mean square error in that it helps in calculating the expected error. A higher conditional probability would result in a lower mean square error, indicating a more accurate prediction or estimation. Conversely, a lower conditional probability would result in a higher mean square error, indicating a less accurate prediction or estimation.

5. Are there any limitations to using conditional probability in mean square error?

While conditional probability is a useful tool in calculating mean square error, it does have some limitations. It assumes that the variables are independent, which may not always be the case in real-world scenarios. Additionally, it may not be suitable for all types of data and may not accurately capture the true error in certain situations.

Similar threads

  • Set Theory, Logic, Probability, Statistics
Replies
4
Views
846
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
740
  • Set Theory, Logic, Probability, Statistics
Replies
9
Views
2K
  • Linear and Abstract Algebra
Replies
19
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
6
Views
675
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
1K
  • Engineering and Comp Sci Homework Help
Replies
1
Views
942
  • Set Theory, Logic, Probability, Statistics
Replies
5
Views
1K
Back
Top