Conditional expectation on an indicator

Click For Summary
The discussion revolves around finding the conditional expectations \(\mathbb{E}[X \vert 1_{\{X+Y=0\}}]\) and \(\mathbb{E}[Y \vert 1_{\{X+Y=0\}}]\) for independent Bernoulli random variables \(X\) and \(Y\). The key point is demonstrating that \(\mathbb{E}[X+Y \vert 1_{\{X+Y=0\}}] = 0\) by showing that the conditional distribution has a mean of zero. It is established that if \(X+Y=0\), then \(\mathbb{P}[X=0 \vert X+Y=0] = 1\), leading to the conclusion that both \(X\) and \(Y\) must be zero. The discussion emphasizes the need for a rigorous proof rather than speculation about the result.
IniquiTrance
Messages
185
Reaction score
0

Homework Statement



Let X and Y be independent Bernoulli RV's with parameter p. Find,
\mathbb{E}[X\vert 1_{\{X+Y=0\}}] and \mathbb{E}[Y\vert 1_{\{X+Y=0\}}]

Homework Equations

The Attempt at a Solution



I'm trying to show that,

\mathbb{E}[X+Y\vert 1_{\{X+Y=0\}}] = 0

by,

<br /> \begin{align*}<br /> \mathbb{E}[X+Y\vert 1_{\{X+Y=0\}}] &amp;= \frac{\mathbb{E}[(X+Y)1_{\{X+Y=0\}}]}{\mathbb{P}[X+Y=0]} \\<br /> &amp;= \frac{0}{(1-p)^2} \\<br /> &amp;= 0<br /> \end{align*}<br />
 
Physics news on Phys.org
IniquiTrance said:
\mathbb{E}[X\vert 1_{\{X+Y=0\}}]
I'm not familiar with the notation for the condition (1_{\{X+Y=0\}}). How does one read it? Can you post a link?
 
haruspex said:
I'm not familiar with the notation for the condition (1_{\{X+Y=0\}}). How does one read it? Can you post a link?

It's the indicator of the event \{X+Y=0\}.

<br /> 1_{\{X+Y=0\}}=\begin{cases}1, \qquad \text{if }X+Y=0; \\ 0, \qquad \text{otherwise.}\end{cases}<br />
 
IniquiTrance said:

Homework Statement



Let X and Y be independent Bernoulli RV's with parameter p. Find,
\mathbb{E}[X\vert 1_{\{X+Y=0\}}] and \mathbb{E}[Y\vert 1_{\{X+Y=0\}}]

Homework Equations

The Attempt at a Solution



I'm trying to show that,

\mathbb{E}[X+Y\vert 1_{\{X+Y=0\}}] = 0

by,

<br /> \begin{align*}<br /> \mathbb{E}[X+Y\vert 1_{\{X+Y=0\}}] &amp;= \frac{\mathbb{E}[(X+Y)1_{\{X+Y=0\}}]}{\mathbb{P}[X+Y=0]} \\<br /> &amp;= \frac{0}{(1-p)^2} \\<br /> &amp;= 0<br /> \end{align*}<br />

Basically, you are trying to show that the conditional distribution ##f(k) \equiv P(X= k|X+Y=0), \; k=0,1,2, \ldots## has mean zero; that is, that ##\sum_{k=0}^{\infty} k f(k) = 0##. How do you suppose that could happen?
 
Ray Vickson said:
Basically, you are trying to show that the conditional distribution ##f(k) \equiv P(X= k|X+Y=0), \; k=0,1,2, \ldots## has mean zero; that is, that ##\sum_{k=0}^{\infty} k f(k) = 0##. How do you suppose that could happen?

I think it is possible since both ##X## and ##Y## are Bernoulli, if their sum is 0, then ##\mathbb{P}[X=0|X+Y=0]=1##. Then, ##\sum_{k=0}^1kf(k)=0*1##.
 
IniquiTrance said:
I think it is possible since both ##X## and ##Y## are Bernoulli, if their sum is 0, then ##\mathbb{P}[X=0|X+Y=0]=1##. Then, ##\sum_{k=0}^1kf(k)=0*1##.

Rather than "thinking it is possible" you need to actually prove it. But, at least you are on the right track---that is, your thinking is OK, and just needs to be firmed up. It is not difficult once you see what needs to be established.
 
Question: A clock's minute hand has length 4 and its hour hand has length 3. What is the distance between the tips at the moment when it is increasing most rapidly?(Putnam Exam Question) Answer: Making assumption that both the hands moves at constant angular velocities, the answer is ## \sqrt{7} .## But don't you think this assumption is somewhat doubtful and wrong?

Similar threads

Replies
1
Views
1K
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 6 ·
Replies
6
Views
2K
Replies
6
Views
1K
Replies
0
Views
860
  • · Replies 8 ·
Replies
8
Views
1K
  • · Replies 4 ·
Replies
4
Views
1K
  • · Replies 15 ·
Replies
15
Views
2K
  • · Replies 3 ·
Replies
3
Views
1K
Replies
3
Views
3K