MHB Proof related to the expected value of cardinality

baiyang11
Messages
8
Reaction score
0
Consider N random variables X_{n} each following a Bernoulli distribution B(r_{n}) with 1 \geq r_{1} \geq r_{2} \geq ... \geq r_{N} \geq 0. If we make following assumptions of sets A and B:

(1) A \subset I and B \subset I with I=\{1,2,3,...,N\}

(2) |A \cap I_{1}| \geq |B \cap I_{1}| with I_{1}=\{1,2,3,...,n\}, n<N

(3) |A|=|B|=n

Do we have \mathbb{E}(\Sigma_{a\in A} X_{a}) \geq\mathbb{E}(\Sigma_{b\in B} X_{b})?

To avoid confusion, \mathbb{E} means expected value.Thanks!
 
Physics news on Phys.org
This question looks really similar to your previous one so I hope I can be of more help this time. I'll give my thoughts about this question.

We have given $N$ random variables with a Bernoulli distribution. That means $\mathbb{P}[X_n = 1] = r_n$ and $\mathbb{P}[X_n = 0]=1-r_n$. Furthermore, $\mathbb{E}[X_i]=r_i$ for $i=1,\ldots,N$.

By condition $(2) |A \cap I_1| \geq |B \cap I_1| \neq 0 $ we know that $A$ has more elements in common with $I_1$. Since the random variables $X_1,\ldots,X_n$ have greater expected values then the random variables $X_j$ with $j \in I \backslash I_1$, the statement
$$\mathbb{E}\left(\sum_{a \in A} X_a \right) \geq \mathbb{E}\left(\sum_{b \in B} X_b \right)$$
probably holds. Of course this is not a formal proof but it should give some intuition.

The only thing we know about $n$ and $N$ is $n<N$ so it could be the case that $N>2n$. In this case we can suppose $A \cap I_1 = \emptyset$ and $B \cap I_1 = \emptyset$. Now it is possible to choose $A$ en $B$ such that the statement does not hold anymore.

To make my point more clear. Suppose $I = \{1,2,\ldots,7\}$ and $I_1 = \{1,2,3\}$, thus $n=3$ and $N = 7$. Set $A = \{5,6,7\}$ and $B=\{4,5,6\}$. Then we have
$$\mathbb{E}\left(\sum_{a \in A} X_a \right) = \mathbb{E}[X_5]+\mathbb{E}[X_6]+\mathbb{E}[X_7] = r_5+r_6+r_7$$
$$\mathbb{E}\left(\sum_{b \in B} X_b \right) = r_4+r_5+r_6$$.

Since we can say $r_4 > r_7$, we have in this case that the statement does not hold.

I hope I made my point clear though I did not give a formal explanation. But I wanted to give my thoughts because in my opinion this is an interesting question.
 
Siron said:
This question looks really similar to your previous one so I hope I can be of more help this time. I'll give my thoughts about this question.

We have given $N$ random variables with a Bernoulli distribution. That means $\mathbb{P}[X_n = 1] = r_n$ and $\mathbb{P}[X_n = 0]=1-r_n$. Furthermore, $\mathbb{E}[X_i]=r_i$ for $i=1,\ldots,N$.

By condition $(2) |A \cap I_1| \geq |B \cap I_1| \neq 0 $ we know that $A$ has more elements in common with $I_1$. Since the random variables $X_1,\ldots,X_n$ have greater expected values then the random variables $X_j$ with $j \in I \backslash I_1$, the statement
$$\mathbb{E}\left(\sum_{a \in A} X_a \right) \geq \mathbb{E}\left(\sum_{b \in B} X_b \right)$$
probably holds. Of course this is not a formal proof but it should give some intuition.

The only thing we know about $n$ and $N$ is $n<N$ so it could be the case that $N>2n$. In this case we can suppose $A \cap I_1 = \emptyset$ and $B \cap I_1 = \emptyset$. Now it is possible to choose $A$ en $B$ such that the statement does not hold anymore.

To make my point more clear. Suppose $I = \{1,2,\ldots,7\}$ and $I_1 = \{1,2,3\}$, thus $n=3$ and $N = 7$. Set $A = \{5,6,7\}$ and $B=\{4,5,6\}$. Then we have
$$\mathbb{E}\left(\sum_{a \in A} X_a \right) = \mathbb{E}[X_5]+\mathbb{E}[X_6]+\mathbb{E}[X_7] = r_5+r_6+r_7$$
$$\mathbb{E}\left(\sum_{b \in B} X_b \right) = r_4+r_5+r_6$$.

Since we can say $r_4 > r_7$, we have in this case that the statement does not hold.

I hope I made my point clear though I did not give a formal explanation. But I wanted to give my thoughts because in my opinion this is an interesting question.
Thank you for reply.
You provide a clear example when $A$ and $B$ have nothing in common with $I_{1}$ at all. But I think my original intention is to make $|A \cap I_{1}| \geq |B \cap I_{1}|>0$. I admit it is my mistake not to make this clear.
So if $|A \cap I_{1}| \geq |B \cap I_{1}|>0$, is the answer "YES" or "NO"?

It may be still No. Someone else gives me the following example.
Let $n \geq 2$, $N=n+1, A=I-\{1\}, B=I-\{2\}$ and $r_{1}>r_{2}$. Then
$$\mathbb{E}\left(\sum_{a \in A} X_a \right)=\sum_{i} r_{i}-r_{1} $$
$$\mathbb{E}\left(\sum_{b \in B} X_b \right)=\sum_{i} r_{i}-r_{2} $$
With $r_{1}>r_{2}$, we get $\mathbb{E}\left(\sum_{a \in A} X_a \right)<\mathbb{E}\left(\sum_{b \in B} X_b \right)$ to make the answer "NO".

In fact, by asking my original question, I what to abstract the following statement which seems to be true intuitively.

If I have better knowledge of what incidents have the top probabilities to happen (like I know set A rather than set B), is the knowledge trustable (A is still favored) when those incidents realize (the expected value thing)?

Now, inspired by the counterexample above, it seems that we can always manipulate different $r_{i}$ to get $B$ favored with $|A|=|B|$.

Now that under the original assumptions $\mathbb{E}\left(\sum_{a \in A} X_a \right) \geq \mathbb{E}\left(\sum_{b \in B} X_b \right)$ does not always hold, I am trying to see under what assumptions, we can always favor A, i.e. $\mathbb{E}\left(\sum_{a \in A} X_a \right) \geq \mathbb{E}\left(\sum_{b \in B} X_b \right)$. But I am a bit lost.
 
baiyang11 said:
Thank you for reply.
You provide a clear example when $A$ and $B$ have nothing in common with $I_{1}$ at all. But I think my original intention is to make $|A \cap I_{1}| \geq |B \cap I_{1}|>0$. I admit it is my mistake not to make this clear.
So if $|A \cap I_{1}| \geq |B \cap I_{1}|>0$, is the answer "YES" or "NO"?

It may be still No. Someone else gives me the following example.
Let $n \geq 2$, $N=n+1, A=I-\{1\}, B=I-\{2\}$ and $r_{1}>r_{2}$. Then
$$\mathbb{E}\left(\sum_{a \in A} X_a \right)=\sum_{i} r_{i}-r_{1} $$
$$\mathbb{E}\left(\sum_{b \in B} X_b \right)=\sum_{i} r_{i}-r_{2} $$
With $r_{1}>r_{2}$, we get $\mathbb{E}\left(\sum_{a \in A} X_a \right)<\mathbb{E}\left(\sum_{b \in B} X_b \right)$ to make the answer "NO".

In fact, by asking my original question, I what to abstract the following statement which seems to be true intuitively.

If I have better knowledge of what incidents have the top probabilities to happen (like I know set A rather than set B), is the knowledge trustable (A is still favored) when those incidents realize (the expected value thing)?

Now, inspired by the counterexample above, it seems that we can always manipulate different $r_{i}$ to get $B$ favored with $|A|=|B|$.

Now that under the original assumptions $\mathbb{E}\left(\sum_{a \in A} X_a \right) \geq \mathbb{E}\left(\sum_{b \in B} X_b \right)$ does not always hold, I am trying to see under what assumptions, we can always favor A, i.e. $\mathbb{E}\left(\sum_{a \in A} X_a \right) \geq \mathbb{E}\left(\sum_{b \in B} X_b \right)$. But I am a bit lost.

Nice counterexample! I did not see it that way. Anyway your question is interesting but not something I can answer right away. I'll think about it but in the mean time I hope some other members can help you out.

Ps: where does this problem come from? I have had some probability theory courses but I've never discussed problems like this before.
 
Siron said:
Nice counterexample! I did not see it that way. Anyway your question is interesting but not something I can answer right away. I'll think about it but in the mean time I hope some other members can help you out.

Ps: where does this problem come from? I have had some probability theory courses but I've never discussed problems like this before.

I appreciate your continuous attention on this problem and I do hope attentions from some other members, maybe from various areas, because it seems to be interdisciplinary between set theory and probability theory.

This abstracted problem comes from a concrete problem in my major, which is electrical engineering (EE). We come up with a concrete method for a EE problem we are researching currently. Then I realize the essence of the concrete method is trying to answer the question, the italic one in my previous post. Then I looked at the EE problem and what we do in the concrete method, and come up with this mathematical problem, which I found interesting as well.
 
Hi all, I've been a roulette player for more than 10 years (although I took time off here and there) and it's only now that I'm trying to understand the physics of the game. Basically my strategy in roulette is to divide the wheel roughly into two halves (let's call them A and B). My theory is that in roulette there will invariably be variance. In other words, if A comes up 5 times in a row, B will be due to come up soon. However I have been proven wrong many times, and I have seen some...
Thread 'Detail of Diagonalization Lemma'
The following is more or less taken from page 6 of C. Smorynski's "Self-Reference and Modal Logic". (Springer, 1985) (I couldn't get raised brackets to indicate codification (Gödel numbering), so I use a box. The overline is assigning a name. The detail I would like clarification on is in the second step in the last line, where we have an m-overlined, and we substitute the expression for m. Are we saying that the name of a coded term is the same as the coded term? Thanks in advance.
Back
Top