Is the Probability Axiom Valid for Mutually Exclusive Events?

  • Thread starter Thread starter Mathman23
  • Start date Start date
  • Tags Tags
    Axiom Probability
Click For Summary

Homework Help Overview

The discussion revolves around the validity of a probability axiom concerning mutually exclusive and independent events. The original poster, Fred, is trying to understand the conditions under which the sum of probabilities of independent events equals one and whether this can be generalized to mutually exclusive events.

Discussion Character

  • Conceptual clarification, Assumption checking

Approaches and Questions Raised

  • Fred attempts to prove that if the union of independent events equals one, then at least one of the events must have a probability of one. Other participants question the validity of this assertion, particularly in relation to the definitions of independence and mutual exclusivity.

Discussion Status

Participants are exploring the definitions and properties of independent and mutually exclusive events. Some have pointed out inconsistencies in Fred's reasoning, while others have clarified the conditions under which certain probability equations hold true. The discussion remains open with various interpretations being examined.

Contextual Notes

There is confusion regarding the terms "independent" and "mutually exclusive," which are being discussed in the context of probability axioms. Participants are also addressing the implications of these definitions on the original theorem proposed by Fred.

Mathman23
Messages
248
Reaction score
0
Hi

I have this here probability axiom which I'm not sure what I have understood correctly.

Let [tex]B_1 \ldots B_n[/tex] be independent events

Then [tex]P(B_1 \mathrm{U} \ldots \mathrm{U} \ B_n) = 1[/tex] which is the same as

[tex]P(B_1) + P(B_2) + \ldots + P(B_n) = 1[/tex]

I would like to show that this only is valid if [tex]1 \leq k \leq n[/tex] such that

P(B_k) = 1.

Proof:

If [tex]1 \leq k \leq n[/tex], then [tex]P(E) = 1[/tex](where E is the probability space).

Thereby it follows that [tex]P(B_1 U \ldots U \ B_n) = P(B_1) + P(B_2) + \ldots + P(B_n) = 1[/tex]

This can be written as the [tex]\sum_{n=1} ^{k} P(B_{n+1}) = P(B_{k}) =1[/tex]

Am I on the right track here?

Best Regards
Fred
 
Last edited:
Physics news on Phys.org
I'm very confused by what you've written.


So I will just state some facts:

[tex]P(B_1 U \ldots U \ B_n) = P(B_1) + P(B_2) + \ldots + P(B_n)[/tex]

is true when the [itex]B_i[/itex] are disjoint -- this equation is usually false when they are independent.


If E is the event consisting of all possible outcomes, then P(E) = 1.

In fact, P(A) cannot be greater than 1 for any event A.
 
Hi Herkyl and thank You for Your answer,

I have looked at it again and come to the conclusion that the proof should have said:

Let B_1 \ldots B_n be independent events. Show that

P(B_1 \mathrm{U} \ldots \mathrm{U} B_n)= 1, if and only if there exists a number 1 \leq k \leq n, such that P(B_k) = 1.


Proof:

If [tex]1 \leq k \leq n[/tex], then [tex]P(E) = 1[/tex](where E is the probability space).

Thereby it follows that [tex]P(B_1 U \ldots U \ B_n) = P(B_1) + P(B_2) + \ldots + P(B_n) = 1[/tex]
Am I on the right path here now?
 
Your theorem is false. "Heads" and "Tails" are independent events, neither P("Heads") nor P("Tails") is 1, but P("Heads" U "Tails") = 1
 
No, heads and tails are not independent. P(Heads and Tails) is certainly unequal to P(Heads)*P(Tails).
 
Hello can I change my original theorem to make it true?

If Yes, how?

Sincerely Fred

Hurkyl said:
No, heads and tails are not independent. P(Heads and Tails) is certainly unequal to P(Heads)*P(Tails).
 
Okay, I guess I confused "independent" with "mutually exclusive".

If you have two independent events B, B', then:

[tex]P(B\cup B')[/tex]

[tex]= P((B\cap B'^C) \sqcup (B\cap B') \sqcup (B' \cap B^C))[/tex]

[tex]= P(B\cap B'^C) + P(B\cap B') + P(B' \cap B^C)[/tex]

[tex]= P(B)P(B'^C) + P(B)P(B') + P(B')P(B^C)[/tex]

[tex]= P(B)(1 - P(B')) + P(B)P(B') + P(B')(1 - P(B))[/tex]

[tex]= 1 - [P(B) - 1][P(B') - 1][/tex]

If [itex]P(B\cup B') = 1[/itex], then [P(B)-1][P(B') - 1] = 0, so either P(B) = 1 or P(B') = 1.
 
Mathman23 said:
Hi

I have this here probability axiom which I'm not sure what I have understood correctly.

Let [tex]B_1 \ldots B_n[/tex] be independent events

Then [tex]P(B_1 \mathrm{U} \ldots \mathrm{U} \ B_n) = 1[/tex] which is the same as

[tex]P(B_1) + P(B_2) + \ldots + P(B_n) = 1[/tex]
If [itex]B_1, B_2, \ldots, B_n[/itex] are mutually exclusive then
[itex]P(B_1 and B_2 and... and B_n)= P(B_1)+ P(B_2)+ \ldots + P(B_n)[/itex]
follows from the definition of "mutually exclusive".
If, in addition, they exhaust all mutually exclusive events, then
[itex]P(B_1 and B_2 and... and B_n)= 1[/itex]

I would like to show that this only is valid if [tex]1 \leq k \leq n[/tex] such that

P(B_k) = 1.

Proof:

If [tex]1 \leq k \leq n[/tex], then [tex]P(E) = 1[/tex](where E is the probability space).

Thereby it follows that [tex]P(B_1 U \ldots U \ B_n) = P(B_1) + P(B_2) + \ldots + P(B_n) = 1[/tex]

This can be written as the [tex]\sum_{n=1} ^{k} P(B_{n+1}) = P(B_{k}) =1[/tex]
? How does that follow from the above? If P(Bk)= 1 then it follows that P(Bi)= 0 for any i not equal to k and so
[itex]\sum_{n=1} ^{k} P(B_{n+1}) = P(B_{k}) =1[/itex][/quote] follows trivially. But you are claiming the converse: if the sum of probabilities is 1 then the probability for each i except one is 1- and that is not, in general, true.

Am I on the right track here?

Best Regards
Fred
 

Similar threads

Replies
2
Views
1K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 11 ·
Replies
11
Views
3K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 2 ·
Replies
2
Views
3K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 37 ·
2
Replies
37
Views
8K
  • · Replies 23 ·
Replies
23
Views
2K