Convergence of Mean in Probability - How to Prove it?

  • Thread starter Thread starter gajohnson
  • Start date Start date
  • Tags Tags
    Convergence Mean
Click For Summary

Homework Help Overview

The discussion revolves around the convergence of the sample mean of a sequence of independent random variables to a specified limit in probability. The original poster presents a problem involving independent random variables with known expected values and queries how to demonstrate that the sample mean converges to the overall mean.

Discussion Character

  • Exploratory, Assumption checking, Conceptual clarification

Approaches and Questions Raised

  • Participants explore whether the problem is as straightforward as it appears, questioning if the convergence can be established simply through the equality of means. There are inquiries about the variances of the random variables and their implications for applying Chebyshev's inequality. Some participants suggest that the proof may be trivial if certain conditions about variance are met, while others note that the proof could be more complex without those conditions.

Discussion Status

The discussion is active, with participants sharing insights and raising questions about the assumptions related to variance and the applicability of Chebyshev's inequality. There is a recognition that different approaches may be necessary depending on the conditions provided in the problem. Some participants express uncertainty about the implications of carrying over variance information from previous problems.

Contextual Notes

There is mention of the Law of Large Numbers (WLLN) and its relevance to the problem at hand. Participants note that the proof may vary in complexity depending on whether the variance is known or finite. The original poster's problem includes specific conditions regarding expected values and variances, which are under discussion.

gajohnson
Messages
70
Reaction score
0

Homework Statement



Let [itex]X_1, X_2...[/itex] be a sequence of independent random variables with [itex]E(X_i)=\mu_i[/itex] and [itex](1/n)\sum(\mu_i)\rightarrow\mu[/itex]

Show that [itex]\overline{X}\rightarrow\mu[/itex] in probability.


Homework Equations



NA

The Attempt at a Solution



I feel as if this shouldn't be too hard, but I'm having some trouble getting started. Any point in the right direction would be helpful. Thanks!
 
Physics news on Phys.org
Or rather, the better question might be. Is this as easy as it looks? Namely, is it as simple as showing that, because of the equality given, X-bar = mu and so they must converge?
 
Is any information given about the variances of ##X_i##? Are they known to be finite, for example?
 
jbunniii said:
Is any information given about the variances of ##X_i##? Are they known to be finite, for example?

Well, this was a little confusing. The previous question also gives that
[itex]Var(X_i)=\sigma^2[/itex].

However, it also says that the sum of the variances converges to 0, which clearly means that [itex]\overline{X}\rightarrow\mu[/itex] as well (and this is the key to the first problem, which asks me to show the same thing, i.e. that [itex]\overline{X}\rightarrow\mu[/itex]).

The second question says simply to "Let [itex]X_i[/itex] be as in Problem 1" and then adds the additional information I included in the OP.

So I guess I assume that [itex]Var(X_i)=\sigma^2[/itex] still applies, but not the bit about the sum of the variances converging to 0 (because this would make the second problem as trivial as the first), and that instead we're supposed to use the new fact about the expected values converging to [itex]\mu[/itex].

What do you think?
 
Last edited:
So they actually ask you to prove the WLLN in a problem??
 
micromass said:
So they actually ask you to prove the WLLN in a problem??

I suppose so, and this is what I found confusing. Because it seems quite trivial and I'm wondering if I am missing something here.

I apply Chebyshev's inequality with the given information and I'm done, right?
 
gajohnson said:
I suppose so, and this is what I found confusing. Because it seems quite trivial and I'm wondering if I am missing something here.

I apply Chebyshev's inequality with the given information and I'm done, right?

Chebyshev requires some knowledge of the variance though. So you got to know it's finite. In that case, the proof is indeed trivial.

However, some versions of the WLLN also holds without any knowledge of the variance. For example, I know that if [itex]X_n[/itex] are iid and if the first moment is finite, then you can also prove that the means converge in probability.
Needless to say that the proof is going to be more difficult, since you can't apply Chebyshev.
 
micromass said:
Chebyshev requires some knowledge of the variance though. So you got to know it's finite. In that case, the proof is indeed trivial.

However, some versions of the WLLN also holds without any knowledge of the variance. For example, I know that if [itex]X_n[/itex] are iid and if the first moment is finite, then you can also prove that the means converge in probability.
Needless to say that the proof is going to be more difficult, since you can't apply Chebyshev.

So carrying over that [itex]Var(X_i)=\sigma^2[/itex] is of no use here? From that can't I just use Chebyshev?

I believe that this fact is supposed to carry over from the previous problem.
 
gajohnson said:
So carrying over that [itex]Var(X_i)=\sigma^2[/itex] is of no use here? From that can't I just use Chebyshev?

I believe that this fact is supposed to carry over from the previous problem.

If you can indeed carry over [itex]Var(X_i)=\sigma^2[/itex], then you can use Chebyshev and your proof is valid.
Whether you can carry over [itex]Var(X_i)=\sigma^2[/itex] is something I don't know, you should ask your professor. But I think that you probably can.
 
  • #10
micromass said:
If you can indeed carry over [itex]Var(X_i)=\sigma^2[/itex], then you can use Chebyshev and your proof is valid.
Whether you can carry over [itex]Var(X_i)=\sigma^2[/itex] is something I don't know, you should ask your professor. But I think that you probably can.

I'll go with that. I envision the proof being quite challenging if this is not possible (and probably beyond the scope of the class). Thanks for your the help, everyone.
 
  • #11
micromass said:
However, some versions of the WLLN also holds without any knowledge of the variance. For example, I know that if [itex]X_n[/itex] are iid and if the first moment is finite, then you can also prove that the means converge in probability.
Needless to say that the proof is going to be more difficult, since you can't apply Chebyshev.
And the proof using characteristic functions, which does not require finite variance, can't be applied directly because we aren't in an iid situation. We will still have ##\phi_{x+y} = \phi_{x}\phi_{y}##, but this won't equal ##\phi_x^2##.
 
  • #12
gajohnson said:
I'll go with that. I envision the proof being quite challenging if this is not possible (and probably beyond the scope of the class). Thanks for your the help, everyone.

The proof without the variance condition is not very difficult too understand. But it wouldn't be suitable as an exercise.
 
  • #13
micromass said:
The proof without the variance condition is not very difficult too understand. But it wouldn't be suitable as an exercise.

Got it. If you don't mind indulging me, what's the gist of it?
 
  • #14
gajohnson said:
Got it. If you don't mind indulging me, what's the gist of it?

The wonderful book W.Feller, "An Introduction to Probability Theory and its Applications", Vol. I, (Wiley) has a great chapter (Chapter X) on this and other limit properties. The way he does it is to look at a method of truncation:
[tex]U_k = X_k, \text{ and }V_k = 0 \text{ if } |X_k| \leq \delta n\\<br /> U_k = 0, \text{ and } V_k = X_k \text{ if } |X_k| > \delta n,[/tex]
where ##\delta > 0## is a parameter to be determined later, and ##k = 1,2,\ldots, n##. This implies ##X_k = U_k + V_k##. He then proves the Law of Large Numbers by showing that for any given ##\epsilon > 0## the constant ##\delta > 0## can be chosen so that as ##n \to \infty##,
[tex]P\{ |U_1 + \cdots + U_n| > \frac{1}{2} \epsilon n \} \to 0\\<br /> \text{ and }\\<br /> P\{ |V_1 + \cdots + V_n | > \frac{1}{2} \epsilon n \} \to 0.[/tex]
The Law of Large Numbers follows from these.

The point of the truncation is to permit some statements to be made and proved even when variance is infinite---it works as long as ##E X## is finite. For proof of the above crucial results, consult Feller.
 
  • #15
Ray Vickson said:
The wonderful book W.Feller, "An Introduction to Probability Theory and its Applications", Vol. I, (Wiley) has a great chapter (Chapter X) on this and other limit properties. The way he does it is to look at a method of truncation:
[tex]U_k = X_k, \text{ and }V_k = 0 \text{ if } |X_k| \leq \delta n\\<br /> U_k = 0, \text{ and } V_k = X_k \text{ if } |X_k| > \delta n,[/tex]
where ##\delta > 0## is a parameter to be determined later, and ##k = 1,2,\ldots, n##. This implies ##X_k = U_k + V_k##. He then proves the Law of Large Numbers by showing that for any given ##\epsilon > 0## the constant ##\delta > 0## can be chosen so that as ##n \to \infty##,
[tex]P\{ |U_1 + \cdots + U_n| > \frac{1}{2} \epsilon n \} \to 0\\<br /> \text{ and }\\<br /> P\{ |V_1 + \cdots + V_n | > \frac{1}{2} \epsilon n \} \to 0.[/tex]
The Law of Large Numbers follows from these.

The point of the truncation is to permit some statements to be made and proved even when variance is infinite---it works as long as ##E X## is finite. For proof of the above crucial results, consult Feller.

This is good stuff; I'll have to take a look at the text. Thanks a lot!
 

Similar threads

Replies
4
Views
2K
Replies
7
Views
2K
  • · Replies 17 ·
Replies
17
Views
2K
  • · Replies 9 ·
Replies
9
Views
1K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 15 ·
Replies
15
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K