Simple problems regarding sum of IID random variables

madilyn
Messages
13
Reaction score
0
Hi! I'm taking my first course in statistics and am hoping to get some intuition for this set of problems...

Suppose I have a bowl of marbles that each weighs m_{marble}=0.01 kg.

For each marble I swallow, there is a chance p=0.53 that it adds m_{marble} to my weight, and chance 1-p that it causes me to puke, therefore losing m_{puke}=0.011 kg of my weight.

1. Assume I religiously swallow n=10^{4} marbles each day. What fraction of the days do I expect to gain weight on?

Let X_{i}denote the random variable for my weight gained for each swallowed marble, indexed by i\in\mathbb{Z}^{+}.

Let Y denote the random variable for my total weight gained each day from swallowing n marbles, Y=\sum_{i=1}^{n}X_{i}. Then, denote
E\left(X\right) := E\left(X_{1}\right)=E\left(X_{2}\right)...=E\left(X_{n}\right)
Var\left(X\right) := Var\left(X_{1}\right)=Var\left(X_{2}\right)=...=Var\left(X_{n}\right)
such that the theoretical distribution of my daily weight gain is approximately normal with mean
E\left(Y\right)=E\left(X_{1}+...+X_{n}\right)=nE\left(X\right)
and variance
Var\left(Y\right)=Var\left(X_{1}+...+X_{n}\right)=nVar\left(X\right)
Then, I expect to gain weight on 1-P\left(Y\leq0\right)\approx0.892 of the days. Is this correct?

2. Why does Y approximately follow the distribution N\left(nE\left(X\right),\sqrt{nVar\left(X\right)}\right)?

Firstly, am I correct that the variance is nVar\left(X\right) and not n^{2}Var\left(X\right)? Can someone refresh me what's the intuitive difference between the random variable Y=X_{1}+...+X_{n} as compared to Y=50X again?

Secondly, it's not immediately obvious to me how the distribution approaches a Gaussian distribution as n\rightarrow\infty? Perhaps I can formulate this in terms of the convolution of a discrete function representing the distribution of my weight gain/loss for each marble swallowed? Will the discrete convolution approach generalize nicely to the sum of any discrete random variable?

3. For finite n, am I correct that this distribution converges faster to a Gaussian distribution near the center and slower in the tails as n increases? Can I quantify this rate of convergence?

Just from my intuition, I think the best strategy to attack this problem is to express the sum of the random variables Y=\sum_{i=1}^{n}X_{i} as a Fourier transform and then investigate the rate of convergence using an asymptotic expansion of the integral in large n i.e. saddle point method?

Thanks!
 
Physics news on Phys.org
madilyn said:
Hi! I'm taking my first course in statistics and am hoping to get some intuition for this set of problems...

This thread will probably be moved to the homework section. (I think the math homework section would be more obvious as a "math homework" subsection of mathematics - so I sympathize with the misplacement.) For homework, you should state the problem - rather than leave the helpers to guess it from reading your reasoning.
Then, I expect to gain weight on 1-P\left(Y\leq0\right)\approx0.892 of the days. Is this correct?

The question of "did gain" vs "did not gain" is a different question that "how much" is gained. "Did gain" can be represented by a random variable that only takes on the values 1 or 0. (A bernoulli random variable.) The theorems about expectation and variance apply to such a random variable but you'd get different answers than you get from answering questions about "how much".
Can someone refresh me what's the intuitive difference between the random variable Y=X_{1}+...+X_{n} as compared to Y=50X again?

Think of a computer simulation. The algorithm for simulating 50 X is to make a random determination for X and then multiply it by 50. The algorithm for simulation the sum of 50 different realizations of X is obviously to make 50 random determinations of X and add them. When you make 50 different random determinations, you have the possibility that opposite extremes will "cancel out". You don't get that with a simulation of 50 X since you only make one random determination for X.
3. For finite n, am I correct that this distribution converges faster to a Gaussian distribution near the center and slower in the tails as n increases? Can I quantify this rate of convergence?

That's a good question! I'll guess there are several ways, but i don't know them. One thing to do is understand the distinction between "pointwise convergence" and "uniform convergence". Convergence of a sequence of functions to another function is more complicated than convergence of a function evaluated at a sequence of points to a single number. There are several different definitions for "convergence" when we deal with sequences of functions converging to a single function.
 
Stephen Tashi said:
This thread will probably be moved to the homework section. (I think the math homework section would be more obvious as a "math homework" subsection of mathematics - so I sympathize with the misplacement.) For homework, you should state the problem - rather than leave the helpers to guess it from reading your reasoning.

Oh sorry, I came up with the questions myself so it isn't a homework problem.
Stephen Tashi said:
The question of "did gain" vs "did not gain" is a different question that "how much" is gained. "Did gain" can be represented by a random variable that only takes on the values 1 or 0. (A bernoulli random variable.) The theorems about expectation and variance apply to such a random variable but you'd get different answers than you get from answering questions about "how much".

I see. What would you intuitively interpret the 89.2% figure as if not the fraction of days I expect to have positive weight gain?

Stephen Tashi said:
Think of a computer simulation. The algorithm for simulating 50 X is to make a random determination for X and then multiply it by 50. The algorithm for simulation the sum of 50 different realizations of X is obviously to make 50 random determinations of X and add them. When you make 50 different random determinations, you have the possibility that opposite extremes will "cancel out". You don't get that with a simulation of 50 X since you only make one random determination for X.

Ah, this made a lot of sense! Thanks!

Stephen Tashi said:
That's a good question! I'll guess there are several ways, but i don't know them. One thing to do is understand the distinction between "pointwise convergence" and "uniform convergence". Convergence of a sequence of functions to another function is more complicated than convergence of a function evaluated at a sequence of points to a single number. There are several different definitions for "convergence" when we deal with sequences of functions converging to a single function.

Yes, I figured this is a difficult problem, but also one of the more interesting ones I've thought of while learning statistics. What's a good starting point for me to learn about the different definitions of "convergence" that you mentioned in your last sentence?

Thanks so much!
 
madilyn said:
3. For finite n, am I correct that this distribution converges faster to a Gaussian distribution near the center and slower in the tails as n increases? Can I quantify this rate of convergence?
A couple of notes. (Talking informally in terms of X=sum of events, not divided by n.)

1) You are talking about approximating a continuous PDF with a discrete PDF so they will always differ significantly at values of X that are not possible for the discrete PDF. You can handle this several ways. I think the best way is to look at convergence of the CDFs instead of the PDF. The CDF of a discrete real random variable is easy to unambiguously extend to all real values.

2) You know that the discrete CDF is 1.0 for all values of X > n. And n+epsilon is where the normal distribution CDF is farthest from 1.0. I suspect that this is where the greatest difference between the two CDFs is and that it gives the rate of uniform convergence. I can not prove it without some work.
 
Hi all, I've been a roulette player for more than 10 years (although I took time off here and there) and it's only now that I'm trying to understand the physics of the game. Basically my strategy in roulette is to divide the wheel roughly into two halves (let's call them A and B). My theory is that in roulette there will invariably be variance. In other words, if A comes up 5 times in a row, B will be due to come up soon. However I have been proven wrong many times, and I have seen some...
Thread 'Detail of Diagonalization Lemma'
The following is more or less taken from page 6 of C. Smorynski's "Self-Reference and Modal Logic". (Springer, 1985) (I couldn't get raised brackets to indicate codification (Gödel numbering), so I use a box. The overline is assigning a name. The detail I would like clarification on is in the second step in the last line, where we have an m-overlined, and we substitute the expression for m. Are we saying that the name of a coded term is the same as the coded term? Thanks in advance.
Back
Top