Insights Blog
-- Browse All Articles --
Physics Articles
Physics Tutorials
Physics Guides
Physics FAQ
Math Articles
Math Tutorials
Math Guides
Math FAQ
Education Articles
Education Guides
Bio/Chem Articles
Technology Guides
Computer Science Tutorials
Forums
General Math
Calculus
Differential Equations
Topology and Analysis
Linear and Abstract Algebra
Differential Geometry
Set Theory, Logic, Probability, Statistics
MATLAB, Maple, Mathematica, LaTeX
Trending
Featured Threads
Log in
Register
What's new
Search
Search
Search titles only
By:
General Math
Calculus
Differential Equations
Topology and Analysis
Linear and Abstract Algebra
Differential Geometry
Set Theory, Logic, Probability, Statistics
MATLAB, Maple, Mathematica, LaTeX
Menu
Log in
Register
Navigation
More options
Contact us
Close Menu
JavaScript is disabled. For a better experience, please enable JavaScript in your browser before proceeding.
You are using an out of date browser. It may not display this or other websites correctly.
You should upgrade or use an
alternative browser
.
Forums
Mathematics
Set Theory, Logic, Probability, Statistics
Show that ##Y_{\infty}=0 ##
Reply to thread
Message
[QUOTE="WMDhamnekar, post: 6869205, member: 696551"] Let ##X_1, X_2, \dots ##be independent, identically distributed random variables with ##P{X_j = 1} = q, P{X_j = −1} = 1 − q.## Let ##S_0 = 0 ##and for n ≥ 1, ##S_n = X_1 + X_2 + \dots + X_n.## Let ##Y_n = e^{S_n}## Let ##Y_{\infty} = \lim\limits_{n\to\infty} Y_n.## Explain why ##Y_{\infty} = 0.## (Hint: there are at least two ways to show this. One is to consider ##\log{Y_n}## and use the law of large numbers. Another is to note that with probability one ##Y_{n+1}/Y_n## does not converge.) [B]My answer:[/B] We can use the law of large numbers to show that ##\log Y_n## converges almost surely to ##-\infty##, which implies that ##Y_n## converges almost surely to 0. First, note that ##Y_n## is a non-negative random variable, so its logarithm ##\log Y_n## is well-defined. Using the definition of ##Y_n##, we have: ##\log{Y_n}= S_n =\displaystyle\sum_{j=1}^n X_j## By the law of large numbers, we know that the sample mean $\frac{1}{n}\sum_{j=1}^{n}X_j$ converges almost surely to the expected value ##E[X_j] = q - (1-q) = 2q-1.## Therefore, we have: ##\frac{1}{n}\displaystyle\sum_{j=1}^n X_j \to (2q-1)## almost surely Using this result, we can show that ##\log{Y_n}## converges almost surely to ##-\infty## as follows: ####\begin{align*} \log Y_n &= \sum_{j=1}^{n} X_j \ \ &= n\cdot\frac{1}{n}\sum_{j=1}^{n} X_j \ \ &\to n(2q-1) \quad \text{almost surely} \ \ &= \infty \quad \text{if } q > \frac{1}{2} \ \ &= -\infty \quad \text{if } q < \frac{1}{2} \end{align*}#### When ##q < \frac{1}{2}##, we have ##\log{Y_n} \to -\infty## almost surely, which implies that ##Y_{\infty} = 0## almost surely. Alternatively, we can use the fact that ##Y_n## is a martingale to show that ##Y_n## does not converge almost surely. By definition, ##Y_n## is a martingale because ##E[Y_{n+1} \mid Y_1,\ldots,Y_n] = E[e^{S_{n+1}} \mid Y_1,\ldots,Y_n] = e^{S_n} = Y_n##. Using the martingale convergence theorem, we know that if ##Y_n## converges almost surely to a limit ##Y_{\infty}##, then ##Y_{\infty}## must be a constant because ##Y_n## is a martingale. However, we can show that ##Y_{n+1}/Y_n## does not converge almost surely, which implies that ##Y_n## does not converge almost surely. To see why ##Y_{n+1}/Y_n## does not converge almost surely, note that: ####\begin{align*}\displaystyle\frac{Y_{n+1}}{Y_n} &= \displaystyle\frac{e^{S_{n+1}}}{e^{S_n}} \ \ &= e^{S_{n+1}-S_n} \ \ &= e^{X_{n+1}} \end{align*}#### Since ##X_{n+1}## takes on the values 1 and -1 with positive probability, we have ##e^{X_{n+1}} = e## or ##1/e## with positive probability. Therefore, ##Y_{n+1}/Y_n## does not converge almost surely, which implies that ##Y_n ##does not converge almost surely.If q = 1/2, what would be your answer? If ##q = 1/2##, then ##X_j## is a symmetric random variable with ##E[X_j] = 0##. In this case, ##S_n## is a random walk with mean zero, which means that ##\log{Y_n} = S_n## also has mean zero. By the law of large numbers, we know that ##\frac{1}{n}\cdot S_n## converges almost surely to zero, which implies that ##S_n## converges with probability one to zero. Using the continuous mapping theorem, we have ##\log{Y_n} \to 0## with probability one, which implies that ## Y_n \to 1##with probability one. Therefore, ##Y_n## converges to 1 with probability one when ##q = 1/2##. Is the above answer correct? [/QUOTE]
Insert quotes…
Post reply
Forums
Mathematics
Set Theory, Logic, Probability, Statistics
Show that ##Y_{\infty}=0 ##
Back
Top