Feynman - Random Walk <D> and coin flipping

Click For Summary
SUMMARY

The discussion centers on the relationship between random walks and coin flipping, specifically the expected value of the absolute distance from the origin after N moves. It is established that the expected distance in a random walk is equal to the square root of N, and this concept parallels the behavior of coin flips, where the average difference between heads and tails increases with the number of flips. The participants clarify that the average difference and expected value are synonymous, and highlight that the lectures on probability in Feynman's work were authored by Matthew Sands, not Feynman himself. Additionally, the distinction between mean average deviation (MAD) and root mean square (RMS) distance is emphasized.

PREREQUISITES
  • Understanding of random walks and their mathematical implications
  • Familiarity with Bernoulli sequences and coin flipping probabilities
  • Knowledge of statistical concepts such as mean average deviation (MAD) and root mean square (RMS)
  • Basic grasp of Feynman's Lectures on Physics, particularly Volume I
NEXT STEPS
  • Study the mathematical foundations of random walks and their applications
  • Explore the concept of expected value in probability theory
  • Learn about the differences between mean average deviation (MAD) and root mean square (RMS) in statistics
  • Review the Feynman Lectures on Physics, focusing on the chapters related to probability and statistics
USEFUL FOR

Students of physics, mathematicians, statisticians, and anyone interested in the principles of probability and their applications in real-world scenarios.

QED-Kasper
Messages
32
Reaction score
0
Hello,

I have read the probability chapter in Feynman's lectures on physics. And got fascinated by the random walk. There is a statement, that in a game where either a vertical distance of +1 or -1 can be walked each move, the expected value of the absolute distance (lets call it <D>) from initial position 0, will be equal to the square root of N if N moves have been made.

For those that don't know and are interested: http://en.wikipedia.org/wiki/Random_walk.

What was fascinating for me for some reason was the fact that this expected distance <D> was becoming ever greater the more moves were made. For some reason I was thinking that the more moves the more likely the person will be at 0.

While I was thinking of this, the ordinary coin-flipping game came to my head. And I perceived an analogy. The more coins you flip the more likely that the fractional amount of tails you get will be closer to 1/2. Which is the probability of getting tails. However as the fractional amount of tails you get comes closer to 1/2, the difference between the amount of coins and tails on the average becomes bigger. Like this: 10 coin flips 4/10 tails 6/10 heads. the difference is only 2. but the fractional amount of tails is 4/10. Compared to 496 333/1000000 tails and 503777/1000000 heads. The fractional amount of tails is much closer to 1/2 but the difference between the amount of tails and heads is several thousands. So on the average you will see much greater difference between the amount of coins and tails the more you throw.
This is my question:
Isn't the average difference the same as the expected value <D> of the random walk?

Thanks for allowing me to share my experience.
 
Physics news on Phys.org
Hello QED-Kasper! :wink:
QED-Kasper said:
Isn't the average difference the same as the expected value <D> of the random walk?

That's right! :smile:

The coin-difference after n flips and the walk-distance after n steps (in 1D) have the same distribution … each process is a model for the other, and in particular, they have the same expected values.
 
Thanks, I appreciate that. And thank you for being extra kind :).
 
The coin-difference after n flips and the walk-distance after n steps (in 1D) have the same distribution … each process is a model for the other, and in particular, they have the same expected values.

I would like to point out two things here:

  1. "Walk-distance" D (generally referred to as "distance from the origin" in the theory of random walks) is defined to be the difference between the number of 'heads' and the number of 'tails' in a (Bernoulli) sequence of 'coin flips,' while the terms "expected value" and "average" have precisely the same meaning. So, "the average difference between heads and tails" and "the expected value of D" are just two ways of saying exactly the same thing. (Stating that "each process is a model for the other" having "the same distribution" and "the same expected values" obscures the fact that they are one and the same process.)
  2. The lecture on Probability in The Feynman Lectures on Physics Volume I, as well as the lecture that precedes it on Time and Distance, were written and delivered by Matthew Sands - Feynman had nothing to do with them (he was called unexpectedly out of town that week).
Mike Gottlieb
Editor, The Feynman Lectures on Physics, Definitive Edition
---
"www.feynmanlectures.info"[/URL]
 
Last edited by a moderator:
Thanks codelieb. I have to add though that I misread the text. In it Sands only mentions the expected distance. Which is also known as the MAD (mean average deviation) in statistics. The "square-root of N rule" applies to the RMS (root mean square) distance, aka standard deviation. This is what is actually being described in the text.
 

Similar threads

  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 21 ·
Replies
21
Views
4K
  • · Replies 41 ·
2
Replies
41
Views
8K
Replies
8
Views
1K
  • · Replies 10 ·
Replies
10
Views
5K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 7 ·
Replies
7
Views
3K
  • · Replies 27 ·
Replies
27
Views
10K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 6 ·
Replies
6
Views
5K