Simple (but interesting) gambling problem

Physics_wiz
Messages
227
Reaction score
0
I thought about this outta no where. I want to make a casino game where the odds are against the house, but the catch is that the game rules force the player to play until the odds change against him.

For example, make the odds 80% against the house (player has 80% chance of winning). If the player puts down a dollar at first and each time he wins, the casino pays him another dollar (doesn't double his money). He is forced to play a certain number of times, but if he loses he loses his initial dollar.

Here's how I thought about it (from the player's POV):

amount of money won = number of times the game was played = X
probability of winning each time = P
amount lost = $1

Number of times the game has to be played in order to break even:

(amount of money won)*(probability of winning) - (amount lost)*(probability of losing) = 0

X*P^X - ($1)*(1- P^X) = 0

which simplifies to: (P^X)*(X+1) - 1 = 0

All the casino has to do to increase the odds in its favor is to force the player to play at least 1 more time than X (which depends on P). Anyone see any problems with my logic or math?

PS. I initially thought this problem would be a lot simpler than this :smile:.
 
Physics news on Phys.org
The net value of the game is still negative for the player.

There are some casino games that can have odds that *slightly* favor the player, like progressive slot machines under the right circumstances.
 
I don't think I made myself clear (sorry about that). The game is supposed to be so that the final edge is against the player. However, I wanted to give the player a chance of more than 50% of making money every time he plays (ie. P>.5).
 
To make this clearer:

I was originally thinking about a game that looks like it favors the player instead of the house when it actually doesn't. So I thought I would give the player an 80% chance of winning but force him to play 4 times, because then his probability of winning would be 0.8^4 or about .41 (less than .5).

This got complicated when I figured out that the house has to pay more when it loses than when it wins. The above rules would actually still be against the house. I then tried to figure out how many times the rules need to force the player to play in order for the game to favor the house.
 
That's easy:
If the chance that the player wins one round is then the chance that the player wins n consecutive rounds is p^n. Assuming that the game starts with a wager of 1, and that the player must win n consecutive rounds to win n+1 dollars, we have the value of the game as:
(n+1)(p^n)<1
for the house to have a edge. Solving for p we get:
p<\frac{1}{\sqrt[n]{n+1}}

so
n=1 \rightarrow p<\frac{1}{2}
n=2 \rightarrow p<\frac{1}{\sqrt{3}}
n=3 \rightarrow p<\frac{1}{\sqrt[3]{4}}
and so on. It should be easy to see that even the n=2 case can have probabilities of winning a round that are greater than .5.

Notably, since the player is wagering his stack (more than 1) the later rounds are clearly not in favor of the player.
 
Last edited:
NateTG said:
That's easy:
If the chance that the player wins one round is then the chance that the player wins n consecutive rounds is p^n. Assuming that the game starts with a wager of 1, and that the player must win n consecutive rounds to win n+1 dollars, we have the value of the game as:<br /> (n+1)(p^n)&amp;lt;1<br /> for the house to have a edge. Solving for p we get:<br /> p&amp;lt;\frac{1}{\sqrt[n]{n+1}}<br /> <br /> so<br /> n=1 \rightarrow p&amp;lt;\frac{1}{2}<br /> n=2 \rightarrow p&amp;lt;\frac{1}{\sqrt{3}}<br /> n=3 \rightarrow p&amp;lt;\frac{1}{\sqrt[3]{4}}<br /> and so on. It should be easy to see that even the n=2 case can have probabilities of winning a round that are greater than .5.<br /> <br /> Notably, since the player is wagering his stack (more than 1) the later rounds are clearly not in favor of the player.
<br /> <br /> That&#039;s very interesting. You solved for the probability given the number of times instead of number of times given the probability (I can&#039;t think of a simple way off the top of my head to solve for n...calculators can always do it though). BTW, I have the the same formula, but your logic in finding the formula was simpler than mine. Thanks! <img src="https://cdn.jsdelivr.net/joypixels/assets/8.0/png/unicode/64/1f600.png" class="smilie smilie--emoji" loading="lazy" width="64" height="64" alt=":biggrin:" title="Big Grin :biggrin:" data-smilie="8"data-shortname=":biggrin:" />
 
Hi all, I've been a roulette player for more than 10 years (although I took time off here and there) and it's only now that I'm trying to understand the physics of the game. Basically my strategy in roulette is to divide the wheel roughly into two halves (let's call them A and B). My theory is that in roulette there will invariably be variance. In other words, if A comes up 5 times in a row, B will be due to come up soon. However I have been proven wrong many times, and I have seen some...
Thread 'Detail of Diagonalization Lemma'
The following is more or less taken from page 6 of C. Smorynski's "Self-Reference and Modal Logic". (Springer, 1985) (I couldn't get raised brackets to indicate codification (Gödel numbering), so I use a box. The overline is assigning a name. The detail I would like clarification on is in the second step in the last line, where we have an m-overlined, and we substitute the expression for m. Are we saying that the name of a coded term is the same as the coded term? Thanks in advance.
Back
Top