exercise:8ae7bbfa06: Difference between revisions
(Created page with "<div class="d-none"><math> \newcommand{\NA}{{\rm NA}} \newcommand{\mat}[1]{{\bf#1}} \newcommand{\exref}[1]{\ref{##1}} \newcommand{\secstoprocess}{\all} \newcommand{\NA}{{\rm NA}} \newcommand{\mathds}{\mathbb}</math></div> A gambler plays a game in which on each play he wins one dollar with probability <math>p</math> and loses one dollar with probability <math>q = 1 - p</math>. The ''Gambler's Ruin problem'' is the problem of finding the probability <math>w_x</math...") |
No edit summary |
||
Line 5: | Line 5: | ||
\newcommand{\secstoprocess}{\all} | \newcommand{\secstoprocess}{\all} | ||
\newcommand{\NA}{{\rm NA}} | \newcommand{\NA}{{\rm NA}} | ||
\newcommand{\mathds}{\mathbb}</math></div> A gambler plays a game in which on each play he wins | \newcommand{\mathds}{\mathbb}</math></div> A gambler plays a game in which on each play he wins one dollar with probability <math>p</math> and loses one dollar with probability <math>q = 1 -p</math>. The ''Gambler's Ruin problem'' is the problem of finding the probability <math>w_x</math> of winning an amount <math>T</math> before losing everything, starting with state <math>x</math>. Show that this problem may be considered to be an absorbing Markov chain with states 0, 1, 2, ..., <math>T</math> with 0 and <math>T</math> absorbing states. Suppose that a gambler has probability <math>p = .48</math> of winning on each play. Suppose, in addition, that the gambler starts with 50 dollars and that <math>T =100</math> dollars. Simulate this game 100 times and see how often the gambler is ruined. This estimates <math>w_{50}</math>. | ||
one dollar with probability <math>p</math> and loses one dollar with probability <math>q = 1 - | |||
p</math>. The ''Gambler's Ruin problem'' is the | |||
problem of | |||
finding the probability <math>w_x</math> of winning an amount <math>T</math> before losing | |||
everything, starting | |||
with state <math>x</math>. Show that this problem may be considered to be an absorbing | |||
Markov chain with states 0, 1, 2, | |||
Suppose that a gambler has probability <math>p = .48</math> of winning on each play. | |||
Suppose, in addition, that the gambler starts with 50 dollars and that <math>T = | |||
100</math> | |||
dollars. Simulate this game 100 times and see how often the gambler is ruined. | |||
This estimates <math>w_{50}</math>. |
Latest revision as of 23:02, 15 June 2024
A gambler plays a game in which on each play he wins one dollar with probability [math]p[/math] and loses one dollar with probability [math]q = 1 -p[/math]. The Gambler's Ruin problem is the problem of finding the probability [math]w_x[/math] of winning an amount [math]T[/math] before losing everything, starting with state [math]x[/math]. Show that this problem may be considered to be an absorbing Markov chain with states 0, 1, 2, ..., [math]T[/math] with 0 and [math]T[/math] absorbing states. Suppose that a gambler has probability [math]p = .48[/math] of winning on each play. Suppose, in addition, that the gambler starts with 50 dollars and that [math]T =100[/math] dollars. Simulate this game 100 times and see how often the gambler is ruined. This estimates [math]w_{50}[/math].