A and B bet on successive flips of a coin. On each flip, A collects 1 unit from B if the coin comes up heads (this has probability p), otherwise A pays B 1 unit. This continues until either of them runs out of money. What's the probability A wins all the money if A starts with i units and B starts with N - i units?
This is a great puzzle! When I first saw it, it reminded me of
a similar problem [essentially the continuous-time equivalent of
coin flipping, where B has unlimited amounts of money, and A gets
shot if his wealth ever falls below H1 or rises above H2. What it
the probability that A is still alive after time t?] I have come
across in finance. That is a rather involved problem.
Yours though, is much easier -just a handful of lines. For equal
probabilities of head/tail, the probability of A winning is
linear on how much they have (relatively) start
with.
I don't want to spoil your fun by giving you the 'trick'. My hint
to you is giving you the answer. Instead, can I show you how I
got there? It was only after going through the agony below that I
found the answer, and realised what the trick must be. The
techniques below are quite advanced (it's basically 'financial
calculus'), so please don't be put off if they are unfamiliar. I
repeat, your problem can be solved with 'O' level maths.
So here goes. I first assumed that the probability of head/tails were equal.
Then I decided to attack the continuous time equivalent of your problem.
Define an asset, S(t) such that
I tried allowing both players to go into debt, then computing
the probability that B has 0 money (A wins) just after the
(k+1)th flip AND neither player had 0 money at any flip
< = k (ie P(A wins at the (k+1)th
flip).
Summing those probabilities has so far produced calculations too
complex for me to believe I won't make an algebraic mistake.
There is a total amount of cash of N. Denote A holding an
amount x of cash after the nth roll by (x,n).
Assume the probability of heads/tails are equal. This means that
the state (x,n) can evolve to (x+1,n+1) and (x-1,n+1) each with
50% probability.
Denote P(x,n) as being the probability that A will win contingent
on not having won/lost to date. So P(N,t)=1 ; P(0,t)=0. Given
that there are just 2 possible states at n+1 (given (x,n)),
P(x,n)=1/2.[P(x+1,n+1)+P(x-1,n+1)]
Repeating the argument for one more step:
P(x,n)=1/4.[P(x+2,n+2)+2.P(x,n+2)+ P(x-2,n+2)]
But what is the difference between P(x,n) and P(x,n+2)? Nothing.
Dice have no memory and we play the game forever (or until A
wins/loses). So
P(x,n+2)=1/4.[P(x+2,n+2)+2.P(x,n+2)+ P(x-2,n+2)]
i.e.
P(x,n+2)=1/2.[P(x+2,n+2) + P(x-2,n+2)]
So these three points are linear, and repeating the argument
justifies linearity from x=0..N.
Andre
That is a nice trick! It is unexpected to use the probability values 1 and 0 in such a way.