Craps principle

From HandWiki

In probability theory, the craps principle is a theorem about event probabilities under repeated iid trials. Let [math]\displaystyle{ E_1 }[/math] and [math]\displaystyle{ E_2 }[/math] denote two mutually exclusive events which might occur on a given trial. Then the probability that [math]\displaystyle{ E_1 }[/math] occurs before [math]\displaystyle{ E_2 }[/math] equals the conditional probability that [math]\displaystyle{ E_1 }[/math] occurs given that [math]\displaystyle{ E_1 }[/math] or [math]\displaystyle{ E_2 }[/math] occur on the next trial, which is

[math]\displaystyle{ \operatorname{P}[ E_1 \, \, \text{before}\,\, E_2]=\operatorname{P}\left[E_1\mid E_1\cup E_2\right]=\frac{\operatorname{P}[E_1]}{\operatorname{P}[E_1]+\operatorname{P}[E_2]} }[/math]

The events [math]\displaystyle{ E_1 }[/math] and [math]\displaystyle{ E_2 }[/math] need not be collectively exhaustive (if they are, the result is trivial).[1][2]

Proof

Let [math]\displaystyle{ A }[/math] be the event that [math]\displaystyle{ E_1 }[/math] occurs before [math]\displaystyle{ E_2 }[/math]. Let [math]\displaystyle{ B }[/math] be the event that neither [math]\displaystyle{ E_1 }[/math] nor [math]\displaystyle{ E_2 }[/math] occurs on a given trial. Since [math]\displaystyle{ B }[/math], [math]\displaystyle{ E_1 }[/math] and [math]\displaystyle{ E_2 }[/math] are mutually exclusive and collectively exhaustive for the first trial, we have

[math]\displaystyle{ \operatorname{P}(A) = \operatorname{P}(E_1)\operatorname{P}(A \mid E_1) + \operatorname{P}(E_2)\operatorname{P}(A \mid E_2) + \operatorname{P}(B) \operatorname{P}(A \mid B) = \operatorname{P}(E_1) + \operatorname{P}(B) \operatorname{P}(A \mid B) }[/math]

and [math]\displaystyle{ \operatorname{P}(B) = 1 - \operatorname{P}(E_1) - \operatorname{P}(E_2) }[/math]. Since the trials are i.i.d., we have [math]\displaystyle{ \operatorname{P}(A \mid B) = \operatorname{P}(A) }[/math]. Using [math]\displaystyle{ \operatorname{P}(A|E_1)=1,\quad \operatorname{P}(A|E_2)=0 }[/math] and solving the displayed equation for [math]\displaystyle{ \operatorname{P}(A) }[/math] gives the formula

[math]\displaystyle{ \operatorname{P}(A) = \frac{\operatorname{P}(E_1)}{\operatorname{P}(E_1)+\operatorname{P}(E_2)} }[/math].

Application

If the trials are repetitions of a game between two players, and the events are

[math]\displaystyle{ E_1:\mathrm{ player\ 1\ wins} }[/math]
[math]\displaystyle{ E_2:\mathrm{ player\ 2\ wins} }[/math]

then the craps principle gives the respective conditional probabilities of each player winning a certain repetition, given that someone wins (i.e., given that a draw does not occur). In fact, the result is only affected by the relative marginal probabilities of winning [math]\displaystyle{ \operatorname{P}[E_1] }[/math] and [math]\displaystyle{ \operatorname{P}[E_2] }[/math] ; in particular, the probability of a draw is irrelevant.

Stopping

If the game is played repeatedly until someone wins, then the conditional probability above is the probability that the player wins the game. This is illustrated below for the original game of craps, using an alternative proof.

Craps example

If the game being played is craps, then this principle can greatly simplify the computation of the probability of winning in a certain scenario. Specifically, if the first roll is a 4, 5, 6, 8, 9, or 10, then the dice are repeatedly re-rolled until one of two events occurs:

[math]\displaystyle{ E_1:\text{ the original roll (called ‘the point’) is rolled (a win) } }[/math]
[math]\displaystyle{ E_2:\text{ a 7 is rolled (a loss) } }[/math]

Since [math]\displaystyle{ E_1 }[/math] and [math]\displaystyle{ E_2 }[/math] are mutually exclusive, the craps principle applies. For example, if the original roll was a 4, then the probability of winning is

[math]\displaystyle{ \frac{3/36}{3/36 + 6/36}=\frac{1}{3} }[/math]

This avoids having to sum the infinite series corresponding to all the possible outcomes:

[math]\displaystyle{ \sum_{i=0}^{\infty}\operatorname{P}[\text{first i rolls are ties,}(i+1)^{\text{th}}\text{roll is ‘the point’}] }[/math]

Mathematically, we can express the probability of rolling [math]\displaystyle{ i }[/math] ties followed by rolling the point:

[math]\displaystyle{ \operatorname{P}[\text{first i rolls are ties, }(i+1)^{\text{th}} \text{roll is ‘the point’}] = (1-\operatorname{P}[E_1]-\operatorname{P}[E_2])^i\operatorname{P}[E_1] }[/math]

The summation becomes an infinite geometric series:

[math]\displaystyle{ \sum_{i=0}^{\infty} (1-\operatorname{P}[E_1]-\operatorname{P}[E_2])^i\operatorname{P}[E_1] = \operatorname{P}[E_1] \sum_{i=0}^{\infty} (1-\operatorname{P}[E_1]-\operatorname{P}[E_2])^i }[/math]
[math]\displaystyle{ = \frac{\operatorname{P}[E_1]}{1-(1-\operatorname{P}[E_1]-\operatorname{P}[E_2])} = \frac{\operatorname{P}[E_1]}{\operatorname{P}[E_1]+\operatorname{P}[E_2]} }[/math]

which agrees with the earlier result.

References

Notes