Conditional probability table

From HandWiki

In statistics, the conditional probability table (CPT) is defined for a set of discrete and mutually dependent random variables to display conditional probabilities of a single variable with respect to the others (i.e., the probability of each possible value of one variable if we know the values taken on by the other variables). For example, assume there are three random variables [math]\displaystyle{ x_1,x_2, x_3 }[/math] where each has [math]\displaystyle{ K }[/math] states. Then, the conditional probability table of [math]\displaystyle{ x_1 }[/math] provides the conditional probability values [math]\displaystyle{ P(x_1=a_k\mid x_2,x_3) }[/math] – where the vertical bar [math]\displaystyle{ | }[/math] means “given the values of” – for each of the K possible values [math]\displaystyle{ a_k }[/math] of the variable [math]\displaystyle{ x_1 }[/math] and for each possible combination of values of [math]\displaystyle{ x_2,\, x_3. }[/math] This table has [math]\displaystyle{ K^3 }[/math] cells. In general, for [math]\displaystyle{ M }[/math] variables [math]\displaystyle{ x_1,x_2,\ldots,x_M }[/math] with [math]\displaystyle{ K_i }[/math] states for each variable [math]\displaystyle{ x_i, }[/math] the CPT for any one of them has the number of cells equal to the product [math]\displaystyle{ K_1K_2\cdots K_M. }[/math][1]

A conditional probability table can be put into matrix form. As an example with only two variables, the values of [math]\displaystyle{ P(x_1=a_k\mid x_2=b_j)=T_{kj}, }[/math] with k and j ranging over K values, create a K×K matrix. This matrix is a stochastic matrix since the columns sum to 1; i.e. [math]\displaystyle{ \sum_k T_{kj} = 1 }[/math] for all j. For example, suppose that two binary variables x and y have the joint probability distribution given in this table:

x=0 x=1 P(y)
y=0 4/9 1/9 5/9
y=1 2/9 2/9 4/9
P(x) 6/9 3/9 1

Each of the four central cells shows the probability of a particular combination of x and y values. The first column sum is the probability that x =0 and y equals any of the values it can have – that is, the column sum 6/9 is the marginal probability that x=0. If we want to find the probability that y=0 given that x=0, we compute the fraction of the probabilities in the x=0 column that have the value y=0, which is 4/9 ÷ 6/9 = 4/6. Likewise, in the same column we find that the probability that y=1 given that x=0 is 2/9 ÷ 6/9 = 2/6. In the same way, we can also find the conditional probabilities for y equalling 0 or 1 given that x=1. Combining these pieces of information gives us this table of conditional probabilities for y:

x=0 x=1
P(y=0 given x) 4/6 1/3
P(y=1 given x) 2/6 2/3
Sum 1 1

With more than one conditioning variable, the table would still have one row for each potential value of the variable whose conditional probabilities are to be given, and there would be one column for each possible combination of values of the conditioning variables.

Moreover, the number of columns in the table could be substantially expanded to display the probabilities of the variable of interest conditional on specific values of only some, rather than all, of the other variables.

References

  1. Murphy, KP (2012). Machine learning: a probabilistic perspective. The MIT Press.