Blackwell-Girshick equation

From HandWiki
Short description: Variance of random sum

The Blackwell-Girshick equation is an equation in probability theory that allows for the calculation of the variance of random sums of random variables.[1] It is the equivalent of Wald's lemma for the expectation of composite distributions.

It is named after David Blackwell and Meyer Abraham Girshick.

Statement

Let [math]\displaystyle{ N }[/math] be a random variable with values in [math]\displaystyle{ \mathbb{Z}_{\ge 0} }[/math], let [math]\displaystyle{ X_1, X_2, X_3, \dots }[/math] be independent and identically distributed random variables, which are also independent of [math]\displaystyle{ N }[/math], and assume that the second moment exists for all [math]\displaystyle{ X_i }[/math] and [math]\displaystyle{ N }[/math]. Then, the random variable defined by

[math]\displaystyle{ Y:=\sum_{i=1}^NX_i }[/math]

has the variance

[math]\displaystyle{ \operatorname{Var}(Y)=\operatorname{Var}(N)\operatorname{E}(X_1)^2+\operatorname{E}(N)\operatorname{Var}(X_1) }[/math].

The Blackwell-Girshick equation can be derived using conditional variance and variance decomposition. If the [math]\displaystyle{ X_i }[/math] are natural number-valued random variables, the derivation can be done elementarily using the chain rule and the probability-generating function.[2]

Proof

For each [math]\displaystyle{ n\ge 0 }[/math], let [math]\displaystyle{ \chi_n }[/math] be the random variable which is 1 if [math]\displaystyle{ N }[/math] equals [math]\displaystyle{ n }[/math] and 0 otherwise, and let [math]\displaystyle{ Y_n:=X_1+\cdots+X_n }[/math]. Then

[math]\displaystyle{ \begin{align} \operatorname{E}(Y^2) & = \sum_{n=0}^\infty \operatorname{E}(\chi_n Y_n^2)\\ &= \sum_{n=0}^\infty \operatorname{P}(N=n) \operatorname{E}(Y_n^2)\\ &= \sum_{n=0}^\infty\operatorname{P}(N=n) (\operatorname{Var}(Y_n)+\operatorname{E}(Y_n)^2)\\ &= \sum_{n=0}^\infty\operatorname{P}(N=n) (n \operatorname{Var}(X_1)+n^2\operatorname{E}(X_1)^2)\\ &= \operatorname{E}(N) \operatorname{Var}(X_1) + \operatorname{E}(N^2) \operatorname{E}(X_1)^2. \end{align} }[/math]

By Wald's equation, under the given hypotheses, [math]\displaystyle{ \operatorname{E}(Y)=\operatorname{E}(N) \operatorname{E}(X_1) }[/math]. Therefore,

[math]\displaystyle{ \begin{align} \operatorname{Var}(Y)&=\operatorname{E}(Y^2)-\operatorname{E}(Y)^2\\ &= \operatorname{E}(N) \operatorname{Var}(X_1) + \operatorname{E}(N^2) \operatorname{E}(X_1)^2 - \operatorname{E}(N)^2 \operatorname{E}(X_1)^2 \\ &= \operatorname{E}(N) \operatorname{Var}(X_1) + \operatorname{Var}(N) \operatorname{E}(X_1)^2, \end{align} }[/math]

as desired.[3]

Example

Let [math]\displaystyle{ N }[/math] have a Poisson distribution with expectation [math]\displaystyle{ \lambda }[/math], and let [math]\displaystyle{ X_1, X_2, \dots }[/math] follow a Bernoulli distribution with parameter [math]\displaystyle{ p }[/math]. In this case, [math]\displaystyle{ Y }[/math] is also Poisson distributed with expectation [math]\displaystyle{ \lambda p }[/math], so its variance must be [math]\displaystyle{ \lambda p }[/math]. We can check this with the Blackwell-Girshick equation: [math]\displaystyle{ N }[/math] has variance [math]\displaystyle{ \lambda }[/math] while each [math]\displaystyle{ X_i }[/math] has mean [math]\displaystyle{ p }[/math] and variance [math]\displaystyle{ p(1-p) }[/math], so we must have

[math]\displaystyle{ \operatorname{Var}(Y)= \lambda p^2+\lambda p (1-p) = \lambda p }[/math].

Application and related concepts

The Blackwell-Girshick equation is used in actuarial mathematics to calculate the variance of composite distributions, such as the compound Poisson distribution. Wald's equation provides similar statements about the expectation of composite distributions.

Literature

  • For an example of an application: Mühlenthaler, M.; Raß, A.; Schmitt, M.; Wanka, R. (2021). "Exact Markov chain-based runtime analysis of a discrete particle swarm optimization algorithm on sorting and OneMax". Natural Computing: 1–27. 

References

  1. Blackwell, D. A.; Girshick, M. A. (1979). Theory of games and statistical decisions. Courier Corporation. 
  2. Achim Klenke (2013), Wahrscheinlichkeitstheorie (3rd ed.), Berlin Heidelberg: Springer-Verlag, p. 109, doi:10.1007/978-3-642-36018-3, ISBN 978-3-642-36017-6 
  3. Probability Theory : A Comprehensive Course, Achim Klenke, London, Heidelberg, New York, Dordrecht: Springer, 2nd ed., 2014, ISBN:978-1-4471-5360-3.