Isserlis' theorem
In probability theory, Isserlis' theorem or Wick's probability theorem is a formula that allows one to compute higher-order moments of the multivariate normal distribution in terms of its covariance matrix. It is named after Leon Isserlis.
This theorem is also particularly important in particle physics, where it is known as Wick's theorem after the work of (Wick 1950).[1] Other applications include the analysis of portfolio returns,[2] quantum field theory[3] and generation of colored noise.[4]
Statement
If [math]\displaystyle{ (X_1,\dots, X_{n}) }[/math] is a zero-mean multivariate normal random vector, then[math]\displaystyle{ \operatorname{E} [\,X_1 X_2\cdots X_{n}\,] = \sum_{p\in P_n^2}\prod_{\{i,j\}\in p} \operatorname{E}[\,X_i X_j\,] = \sum_{p\in P_n^2}\prod_{\{i,j\}\in p} \operatorname{Cov}(\,X_i, X_j\,), }[/math]where the sum is over all the pairings of [math]\displaystyle{ \{1,\ldots,n\} }[/math], i.e. all distinct ways of partitioning [math]\displaystyle{ \{1,\ldots,n\} }[/math] into pairs [math]\displaystyle{ \{i,j\} }[/math], and the product is over the pairs contained in [math]\displaystyle{ p }[/math].[5][6]
More generally, if [math]\displaystyle{ (Z_1,\dots, Z_{n}) }[/math] is a zero-mean complex-valued multivariate normal random vector, then the formula still holds.
The expression on the right-hand side is also known as the hafnian of the covariance matrix of [math]\displaystyle{ (X_1,\dots, X_{n}) }[/math].
Odd case
If [math]\displaystyle{ n=2m+1 }[/math] is odd, there does not exist any pairing of [math]\displaystyle{ \{1,\ldots,2m+1\} }[/math]. Under this hypothesis, Isserlis' theorem implies that[math]\displaystyle{ \operatorname{E}[\,X_1 X_2\cdots X_{2m+1}\,] = 0. }[/math] This also follows from the fact that [math]\displaystyle{ -X=(-X_1,\dots,-X_n) }[/math] has the same distribution as [math]\displaystyle{ X }[/math], which implies that [math]\displaystyle{ \operatorname{E}[\,X_1 \cdots X_{2m+1}\,]=\operatorname{E}[\,(-X_1) \cdots (-X_{2m+1})\,]=-\operatorname{E}[\,X_1 \cdots X_{2m+1}\,] = 0 }[/math].
Even case
In his original paper,[7] Leon Isserlis proves this theorem by mathematical induction, generalizing the formula for the [math]\displaystyle{ 4^{\text{th}} }[/math] order moments,[8] which takes the appearance
- [math]\displaystyle{ \operatorname{E}[\,X_1 X_2 X_3 X_4\,] = \operatorname{E}[X_1X_2]\,\operatorname{E}[X_3X_4] + \operatorname{E}[X_1X_3]\,\operatorname{E}[X_2X_4] + \operatorname{E}[X_1X_4]\,\operatorname{E}[X_2X_3]. }[/math]
If [math]\displaystyle{ n=2m }[/math] is even, there exist [math]\displaystyle{ (2m)!/(2^{m}m!) = (2m-1)!! }[/math] (see double factorial) pair partitions of [math]\displaystyle{ \{1,\ldots,2m\} }[/math]: this yields [math]\displaystyle{ (2m)!/(2^{m}m!) = (2m-1)!! }[/math] terms in the sum. For example, for [math]\displaystyle{ 4^{\text{th}} }[/math] order moments (i.e. [math]\displaystyle{ 4 }[/math] random variables) there are three terms. For [math]\displaystyle{ 6^{\text{th}} }[/math]-order moments there are [math]\displaystyle{ 3\times 5=15 }[/math] terms, and for [math]\displaystyle{ 8^{\text{th}} }[/math]-order moments there are [math]\displaystyle{ 3\times5\times7 = 105 }[/math] terms.
Example
We can evaluate the characteristic function of gaussians by the Isserlis theorem:[math]\displaystyle{ E[e^{-iX}] = \sum_k \frac{(-i)^k}{k!} E[X^k] = \sum_k \frac{(-i)^{2k}}{(2k)!} E[X^{2k}] = \sum_k \frac{(-i)^{2k}}{(2k)!} \frac{(2k)!}{k!2^k} E[X^{2}]^k = e^{-\frac 12 E[X^2]} }[/math]
Proof
Since the formula is linear on both sides, if we can prove the real case, we get the complex case for free.
Let [math]\displaystyle{ \Sigma_{ij} = \operatorname{Cov}(X_i, X_j) }[/math] be the covariance matrix, so that we have the zero-mean multivariate normal random vector [math]\displaystyle{ (X_1, ..., X_n) \sim N(0, \Sigma) }[/math]. Since both sides of the formula are continuous with respect to [math]\displaystyle{ \Sigma }[/math], it suffices to prove the case when [math]\displaystyle{ \Sigma }[/math] is invertible.
Using quadratic factorization [math]\displaystyle{ -x^T\Sigma^{-1}x/2 + v^Tx - v^T\Sigma v/2 = -(x-\Sigma v)^T\Sigma^{-1}(x-\Sigma v)/2 }[/math], we get
[math]\displaystyle{ \frac{1}{\sqrt{(2\pi)^n\det\Sigma}}\int e^{-x^T\Sigma^{-1}x/2 + v^Tx} dx = e^{v^T\Sigma v/2} }[/math]
Differentiate under the integral sign with [math]\displaystyle{ \partial_{v_1, ..., v_n}|_{v_1, ..., v_n=0} }[/math] to obtain
[math]\displaystyle{ E[X_1\cdots X_n] = \partial_{v_1, ..., v_n}|_{v_1, ..., v_n=0}e^{v^T\Sigma v/2} }[/math].
That is, we need only find the coefficient of term [math]\displaystyle{ v_1\cdots v_n }[/math] in the Taylor expansion of [math]\displaystyle{ e^{v^T\Sigma v/2} }[/math].
If [math]\displaystyle{ n }[/math] is odd, this is zero. So let [math]\displaystyle{ n = 2m }[/math], then we need only find the coefficient of term [math]\displaystyle{ v_1\cdots v_n }[/math] in the polynomial [math]\displaystyle{ \frac{1}{m!}(v^T\Sigma v/2)^m }[/math].
Expand the polynomial and count, we obtain the formula. [math]\displaystyle{ \square }[/math]
Generalizations
Gaussian integration by parts
An equivalent formulation of the Wick's probability formula is the Gaussian integration by parts. If [math]\displaystyle{ (X_1,\dots X_{n}) }[/math] is a zero-mean multivariate normal random vector, then
[math]\displaystyle{ \operatorname{E}(X_1 f(X_1,\ldots,X_n))=\sum_{i=1}^{n} \operatorname{Cov}(X_1, X_i)\operatorname{E}(\partial_{X_i}f(X_1,\ldots,X_n)). }[/math]This is a generalization of Stein's lemma.
The Wick's probability formula can be recovered by induction, considering the function [math]\displaystyle{ f:\mathbb{R}^n\to\mathbb{R} }[/math] defined by [math]\displaystyle{ f(x_1,\ldots,x_n)=x_2\ldots x_n }[/math]. Among other things, this formulation is important in Liouville conformal field theory to obtain conformal Ward identities, BPZ equations[9] and to prove the Fyodorov-Bouchaud formula.[10]
Non-Gaussian random variables
For non-Gaussian random variables, the moment-cumulants formula[11] replaces the Wick's probability formula. If [math]\displaystyle{ (X_1,\dots X_{n}) }[/math] is a vector of random variables, then [math]\displaystyle{ \operatorname{E}(X_1 \ldots X_n)=\sum_{p\in P_n} \prod_{b\in p} \kappa\big((X_i)_{i\in b}\big), }[/math]where the sum is over all the partitions of [math]\displaystyle{ \{1,\ldots,n\} }[/math], the product is over the blocks of [math]\displaystyle{ p }[/math] and [math]\displaystyle{ \kappa\big((X_i)_{i\in b}\big) }[/math] is the joint cumulant of [math]\displaystyle{ (X_i)_{i\in b} }[/math].
See also
References
- ↑ Wick, G.C. (1950). "The evaluation of the collision matrix". Physical Review 80 (2): 268–272. doi:10.1103/PhysRev.80.268. Bibcode: 1950PhRv...80..268W.
- ↑ Repetowicz, Przemysław; Richmond, Peter (2005). "Statistical inference of multivariate distribution parameters for non-Gaussian distributed time series". Acta Physica Polonica B 36 (9): 2785–2796. Bibcode: 2005AcPPB..36.2785R. http://th-www.if.uj.edu.pl/acta/vol36/pdf/v36p2785.pdf.
- ↑ Perez-Martin, S.; Robledo, L.M. (2007). "Generalized Wick's theorem for multiquasiparticle overlaps as a limit of Gaudin's theorem". Physical Review C 76 (6): 064314. doi:10.1103/PhysRevC.76.064314. Bibcode: 2007PhRvC..76f4314P.
- ↑ Bartosch, L. (2001). "Generation of colored noise". International Journal of Modern Physics C 12 (6): 851–855. doi:10.1142/S0129183101002012. Bibcode: 2001IJMPC..12..851B.
- ↑ Janson, Svante (June 1997). Gaussian Hilbert Spaces. Cambridge Core. doi:10.1017/CBO9780511526169. ISBN 9780521561280. https://www.cambridge.org/core/books/gaussian-hilbert-spaces/658C87D5A0E7FB440FC34D82B08167FC. Retrieved 2019-11-30.
- ↑ Michalowicz, J.V.; Nichols, J.M.; Bucholtz, F.; Olson, C.C. (2009). "An Isserlis' theorem for mixed Gaussian variables: application to the auto-bispectral density". Journal of Statistical Physics 136 (1): 89–102. doi:10.1007/s10955-009-9768-3. Bibcode: 2009JSP...136...89M.
- ↑ Isserlis, L. (1918). "On a formula for the product-moment coefficient of any order of a normal frequency distribution in any number of variables". Biometrika 12 (1–2): 134–139. doi:10.1093/biomet/12.1-2.134. https://zenodo.org/record/1431593.
- ↑ Isserlis, L. (1916). "On Certain Probable Errors and Correlation Coefficients of Multiple Frequency Distributions with Skew Regression". Biometrika 11 (3): 185–190. doi:10.1093/biomet/11.3.185. https://zenodo.org/record/1431585.
- ↑ Kupiainen, Antti; Rhodes, Rémi; Vargas, Vincent (2019-11-01). "Local Conformal Structure of Liouville Quantum Gravity". Communications in Mathematical Physics 371 (3): 1005–1069. doi:10.1007/s00220-018-3260-3. ISSN 1432-0916. Bibcode: 2019CMaPh.371.1005K.
- ↑ Remy, Guillaume (2020). "The Fyodorov–Bouchaud formula and Liouville conformal field theory". Duke Mathematical Journal 169. doi:10.1215/00127094-2019-0045.
- ↑ Leonov, V. P.; Shiryaev, A. N. (January 1959). "On a Method of Calculation of Semi-Invariants". Theory of Probability & Its Applications 4 (3): 319–329. doi:10.1137/1104031. https://epubs.siam.org/doi/abs/10.1137/1104031.
Further reading
- Koopmans, Lambert G. (1974). The spectral analysis of time series. San Diego, CA: Academic Press.
Original source: https://en.wikipedia.org/wiki/Isserlis' theorem.
Read more |