Marchenko–Pastur distribution

From HandWiki
Short description: Distribution of singular values of large rectangular random matrices
Plot of the Marchenko-Pastur distribution for various values of lambda
Marchenko-Pastur
Notation XMP(λ,σ2)
Parameters λ>0 aspect ratio (λ=m/n);
σ2>0 scale (entry variance)
Support x[λ,λ+], with λ±=σ2(1±λ)2;
plus an atom at 0 of mass 11/λ when λ>1.
PDF f(x)=(λ+x)(xλ)2πσ2λx for λxλ+, and 0 otherwise.
For λ>1 add a point mass 11/λ at 0.
CDF

F(x)=0 for x<λ;
F(x)=12πσ2λ[(λ+x)(xλ)λ+λ+2arccos(2xλλ+λ+λ)+2λλ+arctan(λλ+λ+xxλ)+π2(λ+λ+2λλ+)]
for λxλ+;

F(x)=1 for x>λ+.
For λ>1, add a jump of size 11/λ at 0.
Mean E[X]=σ2
Mode 0 if λ>1
σ2(1λ)21+λ if 0<λ1
Variance σ4λ
Skewness λ
Kurtosis λ+2 (excess λ1)
MGF

MX(t)=12πσ2λλλ+etx(λ+x)(xλ)xdx (continuous part). For λ>1, add 11/λ from the atom at 0.
Bessel form (let α=σ2(1+λ), β=2σ2λ, ρ=min(λ,1/λ), Sρ(z)=I0(z)+2n=1(ρ)nIn(z)):

MX(t)=eαt[1+λ2λI0(βt)1λI1(βt)|1λ|2λSρ(βt)], and if λ>1 add 11λ.



In the mathematical theory of random matrices, the Marchenko–Pastur distribution, or Marchenko–Pastur law, describes the asymptotic behavior of singular values of large rectangular random matrices. The theorem is named after Soviet Ukrainian mathematicians Volodymyr Marchenko and Leonid Pastur who proved this result in 1967.

If X denotes a m×n random matrix whose entries are independent identically distributed random variables with mean 0 and variance σ2<, let

Yn=1nXXT

and let λ1,λ2,,λm be the eigenvalues of Yn (viewed as random variables). Finally, consider the random measure

μm(A)=1m#{λjA},A.

counting the number of eigenvalues in the subset A included in .

Theorem. Assume that m,n so that the ratio m/nλ(0,+). Then μmμ (in weak* topology in distribution), where

μ(A)={(11λ)𝟏0A+ν(A),if λ>1ν(A),if 0λ1,

and

dν(x)=12πσ2(λ+x)(xλ)λx𝟏x[λ,λ+]dx

with

λ±=σ2(1±λ)2.

The Marchenko–Pastur law also arises as the free Poisson law in free probability theory, having rate 1/λ and jump size σ2.

Singular value bounds in the large system limit

As the dimensions of a random matrix 𝐗 grow larger, the max/min singular values converge to 𝐗F(1min(m,n)±1max(m,n)) .

These are useful approximations of singular value bounds for large matrices. For matrices of finite size as are typically encountered, they are more what you'd call "guidelines" than actual rules.

Moments

For each k1, its k-th moment is({{{1}}}, {{{2}}})

r=0k1σ2kr+1(kr)(k1r)λr=σ2kkr=0k1(kr)(kr+1)λr

Some transforms of this law

The Stieltjes transform is given by

s(z)=σ2(1λ)z(zσ2(λ+1))24λσ42λzσ2

for complex numbers z of positive imaginary part, where the complex square root is also taken to have positive imaginary part.({{{1}}}, {{{2}}}) It satisfies the quadratic equationλσ2zs(z)2+(zσ2(1λ))s(z)+1=0The Stieltjes transform can be repackaged in the form of the R-transform, which is given by({{{1}}}, {{{2}}})

R(z)=σ21σ2λz

The S-transform is given by({{{1}}}, {{{2}}})

S(z)=1σ2(1+λz).

For the case of σ=1, the η-transform ({{{1}}}, {{{2}}}) is given by 𝔼11+γX where X satisfies the Marchenko-Pastur law.

η(γ)=1(γ,λ)4γλ

where (x,z)=(x(1+z)2+1x(1z)2+1)2

For exact analysis of high dimensional regression in the proportional asymptotic regime, a convenient form is often T(u):=η(1u) which simplifies to

T(u)=1+λu+(1+uλ)2+4uλ2λ

The following functions B(u):=𝔼(uX+u)2 and V(u):=X(X+u)2, where X satisfies the Marchenko-Pastur law, show up in the limiting Bias and Variance respectively, of ridge regression and other regularized linear regression problems. One can show that B(u)=T(u)uT(u) and V(u)=T(u).

Application to correlation matrices

For the special case of correlation matrices, we know that σ2=1 and λ=m/n. This bounds the probability mass over the interval defined by

λ±=(1±mn)2.

Since this distribution describes the spectrum of random matrices with mean 0, the eigenvalues of correlation matrices that fall inside of the aforementioned interval could be considered spurious or noise. For instance, obtaining a correlation matrix of 10 stock returns calculated over a 252 trading days period would render λ+=(1+10252)21.43. Thus, out of 10 eigenvalues of said correlation matrix, only the values higher than 1.43 would be considered significantly different from random.

See also

References

Template:Random matrix theory