Generalized inverse Gaussian distribution

From HandWiki
Short description: Family of continuous probability distributions
Generalized inverse Gaussian
Probability density function
Probability density plots of GIG distributions
Parameters a > 0, b > 0, p real
Support x > 0
PDF f(x)=(a/b)p/22Kp(ab)x(p1)e(ax+b/x)/2
Mean E[x]=b Kp+1(ab)a Kp(ab)
E[x1]=a Kp+1(ab)b Kp(ab)2pb
E[lnx]=lnba+plnKp(ab)
Mode (p1)+(p1)2+aba
Variance (ba)[Kp+2(ab)Kp(ab)(Kp+1(ab)Kp(ab))2]
MGF (aa2t)p2Kp(b(a2t))Kp(ab)
CF (aa2it)p2Kp(b(a2it))Kp(ab)

In probability theory and statistics, the generalized inverse Gaussian distribution (GIG) is a three-parameter family of continuous probability distributions with probability density function

f(x)=(a/b)p/22Kp(ab)x(p1)e(ax+b/x)/2,x>0,

where Kp is a modified Bessel function of the second kind, a > 0, b > 0 and p a real parameter. It is used extensively in geostatistics, statistical linguistics, finance, etc. This distribution was first proposed by Étienne Halphen.[1][2][3] It was rediscovered and popularised by Ole Barndorff-Nielsen, who called it the generalized inverse Gaussian distribution. Its statistical properties are discussed in Bent Jørgensen's lecture notes.[4]

Properties

Alternative parametrization

By setting θ=ab and η=b/a, we can alternatively express the GIG distribution as

f(x)=12ηKp(θ)(xη)p1eθ(x/η+η/x)/2,

where θ is the concentration parameter while η is the scaling parameter.

Summation

Barndorff-Nielsen and Halgreen proved that the GIG distribution is infinitely divisible.[5]

Entropy

H=12log(ba)+log(2Kp(ab))(p1)[ddνKν(ab)]ν=pKp(ab)+ab2Kp(ab)(Kp+1(ab)+Kp1(ab))

where [ddνKν(ab)]ν=p is a derivative of the modified Bessel function of the second kind with respect to the order ν evaluated at ν=p

Characteristic Function

The characteristic of a random variable XGIG(p,a,b) is given as (for a derivation of the characteristic function, see supplementary materials of [6])

E(eitX)=(aa2it)p2Kp((a2it)b)Kp(ab)

for t where i denotes the imaginary number.

Special cases

The inverse Gaussian and gamma distributions are special cases of the generalized inverse Gaussian distribution for p = −1/2 and b = 0, respectively.[7] Specifically, an inverse Gaussian distribution of the form

f(x;μ,λ)=[λ2πx3]1/2exp(λ(xμ)22μ2x)

is a GIG with a=λ/μ2, b=λ, and p=1/2. A gamma distribution of the form

g(x;α,β)=βα1Γ(α)xα1eβx

is a GIG with a=2β, b=0, and p=α.

Other special cases include the inverse-gamma distribution, for a = 0.[7]

Conjugate prior for Gaussian

The GIG distribution is conjugate to the normal distribution when serving as the mixing distribution in a normal variance-mean mixture.[8][9] Let the prior distribution for some hidden variable, say z, be GIG:

P(za,b,p)=GIG(za,b,p)

and let there be T observed data points, X=x1,,xT, with normal likelihood function, conditioned on z:

P(Xz,α,β)=i=1TN(xiα+βz,z)

where N(xμ,v) is the normal distribution, with mean μ and variance v. Then the posterior for z, given the data is also GIG:

P(zX,a,b,p,α,β)=GIG(za+Tβ2,b+S,pT2)

where S=i=1T(xiα)2.[note 1]

Sichel distribution

The Sichel distribution results when the GIG is used as the mixing distribution for the Poisson parameter λ.[10][11]

Notes

  1. Due to the conjugacy, these details can be derived without solving integrals, by noting that
    P(zX,a,b,p,α,β)P(za,b,p)P(Xz,α,β).
    Omitting all factors independent of z, the right-hand-side can be simplified to give an un-normalized GIG distribution, from which the posterior parameters can be identified.

References

  1. Seshadri, V. (1997). "Halphen's laws". in Kotz, S.; Read, C. B.; Banks, D. L.. Encyclopedia of Statistical Sciences, Update Volume 1. New York: Wiley. pp. 302–306. 
  2. Perreault, L.; Bobée, B.; Rasmussen, P. F. (1999). "Halphen Distribution System. I: Mathematical and Statistical Properties". Journal of Hydrologic Engineering 4 (3): 189. doi:10.1061/(ASCE)1084-0699(1999)4:3(189). 
  3. Étienne Halphen was the grandson of the mathematician Georges Henri Halphen.
  4. Jørgensen, Bent (1982). Statistical Properties of the Generalized Inverse Gaussian Distribution. Lecture Notes in Statistics. 9. New York–Berlin: Springer-Verlag. ISBN 0-387-90665-7. 
  5. Barndorff-Nielsen, O.; Halgreen, Christian (1977). "Infinite Divisibility of the Hyperbolic and Generalized Inverse Gaussian Distributions". Zeitschrift für Wahrscheinlichkeitstheorie und verwandte Gebiete 38: 309–311. doi:10.1007/BF00533162. 
  6. Pal, Subhadip; Gaskins, Jeremy (23 May 2022). "Modified Pólya-Gamma data augmentation for Bayesian analysis of directional data". Journal of Statistical Computation and Simulation 92 (16): 3430–3451. doi:10.1080/00949655.2022.2067853. ISSN 0094-9655. https://www.tandfonline.com/doi/abs/10.1080/00949655.2022.2067853?journalCode=gscs20. 
  7. 7.0 7.1 Johnson, Norman L.; Kotz, Samuel; Balakrishnan, N. (1994), Continuous univariate distributions. Vol. 1, Wiley Series in Probability and Mathematical Statistics: Applied Probability and Statistics (2nd ed.), New York: John Wiley & Sons, pp. 284–285, ISBN 978-0-471-58495-7 
  8. Karlis, Dimitris (2002). "An EM type algorithm for maximum likelihood estimation of the normal–inverse Gaussian distribution". Statistics & Probability Letters 57 (1): 43–52. doi:10.1016/S0167-7152(02)00040-8. 
  9. Barndorf-Nielsen, O. E. (1997). "Normal Inverse Gaussian Distributions and stochastic volatility modelling". Scand. J. Statist. 24 (1): 1–13. doi:10.1111/1467-9469.00045. 
  10. Sichel, Herbert S. (1975). "On a distribution law for word frequencies". Journal of the American Statistical Association 70 (351a): 542-547. doi:10.1080/01621459.1975.10482469. 
  11. Stein, Gillian Z.; Zucchini, Walter; Juritz, June M. (1987). "Parameter estimation for the Sichel distribution and its multivariate extension". Journal of the American Statistical Association 82 (399): 938-944. doi:10.1080/01621459.1987.10478520. 

See also