Hidden Markov random field

From HandWiki

In statistics, a hidden Markov random field is a generalization of a hidden Markov model. Instead of having an underlying Markov chain, hidden Markov random fields have an underlying Markov random field. Suppose that we observe a random variable [math]\displaystyle{ Y_i }[/math], where [math]\displaystyle{ i \in S }[/math]. Hidden Markov random fields assume that the probabilistic nature of [math]\displaystyle{ Y_i }[/math] is determined by the unobservable Markov random field [math]\displaystyle{ X_i }[/math], [math]\displaystyle{ i \in S }[/math]. That is, given the neighbors [math]\displaystyle{ N_i }[/math] of [math]\displaystyle{ X_i, X_i }[/math] is independent of all other [math]\displaystyle{ X_j }[/math] (Markov property). The main difference with a hidden Markov model is that neighborhood is not defined in 1 dimension but within a network, i.e. [math]\displaystyle{ X_i }[/math] is allowed to have more than the two neighbors that it would have in a Markov chain. The model is formulated in such a way that given [math]\displaystyle{ X_i }[/math], [math]\displaystyle{ Y_i }[/math] are independent (conditional independence of the observable variables given the Markov random field).

In the vast majority of the related literature, the number of possible latent states is considered a user-defined constant. However, ideas from nonparametric Bayesian statistics, which allow for data-driven inference of the number of states, have been also recently investigated with success, e.g.[1]

See also

References

  1. ↑ Sotirios P. Chatzis, Gabriel Tsechpenakis, β€œThe Infinite Hidden Markov Random Field Model,” IEEE Transactions on Neural Networks, vol. 21, no. 6, pp. 1004–1014, June 2010. [1]