Ghosh–Pratt identity

From HandWiki
Ghosh-Pratt identity
TypeTheorem
FieldMathematical statistics
StatementEθ0[vol(C(X))]=θθ0Pθ0(θC(X))dθ
First stated byJayanta Kumar Ghosh
John Winsor Pratt
First stated in1961

In mathematical statistics, the Ghosh-Pratt identity is a theorem that establishes a formal relationship between the expected volume of a confidence set and its probability of false coverage. It is a cornerstone of optimal estimation, as it allows the problem of finding the shortest confidence interval to be framed as a problem of maximizing the power of a statistical test.

The identity was independently discovered and published in 1961 by the Indian statistician Jayanta Kumar Ghosh and the American statistician John Winsor Pratt.

Formal statement

Let X be a random variable with a probability distribution indexed by a parameter θΘ. Let C(X) denote a random variable representing a confidence set for θ. The Ghosh-Pratt identity states that the expected volume (or length, if Θ is one dimensional) of the confidence set C(X), calculated under the true parameter value θ0, is equal to the integral of the probabilities of including false values of θ in the set:

Eθ0[vol(C(X))]=θθ0Pθ0(θC(X))dθ

In simpler terms, the expected length of a confidence interval is the sum (integral) of the probabilities of covering all possible incorrect values of θ.[1]

Derivation

The proof of the identity relies on the relationship between the volume of a set and the indicator function. Given a specific realization of data x, let C(x) be an arbitrary confidence set for the parameter θ. The volume of this set, vol(C(x)), can be expressed as:

vol(C(x))=ΘI(θC(x))dθ

where I() is the indicator function. To find the expected volume under the true parameter θ0, we consider all possible confidence sets across the data generating process X and take the expected value with respect to the distribution of X given θ0:

Eθ0[vol(C(X))]=Eθ0[ΘI(θC(X))dθ]

By applying Fubini's theorem, we can reverse the order of expectation:

Eθ0[vol(C(X))]=ΘEθ0[I(θC(X))]dθ

Since the expectation of an indicator function is simply the probability of the event indicated:

Eθ0[I(θC(X))]=Pθ0(θC(X))

Substituting this false coverage probability back into the integral yields the identity:

Eθ0[vol(C(X))]=ΘPθ0(θC(X))dθ

This result shows that the expected volume is the integral of the probability that the confidence set covers any value θ (both the true value θ0 as well as all false values θθ0).

Generalization

While the identity is often presented in the context of a continuous random variable, it can be generalized using measure theory to cover discrete random variables and mixed distributions as well. If μ is a sigma-finite measure on the parameter space Θ, the expected measure of the confidence set C(X) under the distribution Pθ0 is given by:

Eθ0[μ(C(X))]=ΘPθ0(θC(X))μ(dθ)

This general form demonstrates that the identity is independent of the underlying distribution of the data X: whether it is a continuous random variable or a discrete random variable, the relationship between the expected size of the set and the probability of coverage still holds. In other words, the choice of the measure μ is application-specific:

  • If μ is the Lebesgue measure, we obtain the standard formula for the expected volume of an interval in n.
  • If μ is the counting measure, we obtain the formula for the expected number of points in a discrete set.

This mathematical framework ensures the identity is a robust tool across diverse statistical models.[2]

Significance and applications

The identity is significant because it connects estimation theory with hypothesis testing. Since a confidence set is often constructed by inverting a family of hypothesis tests, the identity shows that:

  • to minimize the expected length of a confidence interval, one must minimize the probability of covering false values, and
  • minimizing the probability of covering false values is equivalent to maximizing the statistical power of the underlying hypothesis test.

By applying the Neyman-Pearson lemma, which identifies uniformly most powerful tests, statisticians can use the Ghosh-Pratt identity to construct confidence intervals that are mathematically guaranteed to be the shortest possible on average.[3]

History

The identity was published nearly simultaneously in 1961. J.K. Ghosh published his findings in the Calcutta Statistical Association Bulletin, focusing on the relationship between interval types, while John Pratt published his in the Journal of the American Statistical Association focusing on the decision-theoretic implications. While the two approached the problem from slightly different perspectives, their results were mathematically equivalent.[2]

See also

References