Cramér–Rao bound
In estimation theory and statistics, the Cramér–Rao bound (CRB) relates to estimation of a deterministic (fixed, though unknown) parameter. The result is named in honor of Harald Cramér and C. R. Rao,[1][2][3] but has also been derived independently by Maurice Fréchet,[4] Georges Darmois,[5] and by Alexander Aitken and Harold Silverstone.[6][7] It states that the precision of any unbiased estimator is at most the Fisher information; or (equivalently) the reciprocal of the Fisher information is a lower bound on its variance.
An unbiased estimator that achieves this bound is said to be (fully) efficient. Such a solution achieves the lowest possible mean squared error among all unbiased methods, and is therefore the minimum variance unbiased (MVU) estimator. However, in some cases, no unbiased technique exists which achieves the bound. This may occur either if for any unbiased estimator, there exists another with a strictly smaller variance, or if an MVU estimator exists, but its variance is strictly greater than the inverse of the Fisher information.
The Cramér–Rao bound can also be used to bound the variance of biased estimators of given bias. In some cases, a biased approach can result in both a variance and a mean squared error that are below the unbiased Cramér–Rao lower bound; see estimator bias.
Statement
The Cramér–Rao bound is stated in this section for several increasingly general cases, beginning with the case in which the parameter is a scalar and its estimator is unbiased. All versions of the bound require certain regularity conditions, which hold for most well-behaved distributions. These conditions are listed later in this section.
Scalar unbiased case
Suppose [math]\displaystyle{ \theta }[/math] is an unknown deterministic parameter that is to be estimated from [math]\displaystyle{ n }[/math] independent observations (measurements) of [math]\displaystyle{ x }[/math], each from a distribution according to some probability density function [math]\displaystyle{ f(x;\theta) }[/math]. The variance of any unbiased estimator [math]\displaystyle{ \hat{\theta} }[/math] of [math]\displaystyle{ \theta }[/math] is then bounded[8] by the reciprocal of the Fisher information [math]\displaystyle{ I(\theta) }[/math]:
- [math]\displaystyle{ \operatorname{var}(\hat{\theta}) \geq \frac{1}{I(\theta)} }[/math]
where the Fisher information [math]\displaystyle{ I(\theta) }[/math] is defined by
- [math]\displaystyle{ I(\theta) = n \operatorname{E}_{X;\theta} \left[ \left( \frac{\partial \ell(X;\theta)}{\partial\theta} \right)^2 \right] }[/math]
and [math]\displaystyle{ \ell(x;\theta)=\log (f(x;\theta)) }[/math] is the natural logarithm of the likelihood function for a single sample [math]\displaystyle{ x }[/math] and [math]\displaystyle{ \operatorname{E}_{x;\theta} }[/math] denotes the expected value with respect to the density [math]\displaystyle{ f(x;\theta) }[/math] of [math]\displaystyle{ X }[/math]. If not indicated, in what follows, the expectation is taken with respect to [math]\displaystyle{ X }[/math].
If [math]\displaystyle{ \ell(x;\theta) }[/math] is twice differentiable and certain regularity conditions hold, then the Fisher information can also be defined as follows:[9]
- [math]\displaystyle{ I(\theta) = -n \operatorname{E}_{X;\theta}\left[ \frac{\partial^2 \ell(X;\theta)}{\partial\theta^2} \right] }[/math]
The efficiency of an unbiased estimator [math]\displaystyle{ \hat{\theta} }[/math] measures how close this estimator's variance comes to this lower bound; estimator efficiency is defined as
- [math]\displaystyle{ e(\hat{\theta}) = \frac{I(\theta)^{-1}}{\operatorname{var}(\hat{\theta})} }[/math]
or the minimum possible variance for an unbiased estimator divided by its actual variance. The Cramér–Rao lower bound thus gives
- [math]\displaystyle{ e(\hat{\theta}) \le 1 }[/math].
General scalar case
A more general form of the bound can be obtained by considering a biased estimator [math]\displaystyle{ T(X) }[/math], whose expectation is not [math]\displaystyle{ \theta }[/math] but a function of this parameter, say, [math]\displaystyle{ \psi(\theta) }[/math]. Hence [math]\displaystyle{ E\{T(X)\} - \theta = \psi(\theta) - \theta }[/math] is not generally equal to 0. In this case, the bound is given by
- [math]\displaystyle{ \operatorname{var}(T) \geq \frac{[\psi'(\theta)]^2}{I(\theta)} }[/math]
where [math]\displaystyle{ \psi'(\theta) }[/math] is the derivative of [math]\displaystyle{ \psi(\theta) }[/math] (by [math]\displaystyle{ \theta }[/math]), and [math]\displaystyle{ I(\theta) }[/math] is the Fisher information defined above.
Bound on the variance of biased estimators
Apart from being a bound on estimators of functions of the parameter, this approach can be used to derive a bound on the variance of biased estimators with a given bias, as follows.[10] Consider an estimator [math]\displaystyle{ \hat{\theta} }[/math] with bias [math]\displaystyle{ b(\theta) = E\{\hat{\theta}\} - \theta }[/math], and let [math]\displaystyle{ \psi(\theta) = b(\theta) + \theta }[/math]. By the result above, any unbiased estimator whose expectation is [math]\displaystyle{ \psi(\theta) }[/math] has variance greater than or equal to [math]\displaystyle{ (\psi'(\theta))^2/I(\theta) }[/math]. Thus, any estimator [math]\displaystyle{ \hat{\theta} }[/math] whose bias is given by a function [math]\displaystyle{ b(\theta) }[/math] satisfies[11]
- [math]\displaystyle{ \operatorname{var} \left(\hat{\theta}\right) \geq \frac{[1+b'(\theta)]^2}{I(\theta)}. }[/math]
The unbiased version of the bound is a special case of this result, with [math]\displaystyle{ b(\theta)=0 }[/math].
It's trivial to have a small variance − an "estimator" that is constant has a variance of zero. But from the above equation we find that the mean squared error of a biased estimator is bounded by
- [math]\displaystyle{ \operatorname{E}\left((\hat{\theta}-\theta)^2\right)\geq\frac{[1+b'(\theta)]^2}{I(\theta)}+b(\theta)^2, }[/math]
using the standard decomposition of the MSE. Note, however, that if [math]\displaystyle{ 1+b'(\theta)\lt 1 }[/math] this bound might be less than the unbiased Cramér–Rao bound [math]\displaystyle{ 1/I(\theta) }[/math]. For instance, in the example of estimating variance below, [math]\displaystyle{ 1+b'(\theta)= \frac{n}{n+2} \lt 1 }[/math].
Multivariate case
Extending the Cramér–Rao bound to multiple parameters, define a parameter column vector
- [math]\displaystyle{ \boldsymbol{\theta} = \left[ \theta_1, \theta_2, \dots, \theta_d \right]^T \in \mathbb{R}^d }[/math]
with probability density function [math]\displaystyle{ f(x; \boldsymbol{\theta}) }[/math] which satisfies the two regularity conditions below.
The Fisher information matrix is a [math]\displaystyle{ d \times d }[/math] matrix with element [math]\displaystyle{ I_{m, k} }[/math] defined as
- [math]\displaystyle{ I_{m, k} = \operatorname{E} \left[ \frac{\partial }{\partial \theta_m} \log f\left(x; \boldsymbol{\theta}\right) \frac{\partial }{\partial \theta_k} \log f\left(x; \boldsymbol{\theta}\right) \right] = -\operatorname{E} \left[ \frac{\partial ^2}{\partial \theta_m \, \partial \theta_k} \log f\left(x; \boldsymbol{\theta}\right) \right]. }[/math]
Let [math]\displaystyle{ \boldsymbol{T}(X) }[/math] be an estimator of any vector function of parameters, [math]\displaystyle{ \boldsymbol{T}(X) = (T_1(X), \ldots, T_d(X))^T }[/math], and denote its expectation vector [math]\displaystyle{ \operatorname{E}[\boldsymbol{T}(X)] }[/math] by [math]\displaystyle{ \boldsymbol{\psi}(\boldsymbol{\theta}) }[/math]. The Cramér–Rao bound then states that the covariance matrix of [math]\displaystyle{ \boldsymbol{T}(X) }[/math] satisfies
- [math]\displaystyle{ I\left(\boldsymbol{\theta}\right) \geq \phi(\theta)^T \operatorname{cov}_{\boldsymbol{\theta}}\left(\boldsymbol{T}(X)\right)^{-1}\phi(\theta) }[/math],
- [math]\displaystyle{ \operatorname{cov}_{\boldsymbol{\theta}}\left(\boldsymbol{T}(X)\right) \geq \phi(\theta) I\left(\boldsymbol{\theta}\right)^{-1} \phi(\theta)^T }[/math]
where
- The matrix inequality [math]\displaystyle{ A \ge B }[/math] is understood to mean that the matrix [math]\displaystyle{ A-B }[/math] is positive semidefinite, and
- [math]\displaystyle{ \phi(\theta) := \partial \boldsymbol{\psi}(\boldsymbol{\theta})/\partial \boldsymbol{\theta} }[/math] is the Jacobian matrix whose [math]\displaystyle{ ij }[/math] element is given by [math]\displaystyle{ \partial \psi_i(\boldsymbol{\theta})/\partial \theta_j }[/math].
If [math]\displaystyle{ \boldsymbol{T}(X) }[/math] is an unbiased estimator of [math]\displaystyle{ \boldsymbol{\theta} }[/math] (i.e., [math]\displaystyle{ \boldsymbol{\psi}\left(\boldsymbol{\theta}\right) = \boldsymbol{\theta} }[/math]), then the Cramér–Rao bound reduces to
- [math]\displaystyle{ \operatorname{cov}_{\boldsymbol{\theta}}\left(\boldsymbol{T}(X)\right) \geq I\left(\boldsymbol{\theta}\right)^{-1}. }[/math]
If it is inconvenient to compute the inverse of the Fisher information matrix, then one can simply take the reciprocal of the corresponding diagonal element to find a (possibly loose) lower bound.[12]
- [math]\displaystyle{ \operatorname{var}_{\boldsymbol{\theta}}(T_m(X)) = \left[\operatorname{cov}_{\boldsymbol{\theta}}\left(\boldsymbol{T}(X)\right)\right]_{mm} \geq \left[I\left(\boldsymbol{\theta}\right)^{-1}\right]_{mm} \geq \left(\left[I\left(\boldsymbol{\theta}\right)\right]_{mm}\right)^{-1}. }[/math]
Regularity conditions
The bound relies on two weak regularity conditions on the probability density function, [math]\displaystyle{ f(x; \theta) }[/math], and the estimator [math]\displaystyle{ T(X) }[/math]:
- The Fisher information is always defined; equivalently, for all [math]\displaystyle{ x }[/math] such that [math]\displaystyle{ f(x; \theta) \gt 0 }[/math], [math]\displaystyle{ \frac{\partial}{\partial\theta} \log f(x;\theta) }[/math] exists, and is finite.
- The operations of integration with respect to [math]\displaystyle{ x }[/math] and differentiation with respect to [math]\displaystyle{ \theta }[/math] can be interchanged in the expectation of [math]\displaystyle{ T }[/math]; that is, [math]\displaystyle{
\frac{\partial}{\partial\theta}
\left[
\int T(x) f(x;\theta) \,dx
\right]
=
\int T(x)
\left[
\frac{\partial}{\partial\theta} f(x;\theta)
\right]
\,dx
}[/math] whenever the right-hand side is finite. This condition can often be confirmed by using the fact that integration and differentiation can be swapped when either of the following cases hold:
- The function [math]\displaystyle{ f(x;\theta) }[/math] has bounded support in [math]\displaystyle{ x }[/math], and the bounds do not depend on [math]\displaystyle{ \theta }[/math];
- The function [math]\displaystyle{ f(x;\theta) }[/math] has infinite support, is continuously differentiable, and the integral converges uniformly for all [math]\displaystyle{ \theta }[/math].
Proof
Proof for the general case based on the Chapman–Robbins bound
Proof based on.[13]
First equation:
Let [math]\displaystyle{ \delta }[/math] be an infinitesimal, then for any [math]\displaystyle{ v\in \R^n }[/math], plugging [math]\displaystyle{ \theta' = \theta + \delta v }[/math] in, we have [math]\displaystyle{ (E_{\theta'}[T] - E_{\theta}[T]) = v^T \phi(\theta)\delta; \quad \chi^2(\mu_{\theta'} ; \mu_\theta) = v^T I(\theta) v \delta^2 }[/math]
Plugging this into multivariate Chapman–Robbins bound gives [math]\displaystyle{ I(\theta) \geq \phi(\theta) \operatorname{Cov}_\theta[T]^{-1} \phi(\theta)^T }[/math].
Second equation:
It suffices to prove this for scalar case, with [math]\displaystyle{ h(X) }[/math] taking values in [math]\displaystyle{ \R }[/math] . Because for general [math]\displaystyle{ T(X) }[/math] , we can take any [math]\displaystyle{ v \in \R^m }[/math], then defining [math]\displaystyle{ h:= \sum_j v_j T_j }[/math], the scalar case gives [math]\displaystyle{ \operatorname{Var}_\theta[h] = v^T \operatorname{Cov}_\theta[T]v \geq v^T \phi(\theta) I(\theta)^{-1}\phi(\theta)^T v }[/math]This holds for all [math]\displaystyle{ v \in \R^m }[/math], so we can conclude[math]\displaystyle{ \operatorname{Cov}_\theta[T] \geq \phi(\theta) I(\theta)^{-1}\phi(\theta)^T }[/math]The scalar case states that [math]\displaystyle{ \operatorname{Var}_\theta[h] \geq \phi(\theta)^T I(\theta)^{-1} \phi(\theta) }[/math] with [math]\displaystyle{ \phi(\theta) := \nabla_\theta E_\theta[h] }[/math].
Let [math]\displaystyle{ \delta }[/math] be an infinitesimal, then for any [math]\displaystyle{ v\in \R^n }[/math], taking [math]\displaystyle{ \theta' = \theta + \delta v }[/math] in the single-variate Chapman–Robbins bound gives [math]\displaystyle{ \operatorname{Var}_\theta[h] \geq \frac{\langle v, \phi(\theta)\rangle^2}{v^T I(\theta) v} }[/math].
By linear algebra, [math]\displaystyle{ \sup_{v\neq 0} \frac{\langle w, v\rangle^2}{v^T M v} = w^T M^{-1}w }[/math] for any positive-definite matrix [math]\displaystyle{ M }[/math], thus we obtain [math]\displaystyle{ \operatorname{Var}_\theta[h] \geq \phi(\theta)^T I(\theta)^{-1} \phi(\theta). }[/math]
A standalone proof for the general scalar case
For the general scalar case:
Assume that [math]\displaystyle{ T=t(X) }[/math] is an estimator with expectation [math]\displaystyle{ \psi(\theta) }[/math] (based on the observations [math]\displaystyle{ X }[/math]), i.e. that [math]\displaystyle{ \operatorname{E}(T) = \psi (\theta) }[/math]. The goal is to prove that, for all [math]\displaystyle{ \theta }[/math],
- [math]\displaystyle{ \operatorname{var}(t(X)) \geq \frac{[\psi^\prime(\theta)]^2}{I(\theta)}. }[/math]
Let [math]\displaystyle{ X }[/math] be a random variable with probability density function [math]\displaystyle{ f(x; \theta) }[/math]. Here [math]\displaystyle{ T = t(X) }[/math] is a statistic, which is used as an estimator for [math]\displaystyle{ \psi (\theta) }[/math]. Define [math]\displaystyle{ V }[/math] as the score:
- [math]\displaystyle{ V = \frac{\partial}{\partial\theta} \ln f(X;\theta) = \frac{1}{f(X;\theta)}\frac{\partial}{\partial\theta}f(X;\theta) }[/math]
where the chain rule is used in the final equality above. Then the expectation of [math]\displaystyle{ V }[/math], written [math]\displaystyle{ \operatorname{E}(V) }[/math], is zero. This is because:
- [math]\displaystyle{ \operatorname{E}(V) = \int f(x;\theta)\left[\frac{1}{f(x;\theta)}\frac{\partial }{\partial \theta} f(x;\theta)\right] \, dx = \frac{\partial}{\partial\theta}\int f(x;\theta) \, dx = 0 }[/math]
where the integral and partial derivative have been interchanged (justified by the second regularity condition).
If we consider the covariance [math]\displaystyle{ \operatorname{cov}(V, T) }[/math] of [math]\displaystyle{ V }[/math] and [math]\displaystyle{ T }[/math], we have [math]\displaystyle{ \operatorname{cov}(V, T) = \operatorname{E}(V T) }[/math], because [math]\displaystyle{ \operatorname{E}(V) = 0 }[/math]. Expanding this expression we have
- [math]\displaystyle{ \begin{align} \operatorname{cov}(V,T) & = \operatorname{E} \left( T \cdot\left[\frac{1}{f(X;\theta)}\frac{\partial}{\partial\theta}f(X;\theta) \right] \right) \\[6pt] & = \int t(x) \left[\frac{1}{f(x;\theta)} \frac{\partial}{\partial\theta} f(x;\theta) \right] f(x;\theta)\, dx \\[6pt] & = \frac{\partial}{\partial\theta} \left[ \int t(x) f(x;\theta)\,dx \right] = \frac{\partial}{\partial\theta} E(T) = \psi^\prime(\theta) \end{align} }[/math]
again because the integration and differentiation operations commute (second condition).
The Cauchy–Schwarz inequality shows that
- [math]\displaystyle{ \sqrt{ \operatorname{var} (T) \operatorname{var} (V)} \geq \left| \operatorname{cov}(V,T) \right| = \left | \psi^\prime (\theta) \right | }[/math]
therefore
- [math]\displaystyle{ \operatorname{var} (T) \geq \frac{[\psi^\prime(\theta)]^2}{\operatorname{var} (V)} = \frac{[\psi^\prime(\theta)]^2}{I(\theta)} }[/math]
which proves the proposition.
Examples
Multivariate normal distribution
For the case of a d-variate normal distribution
- [math]\displaystyle{ \boldsymbol{x} \sim \mathcal{N}_d \left( \boldsymbol{\mu}( \boldsymbol{\theta}) , {\boldsymbol C} ( \boldsymbol{\theta}) \right) }[/math]
the Fisher information matrix has elements[14]
- [math]\displaystyle{ I_{m, k} = \frac{\partial \boldsymbol{\mu}^T}{\partial \theta_m} {\boldsymbol C}^{-1} \frac{\partial \boldsymbol{\mu}}{\partial \theta_k} + \frac{1}{2} \operatorname{tr} \left( {\boldsymbol C}^{-1} \frac{\partial {\boldsymbol C}}{\partial \theta_m} {\boldsymbol C}^{-1} \frac{\partial {\boldsymbol C}}{\partial \theta_k} \right) }[/math]
where "tr" is the trace.
For example, let [math]\displaystyle{ w[j] }[/math] be a sample of [math]\displaystyle{ n }[/math] independent observations with unknown mean [math]\displaystyle{ \theta }[/math] and known variance [math]\displaystyle{ \sigma^2 }[/math] .
- [math]\displaystyle{ w[j] \sim \mathcal{N}_{d, n} \left(\theta {\boldsymbol 1}, \sigma^2 {\boldsymbol I} \right). }[/math]
Then the Fisher information is a scalar given by
- [math]\displaystyle{ I(\theta) = \left(\frac{\partial\boldsymbol{\mu}(\theta)}{\partial\theta}\right)^T{\boldsymbol C}^{-1} \left(\frac{\partial\boldsymbol{\mu}(\theta)}{\partial\theta}\right) = \sum^{n}_{i=1} \frac{1}{\sigma^2} = \frac{n}{\sigma^2}, }[/math]
and so the Cramér–Rao bound is
- [math]\displaystyle{ \operatorname{var}\left(\hat \theta\right) \geq \frac{\sigma^2}{n}. }[/math]
Normal variance with known mean
Suppose X is a normally distributed random variable with known mean [math]\displaystyle{ \mu }[/math] and unknown variance [math]\displaystyle{ \sigma^2 }[/math]. Consider the following statistic:
- [math]\displaystyle{ T=\frac{\sum_{i=1}^n (X_i-\mu)^2}{n}. }[/math]
Then T is unbiased for [math]\displaystyle{ \sigma^2 }[/math], as [math]\displaystyle{ E(T)=\sigma^2 }[/math]. What is the variance of T?
- [math]\displaystyle{ \operatorname{var}(T) = \operatorname{var}\left(\frac{\sum_{i=1}^n(X_i-\mu)^2}{n}\right) = \frac{\sum_{i=1}^n\operatorname{var}(X_i-\mu)^2}{n^2} = \frac{n\operatorname{var}(X-\mu)^2}{n^2}=\frac{1}{n} \left[ \operatorname{E}\left\{(X-\mu)^4\right\}-\left(\operatorname{E}\{(X-\mu)^2\}\right)^2 \right] }[/math]
(the second equality follows directly from the definition of variance). The first term is the fourth moment about the mean and has value [math]\displaystyle{ 3(\sigma^2)^2 }[/math]; the second is the square of the variance, or [math]\displaystyle{ (\sigma^2)^2 }[/math]. Thus
- [math]\displaystyle{ \operatorname{var}(T)=\frac{2(\sigma^2)^2}{n}. }[/math]
Now, what is the Fisher information in the sample? Recall that the score [math]\displaystyle{ V }[/math] is defined as
- [math]\displaystyle{ V=\frac{\partial}{\partial\sigma^2}\log\left[ L(\sigma^2,X)\right] }[/math]
where [math]\displaystyle{ L }[/math] is the likelihood function. Thus in this case,
- [math]\displaystyle{ \log\left[L(\sigma^2,X)\right]=\log\left[\frac{1}{\sqrt{2\pi\sigma^2}}e^{-(X-\mu)^2 /{2\sigma^2}}\right] =-\log(\sqrt{2\pi\sigma^2})-\frac{(X-\mu)^2}{2\sigma^2} }[/math]
- [math]\displaystyle{ V=\frac{\partial}{\partial\sigma^2}\log \left[ L(\sigma^2,X) \right]=\frac{\partial}{\partial\sigma^2}\left[-\log(\sqrt{2\pi\sigma^2})-\frac{(X-\mu)^2}{2\sigma^2}\right] =-\frac{1}{2\sigma^2}+\frac{(X-\mu)^2}{2(\sigma^2)^2} }[/math]
where the second equality is from elementary calculus. Thus, the information in a single observation is just minus the expectation of the derivative of [math]\displaystyle{ V }[/math], or
- [math]\displaystyle{ I =-\operatorname{E}\left(\frac{\partial V}{\partial\sigma^2}\right) =-\operatorname{E}\left(-\frac{(X-\mu)^2}{(\sigma^2)^3}+\frac{1}{2(\sigma^2)^2}\right) =\frac{\sigma^2}{(\sigma^2)^3}-\frac{1}{2(\sigma^2)^2} =\frac{1}{2(\sigma^2)^2}. }[/math]
Thus the information in a sample of [math]\displaystyle{ n }[/math] independent observations is just [math]\displaystyle{ n }[/math] times this, or [math]\displaystyle{ \frac{n}{2(\sigma^2)^2}. }[/math]
The Cramér–Rao bound states that
- [math]\displaystyle{ \operatorname{var}(T)\geq\frac{1}{I}. }[/math]
In this case, the inequality is saturated (equality is achieved), showing that the estimator is efficient.
However, we can achieve a lower mean squared error using a biased estimator. The estimator
- [math]\displaystyle{ T=\frac{\sum_{i=1}^n (X_i-\mu)^2}{n+2}. }[/math]
obviously has a smaller variance, which is in fact
- [math]\displaystyle{ \operatorname{var}(T)=\frac{2n(\sigma^2)^2}{(n+2)^2}. }[/math]
Its bias is
- [math]\displaystyle{ \left(1-\frac{n}{n+2}\right)\sigma^2=\frac{2\sigma^2}{n+2} }[/math]
so its mean squared error is
- [math]\displaystyle{ \operatorname{MSE}(T)=\left(\frac{2n}{(n+2)^2}+\frac{4}{(n+2)^2}\right)(\sigma^2)^2 =\frac{2(\sigma^2)^2}{n+2} }[/math]
which is clearly less than what unbiased estimators can achieve according to the Cramér–Rao bound.
When the mean is not known, the minimum mean squared error estimate of the variance of a sample from Gaussian distribution is achieved by dividing by [math]\displaystyle{ n+1 }[/math], rather than [math]\displaystyle{ n-1 }[/math] or [math]\displaystyle{ n+2 }[/math].
See also
References and notes
- ↑ Cramér, Harald (1946). Mathematical Methods of Statistics. Princeton, NJ: Princeton Univ. Press. ISBN 0-691-08004-6. OCLC 185436716. https://archive.org/details/in.ernet.dli.2015.223699.
- ↑ Rao, Calyampudi Radakrishna (1945). "Information and the accuracy attainable in the estimation of statistical parameters". Bulletin of the Calcutta Mathematical Society (Calcutta Mathematical Society) 37: 81–89.
- ↑ Rao, Calyampudi Radakrishna (1994). S. Das Gupta. ed. Selected Papers of C. R. Rao. New York: Wiley. ISBN 978-0-470-22091-7. OCLC 174244259.
- ↑ Fréchet, Maurice (1943). "Sur l'extension de certaines évaluations statistiques au cas de petits échantillons". Rev. Inst. Int. Statist. 11 (3/4): 182–205. doi:10.2307/1401114.
- ↑ Darmois, Georges (1945). "Sur les limites de la dispersion de certaines estimations". Rev. Int. Inst. Statist. 13 (1/4): 9–15. doi:10.2307/1400974.
- ↑ Aitken, A. C.; Silverstone, H. (1942). "XV.—On the Estimation of Statistical Parameters" (in en). Proceedings of the Royal Society of Edinburgh Section A: Mathematics 61 (2): 186–194. doi:10.1017/S008045410000618X. ISSN 2053-5902. https://www.cambridge.org/core/journals/proceedings-of-the-royal-society-of-edinburgh-section-a-mathematics/article/abs/xvon-the-estimation-of-statistical-parameters/74C2B968275EED1A954DB316FBC3635E.
- ↑ Shenton, L. R. (1970). "The so-called Cramer–Rao inequality". The American Statistician 24 (2): 36.
- ↑ Nielsen, Frank (2013). "Cramér-Rao Lower Bound and Information Geometry". Connected at Infinity II. Texts and Readings in Mathematics. 67. Hindustan Book Agency, Gurgaon. p. 18-37. doi:10.1007/978-93-86279-56-9_2. ISBN 978-93-80250-51-9.
- ↑ Suba Rao. "Lectures on statistical inference". http://www.stat.tamu.edu/~suhasini/teaching613/inference.pdf.
- ↑ "Cramér Rao Lower Bound - Navipedia". https://gssc.esa.int/navipedia/index.php/Cram%C3%A9r_Rao_Lower_Bound.
- ↑ "Cramér-Rao Bound". https://theoryandpractice.org/stats-ds-book/statistics/cramer-rao-bound.html.
- ↑ For the Bayesian case, see eqn. (11) of Bobrovsky; Mayer-Wolf; Zakai (1987). "Some classes of global Cramer–Rao bounds". Ann. Stat. 15 (4): 1421–38. doi:10.1214/aos/1176350602.
- ↑ Polyanskiy, Yury (2017). "Lecture notes on information theory, chapter 29, ECE563 (UIUC)". https://people.lids.mit.edu/yp/homepage/data/LN_stats.pdf.
- ↑ Kay, S. M. (1993). Fundamentals of Statistical Signal Processing: Estimation Theory. Prentice Hall. p. 47. ISBN 0-13-042268-1.
Further reading
- Amemiya, Takeshi (1985). Advanced Econometrics. Cambridge: Harvard University Press. pp. 14–17. ISBN 0-674-00560-0. https://archive.org/details/advancedeconomet00amem.
- Bos, Adriaan van den (2007). Parameter Estimation for Scientists and Engineers. Hoboken: John Wiley & Sons. pp. 45–98. ISBN 978-0-470-14781-8.
- Kay, Steven M. (1993). Fundamentals of Statistical Signal Processing, Volume I: Estimation Theory. Prentice Hall. ISBN 0-13-345711-7.. Chapter 3.
- Shao, Jun (1998). Mathematical Statistics. New York: Springer. ISBN 0-387-98674-X.. Section 3.1.3.
- Posterior uncertainty, asymptotic law and Cramér-Rao bound, Structural Control and Health Monitoring 25(1851):e2113 DOI: 10.1002/stc.2113
External links
- FandPLimitTool a GUI-based software to calculate the Fisher information and Cramér-Rao lower bound with application to single-molecule microscopy.
Original source: https://en.wikipedia.org/wiki/Cramér–Rao bound.
Read more |