Chapman–Robbins bound
In statistics, the Chapman–Robbins bound or Hammersley–Chapman–Robbins bound is a lower bound on the variance of estimators of a deterministic parameter. It is a generalization of the Cramér–Rao bound; compared to the Cramér–Rao bound, it is both tighter and applicable to a wider range of problems. However, it is usually more difficult to compute.
The bound was independently discovered by John Hammersley in 1950,{{Citation
| last = Hammersley | first = J. M. |authorlink=John Hammersley
| title = On estimating restricted parameters
| journal = Journal of the Royal Statistical Society, Series B
| volume = 12 | issue = 2 | pages = 192–240 | year = 1950
| doi = 10.1111/j.2517-6161.1950.tb00056.x | mr = 40631
| jstor = 2983981
}} and by Douglas Chapman and Herbert Robbins in 1951.{{Citation
| last1 = Chapman | first1 = D. G.
| last2 = Robbins | first2 = H. | author2-link = Herbert Robbins
| title = Minimum variance estimation without regularity assumptions
| journal = Annals of Mathematical Statistics
| volume = 22 | issue =4 | pages =581–586 | year =1951
| doi = 10.1214/aoms/1177729548
| mr = 44084
| jstor = 2236927
| doi-access = free}}
Statement
Let be the set of parameters for a family of probability distributions on .
For any two , let be the -divergence from to . Then:
{{Math theorem
| math_statement = Given any scalar random variable , and any two , we have
.
}}
A generalization to the multivariable case is:
{{Math theorem
| math_statement = Given any multivariate random variable , and any ,
(E_{\theta'}[\hat g] - E_{\theta}[\hat g])^T \operatorname{Cov}_\theta[\hat g]^{-1} (E_{\theta'}[\hat g] - E_{\theta}[\hat g])
}}
Proof
By the variational representation of chi-squared divergence:{{Cite web |last=Polyanskiy |first=Yury |date=2017 |title=Lecture notes on information theory, chapter 29, ECE563 (UIUC) |url=https://people.lids.mit.edu/yp/homepage/data/LN_stats.pdf |url-status=live |archive-url=https://web.archive.org/web/20220524014051/https://people.lids.mit.edu/yp/homepage/data/LN_stats.pdf |archive-date=2022-05-24 |access-date=2022-05-24 |website=Lecture notes on information theory}}
Plug in , to obtain: Switch the denominator and the left side and take supremum over to obtain the single-variate case. For the multivariate case, we define for any . Then plug in in the variational representation to obtain: Take supremum over , using the linear algebra fact that , we obtain the multivariate case.
Relation to Cramér–Rao bound
Usually, is the sample space of independent draws of a -valued random variable with distribution from a by parameterized family of probability distributions, is its -fold product measure, and is an estimator of . Then, for , the expression inside the supremum in the Chapman–Robbins bound converges to the Cramér–Rao bound of when , assuming the regularity conditions of the Cramér–Rao bound hold. This implies that, when both bounds exist, the Chapman–Robbins version is always at least as tight as the Cramér–Rao bound; in many cases, it is substantially tighter.
The Chapman–Robbins bound also holds under much weaker regularity conditions. For example, no assumption is made regarding differentiability of the probability density function p(x; θ) of . When p(x; θ) is non-differentiable, the Fisher information is not defined, and hence the Cramér–Rao bound does not exist.
See also
References
{{reflist}}
Further reading
- {{Citation
| last1 = Lehmann
| first1 = E. L.
| last2= Casella |first2=G.
| title = Theory of Point Estimation
| year = 1998
| publisher = Springer
| isbn = 0-387-98502-6
| edition = 2nd
| pages = 113–114 }}
{{DEFAULTSORT:Chapman-Robbins bound}}