Score test
{{Short description|Statistical test based on the gradient of the likelihood function}}
{{distinguish|Test score}}
In statistics, the score test assesses constraints on statistical parameters based on the gradient of the likelihood function—known as the score—evaluated at the hypothesized parameter value under the null hypothesis. Intuitively, if the restricted estimator is near the maximum of the likelihood function, the score should not differ from zero by more than sampling error. While the finite sample distributions of score tests are generally unknown, they have an asymptotic χ2-distribution under the null hypothesis as first proved by C. R. Rao in 1948,{{cite journal |first=C. Radhakrishna |last=Rao |title=Large sample tests of statistical hypotheses concerning several parameters with applications to problems of estimation |journal=Mathematical Proceedings of the Cambridge Philosophical Society |volume=44 |issue=1 |year=1948 |pages=50–57 |doi=10.1017/S0305004100023987 |bibcode=1948PCPS...44...50R }} a fact that can be used to determine statistical significance.
Since function maximization subject to equality constraints is most conveniently done using a Lagrangean expression of the problem, the score test can be equivalently understood as a test of the magnitude of the Lagrange multipliers associated with the constraints where, again, if the constraints are non-binding at the maximum likelihood, the vector of Lagrange multipliers should not differ from zero by more than sampling error. The equivalence of these two approaches was first shown by S. D. Silvey in 1959,{{cite journal |first=S. D. |last=Silvey |title=The Lagrangian Multiplier Test |journal=Annals of Mathematical Statistics |volume=30 |issue=2 |year=1959 |pages=389–407 |jstor=2237089 |doi=10.1214/aoms/1177706259|doi-access=free }} which led to the name Lagrange Multiplier (LM) test that has become more commonly used, particularly in econometrics, since Breusch and Pagan's much-cited 1980 paper.{{cite journal |first1=T. S. |last1=Breusch |author-link=Trevor S. Breusch |first2=A. R. |last2=Pagan |author-link2=Adrian Pagan |title=The Lagrange Multiplier Test and its Applications to Model Specification in Econometrics |journal=Review of Economic Studies |volume=47 |issue=1 |year=1980 |pages=239–253 |doi=10.2307/2297111 |jstor=2297111 }}
The main advantage of the score test over the Wald test and likelihood-ratio test is that the score test only requires the computation of the restricted estimator.{{cite book |first1=Ludwig |last1=Fahrmeir |first2=Thomas |last2=Kneib |first3=Stefan |last3=Lang |first4=Brian |last4=Marx |title=Regression : Models, Methods and Applications |url=https://archive.org/details/regressionmodels00fahr |url-access=limited |location=Berlin |publisher=Springer |year=2013 |isbn=978-3-642-34332-2 |pages=[https://archive.org/details/regressionmodels00fahr/page/n677 663]–664 }} This makes testing feasible when the unconstrained maximum likelihood estimate is a boundary point in the parameter space.{{cn|date=March 2019}} Further, because the score test only requires the estimation of the likelihood function under the null hypothesis, it is less specific than the likelihood ratio test about the alternative hypothesis.{{cite book |first=Peter |last=Kennedy |title=A Guide to Econometrics |location=Cambridge |publisher=MIT Press |edition=Fourth |year=1998 |isbn=0-262-11235-3 |page=68 }}
Single-parameter test
=The statistic=
Let be the likelihood function which depends on a univariate parameter and let be the data. The score is defined as
:
U(\theta)=\frac{\partial \log L(\theta \mid x)}{\partial \theta}.
The Fisher information isLehmann and Casella, eq. (2.5.16).
:
I(\theta) = - \operatorname{E} \left[\left. \frac{\partial^2}{\partial\theta^2} \log f(X;\theta)\,\right|\,\theta \right]\,,
where ƒ is the probability density.
The statistic to test is
S(\theta_0) = \frac{U(\theta_0)^2}{I(\theta_0)}
which has an asymptotic distribution of , when is true. While asymptotically identical, calculating the LM statistic using the outer-gradient-product estimator of the Fisher information matrix can lead to bias in small samples.{{cite journal |first1=Russel |last1=Davidson |first2=James G. |last2=MacKinnon |title=Small sample properties of alternative forms of the Lagrange Multiplier test |journal=Economics Letters |volume=12 |issue=3–4 |year=1983 |pages=269–275 |doi=10.1016/0165-1765(83)90048-4 }}
==Note on notation==
Note that some texts use an alternative notation, in which the statistic is tested against a normal distribution. This approach is equivalent and gives identical results.
=As most powerful test for small deviations=
:
\left(\frac{\partial \log L(\theta \mid x)}{\partial \theta}\right)_{\theta=\theta_0} \geq C
where is the likelihood function, is the value of the parameter of interest under the null hypothesis, and is a constant set depending on the size of the test desired (i.e. the probability of rejecting if is true; see Type I error).
The score test is the most powerful test for small deviations from . To see this, consider testing versus . By the Neyman–Pearson lemma, the most powerful test has the form
:
\frac{L(\theta_0+h\mid x)}{L(\theta_0\mid x)} \geq K;
Taking the log of both sides yields
:
\log L(\theta_0 + h \mid x ) - \log L(\theta_0\mid x) \geq \log K.
The score test follows making the substitution (by Taylor series expansion)
:
\log L(\theta_0+h\mid x) \approx \log L(\theta_0\mid x) + h\times \left(\frac{\partial \log L(\theta \mid x)}{\partial \theta}\right)_{\theta=\theta_0}
and identifying the above with .
=Relationship with other hypothesis tests=
If the null hypothesis is true, the likelihood ratio test, the Wald test, and the Score test are asymptotically equivalent tests of hypotheses.{{cite book |title=Handbook of Econometrics |last=Engle |first=Robert F. |editor=Intriligator, M. D. |editor2=Griliches, Z. |publisher=Elsevier |year=1983 |volume=II |pages=796–801 |chapter=Wald, Likelihood Ratio, and Lagrange Multiplier Tests in Econometrics |isbn=978-0-444-86185-6 }}{{cite book|last1=Burzykowski|first1=Andrzej Gałecki, Tomasz|title=Linear mixed-effects models using R : a step-by-step approach|date=2013|publisher=Springer|location=New York, NY|isbn=978-1-4614-3899-1}} When testing nested models, the statistics for each test then converge to a Chi-squared distribution with degrees of freedom equal to the difference in degrees of freedom in the two models. If the null hypothesis is not true, however, the statistics converge to a noncentral chi-squared distribution with possibly different noncentrality parameters.
Multiple parameters
A more general score test can be derived when there is more than one parameter. Suppose that is the maximum likelihood estimate of under the null hypothesis while and are respectively, the score vector and the Fisher information matrix. Then
:
U^T(\widehat{\theta}_0) I^{-1}(\widehat{\theta}_0) U(\widehat{\theta}_0) \sim \chi^2_k
asymptotically under , where is the number of constraints imposed by the null hypothesis and
:
U(\widehat{\theta}_0) = \frac{\partial \log L(\widehat{\theta}_0 \mid x)}{\partial \theta}
and
:
I(\widehat{\theta}_0) = -\operatorname E\left(\frac{\partial^2 \log L(\widehat{\theta}_0 \mid x)}{\partial \theta \, \partial \theta'} \right).
This can be used to test .
The actual formula for the test statistic depends on which estimator of the Fisher information matrix is being used.{{cite web|last1=Taboga|first1=Marco|title=Lectures on Probability Theory and Mathematical Statistics|url=https://www.statlect.com/fundamentals-of-statistics/score-test|website=statlect.com|access-date=31 May 2022}}
Special cases
In many situations, the score statistic reduces to another commonly used statistic.{{cite book |editor-last=Cook |editor-first=T. D. |editor2-last=DeMets |editor2-first=D. L. |year=2007 |title=Introduction to Statistical Methods for Clinical Trials |publisher=Chapman and Hall |isbn=978-1-58488-027-1 |pages=296–297 }}
In linear regression, the Lagrange multiplier test can be expressed as a function of the F-test.{{cite journal |first=Walter |last=Vandaele |title=Wald, likelihood ratio, and Lagrange multiplier tests as an F test |journal=Economics Letters |year=1981 |volume=8 |issue=4 |pages=361–365 |doi=10.1016/0165-1765(81)90026-4 }}
When the data follows a normal distribution, the score statistic is the same as the t statistic.{{clarify|reason=this can't always be true ... eg when null hypothesis is on the variance|date=March 2011}}
When the data consists of binary observations, the score statistic is the same as the chi-squared statistic in the Pearson's chi-squared test.
See also
References
{{Reflist}}
Further reading
- {{cite journal |first=A. |last=Buse |title=The Likelihood Ratio, Wald, and Lagrange Multiplier Tests: An Expository Note |journal=The American Statistician |volume=36 |year=1982 |issue=3a |pages=153–157 |doi=10.1080/00031305.1982.10482817 }}
- {{cite book |first=L. G. |last=Godfrey |author-link=Leslie G. Godfrey |chapter=The Lagrange Multiplier Test and Testing for Misspecification : An Extended Analysis |title=Misspecification Tests in Econometrics |location=New York |publisher=Cambridge University Press |year=1988 |pages=69–99 |isbn=0-521-26616-5 }}
- {{cite book |first1=Jun |last1=Ma |first2=Charles R. |last2=Nelson |chapter=The superiority of the LM test in a class of econometric models where the Wald test performs poorly |title=Unobserved Components and Time Series Econometrics |location= |publisher=Oxford University Press |year=2016 |isbn=978-0-19-968366-6 |pages=310–330 |doi=10.1093/acprof:oso/9780199683666.003.0014 }}
- {{cite book |first=C. R. |last=Rao |chapter=Score Test: Historical Review and Recent Developments |title=Advances in Ranking and Selection, Multiple Comparisons, and Reliability |pages=3–20 |publisher=Birkhäuser |location=Boston |year=2005 |isbn=978-0-8176-3232-8 }}
{{statistics|inference|collapsed}}
{{DEFAULTSORT:Score Test}}