Conway–Maxwell–Poisson distribution
{{Short description|Probability distribution}}
{{Infobox probability distribution
| name = Conway–Maxwell–Poisson
| type = mass
| pdf_image = File:CMP PMF.png
| cdf_image = File:CMP CDF.png
| parameters =
| support =
| pdf =
| cdf =
| mean =
| median = No closed form
| mode = See text
| variance =
| skewness = Not listed
| kurtosis = Not listed
| entropy = Not listed
| mgf =
| char =
|pgf=}}
In probability theory and statistics, the Conway–Maxwell–Poisson (CMP or COM–Poisson) distribution is a discrete probability distribution named after Richard W. Conway, William L. Maxwell, and Siméon Denis Poisson that generalizes the Poisson distribution by adding a parameter to model overdispersion and underdispersion. It is a member of the exponential family,{{cite web|title=Conway–Maxwell–Poisson Regression|url=https://support.sas.com/documentation/cdl/en/etsug/66840/HTML/default/viewer.htm#etsug_countreg_details06.htm |work=SAS Support |publisher=SAS Institute, Inc.|access-date=2 March 2015}} has the Poisson distribution and geometric distribution as special cases and the Bernoulli distribution as a limiting case.
Background
The CMP distribution was originally proposed by Conway and Maxwell in 1962{{Citation|last1=Conway| first1=R. W.| last2=Maxwell| first2= W. L.| year=1962| title=A queuing model with state dependent service rates|journal= Journal of Industrial Engineering| volume=12| pages=132–136}} as a solution to handling queueing systems with state-dependent service rates. The CMP distribution was introduced into the statistics literature by Boatwright et al. 2003 Boatwright, P., Borle, S. and Kadane, J.B. "A model of the joint distribution of purchase quantity and timing." Journal of the American Statistical Association 98 (2003): 564–572. and Shmueli et al. (2005).Shmueli G., Minka T., Kadane J.B., Borle S., and Boatwright, P.B. "A useful distribution for fitting discrete data: revival of the Conway–Maxwell–Poisson distribution." Journal of the Royal Statistical Society: Series C (Applied Statistics) 54.1 (2005): 127–142.[https://dx.doi.org/10.1111/j.1467-9876.2005.00474.x] The first detailed investigation into the
probabilistic and statistical properties of the distribution was published by Shmueli et al. (2005). Some theoretical probability results of COM-Poisson distribution is studied and reviewed by Li et al. (2019),Li B., Zhang H., Jiao H. "Some Characterizations and Properties of COM-Poisson Random Variables." Communications in Statistics - Theory and Methods, (2019).[https://doi.org/10.1080/03610926.2018.1563164] especially the characterizations of COM-Poisson distribution.
Probability mass function and basic properties
The CMP distribution is defined to be the distribution with probability mass function
:
P(X = x) = f(x; \lambda, \nu) = \frac{\lambda^x}{(x!)^\nu}\frac{1}{Z(\lambda,\nu)}.
where :
:
Z(\lambda,\nu) = \sum_{j=0}^\infty \frac{\lambda^j}{(j!)^\nu}.
The function serves as a normalization constant so the probability mass function sums to one. Note that does not have a closed form.
The domain of admissible parameters is , and , .
The additional parameter which does not appear in the Poisson distribution allows for adjustment of the rate of decay. This rate of decay is a non-linear decrease in ratios of successive probabilities, specifically
:
\frac{P(X = x-1)}{P(X = x)} = \frac{x^\nu}{\lambda}.
When , the CMP distribution becomes the standard Poisson distribution and as , the distribution approaches a Bernoulli distribution with parameter . When the CMP distribution reduces to a geometric distribution with probability of success provided .
For the CMP distribution, moments can be found through the recursive formula
:
\operatorname{E}[X^{r+1}] = \begin{cases}
\lambda \, \operatorname{E}[X+1]^{1-\nu} & \text{if } r = 0 \\
\lambda \, \frac{d}{d\lambda}\operatorname{E}[X^r] + \operatorname{E}[X]\operatorname{E}[X^r] & \text{if } r > 0. \\
\end{cases}
Cumulative distribution function
For general , there does not exist a closed form formula for the cumulative distribution function of . If is an integer, we can, however, obtain the following formula in terms of the generalized hypergeometric function:Nadarajah, S. "Useful moment and CDF formulations for the COM–Poisson distribution." Statistical Papers 50 (2009): 617–622.
:
F(n)=P(X\leq n)=1-\frac{ _1F_{\nu-1}(;n+2,\ldots,n+2;\lambda)}{{\{(n+1)!\}^{\nu-1}} _0F_{\nu-1}(;1,\ldots,1;\lambda)}.
The normalizing constant
Many important summary statistics, such as moments and cumulants, of the CMP distribution can be expressed in terms of the normalizing constant . Indeed, The probability generating function is , and the mean and variance are given by
:
\operatorname{E}X=\lambda\frac{d}{d\lambda}\big\{\ln(Z(\lambda,\nu))\big\},
:
\operatorname{var}(X)=\lambda\frac{d}{d\lambda}\operatorname{E}X.
The cumulant generating function is
:
g(t)=\ln(\operatorname{E}[e^{tX}])=\ln(Z(\lambda e^{t},\nu))-\ln(Z(\lambda,\nu)),
and the cumulants are given by
:
\kappa_n=g^{(n)}(0)=\frac{\partial^n}{\partial t^n}\ln(Z(\lambda e^t,\nu)) \bigg|_{t=0}, \quad n\geq1.
Whilst the normalizing constant does not in general have a closed form, there are some noteworthy special cases:
- , where is a modified Bessel function of the first kind.
- For integer , the normalizing constant can expressed as a generalized hypergeometric function: .
Because the normalizing constant does not in general have a closed form, the following asymptotic expansion is of interest. Fix . Then, as ,Gaunt, R.E., Iyengar, S., Olde Daalhuis, A.B. and Simsek, B. "An asymptotic expansion for the normalizing constant of the Conway–Maxwell–Poisson distribution." To appear in Annals of the Institute of Statistical Mathematics (2017+) DOI 10.1007/s10463-017-0629-6
:
Z(\lambda,\nu)=\frac{\exp\left\{\nu\lambda^{1/\nu}\right\}}{\lambda^{(\nu-1)/2\nu}(2\pi)^{(\nu-1)/2}\sqrt{\nu}}\sum_{k=0}^\infty c_k\big(\nu\lambda^{1/\nu}\big)^{-k},
where the are uniquely determined by the expansion
:
\left(\Gamma(t+1)\right)^{-\nu}=\frac{\nu^{\nu (t+1/2)}}{\left(2\pi\right)^{(\nu-1)/2}}\sum_{j=0}^\infty\frac{c_j}{\Gamma(\nu t+(1+\nu)/2+j)}.
In particular, , , . Further coefficients are given in.
Moments for the case of integer <math>\nu</math>
When is an integer explicit formulas for moments can be obtained. The case corresponds to the Poisson distribution. Suppose now that . For ,
:
\operatorname{E}[(X)_m]=\frac{\lambda^{m/2}I_m(2\sqrt{\lambda})}{I_0(2\sqrt{\lambda})},
where is the modified Bessel function of the first kind.
Using the connecting formula for moments and factorial moments gives
:
\operatorname{E}X^m=\sum_{k=1}^m\left\{ {m \atop k} \right\}\frac{\lambda^{k/2}I_k(2\sqrt{\lambda})}{I_0(2\sqrt{\lambda})}.
In particular, the mean of is given by
:
\operatorname{E}X=\frac{\sqrt{\lambda}I_1(2\sqrt{\lambda})}{I_0(2\sqrt{\lambda})}.
Also, since , the variance is given by
:
\mathrm{Var}(X)=\lambda\left(1-\frac{I_1(2\sqrt{\lambda})^2}{I_0(2\sqrt{\lambda})^2}\right).
Suppose now that is an integer. Then
:
\operatorname{E}[(X)_m]=\frac{{\lambda^m}}{{(m!)^{\nu-1}}} \frac{_0F_{\nu-1}(;m+1,\ldots,m+1;\lambda)}{_0F_{\nu-1}(;1,\ldots,1;\lambda)}.
In particular,
:
\operatorname{E}[X]=\lambda \frac{_0F_{\nu-1}(;2,\ldots,2;\lambda)}{_0F_{\nu-1}(;1,\ldots,1;\lambda)},
and
\mathrm{Var}(X)=\frac{{\lambda^2}}{{2^{\nu-1}}} \frac{_0F_{\nu-1}(;3,\ldots,3;\lambda)}{_0F_{\nu-1}(;1,\ldots,1;\lambda)}+\operatorname{E}[X]-(\operatorname{E}[X])^2.
Median, mode and mean deviation
Let . Then the mode of is if
The mean deviation of
:
\operatorname{E}|X^\nu-\lambda| = 2Z(\lambda,\nu)^{-1} \frac{\lambda^{\lfloor\lambda^{1/\nu}\rfloor+1}}{\lfloor\lambda^{1/\nu}\rfloor!}.
No explicit formula is known for the median of
:
m=\lambda^{1/\nu}+\mathcal{O}\left(\lambda^{1/2\nu}\right),
as
Stein characterisation
Let
:
\operatorname{E}[\lambda f(X+1)-X^\nu f(X)]=0.
Conversely, suppose now that
Use as a limiting distribution
Let
:
\mathrm{P}(X = k) =
\frac {{{{(\frac{{\Gamma (r + k)}}{{k! \Gamma (r)}})}^\nu }{p^k}{{(1 - p)}^r}}}
{{\sum\limits_{i = 0}^\infty {{{(\frac{{\Gamma (r + i)}}{{i! \Gamma (r)}})}^\nu }} {p^i}{{(1 - p)}^r}}}
={{\left(\frac{{\Gamma (r + k)}}{{k! \Gamma (r)}}\right)}^\nu }
{{p^k}{{(1 - p)}^r}}
\frac{1}{{C(r,\nu ,p)}},\quad (k = 0,1,2, \ldots ),
convergents to a limiting distribution which is the COM-Poisson, as
Related distributions
X\sim\operatorname{CMP}(\lambda,1) , thenX follows the Poisson distribution with parameter\lambda .- Suppose
\lambda<1 . Then ifX\sim\mathrm{CMP}(\lambda,0) , we have thatX follows the geometric distribution with probability mass functionP(X=k)=\lambda^k(1-\lambda) ,k\geq0 . - The sequence of random variable
X_\nu\sim\mathrm{CMP}(\lambda,\nu) converges in distribution as\nu\rightarrow\infty to the Bernoulli distribution with mean\lambda(1+\lambda)^{-1} .
Parameter estimation
There are a few methods of estimating the parameters of the CMP distribution from the data. Two methods will be discussed: weighted least squares and maximum likelihood. The weighted least squares approach is simple and efficient but lacks precision. Maximum likelihood, on the other hand, is precise, but is more complex and computationally intensive.
= Weighted least squares =
The weighted least squares provides a simple, efficient method to derive rough estimates of the parameters of the CMP distribution and determine if the distribution would be an appropriate model. Following the use of this method, an alternative method should be employed to compute more accurate estimates of the parameters if the model is deemed appropriate.
This method uses the relationship of successive probabilities as discussed above. By taking logarithms of both sides of this equation, the following linear relationship arises
:
\log \frac{p_{x-1}}{p_x} = - \log \lambda + \nu \log x
where
Once the appropriateness of the model is determined, the parameters can be estimated by fitting a regression of
:
\operatorname{var}\left[\log \frac{\hat p_{x-1}}{\hat p_x}\right] \approx \frac{1}{np_x} + \frac{1}{np_{x-1}}
:
\text{cov}\left(\log \frac{\hat p_{x-1}}{\hat p_x}, \log \frac{\hat p_x}{\hat p_{x+1}} \right) \approx - \frac{1}{np_x}
= Maximum likelihood =
The CMP likelihood function is
:
\mathcal{L}(\lambda,\nu\mid x_1,\dots,x_n) = \lambda^{S_1} \exp(-\nu S_2) Z^{-n}(\lambda, \nu)
where
:
\operatorname{E}[X] = \bar X
:
\operatorname{E}[\log X!] = \overline{\log X!}
which do not have an analytic solution.
Instead, the maximum likelihood estimates are approximated numerically by the Newton–Raphson method. In each iteration, the expectations, variances, and covariance of
:
\operatorname{E}[f(x)] = \sum_{j=0}^\infty f(j) \frac{\lambda^j}{(j!)^\nu Z(\lambda, \nu)}.
This is continued until convergence of
Generalized linear model
The basic CMP distribution discussed above has also been used as the basis for a generalized linear model (GLM) using a Bayesian formulation. A dual-link GLM based on the CMP distribution has been developed,Guikema, S.D. and J.P. Coffelt (2008) "A Flexible Count Data Regression Model for Risk Analysis", Risk Analysis, 28 (1), 213–223. {{doi|10.1111/j.1539-6924.2008.01014.x}}
and this model has been used to evaluate traffic accident data.Lord, D., S.D. Guikema, and S.R. Geedipally (2008) "Application of the Conway–Maxwell–Poisson Generalized Linear Model for Analyzing Motor Vehicle Crashes," Accident Analysis & Prevention, 40 (3), 1123–1134. {{doi|10.1016/j.aap.2007.12.003}}Lord, D., S.R. Geedipally, and S.D. Guikema (2010) "Extension of the Application of Conway–Maxwell–Poisson Models: Analyzing Traffic Crash Data Exhibiting Under-Dispersion," Risk Analysis, 30 (8), 1268–1276. {{doi|10.1111/j.1539-6924.2010.01417.x}} The CMP GLM developed by Guikema and Coffelt (2008) is based on a reformulation of the CMP distribution above, replacing
A classical GLM formulation for a CMP regression has been developed which generalizes Poisson regression and logistic regression.Sellers, K. S. and Shmueli, G. (2010), [http://projecteuclid.org/euclid.aoas/1280842147 "A Flexible Regression Model for Count Data"], Annals of Applied Statistics, 4 (2), 943–961 This takes advantage of the exponential family properties of the CMP distribution to obtain elegant model estimation (via maximum likelihood), inference, diagnostics, and interpretation. This approach requires substantially less computational time than the Bayesian approach, at the cost of not allowing expert knowledge to be incorporated into the model. In addition it yields standard errors for the regression parameters (via the Fisher Information matrix) compared to the full posterior distributions obtainable via the Bayesian formulation. It also provides a statistical test for the level of dispersion compared to a Poisson model. Code for fitting a CMP regression, testing for dispersion, and evaluating fit is available.[http://www9.georgetown.edu/faculty/kfs7/research.html Code for COM_Poisson modelling], Georgetown Univ.
The two GLM frameworks developed for the CMP distribution significantly extend the usefulness of this distribution for data analysis problems.
References
External links
- [https://cran.r-project.org/web/packages/compoisson/index.html Conway–Maxwell–Poisson distribution package for R (compoisson) by Jeffrey Dunn, part of Comprehensive R Archive Network (CRAN)]
- [http://alumni.media.mit.edu/~tpminka/software/compoisson Conway–Maxwell–Poisson distribution package for R (compoisson) by Tom Minka, third party package]
{{ProbDistributions|discrete-infinite}}
{{DEFAULTSORT:Conway-Maxwell-Poisson distribution}}