$$\newcommand{\R}{\mathbb{R}}$$ $$\newcommand{\N}{\mathbb{N}}$$ $$\newcommand{\E}{\mathbb{E}}$$ $$\newcommand{\var}{\text{var}}$$ $$\newcommand{\P}{\mathbb{P}}$$ $$\newcommand{\bs}{\boldsymbol}$$
1. Random
2. 4. Special Distributions
3. Power Series Distributions

## Power Series Distributions

Power Series Distributions are discrete distributions on (a subset of) $$\N$$ constructed from power series. This class of distributions is important because most of the special, discrete distributions are power series distributions.

### Basic Theory

#### Power Series

Suppose that $$\bs{a} = (a_0, a_1, a_2, \ldots)$$ is a sequence of nonnegative real numbers. We are interested in the power series with $$\bs{a}$$ as the sequence of coefficients. Recall first that the partial sum of order $$n \in \N$$ is $g_n(\theta) = \sum_{k=0}^n a_k \theta^k, \quad \theta \in \R$ The power series $$g$$ is then defined by $$g(\theta) = \lim_{n \to \infty} g_n(\theta)$$ for $$\theta \in \R$$ for which the limit exists, and is denoted $g(\theta) = \sum_{n=0}^\infty a_n \theta^n$ Note that the series converges when $$\theta = 0$$, and $$g(0) = a_0$$. Beyond this trivial case, recall that there exists $$r \in [0, \infty]$$ such that the series converges absolutely for $$\left|\theta\right| \lt r$$ and diverges for $$\left|\theta\right| \gt r$$. The number $$r$$ is the radius of convergence. From now on, we assume that $$r \gt 0$$. If $$r \lt \infty$$, the series may converge (absolutely) or may diverge to $$\infty$$ at the endpoint $$r$$. At $$-r$$, the series may converge absolutely, may converge conditionally, or may diverge.

#### Distributions

From now on, we restrict $$\theta$$ to the interval $$[0, r)$$; this interval is our parameter space. Some of the results below may hold when $$r \lt \infty$$ and $$\theta = r$$, but dealing with this case explicitly makes the exposition unnecessarily cumbersome.

A random variable $$N$$ with values in $$\N$$ has the power series distribution associated with the function $$g$$ (or equivalently with the sequence $$\bs{a}$$), with parameter $$\theta \in [0, r)$$, if $$N$$ has discrete probability density function $f_\theta(n) = \frac{a_n \theta^n}{g(\theta)}, \quad n \in \N$

Proof:

To show that $$f_\theta$$ is a valid discrete probability density function, note that $$a_n \theta^n$$ is nonnegative for each $$n \in \N$$ and $$g(\theta)$$, by definition, is the normalizing constant for the sequence $$\left(a_n \theta^n: n \in \N\right)$$.

Note that when $$\theta = 0$$, the distribution is simply the point mass distribution at $$0$$; that is, $$f_0(0) = 1$$.

The distribution function $$F_\theta$$ is given by $F_\theta(n) = \frac{g_n(\theta)}{g(\theta)}, \quad n \in \N$

Proof:

This follows immediately from the definitions since $$F_\theta(n) = \sum_{k=0}^n f_\theta(k)$$ for $$n \in \N$$

Of course, the probability density function $$f_\theta$$ is most useful when the power series $$g(\theta)$$ can be given in closed form, and similarly the distribution function $$F_\theta$$ is most useful when the power series and the partial sums can be given in closed form

#### Moments

The moments of $$N$$ can be expressed in terms of the underlying power series function $$g$$, and the nicest expression is for the factorial moments. Recall that the permutation formula is $$t^{(k)} = t (t - 1) \cdots (t - k + 1)$$ for $$t \in \R$$ and $$k \in \N$$, and the factorial moment of $$N$$ of order $$k \in \N$$ is $$\E\left(N^{(k)}\right)$$.

For $$\theta \in [0, r)$$, the factorial moments of $$N$$ are as follows, where $$g^{(k)}$$ is the $$k$$th derivative of $$g$$. $\E\left(N^{(k)}\right) = \frac{\theta^k g^{(k)}(\theta)}{g(\theta)}, \quad k \in \N$

Proof:

Recall that a power series is infinitely differentiable in the open interval of convergence, and that the derivatives can be taken term by term. Thus $\E\left(N^{(k)}\right) = \sum_{n=0}^\infty n^{(k)} \frac{a_n \theta^n}{g(\theta)} = \frac{\theta^k}{g(\theta)} \sum_{n=k}^\infty a_k n^{(k)} \theta^{n-k} = \frac{\theta^k}{g(\theta)} g^{(k)}(\theta)$

The mean and variance of $$N$$ are

1. $$\E(N) = \theta g^\prime(\theta) \big/ g(\theta)$$
2. $$\var(N) = \theta^2 \left(g^{\prime\prime}(\theta) \big/ g(\theta) - \left[g^\prime(\theta) \big/ g(\theta)\right]^2\right)$$

The probability generating function of $$N$$ also has a simple expression in terms of $$g$$.

For $$\theta \in (0, r)$$, the probability generating function $$P$$ of $$N$$ is given by $P(t) = \E\left(t^N\right) = \frac{g(\theta t)}{g(\theta)}, \quad t \lt \frac{r}{\theta}$

Proof:

For $$t \in (0, r / \theta)$$, $P(t) = \sum_{n=0}^\infty t^n f_\theta(n) = \frac{1}{g(\theta)} \sum_{n=0}^\infty a_n (t \theta)^n = \frac{g(t \theta)}{g(\theta)}$

#### Relations

Suppose that $$N_1$$ and $$N_2$$ are independent, and have power series distributions relative to the functions $$g_1$$ and $$g_2$$, respectively, each with parameter value $$\theta \lt \min\{r_1, r_2\}$$. Then $$N_1 + N_2$$ has the power series distribution relative to the function $$g_1 g_2$$, with parameter value $$\theta$$.

Proof:

A direct proof is possible, but there is an easy proof using probability generating functions. Recall that the PGF of the sum of independent variables is the product of the PGFs. Hence the PGF of $$N_1 + N_2$$ is $P(t) = P_1(t) P_2(t) = \frac{g_1(\theta t)}{g_1(\theta)} \frac{g_2(\theta t)}{g_2(\theta)} = \frac{g_1(\theta t) g_2(\theta t)}{g_1(\theta) g_2(\theta)}, \quad t \lt \min\left\{\frac{r_1}{\theta}, \frac{r_2}{\theta}\right\}$ The last expression is the PGF of the power series distribution relative to the function $$g_1 g_2$$, at $$\theta$$.

Suppose that $$(N_1, N_2, \ldots, N_k)$$ is a sequence of independent variables, each with the same power series distribution, relative to the function $$g$$ and with parameter value $$\theta \lt r$$. Then $$N_1 + N_2 + \cdots + N_k$$ has the power series distribution relative to the function $$g^k$$ and with parameter $$\theta$$.

### Examples and Special Cases

#### Special Distributions

The Poisson distribution with rate parameter $$\lambda \in [0, \infty)$$ is a power series distribution relative to the function $$g(\lambda) = e^\lambda$$ for $$\lambda \in [0, \infty)$$.

Proof:

This follows directly from the definition, since the PDF of the Poisson distribution with parameter $$\lambda$$ is $$f(n) = e^{-\lambda} \lambda^n / n!$$ for $$n \in \N$$.

The geometric distribution on $$\N$$ with success parameter $$p \in (0, 1]$$ is a power series distribution relative to the function $$g(\theta) = 1 \big/ (1 - \theta)$$ for $$\theta \in [0, 1)$$, where $$\theta = 1 - p$$.

Proof:

This follows directly from the definition, since the PDF of the geometric distribution on $$\N$$ is $$f(n) = (1 - p)^n p = (1 - \theta) \theta^n$$ for $$n \in \N$$.

For fixed $$k \in (0, \infty)$$, the negative binomial distribution on $$\N$$ with with stopping parameter $$k$$ and success parameter $$p \in (0, 1]$$ is a power series distribution relative to the function $$g(\theta) = 1 \big/ (1 - \theta)^k$$ for $$\theta \in [0, 1)$$, where $$\theta = 1 - p$$.

Proof:

This follows from the result above on sums, but can also be seen directly, since the PDF is $f(n) = \binom{n + k - 1}{k - 1} p^k (1 - p)^{n} = (1 - \theta)^k \binom{n + k - 1}{k - 1} \theta^n, \quad n \in \N$

For fixed $$n \in \N_+$$, the binomial distribution with trial parameter $$n$$ and success parameter $$p \in [0, 1)$$ is a power series distribution relative to the function $$g(\theta) = \left(1 + \theta\right)^n$$ for $$\theta \in [0, \infty)$$, where $$\theta = p \big/ (1 - p)$$.

Proof:

Note that the PDF of the binomial distribution is $f(k) = \binom{n}{k} p^k (1 - p)^{n - k} = (1 - p)^n \binom{n}{k} \left(\frac{p}{1 - p}\right)^k = \frac{1}{(1 + \theta)^n} \binom{n}{k} \theta^k, \quad k \in \{0, 1, \ldots, n\}$ where $$\theta = p / (1 - p)$$. This shows that the distribution is a power series distribution corresponding to the function $$g(\theta) = (1 + \theta)^n$$.

The logarithmic distribution with parameter $$p \in [0, 1)$$ is a power series distribution relative to the function $$g(p) = -\ln(1 - p)$$ for $$p \in [0, 1)$$.

Proof:

This follows directly from the definition, since the PDF is $f(n) = \frac{1}{-\ln(1 - p)} \frac{1}{n} p^n, \quad n \in \N$