In a compound Poisson process, each arrival in an ordinary Poisson process comes with an associated real-valued random variable that represents the value of the arrival in a sense. These variables are independent and identically distributed, and are independent of the underlying Poisson process. Our interest centers on the sum of the random variables for all the arrivals up to a fixed time \( t \), which thus is a Poisson-distributed random sum of random variables. Distributions of this type are said to be compound Poisson distributions, and are important in their own right, particularly since some surprising parametric distributions turn out to be compound Poisson.
We start with a Poisson process with rate \( r \in (0, \infty) \). Recall
Next we attach a random variable to each arrival that represents its value.
For \( n \in \N_+ \), let \( U_n \) denote the value of the \( n \)th arrival. We assume that \( \bs{U} = (U_1, U_2, \ldots) \) is a sequence of independent, identically distributed, real-valued random variables, and that \( \bs{U} \) is independent of the underlying Poisson process. Here are some typical examples:
The common distribution may be discrete or continuous, but in either case, we let \( f \) denote the common probability density function. We will let \( \mu = \E(U_n) \) denote the common mean, \( \sigma^2 = \var(U_n) \) the common variance, and \( G \) the common moment generating function, so that \( G(s) = \E\left[\exp(s U_n)\right] \) for \( s \) in some interval \( I \) about 0. Here is our main definition:
The compound Poisson process associated with the given Poisson process \(\bs{N}\) and the sequence \( \bs{U} \) is the stochastic process \( \bs{V} = \{V_t: t \in [0, \infty)\} \) where \[ V_t = \sum_{n=1}^{N_t} U_n\] So \( V_t \) is the total value for all of the arrivals in \( (0, t] \). For the examples in
Recall that a sum over an empty index set is 0, so \( V_0 = 0 \).
Note that for fixed \( t \), \( V_t \) is a random sum of independent, identically distributed random variables, a topic that we have studied before. In this sense, we have a special case, since the number of terms \( N_t \) has the Poisson distribution with parameter \( r t\). But we also have a new wrinkle, since the process is indexed by the continuous time parameter \( t \), and so we can study its properties as a stochastic process. Our first result is a pair of properties shared by the underlying Poisson process.
\( \bs{V} \) has stationary, independent increments:
Next we consider various moments of the compound process.
For \( t \in [0, \infty) \), the mean and variance of \( V_t \) are
Again, these are special cases of general results for random sums of IID variables, but we give separate proofs for completeness. The basic tool is conditional expected value and conditional variance. Recall also that \( \E(N_t) = \var(N_t) = r t \).
For \( t \in [0, \infty) \), the moment generating function of \( V_t \) is given by \[ \E\left[\exp(s V_t)\right] = \exp\left(r t \left[G(s) - 1\right]\right), \quad s \in I \]
Again, this is a special case of the more general result for random sums of IID variables, but we give a another proof for completeness. As with the last theorem, the key is to condition on \( N_t \) and recall that the MGF of a sum of independent variables is the product of the MGFs. Thus \[ \E\left[\exp(s V_t)\right] = \E\left(\E\left[\exp(s V_t \mid N_t)\right]\right) = \E\left[G^{N_t}(s)\right] = P_t\left[G(s)\right] \] where \( P_t \) is the probability generating function of \( N_t \). But we know from our study of the Poisson distribution that \( P_t(x) = \exp\left[r t (x - 1)\right] \) for \( x \in \R \).
By exactly the same argument, the same relationship holds for characteristic functions and, in the case that the variables in \( \bs{U} \) take values in \( \N \), for probability generating functions.. That is, if the variables in \( \bs{U} \) have generating function \( G \), then the generating function \( H \) of \( V_t \) is given by \[ H(s) = \exp(r t [G(s) - 1]) \] for \( s \) in the domain of \( G \), where generating function can be any of the three types we have discussed: probability, moment, or characteristic.
First note that thinning a Poisson process can be thought of as a special case of a compound Poisson process. Thus, suppose that \( \bs{U} = (U_1, U_2, \ldots) \) is a Bernoulli Trials sequence with success parameter \( p \in (0, 1) \), and as above, that \( \bs{U} \) is independent of the Poisson process \( \bs{N} \). In the usual language of thinning, the arrivals are of two types (1 and 0), and \( U_i \) is the type of the \( i \)th arrival. Thus the compound process \( \bs{V} \) constructed above is the thinned process, so that \( V_t \) is the number of type 1 points up to time \( t \). We know that \( \bs{V} \) is also a Poisson process, with rate \( r p \).
The results for thinning generalize to the case where the values of the arrivals have a discrete distribution. Thus, suppose \( U_i \) takes values in a countable set \( S \subseteq \R \), and as before, let \( f \) denote the common probability density function so that \( f(u) = \P(U_i = u) \) for \( u \in S \) and \( i \in \N_+ \). For \( u \in S \), let \( N^u_t \) denote the number of arrivals up to time \( t \) that have the value \( u \), and let \( \bs{N}^u = \left\{N^u_t: t \in [0, \infty)\right\} \) denote the corresponding stochastic process. Armed with this setup, here is the result:
The compound Poisson process \( \bs{V} \) associated with \( \bs{N} \) and \( \bs{U} \) can be written in the form \[ V_t = \sum_{u \in S} u N^u_t, \quad t \in [0, \infty) \] The processes \( \{\bs{N}^u: u \in S\} \) are independent Poisson processes, and \( \bs{N}^u \) has rate \( r f(u) \) for \( u \in S \).
Note that \( U_i = \sum_{u \in S} u \bs{1}(U_i = u) \) and hence \[ V_t = \sum_{i = 1}^{N_t} U_i = \sum_{i = 1}^{N_t} \sum_{u \in S} u \bs{1}(U_i = u) = \sum_{u \in S} u \sum_{i = 1}^{N_t} \bs{1}(U_i = u) = \sum_{u \in S} u N^u_t \] The fact that \( \{\bs{N}^u: u \in S\} \) are independent Poisson processes, and that \( \bs{N}^u \) has rate \( r f(u) \) for \( u \in S \) follows from our result on thinning.
A compound Poisson random variable can be defined outside of the context of a Poisson process. Here is the formal definition:
Suppose that \( \bs{U} = (U_1, U_2, \ldots) \) is a sequence of independent, identically distributed random variables, and that \( N \) is independent of \( \bs{U} \) and has the Poisson distribution with parameter \( a \in (0, \infty) \). Then \( V = \sum_{i=1}^N U_i \) has a compound Poisson distribution.
But in fact, compound Poisson variables usually do arise in the context of an underlying Poisson process. In any event, the results on the mean and variance in and the generating function in hold with \( r t \) replaced by \( a \). Compound Poisson distributions are infinitely divisible. A famous theorem of William Feller gives a partial converse: an infinitely divisible distribution on \( \N \) must be compound Poisson.
The negative binomial distribution on \( \N \) is infinitely divisible, and hence must be compound Poisson. Here is the construction:
Let \( p, \, k \in (0, \infty) \). Suppose that \( \bs{U} = (U_1, U_2, \ldots) \) is a sequence of independent variables, each having the logarithmic series distribution with shape parameter \( 1 - p \). Suppose also that \( N \) is independent of \( \bs{U} \) and has the Poisson distribution with parameter \( - k \ln(p) \). Then \( V = \sum_{i=1}^N U_i \) has the negative binomial distribution on \( \N \) with parameters \( k \) and \( p \).
From , the probability generating function of \( V \) is \( P(t) = \exp\left( \lambda [Q(t) - 1]\right) \) where \( \lambda \) is the parameter of the Poisson variable \( N \) and \( Q(t) \) is the common PGF of the the terms in the sum. Using the PGF of the logarithmic series distribution, and the particular values of the parameters, we have \[ P(t) = \exp \left[-k \ln(p) \left(\frac{\ln[1 - (1 - p)t]}{\ln(p)} - 1\right)\right], \quad \left|t\right| \lt \frac{1}{1 - p} \] Using properties of logarithms and simple algebra, this reduces to \[ P(t) = \left(\frac{p}{1 - (1 - p)t}\right)^k, \quad \left|t\right| \lt \frac{1}{1 - p} \] which is the PGF of the negative binomial distribution with parameters \( k \) and \( p \).
As a special case (\( k = 1 \)), it follows that the geometric distribution on \( \N \) is also compound Poisson.