Basic Theory
The Memoryless Property
Recall that in the basic model of the Poisson process, we have points
that occur randomly in continuous time, modeled by . The sequence of inter-arrival times is . The strong renewal assumption states that at each arrival time and at each fixed time, the process must probabilistically restart, independent of the past. The first part of that assumption implies that is a sequence of independent, identically distributed variables. The second part of the assumption implies that if the first arrival has not occurred by time , then the time remaining until the arrival occurs must have the same distribution as the first arrival time itself. This is known as the memoryless property and can be stated in terms of a general random variable as follows:
Suppose that takes values in . Then has the memoryless property if the conditional distribution of given is the same as the distribution of for every . Equivalently,
The memoryless property determines the distribution of up to a positive parameter, as we will see now.
Suppose that takes values in and satisfies the memoryless property. Then has a continuous distribution and there exists such that the distribution function of is
Details:
Let denote the denote the right-tail distribution function of (also known as the reliability function), so that for . From the definition of conditional probability, the memoryless property is equivalent to the law of exponents:
Let . Implicit in the memoryless property is for , so . If then
Next, if then
so . Now suppose that and . Then
Thus we have for rational . For , there exists a sequence of rational numbers with as . We have for each . But is continuous from the right, so taking limits gives . Now let . Then for .
The probability density function of is
- is decreasing on .
- is concave upward on .
- as .
Details:
The probability density function follows from [2] since . The properties in parts (a)–(c) are simple.
The probability distribution defined by the distribution function in [2] or equivalently the probability density function in [3] is the exponential distribution with rate parameter . The reciprocal is the scale parameter (as will be justified in [14]). Note that the mode of the distribution is 0, regardless of the parameter , not very helpful as a measure of center.
In the gamma experiment, set so that the simulated random variable has an exponential distribution. Vary with the scrollbar and watch how the shape of the probability density function changes. For selected values of , run the experiment 1000 times and compare the empirical density function to the probability density function.
The quantile function of is
- The median of is
- The first quartile of is
- The third quartile is
- The interquartile range is
Details:
The formula for follows from [2] by solving for in terms of .
In the quantile app, select the exponential distribution. Vary the scale parameter (which is ) and note the shape of the distribution/quantile function. For selected values of the parameter, compute a few values of the distribution function and the quantile function.
Returning to the Poisson model, we have our first formal definition:
A process of random points in time is a Poisson process with rate if and only the interarrvial times are independent, and each has the exponential distribution with rate .
Constant Failure Rate
Suppose now that has a continuous distribution on and is interpreted as the lifetime of a device. If denotes the distribution function of , then is the reliability function of . If denotes the probability density function of then the failure rate function is given by
The property of constant failure rate, like the memoryless property, characterizes the exponential distribution.
Random variable has constant failure rate if and only if has the exponential distribution with parameter .
Details:
If has the exponential distribution with rate , then from [2], the reliability function is and from [3] the probability density function is , so trivially has constant rate . For the converse, recall that in general, the distribution of a lifetime variable is determined by the failure rate function . Specifically, if denotes the reliability function, then , so . Integrating and then taking exponentials gives
In particular, if for , then for .
The memoryless and constant failure rate properties are the most famous characterizations of the exponential distribution, but are by no means the only ones. Indeed, entire books have been written on characterizations of this distribution.
Moments
Suppose again that has the exponential distribution with rate parameter . Naturaly, we want to know the the mean, variance, and various other moments of .
If then .
Details:
By the change of variables theorem for expected value,
Integrating by parts gives for . Of course so the result now follows by induction.
More generally, for every , where is the gamma function.
In particular.
Details:
These results follow from [9] and the compuational formulas for variance, skewness, and kurtosis.
In the context of the Poisson process, the parameter is known as the rate of the process. On average, there are time units between arrivals, so the arrivals come at an average rate of per unit time. The Poisson process is completely determined by the sequence of inter-arrival times, and hence is completely determined by the rate .
Note also that the mean and standard deviation are equal for an exponential distribution, and that the median is always smaller than the mean. Recall also that skewness and kurtosis are standardized measures, and so do not depend on the parameter (which is the reciprocal of the scale parameter). The excess kurtosis is .
In the gamma experiment, set so that the simulated random variable has an exponential distribution. Vary with the scrollbar and watch how the meanstandard deviation bar changes. For various values of , run the experiment 1000 times and compare the empirical mean and standard deviation to the distribution mean and standard deviation, respectively.
Additional Properties
The exponential distribution has a number of interesting and important mathematical properties. First, and not surprisingly, it's a member of the general exponential family.
Suppose that has the exponential distribution with rate parameter . Then has a one parameter general exponential distribution, with natural parameter and natural statistic .
Details:
This follows directly from the form of the PDF in [3] and the definition of the general exponential family.
The Scaling Property
As suggested earlier, the exponential distribution is a scale family, and is the scale parameter. Hence the distribution is trivially closed under scale transformations.
Suppose that has the exponential distribution with rate parameter and that . Then has the exponential distribution with rate parameter .
Details:
For , .
Recall that multiplying a random variable by a positive constant frequently corresponds to a change of units (minutes into hours for a lifetime variable, for example). Thus, the exponential distribution is preserved under such changes of units. In the context of the Poisson process, this has to be the case, since the memoryless property, which led to the exponential distribution in the first place, clearly does not depend on the time units.
In fact, the exponential distribution with rate parameter 1 is referred to as the standard exponential distribution. From [14], if has the standard exponential distribution and , then has the exponential distribution with rate parameter . Conversely, if has the exponential distribution with rate then has the standard exponential distribution.
Similarly, the Poisson process with rate parameter 1 is referred to as the standard Poisson process. If is the th inter-arrival time for the standard Poisson process for , then letting for gives the inter-arrival times for the Poisson process with rate . Conversely if is the th inter-arrival time of the Poisson process with rate for , then for gives the inter-arrival times for the standard Poisson process.
Relation to the Geometric Distribution
In many respects, the geometric distribution is a discrete version of the exponential distribution. In particular, recall that the geometric distribution on is the only distribution on with the memoryless and constant rate properties. So it is not surprising that the two distributions are also connected through various transformations and limits.
Suppose that has the exponential distribution with rate parameter . Then
- has the geometric distributions on with success parameter .
- has the geometric distributions on with success parameter .
Details:
Let denote the distribution function of in [2]
- For note that . Substituting into and simplifying gives .
- For note that . Substituting and simplifying gives .
Our next discussion generalizes part (a) of [15] and also introduces the truncated exponential distribution. In a sense, this discussion is a continuous version of the alternating coin tossing problem studied in the section on the geometric distribution.
Suppose that has the exponential distribution with rate parameter and let . Define and so that . Then
- and are independent.
- has the geometric distribution on with success parameter
- has a continuous distribution on with probability density function given by
Details:
Let and . Then
The independence of and now follows from the factroing theorem since the last expression is a product of a function of only and a function of only. But more explicitly we can rewrite the expression as
The function is the PDF of the geometric distribution on with success parameter while the function is tje cumulative distribution function for a continuous distribution on with PDF given in part (b).
In the context of [16], the distribution of is the same as the conditional distribution of given , and this is the exponential distribution truncated at .
Details:
If denotes the PDF of then the conditional PDF of given is given by for , and of course this is the function in [16].
It's not difficult to find the mean, variance, and moment generating function of the truncated exponential distribution directly from the density function, but the relationship between , and , and the independence of and , provides another clever method.
Suppose that has the exponential distribution with parameter truncated at as in [16]. Then
- The mean of is
- The variance of is
- The moment generating function of is given by
Details:
The crucial facts are that and that and are independent.
- . But and .
- . But and .
- . But for and for
Finally for this discussion, we note some limiting distributions of .
Suppose again that has the exponential distribution with parameter truncated at as in [16]. Then
- The distribution of converges to the exponential distribution with parameter as .
- The distribution of converges to the uniform distribution on as .
Details:
Part (a) is obvious from [16] or [17]. Part (b) also follows from [16] or from part (c) of [18] and L'Hospital's rule.
The following connection between the exponential and geometric distributions is interesting by itself, but will also be very important in the section on splitting Poisson processes. In words, a random, geometrically distributed sum of independent, identically distributed exponential variables is itself exponential.
Suppose that is a sequence of independent variables, each with the exponential distribution with rate . Suppose that has the geometric distribution on with success parameter and is independent of . Then has the exponential distribution with rate .
Details:
Recall that the moment generating function of is where is the common moment generating function of the terms in the sum, and is the probability generating function of the number of terms . But for and for . Thus,
It follows that has the exponential distribution with parameter
The next result explores the connection between the Bernoulli trials process and the Poisson process that was begun in the introduction.
For , suppose that has the geometric distribution on with success parameter , where \) as . Then the distribution of converges to the exponential distribution with parameter as .
Details:
Let denote the CDF of . Then for
But by a famous limit from calculus, as , and hence as . But by definition, or equivalently, so it follows that as . Hence as , which is the CDF of the exponential distribution.
To understand this result more clearly, suppose that we have a sequence of Bernoulli trials processes. In process , we run the trials at a rate of per unit time, with probability of success . Thus, the actual time of the first success in process is . The last result shows that if as , then the sequence of Bernoulli trials processes converges to the Poisson process with rate parameter as . We will return to this point in subsequent sections.
Orderings and Order Statistics
Suppose that and have exponential distributions with parameters and , respectively, and are independent. Then
Details:
This result can be proved in a straightforward way by integrating the joint PDF of over . A more elegant proof uses conditioning and the moment generating function in [11]:
The next result gives an important random version of the memoryless property.
Suppose that and are independent variables with values in and that has the exponential distribution with rate parameter . Then and are conditionally independent given , and the conditional distribution of is also exponential with parameter .
Details:
Suppose that (measurable of course) and . Then
But conditioning on we can write the numerator as
Similarly, conditioning on gives . Thus
Letting we have so given , the variable has the exponential distribution with parameter . Letting , we see that given , variable has the distribution
Finally, because of the factoring, and are conditionally independent given .
For our next discussion, suppose that is a sequence of independent random variables, and that has the exponential distribution with rate parameter for each .
Let . Then has the exponential distribution with parameter .
Details:
Recall that in general, and therefore by independence, for , where is the reliability function of and is the reliability function of for each . When has the exponential distribution with rate for each , we have for .
In the context of reliability, if a series system has independent components, each with an exponentially distributed lifetime, then the lifetime of the system is also exponentially distributed, and the failure rate of the system is the sum of the component failure rates. In the context of random processes, if we have independent Poisson process, then the new process obtained by combining the random points in time is also Poisson, and the rate of the new process is the sum of the rates of the individual processes (we will return to this point latter).
Let . Then has distribution function given by
Details:
Recall that in general, and therefore by independence, for , where is the distribution function of and is the distribution function of for each .
Consider the special case where for each . In statistical terms, is a random sample of size from the exponential distribution with parameter . From [24], the minimum has the exponential distribution with rate while by [25], the maximum has distribution function for . Recall that and are the first and last order statistics, respectively.
In the order statistic experiment, select the exponential distribution.
- Set (this gives the minimum ). Vary with the scrollbar and note the shape of the probability density function. For selected values of , run the simulation 1000 times and compare the empirical density function to the true probability density function.
- Vary with the scrollbar, set each time (this gives the maximum ), and note the shape of the probability density function. For selected values of , run the simulation 1000 times and compare the empirical density function to the true probability density function.
Curiously, the distribution of the maximum of independent, identically distributed exponential variables is also the distribution of the sum of independent exponential variables, with rates that grow linearly with the index.
Suppose that for each where . Then has distribution function given by
Details:
By assumption, has PDF given by for . We want to show that has PDF given by
The PDF of a sum of independent variables is the convolution of the individual PDFs, so we want to show that
The proof is by induction on . Trivially , so suppose the result holds for a given . Then
Now substitute so that or equivalently . After some algebra,
This result has an application to the Yule process, named for George Yule. The Yule process, which has some parallels with the Poisson process, is studied in the chapter on Markov processes. We can now generalize [22]:
For ,
Details:
First, note that for all if and only if . But the minimum on the right is independent of and by [24], has the exponential distribution with parameter . The result now follows from [22].
Suppose that for each , is the time until an event of interest occurs (the arrival of a customer, the failure of a device, etc.) and that these times are independent and exponentially distributed. Then the first time that one of the events occurs is also exponentially distributed, and the probability that the first event to occur is event is proportional to the rate .
The probability of a total ordering is
Details:
Let . then
But from the previous result, and is independent of . Thus we have
so the result follows by induction.
Of course, the probabilities of other orderings can be computed by permuting the parameters appropriately in [29].
Results [24] and [28] are very important in the theory of continuous-time Markov chains. But for that application and others, it's convenient to extend the exponential distribution to two degenerate cases: point mass at 0 and point mass at (so the first is the distribution of a random variable that takes the value 0 with probability 1, and the second the distribution of a random variable that takes the value with probability 1). In terms of the rate parameter and the distribution function , point mass at 0 corresponds to so that for . Point mass at corresponds to so that for . The memoryless property, as expressed in terms of the reliability function , still holds for these degenerate cases on :
We also need to extend some of results above for a finite number of variables to a countably infinite number of variables. So for the remainder of this discussion, suppose that is a countable collection of independent random variables, and that has the exponential distribution with parameter for each .
Let . Then has the exponential distribution with parameter
Details:
The proof is almost the same as in [24]. Note that and so
If then has a proper exponential distribution with the sum as the parameter. If then for all so .
For ,
Details:
First note that since the variables have continuous distributions and is countable,
Next note that for all if and only if where . But is independent of and, by [29], has the exponential distribution with parameter . If , then is 0 with probability 1, and so . If , then and have proper exponential distributions, and so the result now follows from [22].
We need one last result in this setting: a condition that ensures that the sum of an infinite collection of exponential variables is finite with probability one.
Let and . Then and if and only if .
Details:
The result is trivial if is finite, so assume that . Recall that and hence . Trivially if then . Conversely, suppose that . Then and hence . Using independence and the moment generating function in [11],
Next recall that if for then
Hence it follows that
In particular, this means that as and hence as . But then
By the comparison test for infinite series, it follows that
Computational Exercises
Show directly that the exponential probability density function is a valid probability density function.
Details:
Clearly for . Simple integration that
Suppose that the length of a telephone call (in minutes) is exponentially distributed with rate parameter . Find each of the following:
- The probability that the call lasts between 2 and 7 minutes.
- The median, the first and third quartiles, and the interquartile range of the call length.
Details:
Let denote the call length.
- , , ,
Suppose that the lifetime of a certain electronic component (in hours) is exponentially distributed with rate parameter . Find each of the following:
- The probability that the component lasts at least 2000 hours.
- The median, the first and third quartiles, and the interquartile range of the lifetime.
Details:
Let denote the lifetime
- , , ,
Suppose that the time between requests to a web server (in seconds) is exponentially distributed with rate parameter . Find each of the following:
- The mean and standard deviation of the time between requests.
- The probability that the time between requests is less that 0.5 seconds.
- The median, the first and third quartiles, and the interquartile range of the time between requests.
Details:
Let denote the time between requests.
- ,
- , , ,
Suppose that the lifetime of a fuse (in 100 hour units) is exponentially distributed with . Find each of the following:
- The rate parameter.
- The mean and standard deviation.
- The median, the first and third quartiles, and the interquartile range of the lifetime.
Details:
Let denote the lifetime.
- ,
- , , ,
The position of the first defect on a digital tape (in cm) has the exponential distribution with mean 100. Find each of the following:
- The rate parameter.
- The probability that given .
- The standard deviation.
- The median, the first and third quartiles, and the interquartile range of the position.
Details:
Let denote the position of the first defect.
- , , ,
Suppose that are independent, exponentially distributed random variables with respective parameters . Find the probability of each of the 6 orderings of the variables.
Details: