Renewal theory
From Wikipedia, the free encyclopedia
Renewal theory is a branch of probability theory with an interesting and varied range of applications. Applications include calculating the expected time for a monkey who is randomly tapping at a keyboard to type the word Macbeth and comparing the longterm benefits of different insurance policies.
Contents |
[edit] Renewal processes - an introduction
A renewal process is a generalisation of the Poisson process. In essence, the Poisson process is a continuous-time Markov process on the positive integers (usually starting at zero) which has independent identically distributed holding times at each integer i (exponentially distributed) before advancing (with probability 1) to the next integer:i + 1. In the same informal spirit, we may define a renewal process to be the same thing, except that the holding times take on a more general distribution. (Note however that the IID property of the holding times is retained).
[edit] Formal definition
Let be a sequence of independent identically distributed random variables such that
We refer to the random variable Si as the "ith" holding time.
Define for each n > 0 :
each Jn referred to as the "nth" jump time and the intervals
- [Jn,Jn + 1]
being called renewal intervals.
Then the random variable given by
is called a renewal process.
[edit] Interpretation
One may choose to think of the holding times as the time elapsed before a machine breaks for the "ith" time since the last time it broke. (Note this assumes that the machine is immediately fixed and we restart the clock immediately.) Under this interpretation, the jump times record the successive times at which the machine breaks and the renewal process Xt records the number of times the machine has so far had to be repaired at any given time t.
However it is more helpful to understand the renewal process in its abstract form, since it may be used to model a great number of practical situations of interest which do not relate very closely to the operation of machines.
[edit] Renewal-reward processes
Let be a sequence of IID random variables (rewards) satisfying
- .
Then the random variable
is called a renewal-reward process. Note that unlike the Si, each Wi may take negative values as well as positive values.
[edit] Interpretation
In the context of the above interpretation of the holding times as the time between successive malfunctions of a machine, the "rewards" (which in this case happen to be negative) may be viewed as the successive repair costs incurred as a result of the successive malfunctions.
An alternative analogy is that we have a magic goose which lays eggs at intervals (holding times) distributed as Si. Sometimes it lays golden eggs of random weight, and sometimes it lays toxic eggs (also of random weight) which require responsible (and costly) disposal. The "rewards" Wi are the successive (random) financial losses/gains resulting from successive eggs (i = 1,2,3,...) and Yt records the total financial "reward" at time t.
[edit] Properties of renewal processes and renewal-reward processes
We define the renewal function:
[edit] The elementary renewal theorem
The renewal function satisfies
The proof of this statement is non-trivial and therefore omitted.
[edit] The renewal equation
The renewal function satisfies
where FS is the cumulative distribution function of S1 and fS is the corresponding probability density function.
[edit] Proof of the renewal equation
- We may iterate the expectation about the first holding time:
- But by the Markov property
- So
- as required.
[edit] Asymptotic properties
and satisfy
- (strong law of large numbers for renewal processes)
- (strong law of large numbers for renewal-reward processes)
almost surely.
[edit] Proof
- First consider . By definition we have:
- for all and so
- for all t ≥ 0.
- Now since we have:
- as almost surely (with probability 1). Hence:
- almost surely (using the strong law of large numbers); similarly:
- almost surely.
- Thus (since t / Xt is sandwiched between the two terms)
- almost surely.
- Next consider . We have
- almost surely (using the first result and using the law of large numbers on Yt).
[edit] The inspection paradox
A curious feature of renewal processes is that if we wait some predetermined time t and then observe how large the renewal interval containing t is, we should expect it to be typically larger than a renewal interval of average size.
Mathematically the inspection paradox states: for any t > 0 the renewal interval containing t is stochastically larger than the first renewal interval. That is, for all x > 0 and for all t > 0:
where FS is the cumulative distribution function of the IID holding times Si.
[edit] Proof of the inspection paradox
Observe that the last jump-time before t is ; and that the renewal interval containing t is . Then
as required.
[edit] Example applications
[edit] Example 1 - use of the strong law of large numbers
Eric the entrepreneur has n machines, each having an operational lifetime uniformly distributed between zero and two years. Eric may let each machine run until it fails with replacement cost £2600; alternatively he may replace a machine at any time while it is still functional at a cost of £200.
What is his optimal replacement policy?
[edit] Solution
We may model the lifetime of the n machines as n independent concurrent renewal-reward processes, so it is sufficient to consider the case n=1. Denote this process by . The successive lifetimes S of the replacement machines are independent and identically distributed, so the optimal policy is the same for all replacement machines in the process.
If Eric decides at the start of a machine's life to replace it at time 0 < t < 2 but the machine happends to fail before that time then the lifetime S of the machine is uniformly distributed on [0, t] and thus has expectation 0.5t. So the overall expected lifetime of the machine is:
and the expected cost W per machine is:
So by the strong law of large numbers, his longterm average cost per unit time is:
then differentiating with respect to t:
this implies that the turning points satisfy:
- 0 = (4t − t2)(1200) − (4 − 2t)(1200t + 200) = 4800t − 1200t2 − 4800t − 800 + 2400t2 + 400t
-
- = − 800 + 400t + 1200t2,
and thus
- 0 = 3t2 + t − 2 = (3t − 2)(t + 1).
We take the only solution t in [0, 2]: t = 2/3. This is indeed a minimum (and not a maximum) since the cost per unit time tends to infinity as t tends to zero, meaning that the cost is decreasing as t increases, until the point 2/3 where it starts to increase.