Doob's martingale convergence theorems

From Wikipedia, the free encyclopedia

In mathematics — specifically, in stochastic analysisDoob's martingale convergence theorems are a collection of results on the long-time limits of supermartingales, named after the American mathematician Joseph Leo Doob.

Contents

[edit] Statement of the theorems

In the following, (Ω, FFP), F = (Ft)t≥0, will be a filtered probability space and N : [0, +∞) × Ω → R will be a right-continuous supermartingale with respect to the filtration F, so that for all 0 ≤ s ≤ t < +∞,

N_{s} \geq \mathbf{E} \big[ N_{t} \big| F_{s} \big].

[edit] Doob's first martingale convergence theorem

Doob's first martingale convergence theorem provides a sufficient condition for the random variables Nt to have a limit as t → +∞ in a pointwise sense, i.e. for each ω in the sample space Ω individually.

For t ≥ 0, let Nt = max(−Nt, 0) and suppose that

\sup_{t > 0} \mathbf{E} \big[ N_{t}^{-} \big] < + \infty.

Then the pointwise limit

N(\omega) = \lim_{t \to + \infty} N_{t} (\omega)

exists for P-almost all ω ∈ Ω.

[edit] Doob's second martingale convergence theorem

It is important to note that the convergence in Doob's first martingale convergence theorem is pointwise, not uniform, and is unrelated to convergence in mean square, or indeed in any Lp space. In order to obtain convergence in L1 (i.e., convergence in mean), one requires uniform integrability of the random variables Nt. By Chebyshev's inequality, convergence in L1 implies convergence in probability and convergence in distribution.

The following are equivalent:

\lim_{C \to \infty} \sup_{t > 0} \int_{\{ \omega \in \Omega | N_{t} (\omega) > C \}} \big| N_{t} (\omega) \big| \, \mathrm{d} \mathbf{P} (\omega) = 0;
\mathbf{E} \big[ \big| N_{t} - N \big| \big] = \int_{\Omega} \big| N_{t} (\omega) - N (\omega) \big| \, \mathrm{d} \mathbf{P} (\omega) \to 0 \mbox{ as } t \to + \infty.

[edit] Corollary: convergence theorem for continuous martingales

Let M : [0, +∞) × Ω → R be a continuous martingale such that

\sup_{t > 0} \mathbf{E} \big[ \big| M_{t} \big|^{p} \big] < + \infty

for some p > 1. Then there exists a random variable M ∈ L1(Ω, PR) such that Mt → M as t → +∞ both P-almost surely and in L1(Ω, PR).

[edit] Discrete-time results

Similar results can be obtained for discrete-time supermartingales and submartingales, the obvious difference being that no continuity assumptions are required. For example, the result above becomes

Let M : N × Ω → R be a discrete-time martingale such that

\sup_{k \in \mathbf{N}} \mathbf{E} \big[ \big| M_{k} \big|^{p} \big] < + \infty

for some p > 1. Then there exists a random variable M ∈ L1(Ω, PR) such that Mk → M as k → +∞ both P-almost surely and in L1(Ω, PR)

[edit] Convergence of conditional expectations

Doob's martingale convergence theorems imply that conditional expectations also have a convergence property.

Let (Ω, FP) be a probability space and let X be a random variable in L1. Let F = (Fk)kN be any filtration of F, and define F to be the minimal σ-algebra generated by (Fk)kN. Then

\mathbf{E} \big[ X \big| F_{k} \big] \to \mathbf{E} \big[ X \big| F_{\infty} \big] \mbox{ as } k \to \infty

both P-almost surely and in L1.

[edit] References