#jsDisabledContent { display:none; } My Account | Register | Help

# Doob's martingale convergence theorems

Article Id: WHEBN0012261058
Reproduction Date:

 Title: Doob's martingale convergence theorems Author: World Heritage Encyclopedia Language: English Subject: Collection: Publisher: World Heritage Encyclopedia Publication Date:

### Doob's martingale convergence theorems

In mathematics – specifically, in stochastic analysis – Doob's martingale convergence theorems are a collection of results on the long-time limits of supermartingales, named after the American mathematician Joseph L. Doob.

## Contents

• Statement of the theorems 1
• Doob's first martingale convergence theorem 1.1
• Doob's second martingale convergence theorem 1.2
• Corollary: convergence theorem for continuous martingales 1.3
• Discrete-time results 2
• Convergence of conditional expectations: Lévy's zero–one law 3
• References 5

## Statement of the theorems

In the following, (Ω, FFP), F = (Ft)t≥0, will be a filtered probability space and N : [0, +∞) × Ω → R will be a right-continuous supermartingale with respect to the filtration F; in other words, for all 0 ≤ s ≤ t < +∞,

N_{s} \geq \mathbf{E} \big[ N_{t} \big| F_{s} \big].

### Doob's first martingale convergence theorem

Doob's first martingale convergence theorem provides a sufficient condition for the random variables Nt to have a limit as t → +∞ in a pointwise sense, i.e. for each ω in the sample space Ω individually.

For t ≥ 0, let Nt = max(−Nt, 0) and suppose that

\sup_{t > 0} \mathbf{E} \big[ N_{t}^{-} \big] < + \infty.

Then the pointwise limit

N(\omega) = \lim_{t \to + \infty} N_{t} (\omega)

exists finite for P-almost all ω ∈ Ω.

### Doob's second martingale convergence theorem

It is important to note that the convergence in Doob's first martingale convergence theorem is pointwise, not uniform, and is unrelated to convergence in mean square, or indeed in any Lp space. In order to obtain convergence in L1 (i.e., convergence in mean), one requires uniform integrability of the random variables Nt. By Chebyshev's inequality, convergence in L1 implies convergence in probability and convergence in distribution.

The following are equivalent:

\lim_{C \to \infty} \sup_{t > 0} \int_{\{ \omega \in \Omega | N_{t} (\omega) > C \}} \big| N_{t} (\omega) \big| \, \mathrm{d} \mathbf{P} (\omega) = 0;
• there exists an integrable random variable N ∈ L1(Ω, PR) such that Nt → N as t → +∞ both P-almost surely and in L1(Ω, PR), i.e.
\mathbf{E} \big[ \big| N_{t} - N \big| \big] = \int_{\Omega} \big| N_{t} (\omega) - N (\omega) \big| \, \mathrm{d} \mathbf{P} (\omega) \to 0 \mbox{ as } t \to + \infty.

### Corollary: convergence theorem for continuous martingales

Let M : [0, +∞) × Ω → R be a continuous martingale such that

\sup_{t > 0} \mathbf{E} \big[ \big| M_{t} \big|^{p} \big] < + \infty

for some p > 1. Then there exists a random variable M ∈ Lp(Ω, PR) such that Mt → M as t → +∞ both P-almost surely and in Lp(Ω, PR).

## Discrete-time results

Similar results can be obtained for discrete-time supermartingales and submartingales, the obvious difference being that no continuity assumptions are required. For example, the result above becomes

Let M : N × Ω → R be a discrete-time martingale such that

\sup_{k \in \mathbf{N}} \mathbf{E} \big[ \big| M_{k} \big|^{p} \big] < + \infty

for some p > 1. Then there exists a random variable M ∈ Lp(Ω, PR) such that Mk → M as k → +∞ both P-almost surely and in Lp(Ω, PR)

## Convergence of conditional expectations: Lévy's zero–one law

Doob's martingale convergence theorems imply that conditional expectations also have a convergence property.

Let (Ω, FP) be a probability space and let X be a random variable in L1. Let F = (Fk)kN be any filtration of F, and define F to be the minimal σ-algebra generated by (Fk)kN. Then

\mathbf{E} \big[ X \big| F_{k} \big] \to \mathbf{E} \big[ X \big| F_{\infty} \big] \mbox{ as } k \to \infty

both P-almost surely and in L1.

This result is usually called Lévy's zero–one law. The reason for the name is that if A is an event in F, then the theorem says that \mathbf{P}[ A | F_{k} ] \to \mathbf{1}_A almost surely, i.e., the limit of the probabilities is 0 or 1. In plain language, if we are learning gradually all the information that determines the outcome of an event, then we will become gradually certain what the outcome will be. This sounds almost like a tautology, but the result is still non-trivial. For instance, it easily implies Kolmogorov's zero–one law, since it says that for any tail event A, we must have \mathbf{P}[ A ] = \mathbf{1}_A almost surely, hence \mathbf{P}[ A ]\in\{0,1\}.

## References

• (See Appendix C)
• Durrett, Rick (1996). Probability: theory and examples (Second ed.). Duxbury Press.
This article was sourced from Creative Commons Attribution-ShareAlike License; additional terms may apply. World Heritage Encyclopedia content is assembled from numerous content providers, Open Access Publishing, and in compliance with The Fair Access to Science and Technology Research Act (FASTR), Wikimedia Foundation, Inc., Public Library of Science, The Encyclopedia of Life, Open Book Publishers (OBP), PubMed, U.S. National Library of Medicine, National Center for Biotechnology Information, U.S. National Library of Medicine, National Institutes of Health (NIH), U.S. Department of Health & Human Services, and USA.gov, which sources content from all federal, state, local, tribal, and territorial government publication portals (.gov, .mil, .edu). Funding for USA.gov and content contributors is made possible from the U.S. Congress, E-Government Act of 2002.

Crowd sourced content that is contributed to World Heritage Encyclopedia is peer reviewed and edited by our editorial staff to ensure quality scholarly research articles.