# Doob martingale

Jump to navigation Jump to search

A Doob martingale (also known as a Levy martingale) is a mathematical construction of a stochastic process which approximates a given random variable and has the martingale property with respect to the given filtration. It may be thought of as the evolving sequence of best approximations to the random variable based on information accumulated up to a certain time.

When analyzing sums, random walks, or other additive functions of independent random variables, one can often apply the central limit theorem, law of large numbers, Chernoff's inequality, Chebyshev's inequality or similar tools. When analyzing similar objects where the differences are not independent, the main tools are martingales and Azuma's inequality.[clarification needed]

## Definition

A Doob martingale (named after Joseph L. Doob) is a generic construction that is always a martingale. Specifically, consider any set of random variables

${\vec {X}}=X_{1},X_{2},...,X_{n}$ taking values in a set $A$ for which we are interested in the function $f:A^{n}\to \mathbb {R}$ and define:

$B_{i}=E[f({\vec {X}})|X_{1},X_{2},...X_{i}]$ where the above expectation is itself a random quantity since the expectation is only taken over $X_{i+1},X_{i+2},...,X_{n},$ and $X_{1},X_{2},...X_{i}$ are treated as random variables. It is possible to show that $B_{i}$ is always a martingale regardless of the properties of $X_{i}$ .[citation needed]

The sequence ${B_{i}}$ is the Doob martingale for f.

## Application

Thus if one can bound the differences

$|B_{i+1}-B_{i}|$ ,

one can apply Azuma's inequality and show that with high probability $f({\vec {X}})$ is concentrated around its expected value

$E[f({\vec {X}})]=B_{0}.$ ## McDiarmid's inequality

One common way of bounding the differences and applying Azuma's inequality to a Doob martingale is called McDiarmid's inequality.

Suppose $X_{1},X_{2},\dots ,X_{n}$ are independent and assume that $f$ satisfies

$\sup _{x_{1},x_{2},\dots ,x_{n},{\hat {x}}_{i}}|f(x_{1},x_{2},\dots ,x_{n})-f(x_{1},x_{2},\dots ,x_{i-1},{\hat {x}}_{i},x_{i+1},\dots ,x_{n})|\leq c_{i}\qquad {\text{for}}\quad 1\leq i\leq n\;.$ (In other words, replacing the $i$ -th coordinate $x_{i}$ by some other value changes the value of $f$ by at most $c_{i}$ .)

It follows that

$|B_{i+1}-B_{i}|\leq c_{i}$ and therefore Azuma's inequality yields the following McDiarmid inequalities for any $\varepsilon >0$ :

$\Pr \left\{f(X_{1},X_{2},\dots ,X_{n})-E[f(X_{1},X_{2},\dots ,X_{n})]\geq \varepsilon \right\}\leq \exp \left(-{\frac {2\varepsilon ^{2}}{\sum _{i=1}^{n}c_{i}^{2}}}\right)$ and

$\Pr \left\{E[f(X_{1},X_{2},\dots ,X_{n})]-f(X_{1},X_{2},\dots ,X_{n})\geq \varepsilon \right\}\leq \exp \left(-{\frac {2\varepsilon ^{2}}{\sum _{i=1}^{n}c_{i}^{2}}}\right)$ and

$\Pr \left\{|E[f(X_{1},X_{2},\dots ,X_{n})]-f(X_{1},X_{2},\dots ,X_{n})|\geq \varepsilon \right\}\leq 2\exp \left(-{\frac {2\varepsilon ^{2}}{\sum _{i=1}^{n}c_{i}^{2}}}\right).\;$ 