Jump to content
Wikipedia The Free Encyclopedia

Doob decomposition theorem

From Wikipedia, the free encyclopedia
Mathematical theorem in stochastic processes

In the theory of stochastic processes in discrete time, a part of the mathematical theory of probability, the Doob decomposition theorem gives a unique decomposition of every adapted and integrable stochastic process as the sum of a martingale and a predictable process (or "drift") starting at zero. The theorem was proved by and is named for Joseph L. Doob.[1]

The analogous theorem in the continuous-time case is the Doob–Meyer decomposition theorem.

Statement

[edit ]

Let ( Ω , F , P ) {\displaystyle (\Omega ,{\mathcal {F}},\mathbb {P} )} {\displaystyle (\Omega ,{\mathcal {F}},\mathbb {P} )} be a probability space, I = {0, 1, 2, ..., N} with N N {\displaystyle N\in \mathbb {N} } {\displaystyle N\in \mathbb {N} } or I = N 0 {\displaystyle I=\mathbb {N} _{0}} {\displaystyle I=\mathbb {N} _{0}} a finite or countably infinite index set, ( F n ) n I {\displaystyle ({\mathcal {F}}_{n})_{n\in I}} {\displaystyle ({\mathcal {F}}_{n})_{n\in I}} a filtration of  F {\displaystyle {\mathcal {F}}} {\displaystyle {\mathcal {F}}}, and X = (Xn)nI an adapted stochastic process with E[|Xn|] < ∞ for all nI. Then there exist a martingale M = (Mn)nI and an integrable predictable process A = (An)nI starting with A0 = 0 such that Xn = Mn + An for every nI. Here predictable means that An is F n 1 {\displaystyle {\mathcal {F}}_{n-1}} {\displaystyle {\mathcal {F}}_{n-1}}-measurable for every nI \ {0}. This decomposition is almost surely unique.[2] [3] [4]

Remark

[edit ]

The theorem is valid word for word also for stochastic processes X taking values in the d-dimensional Euclidean space R d {\displaystyle \mathbb {R} ^{d}} {\displaystyle \mathbb {R} ^{d}} or the complex vector space C d {\displaystyle \mathbb {C} ^{d}} {\displaystyle \mathbb {C} ^{d}}. This follows from the one-dimensional version by considering the components individually.

Proof

[edit ]

Existence

[edit ]

Using conditional expectations, define the processes A and M, for every nI, explicitly by

A n = k = 1 n ( E [ X k | F k 1 ] X k 1 ) {\displaystyle A_{n}=\sum _{k=1}^{n}{\bigl (}\mathbb {E} [X_{k},円|,円{\mathcal {F}}_{k-1}]-X_{k-1}{\bigr )}} {\displaystyle A_{n}=\sum _{k=1}^{n}{\bigl (}\mathbb {E} [X_{k},円|,円{\mathcal {F}}_{k-1}]-X_{k-1}{\bigr )}} 1

and

M n = X 0 + k = 1 n ( X k E [ X k | F k 1 ] ) , {\displaystyle M_{n}=X_{0}+\sum _{k=1}^{n}{\bigl (}X_{k}-\mathbb {E} [X_{k},円|,円{\mathcal {F}}_{k-1}]{\bigr )},} {\displaystyle M_{n}=X_{0}+\sum _{k=1}^{n}{\bigl (}X_{k}-\mathbb {E} [X_{k},円|,円{\mathcal {F}}_{k-1}]{\bigr )},} 2

where the sums for n = 0 are empty and defined as zero. Here A adds up the expected increments of X, and M adds up the surprises, i.e., the part of every Xk that is not known one time step before. Due to these definitions, An+1 (if n + 1 ∈ I) and Mn are Fn-measurable because the process X is adapted, E[|An|] < ∞ and E[|Mn|] < ∞ because the process X is integrable, and the decomposition Xn = Mn + An is valid for every nI. The martingale property

E [ M n M n 1 | F n 1 ] = 0 {\displaystyle \mathbb {E} [M_{n}-M_{n-1},円|,円{\mathcal {F}}_{n-1}]=0} {\displaystyle \mathbb {E} [M_{n}-M_{n-1},円|,円{\mathcal {F}}_{n-1}]=0}    a.s.

also follows from the above definition (2 ), for every nI \ {0}.

Uniqueness

[edit ]

To prove uniqueness, let X = M' + A' be an additional decomposition. Then the process Y := MM' = A'A is a martingale, implying that

E [ Y n | F n 1 ] = Y n 1 {\displaystyle \mathbb {E} [Y_{n},円|,円{\mathcal {F}}_{n-1}]=Y_{n-1}} {\displaystyle \mathbb {E} [Y_{n},円|,円{\mathcal {F}}_{n-1}]=Y_{n-1}}    a.s.,

and also predictable, implying that

E [ Y n | F n 1 ] = Y n {\displaystyle \mathbb {E} [Y_{n},円|,円{\mathcal {F}}_{n-1}]=Y_{n}} {\displaystyle \mathbb {E} [Y_{n},円|,円{\mathcal {F}}_{n-1}]=Y_{n}}    a.s.

for any nI \ {0}. Since Y0 = A'0A0 = 0 by the convention about the starting point of the predictable processes, this implies iteratively that Yn = 0 almost surely for all nI, hence the decomposition is almost surely unique.

Corollary

[edit ]

A real-valued stochastic process X is a submartingale if and only if it has a Doob decomposition into a martingale M and an integrable predictable process A that is almost surely increasing.[5] It is a supermartingale, if and only if A is almost surely decreasing.

Proof

[edit ]

If X is a submartingale, then

E [ X k | F k 1 ] X k 1 {\displaystyle \mathbb {E} [X_{k},円|,円{\mathcal {F}}_{k-1}]\geq X_{k-1}} {\displaystyle \mathbb {E} [X_{k},円|,円{\mathcal {F}}_{k-1}]\geq X_{k-1}}    a.s.

for all kI \ {0}, which is equivalent to saying that every term in definition (1 ) of A is almost surely positive, hence A is almost surely increasing. The equivalence for supermartingales is proved similarly.

Example

[edit ]

Let X = (Xn)n N 0 {\displaystyle \mathbb {N} _{0}} {\displaystyle \mathbb {N} _{0}} be a sequence in independent, integrable, real-valued random variables. They are adapted to the filtration generated by the sequence, i.e. Fn = σ(X0, . . . , Xn) for all n N 0 . {\displaystyle n\in \mathbb {N} _{0}.} {\displaystyle n\in \mathbb {N} _{0}.} By (1 ) and (2 ), the Doob decomposition is given by

A n = k = 1 n ( E [ X k ] X k 1 ) , n N 0 , {\displaystyle A_{n}=\sum _{k=1}^{n}{\bigl (}\mathbb {E} [X_{k}]-X_{k-1}{\bigr )},\quad n\in \mathbb {N} _{0},} {\displaystyle A_{n}=\sum _{k=1}^{n}{\bigl (}\mathbb {E} [X_{k}]-X_{k-1}{\bigr )},\quad n\in \mathbb {N} _{0},}

and

M n = X 0 + k = 1 n ( X k E [ X k ] ) , n N 0 . {\displaystyle M_{n}=X_{0}+\sum _{k=1}^{n}{\bigl (}X_{k}-\mathbb {E} [X_{k}]{\bigr )},\quad n\in \mathbb {N} _{0}.} {\displaystyle M_{n}=X_{0}+\sum _{k=1}^{n}{\bigl (}X_{k}-\mathbb {E} [X_{k}]{\bigr )},\quad n\in \mathbb {N} _{0}.}

If the random variables of the original sequence X have mean zero, this simplifies to

A n = k = 0 n 1 X k {\displaystyle A_{n}=-\sum _{k=0}^{n-1}X_{k}} {\displaystyle A_{n}=-\sum _{k=0}^{n-1}X_{k}}    and     M n = k = 0 n X k , n N 0 , {\displaystyle M_{n}=\sum _{k=0}^{n}X_{k},\quad n\in \mathbb {N} _{0},} {\displaystyle M_{n}=\sum _{k=0}^{n}X_{k},\quad n\in \mathbb {N} _{0},}

hence both processes are (possibly time-inhomogeneous) random walks. If the sequence X = (Xn)n N 0 {\displaystyle \mathbb {N} _{0}} {\displaystyle \mathbb {N} _{0}} consists of symmetric random variables taking the values +1 and −1, then X is bounded, but the martingale M and the predictable process A are unbounded simple random walks (and not uniformly integrable), and Doob's optional stopping theorem might not be applicable to the martingale M unless the stopping time has a finite expectation.

Application

[edit ]

In mathematical finance, the Doob decomposition theorem can be used to determine the largest optimal exercise time of an American option.[6] [7] Let X = (X0, X1, . . . , XN) denote the non-negative, discounted payoffs of an American option in a N-period financial market model, adapted to a filtration (F0, F1, . . . , FN), and let Q {\displaystyle \mathbb {Q} } {\displaystyle \mathbb {Q} }g denote an equivalent martingale measure. Let U = (U0, U1, . . . , UN) denote the Snell envelope of X with respect to Q {\displaystyle \mathbb {Q} } {\displaystyle \mathbb {Q} }. The Snell envelope is the smallest Q {\displaystyle \mathbb {Q} } {\displaystyle \mathbb {Q} }-supermartingale dominating X[8] and in a complete financial market it represents the minimal amount of capital necessary to hedge the American option up to maturity.[9] Let U = M + A denote the Doob decomposition with respect to Q {\displaystyle \mathbb {Q} } {\displaystyle \mathbb {Q} } of the Snell envelope U into a martingale M = (M0, M1, . . . , MN) and a decreasing predictable process A = (A0, A1, . . . , AN) with A0 = 0. Then the largest stopping time to exercise the American option in an optimal way[10] [11] is

τ max := { N if  A N = 0 , min { n { 0 , , N 1 } A n + 1 < 0 } if  A N < 0. {\displaystyle \tau _{\text{max}}:={\begin{cases}N&{\text{if }}A_{N}=0,\\\min\{n\in \{0,\dots ,N-1\}\mid A_{n+1}<0\}&{\text{if }}A_{N}<0.\end{cases}}} {\displaystyle \tau _{\text{max}}:={\begin{cases}N&{\text{if }}A_{N}=0,\\\min\{n\in \{0,\dots ,N-1\}\mid A_{n+1}<0\}&{\text{if }}A_{N}<0.\end{cases}}}

Since A is predictable, the event {τmax = n} = {An = 0, An+1 < 0} is in Fn for every n ∈ {0, 1, . . . , N − 1}, hence τmax is indeed a stopping time. It gives the last moment before the discounted value of the American option will drop in expectation; up to time τmax the discounted value process U is a martingale with respect to Q {\displaystyle \mathbb {Q} } {\displaystyle \mathbb {Q} }.

Generalization

[edit ]

The Doob decomposition theorem can be generalized from probability spaces to σ-finite measure spaces.[12]

Citations

[edit ]
  1. ^ Doob (1953), see (Doob 1990, pp. 296−298)
  2. ^ Durrett (2010)
  3. ^ (Föllmer & Schied 2011, Proposition 6.1)
  4. ^ (Williams 1991, Section 12.11, part (a) of the Theorem)
  5. ^ (Williams 1991, Section 12.11, part (b) of the Theorem)
  6. ^ (Lamberton & Lapeyre 2008, Chapter 2: Optimal stopping problem and American options)
  7. ^ (Föllmer & Schied 2011, Chapter 6: American contingent claims)
  8. ^ (Föllmer & Schied 2011, Proposition 6.10)
  9. ^ (Föllmer & Schied 2011, Theorem 6.11)
  10. ^ (Lamberton & Lapeyre 2008, Proposition 2.3.2)
  11. ^ (Föllmer & Schied 2011, Theorem 6.21)
  12. ^ (Schilling 2005, Problem 23.11)

References

[edit ]
Discrete time
Continuous time
Both
Fields and other
Time series models
Financial models
Actuarial models
Queueing models
Properties
Limit theorems
Inequalities
Tools
Disciplines

AltStyle によって変換されたページ (->オリジナル) /