Jump to content
Wikipedia The Free Encyclopedia

Optional stopping theorem

From Wikipedia, the free encyclopedia
A martingale's expected value at a stopping time equals its initial expected value
Not to be confused with Optimal stopping.
This article needs additional citations for verification . Please help improve this article by adding citations to reliable sources. Unsourced material may be challenged and removed.
Find sources: "Optional stopping theorem" – news · newspapers · books · scholar · JSTOR
(February 2012) (Learn how and when to remove this message)

In probability theory, the optional stopping theorem (or sometimes Doob's optional sampling theorem, for American probabilist Joseph Doob) says that, under certain conditions, the expected value of a martingale at a stopping time is equal to its initial expected value. Since martingales can be used to model the wealth of a gambler participating in a fair game, the optional stopping theorem says that, on average, nothing can be gained by stopping play based on the information obtainable so far (i.e., without looking into the future). Certain conditions are necessary for this result to hold true. In particular, the theorem applies to doubling strategies.

The optional stopping theorem is an important tool of mathematical finance in the context of the fundamental theorem of asset pricing.

Statement

[edit ]

A discrete-time version of the theorem is given below, with N 0 {\displaystyle \mathbb {N} _{0}} {\displaystyle \mathbb {N} _{0}} denoting the set of natural numbers, including zero.

Let X = ( X t ) t N 0 {\displaystyle X=(X_{t})_{t\in \mathbb {N} _{0}}} {\displaystyle X=(X_{t})_{t\in \mathbb {N} _{0}}} be a discrete-time martingale and τ {\displaystyle \tau } {\displaystyle \tau } a stopping time with values in N 0 { } {\displaystyle \mathbb {N} _{0}\cup \{\infty \}} {\displaystyle \mathbb {N} _{0}\cup \{\infty \}}, both with respect to a filtration ( F t ) t N 0 {\displaystyle ({\mathcal {F}}_{t})_{t\in \mathbb {N} _{0}}} {\displaystyle ({\mathcal {F}}_{t})_{t\in \mathbb {N} _{0}}}. Assume that one of the following three conditions holds:

(a) The stopping time τ {\displaystyle \tau } {\displaystyle \tau } is almost surely bounded, i.e., there exists a constant c N {\displaystyle c\in \mathbb {N} } {\displaystyle c\in \mathbb {N} } such that τ c {\displaystyle \tau \leq c} {\displaystyle \tau \leq c} almost surely
(b) The stopping time τ {\displaystyle \tau } {\displaystyle \tau } has finite expectation and the conditional expectations of the absolute value of the martingale increments are almost surely bounded, more precisely, E [ τ ] < {\displaystyle \mathbb {E} [\tau ]<\infty } {\displaystyle \mathbb {E} [\tau ]<\infty } and there exists a constant c {\displaystyle c} {\displaystyle c} such that E [ | X t + 1 X t | | F t ] c {\displaystyle \mathbb {E} {\bigl [}|X_{t+1}-X_{t}|,円{\big \vert },円{\mathcal {F}}_{t}{\bigr ]}\leq c} {\displaystyle \mathbb {E} {\bigl [}|X_{t+1}-X_{t}|,円{\big \vert },円{\mathcal {F}}_{t}{\bigr ]}\leq c} almost surely on the event { τ > t } {\displaystyle \{\tau >t\}} {\displaystyle \{\tau >t\}} for all t N 0 {\displaystyle t\in \mathbb {N} _{0}} {\displaystyle t\in \mathbb {N} _{0}}.
(c) There exists a constant c {\displaystyle c} {\displaystyle c} such that | X min { t , τ } | c {\displaystyle |X_{\min\{t,\tau \}}|\leq c} {\displaystyle |X_{\min\{t,\tau \}}|\leq c} almost surely for all t N 0 {\displaystyle t\in \mathbb {N} _{0}} {\displaystyle t\in \mathbb {N} _{0}}.

Then X τ {\displaystyle X_{\tau }} {\displaystyle X_{\tau }} is an almost surely well defined random variable and E [ X τ ] = E [ X 0 ] {\displaystyle \mathbb {E} [X_{\tau }]=\mathbb {E} [X_{0}]} {\displaystyle \mathbb {E} [X_{\tau }]=\mathbb {E} [X_{0}]}.

Similarly, if the stochastic process X = ( X t ) t N 0 {\displaystyle X=(X_{t})_{t\in \mathbb {N} _{0}}} {\displaystyle X=(X_{t})_{t\in \mathbb {N} _{0}}} is a submartingale or a supermartingale and one of the above conditions holds, then E [ X τ ] E [ X 0 ] {\displaystyle \mathbb {E} [X_{\tau }]\geq \mathbb {E} [X_{0}]} {\displaystyle \mathbb {E} [X_{\tau }]\geq \mathbb {E} [X_{0}]} for a submartingale, and E [ X τ ] E [ X 0 ] {\displaystyle \mathbb {E} [X_{\tau }]\leq \mathbb {E} [X_{0}]} {\displaystyle \mathbb {E} [X_{\tau }]\leq \mathbb {E} [X_{0}]} for a supermartingale.

Remark

[edit ]

Under condition (c ) it is possible that τ = {\displaystyle \tau =\infty } {\displaystyle \tau =\infty } happens with positive probability. On this event X τ {\displaystyle X_{\tau }} {\displaystyle X_{\tau }} is defined as the almost surely existing pointwise limit of X = ( X t ) t N 0 {\displaystyle X=(X_{t})_{t\in \mathbb {N} _{0}}} {\displaystyle X=(X_{t})_{t\in \mathbb {N} _{0}}}. See the proof below for details.

Applications

[edit ]
This section does not cite any sources . Please help improve this section by adding citations to reliable sources. Unsourced material may be challenged and removed. (May 2024) (Learn how and when to remove this message)
  • The optional stopping theorem can be used to prove the impossibility of successful betting strategies for a gambler with a finite lifetime (which gives condition (a )) or a house limit on bets (condition (b )). Suppose that the gambler can wager up to c {\displaystyle c} {\displaystyle c} dollars on a fair coin flip at times 1, 2, 3, etc., winning their wager if the coin comes up heads and losing it if the coin comes up tails. Suppose further that the gambler can quit whenever they like, but cannot predict the outcome of gambles that have not happened yet. Then the gambler's fortune over time is a martingale, and the time τ {\displaystyle \tau } {\displaystyle \tau } at which they decide to quit (or go broke and are forced to quit) is a stopping time. So the theorem says that E [ X τ ] = E [ X 0 ] {\displaystyle \mathbb {E} [X_{\tau }]=\mathbb {E} [X_{0}]} {\displaystyle \mathbb {E} [X_{\tau }]=\mathbb {E} [X_{0}]}. In other words, the gambler leaves with the same amount of money on average as when they started. (The same result holds if the gambler, instead of having a house limit on individual bets, has a finite limit on their line of credit or how far in debt they may go, though this is easier to show with another version of the theorem.)
  • Suppose a random walk starting at a 0 {\displaystyle a\geq 0} {\displaystyle a\geq 0} that goes up or down by one with equal probability on each step. Suppose further that the walk stops if it reaches 0 or m a {\displaystyle m\geq a} {\displaystyle m\geq a}; the time at which this first occurs is a stopping time. If it is known that the expected time at which the walk ends is finite (say, from Markov chain theory), the optional stopping theorem predicts that the expected stop position is equal to the initial position a {\displaystyle a} {\displaystyle a}. Solving a = p m + ( 1 p ) 0 {\displaystyle a=pm+(1-p)0} {\displaystyle a=pm+(1-p)0} for the probability p {\displaystyle p} {\displaystyle p} that the walk reaches m {\displaystyle m} {\displaystyle m} before 0 gives p = a / m {\displaystyle p=a/m} {\displaystyle p=a/m}.
  • Now consider a random walk X {\displaystyle X} {\displaystyle X} that starts at 0 and stops if it reaches m {\displaystyle -m} {\displaystyle -m} or + m {\displaystyle +m} {\displaystyle +m}, and use the Y n = X n 2 n {\displaystyle Y_{n}=X_{n}^{2}-n} {\displaystyle Y_{n}=X_{n}^{2}-n} martingale from Martingale (probability theory) § Examples of martingales. If τ {\displaystyle \tau } {\displaystyle \tau } is the time at which X {\displaystyle X} {\displaystyle X} first reaches ± m {\displaystyle \pm m} {\displaystyle \pm m}, then 0 = E [ Y 0 ] = E [ Y τ ] = m 2 = E [ τ ] {\displaystyle 0=\mathbb {E} [Y_{0}]=\mathbb {E} [Y_{\tau }]=m^{2}-=\mathbb {E} [\tau ]} {\displaystyle 0=\mathbb {E} [Y_{0}]=\mathbb {E} [Y_{\tau }]=m^{2}-=\mathbb {E} [\tau ]}. This gives E [ τ ] = m 2 {\displaystyle \mathbb {E} [\tau ]=m^{2}} {\displaystyle \mathbb {E} [\tau ]=m^{2}}.
  • Care must be taken, however, to ensure that one of the conditions of the theorem hold. For example, suppose the last example had instead used a 'one-sided' stopping time, so that stopping only occurred at + m {\displaystyle +m} {\displaystyle +m}, not at m {\displaystyle -m} {\displaystyle -m}. The value of X {\displaystyle X} {\displaystyle X} at this stopping time would therefore be m {\displaystyle m} {\displaystyle m}. Therefore, the expectation value E [ τ ] {\displaystyle \mathbb {E} [\tau ]} {\displaystyle \mathbb {E} [\tau ]} must also be m {\displaystyle m} {\displaystyle m}, seemingly in violation of the theorem which would give E [ τ ] = 0 {\displaystyle \mathbb {E} [\tau ]=0} {\displaystyle \mathbb {E} [\tau ]=0}. The failure of the optional stopping theorem shows that all three of the conditions fail.

Proof

[edit ]

Let X τ {\displaystyle X^{\tau }} {\displaystyle X^{\tau }} denote the stopped process, it is also a martingale (or a submartingale or supermartingale, respectively). Under condition (a ) or (b ), the random variable X τ {\displaystyle X^{\tau }} {\displaystyle X^{\tau }} is well defined. Under condition (c ) the stopped process X τ {\displaystyle X^{\tau }} {\displaystyle X^{\tau }} is bounded, hence by Doob's martingale convergence theorem it converges almost surely pointwise to a random variable which we call X τ {\displaystyle X_{\tau }} {\displaystyle X_{\tau }}.

If condition (c ) holds, then the stopped process X τ {\displaystyle X^{\tau }} {\displaystyle X^{\tau }} is bounded by the constant random variable M := c {\displaystyle M:=c} {\displaystyle M:=c}. Otherwise, writing the stopped process as X t τ = X 0 + s = 0 τ 1 t 1 ( X s + 1 X s ) , t N 0 {\displaystyle X_{t}^{\tau }=X_{0}+\sum _{s=0}^{\tau -1\land t-1}(X_{s+1}-X_{s}),\quad t\in {\mathbb {N} }_{0}} {\displaystyle X_{t}^{\tau }=X_{0}+\sum _{s=0}^{\tau -1\land t-1}(X_{s+1}-X_{s}),\quad t\in {\mathbb {N} }_{0}} gives X t τ M {\displaystyle X_{t}^{\tau }\leq M} {\displaystyle X_{t}^{\tau }\leq M} for all t N 0 {\displaystyle t\in \mathbb {N} _{0}} {\displaystyle t\in \mathbb {N} _{0}}, where M := | X 0 | + s = 0 τ 1 | X s + 1 X s | = | X 0 | + s = 0 | X s + 1 X s | 1 { τ > s } . {\displaystyle M:=|X_{0}|+\sum _{s=0}^{\tau -1}|X_{s+1}-X_{s}|=|X_{0}|+\sum _{s=0}^{\infty }|X_{s+1}-X_{s}|\cdot \mathbf {1} _{\{\tau >s\}}.} {\displaystyle M:=|X_{0}|+\sum _{s=0}^{\tau -1}|X_{s+1}-X_{s}|=|X_{0}|+\sum _{s=0}^{\infty }|X_{s+1}-X_{s}|\cdot \mathbf {1} _{\{\tau >s\}}.}

By the monotone convergence theorem E [ M ] = E [ | X 0 | ] + s = 0 E [ | X s + 1 X s | 1 { τ > s } ] . {\displaystyle \mathbb {E} [M]=\mathbb {E} [|X_{0}|]+\sum _{s=0}^{\infty }\mathbb {E} {\bigl [}|X_{s+1}-X_{s}|\cdot \mathbf {1} _{\{\tau >s\}}{\bigr ]}.} {\displaystyle \mathbb {E} [M]=\mathbb {E} [|X_{0}|]+\sum _{s=0}^{\infty }\mathbb {E} {\bigl [}|X_{s+1}-X_{s}|\cdot \mathbf {1} _{\{\tau >s\}}{\bigr ]}.}

If condition (a ) holds, then this series only has a finite number of non-zero terms, hence M {\displaystyle M} {\displaystyle M} is integrable.

If condition (b ) holds, then we continue by inserting a conditional expectation and using that the event { τ > s } {\displaystyle \{\tau >s\}} {\displaystyle \{\tau >s\}} is known at time s {\displaystyle s} {\displaystyle s} (note that τ {\displaystyle \tau } {\displaystyle \tau } is assumed to be a stopping time with respect to the filtration), hence E [ M ] = E [ | X 0 | ] + s = 0 E [ E [ | X s + 1 X s | | F s ] 1 { τ > s } c 1 { τ > s }  a.s. by (b) ] E [ | X 0 | ] + c s = 0 P ( τ > s ) = E [ | X 0 | ] + c E [ τ ] < , {\displaystyle {\begin{aligned}\mathbb {E} [M]&=\mathbb {E} [|X_{0}|]+\sum _{s=0}^{\infty }\mathbb {E} {\bigl [}\underbrace {\mathbb {E} {\bigl [}|X_{s+1}-X_{s}|{\big |}{\mathcal {F}}_{s}{\bigr ]}\cdot \mathbf {1} _{\{\tau >s\}}} _{\leq ,円c,円\mathbf {1} _{\{\tau >s\}}{\text{ a.s. by (b)}}}{\bigr ]}\\&\leq \mathbb {E} [|X_{0}|]+c\sum _{s=0}^{\infty }\mathbb {P} (\tau >s)\\&=\mathbb {E} [|X_{0}|]+c,円\mathbb {E} [\tau ]<\infty ,\\\end{aligned}}} {\displaystyle {\begin{aligned}\mathbb {E} [M]&=\mathbb {E} [|X_{0}|]+\sum _{s=0}^{\infty }\mathbb {E} {\bigl [}\underbrace {\mathbb {E} {\bigl [}|X_{s+1}-X_{s}|{\big |}{\mathcal {F}}_{s}{\bigr ]}\cdot \mathbf {1} _{\{\tau >s\}}} _{\leq ,円c,円\mathbf {1} _{\{\tau >s\}}{\text{ a.s. by (b)}}}{\bigr ]}\\&\leq \mathbb {E} [|X_{0}|]+c\sum _{s=0}^{\infty }\mathbb {P} (\tau >s)\\&=\mathbb {E} [|X_{0}|]+c,円\mathbb {E} [\tau ]<\infty ,\\\end{aligned}}} where a representation of the expected value of non-negative integer-valued random variables is used for the last equality.

Therefore, under any one of the three conditions in the theorem, the stopped process is dominated by an integrable random variable M {\displaystyle M} {\displaystyle M}. Since the stopped process X τ {\displaystyle X^{\tau }} {\displaystyle X^{\tau }} converges almost surely to X τ {\displaystyle X_{\tau }} {\displaystyle X_{\tau }}, the dominated convergence theorem implies E [ X τ ] = lim t E [ X t τ ] . {\displaystyle \mathbb {E} [X_{\tau }]=\lim _{t\to \infty }\mathbb {E} [X_{t}^{\tau }].} {\displaystyle \mathbb {E} [X_{\tau }]=\lim _{t\to \infty }\mathbb {E} [X_{t}^{\tau }].}

By the martingale property of the stopped process, E [ X t τ ] = E [ X 0 ] , t N 0 , {\displaystyle \mathbb {E} [X_{t}^{\tau }]=\mathbb {E} [X_{0}],\quad t\in {\mathbb {N} }_{0},} {\displaystyle \mathbb {E} [X_{t}^{\tau }]=\mathbb {E} [X_{0}],\quad t\in {\mathbb {N} }_{0},} hence E [ X τ ] = E [ X 0 ] . {\displaystyle \mathbb {E} [X_{\tau }]=\mathbb {E} [X_{0}].} {\displaystyle \mathbb {E} [X_{\tau }]=\mathbb {E} [X_{0}].}

Similarly, if X {\displaystyle X} {\displaystyle X} is a submartingale or supermartingale, respectively, change the equality in the last two formulas to the appropriate inequality.

References

[edit ]
  1. Grimmett, Geoffrey R.; Stirzaker, David R. (2001). Probability and Random Processes (3rd ed.). Oxford University Press. pp. 491–495. ISBN 9780198572220.
  2. Bhattacharya, Rabi; Waymire, Edward C. (2007). A Basic Course in Probability Theory. Springer. pp. 43–45. ISBN 978-0-387-71939-9.
[edit ]

AltStyle によって変換されたページ (->オリジナル) /