Conditional Expectation
Suppose , , and
We define
If have a joint probability density function , and
We can easily define continuous analogues,
,
And
is characterized by two properties,
- depends only on the values of (we can write it as for some ). We call measurable with respect to .
- Suppose is an event that depends on and is the indicator function of . .
Measure theoretic treatments of probabilities define conditional expectation as the unique random variable satisfying these two properties
As shorthand, we write , where denotes the information contained in
This gives us
And
If is a function of then
For any , if , then
If is independent of ,
If is any random variable and is a random variable that is measurable with , then
Example:
Suppose are i.i.d. with and let
Let denote the information in and suppose
So
Example:
Assume the same as above, but
Altogether,
Example:
Consider the first example where is Bernoulli and
Consider where is an infinite collection of random variables
Let denote the information in
is -measurable if knowledge of determines , i.e. for some and some finite subcollection, or if is a limit of such random variables
As an example, say are independent random variables and are i.i.d. with the standard normal distribution, while is unknown, define , and let denote the information in
One cannot determine given any finite set , however is measurable since
An event is -measurable if is an -measurable random variable
is defined as the unique -measurable random variable such that for all -measurable events A
I.e.
Martingales
A martingale is a model of a fair game
We let denote an increasing collection of information, a collection of random variables such that if , meaning we don’t lose information
The increasing sequence of is known as a filtration
A sequence of random variables with is a martingale with respect to if
- Each is measurable with respect to and
- for all
The second condition is equivalent to and can be proven by showing
Example:
is a martingale with respect to
Example:
Suppose
Say we double our bet every instance and stop when we win and let denote the winnings or losses up through flips of the coin,
If the first flips have turned up tails,
We can verify , so is a martingale with respect to
We can generalize this, such that we make a bet on the th flip, measurable on and
We can let be negative, corresponding to betting the coin will come up tails
So the general form is also a Martingale
Example:
Consider an urn with a red ball and a green ball. Every time one draws a ball, it is returned along with another of the same color. Let denote the number of red balls in the urn after draws
This is a Markov chain,
is a martingale
Since this is a Markov chain, all relevant information in is contained in ,
is called a submartingale if and a supermartingale if
Example:
Let be a finite Markov chain, with where is the optimal stopping rule
is a supermartingale with respect to
Optional Sampling Theorem
The optional sampling theorem states that you cannot beat a fair game, however it’s a bit subtle
We say is a stopping time with respect to is for each , is measurable with respect to
Under certain conditions,
Suppose is a martingale with respect to and suppose is a stopping time which is bounded, then
Putting it together, we get
We can do this argument again conditioning on to get
This can be iterated until we arrive at
However, is not always bounded, such as in the betting game we first introduced. So when is the optional sampling theorem valid?
Consider
since is bounded, but what about the other terms? The first one approaches 0 essentially because approaches 0 as .
So the problematic term for our nice result is , which does not necessarily approach 0. In our doubling betting example, and , so we get , which does not approach 0.
Optional Sampling Theorem: Suppose is a martingale with respect to and is a stopping time satisfying , , and , then .
Example:
Let be a simple random walk on with absorbing boundaries, let , and let
Since is bounded, we know immediately and
Therefore,
Now let
(using this), so this is a Martingale
Using the same as before, is no longer bounded, however we can show there exists and such that
Since , we can show and
Hence,
Since , we see
Uniform Integrability
is hard to verify, so we’d like to find some easier conditions
Suppose we have with ,
We say is uniformly integrable if for every there is some with for each , where depends on (not )
If is uniformly integrable then for every there is some where for each
To show this, let , so if then
Example:
Consider our martingale betting strategy, with random variables , and the event
This cannot satisfy the conditions for uniform integrability for any
Now suppose is a uniformly integrable martingale (with respect to ) and is a stopping time with
We have
So we therefore have
Optional Sampling Theorem (again): Suppose is a uniformly integrable martingale with respect to and is a stopping time with and , then
If is a sequence of random variable and there is such that for each , then the sequence is uniformly integrable
To show this with our previous definition, let and suppose
Example:
What if we set the the signs of the harmonic series randomly?
This is a martingale, since each has mean 0
So this is uniformly integrable
Martingale Convergence Theorem
This theorem states when a martingale converges to a limiting random variable . Let’s use Polya’s urn as an example
Let , suppose that , let , and
For , the optional sampling theorem says
But
So
This is true for all , so
In other words, with probability the proportion of red balls never gets as high as
If it does go up to (with probability ), the inverse of this argument says it’ll drop back down to with probability
The probability this continues happening is
This tells us the proportion cannot fluctuate infinitely between any two numbers, i.e. exists
It happens to be we can also prove that this setup results in a uniform distribution directly
Martingale Convergence Theorem: Suppose is a martingale with respect to such that there exists with for all . There there exists a random variable such that .
Fix and think of as representing the cumulative results of a fair game
Whenever , keep betting 1 until , then stop until again
Our winnings are then , where is 0 or 1 and is
Note that is a martingale and where denotes the number of “upcrossings” (passes between and )
This holds for all , so the expected number of upcrossings is bounded, meaning the number of upcrossings is always finite
Note that the martingale property does not imply that , such as in the martingale betting strategy
If is a uniformly integrable martingale then exists and
Example:
Let be the proportion of red balls in Polya’s urn and suppose that at time there are red balls and green balls. Since is bounded, it is uniformly integrable, and approaches with
Example:
Let be independent random variables with
Let and for , let
Note that
Therefore, is a martingale with respect to
Since , the conditions of the martingale convergence theorem hold and for some
Is uniformly integrable?
Therefore
, i.e. is not uniformly integrable
Example:
A normal problem in statistics is estimating the mean of a distribution given
In Bayesian statistics, the parameter is taken to be a random variable taken from a prior distribution
Under the prior distribution, , and let and
is a martingale and the conditional distribution on is called the posterior distribution
We also know
The strong law of large numbers says
So
Assume these samples are from a Bernoulli distribution with
This is the beta distribution with parameters and
The mean happens to be
Maximal Inequalities
If is a sequence of random variables, defined the maximum processes as and
Maximal inequalities relate probabilities or expectations for to respectively
Reflection Principle: Suppose are independent random variables whose distributions are symmetric about the origin, and let and . For every ,
To prove this, let
The independence and symmetry of shows that (noting that can be greater than makes this easier to understand)
Doob’s Maximal Inequality: Suppose is a nonnegative submartingale with respect to . Then for every ,
Let denote the -measurable event (same as before),
So
If is not necessarily nonnegative, we cannot immediately use the inequality, but there’s still an extension
If then is a submartingale,
Likewise, if ,
for all
Doob’s Maximal Inequality (again): Suppose is a martingale with respect to , then for every and ,
Example:
Let denote a simple random walk in and let
is a martingale so we have,
Taylor series shows that
Hence,
, i.e. the random walk in is recurrent