Conditional Expectation

Suppose , , and
We define

If have a joint probability density function , and

We can easily define continuous analogues,
,

And

is characterized by two properties,

  1. depends only on the values of (we can write it as for some ). We call measurable with respect to .
  2. Suppose is an event that depends on and is the indicator function of . .

Measure theoretic treatments of probabilities define conditional expectation as the unique random variable satisfying these two properties

As shorthand, we write , where denotes the information contained in

This gives us
And

If is a function of then

For any , if , then

If is independent of ,

If is any random variable and is a random variable that is measurable with , then

Example:
Suppose are i.i.d. with and let
Let denote the information in and suppose




So

Example:
Assume the same as above, but



Altogether,

Example:
Consider the first example where is Bernoulli and


Consider where is an infinite collection of random variables
Let denote the information in
is -measurable if knowledge of determines , i.e. for some and some finite subcollection, or if is a limit of such random variables

As an example, say are independent random variables and are i.i.d. with the standard normal distribution, while is unknown, define , and let denote the information in

One cannot determine given any finite set , however is measurable since

An event is -measurable if is an -measurable random variable
is defined as the unique -measurable random variable such that for all -measurable events A

I.e.

Martingales

A martingale is a model of a fair game

We let denote an increasing collection of information, a collection of random variables such that if , meaning we don’t lose information

The increasing sequence of is known as a filtration

A sequence of random variables with is a martingale with respect to if

  1. Each is measurable with respect to and
  2. for all

The second condition is equivalent to and can be proven by showing

Example:
is a martingale with respect to

Example:
Suppose

Say we double our bet every instance and stop when we win and let denote the winnings or losses up through flips of the coin,

If the first flips have turned up tails,

We can verify , so is a martingale with respect to

We can generalize this, such that we make a bet on the th flip, measurable on and

We can let be negative, corresponding to betting the coin will come up tails


So the general form is also a Martingale

Example:
Consider an urn with a red ball and a green ball. Every time one draws a ball, it is returned along with another of the same color. Let denote the number of red balls in the urn after draws

This is a Markov chain,

is a martingale


Since this is a Markov chain, all relevant information in is contained in ,


is called a submartingale if and a supermartingale if

Example:
Let be a finite Markov chain, with where is the optimal stopping rule

is a supermartingale with respect to

Optional Sampling Theorem

The optional sampling theorem states that you cannot beat a fair game, however it’s a bit subtle

We say is a stopping time with respect to is for each , is measurable with respect to

Under certain conditions,

Suppose is a martingale with respect to and suppose is a stopping time which is bounded, then






Putting it together, we get

We can do this argument again conditioning on to get

This can be iterated until we arrive at

However, is not always bounded, such as in the betting game we first introduced. So when is the optional sampling theorem valid?

Consider

since is bounded, but what about the other terms? The first one approaches 0 essentially because approaches 0 as .

So the problematic term for our nice result is , which does not necessarily approach 0. In our doubling betting example, and , so we get , which does not approach 0.

Optional Sampling Theorem: Suppose is a martingale with respect to and is a stopping time satisfying , , and , then .

Example:
Let be a simple random walk on with absorbing boundaries, let , and let

Since is bounded, we know immediately and

Therefore,

Now let
(using this), so this is a Martingale

Using the same as before, is no longer bounded, however we can show there exists and such that

Since , we can show and

Hence,

Since , we see

Uniform Integrability

is hard to verify, so we’d like to find some easier conditions

Suppose we have with ,

We say is uniformly integrable if for every there is some with for each , where depends on (not )

If is uniformly integrable then for every there is some where for each

To show this, let , so if then

Example:
Consider our martingale betting strategy, with random variables , and the event



This cannot satisfy the conditions for uniform integrability for any

Now suppose is a uniformly integrable martingale (with respect to ) and is a stopping time with

We have
So we therefore have

Optional Sampling Theorem (again): Suppose is a uniformly integrable martingale with respect to and is a stopping time with and , then

If is a sequence of random variable and there is such that for each , then the sequence is uniformly integrable

To show this with our previous definition, let and suppose


Example:
What if we set the the signs of the harmonic series randomly?

This is a martingale, since each has mean 0


So this is uniformly integrable

Martingale Convergence Theorem

This theorem states when a martingale converges to a limiting random variable . Let’s use Polya’s urn as an example

Let , suppose that , let , and
For , the optional sampling theorem says
But
So

This is true for all , so
In other words, with probability the proportion of red balls never gets as high as
If it does go up to (with probability ), the inverse of this argument says it’ll drop back down to with probability

The probability this continues happening is
This tells us the proportion cannot fluctuate infinitely between any two numbers, i.e. exists
It happens to be we can also prove that this setup results in a uniform distribution directly

Martingale Convergence Theorem: Suppose is a martingale with respect to such that there exists with for all . There there exists a random variable such that .

Fix and think of as representing the cumulative results of a fair game
Whenever , keep betting 1 until , then stop until again

Our winnings are then , where is 0 or 1 and is

Note that is a martingale and where denotes the number of “upcrossings” (passes between and )



This holds for all , so the expected number of upcrossings is bounded, meaning the number of upcrossings is always finite

Note that the martingale property does not imply that , such as in the martingale betting strategy

If is a uniformly integrable martingale then exists and

Example:
Let be the proportion of red balls in Polya’s urn and suppose that at time there are red balls and green balls. Since is bounded, it is uniformly integrable, and approaches with

Example:
Let be independent random variables with

Let and for , let
Note that

Therefore, is a martingale with respect to
Since , the conditions of the martingale convergence theorem hold and for some

Is uniformly integrable?



Therefore

, i.e. is not uniformly integrable

Example:
A normal problem in statistics is estimating the mean of a distribution given

In Bayesian statistics, the parameter is taken to be a random variable taken from a prior distribution

Under the prior distribution, , and let and

is a martingale and the conditional distribution on is called the posterior distribution

We also know
The strong law of large numbers says
So

Assume these samples are from a Bernoulli distribution with

This is the beta distribution with parameters and
The mean happens to be


Maximal Inequalities

If is a sequence of random variables, defined the maximum processes as and

Maximal inequalities relate probabilities or expectations for to respectively

Reflection Principle: Suppose are independent random variables whose distributions are symmetric about the origin, and let and . For every ,

To prove this, let


The independence and symmetry of shows that (noting that can be greater than makes this easier to understand)

Doob’s Maximal Inequality: Suppose is a nonnegative submartingale with respect to . Then for every ,

Let denote the -measurable event (same as before),

So

If is not necessarily nonnegative, we cannot immediately use the inequality, but there’s still an extension

If then is a submartingale,

Likewise, if ,

for all

Doob’s Maximal Inequality (again): Suppose is a martingale with respect to , then for every and ,

Example:
Let denote a simple random walk in and let

is a martingale so we have,


Taylor series shows that

Hence,
, i.e. the random walk in is recurrent