Central Limit Theorem

Markov’s inequality: If is a random variable that takes only nonnegative values, then for any value , .

This is simple to prove. Let indicate . We see . Therefore, .

As a corollary,
Chebyshev’s inequality: If is a random variable with finite mean and variance , then for any ,

To get this, we apply Markov’s inequality on with , obtaining . This is directly equivalent to Chebyshev’s.

These bounds are important when we know little about a distribution besides mean (and variance).

Central Limit Theorem: Let be a sequence of independent and identically distributed random variables, with mean and variance . The distribution of tends towards the standard normal as .

That is, for , as

The proof leverages the following,
Lemma: Let be a sequence of random variables with distribution functions and moment generating functions , and let be a random variable having distribution function and moment generating function . If for all , then for all at which is continuous. This makes sense with our previous intuition that is fully representative of a random variable.

If we let be a standard normal random variable, . Therefore, we just need to show that our sequence of random variables will tend towards

Assuming the generating function of , exists and is finite, the moment generating function of is

Since these are independent, we can say that the moment generating function of is

Let . Note that and

We must show that , equivalently that .

Using L’Hôpital,

This proves the central limit theorem on standard variables. The same result can be applied to any variable by considering its standardized version

This theorem states that for each individual ,
It can also be shown that this convergence is uniform in , meaning for all , there is a point where for all

Example:
Suppose an astronomer wants to measure the distance to a star. He has a technique but he knows there’s a a variance of light-years in his observations. He wants to make observations and take the average as his estimation. How many measurements does he need to make sure is with light-years?


The central limit theorem tells us is approximately normal

So we need to find , or
The inverse CDF does not have a closed-form, but a solver would tell us , which means observations

Technically, we don’t know when the normal approximation will begin to be a good approximation of this distribution. If we are especially unsure, we can use Chebyshev’s inequality for a tight bound.

, so observations

General Central Limit Theorem: Let be a sequence of independent random variables with . If are uniformly bounded and then as

It really is a remarkable fact.

Strong Law of Large Numbers

Theorem: Let be a sequence of i.i.d. random variables, with . With probability , as

This implies we can approximate the probability of an event by repeating many trials.

Proof:
We assume the fourth moment of is finite, i.e. (however the theorem can be proven without this)

Assume . Let . Consider, . Expanding these terms yields results in the form , , , , and , where . The terms with single have mean by independence. So expanding yields,

so

Therefore, which implies

, since the series converges. This implies that the series is finite with probability (since the expectation would otherwise be infinite). The convergence of the series implies , which implies

When , we can apply this argument to the random variables to obtain that