Central Limit Theorem
Markov’s inequality: If is a random variable that takes only nonnegative values, then for any value , .
This is simple to prove. Let indicate . We see . Therefore, .
As a corollary,
Chebyshev’s inequality: If is a random variable with finite mean and variance , then for any ,
To get this, we apply Markov’s inequality on with , obtaining . This is directly equivalent to Chebyshev’s.
These bounds are important when we know little about a distribution besides mean (and variance).
Central Limit Theorem: Let be a sequence of independent and identically distributed random variables, with mean and variance . The distribution of tends towards the standard normal as .
That is, for , as
The proof leverages the following,
Lemma: Let be a sequence of random variables with distribution functions and moment generating functions , and let be a random variable having distribution function and moment generating function . If for all , then for all at which is continuous. This makes sense with our previous intuition that is fully representative of a random variable.
If we let be a standard normal random variable, . Therefore, we just need to show that our sequence of random variables will tend towards
Assuming the generating function of , exists and is finite, the moment generating function of is
Since these are independent, we can say that the moment generating function of is
Let . Note that and
We must show that , equivalently that .
Using L’Hôpital,
This proves the central limit theorem on standard variables. The same result can be applied to any variable by considering its standardized version
This theorem states that for each individual ,
It can also be shown that this convergence is uniform in , meaning for all , there is a point where for all
Example:
Suppose an astronomer wants to measure the distance to a star. He has a technique but he knows there’s a a variance of light-years in his observations. He wants to make observations and take the average as his estimation. How many measurements does he need to make sure is with light-years?
The central limit theorem tells us is approximately normal
So we need to find , or
The inverse CDF does not have a closed-form, but a solver would tell us , which means observations
Technically, we don’t know when the normal approximation will begin to be a good approximation of this distribution. If we are especially unsure, we can use Chebyshev’s inequality for a tight bound.
, so observations
General Central Limit Theorem: Let be a sequence of independent random variables with . If are uniformly bounded and then as
It really is a remarkable fact.
Strong Law of Large Numbers
Theorem: Let be a sequence of i.i.d. random variables, with . With probability , as
This implies we can approximate the probability of an event by repeating many trials.
Proof:
We assume the fourth moment of is finite, i.e. (however the theorem can be proven without this)
Assume . Let . Consider, . Expanding these terms yields results in the form , , , , and , where . The terms with single have mean by independence. So expanding yields,
so
Therefore, which implies
, since the series converges. This implies that the series is finite with probability (since the expectation would otherwise be infinite). The convergence of the series implies , which implies
When , we can apply this argument to the random variables to obtain that