Theory 1
In order to show why the CLT is true, we introduce the technique of moment generating functions. Recall that the
Recall the power series for
The function
Given a random variable
This is a function of
It is reasonable to consider
Example - Moment generating function of a standard normal
We compute
where . From the formula for expected value of a function of a random variable, we have: Complete the square in the exponent:
. Thus: The last factor can be taken outside the integral:
Exercise - Moment generating function of an exponential variable
Compute
for .
Moment generating functions have the remarkable property of encoding the distribution itself:
Distributions determined by MGFs
Assume
and both converge. If , then . Moreover, if
for any interval of values , then for all and .
Be careful about moments vs. generating functions!
Sometimes the moments all exist, but they grow so fast that the moment generating function does not converge. For example, the log-normal distribution
for has this property. The fact above does not apply when this happens.
When moment generating functions approximate each other, their corresponding distributions also approximate each other:
Distributions converge when MGFs converge
Suppose that
for all on some interval . (In particular, assume that converges on some such interval.) Then for any , we have:
Exercise Using an MGF
Suppose
is nonnegative and when and when . Find a bound on using (a) Markov’s Inequality, and (b) Chebyshev’s Inequality.
Theory 2
The main role of moment generating functions in the proof of the CLT is to convert the sum
We have
Exchange the sum in the exponent for a product of exponentials:
Now since the
Now expand the exponential in its Taylor series and use linearity of expectation:
We don’t give a complete argument for the final approximation, but a few remarks are worthwhile. For fixed
In any case, the factors of the last line are independent of
But