Reading for Today's Lecture: ?
Goals of Today's Lecture:
Today's notes
Last time we used the Fourier inversion formula to prove the local central limit theorem:
Framework:
We concluded
We differentiated
to obtain
It now follows that
Apply the Fourier inversion formula to deduce
This proof of the central limit theorem is not terribly general
since it requires T to have a bounded continuous density. The
usual central
limit theorem is a statement about cdfs not densities and is
In undergraduate courses we often teach the central limit theorem as follows: if are iid from a population with mean and standard deviation then has approximately a normal distribution. We also say that a Binomial(n,p) random variable has approximately a N(np,np(1-p)) distribution.
To make precise sense of these assertions we need to assign a meaning to statements like ``X and Y have approximately the same distribution''. The meaning we want to give is that X and Y have nearly the same cdf but even here we need some care. If n is a large number is the N(0,1/n) distribution close to the distribution of ? Is it close to the N(1/n,1/n) distribution? Is it close to the distribution? If is the distribution of Xn close to that of ?
The answer to these questions depends in part on how close close needs to be so it's a matter of definition. In practice the usual sort of approximation we want to make is to say that some random variable X, say, has nearly some continuous distribution, like N(0,1). In this case we must want to calculate probabilities like P(X>x) and know that this is nearly P(N(0,1) > x). The real difficulty arises in the case of discrete random variables; in this course we will not actually need to approximate a distribution by a discrete distribution.
When mathematicians say two things are close together: either there is an upper bound on the distance between the two things or they are talking about taking a limit. In this course we do the latter.
Definition: A sequence of random variables Xn converges in
distribution to a random variable X if
Theorem: The following are equivalent:
Now let's go back to the questions I asked:
Here is the message you are supposed to take away from this discussion. You do distributional approximations by showing that a sequence of random variables Xn converges to some X. The limit distribution should be non-trivial, like say N(0,1). We don't say Xn is approximately N(1/n,1/n) but that n1/2 Xn converges to N(0,1) in distribution.
The Central Limit Theorem
If
are iid with mean 0 and variance 1 then
converges in distribution to N(0,1). That is,
Proof: As before
Edgeworth expansions
Suppose that X is a random variable with mean 0, variance 1
and
.
If
is the characteristic
function of X, then
Now apply this calculation to the characteristic
function of
where
is the
mean of a sample of size n. Then
Remarks:
Multivariate convergence in distribution
Definition:
converges in distribution to
if
This is equivalent to either of
Cramér Wold Device: atXn converges in distribution to at X for each
or
Convergence of characteristic functions:
Extensions of the CLT
Slutsky's Theorem: If Xn converges in distribution to Xand Yn converges in distribution (or in probability) to c, a constant, then Xn+Yn converges in distribution to X+c.
Warning: the hypothesis that the limit of Yn be constant is essential.
The delta method: Suppose a sequence Yn of random variables converges to some y a constant and that if we define Xn = an(Yn-y) then Xn converges in distribution to some random variable X. Suppose that f is a differentiable function on the range of Yn. Then an(f(Yn)-f(y)) converges in distribution to . If Xn is in Rp and f maps Rp to Rq then is the matrix of first derivatives of components of f.
Example: Suppose
are a sample from a population with
mean ,
variance ,
and third and fourth central moments
and .
Then
Take
.
Then Yn converges to
.
Take
an = n1/2. Then
Remark: In this sort of problem it is best to learn to recognize that the sample variance is unaffected by subtracting from each X. Thus there is no loss in assuming which simplifies and a.