The Lindeberg central limit theorem
1 Convergence in distribution
We denote by the collection of Borel probability measures on . Unless we say otherwise, we use the narrow topology on : the coarsest topology such that for each , the map
is continuous . Because is a Polish space it follows that is a Polish space.11 1 Charalambos D. Aliprantis and Kim C. Border, Infinite Dimensional Analysis: A Hitchhiker’s Guide, third ed., p. 515, Theorem 15.15; http://individual.utoronto.ca/jordanbell/notes/narrow.pdf (In fact, its topology is induced by the Prokhorov metric.22 2 Onno van Gaans, Probability measures on metric spaces, http://www.math.leidenuniv.nl/~vangaans/jancol1.pdf; Bert Fristedt and Lawrence Gray, A Modern Approach to Probability Theory, p. 365, Theorem 25.)
2 Characteristic functions
For , we define its characteristic function by
Theorem 1.
If has finite th moment, , then, writing :
-
1.
.
-
2.
.
-
3.
is uniformly continuous.
-
4.
.
Proof.
For , define by
For ,
so by the dominated convergence theorem we have for ,
That is,
And, by the dominated convergence, for there is some such that if then
hence if then
showing that is uniformly continuous. As well,
But , i.e. , so
∎
If , Taylor’s theorem tells us that for each ,
and satisfies
Define by and for
with which, for all ,
Because is continuous on , is continuous at each . Moreover,
and as is continuous it follows that is continuous at . Thus is continuous on .
Lemma 2.
If have finite th moment, , and for ,
then there is a continuous function for which
The function satisfies
Proof.
For and , let
Let be the measure on whose density with respect to Lebesgue measure is . We call a Gaussian measure. We calculate that the first moment of is and that its second moment is . We also calculate that
Lévy’s continuity theorem is the following.33 3 http://individual.utoronto.ca/jordanbell/notes/martingaleCLT.pdf, p. 19, Theorem 15.
Theorem 3 (Lévy’s continuity theorem).
Let be a sequence in .
-
1.
If and , then for each converges to pointwise.
-
2.
If there is some function to which converges pointwise and is continuous at , then there is some such that and such that .
3 The Lindeberg condition, the Lyapunov condition, the Feller condition, and asymptotic negligibility
Let be a probability and let , , be independent random variables. We specify when we impose other hypotheses on them; in particular, we specfify if we suppose them to be identically distributed or to belong to for .
For a random variable , write
Write
and, using that the are independent,
and
For and , define
We say that the sequence satisfies the Lindeberg condition if for each ,
For example, if the sequence is identically distributed, then , so
But if is a Borel probability measure on and and is a sequence of compact sets that exhaust , then44 4 V. I. Bogachev, Measure Theory, volume I, p. 125, Proposition 2.6.2.
Hence as , showing that satisfies the Lindeberg condition.
We say that the sequence satisfies the Lyapunov condition if there is some such that the are and
In this case, for , then implies and hence
This is true for each , showing that if satisfies the Lyapunov condition then it satisfies the Lindeberg condition.
For example, if are identically distributed and , then
showing that satisfies the Lyapunov condition.
Another example: Suppose that the sequence is bounded by almost surely and that . almost surely implies that
Therefore almost surely. Let . Then, as ,
showing that satisfies the Lyapunov condition.
We say that a sequence of random variables satisfies the Feller condition when
where and
We prove that if a sequence satisfies the Lindeberg condition then it satisfies the Feller condition.55 5 Heinz Bauer, Probability Theory, p. 235, Lemma 28.2.
Lemma 4.
If a sequence of random variables satisfies the Lindeberg condition, then it satisfies the Feller condition.
Proof.
Let ¿ For and , we calculate
Hence
and so, because the satisfy the Lindeberg condition,
This is true for all , which yields
namely, that the satisfy the Feller condition. ∎
We do not use the following idea of an asymptotically negligible family of random variables elsewhere, and merely take this as an excsuse to write out what it means. A family of random variables , , , is called asymptotically negligible66 6 Heinz Bauer, Probability Theory, p. 225, §27.2. if for each ,
A sequence of random variables converging in probability to is equivalent to it being asymptotically negligible, with for each .
For example, suppose that are random variables each with and that they satisfy
For , by Chebyshev’s inequality,
whence
and so the random variables are asymptotically negligible.
Another example: Suppose that random variables are identically distributed, with . For ,
where . As , . Hence the random variables are asymptotic negligible.
The following is a statement about the characteristic functions of an asymptotically negligible family of random variables.77 7 Heinz Bauer, Probability Theory, p. 227, Lemma 27.3.
Lemma 5.
Suppose that a family , , , of random variables is asymptotically negligible, and write and . For each ,
Proof.
For any real , . For , , , and ,
Hence
Using that the family is asymptotically negligible,
But this is true for all , so
proving the claim. ∎
4 The Lindeberg central limit theorem
We now prove the Lindeberg central limit theorem.88 8 Heinz Bauer, Probability Theory, p. 235, Theorem 28.3.
Theorem 6 (Lindeberg central limit theorem).
If is a sequence of independent random variables that satisfy the Lindeberg condition, then
where
Proof.
The sequence are independent random variables that satisfy the Lindeberg condition and . Proving the claim for the sequence will prove the claim for the sequence , and thus it suffices to prove the claim when , i.e. .
For and , let
The first moment of is
and the second moment of is
for which
For with first moment and second moment , Lemma 2 tells us that
with
But by Lemma 1,
so
For , , so for and , with , when and we have . Thus
Let and , and take . On the one hand, for and , because the first moment of is and its second moment is ,
with, from the above,
On the other hand, the first moment of the Gaussian measure is and its second moment is . Its characteristic function is
with, from the above,
In particular, for all ,
For and for , ,
If further , , then
(1) |
Because the are independent, the distribution of
is the convolution of the distributions of the summands:
whose characteristic function is
since the characteristic function of a convolution of measures is the product of the characteristic functions of the measures. Using and (1), for we have
Therefore, for , , and ,
We calculate
Hence, the fact that the satisfy the Lindeberg condition yields
(2) |
Write
We calculate
Because the sequence satisfies the Lindeberg condition, by Lemma 4 it satisfies the Feller condition, which means that as . Because as , as , hence
as . Thus we get
as . Using this with (2) yields
This is true for all , so
namely, (the characteristic function of ) converges pointwise to . Moreover, is indeed continuous at , and . Therefore, Lévy’s continuity theorem (Theorem 3) tells us that converges narrowly to , which is the claim. ∎