Postscript version of these notes
STAT 804
Lecture 16 Notes
Distribution theory for sample autocovariances
The simplest statistic to consider is
where the sum extends over those for which the data are available.
If the series has mean 0 then the expected value of this statistic
is simply
which differs negligibly for large compared to from .
To compute the variance we begin with the second moment which is
The expectations in question involve the fourth order product
moments of and depend on the distribution of the 's and
not just on . However, for the interesting case of white
noise, we can compute the expected value. For you may assume
that or since the cases can be figured out by swapping
and in the case. For the variable is independent
of all 3 of , and . Thus the expectation factors
into something containing the factor
. For ,
we get
. and so the second
moment is
This is also the variance since, for and for white noise,
.
For and or the expectation is simply
while for we get
.
Thus the variance of the sample variance (when the mean is known
to be 0) is
For the normal distribution the fourth moment is given simply
by .
Having computed the variance it is usual to look at the large
sample distribution theory. For the usual central limit theorem
applies to
(in the case of white noise) to prove that
The presence of in the formula shows that the approximation is
quite sensitive to the assumption of normality.
For the theorem needed is called the -dependent central
limit theorem; it shows that
In each of these cases the assertion is simply that the statistic
in question divided by its standard deviation has an approximate
normal distribution.
The sample autocorrelation at lag is
For we can apply Slutsky's theorem to conclude that
This justifies drawing lines at
to carry
out a 95% test of the hypothesis that the series is white
noise based on the th sample autocorrelation.
It is possible to verify that subtraction of from the
observations before computing the sample covariances does not
change the large sample approximations, although it does affect
the exact formulas for moments.
When the series is actually not white noise the situation is
more complicated. Consider as an example the model
with being white noise. Taking
we find that
The expectation is 0 unless either all 4 indices on the
's are the same or the indices come in two pairs of equal
values. The first case requires and and then
. The second case requires one of three pairs of equalities:
and
or
and
or
and
along with the restriction
that the four indices not all be equal. The actual moment is then
when all four indices are equal and when there
are two pairs. It is now possible to do the sum using geometric
series identities and compute the variance of
.
It is not particularly enlightening to finish the calculation in
detail.
There are versions of the central limit theorem called
mixing central limit theorems which can be used for ARMA() processes
in order to conclude that
has asymptotically a standard normal distribution and that the same
is true when the standard deviation in the denominator is replaced by an
estimate. To get from this to distribution theory for the
sample autocorrelation is easiest when the true autocorrelation is 0.
The general tactic is the method or Taylor expansion. In this
case for each sample size you have two estimates, say and
of two parameters. You want distribution theory for the ratio
. The idea is to write
where
and then make use of the fact that and are
close to the parameters they are estimates of. In our case
is the sample autocovariance at lag which is close to the
true autocovariance while the denominator is the
sample autocovariance at lag 0, a consistent estimator of .
Write
If we can use a central limit theorem to conclude
that
has an approximately bivariate normal distribution
and if we can neglect the remainder term then
has approximately a normal distribution. The notation here is that
denotes differentiation with respect to the th argument
of . For
we have
and
.
When the term involving vanishes and we
simply get the assertion that
has the same asymptotic normal distribution as
.
Similar ideas can be used for the estimated sample partial ACF.
Portmanteau tests
In order to test the hypothesis that a series is white noise using the
distribution theory just given, you have to produce a single statistic
to base youre test on. Rather than pick a single value of the
suggestion has been made to consider a sum of squares or a weighted
sum of squares of the
.
A typical statistic is
which, for white noise, has approximately a distribution.
(This fact relies on an extension of the previous computations to conclude
that
has approximately a standard multivariate distribution. This, in turn, relies
on computation of the covariance between
and
.)
When the parameters in an ARMA() have been estimated by maximum likelihood
the degrees of freedom must be adjusted to . The resulting
test is the Box-Pierce test; a refined version which takes better account
of finite sample properties is the Box-Pierce-Ljung test. S-Plus plots the
-values from these tests for 1 through 10 degrees of freedom as
part of the output of arima.diag.
Richard Lockhart
2001-09-30