LN LinearTSModels 3
LN LinearTSModels 3
Series Analysis
Semester 1, 2015-16
LINEAR TIME SERIES MODELS
1 Model Specification
We now look at statistical inference of ARIM A models.
Specifically we look at how
I how to estimate the parameters of a specific
ARIM A(p; d; q) model:
I how to check on the appropriateness of the fitted
model and improve it if needed.
Broadly, the strategy will first be to decide on reason-
able, but tentative, values of p; d and q: Next we estimate
the 0s; 0s and 2e for that model. Finally, we look crit-
ically at the fitted model to check its adequacy. If the
model appears inadequate in some way, we consider the
nature of the inadequacy to help us select another model.
We then proceed to estimate the new model and check it
for adequacy again.
With a few iterations of this model-building strategy, we
hope to arrive at the best possible model for a given se-
ries.
This strategy is popularized by George E.P. Box and G.M.
Jenkins, commonly referred to as the Box-Jenkins method.
3
1.1 Sample Autocorrelation Function
We have seen that ARM A processes have autocorrela-
tions that exhibit certain patterns.
Our aim is to try to recognise, to the extent possible,
similar patterns in the sample autocorrelations, rk . For
an observed series Y1; Y2; : : : ; Yn, the sample autocorre-
lations can be computed by
P
n
(Yt Y )(Yt k Y)
t=k+1
rk = P
n ; k = 1; 2; ::: (1)
(Yt Y )2
t=1
However, as the rk are only estimates of the k , we need
to investigate their sampling properties to facilitate the
comparison of the estimated correlations with their the-
oretical counterparts.
Sampling properties of rk not easy to derive. Have to
content with a general large-sample result.
Suppose
P
1
Yt = + j et j (2)
j=0
4
2
where the et are iid(0; e ): Further, assume that
P
1 P
1
2
j jj < 1 and j j <1 (3)
j=0 j=0
These will be satisfied by any stationary ARM A model.
Then, for any fixed m, the joint distribution of
p p p
n(r1 1 ); n(r2 2 ); : : : ; n(rm m)
P
1
2
cij = ( k+i k+j + k i k+j 2 i k k+j 2 j k k+j +2 i j k)
k= 1
(4)
For large n, we would say that rk is approximately nor-
mally distributed with E(ri) i , V ar(ri ) cii=n and
p
Corr(ri; rj ) cij = ciicjj
Note that the approximate variance is inversely propor-
tional to the sample size, but Corr(ri; rj ) is approxi-
mately constant.
Consider equation (4) for some special cases. Suppose
fYtg is white noise. Then equation (4) reduces consider-
5
ably and we obtain
1
V ar(ri) and Corr(ri; rj ) 0; i 6= j (5)
n
Next, suppose fYtg is generated by an AR(1) process
with s = s for s > 0: Then, it can be shown that
equation (4) yields
1 (1 + 2)(1 2i
) 2i
V ar(ri) 2 2i (6)
n 1
In particular,
2
1
V ar(r1) (7)
n
Notice that the closer is to 1; the more precise our
estimate of 1(= ) becomes.
For large lags, the term involving 2i in equation (6) may
be ignored, and we have
1 (1 + 2)
V ar(ri) 2 for large i (8)
n 1
Notice here, in contrast to equation (7), values of close
to 1 imply large variances for ri: Thus we should not
6
i
expect nearly as precise estimate of i = for large i
as we for small i:
For AR(1), equation (4) can also be simplified for 0 <
i < j as
j i j+i 2
( )(1 + ) j i j+i
cij = 2 +(j i) (j+i) (9)
1
In particular, we find
s
2
1
Corr(r1; r2) 2 2 4 (10)
1+2 3
For the M A(1) process, equation (4) simplifies as fol-
lows:
2 4 2
c11 = 1 3 1 +4 1 and cii = 1 + 2 1 for i > 1 (11)
2
c12 = 2 1(1 1) (12)
For a general M A(q) process, equation (4) reduces to
P
q
2
cii = 1 + 2 j; for i > q (13)
j=1
7
so that
" #
1 Pq
2
V ar(ri) = 1+2 j ; for i > q (14)
n j=1
j = k1 j 1 + k2 j 2 + + for j = 1; 2; :::; k
kk j k ;
(15)
We can write these k equations more explicitly as
k1 + 1 +
k2 2 k3+ + k 1 = 1
kk
1 k1 + k2 + 1 k3 + +
k 2 kk = 2
.. .. ..
k 1 k1 + k 2 k2 + k 3
k3 + + kk = k
(16)
Express equation (16) in matrix algebra and use Cramer’s
Rule to solve for kk :
9
1 1 2 k 2 1
1 1 1 k 3 2
.. .. .. .. .. ..
k 1 k 2 k 3 1 k
kk = (17)
1 1 2 k 2 k 1
1 1 1 k 3 k 2
.. .. .. .. .. ..
k 1 k 2 k 3 1 1
kP1
k k 1;j k j
j=1
kk = kP1
(18)
1 k 1;j j
j=1
where
2
2 11 1 2 1
22 = = 2 (20)
1 11 1 1 1
3 21 2 22 1
33 = (21)
1 21 2 22 2
12
that looks like this:
MA
AR 0 1 2 3 4 5 6 7
0 X X X X X X X X
1 X O O O O O O O
(23)
2 X X O O O O O O
3 X X X O O O O O
4 X X X X O O O O
5 X X X X X O O O
-130
-130
-140
-140
-140
-140
-140
-140
(Intercept)
test-lag1
test-lag2 Example
test-lag3
test-lag4
test-lag5
test-lag6
test-lag7
test-lag8
test-lag9
test-lag10
test-lag11
test-lag12
test-lag13
16
test-lag14
error-lag1
error-lag2
error-lag3
error-lag4
error-lag5
error-lag6
error-lag7
error-lag8
finding an optimal subset ARM A model.
error-lag9
error-lag10
error-lag11
I Search for best subsets of ARM A(14; 14)
error-lag12
error-lag13
error-lag14
I Simulate a series from Yt = 0:8Yt 12 + et + 0:7et
1
The method of Hannan and Rissanen can be extended to