29
1 Non-Seasonal Box-Jenkins Models

Box Jenkins Methodology

Embed Size (px)

Citation preview

Page 1: Box Jenkins Methodology

1

Non-Seasonal Box-Jenkins Models

Page 2: Box Jenkins Methodology

2

Box-Jenkins (ARIMA) Models

The Box-Jenkins methodology refers to a set of procedures for identifying and estimating time series models within the class of autoregressive integrated moving average (ARIMA) models.

ARIMA models are regression models that use lagged values of the dependent variable and/or random disturbance term as explanatory variables.

ARIMA models rely heavily on the autocorrelation pattern in the data

This method applies to both non-seasonal and seasonal data. In this course, we will only deal with non-seasonal data.

Page 3: Box Jenkins Methodology

3

Box-Jenkins (ARIMA) Models

Three basic ARIMA models for a stationary time series yt : (1) Autoregressive model of order p (AR(p)) i.e., yt depends on its p previous values (2) Moving Average model of order q (MA(q)) i.e., yt depends on q previous random error terms

,2211 tptpttt yyyy εφφφδ +++++= −−−

,2211 qtqtttty −−− −−−−+= εθεθεθεδ

Page 4: Box Jenkins Methodology

4

Box-Jenkins (ARIMA) Models

(3) Autoregressive-moving average model of order p and q (ARMA(p,q))

i.e., yt depends on its p previous values and q

previous random error terms

,2211

2211

qtqttt

ptpttt yyyy

−−−

−−−

−−−−+

++++=

εθεθεθε

φφφδ

Page 5: Box Jenkins Methodology

5

Box-Jenkins (ARIMA) Models

In an ARIMA model, the random disturbance term is typically assumed to be “white noise”; i.e., it

is identically and independently distributed with a mean of 0 and a common variance across all observations.

We write ~ i.i.d.(0, )

tε 2σ

Page 6: Box Jenkins Methodology

6

A five-step iterative procedure

1) Stationarity Checking and Differencing

2) Model Identification

3) Parameter Estimation

4) Diagnostic Checking

5) Forecasting

Page 7: Box Jenkins Methodology

7

Step One: Stationarity checking

Page 8: Box Jenkins Methodology

8

Stationarity

“Stationarity” is a fundamental property underlying almost all time series statistical models.

A time series yt is said to be stationary if it satisfies the following conditions:

2 2

(1) ( ) .

(2) ( ) [( ) ] .

(3) ( , ) .

t y

t t y y

t t k k

E y u for all t

Var y E y u for all tCov y y for all t

σ

γ−

=

= − =

=

Page 9: Box Jenkins Methodology

9

Stationarity

The white noise series satisfies the stationarity condition because

(1) E( ) = 0 (2) Var( ) = (3) Cov( ) = for all s 0

tε t sε −

Page 10: Box Jenkins Methodology

10

Example of a white noise series

3632282420161284

100

80

60

40

20

0

Time

Time Series Plot

Page 11: Box Jenkins Methodology

11

Example of a non-stationary series

2902612322031741451168758291

4000

3900

3800

3700

3600

3500

Time

Time Series Plot of Dow-Jones Index

Page 12: Box Jenkins Methodology

12

Stationarity

Suppose yt follows an AR(1) process without drift. Is yt stationarity? Note that

ot

tttt

ttt

ttt

yy

yy

133

122

111

1211

11

...........

)(

φεφεφεφε

εεφφεφ

+++++=

++=+=

−−−

−−

Page 13: Box Jenkins Methodology

13

Stationarity

Without loss of generality, assume that yo = 0. Then E(yt)=0. Assuming that t is large, i.e., the process started a long time ago, then It can also be shown that provided that the same condition is satisfied,

.1|| that provided ,)1(

)var( 121

2

<−

= φφ

σty

)var()1(

)cov( 121

21

ts

s

stt yyy φφσφ

=−

=−

Page 14: Box Jenkins Methodology

14

Stationarity

Special Case: φ1 = 1 It is a “random walk” process. Now,

Thus,

.1 ttt yy ε+= −

∑−

=−=

1

0.

t

jjtty ε

2

2

(1) ( ) 0 .

(2) ( ) .

(3) ( , ) ( ) .

t

t

t t s

E y for all tVar y t for all tCov y y t s for all t

ε

ε

σ

σ−

=

=

= −

Page 15: Box Jenkins Methodology

15

Stationarity

Suppose the model is an AR(2) without drift, i.e.,

It can be shown that for yt to be stationary,

The key point is that AR processes are not stationary unless appropriate prior conditions are imposed on the parameters.

tttt yyy εφφ ++= −− 2211

1 || and 1 ,1 21221 <<−<+ φφφφφ

Page 16: Box Jenkins Methodology

16

Stationarity

Consider an MA(1) process without drift:

It can be shown, regardless of the value of , that 11 −−= ttty εθε

=−

=

+=

=

− otherwise 01s if

)cov(

)1()var(

0)(

21

21

2

σθ

θσ

stt

t

t

yy

yyE

Page 17: Box Jenkins Methodology

17

Stationarity

For an MA(2) process

2211 −− −−= tttty εθεθε

=−=−−

=

++=

=

otherwise 0 2 s if 1s if )1(

)cov(

)1()var(

0)(

22

22

1

22

21

2

σθθσθ

θθσ

stt

t

t

yy

yyE

Page 18: Box Jenkins Methodology

18

Stationarity

In general, MA processes are stationarity regardless of the values of the parameters, but not necessarily “invertible”.

An MA process is said to be invertible if it can be converted into a stationary AR process of infinite order.

The conditions of invertibility for an MA(k) process is analogous to those of stationarity for an AR(k) process. More on invertibility in tutorial.

Page 19: Box Jenkins Methodology

19

Differencing

Often non-stationary series can be made stationary through differencing.

Examples:

stationary is7.0 but ,stationarynot is 7.07.1 )2

stationary is but ,stationarynot is )1

11

21

1

1

ttttt

tttt

tttt

ttt

wyywyyy

yywyy

εε

εε

+=−=+−=

=−=+=

−−

−−

Page 20: Box Jenkins Methodology

20

Differencing

Differencing continues until stationarity is achieved.

The differenced series has n-1 values after taking the first-difference, n-2 values after taking the second difference, and so on.

The number of times that the original series must be differenced in order to achieve stationarity is called the order of integration, denoted by d.

In practice, it is almost never necessary to go beyond second difference, because real data generally involve only first or second level non-stationarity.

1t t ty y y −∆ = −2

1 1 2( ) ( ) 2t t t t t t ty y y y y y y− − −∆ = ∆ ∆ = ∆ − = − +

Page 21: Box Jenkins Methodology

21

Differencing

Backward shift operator, B

B, operating on yt, has the effect of shifting the data back one period.

Two applications of B on yt shifts the data back two periods.

m applications of B on yt shifts the data back m periods.

1t tBy y −=

22( )t t tB By B y y −= =

mt t mB y y −=

Page 22: Box Jenkins Methodology

22

Differencing

The backward shift operator is convenient for describing the process of differencing.

In general, a dth-order difference can be written as The backward shift notation is convenient because

the terms can be multiplied together to see the combined effect.

1 (1 )t t t t t ty y y y By B y−∆ = − = − = −

2 2 21 22 (1 2 ) (1 )t t t t t ty y y y B B y B y− −∆ = − + = − + = −

(1 )d dt ty B y∆ = −

Page 23: Box Jenkins Methodology

23

Population autocorrelation function

The question is, in practice, how can one tell if the data are stationary?

If the data are non-stationary (i.e., random walk), then = 1

for all values of k If the data are stationary (i.e., ), then the magnitude of

the autocorrelation coefficient “dies down” as k increases.

k. lagat t coefficienation autocorrel theis / thatso )cov(Let

okk

kttk yyγγρ

γ== −

k1k

process, AR(1)an Consider φρ =

1|| 1 <φ

Page 24: Box Jenkins Methodology

24

Population autocorrelation function

Consider an AR(2) process without drift : The autocorrelation coefficients are

.2211 tttt yyy εφφ ++= −−

.21

,1

2211

2

21

22

2

11

≥+=−

+=

−=

−− kforkkk ρφρφρφ

φφρ

φφρ

Page 25: Box Jenkins Methodology

25

Population autocorrelation function

Then the autocorrelation function dies down according to a mixture of damped exponentials and/or damped sine waves. In general, the autocorrelation of a stationary

AR process dies down gradually as k increases.

Page 26: Box Jenkins Methodology

26

Population autocorrelation function

Moving Average (MA) Processes Consider a MA(1) process without drift : Recall that

.11 −−= ttty εθε

>=−

==

+==

=

− .101

),()3(

.)1())2(

.0)()1(

21

21

20

ss

yyCov

tallforVar(ytallforyE

sstt

t

t

ε

ε

σθγ

θσγ

Page 27: Box Jenkins Methodology

27

Population autocorrelation function

Therefore the autocorrelation coefficient of the MA(1) process at lag k is

The autocorrelation function of the MA(1) process

“cuts off” after lag k=1.

>

=+−

=

=

.10

11 2

1

1

0

k

k

kk

θθ

γγρ

Page 28: Box Jenkins Methodology

28

Population autocorrelation function

Similarly, for an MA(2) process : The autocorrelation function of a MA(2)

process cuts off after 2 lags.

.20

,1

,1

)1(

22

21

22

22

21

211

>=++

−=

++−−

=

kforkρθθ

θρ

θθθθρ

Page 29: Box Jenkins Methodology

29

Population autocorrelation function

In general, all stationary AR processes exhibit autocorrelation patterns that “die down” to zero as k increases, while the autocorrelation coefficient of a non-stationary AR process is always 1 for all values of k. MA processes are always stationary with autocorrelation functions that cut off after certain lags.

Question: how are the autocorrelation coefficients “estimated” in practice?