34
Time Series Analysis: 8. Long-memory processes P. F. Góra http://th-www.if.uj.edu.pl/zfs/gora/ 2018

Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Time Series Analysis:

8. Long-memory processes

P. F. Górahttp://th-www.if.uj.edu.pl/zfs/gora/

2018

Page 2: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Joseph effect

According to the Book of Genesis, patriarch Joseph interpreted dreams ofa pharaoh:

Pharaoh dreamed that he was standing by the Nile,and behold, there came up out of the Nile sevencows, attractive and plump, and they fed in the reedgrass. And behold, seven other cows, ugly and thin,came up out of the Nile after them, and stood by theother cows on the bank of the Nile. And the ugly, thincows ate up the seven attractive, plump cows.

Gen. 41

Copyright c© 2009-18 P. F. Góra 8–2

Page 3: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Joseph interpreted this dream that after seven years of abundant harveststhere would be seven years of poor harvests and provisions needed to bemade to feed the people during the years of poor harvests. As harvests inancient Egypt depended entirely on Nile flooding, this can be interpreted asseven years of high flooding to be followed by seven years of low flooding.We can see that good/bad years tend to cluster . Benoit Mandelbrot hascoined the term “Joseph effect” for this phenomenon. The ancient peopleunderstood that this clustering was a very important natural phenomenonand included its description, in a literary form, in their holy book.

Copyright c© 2009-18 P. F. Góra 8–3

Page 4: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Consider a random time series that can assume only two values,{+1,−1}.

The lowest curve: similar states tend to cluster. The middle curve: an uncorrelated series.The upper curve: an anticorrelated series that tends to flip its state at each step (a bluenoise).

Copyright c© 2009-18 P. F. Góra 8–4

Page 5: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

The Aswan High Dam

photo: NASA

Copyright c© 2009-18 P. F. Góra 8–5

Page 6: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

The Aswan Dam history

around 1000s The Arabs first attempt to regulate the flooding of the Nile1889-1902 The British construct the first dam in Aswan1907-1912 The dam is raised1929-1933 The dam is raised for the second time

1946 The dam nearly overflowsDecision is taken to build a new dam

1954 Design of a new dam begins1960 Construction of The Aswan High Dam begins1964 The reservoir starts to fill1970 Construction ends1976 The reservoir reaches its capacity

Copyright c© 2009-18 P. F. Góra 8–6

Page 7: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

The Nile minima

In 1951, a British engineer, Harold Hurst, studied historic records on Nileminima:

900

1000

1100

1200

1300

1400

1500

622 682 742 802 862 922 982 1042 1102 1162 1222 1282

Nile

leve

l

Year

The Nile minima

Copyright c© 2009-18 P. F. Góra 8–7

Page 8: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Moving averages do not flatten the series!

900

1000

1100

1200

1300

1400

1500

622 682 742 802 862 922 982 1042 1102 1162 1222 1282

Poz

iom

Rok

Wylewy Nilu

5 year average25 year average

Wet years cluster, dry years cluster — Joseph effect

The series seems to be self-similar — it behaves similarly on (nearly) allscales

Copyright c© 2009-18 P. F. Góra 8–8

Page 9: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Another example — tree rings

Copyright c© 2009-18 P. F. Góra 8–9

Page 10: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Correlations in the Nile minima

0.01

0.1

1

1 10 100

j

0.0

0.2

0.4

0.6

0.8

1.0

0 10 20 30 40 50 60 70 80

ρ(j)

j

Long-range correlations, decreasing according to a power law(the exponent ∼ −0.35)

Copyright c© 2009-18 P. F. Góra 8–10

Page 11: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

“Ordinary” correlations

The autocorrelation function of a stationary ARMA(p,q) process decreasesrapidly as it is bounded from above

|ρk| 6 C Rk , C > 0 , 0 < R < 1 . (1)

In contrast to that, long-memory processes have autocorrelations that arenot necessarily very large but persist for very long times.

Copyright c© 2009-18 P. F. Góra 8–11

Page 12: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Long Range Dependencies

A stochastic process {ξn} is said to have Long Range Dependencies (or along range memory) if its autocorrelation function ρk =

⟨ξnξn+k

⟩has the

asymptotic form

ρk ∼ k2d−1L(k) as k →∞ (2)

where −12 < d < 1

2 and L(k) is slowly varying at infinity:

∀z > 0: limt→∞

L(zt)

L(t)= 1 (3)

Example: Constant functions and logarithms are slowly varying at infinity.

Copyright c© 2009-18 P. F. Góra 8–12

Page 13: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Examples of Long Range Dependencies

• Internet traffic — can impact heavily on queuing

• Telecommunications

• Hydrological phenomena

• Financial data (at least before the crisis ,)

• Sunspots

• Climate; sequences of dry/wet years

• Movements of foraging animals

• fMRI signals, etc

Remember: GWN and “mild” forms of Gaussian coloured noises implicitlyassumed some forms of thermal equilibrium.

Why should we expect GWN in non-equilibrium systems?No reason. Not the case.

Copyright c© 2009-18 P. F. Góra 8–13

Page 14: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Fractional Gaussian noise (fGn)

{ξn} — a Gaussian process, usually zero-mean, whose autocorrelationfunction satisfies

ρk =⟨ξnξn+k

⟩=σ2

2

(|k − 1|2H − 2|k|2H + |k + 1|2H

)(4a)

and H ∈ [0,1]. For large k and H 6= 12:

ρk ' σ2H(2H − 1)k−(2−2H) (4b)

H is also called the Hurst exponent. H = d+ 12, cf. (2).

Copyright c© 2009-18 P. F. Góra 8–14

Page 15: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Examples of fGn’s

Copyright c© 2009-18 P. F. Góra 8–15

Page 16: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Properties of fGn

• fGn is stationary

• fGn is self-similar

• For H ∈[

12,1

], fGn displays Long Range Dependence

• For H = 12, fGn reduces to GWN

• For H ∈[0, 1

2

], fGn has negative correlations

Copyright c© 2009-18 P. F. Góra 8–16

Page 17: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Power spectrum of fGn

Since fGn is stationary, Wiener-Khinchin theorem holds and we can calcu-late the power spectrum by Fourier transforming (4a). We get

P (f) = Cσ2∣∣∣e2πif − 1

∣∣∣2 ∞∑k=−∞

1

|f + k|2H+1(5a)

for |f | 6 12. If H 6= 1

2,

P (f) ∼ Cσ2|f |1−2H for f → 0 (5b)

fGn displays power-law behaviour for low frequencies (or long waves).

Copyright c© 2009-18 P. F. Góra 8–17

Page 18: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Estimating the Hurst exponent

We can see from (5b) that if a straight line with a slope α (α < 0) (in alog-log scale) is fitted to the power spectrum, then

H =1− α

2(6)

Copyright c© 2009-18 P. F. Góra 8–18

Page 19: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Fractional Brownian Motion (fBm)

Let {ξk} be a fGn. Then

Xn =n∑

k=1

ξk (7)

is called a Fractional Brownian Motion. Observe that

Xn = Xn−1 + ξn (8)

The autocorrelation function

〈XnXm〉 =σ2

2

(|n|2H + |m|2H − |n−m|2H

)(9)

The variance then satisfies ⟨X2n

⟩= σ2|n|2H (10)

Copyright c© 2009-18 P. F. Góra 8–19

Page 20: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Properties of fBm

• fBm is non-stationary

• For H ∈[

12,1

], fBm is a superdiffusion

• For H = 12, fBm is a normal diffusion

• For H ∈[0, 1

2

], fBm is a subdiffusion

• Super- and subdiffusion go under a joint term, anomalous diffusion

� But not all anomalous diffusions are fBm’s!

Copyright c© 2009-18 P. F. Góra 8–20

Page 21: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Detrended Fluctuation Analysis (DFA)

Given a (possibly) fractional series {xn}, do the following:

1. Calculate the average 〈xn〉

2. Convert the series into a “random walk”, i.e., calculate the cumulativesums:

Xn =n∑

k=1

(xk − 〈xn〉) (11)

3. Divide the series into segments of length L and fit a straight line withineach segment. Let the line be Xn = a · n+ b.Copyright c© 2009-18 P. F. Góra 8–21

Page 22: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

-8000

-7000

-6000

-5000

-4000

-3000

-2000

-1000

0

1000

2000

772 804 836 868 900 932 964 996 1028 1060 1092 1124 1156 1188 1220 1252 1284

Cum

ulat

ive

Nile

min

ima

Year

Cumulative Nile minima series divided into segments of length 32.A straight line, representing a local trend, is fitted within each segment.

Copyright c© 2009-18 P. F. Góra 8–22

Page 23: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

4. Within each segment, calculate

F (L) =

√√√√√1

L

i0+L−1∑i=i0

(Xi − a · i− b)2 (12)

5. Calculate the average F̄ (L) over all segments of the same length.

6. If F̄ (L) ∼ Lα, the series is fractional. For fGn, α is the Hurst exponent,H.

Copyright c© 2009-18 P. F. Góra 8–23

Page 24: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

10

100

1000

10000

1 10 100 1000

Results of DFA applied to the Nile minima. The fit gives H ' 0.75.

Copyright c© 2009-18 P. F. Góra 8–24

Page 25: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Another approach - ARFIMA

Remember the time shift operator (cf. Lecture 5, Eq. 2): Bzn = zn−1.With this operator, an ARMA(p,q) process can be written as

φ(B)xn = σ θ(B)ηn , (13)

where {ηn} is GWN, φ, θ are polynomials or orders [, q, respectively,whose roots lie outside the unit circle. Similarly, an ARIMA(p,d,q) can bewritten as

φ(B)(1−B)dxn = σ θ(B)ηn , (14)

where, additionally, d ∈ N.

A presence of long-range correlations in a time series suggests a need fordifferentiating to achieve stationarity, but taking a first difference may betoo extreme. This motivates the notion of fractional differencing.Copyright c© 2009-18 P. F. Góra 8–25

Page 26: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Fractionally integrated processes

Therefore, we consider processes of the form (14), but with −12 < d < 1

2:

φ(B)(1−B)dxn = σ θ(B)ηn , −1

2< d <

1

2(15)

Such processes are called ARFIMA(p,d,q).

For d > −1, the operator (1−B)d can be defined by the binomial expan-sion:

(1−B)d =∞∑j=0

πjBj (16a)

where

π0 = 1 , πj =Γ(j − d)

Γ(j + 1)Γ(−d)=

∏0<k6j

k − 1− dk

(16b)

Copyright c© 2009-18 P. F. Góra 8–26

Page 27: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Fractionally integrated white noise

An ARFIMA(0,d,0) is a fractionally integrated white noise

(1−B)dwn = σ ηn (17a)

or

wn = σ (1−B)−dηn = σ∞∑j=0

ψjηn−j (17b)

where

ψ0 = 1 , ψj =∏

0<k6j

k − 1 + d

k∼

1

Γ(d)jd−1 as j →∞ . (17c)

Copyright c© 2009-18 P. F. Góra 8–27

Page 28: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

0.001

0.01

0.1

1

0 32 64 96 128

ψ j

j

d=0.1d=0.4

d=0.45d=0.499

Coefficients ψj, cf. Eq. (17c), for various values of d. There is no “natural”cutoff, one needs to be taken arbitrarily.

Copyright c© 2009-18 P. F. Góra 8–28

Page 29: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

A cutoff

Formally, the summations in (16) and (17) extend to infinity. In practice, weneed to introduce a cutoff to make the sums finite so that can be handled ina finite time. There are no definite results on how large this cutoff must be,but some research suggest that a cutoff of the order ∼100 in most casesis large enough. Therefore, instead of (16) we take

(1−B)d =100∑j=0

πjBj (18)

and similarly for (17).

Copyright c© 2009-18 P. F. Góra 8–29

Page 30: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

Variance and autocorrelation of the fractionally integrated noise

The variance of this process is

⟨w2n

⟩= σ2 Γ(1− 2d)

[Γ(1− d)]2. (19)

and the autocorrelation function

ρl(w) =Γ(l + d)Γ(1− d)

Γ(l − d+ 1)Γ(d)=

∏0<k6l

k − 1 + d

k − d. (20)

Copyright c© 2009-18 P. F. Góra 8–30

Page 31: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

A popular interpretation

ARFIMA(p,d,q) can be interpreted as as ARMA(p,q) driven by a fractionallyintegrated white noise:

φ(B)(1−B)dxn = σ θ(B)ηn (21a)

therefore

φ(B)xn = σ θ(B)wn (21b)

where

(1−B)dwn = ηn (21c)

Copyright c© 2009-18 P. F. Góra 8–31

Page 32: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

The power spectrum

From the above representation we can see that the power spectrum of ageneral ARFIMA(p,d,q) process is

P (f) = 2σ2∣∣∣1− e−2πif

∣∣∣−2d

∣∣∣θ(e−2πif)∣∣∣2∣∣∣φ(e−2πif)∣∣∣2 (22)

where 0 6 f 6 12. Note that P (f) ∼ f−2d for f → 0 which is a characte-

ristic feature of the spectrum of long memory processes.

Copyright c© 2009-18 P. F. Góra 8–32

Page 33: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

ARFIMA(1,d,0)

In theory, ARFIMA(p,d,q) processes of arbitrary orders can be used. Inpractice, however, processes with small p,q are deemed to be most useful.

Consider ARFIMA(1,d,0):

(1− βB)(1−B)dxn = σ ηn (23a)xn = (1− βB)−1wn (23b)

where wn is (17) and −1 < β < 1. This process has the autocorrelation function

ρl =ρl(w)

1− βF (d+ l,1; 1− d+ l;β) + F (d− l,1; 1− d− l;β)− 1

F (1 + d,1; 1− d;β)(24)

where F (a, b; c;x) is the hypergeometric function.

Copyright c© 2009-18 P. F. Góra 8–33

Page 34: Time Series Analysis: 8. Long-memory processesth- · Long Range Dependencies A stochastic process f˘ngis said to have Long Range Dependencies (or a long range memory) if its autocorrelation

ARFIMA(0,d,1)

(1−B)dxn = σ (1− αB)ηn , −1 < α < 1 . (25)

The autocorrelation function is

ρl = ρl(w)al2 − (1− d)2

l2 − (1− d)2, (26a)

a = (1− α)2[1 + α2 −

2αd

1− d

]−1. (26b)

Copyright c© 2009-18 P. F. Góra 8–34