Comments
Transcript
An introduction to ARMA and GARCH processes 402KB
Introduction to ARMA and GARCH processes Fulvio Corsi SNS Pisa 3 March 2010 Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 1 / 24 Stationarity Strict stationarity: d (X1 , X2 , ..., Xn ) = (X1+k , X2+k , ..., Xn+k ) n > 1, k for any integer Weak/second-order/covariance stationarity: E[xt ] = µ E[Xt − µ]2 = σ 2 < +∞ (i.e. constant and indipendent of t) E[(Xt − µ)(Xt+k − µ)] = γ(|k|) (i.e. indipendent of t for each k) Interpretation: mean and variance are constant mean reversion shocks are transient covariance between Xt and Xt−k tends to 0 as k → ∞ Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 2 / 24 Stationarity Strict stationarity: d (X1 , X2 , ..., Xn ) = (X1+k , X2+k , ..., Xn+k ) n > 1, k for any integer Weak/second-order/covariance stationarity: E[xt ] = µ E[Xt − µ]2 = σ 2 < +∞ (i.e. constant and indipendent of t) E[(Xt − µ)(Xt+k − µ)] = γ(|k|) (i.e. indipendent of t for each k) Interpretation: mean and variance are constant mean reversion shocks are transient covariance between Xt and Xt−k tends to 0 as k → ∞ Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 2 / 24 Stationarity Strict stationarity: d (X1 , X2 , ..., Xn ) = (X1+k , X2+k , ..., Xn+k ) n > 1, k for any integer Weak/second-order/covariance stationarity: E[xt ] = µ E[Xt − µ]2 = σ 2 < +∞ (i.e. constant and indipendent of t) E[(Xt − µ)(Xt+k − µ)] = γ(|k|) (i.e. indipendent of t for each k) Interpretation: mean and variance are constant mean reversion shocks are transient covariance between Xt and Xt−k tends to 0 as k → ∞ Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 2 / 24 Withe noise weak (uncorrelated) E(ǫt ) = 0 V(ǫt ) = σ 2 ρ(ǫt , ǫs ) = 0 ∀t ∀t ∀s 6= t where ρ≡ γ(|t−s|) γ(0) strong (independence) ǫt ∼ I.I.D.(0, σ 2 ) Gaussian (weak=strong) ǫt ∼ N.I.D.(0, σ 2 ) Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 3 / 24 Lag operator the Lag operator is defined as: LXt ≡ Xt−1 is a linear operator: L(βXt ) = L(Xt + Yt ) = β · LXt = βXt−1 LXt + LYt = Xt−1 + Yt−1 and admits power exponent, for instance: L2 Xt = L(LXt ) = LXt−1 = Xt−2 k L Xt = Xt−k L−1 Xt = Xt+1 Some examples: ∆Xt = yt = Xt − Xt−1 = Xt − LXt = (1 − L)Xt (θ1 + θ2 L)LXt = (θ1 L + θ2 L2 )Xt = θ1 Xt−1 + θ2 Xt−2 Expression like (θ0 + θ1 L + θ2 L2 + ... + θn Ln ) with possibly n = ∞, are called lag polynomial and are indicated as θ(L) Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 4 / 24 Lag operator the Lag operator is defined as: LXt ≡ Xt−1 is a linear operator: L(βXt ) = L(Xt + Yt ) = β · LXt = βXt−1 LXt + LYt = Xt−1 + Yt−1 and admits power exponent, for instance: L2 Xt = L(LXt ) = LXt−1 = Xt−2 k L Xt = Xt−k L−1 Xt = Xt+1 Some examples: ∆Xt = yt = Xt − Xt−1 = Xt − LXt = (1 − L)Xt (θ1 + θ2 L)LXt = (θ1 L + θ2 L2 )Xt = θ1 Xt−1 + θ2 Xt−2 Expression like (θ0 + θ1 L + θ2 L2 + ... + θn Ln ) with possibly n = ∞, are called lag polynomial and are indicated as θ(L) Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 4 / 24 Lag operator the Lag operator is defined as: LXt ≡ Xt−1 is a linear operator: L(βXt ) = L(Xt + Yt ) = β · LXt = βXt−1 LXt + LYt = Xt−1 + Yt−1 and admits power exponent, for instance: L2 Xt = L(LXt ) = LXt−1 = Xt−2 k L Xt = Xt−k L−1 Xt = Xt+1 Some examples: ∆Xt = yt = Xt − Xt−1 = Xt − LXt = (1 − L)Xt (θ1 + θ2 L)LXt = (θ1 L + θ2 L2 )Xt = θ1 Xt−1 + θ2 Xt−2 Expression like (θ0 + θ1 L + θ2 L2 + ... + θn Ln ) with possibly n = ∞, are called lag polynomial and are indicated as θ(L) Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 4 / 24 Lag operator the Lag operator is defined as: LXt ≡ Xt−1 is a linear operator: L(βXt ) = L(Xt + Yt ) = β · LXt = βXt−1 LXt + LYt = Xt−1 + Yt−1 and admits power exponent, for instance: L2 Xt = L(LXt ) = LXt−1 = Xt−2 k L Xt = Xt−k L−1 Xt = Xt+1 Some examples: ∆Xt = yt = Xt − Xt−1 = Xt − LXt = (1 − L)Xt (θ1 + θ2 L)LXt = (θ1 L + θ2 L2 )Xt = θ1 Xt−1 + θ2 Xt−2 Expression like (θ0 + θ1 L + θ2 L2 + ... + θn Ln ) with possibly n = ∞, are called lag polynomial and are indicated as θ(L) Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 4 / 24 Moving Average (MA) process The way to construct a stationary process is to use a lag polynomial θ(L) with P ∞simplest 2 j=0 θj < ∞ to construct a sort of “weighted moving average” of withe noises ǫt , i.e. MA(q) Yt = θ(L)ǫt = ǫt + θ1 ǫt−1 + θ2 ǫt−2 + ... + +θq ǫt−q Example, MA(1) Yt = ǫt + θǫt−1 = (1 + θL)ǫt being E[Yt ] = 0 γ(0) = E[Yt Yt ] = E[(ǫt + θǫt−1 )(ǫt + θǫt−1 )] = σ 2 (1 + θ2 ); γ(1) = E[Yt Yt−1 ] = E[(ǫt + θǫt−1 )(ǫt−1 + θǫt−2 )] = σ 2 θ; γ(k) = E[Yt Yt−k ] = E[(ǫt + θǫt−1 )(ǫt−k + θǫt−k−1 )] = 0 and, ρ(1) = ρ(k) = ∀k > 1 θ γ(1) = γ(0) 1 + θ2 γ(k) =0 ∀k > 1 γ(0) hence, while a withe noise is “0-correlated”, MA(1) is 1-correlated (i.e. it has only the first correlation ρ(1) different from zero) Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 5 / 24 Moving Average (MA) process The way to construct a stationary process is to use a lag polynomial θ(L) with P ∞simplest 2 j=0 θj < ∞ to construct a sort of “weighted moving average” of withe noises ǫt , i.e. MA(q) Yt = θ(L)ǫt = ǫt + θ1 ǫt−1 + θ2 ǫt−2 + ... + +θq ǫt−q Example, MA(1) Yt = ǫt + θǫt−1 = (1 + θL)ǫt being E[Yt ] = 0 γ(0) = E[Yt Yt ] = E[(ǫt + θǫt−1 )(ǫt + θǫt−1 )] = σ 2 (1 + θ2 ); γ(1) = E[Yt Yt−1 ] = E[(ǫt + θǫt−1 )(ǫt−1 + θǫt−2 )] = σ 2 θ; γ(k) = E[Yt Yt−k ] = E[(ǫt + θǫt−1 )(ǫt−k + θǫt−k−1 )] = 0 and, ρ(1) = ρ(k) = ∀k > 1 θ γ(1) = γ(0) 1 + θ2 γ(k) =0 ∀k > 1 γ(0) hence, while a withe noise is “0-correlated”, MA(1) is 1-correlated (i.e. it has only the first correlation ρ(1) different from zero) Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 5 / 24 Properties MA(q) In general for a MA(q) process we have γ(0) = σ 2 (1 + θ12 + θ22 + ... + θq2 ) γ(k) = σ2 q−k X θj θj+k j=0 = ∀k ≤ q ∀k > q 0 and ρ(k) = = P q−k θj θj+k Pq 2 j=1 θj j=0 1+ 0 ∀k ≤ q ∀k > q Hence, an MA(q) is q-correlated and it can also be shown that any stationary q-correlated process can be represented as an MA(q). Wold Theorem: any mean zero covariance stationary process can be represented in the form, MA(∞) + deterministic component (the two being uncorrelated). But, given a q-correlated process, is the MA(q) process unique? In general no, indeed it can be shown that for a q-correlated process there are 2q possible MA(q) with same autocovariance structure. However, there is only one MA(q) which is invertible. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 6 / 24 Properties MA(q) In general for a MA(q) process we have γ(0) = σ 2 (1 + θ12 + θ22 + ... + θq2 ) γ(k) = σ2 q−k X θj θj+k j=0 = ∀k ≤ q ∀k > q 0 and ρ(k) = = P q−k θj θj+k Pq 2 j=1 θj j=0 1+ 0 ∀k ≤ q ∀k > q Hence, an MA(q) is q-correlated and it can also be shown that any stationary q-correlated process can be represented as an MA(q). Wold Theorem: any mean zero covariance stationary process can be represented in the form, MA(∞) + deterministic component (the two being uncorrelated). But, given a q-correlated process, is the MA(q) process unique? In general no, indeed it can be shown that for a q-correlated process there are 2q possible MA(q) with same autocovariance structure. However, there is only one MA(q) which is invertible. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 6 / 24 Properties MA(q) In general for a MA(q) process we have γ(0) = σ 2 (1 + θ12 + θ22 + ... + θq2 ) γ(k) = σ2 q−k X θj θj+k j=0 = ∀k ≤ q ∀k > q 0 and ρ(k) = = P q−k θj θj+k Pq 2 j=1 θj j=0 1+ 0 ∀k ≤ q ∀k > q Hence, an MA(q) is q-correlated and it can also be shown that any stationary q-correlated process can be represented as an MA(q). Wold Theorem: any mean zero covariance stationary process can be represented in the form, MA(∞) + deterministic component (the two being uncorrelated). But, given a q-correlated process, is the MA(q) process unique? In general no, indeed it can be shown that for a q-correlated process there are 2q possible MA(q) with same autocovariance structure. However, there is only one MA(q) which is invertible. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 6 / 24 Invertibility conditions for MA first consider the MA(1) case: Yt = (1 + θL)ǫt given the result (1 + θL)−1 = (1 − θL + θ2 L2 − θ3 L3 + θ4 L4 + ...) = ∞ X (−θL)i i=0 inverting the θ(L) lag polynomial, we can write (1 − θL + θ2 L2 − θ3 L3 + θ4 L4 + ...)Yt = ǫt which can be considered an AR(∞) process. If an MA process can be written as an AR(∞) of this type, such MA representation is said to be invertible. For MA(1) process the invertibility condition is given by |θ| < 1. For a general MA(q) process Yt = (1 + θ1 L + θ2 L2 + ... + θq Lq )ǫt the invertibility conditions are that the roots of the lag polynomial 1 + θ1 z + θ2 z2 + ... + θq zq = 0 lie outside the unit circle. Then the MA(q) can be written as an AR(∞) by inverting θ(L). Invertibility also has important practical consequence in application. In fact, given that the ǫt are not observable they have to be reconstructed from the observed Y’s through the AR(∞) representation. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 7 / 24 Invertibility conditions for MA first consider the MA(1) case: Yt = (1 + θL)ǫt given the result (1 + θL)−1 = (1 − θL + θ2 L2 − θ3 L3 + θ4 L4 + ...) = ∞ X (−θL)i i=0 inverting the θ(L) lag polynomial, we can write (1 − θL + θ2 L2 − θ3 L3 + θ4 L4 + ...)Yt = ǫt which can be considered an AR(∞) process. If an MA process can be written as an AR(∞) of this type, such MA representation is said to be invertible. For MA(1) process the invertibility condition is given by |θ| < 1. For a general MA(q) process Yt = (1 + θ1 L + θ2 L2 + ... + θq Lq )ǫt the invertibility conditions are that the roots of the lag polynomial 1 + θ1 z + θ2 z2 + ... + θq zq = 0 lie outside the unit circle. Then the MA(q) can be written as an AR(∞) by inverting θ(L). Invertibility also has important practical consequence in application. In fact, given that the ǫt are not observable they have to be reconstructed from the observed Y’s through the AR(∞) representation. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 7 / 24 Auto-Regressive Process (AR) A general AR process is defined as φ(L)Yt = ǫt It is always invertible but not always stationary. Example: AR(1) (1 − φL)Yt = ǫt or Yt = φYt−1 + ǫt by inverting the lag polynomial (1 − φL) the AR(1) can be written as Yt = (1 − φL)−1 ǫt = ∞ X (φL)i ǫt = i=0 ∞ X φi ǫt−i = MA(∞) i=0 hence the stationarity condition is that |φ| < 1. From this representation we can apply the general formula of MA to compute γ(·) and ρ(·). In particular, ρ(k) = φ|k| ∀k i.e. monotonic exponential decay for φ > 0 and exponentially damped oscillatory decay for φ < 0. In general an AR(p) process Yt = φ1 Yt−1 + φ2 Yt−2 + ... + φp Yt−p + ǫt is stationarity if all the roots of the characteristic equation of the lag polynomial are outside the unit circle. Fulvio Corsi 1 − φ1 z − φ2 z2 − ... − φp zp = 0 Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 8 / 24 Auto-Regressive Process (AR) A general AR process is defined as φ(L)Yt = ǫt It is always invertible but not always stationary. Example: AR(1) (1 − φL)Yt = ǫt or Yt = φYt−1 + ǫt by inverting the lag polynomial (1 − φL) the AR(1) can be written as Yt = (1 − φL)−1 ǫt = ∞ X (φL)i ǫt = i=0 ∞ X φi ǫt−i = MA(∞) i=0 hence the stationarity condition is that |φ| < 1. From this representation we can apply the general formula of MA to compute γ(·) and ρ(·). In particular, ρ(k) = φ|k| ∀k i.e. monotonic exponential decay for φ > 0 and exponentially damped oscillatory decay for φ < 0. In general an AR(p) process Yt = φ1 Yt−1 + φ2 Yt−2 + ... + φp Yt−p + ǫt is stationarity if all the roots of the characteristic equation of the lag polynomial are outside the unit circle. Fulvio Corsi 1 − φ1 z − φ2 z2 − ... − φp zp = 0 Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 8 / 24 State Space Representation of AR(p) to gain more intuition on the AR stationarity conditions write an AR(p) in its state space form Yt φ1 φ2 φ3 . . . φp−1 φp ǫt Yt−1 Yt−1 1 0 0 ... 0 0 Yt−2 0 = . . . . . . . + . .. .. .. .. .. .. .. .. ... Yt−p+1 Xt 0 0 0 = ... 1 Yt−p 0 F Xt−1 0 + vt Hence, the expected value of Xt satisfy, E[Xt ] = F Xt−1 and E[Xt+j ] = F j+1 Xt−1 is a linear map in Rp whose dynamic properties are given by the eigenvalues of the matrix F. The eigenvalues of F are given by solving the characteristic equation λp − φ1 λp−1 − φ2 λp−2 − ... − φp−1 λ − φp = 0. Comparing this with the characteristic equation of the lag polynomial φ(L) 1 − φ1 z − φ2 z2 − ... − φp−1 zp−1 − φp zp = 0 we can see that the roots of the 2 equations are such that z1 = λ−1 1 , Fulvio Corsi z2 = λ−1 2 , ... , zp = λ−1 p Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 9 / 24 State Space Representation of AR(p) to gain more intuition on the AR stationarity conditions write an AR(p) in its state space form Yt φ1 φ2 φ3 . . . φp−1 φp ǫt Yt−1 Yt−1 1 0 0 ... 0 0 Yt−2 0 = . . . . . . . + . .. .. .. .. .. .. .. .. ... Yt−p+1 Xt 0 0 0 = ... 1 Yt−p 0 F Xt−1 0 + vt Hence, the expected value of Xt satisfy, E[Xt ] = F Xt−1 and E[Xt+j ] = F j+1 Xt−1 is a linear map in Rp whose dynamic properties are given by the eigenvalues of the matrix F. The eigenvalues of F are given by solving the characteristic equation λp − φ1 λp−1 − φ2 λp−2 − ... − φp−1 λ − φp = 0. Comparing this with the characteristic equation of the lag polynomial φ(L) 1 − φ1 z − φ2 z2 − ... − φp−1 zp−1 − φp zp = 0 we can see that the roots of the 2 equations are such that z1 = λ−1 1 , Fulvio Corsi z2 = λ−1 2 , ... , zp = λ−1 p Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 9 / 24 ARMA(p,q) An ARMA(p,q) process is defined as φ(L)Yt = θ(L)ǫt where φ(·) and θ(·) are pth and qth lag polynomials. the process is stationary if all the roots of φ(z) ≡ 1 − φ1 z − φ2 z2 − ... − φp−1 zp−1 − φp zp = 0 lie outside the unit circle and, hence, admits the MA(∞) representation: Yt = φ(L)−1 θ(L)ǫt the process is invertible if all the roots of θ(z) ≡ 1 + θ1 z + θ2 z2 + ... + θq zq = 0 lie outside the unit circle and, hence, admits the AR(∞) representation: ǫt = θ(L)−1 φ(L)Yt Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 10 / 24 Estimation For AR(p) Yt = φ1 Yt−1 + φ2 Yt−2 + ... + φp Yt−p + ǫt OLS are consistent and under gaussianity asymptotically equivalent to MLE → asymptotically efficient For example, the full Likelihood of an AR(1) can be written as L(θ) ≡ f (yT , yT−1 , ..., y1 ; θ) = fY1 (y1 ; θ) {z } | · marginal first obs T Y t=2 | fYt |Yt−1 (yt |yt−1 ; θ) {z } conditional likelihood under normality OLS=MLE For a general ARMA(p,q) Yt = φ1 Yt−1 + ... + φp Yt−p + ǫt + θ1 ǫt−1 + ... + θq ǫt−q Yt−1 is correlated with ǫt−1 , ..., ǫt−q ⇒ OLS not consistent. → MLE with numerical optimization procedures. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 11 / 24 Estimation For AR(p) Yt = φ1 Yt−1 + φ2 Yt−2 + ... + φp Yt−p + ǫt OLS are consistent and under gaussianity asymptotically equivalent to MLE → asymptotically efficient For example, the full Likelihood of an AR(1) can be written as L(θ) ≡ f (yT , yT−1 , ..., y1 ; θ) = fY1 (y1 ; θ) {z } | · marginal first obs T Y t=2 | fYt |Yt−1 (yt |yt−1 ; θ) {z } conditional likelihood under normality OLS=MLE For a general ARMA(p,q) Yt = φ1 Yt−1 + ... + φp Yt−p + ǫt + θ1 ǫt−1 + ... + θq ǫt−q Yt−1 is correlated with ǫt−1 , ..., ǫt−q ⇒ OLS not consistent. → MLE with numerical optimization procedures. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 11 / 24 Prediction write the model in its AR(∞) representation: η(L)(Yt − µ) = ǫt then the optimal prediction of Yt+s is given by η(L)−1 E[Yt+s |Yt , Yt−1 , ...] = µ + η(L)(Yt − µ) Ls + with h i Lk = 0 for k < 0 + which is known as Wiener-Kolmogorov prediction formula. In the case of an AR(p) process the prediction formula can also be written as (s) (s) (s) E[Yt+s |Yt , Yt−1 , ...] = µ + f11 (Yt − µ) + f12 (Yt−1 − µ) + ... + f1p (Yt−p+1 − µ) (j) where f11 is the element (1, 1) of the matrix F j . The easiest way to compute prediction from AR(p) model is, however, through recursive methods. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 12 / 24 Prediction write the model in its AR(∞) representation: η(L)(Yt − µ) = ǫt then the optimal prediction of Yt+s is given by η(L)−1 E[Yt+s |Yt , Yt−1 , ...] = µ + η(L)(Yt − µ) Ls + with h i Lk = 0 for k < 0 + which is known as Wiener-Kolmogorov prediction formula. In the case of an AR(p) process the prediction formula can also be written as (s) (s) (s) E[Yt+s |Yt , Yt−1 , ...] = µ + f11 (Yt − µ) + f12 (Yt−1 − µ) + ... + f1p (Yt−p+1 − µ) (j) where f11 is the element (1, 1) of the matrix F j . The easiest way to compute prediction from AR(p) model is, however, through recursive methods. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 12 / 24 Box-Jenkins Approach check for stationarity: if not try different transformation (ex differentiation→ARIMA models) Identification: check the autocorrelation (ACF) function: a q-correlated process is an MA(q) model check the partial autocorrelation (PACF) function: for an AR(p) process, while the k–lag ACF can be interpreted as simple regression Yt = ρ(k)Yt−k + error, the k–lag PACF can be seen as a multiple regression Yt = b1 Yt−1 + b2 Yt−2 + ... + bk Yt−k + error it can be computed by solving the Yule-Walker system: b1 b2 . . . bk = γ(0) γ(1) . . . γ(k − 1) γ(1) γ(0) . . . γ(k − 2) ... ... ... ... γ(k − 1) γ(k − 2) . . . γ(0) γ(1) γ(2) . . . γ(k) −1 Importantly, AR(p) processes are “p–partially correlated” ⇒ identification of AR order Validation: check the appropriateness of the model by some measure of fit. AIC/Akaike = T log σ̂e2 + 2 m BIC/Schwarz = T log σ̂e2 + m log T with σe2 estimation error variance, m = p + q + 1 n◦ of parameters, and T n◦ of obs Diagnostic checking of the residuals. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 13 / 24 Box-Jenkins Approach check for stationarity: if not try different transformation (ex differentiation→ARIMA models) Identification: check the autocorrelation (ACF) function: a q-correlated process is an MA(q) model check the partial autocorrelation (PACF) function: for an AR(p) process, while the k–lag ACF can be interpreted as simple regression Yt = ρ(k)Yt−k + error, the k–lag PACF can be seen as a multiple regression Yt = b1 Yt−1 + b2 Yt−2 + ... + bk Yt−k + error it can be computed by solving the Yule-Walker system: b1 b2 . . . bk = γ(0) γ(1) . . . γ(k − 1) γ(1) γ(0) . . . γ(k − 2) ... ... ... ... γ(k − 1) γ(k − 2) . . . γ(0) γ(1) γ(2) . . . γ(k) −1 Importantly, AR(p) processes are “p–partially correlated” ⇒ identification of AR order Validation: check the appropriateness of the model by some measure of fit. AIC/Akaike = T log σ̂e2 + 2 m BIC/Schwarz = T log σ̂e2 + m log T with σe2 estimation error variance, m = p + q + 1 n◦ of parameters, and T n◦ of obs Diagnostic checking of the residuals. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 13 / 24 Box-Jenkins Approach check for stationarity: if not try different transformation (ex differentiation→ARIMA models) Identification: check the autocorrelation (ACF) function: a q-correlated process is an MA(q) model check the partial autocorrelation (PACF) function: for an AR(p) process, while the k–lag ACF can be interpreted as simple regression Yt = ρ(k)Yt−k + error, the k–lag PACF can be seen as a multiple regression Yt = b1 Yt−1 + b2 Yt−2 + ... + bk Yt−k + error it can be computed by solving the Yule-Walker system: b1 b2 . . . bk = γ(0) γ(1) . . . γ(k − 1) γ(1) γ(0) . . . γ(k − 2) ... ... ... ... γ(k − 1) γ(k − 2) . . . γ(0) γ(1) γ(2) . . . γ(k) −1 Importantly, AR(p) processes are “p–partially correlated” ⇒ identification of AR order Validation: check the appropriateness of the model by some measure of fit. AIC/Akaike = T log σ̂e2 + 2 m BIC/Schwarz = T log σ̂e2 + m log T with σe2 estimation error variance, m = p + q + 1 n◦ of parameters, and T n◦ of obs Diagnostic checking of the residuals. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 13 / 24 ARIMA Integrated ARMA model: ARIMA(p,1,q) denote a nonstationary process Yt for which the first difference Yt − Yt−1 = (1 − L)Yt is a stationary ARMA(p,q) process. ⇓ Yt is said to be integrated of order 1 or I(1). If 2 differentiations of Yt are necessary to get a stationary process i.e. (1 − L)2 Yt ⇓ then the process Yt is said to be integrated of order 2 or I(2). I(0) indicate a stationary process. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 14 / 24 ARIMA Integrated ARMA model: ARIMA(p,1,q) denote a nonstationary process Yt for which the first difference Yt − Yt−1 = (1 − L)Yt is a stationary ARMA(p,q) process. ⇓ Yt is said to be integrated of order 1 or I(1). If 2 differentiations of Yt are necessary to get a stationary process i.e. (1 − L)2 Yt ⇓ then the process Yt is said to be integrated of order 2 or I(2). I(0) indicate a stationary process. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 14 / 24 ARFIMA The k-difference operator (1 − L)n with integer n can be generalized to a fractional difference operator (1 − L)d with 0 < d < 1 defined by the binomial expansion (1 − L)d = 1 − dL + d(d − 1)L2 /2! − d(d − 1)(d − 2)L3 /3! + ... obtaining a fractionally integrated process of order d i.e. I(d). If d < 0.5 the process is cov stationary and admits an AR(∞) representation. The usefulness of a fractional filter (1 − L)d is that it produces hyperbolic decaying autocorrelations i.e. the so called long memory. In fact, for ARFIMA(p,d,q) processes φ(L)(1 − L)d Yt = θ(L)ǫt the autocorrelation functions is proportional to ρ(k) ≈ ck2d−1 Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 15 / 24 ARFIMA The k-difference operator (1 − L)n with integer n can be generalized to a fractional difference operator (1 − L)d with 0 < d < 1 defined by the binomial expansion (1 − L)d = 1 − dL + d(d − 1)L2 /2! − d(d − 1)(d − 2)L3 /3! + ... obtaining a fractionally integrated process of order d i.e. I(d). If d < 0.5 the process is cov stationary and admits an AR(∞) representation. The usefulness of a fractional filter (1 − L)d is that it produces hyperbolic decaying autocorrelations i.e. the so called long memory. In fact, for ARFIMA(p,d,q) processes φ(L)(1 − L)d Yt = θ(L)ǫt the autocorrelation functions is proportional to ρ(k) ≈ ck2d−1 Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 15 / 24 ARCH and GARCH models Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 16 / 24 Basic Structure and Properties of ARMA model standard time series models have: Yt = E[Yt |Ωt−1 ] + ǫt E[Yt |Ωt−1 ] = f (Ωt−1 ; θ) i h Var [Yt |Ωt−1 ] = E ǫ2t |Ωt−1 = σ 2 hence, Conditional mean: varies with Ωt−1 Conditional variance: constant (unfortunately) k-step-ahead mean forecasts: generally depends on Ωt−1 k-step-ahead variance forecasts : depends only on k, not on Ωt−1 (again unfortunately) Unconditional mean: constant Unconditional variance: constant Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 17 / 24 AutoRegressive Conditional Heteroskedasticity (ARCH) model Engle (1982, Econometrica) intruduced the ARCH models: Yt = E[Yt |Ωt−1 ] + ǫt E[Yt |Ωt−1 ] = f (Ωt−1 ; θ) i h Var [Yt |Ωt−1 ] = E ǫ2t |Ωt−1 = σ (Ωt−1 ; θ) ≡ σt2 hence, Conditional mean: varies with Ωt−1 Conditional variance: varies with Ωt−1 k-step-ahead mean forecasts: generally depends on Ωt−1 k-step-ahead variance forecasts : generally depends on Ωt−1 Unconditional mean: constant Unconditional variance: constant Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 18 / 24 ARCH(q) How to parameterize E ǫ2t |Ωt−1 = σ((Ωt−1 ; θ) ≡ σt2 ? ARCH(q) postulated that the conditional variance is a linear function of the past q squared innovations q X αi ǫ2t−i = ω + α(L)ǫ2t−1 σt2 = ω + i=1 Defining vt = ǫ2t − σt2 , the ARCH(q) model can be written as ǫ2t = ω + α(L)ǫ2t−1 + vt Since Et−1 (vt ) = 0, the model corresponds directly to an AR(q) model for the squared innovations, ǫ2t . The process is covariance stationary if and only if the sum of the positive AR parameters is less than 1. Then, the unconditional variance of ǫt is Var(ǫt ) = σ 2 = w/(1 − α1 − α2 − ... − αq ). Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 19 / 24 ARCH(q) How to parameterize E ǫ2t |Ωt−1 = σ((Ωt−1 ; θ) ≡ σt2 ? ARCH(q) postulated that the conditional variance is a linear function of the past q squared innovations q X αi ǫ2t−i = ω + α(L)ǫ2t−1 σt2 = ω + i=1 Defining vt = ǫ2t − σt2 , the ARCH(q) model can be written as ǫ2t = ω + α(L)ǫ2t−1 + vt Since Et−1 (vt ) = 0, the model corresponds directly to an AR(q) model for the squared innovations, ǫ2t . The process is covariance stationary if and only if the sum of the positive AR parameters is less than 1. Then, the unconditional variance of ǫt is Var(ǫt ) = σ 2 = w/(1 − α1 − α2 − ... − αq ). Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 19 / 24 AR(1)-ARCH(1) Example: the AR(1)-ARCH(1) model - Yt = φYt−1 + ǫt σt2 = ω + αǫ2t−1 ǫt ∼ N(0, σt2 ) Conditional mean: E(Yt |Ωt−1 ) = φYt−1 Conditional variance: E([Yt − E(Yt |Ωt−1 )]2 |Ωt−1 = ω + αǫ2t−1 Unconditional mean: E(Yt ) = 0 ω 1 Unconditional variance: E(Yt − E(Yt ))2 = (1−φ 2 ) (1−α) Note that the unconditional distribution of ǫt has Fat Tail. In fact, the unconditional kurtosis E(ǫ4t )/E(ǫ2t )2 is h i h i h i E ǫ4t = E E(ǫ4t |Ωt−1 ) = 3E σt4 = 3[Var(σt2 ) + E(σt2 )2 ] = 3[Var(σt2 ) +E(ǫ2t )2 ] > 3E(ǫ2t )2 . | {z } >0 Hence, Kurtosis(ǫt ) = E(ǫ4t )/E(ǫ2t )2 > 3 Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 20 / 24 AR(1)-ARCH(1) Example: the AR(1)-ARCH(1) model - Yt = φYt−1 + ǫt σt2 = ω + αǫ2t−1 ǫt ∼ N(0, σt2 ) Conditional mean: E(Yt |Ωt−1 ) = φYt−1 Conditional variance: E([Yt − E(Yt |Ωt−1 )]2 |Ωt−1 = ω + αǫ2t−1 Unconditional mean: E(Yt ) = 0 ω 1 Unconditional variance: E(Yt − E(Yt ))2 = (1−φ 2 ) (1−α) Note that the unconditional distribution of ǫt has Fat Tail. In fact, the unconditional kurtosis E(ǫ4t )/E(ǫ2t )2 is h i h i h i E ǫ4t = E E(ǫ4t |Ωt−1 ) = 3E σt4 = 3[Var(σt2 ) + E(σt2 )2 ] = 3[Var(σt2 ) +E(ǫ2t )2 ] > 3E(ǫ2t )2 . | {z } >0 Hence, Kurtosis(ǫt ) = E(ǫ4t )/E(ǫ2t )2 > 3 Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 20 / 24 GARCH(p,q) Problem: empirical volatility very persistent ⇒ Large q i.e. too many α’s Bollerslev (1986, J. of Econometrics) proposed the Generalized ARCH model. The GARCH(p,q) is defined as σt2 = ω + q X i=1 αi ǫ2t−i + p X 2 2 βj σt−j = ω + α(L)ǫ2t−1 + β(L)σt−1 j=1 As before also the GARCH(p,q) can be rewritten as ǫ2t = ω + [α(L) + β(L)] ǫ2t−1 − β(L)vt−1 + vt which defines an ARMA[max(p, q),p] model for et z. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 21 / 24 GARCH(p,q) Problem: empirical volatility very persistent ⇒ Large q i.e. too many α’s Bollerslev (1986, J. of Econometrics) proposed the Generalized ARCH model. The GARCH(p,q) is defined as σt2 = ω + q X i=1 αi ǫ2t−i + p X 2 2 βj σt−j = ω + α(L)ǫ2t−1 + β(L)σt−1 j=1 As before also the GARCH(p,q) can be rewritten as ǫ2t = ω + [α(L) + β(L)] ǫ2t−1 − β(L)vt−1 + vt which defines an ARMA[max(p, q),p] model for et z. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 21 / 24 GARCH(1,1) By far the most commonly used is the GARCH(1,1): 2 σt2 = ω + α ǫ2t−1 + βj σt−1 with ω > 0, α > 0, β > 0. By recursive substitution, the GARCH(1,1) may be written as the following ARCH(∞): σt2 = ω(1 − β) + α ∞ X β i−1 ǫ2t−i i=1 which reduces to an exponentially weighted moving average filter for ω = 0 and α + β = 1 (sometimes referred to as Integrated GARCH or IGARCH(1,1)). Moreover, GARCH(1,1) implies an ARMA(1,1) representation in the ǫ2t ǫ2t = ω + [α + β]ǫ2t−1 − βvt−1 + vt Forecasting. Denoting the unconditional variance σ 2 ≡ ω(1 − α − β)−1 we have: 2 2 σ̂t+h|t = σ 2 + (α + β)h−1 (σt+1 − σ2 ) showing that the forecasts of the conditional variance revert to the long-run unconditional variance at an exponential rate dictated by α + β Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 22 / 24 GARCH(1,1) By far the most commonly used is the GARCH(1,1): 2 σt2 = ω + α ǫ2t−1 + βj σt−1 with ω > 0, α > 0, β > 0. By recursive substitution, the GARCH(1,1) may be written as the following ARCH(∞): σt2 = ω(1 − β) + α ∞ X β i−1 ǫ2t−i i=1 which reduces to an exponentially weighted moving average filter for ω = 0 and α + β = 1 (sometimes referred to as Integrated GARCH or IGARCH(1,1)). Moreover, GARCH(1,1) implies an ARMA(1,1) representation in the ǫ2t ǫ2t = ω + [α + β]ǫ2t−1 − βvt−1 + vt Forecasting. Denoting the unconditional variance σ 2 ≡ ω(1 − α − β)−1 we have: 2 2 σ̂t+h|t = σ 2 + (α + β)h−1 (σt+1 − σ2 ) showing that the forecasts of the conditional variance revert to the long-run unconditional variance at an exponential rate dictated by α + β Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 22 / 24 Asymmetric GARCH News Impact Curve 8 symmetric GARCH asymmetric GARCH σt2 =ω+ 2 αrt−1 + 2 βσt−1 σt2 responds symmetrically to past returns. conditional variance In standard GARCH model: 6 4 2 The so called “news impact curve” is a parabola 0 −5 0 5 standardized lagged shocks Empirically negative rt−1 impact more than positive ones → asymmetric news impact curve GJR or T-GARCH σt2 = 2 2 2 ω + αrt−1 + γrt−1 Dt−1 + βσt−1 with Dt = 1 0 if rt < 0 otherwise - Positive returns (good news): α - Negative returns (bad news): α + γ - Empirically γ > 0 → “Leverage effect” Exponential GARCH (EGARCH) Fulvio Corsi rt−1 + γ rt−1 + βln(σ 2 ) ln(σt2 ) = ω + α t−1 σt−1 σt−1 Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 23 / 24 Asymmetric GARCH News Impact Curve 8 symmetric GARCH asymmetric GARCH σt2 =ω+ 2 αrt−1 + 2 βσt−1 σt2 responds symmetrically to past returns. conditional variance In standard GARCH model: 6 4 2 The so called “news impact curve” is a parabola 0 −5 0 5 standardized lagged shocks Empirically negative rt−1 impact more than positive ones → asymmetric news impact curve GJR or T-GARCH σt2 = 2 2 2 ω + αrt−1 + γrt−1 Dt−1 + βσt−1 with Dt = 1 0 if rt < 0 otherwise - Positive returns (good news): α - Negative returns (bad news): α + γ - Empirically γ > 0 → “Leverage effect” Exponential GARCH (EGARCH) Fulvio Corsi rt−1 + γ rt−1 + βln(σ 2 ) ln(σt2 ) = ω + α t−1 σt−1 σt−1 Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 23 / 24 Estimation A GARCH process with gaussian innovation: rt |Ωt−1 ∼ N(µt (θ), σt2 (θ)) has conditional densities: 1 1 (rt − µ(θ)) f (rt |Ωt−1 ; θ) = √ σt−1 (θ) exp − 2 2 σt (θ) 2π using the “prediction–error” decomposition of the likelihood L(rT , rT−1 , ..., r1 ; θ) = f (rT |ΩT−1 ; θ) × f (rT−1 |ΩT−2 ; θ) × ... × f (r1 |Ω0 ; θ) the log-likelihood becomes: log L(rT , rT−1 , ..., r1 ; θ) = − T T X 1 X (rt − µ(θ)) T log σt (θ) − log(2π) − 2 2 t=1 σt2 (θ) t=1 Non–linear function in θ ⇒ Numerical optimization techiques. Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 24 / 24