Escolar Documentos
Profissional Documentos
Cultura Documentos
Fulvio Corsi
SNS Pisa
3 March 2010
Strict stationarity:
d
(X1 , X2 , ..., Xn ) = (X1+k , X2+k , ..., Xn+k ) for any integer n > 1, k
Weak/second-order/covariance stationarity:
E[xt ] = µ
E[Xt − µ]2 = σ 2 < +∞ (i.e. constant and indipendent of t)
E[(Xt − µ)(Xt+k − µ)] = γ(|k|) (i.e. indipendent of t for each k)
Interpretation:
mean and variance are constant
mean reversion
shocks are transient
covariance between Xt and Xt−k tends to 0 as k → ∞
Strict stationarity:
d
(X1 , X2 , ..., Xn ) = (X1+k , X2+k , ..., Xn+k ) for any integer n > 1, k
Weak/second-order/covariance stationarity:
E[xt ] = µ
E[Xt − µ]2 = σ 2 < +∞ (i.e. constant and indipendent of t)
E[(Xt − µ)(Xt+k − µ)] = γ(|k|) (i.e. indipendent of t for each k)
Interpretation:
mean and variance are constant
mean reversion
shocks are transient
covariance between Xt and Xt−k tends to 0 as k → ∞
Strict stationarity:
d
(X1 , X2 , ..., Xn ) = (X1+k , X2+k , ..., Xn+k ) for any integer n > 1, k
Weak/second-order/covariance stationarity:
E[xt ] = µ
E[Xt − µ]2 = σ 2 < +∞ (i.e. constant and indipendent of t)
E[(Xt − µ)(Xt+k − µ)] = γ(|k|) (i.e. indipendent of t for each k)
Interpretation:
mean and variance are constant
mean reversion
shocks are transient
covariance between Xt and Xt−k tends to 0 as k → ∞
weak (uncorrelated)
E(ǫt ) = 0 ∀t
V(ǫt ) = σ 2 ∀t
γ(|t−s|)
ρ(ǫt , ǫs ) = 0 ∀s 6= t where ρ≡ γ(0)
strong (independence)
ǫt ∼ I.I.D.(0, σ 2 )
Gaussian (weak=strong)
ǫt ∼ N.I.D.(0, σ 2 )
Some examples:
∆Xt = Xt − Xt−1 = Xt − LXt = (1 − L)Xt
yt = (θ1 + θ2 L)LXt = (θ1 L + θ2 L2 )Xt = θ1 Xt−1 + θ2 Xt−2
Expression like
(θ0 + θ1 L + θ2 L2 + ... + θn Ln )
with possibly n = ∞, are called lag polynomial and are indicated as θ(L)
Some examples:
∆Xt = Xt − Xt−1 = Xt − LXt = (1 − L)Xt
yt = (θ1 + θ2 L)LXt = (θ1 L + θ2 L2 )Xt = θ1 Xt−1 + θ2 Xt−2
Expression like
(θ0 + θ1 L + θ2 L2 + ... + θn Ln )
with possibly n = ∞, are called lag polynomial and are indicated as θ(L)
Some examples:
∆Xt = Xt − Xt−1 = Xt − LXt = (1 − L)Xt
yt = (θ1 + θ2 L)LXt = (θ1 L + θ2 L2 )Xt = θ1 Xt−1 + θ2 Xt−2
Expression like
(θ0 + θ1 L + θ2 L2 + ... + θn Ln )
with possibly n = ∞, are called lag polynomial and are indicated as θ(L)
Some examples:
∆Xt = Xt − Xt−1 = Xt − LXt = (1 − L)Xt
yt = (θ1 + θ2 L)LXt = (θ1 L + θ2 L2 )Xt = θ1 Xt−1 + θ2 Xt−2
Expression like
(θ0 + θ1 L + θ2 L2 + ... + θn Ln )
with possibly n = ∞, are called lag polynomial and are indicated as θ(L)
MA(q)
Yt = θ(L)ǫt = ǫt + θ1 ǫt−1 + θ2 ǫt−2 + ... + +θq ǫt−q
Example, MA(1)
Yt = ǫt + θǫt−1 = (1 + θL)ǫt
being E[Yt ] = 0
γ(0) = E[Yt Yt ] = E[(ǫt + θǫt−1 )(ǫt + θǫt−1 )] = σ 2 (1 + θ2 );
γ(1) = E[Yt Yt−1 ] = E[(ǫt + θǫt−1 )(ǫt−1 + θǫt−2 )] = σ 2 θ;
γ(k) = E[Yt Yt−k ] = E[(ǫt + θǫt−1 )(ǫt−k + θǫt−k−1 )] = 0 ∀k > 1
and,
γ(1) θ
ρ(1) = =
γ(0) 1 + θ2
γ(k)
ρ(k) = =0 ∀k > 1
γ(0)
MA(q)
Yt = θ(L)ǫt = ǫt + θ1 ǫt−1 + θ2 ǫt−2 + ... + +θq ǫt−q
Example, MA(1)
Yt = ǫt + θǫt−1 = (1 + θL)ǫt
being E[Yt ] = 0
γ(0) = E[Yt Yt ] = E[(ǫt + θǫt−1 )(ǫt + θǫt−1 )] = σ 2 (1 + θ2 );
γ(1) = E[Yt Yt−1 ] = E[(ǫt + θǫt−1 )(ǫt−1 + θǫt−2 )] = σ 2 θ;
γ(k) = E[Yt Yt−k ] = E[(ǫt + θǫt−1 )(ǫt−k + θǫt−k−1 )] = 0 ∀k > 1
and,
γ(1) θ
ρ(1) = =
γ(0) 1 + θ2
γ(k)
ρ(k) = =0 ∀k > 1
γ(0)
Hence, an MA(q) is q-correlated and it can also be shown that any stationary q-correlated
process can be represented as an MA(q).
Wold Theorem: any mean zero covariance stationary process can be represented in the
form, MA(∞) + deterministic component (the two being uncorrelated).
But, given a q-correlated process, is the MA(q) process unique? In general no, indeed it can
be shown that for a q-correlated process there are 2q possible MA(q) with same
autocovariance structure. However, there is only one MA(q) which is invertible.
Hence, an MA(q) is q-correlated and it can also be shown that any stationary q-correlated
process can be represented as an MA(q).
Wold Theorem: any mean zero covariance stationary process can be represented in the
form, MA(∞) + deterministic component (the two being uncorrelated).
But, given a q-correlated process, is the MA(q) process unique? In general no, indeed it can
be shown that for a q-correlated process there are 2q possible MA(q) with same
autocovariance structure. However, there is only one MA(q) which is invertible.
Hence, an MA(q) is q-correlated and it can also be shown that any stationary q-correlated
process can be represented as an MA(q).
Wold Theorem: any mean zero covariance stationary process can be represented in the
form, MA(∞) + deterministic component (the two being uncorrelated).
But, given a q-correlated process, is the MA(q) process unique? In general no, indeed it can
be shown that for a q-correlated process there are 2q possible MA(q) with same
autocovariance structure. However, there is only one MA(q) which is invertible.
Invertibility also has important practical consequence in application. In fact, given that the ǫt
are not observable they have to be reconstructed from the observed Y’s through the
AR(∞) representation.
Fulvio Corsi Introduction to ARMA
() and GARCH processes SNS Pisa 3 March 2010 7 / 24
Invertibility conditions for MA
first consider the MA(1) case:
Yt = (1 + θL)ǫt
given the result
∞
X
(1 + θL)−1 = (1 − θL + θ2 L2 − θ3 L3 + θ4 L4 + ...) = (−θL)i
i=0
inverting the θ(L) lag polynomial, we can write
(1 − θL + θ2 L2 − θ3 L3 + θ4 L4 + ...)Yt = ǫt
which can be considered an AR(∞) process.
Invertibility also has important practical consequence in application. In fact, given that the ǫt
are not observable they have to be reconstructed from the observed Y’s through the
AR(∞) representation.
Fulvio Corsi Introduction to ARMA
() and GARCH processes SNS Pisa 3 March 2010 7 / 24
Auto-Regressive Process (AR)
A general AR process is defined as
φ(L)Yt = ǫt
It is always invertible but not always stationary.
Example: AR(1)
(1 − φL)Yt = ǫt or Yt = φYt−1 + ǫt
by inverting the lag polynomial (1 − φL) the AR(1) can be written as
∞
X ∞
X
Yt = (1 − φL)−1 ǫt = (φL)i ǫt = φi ǫt−i = MA(∞)
i=0 i=0
hence the stationarity condition is that |φ| < 1.
From this representation we can apply the general formula of MA to compute γ(·) and ρ(·).
In particular,
ρ(k) = φ|k| ∀k
i.e. monotonic exponential decay for φ > 0 and exponentially damped oscillatory decay for
φ < 0.
In general an AR(p) process
Yt = φ1 Yt−1 + φ2 Yt−2 + ... + φp Yt−p + ǫt
is stationarity if all the roots of the characteristic equation of the lag polynomial
1 − φ1 z − φ2 z2 − ... − φp zp = 0
are outside the unit circle.
Fulvio Corsi Introduction to ARMA
() and GARCH processes SNS Pisa 3 March 2010 8 / 24
Auto-Regressive Process (AR)
A general AR process is defined as
φ(L)Yt = ǫt
It is always invertible but not always stationary.
Example: AR(1)
(1 − φL)Yt = ǫt or Yt = φYt−1 + ǫt
by inverting the lag polynomial (1 − φL) the AR(1) can be written as
∞
X ∞
X
Yt = (1 − φL)−1 ǫt = (φL)i ǫt = φi ǫt−i = MA(∞)
i=0 i=0
hence the stationarity condition is that |φ| < 1.
From this representation we can apply the general formula of MA to compute γ(·) and ρ(·).
In particular,
ρ(k) = φ|k| ∀k
i.e. monotonic exponential decay for φ > 0 and exponentially damped oscillatory decay for
φ < 0.
In general an AR(p) process
Yt = φ1 Yt−1 + φ2 Yt−2 + ... + φp Yt−p + ǫt
is stationarity if all the roots of the characteristic equation of the lag polynomial
1 − φ1 z − φ2 z2 − ... − φp zp = 0
are outside the unit circle.
Fulvio Corsi Introduction to ARMA
() and GARCH processes SNS Pisa 3 March 2010 8 / 24
State Space Representation of AR(p)
to gain more intuition on the AR stationarity conditions write an AR(p) in its state space form
Yt φ1 φ2 φ3 . . . φp−1 φp Yt−1 ǫt
Yt−1 1 0 0 ... 0 0 Yt−2 0
. = . . . . . . + .
.. .. .. .. ... .. .. .. ..
Yt−p+1 0 0 0 ... 1 0 Yt−p 0
Xt = F Xt−1 + vt
Hence, the expected value of Xt satisfy,
E[Xt ] = F Xt−1 and E[Xt+j ] = F j+1 Xt−1
is a linear map in Rp whose dynamic properties are given by the eigenvalues of the matrix F.
Xt = F Xt−1 + vt
Hence, the expected value of Xt satisfy,
E[Xt ] = F Xt−1 and E[Xt+j ] = F j+1 Xt−1
is a linear map in Rp whose dynamic properties are given by the eigenvalues of the matrix F.
For AR(p)
For AR(p)
In the case of an AR(p) process the prediction formula can also be written as
(s) (s) (s)
E[Yt+s |Yt , Yt−1 , ...] = µ + f11 (Yt − µ) + f12 (Yt−1 − µ) + ... + f1p (Yt−p+1 − µ)
(j)
where f11 is the element (1, 1) of the matrix F j .
The easiest way to compute prediction from AR(p) model is, however, through recursive
methods.
In the case of an AR(p) process the prediction formula can also be written as
(s) (s) (s)
E[Yt+s |Yt , Yt−1 , ...] = µ + f11 (Yt − µ) + f12 (Yt−1 − µ) + ... + f1p (Yt−p+1 − µ)
(j)
where f11 is the element (1, 1) of the matrix F j .
The easiest way to compute prediction from AR(p) model is, however, through recursive
methods.
Identification:
check the autocorrelation (ACF) function: a q-correlated process is an MA(q) model
check the partial autocorrelation (PACF) function:
for an AR(p) process, while the k–lag ACF can be interpreted as simple regression
Yt = ρ(k)Yt−k + error, the k–lag PACF can be seen as a multiple regression
Yt = b1 Yt−1 + b2 Yt−2 + ... + bk Yt−k + error
it can be computed by solving the Yule-Walker system:
−1
b1 γ(0) γ(1) ... γ(k − 1) γ(1)
b2 γ(1) γ(0) ... γ(k − 2) γ(2)
. = . . . .
. . . . .
. . . ... . .
bk γ(k − 1) γ(k − 2) ... γ(0) γ(k)
Identification:
check the autocorrelation (ACF) function: a q-correlated process is an MA(q) model
check the partial autocorrelation (PACF) function:
for an AR(p) process, while the k–lag ACF can be interpreted as simple regression
Yt = ρ(k)Yt−k + error, the k–lag PACF can be seen as a multiple regression
Yt = b1 Yt−1 + b2 Yt−2 + ... + bk Yt−k + error
it can be computed by solving the Yule-Walker system:
−1
b1 γ(0) γ(1) ... γ(k − 1) γ(1)
b2 γ(1) γ(0) ... γ(k − 2) γ(2)
. = . . . .
. . . . .
. . . ... . .
bk γ(k − 1) γ(k − 2) ... γ(0) γ(k)
Identification:
check the autocorrelation (ACF) function: a q-correlated process is an MA(q) model
check the partial autocorrelation (PACF) function:
for an AR(p) process, while the k–lag ACF can be interpreted as simple regression
Yt = ρ(k)Yt−k + error, the k–lag PACF can be seen as a multiple regression
Yt = b1 Yt−1 + b2 Yt−2 + ... + bk Yt−k + error
it can be computed by solving the Yule-Walker system:
−1
b1 γ(0) γ(1) ... γ(k − 1) γ(1)
b2 γ(1) γ(0) ... γ(k − 2) γ(2)
. = . . . .
. . . . .
. . . ... . .
bk γ(k − 1) γ(k − 2) ... γ(0) γ(k)
If d < 0.5 the process is cov stationary and admits an AR(∞) representation.
ρ(k) ≈ ck2d−1
If d < 0.5 the process is cov stationary and admits an AR(∞) representation.
ρ(k) ≈ ck2d−1
hence,
hence,
How to parameterize E ǫ2t |Ωt−1 = σ((Ωt−1 ; θ) ≡ σt2 ?
ARCH(q) postulated that the conditional variance is a linear function of the past q squared
innovations
Xq
σt2 = ω + αi ǫ2t−i = ω + α(L)ǫ2t−1
i=1
How to parameterize E ǫ2t |Ωt−1 = σ((Ωt−1 ; θ) ≡ σt2 ?
ARCH(q) postulated that the conditional variance is a linear function of the past q squared
innovations
Xq
σt2 = ω + αi ǫ2t−i = ω + α(L)ǫ2t−1
i=1
Yt = φYt−1 + ǫt
σt2 = ω + αǫ2t−1
ǫt ∼ N(0, σt2 )
Hence,
Kurtosis(ǫt ) = E(ǫ4t )/E(ǫ2t )2 > 3
Yt = φYt−1 + ǫt
σt2 = ω + αǫ2t−1
ǫt ∼ N(0, σt2 )
Hence,
Kurtosis(ǫt ) = E(ǫ4t )/E(ǫ2t )2 > 3
Problem: empirical volatility very persistent ⇒ Large q i.e. too many α’s
q
X p
X
σt2 = ω + αi ǫ2t−i + 2
βj σt−j = ω + α(L)ǫ2t−1 + β(L)σt−1
2
i=1 j=1
Problem: empirical volatility very persistent ⇒ Large q i.e. too many α’s
q
X p
X
σt2 = ω + αi ǫ2t−i + 2
βj σt−j = ω + α(L)ǫ2t−1 + β(L)σt−1
2
i=1 j=1
showing that the forecasts of the conditional variance revert to the long-run unconditional
variance at an exponential rate dictated by α + β
showing that the forecasts of the conditional variance revert to the long-run unconditional
variance at an exponential rate dictated by α + β
conditional variance
6
σt2 =ω+ 2
αrt−1 + 2
βσt−1
4
Empirically negative rt−1 impact more than positive ones → asymmetric news impact curve
GJR or T-GARCH
1 if rt < 0
σt2 = 2
ω + αrt−1 2
+ γrt−1 2
Dt−1 + βσt−1 with Dt =
0 otherwise
- Positive returns (good news): α
- Negative returns (bad news): α + γ
- Empirically γ > 0 → “Leverage effect”
conditional variance
6
σt2 =ω+ 2
αrt−1 + 2
βσt−1
4
Empirically negative rt−1 impact more than positive ones → asymmetric news impact curve
GJR or T-GARCH
1 if rt < 0
σt2 = 2
ω + αrt−1 2
+ γrt−1 2
Dt−1 + βσt−1 with Dt =
0 otherwise
- Positive returns (good news): α
- Negative returns (bad news): α + γ
- Empirically γ > 0 → “Leverage effect”
L(rT , rT−1 , ..., r1 ; θ) = f (rT |ΩT−1 ; θ) × f (rT−1 |ΩT−2 ; θ) × ... × f (r1 |Ω0 ; θ)