Processing math: 100%

Tuesday, June 30, 2015

Tsay Ch2 - Linear Time Series Analysis and Its Applications


Ch2: Linear Time Series Analysis and its Applications

We work with log returns rt as a collection of random variables over time, as a time series rt.

Stationarity: The first step in time series analysis should always be checking stationarity.
1) Strictly stationary time series rt if the joint distribution of (rt1,...,rtk) is identical to that of (rt1+t,...,ttk+t) for all t and k, i.e. the joint distribution is invariant under time shift.
2) Weak stationarity: if both mean and covariance between rt and rtl are time-invariant. Weak stationarity implies partial predictability. In other words, the first two moments of rt are finite and constant. A strictly stationary series is weakly stationary but not vice-versa unless rt is normally distributed.

Autocovariance: γl=Cov(rt,rtl) with properties γ0=Var(rt) and γl=γl

Autocorrelation function (ACF): uncorrelated means ρx,y=0. It implies independence under normality. ρl=γl/γ0 giving ρ0=1 and ρl=ρl. Sample autocorrelation is given by
ˆρl=Tt=l+1(rtˉr)(rtlˉr)Tt=1(rtˉr)2
is asymptotically normal with mean zero and variance 1/T, if rt is an iid sequence with finite second moment (Brockwell and Davis 1991).

Bartlett's formula: For a weakly stationary time series of the form rt=μ+q1ψiati, where ψ0=1, q is a non-negative integer, and aj are Gaussian white noise series, then ˆρl is asymptotically normal with mean μ and variance (1+2q1ρ2i)/T for lq (Box, Jenkins, Reinsel 1994). Hence, to test the autocorrelation for a given integer l, we can construct the statistic.

Ljung and Box (1978) test increases the power for finite samples by constructing a chi-squared distribution statistic with m degrees of freedom:
Q(m)=T(T+2)ml=1ˆρ2lTl.
In practice, the selection of m may affect the performance of the Q(m) statistic. mln(T) provides better power performance.

Monthly returns of the value-weighted index seems to have stronger serial dependence than individual stock returns. CAPM suggests there is no autocorrelation in financial series. One has to be careful of the autocorrelation induced by the way index returns are determined or stock prices are determined - these are pseudo relationships.

White noise: time series rt is white noise if the sequence is iid with finite mean and variance. if the mean is 0 and variance σ2, it is called Gaussian white noise. For a white noise series all the ACFs are zero. e.g. monthly individual stocks returns seem to be white noise but not value-weighted index return series.

Wold decomposition: A time series rt is said to be linear if it can be written as Wold decomposition
rt=μ+i=0ψiati,

where μ is the mean of rt, ϕ0=1, and at is white noise series. If rt is weakly stationary, we can obtain the mean, variance and covariance easily. For weakly stationary series ρl converges to zero as l increases.

AR(p) model: Current conditional expectation of returns depend on the last p steps
rt=ϕo+ϕ1rt1+...+ϕprrp+at

The necessary and sufficient condition for weakly stationarity is |ϕ1|<1 for AR(1) process.  The plot of ACF decays exponentially (extends beyond lag 1!). For higher order AR processes the inverse of the solutions to characteristic equation are called characteristic roots. Complex roots correspond to damped oscillations related to business cycles. The stationarity condition is that the characteristic roots are less than one in modulus (or all solutions to characteristic equation greater than one in modulus).

Identification of AR model: order determination of AR models. Two approaches:
1) Partial Autocorrelation Function (PACF): nested AR models with increasing order. For a true order of p, the terms ˆϕl,l converge to zero for all l>p with variance 1/T. These can be plotted to determine the cutoff as a PACF plot.
2) Information criteria:  likelihood based criteria. e.g. the Akaike information criterion (AIC) 1973 is:
AIC=2Tln(likelihood)+2TP

For a Gaussian AR(l) model, AIC reduces to ln(ˆσ2l)+2l/T, where ^σ2l is the maximum likelihood estimate of σ2a, T is the sample size and P is the number of parameters. Another common criteria is Bayesian information criterion (BIC), which for Gaussian AR(l) reduces to ln(ˆσ2l)+ln(T)l/T. BIC tends to select a lower AR model when the sample size is moderate to large.

Parameter estimation: OLS method, or likelihood method. Likelihood estimation is very similar to OLS but gives slightly lower σ2a for finite samples.

Model checking: residual series should be white noise - check ACF of residual and Ljung-Box statistics with mg degrees of chi-squared distribution, where m is approximately ln(T) and g is number of AR coefficients used in the model.

Goodness of fit: For a stationary series R2 can be used as a goodness of fit. For unit-root non-stationary series R2 converges to one regardless of the underlying model. Adjusted R2 should be used to penalize the increased number of parameters.

Forecasting: Longer terms forecasts are more accurate and approach unconditional mean. The speed of mean reversion is measured by half life denoted by k=ln(0.5/|ϕ1|).

MA(q) model: Moving average models. These are simple extension of white noise series, an infinite-order AR model with some parameter constraints. Bid-ask bounce in stock trading may introduce an MA(1) structure in a return series. The general form is
rt=c0+atθ1at1...θqatq

It is a finite memory model with ACF zero beyond the first q. MA(q) is invertible if |θ|<1, otherwise noninvertible. An MA model is always stationary. The coefficient of MA models are also called impulse response function.

Identification of AR model: Identified using ACF, provides information on the nonzero MA lags of the model.

Estimation: Likelihood methods:
1) conditional likelihood method: assumes initial shock to be zero.
2) exact likelihood method: initial shocks are estimated jointly with other parameters.

Forecasting: q step ahead forecast for MA(q) model is the mean and stays there.

ARMA(p,q) model: to reduce higher order parameters in AR or MA model, one can use ARMA model. Generally not used for returns but volatility models. ARMA(1,1) is
rtϕ1rt1=ϕ0+atθ1at1

For the model to be meaningful we need ϕ1θ1. The ACF of an ARMA(1,1) behaves very much like that of an AR(1) model except that the exponential decay starts with lag 2 instead of lag 1. Consequently, the ACF of an ARMA(1,1,) does not cut off at any finite lag. The PACF of an ARMA(1,1) does not cut off at any finite lag either, like of MA(1), except that the exponential decay starts with lag 2. The stationarity condition for an ARMA(1,1) is same as that of an AR(1) model. A general ARMA(p,q) model is
rt=ϕ0+pi=1ϕirti+atqj=1θjatj

There should be no common factors between the AR and MA polynomial, otherwise the order of the model can be reduced. For stationarity all solutions to AR polynomial should be less than 1 in absolute value.

Identification: EACF (Tsay, Tiao 1984) can be used. The first zero corner in the EACF table identifies the order. Estimation can then be done using conditional or exact likelihood. Ljung-Box statistics of the residuals can also be checked for the adequacy of the model.

Unit root nonstationarity: Price series, interest rate, FX rates are generally non stationary. They are called unit root non-stationary  series in time series literature. The best example is random-walk.

Random walk: A time series pt is a random walk if it satisfies
pt=pt1+at

Looking it as and AR(1) process, it is a non-stationary as the coefficient is equal to 1. We call this a unit root nonstationary time series. This is a non-mean reverting model with prediction of price equal to the value at forecast origin, with variance of estimate lσ2a, where l is the number of look ahead steps, which diverge to infinity as l. The series has a strong memory, as the sample ACFs are all approaching 1 as the sample size increases.

Random walk with drift: The market index tends to have a small and positive mean. The model for log price becomes
pt=μ+pt1+at

Both the expected value tμ and variance tσ2a increase with time t.

Trend-Stationary time series: pt=β0+β1t+tt
The mean is β0+β1t and variance is finite and equal to Var(rt), unlike random walk non-stationary model with trend. This can be transformed into a stationary one by removing the time trend via a simple linear regression analysis.

ARIMA models: A conventional way to handle unit-root nonstationarity is to use differencing. A time series yt is said to be ARIMA(p,1,q) process if the change series cT=ytyt1 is stationary and invertible ARMA(p,q) process. For example, log price series is an ARIMA process. Series with multiple roots may need multiple differencing.

Dickey-Fuller unit root test: H0:ϕ1=1 versus Ha:ϕ1<1. The test statistic is the t-ratio of the least squares estimate of ϕ1 from the model pt=ϕ1pt1+et giving
ˆϕ1=T1pt1ptT1p2t1

ˆσ2e=T1(ptˆϕ1pt1)2T1

where p0=0 and T is the sample size. The t-ratio is
ˆϕ11std(ˆϕ1)=T1pt1etˆσeT1p2t1


Left out sections: 2.8 Seasonal models, 2.9 Regression models with time series errors, 2.10 consistent covariance matrix estimation, 2.11 long-memory models

No comments:

Post a Comment