Local trend model
For a univariate time series yt=μt+ϵt and μt+1=μt+ηt, both the error terms are assumed to be normally distributed to distinct variance σ2e and σ2η respectively. Notice the first equation is the observed version of the second trend model with added noise. This model can be used to analyze realized volatility of an asset price if μt is assumed to be the log volatility (which is not directly observable) and yt is the logarithm or realized volatility (which is observable constructed from high-frequency transaction data with microstructure noise).
If there is no measurement error term in the first equation (σe=0) this becomes a ARIMA(0,1,0) model. With the error term it is a ARIMA(0,1,1) model, which is also the simple exponential smoothing model. The form is (1−B)yt=(1−θB)at, θ and σ2a are related to σ2e and σ2η as follows: (1+θ2)σ2a=2σ2e+σ2η and θσ2a=σ2e. The quadratic equation for θ will give two solutions with |θ|<1 chosen. The reverse is also possible for positive θ. Both representations have pros and cons and the objective of data analysis, substantive issues and experience decide which to use.
If there is no measurement error term in the first equation (σe=0) this becomes a ARIMA(0,1,0) model. With the error term it is a ARIMA(0,1,1) model, which is also the simple exponential smoothing model. The form is (1−B)yt=(1−θB)at, θ and σ2a are related to σ2e and σ2η as follows: (1+θ2)σ2a=2σ2e+σ2η and θσ2a=σ2e. The quadratic equation for θ will give two solutions with |θ|<1 chosen. The reverse is also possible for positive θ. Both representations have pros and cons and the objective of data analysis, substantive issues and experience decide which to use.
Statistical Inference
Three types (using reading handwritten note example)
- Filtering - recover state variable μt given Ft to remove the measurement errors from the data. (figuring out the word you are reading based on knowledge accumulated from the beginning to the note).
- Prediction - forecast μt+h or yt+h for h>0 given Ft, where t is the forecast origin. (guess the next word).
- Smoothing - estimate μt given FT, where T>t. (deciphering a particular word once you have read through the note).
The Kalman Filter
Let μt|j=E(μt|Fj) and Σt|j=Var(μt|Fj) be, respectively, the conditional mean and variance of μt given information Fj. Similarly yt|j denotes the conditional mean of yt given Fj. Furthermore let vt=yt−yt|j and Vt=Var(vt|Ft−1) be 1-step ahead forecast error and its variance of yt given Ft−1. Note that Var(vt|Ft−1)=Var(vt), since the forecast error vt is independent of Ft−1. Further, yt|t−1=μt|t−1 giving vt=yt−μt|t−1 and Vt=Σt|t−1+σ2e. Also, E(vt)=0 and Cov(vt,yt)=0 for j<t. The information Ft≡{Ft−1,yt}≡{Ft−1,vt}, hence μt|t=E(μt|Ft−1,vt) and Σt|t=Var(μt|Ft−1,vt).
One can show that Cov(μt,vt|Ft−1)=Σt|t−1 giving, [μtvt]Ft−1∼N([μt|t−10],[Σt|t−1Σt|t−1Σt|t−1Vt]). Applying the multivariate normal theorem we get μt|t=μt|t−1+(V−1tΣt|t−1)vt=μt|t−1+Ktvt, Σt|t=Σt|t−1−Σt|t−1V−1tΣt|t−1=Σt|t−1(1−Kt), where Kt=V−1tΣt|t−1 is referred to as the Kalman gain, which is the regression coefficient of μt on vt, governing the contribution of th enew shock vt to the state variable μt. To predict μt+1 given Ft we have μt+1|t∼N(μt|t,Σt|t+σ2η). once the new data yt+1 is observed, the above procedure can be repeated (obviously once σe and ση are estimated, generally using maximum likelihood method). This is the famous Kalman filter algorithm (1960). The choice of priors μ1|0 and Σ1|0 requires some attention.
Properties of forecast error -
State error recursion -
State smoothing -
Missing Values -
Effect of Initialization -
Estimation -
One can show that Cov(μt,vt|Ft−1)=Σt|t−1 giving, [μtvt]Ft−1∼N([μt|t−10],[Σt|t−1Σt|t−1Σt|t−1Vt]). Applying the multivariate normal theorem we get μt|t=μt|t−1+(V−1tΣt|t−1)vt=μt|t−1+Ktvt, Σt|t=Σt|t−1−Σt|t−1V−1tΣt|t−1=Σt|t−1(1−Kt), where Kt=V−1tΣt|t−1 is referred to as the Kalman gain, which is the regression coefficient of μt on vt, governing the contribution of th enew shock vt to the state variable μt. To predict μt+1 given Ft we have μt+1|t∼N(μt|t,Σt|t+σ2η). once the new data yt+1 is observed, the above procedure can be repeated (obviously once σe and ση are estimated, generally using maximum likelihood method). This is the famous Kalman filter algorithm (1960). The choice of priors μ1|0 and Σ1|0 requires some attention.
Properties of forecast error -
State error recursion -
State smoothing -
Missing Values -
Effect of Initialization -
Estimation -