Variance decomposition of forecast errors

From HandWiki

In econometrics and other applications of multivariate time series analysis, a variance decomposition or forecast error variance decomposition (FEVD) is used to aid in the interpretation of a vector autoregression (VAR) model once it has been fitted.[1] The variance decomposition indicates the amount of information each variable contributes to the other variables in the autoregression. It determines how much of the forecast error variance of each of the variables can be explained by exogenous shocks to the other variables.

Calculating the forecast error variance

For the VAR (p) of form

[math]\displaystyle{ y_t=\nu +A_1y_{t-1}+\dots+A_p y_{t-p}+u_t }[/math] .

This can be changed to a VAR(1) structure by writing it in companion form (see general matrix notation of a VAR(p))

[math]\displaystyle{ Y_t=V+A Y_{t-1}+U_t }[/math] where
[math]\displaystyle{ A=\begin{bmatrix} A_1 & A_2 & \dots & A_{p-1} & A_p \\ \mathbf{I}_k & 0 & \dots & 0 & 0 \\ 0 & \mathbf{I}_k & & 0 & 0 \\ \vdots & & \ddots & \vdots & \vdots \\ 0 & 0 & \dots & \mathbf{I}_k & 0 \\ \end{bmatrix} }[/math] , [math]\displaystyle{ Y=\begin{bmatrix} y_1 \\ \vdots \\ y_p \end{bmatrix} }[/math], [math]\displaystyle{ V=\begin{bmatrix} \nu \\ 0 \\ \vdots \\ 0 \end{bmatrix} }[/math] and [math]\displaystyle{ U_t=\begin{bmatrix} u_t \\ 0 \\ \vdots \\ 0 \end{bmatrix} }[/math]

where [math]\displaystyle{ y_t }[/math], [math]\displaystyle{ \nu }[/math] and [math]\displaystyle{ u }[/math] are [math]\displaystyle{ k }[/math] dimensional column vectors, [math]\displaystyle{ A }[/math] is [math]\displaystyle{ kp }[/math] by [math]\displaystyle{ kp }[/math] dimensional matrix and [math]\displaystyle{ Y }[/math], [math]\displaystyle{ V }[/math] and [math]\displaystyle{ U }[/math] are [math]\displaystyle{ kp }[/math] dimensional column vectors.

The mean squared error of the h-step forecast of variable [math]\displaystyle{ j }[/math] is

[math]\displaystyle{ \mathbf{MSE}[y_{j,t}(h)]=\sum_{i=0}^{h-1}\sum_{l=1}^{k}(e_j'\Theta_ie_l)^2=\bigg(\sum_{i=0}^{h-1}\Theta_i\Theta_i'\bigg)_{jj}=\bigg(\sum_{i=0}^{h-1}\Phi_i\Sigma_u\Phi_i'\bigg)_{jj}, }[/math]

and where

  • [math]\displaystyle{ e_j }[/math] is the jth column of [math]\displaystyle{ I_k }[/math] and the subscript [math]\displaystyle{ jj }[/math] refers to that element of the matrix
  • [math]\displaystyle{ \Theta_i=\Phi_i P , }[/math] where [math]\displaystyle{ P }[/math] is a lower triangular matrix obtained by a Cholesky decomposition of [math]\displaystyle{ \Sigma_u }[/math] such that [math]\displaystyle{ \Sigma_u = PP' }[/math], where [math]\displaystyle{ \Sigma_u }[/math] is the covariance matrix of the errors [math]\displaystyle{ u_t }[/math]
  • [math]\displaystyle{ \Phi_i=J A^{i} J', }[/math] where [math]\displaystyle{ J=\begin{bmatrix} \mathbf{I}_k &0 & \dots & 0\end{bmatrix} , }[/math] so that [math]\displaystyle{ J }[/math] is a [math]\displaystyle{ k }[/math] by [math]\displaystyle{ kp }[/math] dimensional matrix.

The amount of forecast error variance of variable [math]\displaystyle{ j }[/math] accounted for by exogenous shocks to variable [math]\displaystyle{ l }[/math] is given by [math]\displaystyle{ \omega_{jl,h} , }[/math]

[math]\displaystyle{ \omega_{jl,h}=\sum_{i=0}^{h-1}(e_j'\Theta_ie_l)^2/MSE[y_{j,t}(h)] . }[/math]


See also

Notes

  1. Lütkepohl, H. (2007) New Introduction to Multiple Time Series Analysis, Springer. p. 63.