Autoregressive fractionally integrated moving average

From HandWiki

In statistics, autoregressive fractionally integrated moving average models are time series models that generalize ARIMA (autoregressive integrated moving average) models by allowing non-integer values of the differencing parameter. These models are useful in modeling time series with long memory—that is, in which deviations from the long-run mean decay more slowly than an exponential decay. The acronyms "ARFIMA" or "FARIMA" are often used, although it is also conventional to simply extend the "ARIMA(p, d, q)" notation for models, by simply allowing the order of differencing, d, to take fractional values. Fractional differencing and the ARFIMA model were introduced in the early 1980s by Clive Granger, Roselyne Joyeux, and Jonathan Hosking.[1][2][3]

Basics

In an ARIMA model, the integrated part of the model includes the differencing operator (1 − B) (where B is the backshift operator) raised to an integer power. For example,

[math]\displaystyle{ (1-B)^2=1-2B+B^2 \,, }[/math]

where

[math]\displaystyle{ B^2X_t=X_{t-2} \, , }[/math]

so that

[math]\displaystyle{ (1-B)^2X_t = X_t -2X_{t-1} + X_{t-2}. }[/math]

In a fractional model, the power is allowed to be fractional, with the meaning of the term identified using the following formal binomial series expansion

[math]\displaystyle{ \begin{align} (1 - B)^d &= \sum_{k=0}^{\infty} \; {d \choose k} \; (-B)^k \\ & = \sum_{k=0}^{\infty} \; \frac{\prod_{a=0}^{k-1} (d - a)\ (-B)^k}{k!}\\ &=1-dB+\frac{d(d-1)}{2!}B^2 -\cdots \, . \end{align} }[/math]

ARFIMA(0, d, 0)

The simplest autoregressive fractionally integrated model, ARFIMA(0, d, 0), is, in standard notation,

[math]\displaystyle{ (1 - B)^d X_t= \varepsilon_t, }[/math]

where this has the interpretation

[math]\displaystyle{ X_t-dX_{t-1}+\frac{d(d-1)}{2!}X_{t-2} -\cdots = \varepsilon_t . }[/math]

ARFIMA(0, d, 0) is similar to fractional Gaussian noise (fGn): with d = H−​12, their covariances have the same power-law decay. The advantage of fGn over ARFIMA(0,d,0) is that many asymptotic relations hold for finite samples.[4] The advantage of ARFIMA(0,d,0) over fGn is that it has an especially simple spectral density[math]\displaystyle{ f(\lambda) = \frac{1}{2\pi} \left(2 \sin\left(\frac{\lambda}{2}\right)\right)^{-2d} }[/math]

—and it is a particular case of ARFIMA(p, d, q), which is a versatile family of models.[4]

General form: ARFIMA(p, d, q)

An ARFIMA model shares the same form of representation as the ARIMA(p, d, q) process, specifically:

[math]\displaystyle{ \left( 1 - \sum_{i=1}^p \phi_i B^i \right) \left( 1-B \right)^d X_t = \left( 1 + \sum_{i=1}^q \theta_i B^i \right) \varepsilon_t \, . }[/math]

In contrast to the ordinary ARIMA process, the "difference parameter", d, is allowed to take non-integer values.

Enhancement to ordinary ARMA models

The enhancement to ordinary ARMA models is as follows:

  1. Take the original data series and high-pass filter it with fractional differencing enough to make the result stationary, and remember the order d of this fractional difference, d usually between 0 and 1 ... possibly up to 2+ in more extreme cases. Fractional difference of 2 is the 2nd derivative or 2nd difference.
    • note: applying fractional differencing changes the units of the problem. If we started with Prices then take fractional differences, we no longer are in Price units.
    • determining the order of differencing to make a time series stationary may be an iterative, exploratory process.
  2. Compute plain ARMA terms via the usual methods to fit to this stationary temporary data set which is in ersatz units.
  3. Forecast either to existing data (static forecast) or "ahead" (dynamic forecast, forward in time) with these ARMA terms.
  4. Apply the reverse filter operation (fractional integration to the same level d as in step 1) to the forecasted series, to return the forecast to the original problem units (e.g. turn the ersatz units back into Price).
    • Fractional differencing and fractional integration are the same operation with opposite values of d: e.g. the fractional difference of a time series to d = 0.5 can be inverted (integrated) by applying the same fractional differencing operation (again) but with fraction d = -0.5. See GRETL fracdiff function.

The point of the pre-filtering is to reduce low frequencies in the data set which can cause non-stationarities in the data, which non-stationarities ARMA models cannot handle well (or at all)... but only enough so that the reductions can be recovered after the model is built.

Fractional differencing and the inverse operation fractional integration (both directions being used in the ARFIMA modeling and forecasting process) can be thought of as digital filtering and "unfiltering" operations. As such, it is useful to study the frequency response of such filters to know which frequencies are kept and which are attenuated or discarded.[5]

Note that any filtering that would substitute for fractional differencing and integration in this AR(FI)MA model should be similarly invertible as differencing and integration (summing) to avoid information loss. E.g. a high pass filter which completely discards many low frequencies (unlike the fractional differencing high pass filter which only completely discards frequency 0 [constant behavior in the input signal] and merely attenuates other low frequencies, see above PDF) may not work so well, because after fitting ARMA terms to the filtered series, the reverse operation to return the ARMA forecast to its original units would not be able re-boost those attenuated low frequencies, since the low frequencies were cut to zero.

Such frequency response studies may suggest other similar families of (reversible) filters that might be useful replacements for the "FI" part of the ARFIMA modeling flow, such as the well-known, easy to implement, and minimal distortion high-pass Butterworth filter or similar.[6]

See also

References

  1. Granger, C. W. J.; Joyeux, Roselyne (1980). "An Introduction to Long-Memory Time Series Models and Fractional Differencing" (in en). Journal of Time Series Analysis 1 (1): 15–29. doi:10.1111/j.1467-9892.1980.tb00297.x. ISSN 0143-9782. https://onlinelibrary.wiley.com/doi/10.1111/j.1467-9892.1980.tb00297.x. 
  2. Hosking, J. R. M. (1981). "Fractional Differencing". Biometrika 68 (1): 165–176. doi:10.2307/2335817. ISSN 0006-3444. https://www.jstor.org/stable/2335817. 
  3. Robinson, Peter M., ed (2011). Time series with long memory. Advanced texts in econometrics (Repr ed.). Oxford: Oxford Univ. Press. ISBN 978-0-19-925730-0. 
  4. 4.0 4.1 Taqqu, M. S.; Teverovsky, V.; Willinger, W. (1995). "Estimators for Long-Range Dependence: An Empirical Study". Fractals 3 (4): 785–798. doi:10.1142/S0218348X95000692. 
  5. "fracdiff/freqrespfracdiff.pdf at master · diffent/fracdiff" (in en). https://github.com/diffent/fracdiff/blob/master/freqrespfracdiff.pdf. 
  6. Fenga, Livio (2017). Rojas, Ignacio; Pomares, Héctor; Valenzuela, Olga. eds. "Prediction of Noisy ARIMA Time Series via Butterworth Digital Filter" (in en). Advances in Time Series Analysis and Forecasting. Contributions to Statistics (Cham: Springer International Publishing): 173–196. doi:10.1007/978-3-319-55789-2_13. ISBN 978-3-319-55789-2. https://link.springer.com/chapter/10.1007/978-3-319-55789-2_13.