文章

Time Series Notes (4) - Model specification

Model-building strategy

Box-Jenkins method for model-building strategy

graph TD
  A([Start with a time series realization])
  --> B[Identify a preliminary time series model]
  --> C[Estimation of the model parameters]
  --> D{Is the model adequate?}
  D --> |No| B
  D --> |Yes| E([Stop])
No
Yes
Start with a time series realization
Identify a preliminary time series model
Estimation of the model parameters
Is the model adequate?
Stop

Three steps of building models

  1. Model specification. The model is tentative at this moment and it is from the observed part of the time series. The model also should be chosen under the principle of parsimony.
  2. Model fitting.
  3. Model diagnostics. If inadequacies found, another model ought be considered until an acceptable model is found.

Autocorrelation function (ACF)

For a MA(q) process, it has the cut-off property, i.e. ρq0 but ρk=0 for k>q.

For a AR(p) or ARMA(p,q) process, it decays to zero exponentially, but does not enjoy the cut-off property.

Sample autocorrelation function

For a sequence of observations Zt,,Zn the sample autocorrelation function (sample ACF) is defined as follows.

rk=γk^γ0^,  k0

where γk^ is the k-th sample autocovariance:

γk^=1nt=k+1n(ZtZ¯)(ZtkZ¯),  Z¯=1nt=1nZt

Bartlett’s approximation

The 95% approximate confidence interval of rk hence will be (rk denotes sampled ρk here)

±1.96n1+2r12++2rk12

The above confidence interval acts as the accepting region for the test with H0:ρk=0.

In real applications, the original condition of the Bartlett’s approximation is usually ignored. The squared part are usually ignored. And the value of 1.96, the corresponding 95% percentile of the standard normality, is replaced by 2.0 in R.

Partial autocorrelation function (PACF)

Definition of partial autocorrelation

The partial autocorrelation (PACF) at lag k, denoted by ϕkk, is defined as the solution to the system of equations:

(1ρ1ρk1ρ11ρk2ρk1ρk21)(ϕk1ϕk2ϕkk)=(ρ1ρ2ρk)

Or, in other words, the equations are equivalent:

(ϕk1ϕk2ϕkk)=(1ρ1ρk1ρ11ρk2ρk1ρk21)1(ρ1ρ2ρk)

where Zt is a stationary time series with autocorrelation functions ρm,m=1,2,3,

In particular:

  1. k=1: ϕ11=ρ1
  2. k=2: we have
(ϕ21ϕ22)=(1ρ1ρ11)1(ρ1ρ2)

Thus,

ϕ22=ρ2ρ11ρ12

The PACF is to measure the direct impact from Ztk at Zt, with removing the middle terms from Zt1,,Ztk+1. So for AR(p) process, the PACF has a cut-off: ϕkk=0 for k>p. This property mirrors the cut-off property of the ACF of an MA(q) process: ρk=0 for k>q.

For MA(1) process Zt=at+θat1, we have

ϕkk=(1)k+1θk1+θ2+θ4++θ2k,  for k1
  1. Notice that the partial autocorrelations of an MA(1) model are never zero. Except the special case of θ2=1. They decay exponentially to zero, rather like the autocorrelation for an AR(1) series.
  2. The behaviors of PACF for ARMA processes is quite similar to that for MA processes.

Sample PACF

For an observed time series, an obvious method is to estimate the ρ’s by r’s and then to solve the Yule-Walker equations to get estimates of ϕkk, i.e.,

(ϕ^k1ϕ^k2ϕ^kk)=(1r1rk1r11rk2rk1rk21)1(r1r2rk)

Inverse autocorrelation function (IACF) and the sample IACF

  1. The sample IACF plays much the same role in ARIMA modeling as the sample PACF, but it generally indicates subset and seasonal autoregressive models better than the sample PACF.
  2. Additionally, the sample IACF can be useful for detecting over-differencing.
  3. The IACF is defined in the frequency domain. It is not often discussed in textbooks.

Extended autocorrelation function (EACF) and the sample EACF

  1. It can be used to identify the values of p+d and q for an ARIMA(p,d,q).
  2. It is not a very popular method.
  3. There are other similar methods in the literature such as the corner method, the smallest canonical correlation (SCAN) method, etc..

Nonstationarity

How to determine the value of d in ARIMA(p,d,q) models

  1. To determine the order of d is equivalent to determine whether or not a sequence of values are from a stationary time series.
  2. The definition of the sample ACF implicitly assumes stationarity.
  3. However, for a nonstationary series, the sample ACF typically decays slowly as the lags increase.
  4. If the sample ACF decays slowly, we may say the sequence is nonstationary. Otherwise, it is stationary.

Overdifferencing

If a time series Zt is stationary, then the differenced sequence Zt is still stationary. Thus, we sometimes tend to overdifference the sequence.

To avoid overdifferencing, we should look carefully at each difference in succession and always keep in mind the principle of parsimony – models should be as simple as possible.

Other specification methods

Akaike information criterion (AIC)

This criterion is to select the model that minimizes

AIC(k)=2log(maximum likelihood)+2k

where k is the number of parameters. For an ARMA(p,q) model, k=p+q+1 when the model contains an intercept, and p+q otherwise.

Bayesian information criterion (BIC)

This criterion is to select the model that minimizes

BIC(k)=2log(maximum likelihood)+klog(n)

The AIC is more popular in real applications although the BIC has a well justified theoretical basis.

本文由作者按照 CC BY 4.0 进行授权

热门标签