This post mainly uses material from Analysis of Financial Time Series.
Stationarity
Stationarity is the foundation of time series analysis.
Strictly stationary
If the joint distribution of (rt1,...,rtk) is identical to that of (rt1t,...,rtkt), where k is an arbitrary positive integer and (t1,...,tk) is a collection of k positive integers.
This is a really strong condition.
Weak stationary
{rt} is weak stationary if
E[rt]=μ,Cov(rt,rt−l)=γl
In application, weak stationary enables one to make inference concerning future obeservations.
We can assume that the first two moment of a weak stationary series is finite. And if {rt} is normally distributed, weak stationary is equivalent to strict stationary.
Also, γl is called the lag-l autocovariance of rt. And we have
This empirical estimation is called sample autocorrelation function. It plays an important role in linear series analysis.
White Noise
A time series rt is called a white noise if {rt} is a sequence of independent and identically distributed random variables with finite mean and variance. In particular, if rt is normally distributed with mean 0 and variance σ2, the series is called a Gaussian white noise.
For white noise, all ACFs are zero.
Linear Time Series
A time series rt is said to be linear if
rt=μ+i=0∑∞ψiat−i
Where {at} is a white noise sequence. And we define ψ0=1
For weak stationary time series, the variance exist, and therefore we will have ρl→0,(l→∞).
AR
Consider a monthly return rt, the lag-1 will be large and therefore rt−1 might be useful in predicting rt. The simple model for this is:
rt=ϕ0+ϕ1rt−1+at
The model is in the same form as linear regression, therefore we call is autoregressive model of order 1, or simply AR(1). There are also lots of similarity and difference between AR and linear regression, which we will talk about later. For now, we have
E(rt∣rt−1)=ϕ0+ϕ1rt−1,Var(rt∣rt−1)=σa2
This is Markov property. And there will be AR(p) which is
These models are in the form of a multi-dimension linear regression and can be estimated by the least-squares method. The estimated ϕ^1,1 is called the lag-1 sample PACF of rt and ϕ^2,2 is the lag-2 sample PACF and so on.
From the definition, the lag-p sample PACF shows the added contribution of rt−p to an AR(p-1) model. Therefore, for an AR(p) model, the lag-p sample PACF should not be zero and the latter ones should be close to zero.
Information Criteria
There are several information based criteria available to determine the p. All of them are likelihood based, like Akaike information criterion(AIC).
Goodness of Fit
R2=1−totalsumofsquaresresidualsumofsquares
For a stationary AR(p) model, the measure becomes
R2=1−∑t=p+1T(rt−rˉt)2∑t=p+1Ta^t2
For a given data set, it is well known that R2 is a nondecreasing function of the number of parameters used. To overcome this weakness, we could use the adjusted R2:
Radj2=1−σ^r2σ^a2
MA
Think of a special case of AR
rt+θ1rt−1+θ12rt−2+⋯=ϕ0+at
And since
rt−1+θ1rt−2+θ12rt−3+⋯=ϕ0+at−1
We have
rt=ϕ0(1−θ1)+at−θ1at−1
Therefore, MA(1) is:
rt=c0+at−θ1at−1
And MA(q) is
rt=c0+at−θ1at−1−...−θqat−q
Properties
Moving-average models are always weakly stationary because they are finite linear combination of a white noise sequence.
μ=c0,Var(rt)=(1+θ12+θ22+⋯+θ12)σa2
ACF
For simplicity assume c0=0
rt−lrt=rt−lat−θ1rt−lat−1
Therefore
γl={−θ1σa2,l=10,l>1
And
ρl=⎩⎨⎧1,l=0−1+θ22θ1,l=10,l>1
And for MA(q), we could have only the lag-q is not 0 but above are 0. MA(q) is only linearly related to its first 1-lagged values and hence is a "finite memory" model.
How to identify MA model
We could just use the property of ACF to identity q for MA.
ARMA
An ARMA model combines the idea of AR and MA into a compact form so that the number of parameters used is kept small. For the return series in finance the chance of using ARMA is low. However, it is highly relevant in volatility modeling. The simple ARMA(1, 1) is
rt−ϕ1rt−1=ϕ0+at−θ1at−1
to make the function meaningful, we need ϕ1=θ1
Thus, ACF of ARMA(1, 1) behaves very much like that of AR(1).
How to identify MA model
The ACF and PACF are not informative in determining the order of an ARMA model. There is something called extended autocorrelation function (EACF) to specify the order of an ARMA process.
Seasonal model
For seasonal data, there is often strong serial correlation. It is common to do seasonal differencing to it, which is
Δkxt=(1−Bk)xt=xt−xt−k
And some time, we need to do multiple differencing, which leads to
Δk(Δlxt)=(1−Bk)(1−Bl)xt
ARIMA
if
Zt=(1−B)dXt∼ARMA(p,q)
then Xt is ARIMA(p, d, q).
References
Tsay, Ruey S. Analysis of financial time series. Vol. 543. John Wiley & Sons, 2005.