Covariances of ARMA Processes
Covariances of ARMA Processes
Overview
1. Review ARMA models: causality and invertibility
2. AR covariance functions
5. Discussion
φ(z) lie outside the unit circle (i.e., φ(z) 6= 0 for |z| ≤ 1). Since
φ(0) = 1, φ(z) > 0 for |z| ≤ 1. (The unit circle in the complex plane
consists of those x ∈ C for which |z| = 1; the closed unit disc includes
the interior of the unit circle as well as the circle; the open unit disc
consists of |z| < 1.)
If the zeros of φ(z) lie outside the unit circle, then we can invert each
of the factors (1 − B/zj ) that make up φ(B) = pj=1 (1 − B/zj ) one
Q
By definition, φ(z) and θ(z) share no zeros and ψ(z) < ∞ for |z| ≤ 1
(since the ψj are summable). The zeros of φ(z) must be outside the
unit circle else we get a contradiction because the existence of such a
zero implies that φ(z̃) = 0 and θ(z̃) = 0. (For |z| > 1, ψ(z) can grow
arbitrarily large, balancing the zero of φ(z).)
Proposition 3.2 states that the process {Xt }is invertible if and only
if the zeros of the moving average polynomial lie outside the unit circle.
Example 3.6 shows what happens with common zeros in φ(z) and θ(z).
The process is
for which
Hence, the two share a common factor. The initial ARMA(2,2) reduces
to a causal, invertible ARMA(1,1) model.
AR covariance functions
Estimation Given the assumption of stationarity, in most cases we can
easily obtain consistent estimates of the process covariances, such as
PT −h
t=1 (xt+h − x̄)(xt − x̄)
γ̂(h) = . (6)
n
What should such a covariance function resemble? Does the “shape”
of the covariance function or estimated correlation function
γ̂(h)
ρ̂(h) =
γ̂(0)
γ = Γp φ .
(1 + ψ1 z + ψ2 z 2 + . . .)(1 + φ1 z + φ2 z 2 + . . .) = (1 + θ1 z + . . .)
Statistics 910, #9 7
ψ1 + φ1 = θ1
But this only leads to the collection of ψ’s, not the covariances. The
covariances require summing the resulting expressions.
Mixed models Observe that the covariances satisfy the convolution ex-
pression (multiply both sides of (1) by lags Xt−j , j ≥ max(p, q + 1))
p
X
γ(j) − φk γ(j − k) = 0, j ≥ max(p, q + 1),
k=1
Partial regression The reasoning behind the use of the partial corre-
lations resembles the motivation for partial regression residual plots
which show the impact of a variable in regression. If we have the OLS
fit of the two-predictor linear model
Y = b0 + b1 X1 + b2 X2 + residual
r2 = X2 − a0 − a1 X1 , ry = Y − c0 − c1 X1 ,
Key property For an AR(p) process, φhh = 0, h > p so that the partial
correlation function cuts off after the order p of the autoregression.
Also notice that φpp = φp . Since an invertible moving average can be
represented an infinite autoregression, the partial autocorrelations of
a moving average process decay geometrically.
Hence, we have the following table of behaviors (Table 3.1):
Discussion
Weak spots We have left some problems only partially solved, such as
the meaning of infinite sums of random variables. How does one ma-
nipulate these expressions? When are such manipulations valid?