Autoregressive (AR) models form one of the fundamental components of the ARIMA family. The core idea behind an AR model is simple yet powerful: the current value of a time series can be predicted using a linear combination of its own previous values. This directly models the dependency of the series on its past.
An Autoregressive model of order p, denoted as AR(p), predicts the current value Y_t based on the p preceding values (Y_t−1,Y_t−2,dots,Y_t−p). The mathematical representation is:
Yt=c+ϕ1Yt−1+ϕ2Yt−2+⋯+ϕpYt−p+ϵt
Let's break down the terms:
Think of an AR(1) model (p=1). It states that the current value Y_t is primarily dependent on the immediately preceding value Y_t−1, plus a constant and some random noise:
Yt=c+ϕ1Yt−1+ϵt
If phi_1 is positive, a high value yesterday suggests a high value today. If it's negative, a high value yesterday suggests a low value today. An AR(2) model would use the two most recent past values (Y_t−1 and Y_t−2), and so on.
How do we determine the appropriate order p? As discussed in Chapter 3, the Partial Autocorrelation Function (PACF) plot is extremely useful here. For a pure AR(p) process, the PACF plot typically shows:
Observing such a pattern in the PACF of your (stationary) time series suggests that an AR(p) model might be a good starting point.
A sample time series exhibiting AR(1) behavior. Notice how consecutive points tend to be relatively close in value, reflecting the influence of the previous point.
It's important to remember that standard AR models assume the time series Y_t is stationary. Its mean, variance, and autocorrelation structure should not change over time. If your data is non-stationary (e.g., exhibits trends or seasonality), you'll typically need to transform it, often by differencing (as discussed in Chapter 2), before applying an AR model. This differencing step is the "Integrated" part represented by 'I' in ARIMA, which we will combine later in this chapter.
AR models capture one specific type of dependency structure in time series data. They form the basis for more complex models by explaining how past values influence the present. Next, we'll look at Moving Average (MA) models, which focus on the role of past forecast errors.
Was this section helpful?
© 2026 ApX Machine LearningEngineered with