ARIMA (AutoRegressive Integrated Moving Average) is a widely used method for Time Series Forecasting that models the autocorrelations within the data. It is particularly effective for datasets with trends or patterns that are non-seasonal. It is more stochastic, based on past correlations rather than explicit decomposition.

When to use:

  • When shocks, noise structure, or lagged dependencies drive the series more than trend/seasonality.
  • Strong Autocorrelation
  • When seasonality changes over time (SARIMA can capture evolving seasonal relationships better than Holt-Winters).

Requirement: Stationary Time Series

ARIMA Explained

ARIMA consists of three components:

  • AutoRegressive (AR): Uses past values to predict the current value.
  • Integrated (I): Applies differencing to make the series stationary. Differencing in Time Series
  • Moving Average (MA): Uses past forecast errors to improve predictions.

The main ARIMA parameters are:

– Autoregressive (AR) Order

  • Represents the number of lagged observations included in the model.
  • Purpose: AR terms capture how past values influence the current value.
  • Example: the model uses the previous 2 time points to predict the current one.
  • How to pick : Look at the Partial Autocorrelation Function (PACF Plots) plot:
    • The lag after which PACF cuts off (drops to near zero) suggests the p value.
    • Example: PACF is significant at lags 1 and 2, then drops p = 2.

– Differencing Order

  • Represents the number of times the series is differenced (Differencing in Time Series) to achieve stationarity (removing trends).
  • Example: model uses .
  • How to pick :
    • Visual inspection: Check if the time series has a trend or changing mean.
    • ADF Test (Augmented Dickey-Fuller): If p-value > 0.05 series is non-stationary increase .
    • Rule of thumb: Usually . Most series become stationary after 1–2 differences.

– Moving Average (MA) Order

  • Represents the number of lagged forecast errors included in the model.
  • Purpose: MA terms capture short-term shocks (Time Series Shocks) or noise.
  • Example: model uses the previous time step’s error to adjust the prediction.
  • How to pick : Look at the Autocorrelation Function (ACF) plot:
    • The lag after which ACF cuts off suggests the q value.
    • Example: ACF is significant at lag 1 q = 1.

What ARIMA Does

  1. Checks for stationarity and applies differencing ( times) if needed.
  2. Models the relationship between current values and:
    • Past values (AR component)
    • Past forecast errors (MA component)
  3. Fits parameters by minimizing a loss function (typically log-likelihood).
  4. Forecasts future values using the learned structure.

Fine-Tuning and Iterative Refinement

Why ARIMA Isn’t Enough for Seasonal Data

ARIMA does not model repeating patterns (e.g., quarterly or monthly seasonality). For such cases, SARIMA is used.

In ML_Tools see: