0% found this document useful (0 votes)
12 views7 pages

Econometric Methods

The document analyzes quarterly profits from 1971 to 1991, revealing an overall upward trend with cyclical fluctuations and a need for first-order differencing to achieve stationarity for ARIMA modeling. It compares two models, ARIMA (1,1,1) and the more complex auto-selected ARIMA (2,1,2) (0,0,1) [4], finding the latter to provide a better fit based on AIC, BIC, and R² values. Ultimately, the auto ARIMA model is preferred for forecasting due to its superior predictive accuracy and tighter confidence intervals.

Uploaded by

fahimhassan19113
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
12 views7 pages

Econometric Methods

The document analyzes quarterly profits from 1971 to 1991, revealing an overall upward trend with cyclical fluctuations and a need for first-order differencing to achieve stationarity for ARIMA modeling. It compares two models, ARIMA (1,1,1) and the more complex auto-selected ARIMA (2,1,2) (0,0,1) [4], finding the latter to provide a better fit based on AIC, BIC, and R² values. Ultimately, the auto ARIMA model is preferred for forecasting due to its superior predictive accuracy and tighter confidence intervals.

Uploaded by

fahimhassan19113
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 7

Trend Profit from 1971 to 1991

The first plot displays the Quarterly Profits time


series from 1970 to the early 1990s. The overall
trend is clearly upward, indicating that profits
generally increased over the two-decade period.
However, the series also shows noticeable cyclical
fluctuations and a few sharp drops and rebounds,
suggesting short-term volatility and economic
shocks. The pattern appears non-linear, and the
presence of recurring movements hints at possible
seasonal or cyclical behavior, which justifies further
decomposition.

Classical decomposition (corrected plot)


The classical additive decomposition of the quarterly
"Profits" time series breaks the data into three distinct
components: trend, seasonality, and random (residual). The
observed line represents the original time series data, which
shows a general upward movement with periodic
fluctuations over time. The trend component clearly
captures this long-term upward trajectory in profits,
particularly evident from the early 1970s to the late 1980s,
with a slight decline or leveling toward the end of the series.
This confirms the presence of sustained economic growth
in the data.

The seasonal component reveals a consistent and repeating pattern every four quarters, indicating the
presence of strong and regular seasonal effects. Each year, specific quarters systematically show increases
or decreases in profit levels, likely due to underlying economic or fiscal cycles. This seasonal structure
remains stable throughout the entire period, suggesting that the seasonality is not evolving over time.

Lastly, the random or irregular component accounts for short-term, unpredictable fluctuations that are not
explained by either trend or seasonality. These residual variations are relatively small in magnitude
compared to the trend and seasonal components, indicating that the majority of the variability in the time
series is systematic and predictable.

Overall, the classical decomposition reveals that the "Profits" series is driven primarily by a combination
of a strong upward trend and stable seasonality, with limited noise from random effects. This structure
supports the use of time series models such as ARIMA or Seasonal ARIMA for effective forecasting.

Augmented Dickey-Fuller Test to check stationary status


Object Dickey-Fuller Lag order p-value
Original profits 2.7963 4 0.2487
First-differenced profits -3.994 4 0.01356

To determine whether the "Profits" time series is stationary, we applied the Augmented Dickey-Fuller
(ADF) test, which checks for the presence of a unit root a characteristic of non-stationary data. The test was
first applied to the original series. The result yielded a Dickey-Fuller test statistic of –2.7963 with a p-value
of 0.2487. Since the p-value is greater than the conventional 0.05 threshold, we fail to reject the null
hypothesis of a unit root. This indicates that the original series is non-stationary, and therefore unsuitable
for direct modeling with ARIMA.

To address this, we applied a first-order difference to the series and re-ran the ADF test on the differenced
data. This time, the test produced a Dickey-Fuller statistic of -3.9945 and a p-value of 0.01356. As this p-
value is less than 0.05, we reject the null hypothesis and conclude that the differenced series is stationary.

These results confirm that the "Profits" series requires first-order differencing (d = 1) to achieve stationarity,
which is a necessary condition for building a valid ARIMA model.
ACF (Autocorrelation Function) and PACF (Partial Autocorrelation Function)
To forecast the "Profits" variable using the ARIMA model,
the first step in the Box-Jenkins methodology is to identify
the appropriate orders of the autoregressive (p), differencing
(d), and moving average (q) components. The original time
series was first checked for stationarity using the Augmented
Dickey-Fuller (ADF) test. The result of the ADF test
indicated that the series was non-stationary, suggesting the
need for differencing. After applying first-order differencing,
the ADF test confirmed stationarity of the differenced series.
Therefore, the order of differencing was selected as d = 1.

Next, to determine the values of p and q, we examined the


autocorrelation function (ACF) and partial autocorrelation
function (PACF) plots of the differenced series. The ACF plot
showed a sharp cutoff after lag 1, while the PACF plot also
exhibited a significant spike at lag 1 followed by a drop. This
pattern suggests that both AR(1) and MA(1) components may
be present in the model. Based on these observations, we
tentatively identified the model as ARIMA (1,1,1).

Step 2: Parameter Estimation (Customized and Auto Regressive Model)


Model ARIMA (1,1,1) ARIMA (2,1,2) (0,0,1) [4]
Coefficients ar1 = 0.4066 ar1 = 1.3083
ma1 = -0.1311 ar2 = -0.8830
ma1 = -0.8869
ma2 = 0.4897
sma1 = 0.7690
Standard Errors ar1 = 0.5823 ar1 = 0.0889
ma1 = 0.6533 ar2 = 0.0816
ma1 = 0.1339
ma2 = 0.1124
sma1 = 0.0982
σ² (sigma²) 80.63 72.32
Log Likelihood -314.45 -308.13
Training ME 1.1360 0.8727
Training RMSE 8.9284 8.2088
Training MAE 6.5806 6.1264
Training MPE 0.9918 0.8401
Training MAPE 5.1948 4.8564
Training MASE 0.9398 0.3119
Training ACF1 -0.0076 -0.0291

After tentatively identifying the ARIMA (1,1,1) model, we estimated its parameters using maximum
likelihood estimation. The fitted model produced an autoregressive coefficient (ar1) of 0.4066 and a moving
average coefficient (ma1) of –0.1311, with standard errors of 0.5823 and 0.6533, respectively. These
relatively large standard errors suggest that the individual coefficients are not highly significant, though the
model does capture some of the dynamics of the series. The model’s estimated variance (σ²) was 80.63, and
it had a log-likelihood of –314.45, resulting in an AIC of 634.91. The training set error measures, including
RMSE = 8.93, MAE = 6.58, and MAPE = 5.19%, indicate a moderate level of forecasting error.

In contrast, the automated model selected by auto.arima() suggested a more complex seasonal ARIMA
model: ARIMA (2,1,2) (0,0,1) [4], which includes two autoregressive terms, two moving average terms,
and one seasonal moving average term. This model had more statistically significant coefficients, with low
standard errors, and achieved a better overall fit. It yielded a lower estimated variance (σ² = 72.32), a higher
log-likelihood (–308.13), and notably better model selection criteria with AIC = 628.25 and BIC = 643.05.
Furthermore, the error metrics were improved compared to the simpler model, with RMSE = 8.21, MAE =
6.13, and MAPE = 4.86%, indicating enhanced predictive accuracy.

Overall, the ARIMA (2,1,2) (0,0,1) [4] model appears to provide a better fit to the data based on lower AIC,
BIC, and error measures, and its parameters are estimated with greater statistical significance.
Step 3: Diagnostic Checking (Residual Analysis)

Box-Ljung test

Data X-squared df p-value


Residuals (ARIMA (1,1,1)) 41.396 20 0.003312

After fitting the ARIMA (1,1,1) model, the next step in the Box-Jenkins methodology is to evaluate whether
the model's residuals behave like white noise. This means the residuals should have a mean of zero, constant
variance, and no significant autocorrelation. A residual diagnostic plot was generated, which included the
residual time series, the ACF of the residuals, and the PACF of the residuals. From the residual plot, we
observe that the residuals fluctuate randomly around zero without any visible trend, suggesting the absence
of systematic patterns. Additionally, the ACF and PACF plots show that most of the autocorrelations fall
within the 95% confidence bounds, indicating no significant autocorrelation remaining in the residuals.

To statistically verify this observation, we conducted the Box-Ljung test, which tests the null hypothesis
that the residuals are independently distributed. The test produced a Chi-squared value of 41.396 with 20
degrees of freedom and a p-value of 0.0033. Since the p-value is less than the conventional significance
level of 0.05, we reject the null hypothesis. This suggests that the residuals of the ARIMA (1,1,1) model
are not purely random and still exhibit some autocorrelation.

Therefore, although the residual plots appear mostly acceptable visually, the Box-Ljung test indicates a
potential lack of complete model adequacy. This suggests that a more complex model (such as ARIMA
(2,1,2) (0,0,1) [4]) may provide a better fit, as seen in Step 2, where it had both lower AIC/BIC and better
error metrics.

Step 4: Forecasting
In the final step of the Box-Jenkins methodology,
forecasts were generated using both the manually
specified ARIMA (1,1,1) model and the auto-selected
ARIMA (2,1,2) (0,0,1) [4] model. The forecasts cover
the next 8 quarters, providing short-term projections
for the “Profits” variable.

The forecast plot for the ARIMA (1,1,1) model shows


a relatively flat trajectory, indicating that future
profits are expected to remain stable with no
significant upward or downward trend. The shaded
area around the forecast line represents the 80% and
95% confidence intervals, which widen over time.
This widening reflects increasing uncertainty in the
forecast as the time horizon extends. The central
forecast path is smooth, but the wider confidence
bands indicate potential variability in future
outcomes.

In contrast, the forecast from the auto ARIMA model


(ARIMA (2,1,2) (0,0,1) [4]) is more nuanced. While
the point forecast similarly predicts a stable pattern,
the confidence intervals are slightly narrower,
suggesting greater confidence and lower forecast
uncertainty compared to the ARIMA (1,1,1) model. This aligns with earlier findings that the auto ARIMA
model had a better fit, as indicated by lower AIC, BIC, and residual error measures.

Overall, both models predict stability in future profits; however, the auto ARIMA model provides a more
statistically reliable forecast with tighter confidence bands. Based on forecasting accuracy, model
diagnostics, and residual behavior, the auto ARIMA model is preferred for making predictions in this case.
Model Comparison
Model ARIMA Order df AIC BIC R²
model_111 ARIMA (1,1,1) 3 634.9059 642.3036 0.965311
model_auto ARIMA (2,1,2) (0,0,1) [4] 6 628.2536 643.0490 0.9706773

To determine the best-fitting model for forecasting the "Profits" variable, we compared the manually
specified ARIMA (1,1,1) model with the automatically selected ARIMA (2,1,2) (0,0,1) [4] model using
several statistical criteria: AIC (Akaike Information Criterion) and R² (coefficient of determination).

The AIC for ARIMA (1,1,1) was 634.91 with 3 degrees of freedom, while the AIC for the auto ARIMA
model was 628.25 with 6 degrees of freedom. A lower AIC value indicates a better trade-off between model
fit and complexity, suggesting that the auto ARIMA model provides a superior fit despite its higher
complexity.

Additionally, the R² value for ARIMA (1,1,1) was 0.9653, whereas the auto ARIMA model achieved a
slightly higher R² of 0.9707. This indicates that the auto ARIMA model explains a slightly larger proportion
of the variance in the data, further supporting its superior performance.

In summary, both the lower AIC and higher R² values suggest that the auto ARIMA model fits the data
better and should be preferred for forecasting. This conclusion is consistent with the residual diagnostics
and forecasting accuracy observed in earlier steps.

You might also like