Questions tagged [diebold-mariano-test]

A test for superior predictive accuracy of one forecast over another.

The null hypothesis of the Diebold-Mariano test is that the expected forecast loss (e.g. expected mean squared forecast error) is equal for two alternative forecasts. In other words, its tests whether the two forecasts have the same accuracy in population.

The test makes use of the fact that asymptotically, the difference between the loss functions incurred by two forecasts (suitably standardized) approaches a normal distribution.

The original publication was Diebold & Mariano (1995, Journal of Business and Economic Statistics). Twenty years later, Diebold (2015, Journal of Business and Economic Statistics) offered some recollections on the test, its uses and abuses.


In comparing multiple forecasts, we need to address the multiple comparisons problem. In such a case, the standard approach is the "multiple comparisons to the best" (MCB) test originally proposed by Koning et al. (2005) for a re-analysis of the M3 forecasting competition. Most recently it has been applied to submissions in the M5 forecasting competition as well. It is rank-based, so it works with any accuracy measure (and appropriate point forecasts, Kolassa, 2020). A related alternative would be the Friedman-Nemenyi test (Demsar, 2006).

Both the MCB and the Nememyi test are implemented in the TStools package for R. An empirical comparison between the two is given by Hibon et al.'s 2012 ISF presentation.

42 questions
8
votes
2 answers

How can I determine that a forecast is significantly more accurate than another one? (time series)

Reproducible Example Look at this reproducible example: I have a time series that I want to forecast. For the sake of reproducibility, I'll just take AirPassanger. Let's say that I tried to fit 5 models: ETS, (auto)ARIMA, drifted SNAIVE, NNETAR,…
7
votes
1 answer

Diebold-Mariano test for predictive accuracy

I am using the Diebold-Mariano test in the forecast package of R to test the predictive accuracy. In particular, I want to underpin statistically that model 2 has a better accuracy. What I did is the following: These are the squared errors from two…
5
votes
1 answer

Alternatives for Diebold-Mariano test when comparing the best forecast among many against a benchmark

Suppose I encounter a new forecasting method and I wish to evaluate it against a benchmark. I can obtain forecasts from the two methods and compare them to actual realizations and thus obtain the forecast errors. Applying a loss function on the…
Richard Hardy
  • 54,375
  • 10
  • 95
  • 219
5
votes
1 answer

How to test if two RMSE are significantly different?

Say I have two models for a regression task and from each model I get a RMSE. One RMSE is smaller than the other, however I wish to test if the difference is statistically significant in order to be able to say that one model is better than the…
4
votes
2 answers

How to test superior predictive ability over multiple time series?

Suppose you have two models, model A and model B, and let these models forecast 10 time series over a horizon of 12 periods. That is, suppose the time series contain monthly data and your forecasting horizon is 1 year. Can I then statistically test…
4
votes
1 answer

t-test for time series (Diebold Mariano test?)

Currently I am working on the following problem: I want to compare the means of two different time series ts1 and ts2. As the samples are not iid and t-test can't be applied, I thought to use the Diebold Mariano test as an alternative. This test…
Simon
  • 41
  • 1
4
votes
1 answer

Diebold-Mariano with RMSE as loss function?

When applying the Diebold-Mariano test to test for predictive accuracy we need to specify a loss differential. For instance the loss differential $d$ in terms of the mean absolute error (MAE) is $d = abs(e_1) - abs(e_2)$ where $e_1$, $e_2$ are the…
4
votes
1 answer

ANOVA / t-test to compare the errors of different models

I have two forecasting models, moving average and single exponential smoothing. The values of Mean Absolute Percentage Error (MAPE) is 5.2%, 5.8%. Since the difference of MAPE between the models are very close, I am quite confused which model to…
Muk
  • 195
  • 10
3
votes
1 answer

Diebold-Mariano for in-sample model comparisons?

Suppose I have two in-sample forecasts from two different non-nested models. I want to check which one produces the best forecasts. A common way is Diebold-Mariano, GiacominiWhite, ENC-T test. However, these tests were designed for out-of-sample…
Daniel Pinto
  • 613
  • 4
  • 15
3
votes
1 answer

Diebold-Mariano in the context of volatility forecasting: What is the ultimate aim of this test?

Perhaps I'm missing a simple conceptual point here. But do the error statistics (RMSE, MAE) not tell which is the best forecast by presenting the lowest figures between the forecast and the actual figures? Why then do we need to compare two…
3
votes
1 answer

Difference between comparing forecasts and models

I started out looking for a way to test the difference between MSPE between two models (Question here), when (thanks to @Richard Hardy) I ended up reading a paper of Diebold regarding the Diebold-Mariano test (Comparing Predictive Accuracy, Twenty…
2
votes
1 answer

Can I give continuous rank probability score (CRPS) to Diebold-Mariano (DM) test?

I would like to use DM test for probabilistic forecasting case. My initial thinking was to give CRPS of two forecasting methods instead of raw forecast errors, where CRPS is calculated using prediction interval constructed under parametric…
2
votes
0 answers

Can i use Diebold Mariano test for comparison of 2 models across multiple time series?

I have 2 models (for simplicity, let's call them AR(1) and MA(1)) each making 1 day ahead forecasts of time series. If I had only 1 time series I would just use the Diebold-Mariano test to compare the predictive abilities of the models. But let's…
ira
  • 399
  • 2
  • 14
2
votes
1 answer

Tests of Forecast Accuracy for Nested Models

Can anyone explain why "classic" tests of forecast accuracy, (i.e. Diebold-Mariano test, Meese-Rogoff test and Morgan-Granger-Newbold test) are not suited for nested models? I could not find a good explanation in literature.
2
votes
1 answer

Diebold-Mariano test for multiple prediction horizons

I am trying to compare two forecasts using the Mariano Diebold test in R. Both forecasts are for 150 days ahead; that is, on day $t$ I forecast $t+1, t+2, \dotsc, t+150$. I deduced from this post that my forecast horizon $h=150$. Using that, the…
Marieke
  • 33
  • 1
  • 6
1
2 3