A Survey of Forecast Error Measures
A Survey of Forecast Error Measures
net/publication/281718517
CITATIONS READS
264 9,062
6 authors, including:
Anton Tyukov
Volgograd State Technical University
15 PUBLICATIONS 322 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Adriaan Brebels on 24 March 2016.
Submitted: Aug 7, 2013; Accepted: Sep 18, 2013; Published: Sep 25, 2013
Abstract: This article reviews the common used forecast error measurements. All error measurements have been
joined in the seven groups: absolute forecasting errors, measures based on percentage errors, symmetric errors,
measures based on relative errors, scaled errors, relative measures and other error measures. The formulas are
presented and drawbacks are discussed for every accuracy measurements. To reduce the impact of outliers, an
Integral Normalized Mean Square Error have been proposed. Due to the fact that each error measure has the
disadvantages that can lead to inaccurate evaluation of the forecasting results, it is impossible to choose only
one measure, the recommendations for selecting the appropriate error measurements are given.
INTRODUCTION et
= ( yt − ft(m) ) (1)
Different criteria such as forecast error measurements,
the speed of calculation, interpretability and others have where - yt is the measured value at time t, ft
( m) - predicted
been used to assess the quality of forecasting [1-6].
value at time t, obtained from the use of the forecast
Forecast error measures or forecast accuracy are the most
model m. Hereinafter referred to as the index of the model
important in solving practical problems [6]. Typically, the
(m) will be omitted.
common used forecast error measurements are applied for
Mean Absolute Error, MAE is given by:
estimating the quality of forecasting methods and for
choosing the best forecasting mechanism in case of n
1
multiple objects. A set of "traditional" error measurements
in every domain is applied despite on their drawbacks.
MAE
= = ∑
n i =1
ei mean ei ,
i =1, n
(2)
A Review n
∑( ) ( )
Absolute Forecasting Error: The first group is based on = 1
MSE = ei2 mean ei2 , (4)
the absolute error calculation. It includes estimates based n i =1 i =1, n
on the calculation of the value ei
Corresponding Author: Shcherbakov, Volgograd State Technical University, Lenin avenue, 28, 400005, Volgograd, Russia.
171
World Appl. Sci. J., 24 (Information Technologies in Modern Industry, Education & Society): 171-176, 2013
domain. The group of percentage based errors includes and the median mean absolute percentage error
the following errors.
Mean Absolute Percentage Error, MAPE sMdAPE median ( 200 ⋅ si ) .
=
i =1, n (13)
n
1
MAPE
=
n ∑100 ⋅=
pi mean (100 ⋅ pi
i =1, n
) (7) To avoid the problems associated with the division
i =1 by zero, a modified sMAPE - Modified sMAPE, msMAPE
Median Absolute Percentage Error, MdAPE is more has been proposed. Their denominators have an
resistant to outliers and calculated according to the additional member:
formula
n
1 y −f
MdAPE median (100 ⋅ pi
= ) msMAPE =
n ∑ ( yi + ifi ) /i2 + Si , (14)
i =1, n (8) i =1
RMSPE
= mean (100 ⋅ pi
i =1, n
)2 , (9) Developing the idea for the inclusion of an additional
terms, more sophisticated measures was presented [16]:
and the median percentage error of the quadratic KL-N, KL-N1, KL-N2, KL-DE1, KL-DE2, IQR
172
World Appl. Sci. J., 24 (Information Technologies in Modern Industry, Education & Society): 171-176, 2013
The following disadvantages should be noted. If naive model has been chosen then division by zero
error occurs in case of continuous sequence of
Despite its name, this error is also non-symmetric identical values of the time series.
[13].
Furthermore, if the actual value is equal to forecasted Scaled Error: As a basis for calculating the value of the
value, but with opposite sign, or both of these values scaled errors qi is given by
are zero, then a divide by zero error occurs.
These criteria are affected by outliers in analogous et
qt = .
with the percentage errors. 1
n
and Geometric Mean Relative Absolute Error, GMRAE), where MAE and MAE* the mean absolute error for the
which is calculated similarly to (17), but instead of mean(•) analyzed forecasting model and the reference model
the geometric mean is obtained gmean(•). respectively, calculated using the formula (2).
Relative Root Mean Square Error, RelRMSE is
It should be noted the following shortcomings. calculated similarly to (23), except that the right side is
calculated by (5)
Based the formulas (15-18), division by zero error
occurs, if the predictive value obtained by reference RMSE
RRMSE = .
model is equal to the actual value. RMSE* (24)
173
World Appl. Sci. J., 24 (Information Technologies in Modern Industry, Education & Society): 171-176, 2013
In some situations it is reasonable to calculate the over the entire interval or time horizon or defined
logarithm of the ratio (23). In this case, the measure is short interval of observation [18]. However, this
called the Log Mean Squared Error Ratio, (LMR) estimate is affected by influence of outliers, if outlier has
a value much bigger the maximal "normal" value. To
RMSE reduce the impact of outliers, Integral Normalized Mean
LMR = log .
RMSE * (25) Square Error [19] have been proposed, calculated by the
formula:
Syntetos et al. proposed a more complex assessment
of the relative geometric standard deviation Relative
Geometric Root Mean Square Error, RGRMSE [17].
inRSE = n
1
( )
mean ei2 .
({
100% ⋅ mean I MAE < MAE * .
PB ( MAE ) = }) (26) n
∑ ( ei )
2
where I{•} - the operator that yields the value of zero or inRSE = i =1 , (30)
n
∑ ( yi − y )
2
one, in accordance with the expression:
i =1
0, if MAE < MAE *;
I ( MAE ) = (27) n
1. where y=
1
n ∑ yk .
k =1
By analogy with PB (MAE), Percentage Better (MSE)
Other measures are called normalized std_APE and
can be defined.
std_MAPE [20, 21] and calculated by the formula
The disadvantages of these measures are the
n
following.
∑ ( ei − MAE )2
Std _ AE = i =1 (31)
Division by zero error occurs if the reference forecast n −1
error is equal to zero.
These criteria determine the number of cases when
and
the analyzed forecasting model superior to the base
but do not evaluate the value of difference.
n
nRMSE =
1
y
( )
mean ei2 ,
(28)
Recommendations How to Choose Error Measures:
i =1, n One of the most difficult issues is the question of
choosing the most appropriate measures out of the
where y - the normalization factor, which is usually equal groups. Due to the fact that each error measure has the
to either the maximum measured value on the forecast disadvantages that can lead to inaccurate evaluation of
horizon, or the difference between the maximum and the forecasting results, it is impossible to choose only one
minimum values. Normalization factor can be calculated measure [5].
174
World Appl. Sci. J., 24 (Information Technologies in Modern Industry, Education & Society): 171-176, 2013
175
World Appl. Sci. J., 24 (Information Technologies in Modern Industry, Education & Society): 171-176, 2013
11. Shcherbakov, M.V. and A. Brebels, 2011. Outliers and 18. Tyukov, A., M. Shcherbakov and A. Brebels, 2011.
anomalies detection based on neural networks Automatic two way synchronization between server
forecast procedure. In the Proceedings of the 31st and multiple clients for HVAC system. In the
Annual International Symposium on Forecasting, Proceedings of The 13th International Conference on
ISF-2011, pp: 21-22. Information Integration and Web-based Applications
12. Goodwin, P. and R. Lawton, 1999. On the asymmetry & Services, pp: 467-470.
of the symmetric MAPE. International Journal of 19. Brebels, A., M.V. Shcherbakov and V.A. Kamaev,
Forecasting, 15(4): 405-408. 2010. Mathematical and statistical framework for
13. Koehler, A.B., 2001. The asymmetry of the sAPE comparison of neural network models with other
measure and other comments on the M3-competition. algorithms for prediction of Energy consumption in
International Journal of Forecasting, 17: 570-574. shopping centres. In the Proceedings of the 37 Int.
14. Makridakis, S., 1993. Accuracy measures: Theoretical Conf. Information Technology in Science Education
and practical concerns. International Journal of Telecommunication and Business, suppl. to Journal
Forecasting, 9: 527-529. Open Education, pp: 96-97.
15. Kolassa, S. and R. Martin, 2011. Percentage errors 20. Casella, G. and R. Berger, 1990. Statistical inference.
can ruin your day (and rolling the dice shows how). 2nd ed. Duxbury Press, pp: 686.
Foresight, (Fall): 21-27. 21. Kusiak, A., M. Li and Z. Zhang, 2010. A data-driven
16. Assessing Forecast Accuracy Measures. Date View approach for steam load prediction in buildings.
01.08.2013 http:// www.stat.iastate.edu/ preprint/ Applied Energy, 87(3): 925-933.
articles/2004-10.pdf
17. Syntetos, A.A. and J.E. Boylan, 2005. The accuracy
of intermittent demand estimates. International
Journal of Forecasting, 21(2): 303-314.
176