Short-Term Load Forecasting Using Smart Meter Data
Short-Term Load Forecasting Using Smart Meter Data
Article
Short-Term Load Forecasting Using Smart Meter Data:
A Generalization Analysis
Aida Mehdipour Pirbazari *, Mina Farmanbar, Antorweep Chakravorty and Chunming Rong
Department of Electrical and Computer Engineering, University of Stavanger, 4036 Stavanger, Norway;
[email protected] (M.F.); [email protected] (A.C.); [email protected] (C.R.)
* Correspondence: [email protected]
Received: 14 March 2020; Accepted: 15 April 2020; Published: 21 April 2020
Abstract: Short-term load forecasting ensures the efficient operation of power systems besides
affording continuous power supply for energy consumers. Smart meters that are capable of providing
detailed information on buildings energy consumption, open several doors of opportunity to
short-term load forecasting at the individual building level. In the current paper, four machine
learning methods have been employed to forecast the daily peak and hourly energy consumption of
domestic buildings. The utilized models depend merely on buildings historical energy consumption
and are evaluated on the profiles that were not previously trained on. It is evident that developing
data-driven models lacking external information such as weather and building data are of great
importance under the situations that the access to such information is limited or the computational
procedures are costly. Moreover, the performance evaluation of the models on separated house profiles
determines their generalization ability for unseen consumption profiles. The conducted experiments
on the smart meter data of several UK houses demonstrated that if the models are fed with sufficient
historical data, they can be generalized to a satisfactory level and produce quite accurate results
even if they only use past consumption values as the predictor variables. Furthermore, among the
four applied models, the ones based on deep learning and ensemble techniques, display better
performance in predicting daily peak load consumption than those of others.
Keywords: smart meters; short-term load forecasting; machine learning; deep learning; generalization
analysis
1. Introduction
Over the last decade, smart meters have been rapidly deployed around the world.
Around 86 million and 11 million smart meters have been installed by large and small suppliers in
the US and UK respectively by the end of 2018 [1,2]. Almost 90% of these meters were installed
for residential customers. One of the main objectives of smart metering in residential sectors is
to encourage users to consume less energy by raising awareness about their consumption levels.
Smart meters provide information on cost and amount of energy consumption in near real-time for
both suppliers and consumers. Regarding the households, huge amounts of fine-grained data on the
use of energy not only provide them with more accurate bills but also with valuable information on
their electricity consumption habits and time-based pricing rates. This information through demand
response and incentivization programs would help them to reduce their energy usage during peak
hours and schedule their appliances according to electricity prices. High-resolution data generated by
smart meters, on the other hand, provide suppliers with several controlling functions such as power
quality monitoring and power loss identification. Moreover, it opens many doors of opportunities in
electricity load analysis such as load forecasting with high accuracy at lower aggregation levels [3,4].
Electrical load forecasting is the prediction of the load demand that an electricity consumer will
have in the future. Load forecasts help suppliers to balance supply and demand and to ensure the
reliability of power grids at the time of power deficiency. Load forecasts are also important for electricity
traders to balance their electricity purchase and sales [5]. Load forecasting is performed in a wide range
of time horizons aiming at different targets: short-term load forecast (a few minutes to 1 day ahead) to
adjust supply and demand; medium-term load forecast (1 day to 1 year ahead) to plan outage and
maintenance and long-term load forecast (more than 1 year ahead) to plan the development of power
infrastructures. Load forecasting is also performed in various aggregation levels when it is applied to
the areas with different geographical scales such as a country, city, small communities or a building.
The forecasting task becomes more challenging when it comes to lower aggregation levels such as a
building level since, many fluctuating factors affect a building’s energy consumption with varying
degrees such as weather parameters, building properties, Heating, Ventilating and Air-Conditioning
(HAVC) facilities and the consumption behavior of occupants [6,7].
A large number of studies on accurate short-term load forecasting has been reported in recent
years due to its impact on the efficient operation of power systems and the economy. Furthermore,
many studies have benefited from smart metering data to develop more advanced models for load
forecasting at individual building levels. The methods for predicting building energy consumption
generally are classified into two categories: engineering (physical) and data-driven techniques.
Engineering methods use mathematical equations to present the physical components and thermal
performance of buildings. However, they need high details about different parameters of the buildings
that are not always available. Moreover, a high level of expertise is required to perform expensive and
elaborate computations. On the other hand, data-driven approaches do not need such detailed data
about the simulated building and instead learns from real-time or historical data. These approaches
are further classified into two groups: statistical and AI-based techniques [8,9].
Statistical methods use historical data as an aim for correlating energy consumption with the
most important variables as inputs. Therefore, a larger amount of historical data with high quality
plays an important role in the effectiveness of statistical models. Traditional linear statistical models
such as Gaussian mixture models (GMM), Conditional demand analysis (CDA), Regression models
and auto autoregressive moving average (ARMA) and ARIMA, have remained the baseline for time
series prediction with widespread use in many applications [10]. Although it is easy to use statistical
techniques, the basic assumption of such models is based on the fact that time series are considered
linear and therefore follow a specifically known distribution of statistics. Numerous machine learning
models have been developed to overcome these limitations. The models based on Support Vector
Machines (SVM), as well as Classification and Regression Trees (CART), are among the successful
machine learning techniques used in time series forecasting and energy applications.
Over the past decades, many researchers have investigated the application of AI-based techniques
in forecasting problems. Among AI-based techniques, Artificial Neural Networks (ANNs) with
different structures have been widely applied in the load forecasting domain [11]. ANNs similar to
statistical methods use historical data to build a model. However, with hidden layer structures and
learning ability offer several advantages over statistical and classical machine learning techniques
for time series forecasting. They are considered data-driven and self- adaptive methods which can
capture subtle and functional patterns through a training process on historical records of data, even if
the underlying relationship between input and output variables is complex or unknown. Nevertheless,
the neural networks with shallow structures have the disadvantage of assuming that all inputs and
outputs are independent of each other, even when dealing with sequential data [12].
Recent studies in time series forecasting have shown the better performance of prediction models
using neural networks with deep architecture. Long Short-Term Memory (LSTM) network which
was first proposed by [13], is a variation of deep learning concept which was designed specifically to
learn the short-term and long-term dependencies present in sequential data. LSTM has been popular
with excellent accuracy in the realm of sequence learning like natural language translation and speech
Processes 2020, 8, 484 3 of 22
recognition [14]. In recent years there has been an increase in the number of studies on the application
of LSTM networks and their variants in short-term load forecasting.
According to the literature, most classical and AI-based methods specifically deep techniques
developed for load forecasting, require sufficient historical load data for training. At the building level,
they mainly have access to the historical energy consumption of the building under study and utilize it
for training the model. Subsequently, for performance evaluation, they use the future profile of the
same house or building. Similarly, in higher aggregation levels such as a community or a substation
level, the models are trained on aggregated historical consumption of buildings and are tested on the
future profile of the same buildings. Typically, the low testing error of the models guarantees the precise
prediction and the small difference between train and test error ensures the models’ generalization
ability. Moreover, many of the studies use additional information and build multivariate models based
on consumer behavior [15], weather and calendar parameters [16] appliance measurements [17], etc.
to improve the forecasting accuracy.
However, there are still some issues about the forecasting accuracy and generalization ability of
such models which have not been largely addressed. For example, to what extent the generalization
ability can be expanded or trusted and what happens to the model forecasting accuracy if we only
provide them with consumption data. The first question focuses on the scalability of the models;
how successful the forecasting models are when tested on a different profile that they are not previously
trained on. The test profiles could be quite different from the trained ones in terms of consumption
patterns or average daily and yearly consumption. This may happen in scenarios when historical
information on a building’s energy consumption is not accessible and we can still rely on predictive
models trained on available historical profiles. For instance, if a new house profile is added to a
community, or a new smart meter has been installed. The model developed in these situations can also
be less expensive in terms of complexity and computation time. The second question focuses on how
powerful a model can be if we only have access to anonymized data on historical energy consumption
due to privacy issues or lack of other data sources.
This paper investigates the mentioned scenarios with a focus on short-term load forecasting at
an individual building level. For this purpose, we develop four baseline models to predict hourly
residential load demand and evaluates their predictive accuracy and generalization ability in the given
scenarios. The models are chosen from the category of most-widely used machine learning methods for
energy prediction known as ANNs, Support vector Machines, regression trees (CART) and LSTM with
standard architecture. They are trained on consumption data of various load profiles and tested on
unseen houses with different levels of hourly load volatility. Furthermore, the sensitivity of the models
on the size of training data and the number of input variables will be tested and a comprehensive
analysis of forecasting results will be provided. The developed models are expected to learn various
load profiles relying on the built-in information in time series data and are aimed at improving
generalization ability and increasing model robustness. The models that produce low-prediction errors
on multiple houses can further be used as representative predictive models for a group of houses in a
community. In demand response applications, this can potentially remove the need to build separate
forecasting models per house profile within the community of houses.
The paper is structured as follows. Section 2 provides an overview of the literature. Sections 3
and 4 briefly presents the architecture and design considerations of the implemented forecasting
techniques. Then, Section 5 introduces the performance metrics for model evaluations. The dataset
used in our analysis is described in Section 6. The experimental results and discussion are provided in
Section 7. The paper ends with a conclusion in Section 8.
2. Related Work
There are many forecasting models that have been investigated and proposed since the 1970’s for
energy predation. Among them, statistical techniques have been extensively applied in load demand
forecasting problems. For example, in [18] the authors developed one-day-ahead forecasts on hourly
Processes 2020, 8, 484 4 of 22
and daily electricity loads of a house using both simple and multiple regression analyses. They utilized
weather parameters as the predictor variables and concluded that models trained on the daily dataset
provide more accurate forecasting results. S.Sp. Pappas et al. in [19], proposed a method for electricity
consumption and price forecasting using AutoRegressive Moving Average (ARMA) models based on
adaptive multi-model partitioning theory. Their results show that the proposed method could apply to
online modeling and noisy input data.
There have been some hybrid load forecasting approaches that are based on statistical techniques.
For instance, XiaoshuLü et al. [20] presented a hybrid model based on a physical–statistical approach to
improve forecast accuracy in energy and building applications. The physical model was developed to
define the physical concepts of energy streams while the statistical technique was designed to consider
model inconsistencies and specific diversity of buildings.
Support Vector Regression has also been applied in time series prediction as well as power load
demand forecasting. In [21] three Support Regression Models and an improved SVR variant utilizing
optimization algorithms were used to predict the day-ahead electricity load. The models’ effectiveness
centered on the small size of training data and their online learning functionality. In [22] we can find a
comprehensive overview of SVR applications in time series prediction as well as power load demand
forecasting. Many researchers have also attempted to apply the Classification and Regression Trees
(CART) techniques to improve the load forecast accuracy. For instance, Lahouar.A. et al. [23] proposed
a model based on random forests for short term load forecast with special attention to load profile,
holidays and customer behavior. Similarly, researchers in [7,24] utilized environmental and calendar
features to develop a method for electric load forecasting based on Decision Tree and algorithms.
Recent studies have shown the better performance of prediction models using AI-based
techniques due to their ability to learn nonlinearities between inputs and outputs. Among AI-based
techniques, artificial neural networks have been successfully applied in the forecasting domain.
Nasreen K. Ahmed et al. [25] performed an empirical comparison of machine learning models for time
series forecasting. In addition to the classical techniques, they analyzed several variations of artificial
neural networks. The experiments demonstrated that multilayer perceptron and the gaussian process
regression outperform other state-of-the-art models. There are also some studies that discuss the
hybridization of different ANN approaches and are successfully applied to short-term load forecasting.
In [26] Hamid R. Khosravani et al. developed hybrid models based on different neural network
architectures and genetic algorithms with several optimization parameters to predict electric power
demand at the Solar Energy Research Center. The comparison results with an autoregressive baseline
model reveal that the models based on the multi-objective genetic algorithm outperformed the model
based on computational and empirical methods with lower complicity. Kuihe Yang et al. [27] proposed
an ANN-based method with fuzzy logic to develop models with fewer complexities and to improve
the accuracy of forecasts.
There have been numerous studies that utilized optimization algorithms to optimize the structural
and training parameters of ANNs in forecasting problems. For example, Chaturvedi et al. [28] have
demonstrated the effectiveness of training neural networks with a Genetic algorithm as an optimization
strategy. A review study on different variants of artificial neural networks in the field of short-term
load forecasting emphasizes that a combination of neural networks with evolutionary algorithms could
outperform the singles models in terms of forecasting accuracy [29].
Over the last decade, neural networks with deep structure have increasingly attracted the attention
of researches in prediction problems. Compared to shallow networks, they benefit from many hidden
layers, exponentially fewer neurons, better activation functions, and parameter initialization techniques
as well as effective learning algorithms. Different versions of deep neural networks are recently being
employed in energy prediction context, especially LSTM networks and their variance due to their
capability to capture the temporal behavior of time series data. Daniel. M et al. [30] investigated the
effectiveness of LSTM-based architectures for building level energy load forecasting. They applied two
standard and Sequence to Sequence (S2S) architectures for the hourly forecast of a residential load
Processes 2020, 8, 484 5 of 22
dataset with one-minute and one-hour resolutions. Experimental results showed that the standard
LSTM performing better in one-hour resolution data while S2S performed well on both datasets.
In another study by Kong et al. [31] on short-term residential load forecasting an LSTM-based framework
has been assessed for both individual and aggregated prediction levels. The comparison results with
several state-of-the-work approaches demonstrated the superiority of LSTM for individual residential
load forecasting. In terms of aggregated (substation) level, the aggregation of all individual forecasts
yielded better results than the direct forecast of aggregated loads. Agrawal et al. [32] Introduced a
deep-structure RNN-LSTM network at a higher aggregation level; ISO New England energy market
using daily, monthly and weekly features to produce hourly predictions over a one-year period.
Similar to the approaches using shallow ANNs, some studies explored the combination of LSTM with
other models or optimization algorithms. For instance, in [33] a CNN-LSTM neural network was
proposed to predict the energy consumption of residential buildings with higher accuracy. The CNN
layer was used to extract complex features influencing energy consumption and the LSTM layer was fed
with the features to model the temporal information in time series components. Mamun et al. [34] and
Bouktif et al. [35] investigated the effectiveness of hybrid deep neural networks based on LSTM and
genetic algorithms for load forecasting on the energy market and metropolitan’s electricity consumption
data sets. Application of feature selection in [35] proved that using only optimal lagged features
as the input to the LSTM model produces the lowest forecasting error for both medium-term and
long-term horizons.
3. Modeling Techniques
In this paper, four modeling techniques are used for energy load forecasting: Support Vector
Regression (SVR) with Radial Basis Function kernel, Gradient Boosting Regression Trees (GBRT) driven
from Classification and Regression Tree (CART) analysis, feedforward neural networks (FFNNs) and
LSTM networks. The first two methods belong to the category of classical machine learning techniques
and the other two belong to AI-based machine learning techniques with shallow and deep structures
respectively. In the following, the detailed information about each model is provided.
1 T Xl
min w w+C (ξ +ξi∗ ),
2 i=1 i
where xi maps to a higher dimensional space and ςi is the upper training error (ξi∗ is the lower) subject
to the ε-insensitive tube |yi −(wT ϕ (xi )+b)| ≤ε. The parameters that control the output of the regression
are the error cost C, the width of tube ε, and the mapping function, ϕ [37]. The constraints imply that
most data xi are put in the tube |yi −(wT ϕ (xi )+b)|≤ε. If xi is not in the tube, there is an error ςi or ξi∗
which we would like to minimize the objective function. ε is always zero in traditional least-square
Processes 2020, 8, 484 6 of 22
regression and data is not mapped to higher dimensional spaces. The SVR formulation theory is similar
concerning SVM, and once equipped, the SVR will produce predictions using the following formula:
l
X
f(x)= θi ϕ (x, xi )+b.
i=1
We used a Radial Basis Function (RBF) as the kernel function due to its ability to capture non-linear
relationships between inputs and outputs. The RBF kernel on two samples x and x0, represented as
feature vectors in the input space, is defined as:
kx − x0 k2
!
0
K(x, x ) = exp − ,
2 σ2
where kx−x0 k2 is the squared Euclidean distance between the two feature vectors and σ is a
free parameter.
1
l( y, p) = ( yi − F(xi ))2 .
2
The negative gradient is simply computed as follows:
∂ L( yi , F(xi ))
" #
− = yi − Fxi .
∂F(xi )
The simplicity of the gradient computation will facilitate the residual refitting of the GBM
algorithm. The concept behind this loss function is to put penalties on large residuals while neglecting
small deviations from the target outputs.
In the GBM algorithm with decision trees as the base learners, the first step is to construct a
base tree h(x; a) using a training dataset (x1 , y1 ), (x2 , y2 ), . . . , (xN , yN ) with size N. Then for the
iterations from 1 to M, the negative gradients are computed and a new tree h(x; am ) is fitted. Each tree
is further updated according to the best gradient step pm and is added to the ensemble. After M
iterations, all the regression trees which added sequentially to the ensemble form the output of the
algorithm as a combination of weak learners [40].
the error of output every time and feeds back this information to the network to reduce the error up to
an acceptable predefined value. Further, more details on back-propagation algorithms are described
in [41].
The input values in an MLP structure are weighed through weight matrices and the output of
neurons is determined through an activation function. The structure of the Artificial Neural network
that we used in our study illustrated in Figure 1. In the Feed Forward Neural Network illustrated
above, given an input sequence x = (x1 , . . . , xT ) indicating consumption values from previous T time
steps, it computes output y at the next time step yT+1 by the following equations:
T
X
y j = f1 x(i ) wi + bi
i=1
X n
yT+1 = f2 y( j ) w j + b j ,
j=1
where wi denotes the input to hidden weight vector, w j denotes hidden to output weight vector, bi and
b j denote bias vectors, f1 refers to a non-linear hidden activation function while f2 refers to a linear
function and n is the number of neurons in the hidden layer.
Processes 2020, 8, x FOR PEER REVIEW 7 of 23
Figure
Figure 1.
1. Architecture
Architecture of
of applied
applied Artificial
Artificial Neural Network (ANN).
Neural Network (ANN).
Figure
Figure 2. Architecture of
2. Architecture of applied
applied Long
Long Short-Term
Short-Term Memory
Memory (LSTM).
(LSTM).
Processes
In 2020, 8, x FOR
an LSTM an input sequence = (x , . . . , x ), it computes an output as follows:
PEER REVIEW
network, given 8 of 23
In an LSTM network, given an input sequence = (𝑥11 , . . . , 𝑥𝑇T), it computes an output as follows:
+1 =
y𝑦t𝑡+1 =W
𝑊hy h + b 𝑦,,
ℎ𝑦 ℎt𝑡 + 𝑏y
where
where W 𝑊hy denotesthe
ℎ𝑦 denotes thehidden-output
hidden-output weight
weight matrix,
matrix, 𝑏𝑦 denotes
b y denotes bias vector
bias vector and htand ℎ𝑡 denotes
denotes the
the hidden
hidden
vector and is computed from the LSTM cell (block). A common LSTM block illustrated in Figurein
vector and is computed from the LSTM cell (block). A common LSTM block illustrated 3.
Figure
An LSTM 3. An LSTM
cell cell has
has three threean
gates: gates:
inputangate
inputtogate to identify
identify important
important information
information and preserve
and preserve it in
it in a long-term
a long-term memory
memory called
called the cell
the cell statestate 𝐶𝑡 , forget
Ct , an an forget
gategate to decide
to decide what
what information
information needs
needs to
to be
be forgotten from the previous cell sate
forgotten from the previous cell sate Ct−1 and𝐶 and an output gate to decide what to send to the
𝑡−1an output gate to decide what to send to the next sequence. next
sequence.
Once an
Once inputx𝑥t 𝑡enters
an input entersthe
theLSTM
LSTM cell, it isit passed
cell, through
is passed a logistic
through sigmoid
a logistic function
sigmoid σ and𝜎input
function and
gate i t [42]:
input gate 𝑖𝑡 [42]: h i
it = σ Wi ct−1 , h(t−1) , xt + bi .
𝑖𝑡 = 𝜎 (𝑊𝑖 [𝑐𝑡−1 , ℎ(𝑡−1) , 𝑥𝑡 ] + 𝑏𝑖 ).
Then the output of forget gate is computed as:
Then the output of forget gate is computed as:
h i
𝑓ft ==σ𝜎 (𝑊W f [𝑐ct−1,, ℎh(t−1) ,, 𝑥xt] +
(𝑡−1) + b𝑏f ).
.
𝑡 𝑓 𝑡−1 𝑡 𝑓
ℎ𝑡 = 𝑜𝑡 ⨂ 𝑇anh(𝑐𝑡 ).
Processes 2020, 8, 484 9 of 22
Here ⊗ denotes the element-vise multiplication of the vectors. The LSTM output ht at time step t
finally is computed as:
ht = ot ⊗ Tanh(ct ).
During the training process, the weight matrices Wi , W f , Wo and Wd and bias vectors bi , b f , bo
and bd are learned by an optimization algorithm.
For ANN and LSTM as AI-based networks, this transformation is highly recommended to prevent
computational problems and improve the functionality of training algorithms. While the choice of
scaling for the other two models depends on the problem and the scale of features. After testing the
performance of SVR and GBRT on the validation set, with and without data normalization we found
that SVR performs better without scaling while GBRT was not significantly affected, with and without
data normalization. Therefore, we decided to use the scaled data for all models except for the SVR.
The forecasts were rescaled to the initial scale after the prediction process. For the SVR, with RBF
kernel function we tuned the parameters of C, epsilon, and, gamma through a grid search approach.
We considered four candidate values for each parameter; 1,10,50,100 for C, 0.001, 0.01, 0.1, 0.2 for
gamma and 0.1, 0.2, 0.3, 0.4 for epsilon. In total, for each given scenario in the experiments, we tested
4 ∗ 4 ∗ 4 = 64 SVR models on the validation set.
For the GBRT model, one of the important parameters that need to be regularized is the number
of weak learners (trees). The development of a model with a large number of weak learners would
lead to lower regression error, but higher complexity and risk of overfitting. Furthermore, the speed of
learning which scales the contribution of each learner is another influential parameter that needs to be
set to reduce complexity and computation time. The lower learning rate would normally require fewer
learners, thus making the ensemble model simpler with higher generalization ability [40]. To determine
the optimum number of trees and the learning rate we employed a grid search strategy to compare the
generalization error under each combination of these parameters. According to that, for each variant
of the GBRT model, the number of trees was set among the range of 150,200 and 300 and the learning
rate was chosen among a range between 0.04 and 0.1.
The tree depth is also another parameter that requires tuning to avoid the overfitting problem.
The maximum number of variables for decision splits and the minimum number of records for leaf
nodes are the characterizing parameters for defining the tree depth. Generally, the tree depth has a
high impact on the overfitting problem, when decision trees are trained on a few observations with
a large number of attributes. For our problem, the training set has a large number of records and a
few numbers of features which reduces the chance of overfitting. Therefore, we set the maximum
number of features for the best split to the input size and set the leaf size as the default value of ‘one’
Processes 2020, 8, 484 10 of 22
considered in the SkLearn library. Nevertheless, to compute the optimal value for the maximum depth
of each tree we again used a grid search algorithm and set the candidate values in a range of 2 to 4.
Regarding the FFNN we considered a single hidden layer with n input nodes corresponding to
the number of input (lag) variables. To reduce the computation time, we did not tune the number of
nodes in the hidden layer, instead, we considered it as twice as the number of input nodes plus one as
discussed and suggested in [43]. However, we tuned two parameters related to the training process
by grid search: the optimization algorithm and weight initialization technique. We selected ‘Adam’,
‘NAdam’ and ‘RMSprop’ as the candidate values of optimization algorithm along with ‘Uniform’,
‘normal’, ’Golrot Normal’ as the candidates for weight initialization technique. The linear activation
function is used in the Dense layer and The Rectified Linear Units (ReLU) [44] function is used in the
hidden layer. A batch size of 128 training samples and a number of 70 epochs (iterations) were chosen
for the learning process of each variant. Finally, a learning ratio of 0.001 was set in each iteration for
the convergence.
For the LSTM, we did not tune various hyperparameters because of the high computational costs.
However, the number of LSTM units as one of the most important hyperparameters related to the
network structure was tuned using the validation dataset. The optimal value of this parameter for each
LSTM variant was chosen among the candidate values of 5, 10, 15 and 20. The number of features was
set to one and the number of timesteps was chosen as the number of lag variables. Adaptive Moment
Estimation (ADAM) function was used for optimization due to its computational efficiency and its
ability to optimize models with a large number of parameters [45]. The linear activation function is
used in the Dense layer before the output of all units and the ReLU activation function was used for
the recurrent step. ReLU function is monotonic and half rectified, which assigns zero to any negative
values. This has the advantage of not generating vanishing or exploding gradients. However, it can
cause dead neurons; therefore, we used the dropout layer between LSTM and Dense (output) layer
with the rate of 0.2 to reduce the negative effect of dead neurons which may hurt the training phase.
Since LSTM has stronger learning ability than a shallow neural network, higher batch size of 256 and
fewer number of epochs (50) was set for training the network.
5. Error Metrics
To evaluate the performance of a forecasting technique, forecasting error is calculated. The lower
the forecasting error, the higher the performance of the model. The forecasting error is the difference
between the actual observation and the predicted value. There are many error metrics that are
proposed for calculating the forecasting error and comparing the performance of time-series forecasting
techniques. In this study, we used three such metrics—Mean Absolute Error (MAE), Root Mean Square
Error (RMSE) and Mean Average Scaled Error (MASE).
If ŷt is the prediction value and yt is the actual value at time t and n is the number of test
observations, we can define the three metrics as the following:
n
1 X
MAE (t) = yt − ŷt
n
t=1
s
Pn
t=1 ( yt − ŷt 2 )
RMSE (t) =
n
n
1 X yt − ŷt
MASE(t) = .
N 1 n
P
T =1 (n−1) i=2 i
y − yi−1
The MAE calculates the magnitude of the errors on average and ignores whether the prediction
values are higher or lower than the real values. Thus, MAE gives equal importance (weight) to all
Processes 2020, 8, 484 11 of 22
individual differences. The RMSE on the other hand, penalizes large errors by calculating the squared
error before averaging them.
The MASE was proposed by [46] is introduced as a more applicable error metric and as an
alternative to some metrics like Mean Absolute Percentage Error (MAPE) when the observation or
prediction values are zero. The MAPE is commonly used as a loss function in model evaluation
because it can interpret the relative error. However, the problem with MAPE can occur when there
are zero values in the series and there will be a division by zero. For such sequences, MASE is
appropriate as it never produces infinite or unknown values. In this alternative, each actual value
of the series in the MAPE formula can be replaced by the average of all actual values of that series.
In addition to these metrics, we added another error metric to our evaluation which is particularly
defined for demand forecasting problems, applied in [47]—Daily Peak Mean Average Percentage Error
(DpMAPE). The DpMAPE measures how accurate is the model in forecasting daily peak consumption.
The information about peak time and peak consumption values is highly important for energy
management systems for saving grid costs through peak shaving services. The DpMAPE computes the
relative difference (percentage) between the daily peak consumption and predicted daily peak value
expressed by the following equation:
ymax − ŷmax
DpMAPE = ∗ 100 %.
ymax
Finally, since each forecasting model is tested under different scenarios and produces different
values for the defined measures, we need to have a combined metric to assess the best variant of each
model. This metric is also adopted from [47] and calculates a cumulative weighted error (CWE) based
on four defined metrics as follows:
The CWE is further used to compare the prediction performance of the best variants among
different predictive models.
deviation implying flat consumption were filtered out. Furthermore, the house profiles with unusual
total annual consumptions, over 20,000 KWh and less than 2000 KWh as well as total daily consumption
of fewer than 3 KWh for more than a month, were discarded. Finally, from the 180 remaining houses,
we randomly selected 15 houses from 5 acorn groups, and therefore a total of 75 house profiles were
picked for further study. In the final preprocessing step, linear interpolation was performed on the
house profiles containing small gaps from 1 to 24 h. For each building, the energy reading for the year
2013, due to the fewer number of missing records was chosen. Accordingly, the number of observations
in each dataset turned to 356 days ∗ 24 h = 8760 and the total number of observations in all 75 houses
turned to 75 ∗ 8760 = 657,000.
Figure 4 illustrates the energy readings of fifteen sample houses in the dataset belonging to
different acorn groups over one-year. As we can see they demonstrate different amounts of hourly
consumption (ranging between zero and five KW/h), as well as various consumption patterns over the
same year2020,
Processes (2013).
8, x FOR PEER REVIEW 12 of 23
Figure 4. Hourly energy consumption of sample houses in different groups over one year (2013).
Figure 4. Hourly energy consumption of sample houses in different groups over one year (2013).
It is obvious that the short-term load forecasting models are aimed at predicting accurate peak load
It is obvious that the short-term load forecasting models are aimed at predicting accurate peak
or energy consumption. However, one of the main influential factors in short-term load forecasting at
load or energy consumption. However, one of the main influential factors in short-term load
theforecasting
householdatlevel is the load volatility which simply means deviations from the average consumption.
the household level is the load volatility which simply means deviations from the
These
average consumption.frequently
deviations arise in a residential
These deviations house because
arise frequently its energy
in a residential consumption
house because itsisenergy
usually
influenced
consumption is usually influenced by various factors such as temperature, utilized appliances In
by various factors such as temperature, utilized appliances and consumption habits. andthe
context of load habits.
consumption forecasting,
In thehigher
contextload violation
of load increases
forecasting, theload
higher complexity
violationofincreases
the loadthe
profile, thereby
complexity
making
of the load profile, thereby making an accurate load forecasting more complicated. Load analysis ofof
an accurate load forecasting more complicated. Load analysis of the existing profiles in terms
load
thevolatility
existing will assess
profiles in in advance
terms which
of load house profiles
volatility are potentially
will assess in advancemorewhich challenging to forecast.
house profiles are
Figures 5more
potentially and challenging
6 provide more insight into the load volatility of the house profiles using box-plots
to forecast.
statistics [50]. The
Figures 5 andboxplots
6 provide provide
more information on the
insight into the loadmedian value
volatility andhouse
of the variability ofusing
profiles the consumption
box-plots
values. Figure
statistics 5 shows
[50]. how the hourly
The boxplots provide energy load of one
information on house changesvalue
the median duringand
different days ofofa week.
variability the
Forconsumption
example, this values.
houseFigure 5 shows how
has experienced highthe hourly energy
variations in hourly load of one house
consumption changes
during during
the weekend
anddifferent days of
on Thursday a week.
than Forweekdays.
the other example, this house the
Moreover, has median
experienced
valuehigh variations
of energy in hourly
consumption has
consumption during the weekend and on Thursday than the other weekdays.
been increased over the weekend. The bobbles in the plot indicate that a few consumption values areMoreover, the median
value of energy consumption has been increased over the weekend. The bobbles in the plot indicate
that a few consumption values are out of the maximum range; Q3 + 1.5 ∗ (Q3 − Q1) [50] which can
be considered as outliers. For instance, on Saturday, there have been recordings higher than 1 + 1.5 ∗
(1.0 − 0.3) = 2.05 KW/h showing by bobbles.
Processes 2020, 8, 484 13 of 22
out of the maximum range; Q3 + 1.5 ∗ (Q3 − Q1) [50] which can be considered as outliers. For instance,
on Saturday,
Processes 2020, 8,there
x FORhave
PEERbeen recordings higher than 1 + 1.5 ∗ (1.0 − 0.3) = 2.05 KW/h showing by bobbles.
REVIEW 13 of 23
Figure
Figure 5. Boxplot
5. Boxplot statistics
statistics forfor House
House 33 33 over
over oneone week.
week. (4 (4 March
March to to
11 11 March
March 2013).
2013).
Figure 6.
Figure Boxplot statistics
6. Boxplot statistics for
for 75
75 houses
houses in
in aa working
working day
day 2013
2013 March
March 25.
25.
Figure 6 similarly illustrates the hourly load volatility of houses in each customer group over
one working day. We can see that there are certain houses where their hourly consumption patterns
change very little, such as houses 3, 4 and 8 in group A; potentially easier to predict while there are
Processes 2020, 8, 484 14 of 22
Figure 6 similarly illustrates the hourly load volatility of houses in each customer group over
one working day. We can see that there are certain houses where their hourly consumption patterns
change very little, such as houses 3, 4 and 8 in group A; potentially easier to predict while there are
some houses (such as House 15 in group A or House 18 in group B) experiencing major changes in
their hourly load profiles which can make them difficult to forecast.
Table 1 provides summary statistics of house profiles per customer group for the whole test period.
The buildings in group A on average, have the highest mean electricity load over different time slots
(hour, day and week) with the highest deviations from the mean values (standard deviation values of
0.72, 6.5 and 34). Similarly, customers in group D consume high energy but with lower deviations.
On the other hand, the customers in groups B and C behave similarly and on average consume less
electricity over a year. However, the lowest average values and smallest volatilities are recorded for
the households in group E.
As mentioned earlier, for all models, the input variables were considered as the load lags from
As mentioned earlier, for all models, the input variables were considered as the load lags from
previous time steps. The number of lags that were tested in our experiments varies from 1 to 9
previous time steps. The number of lags that were tested in our experiments varies from 1 to 9 e.g.,
e.g., the load consumption for the previous 1 to 3 h and the load consumption at the previous 1 to
the load consumption for the previous 1 to 3 h and the load consumption at the previous 1 to 6 and 1
6 and 1 to 9 h. Regarding the size of the training set, four subsets of training data were considered
to 9 h. Regarding the size of the training set, four subsets of training data were considered for the
for the evaluation: 25%, 50%, 75% and 100% of the total train size represented by D1, D2, D3, and D4
evaluation: 25%, 50%, 75% and 100% of the total train size represented by D1, D2, D3, and D4
respectively. Figure 7 shows the evolution of the average prediction error on the validation set versus
respectively. Figure 7 shows the evolution of the average prediction error on the validation set versus
the number of input variables and the size of the training set for each of the four models studied.
the number of input variables and the size of the training set for each of the four models studied.
According
Accordingto tothe
thebar
barplots,
plots,thetheaverage
average prediction
predictionerrors of all
errors of models
all models decreased between
decreased 1% to1%
between
3% when the size of the training set increased from D1 to D3. However,
to 3% when the size of the training set increased from D1 to D3. However, a further increase in a further increase in thethe
training size has
training size has not
not affected
affected thetheaccuracy
accuracyof ofall
allmodels
modelssimilarly.
similarly.TheTheSVRSVRandandFFNN
FFNNhavehaveshown
shownthe
the
highest accuracy when they are trained on D3 dataset while GBRT and LSTM performed the bestbest
highest accuracy when they are trained on D3 dataset while GBRT and LSTM performed the with
with the largest
the largest size
size of of input
input data data
(D4).(D4).
Furthermore,
Furthermore,the theprediction
predictionerrorerrordoes
doesnotnotfollow
followa aclear
cleartrend
trendregarding
regardingthe thenumber
number ofofinput
input
lags.
lags. For the GBRT and FFNN, it is observed that with the increase of the input number from 69to
For the GBRT and FFNN, it is observed that with the increase of the input number from 6 to
the errors
9 the errorstend
tendtotorise.
rise.This
Thissuggests
suggests that
thatthe hourly
the hourlyconsumption
consumption pattern
pattern can bebe
can captured
captured byby thethe
smaller number of input lags. Thus, in some cases, more load lags can be discarded
smaller number of input lags. Thus, in some cases, more load lags can be discarded in order to reduce in order to reduce
the
the computation
computationcost costandandcomplexity
complexity ofof
thethe
model.
model.TheThe
likely reason
likely is that
reason although
is that the input
although to theto
the input
model (past consumption variables) is highly correlated with the target variable (one-hour ahead
the model (past consumption variables) is highly correlated with the target variable (one-hour ahead
consumption), they are also highly correlated with each other as they are consecutive lags. The
consumption), they are also highly correlated with each other as they are consecutive lags. The mutual
mutual dependence between consecutive lags indicates redundancy of information they convey to
dependence between consecutive lags indicates redundancy of information they convey to the model
the model which would not boost the learning ability, rather it could increase the training time.
which would not boost the learning ability, rather it could increase the training time.
However, for the LSTM on average more lag variables seem more informative to the model. For
However, for the LSTM on average more lag variables seem more informative to the model.
the SVR the forecasting error does not show a clear pattern concerning the input size.
For the SVR the forecasting error does not show a clear pattern concerning the input size.
Processes 2020, 8, 484 16 of 22
Training Time
Best Variant Training Size Parameters
(Minutes)
Kernel: RBF, C:10, Gamma: 0.001, Epsilon: 0.2,
SVR 33 houses 45
input: 6 load lags
Max_Depth: 2, Learning_rate: 0.06,
GBRT 45 houses 15 n_estimatores: 300,
n_features: 6, 6 load lags
Hidden layer: 1, Hidden_neurons: 13,
FFNN 33 houses 30 Optimizer: Adam
Weight_init_mode: Golrot Normal, 6 load lags
LSTM layer: 1, LSTM cells: 10, Activation
LSTM 45 houses 40 Function: ReLu
Dropout rate: 0.1, 6 load lags
The best variants are then tested on the test profiles. To understand the generalization ability of
each model to different profiles, we computed the average error metrics over 15 houses. To estimate
how much the error values, vary from the average, the corresponding standard deviation (SD) for each
error metric and model is further reported. The lower standard deviation values for a model indicates a
narrower range of errors and implicitly more robustness and consistency of the model. Table 3 reports
the average forecasting errors for one-hour ahead predictions. The reported CWE values prove that,
on average, the GBRT, FFNN and LSTM slightly outperform SVR in hourly load predictions.
The AI-based models compared to the CART algorithm (GBRT) obtain better performance
considering the average CWE (0.44 and 0.43 versus 0.45). However, GBRT and LSTM detect better the
daily peak with an average DpMAPE of 17.8 and 17.7 KW/h respectively. In general, all models scale
well and demonstrates robustness with average MAE of at most 0.24 KW/h and standard deviation of
at most 0.06.
Figure 8 illustrates how the models predicted the energy consumption of each test house over one
week during the spring season.
Processes 2020, 8, 484 17 of 22
Processes 2020, 8, x FOR PEER REVIEW 17 of 23
over one
Figure 8. Real consumption of test houses versus predictions over one week.
week.
Except
Except for
for the
the SVR
SVR which
which slightly
slightly overestimates
overestimates the
the real
real consumption values in
consumption values in most
most houses,
houses,
the other three models mainly demonstrate a good match and steady weekly prediction for
the other three models mainly demonstrate a good match and steady weekly prediction for the one-the one-hour
ahead estimation.
hour ahead FigureFigure
estimation. 9 provides more insights
9 provides into theinto
more insights variability or dispersion
the variability of error of
or dispersion metrics
error
which
metricswere averaged
which over all profiles
were averaged over all and compared
profiles among different
and compared ML models.
among different ML models.
The distribution of average DpMAPE errors proves that the median value of the peak prediction
errors of GBRT and LSTM is less than the ones in other techniques. This means that these models
adapt better to changes in daily peak consumption and achieve low perdition errors. The small size of
the boxplot for the GBRT algorithm even shows the least variability in the peak errors, thus higher
robustness. The ability of these models in obtaining high accuracy is also visible through boxplot
statistics of average MASE. The median values in the plots of RMSE and MAE indicate that for all the
Processes 2020, 8, 484 18 of 22
models, half of the errors are around or less than 0.4 and 0.25 KW/h respectively. However, these plots
are not 2020,
Processes informative enough
8, x FOR PEER for comparative analysis.
REVIEW 18 of 23
Figure 9. Comparison
Comparison of
of models
models based on boxplot of forecasting error statistics.
Another
The comparison
distribution is carried
of average DpMAPEout byerrors
considering
provesthe thatdiversity
the median of consumer
value of the groups. Figure 10
peak prediction
shows of
errors theGBRT
averageandmean
LSTM absolute
is less error
than thein five
ones distinct
in other customer
techniques.groups.ThisThe prediction
means that theseerrormodels
for the
house profiles belonging to group A was the highest, while in the other
adapt better to changes in daily peak consumption and achieve low perdition errors. The small size categories it reduced between
2%the
of and 15%. Groups
boxplot B and C
for the GBRT show theeven
algorithm lowestshowsvaluesthe for different
least models
variability outpeak
in the of all the other
errors, thusgroups.
higher
It also confirms
robustness. The the lower
ability of average prediction
these models error as FFNN,
in obtaining and LSTM
high accuracy were
is also used as
visible the prediction
through boxplot
algorithms.
statistics The superiority
of average MASE. The of these
median models
values in in
hourly prediction
the plots of RMSE is and
also MAE
visibleindicate
for the that
test for
houses in
all the
other customer
models, half of groups.
the errorsOverall, we canorconclude
are around less thanthat 0.4the
andforecasting
0.25 KW/htask in group AHowever,
respectively. and groupthese D is
more are
plots challenging due to high
not informative enoughaverage consumptionanalysis.
for comparative over a year as well as high hourly load variations.
The models
Anotherincomparison
groups B and C, on the
is carried outother hand, obtained
by considering higher accuracy
the diversity of consumer due groups.
to the lowerFigure load
10
volatility
shows theand lowermean
average average yearlyerror
absolute consumption of thecustomer
in five distinct house profiles.
groups. These customers error
The prediction have shown
for the
the most
house predictable
profiles profiles.
belonging The A
to group interesting findingwhile
was the highest, is thatinthetheconsumption
other categories behavior
it reducedof thebetween
houses
belonging
2% and 15%. to group
GroupsE Bwithandthe lowest
C show theaverage
lowest consumption
values for different and the lowestout
models daily variations
of all the otherwere still
groups.
difficult
It to predict.
also confirms theThe
lowerprobable
average reason is that error
prediction the majority
as FFNN, of houses
and LSTM in the training
were usedphase,
as theon average,
prediction
have experienced
algorithms. higher hourly
The superiority energy
of these modelsconsumption
in hourly than these test
prediction profiles.
is also visibleTherefore,
for the testthe trained
houses in
models
other could not
customer learn their
groups. consumption
Overall, we can concludebehavior thatproperly.
the forecasting task in group A and group D is
moreThe final analysis
challenging due towas
highperformed to evaluate the
average consumption overeffect
a year ofastemperature
well as highand seasonal
hourly events on
load variations.
the prediction accuracy of various techniques. Figure 11 shows the average
The models in groups B and C, on the other hand, obtained higher accuracy due to the lower load MAE of four techniques
over all test
volatility and houses
lowerin four consecutive
average seasons ofofthe
yearly consumption theYear
house2013. It is evident
profiles. from the chart
These customers havethe SVR
shown
produced the highest error than the other three models over the seasons,
the most predictable profiles. The interesting finding is that the consumption behavior of the houses though its estimation error
reached thetolowest
belonging group (around
E with the 0.07 KW/h)
lowest in the consumption
average summer. The and other thethree
lowestmodels
dailyalso predictwere
variations summer still
load withtothe
difficult highestThe
predict. accuracy.
probable The reason
forecasting error
is that theduring spring
majority of and autumn
houses increased
in the trainingforphase,
the most on
models by around 10%. In winter the SVR and GBRT had another 10%
average, have experienced higher hourly energy consumption than these test profiles. Therefore, the increase, while FFNN and
LSTM seemed
trained modelsmorecouldadaptable
not learn to seasonal
their change and
consumption their error
behavior remained unchanged.
properly.
Processes 2020, 8, 484 19 of 22
Processes 2020, 8, x FOR PEER REVIEW 19 of 23
Figure 10. Comparison of models based on average Mean Absolute Error (MAE) per customer group.
The final analysis was performed to evaluate the effect of temperature and seasonal events on
the prediction accuracy of various techniques. Figure 11 shows the average MAE of four techniques
over all test houses in four consecutive seasons of the Year 2013. It is evident from the chart the SVR
produced the highest error than the other three models over the seasons, though its estimation error
reached the lowest (around 0.07 KW/h) in the summer. The other three models also predict summer
load with the highest accuracy. The forecasting error during spring and autumn increased for the
most Figure
models Comparison
10.by of models
around 10%. basedthe
In winter on average
SVR and Mean
GBRTAbsolute Error (MAE)
had another 10%per customer
increase, group.
while FFNN
Figure 10. Comparison of models based on average Mean Absolute Error (MAE) per customer group.
and LSTM seemed more adaptable to seasonal change and their error remained unchanged.
The final analysis was performed to evaluate the effect of temperature and seasonal events on
the prediction accuracy of various techniques. Figure 11 shows the average MAE of four techniques
over all test houses in four consecutive seasons of the Year 2013. It is evident from the chart the SVR
produced the highest error than the other three models over the seasons, though its estimation error
reached the lowest (around 0.07 KW/h) in the summer. The other three models also predict summer
load with the highest accuracy. The forecasting error during spring and autumn increased for the
most models by around 10%. In winter the SVR and GBRT had another 10% increase, while FFNN
and LSTM seemed more adaptable to seasonal change and their error remained unchanged.
Figure 11.
Figure Comparison of
11. Comparison of models
models based
based on
on average
average MAE
MAE per
per season.
season.
8. Conclusions
8. Conclusions
This paper presents analysis and comparison of hour-ahead load forecasting over one-year period
This paper presents analysis and comparison of hour-ahead load forecasting over one-year
with four data-driven models known as SVR, GBRT, FFNN and LSTM. They were trained on historical
period with four data-driven models known as SVR, GBRT, FFNN and LSTM. They were trained on
load data provided by the UK residential smart meters. Their generalization ability was evaluated
historical load data provided by the UK residential smart meters. Their generalization ability was
on the house profiles which were not previously trained on. The test houses were chosen from five
evaluated on the house profiles which were not previously trained on. The test houses were chosen
customer groups with different levels of load volatility and average yearly consumption. The sensitivity
from five customer groups with different levels of load volatility and average yearly consumption.
of each algorithm was also tested to the number of training houses and the number of input lag
The sensitivity of each algorithm was alsooftested tobased
the number of training
MAE perhouses
season. and the number of
variables. The mainFigure
findings11. Comparison
are summarized models
as follows: on average
input lag variables. The main findings are summarized as follows:
1. Although the models were merely fed with consumption values as the predictors, in general,
8. Conclusions
1 Although the models were merely fed with consumption values as the predictors, in general,
they could provide stable one-hour ahead prediction over a one-year period.
they
This could
paperprovide
presentsstable one-hour
analysis ahead prediction
and comparison over a one-year
of hour-ahead period.
load forecasting over one-year
2. The AI methodologies; LSTM and FFNN compared to the other two techniques, adapted better to
2period
Thewith
AI methodologies;
four data-drivenLSTM modelsand FFNNascompared
known SVR, GBRT, to the other
FFNN two
and techniques,
LSTM. adapted
They were better
trained on
changes while performing predictions, following the trend of real consumption and on average,
to changes
historical while
load data performing
provided by thepredictions,
UK residential following
smart the trend
meters. of real
Their consumption
generalization andwas
ability on
achieving lower prediction errors.
average,
evaluated on achieving
the house lower
profiles prediction
which were errors.
not previously trained on. The test houses were chosen
3. With regard to daily peak load estimations of various profiles, the GBRT in addition to LSTM
3fromWith regard to groups
five customer daily peakwithload estimations
different levels of various profiles,
load volatility andtheaverage
GBRT in addition
yearly to LSTM
consumption.
outperformed other techniques.
outperformed
The sensitivity other
of each techniques.
algorithm was also tested to the number of training houses and the number of
4. As for the computation cost, the GBRT is the fastest algorithm to be trained and fine-tuned
input lag variables. The main findings are summarized as follows:
among the others. On the contrary, the training times of the SVR and LSTM are significantly high,
1 Although
especially the
whenmodels were merely
the training size forfed
thewith
SVR consumption values as
grows or the number of the predictors,
variables in general,
for tuning in the
they
LSTM could provide
network stable one-hour ahead prediction over a one-year period.
increases.
2
5. The AI methodologies;
Increasing the numberLSTM and FFNN
of training houses compared to the other
could improve two techniques,
the accuracy adapted
of forecasts better
as long as
to
thechanges while
additional performing
profile(s) predictions,
raise the following
model knowledge the the
about trend
testofprofile.
real consumption
This implies aand on
larger
average, achieving
training set does notlower prediction
necessarily boosterrors.
the model performance.
3 With regard to daily peak load estimations of various profiles, the GBRT in addition to LSTM
outperformed other techniques.
Processes 2020, 8, 484 20 of 22
6. Increasing the number of inputs does not have a similar effect on the performance of different
variants. Some models perform better with recent information about past consumption and some
need more knowledge on past consumption values.
7. A comparative study among the five customer groups shows that the customers with the
lower average amount of yearly consumption and smaller hourly load volatility generate more
predictable profiles.
8. An analysis of seasonal predictions reveals that the seasons with lower temperatures usually
come with more load violations, thus making forecasting more difficult for almost all models.
Future lines of research in short-term load forecasting at the individual building level aim to
customize forecasting techniques for the consumer groups with large variations in their consumption
patterns. performance evaluation of the studied techniques on the dataset with higher resolution and
for longer forecasting horizons can also be another research direction for the future work.
Author Contributions: A.M.P., designed the study, collected the data and carried out the experiments; A.M.P.,
and M.F., contributed to the the analysis of the results and the writing of the manuscript. A.C. and C.R. supervised
the research. All authors have read and agreed to the published version of the manuscript.
Funding: The project was funded by the Research Council of Norway under project number 267967.
Conflicts of Interest: The authors declare no conflicts of interest.
References
1. Smart Meters, Quarterly Report to end December 2016; Department for Business Energy and Industrial Strategy:
London, UK, 2017.
2. How Many Smart Meters are Installed in the United States, and Who Has Them? Available online:
https://www.eia.gov/tools/faqs/faq.php?id=108&t=3 (accessed on 28 January 2020).
3. Wang, Y.; Chen, Q.; Hong, T.; Kang, C. Review of Smart Meter Data Analytics: Applications, Methodologies,
and Challenges. IEEE Trans. Smart Grid 2019, 10, 3125–3148. [CrossRef]
4. Farmanbar, M.; Parham, K.; Arild, Ø.; Rong, C. A Widespread Review of Smart Grids Towards Smart Cities.
Energies 2019, 12, 4484. [CrossRef]
5. Ryu, S.; Noh, J.; Kim, H. Deep Neural Network Based Demand Side Short Term Load Forecasting. Energies
2016, 10, 3. [CrossRef]
6. Mocanu, E.; Nguyen, P.H.; Gibescu, M.; Kling, W.L. Comparison of machine learning methods for estimating
energy consumption in buildings. In Proceedings of the 2014 International Conference on Probabilistic
Methods Applied to Power Systems (PMAPS), Durham, UK, 7–10 July 2014; pp. 1–6. [CrossRef]
7. Pirbazari, A.M.; Chakravorty, A.; Rong, C. Evaluating Feature Selection Methods for Short-Term Load
Forecasting. In Proceedings of the 2019 IEEE International Conference on Big Data and Smart Computing
(BigComp), Kyoto, Japan, 27 Feburuary–2 March 2019; pp. 1–8. [CrossRef]
8. Zhao, H.; Magoulès, F. A review on the prediction of building energy consumption. Renew. Sustain. Energy
Rev. 2012, 16, 3586–3592. [CrossRef]
9. Pérez-Lombard, L.; Ortiz, J.; Pout, C. A review on buildings energy consumption information. Energy Build.
2008, 40, 394–398. [CrossRef]
10. Hyndman, R.; Koehler, A.; Ord, K.; Snyder, R. Forecasting with Exponential Smoothing; Springer: Berlin/Heidelberg,
Germany, 2008.
11. Ahmad, A.S.; Hassan, M.Y.; Abdullah, M.P.; Rahman, H.A.; Hussin, F.; Abdullah, H.; Saidur, R. A review
on applications of ANN and SVM for building electrical energy consumption forecasting. Renew. Sustain.
Energy Rev. 2014, 33, 102–109. [CrossRef]
12. Khashei, M.; Bijari, M. An artificial neural network (p,d,q) model for timeseries forecasting. Expert Syst. Appl.
2010, 37, 479–489. [CrossRef]
13. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [CrossRef]
14. Sutskever, I.; Vinyals, O.; Le, Q. Sequence to Sequence Learning with Neural Networks. Adv. Neural Inf.
Process. Syst. 2014, 4, 3104–3112.
Processes 2020, 8, 484 21 of 22
15. Quilumba, F.L.; Lee, W.-J.; Huang, H.; Wang, D.Y.; Szabados, R.L. Using Smart Meter Data to Improve the
Accuracy of Intraday Load Forecasting Considering Customer Behavior Similarities. IEEE Trans. Smart Grid
2015, 6, 911–918. [CrossRef]
16. Dahl, M.; Brun, A.; Kirsebom, O.; Andresen, G. Improving Short-Term Heat Load Forecasts with Calendar
and Holiday Data. Energies 2018, 11, 1678. [CrossRef]
17. Kong, W.; Dong, Z.Y.; Hill, D.J.; Luo, F.; Xu, Y. Short-Term Residential Load Forecasting Based on Resident
Behaviour Learning. IEEE Trans. Power Syst. 2018, 33, 1087–1088. [CrossRef]
18. Fumo, N.; Biswas, M.A.R. Regression analysis for prediction of residential energy consumption. Renew. Sustain.
Energy Rev. 2015, 47, 332–343. [CrossRef]
19. Pappas, S.S.; Ekonomou, L.; Karamousantas, D.C.; Chatzarakis, G.E.; Katsikas, S.K.; Liatsis, P. Electricity
demand loads modeling using AutoRegressive Moving Average (ARMA) models. Energy 2008, 33, 1353–1360.
[CrossRef]
20. Lü, X.; Lu, T.; Kibert, C.J.; Viljanen, M. Modeling and forecasting energy consumption for heterogeneous
buildings using a physical–statistical approach. Appl. Energy 2015, 144, 261–275. [CrossRef]
21. Janacek, G. Time Series Analysis Forecasting and Control. J. Time Ser. Anal. 2009. [CrossRef]
22. Thissen, U.; van Brakel, R.; de Weijer, A.P.; Melssen, W.J.; Buydens, L.M.C. Using support vector machines
for time series prediction. Chemom. Intell. Lab. Syst. 2003, 69, 35–49. [CrossRef]
23. Lahouar, A.; Slama, J.B.H. Day-ahead load forecast using random forest and expert input selection.
Energy Convers. Manag. 2015, 103, 1040–1051. [CrossRef]
24. Hambali, A.O.; Akinyemi, M.; JYusuf, N. Electric Power Load Forecast Using Decision Tree Algorithms.
Comput. Inf. Syst. Dev. Inform. Allied Res. J. 2017, 7, 29–42.
25. Ahmed, N.K.; Atiya, A.F.; Gayar, N.E.; El-Shishiny, H. An Empirical Comparison of Machine Learning
Models for Time Series Forecasting. Econom. Rev. 2010, 29, 594–621. [CrossRef]
26. Khosravani, H.; Castilla, M.; Berenguel, M.; Ruano, A.; Ferreira, P. A Comparison of Energy Consumption
Prediction Models Based on Neural Networks of a Bioclimatic Building. Energies 2016, 9, 57. [CrossRef]
27. Yang, K.; Zhao, L. Application of Mamdani Fuzzy System Amendment on Load Forecasting Model.
In Proceedings of the 2009 Symposium on Photonics and Optoelectronics, Wuhan, China, 14–16 August
2009; pp. 1–4. [CrossRef]
28. Chaturvedi, D.; Kumar, R.; Kalra, P.K. Artificial neural network learning using improved genetic algorithm.
J. Inst. Eng. India 2002, 82, 1–8.
29. Baliyan, A.; Gaurav, K.; Mishra, S.K. A Review of Short Term Load Forecasting using Artificial Neural
Network Models. Procedia Comput. Sci. 2015, 48, 121–125. [CrossRef]
30. Marino, D.; Amarasinghe, K.; Manic, M. Building Energy Load Forecasting using Deep Neural Networks.
In Proceedings of the IECON 2016-42nd Annual Conference of the IEEE Industrial Electronics Society,
Florence, Italy, 23–26 October 2016.
31. Kong, W.; Dong, Z.Y.; Jia, Y.; Hill, D.J.; Xu, Y.; Zhang, Y. Short-Term Residential Load Forecasting Based on
LSTM Recurrent Neural Network. IEEE Trans. Smart Grid 2019, 10, 841–851. [CrossRef]
32. Agrawal, R.; Muchahary, F.; Tripathi, M. Long Term Load Forecasting with Hourly Predictions Based on
Long-Short-Term-Memory Networks. In Proceedings of the 2018 IEEE Texas Power and Energy Conference
(TPEC), College Station, TX, USA, 8–9 February 2018.
33. Kim, T.-Y.; Cho, S.-B. Predicting residential energy consumption using CNN-LSTM neural networks. Energy
2019, 182, 72–81. [CrossRef]
34. Mamun, A.; Hoq, M.; Hossain, E.; Bayindir, R. A Hybrid Deep Learning Model with Evolutionary Algorithm
for Short-Term Load Forecasting. In Proceedings of the 2019 8th International Conference on Renewable
Energy Research and Applications (ICRERA), Brasov, Romania, 3–6 November 2019.
35. Bouktif, S.; Fiaz, A.; Ouni, A.; Serhani, M. Optimal Deep Learning LSTM Model for Electric Load Forecasting
using Feature Selection and Genetic Algorithm: Comparison with Machine Learning Approaches. Energies
2018, 11, 1636. [CrossRef]
36. Cristianini, N.; Shawe-Taylor, J. An Introduction to Support Vector Machines and Other Kernel-based Learning
Methods, 1st ed.; Cambridge University Press: New York, NY, USA, 2000.
37. Chang, M.; Chen, B.; Lin, C.-J. EUNITE Network Competition: Electricity Load Forecasting, Dec. 2001.
Available online: http://www.eunite.org/knowledge/Competitions/1st_competition/Scientific_comments_to_
the_competition/Scientific_comments_to_the_competition.htm (accessed on 14 March 2020).
Processes 2020, 8, 484 22 of 22
38. Friedman, J. Stochastic Gradient Boosting. Comput. Stat. Data Anal. 2002, 38, 367–378. [CrossRef]
39. Ruiz-Abellón, M.; Gabaldón, A.; Guillamón, A. Load Forecasting for a Campus University Using Ensemble
Methods Based on Regression Trees. Energies 2018, 11, 2038. [CrossRef]
40. Feng, C.; Cui, M.; Hodge, B.-M.; Zhang, J. A data-driven multi-model methodology with deep feature
selection for short-term wind forecasting. Appl. Energy 2017, 190, 1245–1257. [CrossRef]
41. Géron, A. Hands on Machine Learning with Scikit-Learn and Tensorflow; O’Reilly Media: Sevastopol, CA, USA,
2019; p. 261.
42. Gers, F.A. Learning to forget: Continual prediction with LSTM. In Proceedings of the 9th International Conference
on Artificial Neural Networks: ICANN ’99, Edinburgh, UK, 6 August 1999; Volume 1999, pp. 850–855. [CrossRef]
43. Lippmann, R. An introduction to computing with neural nets. IEEE Assp Mag. 1987, 4, 4–22. [CrossRef]
44. Nair, V.; Hinton, G. Rectified Linear Units Improve Restricted Boltzmann Machines Vinod Nair, vol. 27. 2010.
Available online: https://www.cs.toronto.edu/~{}fritz/absps/reluICML.pdf (accessed on 14 March 2020).
45. Types of Optimization Algorithms used in Neural Networks and Ways to Optimize Gradient
Descent. Available online: https://towardsdatascience.com/types-of-optimization-algorithms-used-in-
neural-networks-and-ways-to-optimize-gradient-95ae5d39529f (accessed on 21 February 2020).
46. Hyndman, R.J.; Koehler, A.B. Another look at measures of forecast accuracy. Int. J. Forecast. 2006, 22, 679–688.
[CrossRef]
47. Asare-Bediako, B.; Kling, W.L.; Ribeiro, P.F. Day-ahead residential load forecasting with artificial neural
networks using smart meter data. In Proceedings of the 2013 IEEE Grenoble Conference, Grenoble, France,
16–20 June 2013; pp. 1–6. [CrossRef]
48. SmartMeter Energy Consumption Data in London Households—London Datastore. Available online: https:
//data.london.gov.uk/dataset/smartmeter-energy-use-data-in-london-households (accessed on 10 February 2020).
49. Acron. Available online: https://acorn.caci.co.uk/downloads/Acorn-User-guide.pdf (accessed on 23 March 2013).
50. Vapnik, V.N. Statistical Learning Theory; John Wiley & Sons, Inc.: New York, NY, USA, 1998.
51. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.;
Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830.
52. Keras: Deep Learning library for Theano and TensorFlow—Data Science Central. Available online: https:
//www.datasciencecentral.com/profiles/blogs/keras-deep-learning-library-for-theano-and-tensorflow
(accessed on 21 February 2020).
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access
article distributed under the terms and conditions of the Creative Commons Attribution
(CC BY) license (http://creativecommons.org/licenses/by/4.0/).