Building Energy Models at Different Time Scales Based On Multi-Output Machine Learning
Building Energy Models at Different Time Scales Based On Multi-Output Machine Learning
Article
Building Energy Models at Different Time Scales Based on
Multi-Output Machine Learning
Guangchen Li 1 , Wei Tian 1,2, * , Hu Zhang 1 and Bo Chen 3
1 College of Mechanical Engineering, Tianjin University of Science and Technology, Tianjin 300222, China
2 Tianjin International Joint Research and Development Center of Low-Carbon Green Process Equipment,
Tianjin 300222, China
3 College of Economics and Management, Tianjin University of Science and Technology, Tianjin 300457, China
* Correspondence: [email protected]; Tel.: +86-(022)-6060-0705
Abstract: Machine learning techniques are widely applied in the field of building energy analysis
to provide accurate energy models. The majority of previous studies, however, apply single-output
machine learning algorithms to predict building energy use. Single-output models are unable to
concurrently predict different time scales or various types of energy use. Therefore, this paper
investigates the performance of multi-output energy models at three time scales (daily, monthly,
and annual) using the Bayesian adaptive spline surface (BASS) and deep neural network (DNN)
algorithms. The results indicate that the multi-output models based on the BASS approach combined
with the principal component analysis can simultaneously predict accurate energy use at three time
scales. The energy predictions also have the same or similar correlation structure as the energy data
from the engineering-based EnergyPlus models. Moreover, the results from the multi-time scale
BASS models have consistent accumulative features, which means energy use at a larger time scale
equals the summation of energy use at a smaller time scale. The multi-output models at various time
Citation: Li, G.; Tian, W.; Zhang, H.; scales for building energy prediction developed in this research can be used in uncertainty analysis,
Chen, B. Building Energy Models at sensitivity analysis, and calibration of building energy models.
Different Time Scales Based on
Multi-Output Machine Learning. Keywords: building energy prediction; machine learning; multi-output model; time scales
Buildings 2022, 12, 2109. https://
doi.org/10.3390/buildings12122109
consumption. The results show that the coefficient of variation of the root mean square
error (CV(RMSE)) of the seven models is less than 10%, in which the lightGBM model has
the best performance with the minimum CV(RMSE) of 4.1%. Jetcheva et al. [12] used the
neural network-based ensemble model method for daily power consumption prediction,
and compared it with the seasonal autoregressive integrated moving average (SARIMA)
model; the results indicate that the neural network model had better accuracy. Ferrantelli
et al. [13] defined an analytical bottom-up model for predicting any daily consumption
given a benchmark daily profile by comparing regression curves obtained with a frequentist
inference to Bayesian inference.
Secondly, the application of a single-output machine learning model in predicting
the monthly energy use of a building is described. Wang et al. [14] proposed a method by
combining a network model with a long-short-term memory learning model to predict
monthly building energy use. Compared with traditional ANN and SVM, the performance
of this model was better; the mean absolute error and root mean square error were 6.66%
and 0.36 kWh/m, respectively. Tran et al. [15] proposed an ensemble model named the
evolutionary neural machine inference model (ENMIM), which consisted of least squares
support vector regression (LSSVR) and the radial basis function neural network (RBFNN).
The model was used to predict the monthly cooling/heating loads and achieved excellent
results. Tian et al. [16] adopted a full linear model to predict monthly power and heating
energy use respectively, and obtained excellent results, providing a reliable statistical
energy model for model calibration. Zhu et al. [17] used five machine learning models
including multivariate adaptive regression splines (MARS) to predict the monthly power
consumption of buildings, which provided a basis for estimating unknown parameters
based on the approximate Bayesian calibration. Koschwitz et al. [18] proposed a novel
recurrent neural network method to predict the monthly thermal loads of buildings, and
the results show that the method is more accurate than support vector machines. Jahani
et al. [19] used the Genetic algorithm-based numerical moment matching (GA-NMM)
method to predict the monthly electricity consumption of buildings. Lin et al. [20] used RF,
SVM, and ANN to predict the monthly electricity consumption of buildings.
Thirdly, the application of a single-output machine learning model to predict the
annual energy use of buildings is presented. Olu-Ajayi et al. [21] used ANN, gradient
boosting (GB), deep neural network (DNN), RF, stacking, K-nearest neighbor (KNN), SVM,
decision tree (DT), and linear regression (LR) methods to predict the annual energy use of
buildings, and the DNN model has the highest prediction accuracy. Tian et al. [22] used
10 machine learning methods to predict the annual energy use of buildings in London, and
conducted variable importance analysis of the learning model and local spatial analysis of
model differences. Tian et al. [23] used five machine learning methods to accurately predict
the annual energy use of buildings for Dempster-Shafer theory uncertainty analysis and
global sensitivity analysis. Building annual energy use forecasts can also use a combination
of machine learning methods and statistical methods to further explore the linear or
nonlinear relationship between annual energy use and certain factors, including building
geometry, occupant, and climate-related factors [24]. A novel multiple linear regression
model can predict building energy use based on changes in climate parameters and the
intensity of the urban heat island effect, which leads to the conclusion that the urban heat
island effect can reduce building energy use in heating-dominated cities [25]. In summary,
energy use at three different time scales (day, month, and year) can be accurately predicted
using different single-output machine learning methods.
There are very limited studies to explore the multi-output models of building energy
and most previous studies use single-output machine learning models. Luo et al. [26]
used heating, cooling, lighting load, and building integrated photovoltaic (BIPV) electrical
power as the output for prediction, and the multi-output model selected SVM and ANN.
The results showed that the ANN model obtained higher accuracy and the SVM model
took less computation time. Liu et al. [27] proposed a multi-input multi-output (MIMO)
strategy based on recurrent neural networks for hourly building energy use prediction. The
Buildings 2022, 12, 2109 3 of 30
absolute percentage error of each time step was analyzed to explore the building’s energy-
saving performance, thus establishing an energy quantification system for the building.
Li et al. [28] developed a MIMO strategy by returning data information in a single-time step,
which avoids accumulated errors compared to recursive strategies and direct strategies.
These previous studies provide useful information on the characteristics of multi-
output models in building energy analysis. However, there is still a lack of sufficient studies
to explore the predictive performance of multi-output machine learning models in building
energy analysis, especially focused on output correlation. Moreover, the accumulative
features are not explored yet in creating multi-output models at various time scales (such
as daily, monthly, and annual).
In order to address these research gaps, this paper is focused on the construction
and performance analysis of multi-output machine learning models for building energy
assessment. Two multi-output machine learning methods are Bayesian adaptive spline
surface (BASS) and deep neural network (DNN). The performance of multi-output models
is evaluated from three aspects: computational time, model accuracy, and output correlation.
The main new contributions of this study are:
(1) This study compares the predictive performance of single-output and multi-output
learning models in building energy analysis. This would provide guidelines on how
to choose the single-output and multi-output models in creating machine learning
models for building energy assessment.
(2) This study explores the performance of two multi-output models (BASS and DNN) in
which the main difference of the two models is whether to maintain output correlation.
This would provide the guidelines on how to choose the learning models with or
without considering output correlation.
(3) The additive or accumulative features are investigated in creating various time scale
models for building energy analysis. This would provide insight on the methods of
obtaining building energy use from a smaller time scale to a larger time scale.
The remaining part of this paper is organized as follows. Section 2 introduces the
research methods, including data acquisition methods, modeling techniques, and perfor-
mance evaluation metrics. In Section 3, the final tuning hyper-parameters from the machine
learning models are first presented. Then, the results and discussion of the multi-output
machine learning model for predicting cooling and heating energy are presented in turn.
2. Method
The research framework based on multi-output machine learning models at different
time scales is shown in Figure 1. The analysis procedure can be divided into four steps: data
preparation, multi-output models, model performance evaluation, and guidelines. The
first step is to prepare the data required to establish the multi-output models as described
in Section 2.1. The second step presents the construction of the two multi-output models–
Bayesian adaptive spline surfaces (BASS) and deep neural networks (DNN). The third step
is model performance evaluation. The fourth step is to provide guidelines for applying the
multi-output models in building energy analysis. All the above steps are implemented in
R language.
The building used in this research is located in Tianjin, China. Therefore, the meteorological
Buildings 2022, 12, x FOR PEER REVIEW
data of
Tianjin in the Chinese standard weather data (CSWD) is used for the calculation 4 of
of 33
the building energy model.
EWU,RU, ,HSP,CSP
Step 1 Meteorological HVAC Envelope
Data preparation parameters Parameters parameters
LHS
n1 n2 n2000
Basic building
energy model
Standard building
energy model
MO-format
data set
Step 3
Performance evaluation CV(RMSE),MAPE,R² Computational time Dendrogram
Correlation
Model accuracy Model efficiency
between outputs
Model choice
Figure
Buildings 2022, 12, x FOR PEER REVIEW
Figure 1. flow
1. A A flow chart
chart of multi-output
of multi-output models
models of building
of building energyenergy at monthly,
at daily, daily, monthly, and5scales
and annual annual
of 33
scales
based onbased on machine
machine learning.learning.
include light power density and equipment power density. Heating and cooling set-point
temperatures are also considered in this study. The construction of the modeling data set is
divided into two steps. The first step is to perform Latin hypercube sampling within the
parameter ranges to obtain 2000 sets of parameter combinations and then generate 2000 idf
(EnergyPlus Input Files) format files. The 2000 sets of energy use data at different time
scales are obtained by running the EnergyPlus (V22.1.0) program, in which 1000 sets of
data are used for model training, 500 sets are used for model validation, and 500 sets are
used for model testing. The second step is data preprocessing to create the daily, monthly,
and annual energy models. The cooling time periods for this office building is from May 1st
to September 30th, and the heating time period is from November 1st to March 31st. The
heating and cooling energy on weekends and holidays in this office building is zero and
thus excluded from this analysis. As a result, the numbers of daily and monthly cooling
energy data are 105 and 5, respectively. The numbers of daily and monthly heating energy
data are 102 and 5, respectively.
where r represents the Pearson correlation coefficient, ai represents the ith sample a, bi
represents the ith sample b, a represents the average of sample a, and b represents the
average of sample b. The PCC value is between −1 and 1. The PCC is close to 1 to indicate
that there is a strong positive correlation. In contrast, a strong negative correlation can be
suggested by the PCC value close to −1.
Two types of machine learning methods– BASS and DNN–are chosen based on these
capability of multi-output models.
The BASS method is similar to Bayesian multivariate adaptive regression splines
used to learn a set of basis functions from historical data, which are tensor products of
polynomial splines. The BASS model can adaptively select the number of basis functions
and adjust the number of variables and nodes in each basis. While most hyper-parameters
of the BASS model are determined automatically, the priors in the BASS model can be
modified to reduce overfitting. The main adjusted parameters include the number of
principal component analysis n.pc and the number of iterations nmcmc. The principal
component analysis (PCA) method is used to calculate the principal components of multiple
outputs and then fit a regression model. The R-BASS package (V1.2.2) is used to implement
Buildings 2022, 12, 2109 6 of 30
the multi-output machine learning technique in this research. The advantages of using this
BASS package include scalable capabilities to handle large numbers of observations and
predictions. In addition, not only regression models but also classification models can be
created using the BASS package [32].
The DNN method is one of the current mainstream methods for analyzing building
energy, including an input layer, multiple hidden layers, and an output layer. The deep
feedforward neural network is selected in this research, and neuron signals are transmitted
forwardly between layers without feedback signal transmission. In other words, each
layer can only receive the signal of the previous layer, and then pass it to the next layer.
The signal transfer between layers depends on the connection of nonlinear functions.
By determining the weight and bias terms of the function, the signal of the previous
layer is effectively transmitted to the next layer. It is worth noting that there is no signal
transmission between neurons in the same layer of a feedforward neural network [33]. The
main hyper-parameters of the feedforward neural network include the number of hidden
layers, the number of hidden layer neurons, the loss function, and the number of iterations.
The DNN parameter adjustment method generally adopts empirical adjustment. The keras
and tensorflow packages in the R environment are used to implement the establishment of
DNN models.
1 n yi − ŷi
n ∑ i =1
MAPE = × 100% (3)
yi
n (yi − ŷi )2
R2 = 1 − ∑i=1 (4)
( y i − y i )2
where n represents the sample size, i represents the ith sample, yi represents the ith true
value, and ŷi represents the ith predicted value. The CV(RMSE), MAPE, and R2 values are
all dimensionless numbers, which are not affected by the order of magnitude of the data
and can more effectively express the accuracy of the model. CV(RMSE) and MAPE reflect
the error of the model–the smaller, the better. The value of R2 ranges from 0 to 1, which
reflects the relative error of the model relative to the direct average value. The closer to 1,
the better.
The computational time of creating machine learning models is also an important
indicator to assess the performance of learning algorithms.. The computing time includes
the modeling time and the prediction time when using the machine learning techniques in
building energy analysis. The computing time is counted using the system.time function in
the R language environment. Simulations were performed on a workstation with an Intel
Xeon CPU (E5-2650 2.3 GHz) and 64 GB RAM.
For the multi-output learning models, it is necessary to investigate whether the ma-
chine learning model maintains the correlation structure between the outputs of the energy
data from the training set. In this research, the cluster dendrogram method is used for
model output correlation analysis [34], which can hierarchically cluster multidimensional
data to aggregate highly similar data into one category. Through this dendrogram, the
clusters and the number of objects belonging to each cluster can be determined [35]. The
dendrogram is plotted by computing the distances between clusters according to the Lance–
Williams dissimilarity update formula by obtaining the dissimilarities prior to forming
Buildings 2022, 12, 2109 7 of 30
the new cluster. The complete linkage method, one of hierarchical classification methods
that uses the distance between the most distant elements from each cluster, is used for
hierarchical clustering [36]. The cluster analysis in this research is implemented using the
R package tidyverse and hclust functions.
Figure3.3.Correlation
Figure Correlationcoefficient
coefficientofofdaily
dailycooling
coolingenergy
energyuse
useininJuly
July(1,
(1,1 1July;
July;. .…
. ; ;31,
31,31
31July).
July).
Figure
Figure44shows
showsthe thepredicted
predictedperformance
performanceof ofthe
themulti-output
multi-outputdailydailyenergy
energycooling
cooling
model. Figure 4a demonstrates that the four models of SO-BASS,
model. Figure 4a demonstrates that the four models of SO-BASS, MO-BASS, MO-BASS, SO-DNN, and
SO-DNN,
MO-DNN
and MO-DNN have have
median CV(RMSE)
median CV(RMSE) of 0.008, 0.006,0.006,
of 0.008, 0.008,0.008,
and 0.015, respectively.
and 0.015, For the
respectively. For
day with the worst prediction accuracy of the four models (SO-BASS,
the day with the worst prediction accuracy of the four models (SO-BASS, MO-BASS, SO- MO-BASS, SO-DNN,
and
DNN,MO-DNN), the largest
and MO-DNN), theCV(RMSE) values are
largest CV(RMSE) 0.059,are
values 0.027, 0.035,
0.059, and0.035,
0.027, 0.038,and
respectively.
0.038, re-
The median and maximum values of the CV(RMSE) from the four models indicate that
spectively. The median and maximum values of the CV(RMSE) from the four models in-
the MO-BASS model performs the best in this case study. Figure 4b shows that the MO-
dicate that the MO-BASS model performs the best in this case study. Figure 4b shows that
BASS model has the smallest median MAPE, followed by SO-BASS and SO-DNN. The
the MO-BASS model has the smallest median MAPE, followed by SO-BASS and SO-DNN.
MO-BASS model has better predictive performance due to the small variations of MAPE
The MO-BASS model has better predictive performance due to the small variations of
in comparison with the SO-BASS although the median MAPE values of the two BASS
MAPE in comparison with the SO-BASS although the median MAPE values of the two
models are similar. This is because the MO-BASS model considers the output correlation
BASS models are similar. This is because the MO-BASS model considers the output cor-
to avoid the extreme errors from the single-output models. Hence, the MO-BASS model
relation to avoid the extreme errors from the single-output models. Hence, the MO-BASS
has the best performance in terms of the MAPE. Figure 4c indicates that the coefficients of
model has the best performance in terms of the MAPE. Figure 4c indicates that the coeffi-
determination (R2 ) are very2 high, except for three data points in SO-BASS. Most of the R2
cients of determination (R ) are very high, except for three data points in SO-BASS. Most
values are l greater than 0.9 to indicate that all these four models have good performance.
of the R2 values are l greater than 0.9 to indicate that all these four models have good
Among them, the two models with the largest R2 are MO-BASS and SO-DNN. Further
performance.
comparison of Among
the median them,andthe two modelsrange
interquartile withshows
the largest R2 R
that the are MO-BASS
2 from and SO-
the MO-BASS
DNN.isFurther
model comparison
larger than of the median
that of SO-DNN. Fromand interquartile
the above analysis,range shows thatmodel
the MO-BASS the R²isfrom
the
the MO-BASS model is larger than that
best daily cooling energy model in this case study. of SO-DNN. From the above analysis, the MO-
BASS model
Table is the
3 lists the best daily cooling
computational timeenergy model in this
for single-output andcase study.
multi-output models of cool-
ing energy for BASS and DNN algorithms at different time scales models. The comparison
of computational time for daily cooing energy use will be described in this section. The
discussion of computational time for monthly and multi-time scales would be presented in
Sections 3.2.2 and 3.2.3, respectively. The computation time for the multi-output models is
less than the computation time for the single-output models for daily cooling energy use.
For the BASS models, the modeling time of the multi-output BASS model for daily cooling
energy is 19.6 times that of the single-output BASS model. This is due to the fact that the
output number in the BASS multi-output model would be same as the principal component
the day with the worst prediction accuracy of the four models (SO-BASS, MO-BASS, SO-
DNN, and MO-DNN), the largest CV(RMSE) values are 0.059, 0.027, 0.035, and 0.038, re-
spectively. The median and maximum values of the CV(RMSE) from the four models in-
dicate that the MO-BASS model performs the best in this case study. Figure 4b shows that
the MO-BASS model has the smallest median MAPE, followed by SO-BASS and SO-DNN.
Buildings 2022,The MO-BASS model has better predictive performance due to the small variations of
12, 2109 9 of 30
MAPE in comparison with the SO-BASS although the median MAPE values of the two
BASS models are similar. This is because the MO-BASS model considers the output cor-
relation to avoid the extreme errors from the single-output models. Hence, the MO-BASS
number, which is much less than the original correlated outputs in the single-output models.
model has the best performance in terms of the MAPE. Figure 4c indicates that the coeffi-
The modeling time of the multi-output DNN model is 27.7 times that of the single-output
cients of determination (R2) are very high, except for three data points in SO-BASS. Most
DNN model for daily cooling energy. Hence, adopting the multi-output models can signifi-
of the R2 values are l greater than 0.9 to indicate that all these four models have good
performance. Among them, the two models time
cantly reduce modeling with and increase
the largest R2the
aremodel’s
MO-BASS efficiency,
and SO-presuming the accuracy
of the model is acceptable. The DNN models require
DNN. Further comparison of the median and interquartile range shows that the R² more computational
from cost compared
with the BASS models. For the single-output models,
the MO-BASS model is larger than that of SO-DNN. From the above analysis, the MO- the computational time for the DNN
model
BASS model is the best is cooling
daily almost energy
two times thatinofthis
model thecase
computational
study. cost for the BASS model.
(a) CV(RMSE)
(b) MAPE
(c) R²
Figure 4. Prediction performance of machineperformance
Figure 4. Prediction learning models for dailylearning
of machine cooling energy
models(SO,
for single
daily cooling energy (SO, single
output; MO, multiple output).
output; MO, multiple output).
Table 3 lists the computational time for single-output and multi-output models of
cooling energy for BASS and DNN algorithms at different time scales models. The com-
parison of computational time for daily cooing energy use will be described in this section.
The discussion of computational time for monthly and multi-time scales would be pre-
sented in Sections 3.2.2 and 3.2.3, respectively. The computation time for the multi-output
Buildings2022,
Buildings 2022,12,
12,2109
x FOR PEER REVIEW 11 10
ofof3330
Table3.3.The
Table Thecomputational
computationaltime
time of
of different
different time-scale
time-scale cooling
cooling energy
energy models
models (unit:
(unit: s).
s).
Models
Models Daily Daily Monthly
Monthly Multi-Time
Multi-TimeScale
Scale
SO-BASS 2032.2 98.7 -
SO-BASS 2032.2 98.7 -
MO-BASS
MO-BASS 103.9 103.9 81.181.1 197.2
197.2
SO-DNN
SO-DNN 3807.2 3807.2 361.7
361.7 - -
MO-DNN
MO-DNN 137.3 137.3 72.272.2 236.3
236.3
Figure 55 shows
Figure shows the
the correlated
correlated structure
structure treetree diagram
diagramfrom fromthe thetraining
trainingset setand
andthethefour
four
machine learning models on the daily cooling energy in July. The
machine learning models on the daily cooling energy in July. The numbers in Figure 5 numbers in Figure 5 repre-
sent a specific
represent day in
a specific July
day inin which
July the weekends
in which the weekendsand the andholidays have been
the holidays haveexcluded since
been excluded
there is no cooling energy use. For example, 1 represents the first day of
since there is no cooling energy use. For example, 1 represents the first day of July. From July. From the top of
Figure 5a, there are two groups of data in which the left tree has 14 data
the top of Figure 5a, there are two groups of data in which the left tree has 14 data and and the right tree has
nineright
the data.tree
Thehas
group
ninecandata.
be further divided
The group canintobesub-groups based on
further divided intothe similarity ofbased
sub-groups cooling
on
the similarity of cooling energy use. The cooling energy in the 3rd and 22nd are theofmost
energy use. The cooling energy in the 3rd and 22nd are the most similar as the height the
link thatasjoins
similar thethem together
height of theislink
the smallest
that joins in them
Figuretogether
5a. By comparing Figure 5a
is the smallest in with
Figurethe5a.
other
By
four dendrograms, Figure 5a–c have the most similar clustering distributions.
comparing Figure 5a with the other four dendrograms, Figure 5a–c have the most similar Thus, the MO-
BASS model
clustering can properly maintain
distributions. Thus, thethe output correlation
MO-BASS model can of the originalmaintain
properly training set thefrom the
output
EnergyPlus models. Figure 5b,d,e show that there are significant differences
correlation of the original training set from the EnergyPlus models. Figure 5b,d,e show from the training
set. there
that Therefore, the MO-BASS
are significant model can
differences from bethe
more in lineset.
training with the training
Therefore, the data set from
MO-BASS the
model
engineering-based building energy models from the perspective of the
can be more in line with the training data set from the engineering-based building energy correlation structure
of building
models fromcooling energy use.of the correlation structure of building cooling energy use.
the perspective
(d) (e)
Figure5.5. Dendrogram
Figure Dendrogram of
of July
July daily
daily cooling
cooling energy
energy for
for the
the training
training set
setand
andfour
fourmachine
machinelearning
learning
models.
models.
Figure 6.
Figure 6. Correlation
Correlation coefficient
coefficient of
of monthly
monthlycooling
coolingenergy.
energy.
Figure
Figure 77 shows
shows thethe predicted
predicted performance
performance of of four
four monthly
monthly cooling
cooling energy
energy machine
machine
learning
learning models.
models. The
The MO-BASS
MO-BASS model model has
has the
the smallest
smallest CV(RMSE), followed by
CV(RMSE), followed by the
the SO-
SO-
BASS model. The CV(RMSE) of both BASS models is below 0.005,
BASS model. The CV(RMSE) of both BASS models is below 0.005, indicating that these indicating that these
two
two models
models have
have good
good performance.
performance. The The CV(RMSE)
CV(RMSE) valuesvalues of
of the
the SO-DNN
SO-DNN modelsmodels are
are
between
between 0.075 and 0.025, slightly larger than the CV(RMSE) of the two BASS models. From
0.075 and 0.025, slightly larger than the CV(RMSE) of the two BASS models. From
Figure
Figure 7b,
7b, similar
similar conclusions
conclusions can can bebe drawn
drawn asas in
in Figure
Figure 7a.7a. Finally,
Finally, itit can
can be
be discovered
discovered
from Figure 7c that all four models have R 2 values higher than 0.93, with two BASS models
from Figure 7c that all four models have R²values higher than 0.93, with two BASS models
having 2
having RR²values
values close
close to 1. In
to 1. In summary,
summary, the best performance
the best performance models
models forfor monthly
monthly cooling
cooling
energy prediction are the MO-BASS in this
energy prediction are the MO-BASS in this case study.case study.
The modeling time of the four monthly cooling energy machine learning models
is listed in Table 3. The modeling time of the two single-output models (SO-BASS and
SO-DNN) is higher than that of the two multi-output models. For the monthly cooling
energy model, the modeling time of the BASS single/multi-output model has no significant
differences since the number of monthly cooling energy data is only 5. There is still a
marked difference in the modeling time of the two DNN models, and the multi-output
model is around five times that of the single-output model. Overall, the multi-output
monthly cooling energy models are more time-saving compared to the single-output
monthly cooling energy models, especially for the DNN models.
Figure 8 shows the correlation structure tree diagram of five monthly energy use
models. Figure 8a shows the dendrogram of engineering-based building energy models,
and the adjacent months of July and August have the highest correlation. Hence, the
hottest two months have similar patterns of cooling energy from the air-conditioning
system in this office building. The two transitional months, May and September, are closely
correlated because they have similar climatic conditions including outdoor temperature
and solar radiation. The two BASS models for monthly cooling energy (Figure 8b,c) can
(a) CV(RMSE)
both retain the same correlation structure as the engineering-based model. In contrast, the
two monthly cooling DNN models (Figure 8d,e) are unable to maintain the same correlation
as the patterns illustrated in Figure 8a. As a result, the multi-output BASS models can
maintain inter-output correlation when predicting both month-by-month and day-by-day
cooling energy in this office building as shown in Figures 5 and 8.
two models have good performance. The CV(RMSE) values of the SO-DNN models are
between 0.075 and 0.025, slightly larger than the CV(RMSE) of the two BASS models. From
Figure 7b, similar conclusions can be drawn as in Figure 7a. Finally, it can be discovered
from Figure 7c that all four models have R²values higher than 0.93, with two BASS models
Buildings 2022, having
12, 2109 R²values close to 1. In summary, the best performance models for monthly cooling 12 of 30
(a) CV(RMSE)
(b) MAPE
(c) R²
Figure 7. Prediction performance
Figure of performance
7. Prediction four machineoflearning models
four machine for monthly
learning cooling
models for energy
monthly (SO,
cooling energy (SO,
single output; MO, multiple
single output).
output; MO, multiple output).
The modeling time of the four monthly cooling energy machine learning models is
listed in Table 3. The modeling time of the two single-output models (SO-BASS and SO-
DNN) is higher than that of the two multi-output models. For the monthly cooling energy
model, the modeling time of the BASS single/multi-output model has no significant dif-
Buildings 2022, 12, x FOR PEER REVIEW 14 of 33
Buildings 2022, 12, 2109 13 of 30
(d) (e)
Figure 8.
Figure 8. Dendrogram
Dendrogramofofmonthly
monthly cooling
cooling energy
energy fortraining
for the the training set
set and and
four four machine
machine learninglearning
models.
models.
3.2.3. Multi-Time Scale Cooling Energy Models
3.2.3.The
Multi-Time Scale
prediction Cooling Energy
performance of the Models
multi-time scale cooling energy models is shown
in Figure 9. The features
The prediction of these of
performance multi-time scale models
the multi-time can predict
scale cooling energy the cooling
models is energy
shown
at
in daily,
Figuremonthly, and annual
9. The features timemulti-time
of these scales, simultaneously.
scale models can First,predict
the performance
the coolingof daily
energy
cooling
at daily,energy
monthly, prediction
and annual from the scales,
time multi-time scale models
simultaneously. is analyzed.
First, The medians
the performance of
of daily
CV(RMSE),
cooling energy MAPE, and R2from
prediction of MO-BASS and MO-DNN
the multi-time scale modelsmodels are 0.007The
is analyzed. andmedians
0.02, 0.005 of
and 0.016, 0.99
CV(RMSE), and and
MAPE, 0.97,R²
respectively.
of MO-BASS Therefore,
and MO-DNN the multi-time
models are scale MO-BASS
0.007 and 0.02,model0.005
has
and a0.016,
better performance
0.99 in predicting
and 0.97, respectively. daily cooling
Therefore, energy. Note
the multi-time scalethat the performance
MO-BASS model has
of
a better performance in predicting daily cooling energy. Note that the performanceof
MO-BASS and MO-DNN models is similar in the case of individual prediction ofdaily
MO-
cooling energy as described in Section 3.2.1. Second, the analysis
BASS and MO-DNN models is similar in the case of individual prediction of daily cooling of the multi-time scale
model
energyisasfocused on the
described monthly3.2.1.
in Section cooling energy
Second, theforecast.
analysisThe CV(RMSE)
of the multi-timeandscale
MAPE values
model is
from
focusedtheonMO-BASS models
the monthly are smaller
cooling energythan thoseThe
forecast. from the MO-DNN
CV(RMSE) models,values
and MAPE and the fromR2
values of MO-BASS
the MO-BASS models areare
larger thanthan
smaller those fromfrom
those the MO-DNN
the MO-DNN models. Hence,
models, andthe
theMO-BASS
R²values
model still performs better than the MO-DNN model for the
of MO-BASS are larger than those from the MO-DNN models. Hence, the MO-BASS monthly cooling energy. Third,
the model performance of annual cooling energy is analyzed.
model still performs better than the MO-DNN model for the monthly cooling energy. The CV(RMSE) and MAPE
values from
Third, the the MO-BASS
model performance models are close
of annual to 0,energy
cooling and the is corresponding R2 is close to
analyzed. The CV(RMSE) and1.
The performance of MO-DNN is worse than that of MO-BASS.
MAPE values from the MO-BASS models are close to 0, and the corresponding R²is close It is worth noting that the
prediction accuracy of of
to 1. The performance theMO-DNN
multi-time is scale
worsemodel increases
than that with theItincrease
of MO-BASS. is worthinnoting
time scale.
that
This is becauseaccuracy
the prediction the data complexity decreases
of the multi-time scalewith
modelthe increases
increase inwithtimethescale. In summary,
increase in time
the MO-BASS model performs better in predicting multiple time scales compared to the
scale. This is because the data complexity decreases with the increase in time scale. In
MO-DNN model.
summary, the MO-BASS model performs better in predicting multiple time scales com-
Table 3 lists the computational time for multi-time scale models of BASS and DNN. The
pared to the MO-DNN model.
MO-DNN model requires 236.3 s, while the MO-BASS model needs 197.2 s. The computation
time MO-DNN is 20% more than that of the MO-BASS model. As a result, the MO-BASS
offers both high accuracy and low modeling costs in comparison with the MO-DNN model
for the simultaneous prediction of daily, monthly, and annual cooling energy.
3.2.4. Performance Analysis of 10 Models for Monthly and Annual Cooling Energy
This section compares the predictive performance and accumulative characteristics
of monthly and annual cooling energy from 10 models as listed in Table 4. The model
Buildings 2022, 12, 2109 14 of 30
names in Table 4 include three parts. The first two letters denote the single-output (SO)
or multi-output (MO) model. The third letter denotes the machine learning algorithm, in
which the B is the Bayesian adaptive spline surface (BASS) and the D is the deep neural
network (DNN). The letters after the hyphen denote the time scale of energy models. The
D is the daily model, the M is the monthly model, and the Mu is the multi-time (daily,
monthly, and annual) scale model. For example, the MOB-Mu is to simultaneously predict
the daily, monthly, and annual energy from the multi-time scale multi-output BASS model.
Note that there are two ways to obtain monthly or annual energy: direct and accumulation.
For instance, the MOD-M or SOD-M models can directly predict the monthly energy based
on the deep neural network models. In contrast, the SOB-D model needs to sum the daily
energy from the single-output daily BASS models to obtain the monthly (or annual) energy
since there is unavailable for monthly (or annual) energy use from the daily BASS model.
The predictive performance of 10 models for monthly cooling energy is shown in
Figure 10. The performance of monthly cooling energy from the direct prediction is better
than those from the accumulative models for the BASS approaches in terms of CV(RMSE),
MAPE, and R2 . The SOB-M performs better than the SOB-D and the MOB-M has better
predictive capability compared to the MOB-D. The multi-time scale BASS model has very
good performance compared to the monthly and daily BASS models for cooling energy
prediction. As for the DNN models, the best prediction model of cooling energy is the
monthly summation from the daily SOD-D models. The multi-time scale MOD-Mu model
has moderate performance compared to the other four DNN models. The BASS models
have better prediction performance in comparison with the DNN models for monthly
cooling energy. All the R2 values for the BASS models are above 0.99, which indicates that
the BASS has very high predictive capability for monthly cooling energy.
It is also interesting to compare the accumulative features for the different time scale
models. As might be expected, the results from the monthly summation of daily cooling
energy models are different from the direct prediction of monthly cooling energy models,
even for the same machine learning algorithm. The computational results from this case
study confirm this statement. For the BASS models, there are five methods to obtain
monthly energy use, which may make it confusing regarding which monthly results should
be selected in the multiple time scale analysis. For instance, if the best daily model is from
the daily BASS single-output models and the best monthly model is from the monthly
BASS single-output model in terms of prediction accuracy, then the monthly energy use
Buildings 2022, 12, 2109 15 of 30
would not be the same as the summation of daily energy use in a specific month. The
only model that can maintain good accumulative features is the multi-time scale BASS
model (MOB-Mu), due to the processing of the principal component analysis. The monthly
2022, 12, x FOR PEER REVIEW cooling energy in July would equal the summation of the July15daily
of 33 prediction of cooling
energy for the MOB-Mu model.
(a) CV(RMSE)
(b) MAPE
(c) R²
Figure 9. PredictionFigure
performance of two performance
9. Prediction machine learning models
of two for learning
machine multi-time scalesfor
models cooling en- scales cooling energy.
multi-time
ergy.
Buildings 2022, 12, 2109 16 of 30
(a) CV(RMSE)
(b) MAPE
(c) R²
Figure 10. Figure
Performance of 10 models
10. Performance of 10for monthly
models cooling energy
for monthly cooling (refer
energyto(refer
Tableto4Table
for model expla-explanation).
4 for model
nation).
models are MOB-Mu, MOB-M, and MOB-D, which all belong to the multi-output BASS
models. The next three models are SOB-M, SOB-D, and SOD-D, which all belong to the
single-output models. All these six models have very high R2 values, above 0.997. The
worst model is the single-out monthly DNN models (SOD-M), in which the CV(RMSE),
MAPE, and R²values are 0.016, 0.013, and 0.970, respectively. The remaining nine models’
Buildings 2022, 12, 2109 17 of 30
CV(RMSE) values are less than 0.008, MAPE values less than 0.0078, and R²values greater
than 0.995, indicating that these models have good predictive performance.
(a) CV(RMSE)
(b) MAPE
(c) R²
Figure 11.11.
Figure Performance ofof
Performance 1010models
modelsfor
forannual
annualcooling
cooling energy (referto
energy (refer toTable
Table44for
formodel
modelexplanation).
explana-
tion).
For the sake of accumulative features, the energy use prediction at a larger scale
canForbethe
obtained
sake offrom the summation
accumulative of the
features, small-scale
energy use energy use foratthe
prediction single-output
a larger scale can or
bemulti-output
obtained from models. However,of
the summation this simple accumulation
small-scale energy use for is not
thenecessarily
single-outputa good method
or multi-
in predicting
output models. the multi-time
However, this scale
simple energy use in buildings.
accumulation There are at
is not necessarily least three
a good method reasons
in
for this. The
predicting the first reason scale
multi-time is thatenergy
the residual
use inerrors from the
buildings. small
There arescale models
at least threemay lead to
reasons
foruncontrolled
this. The firsterrors.
reasonIfisallthat
thethe
residual errors
residual from
errors fromthethe
small scale
small models
scale models aremay
positive
lead or
negative, the final residuals at a large scale can be added to larger residual
to uncontrolled errors. If all the residual errors from the small scale models are positive errors. If there
are negative and positive errors for the small time scale models, the
or negative, the final residuals at a large scale can be added to larger residual errors. Ifresidual errors at a
larger scale may be small due to the offsetting of negative and positive
there are negative and positive errors for the small time scale models, the residual errors values. However,
at the fundamental
a larger scale maylogic for the
be small small
due scale
to the energy of
offsetting models mayand
negative be unreasonable
positive values. in How-
the first
place,
ever, thewhich likely leads
fundamental logictoforunexpected
the small errors in estimating
scale energy modelsenergy
may be use at a larger scale.
unreasonable in the The
second reason is that the model accuracy would be easily improved
first place, which likely leads to unexpected errors in estimating energy use at a larger at a larger time scale.
ThisThe
scale. is because
secondthe efforts
reason of creating
is that the modelmachine learning
accuracy would models at a larger
be easily improvedtimeatscale would
a larger
time scale. This is because the efforts of creating machine learning models at a larger time
scale would be much less compared to the small time scale models due to the decrease in
model number. The third reason is the computational cost. The increase in computational
cost by adding large time scale models is not significant since there are much fewer mod-
els with an increase in time scale. Moreover, the multi-time scale multi-output machine
Buildings 2022, 12, 2109 18 of 30
be much less compared to the small time scale models due to the decrease in model number.
The third reason is the computational cost. The increase in computational cost by adding
large time scale models is not significant since there are much fewer models with an increase
in time scale. Moreover, the multi-time scale multi-output machine learning models that
can maintain the correlation structure of energy data show very good performance, as
discussed in this section.
Figure12.
Figure 12.Correlation
Correlationcoefficient
coefficientofof daily
daily heating
heating energy
energy in January
in January (2, 2(2, 2 January;
January; ... ;…
30, ;30
30,January).
30 Janu-
ary).
Figure 13 shows the predictive performance for daily heating energy from the single-
output (SO) 13
Figure and multi-output
shows (MO)performance
the predictive models based foron theheating
daily BASS and energyDNN fromapproaches,
the single-
respectively.
output (SO)Figure 13a indicates(MO)
and multi-output that the median
models CV(RMSE)
based on the values
BASS and of theDNNboxplots for the
approaches,
four models are
respectively. not significantly
Figure 13a indicates different–all below CV(RMSE)
that the median 0.02. Hence,values
the fourof models all have
the boxplots for
good performance
the four models are regarding daily heating
not significantly energy prediction.
different–all below 0.02.However,
Hence, the there
fourare outliers
models all
for these models as illustrated in Figure 13a. The three models (SO-BASS,
have good performance regarding daily heating energy prediction. However, there are SO-DNN, and
MO-DNN)
outliers forhave
thesethe CV(RMSE)
models values greater
as illustrated than13a.
in Figure 0.15,The
andthree
two points
modelsare even greater
(SO-BASS, SO-
than 0.2. This shows that the model performs poorly on certain
DNN, and MO-DNN) have the CV(RMSE) values greater than 0.15, and two points days of heating energyare
prediction.
even greaterIn than
contrast, the CV(RMSE)
0.2. This shows thatvalues of theperforms
the model MO-BASS models
poorly on are all less
certain than
days 0.15,
of heat-
indicating
ing energy good prediction
prediction. capability.the
In contrast, Moreover,
CV(RMSE) the values
interquartile
of therange
MO-BASSfrom the MO-BASS
models are all
model is the smallest, indicating that the MO-BASS model is the most
less than 0.15, indicating good prediction capability. Moreover, the interquartile range stable in terms of
model errors. The trends of MAPE values shown in Figure 13b are similar
from the MO-BASS model is the smallest, indicating that the MO-BASS model is the most to those shown
stable in terms of model errors. The trends of MAPE values shown in Figure 13b are sim-
ilar to those shown in Figure 13a, which verifies the accuracy results from the CV(RMSE).
The R² values shown in Figure 13c are above 0.95 for the four models, indicating that the
model performance is high. Among them, the R²values of the MO-BASS model are greater
have good performance regarding daily heating energy prediction. However, there are
outliers for these models as illustrated in Figure 13a. The three models (SO-BASS, SO-
DNN, and MO-DNN) have the CV(RMSE) values greater than 0.15, and two points are
even greater than 0.2. This shows that the model performs poorly on certain days of heat-
ing energy prediction. In contrast, the CV(RMSE) values of the MO-BASS models are all
Buildings 2022, 12, 2109
less than 0.15, indicating good prediction capability. Moreover, the interquartile range 19 of 30
from the MO-BASS model is the smallest, indicating that the MO-BASS model is the most
stable in terms of model errors. The trends of MAPE values shown in Figure 13b are sim-
ilar to those shown in Figure 13a, which verifies the accuracy results from the CV(RMSE).
in Figure 13a, which verifies the accuracy results from the CV(RMSE). The R2 values shown
The R² values shown in Figure 13c are above 0.95 for the four models, indicating that the
in Figure 13c are above 0.95 for the four models, indicating that the model performance is
model performance is high. Among them, the2 R²values of the MO-BASS model are greater
high. Among
than 0.985. Therefore, them, the
the MO-BASS R is
model values of the
the most MO-BASS
accurate dailymodel
heatingare greater than 0.985. Therefore,
energy
the MO-BASS
model in this case study. model is the most accurate daily heating energy model in this case study.
(a) CV(RMSE)
(b) MAPE
(c) R²
Figure 13. Prediction
Figureperformance of fourperformance
13. Prediction machine learning models
of four for daily
machine heating models
learning energy (SO,
for daily heating energy (SO,
single output; MO, multiple output).
single output; MO, multiple output).
The modeling time of the four daily heating energy models is shown in Table 5. The
most time-saving model is the MO-DNN model, which requires only 69.7 s. The most
time-consuming model is the SO-DNN, which spends 3722.6 s (more than 1 h)–around 53
times that of the SO-DNN model. The MO-BASS model is also significantly more time
efficient compared to the SO-BASS model. Therefore, the multi-output models save com-
Buildings 2022, 12, 2109 20 of 30
The modeling time of the four daily heating energy models is shown in Table 5. The
most time-saving model is the MO-DNN model, which requires only 69.7 s. The most time-
consuming model is the SO-DNN, which spends 3722.6 s (more than 1 h)–around 53 times
that of the SO-DNN model. The MO-BASS model is also significantly more time efficient
compared to the SO-BASS model. Therefore, the multi-output models save computational
time compared to the single-output models.
Table 5. The computational time of different time-scale heating energy models (unit: s).
The dendrograms for daily heating energy from the training set and the four models
The dendrograms for daily heating energy from the training set and the four models
are shown in Figure 14. The numbers in Figure 14 represent dates in a specific month,
are shown in Figure 14. The numbers in Figure 14 represent dates in a specific month,
such as 2 for 2 January. Figure 14a reflects the dendrogram of the correlation between the
such as 2 for 2 January. Figure 14a reflects the dendrogram of the correlation between the
outputs from the engineering-based EnergyPlus energy models. It can be seen that 2, 5,
outputs from the engineering-based EnergyPlus energy models. It can be seen that 2, 5,
12, 19, and 26 belong to one subgroup, and these days are the first working days after the
12, 19, and 26 belong to one subgroup, and these days are the first working days after the
holidays
holidays oror weekends.
weekends. The The reason
reason for
for the
the aggregation
aggregation is is that
that the
the office
office building
building is
is not
not
heated during non-working days, resulting in a low temperature inside the
heated during non-working days, resulting in a low temperature inside the office building. office build-
ing. The heating
The heating systemsystem
startsstarts to work
to work on theon theworking
first first working day, consumes
day, which which consumes more
more energy
energy than the following days. Figure 14d,e have significantly different correlation
than the following days. Figure 14d,e have significantly different correlation structures struc-
tures from Figure
from Figure 14a since
14a since the non-working
the non-working days days dobelong
do not not belong
to oneto subgroup.
one subgroup. In con-
In contrast,
trast, the correlation structures from Figure 14b,c are similar to that
the correlation structures from Figure 14b,c are similar to that from Figure 14a.from Figure 14a.
(d) (e)
Figure
Figure 14.
14. Dendrogram of January
Dendrogram of Januarydaily
dailyheating
heatingenergy
energyfor
forthe
thetraining
trainingsetset and
and four
four machine
machine learn-
learning
ing models.
models.
Figure
Figure
Figure 15. Correlation 15. Correlation
15. Correlation
coefficient coefficient
heatingof
coefficient
of monthly ofmonthly
monthlyheating
energy. heatingenergy.
energy.
Figure 16 showsFigure
Figure 16
16shows
shows the
the performance the performance
for the indicators
performance
indicators indicators
monthly for
for the
heating the monthly
monthly
energy heating
heating energy
models. energy models.
models.
The
The MO-BASS model MO-BASS
The MO-BASS model has
modelCV(RMSE)
has the smallest the smallest
has the smallest CV(RMSE)
and MAPE CV(RMSE) and
with the and MAPE
MAPE
largest with
R², withthe largest R2 ,R²
the largest
indicat- indicating
, indicat-
that
ing that the MO-BASS theoutperforms
ing that MO-BASS
the MO-BASS outperforms
the other three
outperforms themodels.
other three models.
The three
the other other The
three
models. other
models three three
also
The other models also have
models also
2
have CV(RMSE)CV(RMSE)
less than less
0.13, than
MAPE 0.13,
lessMAPE
than less
0.1, than
and 0.1,
R² and
greater R greater
than 0.92, than 0.92,
indicating
have CV(RMSE) less than 0.13, MAPE less than 0.1, and R² greater than 0.92, indicating indicating that they
also
that they also have have
thatgood goodhave
theypredictive
also predictive
performance. performance.
good predictive that Note
Noteperformance.
the monththat the that
with
Note month
the the
worstwith
pre-
month thewith
worst
theprediction
worst pre-
performance
diction performance for each for each
model ismodel
March, is March,
followed followed
by November. by
diction performance for each model is March, followed by November. What November.
What these What
two these twothese
months
two
months have inhavecommonin common
is that is
theythatare they
both are both
transitionaltransitional
months. months.
Climatic Climatic
conditions
months have in common is that they are both transitional months. Climatic conditions conditions such as
temperature
such as temperature varycan vary considerably theduring the months,
transition months, which can add to the
such can considerably
as temperature can varyduring transition
considerably which
during the transition can months,
add which can add
complexity
to the complexity of the
of the heating heating
energy andenergy
renderand render
it more it more
difficult difficultthe
to predict to heating
predict the heating energy.
to the complexity of the heating energy and render it more difficult to predict the heating
Prediction
energy. Prediction accuracyaccuracy is improved
is improved significantly
significantly when there when there are
are stable stable
changes in changes in climate
energy. Prediction accuracy is improved significantly when there are stable changes in
conditions
climate conditions such as temperatures,
such as outdoor outdoor temperatures,
for example forinexample
Februaryinand February
December.and December.
climate conditions such as outdoor temperatures, for example in February and December.
(a) CV(RMSE)
(b) MAPE
(c) R²
Figure 16. Prediction performance
Figure of four
16. Prediction machine learning
performance of fourmodels for learning
machine monthly models
heating energy.
for monthly heating energy.
points in R2 between 0.875 and 0.95 for the DNN daily models, indicating that the model
performance becomes worse on some days. Note that the interquartile ranges of the BASS
model are smaller than that of the DNN model, indicating that the MO-BASS model is
Buildings 2022, 12, x FOR PEER REVIEW 25 of 33
more stable. Therefore, the multi-time scale BASS model has better predictive capability in
comparison with the DNN model.
(d) (e)
Figure 17. Dendrogram of monthly heating energy for the training set and four machine learning
models.
(a) CV(RMSE)
Figure 18. Cont.
(a) CV(RMSE)
R PEER REVIEW 26 of 33
Buildings 2022, 12, 2109 24 of 30
(b) MAPE
(c) R²
Figure 18. Prediction performance of twoperformance
Figure 18. Prediction machine learning models
of two machine for multi-time
learning scales heating
models for multi-time scales heating
energy. energy.
Table 5 shows the computation time of multi-time scale heating energy models using the
Table 5 shows the computation time of multi-time scale heating energy models using
BASS and DNN approaches. The computational time is quite similar for these two models.
the BASS and DNN Considering
approaches. theThe
modelcomputational
accuracy as showntime is quite
in Figure similar
18, the for these
multi-time twomodels
scale BASS
models. Consideringwould
the model accuracy
be a better as shown
candidate in Figure
for the multi-time 18,prediction
scale the multi-time scale
of building BASS
energy use.
models would be a better candidate for the multi-time scale prediction of building energy
3.3.4. Performance Analysis of Ten Models for Monthly and Annual Heating Energy
use.
Figure 19 shows the performance of 10 models that can obtain monthly heating energy.
The meanings of these 10 models are available in Table 4. The first three best models are the
3.3.4. Performance Analysis of Ten Models
daily multi-output models for Monthly
(MOB-D), andmulti-output
monthly Annual Heating
models Energy
(MOB-M), and multi-
time scale model (MOB-Mu) based on the BASS algorithm.
Figure 19 shows the performance of 10 models that can obtain monthly heating These three modelsen-
have very
low CV(RMSE) and MAPE values. The next two models are the single-output daily BASS
ergy. The meanings of these 10 models are available in Table 4. The first three best models
models (SOB-D) and the multi-output daily DNN models (MOD-D). These two models still
are the daily multi-output models
have very (MOB-D),
high R monthly
2 values–close multi-output
to 1. The following two models
models are(MOB-M), andmonthly
single-output
multi-time scale model (MOB-Mu)
BASS basedand
models (SOB-M) onsingle-output
the BASS algorithm. These (SOD-D).
daily DNN models three models have three
The remaining
models do not have good predictive performance compared
very low CV(RMSE) and MAPE values. The next two models are the single-output dailyto the other seven models. The
BASS models (SOB-D) and the multi-output daily DNN models (MOD-D). These two
models still have very high R2 values–close to 1. The following two models are single-
output monthly BASS models (SOB-M) and single-output daily DNN models (SOD-D).
The remaining three models do not have good predictive performance compared to the
Buildings 2022, 12, 2109 25 of 30
(a) CV(RMSE)
(b) MAPE
(c) R²
Figure 19. Performance of 10 modelsoffor
Figure 19. Performance 10 monthly heating
models for energy
monthly (refer
heating to Table
energy (refer4tofor model
Table 4 forexpla-
model explanation).
nation).
Figure 20 shows the performance of 10 models that can obtain annual energy. The
Figurethree
20 shows the performance
multi-output of 10 models
BASS models (MOB-D, thatMOB-M,
can obtain annual energy.
MOB-Mu) are theThe
first three best
three multi-output
models in BASS
whichmodels (MOB-D, values
the CV(RMSE) MOB-M, areMOB-Mu) arethe
about 0.003, the first three
MAPE valuesbest
are about 0.002,
the R2the
models in which CV(RMSE)
values values
are about are The
0.999. about 0.003, the MAPE
performance of thevalues
SOB-D are about
and 0.002,single-output
SOB-M
the R² values
BASSare models
about 0.999. The performance
is slightly worse than of theofSOB-D
that and SOB-M
the above single-output
three multi-output BASS models.
BASS models is slightly
Among worse
the five DNN than that of
models, thethe above models
MOD-D three multi-output BASS models.
and the MOD-Mu models have similar
Among theperformance
five DNN models,
to thethe MOD-D
SOB-M models
models. and
The the MOD-Mu
SOD-M modelsmodels
and MOD-M have similar
do not have good
performance to the SOB-M
predictive models.for
performance The SOD-M
annual and MOD-M
heating energy. models do not have good
predictive performance for annual heating energy.
Buildings
Buildings 2022, 12, x 2022, 12, 2109
FOR PEER REVIEW 28 of 33 26 of 30
(a) CV(RMSE)
(b) MAPE
(c) R²
Figure 20.Figure
Performance of 10 models
20. Performance of 10for annual
models forheating energy energy
annual heating (refer to(refer
Tableto 4Table
for model explana-
4 for model explanation).
tion).
3.4. Guide and Application of Building Multi-Output Energy Models
3.4. Guide andThis
Application
sectionofdiscusses
Building Multi-Output
the implications, Energy Models and applications of the results
guidelines,
Thispresented in Sections
section discusses the 3.1–3.3 from four
implications, aspects.and
guidelines, The first aspectofdiscusses
applications the results the choice
of machine learning models between the single-output and multiple-output
presented in Sections 3.1 to 3.3 from four aspects. The first aspect discusses the choice of models in
machine learning models between the single-output and multiple-output models in build- models
building energy assessment. The second aspect is the choice of multiple-output
ing energywith or withoutThe
assessment. considering output
second aspect correlation
is the choice ofinmultiple-output
the building energy modelsanalysis.
with orThe third
without aspect is related
considering outputto the choice of
correlation inmulti-output modelsanalysis.
the building energy for various Thetime
thirdscales
aspectby taking
is relatedthe additive
to the choiceorofaccumulative
multi-output features
models for intovarious
account.timeThe fourth
scales aspect the
by taking is focused
addi- on the
application of
tive or accumulative the multi-output
features models
into account. used in
The fourth this paper.
aspect is focused on the application
The multi-output learning
of the multi-output models used in this paper. models are preferred in the case of multiple outputs of
building energy analysis. This is because the computational
The multi-output learning models are preferred in the case of multiple cost can be dramatically
outputs of
reduced in comparison with the single-output models, as discussed
building energy analysis. This is because the computational cost can be dramatically re- in Sections 3.2 and 3.3.
duced inThe benefits with
comparison of using the multi-output
the single-output models
models, are moreinsignificant
as discussed Sections 3.2when there are a
and 3.3.
largeofnumber
The benefits using the of multi-output
building energy outputs–for
models are more instance, over
significant 10 outputs.
when there areMoreover,
a large model
accuracy from the multi-output models is not necessarily less
number of building energy outputs–for instance, over 10 outputs. Moreover, model accu- than that of the single-output
racy from the multi-output models is not necessarily less than that of the single-outputthe multi-
models in building energy assessment. Therefore, it would be recommended that
output learning models should be used when applying the machine learning techniques to
models in building energy assessment. Therefore, it would be recommended that the
building energy assessment, especially for a larger number of model outputs.
multi-output learning models should be used when applying the machine learning tech-
niques to building energy assessment, especially for a larger number of model outputs.
Buildings 2022, 12, 2109 27 of 30
There are two types of machine learning methods depending on whether to consider
the correlation of building energy performance. One example of learning models is the
Bayesian adaptive spline surface in which the correlation of outputs can be taken into
account, while one example of learning models without considering output correlation
is the deep neural network. If there are almost no or very weak correlations among
building energy use, both two types of learning models can be used to create multi-output
models. If the correlation of outputs is significant, it is necessary to implement the learning
methods by considering the correlation of outputs. The model accuracy can be improved
by maintaining the correlation of building energy use.
It is necessary to consider the additive or accumulative features when constructing the
learning models for predicting building energy use in various time scales. This is because
the building energy use at a larger time scale is the sum of building energy at a smaller
time scale. For instance, the monthly electricity use from a building would be the annual
electricity for this building. For single-output models at a smaller time scale, it is natural
to sum up all the energy use at a smaller time scale to obtain total energy use at a larger
time scale. However, the residuals for this simple summation would be hard to manage. If
there are all positive (or negative) residuals from the smaller time scale models, the total
residuals would be very large compared to the single-output model at a larger time scale.
Another issue is the inconsistency between the summation of smaller time scale models
and the larger time scale model, which confuses the estimation of building energy use at a
larger time scale. By considering the correlation of building energy use, the multi-output
learning models can create fast and consistent models at various time scales for building
energy analysis. Therefore, multi-output learning models that can maintain the correlation
of outputs are recommended in multiple time scale predictions of building energy analysis.
The method proposed in this paper can also be applied to performance indicators of
buildings (such as energy, loads, and carbon emissions) in different scenarios although
this paper is concentrated on building energy consumption. It would be interesting to
explore building load profiles at various time scales (hourly, daily, and monthly), which can
provide useful information on the design of district heating and cooling systems. When
designing net-zero emission buildings, it is necessary to provide the matching of demand
and supply at various time scales to guide for the design of PV systems. The multi-output
models at various time scales would be very useful to provide accurate building demand
data at various time scales to compute solar fraction, self-sufficiency, and other indicators.
These results can be used for optimizing the ratio of PV rated power and battery. Moreover,
the method used here can be used in uncertainty analysis, sensitivity analysis, and model
calibration in building energy assessment. For instance, the multi-output models at daily
periods can be used as mathematical models to calibrate these daily models using the
Bayesian analysis. The computational loads would be reduced significantly since there are
only two multi-output learning models for heating and cooling energy use in buildings.
4. Conclusions
This research investigates the predictive performance of multi-output machine learn-
ing models at three time scales (daily, monthly, and annual) for building energy assessment
using two algorithms (Bayesian adaptive spline surface and deep neural network). The
results indicate that the machine learning models, which could consider the correlation of
energy use, can have high model accuracy with fast computation and accumulative features
at different time scales in building energy analysis. The multi-output learning models for
building energy prediction would significantly reduce the computational time for creating
learning models in comparison to the single-output learning models. The ability to main-
tain the correlation structure of energy data for the multi-output learning models is the key
to providing accurate results and accumulative features. The deep neural network models
can simulate the multiple outputs without taking into account the correlation of energy
use. Hence, the deep neural network cannot have accumulative features. In contrast, the
multi-time scale Bayesian adaptive spline surface models can have the same or similar
Buildings 2022, 12, 2109 28 of 30
correlation structure of energy data from the engineering-based building energy models.
This means that the predicted energy data at a larger time scale would equal the summation
of energy data at a smaller time scale in building energy computation. Moreover, it is
found that the simple summation of energy results from smaller time scale learning models
do not necessarily have good predictions of energy use at a larger time scale due to the
uncontrolled offset of large and small residuals, or the final accumulative large errors.
Three guidelines can be obtained for creating machine learning models in building
energy assessment based on this research. The multi-output learning models are preferred
compared to the single-output model in dealing with multiple outputs, especially in the
case of the number of outputs over 10. Multi-output learning models that can consider
output correlation are recommended when the multiple outputs are correlated in building
energy analysis. Compared to the direct summation from the single-output models at the
smaller time scale, multi-output models with a consistent accumulative feature that can
simultaneously predict energy use at various time scales are preferred when dealing with
multiple time scale building energy predictions.
The multi-out learning models used in this paper are only applied to an office building.
More research should be implemented on other building types, for example, residential
buildings, schools, and hospitals, to understand the suitability of these multi-output meth-
ods. More studies should also be conducted to explore the predictive performance of
multi-output models for building energy analysis using different learning algorithms, such
as random forest, support vector machine, and Gaussian process.
Author Contributions: Conceptualization, G.L. and W.T.; methodology, G.L. and W.T.; software, G.L.;
investigation, H.Z. and B.C.; writing—original draft preparation, G.L.; writing—review and editing,
W.T.; visualization, B.C. All authors have read and agreed to the published version of the manuscript.
Funding: This research was supported by the National Natural Science Foundation of China
(No. 51778416) and the Key Projects of Philosophy and Social Sciences Research, Ministry of Education
of China (No. 16JZDH014).
Data Availability Statement: Not applicable.
Conflicts of Interest: The authors declare no conflict of interest.
Abbreviations
ANN artificial neural network
BASS Bayesian adaptive spline surface
BIPV building integrated photovoltaic
BMARS Bayesian multivariate adaptive regression splines
CSP cooling set-point
CSWD Chinese standard weather data
CV(RMSE) coefficient of variation of the root mean square error
DNN deep neural network
DT decision tree
ENMIM ensemble model named evolutionary neural machine inference model
EPD equipment power density
EWU exterior wall U-value
GA-NMM genetic algorithm-based numerical moment matching
GB gradient boosting
HSP heating set-point
HVAC heating, ventilation, and air conditioning
INF infiltration rate
KNN K-nearest neighbor
lightGBM light gradient boosting machine
Buildings 2022, 12, 2109 29 of 30
References
1. UN Environment Programme (UNEP). 2021 Global Status Report for Buildings and Construction; UNEP: Nairobi, Kenya, 2021.
2. Zhang, Y.; Teoh, B.K.; Wu, M.; Chen, J.; Zhang, L. Data-driven estimation of building energy consumption and GHG emissions
using explainable artificial intelligence. Energy 2023, 262, 125468. [CrossRef]
3. Al-Shargabi, A.A.; Almhafdy, A.; Ibrahim, D.M.; Alghieth, M.; Chiclana, F. Buildings’ energy consumption prediction models
based on buildings’ characteristics: Research trends, taxonomy, and performance measures. J. Build. Eng. 2022, 54, 104577.
[CrossRef]
4. Guo, Y.-Y. Revisiting the building energy consumption in China: Insights from a large-scale national survey. Energy Sustain. Dev.
2022, 68, 76–93. [CrossRef]
5. Mohapatra, S.K.; Mishra, S.; Tripathy, H.K.; Alkhayyat, A. A sustainable data-driven energy consumption assessment model for
building infrastructures in resource constraint environment. Sustain. Energy Technol. Assess. 2022, 53, 102697. [CrossRef]
6. Fathi, S.; Srinivasan, R.; Fenner, A.; Fathi, S. Machine learning applications in urban building energy performance forecasting: A
systematic review. Renew. Sustain. Energy Rev. 2020, 133, 110287. [CrossRef]
7. Zhang, L.; Wen, J.; Li, Y.; Chen, J.; Ye, Y.; Fu, Y.; Livingood, W. A review of machine learning in building load prediction.
Appl. Energy 2021, 285, 116452. [CrossRef]
8. Lei, L.; Chen, W.; Wu, B.; Chen, C.; Liu, W. A building energy consumption prediction model based on rough set theory and deep
learning algorithms. Energy Build. 2021, 240, 110886. [CrossRef]
9. Liu, Y.; Chen, H.; Zhang, L.; Wu, X.; Wang, X.-J. Energy consumption prediction and diagnosis of public buildings based on
support vector machine learning: A case study in China. J. Clean. Prod. 2020, 272, 122542. [CrossRef]
10. Alobaidi, M.H.; Chebana, F.; Meguid, M.A. Robust ensemble learning framework for day-ahead forecasting of household based
energy consumption. Appl. Energy 2018, 212, 997–1012. [CrossRef]
11. Wang, Z.; Hong, T.; Li, H.; Piette, M.A. Predicting city-scale daily electricity consumption using data-driven models. Adv. Appl.
Energy 2021, 2, 100025. [CrossRef]
12. Jetcheva, J.G.; Majidpour, M.; Chen, W.-P. Neural network model ensembles for building-level electricity load forecasts.
Energy Build. 2014, 84, 214–223. [CrossRef]
13. Ferrantelli, A.; Kuivjogi, H.; Kurnitski, J.; Thalfeldt, M. Office Building Tenants’ Electricity Use Model for Building Performance
Simulations. Energies 2020, 13, 5541. [CrossRef]
14. Wang, W.; Hong, T.; Xu, X.; Chen, J.; Liu, Z.; Xu, N. Forecasting district-scale energy dynamics through integrating building
network and long short-term memory learning algorithm. Appl. Energy 2019, 248, 217–230. [CrossRef]
15. Tran, D.-H.; Luong, D.-L.; Chou, J.-S. Nature-inspired metaheuristic ensemble model for forecasting energy consumption in
residential buildings. Energy 2020, 191, 116552. [CrossRef]
16. Tian, W.; Yang, S.; Li, Z.; Wei, S.; Pan, W.; Liu, Y. Identifying informative energy data in Bayesian calibration of building energy
models. Energy Build. 2016, 119, 363–376. [CrossRef]
17. Zhu, C.; Tian, W.; Yin, B.; Li, Z.; Shi, J. Uncertainty calibration of building energy models by combining approximate Bayesian
computation and machine learning algorithms. Appl. Energy 2020, 268, 115025. [CrossRef]
18. Koschwitz, D.; Frisch, J.; van Treeck, C. Data-driven heating and cooling load predictions for non-residential buildings based on
support vector machine regression and NARX Recurrent Neural Network: A comparative study on district scale. Energy 2018,
165, 134–142. [CrossRef]
Buildings 2022, 12, 2109 30 of 30
19. Jahani, E.; Cetin, K.; Cho, I.H. City-scale single family residential building energy consumption prediction using genetic
algorithm-based Numerical Moment Matching technique. Build. Environ. 2020, 172, 106667. [CrossRef]
20. Lin, Q.; Liu, K.; Hong, B.; Xu, X.; Chen, J.; Wang, W. A data-driven framework for abnormally high building energy demand
detection with weather and block morphology at community scale. J. Clean. Prod. 2022, 354, 131602. [CrossRef]
21. Olu-Ajayi, R.; Alaka, H.; Sulaimon, I.; Sunmola, F.; Ajayi, S. Building energy consumption prediction for residential buildings
using deep learning and other machine learning techniques. J. Build. Eng. 2022, 45, 103406. [CrossRef]
22. Tian, W.; Zhu, C.; Sun, Y.; Li, Z.; Yin, B. Energy characteristics of urban buildings: Assessment by machine learning. Build. Simul.
2020, 14, 179–193. [CrossRef]
23. Tian, W.; de Wilde, P.; Li, Z.; Song, J.; Yin, B. Uncertainty and sensitivity analysis of energy assessment for office buildings based
on Dempster-Shafer theory. Energy Convers. Manag. 2018, 174, 705–718. [CrossRef]
24. Wang, L.; Lee, E.W.M.; Hussian, S.A.; Yuen, A.C.Y.; Feng, W. Quantitative impact analysis of driving factors on annual residential
building energy end-use combining machine learning and stochastic methods. Appl. Energy 2021, 299, 117303. [CrossRef]
25. Singh, M.; Sharston, R. Quantifying the dualistic nature of urban heat Island effect (UHI) on building energy consumption.
Energy Build. 2022, 255, 111649. [CrossRef]
26. Luo, X.J.; Oyedele, L.O.; Ajayi, A.O.; Akinade, O.O. Comparative study of machine learning-based multi-objective prediction
framework for multiple building energy loads. Sustain. Cities Soc. 2020, 61, 102283. [CrossRef]
27. Liu, J.; Zhang, Q.; Dong, Z.; Li, X.; Li, G.; Xie, Y.; Li, K. Quantitative evaluation of the building energy performance based on
short-term energy predictions. Energy 2021, 223, 120065. [CrossRef]
28. Li, G.; Li, F.; Ahmad, T.; Liu, J.; Li, T.; Fang, X.; Wu, Y. Performance evaluation of sequence-to-sequence-Attention model for
short-term multi-step ahead building energy predictions. Energy 2022, 259, 124915. [CrossRef]
29. U.S. Department of Energy. EnergyPlus V22.1.0; U.S.Department of Energy: Washington, DA, USA, 2021.
30. MOHURD (Ministry of Housing and Urban-Rural Development). China, Energy-Saving Design Standards for Public Buildings;
China Building Industry Press: Beijing, China, 2015.
31. MOHURD (Ministry of Housing and Urban-Rural Development). China, Technical Standards for Near-Zero Energy Buildings; China
Building Industry Press: Beijing, China, 2019.
32. Francom, D.; Sansó, B. BASS: An R Package for Fitting and Performing Sensitivity Analysis of Bayesian Adaptive Spline Surfaces.
J. Stat. Softw. 2020, 94, 1–36. [CrossRef]
33. An, N.; Zhao, W.; Wang, J.; Shang, D.; Zhao, E. Using multi-output feedforward neural network with empirical mode decomposi-
tion based signal filtering for electricity demand forecasting. Energy 2013, 49, 279–288. [CrossRef]
34. Cohen-Addad, V.; Kanade, V.; Mallmann-Trenn, F.; Mathieu, C. Hierarchical Clustering: Objective Functions and Algorithms.
J. Acm. 2019, 66, 1–42. [CrossRef]
35. Kaminskyy, R.; Shakhovska, N.; Kryvinska, N.; Younas, M. Dendrograms-based disclosure method for evaluating cluster analysis
in the IoT domain. Comput. Ind. Eng. 2021, 158, 107402. [CrossRef]
36. Varshney, A.K.; Muhuri, P.K.; Lohani, Q.M.D. PIFHC: The Probabilistic Intuitionistic Fuzzy Hierarchical Clustering Algorithm.
Appl. Soft Comput. 2022, 120, 108584. [CrossRef]