0% found this document useful (0 votes)
4 views30 pages

Building Energy Models at Different Time Scales Based On Multi-Output Machine Learning

This paper explores the use of multi-output machine learning models for building energy prediction at different time scales (daily, monthly, and annual) using Bayesian adaptive spline surface (BASS) and deep neural network (DNN) algorithms. The study finds that multi-output models can accurately predict energy use while maintaining correlation structures similar to engineering-based models, and highlights the importance of these models for uncertainty and sensitivity analysis. The research contributes to the field by comparing single-output and multi-output models, investigating output correlation, and examining accumulative features in energy predictions.

Uploaded by

ferdie marcos
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
4 views30 pages

Building Energy Models at Different Time Scales Based On Multi-Output Machine Learning

This paper explores the use of multi-output machine learning models for building energy prediction at different time scales (daily, monthly, and annual) using Bayesian adaptive spline surface (BASS) and deep neural network (DNN) algorithms. The study finds that multi-output models can accurately predict energy use while maintaining correlation structures similar to engineering-based models, and highlights the importance of these models for uncertainty and sensitivity analysis. The research contributes to the field by comparing single-output and multi-output models, investigating output correlation, and examining accumulative features in energy predictions.

Uploaded by

ferdie marcos
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 30

buildings

Article
Building Energy Models at Different Time Scales Based on
Multi-Output Machine Learning
Guangchen Li 1 , Wei Tian 1,2, * , Hu Zhang 1 and Bo Chen 3

1 College of Mechanical Engineering, Tianjin University of Science and Technology, Tianjin 300222, China
2 Tianjin International Joint Research and Development Center of Low-Carbon Green Process Equipment,
Tianjin 300222, China
3 College of Economics and Management, Tianjin University of Science and Technology, Tianjin 300457, China
* Correspondence: [email protected]; Tel.: +86-(022)-6060-0705

Abstract: Machine learning techniques are widely applied in the field of building energy analysis
to provide accurate energy models. The majority of previous studies, however, apply single-output
machine learning algorithms to predict building energy use. Single-output models are unable to
concurrently predict different time scales or various types of energy use. Therefore, this paper
investigates the performance of multi-output energy models at three time scales (daily, monthly,
and annual) using the Bayesian adaptive spline surface (BASS) and deep neural network (DNN)
algorithms. The results indicate that the multi-output models based on the BASS approach combined
with the principal component analysis can simultaneously predict accurate energy use at three time
scales. The energy predictions also have the same or similar correlation structure as the energy data
from the engineering-based EnergyPlus models. Moreover, the results from the multi-time scale
BASS models have consistent accumulative features, which means energy use at a larger time scale
equals the summation of energy use at a smaller time scale. The multi-output models at various time
Citation: Li, G.; Tian, W.; Zhang, H.; scales for building energy prediction developed in this research can be used in uncertainty analysis,
Chen, B. Building Energy Models at sensitivity analysis, and calibration of building energy models.
Different Time Scales Based on
Multi-Output Machine Learning. Keywords: building energy prediction; machine learning; multi-output model; time scales
Buildings 2022, 12, 2109. https://
doi.org/10.3390/buildings12122109

Academic Editors: Graziano Salvalai,


Shi-Jie Cao, Antonio Caggiano, 1. Introduction
Wei Yu and Christopher In 2020, the construction and operation of buildings already account for 36% of global
Yu-Hang Chao energy use and 37% of global CO2 emissions [1]. The energy use of the construction industry
Received: 6 November 2022
has a significant impact on the environment and the economy. Hence, it is important to
Accepted: 28 November 2022
reduce the energy use of buildings reasonably and efficiently [2,3]. The key to building
Published: 1 December 2022
energy conservation is to accurately compute building energy use, so as to formulate a
reasonable energy conservation plan according to the characteristics of energy use [4,5].
Publisher’s Note: MDPI stays neutral
Machine learning methods have a wide range of applications in the field of building
with regard to jurisdictional claims in
energy prediction [6,7]. Most researchers use single-output models to predict building
published maps and institutional affil-
energy use in which the different energy outputs are predicted by the separately individual
iations.
models. According to time scales, the research in daily, monthly, and annual time periods
would be discussed in the following three paragraphs, respectively [8].
Firstly, the application of a single-output machine learning model to predict the daily
Copyright: © 2022 by the authors.
energy of buildings is introduced. Liu et al. [9] used the support vector machine (SVM)
Licensee MDPI, Basel, Switzerland. method to predict the daily cooling energy use. The average relative error of the model
This article is an open access article was 5.03%, indicating that the method had excellent prediction accuracy, and the authors
distributed under the terms and proposed an energy use diagnosis method based on energy use prediction. Alobaidi
conditions of the Creative Commons et al. [10] proposed an ensemble artificial neural network (ANN) method with a resampling
Attribution (CC BY) license (https:// technique to predict daily building energy use, which has high generalization ability and
creativecommons.org/licenses/by/ robustness. Hong et al. [11] applied seven machine learning methods including light
4.0/). gradient boosting machine (lightGBM) and random forest (RF) to predict daily electricity

Buildings 2022, 12, 2109. https://doi.org/10.3390/buildings12122109 https://www.mdpi.com/journal/buildings


Buildings 2022, 12, 2109 2 of 30

consumption. The results show that the coefficient of variation of the root mean square
error (CV(RMSE)) of the seven models is less than 10%, in which the lightGBM model has
the best performance with the minimum CV(RMSE) of 4.1%. Jetcheva et al. [12] used the
neural network-based ensemble model method for daily power consumption prediction,
and compared it with the seasonal autoregressive integrated moving average (SARIMA)
model; the results indicate that the neural network model had better accuracy. Ferrantelli
et al. [13] defined an analytical bottom-up model for predicting any daily consumption
given a benchmark daily profile by comparing regression curves obtained with a frequentist
inference to Bayesian inference.
Secondly, the application of a single-output machine learning model in predicting
the monthly energy use of a building is described. Wang et al. [14] proposed a method by
combining a network model with a long-short-term memory learning model to predict
monthly building energy use. Compared with traditional ANN and SVM, the performance
of this model was better; the mean absolute error and root mean square error were 6.66%
and 0.36 kWh/m, respectively. Tran et al. [15] proposed an ensemble model named the
evolutionary neural machine inference model (ENMIM), which consisted of least squares
support vector regression (LSSVR) and the radial basis function neural network (RBFNN).
The model was used to predict the monthly cooling/heating loads and achieved excellent
results. Tian et al. [16] adopted a full linear model to predict monthly power and heating
energy use respectively, and obtained excellent results, providing a reliable statistical
energy model for model calibration. Zhu et al. [17] used five machine learning models
including multivariate adaptive regression splines (MARS) to predict the monthly power
consumption of buildings, which provided a basis for estimating unknown parameters
based on the approximate Bayesian calibration. Koschwitz et al. [18] proposed a novel
recurrent neural network method to predict the monthly thermal loads of buildings, and
the results show that the method is more accurate than support vector machines. Jahani
et al. [19] used the Genetic algorithm-based numerical moment matching (GA-NMM)
method to predict the monthly electricity consumption of buildings. Lin et al. [20] used RF,
SVM, and ANN to predict the monthly electricity consumption of buildings.
Thirdly, the application of a single-output machine learning model to predict the
annual energy use of buildings is presented. Olu-Ajayi et al. [21] used ANN, gradient
boosting (GB), deep neural network (DNN), RF, stacking, K-nearest neighbor (KNN), SVM,
decision tree (DT), and linear regression (LR) methods to predict the annual energy use of
buildings, and the DNN model has the highest prediction accuracy. Tian et al. [22] used
10 machine learning methods to predict the annual energy use of buildings in London, and
conducted variable importance analysis of the learning model and local spatial analysis of
model differences. Tian et al. [23] used five machine learning methods to accurately predict
the annual energy use of buildings for Dempster-Shafer theory uncertainty analysis and
global sensitivity analysis. Building annual energy use forecasts can also use a combination
of machine learning methods and statistical methods to further explore the linear or
nonlinear relationship between annual energy use and certain factors, including building
geometry, occupant, and climate-related factors [24]. A novel multiple linear regression
model can predict building energy use based on changes in climate parameters and the
intensity of the urban heat island effect, which leads to the conclusion that the urban heat
island effect can reduce building energy use in heating-dominated cities [25]. In summary,
energy use at three different time scales (day, month, and year) can be accurately predicted
using different single-output machine learning methods.
There are very limited studies to explore the multi-output models of building energy
and most previous studies use single-output machine learning models. Luo et al. [26]
used heating, cooling, lighting load, and building integrated photovoltaic (BIPV) electrical
power as the output for prediction, and the multi-output model selected SVM and ANN.
The results showed that the ANN model obtained higher accuracy and the SVM model
took less computation time. Liu et al. [27] proposed a multi-input multi-output (MIMO)
strategy based on recurrent neural networks for hourly building energy use prediction. The
Buildings 2022, 12, 2109 3 of 30

absolute percentage error of each time step was analyzed to explore the building’s energy-
saving performance, thus establishing an energy quantification system for the building.
Li et al. [28] developed a MIMO strategy by returning data information in a single-time step,
which avoids accumulated errors compared to recursive strategies and direct strategies.
These previous studies provide useful information on the characteristics of multi-
output models in building energy analysis. However, there is still a lack of sufficient studies
to explore the predictive performance of multi-output machine learning models in building
energy analysis, especially focused on output correlation. Moreover, the accumulative
features are not explored yet in creating multi-output models at various time scales (such
as daily, monthly, and annual).
In order to address these research gaps, this paper is focused on the construction
and performance analysis of multi-output machine learning models for building energy
assessment. Two multi-output machine learning methods are Bayesian adaptive spline
surface (BASS) and deep neural network (DNN). The performance of multi-output models
is evaluated from three aspects: computational time, model accuracy, and output correlation.
The main new contributions of this study are:
(1) This study compares the predictive performance of single-output and multi-output
learning models in building energy analysis. This would provide guidelines on how
to choose the single-output and multi-output models in creating machine learning
models for building energy assessment.
(2) This study explores the performance of two multi-output models (BASS and DNN) in
which the main difference of the two models is whether to maintain output correlation.
This would provide the guidelines on how to choose the learning models with or
without considering output correlation.
(3) The additive or accumulative features are investigated in creating various time scale
models for building energy analysis. This would provide insight on the methods of
obtaining building energy use from a smaller time scale to a larger time scale.
The remaining part of this paper is organized as follows. Section 2 introduces the
research methods, including data acquisition methods, modeling techniques, and perfor-
mance evaluation metrics. In Section 3, the final tuning hyper-parameters from the machine
learning models are first presented. Then, the results and discussion of the multi-output
machine learning model for predicting cooling and heating energy are presented in turn.

2. Method
The research framework based on multi-output machine learning models at different
time scales is shown in Figure 1. The analysis procedure can be divided into four steps: data
preparation, multi-output models, model performance evaluation, and guidelines. The
first step is to prepare the data required to establish the multi-output models as described
in Section 2.1. The second step presents the construction of the two multi-output models–
Bayesian adaptive spline surfaces (BASS) and deep neural networks (DNN). The third step
is model performance evaluation. The fourth step is to provide guidelines for applying the
multi-output models in building energy analysis. All the above steps are implemented in
R language.

2.1. Data Preparation


This section describes the process of acquiring the data required to build machine
learning models. Firstly, an L-shaped four-story office building as shown in Figure 2 is
created as the basic building energy model. The choice of this building type is due to its
simplicity and generalizability compared to other building types (residential or commercial
buildings). Each floor of this model has six peripheral zones and two core zones. The
building model has a total of 32 thermal zones with a floor area of 1600 m2 . The fan coil unit
is set to provide ventilation, heating, and cooling for the building. The air-cooled chiller
and gas boiler provide cold water and hot water, respectively [29]. The internal heat gain
schedule of the building is derived from the China national standards of buildings [30].
Buildings 2022, 12, 2109 4 of 30

The building used in this research is located in Tianjin, China. Therefore, the meteorological
Buildings 2022, 12, x FOR PEER REVIEW
data of
Tianjin in the Chinese standard weather data (CSWD) is used for the calculation 4 of
of 33
the building energy model.

Deterministic parameters Uncertain parameters

EWU,RU, ,HSP,CSP
Step 1 Meteorological HVAC Envelope
Data preparation parameters Parameters parameters
LHS

n1 n2 n2000

Basic building
energy model
Standard building
energy model

MO-format
data set

Step 2 Verify the correlation Bayesian adaptive spline surface


Multi-output model between outputs Deep neural network

Daily Model Monthly Model Multi-time Model

Output Output Output

D1&D2& &Dn D1&D2& &Dn


M1&M2& &M5
M1&M2& &M5 M1&M2& &M5
Ann
Ann Ann

Step 3
Performance evaluation CV(RMSE),MAPE,R² Computational time Dendrogram

Correlation
Model accuracy Model efficiency
between outputs

Model choice

Step 4 Output correlation Accumulative feature


• Single output model • Bayesian adaptive spline surface • Direct sum up
Guidelines • Multi-out model • Deep neural network • Multi-output

Figure
Buildings 2022, 12, x FOR PEER REVIEW
Figure 1. flow
1. A A flow chart
chart of multi-output
of multi-output models
models of building
of building energyenergy at monthly,
at daily, daily, monthly, and5scales
and annual annual
of 33
scales
based onbased on machine
machine learning.learning.

2.1. Data Preparation


This section describes the process of acquiring the data required to build machine
learning models. Firstly, an L-shaped four-story office building as shown in Figure 2 is
created as the basic building energy model. The choice of this building type is due to its
simplicity and generalizability compared to other building types (residential or commer-
cial buildings). Each floor of this model has six peripheral zones and two core zones. The
building model has a total of 32 thermal zones with a floor area of 1600 m2. The fan coil
unit is set to provide ventilation, heating, and cooling for the building. The air-cooled
chiller and gas boiler provide cold water and hot water, respectively [29]. The internal
heat gain schedule of the building is derived from the China national standards of build-
Figure 2. AA four-story
ings 2.[30].
Figure four-story officebuilding.
The building
office building.
used in this research is located in Tianjin, China. Therefore, the
meteorological data of Tianjin in the Chinese standard weather data (CSWD) is used for
theAA basic building
calculation of theenergy
building energy model
buildingmodel isis
energy constructed
constructed
model. based
basedon on
deterministic parameters,
deterministic and
parameters,
and
thenthen uncertain
uncertain parameters
parameters usingusing
the the Latin
Latin Hypercube
Hypercube sampling
sampling (LHS)
(LHS) method
method are are
im-
imported
ported into into thebasic
the basicmodel
modeltotoobtain
obtainaa complete
complete standard
standard building energy
energy model.
model. The
The
ranges
ranges of of uncertain
uncertain parameters
parameters areare shown
shown inin Table
Table 1,
1, which
which are
are derived
derived from
from previous
previous
studies [16,17] and Chinese national standards [30,31]. The building envelope
studies [16,17] and Chinese national standards [30,31]. The building envelope parameters parameters
include
include thethe U-values
U-values ofofexternal
externalwalls,
walls,windows,
windows,andandroofs.
roofs. Internal
Internal building
building heat
heat gains
gains
include light power density and equipment power density. Heating and cooling set-point
temperatures are also considered in this study. The construction of the modeling data set
is divided into two steps. The first step is to perform Latin hypercube sampling within the
parameter ranges to obtain 2000 sets of parameter combinations and then generate 2000
Buildings 2022, 12, 2109 5 of 30

include light power density and equipment power density. Heating and cooling set-point
temperatures are also considered in this study. The construction of the modeling data set is
divided into two steps. The first step is to perform Latin hypercube sampling within the
parameter ranges to obtain 2000 sets of parameter combinations and then generate 2000 idf
(EnergyPlus Input Files) format files. The 2000 sets of energy use data at different time
scales are obtained by running the EnergyPlus (V22.1.0) program, in which 1000 sets of
data are used for model training, 500 sets are used for model validation, and 500 sets are
used for model testing. The second step is data preprocessing to create the daily, monthly,
and annual energy models. The cooling time periods for this office building is from May 1st
to September 30th, and the heating time period is from November 1st to March 31st. The
heating and cooling energy on weekends and holidays in this office building is zero and
thus excluded from this analysis. As a result, the numbers of daily and monthly cooling
energy data are 105 and 5, respectively. The numbers of daily and monthly heating energy
data are 102 and 5, respectively.

Table 1. Variation of building input parameters.

No. Parameters Short Names Range Unit


1 Exterior wall U-value EWU 0.1–0.25 W/(m2 K)
2 Roof U-value RU 0.15–0.3 W/(m2 K)
3 Window U-value WU 1–2.4 W/(m2 K)
4 Solar heat gain coefficient SHGC 0.2–0.48 -
5 Infiltration rate INF 0.5–0.8 ACH
6 Lighting power density LPD 5–10 W/m2
7 Equipment power density EPD 9–15 W/m2
8 Occupancy density OPD 9–14 2
m /person
9 Heating set-point HSP 20–22 ◦C
10 Cooling set-point CSP 24–26 ◦C

2.2. Multi-Output Models


This section describes the construction process of multi-output machine learning
models. The Pearson correlation coefficient (PCC) is used to assess the correlation between
the output data before the model is established. The computation of the Pearson correlation
coefficient is as follows:
 
∑ni=1 (ai − a) bi − b
r =q r  2 (1)
n 2 n
∑i=1 (ai − a) ∑i=1 bi − b

where r represents the Pearson correlation coefficient, ai represents the ith sample a, bi
represents the ith sample b, a represents the average of sample a, and b represents the
average of sample b. The PCC value is between −1 and 1. The PCC is close to 1 to indicate
that there is a strong positive correlation. In contrast, a strong negative correlation can be
suggested by the PCC value close to −1.
Two types of machine learning methods– BASS and DNN–are chosen based on these
capability of multi-output models.
The BASS method is similar to Bayesian multivariate adaptive regression splines
used to learn a set of basis functions from historical data, which are tensor products of
polynomial splines. The BASS model can adaptively select the number of basis functions
and adjust the number of variables and nodes in each basis. While most hyper-parameters
of the BASS model are determined automatically, the priors in the BASS model can be
modified to reduce overfitting. The main adjusted parameters include the number of
principal component analysis n.pc and the number of iterations nmcmc. The principal
component analysis (PCA) method is used to calculate the principal components of multiple
outputs and then fit a regression model. The R-BASS package (V1.2.2) is used to implement
Buildings 2022, 12, 2109 6 of 30

the multi-output machine learning technique in this research. The advantages of using this
BASS package include scalable capabilities to handle large numbers of observations and
predictions. In addition, not only regression models but also classification models can be
created using the BASS package [32].
The DNN method is one of the current mainstream methods for analyzing building
energy, including an input layer, multiple hidden layers, and an output layer. The deep
feedforward neural network is selected in this research, and neuron signals are transmitted
forwardly between layers without feedback signal transmission. In other words, each
layer can only receive the signal of the previous layer, and then pass it to the next layer.
The signal transfer between layers depends on the connection of nonlinear functions.
By determining the weight and bias terms of the function, the signal of the previous
layer is effectively transmitted to the next layer. It is worth noting that there is no signal
transmission between neurons in the same layer of a feedforward neural network [33]. The
main hyper-parameters of the feedforward neural network include the number of hidden
layers, the number of hidden layer neurons, the loss function, and the number of iterations.
The DNN parameter adjustment method generally adopts empirical adjustment. The keras
and tensorflow packages in the R environment are used to implement the establishment of
DNN models.

2.3. Performance Evaluation


The predictive performance of learning models is evaluated using three indicators:
CV(RMSE) (coefficient of variation of the root mean square error), MAPE (mean absolute
percentage error), and R2 (coefficient of determination). The calculation formulas for these
three indicators are as follows:
q 2
∑in=1 (yi −ŷi )
n
CV(RMSE) = 1 n
(2)
n ∑ i =1 y i

1 n yi − ŷi
n ∑ i =1
MAPE = × 100% (3)
yi

n (yi − ŷi )2
R2 = 1 − ∑i=1 (4)
( y i − y i )2
where n represents the sample size, i represents the ith sample, yi represents the ith true
value, and ŷi represents the ith predicted value. The CV(RMSE), MAPE, and R2 values are
all dimensionless numbers, which are not affected by the order of magnitude of the data
and can more effectively express the accuracy of the model. CV(RMSE) and MAPE reflect
the error of the model–the smaller, the better. The value of R2 ranges from 0 to 1, which
reflects the relative error of the model relative to the direct average value. The closer to 1,
the better.
The computational time of creating machine learning models is also an important
indicator to assess the performance of learning algorithms.. The computing time includes
the modeling time and the prediction time when using the machine learning techniques in
building energy analysis. The computing time is counted using the system.time function in
the R language environment. Simulations were performed on a workstation with an Intel
Xeon CPU (E5-2650 2.3 GHz) and 64 GB RAM.
For the multi-output learning models, it is necessary to investigate whether the ma-
chine learning model maintains the correlation structure between the outputs of the energy
data from the training set. In this research, the cluster dendrogram method is used for
model output correlation analysis [34], which can hierarchically cluster multidimensional
data to aggregate highly similar data into one category. Through this dendrogram, the
clusters and the number of objects belonging to each cluster can be determined [35]. The
dendrogram is plotted by computing the distances between clusters according to the Lance–
Williams dissimilarity update formula by obtaining the dissimilarities prior to forming
Buildings 2022, 12, 2109 7 of 30

the new cluster. The complete linkage method, one of hierarchical classification methods
that uses the distance between the most distant elements from each cluster, is used for
hierarchical clustering [36]. The cluster analysis in this research is implemented using the
R package tidyverse and hclust functions.

3. Results and Discussion


3.1. Results of Model Hyperparameter Tuning
Table 2 shows the optimal hyper-parameters for machine learning models. The SO-
BASS and MO-BASS denote the single-output models and multiple-output models, respec-
tively, based on the BASS technique. The same rules are also applicable to the SO-DNN
and MO-DNN models. In the BASS models, the degree represents the degree of interaction
between the input parameters. The degree of interaction between the input parameters of
daily cooling energy is lower than that of daily heating energy. For the MO-BASS model,
the n.pc represents the number of principal components used. The daily cooling energy
use can be fitted with fewer principal components than the daily heating energy while
still producing good results. Therefore, it is easier to predict daily cooling energy use. The
nmcmc represents the number of iterations, and all BASS models can achieve excellent
accuracy after 10,000 iterations. Note that the hyper-parameters for cooling and heating
energy are the same for month-to-month and multi-time scale energy models, indicating
the same difficulty in predicting cooling and heating energy use.
The final hyper-parameters from the DNN models are also listed in Table 2. For the
SO-DNN model, there is only one neuron in the output layer. The three layers of the SO-DNN
hidden layer are selected in the daily and monthly energy use prediction. The activation
function of monthly energy use is different from that of daily energy use. For the MO-DNN
model, the output layer neurons should be adjusted according to the number of outputs. For
example, if the building has 105 cooling days throughout the year, then the MO-DNN must
have 105 output layer neurons. The number of hidden layers selected by the MO-DNN model
is four layers for daily energy use and three layers for monthly energy use.

Table 2. The optimal tuning parameters from machine learning models.

Daily Monthly Multi-Time


Model Hyperparameters
Cooling Heating Cooling Heating Cooling Heating
degree 3 4 4
SO-BASS -
nmcmc 10,000
n.pc 7 10 5 7
MO-BASS degree 3 4 4 3
nmcmc 10,000
activation tanh,relu,liner tanh,elu,relu,liner
number of -
SO-DNN 3
hidden layers
output layer neurons 1
activation tanh,elu,relu,liner
MO-DNN number of hidden layers 4 3 4
output layer neurons 105 102 5 111 108

3.2. Results of Multi-Output Cooling Energy Models


Section 3.2.1 presents the results of daily cooling energy models and Section 3.2.2
describes the results of monthly cooling energy models. Section 3.2.3 discusses the results
of multi-time scale models that simultaneously predict daily, monthly, and annual cooling
energy. Section 3.2.4 presents the results from the models in which monthly and annual
energy can be directly or accumulatively obtained.
Buildings 2022, 12, 2109 8 of 30

3.2.1. Daily Cooling Energy Models


The correlation coefficient of daily cooling energy in the work days from the training
set of size 1000 is shown in Figure 3. The correlation coefficients are all above 0.8, and
most of them are above 0.9, as is evident from the color labels in Figure 3. As a result,
daily cooling energy use in this office building is strongly positively correlated, hence, it is
Buildings 2022, 12, x FOR PEER REVIEW 9 of 33
important to include it in the correlation structure account when creating machine learning
models for daily cooling energy use.

Figure3.3.Correlation
Figure Correlationcoefficient
coefficientofofdaily
dailycooling
coolingenergy
energyuse
useininJuly
July(1,
(1,1 1July;
July;. .…
. ; ;31,
31,31
31July).
July).

Figure
Figure44shows
showsthe thepredicted
predictedperformance
performanceof ofthe
themulti-output
multi-outputdailydailyenergy
energycooling
cooling
model. Figure 4a demonstrates that the four models of SO-BASS,
model. Figure 4a demonstrates that the four models of SO-BASS, MO-BASS, MO-BASS, SO-DNN, and
SO-DNN,
MO-DNN
and MO-DNN have have
median CV(RMSE)
median CV(RMSE) of 0.008, 0.006,0.006,
of 0.008, 0.008,0.008,
and 0.015, respectively.
and 0.015, For the
respectively. For
day with the worst prediction accuracy of the four models (SO-BASS,
the day with the worst prediction accuracy of the four models (SO-BASS, MO-BASS, SO- MO-BASS, SO-DNN,
and
DNN,MO-DNN), the largest
and MO-DNN), theCV(RMSE) values are
largest CV(RMSE) 0.059,are
values 0.027, 0.035,
0.059, and0.035,
0.027, 0.038,and
respectively.
0.038, re-
The median and maximum values of the CV(RMSE) from the four models indicate that
spectively. The median and maximum values of the CV(RMSE) from the four models in-
the MO-BASS model performs the best in this case study. Figure 4b shows that the MO-
dicate that the MO-BASS model performs the best in this case study. Figure 4b shows that
BASS model has the smallest median MAPE, followed by SO-BASS and SO-DNN. The
the MO-BASS model has the smallest median MAPE, followed by SO-BASS and SO-DNN.
MO-BASS model has better predictive performance due to the small variations of MAPE
The MO-BASS model has better predictive performance due to the small variations of
in comparison with the SO-BASS although the median MAPE values of the two BASS
MAPE in comparison with the SO-BASS although the median MAPE values of the two
models are similar. This is because the MO-BASS model considers the output correlation
BASS models are similar. This is because the MO-BASS model considers the output cor-
to avoid the extreme errors from the single-output models. Hence, the MO-BASS model
relation to avoid the extreme errors from the single-output models. Hence, the MO-BASS
has the best performance in terms of the MAPE. Figure 4c indicates that the coefficients of
model has the best performance in terms of the MAPE. Figure 4c indicates that the coeffi-
determination (R2 ) are very2 high, except for three data points in SO-BASS. Most of the R2
cients of determination (R ) are very high, except for three data points in SO-BASS. Most
values are l greater than 0.9 to indicate that all these four models have good performance.
of the R2 values are l greater than 0.9 to indicate that all these four models have good
Among them, the two models with the largest R2 are MO-BASS and SO-DNN. Further
performance.
comparison of Among
the median them,andthe two modelsrange
interquartile withshows
the largest R2 R
that the are MO-BASS
2 from and SO-
the MO-BASS
DNN.isFurther
model comparison
larger than of the median
that of SO-DNN. Fromand interquartile
the above analysis,range shows thatmodel
the MO-BASS the R²isfrom
the
the MO-BASS model is larger than that
best daily cooling energy model in this case study. of SO-DNN. From the above analysis, the MO-
BASS model
Table is the
3 lists the best daily cooling
computational timeenergy model in this
for single-output andcase study.
multi-output models of cool-
ing energy for BASS and DNN algorithms at different time scales models. The comparison
of computational time for daily cooing energy use will be described in this section. The
discussion of computational time for monthly and multi-time scales would be presented in
Sections 3.2.2 and 3.2.3, respectively. The computation time for the multi-output models is
less than the computation time for the single-output models for daily cooling energy use.
For the BASS models, the modeling time of the multi-output BASS model for daily cooling
energy is 19.6 times that of the single-output BASS model. This is due to the fact that the
output number in the BASS multi-output model would be same as the principal component
the day with the worst prediction accuracy of the four models (SO-BASS, MO-BASS, SO-
DNN, and MO-DNN), the largest CV(RMSE) values are 0.059, 0.027, 0.035, and 0.038, re-
spectively. The median and maximum values of the CV(RMSE) from the four models in-
dicate that the MO-BASS model performs the best in this case study. Figure 4b shows that
the MO-BASS model has the smallest median MAPE, followed by SO-BASS and SO-DNN.
Buildings 2022,The MO-BASS model has better predictive performance due to the small variations of
12, 2109 9 of 30
MAPE in comparison with the SO-BASS although the median MAPE values of the two
BASS models are similar. This is because the MO-BASS model considers the output cor-
relation to avoid the extreme errors from the single-output models. Hence, the MO-BASS
number, which is much less than the original correlated outputs in the single-output models.
model has the best performance in terms of the MAPE. Figure 4c indicates that the coeffi-
The modeling time of the multi-output DNN model is 27.7 times that of the single-output
cients of determination (R2) are very high, except for three data points in SO-BASS. Most
DNN model for daily cooling energy. Hence, adopting the multi-output models can signifi-
of the R2 values are l greater than 0.9 to indicate that all these four models have good
performance. Among them, the two models time
cantly reduce modeling with and increase
the largest R2the
aremodel’s
MO-BASS efficiency,
and SO-presuming the accuracy
of the model is acceptable. The DNN models require
DNN. Further comparison of the median and interquartile range shows that the R² more computational
from cost compared
with the BASS models. For the single-output models,
the MO-BASS model is larger than that of SO-DNN. From the above analysis, the MO- the computational time for the DNN
model
BASS model is the best is cooling
daily almost energy
two times thatinofthis
model thecase
computational
study. cost for the BASS model.

22, 12, x FOR PEER REVIEW 10 of 33

(a) CV(RMSE)

(b) MAPE

(c) R²
Figure 4. Prediction performance of machineperformance
Figure 4. Prediction learning models for dailylearning
of machine cooling energy
models(SO,
for single
daily cooling energy (SO, single
output; MO, multiple output).
output; MO, multiple output).

Table 3 lists the computational time for single-output and multi-output models of
cooling energy for BASS and DNN algorithms at different time scales models. The com-
parison of computational time for daily cooing energy use will be described in this section.
The discussion of computational time for monthly and multi-time scales would be pre-
sented in Sections 3.2.2 and 3.2.3, respectively. The computation time for the multi-output
Buildings2022,
Buildings 2022,12,
12,2109
x FOR PEER REVIEW 11 10
ofof3330

Table3.3.The
Table Thecomputational
computationaltime
time of
of different
different time-scale
time-scale cooling
cooling energy
energy models
models (unit:
(unit: s).
s).
Models
Models Daily Daily Monthly
Monthly Multi-Time
Multi-TimeScale
Scale
SO-BASS 2032.2 98.7 -
SO-BASS 2032.2 98.7 -
MO-BASS
MO-BASS 103.9 103.9 81.181.1 197.2
197.2
SO-DNN
SO-DNN 3807.2 3807.2 361.7
361.7 - -
MO-DNN
MO-DNN 137.3 137.3 72.272.2 236.3
236.3

Figure 55 shows
Figure shows the
the correlated
correlated structure
structure treetree diagram
diagramfrom fromthe thetraining
trainingset setand
andthethefour
four
machine learning models on the daily cooling energy in July. The
machine learning models on the daily cooling energy in July. The numbers in Figure 5 numbers in Figure 5 repre-
sent a specific
represent day in
a specific July
day inin which
July the weekends
in which the weekendsand the andholidays have been
the holidays haveexcluded since
been excluded
there is no cooling energy use. For example, 1 represents the first day of
since there is no cooling energy use. For example, 1 represents the first day of July. From July. From the top of
Figure 5a, there are two groups of data in which the left tree has 14 data
the top of Figure 5a, there are two groups of data in which the left tree has 14 data and and the right tree has
nineright
the data.tree
Thehas
group
ninecandata.
be further divided
The group canintobesub-groups based on
further divided intothe similarity ofbased
sub-groups cooling
on
the similarity of cooling energy use. The cooling energy in the 3rd and 22nd are theofmost
energy use. The cooling energy in the 3rd and 22nd are the most similar as the height the
link thatasjoins
similar thethem together
height of theislink
the smallest
that joins in them
Figuretogether
5a. By comparing Figure 5a
is the smallest in with
Figurethe5a.
other
By
four dendrograms, Figure 5a–c have the most similar clustering distributions.
comparing Figure 5a with the other four dendrograms, Figure 5a–c have the most similar Thus, the MO-
BASS model
clustering can properly maintain
distributions. Thus, thethe output correlation
MO-BASS model can of the originalmaintain
properly training set thefrom the
output
EnergyPlus models. Figure 5b,d,e show that there are significant differences
correlation of the original training set from the EnergyPlus models. Figure 5b,d,e show from the training
set. there
that Therefore, the MO-BASS
are significant model can
differences from bethe
more in lineset.
training with the training
Therefore, the data set from
MO-BASS the
model
engineering-based building energy models from the perspective of the
can be more in line with the training data set from the engineering-based building energy correlation structure
of building
models fromcooling energy use.of the correlation structure of building cooling energy use.
the perspective

(a) (b) (c)

(d) (e)
Figure5.5. Dendrogram
Figure Dendrogram of
of July
July daily
daily cooling
cooling energy
energy for
for the
the training
training set
setand
andfour
fourmachine
machinelearning
learning
models.
models.

3.2.2. Monthly Cooling Energy Models


The correlation coefficients of monthly cooling energy are shown in Figure 6. All
correlation coefficients are above 0.98, indicating that there is a strong correlation between
monthly cooling energy. Therefore, it is of great significance to create multi-output machine
learning models for monthly cooling energy.
3.2.2. Monthly Cooling Energy Models
The correlation coefficients of monthly cooling energy are shown in Figure 6. All cor-
relation coefficients are above 0.98, indicating that there is a strong correlation between
Buildings 2022, 12, 2109 11 of 30
monthly cooling energy. Therefore, it is of great significance to create multi-output ma-
chine learning models for monthly cooling energy.

Figure 6.
Figure 6. Correlation
Correlation coefficient
coefficient of
of monthly
monthlycooling
coolingenergy.
energy.

Figure
Figure 77 shows
shows thethe predicted
predicted performance
performance of of four
four monthly
monthly cooling
cooling energy
energy machine
machine
learning
learning models.
models. The
The MO-BASS
MO-BASS model model has
has the
the smallest
smallest CV(RMSE), followed by
CV(RMSE), followed by the
the SO-
SO-
BASS model. The CV(RMSE) of both BASS models is below 0.005,
BASS model. The CV(RMSE) of both BASS models is below 0.005, indicating that these indicating that these
two
two models
models have
have good
good performance.
performance. The The CV(RMSE)
CV(RMSE) valuesvalues of
of the
the SO-DNN
SO-DNN modelsmodels are
are
between
between 0.075 and 0.025, slightly larger than the CV(RMSE) of the two BASS models. From
0.075 and 0.025, slightly larger than the CV(RMSE) of the two BASS models. From
Figure
Figure 7b,
7b, similar
similar conclusions
conclusions can can bebe drawn
drawn asas in
in Figure
Figure 7a.7a. Finally,
Finally, itit can
can be
be discovered
discovered
from Figure 7c that all four models have R 2 values higher than 0.93, with two BASS models
from Figure 7c that all four models have R²values higher than 0.93, with two BASS models
having 2
having RR²values
values close
close to 1. In
to 1. In summary,
summary, the best performance
the best performance models
models forfor monthly
monthly cooling
cooling
energy prediction are the MO-BASS in this
energy prediction are the MO-BASS in this case study.case study.
The modeling time of the four monthly cooling energy machine learning models
is listed in Table 3. The modeling time of the two single-output models (SO-BASS and
SO-DNN) is higher than that of the two multi-output models. For the monthly cooling
energy model, the modeling time of the BASS single/multi-output model has no significant
differences since the number of monthly cooling energy data is only 5. There is still a
marked difference in the modeling time of the two DNN models, and the multi-output
model is around five times that of the single-output model. Overall, the multi-output
monthly cooling energy models are more time-saving compared to the single-output
monthly cooling energy models, especially for the DNN models.
Figure 8 shows the correlation structure tree diagram of five monthly energy use
models. Figure 8a shows the dendrogram of engineering-based building energy models,
and the adjacent months of July and August have the highest correlation. Hence, the
hottest two months have similar patterns of cooling energy from the air-conditioning
system in this office building. The two transitional months, May and September, are closely
correlated because they have similar climatic conditions including outdoor temperature
and solar radiation. The two BASS models for monthly cooling energy (Figure 8b,c) can
(a) CV(RMSE)
both retain the same correlation structure as the engineering-based model. In contrast, the
two monthly cooling DNN models (Figure 8d,e) are unable to maintain the same correlation
as the patterns illustrated in Figure 8a. As a result, the multi-output BASS models can
maintain inter-output correlation when predicting both month-by-month and day-by-day
cooling energy in this office building as shown in Figures 5 and 8.
two models have good performance. The CV(RMSE) values of the SO-DNN models are
between 0.075 and 0.025, slightly larger than the CV(RMSE) of the two BASS models. From
Figure 7b, similar conclusions can be drawn as in Figure 7a. Finally, it can be discovered
from Figure 7c that all four models have R²values higher than 0.93, with two BASS models
Buildings 2022, having
12, 2109 R²values close to 1. In summary, the best performance models for monthly cooling 12 of 30

energy prediction are the MO-BASS in this case study.

2, 12, x FOR PEER REVIEW 13 of 33

(a) CV(RMSE)

(b) MAPE

(c) R²
Figure 7. Prediction performance
Figure of performance
7. Prediction four machineoflearning models
four machine for monthly
learning cooling
models for energy
monthly (SO,
cooling energy (SO,
single output; MO, multiple
single output).
output; MO, multiple output).

The modeling time of the four monthly cooling energy machine learning models is
listed in Table 3. The modeling time of the two single-output models (SO-BASS and SO-
DNN) is higher than that of the two multi-output models. For the monthly cooling energy
model, the modeling time of the BASS single/multi-output model has no significant dif-
Buildings 2022, 12, x FOR PEER REVIEW 14 of 33
Buildings 2022, 12, 2109 13 of 30

(a) (b) (c)

(d) (e)
Figure 8.
Figure 8. Dendrogram
Dendrogramofofmonthly
monthly cooling
cooling energy
energy fortraining
for the the training set
set and and
four four machine
machine learninglearning
models.
models.
3.2.3. Multi-Time Scale Cooling Energy Models
3.2.3.The
Multi-Time Scale
prediction Cooling Energy
performance of the Models
multi-time scale cooling energy models is shown
in Figure 9. The features
The prediction of these of
performance multi-time scale models
the multi-time can predict
scale cooling energy the cooling
models is energy
shown
at
in daily,
Figuremonthly, and annual
9. The features timemulti-time
of these scales, simultaneously.
scale models can First,predict
the performance
the coolingof daily
energy
cooling
at daily,energy
monthly, prediction
and annual from the scales,
time multi-time scale models
simultaneously. is analyzed.
First, The medians
the performance of
of daily
CV(RMSE),
cooling energy MAPE, and R2from
prediction of MO-BASS and MO-DNN
the multi-time scale modelsmodels are 0.007The
is analyzed. andmedians
0.02, 0.005 of
and 0.016, 0.99
CV(RMSE), and and
MAPE, 0.97,R²
respectively.
of MO-BASS Therefore,
and MO-DNN the multi-time
models are scale MO-BASS
0.007 and 0.02,model0.005
has
and a0.016,
better performance
0.99 in predicting
and 0.97, respectively. daily cooling
Therefore, energy. Note
the multi-time scalethat the performance
MO-BASS model has
of
a better performance in predicting daily cooling energy. Note that the performanceof
MO-BASS and MO-DNN models is similar in the case of individual prediction ofdaily
MO-
cooling energy as described in Section 3.2.1. Second, the analysis
BASS and MO-DNN models is similar in the case of individual prediction of daily cooling of the multi-time scale
model
energyisasfocused on the
described monthly3.2.1.
in Section cooling energy
Second, theforecast.
analysisThe CV(RMSE)
of the multi-timeandscale
MAPE values
model is
from
focusedtheonMO-BASS models
the monthly are smaller
cooling energythan thoseThe
forecast. from the MO-DNN
CV(RMSE) models,values
and MAPE and the fromR2
values of MO-BASS
the MO-BASS models areare
larger thanthan
smaller those fromfrom
those the MO-DNN
the MO-DNN models. Hence,
models, andthe
theMO-BASS
R²values
model still performs better than the MO-DNN model for the
of MO-BASS are larger than those from the MO-DNN models. Hence, the MO-BASS monthly cooling energy. Third,
the model performance of annual cooling energy is analyzed.
model still performs better than the MO-DNN model for the monthly cooling energy. The CV(RMSE) and MAPE
values from
Third, the the MO-BASS
model performance models are close
of annual to 0,energy
cooling and the is corresponding R2 is close to
analyzed. The CV(RMSE) and1.
The performance of MO-DNN is worse than that of MO-BASS.
MAPE values from the MO-BASS models are close to 0, and the corresponding R²is close It is worth noting that the
prediction accuracy of of
to 1. The performance theMO-DNN
multi-time is scale
worsemodel increases
than that with theItincrease
of MO-BASS. is worthinnoting
time scale.
that
This is becauseaccuracy
the prediction the data complexity decreases
of the multi-time scalewith
modelthe increases
increase inwithtimethescale. In summary,
increase in time
the MO-BASS model performs better in predicting multiple time scales compared to the
scale. This is because the data complexity decreases with the increase in time scale. In
MO-DNN model.
summary, the MO-BASS model performs better in predicting multiple time scales com-
Table 3 lists the computational time for multi-time scale models of BASS and DNN. The
pared to the MO-DNN model.
MO-DNN model requires 236.3 s, while the MO-BASS model needs 197.2 s. The computation
time MO-DNN is 20% more than that of the MO-BASS model. As a result, the MO-BASS
offers both high accuracy and low modeling costs in comparison with the MO-DNN model
for the simultaneous prediction of daily, monthly, and annual cooling energy.

3.2.4. Performance Analysis of 10 Models for Monthly and Annual Cooling Energy
This section compares the predictive performance and accumulative characteristics
of monthly and annual cooling energy from 10 models as listed in Table 4. The model
Buildings 2022, 12, 2109 14 of 30

names in Table 4 include three parts. The first two letters denote the single-output (SO)
or multi-output (MO) model. The third letter denotes the machine learning algorithm, in
which the B is the Bayesian adaptive spline surface (BASS) and the D is the deep neural
network (DNN). The letters after the hyphen denote the time scale of energy models. The
D is the daily model, the M is the monthly model, and the Mu is the multi-time (daily,
monthly, and annual) scale model. For example, the MOB-Mu is to simultaneously predict
the daily, monthly, and annual energy from the multi-time scale multi-output BASS model.
Note that there are two ways to obtain monthly or annual energy: direct and accumulation.
For instance, the MOD-M or SOD-M models can directly predict the monthly energy based
on the deep neural network models. In contrast, the SOB-D model needs to sum the daily
energy from the single-output daily BASS models to obtain the monthly (or annual) energy
since there is unavailable for monthly (or annual) energy use from the daily BASS model.
The predictive performance of 10 models for monthly cooling energy is shown in
Figure 10. The performance of monthly cooling energy from the direct prediction is better
than those from the accumulative models for the BASS approaches in terms of CV(RMSE),
MAPE, and R2 . The SOB-M performs better than the SOB-D and the MOB-M has better
predictive capability compared to the MOB-D. The multi-time scale BASS model has very
good performance compared to the monthly and daily BASS models for cooling energy
prediction. As for the DNN models, the best prediction model of cooling energy is the
monthly summation from the daily SOD-D models. The multi-time scale MOD-Mu model
has moderate performance compared to the other four DNN models. The BASS models
have better prediction performance in comparison with the DNN models for monthly
cooling energy. All the R2 values for the BASS models are above 0.99, which indicates that
the BASS has very high predictive capability for monthly cooling energy.

Table 4. Description of 10 models for monthly or annual energy prediction.

Machine Learning Model Description


Sum the daily energy from the single-output daily BASS models to obtain the monthly or
SOB-D
annual energy
Sum the daily energy from the multi-output daily BASS models to obtain monthly or
MOB-D
annual energy
BASS Monthly predictions or annual prediction (sum of monthly predictions) from the
SOB-M
single-output monthly BASS models
Monthly predictions or annual prediction (sum of monthly predictions) from the
MOB-M
multi-output monthly BASS models
MOB-Mu Monthly or annual predictions from the multi-output multi-time scale BASS models
Sum the daily energy from the single-output daily DNN models to obtain the monthly or
SOD-D
annual energy
Sum the daily energy from the multi-output daily DNN models to obtain monthly or
MOD-D
annual energy
DNN Monthly predictions or annual prediction (sum of monthly predictions) from the
SOD-M
single-output monthly DNN models
Monthly predictions or annual prediction (sum of monthly predictions) from the
MOD-M
multi-output monthly DNN models
MOD-Mu Monthly or annual predictions from the multi-output multi-time scale DNN models

It is also interesting to compare the accumulative features for the different time scale
models. As might be expected, the results from the monthly summation of daily cooling
energy models are different from the direct prediction of monthly cooling energy models,
even for the same machine learning algorithm. The computational results from this case
study confirm this statement. For the BASS models, there are five methods to obtain
monthly energy use, which may make it confusing regarding which monthly results should
be selected in the multiple time scale analysis. For instance, if the best daily model is from
the daily BASS single-output models and the best monthly model is from the monthly
BASS single-output model in terms of prediction accuracy, then the monthly energy use
Buildings 2022, 12, 2109 15 of 30

would not be the same as the summation of daily energy use in a specific month. The
only model that can maintain good accumulative features is the multi-time scale BASS
model (MOB-Mu), due to the processing of the principal component analysis. The monthly
2022, 12, x FOR PEER REVIEW cooling energy in July would equal the summation of the July15daily
of 33 prediction of cooling
energy for the MOB-Mu model.

(a) CV(RMSE)

(b) MAPE

(c) R²
Figure 9. PredictionFigure
performance of two performance
9. Prediction machine learning models
of two for learning
machine multi-time scalesfor
models cooling en- scales cooling energy.
multi-time
ergy.
Buildings 2022, 12, 2109 16 of 30

Buildings 2022, 12, x FOR PEER REVIEW 17 of 33


Figure 11 demonstrates the predictive performance of 10 models for the annual cooling
energy use. The meaning of these 10 models is described in Table 4. The three best models
are MOB-Mu,
good performance MOB-M,
compared and
to the MOB-D,
monthly which
and dailyallBASS
belong to thefor
models multi-output BASS models.
cooling energy
The next three models are SOB-M, SOB-D,
prediction. As for the DNN models, the best prediction and SOD-D, which all belong to the
model of cooling energy is the single-output
models. All these six models have very high R 2 values, above 0.997. The worst model is the
monthly summation from the daily SOD-D models. The multi-time scale MOD-Mu model 2
single-out
has moderate monthlycompared
performance DNN models (SOD-M),
to the in which
other four the CV(RMSE),
DNN models. The BASSMAPE, and R values
models
areprediction
have better 0.016, 0.013, and 0.970, respectively.
performance in comparison Thewith
remaining ninemodels
the 2DNN models’forCV(RMSE)
monthly values are
cooling energy. All the R values for the BASS models are above 0.99, which indicates0.995,
less than 0.008, MAPE
2 values less than 0.0078, and R values greater than that indicating
that these models have good predictive performance.
the BASS has very high predictive capability for monthly cooling energy.

(a) CV(RMSE)

(b) MAPE

(c) R²
Figure 10. Figure
Performance of 10 models
10. Performance of 10for monthly
models cooling energy
for monthly cooling (refer
energyto(refer
Tableto4Table
for model expla-explanation).
4 for model
nation).
models are MOB-Mu, MOB-M, and MOB-D, which all belong to the multi-output BASS
models. The next three models are SOB-M, SOB-D, and SOD-D, which all belong to the
single-output models. All these six models have very high R2 values, above 0.997. The
worst model is the single-out monthly DNN models (SOD-M), in which the CV(RMSE),
MAPE, and R²values are 0.016, 0.013, and 0.970, respectively. The remaining nine models’
Buildings 2022, 12, 2109 17 of 30
CV(RMSE) values are less than 0.008, MAPE values less than 0.0078, and R²values greater
than 0.995, indicating that these models have good predictive performance.

(a) CV(RMSE)

Buildings 2022, 12, x FOR PEER REVIEW 19 of 33

(b) MAPE

(c) R²
Figure 11.11.
Figure Performance ofof
Performance 1010models
modelsfor
forannual
annualcooling
cooling energy (referto
energy (refer toTable
Table44for
formodel
modelexplanation).
explana-
tion).
For the sake of accumulative features, the energy use prediction at a larger scale
canForbethe
obtained
sake offrom the summation
accumulative of the
features, small-scale
energy use energy use foratthe
prediction single-output
a larger scale can or
bemulti-output
obtained from models. However,of
the summation this simple accumulation
small-scale energy use for is not
thenecessarily
single-outputa good method
or multi-
in predicting
output models. the multi-time
However, this scale
simple energy use in buildings.
accumulation There are at
is not necessarily least three
a good method reasons
in
for this. The
predicting the first reason scale
multi-time is thatenergy
the residual
use inerrors from the
buildings. small
There arescale models
at least threemay lead to
reasons
foruncontrolled
this. The firsterrors.
reasonIfisallthat
thethe
residual errors
residual from
errors fromthethe
small scale
small models
scale models aremay
positive
lead or
negative, the final residuals at a large scale can be added to larger residual
to uncontrolled errors. If all the residual errors from the small scale models are positive errors. If there
are negative and positive errors for the small time scale models, the
or negative, the final residuals at a large scale can be added to larger residual errors. Ifresidual errors at a
larger scale may be small due to the offsetting of negative and positive
there are negative and positive errors for the small time scale models, the residual errors values. However,
at the fundamental
a larger scale maylogic for the
be small small
due scale
to the energy of
offsetting models mayand
negative be unreasonable
positive values. in How-
the first
place,
ever, thewhich likely leads
fundamental logictoforunexpected
the small errors in estimating
scale energy modelsenergy
may be use at a larger scale.
unreasonable in the The
second reason is that the model accuracy would be easily improved
first place, which likely leads to unexpected errors in estimating energy use at a larger at a larger time scale.
ThisThe
scale. is because
secondthe efforts
reason of creating
is that the modelmachine learning
accuracy would models at a larger
be easily improvedtimeatscale would
a larger
time scale. This is because the efforts of creating machine learning models at a larger time
scale would be much less compared to the small time scale models due to the decrease in
model number. The third reason is the computational cost. The increase in computational
cost by adding large time scale models is not significant since there are much fewer mod-
els with an increase in time scale. Moreover, the multi-time scale multi-output machine
Buildings 2022, 12, 2109 18 of 30

be much less compared to the small time scale models due to the decrease in model number.
The third reason is the computational cost. The increase in computational cost by adding
large time scale models is not significant since there are much fewer models with an increase
in time scale. Moreover, the multi-time scale multi-output machine learning models that
can maintain the correlation structure of energy data show very good performance, as
discussed in this section.

3.3. Results of Multi-Output Heating Energy Models


Section 3.3.1 discusses the results of daily heating energy models. Section 3.3.2 presents
the results of monthly heating energy models. Section 3.3.3 discusses the results of multi-
time scale models to simultaneously predict daily, monthly, and annual heating energy.
Section 3.2.4 presents the results from the models in which monthly and annual energy can
be obtained by using the direct simulation of energy models or the summation of energy
use at a smaller time scale.

3.3.1. Daily Heating Energy Models


The correlation coefficient between the daily heating energy in January is shown in
Figure 12. The daily heating energy is selected in January as a representative cold month
for analysis, which shows the correlation coefficients between the energy use of all working
days in January. The correlation coefficient between 0.8–0.85 is orange and brown, which
only exists between the 15th and 16th of January. The color labels for seven correlation
coefficients are in green, representing the correlation coefficients between 0.85–0.9. The
Buildings 2022, 12, x FOR PEER REVIEW 20 of
correlation coefficients between all the remaining days are above 0.9. Therefore, there is a33
strong correlation between daily heating energy.

Figure12.
Figure 12.Correlation
Correlationcoefficient
coefficientofof daily
daily heating
heating energy
energy in January
in January (2, 2(2, 2 January;
January; ... ;…
30, ;30
30,January).
30 Janu-
ary).
Figure 13 shows the predictive performance for daily heating energy from the single-
output (SO) 13
Figure and multi-output
shows (MO)performance
the predictive models based foron theheating
daily BASS and energyDNN fromapproaches,
the single-
respectively.
output (SO)Figure 13a indicates(MO)
and multi-output that the median
models CV(RMSE)
based on the values
BASS and of theDNNboxplots for the
approaches,
four models are
respectively. not significantly
Figure 13a indicates different–all below CV(RMSE)
that the median 0.02. Hence,values
the fourof models all have
the boxplots for
good performance
the four models are regarding daily heating
not significantly energy prediction.
different–all below 0.02.However,
Hence, the there
fourare outliers
models all
for these models as illustrated in Figure 13a. The three models (SO-BASS,
have good performance regarding daily heating energy prediction. However, there are SO-DNN, and
MO-DNN)
outliers forhave
thesethe CV(RMSE)
models values greater
as illustrated than13a.
in Figure 0.15,The
andthree
two points
modelsare even greater
(SO-BASS, SO-
than 0.2. This shows that the model performs poorly on certain
DNN, and MO-DNN) have the CV(RMSE) values greater than 0.15, and two points days of heating energyare
prediction.
even greaterIn than
contrast, the CV(RMSE)
0.2. This shows thatvalues of theperforms
the model MO-BASS models
poorly on are all less
certain than
days 0.15,
of heat-
indicating
ing energy good prediction
prediction. capability.the
In contrast, Moreover,
CV(RMSE) the values
interquartile
of therange
MO-BASSfrom the MO-BASS
models are all
model is the smallest, indicating that the MO-BASS model is the most
less than 0.15, indicating good prediction capability. Moreover, the interquartile range stable in terms of
model errors. The trends of MAPE values shown in Figure 13b are similar
from the MO-BASS model is the smallest, indicating that the MO-BASS model is the most to those shown
stable in terms of model errors. The trends of MAPE values shown in Figure 13b are sim-
ilar to those shown in Figure 13a, which verifies the accuracy results from the CV(RMSE).
The R² values shown in Figure 13c are above 0.95 for the four models, indicating that the
model performance is high. Among them, the R²values of the MO-BASS model are greater
have good performance regarding daily heating energy prediction. However, there are
outliers for these models as illustrated in Figure 13a. The three models (SO-BASS, SO-
DNN, and MO-DNN) have the CV(RMSE) values greater than 0.15, and two points are
even greater than 0.2. This shows that the model performs poorly on certain days of heat-
ing energy prediction. In contrast, the CV(RMSE) values of the MO-BASS models are all
Buildings 2022, 12, 2109
less than 0.15, indicating good prediction capability. Moreover, the interquartile range 19 of 30
from the MO-BASS model is the smallest, indicating that the MO-BASS model is the most
stable in terms of model errors. The trends of MAPE values shown in Figure 13b are sim-
ilar to those shown in Figure 13a, which verifies the accuracy results from the CV(RMSE).
in Figure 13a, which verifies the accuracy results from the CV(RMSE). The R2 values shown
The R² values shown in Figure 13c are above 0.95 for the four models, indicating that the
in Figure 13c are above 0.95 for the four models, indicating that the model performance is
model performance is high. Among them, the2 R²values of the MO-BASS model are greater
high. Among
than 0.985. Therefore, them, the
the MO-BASS R is
model values of the
the most MO-BASS
accurate dailymodel
heatingare greater than 0.985. Therefore,
energy
the MO-BASS
model in this case study. model is the most accurate daily heating energy model in this case study.

uildings 2022, 12, x FOR PEER REVIEW 21 of 33

(a) CV(RMSE)

(b) MAPE

(c) R²
Figure 13. Prediction
Figureperformance of fourperformance
13. Prediction machine learning models
of four for daily
machine heating models
learning energy (SO,
for daily heating energy (SO,
single output; MO, multiple output).
single output; MO, multiple output).
The modeling time of the four daily heating energy models is shown in Table 5. The
most time-saving model is the MO-DNN model, which requires only 69.7 s. The most
time-consuming model is the SO-DNN, which spends 3722.6 s (more than 1 h)–around 53
times that of the SO-DNN model. The MO-BASS model is also significantly more time
efficient compared to the SO-BASS model. Therefore, the multi-output models save com-
Buildings 2022, 12, 2109 20 of 30

The modeling time of the four daily heating energy models is shown in Table 5. The
most time-saving model is the MO-DNN model, which requires only 69.7 s. The most time-
consuming model is the SO-DNN, which spends 3722.6 s (more than 1 h)–around 53 times
that of the SO-DNN model. The MO-BASS model is also significantly more time efficient
compared to the SO-BASS model. Therefore, the multi-output models save computational
time compared to the single-output models.

Table 5. The computational time of different time-scale heating energy models (unit: s).

Model Daily Monthly Multi-Time Scale


SO-BASS 2293.0 128.8 -
Buildings 2022, 12, x FOR PEER REVIEW MO-BASS 216.6 101.6 150.4 22 of 33
SO-DNN 3722.6 373.6 -
MO-DNN 69.7 65.4 165.9

The dendrograms for daily heating energy from the training set and the four models
The dendrograms for daily heating energy from the training set and the four models
are shown in Figure 14. The numbers in Figure 14 represent dates in a specific month,
are shown in Figure 14. The numbers in Figure 14 represent dates in a specific month,
such as 2 for 2 January. Figure 14a reflects the dendrogram of the correlation between the
such as 2 for 2 January. Figure 14a reflects the dendrogram of the correlation between the
outputs from the engineering-based EnergyPlus energy models. It can be seen that 2, 5,
outputs from the engineering-based EnergyPlus energy models. It can be seen that 2, 5,
12, 19, and 26 belong to one subgroup, and these days are the first working days after the
12, 19, and 26 belong to one subgroup, and these days are the first working days after the
holidays
holidays oror weekends.
weekends. The The reason
reason for
for the
the aggregation
aggregation is is that
that the
the office
office building
building is
is not
not
heated during non-working days, resulting in a low temperature inside the
heated during non-working days, resulting in a low temperature inside the office building. office build-
ing. The heating
The heating systemsystem
startsstarts to work
to work on theon theworking
first first working day, consumes
day, which which consumes more
more energy
energy than the following days. Figure 14d,e have significantly different correlation
than the following days. Figure 14d,e have significantly different correlation structures struc-
tures from Figure
from Figure 14a since
14a since the non-working
the non-working days days dobelong
do not not belong
to oneto subgroup.
one subgroup. In con-
In contrast,
trast, the correlation structures from Figure 14b,c are similar to that
the correlation structures from Figure 14b,c are similar to that from Figure 14a.from Figure 14a.

(a) (b) (c)

(d) (e)
Figure
Figure 14.
14. Dendrogram of January
Dendrogram of Januarydaily
dailyheating
heatingenergy
energyfor
forthe
thetraining
trainingsetset and
and four
four machine
machine learn-
learning
ing models.
models.

3.3.2. Monthly Heating Energy Models


The correlation coefficient of monthly heating energy is shown in Figure 15. The
smallest correlation coefficient is 0.969 between November and March, and the largest
correlation coefficient is 0.999 between January and December. As a result, there is a sig-
nificant correlation between the five months of heating energy data. Therefore, it is of
great significance to construct heating energy models by considering the correlation struc-
ture of energy use.
Buildings 2022, 12, 2109 21 of 30

3.3.2. Monthly Heating Energy Models


The correlation coefficient of monthly heating energy is shown in Figure 15. The
smallest correlation coefficient is 0.969 between November and March, and the largest
correlation coefficient is 0.999 between January and December. As a result, there is a
significant correlation between the five months of heating energy data. Therefore, it is
Buildings 2022, 12, x FOR PEER REVIEW 23 of 33
of great significance to construct heating energy models by considering the correlation
ngs 2022, 12, x FOR PEER REVIEW 23 of 33
structure of energy use.

Figure
Figure
Figure 15. Correlation 15. Correlation
15. Correlation
coefficient coefficient
heatingof
coefficient
of monthly ofmonthly
monthlyheating
energy. heatingenergy.
energy.

Figure 16 showsFigure
Figure 16
16shows
shows the
the performance the performance
for the indicators
performance
indicators indicators
monthly for
for the
heating the monthly
monthly
energy heating
heating energy
models. energy models.
models.
The
The MO-BASS model MO-BASS
The MO-BASS model has
modelCV(RMSE)
has the smallest the smallest
has the smallest CV(RMSE)
and MAPE CV(RMSE) and
with the and MAPE
MAPE
largest with
R², withthe largest R2 ,R²
the largest
indicat- indicating
, indicat-
that
ing that the MO-BASS theoutperforms
ing that MO-BASS
the MO-BASS outperforms
the other three
outperforms themodels.
other three models.
The three
the other other The
three
models. other
models three three
also
The other models also have
models also
2
have CV(RMSE)CV(RMSE)
less than less
0.13, than
MAPE 0.13,
lessMAPE
than less
0.1, than
and 0.1,
R² and
greater R greater
than 0.92, than 0.92,
indicating
have CV(RMSE) less than 0.13, MAPE less than 0.1, and R² greater than 0.92, indicating indicating that they
also
that they also have have
thatgood goodhave
theypredictive
also predictive
performance. performance.
good predictive that Note
Noteperformance.
the monththat the that
with
Note month
the the
worstwith
pre-
month thewith
worst
theprediction
worst pre-
performance
diction performance for each for each
model ismodel
March, is March,
followed followed
by November. by
diction performance for each model is March, followed by November. What November.
What these What
two these twothese
months
two
months have inhavecommonin common
is that is
theythatare they
both are both
transitionaltransitional
months. months.
Climatic Climatic
conditions
months have in common is that they are both transitional months. Climatic conditions conditions such as
temperature
such as temperature varycan vary considerably theduring the months,
transition months, which can add to the
such can considerably
as temperature can varyduring transition
considerably which
during the transition can months,
add which can add
complexity
to the complexity of the
of the heating heating
energy andenergy
renderand render
it more it more
difficult difficultthe
to predict to heating
predict the heating energy.
to the complexity of the heating energy and render it more difficult to predict the heating
Prediction
energy. Prediction accuracyaccuracy is improved
is improved significantly
significantly when there when there are
are stable stable
changes in changes in climate
energy. Prediction accuracy is improved significantly when there are stable changes in
conditions
climate conditions such as temperatures,
such as outdoor outdoor temperatures,
for example forinexample
Februaryinand February
December.and December.
climate conditions such as outdoor temperatures, for example in February and December.

(a) CV(RMSE)

Figure 16. Cont. (a) CV(RMSE)


ngs 2022, 12,Buildings 12, 2109
2022,REVIEW
x FOR PEER 24 of 33 22 of 30

(b) MAPE

(c) R²
Figure 16. Prediction performance
Figure of four
16. Prediction machine learning
performance of fourmodels for learning
machine monthly models
heating energy.
for monthly heating energy.

The modeling timeThe of the fourtime


modeling monthly
of theheating energy heating
four monthly models is listedmodels
energy in Tableis5.listed in Table 5. The
The difference between
difference thebetween
two BASS themodels
two BASSis similar sinceisthere
models aresince
similar only five
there months
are only five months for
for heating energy. The energy.
heating computing Thetime of the SO-DNN
computing time of model is aboutmodel
the SO-DNN 5.7 times that of5.7 times that of the
is about
the MO-DNN model.
MO-DNN In general,
model.the In modeling
general, theof the multi-output
modeling of the models is faster
multi-output com- is faster compared
models
pared to the single-output models. models.
to the single-output
The dendrograms of monthly heating energy for the training set and the four machine
The dendrograms
learningofmodels
monthly areheating
shown in energy
Figure for17.
theThe
training set months
adjacent and theof four ma- and December are
January
chine learning models are shown in Figure 17. The adjacent months of January
grouped together, as is shown in Figure 17a, due to their similar weather and De- conditions. The
cember are grouped together, as is shown in Figure 17a, due to their similar weather con-
two transition months, November and March, are classified into one category, indicating
ditions. The twothat
transition months,
there is a largeNovember
gap betweenand March, are classified
the weather into one
conditions andcategory,
the rest of the months. The
indicating that there is a large gap between the weather conditions and the rest of the
single-output and multi-output BASS models can maintain the same correlation with the
months. The single-output and multi-output BASS models can maintain the same corre-
EnergyPlus models, while the single-output and multi-output DNN models are different
lation with the EnergyPlus models, while the single-output and multi-output DNN mod-
from the EnergyPlus models. As shown in Figure 16, these correlation structures are helpful
els are different from the EnergyPlus models. As shown in Figure 16, these correlation
to improve the prediction capability of learning models.
structures are helpful to improve the prediction capability of learning models.
3.3.3. Multi-Time Scale Heating Energy Models
Figure 18 shows the prediction performance of the heating energy model on multiple
time scales. The model accuracy increases as the time scale increases. This is directly related
to the complexity of the energy data in different time scales. Figure 18a demonstrates that
the median CV(RMSE) values of the daily heating energy of the two models are less than
0.05, although there are a few outliers. In all three time scales, the BASS models perform
better than the DNN models in terms of the CV(RMSE). The same conclusions can be
obtained from Figure 18b in terms of the MAPE. Figure 18c shows that there are a few
(a) (b) (c)
Buildings 2022, 12, 2109 23 of 30

points in R2 between 0.875 and 0.95 for the DNN daily models, indicating that the model
performance becomes worse on some days. Note that the interquartile ranges of the BASS
model are smaller than that of the DNN model, indicating that the MO-BASS model is
Buildings 2022, 12, x FOR PEER REVIEW 25 of 33
more stable. Therefore, the multi-time scale BASS model has better predictive capability in
comparison with the DNN model.

(d) (e)
Figure 17. Dendrogram of monthly heating energy for the training set and four machine learning
models.

3.3.3. Multi-Time Scale Heating Energy Models


Figure 18(a)shows the prediction performance (b) of the heating energy model (c)on multiple
time scales. The model accuracy increases as the time scale increases. This is directly re-
lated to the complexity of the energy data in different time scales. Figure 18a demonstrates
that the median CV(RMSE) values of the daily heating energy of the two models are less
than 0.05, although there are a few outliers. In all three time scales, the BASS models per-
form better than the DNN models in terms of the CV(RMSE). The same conclusions can
be obtained from Figure 18b in terms of the MAPE. Figure 18c shows that there are a few
points in R² between 0.875 and 0.95 for the DNN daily models, indicating that the model
performance becomes worse on some days. Note that the interquartile ranges of the BASS
(d) that of the DNN model, indicating that the (e)
model are smaller than MO-BASS model is
more stable. Therefore, Figure
the 17. Dendrogram
multi-time of monthly
scale BASS heating energy
model for
has the training
better
Figure 17. Dendrogram of monthly heating energy for the training setset
and
predictive four
and machine learning
capability
four machine learning
models.
in comparison with the DNN model.
models.
3.3.3. Multi-Time Scale Heating Energy Models
Figure 18 shows the prediction performance of the heating energy model on multiple
time scales. The model accuracy increases as the time scale increases. This is directly re-
lated to the complexity of the energy data in different time scales. Figure 18a demonstrates
that the median CV(RMSE) values of the daily heating energy of the two models are less
than 0.05, although there are a few outliers. In all three time scales, the BASS models per-
form better than the DNN models in terms of the CV(RMSE). The same conclusions can
be obtained from Figure 18b in terms of the MAPE. Figure 18c shows that there are a few
points in R² between 0.875 and 0.95 for the DNN daily models, indicating that the model
performance becomes worse on some days. Note that the interquartile ranges of the BASS
model are smaller than that of the DNN model, indicating that the MO-BASS model is
more stable. Therefore, the multi-time scale BASS model has better predictive capability
in comparison with the DNN model.

(a) CV(RMSE)
Figure 18. Cont.

(a) CV(RMSE)
R PEER REVIEW 26 of 33
Buildings 2022, 12, 2109 24 of 30

(b) MAPE

(c) R²
Figure 18. Prediction performance of twoperformance
Figure 18. Prediction machine learning models
of two machine for multi-time
learning scales heating
models for multi-time scales heating
energy. energy.

Table 5 shows the computation time of multi-time scale heating energy models using the
Table 5 shows the computation time of multi-time scale heating energy models using
BASS and DNN approaches. The computational time is quite similar for these two models.
the BASS and DNN Considering
approaches. theThe
modelcomputational
accuracy as showntime is quite
in Figure similar
18, the for these
multi-time twomodels
scale BASS
models. Consideringwould
the model accuracy
be a better as shown
candidate in Figure
for the multi-time 18,prediction
scale the multi-time scale
of building BASS
energy use.
models would be a better candidate for the multi-time scale prediction of building energy
3.3.4. Performance Analysis of Ten Models for Monthly and Annual Heating Energy
use.
Figure 19 shows the performance of 10 models that can obtain monthly heating energy.
The meanings of these 10 models are available in Table 4. The first three best models are the
3.3.4. Performance Analysis of Ten Models
daily multi-output models for Monthly
(MOB-D), andmulti-output
monthly Annual Heating
models Energy
(MOB-M), and multi-
time scale model (MOB-Mu) based on the BASS algorithm.
Figure 19 shows the performance of 10 models that can obtain monthly heating These three modelsen-
have very
low CV(RMSE) and MAPE values. The next two models are the single-output daily BASS
ergy. The meanings of these 10 models are available in Table 4. The first three best models
models (SOB-D) and the multi-output daily DNN models (MOD-D). These two models still
are the daily multi-output models
have very (MOB-D),
high R monthly
2 values–close multi-output
to 1. The following two models
models are(MOB-M), andmonthly
single-output
multi-time scale model (MOB-Mu)
BASS basedand
models (SOB-M) onsingle-output
the BASS algorithm. These (SOD-D).
daily DNN models three models have three
The remaining
models do not have good predictive performance compared
very low CV(RMSE) and MAPE values. The next two models are the single-output dailyto the other seven models. The
BASS models (SOB-D) and the multi-output daily DNN models (MOD-D). These two
models still have very high R2 values–close to 1. The following two models are single-
output monthly BASS models (SOB-M) and single-output daily DNN models (SOD-D).
The remaining three models do not have good predictive performance compared to the
Buildings 2022, 12, 2109 25 of 30

ildings 2022, 12, x FOR PEER REVIEW 27 of 33 models for


multi-output models have better performance compared to the single-output
monthly heating energy.

(a) CV(RMSE)

(b) MAPE

(c) R²
Figure 19. Performance of 10 modelsoffor
Figure 19. Performance 10 monthly heating
models for energy
monthly (refer
heating to Table
energy (refer4tofor model
Table 4 forexpla-
model explanation).
nation).
Figure 20 shows the performance of 10 models that can obtain annual energy. The
Figurethree
20 shows the performance
multi-output of 10 models
BASS models (MOB-D, thatMOB-M,
can obtain annual energy.
MOB-Mu) are theThe
first three best
three multi-output
models in BASS
whichmodels (MOB-D, values
the CV(RMSE) MOB-M, areMOB-Mu) arethe
about 0.003, the first three
MAPE valuesbest
are about 0.002,
the R2the
models in which CV(RMSE)
values values
are about are The
0.999. about 0.003, the MAPE
performance of thevalues
SOB-D are about
and 0.002,single-output
SOB-M
the R² values
BASSare models
about 0.999. The performance
is slightly worse than of theofSOB-D
that and SOB-M
the above single-output
three multi-output BASS models.
BASS models is slightly
Among worse
the five DNN than that of
models, thethe above models
MOD-D three multi-output BASS models.
and the MOD-Mu models have similar
Among theperformance
five DNN models,
to thethe MOD-D
SOB-M models
models. and
The the MOD-Mu
SOD-M modelsmodels
and MOD-M have similar
do not have good
performance to the SOB-M
predictive models.for
performance The SOD-M
annual and MOD-M
heating energy. models do not have good
predictive performance for annual heating energy.
Buildings
Buildings 2022, 12, x 2022, 12, 2109
FOR PEER REVIEW 28 of 33 26 of 30

(a) CV(RMSE)

(b) MAPE

(c) R²
Figure 20.Figure
Performance of 10 models
20. Performance of 10for annual
models forheating energy energy
annual heating (refer to(refer
Tableto 4Table
for model explana-
4 for model explanation).
tion).
3.4. Guide and Application of Building Multi-Output Energy Models
3.4. Guide andThis
Application
sectionofdiscusses
Building Multi-Output
the implications, Energy Models and applications of the results
guidelines,
Thispresented in Sections
section discusses the 3.1–3.3 from four
implications, aspects.and
guidelines, The first aspectofdiscusses
applications the results the choice
of machine learning models between the single-output and multiple-output
presented in Sections 3.1 to 3.3 from four aspects. The first aspect discusses the choice of models in
machine learning models between the single-output and multiple-output models in build- models
building energy assessment. The second aspect is the choice of multiple-output
ing energywith or withoutThe
assessment. considering output
second aspect correlation
is the choice ofinmultiple-output
the building energy modelsanalysis.
with orThe third
without aspect is related
considering outputto the choice of
correlation inmulti-output modelsanalysis.
the building energy for various Thetime
thirdscales
aspectby taking
is relatedthe additive
to the choiceorofaccumulative
multi-output features
models for intovarious
account.timeThe fourth
scales aspect the
by taking is focused
addi- on the
application of
tive or accumulative the multi-output
features models
into account. used in
The fourth this paper.
aspect is focused on the application
The multi-output learning
of the multi-output models used in this paper. models are preferred in the case of multiple outputs of
building energy analysis. This is because the computational
The multi-output learning models are preferred in the case of multiple cost can be dramatically
outputs of
reduced in comparison with the single-output models, as discussed
building energy analysis. This is because the computational cost can be dramatically re- in Sections 3.2 and 3.3.
duced inThe benefits with
comparison of using the multi-output
the single-output models
models, are moreinsignificant
as discussed Sections 3.2when there are a
and 3.3.
largeofnumber
The benefits using the of multi-output
building energy outputs–for
models are more instance, over
significant 10 outputs.
when there areMoreover,
a large model
accuracy from the multi-output models is not necessarily less
number of building energy outputs–for instance, over 10 outputs. Moreover, model accu- than that of the single-output
racy from the multi-output models is not necessarily less than that of the single-outputthe multi-
models in building energy assessment. Therefore, it would be recommended that
output learning models should be used when applying the machine learning techniques to
models in building energy assessment. Therefore, it would be recommended that the
building energy assessment, especially for a larger number of model outputs.
multi-output learning models should be used when applying the machine learning tech-
niques to building energy assessment, especially for a larger number of model outputs.
Buildings 2022, 12, 2109 27 of 30

There are two types of machine learning methods depending on whether to consider
the correlation of building energy performance. One example of learning models is the
Bayesian adaptive spline surface in which the correlation of outputs can be taken into
account, while one example of learning models without considering output correlation
is the deep neural network. If there are almost no or very weak correlations among
building energy use, both two types of learning models can be used to create multi-output
models. If the correlation of outputs is significant, it is necessary to implement the learning
methods by considering the correlation of outputs. The model accuracy can be improved
by maintaining the correlation of building energy use.
It is necessary to consider the additive or accumulative features when constructing the
learning models for predicting building energy use in various time scales. This is because
the building energy use at a larger time scale is the sum of building energy at a smaller
time scale. For instance, the monthly electricity use from a building would be the annual
electricity for this building. For single-output models at a smaller time scale, it is natural
to sum up all the energy use at a smaller time scale to obtain total energy use at a larger
time scale. However, the residuals for this simple summation would be hard to manage. If
there are all positive (or negative) residuals from the smaller time scale models, the total
residuals would be very large compared to the single-output model at a larger time scale.
Another issue is the inconsistency between the summation of smaller time scale models
and the larger time scale model, which confuses the estimation of building energy use at a
larger time scale. By considering the correlation of building energy use, the multi-output
learning models can create fast and consistent models at various time scales for building
energy analysis. Therefore, multi-output learning models that can maintain the correlation
of outputs are recommended in multiple time scale predictions of building energy analysis.
The method proposed in this paper can also be applied to performance indicators of
buildings (such as energy, loads, and carbon emissions) in different scenarios although
this paper is concentrated on building energy consumption. It would be interesting to
explore building load profiles at various time scales (hourly, daily, and monthly), which can
provide useful information on the design of district heating and cooling systems. When
designing net-zero emission buildings, it is necessary to provide the matching of demand
and supply at various time scales to guide for the design of PV systems. The multi-output
models at various time scales would be very useful to provide accurate building demand
data at various time scales to compute solar fraction, self-sufficiency, and other indicators.
These results can be used for optimizing the ratio of PV rated power and battery. Moreover,
the method used here can be used in uncertainty analysis, sensitivity analysis, and model
calibration in building energy assessment. For instance, the multi-output models at daily
periods can be used as mathematical models to calibrate these daily models using the
Bayesian analysis. The computational loads would be reduced significantly since there are
only two multi-output learning models for heating and cooling energy use in buildings.

4. Conclusions
This research investigates the predictive performance of multi-output machine learn-
ing models at three time scales (daily, monthly, and annual) for building energy assessment
using two algorithms (Bayesian adaptive spline surface and deep neural network). The
results indicate that the machine learning models, which could consider the correlation of
energy use, can have high model accuracy with fast computation and accumulative features
at different time scales in building energy analysis. The multi-output learning models for
building energy prediction would significantly reduce the computational time for creating
learning models in comparison to the single-output learning models. The ability to main-
tain the correlation structure of energy data for the multi-output learning models is the key
to providing accurate results and accumulative features. The deep neural network models
can simulate the multiple outputs without taking into account the correlation of energy
use. Hence, the deep neural network cannot have accumulative features. In contrast, the
multi-time scale Bayesian adaptive spline surface models can have the same or similar
Buildings 2022, 12, 2109 28 of 30

correlation structure of energy data from the engineering-based building energy models.
This means that the predicted energy data at a larger time scale would equal the summation
of energy data at a smaller time scale in building energy computation. Moreover, it is
found that the simple summation of energy results from smaller time scale learning models
do not necessarily have good predictions of energy use at a larger time scale due to the
uncontrolled offset of large and small residuals, or the final accumulative large errors.
Three guidelines can be obtained for creating machine learning models in building
energy assessment based on this research. The multi-output learning models are preferred
compared to the single-output model in dealing with multiple outputs, especially in the
case of the number of outputs over 10. Multi-output learning models that can consider
output correlation are recommended when the multiple outputs are correlated in building
energy analysis. Compared to the direct summation from the single-output models at the
smaller time scale, multi-output models with a consistent accumulative feature that can
simultaneously predict energy use at various time scales are preferred when dealing with
multiple time scale building energy predictions.
The multi-out learning models used in this paper are only applied to an office building.
More research should be implemented on other building types, for example, residential
buildings, schools, and hospitals, to understand the suitability of these multi-output meth-
ods. More studies should also be conducted to explore the predictive performance of
multi-output models for building energy analysis using different learning algorithms, such
as random forest, support vector machine, and Gaussian process.

Author Contributions: Conceptualization, G.L. and W.T.; methodology, G.L. and W.T.; software, G.L.;
investigation, H.Z. and B.C.; writing—original draft preparation, G.L.; writing—review and editing,
W.T.; visualization, B.C. All authors have read and agreed to the published version of the manuscript.
Funding: This research was supported by the National Natural Science Foundation of China
(No. 51778416) and the Key Projects of Philosophy and Social Sciences Research, Ministry of Education
of China (No. 16JZDH014).
Data Availability Statement: Not applicable.
Conflicts of Interest: The authors declare no conflict of interest.

Abbreviations
ANN artificial neural network
BASS Bayesian adaptive spline surface
BIPV building integrated photovoltaic
BMARS Bayesian multivariate adaptive regression splines
CSP cooling set-point
CSWD Chinese standard weather data
CV(RMSE) coefficient of variation of the root mean square error
DNN deep neural network
DT decision tree
ENMIM ensemble model named evolutionary neural machine inference model
EPD equipment power density
EWU exterior wall U-value
GA-NMM genetic algorithm-based numerical moment matching
GB gradient boosting
HSP heating set-point
HVAC heating, ventilation, and air conditioning
INF infiltration rate
KNN K-nearest neighbor
lightGBM light gradient boosting machine
Buildings 2022, 12, 2109 29 of 30

LPD lighting power density


LR linear regression
LSSVR least squares support vector regression
MAPE mean absolute percentage error
MARS multivariate adaptive regression splines
MIMO multi-input multi-output
MO multiple outputs
OPD occupancy density
PCA principal component analysis
PCC Pearson’s correlation coefficient
PV photovoltaic
R2 coefficient of determination
RBFNN radial basis function neural network
RF random forest
RU roof U-value
SARIMA seasonal autoregressive integrated moving average
SHGC solar heat gain coefficient
SO single output
SVM support vector machine
WU window U-value

References
1. UN Environment Programme (UNEP). 2021 Global Status Report for Buildings and Construction; UNEP: Nairobi, Kenya, 2021.
2. Zhang, Y.; Teoh, B.K.; Wu, M.; Chen, J.; Zhang, L. Data-driven estimation of building energy consumption and GHG emissions
using explainable artificial intelligence. Energy 2023, 262, 125468. [CrossRef]
3. Al-Shargabi, A.A.; Almhafdy, A.; Ibrahim, D.M.; Alghieth, M.; Chiclana, F. Buildings’ energy consumption prediction models
based on buildings’ characteristics: Research trends, taxonomy, and performance measures. J. Build. Eng. 2022, 54, 104577.
[CrossRef]
4. Guo, Y.-Y. Revisiting the building energy consumption in China: Insights from a large-scale national survey. Energy Sustain. Dev.
2022, 68, 76–93. [CrossRef]
5. Mohapatra, S.K.; Mishra, S.; Tripathy, H.K.; Alkhayyat, A. A sustainable data-driven energy consumption assessment model for
building infrastructures in resource constraint environment. Sustain. Energy Technol. Assess. 2022, 53, 102697. [CrossRef]
6. Fathi, S.; Srinivasan, R.; Fenner, A.; Fathi, S. Machine learning applications in urban building energy performance forecasting: A
systematic review. Renew. Sustain. Energy Rev. 2020, 133, 110287. [CrossRef]
7. Zhang, L.; Wen, J.; Li, Y.; Chen, J.; Ye, Y.; Fu, Y.; Livingood, W. A review of machine learning in building load prediction.
Appl. Energy 2021, 285, 116452. [CrossRef]
8. Lei, L.; Chen, W.; Wu, B.; Chen, C.; Liu, W. A building energy consumption prediction model based on rough set theory and deep
learning algorithms. Energy Build. 2021, 240, 110886. [CrossRef]
9. Liu, Y.; Chen, H.; Zhang, L.; Wu, X.; Wang, X.-J. Energy consumption prediction and diagnosis of public buildings based on
support vector machine learning: A case study in China. J. Clean. Prod. 2020, 272, 122542. [CrossRef]
10. Alobaidi, M.H.; Chebana, F.; Meguid, M.A. Robust ensemble learning framework for day-ahead forecasting of household based
energy consumption. Appl. Energy 2018, 212, 997–1012. [CrossRef]
11. Wang, Z.; Hong, T.; Li, H.; Piette, M.A. Predicting city-scale daily electricity consumption using data-driven models. Adv. Appl.
Energy 2021, 2, 100025. [CrossRef]
12. Jetcheva, J.G.; Majidpour, M.; Chen, W.-P. Neural network model ensembles for building-level electricity load forecasts.
Energy Build. 2014, 84, 214–223. [CrossRef]
13. Ferrantelli, A.; Kuivjogi, H.; Kurnitski, J.; Thalfeldt, M. Office Building Tenants’ Electricity Use Model for Building Performance
Simulations. Energies 2020, 13, 5541. [CrossRef]
14. Wang, W.; Hong, T.; Xu, X.; Chen, J.; Liu, Z.; Xu, N. Forecasting district-scale energy dynamics through integrating building
network and long short-term memory learning algorithm. Appl. Energy 2019, 248, 217–230. [CrossRef]
15. Tran, D.-H.; Luong, D.-L.; Chou, J.-S. Nature-inspired metaheuristic ensemble model for forecasting energy consumption in
residential buildings. Energy 2020, 191, 116552. [CrossRef]
16. Tian, W.; Yang, S.; Li, Z.; Wei, S.; Pan, W.; Liu, Y. Identifying informative energy data in Bayesian calibration of building energy
models. Energy Build. 2016, 119, 363–376. [CrossRef]
17. Zhu, C.; Tian, W.; Yin, B.; Li, Z.; Shi, J. Uncertainty calibration of building energy models by combining approximate Bayesian
computation and machine learning algorithms. Appl. Energy 2020, 268, 115025. [CrossRef]
18. Koschwitz, D.; Frisch, J.; van Treeck, C. Data-driven heating and cooling load predictions for non-residential buildings based on
support vector machine regression and NARX Recurrent Neural Network: A comparative study on district scale. Energy 2018,
165, 134–142. [CrossRef]
Buildings 2022, 12, 2109 30 of 30

19. Jahani, E.; Cetin, K.; Cho, I.H. City-scale single family residential building energy consumption prediction using genetic
algorithm-based Numerical Moment Matching technique. Build. Environ. 2020, 172, 106667. [CrossRef]
20. Lin, Q.; Liu, K.; Hong, B.; Xu, X.; Chen, J.; Wang, W. A data-driven framework for abnormally high building energy demand
detection with weather and block morphology at community scale. J. Clean. Prod. 2022, 354, 131602. [CrossRef]
21. Olu-Ajayi, R.; Alaka, H.; Sulaimon, I.; Sunmola, F.; Ajayi, S. Building energy consumption prediction for residential buildings
using deep learning and other machine learning techniques. J. Build. Eng. 2022, 45, 103406. [CrossRef]
22. Tian, W.; Zhu, C.; Sun, Y.; Li, Z.; Yin, B. Energy characteristics of urban buildings: Assessment by machine learning. Build. Simul.
2020, 14, 179–193. [CrossRef]
23. Tian, W.; de Wilde, P.; Li, Z.; Song, J.; Yin, B. Uncertainty and sensitivity analysis of energy assessment for office buildings based
on Dempster-Shafer theory. Energy Convers. Manag. 2018, 174, 705–718. [CrossRef]
24. Wang, L.; Lee, E.W.M.; Hussian, S.A.; Yuen, A.C.Y.; Feng, W. Quantitative impact analysis of driving factors on annual residential
building energy end-use combining machine learning and stochastic methods. Appl. Energy 2021, 299, 117303. [CrossRef]
25. Singh, M.; Sharston, R. Quantifying the dualistic nature of urban heat Island effect (UHI) on building energy consumption.
Energy Build. 2022, 255, 111649. [CrossRef]
26. Luo, X.J.; Oyedele, L.O.; Ajayi, A.O.; Akinade, O.O. Comparative study of machine learning-based multi-objective prediction
framework for multiple building energy loads. Sustain. Cities Soc. 2020, 61, 102283. [CrossRef]
27. Liu, J.; Zhang, Q.; Dong, Z.; Li, X.; Li, G.; Xie, Y.; Li, K. Quantitative evaluation of the building energy performance based on
short-term energy predictions. Energy 2021, 223, 120065. [CrossRef]
28. Li, G.; Li, F.; Ahmad, T.; Liu, J.; Li, T.; Fang, X.; Wu, Y. Performance evaluation of sequence-to-sequence-Attention model for
short-term multi-step ahead building energy predictions. Energy 2022, 259, 124915. [CrossRef]
29. U.S. Department of Energy. EnergyPlus V22.1.0; U.S.Department of Energy: Washington, DA, USA, 2021.
30. MOHURD (Ministry of Housing and Urban-Rural Development). China, Energy-Saving Design Standards for Public Buildings;
China Building Industry Press: Beijing, China, 2015.
31. MOHURD (Ministry of Housing and Urban-Rural Development). China, Technical Standards for Near-Zero Energy Buildings; China
Building Industry Press: Beijing, China, 2019.
32. Francom, D.; Sansó, B. BASS: An R Package for Fitting and Performing Sensitivity Analysis of Bayesian Adaptive Spline Surfaces.
J. Stat. Softw. 2020, 94, 1–36. [CrossRef]
33. An, N.; Zhao, W.; Wang, J.; Shang, D.; Zhao, E. Using multi-output feedforward neural network with empirical mode decomposi-
tion based signal filtering for electricity demand forecasting. Energy 2013, 49, 279–288. [CrossRef]
34. Cohen-Addad, V.; Kanade, V.; Mallmann-Trenn, F.; Mathieu, C. Hierarchical Clustering: Objective Functions and Algorithms.
J. Acm. 2019, 66, 1–42. [CrossRef]
35. Kaminskyy, R.; Shakhovska, N.; Kryvinska, N.; Younas, M. Dendrograms-based disclosure method for evaluating cluster analysis
in the IoT domain. Comput. Ind. Eng. 2021, 158, 107402. [CrossRef]
36. Varshney, A.K.; Muhuri, P.K.; Lohani, Q.M.D. PIFHC: The Probabilistic Intuitionistic Fuzzy Hierarchical Clustering Algorithm.
Appl. Soft Comput. 2022, 120, 108584. [CrossRef]

You might also like