Modeling and Analysis of Commercial Building Electrical Loads For Demand Side Management
Modeling and Analysis of Commercial Building Electrical Loads For Demand Side Management
Management
A Thesis
of
Drexel University
by
Jonathan Berardino
of
Doctor of Philosophy
December 2016
© Copyright 2016
Jonathan Berardino. All Rights Reserved.
iii
ACKNOWLEDGMENTS
First, I would like to thank my advisor Dr. Chika Nwankpa for your time,
guidance, and support in performing the research that led to this thesis. I am forever
grateful to you for accepting me as a student and providing me with the opportunity to
I would like to thank Dr. Miu, Dr. Niebur, Dr. Benson, and Dr. Kwatny for
serving on my thesis committee. Your comments have greatly improved the final version
of this thesis and I appreciate the advice and insights you have provided that will enhance
Thank you to my fellow members of the Center for Electric Power Engineering. I
have been very fortunate to work with a group of people who have been valuable
colleagues and (more importantly) friends. I wish you all the best and look forward to
Lastly, and only because the last paragraph is always the most important, I would
like to thank my wife Lindsay. Pursuing this degree has taken up enormous amounts of
my time and nobody has felt that impact more than you. Thank you for your patience and
selflessness over the years. This would not have happened without you.
iv
TABLE OF CONTENTS
1. INTRODUCTION ....................................................................................................... 1
RESOURCES ..................................................................................................................... 6
MANAGEMENT RESEARCH................................................................................. 13
16
INTERVALS .................................................................................................................... 69
PROCEDURE............................................................................................................ 72
LIST OF TABLES
Table 3.1. Indicator variables representing the day of the week ...................................... 26
Table 3.3. Quantiles of day-ahead MAPE performance across all out-of-sample forecasts
........................................................................................................................................... 38
Table 3.5. MAPE (%) performance broken up by weekdays and weekends/holidays ..... 39
Table 3.6. Hourly MAPE (%) across all out-of-sample forecasts .................................... 41
Table 3.8. Hourly SDE (kW) Across all out-of-sample forecasts .................................... 51
Table 4.1. % Deviation in empirical coverage compared to the nominal coverage rate . 85
Table 5.1 Parameter estimation results for the test described in Section 5.2.1............... 102
Table 6.4. Load footprint and flexibility for each building used in this case study........ 118
Table 6.5 Base building parameters values used in the simulations ............................... 120
Table 6.6 Optimal load schedule in kW for Test 1 (Scenario 1) .................................... 123
Table 6.7. Optimal load schedule for in kW Test 2 (Scenario 1 with reduced load
Table 6.8. Optimal load schedule in kW for Test 4 (Scenario 2) ................................... 127
ix
thresh
Table 6.9 Optimal load schedule in kW for Test 5 (Scenario 3 – P =150kW) ......... 127
Table 6.10 Optimal load schedule in kW for Test 6 (Scenario 3 – Pthresh=75kW) ......... 132
Table 6.11 Optimal load schedule in kW for Test 6 (Scenario 4) .................................. 132
Table 6.12 Optimal load schedule in kW for Test 8: +10kW uniform load forecast
uncertainty....................................................................................................................... 133
Table B.1 MAPE results at 5 minute resolution for all 116 day-ahead forecasts ........... 150
x
LIST OF FIGURES
Figure 2.2. Breakdown of commercial building energy use in the United States (2014)
[19] ...................................................................................................................................... 9
Figure 2.3. Linear correlation between demand and several variables: OAT (top), OAT-
Figure 2.4. Temporal correlation for 24 hours of building demand. Sample size is 4
Figure 2.5. Example of a baseline demand profile during a DSM event .......................... 14
Figure 3.2. Breakup of data set and how each portion is used in the forecasting process 24
Figure 3.3. Daily building demand curves for the 116 days used for out of sample
testing. The thicker line shows the mean daily profile .................................................... 25
Figure 3.5. A general multi-layer, feed-forward artificial neural network with N hidden
layers ................................................................................................................................. 28
Figure 3.7. Hourly MAPE for all forecasts generated by the MLR models ..................... 42
Figure 3.8. Hourly MAPE for all forecasts generated by the NN models ........................ 43
Figure 3.9. Hourly MAPE for all forecasts generated by the SVM models ..................... 43
Figure 3.10. Hourly MAPE for all forecasts generated by the SAM models ................... 44
Figure 3.11. Box plots of hourly MAPE results. (top) MLR without building
Figure 3.13. Box plots of hourly MAPE results. (top) SVM without building
Figure 3.14. Box plots of hourly MAPE results. (top) SAM without building
Figure 3.15. Hourly Bias for all forecasts generated by the MLR models ....................... 49
Figure 3.16. Hourly Bias for all forecasts generated by the NN models .......................... 49
Figure 3.17. Hourly Bias for all forecasts generated by the SVM models ....................... 50
Figure 3.18. Hourly Bias for all forecasts generated by the SAM models ....................... 50
Figure 3.19. Hourly standard deviation of the forecast error for all forecasts generated by
Figure 3.20. Hourly standard deviation of the forecast error for all forecasts generated by
Figure 3.21. Hourly standard deviation of the forecast error for all forecasts generated by
Figure 3.22. Hourly standard deviation of the forecast error for all forecasts generated by
Figure 3.23. Empirical distribution of forecast errors for the NN model without building
Figure 3.24. Empirical distribution of forecast errors for the NN model with building
models ............................................................................................................................... 56
Figure 3.26. Frequency of forecast errors within a ±30 kW error margin for the MLR
models ............................................................................................................................... 57
Figure 3.27. Frequency of forecast errors within a ±10 kW error margin for the NN
models ............................................................................................................................... 57
Figure 3.28. Frequency of forecast errors within a ±30 kW error margin for the NN
models ............................................................................................................................... 58
Figure 3.29. Frequency of forecast errors within a ±10 kW error margin for the SVM
models ............................................................................................................................... 58
Figure 3.30. Frequency of forecast errors within a ±30 kW error margin for the SVM
models ............................................................................................................................... 59
Figure 3.31. Frequency of forecast errors within a ±10 kW error margin for the SAM
models ............................................................................................................................... 59
Figure 3.32. Frequency of forecast errors within a ±30 kW error margin for the SAM
models ............................................................................................................................... 60
Figure 3.33. Hourly skew of the forecast error for all forecasts generated by the MLR
models ............................................................................................................................... 62
Figure 3.34. Hourly skew of the forecast error for all forecasts generated by the NN
models ............................................................................................................................... 62
Figure 3.35. Hourly skew of the forecast error for all forecasts generated by the SVM
models ............................................................................................................................... 63
xiii
Figure 3.36. Hourly skew of the forecast error for all forecasts generated by the SAM
models ............................................................................................................................... 63
Figure 3.37. Hourly kurtosis of the forecast error for all forecasts generated by the MLR
models ............................................................................................................................... 65
Figure 3.38. Hourly kurtosis of the forecast error for all forecasts generated by the NN
models ............................................................................................................................... 65
Figure 3.39. Hourly kurtosis of the forecast error for all forecasts generated by the SVM
models ............................................................................................................................... 66
Figure 3.40. Hourly kurtosis of the forecast error for all forecasts generated by the SAM
models ............................................................................................................................... 66
Figure 4.1. Example of the difference between a 95% confidence interval (CI) and 95%
Figure 4.2. Seasonal block segmentation (a) and double seasonal block segmentation (b)
........................................................................................................................................... 76
Figure 4.3. Block bootstrapped time series (a) and double seasonal block bootstrapped
Figure 4.9. Average hourly skill score for the MLR models. Nominal coverage rate =
25% ................................................................................................................................... 86
xiv
Figure 4.10. Average hourly skill score for the MLR models. Nominal coverage rate =
75% ................................................................................................................................... 86
Figure 4.11. Average hourly skill score for the NN models. Nominal coverage rate = 25%
........................................................................................................................................... 87
Figure 4.12. Average hourly skill score for the NN models. Nominal coverage rate = 75%
........................................................................................................................................... 87
Figure 4.13. Average hourly skill score for the SVM models. Nominal coverage rate =
25% ................................................................................................................................... 88
Figure 4.14. Average hourly skill score for the SVM models. Nominal coverage rate =
75% ................................................................................................................................... 88
Figure 4.15. Average hourly skill score for the SAM models. Nominal coverage rate =
25% ................................................................................................................................... 89
Figure 4.16. Average hourly skill score for the SAM models. Nominal coverage rate =
75% ................................................................................................................................... 89
Figure 5.1. Electrical load data in %FLA and chilled water outlet temperature data in oF.
Figure 5.2. Electric load (blue line) and temperature (red line) response to raising the chill
Figure 5.3. Response to a step change in the chilled water temperature setpoint value
Figure 5.4. Application of the curve fit approach on the collected data for parameter
Figure 6.1. Sample building load forecast and dispatched load ..................................... 105
xv
Figure 6.2. Relationship between load margin and building load forecast uncertainty
Figure 6.4. LMP data for Monday July 21st, 2014 .......................................................... 120
Figure 6.5. Dispatch schedule for buildings 7 with decreasing load flexibility (No
Figure 6.6 Dispatch schedule for buildings 7 with decreasing load footprint (No dispatch
Figure 6.7 Cost savings as a function of load uncertainty. Break even line shown in blue
......................................................................................................................................... 130
Figure 6.8. Decreasing number of intervals where the DR revenue constraint can be met
......................................................................................................................................... 131
Figure B.1 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.2 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.3 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.4 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.5 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.7 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.8 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.9 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.10 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.11 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.12 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.13 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.14 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.15 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.16 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.18 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.19 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.20 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.21 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.22 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.23 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.24 MLR Empirical error distribution (L) No building measurements (R) With
Figure B.25 NN Empirical error distribution (L) No building measurements (R) With
Figure B.26 NN Empirical error distribution (L) No building measurements (R) With
Figure B.27 NN Empirical error distribution (L) No building measurements (R) With
Figure B.29 NN Empirical error distribution (L) No building measurements (R) With
Figure B.30 NN Empirical error distribution (L) No building measurements (R) With
Figure B.31 NN Empirical error distribution (L) No building measurements (R) With
Figure B.32 NN Empirical error distribution (L) No building measurements (R) With
Figure B.33 NN Empirical error distribution (L) No building measurements (R) With
Figure B.34 NN Empirical error distribution (L) No building measurements (R) With
Figure B.35 NN Empirical error distribution (L) No building measurements (R) With
Figure B.36 NN Empirical error distribution (L) No building measurements (R) With
Figure B.37 NN Empirical error distribution (L) No building measurements (R) With
Figure B.38 NN Empirical error distribution (L) No building measurements (R) With
Figure B.40 NN Empirical error distribution (L) No building measurements (R) With
Figure B.41 NN Empirical error distribution (L) No building measurements (R) With
Figure B.42 NN Empirical error distribution (L) No building measurements (R) With
Figure B.43 NN Empirical error distribution (L) No building measurements (R) With
Figure B.44 NN Empirical error distribution (L) No building measurements (R) With
Figure B.45 NN Empirical error distribution (L) No building measurements (R) With
Figure B.46 NN Empirical error distribution (L) No building measurements (R) With
Figure B.47 NN Empirical error distribution (L) No building measurements (R) With
Figure B.48 NN Empirical error distribution (L) No building measurements (R) With
Figure B.49 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.51 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.52 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.53 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.54 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.55 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.56 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.57 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.58 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.59 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.60 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.62 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.63 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.64 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.65 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.66 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.67 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.68 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.69 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.70 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.71 SVM Empirical error distribution (L) No building measurements (R) With
Figure B.73 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.74 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.75 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.76 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.77 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.78 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.79 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.80 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.81 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.82 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.84 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.85 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.86 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.87 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.88 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.89 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.90 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.91 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.92 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.93 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.95 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.96 SAM Empirical error distribution (L) No building measurements (R) With
Figure B.97 Average hourly skill score for the MLR Models. Nominal coverage rate =
5% ................................................................................................................................... 190
Figure B.98 Average hourly skill score for the MLR Models. Nominal coverage rate =
Figure B.99 Average hourly skill score for the MLR Models. Nominal coverage rate =
Figure B.100 Average hourly skill score for the MLR Models. Nominal coverage rate =
Figure B.101 Average hourly skill score for the MLR Models. Nominal coverage rate =
Figure B.102 Average hourly skill score for the MLR Models. Nominal coverage rate =
Figure B.103 Average hourly skill score for the MLR Models. Nominal coverage rate =
Figure B.104 Average hourly skill score for the MLR Models. Nominal coverage rate =
Figure B.106 Average hourly skill score for the MLR Models. Nominal coverage rate =
Figure B.107 Average hourly skill score for the MLR Models. Nominal coverage rate =
Figure B.108 Average hourly skill score for the MLR Models. Nominal coverage rate =
Figure B.109 Average hourly skill score for the MLR Models. Nominal coverage rate =
Figure B.110 Average hourly skill score for the MLR Models. Nominal coverage rate =
Figure B.111 Average hourly skill score for the MLR Models. Nominal coverage rate =
Figure B.112 Average hourly skill score for the MLR Models. Nominal coverage rate =
Figure B.113 Average hourly skill score for the MLR Models. Nominal coverage rate =
Figure B.114 Average hourly skill score for the MLR Models. Nominal coverage rate =
Figure B.115 Average hourly skill score for the MLR Models. Nominal coverage rate =
5% ................................................................................................................................... 196
Figure B.117 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.118 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.119 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.120 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.121 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.122 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.123 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.124 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.125 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.126 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.128 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.129 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.130 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.131 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.132 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.133 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.134 Average hourly skill score for the NN Models. Nominal coverage rate =
Figure B.135 Average hourly skill score for the SVM Models. Nominal coverage rate =
5% ................................................................................................................................... 202
Figure B.136 Average hourly skill score for the SVM Models. Nominal coverage rate =
Figure B.137 Average hourly skill score for the SVM Models. Nominal coverage rate =
Figure B.139 Average hourly skill score for the SVM Models. Nominal coverage rate =
Figure B.140 Average hourly skill score for the SVM Models. Nominal coverage rate =
Figure B.141 Average hourly skill score for the SVM Models. Nominal coverage rate =
Figure B.142 Average hourly skill score for the SVM Models. Nominal coverage rate =
Figure B.143 Average hourly skill score for the SVM Models. Nominal coverage rate =
Figure B.144 Average hourly skill score for the SVM Models. Nominal coverage rate =
Figure B.145 Average hourly skill score for the SVM Models. Nominal coverage rate =
Figure B.146 Average hourly skill score for the SVM Models. Nominal coverage rate =
Figure B.147 Average hourly skill score for the SVM Models. Nominal coverage rate =
Figure B.148 Average hourly skill score for the SVM Models. Nominal coverage rate =
Figure B.150 Average hourly skill score for the SVM Models. Nominal coverage rate =
Figure B.151 Average hourly skill score for the SVM Models. Nominal coverage rate =
Figure B.152 Average hourly skill score for the SVM Models. Nominal coverage rate =
Figure B.153 Average hourly skill score for the SVM Models. Nominal coverage rate =
Figure B.154 Average hourly skill score for the SAM Models. Nominal coverage rate =
5% ................................................................................................................................... 209
Figure B.155 Average hourly skill score for the SAM Models. Nominal coverage rate =
Figure B.156 Average hourly skill score for the SAM Models. Nominal coverage rate =
Figure B.157 Average hourly skill score for the SAM Models. Nominal coverage rate =
Figure B.158 Average hourly skill score for the SAM Models. Nominal coverage rate =
Figure B.159 Average hourly skill score for the SAM Models. Nominal coverage rate =
Figure B.161 Average hourly skill score for the SAM Models. Nominal coverage rate =
Figure B.162 Average hourly skill score for the SAM Models. Nominal coverage rate =
Figure B.163 Average hourly skill score for the SAM Models. Nominal coverage rate =
Figure B.164 Average hourly skill score for the SAM Models. Nominal coverage rate =
Figure B.165 Average hourly skill score for the SAM Models. Nominal coverage rate =
Figure B.166 Average hourly skill score for the SAM Models. Nominal coverage rate =
Figure B.167 Average hourly skill score for the SAM Models. Nominal coverage rate =
Figure B.168 Average hourly skill score for the SAM Models. Nominal coverage rate =
Figure B.169 Average hourly skill score for the SAM Models. Nominal coverage rate =
Figure B.170 Average hourly skill score for the SAM Models. Nominal coverage rate =
Figure B.172 Average hourly skill score for the SAM Models. Nominal coverage rate =
ABSTRACT
Modeling and Analysis of Commercial Building Electrical Loads for Demand Side
Management
Jonathan Berardino
Chikaodinaka Nwankpa, Ph.D.
In recent years there has been a push in the electric power industry for more
customer involvement in the electricity markets. Traditionally the end user has played a
passive role in the planning and operation of the power grid. However, many energy
markets have begun opening up opportunities to consumers who wish to commit a certain
amount of their electrical load under various demand side management programs. The
potential benefits of more demand participation include reduced operating costs and new
revenue opportunities for the consumer, as well as more reliable and secure operations for
the utilities. The management of these load resources creates challenges and
opportunities to the end user that were not present in previous market structures.
This work examines the behavior of commercial-type building electrical loads and
their capacity for supporting demand side management actions. This work is motivated
by the need for accurate and dynamic tools to aid in the advancement of demand side
operations. A dynamic load model is proposed for capturing the response of controllable
information. These approaches are tested using Drexel University building data. The
application of building-specific load forecasts and dynamic load modeling to the optimal
1. INTRODUCTION
1.1. OVERVIEW
This thesis presents a study of commercial buildings as potential resources for demand
side management. Within this work, methods for modeling and forecasting building electric load
behavior are presented. These problems are examined through extensive case studies involving
Drexel University building data. The resulting modeling and forecasting approaches are applied
to the problem of optimally scheduling building loads as part of several demand side
building loads, subject to real-world building operational constraints and building load
variability.
Power system operators are tasked with ensuring adequate generation is available to meet
projected demand. Traditionally this requirement is met through controlling the supply side (i.e.
- generators) of the power system; limited priority has been placed on involving the demand side.
However, increased investment in the development of a “Smart Grid” [1] is driving new
opportunities for the demand side to take a more active role in power system planning and
operations. Demand Side Management (DSM), also referred to often as Demand Response
(DR), is a class of programs that are designed to motivate end-use customers to modify their
2
electricity usage. This could be through the shifting of electricity use to another time (e.g. – off
The idea of managing demand side resources has been discussed since the 1890’s, as
described in detail in [2]. However, it was not until the restructuring and deregulation of electric
utilities in the 1990’s and subsequent issues that began to arise in the new wholesale markets that
a concerted effort was made to include DSM as an essential aspect of these new markets [3]. In
support of this, the United States government has issued a number of policies in an effort to
remove potential barriers to DSM participants [4] [5] [6] [7]. These programs have been
identified as having the potential to provide a wide range of benefits to both the power system
operator and the end-user [8] [9] [10]. Potential benefits include:
Reduction in grid demand during peak hours and subsequent reduction in the
reliance on expensive peaking units [11]. In the United States, DSM potential is
estimated to be between 38GW and 188GW by 2019 [8]. Peak reduction also has
the added potential to defer transmission system infrastructure upgrades that may
Reduction in wholesale energy prices and decreased price volatility. Even a small
increase in demand elasticity can offset the extreme increase in generation cost at
Reduced electricity costs and new revenue opportunities through the local
supporting power system operations in owing to a high load footprint [19] and controllability that
is at least on par with that of generators [20]. The future involvement of building loads in
infrastructure (AMI). Such advancements bring a new level of two-way communication and
The motivation for this thesis is the need for improved methods to assess building-
specific DSM opportunities and support increased involvement of building loads in various DSM
programs. In order for building participation in DSM programs to continue to evolve, the
that end, this thesis will consider what tools are needed from the perspective of the end-user in
order to understand and plan their own building loads as part of a DSM program.
1.3. OBJECTIVES
Describe the existing standard of DSM programs and the potential for using
o Drexel specific load forecasting case study with results and observations
RESOURCES
2.1. OVERVIEW
This chapter examines the potential of using commercial building loads as controllable
building demand
building loads and evaluating building DSM capabilities. These methods will be
As mentioned previously, there are a wide variety of DSM programs that may be
available to the end-user. These programs are generally grouped into two basic types: price-
based and incentive based programs [10]. Figure 2.1 shows this classification of various DSM
programs. Price-based programs, as the name would imply, use variable electricity pricing
provide variable rate plans to customers where the rates are defined well in advance of pre-
7
defined usage periods [21]. These usage periods can include, for example, daily peak or off
peak hours or seasonal time windows where each period is priced differently.
In contrast to TOU pricing, dynamic pricing sets electricity rates that are not known to
the end-user so far in advance. These rates may be set on a day-ahead or an hour-ahead time
scale. Dynamic pricing is a key factor in both Critical Peak Pricing (CPP) and Real Time Pricing
(RTP) programs. During CPP events which occur during high system demand days several
times a year, enrolled participants receive higher electricity rates to encourage reduced
rates [22]. In RTP programs, the cost of electricity is variable from day to day or hour to hour.
At present time, RTP programs are not very common. These types of programs rely on two-way
communication between the utility and end-user for pricing information. As more AMI projects
and other communication and control advancements are put into operation, RTP programs
should become more widespread. It is arguable that the evolution of DSM will rely heavily on
market based programs. Direct control refers to programs where the end-user allows the utility
to control their load in exchange for an agreed upon payment or credit. Interruptible Load
Management (ILM), which targets mainly large industrial customers, and Direct Load Control
(DLC) programs, which focus mainly on residential customers, have been around since the
1970’s [2] to support peak load management. These programs continue to be employed by
utilities today.
Market based programs are a more recent expansion of DSM opportunities which allow
the end-user to participate in the wholesale electricity market. In Demand Bidding and Capacity
Market programs, the end-user bids demand reductions and, if the bids are accepted, must
provide demand reductions at the specified time [23]. In these programs the demand-side bids
are optimized with the supply-side bids, effectively treating loads and generators as equal players
in the market. Participation in the Ancillary services market includes, for example, registering
controllable demand as resources for regulation or spinning reserve applications [14] [15] [16]
[17] [18].
The DSM programs described above continue to grow in participation as time goes on. It
is expected these markets will continue to advance and that building loads will be increasingly
utilized as DSM sources in all programs [8]. While building loads make attractive DSM
participants there are still practical challenges that remain in order to evaluate participation in
DSM applications. Several qualities of the loads themselves need to be better understood: what
factors drive building demand, what temporal behavior do they present, can the behavior be
described accurately in a form conducive to automated decision making, and what uncertainties
One of the most appealing aspects of using commercial building loads for DSM purposes
is the large electrical footprint and potential for controllability. In 2014, commercial building
energy use made up 34% of the total United States electricity demand [19]. Within these
buildings, over 50% of the electricity was consumed by the HVAC and lighting systems as
Figure 2.2. Breakdown of commercial building energy use in the United States (2014) [19]
It stands to reason given this breakdown that improved operation of HVAC systems has
been identified as a means of potential energy savings and increased energy efficiency in
commercial buildings [24]. It also indicates a strongly coupled relationship between the building
thermal mass and building demand. By exploring this relationship, important information can be
obtained in regards to factors driving building load and the dynamics of building load behavior.
The approach taken in this thesis is to focus on information available from building
management systems (BMS) typical of commercial buildings that influence building electrical
behavior. These measurements are directly linked to the thermal-electrical behavior that drives a
large portion of the building demand. Additionally, this approach can lead to the development of
10
data-driven building models that are well suited for integration in a DSM scheduling tool
Many medium to large commercial and industrial buildings operate a BMS that is
responsible for the monitoring and control of various building mechanical and electrical system
performance, such as the heating and cooling systems. These BMS are typically configured as a
distributed control system, with a software layer managing the functions of hardware
order to study building load features, electrical and thermal data has been collected using the
Drexel University BMS over the last several years. Three main variables are considered from
this data:
3. Target Temperature Gradient: Outside Air Temp. – Indoor Air Temp. Setpoint (OAT-
Stpt)
For the studies presented in this chapter, the data used is from the summer months (May-
August) of 2011 and 2012. A large portion of the load for campus buildings is the HVAC
system, making them good candidates as controllable loads during the typically warm and humid
summers. Consequently, that makes these good months for studying building behavior to
support DSM planning activities. It should be noted that Drexel does not have electric heating in
the majority of buildings on campus (and none where data recording capabilities exist).
However the approach employed in this thesis could easily be employed in characterizing the
building load behavior if heating loads were also a potential controllable resource.
Figure 2.3 shows the linear correlation between demand and each BMS measured
Figure 2.3. Linear correlation between demand and several variables: OAT (top), OAT-Temp (middle), and
OAT-Stpt (bottom)
Particularly strong correlation is noted between demand and the target temperature
gradient. This relationship makes sense intuitively, as the building HVAC system needs to draw
more energy to maintain a given setpoint inside the building when temperature outside increases.
These results indicate a reasonable correlation, particularly when considering the relationship
between demand and temperature is inherently nonlinear. Additional correlation studies have
In addition to correlation with measured BMS variables, there are temporal dependencies
that arise due to the thermal-electrical coupling present in large HVAC electrical loads [26] [27],
which make up a considerable portion of the load base for commercial buildings [28] [19].
Figure 2.4 shows an example of the correlation between measured demand values with future
demand levels.
12
Linear Correlation Between Demand Readings (Summer 2011)
0 1
3 0.9
0.8
6
0.7
9
0.6
12
0.5
15
0.4
18
0.3
21
0.2
24
0 3 6 9 12 15 18 21 24
Figure 2.4. Temporal correlation for 24 hours of building demand. Sample size is 4 months of weekdays (88
days – 2011)
Figure 2.4 depicts how the demand at each point in time is correlated to the demand at a
future point in time over 24 hours, with samples taken every 5 minutes. For this example there
are large pockets of high correlation shown, particularly between 4am and 8am and 1pm and
8pm. This serial correlation between demand observations is representative of the temporal
dependencies found in building load behavior. The HVAC consumption, and indeed that of
other load components, is driven not only by environmental factors but also building operational
schedules and occupant behaviors. When compared to the residential sector, commercial
buildings experience less variance day in and day out due to more consistent schedules and
occupant behaviors. However, from the demand usage patterns, important intraday
develop improved, data-informed building demand profiles [29]. These patterns are inherently
It has now been established that this thesis will focus on applying information available
building resources for DSM are reviewed and contrasted with the approach taken in this thesis.
RESEARCH
Before buildings can be deployed as part of a DSM program, two questions must be
addressed: how accurately can the load of the building be predicted and how can a building load
be integrated in a DSM application. This thesis argues that the answer to both questions is more
realistic, and consequently more accurate, if measurements from the BMS play a role in the
Given the potential of buildings as DSM resources it is not surprising that there is
considerable research that has been conducted in attempting to answer these questions. The next
two subsections look at the state-of-the-art in predicting building loads and modeling and
The load shed during a DSM event must be measured against a forecast of what the
demand would be if no adjustment occurred. This forecast is often referred to as the baseline
demand. Figure 2.5 shows an example of the baseline demand and actual metered load during a
DSM event. The area between the magenta and blue vertical lines is the period of time in which
Baseline
Metered Load
220
200
160
140
120
0 4 8 12 16 20 24
Time of Day (hours)
Utilities generally use simple baseline models that involve averaging the daily demand
over several days (excluding DSM event days) [30]. For example, CAISO introduced a “3-in-
10” method where the baseline is based on the hourly average of the three highest energy usages
in the past ten similar days [31]. PJM used the same approach except they look at four out of the
past 5 similar days [23]. Recent adjustments to both the CAISO and PJM models have been
made to improve the accuracy. Observations of metered demand are collected several hours
prior to the beginning of a DSM event and the original baseline calculations are adjusted up or
down based on these more recent demand readings. This adjustment has improved the accuracy
in quantifying demand behavior for economic settlement post event. However, this adjustment
provides no benefit for forecasting building loads except in near real-time applications since it
to the averaging methods discussed above [32] [33] [34] [35]. These methods include predictor
variables in the model such as weather and calendar information in an attempt to better predict
load. However, a comprehensive analysis of baseline calculation methods prepared for PJM
15
showed only 2 of 13 unadjusted methods applied a regression approach [36] and only one
method was actually employed by a utility (ERCOT). The results of this study also indicated the
regression models, as currently designed, did not necessarily outperform the other models in the
Beyond the averaging and regression-based baseline calculation approaches that have
been discussed, there is a surprising lack of more sophisticated approaches in the technical
literature [37]. This can be attributed more to terminology rather than research effort however.
The DSM community tends to use the term baseline as opposed to forecast. Baseline has the
very specific connotation as the load profile against which incentive-based DSM performance is
measured when the utility and end-user reach settlement following a DSM event. There are
other research efforts that focus on building-specific forecasts for studying building load
behavior. The models in these works tend to be more complex and less intuitive than the
baseline models but they more thoroughly capture building load behavior and produce more
consistently accurate load forecasts. These works are not solely concerned with quantifying
building behavior for DSM, although that is still often a motivation. Going forward, this thesis
will use the term forecast to refer to any approach concerned with predicting the future building
electricity demand.
Methods for forecasting end-user facility demand are presented in [30] [25] [38] [39] [40]
[41] [42] [43] [44] [45] [46] [47] [48] [49]. These works have applied a wide variety of methods
that are often used in short term load forecasting (STLF) studies but at a campus or building
level. In [30] [25] [38] [39] regression models are used. [40] [41] [42] [43] [44] employ several
different forms of neural network models. [45] proposes a new day-ahead probabilistic model
based on Gaussian processes for an industrial facility. Support vector machines (SVM) are used
in [46] to forecast monthly demand at four commercial sites in Singapore and in a short-term
16
application in [47]. In [48] and [49], several forecasting models are combined to create
ensemble models for forecasting building demand. All of these works recognize that forecasting
at the building level must deal with a higher degree of variability, a feature that becomes washed
out to a degree when forecasting at a transmission substation level with aggregated loads.
However, of these references, only [47] included internal building measurements in their forecast
models. The measurements were collected from a limited sensor network installed for testing
purposes. These measurements included occupancy information from sensors at the two
building entrances and four temperature sensors distributed through the building.
In contrast to the works above, the approach taken in this thesis is to focus on including
buildings that influence building electrical behavior. As discussed earlier in this chapter, these
measurements are directly linked to the thermal-electrical behavior that drives a large portion of
the building demand. The goal is to enhance existing STLF techniques with these building-
specific measurements and compare the performance against similarly trained models that do not
come from the HVAC engineer community. This research tends to provide highly detailed
models of the building thermal dynamic behavior [50]. Often these models require a thorough
knowledge of the building construction and equipment profiles. In general these models are far
too complex to be used for evaluating DSM applications for a given building, let alone a group
Power systems researchers are increasingly interested in how building loads can be
integrated in the electric grid through DSM. However, the majority this research has centered
17
around the integration of bulk demand response by the independent system operator (ISO)
when scheduling system resources and performing system security analysis [51] [52] [53] [54]
[55]. It is assumed that building loads can play a role as DSM resources in these scenarios but
information can be derived from these processes that would help the end-user achieve in practice
From the perspective of the end-user it is crucial to have a method of evaluating one’s
own potential for DSM activity; this includes models of building behavior and formal method of
determining when and how to dispatch loads. This is particularly true when demand
modifications are driven through control of the HVAC system given the potential for significant
impact to the comfort levels of the building. Modeling the behavior of thermostatically
controllable loads (TCL) for DSM has been studied in [14] [56] [57] [58]. While there is
potential applications for these models they are better suited for the aggregation of many
residential or small commercial loads [14] and overlook the fact that cycling HVAC systems
between on and off states is undesirable for reliability and efficiency reasons [59].
There remains a need for a model that captures the thermal-electrical dynamics of
commercial building loads with an eye towards controllability and simple implementation in a
DSM scheduling for the consumer. In this thesis such a model is developed in Chapter 5 using
methods applied previously for the development of power system dynamic load models and
again leveraging information collected from the BMS. Chapter 6 evaluates how this model,
combined with the building-specific forecasts from Chapters 3 and 4, can be used in a method of
planning commercial building load resources for DSM directed at the end-user.
18
3.1. OVERVIEW
In this chapter, the problem of building-specific load forecasting is studied. The main
techniques and evaluating the performance. This chapter includes the following:
A case study where conventional forecasting models are trained on data collected
from the Drexel University BMS and a series of day-ahead load forecasts are
performed.
Results from the case study are used to evaluate building-specific forecasting
Historically, load forecasting (in particular short term load forecasting (STLF) ) has played a
critical role in ensuring power system dispatchers schedule adequate generating capacity in the
most economical way possible. The importance of having access to accurate and consistent
methods of load forecasting is obvious. Consequently, a considerable amount of work has been
done to develop and examine methods of forecasting power system loads over varying time
horizons [60] [61]. In this thesis the focus is on day-ahead forecasts, which fall into the STLF
window.
19
Load forecasting approaches generally fall into two categories: statistical approaches
such as classical time series analysis and regression–based models, and artificial intelligence
(AI) models such as artificial neural networks and fuzzy logic models [62]. Methods from both
of these categories have been applied to solving the STLF problem. A very thorough review and
critique of STLF techniques introduced prior to 2010 is presented in [63]. The popularity of AI
methods has led to a lot of research resources being devoted to applying newly developed AI
techniques to STLF. Far less attention is placed on applying modern statistical methods to this
problem. This is despite, as noted in [63], that statistical methods are much more widely used to
An important contribution from [63] is the focus is not only on the variety of techniques
but also what predictor variables are employed in the forecasting methods and how said variables
can lead to conclusions regarding the causality of load consumption. Nearly all of the research
surveyed included some predictor variables and this thesis will consider their impact on
variables”). One of the conclusions is that successful inclusion of such variables depends on an
understanding of the geography of the power system under study and time frame of interest as
much as the method used to integrate these variables into a model. Without understanding this
behavior the chosen method will not successfully capture the link between demand and any
In the years since the review in [63], the technical literature related to STLF has explored
additional methods for capturing the relationship between demand and additional predictor
variables. Artificial neural network models continue to be widely used owing to the inherent
capability of being able to learn and capture complex linear and nonlinear relationships from the
20
data to be modeled [64]. Such models have been recently applied in a variety of forms [65]
[66] [67] [68] [69] [70] to take advantage of this property. Additional AI techniques such as
Support Vector Machines (SVM) and other supervised learning algorithms [71] [72] [73] [74]
have also been used to achieve suitable results. On the statistical method front, semi-parametric
additive models have recently shown very good results at both the transmission and distribution
system level [75] [76] [77] [78]. These works employ a regression-based structure with smooth,
nonlinear functions used to capture the link between demand and a number of covariates.
variable given a set of information about said variable. This information may include only
historical observations of the variable of interest (univariate case) or observations and predictions
of related predictor variables as well (multivariate case). There is an implicit assumption that
For this work, we are interested in predicting the demand level p for a given building.
Although electric power is consumed by a building continuously, the observations are discrete
based on the sampling of the electric meter. Therefore, demand is represented by a discrete time-
series pt t 1,.., where pt is the metered demand at time t and T is the collection of time
The general load forecasting problem can be defined as a function f of the available
pˆ t k|t f Φt (3.1)
where
21
K: Forecast horizon
k 1,..., K : Time index within forecast horizon
pˆ t k |t : Estimate of p given observations up to time t looking
k steps ahead
Φt : Collection of historical demand and predictor variables
with
where
Figure 3.1 shows a sample load forecast. In the plot, the division is shown between
historical observations pt-m (blue line) and estimated demand pˆ t k|t (grey line), also noting the
The forecast horizon K, also referred to as the look-ahead time in the forecasting
literature, defines how far into the future one is looking to forecast. If K > 1, the forecast is
called a multiple step ahead forecast. The estimated model remains fixed for the duration of the
forecast horizon. The predictor variable estimates xˆ t k |t are used as inputs to the model in order
to generate the final forecast pˆ t k|t . An iterative method can also be used which effectively turns
the problem into a single step ahead forecast. At each step k, the estimate pˆ t k|t is found. This
estimate is then treated as a previous observation, becoming an input to the same model in order
to forecast the subsequent point. The process continues in this manner until reaching the end of
the forecast horizon. This approach is more often used for univariate forecasts and has the
disadvantage that the errors in the predicted values are accumulated into the next predictions.
23
The predictor intervals contained in Φ will ideally include only variables that have a
causal relationship with the demand p. Identification of such variables can be done prior to
estimation of the function f by observing correlation plots such as the one shown in Figure 2.3.
Variable identification can also be done by training several forecast models with different
observations, and selecting the “best” model. In this thesis, two groups of predictor variables
will be used: one which includes building-specific measurements and one that does not. The
specifics of this are described in the model building procedure in subsection 3.4.3.
In order to test the idea of enhancing conventional load forecasting methods with
building-specific information, several STLF techniques will be used to forecast the day-ahead
demand of a building on Drexel University’s campus. The following subsections will describe
the data set used in this study, introduce the selected forecasting methods and describe the model
building procedure.
Data has been collected using the Drexel University building management system (BMS)
over the last several years. For the studies presented in this thesis, the data used is from the
summer months (May-August) of 2011, 2012, and 2014. As mentioned previously, the summer
weather conditions and large HVAC building load make these good months to study load
forecasting performance in support of demand side management activities. This case study will
focus on the Hagerty Library however the approach is generic to any similar commercial-type
building.
24
The raw data includes four variables, all recorded at 5 minute intervals:
The summer months (May-August) of 2011 and 2012 are used as the training set and
2014 held out to be used for out-of-sample testing. This distinction is shown in Figure 3.2
below. By training and testing the models on separate data sets the problem of overfitting can be
avoided. The training set can be broken into a sample used for estimating model parameters and
a sample used to validate these parameters. This is not required for all methods in this study.
Instances where this is the case will be discussed in the methods section 3.4.2.
Figure 3.2. Breakup of data set and how each portion is used in the forecasting process
Unfortunately the 2013 data had to be excluded from this study due to a large number of
bad measurements in this data set. These issues are attributed to BMS data collection
functionality problems. However, the building involved in this study is known to have
acceptable to leave out this data set for the purposes of this research effort.
Figure 3.3 below shows the daily building demand measurements for the test set portion
of the data (2014). The superimposed dark line represents the mean load profile for the test set.
It is these measurements against which the forecasts will be compared in this case study.
25
Daily Building Demand (116 days) with Mean Profile
300
250
200
Demand (kW)
150
100
50
0 4 8 12 16 20 24
Hours
Figure 3.3. Daily building demand curves for the 116 days used for out of sample testing. The thicker line
shows the mean daily profile
This work examines the performance of four popular forecasting methods used in a
building-specific application. The selected methods were chosen to represent a broad cross
section of techniques. Detailed descriptions of each approach are well established in other works
but the general problem formulations are outlined below. Information specific to how these
A multiple linear regression model relates a dependent (or response) variable p to two or
more independent (or predictor) variables x. The general model is shown in (3.3) [79]:
In many cases, the predictor variables are quantitative such as temperature or wind speed.
However, the formulation in (3.3) does not limit the predictor variables to quantitative ones.
Qualitative predictor variables, often referred to as indicator or dummy variables, can also be
included in the model and are particularly important in forecasting applications. Indicator
variables with values 0 and 1 can be used to identify the category of a quantitative variable, for
example indicating a weekday versus a weekend. A qualitative variable with c categories must
representing the day of the week has 7 categories (i.e. - Sunday, Monday, … , Saturday) and is
represented by 6 indicator variables as shown in Table 3.1. Choosing to use c variables will
The coefficients β represent the partial effect of one predictor variable when all others are
held constant [79]. In other words they represent the marginal effect of each predictor variable.
Given a set of training data, the β values are estimated using the ordinary least squares estimation
method. The models used in this thesis were implemented in “R” using the ‘lm’ function. This
Artificial neural networks (NN) are mathematical tools inspired by the way the human
brain processes information. The most basic computational unit of an NN is the neuron. The
neuron receives information, processes it internally, and provides a response. Figure 3.4 shows a
In general, the information is processed in two stages. First, the input values are linearly
combined. Each value of the input array is associated with a weight value wi. An additional
input, a constant bias term θ, with a weight value equal to 1 is also applied. Second, this
combination becomes the argument of a non-linear activation function. There are a number of
possible functions but a very common choice (and the one implemented in this thesis) is the
1
f x (3.4)
1 e x
The organization of the neurons defines the architecture of the NN. A feed-forward
multilayer perceptron (MLP) neural network is a typical NN architecture employed for STLF. In
this architecture, the neurons are organized in layers where no neuron in a given layer is
connected to another neuron in the same layer, though they can share inputs. The term feed-
28
forward means that the outputs of one layer become the inputs to the following layer. An
Figure 3.5. A general multi-layer, feed-forward artificial neural network with N hidden layers
The parameters of this network are the matrices of weights between each neuron and its
associated input. Input in this case also means connections between neurons in different layers.
Estimation of these weight matrices is referred to as “training” the network. The most widely
applied approach to training in STLF is supervised learning. For this approach, sets of inputs
and matched outputs are used as teaching patterns and the network weights that provide the best
fit between the network output and teaching output are found. Best fit is determined through
minimization of a loss function. The available training algorithms and loss functions are varied,
but historically the back-propagation method and mean squared error criterion are common in
The neural network used in this thesis is a three-layer feed-forward MLP implemented
using the Matlab Neural Network Toolbox. The network architecture and training method used
are as described in the preceding paragraphs. The model is used to forecast multiple steps ahead;
that is to say 288 point forecasts, corresponding to a single day-ahead forecast at 5 minute
29
resolution. This results in a large neural network model that may be computationally
undesirable for implementation in a building level energy management system. However for the
purposes of this study it is sufficient and more computationally sensitive techniques for building
One other item of note when constructing the NN model used in this work is that there
are no hard and fast rules guiding the selection of the number of hidden neurons [64]. The
hidden neurons are the neurons in the layer (or layers) between the input and output layers. The
model in this work uses 25 hidden neurons. This number was determined by estimating the
model on a portion of the training set as shown in Figure 3.2 with a varying the number of
neurons and observing the accumulated error on the validation set. Varying the number of
neurons did not appear to significantly affect the error results, which is not uncommon [64], and
25 was selected.
Support vectors and support vector machines (SVM) are a machine learning technique
used for data classification and regression [80]. Assume given training data (x1 , y1),…,(xn , yn),
where xi are the inputs and yi the corresponding outputs, the support vector regression solves the
1 n
min * 2
wT w C i i* (3.5)
w , b , , i 1
subject to
yi wT xi b i* (3.6)
w x b y
T
i i
i (3.7)
30
i , 0, i 1,.., n
i
*
(3.8)
where xi is mapped to a higher dimensional space by the function ϕ, i* is the upper
training error and i is the lower limit subject to the δ-insensitive tube y wT x b .
The parameters which control regression quality are the error cost C, the width of the tube δ, and
As noted in [81], since the function ϕ can map xi to a high or even infinite dimensional
1 n n
* Q * i i* + yi i i*
T
min 2
(3.9)
, * i 1 i 1
subject to
i
i* 0 (3.10)
i 1
0 i , i* C , i 1,.., n (3.11)
where Qij= ϕ(xi)Tϕ(xj). To solve this inner product, which may be computationally
difficult, a “kernel trick” is implemented to do the mapping implicitly. In other words, the
application of special forms which are inner products in a higher dimensional space yet can be
calculated in the original space. There are many options for kernel functions but for this work
the radial basis function (RBF) kernel is used. The expression for this kernel is shown in (3.13).
xi x j e
2
T x x i j
(3.12)
31
The RBF kernel can handle nonlinear relationships between xi and yi unlike the linear
kernel yet has fewer parameters than a polynomial kernel, reducing model complexity. This
kernel selection has been used to great effect in STLF applications [80] [73].
There are two key parameters that must be found when training the SVM models: the error
cost C and the RBF parameter γ. To decide the proper parameter values, the training set is
segmented such that the model performance can be evaluated. A portion of the training set is
used for updating the model parameters while the validation set is used to observe the
corresponding model performance. To determine suitable values for C and γ, v-fold cross
validation [82] is performed by dividing the training set into v equally sized subsets. With one
subset held out for validation, the model is trained on the other v-1 subsets. This process is
repeated using each subset as a validation set and the model performance aggregated. In this
thesis the models are implemented using the LIBSVM extension in R [81]. This package
automatically handles the parameter estimation procedure by efficiently solving (3.9) while
In a manner similar to the regression structure established for the multiple linear
regression model, the semi-parametric additive models capture nonlinear relationships using the
The functions fk are non-linear, smooth functions that can be well estimated from
observed data. These functions can be multivariate (e.g. – f(x1,x2) ) but for this thesis they will
only be a function of a single variable. It is common for these functions to be estimated via
penalized regression in a spline basis. If bi(x) is the ith basis function, then f can be represented
as in (3.14).
d
f k xk bi xk i (3.14)
i 1
where d is the dimension of the spline basis and bi(xk) are the corresponding spline
functions. There are many possible spline functions available. This work uses cubic regression
splines as was applied in previous STLF works [ [75] [76]]. In order to estimate f a penalized
kv
min p Bβ q βT S q β
2
β ,
(3.15)
q 1
where
model smoothness. The models used in this work are implemented in R using the mgcv package
[84]. The problem in (3.15) is solved using the methodology presented in [85] [86] which
involves minimization of the Generalize Cross Validation (GCV) criteria. This is done
automatically in mgcv with the appropriate spline function and training set specified.
The overall STLF process is summarized in Figure 3.6. (Modified from [63]).
The first step is to use historical information to estimate the model. The predictor
variables contained in Φ includes the data required to estimate the necessary model parameters.
When training each model, the considered predictor variables are shown in Table 3.2.
34
Table 3.2. General and Building-specific predictor variables
For each forecasting method tested in this thesis, two models are built: one which
includes the general variables and building-specific variables and one that only includes the
general variables. The 2011 and 2012 data is used for model estimation. For those forecasting
methods that use a validation set (refer to section 3.4.2 for details), the August 2012 data is set
Once the models are built the forecasts can be generated. Each model is used to forecast
the day-ahead demand for the 2014 test set. The forecasts are generated in 5 minute steps
matching the sample rate of the historical data. A total of 116 day-ahead forecasts were
performed (May 8th - Aug. 31st 2014). The first week of May was excluded as no historical
lagged measurement observations were available prior to May 1st to support forecast generation.
the predictor variables xˆ t k |t into the model. For example, a weather forecast for the next day
would be used as an input as opposed to actual temperature information. Using only information
known in advance, these forecasts are known as ex-ante forecasts. These are the only “true”
forecasts since all future information is unknown. In this thesis, because the goal is model
35
analysis, estimates of the predictor variables xˆ t k |t are replaced with observed values xt k |t , and
the generated forecast pˆ t k |t is compared to the known value pt k |t . These types of forecasts are
known as ex-post forecasts and are useful in demonstrating the potential accuracy of a model and
measurements improve the forecast accuracy, how consistent is this performance, and how does
the forecast uncertainty change when including building information. Most of the forecasting
literature is focused solely on the forecast accuracy. The majority of papers follow the same
general formula: develop a forecast model and then precede to use an accuracy metric that
demonstrates how great a particular model is for that given test. This type of analysis is fine but
a bit limited. It struggles to address consistency of the forecast performance and provides little
this thesis will address not only the overall accuracy for the forecasts but focus particularly on
Forecast error is defined as the difference between the predicted power value and the
The error is composed of a systematic part te k |t and a random part te k |t .
Ideally, the systematic error is zero and the random part is white noise (zero mean,
Gaussian random variable). However, in practice these conditions are rarely the case and
36
examining each part of the error is necessary to understand the impact of building
The standard metric for defining the forecast accuracy in the load forecasting literature is
1 N et k |t
MAPE 100 (3.18)
N k 1 pt k |t
where N is the number of forecasted points in the interval of interest. By varying the timeframe
over which the MAPE is calculated certain behaviors of the forecast method of interest can be
observed. This metric is an overall accuracy measure and both the systematic and random errors
Another basic metric is the forecast method’s bias, given by the mean error over a
specified interval.
N
1
bias ke
N
e
k 1
t k |t (3.19)
where N is the number of forecasted points in the interval of interest. The bias shows if the
method tends to under- or overestimate the forecast. It corresponds to the systematic part of the
forecast error et+k|t. Ideally the forecasts will be unbiased but in practice this is often unrealistic.
The variability of the forecast performance can observed by calculating the standard
1 N
et k |t ke
2
ke (3.20)
N k 1
where N is the number of forecasted points in the interval of interest. The standard deviation
forecast errors represent the empirical distributions of these errors. This type of analysis is
important for characterizing the consistency of the forecast performance by answering the
question “How often does a given forecasting method result in a specific error level?”. For
example, two methods might result in nearly identical MAPE values over a given time window
but have radically different error distribution shapes, thus different frequencies of large errors.
Evaluating the moments of the forecast error distributions will shed light on several important
characteristics:
The mean ke is a measure of the central tendency of the error distribution. As mentioned
The skewness ske describes the lack of symmetry of a distribution, indicating the most
likely direction of expected forecast errors. The skewness is calculated using (3.21) [87]:
N 3
1
N
e k t |t e
k
s
e
k
k 1
(3.21)
N 3
1
N
e
k 1
k t |t e
k
The kurtosis represents the “tailedness” of the distribution, indicating the propensity for
outliers in the forecast errors. The kurtosis is calculated using (3.21) [87]:
38
4
1 N
ek t|t ke
N k 1
ke 2 (3.22)
1 N 2
k t|t k
e
e
N k 1
Error margin plots will be used to shed more light on the error histograms. These plots
show what proportion of forecast errors fall above a certain threshold for a given window of
time. This will help answer the question of how often the forecasts result in unacceptable errors
The following subsection will use the above metrics and distribution-based methods to
evaluate the 116 day-ahead building-specific load forecasts. The forecasts correspond to 33408
total point predictions. Comparisons are made between the models that include building
measurements and those that do not. The forecasts are observed over several different time
3.4.5. RESULTS
Before taking a more detailed look at the forecast performance it is typical of forecasting
studies to observe the overall accuracy results. First, the day-ahead MAPE for each of the 116
Table 3.3. Quantiles of day-ahead MAPE performance across all out-of-sample forecasts
MLR NN SVM SAM
quantile No Bldg Bldg No Bldg Bldg No Bldg Bldg No Bldg Bldg
10% 3.06 2.87 3.88 3.58 2.72 2.53 3.36 3.04
25% 3.82 3.64 4.74 4.25 3.15 2.97 4.01 3.94
50% 5.03 4.52 5.65 5.36 4.18 4.08 5.34 5.29
75% 6.74 5.93 8.55 6.68 6.44 6.39 7.72 7.62
90% 8.75 7.49 11.99 9.02 9.26 9.60 11.78 10.86
39
The universal improvement when building-specific measurements are included is
apparent for all 4 forecasting methods. The only deviation is for the SVM model in the 90th
percentile of forecasted days. It is important to note that the MAPE values of the models without
building measurements are already consistently good (and in some cases excellent) by today’s
forecasting standards. Improving upon a poorly performing model by feeding it new variables
would say little about the actual impact. That the models with building information consistently
outperformed the other set in this case study demonstrates the predictive capacity of these
variables.
Other breakdowns of forecast performance are shown in Table 3.4 and Table 3.5. Table
3.4 shows the MAPE calculated on a monthly basis. Table 3.5 shows the MAPE calculated for
demonstrated for all methods. It is worth noting the improvement in July and August vs. May
and June. Although classes at Drexel University are in session year-round, the period between
40
July and September tends to have fewer people on campus. The forecast improvement during
these months could be attributed to diminished volatility of occupancy effects that are more
One of the objectives of this thesis is to study a tool for improving end-user energy
management decisions. While the results in Table 3.3 – Table 3.5 indicate solid performance
improvement, good daily or monthly performance is not sufficient to confidently inform DSM
operations and could in fact be misleading. The performance during hourly (or even shorter)
intervals must be explored to provide value for demand side management planning.
For this study the forecasts have been generated in 5 minute steps as stated previously.
The error statistics presented going forward in this chapter are aggregated as to be reported
hourly. The decision to present error statistics hourly is motivated by typical settlement period
lengths for day-ahead demand response programs [23]. This decision will also make
observations of the forecast performance a little more straightforward, but no less accurate, with
a condensed set of results. A full set of results at 5 minute resolution can be found in Appendix
A2.
Table 3.6 shows the hourly MAPE for all 4 methods across all 116 day-ahead forecasts.
When looking at the hourly performance we again see an overall measurable improvement when
building measurements are included in the models. There are also a number of additional
observations that can be made about the consistency of forecast performance from these results.
41
Table 3.6. Hourly MAPE (%) across all out-of-sample forecasts
MLR NN SVM SAM
Hour No Bldg Bldg No Bldg Bldg No Bldg Bldg No Bldg Bldg
1 5.16 5.12 8.06 6.54 5.49 5.55 7.03 6.31
2 5.37 5.33 7.85 5.62 6.86 6.81 7.51 6.62
3 8.53 8.61 8.61 7.95 6.47 6.24 10.80 9.39
4 8.37 8.56 12.50 8.08 6.72 7.02 11.52 10.16
5 8.18 8.26 10.23 6.85 6.11 6.35 10.56 9.38
6 10.01 9.61 10.38 7.73 8.66 9.18 10.01 10.31
7 6.55 6.61 6.77 6.14 5.95 5.93 7.26 6.66
8 5.27 5.31 6.43 5.23 5.09 5.17 6.05 5.76
9 5.47 5.46 6.94 5.52 6.30 6.15 6.19 6.17
10 5.25 5.19 5.96 5.40 5.65 5.55 5.94 5.86
11 5.00 4.89 5.99 5.25 5.49 5.23 5.42 5.43
12 4.93 4.84 6.37 4.89 4.87 4.84 5.62 5.36
13 4.41 4.33 5.64 4.81 4.43 4.60 5.06 4.86
14 4.43 4.42 6.24 5.12 4.08 4.01 5.05 4.70
15 4.89 4.71 5.59 4.70 3.99 3.83 5.30 5.03
16 5.06 4.86 5.77 5.06 4.20 3.93 5.41 5.13
17 4.89 4.78 5.68 5.64 4.23 3.88 5.58 4.99
18 4.56 4.53 5.06 5.11 4.00 3.35 5.24 4.61
19 4.41 4.22 4.83 6.98 4.15 3.72 5.32 4.64
20 4.68 4.48 4.83 6.63 4.13 3.87 5.75 5.03
21 4.57 4.62 5.05 4.77 3.96 3.85 5.80 4.85
22 4.98 5.30 7.62 5.18 4.41 4.48 6.51 5.27
23 5.09 5.36 8.25 5.20 4.38 4.49 6.62 5.46
24 4.98 5.10 6.27 5.52 4.70 4.94 6.58 5.78
Table 3.6 describes the day-ahead MAPE on an hourly interval. The hourly MAPE is
generally higher during the early morning hours. This is a time of lower building demand (see
Figure 3.3) so any error in the forecast will reflect more heavily on a percentage metric. During
a number of these early morning time intervals the MLR and SVM models actually perform
slightly better without the building measurements included. This is in contrast to the neural
network and additive models, and also to how all of the models generally behave over the rest of
the day.
The hourly MAPE values tend to drop as the day progresses and the building demand
increases considerably (Figure 3.3). More importantly, during this time window the models
which include building measurements improve the forecast in nearly all intervals, for all
42
methods, and by a greater amount. This is important for DSM planning purposes. This is the
time window when building demand resources will most often be dispatched and having a better
estimate of the building load is essential. Figure 3.7 - Figure 3.10 below show the hourly MAPE
for each forecasting method to visualize the results from Table 3.6.
13
W/O Bldg Meas.
12 WITH Bldg Meas
11
10
9
MAPE (%)
3
0 5 10 15 20 25
Hours
Figure 3.7. Hourly MAPE for all forecasts generated by the MLR models
43
13
W/O Bldg Meas
12 WITH Bldg Meas
11
10
9
MAPE (%)
3
0 5 10 15 20 25
Hours
Figure 3.8. Hourly MAPE for all forecasts generated by the NN models
13
W/O Bldg Meas
12 WITH Bldg Meas
11
10
9
MAPE (%)
3
0 5 10 15 20 25
Hours
Figure 3.9. Hourly MAPE for all forecasts generated by the SVM models
44
13
W/O Bldg Meas
12 WITH Bldg Meas
11
10
9
MAPE (%)
3
0 5 10 15 20 25
Hours
Figure 3.10. Hourly MAPE for all forecasts generated by the SAM models
helpful to visualize the range of MAPE results on an hourly basis. The box plots in Figure 3.11 -
Figure 3.14 look at the hourly MAPE across all out-of-sample forecasts. Each box plot displays
a notched mark at the median value with the box edges showing the 25th and 75th percentiles
and the whiskers capturing the most extreme results. These plots help identify during what
intervals of the day the models perform less consistently (in terms of MAPE values) and how this
MAPE (%)
20
15
10
5
0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
Hour
20
15
10
5
0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
Hour
Figure 3.11. Box plots of hourly MAPE results. (top) MLR without building measurements (bottom) MLR
with building measurements
20
15
10
5
0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
Hour
20
15
10
5
0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
Hour
Figure 3.12. Box plots of hourly MAPE results. (top) NN without building measurements (bottom) NN with
building measurements
46
Support Vector Machine Model without Building Measurements
35
30
25
MAPE (%)
20
15
10
5
0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
Hour
20
15
10
5
0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
Hour
Figure 3.13. Box plots of hourly MAPE results. (top) SVM without building measurements (bottom) SVM
with building measurements
20
15
10
5
0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
Hour
20
15
10
5
0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
Hour
Figure 3.14. Box plots of hourly MAPE results. (top) SAM without building measurements (bottom) SAM
with building measurements
The plots in Figure 3.11 - Figure 3.14 reinforce the results from Table 3.6. Including the
building measurements generally improves the forecast performance overall with a couple
exceptions. Again, the hourly MAPE performance is more variable in the early morning hours
47
for all models and improves later in the day. Of the four methods studied in this work, the
neural network model overall performance is the most greatly improved with building
measurements included in the model. This is especially true in terms of reducing the day-ahead
forecast variability. The SVM models are generally better when BMS measurements are
included but far less consistently compared to the other approaches. It would be interesting in
future work to try and identify what methods can be employed to optimize how building-specific
measurements can be included in forecast models. The variability in Figure 3.11 - Figure 3.14
While MAPE is the most commonly applied metric for quantifying forecast accuracy, it
is also helpful to look at other error metrics. Table 3.7 shows the hourly bias for all 4 methods
across all 116 day-ahead forecasts. As stated earlier, the ideal bias value is zero.
48
Table 3.7. Hourly bias (kW) across all out-of-sample forecasts
MLR NN SVM SAM
Hour No Bldg Bldg No Bldg Bldg No Bldg Bldg No Bldg Bldg
1 ‐2.53 ‐2.97 ‐7.15 ‐8.06 ‐5.50 ‐6.23 ‐1.49 ‐1.02
2 ‐2.67 ‐3.08 ‐7.61 0.59 ‐8.94 ‐9.06 ‐1.20 ‐1.07
3 ‐2.03 ‐1.01 0.04 4.53 0.20 ‐1.17 2.10 ‐0.35
4 ‐3.26 ‐0.66 ‐12.57 ‐5.76 ‐3.47 ‐3.98 2.59 ‐2.41
5 ‐4.85 ‐2.93 ‐7.32 ‐1.93 ‐1.89 ‐2.64 1.17 ‐2.78
6 ‐6.50 ‐6.57 ‐2.65 2.93 ‐5.03 ‐7.63 ‐2.78 ‐4.72
7 0.77 0.59 1.75 3.08 ‐1.28 ‐1.94 2.35 1.08
8 ‐0.56 ‐0.70 ‐0.30 1.26 ‐4.42 ‐4.95 0.87 ‐0.50
9 ‐1.73 ‐1.69 ‐3.85 ‐2.16 ‐5.64 ‐5.78 0.57 ‐0.81
10 ‐0.78 ‐0.64 ‐0.57 2.27 ‐4.22 ‐4.73 1.80 0.98
11 ‐2.10 ‐2.25 ‐0.81 ‐0.19 ‐5.42 ‐5.35 0.56 0.28
12 1.08 0.74 ‐6.88 2.42 ‐3.68 ‐4.34 3.48 3.01
13 0.91 0.11 ‐1.67 3.42 ‐3.63 ‐5.39 2.70 2.51
14 1.81 0.67 ‐3.39 3.84 ‐2.75 ‐3.83 3.69 3.27
15 3.40 1.67 0.46 3.10 0.38 ‐1.50 4.80 5.17
16 4.24 2.15 ‐2.79 6.63 0.98 ‐0.71 5.45 5.98
17 3.70 1.32 ‐3.33 8.51 2.06 0.47 4.63 5.26
18 2.76 0.23 ‐3.50 6.68 3.26 0.53 3.21 3.63
19 4.10 1.74 ‐2.06 12.39 5.20 3.29 3.79 4.77
20 3.65 1.79 ‐1.41 10.41 3.82 2.67 3.06 4.31
21 0.76 ‐1.00 ‐4.29 3.58 ‐0.39 ‐1.00 ‐0.14 1.29
22 ‐2.17 ‐4.15 ‐12.89 3.17 ‐3.07 ‐3.85 ‐3.57 ‐1.55
23 ‐3.89 ‐4.91 ‐13.94 ‐2.91 ‐2.25 ‐3.35 ‐4.35 ‐2.99
24 ‐3.71 ‐4.46 ‐9.54 2.83 ‐3.00 ‐4.76 ‐3.82 ‐2.46
Figure 3.15 - Figure 3.18 below show the hourly bias for each forecasting method to
10
Bias (kW) 5
-5
-10
-15
0 5 10 15 20 25
Hours
Figure 3.15. Hourly Bias for all forecasts generated by the MLR models
15
W/O Bldg Meas
WITH Bldg Meas
10
5
Bias (kW)
-5
-10
-15
0 5 10 15 20 25
Hours
Figure 3.16. Hourly Bias for all forecasts generated by the NN models
50
15
W/O Bldg Meas
WITH Bldg Meas
10
Bias (kW) 5
-5
-10
-15
0 5 10 15 20 25
Hours
Figure 3.17. Hourly Bias for all forecasts generated by the SVM models
15
W/O Bldg Meas
WITH Bldg Meas
10
5
Bias (kW)
-5
-10
-15
0 5 10 15 20 25
Hours
Figure 3.18. Hourly Bias for all forecasts generated by the SAM models
These results do not provide any conclusive evidence that building measurements affect
the bias of the forecasts. In fact, the shape of the bias as it evolves over time is basically the
51
same whether building measurements are included or not. The only exception to this is a few
intervals for the neural network models. This implies that the bias is a function of the model
building process more so than a function of the input variables. It is also interesting that the
jump in bias for the NN model without building measurements seen between 18:00 and 21:00 in
Figure 3.16 corresponds to the diminished accuracy during the same time period seen in the
The next metric presented in this section is the standard deviation of the forecast errors
(SDE). This metric is very important for understanding the consistency of the forecast
performance. Table 3.8 shows the hourly SDE for all 4 methods across all 116 day-ahead
forecasts.
Figure 3.19 - Figure 3.22 below show the hourly SDE for each forecasting method to
17
W/O Bldg Meas
WITH Bldg Meas
16
15
Standard Deviation (kW)
14
13
12
11
10
9
0 5 10 15 20 25
Hours
Figure 3.19. Hourly standard deviation of the forecast error for all forecasts generated by the MLR models
53
17
W/O Bldg Meas
WITH Bldg Meas
16
15
13
12
11
10
9
0 5 10 15 20 25
Hours
Figure 3.20. Hourly standard deviation of the forecast error for all forecasts generated by the NN models
17
W/O Bldg Meas
WITH Bldg Meas
16
15
Standard Deviation (kW)
14
13
12
11
10
9
0 5 10 15 20 25
Hours
Figure 3.21. Hourly standard deviation of the forecast error for all forecasts generated by the SVM models
54
17
W/O Bldg Meas
WITH Bldg Meas
16
15
13
12
11
10
9
0 5 10 15 20 25
Hours
Figure 3.22. Hourly standard deviation of the forecast error for all forecasts generated by the SAM models
Unlike with the bias, the building measurements have a definite overall positive impact
on the SDE. This improvement is most pronounced in the NN models where there is a
significant decrease over the majority of the day. The SDE for the SVM and SAM models also
decreases but more substantially later in the day than during the morning hours. Only the MLR
The next step is to analyze the shape of the error distributions. As an example, the
following two histograms show the forecast errors during the 09:00-10:00 time window for the
neural network models. The model in Figure 3.23 does not include building measurements while
120
100
Number of Events
80
60
40
20
0
-80 -60 -40 -20 0 20 40 60 80
Error (kW)
Figure 3.23. Empirical distribution of forecast errors for the NN model without building measurements
included (09:00-10:00)
120
100
Number of Events
80
60
40
20
0
-80 -60 -40 -20 0 20 40 60 80
Error (kW)
Figure 3.24. Empirical distribution of forecast errors for the NN model with building measurements
included (09:00-10:00)
56
The overall change in shape of the distribution is noticeable: the tail ends are `reduced
in both directions in Figure 3.24, as are a number of the peaks. All hourly plots of the empirical
distributions for each forecast model are located in Appendix A2. It is not necessary to show all
the distributions here as the characteristics of these distributions will be measured in several
ways below.
First, the error margin plots corresponding to two error thresholds are shown in Figure
3.25 - Figure 3.26. The definition of “unacceptable error” is subjective and must be determined
on a case by case basis. This study will look at the 10kW and 30kW thresholds. These plots are
80
W/O Bldg Meas
WITH Bldg Meas
75
70
65
60
Frequency (%)
55
50
45
40
35
30
0 5 10 15 20 25
Hours
Figure 3.25. Frequency of forecast errors within a ±10 kW error margin for the MLR models
57
100
W/O Bldg Meas
WITH Bldg Meas
98
Frequency (%) 96
94
92
90
88
0 5 10 15 20 25
Hours
Figure 3.26. Frequency of forecast errors within a ±30 kW error margin for the MLR models
80
W/O Bldg Meas
WITH Bldg Meas
75
70
65
60
Frequency (%)
55
50
45
40
35
30
0 5 10 15 20 25
Hours
Figure 3.27. Frequency of forecast errors within a ±10 kW error margin for the NN models
58
100
W/O Bldg Meas
WITH Bldg Meas
98
Frequency (%) 96
94
92
90
88
0 5 10 15 20 25
Hours
Figure 3.28. Frequency of forecast errors within a ±30 kW error margin for the NN models
80
W/O Bldg Meas
WITH Bldg Meas
75
70
65
60
Frequency (%)
55
50
45
40
35
30
0 5 10 15 20 25
Hours
Figure 3.29. Frequency of forecast errors within a ±10 kW error margin for the SVM models
59
100
W/O Bldg Meas
WITH Bldg Meas
98
Frequency (%) 96
94
92
90
88
0 5 10 15 20 25
Hours
Figure 3.30. Frequency of forecast errors within a ±30 kW error margin for the SVM models
80
W/O Bldg Meas
WITH Bldg Meas
75
70
65
60
Frequency (%)
55
50
45
40
35
30
0 5 10 15 20 25
Hours
Figure 3.31. Frequency of forecast errors within a ±10 kW error margin for the SAM models
60
100
W/O Bldg Meas
WITH Bldg Meas
98
Frequency (%) 96
94
92
90
88
0 5 10 15 20 25
Hours
Figure 3.32. Frequency of forecast errors within a ±30 kW error margin for the SAM models
The error margin plots reflect similar performance as in the previous tests. The MLR
models perform nearly identically regardless of the inclusion of building measurements. The
frequency at both the 10kW and 30kW thresholds. One observation of note is a sharp decrease
in frequency at the 10kW threshold in Figure 3.27, corresponding to the jump in bias for the NN
model with building measurements seen previously between 18:00 and 21:00. Again the SVM
and SAM models also perform better with building measurements but more substantially later in
The next metric presented in this section is the skewness of the forecast error
distributions. Table 3.9 shows the hourly skewness for all 4 methods across all 116 day-ahead
forecasts.
61
Table 3.9. Hourly skewness across all out-of-sample forecasts
MLR NN SVM SAM
Hour No Bldg Bldg No Bldg Bldg No Bldg Bldg No Bldg Bldg
1 -0.84 -0.74 -0.97 -0.84 -0.75 -0.90 0.41 -0.41
2 -0.95 -0.89 -0.99 -0.92 -0.42 -0.64 0.53 -0.20
3 0.00 -0.25 -0.39 -0.04 -0.01 -0.26 -0.14 -0.39
4 -0.13 -0.05 -1.19 -0.22 -0.69 -0.75 -0.02 -0.60
5 0.03 0.30 -1.02 0.40 0.10 -0.10 0.39 -0.59
6 -0.11 -0.12 -0.16 -0.46 0.61 0.62 0.26 -0.46
7 -0.55 -0.54 -0.10 0.11 -0.20 -0.25 0.06 -0.44
8 -0.38 -0.33 0.30 -0.03 -0.02 -0.12 0.30 -0.52
9 0.26 0.31 -0.32 0.27 -0.08 -0.22 0.49 -0.11
10 0.15 0.20 0.51 0.06 -0.02 -0.01 0.17 -0.32
11 -0.43 -0.41 -0.57 -0.70 -1.18 -1.22 -0.10 -0.84
12 0.51 0.59 0.33 -0.18 -0.14 -0.04 1.23 -0.18
13 -0.23 -0.15 -0.22 -0.53 -0.71 -0.61 0.31 -0.32
14 -0.12 -0.03 -0.02 -0.47 -0.47 -0.55 0.33 -0.09
15 0.03 0.15 -0.13 -0.22 -0.34 -0.59 0.50 0.24
16 -0.07 0.10 0.01 -0.38 -0.71 -0.65 0.47 0.13
17 0.04 0.15 -0.14 -0.45 -0.47 -0.52 0.49 0.23
18 -0.03 0.16 -0.20 -0.29 0.04 -0.26 0.56 0.08
19 -0.08 0.02 0.06 -0.72 0.12 -0.11 0.27 -0.14
20 -0.91 -0.78 -0.84 -1.65 -0.74 -1.11 -0.36 -0.75
21 -0.66 -0.54 -0.85 -0.99 -0.70 -0.99 -0.21 -0.54
22 -0.50 -0.35 -0.51 -1.00 -0.40 -0.68 0.02 -0.37
23 -0.46 -0.29 -0.19 -1.20 -0.35 -0.53 0.11 -0.24
24 -0.66 -0.54 -0.89 -1.01 -0.12 -0.37 0.06 -0.10
Figure 3.33 - Figure 3.36 below show the hourly skew for each forecasting method to
0.5
0
Skew
-0.5
-1
-1.5
-2
0 5 10 15 20 25
Hours
Figure 3.33. Hourly skew of the forecast error for all forecasts generated by the MLR models
1.5
W/O Bldg Meas
WITH Bldg Meas
1
0.5
0
Skew
-0.5
-1
-1.5
-2
0 5 10 15 20 25
Hours
Figure 3.34. Hourly skew of the forecast error for all forecasts generated by the NN models
63
1.5
W/O Bldg Meas
WITH Bldg Meas
1
0.5
0
Skew
-0.5
-1
-1.5
-2
0 5 10 15 20 25
Hours
Figure 3.35. Hourly skew of the forecast error for all forecasts generated by the SVM models
1.5
W/O Bldg Meas
WITH Bldg Meas
1
0.5
0
Skew
-0.5
-1
-1.5
-2
0 5 10 15 20 25
Hours
Figure 3.36. Hourly skew of the forecast error for all forecasts generated by the SAM models
Much like the bias measurement, there is little that can be drawn from the skewness of
the error distributions. The evolution of skewness over the course of the day is roughly the same
whether or not the forecast models included building measurements. One interesting observation
64
from Figure 3.36 is that the SAM model without building measurements tends to overestimate
the load while the SAM model with building measurements tends to underestimate the load, both
The last metric presented in this section is the kurtosis of the forecast error distributions.
If the errors are normally distributed then the kurtosis is equal to 3. Table 3.10 shows the hourly
Figure 3.37 - Figure 3.40 below show the hourly skew for each forecasting method to
7
Kurtosis
2
0 5 10 15 20 25
Hours
Figure 3.37. Hourly kurtosis of the forecast error for all forecasts generated by the MLR models
10
W/O Bldg Meas
WITH Bldg Meas
9
7
Kurtosis
2
0 5 10 15 20 25
Hours
Figure 3.38. Hourly kurtosis of the forecast error for all forecasts generated by the NN models
66
10
W/O Bldg Meas
WITH Bldg Meas
9
7
Kurtosis
2
0 5 10 15 20 25
Hour
Figure 3.39. Hourly kurtosis of the forecast error for all forecasts generated by the SVM models
10
W/O Bldg Meas
WITH Bldg Meas
9
7
Kurtosis
2
0 5 10 15 20 25
Hours
Figure 3.40. Hourly kurtosis of the forecast error for all forecasts generated by the SAM models
67
For the MLR and SVM models the kurtosis of the error distributions is nearly identical
throughout the day whether models include building measurements or not. The only method
where building measurements tend to make the errors more “normal” is the SAM method.
3.4.6. REMARKS
There are a number of observations that can be made from the analysis above. Most
prominently, there is a clear trend in forecast performance throughout the day. Early in the day
the forecast error is higher and more variable. Heading into the afternoon and evening hours the
forecasts become more accurate and with a greater frequency. This is true in general for all
methods and all models. The impact of building measurements to both the accuracy and
consistency also follows this performance trend. Improvements are lower during the morning
hours and more pronounced later on. Knowing the operational patterns of the building can make
For energy efficiency reasons, large commercial HVAC systems operate under multiple
control schemes [88]. The Hagerty Library operates with a “nighttime setback” where the
building setpoint is reset (e.g. – raised) from 02:00-06:00 which is considered an “unoccupied”
time period. Referring back to Figure 3.3, this time window also corresponds to the low point in
the daily load profile. This low level of demand approaches the base load of the building which
is almost completely thermally insensitive [89]. It stands to reason that including BMS
measurements tied directly to the HVAC system will have less of an impact on the forecasts
during this period. Conversely, later in the day the building HVAC equipment is required to
maintain the building internal temperature in the face of much higher ambient temperature and
increased internal heat gains related to occupancy. This increased cooling load corresponds to
specific forecasting applications, several interesting observations can be made about the
performance of each model. The characteristics of the MLR forecasts are impacted less heavily
by the inclusion of building measurements than the other methods. It is however the one of the
most accurate and consistent models overall. In contrast, the NN forecasts when the models
include building measurements are the most greatly improved by nearly every metric observed.
The SAM models also see a similar level of improvement. It makes sense that the only linear
model of those tested (MLR) would see the smallest impact from the addition of these variables.
The good overall MLR model forecast performance is a result of well-developed models. The
nonlinear models are better able to capture the underlying relationship between demand and the
building measurements but they are more complicated to build. It will be an area of future
work to fine tune the model building approach to try and determine if there is an ideal model for
building specific applications. The examination of several models in this thesis is an important
INTERVALS
4.1. OVERVIEW
The previous chapter focused on the generation of point forecasts. No matter how
accurate the selected method is, point forecasts will always be incorrect to a certain degree. The
estimation of forecasting distributions, and the prediction intervals that are obtained from them,
can provide information that is more relevant to risk assessment for the end-user than simple
point forecasts. This Chapter will apply a nonparametric method of estimating prediction
intervals for building-specific load forecasting. The same case study from Chapter 3 will provide
the basis for assessing the impact of building measurements on the resulting prediction intervals.
Traditionally, load forecasts are presented in the form of point forecasts. This form has
the advantage of being very easy to interpret: each number corresponds directly to an estimate of
the load at a future time. The majority of the forecasting literature focuses on refining the
models and methods that generate these point forecasts. However, one shortcoming of point
forecasts is the lack of any information about the forecast uncertainty. The error analysis
performed in Section 3.4.5 can describe the historical performance of a given method but cannot
fully describe the uncertainty related to a given prediction. Conservative demand side planning
activities require some measure of understanding of the forecast uncertainty. This uncertainty
can be expressed in the form of a prediction interval, also referred to as an interval forecast. The
prediction interval is a range in which a future observation is expected to lie with a pre-assigned
probability [90].
70
It bears mentioning here that there is some confusion surrounding the use of
confidence intervals versus prediction intervals. While these quantities are often used
interchangeably in the literature, they do not describe the same thing. A confidence interval (CI)
interval (PI) is an interval associated with future observations of a random variable, where said
observations lie within the interval with a specified probability. Figure 4.1 below shows an
example of the difference between a 95% CI and 95% PI for a simple regression example
Figure 4.1. Example of the difference between a 95% confidence interval (CI) and 95% prediction interval
(PI)
The confidence interval tells us with a 95% probability that the fitted line lies within the
interval. The prediction interval tells us with a 95% probability that the data points will be
within the given interval. Figure 4.1 shows the prediction interval encloses the confidence
interval and it was proven in [92] this is necessarily the case. This thesis is interested in the
accuracy of building-specific forecasts in relation to the actual metered load. Therefore this
71
thesis will look at prediction intervals rather than confidence intervals to assess the impact of
Although traditionally the forecasting community has focused on point forecasts, there
has been a more recent push towards methods that can provide valuable formation about the
forecast uncertainty. Uncertain forecasts are a necessary input to a number of power systems
analyses such as unit commitment [93] [94] and reliability planning [95] [96] [97]. This type of
forecasting research falls into the area broadly classified as probabilistic forecasting. In [98], a
very recent and thorough review of probabilistic load forecasting (PLF) is presented. Related
review papers have been written on probabilistic wind forecasting [99] and probabilistic
electricity price forecasting [100]. Similar concepts are shared among all three applications.
As defined in [98] there are a number of techniques employed in the PLF problem. The
PLF process addresses the stochastic nature of the problem by focusing on one of three parts: the
input, the model, or the output. Approaches that focus on the input are concerned with scenario
generation with simulated predictors. These types of works are most often concerned with long-
term forecasting (LTF). Modeling techniques for producing PLFs either extend existing point
probabilistic forecasts. The last category includes PLF methods that post-process the outputs of
point forecasts. This is done either by producing density functions of the point forecast residuals
This thesis will employ a modified version of the method from [75] where forecasting
residuals are simulated via a modified bootstrap method in order to produce density functions of
the electrical demand. The details of this procedure are laid out in the following section. There
are several reasons why this method is chosen. By focusing on the out-of-sample test residuals
this method takes advantage of the existing accurate point forecasting methods developed in the
72
previous chapter. Additionally this method assumes no underlying distribution of the forecast
errors. Typical parametric methods assume an underlying, often Gaussian, distribution. It has
been shown in [101] [75] and many other works that this assumption is usually incorrect. The
for methods that rely on forecasts of predictor variables, the exact distribution of the load
forecast depends on the distribution of the predictor variables. This makes such distributions
used to construct empirical forecasting distributions and the desired prediction intervals can be
obtained from these distributions. The next section will describe what defines a prediction
interval applied to the load forecasting problem and the details of the chosen method.
The formal definition of a forecast prediction interval estimated at time t looking k steps
P pt k ˆIt k |t P pt k Lˆt k |t ,Uˆ t k|t 1 (4.1)
where
A prediction interval is defined by its lower and upper bounds. The probability (1-α)
within which the observed value is expected to lie is called the nominal confidence rate [90].
73
This term is preferred to confidence level which is often used in the literature but may lead to
confusion with the confidence interval. The lower and upper bounds are defined in (4.2) - (4.3)
below.
Lˆt k |t qˆtk/|t2 , P pt k Lˆt k |t / 2 (4.2)
Uˆ tk|t qˆt1k|t / 2 , P pt k Uˆ tk|t 1 / 2 (4.3)
The boundaries correspond to the quantiles qˆt k|t of the estimated prediction distribution
Fˆt pk|t . This defines a central prediction interval centered on the median of Fˆt pk|t . If a Gaussian
distribution is assumed for Fˆt pk|t then a central prediction interval is sufficient since the mean and
median are the same, and thus the point forecast will lie within the prediction interval. However,
for an asymmetric distribution the mean and median can differ significantly. For small nominal
confidence rates the point forecasts may not even lie within the prediction interval. Therefore,
when the distribution is expected to be non-Gaussian, prediction intervals can be constructed that
P pt k Lˆt k |t , pˆ t k |t P pt k pˆ t k |t ,Uˆ t k|t 1 / 2
(4.4)
step ahead forecast errors denoted St,k. It is assumed that these error samples are representative
of the true predictive distribution Ft pk |t . For a well-developed point forecasting method this is a
reasonable assumption. Thus there is an implicit assumption that the future uncertainty can be
expressed based on the recent performance of the point prediction method. To find an estimate
Fˆt pk|t of the distribution, the cumulative distribution function is introduced in (4.5).
74
1
Gˆ t,k # i St , k | i (4.5)
n
where the cumulative distribution function gives the fraction of collected errors εi less
Recall from (4.2) and (4.3) that the prediction interval bounds with nominal coverage rate
(1-α) correspond to the quantiles of the estimated prediction distribution Fˆt pk|t . Thus the final
form of the prediction interval upper and lower bounds are defined in (4.6) and (4.7) below.
where Gˆ t,k 1 are the quantile functions. Effectively, the prediction interval bounds are
calculated by “dressing” the point predictions with estimates of the prediction distribution.
In order to construct the prediction intervals as defined in (4.6) and (4.7), this thesis will
generally employ the modified bootstrap method from [75] as stated previously. The main idea
distributions from which prediction intervals will be obtained. The steps for this process are
In practice, a rolling window of forecasts and corresponding measured load values would
be collected. For this analysis the out-of-sample load forecasts that are described in section 3.4.3
are used. Of these 116 day-ahead forecasts, the first 3 months (May-July) are collected to
estimate the prediction intervals. The August forecasts are held out to evaluate the prediction
intervals. Data from the Chapter 3 training set (2011 and 2012 data) cannot be used in order to
The residuals are calculated as per equation (3.16). This forms the collection of residuals
The standard method for bootstrapping time series data is the block bootstrap [102]. This
process involves taking random segments of the historical time series and rearranging them to
form new artificial series. The length of each random segment (or block) must be long enough to
capture any serial correlations in the data but short enough to allow for a large number of
possible simulated series. At this step of the process this thesis will deviate slightly from the
procedure used in [75]. Rather than use a single block bootstrap, a double seasonal block
bootstrap as used. Seasonality in this case refers to periodic correlations in the data, such as
annual or daily patterns. This method was introduced in [102] to address the double seasonality
(annual and daily) contained in historical temperature observations. Building loads express
similar double seasonality, both daily and hourly, which makes this approach more appropriate.
The difference in how the residual time series are segmented is highlighted in Figure 4.2. The
diagram (a) shows a single seasonal block segmentation and (b) the double seasonal block
segmentation.
76
Figure 4.2. Seasonal block segmentation (a) and double seasonal block segmentation (b)
The original residual time series is double seasonal bootstrapped to obtain simulated
forecast errors. The rearranged time series are shown in Figure 4.3. The diagram (a) shows an
example of a single seasonal block bootstrap and (b) the double seasonal block bootstrap. Note
for this work only (b) is implemented. A total of 500 bootstrap replications are made.
Figure 4.3. Block bootstrapped time series (a) and double seasonal block bootstrapped time series (b)
77
Using the accumulated simulated forecast errors, empirical forecasting distributions can
be constructed. Distributions for each hour of the day are estimated via kernel density
estimation. Using the definition of the lower and upper bounds of the prediction interval from
(4.6) and (4.7), the prediction interval can be determined from the constructed distributions. The
nominal coverage rate 1- α is varied in order to evaluate the forecasts over a range of coverages.
78
It is simple to evaluate a point forecast: how large was the deviation between the
estimated value and the true value defines whether or not a forecast is acceptable. Evaluating
probabilistic forecasts is far more difficult. Research into establishing methods for evaluating
probabilistic load forecasts is almost nonexistent [98]. There are a few commonly used attributes
for probabilistic wind forecasting evaluation, which have been borrowed from the meteorological
forecast world, that are of interest to the probabilistic load forecasting problem and will be
Three main attributes are used to characterize the prediction intervals developed using the
procedure laid out in the previous section. These attributes are called reliability, sharpness, and
resolution. Reliability refers to how close the predicted distribution is to the actual. In other
words it is a comparison of the empirical coverage to the nominal coverage rate 1-α. To check
the reliability of a given prediction interval, it is first necessary to introduce an indicator variable
If the true power value pt+k lies in the prediction interval for a given α, then the indicator
value equals 1. Otherwise it is zero. The collection of “hits” and “misses” over the interval of
interest are used to assess the reliability of the estimated prediction intervals.
NT
nk ,1 t,k
(4.9)
t 1
nk ,0 :
Number of "misses" over the interval of interest
nk,1 : Number of "hits" over the interval of interest
NT : Number of realizations over the interval of interest
Given the above expressions, the reliability of a prediction interval is obtained using
(4.11).
1 NT
nk,1
aˆk
NT
t,k
t 1 nk,0 nk,1
(4.11)
Sharpness and resolution are concerned with describing the shape of the forecast
distributions. Sharpness refers to how tightly the estimated distribution covers the actual one.
Narrower distributions are preferable from a decision making context (provided they still
maintain the desirable reliability). For example, if the lower and upper bounds of the prediction
interval are very close to the predicted 99% interval then the interval is said to be very sharp.
NT
1
t ,k
NT
t 1
t ,k (4.13)
As a complement to the sharpness (the mean prediction interval size) the variation of the
interval size over time (e.g. – resolution) can be expressed as the standard deviation of the
interval size for a future time k and a given nominal coverage 1-α.
81
1 N T
2
k t ,k k
NT 1 t 1
(4.14)
beneficial to have a single numerical value that can be calculated that includes all relevant
information about the probabilistic forecast. Using an appropriate skill score provides a
comprehensive measure for comparing rival forecasting approaches. As noted previously, this
thesis is focused on comparing forecast performance with and without building measurements
included in the models so such a metric will be a helpful tool. In addition to providing a
convenient measure for comparison, a skill score presented in tandem with observations of the
reliability allows for the assessment of both the sharpness and resolution without having to
calculate these values separately [90]. A separate analysis of the reliability is required to divorce
its influence on the skill score from that of the sharpness and resolution.
extremely limited in the load forecasting literature. The pinball loss function was used as a
criteria in the GEFCom 2014 competition [104]. This skill score is formulated specifically to
assess quantile regression forecasts. More appropriate for this work is the Winkler score [105]
shown in (4.15), which has been used for evaluating wind power forecast prediction intervals in
[90] and for probabilistic load forecasting in a slightly modified form in [106]. It can be seen in
(4.15) that the Winkler score will reward narrow prediction intervals while penalizing
2 t,k 4 Lˆt k |t pt k , if pt k Lˆt k |t
Sct ,k 2 t,k , if pt k Iˆtk|t (4.15)
ˆ ˆ
2 t ,k 4 pt k U t k |t , if pt k U t k |t
following approach. First, the reliability will be calculated. Results will be presented as
reliability diagrams. These plots will show the deviations from the nominal coverage as a
function of the nominal coverage, indicating whether the given method over or underestimates
the uncertainty. In addition, the skill score defined in (4.15) will be calculated and presented
hourly for a range of nominal coverage rates. This analysis will again focus on evaluating the
impact to the forecast when building measurements are included in the forecast models.
Between the reliability and skill score metrics a comprehensive comparison of the forecast
4.5. RESULTS
The reliability results for the August 2014 test data is presented first. To demonstrate the
impact of building measurements on reliability, reliability diagrams are presented in Figure 4.5 -
Figure 4.8.
83
100
90
80
60
50
40
30
20
10
0
0 10 20 30 40 50 60 70 80 90 100
Required Probability (%)
100
90
80
W/O Bldg Meas
WITH Bldg Meas
70
Empirical Coverage (%)
60
50
40
30
20
10
0
0 10 20 30 40 50 60 70 80 90 100
Required Probability (%)
90
80
W/O Bldg Meas
70 WITH Bldg Meas
Empirical Coverage (%)
60
50
40
30
20
10
0
0 10 20 30 40 50 60 70 80 90 100
Required Probability (%)
100
90
80
W/O Bldg Meas
WITH Bldg Meas
70
Empirical Coverage (%)
60
50
40
30
20
10
0
0 10 20 30 40 50 60 70 80 90 100
Required Probability (%)
For the above reliability diagrams, the x-axis gives the required probability (i.e. the
nominal coverage rate of the prediction intervals). The y-axis shows the empirical coverage
calculated from (4.11). The deviation from the dashed blue line represents the deviation from
85
the ‘perfect reliability’ situation for which the empirical coverage would equal the nominal
one. For example a 1% deviation for a required 50% coverage rate means the empirical
coverage rate is 51%. These plots indicate a systematic overestimation of the forecast
uncertainty across all methods. There is also a general trend where the deviation is greater for
mid-range coverage rates. Inclusion of building-specific variables in the models does not stray
too much from these observed trends. However, for all but the MLR models, the inclusion of
building-specific measurements drives the reliability closer to the nominal coverage. This effect
is most pronounced for the NN and SAM models. Table 4.1 below quantifies the deviation in %
Table 4.1. % Deviation in empirical coverage compared to the nominal coverage rate
MLR NN SVM SAM
Coverage No Bldg Bldg No Bldg Bldg No Bldg Bldg No Bldg Bldg
5% 3.92 4.06 2.72 3.12 4.33 4.19 4.06 4.06
10% 7.31 7.31 5.43 4.76 7.31 7.31 6.37 7.04
15% 9.89 10.03 5.59 5.05 8.68 8.68 8.68 9.49
20% 11.26 11.26 6.56 5.35 9.25 8.98 10.86 10.73
25% 12.50 12.50 7.53 5.24 11.16 9.95 13.44 11.96
30% 12.80 12.66 7.02 5.27 12.12 11.45 15.35 11.85
35% 12.96 13.49 8.25 5.43 11.88 10.94 15.78 12.42
40% 14.60 14.19 7.34 5.46 11.77 11.64 15.13 12.04
45% 14.76 15.03 6.69 5.48 11.26 12.20 16.51 12.07
50% 13.71 15.05 5.24 5.78 11.29 10.75 17.47 12.37
55% 14.01 14.27 4.60 4.06 9.30 9.84 19.11 12.12
60% 11.75 13.76 5.56 2.88 9.33 5.83 17.39 9.19
65% 11.64 12.04 5.86 2.37 8.28 5.59 15.00 8.28
70% 10.05 12.07 5.62 2.53 5.48 6.29 14.35 8.17
75% 7.80 9.95 3.23 2.69 5.24 4.03 11.42 6.72
80% 5.94 7.82 2.45 2.18 3.52 4.06 9.70 5.13
85% 4.49 6.24 2.47 1.26 1.40 3.15 6.91 4.35
90% 3.44 5.19 1.02 0.62 1.69 1.02 5.59 2.77
95% 1.32 2.26 0.65 0.38 0.78 0.24 3.20 1.45
The average hourly skill score for the August test data is calculated for each nominal
coverage rate. A total of 19 coverage rates are considered (5% to 95% in 5% increments). A full
86
set of the results is located in Appendix A2. A subset is shown here in Figure 4.9 - Figure
4.16. Each forecast method is shown for nominal coverage rates of 25% and 75%.
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
0 5 10 15 20 25
Hours
Figure 4.9. Average hourly skill score for the MLR models. Nominal coverage rate = 25%
-5
W/O Bldg Meas
WITH Bldg Meas
-10
-15
-20
Skill Score
-25
-30
-35
-40
0 5 10 15 20 25
Hours
Figure 4.10. Average hourly skill score for the MLR models. Nominal coverage rate = 75%
87
0
-10
-20
-30
Skill Score
-40
-50
-60
-80
0 5 10 15 20 25
Hours
Figure 4.11. Average hourly skill score for the NN models. Nominal coverage rate = 25%
-5
-10
-15
-20
Skill Score
-25
-30
-35
W/O Bldg Meas
WITH Bldg Meas
-40
0 5 10 15 20 25
Hours
Figure 4.12. Average hourly skill score for the NN models. Nominal coverage rate = 75%
88
0
-10
-20
-30
Skill Score
-40
-50
-60
-80
0 5 10 15 20 25
Hours
Figure 4.13. Average hourly skill score for the SVM models. Nominal coverage rate = 25%
-5
-10
-15
-20
Skill Score
-25
-30
-35
W/O Bldg Meas
WITH Bldg Meas
-40
0 5 10 15 20 25
Hours
Figure 4.14. Average hourly skill score for the SVM models. Nominal coverage rate = 75%
89
0
-20
-30
Skill Score
-40
-50
-60
-70
-80
0 5 10 15 20 25
Hours
Figure 4.15. Average hourly skill score for the SAM models. Nominal coverage rate = 25%
-5
-15
-20
Skill Score
-25
-30
-35
-40
0 5 10 15 20 25
Hours
Figure 4.16. Average hourly skill score for the SAM models. Nominal coverage rate = 75%
The closer to zero the skill score, the better the performance. Figure 4.9 and Figure 4.10
demonstrate that the skill of the MLR model forecasts is not sensitive to inclusion of building-
specific measurements. The other three methods however are measurably improved. One
90
notable characteristic common between the point forecast accuracy results and the skill scores
above is the drop in average performance during the morning hours compared to later in the day.
4.6. REMARKS
recent point forecast errors in order to assess the forecast uncertainty. Applying the same
method for generating probabilistic forecasts to a set of different point forecasting methods
allows for a comprehensive statistical comparison of the resulting prediction intervals. There are
a number of observations that can be made from this analysis. First, the reliability diagrams
underestimating as well) the uncertainty is not a bad thing in and of itself. That determination
must be made by the forecaster and is subjective. However it is necessary to characterize the
evaluating reliability based on statistical hypothesis testing has been considered in [103], [107].
These tests rely on the assumption of independence in the forecasting residuals. It is widely
understood in load forecasting (as well as wind and electricity price forecasting) that this
assumption generally does not hold. Therefore the reliability as presented in this thesis provides
The reliability of a probabilistic forecast is similar to the bias metric for point forecasts.
It is interesting then to compare these reliability results to the bias results presented in Section
3.4.5 for point forecasts. The bias of the point forecasts followed no particular trend across each
of the four methods. This is in contrast to the reliability plots which show a distinct trend that is
common across each of the four methods. Notably for both the bias and reliability results the
inclusion of building specific measurements had very little impact. It is reasonable to conclude
91
then that the point forecasting method drives the error bias and the method of prediction
Unlike with the reliability results, the skill score is measurably improved when building
measurements are included for all methods except the MLR forecasts. Similar to the conclusions
made in Section 3.4.5 for point forecasts, the skill of the probabilistic forecasts is more positively
impacted for the nonlinear models. As stated in Section 4.4, the combination of the reliability
and skill score metrics fully describes a comprehensive measure of the uncertain forecast
performance. The reliability has been shown to be driven by the method employed to construct
the probabilistic forecast. Thus the improvement in skill observed for the models including
building measurements can be attributed to these variables. This improvement in skill reflects
tighter, more consistently sized prediction intervals (i.e. – improved sharpness and resolution).
Forecasts that produce narrower intervals are more valuable in a decision making context. This
is important where building-specific load forecasts are used to assess demand side management
capabilities.
It stands to reason that the quality of the method for constructing prediction intervals is
tied to the quality of the related point forecasts themselves. The sharper the error distributions,
the higher the quality of the resulting probabilistic forecasts. It is then an area of future interest
to develop direct probabilistic forecasts of building demand, removing the reliance on point
forecasts in order to possibly improve the quality of the probabilistic forecasts. For this thesis
however it is an important first contribution to the area of building-specific forecasting that the
5.1. OVERVIEW
In this chapter a dynamic load model for the controllable load of a building is developed.
In this work the controllable load is considered to be the HVAC chiller. The coupling of a
building’s electrical and thermal characteristics is not adequately captured using existing static
models. Therefore a dynamic exponential load model is proposed. This allows for more
accurate planning of the dispatching of load for DSM. The model in this thesis is derived from
Traditionally in power systems, loads are represented in aggregate. Loads are grouped by
bus, or substation, and modeled as one complex power injection (for a given bus) of the form
shown in (5.1).
While this structure is convenient in that it reduces model complexity when performing
power system analysis, studies over the last 20 years have shown such static models are not
appropriate for studies where load dynamics have critical impacts [108] [109]. In terms of DSM
applications, the model in (5.1) neglects dynamic behavior and lacks the resolution required to
analyze and optimize control of individual components. An alternative approach is one where a
direct mathematical model is developed based on measurements. The advantages of this model
will be two-fold:
93
The increased granularity will provide more opportunities for customers to
The dynamic response of HVAC loads due to the natural coupling of the electric
demand and the thermal response of the building can be clearly understood.
At the power system level, a measurement-based approach is used in [110] [111] [112]
[113] to study the impact of voltage changes on real power and reactive power respectively.
Measured data is fit to an assumed model of the dynamic behavior. The model used in these
works is a dynamic exponential load model. In this thesis the same model will be used to
capture the thermal electrical behavior of an HVAC chiller. As noted in [110] [111] this model
was proposed as a means of capturing the unique effects of large electrical heating loads on the
If the end purpose is to use this model for DSM, characterizing the electrical load response
to a control action is essential. In the case of an HVAC chiller, raising or lowering the outlet
chilled water temperature is a straightforward action a facilities manager can take to increase or
decrease the electric power drawn by the machine. Therefore, in the following sections, a
mathematical model for the HVAC chiller electrical load response to a change in outlet water
Data was collected during tests performed in the fall of 2009 for a single HVAC chiller
on Drexel University’s campus whose cooling load was carried on a single chiller. For this
particular chiller, the nominal outlet water temperature is 44oF. At approximately 09:00, the
temperature setpoint was raised to 51oF and around 15:00 it was lowered back to 44oF. These
setpoint changes were the only change made to the HVAC system. All other system operations
were allowed to operate as normal (i.e. fan speeds, damper positions, etc.). While this may not
94
be the most effective action in terms of a DSM activity since the overall power draw of the
system might not decrease as much as desired, this allowed us to observe the chiller electrical
response to the setpoint change isolated from any other HVAC actions. Chiller load data and
outlet water temperature was recorded. These values are shown in Figure 5.1.
Full Load Amps (%)
100
80
60
40
54
52
50
48
46
44
42
6:08 8:58 11:48 14:35 18:28
Time
Figure 5.1. Electrical load data in %FLA and chilled water outlet temperature data in oF. Data was collected
over a 12 hour window
Note that this data was recorded by the Drexel BMS which measures the chiller electrical
load in terms of percentage of full load amps (FLA). Because there are no local voltage
measurements available in the BMS, it is assumed that there is a constant voltage profile and the
% FLA reading is a direct reflection of the real power drawn by the chiller motor. Another
advantage of plotting the electrical power in this manner is that we can easily view the electrical
response and outlet temperature response on the same graph. This is demonstrated more clearly
in Figure 5.2. This plot shows the electrical response versus the temperature response, and in
looking to develop a mathematical model for this component it is more telling to examine this
time period.
95
Figure 5.2. Electric load (blue line) and temperature (red line) response to raising the chill water temp.
Several important points can be drawn from observing the response in Figure 5.2. The
recovery of the electrical load does not exhibit the “fast” response normally associated with
electrical systems. Due to the dependence of the electrical load on the temperature of the system,
there are long time delays involved in the load recovery. While the speed of the response is
unique to this type of system, the general electrical power response has been exhibited before in
dynamic load studies of transmission systems as noted previously [110] [113] [114]. The
exponential recovery model used in these can be used for the HVAC chiller model as well.
To express the electric power response of the chiller, a similar set of equations as in [110]
[113] [114] is applied, the difference being the chiller is a function of time and outlet water
temperature change in lieu of time and voltage. A set of equations describing this model is
t
t
(5.3)
Pt Po c
o
where
with
dPd t dPt
Tr Pd t Ps Tr (5.4)
dt dt
where
From [110], a closed form solution may be derived to Equation (5.4) in the event of a
t t
c
Pd t Ps Pt Ps e Tr (5.5)
The equations (5.4)-(5.5) describing this response are plotted in Figure 5.3 for an
arbitrary increase in the chilled water temperature setpoint. Subplots (a) and (b) show the
temperature response. Of note is the time delay tc between when the temperature setpoint is
97
raised and when the actual chilled water temperature begins to increase, as well as the time
required for the temperature to reach its new value. The electric power response is shown in
subplots (c) – (e). The transient model Pt and the static model Ps show the minimum power
value and new steady state operating power value respectively. The dynamic load function Pd
shows the full dynamic power response. It is important to observe that the power level recovers
to a new steady state power value that is less than the initial power value Po. The level of steady-
state power recovery is dictated by the static load model index βs.
98
Figure 5.3. Response to a step change in the chilled water temperature setpoint value according to Equations
(5.4)-(5.5)
99
5.2.3. DREXEL-SPECIFIC EXAMPLE
Given the mathematical model presented here, the next step is to estimate the model
parameters based on the observed test data. The specific method of parameter estimation is not a
focus of this thesis. As noted in [111] and [115] the model parameters can be estimated from
N
min p y k yˆ k
2
pZ
(5.6)
k 1
where
Another approach is to use curve fitting and manual inspection [113] [116]. While this
approach is not well suited for automation, when considering a single test a greater context as to
The set of model parameters from (5.4)-(5.5) that need to be estimated includes θo, θs, Po,
tc, TR, s, and t. These parameters will be solved for by directly observing the data to measure
the values, as well as applying appropriate curve fits. The parameters will be solved for in the
following order:
1. Determination of tc
The time delay tc is measured as the time delay between raising the chilled water
temperature setpoint and when the actual temperature begins to rise. This delay is measured
directly from the data. The time delay is measured as the point in time after the setpoint change
100
where the observed temperature is greater than the maximum temp. from the previous minute
in order to distinguish the response from measurement noise. Through several tests performed at
Drexel, this time delay was observed to be on the order of 100 – 1000 seconds. These numbers
seem reasonable but more testing or research into the chiller control system is warranted to better
Each of these parameters was estimated by curve fitting the collected data. Figure 5.4
shows the curve fit approach to the measured test data. As shown in this figure, the fit can be
broken into three sections. Line segment 1 is applied to the initial load behavior prior to raising
the temperature setpoint. This defines the initial operating power Po and temperature θo. Line
segment 2 is applied to the drop in load following the setpoint change. This line segment runs
between the initial power Po and the minimum transient power Pt. The load recovery response is
fit using an exponential regression of the general form Y A 1 e BX . The recovery time is an
estimate of TR and this segment settles out to the new steady-state power level Ps. This
regression analysis was performed using both the Excel regression tool as well as the Matlab
Figure 5.4. Application of the curve fit approach on the collected data for parameter estimation
3. Determination of s and t
P
s ln s / ln s (5.7)
Po o
P
t ln t / ln s (5.8)
Po o
The parameter estimation method outlined in this section was applied to the test data shown in Figure 5.4.
Table 5.1 shows the resulting parameter estimates for the Drexel test described in Section
5.2.1. Note that power values are presented as a percentage of full load.
102
Table 5.1 Parameter estimation results for the test described in Section 5.2.1
Parameter Estimated Value
Po 49.75%
Ps 46.81%
Pt 28.25%
o
θo 44.11 F
o
θs 51.27 F
s -0.405
t -3.762
tc 143 sec.
TR 1690 sec.
One important observation from Table 5.1 shows that the load power recovery time TR is
approximately 28 minutes. This demonstrates the effects of the thermal time constant of the
building on the electrical response. Looking back at Figure 5.1, Figure 5.2, and Figure 5.4 there
is a significant amount of noise contained in the data. There is a concern that this amount of
noise can lead to inaccurate parameter estimates regardless of the parameter estimation method
used. In addition to the noise, there is a known hunting issue with this particular chiller. This
issue is believed to be a considerable source of noise in this data. Correcting this problem prior
to future testing may help achieve test data reflecting more regular behavior.
5.3. REMARKS
The HVAC chiller dynamic load model presented above can be used as a tool for
implementing DSM plans for any customer where their HVAC load is a significant portion of
their total electrical load. The model allows a good amount of flexibility to the end-user
depending on the time window of interest. In a simple case, the model can be used to accurately
predict a specific amount of a facility’s electric load being shed through a temperature setpoint
adjustment. Provided the dispatch interval is sufficiently long the static load model in (5.2) can
Even when using the static model, the impact of the load dynamics can still be applied.
Knowing the recovery time TR, the time-varying properties can be programmed into ramp
constraints that ensure load recovery response times meet the required scheduling windows. In
a more dynamic case, a customer can use the full dynamic exponential load model to look at the
behavior during the transient period in addition to how long it will take to reach a new steady
state load level. In the next section of this thesis, this load model will be used in a method of
planning commercial building load resources for DSM. The end goal is to determine the optimal
dispatch schedule for a group of buildings and including the load dynamics established here is
essential.
104
6.1. OVERVIEW
The previous chapters of this thesis focused on modeling and predicting building
electrical demand behavior. The study of these topics is an essential piece in determining how to
use building load resources for demand side management. Given a method to predict building
load and a model of the controllable HVAC load, we can now develop a systematic way of
Define the optimal load scheduling problem, including the variables, assumptions,
problems will address the impacts of real-time pricing, demand response revenue,
6.2.1. GOALS
The goal of the building load scheduling problem is to find the optimal day-ahead, hourly
load dispatch schedule for a multi-building campus, subject to a set of operational constraints. In
this work optimal refers to a load schedule that minimizes total electricity cost through usage
reduction and (in some cases) demand response (DR) participation. The constraints will include
electrical and thermal operating limits of the buildings. The output of the scheduling problem is
105
the optimal reduction of the controllable load for each building. Simulations will focus on
the impact of several key practical features of this problem. This includes the role of several
DSM programs and sensitivity to a number of model parameters defined for this problem.
Several terms need to be defined in order to formulate the load scheduling problem. This
work is concerned with finding optimal schedules that will deviate from the normal demand
required to assess the performance resulting from any control actions. The baseline is a forecast
of the building load assuming no demand side management actions are taken and this is the load
profile against which DSM activities are measured. The controllable load for this work is
considered to be the building HVAC chiller. This is the only portion of the load that can be
adjusted. Figure 6.1 below shows the relationship between the forecast PF and dispatched
Often a campus may have a variety of building stock available as potential DSM
resources. To characterize several features that describe potential variety of available buildings
1) Load Footprint
First, the size (electrically speaking) of a given building is referred to as the load
footprint. The load footprint of a building is defined by the maximum power level of the
106
building. In practice this value can be derived from observed historical data or the electrical
2) Load Flexibility
Some buildings will act as more responsive load resources. This responsiveness is
defined as the load flexibility. Load flexibility refers to the amount of the controllable load ∆PC
P c
changes to a building temperature setpoint change . This responsiveness is captured by
3) Load Margin
The last term that needs to be defined for this problem is the load margin. Load margin
refers to the amount of load that is available for dispatch during a given interval. This parameter
captures two important issues with dispatching building loads. First, the controllable load
margin is defined by how close the controllable nominal operating level is to any operating
limits. This margin will constrain the amount of load that can be dispatched at any given time.
This allowable range is defined in (6.1) where it is assumed the building is designed such that the
The overall building load margin is also impacted by any uncertainty in the load forecast.
The margin as defined above constrains how much the controllable load can be changed.
However, the total load change ∆P observed in relation to the forecast may be much lower.
Recall that this comparison is how settlement is done in determining DSM savings. This
situation is shown in Figure 6.2. The bounds of the load forecast Û and L̂ are depicted with
dashed lines. For identical amounts of ∆PC, the amount of load shed that qualifies as a DSM
action can be quite different. This distinction in available load margin becomes particularly
107
important in situations where the utility may impose penalties for not achieving the promised
Figure 6.2. Relationship between load margin and building load forecast uncertainty (expected worst case)
The influence of these 3 new parameters (load footprint, load flexibility, and load
margin) on the optimal load dispatch schedule will be studied in this Chapter. Each will be
addressed within the context of several DSM scenarios defined in the next section.
6.2.3. SCENARIOS
Four different scenarios are considered in this work. Each scenario corresponds to a
different demand side management program. All four scenarios are defined below:
among large electricity users who typically negotiate annually a constant cost per kWh,
provided the user stays within agreed upon minimum and maximum total electrical
energy limits. In this scenario, the controllable load for each building will be dispatched
during each interval as determined to achieve the greatest electricity cost savings.
108
Scenario 2 – Cost Minimization with Variable Pricing
As discussed in Chapter 2, there has been a push for dynamic pricing structures to
be implemented in the retail electricity market. These changes will serve to incentive
customers to take active control of their electricity usage. The only difference between
Scenario 2 and Scenario 1 is that electricity cost is allowed to vary throughout the day.
Minimizing electricity costs over the day by dispatching controllable building loads
Scenario 3 – Cost Minimization with Constant Pricing and Demand Response Revenue
Many electricity markets now allow large end-users or demand aggregators to bid
a certain amount of load into the market. Generally a threshold exists for the amount of
load that needs to be shed in order to qualify for a market payout [23]. In this scenario
the goal remains to find the load schedule that minimizes total daily electricity cost. For
each interval the possibility of earning revenue is considered as in input to the overall
cost function.
The only difference between Scenario 4 and Scenario 3 is that electricity cost is
again allowed to vary thought the day. The distinctions between each scenario are
6.2.4. ASSUMPTIONS
assumed that the electric meters where demand is measured are located directly at
the building service entry and thus the effects of the network do not need to be
considered.
A2: Day-ahead load forecasts and price forecasts are available for all buildings. It is
load forecasting (such as those in Chapter 3) and that day-ahead price forecasts
are provided and they are correct. Price volatility is not considered explicitly in
this work, though the role of price plays in the solution will be examined.
A3: The nominal operating conditions (controllable load level and temperature) for all
buildings are known. The nominal operating level PC,Nom of the building HVAC
chiller varies throughout the day as a function of the cooling requirements of the
building. Methods exist for accurately estimating this value and this work will
assume the value is readily available [89]. Therefore these quantities are assumed
A4: Load variability within the dispatch intervals is ignored. All analysis is concerned
behavior is of interest during real-time operations but is outside the scope of this
study.
A5: The dispatch interval is sufficiently long to ignore controllable load dynamics.
This assumption does not mean the dynamics of the load are ignored entirely. As
110
will be shown in the optimization problem formulations in the next section,
load recovery rates are accounted by applying a ramp rate constraint. This
assumptions, like A5 before it, means that the intra-interval transient behavior of
The optimization problem for scenario 1 and 2 is constructed as a mixed integer linear
programming (MILP) problem and is presented below. This formulation assumes for now a
perfect load forecast is available for each building. This assumption will be relaxed later on.
T M
min t , m Pt C,m (6.2)
t 1 m 1
s
t 1
t ,m mon (6.7)
where
111
PtC,m : Dispatched controllable load at time t for building m
t ,m : Price of electricity at time t for building m
st ,m : State (on/off) of controllable load at time t for building m
PtC,m,nom : Nominal controllable load level at time t for building m
PtC,m,min : Minimum controllable load level at time t for building m
PtC,m,max : Maximum controllable load level at time t for building m
tdown
,m : Ramp rate limit (down) at time t for building m
t ,m :
up
Ramp rate limit (up) at time t for building m
on
m : Limit on number of intervals PC can deviate from PC ,nom
for building m
M 0 Total number of buildings
T 0 Total number of dispatch intervals
with
where
tmax
,m : Maximum allowable temperature at time t for building m
t , m :
min
Minimum allowable temperature at time t for building m
Pt max
,m : Maximum allowable load level at time t for building m
Pt min
,m : Minimum allowable load level at time t for building m
This problem seeks to minimize the cost associated with dispatching building demand for
all buildings M over every time interval T. The optimal value will be less than or equal to zero,
price varies at each interval t but is the same for each building m.
These constraints do two things. First, the quantity of dispatched load must not cause the
building to operate above or below practical limits (i.e.- controllable load margin). These
minimum and maximum values (determined via (6.8) - (6.9)) capture both the electrical limits
(e.g. – HVAC machine ratings) and thermal comfort limits. Whichever is more restrictive at
The second function of these constraints is to force Pt C,m =0 during intervals when the
load is not being dispatched. Inserting the operating state st , m in these inequality constraints
avoids the situation where Pt C,m and st , m would be multiplied in the objective function, thus
These linear constraints capture the controllable load recovery dynamics. The ramp rate
limits tdown
,m and upt, m are derived from the recovery time of the nonlinear load model. These
constraints ensure raising or lowering of the load to a new operating point is achievable within
It is necessary to limit the number of times the controllable load is operated off of its
normal operating point. Even though the thermal comfort issue is managed by the min/max
constraints, there is a practical need to limit the time the HVAC system operates off its normal
level. As was discussed earlier, repeated cycling of the HVAC system is not good for reliability
reasons. Additionally, HVAC systems are designed for specific operating conditions for
113
efficiency reasons. Therefore it is prudent to limit the total number of times a day the HVAC
Decision variables
The decision variables are Pt C, m and st , m . There are (M T) of each variable: one for each
building at every time step. The dispatched power Pt C, m is a real valued number and the load
state is a binary on/off indicator {0,1}. The total number of decision variables is therefore 2
(MT).
The optimization problem for scenario 3 and 4 is constructed as a mixed integer linear
T M T M
min t , m Pt C, m t , m Pt ,DR
m (6.10)
t 1 m 1 t 1 m 1
P
m 1
C
t ,m BstDR PtThresh (6.11)
m B 1 st
Pt ,DR DR (6.15)
where
An additional term is added to the cost function to capture potential revenue earned
through demand response (DR) programs. A new decision variable Pt ,DRm is introduced to
quantify load shed that qualifies for DR revenue. This variable equals Pt C, m if the required
The price of electricity t ,m is kept constant t,m in scenario 3. In scenario 4 the price
varies at each interval t but is the same for each building m. For both scenarios, the LMP t ,m
In order to qualify for DR revenue the amount of load shed must exceed a certain
threshold PtThresh . Another decision variable for the state of DR revenue is introduced ( stDR ). This
constraint ensures that stDR takes the value of 1 if the threshold is met and zero otherwise.
The variable Pt ,DRm does not represent a real quantity of load separate from Pt C, m . It is a
M
variable that is used solely to account for DR revenue if the condition P
m 1
C
t ,m PtThresh exists. To
that end, these constraints ensure that if the above condition is true, Pt ,DRm Pt C, m . If it does not,
Decision variables
The decision variables Pt C, m and st , m from scenario 1 and 2 are the same here. The two
new decision variables introduced in scenario 3 and 4 are Pt ,DRm and stDR . There are (M x T)
variables for Pt ,DRm - one for each building at every time interval. There are T stDR variables – one
for each time interval. The total number of decision variables is therefore 3 (MT)+T.
The goal remains finding the optimal Pt C, m that will lead to the greatest electricity cost
savings. However now the difference between the dispatched controllable load Pt C, m and the
deviation from the total predicted load Pt , m becomes relevant. An error parameter ξ which in
116
practice can be derived from a probabilistic forecast will be added to the constraints to
T M
min t , m Pt C,m (6.16)
t 1 m 1
s
t 1
t ,m mon (6.21)
Notice that the objective function remains unchanged. There is no need to add an
additional cost term associated with the load uncertainty. As the objective is written, the
resulting optimal day-ahead dispatch schedule will still minimize costs with respect to how much
Just like with scenario 1 and 2, a new parameter ξ is introduced. Following the same
T M T M
min t , m Pt C, m t , m Pt ,DR
m (6.22)
t 1 m 1 t 1 m 1
117
s.t. t 1,..., T m 1,..., M
P
m 1
C
t ,m t , m BstDR PtThresh (6.23)
m B 1 st
Pt ,DR DR (6.26)
m B 1 st
Pt ,DR DR (6.27)
As with the deterministic case, the variable Pt ,DRm does not represent a real quantity of
load separate from Pt C, m . It is a quantity that is used solely to account for DR revenue if the
M
condition P
m 1
C
t ,m PtThresh exists. Therefore no uncertainty term is associated with that variable.
6.4. SIMULATIONS
The proposed load scheduling methods are tested on a fictitious 10 building system. The
relevant parameters for each building are based on observations of Drexel University buildings
so while the test system does not actually exist it is representative of a real world campus. It is
desirable to have a variety of building stock to observe the impact each building type has in the
proposed load scheduling approaches. Variety in this case refers to both the load footprint as
well as load flexibility. For this study the following range values are presented. Building load
footprint classification is shown in Table 6.2. Building controllable load flexibility is shown in
118
Table 6.3. A summary of the building types that form the base of this study are presented in
Table 6.4.
Table 6.4. Load footprint and flexibility for each building used in this case study
Building # Footprint Flexibility
1 Small Medium
2 Medium Medium
3 Medium High
4 Small High
5 Large Medium
6 Large Low
7 Large High
8 Medium Low
9 Medium High
10 Large Medium
The forecasted building load profiles are shown in Figure 6.3. These profiles define the
nominal operating behavior against which the optimal dispatched load is compared.
119
900
Bldg 1
Bldg 2
800 Bldg 3
Bldg 4
Bldg 5
Bldg 6
700
Bldg 7
Bldg 8
Bldg 9
600
Bldg 10
Demand (kW)
500
400
300
200
100
0 5 10 15 20 25
Hour
Electricity price information for a single day has been obtained from the PJM website
[117] for Monday, July 21, 2014. These prices are representative of the LMP prices at the nodes
in the system where Drexel University connects to the grid. For the variable pricing scenarios
and for the DR revenue pricing, these LMP prices are used as the cost of electricity. For the
constant pricing scenarios, the price of electricity is assumed to be the average of the daily LMP.
65
60 LMP
mean(LMP)
55
50
LMP ($/MWh)
45
40
35
30
25
20
0 5 10 15 20 25
Hour
Finally, the parameters that define the building operating behavior and constraints is
The impact to the load schedule solution from changing some of these parameters is one of the
main goals of this work. Therefore some of the values presented here may be changed during
various tests. Any deviations from these numbers will be specifically noted.
It bears mentioning that the following simulations will not compare the performance of
the proposed load scheduling algorithms against any established methods. This is partly because
we are more interested in observing the impact of various DSM scenarios and model parameters
on the solution and less on the value of the solution itself. It is also partly because there is no
All of the following simulations are carried out using the OPTI Toolbox for Matlab
[118]. Specifically the GLPK solver is used to solve each MILP problem. This solver employs a
version of the simplex method. All simulations are performed on a personal computer with 2
processors clocking at 2.0 GHz and 8 GB of RAM. The computational aspects of the problem
are not considered for this work but it should be noted that none of these simulations required
more than 30s to complete. This is perfectly suitable for a day-ahead scheduling algorithm.
The optimal dispatch schedule is given in Table 6.6 on the next page. The total savings
achieved through this dispatch schedule is $33212.00. Several key observations can be made
the day, the most potential value occurs where the greatest amount of load margin
exists. Considering the load forecasts in Figure 6.3, it makes sense that most of
122
the dispatch is scheduled during the middle of the day when building load is
highest.
o The amount of load dispatched per building is clearly driven by the size of the
load footprint as well as the load flexibility. The next two tests will look at how
Test 2: Examine the impact of load sensitivity to the Scenario 1 dispatch schedule
For this test the β value for building 7 is varied from -1 to -0.3 in steps of -0.1. This is a
major reduction in load flexibility. No other changes are made. The optimal dispatch
schedule for β = -0.3 is given in Table 6.7 on the previous page. Several key observations
o The savings is reduced by $4832 when compared to the schedule in Table 6.6
building 7 by such a large amount the capacity for shedding load (and hence
o It is notable that the rest of the dispatch schedule is unchanged for each building.
6.5. The load sensitivity does not impact the timing of the schedule in this
Scenario, only the amount of power that is dispatched and only for a specific
building. This result should be intuitive. Load will be dispatched when the
largest margin for shedding exists. This is driven by the nominal operating level
constant pricing DSM scenario, impacts how much load is dispatched for a
specific building only, and no impact not when the load is dispatched.
123
Table 6.7. Optimal load schedule for in kW Test 2 (Scenario 1 with reduced load flexibility)
Interval Number
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
1 0 0 0 0 0 0 0 -10.94 0 -10.94 0 0 -10.94 0 0 -10.94 0 0 0 0 0 0 0 0
2 0 0 0 0 0 0 0 0 0 0 0 0 0 -11.91 -11.57 0 0 0 0 0 0 0 0 0
Building Number
Figure 6.5. Dispatch schedule for buildings 7 with decreasing load flexibility (No dispatch during the
intervals not shown)
Test 3: Examine the impact of load footprint to the Scenario 1 dispatch schedule
For this test the load footprint of building 7 is varied. The nominal operating point across the
whole day is reduced to 75%, 50%, and 25% of the normal level. No other changes are
made. The complete load schedule is omitted for this test as the overall results are very
similar to Table 6.7. Several observations about the impact of varying the load footprint are
made:
o The savings is reduced from $33212.00 to $28512 as the load footprint goes
down. The response is similar to Test 2 where load flexibility is decreased. This
Figure 6.6 Dispatch schedule for buildings 7 with decreasing load footprint (No dispatch during the intervals
not shown
The optimal dispatch schedule for scenario 2 is shown in Table 6.9. The total savings
achieved through this dispatch schedule is $47164.00. Several key observations can be made
o Comparing the optimal schedule under dynamic pricing (Table 6.9) to the
optimal schedule under constant pricing (Table 6.6) the savings increases by
$28496 with dynamic pricing. Simply observing the price curve in Figure 6.4
o More interesting than the overall savings, the timing of the dispatch schedule
changes drastically with dynamic pricing. This shift in dispatch timing can be
seen in Table 6.9. Load is shed during periods of higher prices even though
126
these periods often have lower load available for dispatch. In fact, the
consumer participation.
Test 5: Find the optimal dispatch schedule for Scenario 3 in the deterministic case
This is the first test where a market entry constraint is present. In this test, PtThresh 150kW .
The optimal dispatch schedule for scenario 3 is shown in Table 6.8. The total savings
achieved through this dispatch schedule is $71417. Several key observations can be made
Breaking down the savings, it can be seen that the overall load reduction saves
schedule. Looking at Table 6.8 compared to Table 6.6 it is apparent that the
optimal solution shifts load reductions to intervals where the total can meet
o Similar to the results for Test 4, the overall load reduction is less for Scenario
3 than Scenario 1 (880kW vs 931.00 kW). This underlines the point that
In this test, the market entry constraint threshold is cut in half ( PtThresh 75kW ). The optimal
dispatch schedule is shown in Table 6.10. The total savings achieved through this dispatch
schedule is $77538. Reducing the market entry constraint threshold allows for more
as the dispatch schedule shifts to make the most of the lower market entry threshold.
DR revenue
The optimal dispatch schedule for scenario 4 is shown in Table 6.11. The total savings
achieved through this dispatch schedule is $83666.66. Several key observations can be made
This can be attributed to the fact that higher prices occur during intervals
Test 8: Find the optimal dispatch schedule for Scenario 1 in the uncertain case
This is the first test to include uncertainty related to the load forecast. As a first cut, the error
parameter ξ will be held constant at 10kW for all buildings and all intervals. The resulting
o In this simple example, the dispatch schedule differs from Table 6.6 in all
intervals by +10kW (in other words, 10kW less load shed). While this seems
a trivial result, it is important to note that the results are showing the load shed
in relation to the baseline forecast, not just dispatched controllable load. This
129
shows the expected total building load reduction ∆P. The load reduction
o For building 4 and 8, which have very small load footprints, the forecast
uncertainty.
Test 9: Observe cost savings as forecast uncertainty varies from best case to worst case
In order to see how the overall expected cost saving are impacted with increasing forecast
uncertainty the error parameter ξ will be varied over a range of values. Unlike the previous
test, the load footprint of the building will be taken into account by scaling ξ. For each
building m, ξ will be varied at each interval t from -10% to 10% of the forecasted building
demand. Total campus expected cost savings are plotted against the range of ξ values in
Figure 6.7
point at slightly less than 5% uncertainty where the expected savings becomes
an expected expenditure.
o This plot shows both the worst case and best case in total building load
dispatch. When the forecast overestimates the load (ξ<0) then the consumer
“sheds” extra load. However in the worst case, the amount of dispatchable
10
8
Total Expected Campus ($)
Savings
0
Expenditure
-2
-4
-10 -8 -6 -4 -2 0 2 4 6 8 10
Load Uncertainty (% of Forecast)
Figure 6.7 Cost savings as a function of load uncertainty. Break even line shown in blue
Test 11: Examine the impact of forecast uncertainty to the Scenario 3 dispatch
schedule
For each building m, the load uncertainty parameter ξ will be varied at each interval t from -
10% to 10% of the forecasted building demand. The impact of the load uncertainty is
reflected in the number of intervals where the DR revenue threshold can be met
( PtThresh 100kW ). These results are shown in Figure 6.8. A small deviation from the perfect
forecast can results in a great misunderstanding of how often the DR revenue constraint can
be met. A 3% underestimation of the forecast gives the impression that the DR revenue
constraint can never be met, while a 3% overestimation makes it look as though it can be met
25
Increasing underestimation
DR Constraint Met (# of Intervals) 20
15
10
Increasing overestimation
-10 -8 -6 -4 -2 0 2 4 6 8 10
Load Uncertainty (% of Forecast)
Figure 6.8. Decreasing number of intervals where the DR revenue constraint can be met
132
Table 6.12 Optimal load schedule in kW for Test 8: +10kW uniform load forecast uncertainty
Interval Number
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
1 0 0 0 0 0 0 0 -0.94 0 -0.94 0 0 -0.94 0 0 -0.94 0 0 0 0 0 0 0 0
2 0 0 0 0 0 0 0 0 0 0 0 0 0 -1.91 -1.57 0 0 0 0 0 0 0 0 0
Building Number
7.1. OVERVIEW
A discussion of how the work presented in this thesis can be extended in the
future
The overall vision for this work was to study demand side management (DSM) using
commercial buildings with a focus on addressing practical issues of interest to the end-user.
Specifically these contributions fall into two main areas. The first focuses on the forecasting of
building electrical demand. The second is concerned with methods for scheduling commercial
In this work a case study involving Drexel University building demand provides the basis
for studying building specific load forecasting techniques. Four popular methods for load
forecasting are implemented and tested on actual building data. These traditional approaches are
enhanced by measurements collected via a BMS which is a new contribution to this area. In
addition to point forecasts and method for generating probabilistic forecasts of building demand
is presented. These studies demonstrate the overall improved performance (in terms of accuracy,
consistency, and overall skill) when building measurements are included in the models.
Additionally these tests have examined some performance characteristics that are important
understand how building-specific behavior may be valuable in future efforts to better forecast
building loads.
135
In addition to the forecasting studies, the problem of scheduling building electrical
loads for DSM has been presented. A new model for the electrical demand of an HVAC chiller
was derived from Drexel building data, leveraging existing dynamic load modeling techniques.
This model captures the dynamics exhibited by the HVAC system that arises from the strong link
between the thermal and electrical behavior of the building and has the flexibility to be used
across a number of time windows (i.e.- static vs transient studies). This model has the added
benefit of being well suited for integration in a load scheduling algorithm. To that end, several
load scheduling optimizations problems have been presented for determining the optimal
cover several different potential DSM programs, including simple minimization of usage,
dynamic pricing, and the potential to bid DR resources into the market. Three parameters, load
footprint, load flexibility, and load margin, have been introduced to help analyze building
dispatch capabilities. The analysis performed has identified the important sensitivities of the
solution to variations in model parameters and DSM scenarios. Lastly, the uncertainty related to
building forecasts is introduced into the problem. To get a feel for the impact such uncertainty
has on the dispatch schedule, potential best and worst case uncertainty scenarios are presented.
This thesis has laid the foundation for several areas of future research. First, it would be
beneficial to develop a load forecasting model that is “best” for building-specific applications.
This model would ideally account for the building-specific characteristics that have been
demonstrated in this thesis as being highly impactful. Some preliminary work has been done in
this area in [119]. In that work, a forecast model is developed that considers resampling from
Gaussian and non-Gaussian copulas are used in tandem with a boosting procedure to improve
136
forecast accuracy. This approach captures some of the temporal dependencies of building
load forecast model. Data on occupancy levels and other building usage information may lead to
better load characterization. This information would need to be balanced with keeping the
There is also a number of aspects of the load scheduling problem that can enhanced.
First is a more systematic way of including forecast uncertainty in the optimization formulation.
The work in this thesis has laid the groundwork for a scheduling method that provides practical
schedules that are sensitive to real-world building constraints. However the uncertainty of the
forecast can be better balanced in terms of the conservativeness of the solution. In addition to
the forecast uncertainty it would be good to consider price forecast uncertainty. This thesis
showed how price-dependent the dispatch schedule is under various DSM programs. However a
forecast of day-ahead prices will, just like the load forecast, include some uncertainty. Lastly, it
would be interesting to explore extensions of these formulations into other DSM scenarios. For
example, this can include the ancillary services market, where the model could try and balance
the benefits of load reduction with potential rewards for raising building load to support the
power system operation. Real-time load dispatch is also an important possible extension where
the full dynamics of the load would need to be incorporated to model the scenario accurately.
137
List of References
[1] "H.R. 6 (110th): Energy Independence and Security Act of 2007," 2007. [Online]. Available:
https://www.govtrack.us/congress/bills/110/hr6. [Accessed 2010].
[2] P. Cappers, C. Goldman and D. Kathan, "Demand response in U.S. electricity markets: Empirical
evidence," Energy, vol. 35, no. 4, pp. 1526-1535, 2010.
[3] S. Borenstein, "The trouble with electricity markets: Understanding California's restructuring
disaster," Journal of Economic Perspectives, vol. 16, no. 1, pp. 191-211, 2002.
[4] U.S. Congress, "Energy Policy Act of 2005," 2005. [Online]. Available:
https://www.gpo.gov/fdsys/pkg/BILLS-109hr6enr/pdf/BILLS-109hr6enr.pdf. [Accessed 2016].
[8] FERC, "A national assessment of demand response potential," Prepared by the Brattle Group;
Freeman, Sullivan & Co.; and Global Energy Partners, 2009. [Online]. Available:
https://www.ferc.gov/legal/staff-reports/06-09-demand-response.pdf. [Accessed 2016].
[9] U.S. DOE, "Benefits of demand response in electricity markets and recommendations for
achieving them," Technical report to the United States Congress, 2006. [Online]. Available:
http://energy.gov/sites/prod/files/oeprod/DocumentsandMedia/DOE_Benefits_of_Demand_Response_in_E
lectricity_Markets_and_Recommendations_for_Achieving_Them_Report_to_Congress.pdf. [Accessed
2016].
[10] M. Albadi and E. El-Saadany, "A Summary of Demand Response in Electricity Markets," Electric
Power Systems Research, vol. 78, pp. 1989-1996, 2008.
[11] EPRI, "The green grid: Energy savings and carbon reductions enabled by a smart grid.," Electric
Power Research Institute, 2008.
[12] D. Chassin and J. Fuller, "On the Equilibrium Dynamics of Demand Response in Thermostatic
Loads," in Hawaii International Conference on System Sciences (HICSS), 2011.
[13] D. Kirschen, "Demand-side View of Electricity Markets," IEEE Transactions on Power Systems,
vol. 100, no. 2, pp. 1104-1118, 2003.
[14] N. Liu, "An Evaluation of the HVAC Load Potential for Providing Load Balancing Service,"
IEEE Transactions on Smart Grid, vol. 3, no. 3, pp. 1263-1270, 2012.
[15] D. S. Callaway, "Tapping the energy storage potential in electric loads to deliver load following
and regulation, with application to wind energy," Energy Conversion and Management, vol. 50, no. 5, pp.
1389-1400, 2009.
[16] NERC, "Long-term reliability assessment 2008-2017," North American Electric Reliability
Corporation, 2008.
[17] C. De Jonghe and B. F. B. R. Hobbs, "Optimal Generation Mix With Short-Term Demand
Response and Wind Penetration," IEEE Transactions on Power Systems, vol. 27, no. 2, pp. 830-839, 2012.
138
[18] J. Eto, "Innovative developments in load as a reliability resource," in Proceedings of the IEEE PES
Winter Meeting, 2002.
[19] DOE/EIA, "Annual Energy Outlook 2015 with projections to 2040," U.S. Energy Information
Administration (EIA), 2015.
[20] D. Callaway and I. Hiskens, "Achieving controllability of electric loads," Proceedings of the
IEEE, vol. 99, no. 1, pp. 184-199, 2011.
[21] PG&E, "Explore the PG&E Time-of-Use plans," 2016. [Online]. Available:
https://www.pge.com/en_US/residential/rate-plans/rate-plan-options/time-of-use-base-plan/time-of-use-
plan.page. [Accessed 2016].
[22] PG&E, "Find out if Peak Day Pricing is right for your business," PG&E, 2016. [Online].
Available: https://www.pge.com/en_US/business/rate-plans/rate-plans/peak-day-pricing/peak-day-
pricing.page. [Accessed 2016].
[23] PJM, "PJM Manual 11: Energy & Ancillary Services Market Operations," August 2016. [Online].
Available: http://www.pjm.com/~/media/documents/manuals/m11.ashx. [Accessed 2016].
[26] J. Berardino and C. Nwankpa, "Dynamic Load Modeling of an HVAC Chiller for Demand
Response Applications," in Proceedings of the IEEE Smart Grid Communications Conference, 2010.
[27] D. Hill, "Nonlinear Dynamic Load Models with Recovery for voltage Stability Studies," IEEE
Transactions on Power Systems, vol. 8, no. 1, pp. 166-176, 1993.
[28] J. Lam and e. al, "An analysis of electricity end-use in air-conditioned office buildings in Hong
Kong," Building and Environment, vol. 38, no. 3, pp. 493-498, 2003.
[31] CAISO, "Baselines for Retail Demand Response," 2009. [Online]. Available:
https://www.caiso.com/Documents/Presentation-Baselines_RetailDemandResponsePrograms.pdf.
[Accessed 2016].
[32] M. Goldberg and G. Agnew, "Protocol development for demand response calculation - findings
and recommendations," California Energy Commission CEC-400-02-017F, 2003.
[33] K. Coughlin, M. Piette, C. Goldman and S. Kiliccote, "Statistical analysis of baseline load models
for non-residential buildings," Energy and Buildings, vol. 41, no. 4, 2009.
[34] S. Kiliccote, M. Piette and D. Hansen, "Advanced controls and com- munications for demand
response and energy efficiency in commercial buildings," in Proceedings of Second Carnegie Mellon
Conference in Electric Power Systems, 2006.
139
[35] S. Braithwait, D. Hansen and J. Reaser, "2009 load impact evaluation of California statewide
critical-peak pricing rates for non-residential customers: Ex post and ex ante report," Christensen
Associates Energy Consulting CALMAC Study ID: SDG0244.01, 2010.
[36] KEMA, "PJM Empirical Analysis of Demand Response Baseline Methods," Prepared for the PJM
Market Implementation Committee, 2011.
[38] J. Berardino and C. Nwankpa, "Interval-specific building load models for demand resouce
planning," in Proceedings of the IEEE PowerCon 2012 Conference, Auckland, 2012.
[39] J. Berardino and C. Nwankpa, "Inclusion of temporal effects in forecasting building electrical
loads for demand resource planning," in Proceedings of the IEEE PowerTech 2013 Conference, Grenoble,
2013.
[40] D. Palchak, S. Suryanarayanan and D. Zimmerle, "An Artificial Neural Network in Short-Term
Electrical Load Forecasting of a University Campus: A Case Study," Journal of Energy Resources
Technology, vol. 135, no. 3, 2013.
[41] W. Mai, C. Y. Chung, T. Wu and H. Huang, "Electric load forecasting for large office building
based on radial basis function neural network," in Proceedings of the 2014 IEEE PES General Meeting,
National Harbor, MD, 2014.
[42] H. Chitsaz, H. Shker, H. Zareipour, D. Wood and N. Amjady, "Short-term electricity load
forecasting of buildings in microgrids," Energy and Buildings, vol. 99, pp. 50-60, 2015.
[43] P. A. Gonzalez and J. M. Zamarreno, "Prediction of hourly energy consumption in buildings based
on a feedback artificial neural network," Energy and Buildings, vol. 37, no. 6, pp. 595-601, 2005.
[44] S. Karatasou, M. Santamouris and V. Geros, "Modeling and predicting building's energy use with
artificial neural networks: Methods and results," Energy and Buildings, vol. 38, no. 8, pp. 949-958, 2006.
[45] P. Kou and Gao, "A sparse heteroscedastic model for the probabilistic load forecasting in energy-
intensive enterprises," International Journal of Electric Power Energy Systems, vol. 55, pp. 144-154, 2014.
[46] B. Dong, C. Cao and S. E. Lee, "Applying support vector machines to predict building energy
consumption in tropical region," Energy and Buildings, vol. 37, no. 5, pp. 545-553, 2005.
[47] J. Massana, C. Pous, L. Burgas, J. Melendez and J. Colomer, "Short-term load forecasting in a
non-residential building contrasting models and attributes," Energy and Buildings, vol. 92, pp. 322-330,
2015.
[48] C. E. Borges, Y. K. Penya and I. Fernandez, "Optimal combined short-term building load
forecasting," in Proceedings of the 2011 IEEE PES Innovative Smart Grid Technologies (ISGT)
Conference, 2001.
[49] J. G. Jetcheva, M. Majidpour and W.-P. Chenwei, "Neural Network Model Ensembles for
Building-Level Electricity Load Forecasts," Energy and Buildings, vol. 84, pp. 214-223, 2014.
[50] A. Afram and F. Janabi-Sharifi, "Review of modeling methods for HVAC systems," Applied
Thermal Engineering, vol. 67, no. 1-2, pp. 507-519, 2014.
140
[51] A. Khodaei, M. Shahidehpour and S. Bahramirad, "SCUC With Hourly Demand Response
Considering Intertemporal Load Characteristics," IEEE Transactions on Smart Grid, vol. 2, no. 2, pp. 564-
571, 2011.
[53] P. Palensky and D. Dietrich, "Demand Side Management : Demand Response , Intelligent Energy
Systems , and Smart Loads," IEEE Transactions on Industrial Informatics, vol. 7, no. 3, pp. 381-388, 2011.
[54] J. S. Vardakas, N. Zorba and C. V. Verikoukis, "A Survey on demand response programs in smart
grids: pricing methods and optimization algorithms," IEEE Communications Surveys & Tutorials, vol. 17,
no. 1, pp. 152-178, 2014.
[55] R. Deng, Z. Yang, M.-Y. Chow and J. Chen, "A Survey on Demand Response in Smart Grids:
Mathematical Models and Approaches," IEEE Transactions on Industrial Informatics, vol. 11, no. 3, 2015.
[56] S. Bashash and H. Fathy, ""Modeling and control insights into demand-side energy management
through setpoint control of thermostatic loads," in Proceedings of the American Control Conference
(ACC), 2011.
[57] N. Lu and D. Chassin, "A state queuing model of thermostatically controlled appliances," IEEE
Transactions on Power Systems, vol. 19, no. 3, pp. 1666-1673, 2004.
[58] J. Mathieu and D. Callaway, "State estimation and control of heterogeneous thermostatically
controlled loads for load following," in Proceedings of the Hawaii International Conference on Systems
Science, 2012.
[59] W. Burke and D. Auslander, "Development of an HVAC Load Model for Aggregates of Homes,"
California Energy Commission Multi-Campus Award No. C-08-19, 2010.
[60] I. Moghram and S. Rahman, "Analysis of five short-term load forecasting techniques," IEEE
Transactions on Power Systems, vol. 4, no. 4, pp. 1484-1491, 1989.
[61] H. Alfares and M. Nazeeruddin, "Electric load forecasting: literature survey and classification of
methods," International Journal of System Science, vol. 33, no. 1, pp. 23-34, 2002.
[62] H. Hahn, S. Meyer-Nieburg and S. Pickl, "Electric load forecasting methods: tools for decision
making," European Journal of Operational Research, vol. 199, no. 3, pp. 902-907, 2009.
[63] T. Hong, "Short term electric load forecasting," Ph.D. dissertation, North Carolina State
University, 2010.
[64] H. Hippert, C. Pedreira and R. Souza, "Neural networks for short-term load forecasting: a review
and evaluation," IEEE Transactions on Power Systems, vol. 16, no. 1, pp. 44-55, 2001.
[66] S. Li, P. Wang and L. Goel, "A Novel Wavelet-Based Ensemble Method for Short-Term Load
Forecasting with Hybrid Neural Networks and Feature Selection," IEEE Transactions on Power Systems,
vol. 31, no. 3, pp. 1788 - 1798, 2016.
[67] A. Deihimi and H. Showkati, "Application of echo state networks in short-term electric load
forecasting," Energy, vol. 39, no. 1, pp. 327-340, 2012.
141
[68] W.-C. Chu, Y.-P. Chen, Z.-W. Xu and W.-J. Lee, "Multiregion Short-Term Load Forecasting in
Consideration of HI and Load/Weather Diversity," IEEE Transactions on Industry Applications, vol. 47,
no. 1, pp. 232-237, 2011.
[69] N. Amjady, "Short-Term Bus Load Forecasting of Power Systems by a New Hybrid Method,"
IEEE Transactions on Power Systems, vol. 22, no. 1, pp. 333-341, 2007.
[70] R.-J. Wai, Y.-C. Chen and Y.-R. Chang, "Short-term load forecasting via fuzzy neural network
with varied learning rates," in Proceedings of the 2011 IEEE International Conference on Fuzzy Systems
(FUZZ-IEEE 2011), 2011.
[71] B.-J. Chen, M.-W. Chang and C.-J. Lin, "Load Forecasting Using Support Vector Machines: A
Study on EUNITE Competition 2001," IEEE Transactions on Power Systems, vol. 19, no. 4, pp. 1821-
1830, 2004.
[72] D. Niu, Y. Wang and D. D. Wu, "Power load forecasting using support vector machine and ant
colony optimization," Expert Systems with Applications, vol. 37, no. 3, pp. 2531-2539, 2010.
[73] S. Fan and L. Chen, "Short-term load forecasting based on an adaptive hybrid method," IEEE
Transactions on Power Systems, vol. 21, no. 1, pp. 392-401, 2006.
[74] A. Hinojosa and V. Hoese, "Short-term load forecasting using fuzzy inductive reasoning and
evolutionary algorithms," IEEE Transactions on Power Systems, vol. 25, p. 565–574, 2010.
[75] S. Fan and R. J. Hyndman, "Short-term load forecasting based on a semi-parametric additive
model," IEEE Transactions on Power Systems, vol. 27, no. 1, pp. 134-141, 2012.
[76] Y. Goude, R. Nedellec and N. Kong, "Local short and middle term electricity load forecasting
with semi-parametric additive models," IEEE Transactions on Smart Grid, vol. 5, no. 1, pp. 440-446, 2014.
[77] R. Nedellec, J. Cugliari and Y. Goude, "Gefcom2012: Electricity load forecasting and backcasting
with semi-parametric models," International Journal of Forecasting, vol. 30, no. 2, p. 375–381, 2014.
[78] T. Hong and P. F. S. Pinson, "Global Energy Forecasting Competition 2012," International Journal
of Forecasting, vol. 30, no. 2, pp. 357-363, 2014.
[79] M. Kutner, C. Nachtsheim, J. Neter and W. Li, Applied Linear Statistical Models, McGraw-
Hill/Irwin, 2004.
[80] C. Cortes and V. Vapnik, "Support-vector Network," Machine Learning, vol. 20, pp. 273-297,
1995.
[81] C.-C. Chang and C.-J. Lin, "LIBSVM: A Library for Support Vector Machines," 2001. [Online].
Available: https://www.csie.ntu.edu.tw/~cjlin/libsvm/. [Accessed 2016].
[82] S. Arlot and A. Celisse, "A survey of cross-validation procedures for model selection," Statistical
Surveys, vol. 4, pp. 40-79, 2010.
[83] T. Hastie and R. Tibshirani, Generalized Additive Models, London, U.K.: Chapman & Hall, 2006.
[84] S. Wood, "mgcv: GAMs and generalized ridge regression of R," R News, vol. 1, no. 2, pp. 20-25,
2001.
[85] S. Wood, "Stable and efficient multiple smoothing parameter estimation for generalized additive
models," Journal of the American Statistical Association, vol. 99, p. 673–686, 2004.
142
[86] S. Wood, "Fast stable restricted maximum likelihood and marginal likelihood estimation of
semiparametric generalized linear models," Journal of the Royal Statistical Society B, vol. 73, no. 1, p. 3–
36, 2011.
[88] S. Katipamula, R. Underhill, J. Goddard, G. Liu and S. Eaton, "Instructor Manual: Re-tuning
Large Laboratory (PNNL-SA-82896)," U. S. Department of Energy, Pacific Northwest National
Laboratory , 2011.
[89] M. Muthalib and C. Nwankpa, "Physically Based Building Load Model for Electric Grid
Operation and Planning," IEEE Transactions on Smart Grid, 2016.
[90] P. Pinson, "Estimation of the uncertainty in wind power forecasting," Ph.D. dissertation, Centre
Energétique et Procédés–Ecole des Mines de Paris, Paris, France, 2006.
[91] G. Verschuuren, "Confidence Intervals vs. Prediction Intervals," 2014. [Online]. Available:
https://www.youtube.com/watch?v=o0UESA3UZss. [Accessed 2016].
[92] T. Heskes, "Practical Confidence and Prediction Intervals," in Advances in Neural Information
Processing Systems 9, M. Mozer, M. Jordan and T. Petsche, Eds., Cambridge, MIT Press, 1997.
[93] L. Wu, M. Shahidehpour and T. Li, "Stochastic security-constrained unit commitment," IEEE
Transactions on Power Systems, vol. 22, no. 2, pp. 800-811, 2007.
[94] Y. Wang, Q. Xia and C. Kang, "Unit commitment with volatile node injections by using interval
optimization," IEEE Transactions on Power Systems, vol. 26, no. 3, pp. 1705-1713, 2011.
[95] J. Hoffer and P. Dorfner, "Reliability and production cost calculation with peak load forecast
uncertainty," International Journal of Electrical Power & Energy Systems, vol. 13, no. 4, pp. 223-229,
1991.
[97] R. Billinton and D. Huang, "Effects of load forecast uncertainty on bulk electric system reliability
evaluation," IEEE Transactions on Power Systems, vol. 23, no. 2, pp. 418-425, 2008.
[98] T. Hong and S. Fan, "Probabilistic electric load forecasting: A tutorial review," International
Journal of Forecasting, vol. 32, pp. 914-938, 2016.
[100] R. Weron, "Electricity price forecasting: A review of the state-of-the-art with a look into the
future," International Journal of Forecasting, vol. 30, no. 4, pp. 1030-1081, 2014.
[101] T. Hong, J. Wilson and J. Xie, "Long term probabilistic forecasting and normalization with hourly
information," IEEE Transactions on Smart Grid, vol. 5, no. 1, pp. 456-462, 2014.
[102] R. Hyndman and S. Fan, "Density Forecasting for long-term peak electricity demand," IEEE
Transactions on Power Systems, vol. 25, pp. 1142-1153, 2010.
[103] J. Bremnes, "Probabilistic wind power forecasts using local quantile regression," Wind Energy,
vol. 7, no. 1, pp. 47-54, 2004.
143
[104] T. Hong, P. Pinson, S. Fan, H. Zareipour, A. Troccoli and R. Hyndman, "Probabilistic energy
forecasting: Global Energy Forecasting Competition 2014 and beyond," International Journal of
Forecasting, vol. 32, no. 3, pp. 896-913, 2016.
[105] R. Winkler, "A decision-theoretic approach to interval estimation," Journal of the American
Statistical Association, vol. 67, pp. 187-191, 1972.
[106] B. Liu, J. Nowotarski, T. Hong and R. Weron, "Probabilistic load forecasting via quantile
regression averaging on sister forecasts," IEEE Transactions on Smart Grid, In Press..
[107] P. F. Christoffersen, "Evaluating interval forecasts," International Economic Review, vol. 39, no.
4, p. 841–862, 1998.
[108] "IEEE Stability Special: Voltage Stability of Power Systems: Concepts, Analytical Tools and
Industry Experience," IEEE Special Publication, 1990.
[109] C. Taylor, Power System Voltage Stability, Electric Power Research Institute, McGraw-Hill,
1994.
[110] D. Hill, "Nonlinear dynamic load models with recovery for voltage stability studies," IEEE
Transactions on Power Systems, vol. 8, no. 1, pp. 166- 176, 1993.
[111] D. Karlsson and D. Hill, "Modeling and identification of nonlinear dynamic loads in power
systems," IEEE Transactions on Power Systems, vol. 9, no. 1, pp. 157-166, 1994.
[112] S. Casper, C. Nwankpa, R. Fischl, A. Devito and S. Readinger, "On substation tests for load
modeling," in Proceedings of the 26th North American Power Symposium (NAPS) Part II, Manhattan, KS,
1995.
[113] Y. Liang, R. Fischl, A. DeVito and S. Readinger, "Dynamic reactive load model," IEEE
Transactions on Power Systems, vol. 13, no. 4, pp. 1365-1372, 1998.
[114] IEEE Task Force on Load Model Representation for Voltage Stability Studies, "Bibliography on
Load Models for Power Flow and Dynamic Performance Simulation," IEEE Transactions on Power
Systems, vol. 10, no. 1, pp. 523-538, 1995.
[115] B.-K. Choi, H.-D. Chiang, Y. Li, H. Li, Y.-T. Chen, D.-H. Huang and M. Lauby, "Measurement-
Based Dynamic Load Models: derivation, comparison, and validation," IEEE Transactions on Power
Systems, vol. 21, no. 3, pp. 1276 - 1283, 2006.
[116] I. R. Navarro, "Dynamic Load Models for Power Systems: Estimation of Time-Varying
Parameters During Normal Operation," Ph.D dissertation, Department of Industrial Electrical Engineering
and Automation, Lund University, Lund, Sweden, 2002.
[118] J. Currie and D. Wilson, "Opti: Lowering the Barrier Between Open Source Optimizers and the
Industrial MATLAB User," in Foundations of Computer-Aided Process Operations, Savannah, Georgia,
2012.
[119] B. Stephen, J. Shaw, J. Berardino, S. Galloway and C. Nwankpa, "Intra-day Error Dependency
Boosting for Short Term Building and LV Network Load Forecast Models," IEEE Transactions on Power
Systems.
144
145
APPENDICES
146
Appendix A. LIST OF NOMENCLATURE
In order of appearance:
PREDICTION INTERVALS
This Appendix presents the additional forecasting case study results from Chapter
3 and Chapter 4. These results are presented here without commentary. Refer to the
Table B.1 MAPE results at 5 minute resolution for all 116 day-ahead forecasts
MLR NN SVM SAM
Time No Bldg Bldg No Bldg Bldg No Bldg Bldg No Bldg Bldg
0:00 4.46 4.34 6.45 4.50 3.80 3.62 5.64 5.42
0:05 4.14 3.99 6.62 5.02 3.88 3.81 5.67 5.34
0:10 4.35 4.22 7.00 5.29 4.13 4.14 5.92 5.35
0:15 4.51 4.36 7.21 5.51 4.39 4.39 6.13 5.44
0:20 4.94 4.86 7.84 6.28 5.09 5.13 6.77 6.25
0:25 5.09 5.04 7.93 6.67 5.35 5.38 7.12 6.43
0:30 5.34 5.35 8.28 7.08 5.87 6.02 7.46 6.70
0:35 5.54 5.53 8.35 7.11 5.96 6.10 7.78 6.88
0:40 5.79 5.76 8.84 7.48 6.57 6.69 7.85 7.02
0:45 6.13 6.16 9.31 7.97 7.04 7.18 8.19 7.25
0:50 5.72 5.84 9.47 7.76 6.89 7.05 7.90 6.79
0:55 5.89 5.96 9.42 7.79 6.94 7.14 7.98 6.85
1:00 5.54 5.48 8.44 5.81 7.08 7.13 7.59 6.70
1:05 5.45 5.40 8.37 5.80 6.95 6.98 7.61 6.74
1:10 5.38 5.32 7.96 5.84 6.80 6.80 7.53 6.74
1:15 5.25 5.18 8.00 5.74 6.69 6.66 7.29 6.49
1:20 5.39 5.29 7.75 5.74 6.74 6.71 7.41 6.62
1:25 5.23 5.19 7.78 5.74 6.68 6.63 7.38 6.60
1:30 5.32 5.25 7.74 5.58 6.74 6.71 7.49 6.60
1:35 5.24 5.21 7.75 5.54 6.73 6.70 7.37 6.47
1:40 5.16 5.16 7.68 5.28 6.78 6.66 7.21 6.32
1:45 5.60 5.57 7.56 5.51 7.01 6.92 7.70 6.88
1:50 5.36 5.33 7.40 5.33 6.79 6.73 7.41 6.57
1:55 5.52 5.62 7.80 5.51 7.39 7.12 8.09 6.70
2:00 10.40 7.66 8.46 10.10 6.40 5.65 9.85 9.41
2:05 8.49 9.87 8.33 10.66 6.48 5.98 11.37 8.48
2:10 6.93 7.99 8.01 9.32 6.10 5.44 10.42 8.05
2:15 5.87 6.59 7.77 7.89 5.76 5.38 9.89 8.04
2:20 6.20 6.46 7.77 7.09 5.53 5.29 9.61 8.07
2:25 7.37 7.55 8.02 6.80 6.03 5.93 10.17 9.05
2:30 8.76 8.81 8.57 7.02 6.60 6.59 10.72 9.56
2:35 9.24 9.18 8.63 7.20 6.75 6.70 11.16 10.12
151
MLR NN SVM SAM
Time No Bldg Bldg No Bldg Bldg No Bldg Bldg No Bldg Bldg
2:40 9.40 9.49 9.35 7.19 6.71 6.63 11.34 10.45
2:45 9.73 9.64 9.31 7.29 6.93 7.02 11.62 10.41
2:50 9.98 10.16 9.66 7.62 7.21 7.26 11.90 10.69
2:55 9.98 9.90 9.47 7.25 7.15 7.01 11.53 10.39
3:00 8.22 8.41 12.92 7.99 6.95 7.27 11.52 10.00
3:05 8.09 8.43 12.53 7.89 6.69 7.01 11.92 10.10
3:10 8.33 8.45 12.93 8.39 7.14 7.45 11.94 10.46
3:15 8.90 8.91 13.10 8.21 7.07 7.34 12.10 10.50
3:20 8.72 8.76 12.74 8.05 6.97 7.18 11.64 10.27
3:25 8.78 8.70 12.51 8.13 6.99 7.24 11.79 10.36
3:30 8.50 8.71 12.25 7.92 6.57 6.79 11.78 10.32
3:35 8.58 8.60 12.85 8.29 6.86 7.18 11.22 10.17
3:40 8.16 8.30 12.19 7.93 6.48 6.76 11.14 9.91
3:45 8.00 8.33 11.83 7.60 6.15 6.49 11.24 9.78
3:50 8.15 8.44 12.23 8.24 6.47 6.93 11.06 9.96
3:55 8.06 8.67 11.87 8.25 6.24 6.65 10.91 10.05
4:00 8.38 7.81 10.36 6.20 5.87 5.93 10.59 9.54
4:05 8.19 7.77 9.73 5.97 5.62 5.67 10.69 9.38
4:10 8.24 8.02 9.73 6.17 5.87 5.88 10.80 9.52
4:15 8.32 7.98 9.94 6.46 6.03 6.03 10.45 9.14
4:20 8.14 7.89 9.85 6.56 6.02 6.05 10.35 8.87
4:25 8.09 7.95 9.62 6.60 5.98 6.18 10.50 8.78
4:30 8.33 8.10 10.25 6.83 6.27 6.46 10.61 8.99
4:35 8.61 8.54 10.46 7.16 6.32 6.55 11.00 9.40
4:40 8.37 8.72 10.63 7.30 6.57 6.97 10.62 9.36
4:45 8.02 8.83 10.64 7.51 6.17 6.73 10.43 9.49
4:50 7.75 8.72 10.63 7.56 6.22 6.81 10.30 9.90
4:55 7.75 8.75 10.97 7.90 6.41 6.99 10.41 10.17
5:00 13.51 11.53 13.30 7.89 11.77 12.03 11.70 12.91
5:05 12.43 10.96 12.42 7.64 10.96 11.55 10.91 12.04
5:10 11.73 10.64 11.59 7.52 10.26 11.06 10.54 11.37
5:15 11.03 10.30 10.95 7.56 9.90 10.58 10.33 11.24
5:20 10.42 10.04 10.59 7.32 9.07 9.81 10.26 10.90
5:25 10.05 9.94 10.25 7.49 8.73 9.41 9.83 10.72
5:30 9.53 9.65 9.97 7.52 8.10 8.91 9.80 10.24
5:35 8.92 9.18 9.22 7.77 7.77 8.43 9.89 9.87
5:40 8.76 8.98 9.21 7.75 7.41 7.99 9.68 9.49
5:45 8.28 8.50 8.92 7.51 6.92 7.36 9.20 8.86
5:50 7.88 8.04 8.88 7.97 6.62 6.80 9.08 8.23
5:55 7.63 7.59 9.19 8.75 6.44 6.21 8.87 7.88
6:00 8.56 8.77 9.13 7.74 9.07 9.34 8.84 8.32
152
MLR NN SVM SAM
Time No Bldg Bldg No Bldg Bldg No Bldg Bldg No Bldg Bldg
6:05 8.01 8.14 8.27 7.09 7.98 8.20 8.25 8.00
6:10 7.66 7.75 7.29 6.75 7.14 7.26 7.75 7.58
6:15 6.70 6.75 6.27 5.92 5.93 6.02 7.11 6.77
6:20 6.42 6.45 6.21 5.90 5.43 5.42 7.03 6.39
6:25 6.33 6.38 6.50 6.08 5.45 5.34 7.04 6.30
6:30 6.06 6.11 6.65 6.18 5.28 5.15 7.17 6.36
6:35 6.07 6.13 6.58 6.02 5.37 5.20 7.23 6.44
6:40 5.77 5.81 6.12 5.65 5.09 4.95 7.00 6.09
6:45 5.70 5.71 6.18 5.68 5.01 4.85 6.83 5.99
6:50 5.66 5.65 5.95 5.36 4.84 4.74 6.54 5.99
6:55 5.61 5.66 6.04 5.36 4.84 4.72 6.35 5.73
7:00 5.27 5.26 6.33 5.25 5.20 5.18 5.96 5.47
7:05 5.35 5.34 6.41 5.48 5.51 5.55 6.33 5.69
7:10 5.45 5.40 6.50 5.52 5.58 5.64 6.29 5.89
7:15 5.45 5.51 6.39 5.43 5.59 5.62 6.18 5.83
7:20 5.08 5.12 6.12 5.17 5.15 5.17 5.94 5.55
7:25 5.10 5.11 6.23 5.15 5.05 5.15 5.94 5.65
7:30 5.38 5.46 6.32 5.38 5.20 5.30 6.30 5.93
7:35 5.07 5.12 6.36 4.95 4.77 4.95 5.79 5.64
7:40 5.35 5.41 6.58 5.30 4.96 5.06 5.90 5.91
7:45 5.25 5.30 6.72 5.05 4.84 5.00 5.95 5.90
7:50 5.20 5.26 6.45 4.95 4.62 4.72 6.00 5.81
7:55 5.36 5.45 6.78 5.14 4.64 4.71 6.04 5.89
8:00 5.32 5.42 7.49 5.59 7.20 7.05 5.90 5.83
8:05 5.66 5.73 7.66 5.99 7.30 7.07 6.39 6.36
8:10 5.24 5.36 7.30 5.56 6.66 6.48 6.34 6.07
8:15 5.32 5.33 7.31 5.70 6.82 6.62 6.38 6.16
8:20 5.43 5.44 7.45 5.65 6.74 6.57 6.04 6.20
8:25 5.15 5.10 6.79 5.44 6.18 6.02 5.74 5.84
8:30 5.27 5.28 6.52 5.30 5.79 5.65 5.83 5.92
8:35 5.14 5.14 6.17 5.26 5.58 5.44 5.77 5.76
8:40 5.57 5.48 6.55 5.37 5.89 5.78 6.15 6.17
8:45 5.92 5.75 6.90 5.58 5.91 5.83 6.44 6.56
8:50 5.69 5.62 6.66 5.37 5.73 5.65 6.61 6.57
8:55 5.86 5.82 6.43 5.49 5.81 5.66 6.71 6.64
9:00 5.97 5.87 6.35 5.82 6.74 6.61 6.53 6.36
9:05 5.78 5.65 6.40 5.65 6.51 6.42 6.28 6.20
9:10 5.46 5.32 6.06 5.36 6.08 6.00 6.07 5.97
9:15 5.27 5.16 5.78 5.25 5.92 5.88 6.04 5.88
9:20 5.07 5.02 5.82 5.31 5.67 5.53 5.65 5.49
9:25 5.17 5.15 6.06 5.26 5.73 5.64 5.79 5.68
153
MLR NN SVM SAM
Time No Bldg Bldg No Bldg Bldg No Bldg Bldg No Bldg Bldg
9:30 4.97 4.95 5.98 5.25 5.49 5.37 5.78 5.71
9:35 4.99 5.01 6.11 5.31 5.56 5.49 5.79 5.87
9:40 5.11 5.09 6.06 5.41 5.36 5.26 5.80 5.94
9:45 5.04 5.00 5.88 5.30 5.13 5.04 5.88 5.84
9:50 4.99 5.02 5.59 5.49 4.97 4.83 5.77 5.64
9:55 5.13 5.05 5.39 5.34 4.67 4.55 5.88 5.79
10:00 5.37 5.29 6.49 5.75 6.44 6.14 5.43 5.58
10:05 5.30 5.18 6.19 5.46 6.08 5.83 5.45 5.44
10:10 5.17 5.10 6.05 5.54 5.95 5.66 5.40 5.53
10:15 4.95 4.86 6.11 5.19 5.65 5.35 5.38 5.43
10:20 5.03 4.93 6.19 5.20 5.55 5.33 5.41 5.42
10:25 4.85 4.76 6.06 5.06 5.41 5.14 5.27 5.22
10:30 4.83 4.74 5.87 4.88 5.18 4.90 5.34 5.20
10:35 4.72 4.60 5.76 4.95 5.15 4.86 5.24 5.31
10:40 4.70 4.59 5.65 5.09 5.04 4.79 5.37 5.34
10:45 4.79 4.68 5.80 5.16 5.05 4.81 5.40 5.38
10:50 5.01 4.90 5.91 5.39 5.23 5.02 5.52 5.52
10:55 5.22 5.09 5.82 5.32 5.16 4.93 5.82 5.74
11:00 5.43 5.34 6.82 5.11 5.66 5.71 5.97 5.72
11:05 5.39 5.26 6.77 5.06 5.46 5.47 5.85 5.61
11:10 5.10 4.98 6.40 4.90 5.09 5.12 5.81 5.56
11:15 4.97 4.86 6.44 4.79 4.97 4.97 5.54 5.30
11:20 5.02 4.97 6.39 5.03 4.87 4.89 5.71 5.48
11:25 4.95 4.90 6.42 5.01 4.69 4.70 5.64 5.47
11:30 4.89 4.84 6.44 4.86 4.76 4.74 5.58 5.38
11:35 4.84 4.82 6.09 4.95 4.69 4.62 5.68 5.37
11:40 4.77 4.72 6.42 4.94 4.78 4.71 5.58 5.26
11:45 4.78 4.68 6.28 4.76 4.56 4.50 5.57 5.28
11:50 4.53 4.38 5.99 4.68 4.33 4.21 5.29 4.99
11:55 4.45 4.31 6.02 4.57 4.56 4.42 5.19 4.92
12:00 4.50 4.41 5.82 4.82 4.35 4.54 5.15 5.00
12:05 4.41 4.31 5.82 4.83 4.46 4.69 4.99 4.89
12:10 4.51 4.45 5.62 4.75 4.49 4.68 5.07 4.88
12:15 4.49 4.40 5.80 4.82 4.52 4.75 5.14 4.95
12:20 4.45 4.34 5.78 4.77 4.46 4.60 4.95 4.79
12:25 4.38 4.31 5.57 4.88 4.35 4.49 5.07 4.94
12:30 4.25 4.15 5.48 4.66 4.49 4.66 4.87 4.75
12:35 4.42 4.29 5.65 4.87 4.39 4.59 4.89 4.85
12:40 4.21 4.06 5.55 4.79 4.30 4.45 4.90 4.83
12:45 4.46 4.41 5.52 5.00 4.40 4.58 5.20 4.95
12:50 4.37 4.34 5.43 4.76 4.35 4.55 5.38 4.81
154
MLR NN SVM SAM
Time No Bldg Bldg No Bldg Bldg No Bldg Bldg No Bldg Bldg
12:55 4.46 4.45 5.67 4.73 4.54 4.68 5.11 4.71
13:00 4.31 4.33 6.12 5.25 4.13 4.20 5.03 4.76
13:05 4.43 4.46 6.29 5.20 4.11 4.07 5.04 4.76
13:10 4.45 4.50 6.41 5.02 4.07 4.04 5.00 4.72
13:15 4.46 4.47 6.35 5.04 4.17 4.10 4.97 4.74
13:20 4.44 4.42 6.22 5.10 4.13 4.00 5.13 4.80
13:25 4.44 4.44 6.12 4.88 3.89 3.78 5.25 4.70
13:30 4.49 4.49 6.25 5.24 4.12 3.98 5.09 4.64
13:35 4.55 4.57 6.16 5.16 4.16 4.04 5.09 4.66
13:40 4.42 4.36 6.12 5.12 4.12 4.01 5.01 4.64
13:45 4.35 4.33 6.23 5.06 4.05 3.96 5.00 4.60
13:50 4.43 4.33 6.31 5.10 4.02 3.99 4.93 4.63
13:55 4.44 4.34 6.29 5.21 3.98 3.89 5.09 4.71
14:00 4.66 4.59 5.79 4.60 4.02 3.86 5.18 4.86
14:05 4.79 4.64 5.72 4.70 4.01 3.80 5.21 4.96
14:10 4.65 4.51 5.46 4.65 4.02 3.95 5.21 4.90
14:15 4.91 4.74 5.63 4.67 4.05 3.93 5.40 5.03
14:20 4.94 4.79 5.66 4.64 4.09 3.90 5.32 5.14
14:25 4.84 4.59 5.46 4.50 3.91 3.72 5.30 5.04
14:30 4.90 4.69 5.53 4.59 3.81 3.63 5.25 4.97
14:35 4.83 4.72 5.48 4.69 3.87 3.71 5.31 4.83
14:40 4.89 4.76 5.59 4.99 4.09 3.95 5.28 5.04
14:45 4.91 4.68 5.58 4.81 4.18 3.96 5.18 5.08
14:50 5.14 4.86 5.47 4.66 3.89 3.77 5.35 5.24
14:55 5.20 4.95 5.69 4.91 3.98 3.75 5.56 5.31
15:00 4.84 4.71 5.67 4.99 4.27 3.96 5.23 5.06
15:05 4.86 4.65 5.65 5.07 4.08 3.80 5.26 4.99
15:10 4.93 4.67 5.74 5.13 4.12 3.93 5.31 4.96
15:15 5.07 4.80 5.65 5.07 4.06 3.81 5.40 5.07
15:20 5.08 4.88 5.63 5.22 4.30 3.94 5.55 5.24
15:25 5.22 5.04 5.60 5.21 4.37 4.05 5.57 5.38
15:30 5.26 5.07 5.83 5.10 4.19 4.00 5.52 5.21
15:35 5.12 4.96 5.91 5.06 4.54 4.24 5.44 5.26
15:40 5.18 5.02 5.81 5.06 4.23 3.93 5.48 5.17
15:45 5.14 4.99 5.87 5.07 4.26 3.97 5.50 5.22
15:50 5.05 4.80 5.93 4.92 4.10 3.86 5.35 5.04
15:55 4.92 4.68 5.90 4.76 3.86 3.72 5.34 4.97
16:00 4.85 4.69 5.59 5.57 4.10 3.78 5.34 4.87
16:05 5.01 4.80 5.52 5.70 4.10 3.76 5.47 4.95
16:10 5.05 4.86 5.44 5.70 4.27 3.92 5.58 4.99
16:15 4.80 4.64 5.41 5.57 4.22 3.82 5.57 4.91
155
MLR NN SVM SAM
Time No Bldg Bldg No Bldg Bldg No Bldg Bldg No Bldg Bldg
16:20 4.81 4.72 5.57 5.57 4.17 3.81 5.58 4.88
16:25 4.77 4.74 5.68 5.61 4.28 3.95 5.60 4.89
16:30 4.83 4.72 5.83 5.57 4.37 3.96 5.65 5.08
16:35 4.91 4.82 5.88 5.89 4.40 4.02 5.60 5.23
16:40 4.96 4.93 5.96 5.87 4.37 4.00 5.75 5.08
16:45 4.94 4.88 5.79 5.62 4.30 3.96 5.69 5.05
16:50 4.88 4.75 5.77 5.55 4.07 3.70 5.66 5.04
16:55 4.89 4.78 5.67 5.42 4.14 3.88 5.48 4.92
17:00 5.03 4.89 4.93 5.77 4.41 3.71 5.42 4.92
17:05 4.64 4.45 4.85 5.58 4.22 3.51 5.34 4.75
17:10 4.75 4.55 4.90 5.43 4.06 3.42 5.23 4.76
17:15 4.70 4.60 4.93 5.43 4.04 3.41 5.33 4.73
17:20 4.69 4.59 5.06 5.17 4.21 3.57 5.20 4.65
17:25 4.60 4.53 4.92 5.18 3.99 3.21 5.30 4.50
17:30 4.49 4.52 4.96 5.01 3.93 3.29 5.26 4.52
17:35 4.41 4.38 5.00 4.88 3.80 3.06 5.14 4.49
17:40 4.51 4.54 5.36 4.92 3.91 3.29 5.34 4.65
17:45 4.36 4.42 5.22 4.83 3.88 3.26 5.18 4.53
17:50 4.39 4.57 5.27 4.52 3.79 3.21 5.22 4.51
17:55 4.18 4.31 5.29 4.56 3.78 3.31 4.93 4.37
18:00 4.55 4.22 5.01 7.36 4.56 3.93 5.32 4.74
18:05 4.59 4.27 5.02 7.26 4.65 4.07 5.21 4.62
18:10 4.37 4.07 4.89 7.32 4.57 4.03 4.99 4.46
18:15 4.49 4.23 4.70 7.12 4.34 3.87 5.03 4.53
18:20 4.51 4.29 4.85 7.29 4.29 3.83 5.30 4.62
18:25 4.45 4.16 4.36 7.04 4.07 3.52 5.18 4.62
18:30 4.57 4.42 4.67 6.99 4.21 3.74 5.49 4.92
18:35 4.32 4.17 4.66 6.91 3.94 3.56 5.47 4.81
18:40 4.37 4.37 4.90 6.66 3.96 3.72 5.59 4.79
18:45 4.10 4.00 4.64 6.60 3.65 3.35 5.31 4.47
18:50 4.28 4.16 4.96 6.65 3.75 3.43 5.40 4.56
18:55 4.31 4.30 5.26 6.53 3.79 3.59 5.56 4.61
19:00 4.60 4.28 4.47 6.87 4.18 3.84 5.54 4.87
19:05 4.65 4.39 4.79 6.83 4.22 3.90 5.67 5.06
19:10 4.58 4.29 4.62 7.02 4.06 3.80 5.63 4.86
19:15 4.92 4.66 4.80 6.75 4.26 3.93 5.98 5.24
19:20 4.79 4.51 4.73 6.68 4.08 3.82 5.78 5.14
19:25 4.88 4.64 4.95 6.79 4.30 4.01 6.05 5.21
19:30 4.83 4.65 5.03 6.77 4.23 3.96 6.02 5.24
19:35 4.80 4.67 5.04 6.55 4.17 3.98 5.90 5.11
19:40 4.69 4.54 4.97 6.58 4.14 3.90 5.67 5.02
156
MLR NN SVM SAM
Time No Bldg Bldg No Bldg Bldg No Bldg Bldg No Bldg Bldg
19:45 4.58 4.48 4.86 6.42 4.16 3.95 5.58 4.92
19:50 4.46 4.32 4.79 6.29 3.93 3.71 5.58 4.79
19:55 4.40 4.33 4.87 6.05 3.81 3.67 5.62 4.91
20:00 4.60 4.53 4.86 5.06 3.89 3.73 5.60 5.04
20:05 4.35 4.25 4.58 5.01 3.85 3.58 5.44 4.83
20:10 4.49 4.48 4.70 5.00 3.72 3.55 5.68 4.86
20:15 4.60 4.57 4.79 4.87 3.62 3.45 5.82 4.88
20:20 4.50 4.49 4.86 4.82 3.70 3.58 5.73 4.79
20:25 4.62 4.64 5.17 4.78 3.95 3.85 5.66 4.83
20:30 4.69 4.74 4.96 4.61 4.06 3.99 5.97 4.91
20:35 4.50 4.60 5.06 4.66 4.01 4.03 5.85 4.88
20:40 4.51 4.66 5.10 4.51 3.83 3.78 5.76 4.76
20:45 4.64 4.80 5.40 4.65 4.24 4.19 5.96 4.82
20:50 4.74 4.98 5.58 4.70 4.41 4.35 6.14 4.82
20:55 4.56 4.69 5.54 4.57 4.27 4.11 5.95 4.83
21:00 4.78 4.95 6.44 5.41 4.12 4.17 6.19 5.09
21:05 4.98 5.15 6.75 5.53 4.20 4.23 6.36 5.24
21:10 5.10 5.24 6.71 5.54 4.40 4.48 6.44 5.40
21:15 4.92 5.10 6.79 5.08 4.05 4.17 6.43 5.31
21:20 4.97 5.19 6.96 5.16 4.15 4.26 6.40 5.25
21:25 4.86 5.19 7.85 4.98 4.32 4.41 6.46 5.27
21:30 4.74 5.06 7.60 5.24 4.19 4.33 6.34 5.26
21:35 4.90 5.32 8.05 4.85 4.44 4.52 6.55 5.11
21:40 5.16 5.60 8.38 4.92 4.82 4.85 6.72 5.26
21:45 4.99 5.46 8.42 5.05 4.66 4.68 6.62 5.23
21:50 5.17 5.58 8.75 5.12 4.82 4.84 6.78 5.43
21:55 5.21 5.71 8.81 5.31 4.77 4.79 6.83 5.40
22:00 5.07 5.25 7.48 5.11 4.30 4.40 6.49 5.17
22:05 4.97 5.24 7.71 5.04 4.15 4.23 6.51 5.27
22:10 5.09 5.32 7.62 5.23 4.21 4.28 6.58 5.21
22:15 5.22 5.50 8.03 5.20 4.24 4.34 6.77 5.38
22:20 4.99 5.20 7.79 4.84 4.17 4.24 6.56 5.37
22:25 5.00 5.25 7.91 5.08 4.25 4.41 6.65 5.41
22:30 5.20 5.51 8.38 5.10 4.44 4.57 6.87 5.56
22:35 5.06 5.35 8.28 5.07 4.47 4.50 6.51 5.40
22:40 5.04 5.34 8.83 5.26 4.57 4.65 6.63 5.68
22:45 5.15 5.46 8.77 5.27 4.66 4.78 6.68 5.62
22:50 5.19 5.49 9.03 5.50 4.60 4.78 6.54 5.69
22:55 5.10 5.46 9.18 5.68 4.53 4.68 6.69 5.71
23:00 4.85 4.94 5.74 5.65 4.59 4.81 6.39 5.57
23:05 4.84 4.97 5.85 5.32 4.54 4.74 6.50 5.55
157
MLR NN SVM SAM
Time No Bldg Bldg No Bldg Bldg No Bldg Bldg No Bldg Bldg
23:10 5.03 5.07 6.14 5.75 4.70 5.00 6.49 5.83
23:15 5.01 5.05 6.03 5.53 4.53 4.76 6.56 5.83
23:20 5.15 5.22 6.25 5.63 4.72 4.99 6.85 5.93
23:25 4.86 4.96 6.16 5.33 4.69 4.83 6.53 5.66
23:30 4.80 4.95 6.22 5.54 4.69 4.94 6.54 5.60
23:35 5.00 5.11 6.68 5.62 4.75 5.13 6.49 5.87
23:40 4.97 5.09 6.38 5.60 4.67 4.85 6.41 5.69
23:45 5.11 5.27 6.45 5.54 4.86 5.05 6.73 5.96
23:50 5.17 5.38 6.47 5.28 4.87 5.03 6.87 5.88
23:55 5.04 5.20 6.89 5.44 4.81 5.11 6.67 5.94
158
Hourly empirical error distributions for the MLR models
Figure B.1 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (00:00-01:00)
Figure B.2 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (01:00-02:00)
Figure B.3 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (02:00-03:00)
159
Figure B.4 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (03:00-04:00)
Figure B.5 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (04:00-05:00)
Figure B.6 Empirical error distribution (L) No building measurements (R) With building
measurements (05:00-06:00)
160
Figure B.7 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (06:00-07:00)
Figure B.8 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (07:00-08:00)
Figure B.9 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (08:00-09:00)
161
Figure B.10 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (09:00-10:00)
Figure B.11 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (10:00-11:00)
Figure B.12 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (11:00-12:00)
162
Figure B.13 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (12:00-13:00)
Figure B.14 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (13:00-14:00)
Figure B.15 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (14:00-15:00)
163
Figure B.16 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (15:00-16:00)
Figure B.17 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (16:00-17:00)
Figure B.18 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (17:00-18:00)
164
Figure B.19 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (18:00-19:00)
Figure B.20 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (19:00-20:00)
Figure B.21 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (20:00-21:00)
165
Figure B.22 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (21:00-22:00)
Figure B.23 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (22:00-23:00)
Figure B.24 MLR Empirical error distribution (L) No building measurements (R) With building
measurements (23:00-24:00)
166
Hourly empirical error distributions for the NN models
Figure B.25 NN Empirical error distribution (L) No building measurements (R) With building
measurements (00:00-01:00)
Figure B.26 NN Empirical error distribution (L) No building measurements (R) With building
measurements (01:00-02:00)
Figure B.27 NN Empirical error distribution (L) No building measurements (R) With building
measurements (02:00-03:00)
167
Figure B.28 NN Empirical error distribution (L) No building measurements (R) With building
measurements (03:00-04:00)
Figure B.29 NN Empirical error distribution (L) No building measurements (R) With building
measurements (04:00-05:00)
Figure B.30 NN Empirical error distribution (L) No building measurements (R) With building
measurements (05:00-06:00)
168
Figure B.31 NN Empirical error distribution (L) No building measurements (R) With building
measurements (06:00-07:00)
Figure B.32 NN Empirical error distribution (L) No building measurements (R) With building
measurements (07:00-08:00)
Figure B.33 NN Empirical error distribution (L) No building measurements (R) With building
measurements (08:00-09:00)
169
Figure B.34 NN Empirical error distribution (L) No building measurements (R) With building
measurements (09:00-10:00)
Figure B.35 NN Empirical error distribution (L) No building measurements (R) With building
measurements (10:00-11:00)
Figure B.36 NN Empirical error distribution (L) No building measurements (R) With building
measurements (11:00-12:00)
170
Figure B.37 NN Empirical error distribution (L) No building measurements (R) With building
measurements (12:00-13:00)
Figure B.38 NN Empirical error distribution (L) No building measurements (R) With building
measurements (13:00-14:00)
Figure B.39 NN Empirical error distribution (L) No building measurements (R) With building
measurements (14:00-15:00)
171
Figure B.40 NN Empirical error distribution (L) No building measurements (R) With building
measurements (15:00-16:00)
Figure B.41 NN Empirical error distribution (L) No building measurements (R) With building
measurements (16:00-17:00)
Figure B.42 NN Empirical error distribution (L) No building measurements (R) With building
measurements (17:00-18:00)
172
Figure B.43 NN Empirical error distribution (L) No building measurements (R) With building
measurements (18:00-19:00)
Figure B.44 NN Empirical error distribution (L) No building measurements (R) With building
measurements (19:00-20:00)
Figure B.45 NN Empirical error distribution (L) No building measurements (R) With building
measurements (20:00-21:00)
173
Figure B.46 NN Empirical error distribution (L) No building measurements (R) With building
measurements (21:00-22:00)
Figure B.47 NN Empirical error distribution (L) No building measurements (R) With building
measurements (22:00-23:00)
Figure B.48 NN Empirical error distribution (L) No building measurements (R) With building
measurements (23:00-24:00)
174
Hourly empirical error distributions for the SVM models
Figure B.49 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (00:00-01:00)
Figure B.50 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (01:00-02:00)
Figure B.51 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (02:00-03:00)
175
Figure B.52 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (03:00-04:00)
Figure B.53 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (04:00-05:00)
Figure B.54 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (05:00-06:00)
176
Figure B.55 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (06:00-07:00)
Figure B.56 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (07:00-08:00)
Figure B.57 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (08:00-09:00)
177
Figure B.58 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (09:00-10:00)
Figure B.59 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (10:00-11:00)
Figure B.60 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (11:00-12:00)
178
Figure B.61 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (12:00-13:00)
Figure B.62 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (13:00-14:00)
Figure B.63 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (14:00-15:00)
179
Figure B.64 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (15:00-16:00)
Figure B.65 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (16:00-17:00)
Figure B.66 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (17:00-18:00)
180
Figure B.67 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (18:00-19:00)
Figure B.68 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (19:00-20:00)
Figure B.69 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (20:00-21:00)
181
Figure B.70 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (21:00-22:00)
Figure B.71 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (22:00-23:00)
Figure B.72 SVM Empirical error distribution (L) No building measurements (R) With building
measurements (23:00-24:00)
182
Hourly empirical error distributions for the SAM models
Figure B.73 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (00:00-01:00)
Figure B.74 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (01:00-02:00)
Figure B.75 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (02:00-03:00)
183
Figure B.76 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (03:00-04:00)
Figure B.77 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (04:00-05:00)
Figure B.78 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (05:00-06:00)
184
Figure B.79 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (06:00-07:00)
Figure B.80 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (07:00-08:00)
Figure B.81 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (08:00-09:00)
185
Figure B.82 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (09:00-10:00)
Figure B.83 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (10:00-11:00)
Figure B.84 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (11:00-12:00)
186
Figure B.85 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (12:00-13:00)
Figure B.86 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (13:00-14:00)
Figure B.87 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (14:00-15:00)
187
Figure B.88 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (15:00-16:00)
Figure B.89 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (16:00-17:00)
Figure B.90 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (17:00-18:00)
188
Figure B.91 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (18:00-19:00)
Figure B.92 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (19:00-20:00)
Figure B.93 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (20:00-21:00)
189
Figure B.94 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (21:00-22:00)
Figure B.95 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (22:00-23:00)
Figure B.96 SAM Empirical error distribution (L) No building measurements (R) With building
measurements (23:00-24:00)
190
Average hourly skill score for the MLR models
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.97 Average hourly skill score for the MLR Models. Nominal coverage rate = 5%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.98 Average hourly skill score for the MLR Models. Nominal coverage rate = 10%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.99 Average hourly skill score for the MLR Models. Nominal coverage rate = 15%
191
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.100 Average hourly skill score for the MLR Models. Nominal coverage rate = 20%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.101 Average hourly skill score for the MLR Models. Nominal coverage rate = 25%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.102 Average hourly skill score for the MLR Models. Nominal coverage rate = 30%
192
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.103 Average hourly skill score for the MLR Models. Nominal coverage rate = 35%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.104 Average hourly skill score for the MLR Models. Nominal coverage rate = 40%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.105 Average hourly skill score for the MLR Models. Nominal coverage rate = 45%
193
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.106 Average hourly skill score for the MLR Models. Nominal coverage rate = 50%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.107 Average hourly skill score for the MLR Models. Nominal coverage rate = 55%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.108 Average hourly skill score for the MLR Models. Nominal coverage rate = 60%
194
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.109 Average hourly skill score for the MLR Models. Nominal coverage rate = 65%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.110 Average hourly skill score for the MLR Models. Nominal coverage rate = 70%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.111 Average hourly skill score for the MLR Models. Nominal coverage rate = 75%
195
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.112 Average hourly skill score for the MLR Models. Nominal coverage rate = 80%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.113 Average hourly skill score for the MLR Models. Nominal coverage rate = 85%
0
-10
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.114 Average hourly skill score for the MLR Models. Nominal coverage rate = 90%
196
0
-10
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.115 Average hourly skill score for the MLR Models. Nominal coverage rate = 95%
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.116 Average hourly skill score for the NN Models. Nominal coverage rate = 5%
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.117 Average hourly skill score for the NN Models. Nominal coverage rate = 10%
197
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.118 Average hourly skill score for the NN Models. Nominal coverage rate = 15%
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.119 Average hourly skill score for the NN Models. Nominal coverage rate = 20%
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.120 Average hourly skill score for the NN Models. Nominal coverage rate = 25%
198
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.121 Average hourly skill score for the NN Models. Nominal coverage rate = 30%
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.122 Average hourly skill score for the NN Models. Nominal coverage rate = 35%
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.123 Average hourly skill score for the NN Models. Nominal coverage rate = 40%
199
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.124 Average hourly skill score for the NN Models. Nominal coverage rate = 45%
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.125 Average hourly skill score for the NN Models. Nominal coverage rate = 50%
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.126 Average hourly skill score for the NN Models. Nominal coverage rate = 55%
200
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.127 Average hourly skill score for the NN Models. Nominal coverage rate = 60%
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.128 Average hourly skill score for the NN Models. Nominal coverage rate = 65%
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.129 Average hourly skill score for the NN Models. Nominal coverage rate = 70%
201
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.130 Average hourly skill score for the NN Models. Nominal coverage rate = 75%
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.131 Average hourly skill score for the NN Models. Nominal coverage rate = 80%
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.132 Average hourly skill score for the NN Models. Nominal coverage rate = 85%
202
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.133 Average hourly skill score for the NN Models. Nominal coverage rate = 90%
0
-10
-20
-30
Skill Score
-40
-50
-60
-70
-90
0 5 10 15 20 25
Hours
Figure B.134 Average hourly skill score for the NN Models. Nominal coverage rate = 95%
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.135 Average hourly skill score for the SVM Models. Nominal coverage rate = 5%
203
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.136 Average hourly skill score for the SVM Models. Nominal coverage rate = 10%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.137 Average hourly skill score for the SVM Models. Nominal coverage rate = 15%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.138 Average hourly skill score for the SVM Models. Nominal coverage rate = 20%
204
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.139 Average hourly skill score for the SVM Models. Nominal coverage rate = 25%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.140 Average hourly skill score for the SVM Models. Nominal coverage rate = 30%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.141 Average hourly skill score for the SVM Models. Nominal coverage rate = 35%
205
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.142 Average hourly skill score for the SVM Models. Nominal coverage rate = 40%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.143 Average hourly skill score for the SVM Models. Nominal coverage rate = 45%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.144 Average hourly skill score for the SVM Models. Nominal coverage rate = 50%
206
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.145 Average hourly skill score for the SVM Models. Nominal coverage rate = 55%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.146 Average hourly skill score for the SVM Models. Nominal coverage rate = 60%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.147 Average hourly skill score for the SVM Models. Nominal coverage rate = 65%
207
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.148 Average hourly skill score for the SVM Models. Nominal coverage rate = 70%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.149 Average hourly skill score for the SVM Models. Nominal coverage rate = 75%
0
-10
-20
W/O Bldg Meas
WITH Bldg Meas
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.150 Average hourly skill score for the SVM Models. Nominal coverage rate = 80%
208
0
-10
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.151 Average hourly skill score for the SVM Models. Nominal coverage rate = 85%
0
-10
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.152 Average hourly skill score for the SVM Models. Nominal coverage rate = 90%
0
-10
W/O Bldg Meas
WITH Bldg Meas
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.153 Average hourly skill score for the SVM Models. Nominal coverage rate = 95%
209
Average hourly skill score for the SAM models
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.154 Average hourly skill score for the SAM Models. Nominal coverage rate = 5%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.155 Average hourly skill score for the SAM Models. Nominal coverage rate = 10%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.156 Average hourly skill score for the SAM Models. Nominal coverage rate = 15%
210
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.157 Average hourly skill score for the SAM Models. Nominal coverage rate = 20%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.158 Average hourly skill score for the SAM Models. Nominal coverage rate = 25%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.159 Average hourly skill score for the SAM Models. Nominal coverage rate = 30%
211
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.160 Average hourly skill score for the SAM Models. Nominal coverage rate = 35%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.161 Average hourly skill score for the SAM Models. Nominal coverage rate = 40%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.162 Average hourly skill score for the SAM Models. Nominal coverage rate = 45%
212
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.163 Average hourly skill score for the SAM Models. Nominal coverage rate = 50%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.164 Average hourly skill score for the SAM Models. Nominal coverage rate = 55%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.165 Average hourly skill score for the SAM Models. Nominal coverage rate = 60%
213
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.166 Average hourly skill score for the SAM Models. Nominal coverage rate = 65%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.167 Average hourly skill score for the SAM Models. Nominal coverage rate = 70%
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.168 Average hourly skill score for the SAM Models. Nominal coverage rate = 75%
214
0
W/O Bldg Meas
WITH Bldg Meas
-10
-20
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.169 Average hourly skill score for the SAM Models. Nominal coverage rate = 80%
0
-10
-20
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.170 Average hourly skill score for the SAM Models. Nominal coverage rate = 85%
0
-10
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.171 Average hourly skill score for the SAM Models. Nominal coverage rate = 90%
215
0
-10
-30
Skill Score
-40
-50
-60
-70
-80
-90
0 5 10 15 20 25
Hours
Figure B.172 Average hourly skill score for the SAM Models. Nominal coverage rate = 95%
216
VITA
SELECT PUBLICATIONS:
[J1]. Jonathan Berardino, Chika Nwankpa, “Enhanced Demand Forecasting Through
Inclusion of Building-Specific Measurements”, Submitted to IEEE Transactions on
Power Systems
[J2]. Jonathan Berardino and Chika Nwankpa, "Demand Scheduling for Multi-
Building Campuses Considering Building Dynamic Behavior", (In preparation)
[J3]. Bruce Stephen, Jennifer Shaw, Jonathan Berardino, Stuart Galloway, Chika
Nwankpa, “Intra-day Error Dependency Boosting for Short Term Building and LV
Network Load Forecast Models” Submitted to IEEE Transactions on Smart Grid
[C1]. Jonathan Berardino and Chika Nwankpa, "Statistics of Building-Specific Load
Forecasting Models", Proceedings of the North American Power Symposium 2013,
September 2013, Manhattan, KS
[C2]. Jonathan Berardino and Chika Nwankpa, " Inclusion of Temporal Effects in
Forecasting Building Electrical Loads for Demand Resource Planning", Proceedings of
the IEEE PowerTech 2013 Conference, June 2013, Grenoble, France
[C3]. Jonathan Berardino and Chika Nwankpa, "Economic Demand Dispatch of
Controllable Building Electrical Loads Incorporating Delayed Response Times",
Innovative Smart Grid Technologies (ISGT), 2013, Washington DC
[C4]. Jonathan Berardino and Chika Nwankpa, "Interval-Specific Building Load
Forecasting Models for Demand Resource Planning", Proceedings of the IEEE
PowerCon 2012 Conference, October 2012, Auckland, New Zealand
[C5]. Jonathan Berardino, Mohammed Muthalib, Chika Nwankpa, “Network
Constrained Economic Demand Dispatch of Controllable Building Electric Loads”,
Innovative Smart Grid Technologies (ISGT), 2012, Washington DC
[C6]. Jonathan Berardino, Chika Nwankpa, Karen Miu, “Economic Demand Dispatch
of Controllable Building Electric Loads for Demand Response”, Proceedings of the IEEE
PowerTech 2011 Conference, June 2011, Trondheim, Norway
[C7] Jonathan Berardino and Chika Nwankpa, “Dynamic Load Modeling of an
HVAC Chiller for Demand Response Applications”, IEEE International Conference on
Smart Grid Communications, 2010, Gaithersburg, MD