CP Manual v6.6 - Part 2
CP Manual v6.6 - Part 2
v6.6 1
Module Topics
Continuous Cooling
The X Factor
2
Hidden Risk with High Density
3
Temperature rise example, no inertia
4
Actual recorded rise
C/ S
5
Actual recorded rise
6
Thermal Storage
• To keep Continuous Cooling through changeover between mains power
and generator power you can either provide UPS power to the aircon
system or use thermal storage
• This is not to be confused with the (similar) thermal storage used in e.g.
district cooling schemes – the cooling for Data Centre must always be
available, so the storage is not used day-to-day but only in emergency
7
Module Topics
Continuous Cooling
The X Factor
8
Why are we “Cooling”?
9
Trends Influencing Data Center Cooling Design
New Approach
• Internet Search, Social Media & Web Services
– Different Availability, Reliability, and Fault-Tolerance Criteria issues allows
testing and deploying “unconventional” cooling strategies
• Moves Physical Reliability of a site to IT architecture for overall system
availability (shift loads and backups to other sites)
• Internet Search, Social Media & Web Services are the Missionaries:
Google - Yahoo – Facebook – Microsoft - Amazon
• Commercial – Leading Edge DC designs
• However, even some conventional DCs (Enterprise, Colo & Hosting)
are adopting some of the technologies and practices
– NetApp – Free cooling – Airside Economization - Energy Star rated
– Deutsche Bank - Adiabatic cooled
10
Cooling Trends
11
ASHRAE TC 9.9 2015 Class A1- A4 Summary
12
Which Group do you belong to?
• Keep the temperature at 65°F (18.5°C) and humidity tightly
controlled at 50% RH
– “I won’t get fired if the utility bill is high, but I will if the application
goes down.”
• Keep the temperature at 72-75°F (22-24°C) at and humidity
at 50% ±5% RH
– “We typically run at moderate temperatures 22-24°C (72-75°F) now
and may consider going a little higher, and loosening the humidity
range, to save energy …. but waiting to see what others are doing”
• I am a leading edge type and will do anything to show how
much energy I can save. I want to run at hot as possible and
ignore humidity controls
– “Facebook and Google are too conservative”
– “In fact I want to use some of my racks as coffee roasters”
– “I want the most effective and efficient cooling system for my DC”
13
Best Energy Practice - “Free Cooling”
• Economizers
– Water - Effective in Colder Climates– requires additional equipment
– Air - More Hours - Simplest – Requires Filtration
14
Module Topics
Continuous Cooling
The X Factor
15
Air-Side Free Cooling Techniques
16
Yahoo’s “Chicken Coop”
Lockport NY
17
Simple Ventilation Only – Yahoo!
18
Facebook – Direct Evaporative Cooling
(no mechanical supplement)
19
Facebook – Direct Evaporative Cooling
(no mechanical supplement)
20
Facebook
(OpenCompute.org)
21
Adiabatic cooling principle
• Water is atomised into air stream
22
Evaporative Air-Side Economizer
(with Supplementary Mechanical Cooling)
[24C @ 13C]
[13C @ 13C]
[13C]
23
Air-Side Economizer Summer / Hot Day Operation
Modulating
Closed
Outside Air Cooling
Coil
[35C]
95°F
Supply Air to Datacenter
[18C]
Economizer dampers 65°F
Closed
24
Air-Side Economizer Mild Temperature Operation
Modulating
25
Air-Side Economizer Cold Day Operation
Closed
26
Sample Energy Saving Air Side Economizer
Outside Air Availability Outside Air Recir/Mix Mech. Cooling Saving
F Hrs/Yr %/year % % % %
27
Case Studies
Air Side Economization
(with supplementary mechanical cooling)
• NetApp
– First “Energy Star” Data Center
– Fully operating data center
– Located in North Carolina
• Deutsche Bank
– Major Financial Firm
– Proof of Concept – in operation May 2011
– Located in New York City Metro Area
28
Air Side Economization
Case Study – NetApp, Inc.
© NetApp Inc.
29
Air Side Economization
Case Study – NetApp, Inc.
• NetApp Inc. (Raleigh, NC) was the first to achieve the EPA’s Energy
Star® for Data Centers rating. (July 2010)
30
Air Side Economization
Case Study – NetApp, Inc.
The Result:
• They are able to utilize only outside air 67% of the time
resulting in:
31
Air Side Economization
Case Study Deutsche Bank
32
Air Side Economization
Case Study Deutsche Bank
This case study is noteworthy on several fronts:
• It was designed to be cooled using only outside air nearly 100
percent of the year
33
Air Side Economization “Free Cooling” –
Other Costs..
• Direct Outside Air - Requires high MERV Filters
(typically MERV 13)
– High Frequency of Replacement – Cost – Sustainability may require
re-usable (Washable ) filters “Greeness” (Carbon Footprint)
• Trading-off: energy use vs sustainably (land fill full of filters)
• High MERV filtering of outdoor air increases fan energy
• ASHRAE TC9.9 Particulate and Gaseous Contamination
Guidelines for Data Centers
34
Direct Air-Side Economisers –
Problems
• The “Y-Factor” …. Humidity
35
Indirect Evaporative Air Cooling
• 3 Modes
– Air-to-Air HX
– Evaporative Cooling
(indirect- water sprayed on tubes)
– DX Cooling
• Modular
– 250 kW or 500kW per Module
36
Ecoflair
37
Ecoflair
38
EcoBreeze
39
Air-To-Air Heat Exchangers
Outside Air Does Not Enter Data Center
• Kyoto Heat Wheel Design
40
Heat Wheel Cooling - Illustration
Heated air flows through the
heatwheel and is cooled down
to a temperature of 18-27°C
(adjustable)
Physical separation of
hot and cold air
Heatwheel
‘Cool‘
outside air
41
Indirect Air Economiser with Adiabatic Cooling
43
Air-To-Air Heat Exchangers vs
Direct Air-Side Economizers
• Direct Air-side Economizers introduce outside air into the
Data Center
– This requires filters – in many cases 2 levels –
• First level pre-filter for larger particles (MERV 8) and
• Secondary level filter for smaller particulate and dust (MERV 13)
– While Air Side Economizers generally are the most energy efficient heat
removal systems, the introduction of outside air may not be acceptable to
some data center operators
• Air-to-Air Heat Exchangers offer energy savings without introducing
outside air in the data center
– Can be used without additional cooling up to 24˚C (75F), i.e. differential of
3C/5F
– They require twice the fan energy at peak outside temperatures, since
they use separate fans
– Need a lot of space and $$
• Both Types need a particular building shape for large air intake/exhaust
44
Humidity Control ?
Do we still need it?
– Geographic Location Dependant
– Broader Range as of TC 9.9 2015
• A1 “Allowable” 8-80% RH
• Room level – Not in each CRAC / CRAH
– Consider separate Ultrasonic system, not electric heated Steam
– Saves direct energy – no electric heating
– Also provides Cooling- Adiabatic effect
Concerns
• High Humidity
– Corrosion Issues?
– The Y-Factor
• Low Humidity, Static Discharge Damage “ESD” ?
– Now less of a problem – Source ESDA (see www.esda.org )
– ASHRAE TC9.9 allows down to 8%
45
Water-Side Economizer Hours with no required mechanical cooling1,2
Air-Side Economizers: Energy Savings Los Angeles San Jose Denver Co Chicago Boston MA Atlanta Seattle
Potential
Outdoor Air Supply Air
% of Yr below % of Yr below % of Yr below % of Yr below % of Yr below % of Yr below % of Yr below
Wetbulb Bin CWS1 oF(oC) Temp2
o o o o wetbulb wetbulb wetbulb wetbulb wetbulb wetbulb wetbulb
F( C) F( C)
59 (15) 66 (19) 70 (21) 68% 78% 93% 75% 75% 56% 90%
53 (12) 60 (16) 64 (18) 36% 46% 77% 64% 63% 44% 68%
47 (8) 54 (12) 58 (14) 13% 21% 63% 55% 52% 33% 45%
41 (5) 48 (9) 52 (11) 3% 6% 51% 46% 41% 22% 21%
Adiabatically Humidified/ Cooled Air-Side Economizer Hours with no required mechanical cooling3,4,5
Los Angeles San Jose Denver Co Chicago Boston MA Atlanta Seattle
Relative
Supply Air
availability of air-side economizer hours
Outdoor Air
Wetbulb Bin
o
for selected
Temp US cities as a function of supply air temperature
% of Yr below % of Yr below % of Yr below % of Yr below % of Yr below % of Yr below
3
wetbulb wetbulb wetbulb wetbulb wetbulb wetbulb
% of Yr below
wetbulb
F (oC) o o
F ( C)
Prepared by DLB Associates
69 (21) 70 (21) 99% 100% 100% 93% 95% 82% 100%
63 (17) 64 (18) 87% 93% 99% 83% 85% 65% 98%
57 (14) 58 (14) 53% 70% 89% 71% 72% 51% 85%
51 (11) 52 (11) 23% 39% 73% 60% 60% 39% 62% 46
“That’s all very well, but we can’t do air-side
free cooling in a place like Singapore…”
“… can we??”
47
“That’s all very well, but we can’t do air-side
free cooling in a place like Singapore…”
“… can we??”
Continuous Cooling
The X Factor
49
Water-Side Free Cooling
Air-Side Free cooling depends on moving huge volumes of air into the
data centre and back out again – that means it can only be done with
certain building shapes and without neighbouring blockages.
On the other hand, if we can get a water source that is cool enough,
we can use that cool water as the ‘chilled water’ supply to the CRAH
units, thus skipping the chillers.
Note that the warmer the data centre, the warmer we can use ‘cool’
water, and so the more economisation we can do.
50
Water-Side Free Cooling
Data Center
Future
w/Pump & Tank
Chiller
Chiller
Chiller
Chiller
55 Ton
55 Ton
55 Ton
55 Ton
51
Evaporative cooling principle
• Water is applied to the heat exchanger which evaporates, increasing
the cooling effect
• Common applications include cooling towers and evaporative cooling of
air heat exchangers (e.g. Ecobreeze coil)
• We can use this to enhance the water-side economisation
52
Water side economiser options
Combination system
53
Water-Cooled DX CRAC
with Economizer Coil
Refrigerant
Ambient Air Condenser Supply Air Fan
54
“Free Cooling”
55
“Free Cooling”
56
Annual cooling degree days
London 171 7 0
San 254 11 0
Francisco
Tokyo 1004 184 0
Moscow 270 37 0
Singapore 3592 1054 0
Sydney 909 94 2
Cooling degree days are a measure of how much (in degrees), and for how long (in
days), the outside air temperature was above a certain level.
http://www.degreedays.net/
57
Air-Side Free Cooling
Maps Class A2
58
Water-Side Free
Cooling Maps
59
Module Topics
Continuous Cooling
The X Factor
60
ASHRAE TC 9.9
Whitepaper Excerpts
• the purpose of the recommended envelope was to give guidance to data
center operators on maintaining high reliability and also operating their
data centers in the most energy efficient manner.
61
ASHRAE TC 9.9
X Factor
62
ASHRAE TC9.9
The X-Factor
63
The “X” Factor - IT Equipment Reliability
The values in the columns marked “x-factor” are the relative failure rate for
that temperature bin.
64
The “X” Factor - IT Equipment Reliability
Major US Cities – Air Side Economizer
• Time weighted failure rate x-factor calculations for Class A2 for air side
economization for selected major US cities.
65
The “X” Factor - IT Equipment Reliability
Major US Cities – Water Side Economizer
• Time weighted failure rate x-factor calculations for Class A2 for water side
dry-cooler type economization for selected major US cities.
• The data assumes a 21.6°F (12°C) temperature rise between the outdoor
air temperature and the equipment inlet air temperature.
66
The “X” Factor - IT Equipment Reliability
Worldwide Cities – Air Side Economizer
• Time weighted failure rate x-factor calculations for Class A2 for air side
economization for selected major Worldwide cities.
67
The “X” Factor - IT Equipment Reliability
Worldwide Cities – Water Side Economizer
• Time weighted failure rate x-factor calculations for Class A2 for water side
dry-cooler type economization for selected Worldwide cities.
• The data assumes a 21.6°F (12°C) temperature rise between the outdoor
air temperature and the equipment inlet air temperature.
68
Module Topics
Continuous Cooling
The X Factor
69
Fan Laws
• Fan Power Commonly Expressed in Horse Power “HP”
• 1 HP = 746 watts (direct conversion shaft /blade HP
– excludes motor efficiency 80-90%)
• VFD – for AC motors – Better Efficiency with EC Fans
• Efficiency can also be expressed as CFM per Watt
– (Higher is more Efficient)
70
Fan Laws
• Airflow varies directly with the fan speed (RPM)
𝐴𝑖𝑟𝑓𝑙𝑜𝑤2 𝑟𝑝𝑚2
≈
𝐴𝑖𝑟𝑓𝑙𝑜𝑤1 𝑟𝑝𝑚1
71
Fan Affinity Laws
These “laws” only help to estimate actual power vs Speed/CFM and only
approximates real world results
– Static Pressure will impact actual performance
– Variation in fan aerodynamic efficiency at difference speeds
– Actual fan performance curve may be closer to “square” law
In CRAC/CRAH
– EC Fans are generally more efficient than Centrifugal Fans/Blowers
– Additional frictional energy losses for Belt Driven fans/blowers
72
Fan Energy – Delta-T
• Q = M x Cp x dT, so a smaller dT will need we need bigger M i.e. bigger
air flow, thus more fan power
• Fan Energy Impact of low Delta-T across CRAC/CRAH
• Example
• 10 KW Fan Power (@ 100% Speed) w/VFD drive or EC fan
Air-Side Delta-T = 8°F [4.67C], CRAH supply fans with VFD
Total CRAH Fan Power = 10 kW
Assume New Air-Side Delta-T = 12°F [7C] and fan VFDs turn down
New CFM = (8/12) = 0.67 i.e. 2/3 of the previous air flow
New FAN Power = 10KW x (0.67 x 0.67 x 0.67) = 3KW (up to 70% Savings!)
(actual fan power expected in the range 3-to-4 kW)
73
Delta-T vs Airflow vs Fan Power
BTU Fan Fan Percent
(1KW) 1.08 ∆T (F) CFM Speed Power Change Notes:
REFERENCE
3413 1.08 20 158.0 100% 47.4 100% (CFM-Spd-Power)
74
Module Topics
Continuous Cooling
The X Factor
75
Energy Monitoring is Key to Improving
Energy Efficiency
• Industry research has shown that after IT equipment, a data center’s
cooling system is usually the 2nd-largest energy consumer (and is
sometimes the largest!)
• The irony of this is that many data center managers have no real visibility
into their cooling system’s effectiveness or efficiency
• To begin to review a data center’s cooling system efficiency, the first step is
to measure how much energy the entire cooling system uses
• To analyse a data center’s cooling system efficiency, the second level is to
measure how much energy each sub-system of the cooling system uses:
– Fans (CRAC or CRAH) within the Data Center
– Compressors (internal to CRAC or External Chillers)
– Pumps (Chilled Water and/or Condenser Water
– External Condensers, Dry Coolers, or Tower Fans
• To fully optimize a data center’s cooling system, you need granular data,
including temperatures, humidity, fan speed, etc
76
Energy Monitoring
77
Basic! (you only need two numbers)
78
Going One Step Further – Energy Monitoring
79
PUE
• PUE has established itself as a data center standard, but it’s not
perfect, and there is concerns that it may not report properly:
– Originally based on power (kW) as instantaneous measurements
– (easy to fool)
80
PUE - Category 1-3 Points of Measurement
81
PUE Version 2
source: Recommendations for Measuring and Reporting Overall Data Center Efficiency
Version 2 – Measuring PUE for Data Centers, 17th May 2011
82
The Green Grid – NEW Metrics
83
pPUE: Partial PUE
84
pPUE: Partial PUE
• Partial PUE is for energy use within a boundary (such as a Container)
• pPUE = Total Energy within a boundary area divided by the IT
Equipment Energy within that boundary area
85
WUE: Water Usage Efficiency
86
WUE: Water Usage Effectiveness
• WUEsource, a source-based metric that includes water used on-site and water
used off-site in the production of the energy used on-site. Typically this adds
the water used at the power-generation source to the water used on-site
87
CUE: Carbon Usage Effectiveness
88
CUE: Carbon Usage Effectiveness
• Total CO2 Emissions
– This component includes the CO2 emissions from local and
energy grid–based energy sources
• CO2 emissions should be determined for the mix of
energy delivered to the site
– e.g. the electricity may have been generated from varying
CO2-intensive plants — coal or gas generate more CO2
than hydro or wind.
• The mix also must include other energy sources such as
natural gas, diesel fuel, etc.
• Total CO2 emissions value will include all GHGs, such as
CO2 and methane (CH4)
• All emissions will need to be converted to ―CO2
equivalents
89
Module Topics
Continuous Cooling
The X Factor
90
Impact of Cooling Systems Choices and
Designs - Data Center Size and Locations
• Dedicated Purpose-Built Data Center
– Can be design-optimized to maximize cooling efficiency
– State of the art systems
– Best Practices
91
Design Goals and Assumptions
92
Consider Segregated Zones
• Servers have broadest (and widening) environmental envelopes
• Storage (disk based) have more moderate environmental envelopes
• Tape has tightest environmental requirements, but lowest heat load
and airflow
• Who should decide what is the proper environmental parameters, and
any possible increased equipment failures?
– Facilities or IT, or both ?
SSD Servers Tape
94
Temperature Control
95
Humidity Control
Impacts Cooling System Efficiency
• Beside Cooling, a presumed function of a CRAC is Humidity Control and the CRAC
Humidification & Reheat Process is thought important for maintaining “proper”
levels of humidity. This is still a very common mindset. However this consumes
significant additional energy and the reheat may not be effective.
• In most cases each individual CRAC/CRAHs monitor’s it own return airflow and uses
it to try to control humidity
– This is the most common practice, but is a very poor practice
– Most often results in “battling CRACs”
– Consider Disabling the humidification unit on some of the CRACs
….. better yet, disable them all
– ASHRAE 2004 40-55% RH
– ASHRAE 2008 Slightly Broader DP 8°F/5.5°C - 60% RH
– ASHRAE 2011 – “Allowable” 20-80%
– ASHRAE 2015 – “Allowable” 8% - 80%
96
Humidity Control
Impacts Cooling System Efficiency
• Consider a moderate level of humidity control Hi-Low set points:
– Broadening the set points i.e. 35% - 70% RH will lower energy use
– Above “Recommended” (15°C dew point) – but less than 80% RH
“Allowable”*
(but should ensure stay below 17°C dew point allowable
98
Cooling or Ventilating other Areas..
continued
• Transformer Rooms can use outside air ventilation
– Note temp rise specs of transformers
99
Raised Floor Design?
• Hiding cabling under raised floors has been one of the traditional reasons
to have raised floors since its inception. In practice it has been proven
that the raised floor plenum performs best when empty (no cables).
However, it can also contain chilled or condenser water piping, and thus
offers protection of the IT gear in the event of leaks
• Perforated floor tiles and even high flow floor grates are usually still
limited to the 1 cold tile per rack which limits the practical maximum
airflow rates
• Raised floor is an added capital and maintenance expense
• Higher cabinet equipment densities result in cabinets that can range
from 900-1360kg (2000-3000lbs) and thus require a higher weight rating
resulting in a more expensive raised floor system
• Ramp can use a significant portion of the whitespace
Alternatively consider a no-raised floor design, where the room is flooded
with cool air e.g. with hot aisle containment
100
Solar Heat Load Summary
101
Reducing Roof Heat Loads …after Insulation & White Paint
Solar heat
Average Return
24C
∆ 8C
Supply 16C
102
Reducing Roof Heat Loads
Thermal Barrier Utilisation
Solar heat
Venting Venting
Average Return
24C
∆ 6C
Supply 18C
103
Reducing Roof Heat Loads
Maximising the Thermal Barrier
Solar heat
Average Return
24C
∆ 6C
Supply 18C
104
Hybrid Approach
105
Hybrid Design
106
Any Questions?
107
EXERCISE
108
BONUS EXERCISE
109
Module 5
The Future – Looking Forward
110
What do YOU see?
111
Module Topics
In this module we will cover:
Liquid Cooling
Containers
Energy Re-use
Thinking Forward
112
How Much Air Flow Do I Need?
Q = M x Cp x dT
So you think you know how much air flow you need?
113
How Much Air Flow Do I Need?
Q = M x Cp x dT
So you think you know how much air flow you need?
114
Typical Server Airflow vs Intake Temperature
115
Power & Airflow of 1U Energy Star Server
CFM
5 to 25
116
IT Equipment Intake Temps and Airflow
• Higher Intake Air Causes the Fans in the IT gear to Speed Up
• The CRAC/CRAH must then also raise its own Fan Speed/Airflow
to meet the IT gear’s need for more airflow and deliver the
required cooling
• The overall result of higher intake air temps is that both the IT
fan and the CRAC/CRAH fans increase in speed – using more
energy.
• It is therefore important to “fine tune” the data center
temperature to find the ideal intake temperature vs total energy
point – beyond a certain point, the total energy may rise with
the temperature instead of decreasing!
• Studies have found that this begins to
occur at approx. 25-25.5°C (77-78°F)
for typical IT equipment, however,
each data center is different
(see “knee” in chart)
117
Balancing Act – if the IT Fans Speed Up
1KW = Power = 1KW
Recirculation 3412 = (Heat-BTU) = 3412
CRAC IT
--------------------------------
118
Balancing Act – if Excess Cool Air
1KW = Power = 1KW
Bypass (loss of Efficiency) 3412 = (Heat-BTU) = 3412
CRAC IT
--------------------------------
119
Balancing Act - Ideal
1KW = Power = 1KW
3412 = (Heat-BTU) = 3412
CRAC IT
--------------------------------
121
Strategies for Mis-Matched Airflows
122
Module Topics
Liquid Cooling
Containers
Energy Re-use
Thinking Forward
123
Water vs Air
• Air Cooling
– Universal - Easy … Common Standard
– Easy to install, service and remove
– Lower Heat Transfer – Far More Fan Energy Required than water pumps
• Water is ~3,500 times more effective than air as a heat transfer medium! (by
volume and delta T)
– Requires Plumbing
– CRAH Computer Room Air Handler - very common
– Air to Water heat exchangers (really air cooling)
(most large data centers use “water via CRAH – via raised floor”)
“Close Coupled” Inrow – sometimes referred to a Liquid Cooling – Misnomer
124
Liquid Cooling?
Is this a liquid-cooled server?
125
Liquid Cooling?
Is this a liquid-cooled server?
128
ASHRAE Latest Release “Liquid Cooling”
Thermal Guidelines for Liquid Cooled Data Processing Environments -
Released September 2011 by ASHRAE TC 9.9, Updated 2015
• Describes classes for the temperature ranges of the facility supply of
water to liquid cooled IT equipment.
• Also reinforces some of the information provided in the Liquid Cooling
Guidelines book on the interface between the IT equipment and
infrastructure in support of the liquid cooled IT equipment.
• Because of the energy densities found in many high performance
computing (HPC) applications, liquid cooling can be a very appropriate
technology.
• 30 kW racks are typical with densities extending as high as 80 to 120 kW.
Without some implementation of liquid cooling these higher powers
would be very difficult if not impossible to cool.
• Advantages of liquid cooling increase as the load densities increase.
• More details can be found in “Liquid Cooling Guidelines for Datacom
Equipment Centers”, part of the ASHRAE Datacom Series.
129
ASHRAE Thermal Guidelines for
Liquid Cooled Data Processing Environments
130
Highlights of ASHRAE “Liquid Cooling”
Equipment Classes
• Class W1/W2
– Typically a data center that is traditionally cooled using chillers and a
cooling tower but with an optional water side economizer to improve on
energy efficiency depending on the location of the data center. See Figure
3a slide 51/52
• Class W3
– For most locations these data centers may be operated without chillers.
Some locations will still require chillers. See Figure 3a slide 51/52
• Class W4
– To take advantage of energy efficiency and reduce capital expense, these
data centers are operated without chillers. See Figure 3b slide 51/52
• Class W5
– To take advantage of energy efficiency, reduce capital expense with chiller-
less operation and also make use of the waste energy, the water
temperature is high enough to make use of the water exiting the IT
equipment for heating local buildings. See Figure 3c slide 51/52
131
Highlights of ASHRAE “Liquid Cooling”
Equipment Classes
133
Highlights of ASHRAE “Liquid Cooling”
Equipment Classes (Imperial Units)
134
Highlights of ASHRAE “Liquid Cooling”
135
Temperature Stack
• Figure B-2 shows the stack of
approach temperatures, figuratively
depicting the temperature rise from
ambient (cooling tower or dry
cooler) through to the liquid
provided to the IT equipment.
• The left stack is for a dry cooler
from the 99.5°F value found in the
lower chart in Figure B-1 through a
Cooling Distribution Unit (CDU) and
any system preheat (e.g. series cooling inside
the IT) and finally the temperature rise
through to the case temperature
136
IBM Hydro Cluster
Liquid Cooled CPUs
IBM Corp
137
IBM Hydro Cluster
IBM Corp
138
IBM
Hydro
Cluster
IBM Corp
139
LIQUID COOLING EXAMPLE : AIST
USES NOVEC
140
LIQUID COOLING EXAMPLE : SUBMER
Liquid Cooling
Containers
Energy Re-use
Thinking Forward
142
Containers
143
Container and Modular Data Centers
Fans
144
Container with Isolated Air Side
Economizer Air-to-Air Heat
Exchanger and Exterior Adiabatic Cooling
145
Module Topics
Liquid Cooling
Containers
Energy Re-use
Thinking Forward
146
Thinking Forward
147
Thinking Future
148
Thinking Further
– Heat water
149
Moving Beyond PUE
ERE: Energy Reuse Effectiveness
I am going to be re-using waste heat from my data
center in another part of my site:
150
ERE: Energy Reuse Effectiveness
– ERE range:
• Minimum “0” (All energy Re-used)
• NO Maximum (little or NO energy Re-used)
151
ERE: Energy Reuse Effectiveness
152
153
ERE: Energy Reuse Effectiveness
• Under the definition of ERE, the “Waste Heat” must be re-used outside
of the *data center boundary area, such as office space, other
buildings. The Waste Heat can be re-used directly as Heat or converted
to:
– Cooling (heat converted to cooling via an absorption chiller)
– Electrical Energy (converted via a technology)
• If the Waste Heat is re-used within the data center boundary area, then
it is not considered as ERE, but it will still beneficial and help lower the
PUE, since it would lower the “Total Energy” within the Data Center
boundary area
*(Data center whitespace or directly related support areas – UPS. Battery, mechanical or electrical rooms)
154
155
ERE: Energy Reuse Effectiveness
Reuse Energy
ERF = Total Energy
– ERF range:
– Minimum 0 (no Reuse)
– Maximum of 1.0 (all energy Re-used)
• Complexity:
– While in some cases some of the heat can be used to simply warm office
space during the winter, it still must be rerouted to the outside the rest of
the year, adding to the complexity and cost
157
ERE: Energy Reuse Effectiveness
Potential technologies to allow for more effective Reuse of waste heat
• Low Grade Heat (simplest)
– Warm other space or heat water (even partially)
158
IBM HydroCluster
• Envisions Re-using Heat from Servers to heat homes or other buildings
159
Spot the Data Centre in this Picture
160
Energy Energy Energy
• The next challenge is to make more common and effective use of the
major underlying byproduct - the waste heat
• Rather than simply being happy that we have improved the cooling
system efficiencies to the point where only minor improvements can be
made
161
Module Topics
Liquid Cooling
Containers
Energy Re-use
Thinking Forward
162
Thinking Forward
• In that time, the IT gear will have moved ahead 1-2 generations
163
Sea, River and Lake Water Cooling
• While Utility Power plants have used natural bodies of water to cool power
plants for many years, only recently have data centers begun to test this.
• While it can improve PUE since it eliminates the need to have or run
mechanical chillers, it still “dumps” waste heat into the environment
164
Microsoft Seawater Cooling
165
Standards are Developing
BS EN 50600-2-3:2014
166
If Your Data Center Goes Down and
Nobody Notices – Does it Matter?
Are Two Tier 1 sites better than one Tier 4 Data Center?
• Part of the inherent IT design philosophy of Internet Search, Social Mediaand
Web Services sites is the concept of site redundancy and overlap
• By being able to failover and transfer the compute request in the event of
the loss of:
– A Single Server
– A Cluster of Servers
– An Entire Data Center
• This has lessened their reliance on Infrastructure redundancy significantly
improving energy efficiency and placed on the IT architectural design.
• This allows some sites to minimize the risk of a Tier 1 (N) design, because
there are 2 or more data centers geographically separated, each capable of
sharing or servicing the compute load.
167
IT equipment becoming more robust,
can operate at a much wider environmental window
168
Current trend
Which free air cooling methods are you currently using in your data centre(s)?
169
Future trend
Which do you plan to introduce into your datacenter(s)?
170
The Ultimate “Cooling” System Design Goals
171
Final Thoughts
• Q> Which is the “Best” Cooling System
– A> It depends.
• Q> Which is the most Efficient Cooling System
– A> It depends.
• What is the most Cost Effective Cooling System
– A> It depends.
• What is the Worst Cooling System
– A> It depends.
• Will the power densities continue to Rise?
– A> It depends
• Make sure that you design for flexibility, because when it
comes to IT, the only constant is … change!
172
Don’t keep your head in the same old sand when designing cooling systems
173
Any Questions?
174
EXERCISE
175
Module 6
CFD Fundamentals
176
Module Topics
Introduction to CFD
Case Studies
177
CFD
178
CFD as an Option
• Benefits
• Concerns
179
CFD – History
Space Industry
1960s
Data Centres
2000 onwards
It’s not new! The application of CFD has been evolving for over 50 years now
180
CFD – Where Did it Come from?
181
CFD – How does it work?
A 3 dimensional model is firstly created, incorporating all of the important objects
that contribute to blockage of airflow and thermal conditions within the room. The
fewer assumptions that are made, the more accurate the predictions will be gained.
182
CFD – How does it work?
The model is then divided into many cells (gridding) to accurately represent
each individual item within the model.
183
CFD – How does it work?
The discretised model can now be solved iteratively. For every iteration,
values are calculated for each cell. Error between the cells for any variable
imbalance is then summated to provide a total error in the solution (see graph
below). If the residual error number is too large, the process begins again
(iteration)until the solution error is reduced to an acceptable level
Starting Error
Residual Error
Acceptable Error
Iteration Number
184
CFD – How does it work?
Once complete, the predicted thermal information can then be extracted in
many different forms:
Block Colours
Tabular Reporting
185
The Application of CFD to the Data Centre
Industry
Equipment Manufacturer IT / Procurement Facilities Manager
Simulation has the ability to bring together all of the thermal aspects from
the chip to the data centre. (IT Supply Chain)
This in turn provides the opportunity to utilise not only as an analysis tool
but also a communication facilitation tool bringing together all associated
departments from IT, to facilities, including management.
186
The Application of CFD to the Data Centre
Industry
IT Equipment Simulation tools have been used in the
production of IT equipment for over 30 years
(Every electronic device that even you personally
purchase today will have had some form of simulation
applied to it!)
Market competition, the pressures of time to
market, cost, and resilience of product have all
been factors in the successful adoption of CFD
into a standard design process
Extremely fast technology changes - Server
model refresh can be as little as 6 months! No
place for physical prototypes
Usually a single department with experienced
design engineers controlling all parameters
187
The Application of CFD to the Data Centre
Industry
Data Centre Seen in the data centre arena around the year
2000 when chip densities started to become an
issue
Utilised predominantly in the design phase or
one-off troubleshooting exercises.
Most simulations are carried out by external
consultants to the end customer as they feel
they do not have the engineering capability in-
house
Not recognised as part of a standard process so
seen by many as time consuming and costly
(This is changing slowly!)
Multiple departments have input into the space
Still evolving
188
The Application of CFD to the Data Centre
Industry
189
Section Topics
Introduction to CFD
Case Studies
190
Operational Intent
Design Capacity
Utilized
Capacity
• Space
• Power
• Cooling
Utilization
• Network
191
The Reality…
Design Capacity
Lost Capacity
Utilized
Capacity
• Space
• Power
• Cooling
Utilization
• Network
192
Data Centre Capacity
Assumption Reality
1
1 2 2
4
Capacity is made up of four components: Space (U), Power (kW), Cooling (kW), Network (ports).
Capacity is lost when there is no location in the data centre where all the components are
available together in sufficient capacity for a proposed installation of equipment.
193
Data Centre Fragmentation
Total Space
Available U: 33 Available U: 33 Available U: 33 Available U: 33 Available U: 33
Available kW:4 Available kW:4 Available kW:4 Available kW:4 Available kW:4
33U 33U
33U 33U
33U
Total Power
4kW 4kW
4kW 4kW
4kW
All cabinets start of with 100% available capacity
194
Data Centre Fragmentation
Capacity is lost when there is no location in the data centre where all the
components are available together in sufficient capacity for the proposed install.
Available U: 13 Available U: 15 Available U: 26 Available U: 17 Available U: 17
Available kW:
1.2 Available kW:
1.7 Available kW:0 Available kW:2 Available kW:
1.7
195
Data Centre Fragmentation
Total Space
Available U: 13 Available U: 15 Available U: 26 Available U: 17 Available U: 17
Available kW:
1.2 Available kW:
1.7 Available kW:0 Available kW:2 Available kW:
1.7
17U
13U
17U
15U
26U
Total Power
1.7kW
1.2kW
2.0kW
1.7kW
196
Lost (Unavailable) Capacity – it all adds up
With the advent of DCIM tools, it is becoming easier to handle space, power, and
networking to minimise lost capacity as these are visible and power/network may
have flexibility – and once you install, you know how much you’ve got
Imagine how much more difficult it would be if power, data and network cables
were invisible and unsure in value...
197
Cooling is Invisible!
• Most IT Equipment is air cooled
– There are no standards for cooling design.
– IT equipment designs change over time.
– IT deployment decisions often do not
consider air flow or cooling.
– Data Center cooling infrastructure is fixed
over its lifetime.
– The owner/operator has the responsibility to
deliver sufficient airflow and cooling to
every piece of IT equipment at all times.
• But air is invisible!
– Hotspots can lead to lost capacity.
– Do you really know how much effective
cooling capacity you have available
198
Planning an IT layout without considering
airflow causes capacity to be lost
Mid-range equipment,
2 layouts, one works... one fails.
199
Example: IT demands do not match design
assumptions
• We will consider a simple small example data hall:
• Design Specification
– 200kW data hall
– Raised Floor
– Range of 3kw and 4kw Cabinets
– N+1 cooling redundancy
• Design Assumptions
– Hot aisle cold aisle arrangement
– All IT equipment breathes front to back
– Temperature rise across the rack is 12°C
200
Confirming Design Assumptions through CFD
Cooling Capacity
40kW
40kW 100%
30kW
30kW
100 %
30kW
30kW
Utilization
Cooling
Total Load 200kW
Time
201
Accommodating IT Demand
• Design Specification
– 200kW data hall
– Footprint
– N+1 cooling redundancy
• Design Assumptions
– Hot aisle cold aisle arrangement
– All IT equipment breathes front to back
– Temperature rise across the rack is 12°C
202
The mid-range equipment fits into the room
based on design and installation
specifications Cooling Capacity
80kW
60%
0kW
0kW 100 %
0kW
0kW 40 %
203
The other four rows are loaded up according
to the original design
Cooling Capacity
80kW 40%
10kW
10kW 100 %
10kW
60 %
10kW
204
Lost Capacity is only realised when things go
wrong
Cooling Capacity
30%
80kW
15kW
15kW 100 %
30% Lost Capacity
15kW
70 %
15kW
205
But this Lost Capacity could have been
predicted when the problem was caused
206
The Lost Capacity was caused by the
installation of the first 40%
Immediate effect Assumption
80kW
60%
0kW
0kW
Simulation shows that the
0kW original deployment was ok
at the time, but what are
0kW the long term effects?
207
A simulation of the long term effects would
have shown that this deployment sacrifices
30% of the capacity of the data center
The Future Assumption
40%
80kW
60%
15kW
15kW
Reality
15kW
30% 40%
15kW
208
Utilisation and Stranded Capacity - Summary
Problem:
• Any IT deployed in a data centre will disrupt the airflow and
cooling even in empty zones of the room.
Symptom:
• The simple example we just looked at illustrates a common
feature of data centres that is not well understood:
• If your data centre is running at 40% of design load, you do not
have 60% capacity left!
• Thermal hotspots will occur before the data centre reaches
capacity.
Reaction:
• Resilience concerns prevent further installation of IT devices
• Data Centre is effectively ‘closed’.
• More data centres are built to gain more capacity.
209
Foresight is necessary to avoid lost capacity
210
First simulation is to check the initial
resilience
Cooling Capacity
0kW
60%
0kW
80kW
211
Simulating the future shows that the room
will be loaded to 90% before warnings
Cooling Capacity
10%
25kW
25kW
80kW
100 %
10% Lost Capacity
90 %
25kW
10kW
15kW
25kW
212
The Business Case: Build a new data centre?
Or spending $ ?? Million on trying to upgrade aircon for the other data centres?
+
Utilization
Utilization
60% 60% 60% 60%
Time Time
$??M*
Build new Data Centre?
* Annualised Cost Estimates based on Uptime Institute true TCO calculator based on a 1000m2
Cost of IT equipment ignored
213
The Business Case : Better Data Centre Management
=
Utilization
Utilization
60% 60% 60% 60%
Time Time
$??M*
Build new Data Centre
214
Business Case : Energy Savings
215
Frequency of change in an operational data
centre
216
The Virtual Model – Predict before you
commit
Real Facility
Projects &
Facilities IT Virtual Model
Engineering
217
The Virtual Model (or “Digital Twin”)
218
Take a Phased Approach Design
3. Change Management /
1. Validate M&E Design 2. Validate IT Layout
Revalidation
Deploy Deploy
Designing a New Data Centre
219
Take a Phased Approach Existing Facility
1. Improve Facility 3. Change Management /
2. Maximise IT Utilisation
Resilience & Efficiency Revalidation
Deploy Deploy
Improving an existing Data Centre
220
Section Topics
Introduction to CFD
Case Studies
221
Case Study 1 – Troubleshooting/Assessing Options
• 100 m2
• Underfloor Obstructions
• Chilled Water Pipes
• Power Cables
• Cable Trays
• 25 Floor Grilles
• 4 ACUs
• 120 kW of cooling capacity
• 24 Cabinets
• 40 kW of IT load
• IBM Blades – 4 Units
• Rackable C3106 – 12 Units
• HP Proliants DL360 G4 – 15 Units
• NetApp DS14 MK2 - 12 Units
• Cisco Systems Catalyst 6509 – 4
Units
222
Case Study – Simulation Results
Cooling requirement Key results Key conclusions
ACU Supply to Grille • 16,950cfm delivered, 7,633cfm required • Air supply can be
• Air supply is 2x more than is needed reduced
Grille to Equipment Inlet • Air supply can be
73% of
cooling air is reduced
bypassing • ACU performance
limited by short
circuiting of supply
air
Exhaust to ACU Return 44% of exhaust air is • Cooling paths are ill
recirculating defined leading to
potential over heat
of equipment
Doubled-up • Overheating must
Equipment Thermal Safety
networking switches
be controlled
are overheating
before the cooling
system can be
optimized
223
Case Study – Design Options
Option 1
ACU- Network
02 Switches
• Option 1+
• Blanking &
Baffling
Blanking
224
Case Study – Design Options
Option 3 False
Ceiling
• Option 1 +
• Option 2 +
• Chimney Cabinets
Chimney • Addition of a false ceiling
Cabinets
Option 4
Hot-Aisle
Containment
225
Case Study - Results
226
Case Study - Results
227
Case Study 2 : Innovative Design
New Build, Turkish Financial Institution with Economiser Cooling System
• 3 Aisles, 72 Racks,
4-10kW per Rack
• Contained Hot Aisle
x x
x
• Downtime through maintenance
or failure of the central ACU will
raise inlet temperatures past
ASHRAE limits
229
Design Verification New Build, Turkish Financial Institution
Analysis
38°C Air
x 34°C Air
x
there is adequate cooling capacity
in the system
230
Design Verification New Build, Turkish Financial Institution
Solution
231
Design Verification New Build, Turkish Financial Institution
Verification
x x
x
• Downtime through maintenance
or failure of the central ACU will
now have little effect on the
server inlet temperatures
232
Section Topics
Introduction to CFD
Case Studies
233
Calibration & Validation against existing
installations
When using a model to fix/upgrade/optimise/manage an existing data
centre, it is best to calibrate the model against actual measurements
234
Calibration & Validation against existing
installations Survey Equipment
• Become familiar with the airflow system Vane Anemometer
• Server Airflow
• Everything starts and finishes at the cooling unit • Grille Airflow
Temperature Sensors
Equipment • ACU Temperature
Air Paths • IT Temperature
• Aisle Temperature
235
Calibration & Validation against existing
installations
236
Calibration & Validation against existing
installations
• Ensure your measurement techniques are providing the correct numbers!
The hood on a
balometer adds an
extra obstruction to
the air which finds it
easier to leave
through an adjacent
grille. The recorded
number will
therefore be wrong.
An alternative
method would be to
use a vane
anemometer in
conjunction with
the balometer.
237
Section Topics
Introduction to CFD
Case Studies
238
CFD Summary - Benefits
• Design/Optimise the architectural form (e.g. test raised floor height)
239
CFD Summary - Concerns
• $ Cost:
– Is it in the budget?
– ROI is unclear
• Quality/Accuracy of Software (not all CFD programs apply the equations in the same way)
• Requires expert knowledge to utilise interpret
– Complex software
– Qualified staff required for model checking and result interpretation
• Rubbish in, rubbish out
– Bad model design
– Data collection issues
– Easy to make mistakes
• Time consuming
– Models take time to build and then to run (and adjust/rerun)
– Not part of previous project process, there for deemed an overhead
240
CFD Summary - Evaluate your needs
241
CFD Summary - Evaluate your needs
• Cost
– Utilisation of external consultants
– Bring in-house
• Software
– License Costs
– Level of Accuracy
– User Friendliness
– Customer Support
– Productivity Gains
– Integration into existing company processes / software
• Resourcing Costs
242
CFD – What The Future Holds
243
Any Questions?
244