If you can't read please download the document
Upload
coregroup1
View
229
Download
3
Embed Size (px)
Citation preview
Innovative Best Practice Data Centre Cooling
Dr. Peter Koch - Sr. VP Engineering and Product Management
Knrr Racks & Solutions
Emerson Network Power Systems EMEA
by Emerson Network Power March 2012 2
How do you realize
that you are in an energy efficient data center?
3 by Emerson Network Power September 2012
No Datacenter is Like Another
4 by Emerson Network Power September 2012
Key Requirements Are The Same
Availability - Capacity - Efficiency
TCO - Agility
5 by Emerson Network Power September 2012
Cooling
6 by Emerson Network Power September 2012
Cooling Needs of IT Equipment
ASHRAE TC 9.9 Thermal Guidelines 2008 / modified 2011 for
Class A1*)
Recommended: 18 - 27C, 5,5C DP - 60% RH and 15C DP
Allowable: 15 - 32C, 20% RH - 80% RH (Class A1)
10 - 35C, 20% RH - 80% RH (Class A2)
Temperature and humidity at equipment intake:
T 7K 25C 32C 470 m/h per kW
25C 40C 220 m/h per kW T 15K
25C 50C 130 m/h per kW T 25K
*) Quelle: 2011 Thermal Guidelines for Data Center Processing Environments -
Expanded Data Center Classes and Usage Guidance (ASHRAE Whitepaper)
7 by Emerson Network Power September 2012
Server Power Draw vs. Intake Temperature
1,00
1,05
1,20
1,15
1,10
15 20 25 30 35C
ca. + 12%
Quelle: 2011 Thermal Guidelines for Data Center Processing Environments -
Expanded Data Center Classes and Usage Guidance (ASHRAE Whitepaper)
Po
we
r D
ra R
atio
Air Intake Temperature
8 by Emerson Network Power September 2012
Standard Room Cooling System
25C
35 - 50C IT Room
Cooling Plant
9 by Emerson Network Power September 2012
Cooling Alternatives
25C
35 - 50C
+ PIT Equipment Pair movement
chilled water (CW)
for larger sites
compressed refrigerant (DX)
for smaller sites
air?
10 by Emerson Network Power September 2012
Outside Air: Too Hot - Too Cold (+ too humid - too dry, too dusty, polluted)
40C
35C
20C
25C
30C
15C
10C
45C
50C
32C
27C
18C C
lass A
1
Cla
ss A
2
allowed envelope
recommended envelope
Server inlet temperature
ASHRAE TC 9.9*
*) Quelle: 2011 Thermal Guidelines for Data Center Processing Environments -
Expanded Data Center Classes and Usage Guidance (ASHRAE Whitepaper)
am
bie
nt te
mp
era
ture
ra
ng
e
11 by Emerson Network Power September 2012
Direct Free Cooling = Fresh Air Cooling?
Temperature not appropriate most of the time
heating / hot air recirculation in winter
additional cooling in summer
Humidification / dehumidification required (energy consumption!)
Filtering required (energy consumption, maintenance)
Gaseous pollutants can not be filtered corrosion risk!
Large air ducts required, strong restrictions to building design
Difficult to protect against intruders / vandalism / terrorism
Difficult to protect against fire
12 by Emerson Network Power September 2012
Indirect Free Cooling
ambient IT Room
chilled water intermediate circuit
Inside air is kept separate from outside air.
Heat transport by chilled water enables flexible building design
and a high level of protection.
About 10 C are "lost" by the heat exchangers.
Refrigeration required when outside temperature is to high
25C e.g.
20C max. 15C
- dry bulb
- wet bulb
13 by Emerson Network Power September 2012
Air-to-Air Heat Exchangers
Indirect free cooling
Don't do any better!
Take valuable space.
Complicate building design.
4 x 300 m
IT-space on
2 floors
4 x 470 qm IT-
Flche
auf 2
Geschossen
4 x 470 m
IT-space on
2 floors
14 by Emerson Network Power September 2012
Capacity
15 by Emerson Network Power September 2012
A Brief History of Cooling Painpoints
2002: Heat Density Closed Watercooled Racks
16 by Emerson Network Power September 2012
Cooling Solutions - Capacity Ranges
Room
w/o with containment
Row Rack
closed open
Server
chip - cooling
cold plate
< 5
-
around 10
+
around 15
+
around 25
+
around 30
++
up to 75
+++
kW per rack
efficiency
17 by Emerson Network Power September 2012
Power Density Matters
4 kW per rack
80% empty !
18 by Emerson Network Power September 2012
Efficiency
Availability
19 by Emerson Network Power September 2012
A Brief History of Cooling Painpoints
50%
25%
12%
11% 3%
IT Load
Pow er Cooling
Pow er Fans
Electrical Losses
Pow er Light
80%
8%
2% 7%2%
IT Load
Pow er Cooling
Pow er Fans
Electrical Losses
Pow er Light
PUE = 2.0
PUE = 1.25
2007: Energy Efficiency Aisle Containments
20 by Emerson Network Power September 2012
PUE - How to Measure it Correctly
PUE = Total Facility Power
IT Equipment Power
Power Usage Effectiveness
medium voltage
genset
CR
AC
UPS
Total Facility Power
IT Equipment
Power
PDU
PDU
Co
oli
ng
Pla
nt
transformer here if possible
or here
metered rack PDUs
make sure to capture all electrical losses
cooling) may be on UPS
r
one full year averaging
WUE for water consumption
21 by Emerson Network Power September 2012
Power is Money
1 MW for 1 year
1 Mio.
5.000 t CO2
8,76 Mio. kWh @ 0.12 /kWh
German power mix
22 by Emerson Network Power September 2012
Room Cooling: Cold Aisle Containment
23 by Emerson Network Power September 2012
Control Principle
p 0
24 by Emerson Network Power September 2012
Huge Savings From Fan Speed Control 3
3%
50
%
75
%
10
0%
0,00
1,00
2,00
3,00
4,00
5,00
6,00
9,00
Po
wer
dra
w [
kW
]
% fan speed & air flow
EC fans speed control
HPM L15EC (EC Fans)
"sweet spot"
8,00
7,00
Fan speed control bears
enormous savings
1/2 speed 1/8 power
25 by Emerson Network Power September 2012
Containment Benefits
Capacity
more power / heat density per rack
more cooling capacity per cooling unit
suitable for greenfield and retrofit installations
Efficiency
optimized floor space utilization
utmost energy efficiency by system control "from rack to roof"
easy upgrade of existing sites
Availability
no hot spots
resilient response against cooling system failures
predictable system behavior
controllable capacity limits
easy to monitor / manage / operate
26 by Emerson Network Power September 2012
Containment Benefits
TCO
huge energy cost and floor space savings
upgrade amortization typically between 6 and 18 months
simplified design process
easy deployment and installation
based on standard components from industrial production
Agility
easy deployment of IT components
dynamic system control
early identification of capacity limits
modular expandability
customizable containment designs for virtually any situation
27 by Emerson Network Power September 2012
Alternative Setups
28 by Emerson Network Power September 2012
Consistent cold/hot air separation
Dynamc system control from rack to roof
High temperature level, free cooling
Intelligent utilization of redundancies
Modularity
29 by Emerson Network Power September 2012
The Classical Approach Still Going Strong !
29
25C
-20 ... +40C
20C 25C
35 - 50C
Highly efficient components
Aisle containment
Fan speed control
High temperature level
Extensive utilization of indirect free cooling
Optimized system, dynamic system control
"from rack to roof"
ASHRAE recommended envelope all year round
PUE better then 1.25 achievable
30 by Emerson Network Power September 2012
Modularity
Racks, Rows
Aisles
Containers, Pods
Buildings
Rooms / 300 500 sqm
In-room modularity
31 by Emerson Network Power September 2012
In-Room Modularity
32 by Emerson Network Power September 2012
In-Room Modularity
33 by Emerson Network Power September 2012
In-Room Modularity
34 by Emerson Network Power September 2012
In-Room Modularity
35 by Emerson Network Power September 2012
In-Room Modularity
36 by Emerson Network Power September 2012
A Brief History of Cooling Painpoints
From 35 to 50
2012: Hot Side Temperature
37 by Emerson Network Power September 2012
Hot Side Temperature Up to 50C
32C
27C
18C
Cla
ss A
1
Cla
ss A
2
40C
35C
20C
25C
30C
15C
10C
45C
50C
allowed envelope
recommended envelope
Server inlet temperature range
ASHRAE TC 9.9*
Exhaust temperature range
High-end blade servers
full load
idling
*) Quelle: 2011 Thermal Guidelines for Data Center Processing Environments -
Expanded Data Center Classes and Usage Guidance (ASHRAE Whitepaper)
38 by Emerson Network Power September 2012
What Can Be Done?
Excess air supply, cooling of hot zones
Reduced cold air temperature
Gliding operation mode
Enclose the hot air
Rack Cooling
Row Cooling
Efficiency !
Efficiency !
Helps temporarily
39 by Emerson Network Power September 2012
Closed Rack Cooling
Closed watercooled serverracks
12 - 30 kW nominal cooling
Vertical or horizontal air flow
40 by Emerson Network Power September 2012
Open Rack Cooling
Open watercooled serverrack
35 kW cooling capacity
Passive rear door heat exchanger
Air movement by server fans
Currently the most efficient solution
Goethe University Frankfurt / Germany
41 by Emerson Network Power September 2012
In The Row Cooling
Water cooled units
30 kW nominal capacity
One or more per row
Closed or "hybrid"
Solutions from 6 to 60 kW per rack
42 by Emerson Network Power September 2012
7
1
2
3
4 4
5 5
6 6