Switch to Perfection
Traditional Data Centre - Temperatures Ideal Air Temperature: 22 C Chilled Water: 7-9 C Heat Load per Rack: Up to 4,5 Chilling & Pump Set
The challenge a thermal image Front View of Rack Actual Air Temperature: 10-40 C Cold Aisle Hot Aisle Cold Aisle Hot Aisle Cold Aisle
Hot Spot Study 90% of data centers have far more cooling capacity than they need. Although the data centres studied had 2.6 times as many cooling units as they required, hot spots could still be identified, making up 10% of the total floor space. ~ Uptime Institute Increase in room temperature 1 K 2 K 3 K 4 K 5 K Energy saving for climate control 4 % 8 % 12 % 16 % 20 % Source: Swiss Federal Energy Supply Office
Electronic Failures List Source: BBC, Inc, Report GB-185R
Solving Hot Spots Saving Energy The Challenge of the modern Data Center: 1. Eliminating Existing Hot Spots whilst conserving energy. 2. Increasing Heat Density (See below)
Future Data Centre Server Heat
So Much Confusion
Facts about Data Centres and Power Power capacity is not only an Eskom / SA Problem. One BT Data Centre has a larger carbon foot print than all the Scottish distilleries put together The new Norwich Union Data Centre pulls 10% of the Norwich's overall power demand We need to make the most out of the power that we have available It is now estimated that Data Centres are responsible for a bigger carbon footprint than the Airline Industry
Let s look at the energy in a Data Centre Cooling IT Components Power Distribution
Add Cold Aisle Containment Very simple to deploy / retrofit Hot and cold aisles physically separated Greater watts per rack approx 10 or even more Over sizing of the CRAC is reduced CRAC efficiency is increased due to a higher delta T Increase air inlet temperatures CRAC fan speed can be reduced which provides:- Reduced running costs Increased MTBF
Tuning the CRAC with Cold Aisle Containment 110 68 49 49 Fan Power Fan Power Fan Power Delta TDelta T Delta T Air Inlet Temperatur e Air Outlet Temperatur e Fan Power Standard Cold Aisle Hot Aisle Containment Cold Aisle Cold Aisle increases Containment Configuration Delta T by increases 10 o C but Delta by Delta 10 increased 15 o T C5 o C Outlet Temperature CRAC Rating Fan Power 20 o C 15 o C 49 3 25 o C 15 o C 68 3 30 o C 15 o C 110 3 30 o C 20 o C 49 1.5 Delta T Hot Aisle Cold Aisle Cold Aisle Containment
Summary of tuning the CRAC Inlet Temperature Outlet Temperature Rating Fan Power 20 o C 15 o C 49 3 25 o C 15 o C 68 3 30 o C 15 o C 110 3 30 o C 20 o C 49 1.5 Increasing delta T can increase cooling capacity of the CRAC, or reduce CRAC fan speed Increasing air outlet temperature/ increases water temp which in turn helps improves the effectiveness of Free Cooling
Advantages of Cold Aisle Containment Cold Aisle Containment You have a choice, either:- Same capacity but greater efficiency, or Greater capacity, but same efficiency Even airflow of air means a reduction in hot spots Greater heat load per cabinet, 10 plus Reduced foot print compared to hot aisle cold aisle? Reduced noise levels Greater MTBF on the CRAC/ IT Cold Aisle Containment really allows the existing infrastructure to be stretched even further
New Build Medium to High Density
If the predicted demand is less than 10 per enclosure Implement or have the ability to upgrade to Cold Aisle Containment
Depending on the site Cold Aisle Containment with LCP In line For high density applications without raised floors Cooling capacity up to 30 (10kw per rack) Ambient air is extracted from the warm aisle or room through the perforated rear door, cooled and blown into the cold aisle Supports existing room climate control In conjunction with Cold Aisle Containment, provides significant increase in cooling capacity per server rack
Liquid cooling systems Combination of LCP Inline + Cold Aisle Containment LCP Inline Top view Cold aisle containment LCP In line either assists existing CRACs or removes the need
Benefits of Free Cooling Liquid Cooling Packages
Benefits of adding Free Cooling? Johannesburg Temperatures 40.0 35.0 30.0 25.0 20.0 15.0 10.0 5.0 0.0 30% of the year your temperature is less or equal than 10 degrees Max C Mean C Min C
Now the cost of running the system for a year 100% free cooling 30% of the year Chiller capacity Energy needed to run the chiller Not using the Free Cooler Chiller capacity 150 Energy needed to run the chiller 62 Numbers of hours running per year 8784 Cost per h (day /night) R 0.5 Total cost of running per year R272,304 150 62 Numbers of hours running per year 6228 Cost per h (day/night) R 0.5 Cost of running the chiller R 193,068 Cost of running Free Cooling (10.4kw) R 13,291 Total cost of running per year R 206,359 Incorporating Free Cooling reduces energy costs by 25%
There are additional benefits of Free Cooling Even when temp is between 20 0 C and 16 0 C Free Cooling still reduces chiller usage It also improved the MTBF of the chiller. Which in turn lowers maintenance costs Increase your Return on Investment More importantly it gives you a degree of redundancy If you are using water Free Cooling is a logical solution
High Density Data Centres and Liquid Cooling When going above 10 per rack a new cooling method needs to be deployed. Direct water cooling to the cabinet is the next step Increased load up to 30 Increased efficiency 35-45% reduction in required real estate 10-20% reduction on total annual fan power consumption 30-45% reduction in light related cost. 12-14% reduction in power delivered to mechanical chilled water plant Higher cooling capacity, with reduction in hotspots
So how does it work? Single Cabinet Cooling LCP+ unit in its own right and can be connected to a standard Rittal TS8 enclosure. Water is physically separated away from the server. 30 of cooling Dual Cabinet Cooling LCP+ can be placed between two racks and share the cooling capability. 15 per enclosure
System Resilience So how does it work? LCP+ can be mounted to both sides of the TS 8 to provide an N+1 resilience. In case of a chiller failure the second LCP+ has the ability to still maintain the cabinet temperature Fan Resilience Each fan is a single phase fan but fed from a three phase source. In the event of a phase failure, 2/3rds of your fans are able to operate.
Space Comparison 60 load Conventional Hot Aisle Cold Aisle 5 5 5 5 5 5 5 5 5 5 5 5 Cold Aisle Containment 10 10 10 10 10 10 Water Cooled 30 30 Change starts when someone sees the next step William Drayton
Summary of Stage 3 Making Legacy more efficient 9% Saving 25% Saving Cooling IT Components Power Distribution New Build High Density Increased load up to 30 Increased efficiency Better utilisation of Free Cooling 35-45% reduction in required real estate 10-20% reduction on total annual fan power consumption 30-45% reduction in light related cost. 12-14% reduction in power delivered to mechanical chilled water plant 25% in overall energy consumption
Lets look at the energy in a Data Centre Cooling IT Components Power Distribution
Transformer-less UPS & Modularity
Advantages of Modularity - Redundancy - Easy Serviceability 40 kva (N+1) 40kVA (2+1) 48kVA (4+1)
Advantages of Modularity - Scalability Scalability Pay as you grow Easy budgeting Maximises inverter efficiencies (Full load is 95%) Reduces long-term energy costs Available space Scalability Demmand
Advantages of Modularity - Efficiency 120kVA conventional UPS (N+1) At full load, each system is only at 50% Traditional UPS systems are designed for maximum efficiency at >95% load There is a waste in energy PMC 200 128kVA UPS (N+1) At full load, each system is only at 75% At 75% load, UPS is still 95% efficient Therefore saving energy
Advantages of Modularity - Availability Traditional Free Standing 2x120kVA (1+1) Redundant Configuration Advanced Modular (3+1) Redundant Configuration Spare 40kVA Module 40kVA Module 40kVA Module 40kVA Module 120kVA 120kVA LOAD 120kVA LOAD 120kVA Availability: A=MTBFUPS/(MTBFUPS+MTTRUPS) MTBF MTTR Availability Non-Modular (1+1) Redundant Configuration 600,000 6h 0,999990 (5 nines) Non-Modular (3+1) Redundant Configuration 500,000 0.5h 0,999999 (6 nines)
"Total Cost of Ownership" Rittal PMC 200 PMC 200 Efficiency of consumers in 160 160 UPS efficiency in % (with partial load) 95.0% Input in 168.4 UPS heat loss in 8.4 /h per year 73,768 Electricity costs in EUR per /h 0.06 0.06 Additional costs owing to heat loss per year (without cooling) 4,124 /h per year with cooling 103,276 Cooling required for heat loss, factor 0.4 1,649 Additional cost per year including cooling 5,773 A Nother UPS 160 93.0% 172.0 12.0 105,497 0.06 5,897 147,695 2,359 8,256 During 5 years 28,866 More than 10 years 57,731 Additional cost compared with PMC 200 UPS during 10 years 41,281 82,562 24,831 CO2 emissions per year 1 Kilowatt hour (h) of electr. = 0.6 kg CO2 0.6 61,965 88,617 kg Source: Internationales Wirtschaftsforum Regenerative Energien (IWR) http://www.iwr.de/re/eu/co2/co2.html Difference in CO2 per year 26,652 kg Equals an annual car mileage with a Golf FSI (1.6 engine petro): 165,539 km 102,861 miles Difference in CO2 throughout 10 years of operation 266,518 kg
Summary of Power Savings 9% Saving 25% Saving Cooling IT Components Power Distribution 30% Saving 15% in overall energy consumption 47% Saving30%
15% overall saving... is it worth it? Yes Design life of the data centre is 10 years Energy costs are rising 24% increase last year 31% increase this year What is happening next year? Long-term operating costs now out weigh initial Data centre Capex,
Summary In Legacy Sites:- Making it Greener Simple reduce the amount of air mixing Block of air leakages Incorporate aisle containment Install new lower energy CRAC systems Implement Free Cooling In Legacy Sites - Getting more Cooling Capacity Incorporate Aisle containment In New Builds:- Making it Greener and Future Proof Learn from the Legacy Deploy water cooling Use Free Cooling Reducing Power Consumption Transformer less UPS are more efficient Modularity future proofs Little things do make a difference Change the colour of your racks from black to grey, gives you approx a 40% saving on your lighting circuits Some people change when they see the light, others when they feel the heat. Caroline Schoeder
Rittal together + + = Complete IT Competence Thank you for your Time Quintin Lamprecht