MEASURE MONITOR UNDERSTAND

Similar documents
Cooling on Demand - scalable and smart cooling solutions. Marcus Edwards B.Sc.

Evaporative free air cooling technology Providing evaporative free air cooling solutions.

Thermal management. Thermal management

MSYS 4480 AC Systems Winter 2015

Virtualization and consolidation

18 th National Award for Excellence in Energy Management. July 27, 2017

Recapture Capacity for Existing. and Airflow Optimization

Energy Efficiency Best Practice Guide Data Centre and IT Facilities

PROCESS & DATA CENTER COOLING. How and Why To Prepare For What s To Come JOHN MORRIS REGIONAL DIRECTOR OF SALES

Smart Data Centres. Robert M Pe, Data Centre Consultant HP Services SEA

Efficiency of Data Center cooling

Switch to Perfection

Green Data Centers A Guideline

Future of Cooling High Density Equipment. Steve Madara Vice President and General Manager Liebert Precision Cooling Business Emerson Network Power

Industry leading low energy evaporative fresh air cooling.

Green Sustainable Data Centres. Checklist - EU Code of Conduct on Data Centres. Annex

Server Room & Data Centre Energy Efficiency. Technology Paper 003

Data Centre Energy Efficiency. Duncan Clubb EMEA Data Centre Practice Director CS Technology

Google s Green Data Centers: Network POP Case Study

Data Centre Energy & Cost Efficiency Simulation Software. Zahl Limbuwala

Survey and Audit Service Schedule. Airflow and Thermal Imaging Survey Service Schedule. Data Centre Solutions Expertly Engineered

IT Air Conditioning Units

Case Study: The design of a modern data centre Jeffrey Brindle Technical Director, HP Critical Facilities Services

Mission Critical Facilities & Technology Conference November 3, 2011 Cooling 101. Nick Gangemi Regional Sales Manager Data Aire

7 Best Practices for Increasing Efficiency, Availability and Capacity. XXXX XXXXXXXX Liebert North America

Excool Indirect Adiabatic and Evaporative Data Centre Cooling World s Leading Indirect Adiabatic and Evaporative Data Centre Cooling

Data Center Energy Savings By the numbers

Where s the Heat Coming From. Leland Sparks Global Account Technical Consultant Belden

Passive RDHx as a Cost Effective Alternative to CRAH Air Cooling. Jeremiah Stikeleather Applications Engineer

HIGH DENSITY DATACOM INTEGRATED COOLING SYSTEM.

How much power?! Data Centre Efficiency. How much Money?! Carbon Reduction Commitment (CRC)

1.0 Executive Summary. 2.0 Available Features & Benefits

Application of Green Technology For Data Centre

Total Modular Data Centre Solutions

Case study on Green Data Center. Reliance Jio Infocomm Limited

Moving Containment Inside the Enclosure. Jeff Markle Great Lakes Case & Cabinet

Overcoming the Challenges of Server Virtualisation

TWO FINANCIAL IMPACT STUDIES. Seal IT Equipment Cabinets for Significant Annual Cost Savings and Simple Payback in a Few Short Months

DATA CENTER EFFICIENCY: Data Center Temperature Set-point Impact on Cooling Efficiency

How to Increase Data Center Efficiency

Power & Cooling Considerations for Virtualized Environments

Innovative Data Center Energy Efficiency Solutions

Beyond BREEAM Excellent DCs. Cobalt Data Centre Campus

brightsolid Data Hall 1 Keysource Technical Specifications

84 kw, Tier 3, Direct Expansion, 937 ft 2

Impact of Air Containment Systems

Data center cooling evolution. The hottest way to save energy. Lucio Bustaffa: Product Development Manager BlueBox Group

DCIM Data Center Infrastructure Management Measurement is the first step

How Liquid Cooling Helped Two University Data Centers Achieve Cooling Efficiency Goals. Michael Gagnon Coolcentric October

Green IT and Green DC

Planning a Green Datacenter

3300 kw, Tier 3, Chilled Water, 70,000 ft 2

Munters Data Center Cooling Solutions. Simon Young & Craig MacFadyen Munters Data Centers

Datacenter Efficiency Trends. Cary Roberts Tellme, a Microsoft Subsidiary

Welcome Association of Energy Engineers to the Microsoft Technology Center

Power Monitoring in the Data Centre

Reducing Energy Consumption with

HIGHLY EFFICIENT COOLING FOR YOUR DATA CENTRE

The Efficient Enterprise. All content in this presentation is protected 2008 American Power Conversion Corporation

Ten Cooling Solutions to Support High-density Server Deployment. Schneider Electric Data Center Science Center White Paper #42

Trading TCO for PUE? A Romonet White Paper: Part I

Partner Introduction

REPORT. Energy Efficiency of the San Diego Supercomputer Center and Distributed Data Centers at UCSD. Prepared for:

A Data Center Heat Removal System That Saves... 80% Discover Inertech s Award-winning Data Center Heat Removal System

Increasing Data Center Efficiency through Metering and Monitoring Power Usage

ZTE ZXCR Series DX In-Row Air Conditioning Product Description

Energy Logic: Emerson Network Power. A Roadmap for Reducing Energy Consumption in the Data Center. Ross Hammond Managing Director

Product Brochure DENCO Close Control Units Row-DENCO. High Density Cooling for Data Centres

White paper adaptive IT infrastructures

DCIM Solutions to improve planning and reduce operational cost across IT & Facilities

Data Center Enclosures Best Practices. Maximize the Efficiency of the Enclosure within a Data Center

High Efficient Data Center Cooling

Data Center Optimized Controls

Data Center Infrastructure Management (D.C.I.M.)

Retro-Commissioning Report

Introducing the Heat Wheel to the Data Center

Combining Cold Aisle Containment with Intelligent Control to Optimize Data Center Cooling Efficiency

Scalable, Secure. Jason Rylands

HIGH DENSITY RACK COOLING CHOICES

survey, 71% of respondents listed Improving capacity planning as a top driver for buying DCIM 1.

FACT SHEET: IBM NEW ZEALAND DATA CENTRE

How High-Density Colocation and Enhanced Visibility Improve Business Performance

E e n rgy g y St S ud u y y Pr P og o r g ams December 2011

A Green Approach. Thermal

Measurement and Management Technologies (MMT)

Is Data Center Free Cooling Feasible in the Middle East? By Noriel Ong, ASEAN Eng., PMP, ATD, PQP

Five Strategies for Cutting Data Center Energy Costs Through Enhanced Cooling Efficiency

DC Energy, Efficiency Challenges

Optimization in Data Centres. Jayantha Siriwardana and Saman K. Halgamuge Department of Mechanical Engineering Melbourne School of Engineering

FREEING HPC FROM THE DATACENTRE

Data Center Infrastructure Management ( DCiM )

White Paper RiZone The Rittal Management Software for IT infrastructures

Measurement and Management Technologies (MMT) for more energy efficient Data Center and Telecommunication Facilities

Ecolibrium. Cloud control. Keeping data centres cool. OCTOBER 2018 VOLUME 17.9 RRP $14.95 PRINT POST APPROVAL NUMBER PP352532/00001

» BEST PRACTICES FOR AIRFLOW MANAGEMENT AND COOLING OPTIMISATION IN THE DATA CENTRE

<Insert Picture Here> Austin Data Center 6Sigma DC CFD Model

Airflow Management s Role in Data Center Cooling Capacity

Ecotel Free Cool 5-15kW

Reducing Data Center Cooling Costs through Airflow Containment

Cooling. Highly efficient cooling products for any IT application. For More Information: (866) DATA CENTER SOLUTIONS

Transcription:

MEASURE MONITOR UNDERSTAND Presented by Alistair Hunt Energy Monitoring, Measurement metrics Where and What 1

History & Background Company Timeline 18 Years Comrac & Sinetica founded 15 Years Raritan Integration Partner 11 Years First generic SNMP Adaptor 7 Years Hawk-i Platform 7 Years Rack Level Power Monitoring 6 Years First 1000+ rack Deployment at BP 4 Years Second Generation Products 4 Years Partner Programme 3 Years Unite Technologies Formed 3 Years Carbon Trust Listing 2 Year Third Generation Products 1 Year SEI ACA Accreditation 2

History & Background Client Base Finance Blue Chip Universities Local Govt. Blue Light 3

Original Energy Usage Case study Annual energy usage of 1.4 MWHrs, 1,260,000 Annual CO2 consumption 7,500 Tonnes/Yr Original PUE calculated at 2.5 No of racks 273, Ave KVA /rack = 2.34KVA Energy Savings opportunity qualified against a PUE target of 1.8

Energy savings no Change in IT Load 5

Observations Observations and recommendations Supporting services Cooling in areas with little IT pay load in Data hall Condenser cooling water temperature at 12degrees, [8 degrees ambient air temperature] Condenser suction and Head end pressure setting sub optimal Recommendations Optimise Under floor air movement to data racks, seal off un-used areas Raise Condenser cooling water temperature to 14 degrees Celsius [10 degrees Celsius ambient air temperature], more free cooling hours Optimise BMS settings Small refrigeration systems were used with lower efficiency centralised refrigeration systems are more efficient at high load, better condenser performance, efficient compressors, better expansion valves and large evaporators better part load performance due to shared condensing and evaporating surface floating head pressures/suction pressures variable speed compressor control better level of free cooling by sharing available heat exchanger surface

Observations Observations and recommendations Data Hall Recommendations Cooling in areas with little IT pay load Data Hall Optimise CRAC use Optimise air flow to data racks and under floor, seal off un-used areas Increase to ASHRAE recommendations, note 23 Cold Aisle temperatures measured at 19 degrees degrees Celsius is the top limit due to server fans Celsius being continuously on Increase Free Air Cooling Cold Aisle ventilated tiles less than 50 % Rack Hotspots Cold Aisle and Hot Air mixing Deploy 80% efficient Ventilation Tiles Rack level Temperature Monitoring. Optimise air flow within the racks and under floor, high Pay load operations Deploy Cold Aisle Containment install Blanking plates and Floor grommets to improve CRAC operational efficiency,, switch off 3 out of 9 CRAC Units Oversized, utilised below optimum efficiency UPS Identify and remove, Virtulisation Non operational IT Equipment powered up

Energy Management Solutions [no change in IT load] 8

Energy Management Solutions [with a 5% reduction in IT Load] 9

WHERE THE SAVINGS COME FROM Source PUE 2.5 Recorded Energy spend PUE 1.8 Energy spend Saving PUE 1.8 Energy spend With 5% IT saving IT Pay load 504,000 504,000 0.00 476,000 Refrigeration 529,000 247,000 302,000 236,000 UPS Losses 57,200 42,000 15,200 42,000 Lighting 32,000 18,000 14,000 18,000 AHU s 84,800 56,000 48,800 51,000 Other 52,000 40,000 12,000 40,000 1,260,000 907,000 353,000 857,000 70% savings in Chillers 1 degree Celsius = 3-4% Energy costs Maximise Free cooling Saving of 28% Total Energy spend 46% Total Facilities spend 10

5% in IT Energy savings Switching off unused servers consolidation of applications where possible implementation of VM technology 5%,[ 28,000] savings can be achieved by turning off Servers. eg HP Proliant 585 G5 Server Max power is 1300W, running power consumption is 650W this equates to an 5694KWh Annually. At 0.09 a KW, Annual cost of 512.00. To make these savings, would mean switching off 55 servers. 11

Capital Expenditure costs Activity Related costs Comments Deployment of power and environmental monitoring hardware including sensors Airflow management and miscellaneous items 163,800 Non invasive technology Temperature & Humidity Data Power and Facility Monitoring 18,000 Blanking of open U space Vertical infill's (800mm wide Cabs) Optional Under floor solutions (not inc.) Installation and commissioning 9.,000 Site services Commissioning and test Deployment of cold aisle containment including sensor in 273 Racks 90,000 Includes supply and installation Enterprise Management Software including standard reporting package 35,000 Dynamic PUE Environmental kwhrs, kva, CO2 Tonnes Costs per rack, group, total Total 315,800 12

Return on investment Return on investment Total Energy spend PUE: 2.50 1,260,000.00 Total Energy spend target PUE:1.8,907.00 Energy cost saving target (year 1) 353,000.00 Total cost to deploy legacy EMS solution 315,800.00 Average cost per cabinet 1,290.00 Return on investment 11 months 13

WHERE FURTHER SAVINGS CAN BE MADE Source Saving Capital cost ROI CRC Carbon credit allowance reduction 25,200 New Centralised Chiller 100,000 250,000 2.5 Yrs National Grid Power Rental 3MW Back up Generator @ 50,000/MW Capacity 150,000 5,000 NA [New Revenue Stream] TOTAL 275,200 255,000 14

National Grid has spike demands Backup/Stand by Power Rental to the National Grid To cater for this the National Grid has to supply a sharp increase in power for a Short period of time. This is done by firing up Power Stations 12 hours in advance. This is inefficient. by having the ability to use customer Back up / Standby Generators, high peak demand can be efficiently managed. Rental is based on Capacity, 50,000 / MW. Use of capacity can be 1 hour per year to a few hours per month. Non Intrusive to the customers business activity. CRC Carbon credits are also given for the full capacity to the customer 15

Monitoring Metrics 16

Data Centre Metrics PUE [Power Usage Effectiveness] Total Facility Energy Usage/ IT Energy usage DCiE Reciprocal of PUE IT Energy Usage/ Total Facility Energy Usage However PUE and DCiE does not distinguish between fixed and variable Energy costs Fixed Energy Overhead Facility Power drawn if no IT equipment is on load, ie UPS s. CRAC Units Variable Energy Overhead Facility Power drawn in response to IT load, ie Chiller compressor pumps I²T Heating affect with current draw, ie Caballing and Transformers 17

Dynamic PUE 18

Data Centre Metrics PUE / DCiE Not the be all and end all Data Centre metrics such as PUE and DCiE are commonly Used as a measure of Data Centre Efficiency. There are however some issues with these metrics If an energy efficiency programme is being carried out, which includes reducing the IT load, deployment of more efficient IT equipment, these metrics could get worse Operators with older designed Data Centres Operators with high resilience levels Both will have higher fixed energy overheads. The Fixed Energy overhead has a significant effect on PUE and DCiE, Deploying Fixed and Variable energy overhead metrics can be used to develop Data Centre Efficiency performance Fixed Energy overhead; Base line performance Variable overhead Operational performance 19

Data Centre Metrics DCEe and PUE Metrics will identify efficiency issues They need to be monitored dynamically. Manual measurements at 1] Max IT Load, low Utility load = Optimum DCiE/ PUE 2] Min it Load, High Utility Load = Worst DCiE / PUE Neither are correct Snap shot in time Information not accurate to deliver effective cost savings analysis Dynamic PUE /DCiE, Fixed and Variable Energy overhead metrics measured at the right level of granularity, will high level information to help make practical decisions to improve Energy efficiency 20

Equipment Loss Parameters Device Caballing & Switchgear Rated power (MW) Fixed Losses Variable Losses I²T Losses 1 0% 0% 1.5% PDU s 1.6 0.5% 0% 0.5% UPS s 1.1 2% 2.5% 5% CRAC s 1.2 10% 1.5% 0% Chiller Plant 1.5 5% 30% 0% Transformer 2.1 0.5% 0% 2.5% 21

Data Centre Metrics Calculating / Measuring Fixed and Variable Energy Overhead for New build can be carried out at the Design Test stage of the deployment For an existing Legacy Data centres this is not so easy. Estimations need to be made. This can be done by monitoring the Utility load and IT Load in hourly intervals over several days. Rack level and External temperature should also be measured to consider the effect of cooling/ Free air cooling with external temperature A scatter graph with regression analysis can be plotted to calculate the Fixed Energy Overhead. 22

Fixed and Proportional Energy Overhead 23

Fixed and Proportional Energy Overhead 24

Fixed and Proportional Energy Overhead Variable Energy Overhead 25

Monitoring points for DC Metrics 26

Monitoring points for DC Metrics 27

Conclusion Modular flexible Data Centre design to optimise Fixed Energy overhead, particularly for New Build Data Centres, not fully populated Monitoring of Facilities and IT Load equipment to provide accurate operation performance measurements, DCiE / PUE, Fixed and Variable Energy overheads Monitoring Parameters should include External air Temperature Cold Aisle air Temperature Rack temp monitoring Rack IT load, Utility load dynamic Ideally to specific utilities to monitor optimal performance Monitoring the above parameters at the right level of granularity will allow optimisation of Free air cooling CRAC, Chiller, UPS optimisation Dynamic PUE / DCiE metrics Identify and reduce Hot spots which determine cooling requirements Rack/server KVA duty Phase Balancing 28

Reporting 29

Dash board view of estate to a rack 30

31

32

33

34

35

36

37