LRZ SuperMUC One year of Operation
|
|
- Beatrix Lindsey
- 5 years ago
- Views:
Transcription
1 LRZ SuperMUC One year of Operation IBM Deep Computing Klaus Gottschalk IBM HPC Architect
2 Leibniz Computing Center s new HPC System is now installed and operational 2
3 SuperMUC Technical Highlights 3 PFLOP Computer in Germany in Gauß-Center 9414 Nodes with 2 Intel Sandy Bridge EP 209 Nodes with 4 Intel Westmere EX 3 PFLOP/s Peak Performance 327 TB Memory Infiniband Interconnect Large File Space for multiple purpose 10 PByte File Space based on IBM GPFS with 200GByte/s aggregated I/O Bandwidth 2 PByte NAS Storage with 10GByte/s aggregated I/O Bandwidth No GPGPUs or other Accelerator Technology Innovative Technology for Energy Efficient Computing Hot Water Cooling Energy Aware Scheduling Most Energy and Cooling Efficient high End HPC System: PUE 1.1
4 SuperMUC Energy Efficiency Goals Stable, highly scalable, efficient Hardware based on standard x86 components Save 40% of Energy compared to air-cooled HPC systems Hot Water cooling allowing for Fee Cooling all year around Using standard components Easy serviceable Frequency controlled nodes Optimize Application Energy consumption during use Energy saving if not in use End of a LINPack Run on 240 nodes Power Aware Job Scheduling Run Application at optimal clock rate according to predefined policies Deliver an energy report after job run
5
6 idataplex dx360 M4 water cooled - with Intel Sandy Bridge CPU
7 Island Architecture Muticluster GPFS Island 2 Core SW N1 N2 GPFS Server Cluster N3 Island 2 Core SW Spine Switch 1 Spine Switch 2 Spine Switch 3 Spine Switch 124 Spine Switch 125 I/O Island Core Switch 2 N4 N5 N6 N7 N8 N65 N66 IB P2P SFA 12k Spine Switch N67 N68 N69 SFA 12k N70 Island 18 Core SW N71 N Martin W Hiegl / Uwe Tron
8 Option: Direct Water Cooling (DWC) Direct cooling in water at heat source (95%) - no media change Less noise in machine room no spinning fans Cooling of system without need for Chillers PUE Node inlet temperature between C Inlet temperature can vary with seasons based on achievable temperature Lower and more stable CPU Core temperature (max 70 C) About 10% less leak current compared to air cooled systems Similar pipework requirements as rear door heat exchangers* Clear Advantages: Less energy consumption for cooling of the system (about 40%) Less energy consumption of the CPU (10%) Enables usage of Turbo Mode with all Cores Better TCO and higher efficiency of compute power usage (*) see ASHRAE Technical Committee 9.9 Whitepaper: 2011 Thermal Guidelines for Liquid Cooled Data Processing Environments 8
9 Option: Energy Aware Scheduling (EAS) Policy based steering of node CPU clock for user batch jobs Batch scheduler estimates application run time based on clock rate Admin defined policies determine node clock rate at application execution time Currently unused nodes will be powered down EAS is part of IBM LoadLeveler and xcat and will be ported to LSF Clear Advantages: Less power consumption of application that cannot gain performance from high clock rates Reduction of power consumption of idle nodes Observation of operational limits Example SuperMUC : Default clock rate is 2.2GHz Higher rates up to 2.7 GHz (or Turbo Mode) for applications that will gain performance LINPACK measurement done with Intel Turbomode 9
10 LINPACK on May 31, :================================================================================ 0:T/V N NB P Q Time Gflops 0: :WR01C2R e+06 0:--VVV--VVV--VVV--VVV--VVV--VVV--VVV--VVV--VVV--VVV--VVV--VVV--VVV--VVV--VVV- 0:Max aggregated wall time rfact... : :+ Max aggregated wall time pfact.. : :+ Max aggregated wall time mxswp.. : :Max aggregated wall time update.. : :+ Max aggregated wall time laswp.. : :Max aggregated wall time up tr sv. : : : Ax-b _oo/(eps*( A _oo* x _oo+ b _oo)*n)= PASSED 0:============================================================================ 0: 0:Finished 1 tests with the following results: 0: 1 tests completed and passed residual checks, 0: 0 tests completed and failed residual checks, 0: 0 tests skipped because of illegal input values. 0: : 0:End of Tests. 0:============================================================================ done running linpack at :43:
11 Value of the SuperMUC System SuperMUC represents tightly integrated innovative solution, with a value proposition which reduces client s total cost of ownership and which address growth areas of x86 and green computing. Energy- and cooling efficiency characteristics of hardware and HPC Software Stack provides quantifiable cost reduction PUE 1.1 (SuperMUC incl. Cooling) Holistic view of the Supercomputer Hardware, -Software and Applications Running cost of client reduced by 40% compared to HPC standard system of similar size Scalability, functionality and quality of hardware, software and service provide a qualifiable cost advantage Fewer problems because of leveraging experience from other platforms Faster problem resolution because of integrating development and support Running cost of client reduced by less downtime Running cost of client reduced by less management effort
12 One Year of Operation Direct water cooling is relievable and stable Summer and Winter LRZ Decision: Inlet Temperature varies with outdoor temperature between C Energy Saving Goal of LRZ and IBM is achieved Hardware Failures are below the expected Range Island based architecture for Infiniband, xcat, GPFS, LoadLeveler proves its scalability for large systems Power Consumption metering based on ipdus down to outlet level Hardware and Service Monitoring based on Icinga Automated Call Home on Failure for all system parts Log file analysis based on Splunk Martin W Hiegl / Uwe Tron
13
Co-designing an Energy Efficient System
Co-designing an Energy Efficient System Luigi Brochard Distinguished Engineer, HPC&AI Lenovo lbrochard@lenovo.com MaX International Conference 2018 Trieste 29.01.2018 Industry Thermal Challenges NVIDIA
More informationThe Energy Challenge in HPC
ARNDT BODE Professor Arndt Bode is the Chair for Computer Architecture at the Leibniz-Supercomputing Center. He is Full Professor for Informatics at TU Mü nchen. His main research includes computer architecture,
More informationupdate: HPC, Data Center Infrastructure and Machine Learning HPC User Forum, February 28, 2017 Stuttgart
GCS@LRZ update: HPC, Data Center Infrastructure and Machine Learning HPC User Forum, February 28, 2017 Stuttgart Arndt Bode Chairman of the Board, Leibniz-Rechenzentrum of the Bavarian Academy of Sciences
More informationLenovo Technical / Strategy Update
Lenovo Technical / Strategy Update Martin W Hiegl October 1st, 2018 Bio or: Why should you believe a word I say? Based in Stuttgart, Germany, home of the Lenovo HPC & AI Innovation Center HPC & AI WW Offerings
More informationCAS 2K13 Sept Jean-Pierre Panziera Chief Technology Director
CAS 2K13 Sept. 2013 Jean-Pierre Panziera Chief Technology Director 1 personal note 2 Complete solutions for Extreme Computing b ubullx ssupercomputer u p e r c o p u t e r suite s u e Production ready
More informationData Center Trends and Challenges
Data Center Trends and Challenges, IBM Fellow Chief Engineer - Data Center Energy Efficiency Customers challenges are driven by the following Increasing IT demand Continued cost pressure Responsive to
More informationTECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 6 th CALL (Tier-0)
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 6 th CALL (Tier-0) Contributing sites and the corresponding computer systems for this call are: GCS@Jülich, Germany IBM Blue Gene/Q GENCI@CEA, France Bull Bullx
More informationUpdate on LRZ Leibniz Supercomputing Centre of the Bavarian Academy of Sciences and Humanities. 2 Oct 2018 Prof. Dr. Dieter Kranzlmüller
Update on LRZ Leibniz Supercomputing Centre of the Bavarian Academy of Sciences and Humanities 2 Oct 2018 Prof. Dr. Dieter Kranzlmüller 1 Leibniz Supercomputing Centre Bavarian Academy of Sciences and
More informationAnalyzing the High Performance Parallel I/O on LRZ HPC systems. Sandra Méndez. HPC Group, LRZ. June 23, 2016
Analyzing the High Performance Parallel I/O on LRZ HPC systems Sandra Méndez. HPC Group, LRZ. June 23, 2016 Outline SuperMUC supercomputer User Projects Monitoring Tool I/O Software Stack I/O Analysis
More informationTECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 11th CALL (T ier-0)
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 11th CALL (T ier-0) Contributing sites and the corresponding computer systems for this call are: BSC, Spain IBM System X idataplex CINECA, Italy The site selection
More informationTECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 13 th CALL (T ier-0)
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 13 th CALL (T ier-0) Contributing sites and the corresponding computer systems for this call are: BSC, Spain IBM System x idataplex CINECA, Italy Lenovo System
More informationPlanning for Liquid Cooling Patrick McGinn Product Manager, Rack DCLC
Planning for Liquid Cooling -------------------------------- Patrick McGinn Product Manager, Rack DCLC February 3 rd, 2015 Expertise, Innovation & Delivery 13 years in the making with a 1800% growth rate
More informationTSUBAME-KFC : Ultra Green Supercomputing Testbed
TSUBAME-KFC : Ultra Green Supercomputing Testbed Toshio Endo,Akira Nukada, Satoshi Matsuoka TSUBAME-KFC is developed by GSIC, Tokyo Institute of Technology NEC, NVIDIA, Green Revolution Cooling, SUPERMICRO,
More informationLeibniz Supercomputer Centre. Movie on YouTube
SuperMUC @ Leibniz Supercomputer Centre Movie on YouTube Peak Performance Peak performance: 3 Peta Flops 3*10 15 Flops Mega 10 6 million Giga 10 9 billion Tera 10 12 trillion Peta 10 15 quadrillion Exa
More informationWater-Cooling and its Impact on Performance and TCO
Water-Cooling and its Impact on Performance and TCO Matthew T. Ziegler Director, HPC and AI WW BU System Strategy and Architecture mziegler@lenovo.com 2016 Lenovo Internal. All rights reserved. 1 Agenda
More informationEnergy Efficiency and WCT Innovations
Energy Efficiency and WCT Innovations Zeeshan Kamal Siddiqi HPC Leader Middle East, Turkey and Africa (META) Lenovo 2017 Lenovo. All rights reserved. Why do we have a problem? Higher TDP Processors Data
More informationThread and Data parallelism in CPUs - will GPUs become obsolete?
Thread and Data parallelism in CPUs - will GPUs become obsolete? USP, Sao Paulo 25/03/11 Carsten Trinitis Carsten.Trinitis@tum.de Lehrstuhl für Rechnertechnik und Rechnerorganisation (LRR) Institut für
More informationEnergy Efficiency and Water-Cool-Technology Innovations
Energy Efficiency and Water-Cool-Technology Innovations Karsten Kutzer April 10th 2018 Swiss Conference 2018 Acknowledgments: Luigi Brochard, Vinod Kamath, Martin Hiegl (Lenovo) Julita Corbalan (BSC) Why
More informationPrototyping in PRACE PRACE Energy to Solution prototype at LRZ
Prototyping in PRACE PRACE Energy to Solution prototype at LRZ Torsten Wilde 1IP-WP9 co-lead and 2IP-WP11 lead (GSC-LRZ) PRACE Industy Seminar, Bologna, April 16, 2012 Leibniz Supercomputing Center 2 Outline
More informationOpenFOAM Performance Testing and Profiling. October 2017
OpenFOAM Performance Testing and Profiling October 2017 Note The following research was performed under the HPC Advisory Council activities Participating vendors: Huawei, Mellanox Compute resource - HPC
More informationHPC projects. Grischa Bolls
HPC projects Grischa Bolls Outline Why projects? 7th Framework Programme Infrastructure stack IDataCool, CoolMuc Mont-Blanc Poject Deep Project Exa2Green Project 2 Why projects? Pave the way for exascale
More informationTECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 16 th CALL (T ier-0)
PRACE 16th Call Technical Guidelines for Applicants V1: published on 26/09/17 TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 16 th CALL (T ier-0) The contributing sites and the corresponding computer systems
More informationInspur AI Computing Platform
Inspur Server Inspur AI Computing Platform 3 Server NF5280M4 (2CPU + 3 ) 4 Server NF5280M5 (2 CPU + 4 ) Node (2U 4 Only) 8 Server NF5288M5 (2 CPU + 8 ) 16 Server SR BOX (16 P40 Only) Server target market
More information100% Warm-Water Cooling with CoolMUC-3
100% Warm-Water Cooling with CoolMUC-3 No chillers, no fans, no problems Axel Auweter MEGWARE CTO Workshop on Energy Efficiency in HPC May 30, 2018, Ljubljana!1 MEGWARE Computer Vertrieb und Service GmbH
More informationI/O Monitoring at JSC, SIONlib & Resiliency
Mitglied der Helmholtz-Gemeinschaft I/O Monitoring at JSC, SIONlib & Resiliency Update: I/O Infrastructure @ JSC Update: Monitoring with LLview (I/O, Memory, Load) I/O Workloads on Jureca SIONlib: Task-Local
More informationFujitsu Petascale Supercomputer PRIMEHPC FX10. 4x2 racks (768 compute nodes) configuration. Copyright 2011 FUJITSU LIMITED
Fujitsu Petascale Supercomputer PRIMEHPC FX10 4x2 racks (768 compute nodes) configuration PRIMEHPC FX10 Highlights Scales up to 23.2 PFLOPS Improves Fujitsu s supercomputer technology employed in the FX1
More informationTECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 14 th CALL (T ier-0)
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 14 th CALL (T ier0) Contributing sites and the corresponding computer systems for this call are: GENCI CEA, France Bull Bullx cluster GCS HLRS, Germany Cray
More informationSession 6: Data Center Energy Management Strategies
Session 6: Data Center Energy Management Strategies Data Center Energy Use and Opportunities: Applying Best Practices to meet Executive Order Requirements Dale Sartor, PE Lawrence Berkeley National Laboratory
More informationPackaging of New Servers - energy efficiency aspects-
Packaging of New Servers - energy efficiency aspects- Roger Schmidt, IBM Fellow Chief Engineer, Data Center Energy Efficiency Many slides provided by Ed Seminaro Chief Architect of HPC Power6 & 7 1 st
More informationLS-DYNA Performance Benchmark and Profiling. October 2017
LS-DYNA Performance Benchmark and Profiling October 2017 2 Note The following research was performed under the HPC Advisory Council activities Participating vendors: LSTC, Huawei, Mellanox Compute resource
More informationAlgorithms, System and Data Centre Optimisation for Energy Efficient HPC
2015-09-14 Algorithms, System and Data Centre Optimisation for Energy Efficient HPC Vincent Heuveline URZ Computing Centre of Heidelberg University EMCL Engineering Mathematics and Computing Lab 1 Energy
More informationHot vs Cold Energy Efficient Data Centers. - SVLG Data Center Center Efficiency Summit
Hot vs Cold Energy Efficient Data Centers - SVLG Data Center Center Efficiency Summit KC Mares November 2014 The Great Debate about Hardware Inlet Temperature Feb 2003: RMI report on high-performance data
More informationCHILLED WATER. HIGH PRECISION AIR CONDITIONERS, FROM 7 TO 211 kw
IT COOLING CLOSE CONTROL AIR CONDITIONERS CHILLED WATER HIGH PRECISION AIR CONDITIONERS, FROM 7 TO 211 kw 3 Versions available Single chilled water coil Double chilled water coil High temperature IT COOLING
More informationSaving Energy with Free Cooling and How Well It Works
Saving Energy with Free Cooling and How Well It Works Brent Draney HPC Facility Integration Lead Data Center Efficiency Summit 2014-1 - NERSC is the primary computing facility for DOE Office of Science
More information2014 LENOVO INTERNAL. ALL RIGHTS RESERVED.
2014 LENOVO INTERNAL. ALL RIGHTS RESERVED. Who is Lenovo? A $39 billion, Fortune 500 technology company - Publicly listed/traded on the Hong Kong Stock Exchange - 54,000 employees serving clients in 160+
More informationEfficiency of application the liquid cooling system in a personal hybrid computing system based on graphics processors
Efficiency of application the liquid cooling system in a personal hybrid computing system based on graphics processors D. Akhmedov, S. Yelubayev, F. Abdoldina, T. Bopeyev, D. Muratov Institute of space
More information19. prosince 2018 CIIRC Praha. Milan Král, IBM Radek Špimr
19. prosince 2018 CIIRC Praha Milan Král, IBM Radek Špimr CORAL CORAL 2 CORAL Installation at ORNL CORAL Installation at LLNL Order of Magnitude Leap in Computational Power Real, Accelerated Science ACME
More informationSun Lustre Storage System Simplifying and Accelerating Lustre Deployments
Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments Torben Kling-Petersen, PhD Presenter s Name Principle Field Title andengineer Division HPC &Cloud LoB SunComputing Microsystems
More informationArchitecting High Performance Computing Systems for Fault Tolerance and Reliability
Architecting High Performance Computing Systems for Fault Tolerance and Reliability Blake T. Gonzales HPC Computer Scientist Dell Advanced Systems Group blake_gonzales@dell.com Agenda HPC Fault Tolerance
More informationInfiniBand Strengthens Leadership as the Interconnect Of Choice By Providing Best Return on Investment. TOP500 Supercomputers, June 2014
InfiniBand Strengthens Leadership as the Interconnect Of Choice By Providing Best Return on Investment TOP500 Supercomputers, June 2014 TOP500 Performance Trends 38% CAGR 78% CAGR Explosive high-performance
More informationPerformance Optimizations via Connect-IB and Dynamically Connected Transport Service for Maximum Performance on LS-DYNA
Performance Optimizations via Connect-IB and Dynamically Connected Transport Service for Maximum Performance on LS-DYNA Pak Lui, Gilad Shainer, Brian Klaff Mellanox Technologies Abstract From concept to
More informationOutline. March 5, 2012 CIRMMT - McGill University 2
Outline CLUMEQ, Calcul Quebec and Compute Canada Research Support Objectives and Focal Points CLUMEQ Site at McGill ETS Key Specifications and Status CLUMEQ HPC Support Staff at McGill Getting Started
More informationFujitsu s Approach to Application Centric Petascale Computing
Fujitsu s Approach to Application Centric Petascale Computing 2 nd Nov. 2010 Motoi Okuda Fujitsu Ltd. Agenda Japanese Next-Generation Supercomputer, K Computer Project Overview Design Targets System Overview
More informationSuperMUC. PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Top 500 Supercomputer (Juni 2012)
SuperMUC PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller Munich Network Management Team Ludwig Maximilians Universität München (LMU) & Leibniz Supercomputing Centre of the
More informationThe Mont-Blanc Project
http://www.montblanc-project.eu The Mont-Blanc Project Daniele Tafani Leibniz Supercomputing Centre 1 Ter@tec Forum 26 th June 2013 This project and the research leading to these results has received funding
More informationInvestigating and Vectorizing IFS on a Cray Supercomputer
Investigating and Vectorizing IFS on a Cray Supercomputer Ilias Katsardis (Cray) Deborah Salmond, Sami Saarinen (ECMWF) 17th Workshop on High Performance Computing in Meteorology 24-28 October 2016 Introduction
More informationLS-DYNA Performance Benchmark and Profiling. April 2015
LS-DYNA Performance Benchmark and Profiling April 2015 2 Note The following research was performed under the HPC Advisory Council activities Participating vendors: Intel, Dell, Mellanox Compute resource
More information8233-E8B 3x6-core ENERGY STAR Power and Performance Data Sheet
8233-E8B 3x6-core ENERGY STAR Power and Performance Data Sheet ii 8233-E8B 3x6-core ENERGY STAR Power and Performance Data Sheet Contents 8233-E8B 3x6-core ENERGY STAR Power and Performance Data Sheet...
More informationAltair OptiStruct 13.0 Performance Benchmark and Profiling. May 2015
Altair OptiStruct 13.0 Performance Benchmark and Profiling May 2015 Note The following research was performed under the HPC Advisory Council activities Participating vendors: Intel, Dell, Mellanox Compute
More informationBeyond PetaFlops: Scalable, Energy Efficient IBM System x idataplex dx360 M4 powered by Intel Xeon processor E Product Family
Beyond PetaFlops: Scalable, Energy Efficient IBM System x idataplex dx360 M4 powered by Intel Xeon processor E5-2600 Product Family Cabot Partners Group, Inc. 100 Woodcrest Lane, Danbury CT 06810, www.cabotpartners.com
More informationResults of Study Comparing Liquid Cooling Methods
School of Mechanical Engineering FACULTY OF ENGINEERING Results of Study Comparing Liquid Cooling Methods Dr Jon Summers (J.L.Summers@Leeds.Ac.UK) Institute of ThermoFluids (itf) 54th HPC User Forum, September
More informationUniversity at Buffalo Center for Computational Research
University at Buffalo Center for Computational Research The following is a short and long description of CCR Facilities for use in proposals, reports, and presentations. If desired, a letter of support
More informationEN2910A: Advanced Computer Architecture Topic 06: Supercomputers & Data Centers Prof. Sherief Reda School of Engineering Brown University
EN2910A: Advanced Computer Architecture Topic 06: Supercomputers & Data Centers Prof. Sherief Reda School of Engineering Brown University Material from: The Datacenter as a Computer: An Introduction to
More informationInfiniBand-based HPC Clusters
Boosting Scalability of InfiniBand-based HPC Clusters Asaf Wachtel, Senior Product Manager 2010 Voltaire Inc. InfiniBand-based HPC Clusters Scalability Challenges Cluster TCO Scalability Hardware costs
More informationS8765 Performance Optimization for Deep- Learning on the Latest POWER Systems
S8765 Performance Optimization for Deep- Learning on the Latest POWER Systems Khoa Huynh Senior Technical Staff Member (STSM), IBM Jonathan Samn Software Engineer, IBM Evolving from compute systems to
More informationMission Critical Facilities & Technology Conference November 3, 2011 Cooling 101. Nick Gangemi Regional Sales Manager Data Aire
Mission Critical Facilities & Technology Conference November 3, 2011 Cooling 101 Nick Gangemi Regional Sales Manager Data Aire Agenda Exactly what are we talking about ASHRAE TC 9.9 Design best practices
More informationGreen IT. Finding Practical Efficiencies. Infrastructure Planning and Facilities. Green IT: Finding Practical Efficiencies
Green IT Finding Practical Infrastructure Planning and Facilities Dan Francis dfrancis@ipf.msu.edu David Graff dsgraff@ipf.msu.edu William Lakos wjlakos@ipf.msu.edu 1 2 MSU Historical Energy Efficiency
More informationIndirect Adiabatic and Evaporative Data Centre Cooling
Indirect Adiabatic and Evaporative Data Centre Cooling Developed in response to the broadening of the ASHRAE recommended server inlet temperature envelope in 2008, Excool represents the most advanced
More informationGreen IT and Green DC
Green IT and Green DC Alex SL Tay Regional Service Product Line Manager Site & Facilities Services, IBM ASEAN 1 What our clients are telling us We are running out of space in our data center We have environmental
More informationHighest Levels of Scalability Simplified Network Manageability Maximum System Productivity
InfiniBand Brochure Highest Levels of Scalability Simplified Network Manageability Maximum System Productivity 40/56/100/200Gb/s InfiniBand Switch System Family MELLANOX SMART INFINIBAND SWITCH SYSTEMS
More informationBrand-New Vector Supercomputer
Brand-New Vector Supercomputer NEC Corporation IT Platform Division Shintaro MOMOSE SC13 1 New Product NEC Released A Brand-New Vector Supercomputer, SX-ACE Just Now. Vector Supercomputer for Memory Bandwidth
More informationData Sheet FUJITSU Server PRIMERGY CX2550 M1 Dual Socket Server Node
Data Sheet FUJITSU Server PRIMERGY CX2550 M1 Dual Socket Server Node Data Sheet FUJITSU Server PRIMERGY CX2550 M1 Dual Socket Server Node Standard server node for PRIMERGY CX400 M1 multi-node server system
More informationIndirect Adiabatic and Evaporative Data Center Cooling
Indirect Adiabatic and Evaporative Data Center Cooling Developed in response to the broadening of the ASHRAE recommended server inlet temperature envelope in 2008, Excool represents the most advanced technology
More informationCS500 SMARTER CLUSTER SUPERCOMPUTERS
CS500 SMARTER CLUSTER SUPERCOMPUTERS OVERVIEW Extending the boundaries of what you can achieve takes reliable computing tools matched to your workloads. That s why we tailor the Cray CS500 cluster supercomputer
More informationThink Blue Go Green. Javier Suarez e-techservices
Think Blue Go Green Javier Suarez e-techservices jsuarez@e-techservices.com IBM Systems Director Total cost of ownership (TCO) for servers continues to rise, even as total server spending remains flat
More informationIBM Power Systems HPC Cluster
IBM Power Systems HPC Cluster Highlights Complete and fully Integrated HPC cluster for demanding workloads Modular and Extensible: match components & configurations to meet demands Integrated: racked &
More informationHabanero Operating Committee. January
Habanero Operating Committee January 25 2017 Habanero Overview 1. Execute Nodes 2. Head Nodes 3. Storage 4. Network Execute Nodes Type Quantity Standard 176 High Memory 32 GPU* 14 Total 222 Execute Nodes
More informationThe NCAR Yellowstone Data Centric Computing Environment. Rory Kelly ScicomP Workshop May 2013
The NCAR Yellowstone Data Centric Computing Environment Rory Kelly ScicomP Workshop 27 31 May 2013 Computers to Data Center EVERYTHING IS NEW 2 NWSC Procurement New facility: the NWSC NCAR Wyoming Supercomputing
More informationData storage services at KEK/CRC -- status and plan
Data storage services at KEK/CRC -- status and plan KEK/CRC Hiroyuki Matsunaga Most of the slides are prepared by Koichi Murakami and Go Iwai KEKCC System Overview KEKCC (Central Computing System) The
More informationBENEFITS OF ASETEK LIQUID COOLING FOR DATA CENTERS
BENEFITS OF ASETEK LIQUID COOLING FOR DATA CENTERS Asetek has leveraged its expertise as the world-leading provider of efficient liquid cooling systems to create its RackCDU and ISAC direct-to-chip liquid
More information18 th National Award for Excellence in Energy Management. July 27, 2017
18 th National Award for Excellence in Energy Management July 27, 2017 ICICI Bank, Data Centre, August Hyderabad 30 & 31, 2017 ICICI Bank - Revolution in Digital Banking Over 100,000 Digital POS in 4 months
More informationHPC Technology Trends
HPC Technology Trends High Performance Embedded Computing Conference September 18, 2007 David S Scott, Ph.D. Petascale Product Line Architect Digital Enterprise Group Risk Factors Today s s presentations
More informationMapping MPI+X Applications to Multi-GPU Architectures
Mapping MPI+X Applications to Multi-GPU Architectures A Performance-Portable Approach Edgar A. León Computer Scientist San Jose, CA March 28, 2018 GPU Technology Conference This work was performed under
More informationENERGY AWARE COMPUTING
ENERGY AWARE COMPUTING Luigi Brochard, Lenovo Distinguished Engineer, WW HPC & AI HPC Knowledge June 15 2017, Agenda Different metrics for energy efficiency Lenovo cooling solutions Lenovo software for
More informationCooling Solutions & Considerations for High Performance Computers
Cooling Solutions & Considerations for High Performance Computers White Paper 6 Revision 1 By Rich Whitmore Executive summary The recent advances in high performance computing have driven innovations not
More informationNCAR s Data-Centric Supercomputing Environment Yellowstone. November 29, 2011 David L. Hart, CISL
NCAR s Data-Centric Supercomputing Environment Yellowstone November 29, 2011 David L. Hart, CISL dhart@ucar.edu Welcome to the Petascale Yellowstone hardware and software Deployment schedule Allocations
More informationTECHNOLOGIES CO., LTD.
A Fresh Look at HPC HUAWEI TECHNOLOGIES Francis Lam Director, Product Management www.huawei.com WORLD CLASS HPC SOLUTIONS TODAY 170+ Countries $74.8B 2016 Revenue 14.2% of Revenue in R&D 79,000 R&D Engineers
More information8205-E6C ENERGY STAR Power and Performance Data Sheet
8205-E6C ENERGY STAR Power and Performance Data Sheet ii 8205-E6C ENERGY STAR Power and Performance Data Sheet Contents 8205-E6C ENERGY STAR Power and Performance Data Sheet........ 1 iii iv 8205-E6C ENERGY
More informationData Sheet FUJITSU Server PRIMERGY CX400 M1 Scale out Server
Data Sheet FUJITSU Server Scale out Server Data Sheet FUJITSU Server Scale out Server Scale-Out Smart for HPC, Cloud and Hyper-Converged Computing FUJITSU Server PRIMERGY will give you the servers you
More informationOpen Compute Solutions Guide
Open Compute Solutions Guide Delivering density, performance and serviceability for demanding and extraordinary customers www.penguincomputing.com 1-888-PENGUIN (736-4846) twitter: @Penguin HPC Tundra
More informationSmarter Clusters from the Supercomputer Experts
Smarter Clusters from the Supercomputer Experts Maximize Your Results with Flexible, High-Performance Cray CS500 Cluster Supercomputers In science and business, as soon as one question is answered another
More informationDiamond Networks/Computing. Nick Rees January 2011
Diamond Networks/Computing Nick Rees January 2011 2008 computing requirements Diamond originally had no provision for central science computing. Started to develop in 2007-2008, with a major development
More informationCOST EFFICIENCY VS ENERGY EFFICIENCY. Anna Lepak Universität Hamburg Seminar: Energy-Efficient Programming Wintersemester 2014/2015
COST EFFICIENCY VS ENERGY EFFICIENCY Anna Lepak Universität Hamburg Seminar: Energy-Efficient Programming Wintersemester 2014/2015 TOPIC! Cost Efficiency vs Energy Efficiency! How much money do we have
More informationHybrid Warm Water Direct Cooling Solution Implementation in CS300-LC
Hybrid Warm Water Direct Cooling Solution Implementation in CS300-LC Roger Smith Mississippi State University Giridhar Chukkapalli Cray, Inc. C O M P U T E S T O R E A N A L Y Z E 1 Safe Harbor Statement
More informationHOURLY ANALYSIS PROGRAM 5.10
HOURLY ANALYSIS PROGRAM Carrier Software Systems Carrier Corporation Syracuse, New York Rev. August 21, 2017 Copyright 2017 Carrier Corporation Contents LEED v4 ASHRAE Standards Updates Modeling Capabilities
More informationData Sheet FUJITSU Server PRIMERGY CX400 M4 Scale out Server
Data Sheet FUJITSU Server PRIMERGY CX400 M4 Scale out Server Data Sheet FUJITSU Server PRIMERGY CX400 M4 Scale out Server Workload-specific power in a modular form factor FUJITSU Server PRIMERGY will give
More informationAcuSolve Performance Benchmark and Profiling. October 2011
AcuSolve Performance Benchmark and Profiling October 2011 Note The following research was performed under the HPC Advisory Council activities Participating vendors: Intel, Dell, Mellanox, Altair Compute
More informationServer room guide helps energy managers reduce server consumption
Server room guide helps energy managers reduce server consumption Jan Viegand Viegand Maagøe Nr. Farimagsgade 37 1364 Copenhagen K Denmark jv@viegandmaagoe.dk Keywords servers, guidelines, server rooms,
More informationCase Study: Energy Systems Integration Facility (ESIF)
Case Study: Energy Systems Integration Facility (ESIF) at the U.S. Department of Energy s National Renewable Energy Laboratory (NREL) in Golden, Colorado OFFICE HIGH BAY LABORATORIES DATA CENTER 1 Case
More informationSystem Overview. Liquid Cooling for Data Centers. Negative Pressure Cooling Without Risk of Leaks
Liquid Cooling for Data Centers System Overview The Chilldyne Cool-Flo System is a direct-to-chip liquid cooling system that delivers coolant under negative pressure. Chilldyne s technologies were designed
More informationPower Bounds and Large Scale Computing
1 Power Bounds and Large Scale Computing Friday, March 1, 2013 Bronis R. de Supinski 1 Tapasya Patki 2, David K. Lowenthal 2, Barry L. Rountree 1 and Martin Schulz 1 2 University of Arizona This work has
More informationVirtualization and consolidation
Virtualization and consolidation Choosing a density strategy Implementing a high-density environment Maximizing the efficiency benefit Anticipating the dynamic data center Schneider Electric 1 Topical
More informationAdvanced GPU Computing References with HP Workstations
Advanced GPU Computing References with HP Workstations Woon Yung Chung 鄭運永 Segment Marketing Manager HP Asia Pacific Workstations Dec 14~ 15, 2011. Beijing Full-fledged Use Increase, Application Area Expand
More informationInnovative Data Center Energy Efficiency Solutions
Innovative Data Center Energy Efficiency Solutions Dr. Hendrik F. Hamann IBM T.J. Watson Research Center 2/8/2009 2007 IBM Corporation A holistic Challenge: Energy & Thermal Management Energy / thermal
More informationSTAR-CCM+ Performance Benchmark and Profiling. July 2014
STAR-CCM+ Performance Benchmark and Profiling July 2014 Note The following research was performed under the HPC Advisory Council activities Participating vendors: CD-adapco, Intel, Dell, Mellanox Compute
More informationReducing Data Center Cooling Costs through Airflow Containment
FEATURE Reducing Data Center Cooling Costs through Airflow Containment Implementing a containment solution in the data center costs much less than the alternatives of adding air handlers or other supplemental
More informationIBM Power Advanced Compute (AC) AC922 Server
IBM Power Advanced Compute (AC) AC922 Server The Best Server for Enterprise AI Highlights IBM Power Systems Accelerated Compute (AC922) server is an acceleration superhighway to enterprise- class AI. A
More informationLCP Hybrid Efficient performance with heat pipe technology
LCP Hybrid Efficient performance with heat pipe technology 2 LCP Hybrid The whole is more than the sum of its parts The same is true of "Rittal The System." With this in mind, we have bundled our innovative
More informationGreen Data Centers A Guideline
Green Data Centers A Guideline Sekhar Kondepudi Ph.D Associate Professor Smart Buildings & Smart Cities Vice Chair Focus Group on Smart Sustainable Cities ITU-TRCSL Workshop on Greening the Future: Bridging
More informationNCAR s Data-Centric Supercomputing Environment Yellowstone. November 28, 2011 David L. Hart, CISL
NCAR s Data-Centric Supercomputing Environment Yellowstone November 28, 2011 David L. Hart, CISL dhart@ucar.edu Welcome to the Petascale Yellowstone hardware and software Deployment schedule Allocations
More information