COMPUTATION RESEARCH COMPUTING STRENGTH. Since Steele in 2008, Research Computing has deployed many world-class offerings in computation

Similar documents
BETTER THAN REMOVING YOUR APPENDIX WITH A SPORK: DEVELOPING FACULTY RESEARCH PARTNERSHIPS

RESEARCH DATA DEPOT AT PURDUE UNIVERSITY

High Performance Computing Advisory Group May 23, 2016

High Performance Computing Resources at MSU

NERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber

Organizational Update: December 2015

Advanced Research Compu2ng Informa2on Technology Virginia Tech

Introduction to High Performance Computing. Shaohao Chen Research Computing Services (RCS) Boston University

unleashed the future Intel Xeon Scalable Processors for High Performance Computing Alexey Belogortsev Field Application Engineer

WVU RESEARCH COMPUTING INTRODUCTION. Introduction to WVU s Research Computing Services

The Stampede is Coming Welcome to Stampede Introductory Training. Dan Stanzione Texas Advanced Computing Center

HETEROGENEOUS HPC, ARCHITECTURAL OPTIMIZATION, AND NVLINK STEVE OBERLIN CTO, TESLA ACCELERATED COMPUTING NVIDIA

Umeå University

Umeå University

The Effect of In-Network Computing-Capable Interconnects on the Scalability of CAE Simulations

SuperMike-II Launch Workshop. System Overview and Allocations

Comet Virtualization Code & Design Sprint

Gateways to Discovery: Cyberinfrastructure for the Long Tail of Science

Interconnect Your Future Enabling the Best Datacenter Return on Investment. TOP500 Supercomputers, November 2017

The Cambridge Bio-Medical-Cloud An OpenStack platform for medical analytics and biomedical research

Flux: The State of the Cluster

High Performance Computing and Data Resources at SDSC

TECHNICAL OVERVIEW ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING

Journey Towards Science DMZ. Suhaimi Napis Technical Advisory Committee (Research Computing) MYREN-X Universiti Putra Malaysia

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

GPU ACCELERATED COMPUTING. 1 st AlsaCalcul GPU Challenge, 14-Jun-2016, Strasbourg Frédéric Parienté, Tesla Accelerated Computing, NVIDIA Corporation

Leonhard: a new cluster for Big Data at ETH

The Impact of Inter-node Latency versus Intra-node Latency on HPC Applications The 23 rd IASTED International Conference on PDCS 2011

Emerging Technologies for HPC Storage

Information Technology Services. Informational Report for the Board of Trustees October 11, 2017 Prepared effective August 31, 2017

Technologies and application performance. Marc Mendez-Bermond HPC Solutions Expert - Dell Technologies September 2017

Habanero Operating Committee. January

Resources Current and Future Systems. Timothy H. Kaiser, Ph.D.

TECHNICAL OVERVIEW ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING

Interconnect Your Future

Annual Report of the Shared Research Computing Policy Advisory Committee (SRCPAC)

PRACE Project Access Technical Guidelines - 19 th Call for Proposals

Deep Learning mit PowerAI - Ein Überblick

Welcome to the XSEDE Big Data Workshop

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

TECHNICAL OVERVIEW ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING

NVIDIA Update and Directions on GPU Acceleration for Earth System Models

Big Data 2015: Sponsor and Participants Research Event ""

HPC Saudi Jeffrey A. Nichols Associate Laboratory Director Computing and Computational Sciences. Presented to: March 14, 2017

Experiences in Optimizing a $250K Cluster for High- Performance Computing Applications

The BioHPC Nucleus Cluster & Future Developments

MARCC: On Data Subject to Restrictions

Introduc)on to High Performance Compu)ng Advanced Research Computing

Library. Summary Report

Overview of HPC at LONI

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 16 th CALL (T ier-0)

HPC 2 Informed by Industry

The Center for High Performance Computing. Dell Breakfast Events 20 th June 2016 Happy Sithole

Welcome to the XSEDE Big Data Workshop

Hybrid KAUST Many Cores and OpenACC. Alain Clo - KAUST Research Computing Saber Feki KAUST Supercomputing Lab Florent Lebeau - CAPS

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 13 th CALL (T ier-0)

Using an HPC Cloud for Weather Science

UB CCR's Industry Cluster

Undergraduate Admission File

GPFS Experiences from the Argonne Leadership Computing Facility (ALCF) William (Bill) E. Allcock ALCF Director of Operations

CSD3 The Cambridge Service for Data Driven Discovery. A New National HPC Service for Data Intensive science

Welcome to the XSEDE Big Data Workshop

Checklist for Selecting and Deploying Scalable Clusters with InfiniBand Fabrics

NCAR s Data-Centric Supercomputing Environment Yellowstone. November 28, 2011 David L. Hart, CISL

The Architecture and the Application Performance of the Earth Simulator

Oak Ridge National Laboratory Computing and Computational Sciences

Tools for Handling Big Data and Compu5ng Demands. Humani5es and Social Science Scholars

Monash High Performance Computing

Recent Advances in ANSYS Toward RDO Practices Using optislang. Wim Slagter, ANSYS Inc. Herbert Güttler, MicroConsult GmbH

S THE MAKING OF DGX SATURNV: BREAKING THE BARRIERS TO AI SCALE. Presenter: Louis Capps, Solution Architect, NVIDIA,

ACCELERATED COMPUTING: THE PATH FORWARD. Jen-Hsun Huang, Co-Founder and CEO, NVIDIA SC15 Nov. 16, 2015

Sharing High-Performance Devices Across Multiple Virtual Machines

HPC Capabilities at Research Intensive Universities

NAMD GPU Performance Benchmark. March 2011

Fulton Schools of Engineering Career Fair, Tempe Fall 2018 Masters & PhD Recruiting Day Pages 2-3

VxRail: Level Up with New Capabilities and Powers GLOBAL SPONSORS

Leveraging Software-Defined Storage to Meet Today and Tomorrow s Infrastructure Demands

Batch Scheduling on XT3

INSPUR and HPC Innovation. Dong Qi (Forrest) Oversea PM

JÜLICH SUPERCOMPUTING CENTRE Site Introduction Michael Stephan Forschungszentrum Jülich

LAMMPSCUDA GPU Performance. April 2011

UAntwerpen, 24 June 2016

2014 Bond Technology Update Progress of the Technology Network Infrastructure Upgrades Long Range Planning Committee March 4, 2015

IBM Spectrum Scale IO performance

Illinois Proposal Considerations Greg Bauer

HPC Solution. Technology for a New Era in Computing

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments

Overview of the Texas Advanced Computing Center. Bill Barth TACC September 12, 2011

TACC s Stampede Project: Intel MIC for Simulation and Data-Intensive Computing

Managing HPC Active Archive Storage with HPSS RAIT at Oak Ridge National Laboratory

Building Bridges: A System for New HPC Communities

HPC and AI Solution Overview. Garima Kochhar HPC and AI Innovation Lab

National Level Computing at UTK. Jim Ferguson NICS Director of Education, Outreach & Training August 19, 2011

PowerServe HPC/AI Servers

BRC HPC Services/Savio

Proven video conference management software for Cisco Meeting Server

Supercomputer and grid infrastructure! in Poland!

HPC and IT Issues Session Agenda. Deployment of Simulation (Trends and Issues Impacting IT) Mapping HPC to Performance (Scaling, Technology Advances)

Jetstream: Adding Cloud-based Computing to the National Cyberinfrastructure

The Stampede is Coming: A New Petascale Resource for the Open Science Community

Transcription:

Preston Smith Director of Research Computing Services Alex Younts Senior Research Engineer 2017 COMMUNITY 8/29/2017 HPC FACULTY MEETING CLUSTER

COMPUTATION RESEARCH COMPUTING STRENGTH Since Steele in 2008, Research Computing has deployed many world-class offerings in computation

9 HPC SYSTEMS STEELE 7,216 cores, Installed May 2008 Retired Nov. 2013 COATES 8,032 cores, Installed May 2008 24 departments, 61 faculty investors Retired Sep. 2014 ROSSMANN 11,088 cores, Installed Sept. 2010 17 departments, 37 faculty investors Retired Sep. 2015 HANSEN 9,120 cores, Installed Sept. 2011 13 departments, 26 faculty investors Retiring Oct. 2016 CARTER 10,368 cores Installed April 2012 26 departments, 60 faculty investors #54 on June 2012 Top 500 CONTE 9,280 Xeon cores (69,900 Xeon Phi) Installed August 2013 26 departments, 62 faculty investors #28 on June 2013 Top 500 DATA DEPOT RICE HALSTEAD 2.5 PB of disk storage Installed Nov. 2014 400+ faculty investors from every academic college 13,200 cores Installed May 2015 23 departments 69 faculty investors 10,160 cores Installed December 2016 25 departments 62 faculty investors Halstead-GPU Expansion May 2017

#166 RICE 2015

PARTNERS FACULTY DEMOGRAPHICS Department Cores Aeronautics and Astronautics 5740 Mechanical Engineering 5556 CMS Tier2 5440 Electrical and Computer Engineering 4344 Earth, Atmospheric, and Planetary Sciences 2540 Materials Engineering 2064 Nuclear Engineering 1564 Other College of Engineering 980 Chemistry 824 Physics and Astronomy 820 Biomedical Engineering 640 Other Executive Vice President for Research and Partnerships 600 Statistics 512 Chemical Engineering 424 Agricultural and Biological Engineering (Biological Engineering) 368 Biological Sciences 356 Industrial Engineering 296 Civil Engineering 276 Computer and Information Technology 248 Medicinal Chemistry and Molecular Pharmacology 248 Mathematics 232 Bioinformatics Core 200 Agronomy 180 ITaP 176 Computer Science 156 Horticulture and Landscape Architecture 156 Cancer Center 96 Forestry and Natural Resources 96 Biochemistry 40 Botany and Plant Pathology 40 Industrial and Physical Pharmacy 40 Brian Lamb School of Communication 32 Agricultural Economics 20 Animal Sciences 20 Food Science 20 Health Sciences 20 Other College of Pharmacy 20 Agricultural and Biological Engineering (Agricultural Systems Mgmt) 16 287M hours delivered in 2016 Almost 200 investors from 36 departments, from every College, and 3 Purdue campuses The gold standard for condo-style computing Today, the program is part of departments faculty recruiting process. A selling point to attract people to Purdue!

BID PROCESS Open bid process requesting: Quantity 500 nodes Included various interconnects (EDR/HDR Infiniband, OmniPath) Broadwell or Sky Lake processors, 2.4 GHz or better At least 128GB memory per node SSD boot drive, 250G or better Optional uplift for one Nvidia P100 GPU per node Responses ranged from $4,400-4,800 for a node like Halstead, or $6,000-7,700 per node for the latest Intel CPU. Vendor Cores/Node CPU GB RAM/Node $/node RDMA Network $ per GF Vendor 1 24 Sky Lake 192 $6,050.00 EDR IB $3.03 Vendor 2 24 Sky Lake 192 $7,675.13 EDR IB $3.88 Vendor 3 24 Sky Lake 192 $7,754.82 EDR IB $3.84 Vendor 4 20 Broadwell 128 $4,756.70 EDR IB $6.19 Vendor 5 20 Broadwell 128 $4,456.84 EDR IB $5.80 Halstead 20 Haswell 128 $3,600.00 EDR IB $4.80

FINAL NODE SPECS BID RESULTS Base node: Dell R640 24-core node, 2.6 GHz Intel Xeon Gold Sky Lake processors (Xeon Gold 6126) Higher node price, but 50% faster processors 32 Flops per cycle! 96 GB DDR4 memory 384 GB, 768 GB & 1536 GB options Memory prices are high world-wide (More on this later!) EDR Infiniband interconnect 100 Gbps, 3:1 fat tree very similar in speed to Halstead Converged fabric IP traffic uses Infiniband rail Also bid systems for provost equipment program, leading to substantial savings!

STORAGE CPU IS ONLY PART OF THE PICTURE Halstead 1 PB GPFS 22 GB/sec bandwidth 100k IOPS Brown 3 PB Lustre 40 GB/sec bandwidth 400k IOPS

THE BOTTOM LINE XEON GOLD VS HASWELL (HALSTEAD) For codes used by community cluster partners Ansys Fluent: 1.6x faster Converge CFD: up to 1.29x faster Gaussian: 1.25x faster LAMMPS: Up to 2.4x faster GROMACS: Up to 2x faster VASP: Up to 1.9x faster AMBER: Up to 1.73x faster NAMD: 1.67x faster HOMME: Up to 1.67x faster WRF: Up to 1.41x faster HS06: 675.27

MEMORY WHAT IS THE IDEAL AMOUNT? Halstead had 128G RAM per node - Why 96GB for Brown? Cluster 95th Percentile 99th Percentile Rice 17.2 36.6 Conte 27.3 44.7 Snyder 184 510.4 GB/node per job

FOR THE MAJORITY: HPC SYSTEM Brown: A traditional HPC system The same, familiar model: New cluster acquisition every year Service through October of 2022 Improved floating point performance Vastly improved I/O subsystem Nobel Laureate Herbert C. Brown

550 Nodes of Brown: 1.1 PF peak #166 RICE #150 BROWN 2015 Equivalent to all of Conte, including Xeon Phi Accelerators! 2017 (Est)

NODE PRICES OPTIONS Base node option, plus Node with 3 Nvidia GPUs Large memory options, also with 768 GB and 1.5 TB configurations available. 96G Brown Node 384G Snyder Node 3-GPU Brown Node $5,599.00 $9,500.00 $18,325.00

GPU PARTITION GPUS Requests for GPU for accelerating simulations, computer architecture research, or deep learning. 12 researchers are currently evaluating Halstead s GPU partition With Brown: Plans are for a 16-node GPU partition, with 3 Nvidia P100 GPUs per node. 250 TF of GPU! Annual subscription to access to GPU nodes: $2,500 per year

INCREASING VALUE TRENDS Community Cluster Program - Price per Node and Dollars per GFlop 8000 30 7000 25 6000 Dollars per Node 5000 4000 3000 20 15 10 Dollars per GFlop 2000 1000 5 0 Steele Coates Rossmann Hansen Carter Conte Rice Halstead Brown 0 Cost per Node Dollars per Gflop

IMPLEMENTATION General implementation schedule: Facilities preparation already in progress - in the POD modular datacenter. Plan is for early access and burn-in early October Top 500 benchmark run week of Oct 16 General availability Oct 30 No install event this year.

RESEARCH SECTION TITLE DATA STORAGE

DATA IS HUGE! AND GROWING Fortress Archive Growth TB 9,000.00 8,000.00 8,331.98 7,000.00 6,000.00 5,000.00 4,000.00 TB 3,000.00 3,273.25 2,000.00 1,000.00-2,109.91 1,002.30 550.15 0.24 110.91 Mar-97 Dec-97 Sep-98 Jun-99 Mar-00 Dec-00 Sep-01 Jun-02 Mar-03 Dec-03 Sep-04 Jun-05 Mar-06 Dec-06 Sep-07 Jun-08 Mar-09 Dec-09 Sep-10 Jun-11 Mar-12 Dec-12 Sep-13 Jun-14 Mar-15 Dec-15 Sep-16 Jun-17

RESEARCH DATA DEPOT THE 2014 HPC SYSTEM At $75/TB per year: New price! Storage oriented around your research lab, with Snapshots Multi-site copies of your data Disaster protection A scalable, expandable storage resource optimized for HPC Access to Globus data transfer service, and endpoint sharing

GIVE IT A TRY! GET ACCESS To buy 1 or more TB of space, Or to set up a trial for your lab Order online: https://www.rcac.purdue.edu/purchase/

STATISTICS ADOPTION A hit! Over 400 research labs are Depot partners! Many are not HPC users! Thousands of individual users 1.25 PB sold A research group purchasing space has purchased, on average, nearly 10 TB. Other institutions looking to Purdue for leadership from our Depot storage service

GLOBUS STATISTICS Transfer and share large datasets. https://transfer.rcac.purdue.edu. With dropbox-like characteristics.. Directly from your own storage system!

GLOBUS ADOPTION Terabytes Transferred per Month per Managed Endpont Terabytes Transferred 300 250 200 150 100 50 0 Jun-14 Aug-14 Oct-14 Dec-14 Feb-15 Apr-15 Jun-15 Aug-15 Oct-15 Dec-15 Feb-16 Apr-16 Jun-16 Aug-16 Oct-16 Dec-16 Feb-17 Apr-17 Jun-17 Purdue Carter Cluster Purdue Conte Cluster Purdue Fortress HPSS Archive Purdue Halstead Cluster Purdue Hammer Cluster Purdue Hansen Cluster Purdue Hathi Cluster Month Purdue Peregrine1 cluster Fortress: 1.74PB since Aug 15! Last 12 months: 2.4 PB transferred (388TB in March!) Average of 185 TB, 84 unique users per month

RESEARCH SECTION TITLE SERVICES OTHER SERVICES YOU MIGHT BE INTERESTED IN

GROWTH NON-TRADITIONAL HPC RESEARCHERS College 2014 2017 Growth 2014-2017 Growth Rate 2014-17 Agriculture 48 111 63 131% Engineering 161 265 104 65% Science 199 227 28 14% Education 1 7 6 600% Liberal Arts 1 9 8 800% Management 20 24 4 20% Pharmacy 5 9 4 80% Polytechnic 13 21 8 62% Heath and Human Sciences 14 28 14 100% Veterinary Medicine 0 5 5

DATA WORKBENCH But not wrapped up in a package WE HAVE THE PARTS

COMPUTATIONAL DESKTOPS THINLINC

DATA ANALYTICS IN THE BROWSER Jupyterhub (Python) Rstudio

FEATURES DATA WORKBENCH Approx. $150 annual charge for access Easy access to data analysis tools Run virtual private Windows desktops Run virtual private Linux images in containers Integrate with RCAC services Depot, Fortress, Globus, Github, Self Service, etc. Grow to batch HPC as your needs grow! Add same interactive capabilities to community cluster frontends

COMPUTATIONAL SUPPORT SCIENCE EXPERTISE DOMAINS Chemistry Physics Astrophysics Earth and Atmospheric Sciences Computer Science Chemical Engineering Electrical and Computer Engineering Cell and Molecular Biology Agriculture APPLICATION SPACES Molecular Dynamics Image Processing Quantum Chemistry Weather Modeling Machine Learning Big Data Computer Architecture Finite Element Analysis Statistics Bioinformatics Geospatial Remote Sensing Visualization

NEED HELP? COFFEE BREAK CONSULTATIONS Hard to solve problems with HPC? Need help building your software or optimizing your workflow? Need to learn what resources are available? RESEARCH COMPUTING COFFEE BREAK CONSULTATIONS Meet up with ITaP research computing staff and other researchers who use or are interested in High Performance Computing at Purdue. Join us for informal discussions of scientific computing along with any other topic that might spring up. We ll be meeting at different coffee shops around campus each week. Check the coffee web page to see this week s location and time. rcac.purdue.edu/coffee

SCHOLAR HPC FOR INSTRUCTION Need to teach students to use HPC or work with big data in a course? Scholar cluster is available to any instructor at no cost. Spring 2016: CS STAT CHEM EAPS AAE ANSC ME

VERSION CONTROL COLLABORATE ON CODE OR DOCUMENTATION Local-to-Purdue Github repositories for your lab, managed by you!

WOMEN IN HPC Fall kick-off meeting Wednesday, August 30 th 12:00-1:00PM in RAWL 2082. Our invited speaker: Dr. Beth M. Holloway, Director of the Women in Engineering program and Assistant Dean of Undergraduate Education, College of Engineering. Join WHPC for computing scholarship, networking, and mentorship opportunities!

COMING EVENTS TRAINING Big data research with Hadoop and Spark XSEDE Workshop September 12 & 13, 2017 11:00am 5:00pm Unix 101 September 19 & 21, 2017 Jetstream cloud resource for science and engineering September 20, 2017 1:30pm 4:30pm Unix 201 October 3 & 5, 2017 XSEDE HPC Monthly Workshop Series - MPI Day 1 October 3& 4, 2017 11:00am 5:00pm Clusters 101 October 17& 19, 2017 XSEDE HPC Monthly Workshop Series - GPU Programming Using OpenACC November 7, 2017 11:00am 5:00pm Upcoming but not yet scheduled: NVIDIA GPU - Deep learning and CUDA programming Software Carpentry Programming with Python Software Carpentry - Programming with R Software Carpentry - R for Reproducible Scientific Analysis Software Carpentry - Programming with MATLAB

THE END QUESTIONS? rcac.purdue.edu/purchase

SPARE SLIDES EVERYTHING PAST HERE IS SPARE