HPC Saudi Jeffrey A. Nichols Associate Laboratory Director Computing and Computational Sciences. Presented to: March 14, 2017
|
|
- Ethan Elliott
- 5 years ago
- Views:
Transcription
1 Creating an Exascale Ecosystem for Science Presented to: HPC Saudi 2017 Jeffrey A. Nichols Associate Laboratory Director Computing and Computational Sciences March 14, 2017 ORNL is managed by UT-Battelle for the US Department of Energy
2 2 Our vision Sustain leadership and scientific impact in computing and computational sciences Provide world s most powerful open resources for scalable computing and simulation, data and analytics at any scale, and scalable infrastructure for science Follow a well-defined path for maintaining world leadership in these critical areas Attract the brightest talent and partnerships from all over the world Deliver leading-edge science relevant to missions of DOE and key federal and state agencies Invest in cross-cutting partnerships with industry Provide unique opportunity for innovation based on multiagency collaboration Invest in education and training
3 Oak Ridge Leadership Computing Facility (OLCF) is one of the world s most powerful computing facilities Titan Peak performance 27 PF/s Data storage Memory Disk bandwidth 710 TB 240 GB/s Square feet 5,000 Power 8.8 MW Spider file system 40 PB capacity >1 TB/s bandwidth HPSS archive 240 PB capacity 6 tape libraries Gaea Peak performance 1.1 PF/s Memory 240 TB Disk bandwidth 104 GB/s Square feet 1,600 Power 2.2 MW Data analytics/visualization LENS cluster Ewok cluster EVEREST visualization facility urika data appliance Beacon Peak performance 210 TF/s Memory Disk bandwidth 12 TB 56 GB/s Networks ESnet, 100 Gbps Darter Peak performance TF/s Internet2, 100 Gbps Memory 22.6 TB Private dark fibre 3 Disk bandwidth 30 GB/s
4 Our Compute and Data Environment for Science (CADES) provides a shared infrastructure to help solve big science problems Spallation Neutron Source Leadership Computing Facility Atmospheric Radiation Measurement Basic Energy Sciences Center for Nanophase Materials Sciences UT- CADES CADES ALICE, etc. Graph Analytics Cray GX Infrastructure (File systems/ Networking / etc.) XK7 Condos, Clusters, and Hybrid Clouds Shared- Memory Ultraviolet (UV) Future Technologies, Beyond Moore s Law 4
5 CADES connects the modes of discovery In silico investigation 1 Model Capturing the physical processes 2 Formulation Mapping to solvers and computing 3 Execution Forward process in simulation or iteration for convergence C Analytics at scale Machine learning B Alignment Data capture into staged structures A Experiment design Synthesis or control Empirical/experimental investigation 5
6 CADES Deployment.. and several ORNL projects on OIC.. and several other smaller projects. ~5000 Cores of Integrated Condos on Infiniband ~10,000 OIC Cores Attested PHI Enclave Integrated with UCAMS and XCAMS ~6000 Cores of Integrated Condos on Infiniband ~5000 Cores of Hybrid, Expandable Cloud SGI UV, Urika-GD/XA: GX 5PB+ High-Speed Storage ~3000 Cores of XK7 OIC CADES Moderate CADES Open Cray Condos Hybrid Cloud Unique Heterogeneous Platforms PHI Enclave Object store Large-Scale Storage 6 High-Speed Interconnects
7 Big Compute + Analytics (OLCF and CADES) coupled to Big Science Data Beam user tier BEAM web and data tier CADES Cades cluster computing HTTPS Storage MySQL Database Data/ artifacts Local High speed secure data transfer Scientific instrument tier Supercomputing tier Scanning Transmission Electron Microscopy (STEM) Scanning Tunneling Microscopy (STM) Scanning Probe Microscopy (SPM) IFIR/CNMS resources 7
8 Distributed cloud-based architecture DOE HPC cloud Titan, Edison, and Hopper Scanning probe microscope CADES VM web/data server CADES compute clusters CADES data storage 8
9 ORNL s computing ecosystem must integrate data analysis and simulation capabilities Simulation and data are critical to DOE Both need more computing capability Both have similar hardware technology requirements High bandwidth to memory Efficient processing Very fast I/O Different machine balance may be required Experiment Computing Theory Big data Analyzing and managing large complex data sets from experiments, observation, or simulation and sharing them with a community Simulation Used to implement theory; helps with understanding and prediction 9
10 2017 OLCF Leadership System The Smartest Supercomputer on the Planet Hybrid CPU/GPU architecture Vendor: IBM (Prime) / NVIDIA / Mellanox Technologies At least 5X Titan s Application Performance Total System Memory >6 PB HBM, DDR, and non-volatile Dual-rail Mellanox Infiniband full, non-blocking fat-tree interconnect IBM Elastic Storage (GPFS ) 2.5 TB/s I/O and 250 PB disk capacity Approximately 4,600 nodes, each with: Multiple IBM POWER9 CPUs and multiple NVIDIA Tesla GPUs using the NVIDIA Volta architecture CPUs and GPUs connected with high speed NVLink Large coherent memory: over 512 GB (HBM + DDR4) all directly addressable from the CPUs and GPUs An additional 800 GB of NVRAM, which can be configured as either a burst buffer or as extended memory over 40 TF peak performance 10
11 11 Summit will replace Titan as the OLCF s leadership supercomputer Many fewer nodes Much more powerful nodes Much more memory per node and total system memory Faster interconnect Much higher bandwidth between CPUs and GPUs Much larger and faster file system Feature Titan Summit Application performance Baseline 5-10x Titan Number of nodes 18,688 ~4,600 Node performance 1.4 TF > 40 TF Memory per node 38GB DDR3 + 6GB GDDR5 512 GB DDR4 + HBM NV memory per node GB Total system memory System interconnect (node injection bandwidth) 710 TB Gemini (6.4 GB/s) >6 PB DDR4 + HBM + non-volatile Dual Rail EDR-IB (23 GB/s) Or Dual Rail HDR-IB (48 GB/s) Interconnect topology 3D Torus Non-blocking Fat Tree Processors 1 AMD Opteron 1 NVIDIA Kepler 2 IBM POWER9 6 NVIDIA Volta File system 32 PB, 1 TB/s, Lustre 250 PB, 2.5 TB/s, GPFS Peak power consumption 9 MW 13 MW
12 12 ECP aims to transform the HPC ecosystem and make major contributions to the nation Develop applications that will tackle a broad spectrum of mission critical problems of unprecedented complexity with unprecedented performance Contribute to the economic competitiveness of the nation Support national security Develop a software stack, in collaboration with vendors, that is exascalecapable and is usable on smaller systems by industry and academia Train a large cadre of computational scientists, engineers, and computer scientists who will be an asset to the nation long after the end of ECP Partner with vendors to develop computer architectures that support exascale applications Revitalize the US HPC vendor industry Demonstrate the value of comprehensive co-design
13 The ECP Plan of Record A 7-year project that follows the holistic/co-design approach, which runs through 2023 (including 12 months of schedule contingency) Enable an initial exascale system based on advanced architecture and delivered in 2021 Enable capable exascale systems, based on ECP R&D, delivered in 2022 and deployed in 2023 as part of an NNSA and SC facility upgrades Acquisition of the exascale systems is outside of the ECP scope, will be carried out by DOE-SC and NNSA-ASC supercomputing facilities 13
14 Transition to higher trajectory with advanced architecture Computing Capability First exascale advanced architecture system Capable exascale systems 10X 5X Time
15 Reaching the elevated trajectory will require advanced and innovative architectures In order to reach the elevated trajectory, advanced architectures must be developed that make a big leap in Parallelism Memory and Storage Reliability Energy Consumption The exascale advanced architecture developments benefit all future U.S. systems on the higher trajectory In addition, the exascale advanced architecture will need to solve emerging data science and machine learning problems in addition to the traditional modeling and simulations applications. 15
16 ECP follows a holistic approach that uses co-design and integration to achieve capable exascale Resilience Workflows Application Development Software Technology Hardware Technology Exascale Systems Science and mission applications Scalable and productive software stack Hardware technology elements Integrated exascale supercomputers Correctness Applications Programming models, development environment, and System runtimes Software, resource management threading, scheduling, monitoring, and control Node OS, runtimes Visualization Math libraries and Frameworks Memory and Burst buffer Hardware interface Data Analysis Co-Design Tools Data managem ent I/O and file system ECP s work encompasses applications, system software, hardware technologies and architectures, and workforce development 16
17 Planned outcomes of the ECP Important applications running at exascale in 2021, producing useful results A full suite of mission and science applications ready to run on the 2023 capable exascale systems A large cadre of computational scientists, engineers, and computer scientists with deep expertise in exascale computing, who will be an asset to the nation long after the end of ECP An integrated software stack that supports exascale applications Results of PathForward R&D contracts with vendors that are integrated into exascale systems and are in vendors product roadmaps Industry and mission critical applications have been prepared for a more diverse and sophisticated set of computing technologies, carrying U.S. supercomputing well into the future 17
18 The Oak Ridge Leadership Computing Facility is on a well-defined path to exascale Since clock-rate scaling ended in 2003: HPC performance achieved through increased parallelism Titan and beyond: hierarchical parallelism with very powerful nodes Jaguar scaled to 300,000 cores MPI plus thread-level parallelism through OpenACC or OpenMP plus vectors Jaguar 2.3 PF Multi-core CPU 7 MW Titan: 27 PF Hybrid GPU/CPU 9 MW Summit 200PF Titan Hybrid GPU/CPU 13 MW OLCF Summit ~20-50 MW 18
19 Summary ORNL has a long history in highperformance computing for science, delivering many first-ofa-kind systems that were among the world s most powerful computers. We will continue this as a core-competency of the laboratory Delivering an ecosystem focused on the integration of computing and data into instruments of science and engineering This ecosystem delivers important, time-critical science with enormous impacts 19
20 20 Questions?
Preparing GPU-Accelerated Applications for the Summit Supercomputer
Preparing GPU-Accelerated Applications for the Summit Supercomputer Fernanda Foertter HPC User Assistance Group Training Lead foertterfs@ornl.gov This research used resources of the Oak Ridge Leadership
More informationOak Ridge National Laboratory Computing and Computational Sciences
Oak Ridge National Laboratory Computing and Computational Sciences OFA Update by ORNL Presented by: Pavel Shamis (Pasha) OFA Workshop Mar 17, 2015 Acknowledgments Bernholdt David E. Hill Jason J. Leverman
More informationPresent and Future Leadership Computers at OLCF
Present and Future Leadership Computers at OLCF Al Geist ORNL Corporate Fellow DOE Data/Viz PI Meeting January 13-15, 2015 Walnut Creek, CA ORNL is managed by UT-Battelle for the US Department of Energy
More informationManaging HPC Active Archive Storage with HPSS RAIT at Oak Ridge National Laboratory
Managing HPC Active Archive Storage with HPSS RAIT at Oak Ridge National Laboratory Quinn Mitchell HPC UNIX/LINUX Storage Systems ORNL is managed by UT-Battelle for the US Department of Energy U.S. Department
More informationExploring Emerging Technologies in the Extreme Scale HPC Co- Design Space with Aspen
Exploring Emerging Technologies in the Extreme Scale HPC Co- Design Space with Aspen Jeffrey S. Vetter SPPEXA Symposium Munich 26 Jan 2016 ORNL is managed by UT-Battelle for the US Department of Energy
More informationUCX: An Open Source Framework for HPC Network APIs and Beyond
UCX: An Open Source Framework for HPC Network APIs and Beyond Presented by: Pavel Shamis / Pasha ORNL is managed by UT-Battelle for the US Department of Energy Co-Design Collaboration The Next Generation
More informationTitan - Early Experience with the Titan System at Oak Ridge National Laboratory
Office of Science Titan - Early Experience with the Titan System at Oak Ridge National Laboratory Buddy Bland Project Director Oak Ridge Leadership Computing Facility November 13, 2012 ORNL s Titan Hybrid
More informationIBM HPC Technology & Strategy
IBM HPC Technology & Strategy Hyperion HPC User Forum Stuttgart, October 1st, 2018 The World s Smartest Supercomputers Klaus Gottschalk gottschalk@de.ibm.com HPC Strategy Deliver End to End Solutions for
More informationOak Ridge Leadership Computing Facility: Summit and Beyond
Oak Ridge Leadership Computing Facility: Summit and Beyond Justin L. Whitt OLCF-4 Deputy Project Director, Oak Ridge Leadership Computing Facility Oak Ridge National Laboratory March 2017 ORNL is managed
More informationNERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber
NERSC Site Update National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory Richard Gerber NERSC Senior Science Advisor High Performance Computing Department Head Cori
More informationMapping MPI+X Applications to Multi-GPU Architectures
Mapping MPI+X Applications to Multi-GPU Architectures A Performance-Portable Approach Edgar A. León Computer Scientist San Jose, CA March 28, 2018 GPU Technology Conference This work was performed under
More informationIBM CORAL HPC System Solution
IBM CORAL HPC System Solution HPC and HPDA towards Cognitive, AI and Deep Learning Deep Learning AI / Deep Learning Strategy for Power Power AI Platform High Performance Data Analytics Big Data Strategy
More informationPreparing Scientific Software for Exascale
Preparing Scientific Software for Exascale Jack Wells Director of Science Oak Ridge Leadership Computing Facility Oak Ridge National Laboratory Mini-Symposium on Scientific Software Engineering Monday,
More informationCenter for Accelerated Application Readiness. Summit. Tjerk Straatsma. Getting Applications Ready for. OLCF Scientific Computing Group
Center for Accelerated Application Readiness Getting Applications Ready for Summit Tjerk Straatsma OLCF Scientific Computing Group ORNL is managed by UT-Battelle for the US Department of Energy OLCF on
More information19. prosince 2018 CIIRC Praha. Milan Král, IBM Radek Špimr
19. prosince 2018 CIIRC Praha Milan Král, IBM Radek Špimr CORAL CORAL 2 CORAL Installation at ORNL CORAL Installation at LLNL Order of Magnitude Leap in Computational Power Real, Accelerated Science ACME
More informationHETEROGENEOUS HPC, ARCHITECTURAL OPTIMIZATION, AND NVLINK STEVE OBERLIN CTO, TESLA ACCELERATED COMPUTING NVIDIA
HETEROGENEOUS HPC, ARCHITECTURAL OPTIMIZATION, AND NVLINK STEVE OBERLIN CTO, TESLA ACCELERATED COMPUTING NVIDIA STATE OF THE ART 2012 18,688 Tesla K20X GPUs 27 PetaFLOPS FLAGSHIP SCIENTIFIC APPLICATIONS
More informationAllowing Users to Run Services at the OLCF with Kubernetes
Allowing Users to Run Services at the OLCF with Kubernetes Jason Kincl Senior HPC Systems Engineer Ryan Adamson Senior HPC Security Engineer This work was supported by the Oak Ridge Leadership Computing
More informationLeadership Computing Directions at Oak Ridge National Laboratory: Navigating the Transition to Heterogeneous Architectures
Leadership Computing Directions at Oak Ridge National Laboratory: Navigating the Transition to Heterogeneous Architectures International Symposium on Computing in Atmospheric Sciences September 8-12, 2013
More informationOrganizational Update: December 2015
Organizational Update: December 2015 David Hudak Doug Johnson Alan Chalker www.osc.edu Slide 1 OSC Organizational Update Leadership changes State of OSC Roadmap Web app demonstration (if time) Slide 2
More informationThe Stampede is Coming Welcome to Stampede Introductory Training. Dan Stanzione Texas Advanced Computing Center
The Stampede is Coming Welcome to Stampede Introductory Training Dan Stanzione Texas Advanced Computing Center dan@tacc.utexas.edu Thanks for Coming! Stampede is an exciting new system of incredible power.
More informationResponsive Large Data Analysis and Visualization with the ParaView Ecosystem. Patrick O Leary, Kitware Inc
Responsive Large Data Analysis and Visualization with the ParaView Ecosystem Patrick O Leary, Kitware Inc Hybrid Computing Attribute Titan Summit - 2018 Compute Nodes 18,688 ~3,400 Processor (1) 16-core
More informationThe Cambridge Bio-Medical-Cloud An OpenStack platform for medical analytics and biomedical research
The Cambridge Bio-Medical-Cloud An OpenStack platform for medical analytics and biomedical research Dr Paul Calleja Director of Research Computing University of Cambridge Global leader in science & technology
More informationCRAY XK6 REDEFINING SUPERCOMPUTING. - Sanjana Rakhecha - Nishad Nerurkar
CRAY XK6 REDEFINING SUPERCOMPUTING - Sanjana Rakhecha - Nishad Nerurkar CONTENTS Introduction History Specifications Cray XK6 Architecture Performance Industry acceptance and applications Summary INTRODUCTION
More informationProgramming NVM Systems
Programming NVM Systems Random Access Talk Jeffrey S. Vetter Seyong Lee, Joel Denny, Jungwon Kim, et al. Presented to Salishan Conference on High Speed Computing Gleneden Beach, Oregon 27 Apr 2016 ORNL
More informationInterconnect Your Future
#OpenPOWERSummit Interconnect Your Future Scot Schultz, Director HPC / Technical Computing Mellanox Technologies OpenPOWER Summit, San Jose CA March 2015 One-Generation Lead over the Competition Mellanox
More informationIBM Spectrum Scale IO performance
IBM Spectrum Scale 5.0.0 IO performance Silverton Consulting, Inc. StorInt Briefing 2 Introduction High-performance computing (HPC) and scientific computing are in a constant state of transition. Artificial
More informationMICROWAY S NVIDIA TESLA V100 GPU SOLUTIONS GUIDE
MICROWAY S NVIDIA TESLA V100 GPU SOLUTIONS GUIDE LEVERAGE OUR EXPERTISE sales@microway.com http://microway.com/tesla NUMBERSMASHER TESLA 4-GPU SERVER/WORKSTATION Flexible form factor 4 PCI-E GPUs + 3 additional
More informationGPU Architecture. Alan Gray EPCC The University of Edinburgh
GPU Architecture Alan Gray EPCC The University of Edinburgh Outline Why do we want/need accelerators such as GPUs? Architectural reasons for accelerator performance advantages Latest GPU Products From
More informationOverview. CS 472 Concurrent & Parallel Programming University of Evansville
Overview CS 472 Concurrent & Parallel Programming University of Evansville Selection of slides from CIS 410/510 Introduction to Parallel Computing Department of Computer and Information Science, University
More informationThe Titan Tools Experience
The Titan Tools Experience Michael J. Brim, Ph.D. Computer Science Research, CSMD/NCCS Petascale Tools Workshop 213 Madison, WI July 15, 213 Overview of Titan Cray XK7 18,688+ compute nodes 16-core AMD
More informationIBM Power Systems HPC Cluster
IBM Power Systems HPC Cluster Highlights Complete and fully Integrated HPC cluster for demanding workloads Modular and Extensible: match components & configurations to meet demands Integrated: racked &
More informationMonash High Performance Computing
MONASH eresearch Monash High Performance Computing Gin Tan Senior HPC Consultant MeRC (Monash eresearch) Monash HPC Infrastructure MASSIVE MonARCH Characterisation VL and Instruments MASSIVE-3 MeRC Infrastructure
More informationPhilip C. Roth. Computer Science and Mathematics Division Oak Ridge National Laboratory
Philip C. Roth Computer Science and Mathematics Division Oak Ridge National Laboratory A Tree-Based Overlay Network (TBON) like MRNet provides scalable infrastructure for tools and applications MRNet's
More informationoutthink limits Spectrum Scale Enhancements for CORAL Sarp Oral, Oak Ridge National Laboratory Gautam Shah, IBM
outthink limits Spectrum Scale Enhancements for CORAL Sarp Oral, Oak Ridge National Laboratory Gautam Shah, IBM What is CORAL Collaboration of DOE Oak Ridge, Argonne, and Lawrence Livermore National Labs
More informationIllinois Proposal Considerations Greg Bauer
- 2016 Greg Bauer Support model Blue Waters provides traditional Partner Consulting as part of its User Services. Standard service requests for assistance with porting, debugging, allocation issues, and
More informationOak Ridge National Laboratory Computing and Computational Sciences
Oak Ridge National Laboratory Computing and Computational Sciences Preparing OpenSHMEM for Exascale Presented by: Pavel Shamis (Pasha) HPC Advisory Council Stanford Conference California Feb 2, 2015 Outline
More informationNVIDIA Update and Directions on GPU Acceleration for Earth System Models
NVIDIA Update and Directions on GPU Acceleration for Earth System Models Stan Posey, HPC Program Manager, ESM and CFD, NVIDIA, Santa Clara, CA, USA Carl Ponder, PhD, Applications Software Engineer, NVIDIA,
More informationStore Process Analyze Collaborate Archive Cloud The HPC Storage Leader Invent Discover Compete
Store Process Analyze Collaborate Archive Cloud The HPC Storage Leader Invent Discover Compete 1 DDN Who We Are 2 We Design, Deploy and Optimize Storage Systems Which Solve HPC, Big Data and Cloud Business
More informationABySS Performance Benchmark and Profiling. May 2010
ABySS Performance Benchmark and Profiling May 2010 Note The following research was performed under the HPC Advisory Council activities Participating vendors: AMD, Dell, Mellanox Compute resource - HPC
More informationJohn Unthank IBM Federal Sales IBM HPC Topics IBM Corporation
John Unthank IBM Federal Sales unthank@us.ibm.com IBM HPC Topics IBM Imperatives* Transform industries and professions with data Remake enterprise IT for the cloud Reimagine work through mobile and social
More informationGPFS Experiences from the Argonne Leadership Computing Facility (ALCF) William (Bill) E. Allcock ALCF Director of Operations
GPFS Experiences from the Argonne Leadership Computing Facility (ALCF) William (Bill) E. Allcock ALCF Director of Operations Argonne National Laboratory Argonne National Laboratory is located on 1,500
More informationInauguration Cartesius June 14, 2013
Inauguration Cartesius June 14, 2013 Hardware is Easy...but what about software/applications/implementation/? Dr. Peter Michielse Deputy Director 1 Agenda History Cartesius Hardware path to exascale: the
More informationThe Stampede is Coming: A New Petascale Resource for the Open Science Community
The Stampede is Coming: A New Petascale Resource for the Open Science Community Jay Boisseau Texas Advanced Computing Center boisseau@tacc.utexas.edu Stampede: Solicitation US National Science Foundation
More informationInterconnect Your Future
Interconnect Your Future Gilad Shainer 2nd Annual MVAPICH User Group (MUG) Meeting, August 2014 Complete High-Performance Scalable Interconnect Infrastructure Comprehensive End-to-End Software Accelerators
More informationStepping up to Summit
DEPARTMENT: Leadership Computing Stepping up to Summit Jonathan Hines Oak Ridge National Laboratory Editors: James J. Hack, jhack@ornl.gov; Michael E. Papka, papka@anl.gov In November 2014, the Oak Ridge
More informationCS500 SMARTER CLUSTER SUPERCOMPUTERS
CS500 SMARTER CLUSTER SUPERCOMPUTERS OVERVIEW Extending the boundaries of what you can achieve takes reliable computing tools matched to your workloads. That s why we tailor the Cray CS500 cluster supercomputer
More informationOverlapping Computation and Communication for Advection on Hybrid Parallel Computers
Overlapping Computation and Communication for Advection on Hybrid Parallel Computers James B White III (Trey) trey@ucar.edu National Center for Atmospheric Research Jack Dongarra dongarra@eecs.utk.edu
More informationCryogenic Computing Complexity (C3) Marc Manheimer December 9, 2015 IEEE Rebooting Computing Summit 4
Cryogenic Computing Complexity (C3) Marc Manheimer marc.manheimer@iarpa.gov December 9, 2015 IEEE Rebooting Computing Summit 4 C3 for the Workshop Review of the C3 program Motivation Technical challenges
More informationToward portable I/O performance by leveraging system abstractions of deep memory and interconnect hierarchies
Toward portable I/O performance by leveraging system abstractions of deep memory and interconnect hierarchies François Tessier, Venkatram Vishwanath, Paul Gressier Argonne National Laboratory, USA Wednesday
More informationThe Future of High- Performance Computing
Lecture 26: The Future of High- Performance Computing Parallel Computer Architecture and Programming CMU 15-418/15-618, Spring 2017 Comparing Two Large-Scale Systems Oakridge Titan Google Data Center Monolithic
More informationInfiniBand Strengthens Leadership as the Interconnect Of Choice By Providing Best Return on Investment. TOP500 Supercomputers, June 2014
InfiniBand Strengthens Leadership as the Interconnect Of Choice By Providing Best Return on Investment TOP500 Supercomputers, June 2014 TOP500 Performance Trends 38% CAGR 78% CAGR Explosive high-performance
More informationPredicting Service Outage Using Machine Learning Techniques. HPE Innovation Center
Predicting Service Outage Using Machine Learning Techniques HPE Innovation Center HPE Innovation Center - Our AI Expertise Sense Learn Comprehend Act Computer Vision Machine Learning Natural Language Processing
More informationMining Supercomputer Jobs' I/O Behavior from System Logs. Xiaosong Ma
Mining Supercomputer Jobs' I/O Behavior from System Logs Xiaosong Ma OLCF Architecture Overview Rhea node Development Cluster Eos 76 Node Cray XC Cluster Scalable IO Network (SION) - Infiniband Servers
More informationShort Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy
Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy François Tessier, Venkatram Vishwanath Argonne National Laboratory, USA July 19,
More informationUmeå University
HPC2N @ Umeå University Introduction to HPC2N and Kebnekaise Jerry Eriksson, Pedro Ojeda-May, and Birgitte Brydsö Outline Short presentation of HPC2N HPC at a glance. HPC2N Abisko, Kebnekaise HPC Programming
More informationUmeå University
HPC2N: Introduction to HPC2N and Kebnekaise, 2017-09-12 HPC2N @ Umeå University Introduction to HPC2N and Kebnekaise Jerry Eriksson, Pedro Ojeda-May, and Birgitte Brydsö Outline Short presentation of HPC2N
More informationLeveraging Software-Defined Storage to Meet Today and Tomorrow s Infrastructure Demands
Leveraging Software-Defined Storage to Meet Today and Tomorrow s Infrastructure Demands Unleash Your Data Center s Hidden Power September 16, 2014 Molly Rector CMO, EVP Product Management & WW Marketing
More informationTrends in HPC (hardware complexity and software challenges)
Trends in HPC (hardware complexity and software challenges) Mike Giles Oxford e-research Centre Mathematical Institute MIT seminar March 13th, 2013 Mike Giles (Oxford) HPC Trends March 13th, 2013 1 / 18
More informationIBM Deep Learning Solutions
IBM Deep Learning Solutions Reference Architecture for Deep Learning on POWER8, P100, and NVLink October, 2016 How do you teach a computer to Perceive? 2 Deep Learning: teaching Siri to recognize a bicycle
More informationThe Future of High- Performance Computing
Lecture 24: The Future of High- Performance Computing Parallel Computer Architecture and Programming CMU 15-418/15-618, Spring 2016 Executive Order July 29, 2015 EXECUTIVE ORDER - - - - - - - CREATING
More informationPaving the Road to Exascale
Paving the Road to Exascale Gilad Shainer August 2015, MVAPICH User Group (MUG) Meeting The Ever Growing Demand for Performance Performance Terascale Petascale Exascale 1 st Roadrunner 2000 2005 2010 2015
More informationGateways to Discovery: Cyberinfrastructure for the Long Tail of Science
Gateways to Discovery: Cyberinfrastructure for the Long Tail of Science ECSS Symposium, 12/16/14 M. L. Norman, R. L. Moore, D. Baxter, G. Fox (Indiana U), A Majumdar, P Papadopoulos, W Pfeiffer, R. S.
More informationAutomatic Identification of Application I/O Signatures from Noisy Server-Side Traces. Yang Liu Raghul Gunasekaran Xiaosong Ma Sudharshan S.
Automatic Identification of Application I/O Signatures from Noisy Server-Side Traces Yang Liu Raghul Gunasekaran Xiaosong Ma Sudharshan S. Vazhkudai Instance of Large-Scale HPC Systems ORNL s TITAN (World
More informationJÜLICH SUPERCOMPUTING CENTRE Site Introduction Michael Stephan Forschungszentrum Jülich
JÜLICH SUPERCOMPUTING CENTRE Site Introduction 09.04.2018 Michael Stephan JSC @ Forschungszentrum Jülich FORSCHUNGSZENTRUM JÜLICH Research Centre Jülich One of the 15 Helmholtz Research Centers in Germany
More informationExecution Models for the Exascale Era
Execution Models for the Exascale Era Nicholas J. Wright Advanced Technology Group, NERSC/LBNL njwright@lbl.gov Programming weather, climate, and earth- system models on heterogeneous muli- core plajorms
More informationPowering)the)Road)to)National)HPC)Leadership)
Powering)the)Road)to)National)HPC)Leadership) Jack%C.%Wells,%Director%of%Science Oak$Ridge$Leadership$Computing$Facility/Oak$Ridge$National$Laboratory Join%the%Conversation%#OpenPOWERSummit Powering$the$Road$to$National$HPC$
More informationHPC future trends from a science perspective
HPC future trends from a science perspective Simon McIntosh-Smith University of Bristol HPC Research Group simonm@cs.bris.ac.uk 1 Business as usual? We've all got used to new machines being relatively
More informationDeep Learning mit PowerAI - Ein Überblick
Stephen Lutz Deep Learning mit PowerAI - Open Group Master Certified IT Specialist Technical Sales IBM Cognitive Infrastructure IBM Germany Ein Überblick Stephen.Lutz@de.ibm.com What s that? and what s
More informationIntroduction to FREE National Resources for Scientific Computing. Dana Brunson. Jeff Pummill
Introduction to FREE National Resources for Scientific Computing Dana Brunson Oklahoma State University High Performance Computing Center Jeff Pummill University of Arkansas High Peformance Computing Center
More informationIntroduction to HPC Parallel I/O
Introduction to HPC Parallel I/O Feiyi Wang (Ph.D.) and Sarp Oral (Ph.D.) Technology Integration Group Oak Ridge Leadership Computing ORNL is managed by UT-Battelle for the US Department of Energy Outline
More informationCori (2016) and Beyond Ensuring NERSC Users Stay Productive
Cori (2016) and Beyond Ensuring NERSC Users Stay Productive Nicholas J. Wright! Advanced Technologies Group Lead! Heterogeneous Mul-- Core 4 Workshop 17 September 2014-1 - NERSC Systems Today Edison: 2.39PF,
More informationEarly Experiences Writing Performance Portable OpenMP 4 Codes
Early Experiences Writing Performance Portable OpenMP 4 Codes Verónica G. Vergara Larrea Wayne Joubert M. Graham Lopez Oscar Hernandez Oak Ridge National Laboratory Problem statement APU FPGA neuromorphic
More informationS THE MAKING OF DGX SATURNV: BREAKING THE BARRIERS TO AI SCALE. Presenter: Louis Capps, Solution Architect, NVIDIA,
S7750 - THE MAKING OF DGX SATURNV: BREAKING THE BARRIERS TO AI SCALE Presenter: Louis Capps, Solution Architect, NVIDIA, lcapps@nvidia.com A TALE OF ENLIGHTENMENT Basic OK List 10 for x = 1 to 3 20 print
More informationS8765 Performance Optimization for Deep- Learning on the Latest POWER Systems
S8765 Performance Optimization for Deep- Learning on the Latest POWER Systems Khoa Huynh Senior Technical Staff Member (STSM), IBM Jonathan Samn Software Engineer, IBM Evolving from compute systems to
More informationBuilding NVLink for Developers
Building NVLink for Developers Unleashing programmatic, architectural and performance capabilities for accelerated computing Why NVLink TM? Simpler, Better and Faster Simplified Programming No specialized
More informationOverview of Tianhe-2
Overview of Tianhe-2 (MilkyWay-2) Supercomputer Yutong Lu School of Computer Science, National University of Defense Technology; State Key Laboratory of High Performance Computing, China ytlu@nudt.edu.cn
More informationThe Future of High Performance Interconnects
The Future of High Performance Interconnects Ashrut Ambastha HPC Advisory Council Perth, Australia :: August 2017 When Algorithms Go Rogue 2017 Mellanox Technologies 2 When Algorithms Go Rogue 2017 Mellanox
More informationDDN About Us Solving Large Enterprise and Web Scale Challenges
1 DDN About Us Solving Large Enterprise and Web Scale Challenges History Founded in 98 World s Largest Private Storage Company Growing, Profitable, Self Funded Headquarters: Santa Clara and Chatsworth,
More informationDVS, GPFS and External Lustre at NERSC How It s Working on Hopper. Tina Butler, Rei Chi Lee, Gregory Butler 05/25/11 CUG 2011
DVS, GPFS and External Lustre at NERSC How It s Working on Hopper Tina Butler, Rei Chi Lee, Gregory Butler 05/25/11 CUG 2011 1 NERSC is the Primary Computing Center for DOE Office of Science NERSC serves
More informationNVIDIA DGX SYSTEMS PURPOSE-BUILT FOR AI
NVIDIA DGX SYSTEMS PURPOSE-BUILT FOR AI Overview Unparalleled Value Product Portfolio Software Platform From Desk to Data Center to Cloud Summary AI researchers depend on computing performance to gain
More informationBig Data 2015: Sponsor and Participants Research Event ""
Big Data 2015: Sponsor and Participants Research Event "" Center for Large-scale Data Systems Research, CLDS! San Diego Supercomputer Center! UC San Diego! Agenda" Welcome and introductions! SDSC: Who
More informationA More Realistic Way of Stressing the End-to-end I/O System
A More Realistic Way of Stressing the End-to-end I/O System Verónica G. Vergara Larrea Sarp Oral Dustin Leverman Hai Ah Nam Feiyi Wang James Simmons CUG 2015 April 29, 2015 Chicago, IL ORNL is managed
More informationPower Systems AC922 Overview. Chris Mann IBM Distinguished Engineer Chief System Architect, Power HPC Systems December 11, 2017
Power Systems AC922 Overview Chris Mann IBM Distinguished Engineer Chief System Architect, Power HPC Systems December 11, 2017 IBM POWER HPC Platform Strategy High-performance computer and high-performance
More informationThe Role of InfiniBand Technologies in High Performance Computing. 1 Managed by UT-Battelle for the Department of Energy
The Role of InfiniBand Technologies in High Performance Computing 1 Managed by UT-Battelle Contributors Gil Bloch Noam Bloch Hillel Chapman Manjunath Gorentla- Venkata Richard Graham Michael Kagan Vasily
More informationOLCF's next- genera0on Spider file system
OLCF's next- genera0on Spider file system Sarp Oral, PhD File and Storage Systems Team Lead Technology Integra0on Group Oak Ridge Leadership Compu0ng Facility Oak Ridge Na0onal Laboratory April 18, 2013
More informationTACC s Stampede Project: Intel MIC for Simulation and Data-Intensive Computing
TACC s Stampede Project: Intel MIC for Simulation and Data-Intensive Computing Jay Boisseau, Director April 17, 2012 TACC Vision & Strategy Provide the most powerful, capable computing technologies and
More informationMass Big Data: Progressive Growth through Strategic Collaboration
Massachusetts Technology Collaborative Mass Big Data: Progressive Growth through Strategic Collaboration Patrick Larkin, Executive Director The Innovation Institute at the Massachusetts Technology Collaborative
More informationComputational Challenges and Opportunities for Nuclear Astrophysics
Computational Challenges and Opportunities for Nuclear Astrophysics Bronson Messer Acting Group Leader Scientific Computing Group National Center for Computational Sciences Theoretical Astrophysics Group
More informationin Action Fujitsu High Performance Computing Ecosystem Human Centric Innovation Innovation Flexibility Simplicity
Fujitsu High Performance Computing Ecosystem Human Centric Innovation in Action Dr. Pierre Lagier Chief Technology Officer Fujitsu Systems Europe Innovation Flexibility Simplicity INTERNAL USE ONLY 0 Copyright
More informationPortable and Productive Performance with OpenACC Compilers and Tools. Luiz DeRose Sr. Principal Engineer Programming Environments Director Cray Inc.
Portable and Productive Performance with OpenACC Compilers and Tools Luiz DeRose Sr. Principal Engineer Programming Environments Director Cray Inc. 1 Cray: Leadership in Computational Research Earth Sciences
More informationMELLANOX EDR UPDATE & GPUDIRECT MELLANOX SR. SE 정연구
MELLANOX EDR UPDATE & GPUDIRECT MELLANOX SR. SE 정연구 Leading Supplier of End-to-End Interconnect Solutions Analyze Enabling the Use of Data Store ICs Comprehensive End-to-End InfiniBand and Ethernet Portfolio
More informationThe Arm Technology Ecosystem: Current Products and Future Outlook
The Arm Technology Ecosystem: Current Products and Future Outlook Dan Ernst, PhD Advanced Technology Cray, Inc. Why is an Ecosystem Important? An Ecosystem is a collection of common material Developed
More informationThe Hopper System: How the Largest* XE6 in the World Went From Requirements to Reality! Katie Antypas, Tina Butler, and Jonathan Carter
The Hopper System: How the Largest* XE6 in the World Went From Requirements to Reality! Katie Antypas, Tina Butler, and Jonathan Carter CUG 2011, May 25th, 2011 1 Requirements to Reality Develop RFP Select
More informationExascale: challenges and opportunities in a power constrained world
Exascale: challenges and opportunities in a power constrained world Carlo Cavazzoni c.cavazzoni@cineca.it SuperComputing Applications and Innovation Department CINECA CINECA non profit Consortium, made
More informationTuning I/O Performance for Data Intensive Computing. Nicholas J. Wright. lbl.gov
Tuning I/O Performance for Data Intensive Computing. Nicholas J. Wright njwright @ lbl.gov NERSC- National Energy Research Scientific Computing Center Mission: Accelerate the pace of scientific discovery
More informationHybrid KAUST Many Cores and OpenACC. Alain Clo - KAUST Research Computing Saber Feki KAUST Supercomputing Lab Florent Lebeau - CAPS
+ Hybrid Computing @ KAUST Many Cores and OpenACC Alain Clo - KAUST Research Computing Saber Feki KAUST Supercomputing Lab Florent Lebeau - CAPS + Agenda Hybrid Computing n Hybrid Computing n From Multi-Physics
More informationPorting Scientific Applications to OpenPOWER
Porting Scientific Applications to OpenPOWER Dirk Pleiter Forschungszentrum Jülich / JSC #OpenPOWERSummit Join the conversation at #OpenPOWERSummit 1 JSC s HPC Strategy IBM Power 6 JUMP, 9 TFlop/s Intel
More informationThe Future of Interconnect Technology
The Future of Interconnect Technology Michael Kagan, CTO HPC Advisory Council Stanford, 2014 Exponential Data Growth Best Interconnect Required 44X 0.8 Zetabyte 2009 35 Zetabyte 2020 2014 Mellanox Technologies
More informationImproved Solutions for I/O Provisioning and Application Acceleration
1 Improved Solutions for I/O Provisioning and Application Acceleration August 11, 2015 Jeff Sisilli Sr. Director Product Marketing jsisilli@ddn.com 2 Why Burst Buffer? The Supercomputing Tug-of-War A supercomputer
More informationIt s a Multicore World. John Urbanic Pittsburgh Supercomputing Center
It s a Multicore World John Urbanic Pittsburgh Supercomputing Center Waiting for Moore s Law to save your serial code start getting bleak in 2004 Source: published SPECInt data Moore s Law is not at all
More information