HECToR. UK National Supercomputing Service. Andy Turner & Chris Johnson
|
|
- Linette Spencer
- 5 years ago
- Views:
Transcription
1 HECToR UK National Supercomputing Service Andy Turner & Chris Johnson
2 Outline EPCC HECToR Introduction HECToR Phase 3 Introduction to AMD Bulldozer Architecture Performance Application placement the hardware really matters CP2K in PRACE Simulation at exascale What software is used on HECToR? Exascale software challenges What does this mean for a HPC users? 2
3 EPCC Founded in 1990 Based at The University of Edinburgh, within the School of Physics and Astronomy leading European centre of expertise in advanced research training and technology transfer supercomputer services to academia and business 95% of our funding from external sources Heavily involved in European projects such as PRACE and HPC-Europa
4 PRACE and HPC-Europa PRACE DECI - a resource exchange programme projects can access several million CPU-hours of compute resource on machines across Europe HPC-Europa visitor programme Visitors can visit one of 7 countries: Italy, UK, Spain, Germany, France, The Netherlands or Finland Find a host in an academic department and HPC-Europa provides travel, subsistence and access to HPC resources
5 HECToR 5
6 HECToR Partners RCUK UK Research funding councils UoE HPCx Ltd./EPCC System host and operator Cray Inc. System provider NAG Ltd. Computational science and engineering support 6
7 HECToR Details UK National HPC Service PRACE Tier-1 machine Currently 30-cabinet Cray XE6 system 2816 nodes, 90,112 cores Each node has 2 16-core AMD Opterons (2.3GHz Interlagos) 32 GB memory Peak of over 800 TF 90 TB of memory 7
8 HECToR Service Compute nodes Login nodes Lustre OSS Lustre MDS NFS Server Boot/SDB node Cray XE6 Supercomputer 1 GigE Backbone Infiniband Switch 10 GigE Backup and Archive Servers esfs Lustre high-performance, parallel filesystem 8
9 HECToR Compute Nodes All dies link to memory, interconnect and each other by HyperTransport. Nodes arranged in 3D torus. Interconnect supports message passing and DRMA in hardware. Interconnect supports MPI, SHMEM, PGAS, ARMCI. Image courtesy of NAG. 9
10 AMD Bulldozer Architecture Image courtesy of Wikipedia 10
11 Dual-core Interlagos module Image courtesy of NAG.
12 Phase 3 Performance Comparison 12
13 Task placement matters 13
14 Taks placement
15 CP2K Improving scaling 15
16 CP2K: Overview CP2K is a freely available (GPL) Density Functional Theory code (+ support for classical, empirical potentials) can perform MD, MC, geometry optimisation, normal mode calculations The Swiss Army Knife of Molecular Simulation (VandeVondele) c.f. CASTEP, VASP, CPMD etc. 16
17 CP2K: Overview CP2K is a freely available (GPL) Density Functional Theory code (+ support for classical, empirical potentials) can perform MD, MC, geometry optimisation, normal mode calculations The Swiss Army Knife of Molecular Simulation (VandeVondele) c.f. CASTEP, VASP, CPMD etc. 17
18 CP2K million atom KS-DFT Focussing on CP2K on BlueGene/P (reducing memory usage) scaling to 1,000,000 atoms (estimated as 200,000 cores) Led by Iain Bethune at EPCC Supported by Dr. Joost VandeVondele et al, CP2K Developers at Physical Chemistry Institute, University of Zurich Work done under dcse and PRACE Improved scaling via increased use of OpenMP directives
19 CP2K mixed mode Performance improvement due to: Reduce impact of algorithms which scale poorly with number of MPI tasks E.g. When using T threads, switchover point from 1D decomposed FFT (more efficient) to 2D decomposed FFT (less efficient) is increased by a factor of T Improved load balancing Existing MPI load balancing algorithms do a coarser load balance, finegrained balance done over OpenMP threads Reduced number of messages significantly Especially on pre-gemini networks For all-to-all communications, message count reduced by factor oft2
20 CP2K: Functional Evaluation 93% efficiency with 6 threads, 74% with 24 threads Mixed Mode Parallelism in CP2K: A Case Study 20
21 CP2K: Fast Fourier Transforms CP2K uses a 3D Fourier Transform to turn real data on the plane wave grids into g-space data on the plane wave grids. The grids may be distributed as planes, or rays (pencils) so the FFT may involve one or two transpose steps between the 3 1D FFT operations The 1D FFTs are performed via an interface which supports many libraries e.g. FFTW 2/3 ESSL, ACML, CUDA, FFTSG (in-built) 21
22 CP2K: Fast Fourier Transforms We can parallelise two parts with OpenMP 1D FFT assign each thread a subset of rows to FFT Buffer packing threads cooperatively pack the buffers which are passed to MPI Communication still handled outside a the parallel regions 22
23 Simulation at Exascale Software Edinburgh/Tsukuba Workshop, February
24 Scientific Software Chemistry, materials science, climate, oceanography, engineering, plasma physics, paleontology Dye-sensitised solar cells F. Schiffmann and J. VandeVondele University of Zurich Modelling dinosaur gaits Dr Bill Sellers, University of Manchester Fractal-based models of turbulent flows Christos Vassilicos & Sylvain Laizet, Imperial College Edinburgh/Tsukuba Workshop, February 2012
25 Scientific Usage Profile HECToRXT4 Chemistry/Materials Science, Chemistry/Materials Science her/unknown, Earth Science/Climate Physics Engineering Other/Unknown Engineering, 1.91 Physics, 6 Earth Science/Climate, 16.6 Edinburgh/Tsukuba Workshop, February 2012
26 HELIUM 2.9% NAMD 3.3% CP2KNEMO (Hybrid) LAMMPS Fluidity Quantum Espresso DL_POLY ChemShell 0.1% 0.1% 0.7% SENGA 1.0% 0.9% 1.2% 1.7% Terra 1.9% Shelf 2.3% 2.7% CP2K (MPI) 4.5% Others 45.7% CASTEP 5.9% UM 6.4% VASP 17.5% HECToRXT4
27 Future Look What does the future hold for HPC and the national facility? System Perf. 20 PFlops PFlops 1 EFlops Memory 1 PB 5 PB 10 PB Node Perf. 200 GFlops 400 GFlops 1-10 TFlops Concurrency 32 O(100) O(1000) Interconnect BW 40 GB/s 100 GB/s GB/s Nodes 100, ,000 O(Million) I/O 2 TB/s 10 TB/s 20 TB/s MTTI Days Days O(1 Day) Power 10 MW 10 MW 20 MW Accelerators: GPGPUs Edinburgh/Tsukuba Workshop, February 2012
28 Application sustainability National-scale HPC facilities provide a capability resource. For users who want to run calculations that are too large for other resources In reality, in the UK, also gets used for smaller-scale calculations The future of national-scale HPC (as for everyone else): Lots of cores per node (CPU + co-processor) Little memory per core Lots of compute power per network interface The balance of compute to communication power and compute to memory are both radically different to now Need to ensure UK researchers have software that can exploit these resources effectively Edinburgh/Tsukuba Workshop, February
29 Application sustainability Requirements for software on future capability HPC resources: Probably cannot be pure message passing parallel This will not scale on nodes with high amount of compute Must exploit all parallelism at all levels vectorisation, shared-memory, message-passing Must exploit memory hierarchy efficiently Must harness the co-processors/lightweight cores Must be fault-tolerant None of today s large codes meet all these requirements Edinburgh/Tsukuba Workshop, February
High Performance Computing. What is it used for and why?
High Performance Computing What is it used for and why? Overview What is it used for? Drivers for HPC Examples of usage Why do you need to learn the basics? Hardware layout and structure matters Serial
More informationHPC Issues for DFT Calculations. Adrian Jackson EPCC
HC Issues for DFT Calculations Adrian Jackson ECC Scientific Simulation Simulation fast becoming 4 th pillar of science Observation, Theory, Experimentation, Simulation Explore universe through simulation
More informationPORTING CP2K TO THE INTEL XEON PHI. ARCHER Technical Forum, Wed 30 th July Iain Bethune
PORTING CP2K TO THE INTEL XEON PHI ARCHER Technical Forum, Wed 30 th July Iain Bethune (ibethune@epcc.ed.ac.uk) Outline Xeon Phi Overview Porting CP2K to Xeon Phi Performance Results Lessons Learned Further
More informationTitan - Early Experience with the Titan System at Oak Ridge National Laboratory
Office of Science Titan - Early Experience with the Titan System at Oak Ridge National Laboratory Buddy Bland Project Director Oak Ridge Leadership Computing Facility November 13, 2012 ORNL s Titan Hybrid
More informationInauguration Cartesius June 14, 2013
Inauguration Cartesius June 14, 2013 Hardware is Easy...but what about software/applications/implementation/? Dr. Peter Michielse Deputy Director 1 Agenda History Cartesius Hardware path to exascale: the
More informationPreparing GPU-Accelerated Applications for the Summit Supercomputer
Preparing GPU-Accelerated Applications for the Summit Supercomputer Fernanda Foertter HPC User Assistance Group Training Lead foertterfs@ornl.gov This research used resources of the Oak Ridge Leadership
More informationHPC IN EUROPE. Organisation of public HPC resources
HPC IN EUROPE Organisation of public HPC resources Context Focus on publicly-funded HPC resources provided primarily to enable scientific research and development at European universities and other publicly-funded
More informationCALMIP : HIGH PERFORMANCE COMPUTING
CALMIP : HIGH PERFORMANCE COMPUTING Nicolas.renon@univ-tlse3.fr Emmanuel.courcelle@inp-toulouse.fr CALMIP (UMS 3667) Espace Clément Ader www.calmip.univ-toulouse.fr CALMIP :Toulouse University Computing
More informationAn Introduction to OpenACC
An Introduction to OpenACC Alistair Hart Cray Exascale Research Initiative Europe 3 Timetable Day 1: Wednesday 29th August 2012 13:00 Welcome and overview 13:15 Session 1: An Introduction to OpenACC 13:15
More informationExascale Challenges and Applications Initiatives for Earth System Modeling
Exascale Challenges and Applications Initiatives for Earth System Modeling Workshop on Weather and Climate Prediction on Next Generation Supercomputers 22-25 October 2012 Tom Edwards tedwards@cray.com
More informationIt s a Multicore World. John Urbanic Pittsburgh Supercomputing Center
It s a Multicore World John Urbanic Pittsburgh Supercomputing Center Waiting for Moore s Law to save your serial code start getting bleak in 2004 Source: published SPECInt data Moore s Law is not at all
More informationHECToR. The new UK National High Performance Computing Service. Dr Mark Parsons Commercial Director, EPCC
HECToR The new UK National High Performance Computing Service Dr Mark Parsons Commercial Director, EPCC m.parsons@epcc.ed.ac.uk +44 131 650 5022 Summary Why we need supercomputers The HECToR Service Technologies
More informationHECToR Annual Report January 31 December Issue: 1.0. HECToR_Annual_2011 Page 1 of 59
HECToR Annual Report 2011 01 January 31 December 2011 Issue: 1.0 HECToR_Annual_2011 Page 1 of 59 Contents 1 Introduction... 5 2 Executive Summary... 5 3 Quantitative Metrics... 7 3.1 Reliability... 7 3.1.1
More informationUmeå University
HPC2N @ Umeå University Introduction to HPC2N and Kebnekaise Jerry Eriksson, Pedro Ojeda-May, and Birgitte Brydsö Outline Short presentation of HPC2N HPC at a glance. HPC2N Abisko, Kebnekaise HPC Programming
More informationUmeå University
HPC2N: Introduction to HPC2N and Kebnekaise, 2017-09-12 HPC2N @ Umeå University Introduction to HPC2N and Kebnekaise Jerry Eriksson, Pedro Ojeda-May, and Birgitte Brydsö Outline Short presentation of HPC2N
More informationHigh Performance Computing : Code_Saturne in the PRACE project
High Performance Computing : Code_Saturne in the PRACE project Andy SUNDERLAND Charles MOULINEC STFC Daresbury Laboratory, UK Code_Saturne User Meeting Chatou 1st-2nd Dec 28 STFC Daresbury Laboratory HPC
More informationGreen Supercomputing
Green Supercomputing On the Energy Consumption of Modern E-Science Prof. Dr. Thomas Ludwig German Climate Computing Centre Hamburg, Germany ludwig@dkrz.de Outline DKRZ 2013 and Climate Science The Exascale
More informationThe View from the High End Fortran, Parallelism and the HECToR Service
Introduction The View from the High End Fortran, Parallelism and the HECToR Service Ian J. Bush NAG Ltd. Wilkinson House Jordan Hill Road Oxford OX2 8DR Since its inception in 1956 Fortran, and indeed
More informationHigh Performance Computing. What is it used for and why?
High Performance Computing What is it used for and why? Overview What is it used for? Drivers for HPC Examples of usage Why do you need to learn the basics? Hardware layout and structure matters Serial
More informationVýpočetní zdroje IT4Innovations a PRACE pro využití ve vědě a výzkumu
Výpočetní zdroje IT4Innovations a PRACE pro využití ve vědě a výzkumu Filip Staněk Seminář gridového počítání 2011, MetaCentrum, Brno, 7. 11. 2011 Introduction I Project objectives: to establish a centre
More informationExascale: challenges and opportunities in a power constrained world
Exascale: challenges and opportunities in a power constrained world Carlo Cavazzoni c.cavazzoni@cineca.it SuperComputing Applications and Innovation Department CINECA CINECA non profit Consortium, made
More informationTECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 6 th CALL (Tier-0)
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 6 th CALL (Tier-0) Contributing sites and the corresponding computer systems for this call are: GCS@Jülich, Germany IBM Blue Gene/Q GENCI@CEA, France Bull Bullx
More informationPractical Scientific Computing
Practical Scientific Computing Performance-optimized Programming Preliminary discussion: July 11, 2008 Dr. Ralf-Peter Mundani, mundani@tum.de Dipl.-Ing. Ioan Lucian Muntean, muntean@in.tum.de MSc. Csaba
More information2DECOMP&FFT The Library Behind Incompact3D
2DECOMP&FFT The Library Behind Incompact3D Ning Li NAG Incompact3D User Group Meeting Imperial College London 24/04/2014 Experts in numerical algorithms and HPC services About the Speaker PhD in Mechanical
More informationHPC Architectures. Types of resource currently in use
HPC Architectures Types of resource currently in use Reusing this material This work is licensed under a Creative Commons Attribution- NonCommercial-ShareAlike 4.0 International License. http://creativecommons.org/licenses/by-nc-sa/4.0/deed.en_us
More informationIt s a Multicore World. John Urbanic Pittsburgh Supercomputing Center Parallel Computing Scientist
It s a Multicore World John Urbanic Pittsburgh Supercomputing Center Parallel Computing Scientist Waiting for Moore s Law to save your serial code started getting bleak in 2004 Source: published SPECInt
More informationPresent and Future Leadership Computers at OLCF
Present and Future Leadership Computers at OLCF Al Geist ORNL Corporate Fellow DOE Data/Viz PI Meeting January 13-15, 2015 Walnut Creek, CA ORNL is managed by UT-Battelle for the US Department of Energy
More informationIt s a Multicore World. John Urbanic Pittsburgh Supercomputing Center Parallel Computing Scientist
It s a Multicore World John Urbanic Pittsburgh Supercomputing Center Parallel Computing Scientist Waiting for Moore s Law to save your serial code started getting bleak in 2004 Source: published SPECInt
More informationUpdate on Cray Activities in the Earth Sciences
Update on Cray Activities in the Earth Sciences Presented to the 13 th ECMWF Workshop on the Use of HPC in Meteorology 3-7 November 2008 Per Nyberg nyberg@cray.com Director, Marketing and Business Development
More informationUCX: An Open Source Framework for HPC Network APIs and Beyond
UCX: An Open Source Framework for HPC Network APIs and Beyond Presented by: Pavel Shamis / Pasha ORNL is managed by UT-Battelle for the US Department of Energy Co-Design Collaboration The Next Generation
More informationCP2K Performance Benchmark and Profiling. April 2011
CP2K Performance Benchmark and Profiling April 2011 Note The following research was performed under the HPC Advisory Council activities Participating vendors: AMD, Dell, Mellanox Compute resource - HPC
More informationUniversity at Buffalo Center for Computational Research
University at Buffalo Center for Computational Research The following is a short and long description of CCR Facilities for use in proposals, reports, and presentations. If desired, a letter of support
More informationHybrid KAUST Many Cores and OpenACC. Alain Clo - KAUST Research Computing Saber Feki KAUST Supercomputing Lab Florent Lebeau - CAPS
+ Hybrid Computing @ KAUST Many Cores and OpenACC Alain Clo - KAUST Research Computing Saber Feki KAUST Supercomputing Lab Florent Lebeau - CAPS + Agenda Hybrid Computing n Hybrid Computing n From Multi-Physics
More informationThe NEMO Ocean Modelling Code: A Case Study
The NEMO Ocean Modelling Code: A Case Study CUG 24 th 27 th May 2010 Dr Fiona J. L. Reid Applications Consultant, EPCC f.reid@epcc.ed.ac.uk +44 (0)131 651 3394 Acknowledgements Cray Centre of Excellence
More informationJÜLICH SUPERCOMPUTING CENTRE Site Introduction Michael Stephan Forschungszentrum Jülich
JÜLICH SUPERCOMPUTING CENTRE Site Introduction 09.04.2018 Michael Stephan JSC @ Forschungszentrum Jülich FORSCHUNGSZENTRUM JÜLICH Research Centre Jülich One of the 15 Helmholtz Research Centers in Germany
More informationAnalyzing and Optimizing Global Array Toolkit for Cray Gemini Interconnect
Analyzing and Optimizing Global Array Toolkit for Cray Gemini Interconnect Vairavan Murugappan August 27, 2010 MSc in High Performance Computing The University of Edinburgh Year of Presentation: 2010 Abstract
More informationHigh-Performance Computing & Simulations in Quantum Many-Body Systems PART I. Thomas Schulthess
High-Performance Computing & Simulations in Quantum Many-Body Systems PART I Thomas Schulthess schulthess@phys.ethz.ch What exactly is high-performance computing? 1E10 1E9 1E8 1E7 relative performance
More informationThe IBM Blue Gene/Q: Application performance, scalability and optimisation
The IBM Blue Gene/Q: Application performance, scalability and optimisation Mike Ashworth, Andrew Porter Scientific Computing Department & STFC Hartree Centre Manish Modani IBM STFC Daresbury Laboratory,
More informationOak Ridge National Laboratory Computing and Computational Sciences
Oak Ridge National Laboratory Computing and Computational Sciences OFA Update by ORNL Presented by: Pavel Shamis (Pasha) OFA Workshop Mar 17, 2015 Acknowledgments Bernholdt David E. Hill Jason J. Leverman
More informationMesh reordering in Fluidity using Hilbert space-filling curves
Mesh reordering in Fluidity using Hilbert space-filling curves Mark Filipiak EPCC, University of Edinburgh March 2013 Abstract Fluidity is open-source, multi-scale, general purpose CFD model. It is a finite
More informationParallel FFT Libraries
Parallel FFT Libraries Evangelos Brachos August 19, 2011 MSc in High Performance Computing The University of Edinburgh Year of Presentation: 2011 Abstract The focus of this project is the area of the fast
More informationPerformance Comparison of Capability Application Benchmarks on the IBM p5-575 and the Cray XT3. Mike Ashworth
Performance Comparison of Capability Application Benchmarks on the IBM p5-575 and the Cray XT3 Mike Ashworth Computational Science and CCLRC Daresbury Laboratory Warrington UK http://www.cse.clrc.ac.uk/arc/
More informationIntel Many Integrated Core (MIC) Architecture
Intel Many Integrated Core (MIC) Architecture Karl Solchenbach Director European Exascale Labs BMW2011, November 3, 2011 1 Notice and Disclaimers Notice: This document contains information on products
More informationCRAY XK6 REDEFINING SUPERCOMPUTING. - Sanjana Rakhecha - Nishad Nerurkar
CRAY XK6 REDEFINING SUPERCOMPUTING - Sanjana Rakhecha - Nishad Nerurkar CONTENTS Introduction History Specifications Cray XK6 Architecture Performance Industry acceptance and applications Summary INTRODUCTION
More informationComputational Challenges and Opportunities for Nuclear Astrophysics
Computational Challenges and Opportunities for Nuclear Astrophysics Bronson Messer Acting Group Leader Scientific Computing Group National Center for Computational Sciences Theoretical Astrophysics Group
More informationPiz Daint: Application driven co-design of a supercomputer based on Cray s adaptive system design
Piz Daint: Application driven co-design of a supercomputer based on Cray s adaptive system design Sadaf Alam & Thomas Schulthess CSCS & ETHzürich CUG 2014 * Timelines & releases are not precise Top 500
More informationHPC Resources & Training
www.bsc.es HPC Resources & Training in the BSC, the RES and PRACE Montse González Ferreiro RES technical and training coordinator + Facilities + Capacity How fit together the BSC, the RES and PRACE? TIER
More informationCloverLeaf: Preparing Hydrodynamics Codes for Exascale
CloverLeaf: Preparing Hydrodynamics Codes for Exascale Andrew Mallinson Andy.Mallinson@awe.co.uk www.awe.co.uk British Crown Owned Copyright [2013]/AWE Agenda AWE & Uni. of Warwick introduction Problem
More informationSami Saarinen Peter Towers. 11th ECMWF Workshop on the Use of HPC in Meteorology Slide 1
Acknowledgements: Petra Kogel Sami Saarinen Peter Towers 11th ECMWF Workshop on the Use of HPC in Meteorology Slide 1 Motivation Opteron and P690+ clusters MPI communications IFS Forecast Model IFS 4D-Var
More informationComparison of PRACE prototypes and benchmarks. Axel Berg (SARA, NL), ISC 10 Hamburg June 1 st 2010
Comparison of PRACE prototypes and benchmarks Axel Berg (SARA, NL), ISC 10 Hamburg June 1 st 2010 What is a prototype? 2 The prototype according to Wikipedia A prototype is an original type, form, or instance
More informationHECToR to ARCHER. An Introduction from Cray. 10/3/2013 Cray Inc. Property
HECToR to ARCHER An Introduction from Cray 10/3/2013 Cray Inc. Property 1 HECToR High-End Computing Terascale Resource HECToR has been the UK s front-line national supercomputing service since 2007. The
More informationTechnologies and application performance. Marc Mendez-Bermond HPC Solutions Expert - Dell Technologies September 2017
Technologies and application performance Marc Mendez-Bermond HPC Solutions Expert - Dell Technologies September 2017 The landscape is changing We are no longer in the general purpose era the argument of
More informationCP2K: HIGH PERFORMANCE ATOMISTIC SIMULATION
CP2K: HIGH PERFORMANCE ATOMISTIC SIMULATION Iain Bethune ibethune@epcc.ed.ac.uk http://tinyurl.com/mcc-ukcp-2016 CP2K Overview CP2K is a program to perform atomistic and molecular simulations of solid
More informationHETEROGENEOUS HPC, ARCHITECTURAL OPTIMIZATION, AND NVLINK STEVE OBERLIN CTO, TESLA ACCELERATED COMPUTING NVIDIA
HETEROGENEOUS HPC, ARCHITECTURAL OPTIMIZATION, AND NVLINK STEVE OBERLIN CTO, TESLA ACCELERATED COMPUTING NVIDIA STATE OF THE ART 2012 18,688 Tesla K20X GPUs 27 PetaFLOPS FLAGSHIP SCIENTIFIC APPLICATIONS
More informationCompute Node Linux: Overview, Progress to Date & Roadmap
Compute Node Linux: Overview, Progress to Date & Roadmap David Wallace Cray Inc ABSTRACT: : This presentation will provide an overview of Compute Node Linux(CNL) for the CRAY XT machine series. Compute
More informationScheduling Strategies for HPC as a Service (HPCaaS) for Bio-Science Applications
Scheduling Strategies for HPC as a Service (HPCaaS) for Bio-Science Applications Sep 2009 Gilad Shainer, Tong Liu (Mellanox); Jeffrey Layton (Dell); Joshua Mora (AMD) High Performance Interconnects for
More informationPorting and Optimisation of UM on ARCHER. Karthee Sivalingam, NCAS-CMS. HPC Workshop ECMWF JWCRP
Porting and Optimisation of UM on ARCHER Karthee Sivalingam, NCAS-CMS HPC Workshop ECMWF JWCRP Acknowledgements! NCAS-CMS Bryan Lawrence Jeffrey Cole Rosalyn Hatcher Andrew Heaps David Hassell Grenville
More informationCP2K Performance Benchmark and Profiling. April 2011
CP2K Performance Benchmark and Profiling April 2011 Note The following research was performed under the HPC Advisory Council HPC works working group activities Participating vendors: HP, Intel, Mellanox
More informationBlueGene/L. Computer Science, University of Warwick. Source: IBM
BlueGene/L Source: IBM 1 BlueGene/L networking BlueGene system employs various network types. Central is the torus interconnection network: 3D torus with wrap-around. Each node connects to six neighbours
More informationOur Workshop Environment
Our Workshop Environment John Urbanic Parallel Computing Scientist Pittsburgh Supercomputing Center Copyright 2015 Our Environment Today Your laptops or workstations: only used for portal access Blue Waters
More informationHigh Performance Computing from an EU perspective
High Performance Computing from an EU perspective DEISA PRACE Symposium 2010 Barcelona, 10 May 2010 Kostas Glinos European Commission - DG INFSO Head of Unit GÉANT & e-infrastructures 1 "The views expressed
More informationARCHER/RDF Overview. How do they fit together? Andy Turner, EPCC
ARCHER/RDF Overview How do they fit together? Andy Turner, EPCC a.turner@epcc.ed.ac.uk www.epcc.ed.ac.uk www.archer.ac.uk Outline ARCHER/RDF Layout Available file systems Compute resources ARCHER Compute
More informationMaxwell: a 64-FPGA Supercomputer
Maxwell: a 64-FPGA Supercomputer Copyright 2007, the University of Edinburgh Dr Rob Baxter Software Development Group Manager, EPCC R.Baxter@epcc.ed.ac.uk +44 131 651 3579 Outline The FHPCA Why build Maxwell?
More informationLecture 20: Distributed Memory Parallelism. William Gropp
Lecture 20: Distributed Parallelism William Gropp www.cs.illinois.edu/~wgropp A Very Short, Very Introductory Introduction We start with a short introduction to parallel computing from scratch in order
More informationExperiences with HP SFS / Lustre in HPC Production
Experiences with HP SFS / Lustre in HPC Production Computing Centre (SSCK) University of Karlsruhe Laifer@rz.uni-karlsruhe.de page 1 Outline» What is HP StorageWorks Scalable File Share (HP SFS)? A Lustre
More informationOP2 FOR MANY-CORE ARCHITECTURES
OP2 FOR MANY-CORE ARCHITECTURES G.R. Mudalige, M.B. Giles, Oxford e-research Centre, University of Oxford gihan.mudalige@oerc.ox.ac.uk 27 th Jan 2012 1 AGENDA OP2 Current Progress Future work for OP2 EPSRC
More informationThe Center for High Performance Computing. Dell Breakfast Events 20 th June 2016 Happy Sithole
The Center for High Performance Computing Dell Breakfast Events 20 th June 2016 Happy Sithole Background: The CHPC in SA CHPC User Community: South Africa CHPC Existing Users Future Users Introduction
More informationTrends in HPC (hardware complexity and software challenges)
Trends in HPC (hardware complexity and software challenges) Mike Giles Oxford e-research Centre Mathematical Institute MIT seminar March 13th, 2013 Mike Giles (Oxford) HPC Trends March 13th, 2013 1 / 18
More informationHigh Performance Computing and Data Resources at SDSC
High Performance Computing and Data Resources at SDSC "! Mahidhar Tatineni (mahidhar@sdsc.edu)! SDSC Summer Institute! August 05, 2013! HPC Resources at SDSC Hardware Overview HPC Systems : Gordon, Trestles
More informationUser Training Cray XC40 IITM, Pune
User Training Cray XC40 IITM, Pune Sudhakar Yerneni, Raviteja K, Nachiket Manapragada, etc. 1 Cray XC40 Architecture & Packaging 3 Cray XC Series Building Blocks XC40 System Compute Blade 4 Compute Nodes
More informationIBM HPC DIRECTIONS. Dr Don Grice. ECMWF Workshop November, IBM Corporation
IBM HPC DIRECTIONS Dr Don Grice ECMWF Workshop November, 2008 IBM HPC Directions Agenda What Technology Trends Mean to Applications Critical Issues for getting beyond a PF Overview of the Roadrunner Project
More informationIntroduction to National Supercomputing Centre in Guangzhou and Opportunities for International Collaboration
Exascale Applications and Software Conference 21st 23rd April 2015, Edinburgh, UK Introduction to National Supercomputing Centre in Guangzhou and Opportunities for International Collaboration Xue-Feng
More informationThe Hopper System: How the Largest* XE6 in the World Went From Requirements to Reality! Katie Antypas, Tina Butler, and Jonathan Carter
The Hopper System: How the Largest* XE6 in the World Went From Requirements to Reality! Katie Antypas, Tina Butler, and Jonathan Carter CUG 2011, May 25th, 2011 1 Requirements to Reality Develop RFP Select
More informationCray events. ! Cray User Group (CUG): ! Cray Technical Workshop Europe:
Cray events! Cray User Group (CUG):! When: May 16-19, 2005! Where: Albuquerque, New Mexico - USA! Registration: reserved to CUG members! Web site: http://www.cug.org! Cray Technical Workshop Europe:! When:
More informationThe Stampede is Coming Welcome to Stampede Introductory Training. Dan Stanzione Texas Advanced Computing Center
The Stampede is Coming Welcome to Stampede Introductory Training Dan Stanzione Texas Advanced Computing Center dan@tacc.utexas.edu Thanks for Coming! Stampede is an exciting new system of incredible power.
More informationAchieving Efficient Strong Scaling with PETSc Using Hybrid MPI/OpenMP Optimisation
Achieving Efficient Strong Scaling with PETSc Using Hybrid MPI/OpenMP Optimisation Michael Lange 1 Gerard Gorman 1 Michele Weiland 2 Lawrence Mitchell 2 Xiaohu Guo 3 James Southern 4 1 AMCG, Imperial College
More informationD1-1 Optimisation of R bootstrapping on HECToR with SPRINT
D1-1 Optimisation of R bootstrapping on HECToR with SPRINT Project Title Document Title Authorship ESPRC dcse "Bootstrapping and support vector machines with R and SPRINT" D1-1 Optimisation of R bootstrapping
More informationThe Cray Rainier System: Integrated Scalar/Vector Computing
THE SUPERCOMPUTER COMPANY The Cray Rainier System: Integrated Scalar/Vector Computing Per Nyberg 11 th ECMWF Workshop on HPC in Meteorology Topics Current Product Overview Cray Technology Strengths Rainier
More informationAim High. Intel Technical Update Teratec 07 Symposium. June 20, Stephen R. Wheat, Ph.D. Director, HPC Digital Enterprise Group
Aim High Intel Technical Update Teratec 07 Symposium June 20, 2007 Stephen R. Wheat, Ph.D. Director, HPC Digital Enterprise Group Risk Factors Today s s presentations contain forward-looking statements.
More informationCPMD Performance Benchmark and Profiling. February 2014
CPMD Performance Benchmark and Profiling February 2014 Note The following research was performed under the HPC Advisory Council activities Special thanks for: HP, Mellanox For more information on the supporting
More informationGREAT PERFORMANCE FOR TINY PROBLEMS: BATCHED PRODUCTS OF SMALL MATRICES. Nikolay Markovskiy Peter Messmer
GREAT PERFORMANCE FOR TINY PROBLEMS: BATCHED PRODUCTS OF SMALL MATRICES Nikolay Markovskiy Peter Messmer ABOUT CP2K Atomistic and molecular simulations of solid state From ab initio DFT and Hartree-Fock
More informationWhat does Heterogeneity bring?
What does Heterogeneity bring? Ken Koch Scientific Advisor, CCS-DO, LANL LACSI 2006 Conference October 18, 2006 Some Terminology Homogeneous Of the same or similar nature or kind Uniform in structure or
More informationThe Red Storm System: Architecture, System Update and Performance Analysis
The Red Storm System: Architecture, System Update and Performance Analysis Douglas Doerfler, Jim Tomkins Sandia National Laboratories Center for Computation, Computers, Information and Mathematics LACSI
More informationBuilding supercomputers from embedded technologies
http://www.montblanc-project.eu Building supercomputers from embedded technologies Alex Ramirez Barcelona Supercomputing Center Technical Coordinator This project and the research leading to these results
More informationHPC Saudi Jeffrey A. Nichols Associate Laboratory Director Computing and Computational Sciences. Presented to: March 14, 2017
Creating an Exascale Ecosystem for Science Presented to: HPC Saudi 2017 Jeffrey A. Nichols Associate Laboratory Director Computing and Computational Sciences March 14, 2017 ORNL is managed by UT-Battelle
More informationPerformance Study of Popular Computational Chemistry Software Packages on Cray HPC Systems
Performance Study of Popular Computational Chemistry Software Packages on Cray HPC Systems Junjie Li (lijunj@iu.edu) Shijie Sheng (shengs@iu.edu) Raymond Sheppard (rsheppar@iu.edu) Pervasive Technology
More informationABySS Performance Benchmark and Profiling. May 2010
ABySS Performance Benchmark and Profiling May 2010 Note The following research was performed under the HPC Advisory Council activities Participating vendors: AMD, Dell, Mellanox Compute resource - HPC
More informationThe Future of GPU Computing
The Future of GPU Computing Bill Dally Chief Scientist & Sr. VP of Research, NVIDIA Bell Professor of Engineering, Stanford University November 18, 2009 The Future of Computing Bill Dally Chief Scientist
More informationArm in HPC. Toshinori Kujiraoka Sales Manager, APAC HPC Tools Arm Arm Limited
Arm in HPC Toshinori Kujiraoka Sales Manager, APAC HPC Tools Arm 2019 Arm Limited Arm Technology Connects the World Arm in IOT 21 billion chips in the past year Mobile/Embedded/IoT/ Automotive/GPUs/Servers
More informationBarcelona Supercomputing Center
www.bsc.es Barcelona Supercomputing Center Centro Nacional de Supercomputación EMIT 2016. Barcelona June 2 nd, 2016 Barcelona Supercomputing Center Centro Nacional de Supercomputación BSC-CNS objectives:
More informationCS500 SMARTER CLUSTER SUPERCOMPUTERS
CS500 SMARTER CLUSTER SUPERCOMPUTERS OVERVIEW Extending the boundaries of what you can achieve takes reliable computing tools matched to your workloads. That s why we tailor the Cray CS500 cluster supercomputer
More informationNERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber
NERSC Site Update National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory Richard Gerber NERSC Senior Science Advisor High Performance Computing Department Head Cori
More information: A new version of Supercomputing or life after the end of the Moore s Law
: A new version of Supercomputing or life after the end of the Moore s Law Dr.-Ing. Alexey Cheptsov SEMAPRO 2015 :: 21.07.2015 :: Dr. Alexey Cheptsov OUTLINE About us Convergence of Supercomputing into
More informationPART I - Fundamentals of Parallel Computing
PART I - Fundamentals of Parallel Computing Objectives What is scientific computing? The need for more computing power The need for parallel computing and parallel programs 1 What is scientific computing?
More informationHigh Performance Computing with Accelerators
High Performance Computing with Accelerators Volodymyr Kindratenko Innovative Systems Laboratory @ NCSA Institute for Advanced Computing Applications and Technologies (IACAT) National Center for Supercomputing
More information"On the Capability and Achievable Performance of FPGAs for HPC Applications"
"On the Capability and Achievable Performance of FPGAs for HPC Applications" Wim Vanderbauwhede School of Computing Science, University of Glasgow, UK Or in other words "How Fast Can Those FPGA Thingies
More informationOverview. Idea: Reduce CPU clock frequency This idea is well suited specifically for visualization
Exploring Tradeoffs Between Power and Performance for a Scientific Visualization Algorithm Stephanie Labasan & Matt Larsen (University of Oregon), Hank Childs (Lawrence Berkeley National Laboratory) 26
More informationOverlapping Computation and Communication for Advection on Hybrid Parallel Computers
Overlapping Computation and Communication for Advection on Hybrid Parallel Computers James B White III (Trey) trey@ucar.edu National Center for Atmospheric Research Jack Dongarra dongarra@eecs.utk.edu
More informationBlue Waters System Overview. Greg Bauer
Blue Waters System Overview Greg Bauer The Blue Waters EcoSystem Petascale EducaIon, Industry and Outreach Petascale ApplicaIons (CompuIng Resource AllocaIons) Petascale ApplicaIon CollaboraIon Team Support
More informationPractical Scientific Computing
Practical Scientific Computing Performance-optimised Programming Preliminary discussion, 17.7.2007 Dr. Ralf-Peter Mundani, mundani@tum.de Dipl.-Ing. Ioan Lucian Muntean, muntean@in.tum.de Dipl.-Geophys.
More information