Exascale Challenges and Applications Initiatives for Earth System Modeling

Size: px
Start display at page:

Download "Exascale Challenges and Applications Initiatives for Earth System Modeling"

Transcription

1 Exascale Challenges and Applications Initiatives for Earth System Modeling Workshop on Weather and Climate Prediction on Next Generation Supercomputers October 2012 Tom Edwards

2 An Introduction HECToR UK s National Academic Supercomputing Service Cray XE6 AMD Interlagos + Gemini Network Service based in Edinburgh managed by EPCC users, mixture of Academics and Industry Part of the PRACE consortium Cray Centre of Excellence Two Cray staff working on scaling user applications Addressing high end scaling and exploitation of Cray systems Covers a wide variety of applications Climate, CFD, Turbulence, Chemistry Cray Exascale Research Initiative, Europe Another two Cray staff working in partnership with EPCC Addressing software and co-design challenges for Exascale 24 October

3 Topics The Challenge of Exascale From Petascale towards Exascale Key challenges Macro level trends Impacts on architectures and applications Investing in Applications Impact and objectives of Blue Waters and Titan (pre-exascale) Community R&D efforts CRESTA UPSCALE A Case study in Petascale climate The role of decomposition in performance Optimizing task placement 3

4 The Challenge of Exascale 4

5 Concurrency is growing Not just from the number of cores Longer vector lengths Probably fixed length Systems made from different types of processors Hybrids of scalar and vectors Requires new methods of programming to fully exploit the hardware 47,309 79, ,306 Average Number of Processor Cores per Supercomputer (Top20 of Top500) Source: ,245 1,073 1,644 1,847 2,230 2,827 3,093 3,518 10,073 16,316 20, October 2012 AMMW03 5

6 Energy is now the limiting factor Which requires more energy? Performing a 64-bit floating-point FMA: 893, x Or moving the three 64-bit operands 20 mm across the die: = 39,226, = 39,226, This one takes over 3x the energy! And loading the data from off chip takes > 10x more yet Flops are cheap, communication is expensive Exploiting data locality is critical for energy efficiency 6

7 Key Challenges to Get to Exascale Power Traditional voltage scaling is over Power now a major design constraint Cost of ownership Driving significant changes in architecture Concurrency A billion operations per clock Billions of refs in flight at all times Will require huge problems Need to exploit all available parallelism Programming Difficulty Concurrency and new microarchitectures will significantly complicate software Need to hide this complexity from the users Resiliency Many more components Components getting less reliable Checkpoint bandwidth not scaling Impacts both systems and storage 7

8 Macro Level Trends (1) Major changes to applications and algorithms to address and extend scalability O(10K) increase in concurrency for exascale Heterogeneous computing Defined as accelerators today Processor architecture in 5+ years remained undefined Software is the barrier to successful and efficient usage Architectural imbalances Memory and I/O speeds are not keeping up Expectation of deeper memory hierarchies Communication will continue to be expensive Exploiting data locality is critical for computational efficiency Very much defined in terms of Exascale, but equally of relevance today 8

9 Macro Level Trends (2) Application Resiliency Ability for an application to ride through system faults needs to be solved at the system software and application level Energy aware computing Architectural perspective: Driving heterogeneous approaches Driving the discussion of low power, low frequency processors Application perspective: Energy aware optimization of applications Although today there is little/no control by the system admins and programmers Exploiting data locality is critical for energy efficiency 9

10 From Petascale towards Exascale Adapting codes to petascale and beyond 10

11 Petascale towards Exascale Current Petascale systems have been key: In enabling transformational capabilities across a broad range of science and engineering, including industry As development platforms for preparing earth system models for extreme scale capabilities. Systems Such as NCSA Blue Waters and ORNL Titan will further push the boundary of Petascale computing. Fully coupled biogeochemistryphysical climate simulation Kinetic simulations of the magnetosphere HIV immature virion simulations Climate Specialists Workshop Earthquake simulation of San Andeas fault 11

12 Blue Waters System 11.5 PF Sustained Petascale Performance Cray System & Storage cabinets: Compute nodes: Usable Storage Bandwidth: System Memory: Memory per core: Gemini Interconnect Topology: Usable Storage: Peak performance: Number of AMD processors: Number of AMD x86 core: Number of NVIDIA GPUs: >300 >25,000 >1 TB/s >1.5 Petabytes 4 GB 3D Torus >25 Petabytes >11.5 Petaflops >49,000 >380,000 >3,000 Cray Sonexion Storage 12

13 Blue Waters Science More than 25 science teams 25 distinct research fields selected to run on the new Blue Waters Nanotechnology Astronomy Earthquakes and the damage they cause Viruses entering cells Severe storms Climate change 13

14 ORNL Titan PF Hybrid System Upgrading Jaguar (XT5) to Titan (XK6) - $97M Contract Peak performance 10 to 20 petaflops Upgrade will be completed by the end of 2012 Efficiency & scalability of unprecedented scale Hybrid of NVIDIA GPUs, AMD CPUs and Gemini interconnect ORNL step towards pre-exascale architecture Focus on programming environment: To exploit all levels of parallelism. Compiler directives hold the promise of allowing developers to efficiently exploit the GPUs while preserving portability to other platforms. Partnership with Cray. 14

15 Name Titan Configuration Architecture Processor Titan XK6 Cabinets 200 Nodes 18,688 CPU Memory/Node GPU Memory/Node Interconnect GPUs AMD Interlagos 32 GB 6 GB Gemini Nvidia Kepler 15

16 Investing in Applications

17 Titan: Early Science Applications 17

18 10/31/

19 10/31/

20 10/31/

21 Applications R&D for the Extreme Scale A number of applications focused petascale and exascale activities involving the earth system modeling community: Swiss Platform for High-Performance and High-Productivity Computing - HP2C UK Met Office / NCAS Project UPSCALE ICOsahedral-grid Models for EXascale Earth system simulations (ICOMEX) Titan Early Science Applications Blue Waters Petascale Computing Resource Allocations (PRAC) Teams CRESTA Project EU funded exascale co-design 21

22 The CRESTA Project Three year EU-funded CREST collaborative project, 13 partners, 12 million costs, 8.5 million funding Collaborative Research into Exascale Systemware, Tools and Applications Project coordinator: EPCC at The University of Edinburgh CRESTA has a very strong focus on exascale software challenges Uses a co-design model of applications with exascale potential interacting with CREST systemware and tools activities The hardware partner is Cray Applications represent broad spectrum from science and engineering CRESTA will compare and contrast incremental and disruptive solutions to Exascale challenges CREST

23 Consortium & Applications Leading European HPC centres EPCC, HLRS, CSC, PDC A world leading vendor Cray World leading tools providers TUD (Vampir), Allinea (DDT) Exascale application owners and specialists ABO, JYU, UCL, ECMWF, ECP, DLR epcc cresta Visual Identity Designs CREST CREST Application Grand challenge Partner responsible GROMACS Biomolecular systems KTH (Sweden) ELMFIRE Fusion energy ABO (Finland) HemeLB Virtual Physiological Human UCL (UK) IFS Numerical weather prediction ECMWF (International) OpenFOAM Engineering EPCC / HLRS / ECP Nek5000 Engineering KTH (Sweden)

24 Courtesy: George Mozdzynski, ECMWF

25 UPSCALE A Case study in Petascale Climate 25

26 The UPSCALE project Aim is to increase the fidelity of global climate simulations and understanding of weather and climate risk, by representing fundamental weather and climate processes more completely. Hermit XE6 is being used to conduct a series of 25-year simulations: Under both current climate conditions and a climate change scenario, using HadGEM3 at 25km horizontal resolution. Essentially the same resolution as used for the Met Office's global weather forecasts. Data set produced will be widely studied over next decade. Key component of project is to extend the scalability of HadGEM (and UM) to ~ 25K cores. 26

27 Project UPSCALE: UK Climate Research Team Wins Major Computing Award on the HLRS Cray XE6 UPSCALE project: UK on PRACE: weather-resolving Simulations of Climate for global Environmental risk The team comprises climate scientists at the UK Met Office and National Centre for Atmospheric Science (NCAS) Led by Pier Luigi Vidale, Willis Professor of Climate System Science and Climate Hazards at the University of Reading's Meteorology department and Director of the Weather and Climate Hazards Laboratory. Granted substantial computing resources on the HLRS Hermit XE6 by the Partnership for Advanced Computing in Europe (PRACE) Access Committee. 27

28 HLRS Configuration Value Number of Cray XE6 LC Cabinets 38 Number of Internal Chassis 114 Number of XDPs 8 Compute Partition Value Number of Compute Nodes (incl. 480 "fat" nodes and 16 XK6 nodes) 3,552 Number of Compute Processors 7,104(*) Number of Compute Cores 113,664(*) Compute Processor Type AMD 2.3 GHz Compute peak TF 1,046 TF Memory Configuration per Node (3072 nodes) 32 GB, DDR3 Memory Configuration per "fat" Node (480 nodes) Compute Partition Memory Capacity 64 GB, DDR TB (*) Currently there are 16 Tesla based XK6, will be upgraded to 32 Kepler based nodes at the end of the year 28

29 Optimising UPSCALE for Petascale (and beyond) Running production climate simulations at relatively untested resolutions (for climate) Potential stability issues CPU grant for 1 year only => need to be productive quickly Time to solution important (without wasting resources) No time for in-depth code modifications, look for optimisation in existing application and environmental parameters Large volumes of output data (~1TB per model per wallclock day) Potentially 5 models running consecutively (with favourable stability and queues) Tuning IO and the new IOS 29

30 Two Lessons from UPSCALE 31 October

31 Choices of Parallel Decomposition are growing More parallelism offers greater choices of decomposition Each parallel paradigm (distributed memory, shared memory, SIMT, vectorisation) offers even greater choices Applications need to make optimal choices to get best performance UPSCALE Choice of how to perform the 2D MPI Decomposition Plus number of OpenMP threads. Investigated a large number of decompositions over a limited range of total cores Found a wide range in performance 31

32 Maximum # of 2D Decompositions Max # of 2D Decompositions vs Processor Count Number of Ranks 32

33 Number of decopmsotions The best decompositions are rare Only a small proportion of decompositions giving the better performance % 5-10% 10-15% 15-20% 20-25% 25-30% 30-35% 35-40% 40-45% % Slower than fastest decomposition 45% between best and worst 33

34 Specialising cores to improve performance Using MPMD models to improve overall application performance 34

35 MPMD and Exascale It is probable that MPMD programs will become more prevalent at Exascale. Applications may be designed to use specialised subsets of processors for specific tasks (e.g. IO Servers, novel solvers, accelerated nodes) Coupling of multiple models is already happening (e.g. Atmosphere and Ocean), and is likely to be extended to other fields (e.g. whole body/brain simulations, multi-phase flows, combustion/chemistry) We will not be able to rely on there being corresponding symmetry in hardware and software layouts Instead, for optimal performance application placement must be guided by the software architecture in the future. UPSCALE s UM8.0 includes IO Servers for optimising data transfer to disk 35

36 UM Standard MPI Decomposition Node 3 Node 0 Node 1 Node 2 Domain decomposition distributed over the cores on each processor Deep East-West halos favour rows using intra-node comms (e.g. shared memory) Best performance achieved when processor E-W decomposition is a factor or domain E-W decomposition 36

37 IO Servers MPMD providing net performance gains Originally codes treat compute and IO as serial tasks to be performed by all nodes Compute+IO Compute IO Compute IO Compute IO Compute IO IO costs have grown so codes (e.g. UM) have been extended to include IO Server ranks These ranks are dedicated to performing the IO operations asynchronously of compute. Compute Compute Compute Compute Compute IO Servers IO IO IO IO Typically adding an additional 1% of nodes to act as IO servers can eliminate almost all IO from runtime. Requires data to be double buffered, so can increase overall memory overhead. Essentially a form of MPMD 37

38 Basic distribution of IO servers (1 per node) Node 4 Node 3 Node 0 Node 1 Node 2 38

39 Basic distribution of IO servers (pt 2) Advantages Easier implementation Efficient when number nodes number of IO servers Do not have to change the distribution of ranks across nodes E.g, Keep 16 ranks per node, just add to the total number of ranks Allows for much larger buffers on IO Server tasks Distributes IO traffic across the network. Disadvantages Disrupts the nice alignment between decomposition and nodes IO Servers restricted to the same memory limits as compute ranks IO Servers likely to require more memory, far less compute. 39

40 Rank Reordered Decomposition (IO Nodes) Node 4 Node 3 Node 0 Node 1 Node 2 23 October 2012 AMMW03 40

41 Rank Reordered Decomposition (pt 2) Advantages Keeps the nice alignment between proc decomposition and nodes Can change the distribution of ranks across nodes keep large numbers of ranks per node for compute nodes use fewer ranks per node on IO nodes Can be implemented at runtime on Cray architectures Efficient when number compute nodes >> number of IO servers Disadvantages Concentrates IO traffic on a few nodes on the systems Network bandwith > IO Bandwidth IO Servers should hide any IO delays anyway. UPSCALE recorded 6.5% improvement in runtime performance from rank reordering compared to default IO server distribution. 41

42 Summary Extreme scale computing will require the successful solution to a collection of interrelated science and technology challenges. A number of applications focused petascale and exascale activities involving the earth system modeling community: Several co-design initiatives are progressing well. Those that can scale have benefited from a focused, iterative multi-year algorithmic optimization effort: Optimization strategies do not remain stagnant and must take advantage of evolving hardware and software technologies. The successful groups have had ongoing access to scalable systems and support. 42

43 For More Information 43

An Introduction to OpenACC

An Introduction to OpenACC An Introduction to OpenACC Alistair Hart Cray Exascale Research Initiative Europe 3 Timetable Day 1: Wednesday 29th August 2012 13:00 Welcome and overview 13:15 Session 1: An Introduction to OpenACC 13:15

More information

Preparing GPU-Accelerated Applications for the Summit Supercomputer

Preparing GPU-Accelerated Applications for the Summit Supercomputer Preparing GPU-Accelerated Applications for the Summit Supercomputer Fernanda Foertter HPC User Assistance Group Training Lead foertterfs@ornl.gov This research used resources of the Oak Ridge Leadership

More information

Titan - Early Experience with the Titan System at Oak Ridge National Laboratory

Titan - Early Experience with the Titan System at Oak Ridge National Laboratory Office of Science Titan - Early Experience with the Titan System at Oak Ridge National Laboratory Buddy Bland Project Director Oak Ridge Leadership Computing Facility November 13, 2012 ORNL s Titan Hybrid

More information

Dr.-Ing. Achim Basermann DLR, Simulation and Software Technology (SC)

Dr.-Ing. Achim Basermann DLR, Simulation and Software Technology (SC) www.dlr.de Chart 1 > HPC@SC > Achim Basermann 20131125_HPCatSC_Basermann.pptx > 25.11.2013 HPC@SC Dr.-Ing. Achim Basermann DLR, Simulation and Software Technology (SC) www.dlr.de Chart 2 > HPC@SC > Achim

More information

Portable and Productive Performance with OpenACC Compilers and Tools. Luiz DeRose Sr. Principal Engineer Programming Environments Director Cray Inc.

Portable and Productive Performance with OpenACC Compilers and Tools. Luiz DeRose Sr. Principal Engineer Programming Environments Director Cray Inc. Portable and Productive Performance with OpenACC Compilers and Tools Luiz DeRose Sr. Principal Engineer Programming Environments Director Cray Inc. 1 Cray: Leadership in Computational Research Earth Sciences

More information

Update on Cray Activities in the Earth Sciences

Update on Cray Activities in the Earth Sciences Update on Cray Activities in the Earth Sciences Presented to the 13 th ECMWF Workshop on the Use of HPC in Meteorology 3-7 November 2008 Per Nyberg nyberg@cray.com Director, Marketing and Business Development

More information

Towards Exascale Programming Models HPC Summit, Prague Erwin Laure, KTH

Towards Exascale Programming Models HPC Summit, Prague Erwin Laure, KTH Towards Exascale Programming Models HPC Summit, Prague Erwin Laure, KTH 1 Exascale Programming Models With the evolution of HPC architecture towards exascale, new approaches for programming these machines

More information

Steve Scott, Tesla CTO SC 11 November 15, 2011

Steve Scott, Tesla CTO SC 11 November 15, 2011 Steve Scott, Tesla CTO SC 11 November 15, 2011 What goal do these products have in common? Performance / W Exaflop Expectations First Exaflop Computer K Computer ~10 MW CM5 ~200 KW Not constant size, cost

More information

CRAY XK6 REDEFINING SUPERCOMPUTING. - Sanjana Rakhecha - Nishad Nerurkar

CRAY XK6 REDEFINING SUPERCOMPUTING. - Sanjana Rakhecha - Nishad Nerurkar CRAY XK6 REDEFINING SUPERCOMPUTING - Sanjana Rakhecha - Nishad Nerurkar CONTENTS Introduction History Specifications Cray XK6 Architecture Performance Industry acceptance and applications Summary INTRODUCTION

More information

Hybrid multicore supercomputing and GPU acceleration with next-generation Cray systems

Hybrid multicore supercomputing and GPU acceleration with next-generation Cray systems Hybrid multicore supercomputing and GPU acceleration with next-generation Cray systems Alistair Hart Cray Exascale Research Initiative Europe T-Systems HPCN workshop DLR Braunschweig 26.May.11 ahart@cray.com

More information

HPC Saudi Jeffrey A. Nichols Associate Laboratory Director Computing and Computational Sciences. Presented to: March 14, 2017

HPC Saudi Jeffrey A. Nichols Associate Laboratory Director Computing and Computational Sciences. Presented to: March 14, 2017 Creating an Exascale Ecosystem for Science Presented to: HPC Saudi 2017 Jeffrey A. Nichols Associate Laboratory Director Computing and Computational Sciences March 14, 2017 ORNL is managed by UT-Battelle

More information

NVIDIA Update and Directions on GPU Acceleration for Earth System Models

NVIDIA Update and Directions on GPU Acceleration for Earth System Models NVIDIA Update and Directions on GPU Acceleration for Earth System Models Stan Posey, HPC Program Manager, ESM and CFD, NVIDIA, Santa Clara, CA, USA Carl Ponder, PhD, Applications Software Engineer, NVIDIA,

More information

Timothy Lanfear, NVIDIA HPC

Timothy Lanfear, NVIDIA HPC GPU COMPUTING AND THE Timothy Lanfear, NVIDIA FUTURE OF HPC Exascale Computing will Enable Transformational Science Results First-principles simulation of combustion for new high-efficiency, lowemision

More information

GPU Debugging Made Easy. David Lecomber CTO, Allinea Software

GPU Debugging Made Easy. David Lecomber CTO, Allinea Software GPU Debugging Made Easy David Lecomber CTO, Allinea Software david@allinea.com Allinea Software HPC development tools company Leading in HPC software tools market Wide customer base Blue-chip engineering,

More information

Productive Performance on the Cray XK System Using OpenACC Compilers and Tools

Productive Performance on the Cray XK System Using OpenACC Compilers and Tools Productive Performance on the Cray XK System Using OpenACC Compilers and Tools Luiz DeRose Sr. Principal Engineer Programming Environments Director Cray Inc. 1 The New Generation of Supercomputers Hybrid

More information

The IBM Blue Gene/Q: Application performance, scalability and optimisation

The IBM Blue Gene/Q: Application performance, scalability and optimisation The IBM Blue Gene/Q: Application performance, scalability and optimisation Mike Ashworth, Andrew Porter Scientific Computing Department & STFC Hartree Centre Manish Modani IBM STFC Daresbury Laboratory,

More information

HPC Issues for DFT Calculations. Adrian Jackson EPCC

HPC Issues for DFT Calculations. Adrian Jackson EPCC HC Issues for DFT Calculations Adrian Jackson ECC Scientific Simulation Simulation fast becoming 4 th pillar of science Observation, Theory, Experimentation, Simulation Explore universe through simulation

More information

HECToR. UK National Supercomputing Service. Andy Turner & Chris Johnson

HECToR. UK National Supercomputing Service. Andy Turner & Chris Johnson HECToR UK National Supercomputing Service Andy Turner & Chris Johnson Outline EPCC HECToR Introduction HECToR Phase 3 Introduction to AMD Bulldozer Architecture Performance Application placement the hardware

More information

Achieving Efficient Strong Scaling with PETSc Using Hybrid MPI/OpenMP Optimisation

Achieving Efficient Strong Scaling with PETSc Using Hybrid MPI/OpenMP Optimisation Achieving Efficient Strong Scaling with PETSc Using Hybrid MPI/OpenMP Optimisation Michael Lange 1 Gerard Gorman 1 Michele Weiland 2 Lawrence Mitchell 2 Xiaohu Guo 3 James Southern 4 1 AMCG, Imperial College

More information

Present and Future Leadership Computers at OLCF

Present and Future Leadership Computers at OLCF Present and Future Leadership Computers at OLCF Al Geist ORNL Corporate Fellow DOE Data/Viz PI Meeting January 13-15, 2015 Walnut Creek, CA ORNL is managed by UT-Battelle for the US Department of Energy

More information

The Hopper System: How the Largest* XE6 in the World Went From Requirements to Reality! Katie Antypas, Tina Butler, and Jonathan Carter

The Hopper System: How the Largest* XE6 in the World Went From Requirements to Reality! Katie Antypas, Tina Butler, and Jonathan Carter The Hopper System: How the Largest* XE6 in the World Went From Requirements to Reality! Katie Antypas, Tina Butler, and Jonathan Carter CUG 2011, May 25th, 2011 1 Requirements to Reality Develop RFP Select

More information

Trends in HPC (hardware complexity and software challenges)

Trends in HPC (hardware complexity and software challenges) Trends in HPC (hardware complexity and software challenges) Mike Giles Oxford e-research Centre Mathematical Institute MIT seminar March 13th, 2013 Mike Giles (Oxford) HPC Trends March 13th, 2013 1 / 18

More information

Challenges in High Performance Computing. William Gropp

Challenges in High Performance Computing. William Gropp Challenges in High Performance Computing William Gropp www.cs.illinois.edu/~wgropp 2 What is HPC? High Performance Computing is the use of computing to solve challenging problems that require significant

More information

Debugging CUDA Applications with Allinea DDT. Ian Lumb Sr. Systems Engineer, Allinea Software Inc.

Debugging CUDA Applications with Allinea DDT. Ian Lumb Sr. Systems Engineer, Allinea Software Inc. Debugging CUDA Applications with Allinea DDT Ian Lumb Sr. Systems Engineer, Allinea Software Inc. ilumb@allinea.com GTC 2013, San Jose, March 20, 2013 Embracing GPUs GPUs a rival to traditional processors

More information

Intel Many Integrated Core (MIC) Architecture

Intel Many Integrated Core (MIC) Architecture Intel Many Integrated Core (MIC) Architecture Karl Solchenbach Director European Exascale Labs BMW2011, November 3, 2011 1 Notice and Disclaimers Notice: This document contains information on products

More information

Oak Ridge National Laboratory Computing and Computational Sciences

Oak Ridge National Laboratory Computing and Computational Sciences Oak Ridge National Laboratory Computing and Computational Sciences OFA Update by ORNL Presented by: Pavel Shamis (Pasha) OFA Workshop Mar 17, 2015 Acknowledgments Bernholdt David E. Hill Jason J. Leverman

More information

The Cray Rainier System: Integrated Scalar/Vector Computing

The Cray Rainier System: Integrated Scalar/Vector Computing THE SUPERCOMPUTER COMPANY The Cray Rainier System: Integrated Scalar/Vector Computing Per Nyberg 11 th ECMWF Workshop on HPC in Meteorology Topics Current Product Overview Cray Technology Strengths Rainier

More information

Determining Optimal MPI Process Placement for Large- Scale Meteorology Simulations with SGI MPIplace

Determining Optimal MPI Process Placement for Large- Scale Meteorology Simulations with SGI MPIplace Determining Optimal MPI Process Placement for Large- Scale Meteorology Simulations with SGI MPIplace James Southern, Jim Tuccillo SGI 25 October 2016 0 Motivation Trend in HPC continues to be towards more

More information

Cray XC Scalability and the Aries Network Tony Ford

Cray XC Scalability and the Aries Network Tony Ford Cray XC Scalability and the Aries Network Tony Ford June 29, 2017 Exascale Scalability Which scalability metrics are important for Exascale? Performance (obviously!) What are the contributing factors?

More information

HPC state of play. HPC Ecosystem. HPC system supply. HPC use 24% EU 4,3% EU. Application software & tools. Academia 23% Bio-sciences 22% CAE 21%

HPC state of play. HPC Ecosystem. HPC system supply. HPC use 24% EU 4,3% EU. Application software & tools. Academia 23% Bio-sciences 22% CAE 21% Only half of US GDP spending HPC state of play HPC Ecosystem HPC system supply 4,3% EU EU market: 630 M/yr (EU lost 10% capacity from 2007 to 2009) 24% EU HPC use Academia 23% Bio-sciences 22% CAE 21%

More information

Introduction to FREE National Resources for Scientific Computing. Dana Brunson. Jeff Pummill

Introduction to FREE National Resources for Scientific Computing. Dana Brunson. Jeff Pummill Introduction to FREE National Resources for Scientific Computing Dana Brunson Oklahoma State University High Performance Computing Center Jeff Pummill University of Arkansas High Peformance Computing Center

More information

Debugging at Scale Lindon Locks

Debugging at Scale Lindon Locks Debugging at Scale Lindon Locks llocks@allinea.com Debugging at Scale At scale debugging - from 100 cores to 250,000 Problems faced by developers on real systems Alternative approaches to debugging and

More information

Overlapping Computation and Communication for Advection on Hybrid Parallel Computers

Overlapping Computation and Communication for Advection on Hybrid Parallel Computers Overlapping Computation and Communication for Advection on Hybrid Parallel Computers James B White III (Trey) trey@ucar.edu National Center for Atmospheric Research Jack Dongarra dongarra@eecs.utk.edu

More information

NERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber

NERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber NERSC Site Update National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory Richard Gerber NERSC Senior Science Advisor High Performance Computing Department Head Cori

More information

Deutscher Wetterdienst

Deutscher Wetterdienst Porting Operational Models to Multi- and Many-Core Architectures Ulrich Schättler Deutscher Wetterdienst Oliver Fuhrer MeteoSchweiz Xavier Lapillonne MeteoSchweiz Contents Strong Scalability of the Operational

More information

GPU Consideration for Next Generation Weather (and Climate) Simulations

GPU Consideration for Next Generation Weather (and Climate) Simulations GPU Consideration for Next Generation Weather (and Climate) Simulations Oliver Fuhrer 1, Tobias Gisy 2, Xavier Lapillonne 3, Will Sawyer 4, Ugo Varetto 4, Mauro Bianco 4, David Müller 2, and Thomas C.

More information

PLAN-E Workshop Switzerland. Welcome! September 8, 2016

PLAN-E Workshop Switzerland. Welcome! September 8, 2016 PLAN-E Workshop Switzerland Welcome! September 8, 2016 The Swiss National Supercomputing Centre Driving innovation in computational research in Switzerland Michele De Lorenzi (CSCS) PLAN-E September 8,

More information

Managing HPC Active Archive Storage with HPSS RAIT at Oak Ridge National Laboratory

Managing HPC Active Archive Storage with HPSS RAIT at Oak Ridge National Laboratory Managing HPC Active Archive Storage with HPSS RAIT at Oak Ridge National Laboratory Quinn Mitchell HPC UNIX/LINUX Storage Systems ORNL is managed by UT-Battelle for the US Department of Energy U.S. Department

More information

The Future of High Performance Computing

The Future of High Performance Computing The Future of High Performance Computing Randal E. Bryant Carnegie Mellon University http://www.cs.cmu.edu/~bryant Comparing Two Large-Scale Systems Oakridge Titan Google Data Center 2 Monolithic supercomputer

More information

Early Experiences Writing Performance Portable OpenMP 4 Codes

Early Experiences Writing Performance Portable OpenMP 4 Codes Early Experiences Writing Performance Portable OpenMP 4 Codes Verónica G. Vergara Larrea Wayne Joubert M. Graham Lopez Oscar Hernandez Oak Ridge National Laboratory Problem statement APU FPGA neuromorphic

More information

The challenges of new, efficient computer architectures, and how they can be met with a scalable software development strategy.! Thomas C.

The challenges of new, efficient computer architectures, and how they can be met with a scalable software development strategy.! Thomas C. The challenges of new, efficient computer architectures, and how they can be met with a scalable software development strategy! Thomas C. Schulthess ENES HPC Workshop, Hamburg, March 17, 2014 T. Schulthess!1

More information

Porting and Optimisation of UM on ARCHER. Karthee Sivalingam, NCAS-CMS. HPC Workshop ECMWF JWCRP

Porting and Optimisation of UM on ARCHER. Karthee Sivalingam, NCAS-CMS. HPC Workshop ECMWF JWCRP Porting and Optimisation of UM on ARCHER Karthee Sivalingam, NCAS-CMS HPC Workshop ECMWF JWCRP Acknowledgements! NCAS-CMS Bryan Lawrence Jeffrey Cole Rosalyn Hatcher Andrew Heaps David Hassell Grenville

More information

Adapting Numerical Weather Prediction codes to heterogeneous architectures: porting the COSMO model to GPUs

Adapting Numerical Weather Prediction codes to heterogeneous architectures: porting the COSMO model to GPUs Adapting Numerical Weather Prediction codes to heterogeneous architectures: porting the COSMO model to GPUs O. Fuhrer, T. Gysi, X. Lapillonne, C. Osuna, T. Dimanti, T. Schultess and the HP2C team Eidgenössisches

More information

GPU Developments for the NEMO Model. Stan Posey, HPC Program Manager, ESM Domain, NVIDIA (HQ), Santa Clara, CA, USA

GPU Developments for the NEMO Model. Stan Posey, HPC Program Manager, ESM Domain, NVIDIA (HQ), Santa Clara, CA, USA GPU Developments for the NEMO Model Stan Posey, HPC Program Manager, ESM Domain, NVIDIA (HQ), Santa Clara, CA, USA NVIDIA HPC AND ESM UPDATE TOPICS OF DISCUSSION GPU PROGRESS ON NEMO MODEL 2 NVIDIA GPU

More information

PART I - Fundamentals of Parallel Computing

PART I - Fundamentals of Parallel Computing PART I - Fundamentals of Parallel Computing Objectives What is scientific computing? The need for more computing power The need for parallel computing and parallel programs 1 What is scientific computing?

More information

Experiences with GPGPUs at HLRS

Experiences with GPGPUs at HLRS ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: Experiences with GPGPUs at HLRS Stefan Wesner, Managing Director High

More information

The Uintah Framework: A Unified Heterogeneous Task Scheduling and Runtime System

The Uintah Framework: A Unified Heterogeneous Task Scheduling and Runtime System The Uintah Framework: A Unified Heterogeneous Task Scheduling and Runtime System Alan Humphrey, Qingyu Meng, Martin Berzins Scientific Computing and Imaging Institute & University of Utah I. Uintah Overview

More information

Porting COSMO to Hybrid Architectures

Porting COSMO to Hybrid Architectures Porting COSMO to Hybrid Architectures T. Gysi 1, O. Fuhrer 2, C. Osuna 3, X. Lapillonne 3, T. Diamanti 3, B. Cumming 4, T. Schroeder 5, P. Messmer 5, T. Schulthess 4,6,7 [1] Supercomputing Systems AG,

More information

Toward portable I/O performance by leveraging system abstractions of deep memory and interconnect hierarchies

Toward portable I/O performance by leveraging system abstractions of deep memory and interconnect hierarchies Toward portable I/O performance by leveraging system abstractions of deep memory and interconnect hierarchies François Tessier, Venkatram Vishwanath, Paul Gressier Argonne National Laboratory, USA Wednesday

More information

GPU Architecture. Alan Gray EPCC The University of Edinburgh

GPU Architecture. Alan Gray EPCC The University of Edinburgh GPU Architecture Alan Gray EPCC The University of Edinburgh Outline Why do we want/need accelerators such as GPUs? Architectural reasons for accelerator performance advantages Latest GPU Products From

More information

Hybrid KAUST Many Cores and OpenACC. Alain Clo - KAUST Research Computing Saber Feki KAUST Supercomputing Lab Florent Lebeau - CAPS

Hybrid KAUST Many Cores and OpenACC. Alain Clo - KAUST Research Computing Saber Feki KAUST Supercomputing Lab Florent Lebeau - CAPS + Hybrid Computing @ KAUST Many Cores and OpenACC Alain Clo - KAUST Research Computing Saber Feki KAUST Supercomputing Lab Florent Lebeau - CAPS + Agenda Hybrid Computing n Hybrid Computing n From Multi-Physics

More information

ECMWF's Next Generation IO for the IFS Model and Product Generation

ECMWF's Next Generation IO for the IFS Model and Product Generation ECMWF's Next Generation IO for the IFS Model and Product Generation Future workflow adaptations Tiago Quintino, B. Raoult, S. Smart, A. Bonanni, F. Rathgeber, P. Bauer ECMWF tiago.quintino@ecmwf.int ECMWF

More information

John Levesque Nov 16, 2001

John Levesque Nov 16, 2001 1 We see that the GPU is the best device available for us today to be able to get to the performance we want and meet our users requirements for a very high performance node with very high memory bandwidth.

More information

Green Supercomputing

Green Supercomputing Green Supercomputing On the Energy Consumption of Modern E-Science Prof. Dr. Thomas Ludwig German Climate Computing Centre Hamburg, Germany ludwig@dkrz.de Outline DKRZ 2013 and Climate Science The Exascale

More information

The Icosahedral Nonhydrostatic (ICON) Model

The Icosahedral Nonhydrostatic (ICON) Model The Icosahedral Nonhydrostatic (ICON) Model Scalability on Massively Parallel Computer Architectures Florian Prill, DWD + the ICON team 15th ECMWF Workshop on HPC in Meteorology October 2, 2012 ICON =

More information

The DEEP (and DEEP-ER) projects

The DEEP (and DEEP-ER) projects The DEEP (and DEEP-ER) projects Estela Suarez - Jülich Supercomputing Centre BDEC for Europe Workshop Barcelona, 28.01.2015 The research leading to these results has received funding from the European

More information

Compute Node Linux (CNL) The Evolution of a Compute OS

Compute Node Linux (CNL) The Evolution of a Compute OS Compute Node Linux (CNL) The Evolution of a Compute OS Overview CNL The original scheme plan, goals, requirements Status of CNL Plans Features and directions Futures May 08 Cray Inc. Proprietary Slide

More information

The ECMWF forecast model, quo vadis?

The ECMWF forecast model, quo vadis? The forecast model, quo vadis? by Nils Wedi European Centre for Medium-Range Weather Forecasts wedi@ecmwf.int contributors: Piotr Smolarkiewicz, Mats Hamrud, George Mozdzynski, Sylvie Malardel, Christian

More information

Our Workshop Environment

Our Workshop Environment Our Workshop Environment John Urbanic Parallel Computing Scientist Pittsburgh Supercomputing Center Copyright 2015 Our Environment Today Your laptops or workstations: only used for portal access Blue Waters

More information

ICON for HD(CP) 2. High Definition Clouds and Precipitation for Advancing Climate Prediction

ICON for HD(CP) 2. High Definition Clouds and Precipitation for Advancing Climate Prediction ICON for HD(CP) 2 High Definition Clouds and Precipitation for Advancing Climate Prediction High Definition Clouds and Precipitation for Advancing Climate Prediction ICON 2 years ago Parameterize shallow

More information

CloverLeaf: Preparing Hydrodynamics Codes for Exascale

CloverLeaf: Preparing Hydrodynamics Codes for Exascale CloverLeaf: Preparing Hydrodynamics Codes for Exascale Andrew Mallinson Andy.Mallinson@awe.co.uk www.awe.co.uk British Crown Owned Copyright [2013]/AWE Agenda AWE & Uni. of Warwick introduction Problem

More information

HPC future trends from a science perspective

HPC future trends from a science perspective HPC future trends from a science perspective Simon McIntosh-Smith University of Bristol HPC Research Group simonm@cs.bris.ac.uk 1 Business as usual? We've all got used to new machines being relatively

More information

HPC IN EUROPE. Organisation of public HPC resources

HPC IN EUROPE. Organisation of public HPC resources HPC IN EUROPE Organisation of public HPC resources Context Focus on publicly-funded HPC resources provided primarily to enable scientific research and development at European universities and other publicly-funded

More information

Aim High. Intel Technical Update Teratec 07 Symposium. June 20, Stephen R. Wheat, Ph.D. Director, HPC Digital Enterprise Group

Aim High. Intel Technical Update Teratec 07 Symposium. June 20, Stephen R. Wheat, Ph.D. Director, HPC Digital Enterprise Group Aim High Intel Technical Update Teratec 07 Symposium June 20, 2007 Stephen R. Wheat, Ph.D. Director, HPC Digital Enterprise Group Risk Factors Today s s presentations contain forward-looking statements.

More information

Blue Waters I/O Performance

Blue Waters I/O Performance Blue Waters I/O Performance Mark Swan Performance Group Cray Inc. Saint Paul, Minnesota, USA mswan@cray.com Doug Petesch Performance Group Cray Inc. Saint Paul, Minnesota, USA dpetesch@cray.com Abstract

More information

A PCIe Congestion-Aware Performance Model for Densely Populated Accelerator Servers

A PCIe Congestion-Aware Performance Model for Densely Populated Accelerator Servers A PCIe Congestion-Aware Performance Model for Densely Populated Accelerator Servers Maxime Martinasso, Grzegorz Kwasniewski, Sadaf R. Alam, Thomas C. Schulthess, Torsten Hoefler Swiss National Supercomputing

More information

PERFORMANCE PORTABILITY WITH OPENACC. Jeff Larkin, NVIDIA, November 2015

PERFORMANCE PORTABILITY WITH OPENACC. Jeff Larkin, NVIDIA, November 2015 PERFORMANCE PORTABILITY WITH OPENACC Jeff Larkin, NVIDIA, November 2015 TWO TYPES OF PORTABILITY FUNCTIONAL PORTABILITY PERFORMANCE PORTABILITY The ability for a single code to run anywhere. The ability

More information

The Red Storm System: Architecture, System Update and Performance Analysis

The Red Storm System: Architecture, System Update and Performance Analysis The Red Storm System: Architecture, System Update and Performance Analysis Douglas Doerfler, Jim Tomkins Sandia National Laboratories Center for Computation, Computers, Information and Mathematics LACSI

More information

Execution Models for the Exascale Era

Execution Models for the Exascale Era Execution Models for the Exascale Era Nicholas J. Wright Advanced Technology Group, NERSC/LBNL njwright@lbl.gov Programming weather, climate, and earth- system models on heterogeneous muli- core plajorms

More information

Dynamic Load Balancing for Weather Models via AMPI

Dynamic Load Balancing for Weather Models via AMPI Dynamic Load Balancing for Eduardo R. Rodrigues IBM Research Brazil edrodri@br.ibm.com Celso L. Mendes University of Illinois USA cmendes@ncsa.illinois.edu Laxmikant Kale University of Illinois USA kale@cs.illinois.edu

More information

GPUs and Emerging Architectures

GPUs and Emerging Architectures GPUs and Emerging Architectures Mike Giles mike.giles@maths.ox.ac.uk Mathematical Institute, Oxford University e-infrastructure South Consortium Oxford e-research Centre Emerging Architectures p. 1 CPUs

More information

Aggregation of Real-Time System Monitoring Data for Analyzing Large-Scale Parallel and Distributed Computing Environments

Aggregation of Real-Time System Monitoring Data for Analyzing Large-Scale Parallel and Distributed Computing Environments Aggregation of Real-Time System Monitoring Data for Analyzing Large-Scale Parallel and Distributed Computing Environments Swen Böhm 1,2, Christian Engelmann 2, and Stephen L. Scott 2 1 Department of Computer

More information

Topology and affinity aware hierarchical and distributed load-balancing in Charm++

Topology and affinity aware hierarchical and distributed load-balancing in Charm++ Topology and affinity aware hierarchical and distributed load-balancing in Charm++ Emmanuel Jeannot, Guillaume Mercier, François Tessier Inria - IPB - LaBRI - University of Bordeaux - Argonne National

More information

Porting The Spectral Element Community Atmosphere Model (CAM-SE) To Hybrid GPU Platforms

Porting The Spectral Element Community Atmosphere Model (CAM-SE) To Hybrid GPU Platforms Porting The Spectral Element Community Atmosphere Model (CAM-SE) To Hybrid GPU Platforms http://www.scidacreview.org/0902/images/esg13.jpg Matthew Norman Jeffrey Larkin Richard Archibald Valentine Anantharaj

More information

GPU Performance Optimisation. Alan Gray EPCC The University of Edinburgh

GPU Performance Optimisation. Alan Gray EPCC The University of Edinburgh GPU Performance Optimisation EPCC The University of Edinburgh Hardware NVIDIA accelerated system: Memory Memory GPU vs CPU: Theoretical Peak capabilities NVIDIA Fermi AMD Magny-Cours (6172) Cores 448 (1.15GHz)

More information

It s a Multicore World. John Urbanic Pittsburgh Supercomputing Center

It s a Multicore World. John Urbanic Pittsburgh Supercomputing Center It s a Multicore World John Urbanic Pittsburgh Supercomputing Center Waiting for Moore s Law to save your serial code start getting bleak in 2004 Source: published SPECInt data Moore s Law is not at all

More information

Introduction CPS343. Spring Parallel and High Performance Computing. CPS343 (Parallel and HPC) Introduction Spring / 29

Introduction CPS343. Spring Parallel and High Performance Computing. CPS343 (Parallel and HPC) Introduction Spring / 29 Introduction CPS343 Parallel and High Performance Computing Spring 2018 CPS343 (Parallel and HPC) Introduction Spring 2018 1 / 29 Outline 1 Preface Course Details Course Requirements 2 Background Definitions

More information

OpenFOAM Scaling on Cray Supercomputers Dr. Stephen Sachs GOFUN 2017

OpenFOAM Scaling on Cray Supercomputers Dr. Stephen Sachs GOFUN 2017 OpenFOAM Scaling on Cray Supercomputers Dr. Stephen Sachs GOFUN 2017 Safe Harbor Statement This presentation may contain forward-looking statements that are based on our current expectations. Forward looking

More information

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 6 th CALL (Tier-0)

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 6 th CALL (Tier-0) TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 6 th CALL (Tier-0) Contributing sites and the corresponding computer systems for this call are: GCS@Jülich, Germany IBM Blue Gene/Q GENCI@CEA, France Bull Bullx

More information

Maxwell: a 64-FPGA Supercomputer

Maxwell: a 64-FPGA Supercomputer Maxwell: a 64-FPGA Supercomputer Copyright 2007, the University of Edinburgh Dr Rob Baxter Software Development Group Manager, EPCC R.Baxter@epcc.ed.ac.uk +44 131 651 3579 Outline The FHPCA Why build Maxwell?

More information

Illinois Proposal Considerations Greg Bauer

Illinois Proposal Considerations Greg Bauer - 2016 Greg Bauer Support model Blue Waters provides traditional Partner Consulting as part of its User Services. Standard service requests for assistance with porting, debugging, allocation issues, and

More information

GRIDS INTRODUCTION TO GRID INFRASTRUCTURES. Fabrizio Gagliardi

GRIDS INTRODUCTION TO GRID INFRASTRUCTURES. Fabrizio Gagliardi GRIDS INTRODUCTION TO GRID INFRASTRUCTURES Fabrizio Gagliardi Dr. Fabrizio Gagliardi is the leader of the EU DataGrid project and designated director of the proposed EGEE (Enabling Grids for E-science

More information

Accelerating HPL on Heterogeneous GPU Clusters

Accelerating HPL on Heterogeneous GPU Clusters Accelerating HPL on Heterogeneous GPU Clusters Presentation at GTC 2014 by Dhabaleswar K. (DK) Panda The Ohio State University E-mail: panda@cse.ohio-state.edu http://www.cse.ohio-state.edu/~panda Outline

More information

QLogic TrueScale InfiniBand and Teraflop Simulations

QLogic TrueScale InfiniBand and Teraflop Simulations WHITE Paper QLogic TrueScale InfiniBand and Teraflop Simulations For ANSYS Mechanical v12 High Performance Interconnect for ANSYS Computer Aided Engineering Solutions Executive Summary Today s challenging

More information

Memory Systems IRAM. Principle of IRAM

Memory Systems IRAM. Principle of IRAM Memory Systems 165 other devices of the module will be in the Standby state (which is the primary state of all RDRAM devices) or another state with low-power consumption. The RDRAM devices provide several

More information

Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy

Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy François Tessier, Venkatram Vishwanath Argonne National Laboratory, USA July 19,

More information

Hybrid Architectures Why Should I Bother?

Hybrid Architectures Why Should I Bother? Hybrid Architectures Why Should I Bother? CSCS-FoMICS-USI Summer School on Computer Simulations in Science and Engineering Michael Bader July 8 19, 2013 Computer Simulations in Science and Engineering,

More information

CUDA. Matthew Joyner, Jeremy Williams

CUDA. Matthew Joyner, Jeremy Williams CUDA Matthew Joyner, Jeremy Williams Agenda What is CUDA? CUDA GPU Architecture CPU/GPU Communication Coding in CUDA Use cases of CUDA Comparison to OpenCL What is CUDA? What is CUDA? CUDA is a parallel

More information

Introduction to GPU hardware and to CUDA

Introduction to GPU hardware and to CUDA Introduction to GPU hardware and to CUDA Philip Blakely Laboratory for Scientific Computing, University of Cambridge Philip Blakely (LSC) GPU introduction 1 / 35 Course outline Introduction to GPU hardware

More information

Compute Node Linux: Overview, Progress to Date & Roadmap

Compute Node Linux: Overview, Progress to Date & Roadmap Compute Node Linux: Overview, Progress to Date & Roadmap David Wallace Cray Inc ABSTRACT: : This presentation will provide an overview of Compute Node Linux(CNL) for the CRAY XT machine series. Compute

More information

HPC and IT Issues Session Agenda. Deployment of Simulation (Trends and Issues Impacting IT) Mapping HPC to Performance (Scaling, Technology Advances)

HPC and IT Issues Session Agenda. Deployment of Simulation (Trends and Issues Impacting IT) Mapping HPC to Performance (Scaling, Technology Advances) HPC and IT Issues Session Agenda Deployment of Simulation (Trends and Issues Impacting IT) Discussion Mapping HPC to Performance (Scaling, Technology Advances) Discussion Optimizing IT for Remote Access

More information

Overview. CS 472 Concurrent & Parallel Programming University of Evansville

Overview. CS 472 Concurrent & Parallel Programming University of Evansville Overview CS 472 Concurrent & Parallel Programming University of Evansville Selection of slides from CIS 410/510 Introduction to Parallel Computing Department of Computer and Information Science, University

More information

Introduction to parallel Computing

Introduction to parallel Computing Introduction to parallel Computing VI-SEEM Training Paschalis Paschalis Korosoglou Korosoglou (pkoro@.gr) (pkoro@.gr) Outline Serial vs Parallel programming Hardware trends Why HPC matters HPC Concepts

More information

HETEROGENEOUS HPC, ARCHITECTURAL OPTIMIZATION, AND NVLINK STEVE OBERLIN CTO, TESLA ACCELERATED COMPUTING NVIDIA

HETEROGENEOUS HPC, ARCHITECTURAL OPTIMIZATION, AND NVLINK STEVE OBERLIN CTO, TESLA ACCELERATED COMPUTING NVIDIA HETEROGENEOUS HPC, ARCHITECTURAL OPTIMIZATION, AND NVLINK STEVE OBERLIN CTO, TESLA ACCELERATED COMPUTING NVIDIA STATE OF THE ART 2012 18,688 Tesla K20X GPUs 27 PetaFLOPS FLAGSHIP SCIENTIFIC APPLICATIONS

More information

Overview of Tianhe-2

Overview of Tianhe-2 Overview of Tianhe-2 (MilkyWay-2) Supercomputer Yutong Lu School of Computer Science, National University of Defense Technology; State Key Laboratory of High Performance Computing, China ytlu@nudt.edu.cn

More information

ECMWF s Next Generation IO for the IFS Model

ECMWF s Next Generation IO for the IFS Model ECMWF s Next Generation IO for the Model Part of ECMWF s Scalability Programme Tiago Quintino, B. Raoult, P. Bauer ECMWF tiago.quintino@ecmwf.int ECMWF January 14, 2016 ECMWF s HPC Targets What do we do?

More information

Dynamical Exascale Entry Platform

Dynamical Exascale Entry Platform DEEP Dynamical Exascale Entry Platform 2 nd IS-ENES Workshop on High performance computing for climate models 30.01.2013, Toulouse, France Estela Suarez The research leading to these results has received

More information

Center Extreme Scale CS Research

Center Extreme Scale CS Research Center Extreme Scale CS Research Center for Compressible Multiphase Turbulence University of Florida Sanjay Ranka Herman Lam Outline 10 6 10 7 10 8 10 9 cores Parallelization and UQ of Rocfun and CMT-Nek

More information

Breakthrough Science via Extreme Scalability. Greg Clifford Segment Manager, Cray Inc.

Breakthrough Science via Extreme Scalability. Greg Clifford Segment Manager, Cray Inc. Breakthrough Science via Extreme Scalability Greg Clifford Segment Manager, Cray Inc. clifford@cray.com Cray s focus The requirement for highly scalable systems Cray XE6 technology The path to Exascale

More information

The Use of Cloud Computing Resources in an HPC Environment

The Use of Cloud Computing Resources in an HPC Environment The Use of Cloud Computing Resources in an HPC Environment Bill, Labate, UCLA Office of Information Technology Prakashan Korambath, UCLA Institute for Digital Research & Education Cloud computing becomes

More information