Scientific Computing at Million-way Parallelism - Blue Gene/Q Early Science Program

Size: px
Start display at page:

Download "Scientific Computing at Million-way Parallelism - Blue Gene/Q Early Science Program"

Transcription

1 Scientific Computing at Million-way Parallelism - Blue Gene/Q Early Science Program Implementing Hybrid Parallelism in FLASH Christopher Daley 1 2 Vitali Morozov 1 Dongwook Lee 2 Anshu Dubey 1 2 Jonathon Gallagher 2 Don Lamb 2 Klaus Weide 2 1 Argonne National Laboratory 2 The Flash Center for Computational Science at the University of Chicago July 10, / 26

2 1 Introduction 2 Multithreading FLASH 3 FLASH on BG/Q RTFlame White Dwarf 4 Optimizations 5 Conclusion 2 / 26

3 FLASH overview FLASH simulates problems from astrophysics, cosmology, HEDP, incompressible fluid dynamics Collection of code units which a user assembles into a custom application Portable and scalable 1/2 million lines of code Written in Fortran90 and C Parallelised with MPI and (recently) OpenMP Adaptive Mesh Refinement (AMR) with Paramesh or Chombo Parallel I/O with HDF5 or PnetCDF. 3 / 26

4 Early science goals Improve understanding of Type Ia supernova Key physical processes will be studied in various FLASH simulations The early science applications include the following two applications as well as other applications with functionality in-between the two RTFlame (flame in a rectilinear domain with constant gravity) White Dwarf (full supernova problem) 4 / 26

5 BG/P to BG/Q transition Intrepid BG/P 4 cores/node, 2 GB/node, 40,960 nodes FLASH has been run on Intrepid for the last several years Scales to the whole machine MPI-only is sufficient Run in VN mode (4 MPI ranks/node) Mira BG/Q 4 hw threads/core, 16 cores/node, 16 GB/node, 49,152 nodes MPI-only approach not suitable for BG/Q OpenMP directives have been added to FLASH to take advantage of the additional intra-node parallelism 5 / 26

6 Paramesh There exists an underlying mesh in every FLASH application We will use Paramesh during our early science time Mesh is divided into blocks of fixed size (typically 16 3 cells) Hierarchy of blocks organized as an Oct-Tree (3D) Blocks assigned to MPI ranks 6 / 26

7 MPI parallelism in Paramesh The thick black lines show blocks being assigned to a single MPI rank 6 total blocks 5 leaf blocks 1 parent block FLASH solvers update the solution on local leaf blocks We use multiple threads to speed up the solution update (details in next slides) 7 / 26

8 Multithreading strategy 1 Assign different blocks to different threads Assuming 2 threads per MPI rank Thread 0 (blue) updates 3 full blocks - 72 cells Thread 1 (yellow) updates 2 full blocks - 48 cells This will be referred to as thread block list 8 / 26

9 Multithreading strategy 2 Assign different cells from the same block to different threads Assuming 2 threads per MPI rank Thread 0 (blue) updates 5 partial blocks - 60 cells Thread 1 (yellow) updates 5 partial blocks - 60 cells This will be referred to as thread within block 9 / 26

10 RTFlame: Finding the best configuration Run a fixed problem on 128 nodes of BG/Q and find the fastest time to solution. Vary the number of MPI ranks per node OpenMP threads per MPI rank We want a configuration which gives A fast time to solution Enough run-time flexibility in terms of memory per MPI rank The figure on the next slide shows the time to solution for the fastest multithreaded strategy only This is thread within block (investigated in later slides) 10 / 26

11 RTFlame: FLASH performance on BG/Q 11 / 26

12 RTFlame: FLASH performance on BG/Q 12 / 26

13 RTFlame: FLASH performance on BG/Q 13 / 26

14 RTFlame: FLASH performance on BG/Q summary Best performance: 32 MPI ranks/node, 2 threads/mpi rank BG/Q to BG/P node-to-node ratio of 8.9x Tricky to fit application in 512MB/MPI rank The unsplit hydro solver is more memory hungry than the split hydro solver used for science runs on BG/P in previous years Best compromise: 16 MPI ranks/node, 4 threads/mpi rank BG/Q to BG/P node-to-node ratio of 8.6x Comfortable to fit application in 1GB/MPI rank. Buffers can be sized larger to accommodate Rapid refinement of problem Congregation of many tracer particles on some MPI ranks All subsequent results use this configuration 14 / 26

15 RTFlame: Strong scaling on BG/Q 15 / 26

16 RTFlame: Strong scaling on BG/Q summary Good strong scaling for both multithreading strategies Block count/mpi rank is varied by a factor of 10 Only 3 blocks of 16 3 cells on some MPI ranks for the 4096 MPI rank (256 node) data point! Finer grained threading performs slightly better Better load balancing within an MPI rank Performance advantage increases as work becomes more finely distributed. 16 / 26

17 White Dwarf: Steps towards a successful run Difficulty running the full supernova problem on BG/Q Controlled FLASH aborts because of Imaginary sound speed in unsplit hydro Non-convergence in iterative equation of state (EOS) No nice segfault to debug :( Pre-V1R1M1 driver upgrade Only able to run FLASH by turning off compiler optimization Post-V1R1M1 driver upgrade Able to run MPI+OpenMP FLASH with compiler optimization, but only by adopting a custom build strategy 17 / 26

18 White Dwarf: Custom build strategy White Dwarf application aborts when compiling all source files with the OpenMP compiler option -qsmp=omp:noauto. XLF OpenMP compiler option known to introduce additional code transformations Adopt a build strategy where we use the OpenMP compiler option only if a file contains OpenMP Strategy initially failed when using GNU compiler Need to use -frecursive option on files without OpenMP Prevents compiler placing local arrays in static memory Keeps application thread-safe We use -qnosave in case this happens with XLF compiler too 18 / 26

19 White Dwarf: Weak scaling on BG/Q 19 / 26

20 White Dwarf: Weak scaling on BG/Q summary Good weak scaling on Intrepid BG/P (up to 131K MPI ranks) and Vesta BG/Q (up to 16K MPI ranks - whole machine) Hard to create perfect weak scaling adaptive grid tests Oscillations in evolution time happen because the work per MPI rank does not stay exactly constant BG/Q to BG/P node-to-node ratio between 7.4x and 7.9x Less than RTFlame node-to-node ratio White Dwarf application includes additional physics and so more guard cell fills per time step Does not include multipole solver multithreading - removed to avoid a strange crash during access of threadprivate data 20 / 26

21 Optimizations Easy optimization opportunities found by collecting performance measurements with libmpihpm smp.a Supports sampling of program counter through vprof one sample every 0.01 seconds The 128 node RTFlame test problem originally took 202 seconds Many samples in glibc log function Linking against MASS library reduced time to 184 seconds Many samples on function call lines in unsplit hydro Reordering arrays in unsplit hydro reduced time to 156 seconds (example function call improvement in next slide) 21 / 26

22 Unsplit hydro array layout optimization c a l l h y u h d d a t a R e c o n s t O n e s t e p & l e i g ( 1 : NDIM, i, j, k, 1 :HY WAVENUM, 1 : HY VARINUM),& 712 r e i g ( 1 : NDIM, i, j, k, 1 : HY VARINUM, 1 :HY WAVENUM) ) 479 counts (approximately 4.79 seconds) 22 / 26

23 Unsplit hydro array layout optimization c a l l h y u h d d a t a R e c o n s t O n e s t e p & l e i g ( 1 : NDIM, i, j, k, 1 :HY WAVENUM, 1 : HY VARINUM),& 712 r e i g ( 1 : NDIM, i, j, k, 1 : HY VARINUM, 1 :HY WAVENUM) ) 479 counts (approximately 4.79 seconds) Reorder arrays so that i,j,k are the slowest varying dimensions allows us to pass a memory address instead of creating a temporary array 23 / 26

24 Unsplit hydro array layout optimization c a l l h y u h d d a t a R e c o n s t O n e s t e p & l e i g ( 1 : NDIM, i, j, k, 1 :HY WAVENUM, 1 : HY VARINUM),& 712 r e i g ( 1 : NDIM, i, j, k, 1 : HY VARINUM, 1 :HY WAVENUM) ) 479 counts (approximately 4.79 seconds) Reorder arrays so that i,j,k are the slowest varying dimensions allows us to pass a memory address instead of creating a temporary array c a l l h y u h d d a t a R e c o n s t O n e s t e p & l e i g ( 1, 1, 1, i, j, k ),& 712 r e i g ( 1, 1, 1, i, j, k ) ) 37 counts (approximately 0.37 seconds) 24 / 26

25 Conclusion FLASH early science applications run correctly on BG/Q Best compromise between performance and usability happens when using 16 MPI ranks / node and 4 threads / MPI rank Finer grained multithreading performs slightly better Custom build process allows us to run the more complicated White Dwarf application What is -qsmp=omp:noauto doing to MPI-only code? 25 / 26

26 Any questions? 26 / 26

27 Backup slides Vary block size in RTFlame test problem 27 / 26

Histogram sort with Sampling (HSS) Vipul Harsh, Laxmikant Kale

Histogram sort with Sampling (HSS) Vipul Harsh, Laxmikant Kale Histogram sort with Sampling (HSS) Vipul Harsh, Laxmikant Kale Parallel sorting in the age of Exascale Charm N-body GrAvity solver Massive Cosmological N-body simulations Parallel sorting in every iteration

More information

A Scalable Adaptive Mesh Refinement Framework For Parallel Astrophysics Applications

A Scalable Adaptive Mesh Refinement Framework For Parallel Astrophysics Applications A Scalable Adaptive Mesh Refinement Framework For Parallel Astrophysics Applications James Bordner, Michael L. Norman San Diego Supercomputer Center University of California, San Diego 15th SIAM Conference

More information

Parallel Algorithms: Adaptive Mesh Refinement (AMR) method and its implementation

Parallel Algorithms: Adaptive Mesh Refinement (AMR) method and its implementation Parallel Algorithms: Adaptive Mesh Refinement (AMR) method and its implementation Massimiliano Guarrasi m.guarrasi@cineca.it Super Computing Applications and Innovation Department AMR - Introduction Solving

More information

RAMSES on the GPU: An OpenACC-Based Approach

RAMSES on the GPU: An OpenACC-Based Approach RAMSES on the GPU: An OpenACC-Based Approach Claudio Gheller (ETHZ-CSCS) Giacomo Rosilho de Souza (EPFL Lausanne) Romain Teyssier (University of Zurich) Markus Wetzstein (ETHZ-CSCS) PRACE-2IP project EU

More information

Productive Performance on the Cray XK System Using OpenACC Compilers and Tools

Productive Performance on the Cray XK System Using OpenACC Compilers and Tools Productive Performance on the Cray XK System Using OpenACC Compilers and Tools Luiz DeRose Sr. Principal Engineer Programming Environments Director Cray Inc. 1 The New Generation of Supercomputers Hybrid

More information

Results from the Early Science High Speed Combus:on and Detona:on Project

Results from the Early Science High Speed Combus:on and Detona:on Project Results from the Early Science High Speed Combus:on and Detona:on Project Alexei Khokhlov, University of Chicago Joanna Aus:n, University of Illinois Charles Bacon, Argonne Na:onal Laboratory Andrew Knisely,

More information

Parallel Programming with MPI on Clusters

Parallel Programming with MPI on Clusters Parallel Programming with MPI on Clusters Rusty Lusk Mathematics and Computer Science Division Argonne National Laboratory (The rest of our group: Bill Gropp, Rob Ross, David Ashton, Brian Toonen, Anthony

More information

Experiences with ENZO on the Intel Many Integrated Core Architecture

Experiences with ENZO on the Intel Many Integrated Core Architecture Experiences with ENZO on the Intel Many Integrated Core Architecture Dr. Robert Harkness National Institute for Computational Sciences April 10th, 2012 Overview ENZO applications at petascale ENZO and

More information

Designing Parallel Programs. This review was developed from Introduction to Parallel Computing

Designing Parallel Programs. This review was developed from Introduction to Parallel Computing Designing Parallel Programs This review was developed from Introduction to Parallel Computing Author: Blaise Barney, Lawrence Livermore National Laboratory references: https://computing.llnl.gov/tutorials/parallel_comp/#whatis

More information

Performance of the hybrid MPI/OpenMP version of the HERACLES code on the Curie «Fat nodes» system

Performance of the hybrid MPI/OpenMP version of the HERACLES code on the Curie «Fat nodes» system Performance of the hybrid MPI/OpenMP version of the HERACLES code on the Curie «Fat nodes» system Edouard Audit, Matthias Gonzalez, Pierre Kestener and Pierre-François Lavallé The HERACLES code Fixed grid

More information

Enzo-P / Cello. Formation of the First Galaxies. San Diego Supercomputer Center. Department of Physics and Astronomy

Enzo-P / Cello. Formation of the First Galaxies. San Diego Supercomputer Center. Department of Physics and Astronomy Enzo-P / Cello Formation of the First Galaxies James Bordner 1 Michael L. Norman 1 Brian O Shea 2 1 University of California, San Diego San Diego Supercomputer Center 2 Michigan State University Department

More information

Peta-Scale Simulations with the HPC Software Framework walberla:

Peta-Scale Simulations with the HPC Software Framework walberla: Peta-Scale Simulations with the HPC Software Framework walberla: Massively Parallel AMR for the Lattice Boltzmann Method SIAM PP 2016, Paris April 15, 2016 Florian Schornbaum, Christian Godenschwager,

More information

Computational Astrophysics 5 Higher-order and AMR schemes

Computational Astrophysics 5 Higher-order and AMR schemes Computational Astrophysics 5 Higher-order and AMR schemes Oscar Agertz Outline - The Godunov Method - Second-order scheme with MUSCL - Slope limiters and TVD schemes - Characteristics tracing and 2D slopes.

More information

Adaptive-Mesh-Refinement Hydrodynamic GPU Computation in Astrophysics

Adaptive-Mesh-Refinement Hydrodynamic GPU Computation in Astrophysics Adaptive-Mesh-Refinement Hydrodynamic GPU Computation in Astrophysics H. Y. Schive ( 薛熙于 ) Graduate Institute of Physics, National Taiwan University Leung Center for Cosmology and Particle Astrophysics

More information

Enzo-P / Cello. Scalable Adaptive Mesh Refinement for Astrophysics and Cosmology. San Diego Supercomputer Center. Department of Physics and Astronomy

Enzo-P / Cello. Scalable Adaptive Mesh Refinement for Astrophysics and Cosmology. San Diego Supercomputer Center. Department of Physics and Astronomy Enzo-P / Cello Scalable Adaptive Mesh Refinement for Astrophysics and Cosmology James Bordner 1 Michael L. Norman 1 Brian O Shea 2 1 University of California, San Diego San Diego Supercomputer Center 2

More information

Experiences with ENZO on the Intel R Many Integrated Core (Intel MIC) Architecture

Experiences with ENZO on the Intel R Many Integrated Core (Intel MIC) Architecture Experiences with ENZO on the Intel R Many Integrated Core (Intel MIC) Architecture 1 Introduction Robert Harkness National Institute for Computational Sciences Oak Ridge National Laboratory The National

More information

Portable and Productive Performance on Hybrid Systems with libsci_acc Luiz DeRose Sr. Principal Engineer Programming Environments Director Cray Inc.

Portable and Productive Performance on Hybrid Systems with libsci_acc Luiz DeRose Sr. Principal Engineer Programming Environments Director Cray Inc. Portable and Productive Performance on Hybrid Systems with libsci_acc Luiz DeRose Sr. Principal Engineer Programming Environments Director Cray Inc. 1 What is Cray Libsci_acc? Provide basic scientific

More information

ENERGY-224 Reservoir Simulation Project Report. Ala Alzayer

ENERGY-224 Reservoir Simulation Project Report. Ala Alzayer ENERGY-224 Reservoir Simulation Project Report Ala Alzayer Autumn Quarter December 3, 2014 Contents 1 Objective 2 2 Governing Equations 2 3 Methodolgy 3 3.1 BlockMesh.........................................

More information

About Phoenix FD PLUGIN FOR 3DS MAX AND MAYA. SIMULATING AND RENDERING BOTH LIQUIDS AND FIRE/SMOKE. USED IN MOVIES, GAMES AND COMMERCIALS.

About Phoenix FD PLUGIN FOR 3DS MAX AND MAYA. SIMULATING AND RENDERING BOTH LIQUIDS AND FIRE/SMOKE. USED IN MOVIES, GAMES AND COMMERCIALS. About Phoenix FD PLUGIN FOR 3DS MAX AND MAYA. SIMULATING AND RENDERING BOTH LIQUIDS AND FIRE/SMOKE. USED IN MOVIES, GAMES AND COMMERCIALS. Phoenix FD core SIMULATION & RENDERING. SIMULATION CORE - GRID-BASED

More information

Adaptive Mesh Refinement in Titanium

Adaptive Mesh Refinement in Titanium Adaptive Mesh Refinement in Titanium http://seesar.lbl.gov/anag Lawrence Berkeley National Laboratory April 7, 2005 19 th IPDPS, April 7, 2005 1 Overview Motivations: Build the infrastructure in Titanium

More information

Simulation-time data analysis and I/O acceleration at extreme scale with GLEAN

Simulation-time data analysis and I/O acceleration at extreme scale with GLEAN Simulation-time data analysis and I/O acceleration at extreme scale with GLEAN Venkatram Vishwanath, Mark Hereld and Michael E. Papka Argonne Na

More information

LAMMPS, FLASH and MESA

LAMMPS, FLASH and MESA , and 22 June 2017 Westgrid Summer School, and What is? is a classical molecular dynamics simulator. Serial or parallel. GPU or CPU Not just classical: PIMD, Lattice Boltzmann Fluid,... Adapted from http://lammps.sandia.gov/tutorials.html,

More information

Introduction to High Performance Parallel I/O

Introduction to High Performance Parallel I/O Introduction to High Performance Parallel I/O Richard Gerber Deputy Group Lead NERSC User Services August 30, 2013-1- Some slides from Katie Antypas I/O Needs Getting Bigger All the Time I/O needs growing

More information

Intro to Parallel Computing

Intro to Parallel Computing Outline Intro to Parallel Computing Remi Lehe Lawrence Berkeley National Laboratory Modern parallel architectures Parallelization between nodes: MPI Parallelization within one node: OpenMP Why use parallel

More information

Advanced Parallel Programming. Is there life beyond MPI?

Advanced Parallel Programming. Is there life beyond MPI? Advanced Parallel Programming Is there life beyond MPI? Outline MPI vs. High Level Languages Declarative Languages Map Reduce and Hadoop Shared Global Address Space Languages Charm++ ChaNGa ChaNGa on GPUs

More information

Hybrid programming with MPI and OpenMP On the way to exascale

Hybrid programming with MPI and OpenMP On the way to exascale Institut du Développement et des Ressources en Informatique Scientifique www.idris.fr Hybrid programming with MPI and OpenMP On the way to exascale 1 Trends of hardware evolution Main problematic : how

More information

HANDLING LOAD IMBALANCE IN DISTRIBUTED & SHARED MEMORY

HANDLING LOAD IMBALANCE IN DISTRIBUTED & SHARED MEMORY HANDLING LOAD IMBALANCE IN DISTRIBUTED & SHARED MEMORY Presenters: Harshitha Menon, Seonmyeong Bak PPL Group Phil Miller, Sam White, Nitin Bhat, Tom Quinn, Jim Phillips, Laxmikant Kale MOTIVATION INTEGRATED

More information

Hybrid Model Parallel Programs

Hybrid Model Parallel Programs Hybrid Model Parallel Programs Charlie Peck Intermediate Parallel Programming and Cluster Computing Workshop University of Oklahoma/OSCER, August, 2010 1 Well, How Did We Get Here? Almost all of the clusters

More information

Development of a Computational Framework for Block-Based AMR Simulations

Development of a Computational Framework for Block-Based AMR Simulations Procedia Computer Science Volume 29, 2014, Pages 2351 2359 ICCS 2014. 14th International Conference on Computational Science Development of a Computational Framework for Block-Based AMR Simulations Hideyuki

More information

Introducing Overdecomposition to Existing Applications: PlasComCM and AMPI

Introducing Overdecomposition to Existing Applications: PlasComCM and AMPI Introducing Overdecomposition to Existing Applications: PlasComCM and AMPI Sam White Parallel Programming Lab UIUC 1 Introduction How to enable Overdecomposition, Asynchrony, and Migratability in existing

More information

From Physics Model to Results: An Optimizing Framework for Cross-Architecture Code Generation

From Physics Model to Results: An Optimizing Framework for Cross-Architecture Code Generation From Physics Model to Results: An Optimizing Framework for Cross-Architecture Code Generation Erik Schnetter, Perimeter Institute with M. Blazewicz, I. Hinder, D. Koppelman, S. Brandt, M. Ciznicki, M.

More information

Developing the TELEMAC system for HECToR (phase 2b & beyond) Zhi Shang

Developing the TELEMAC system for HECToR (phase 2b & beyond) Zhi Shang Developing the TELEMAC system for HECToR (phase 2b & beyond) Zhi Shang Outline of the Talk Introduction to the TELEMAC System and to TELEMAC-2D Code Developments Data Reordering Strategy Results Conclusions

More information

Transport Simulations beyond Petascale. Jing Fu (ANL)

Transport Simulations beyond Petascale. Jing Fu (ANL) Transport Simulations beyond Petascale Jing Fu (ANL) A) Project Overview The project: Peta- and exascale algorithms and software development (petascalable codes: Nek5000, NekCEM, NekLBM) Science goals:

More information

Accelerating Octo-Tiger: Stellar Mergers on Intel Knights Landing with HPX

Accelerating Octo-Tiger: Stellar Mergers on Intel Knights Landing with HPX Accelerating Octo-Tiger: Stellar Mergers on Intel Knights Landing with HPX David Pfander*, Gregor Daiß*, Dominic Marcello**, Hartmut Kaiser**, Dirk Pflüger* * University of Stuttgart ** Louisiana State

More information

Portable and Productive Performance with OpenACC Compilers and Tools. Luiz DeRose Sr. Principal Engineer Programming Environments Director Cray Inc.

Portable and Productive Performance with OpenACC Compilers and Tools. Luiz DeRose Sr. Principal Engineer Programming Environments Director Cray Inc. Portable and Productive Performance with OpenACC Compilers and Tools Luiz DeRose Sr. Principal Engineer Programming Environments Director Cray Inc. 1 Cray: Leadership in Computational Research Earth Sciences

More information

Fault tolerant issues in large scale applications

Fault tolerant issues in large scale applications Fault tolerant issues in large scale applications Romain Teyssier George Lake, Ben Moore, Joachim Stadel and the other members of the project «Cosmology at the petascale» SPEEDUP 2010 1 Outline Computational

More information

Adaptive Refinement Tree (ART) code. N-Body: Parallelization using OpenMP and MPI

Adaptive Refinement Tree (ART) code. N-Body: Parallelization using OpenMP and MPI Adaptive Refinement Tree (ART) code N-Body: Parallelization using OpenMP and MPI 1 Family of codes N-body: OpenMp N-body: MPI+OpenMP N-body+hydro+cooling+SF: OpenMP N-body+hydro+cooling+SF: MPI 2 History:

More information

Damaris. In-Situ Data Analysis and Visualization for Large-Scale HPC Simulations. KerData Team. Inria Rennes,

Damaris. In-Situ Data Analysis and Visualization for Large-Scale HPC Simulations. KerData Team. Inria Rennes, Damaris In-Situ Data Analysis and Visualization for Large-Scale HPC Simulations KerData Team Inria Rennes, http://damaris.gforge.inria.fr Outline 1. From I/O to in-situ visualization 2. Damaris approach

More information

Cosmological Particle Mesh Simulations in Chapel. Nikhil Padmanabhan & Ben Albrecht Yale University Cray Inc PAW 2017

Cosmological Particle Mesh Simulations in Chapel. Nikhil Padmanabhan & Ben Albrecht Yale University Cray Inc PAW 2017 Cosmological Particle Mesh Simulations in Chapel Nikhil Padmanabhan & Ben Albrecht Yale University Cray Inc PAW 2017 My perspective Astrophysicist who occassionally writes code. Need to be able to easily

More information

Software and Performance Engineering for numerical codes on GPU clusters

Software and Performance Engineering for numerical codes on GPU clusters Software and Performance Engineering for numerical codes on GPU clusters H. Köstler International Workshop of GPU Solutions to Multiscale Problems in Science and Engineering Harbin, China 28.7.2010 2 3

More information

Optimization of PIERNIK for the Multiscale Simulations of High-Redshift Disk Galaxies

Optimization of PIERNIK for the Multiscale Simulations of High-Redshift Disk Galaxies Available online at www.prace-ri.eu Partnership for Advanced Computing in Europe Optimization of PIERNIK for the Multiscale Simulations of High-Redshift Disk Galaxies Kacper Kowalik a, Artur Gawryszczak

More information

Ateles performance assessment report

Ateles performance assessment report Ateles performance assessment report Document Information Reference Number Author Contributor(s) Date Application Service Level Keywords AR-4, Version 0.1 Jose Gracia (USTUTT-HLRS) Christoph Niethammer,

More information

Parallel Mesh Partitioning in Alya

Parallel Mesh Partitioning in Alya Available online at www.prace-ri.eu Partnership for Advanced Computing in Europe Parallel Mesh Partitioning in Alya A. Artigues a *** and G. Houzeaux a* a Barcelona Supercomputing Center ***antoni.artigues@bsc.es

More information

Mixed Mode MPI / OpenMP Programming

Mixed Mode MPI / OpenMP Programming Mixed Mode MPI / OpenMP Programming L.A. Smith Edinburgh Parallel Computing Centre, Edinburgh, EH9 3JZ 1 Introduction Shared memory architectures are gradually becoming more prominent in the HPC market,

More information

Evaluation of PGAS Communication Paradigms With Geometric Multigrid

Evaluation of PGAS Communication Paradigms With Geometric Multigrid Lawrence Berkeley National Laboratory Evaluation of PGAS Communication Paradigms With Geometric Multigrid Hongzhang Shan, Amir Kamil, Samuel Williams, Yili Zheng, and Katherine Yelick Lawrence Berkeley

More information

LARGE-EDDY EDDY SIMULATION CODE FOR CITY SCALE ENVIRONMENTS

LARGE-EDDY EDDY SIMULATION CODE FOR CITY SCALE ENVIRONMENTS ARCHER ECSE 05-14 LARGE-EDDY EDDY SIMULATION CODE FOR CITY SCALE ENVIRONMENTS TECHNICAL REPORT Vladimír Fuka, Zheng-Tong Xie Abstract The atmospheric large eddy simulation code ELMM (Extended Large-eddy

More information

Splotch: High Performance Visualization using MPI, OpenMP and CUDA

Splotch: High Performance Visualization using MPI, OpenMP and CUDA Splotch: High Performance Visualization using MPI, OpenMP and CUDA Klaus Dolag (Munich University Observatory) Martin Reinecke (MPA, Garching) Claudio Gheller (CSCS, Switzerland), Marzia Rivi (CINECA,

More information

ORAP Forum October 10, 2013

ORAP Forum October 10, 2013 Towards Petaflop simulations of core collapse supernovae ORAP Forum October 10, 2013 Andreas Marek 1 together with Markus Rampp 1, Florian Hanke 2, and Thomas Janka 2 1 Rechenzentrum der Max-Planck-Gesellschaft

More information

Introduction to Parallel Programming for Multicore/Manycore Clusters Part II-3: Parallel FVM using MPI

Introduction to Parallel Programming for Multicore/Manycore Clusters Part II-3: Parallel FVM using MPI Introduction to Parallel Programming for Multi/Many Clusters Part II-3: Parallel FVM using MPI Kengo Nakajima Information Technology Center The University of Tokyo 2 Overview Introduction Local Data Structure

More information

Parallel Computing Using OpenMP/MPI. Presented by - Jyotsna 29/01/2008

Parallel Computing Using OpenMP/MPI. Presented by - Jyotsna 29/01/2008 Parallel Computing Using OpenMP/MPI Presented by - Jyotsna 29/01/2008 Serial Computing Serially solving a problem Parallel Computing Parallelly solving a problem Parallel Computer Memory Architecture Shared

More information

Particle-in-Cell Simulations on Modern Computing Platforms. Viktor K. Decyk and Tajendra V. Singh UCLA

Particle-in-Cell Simulations on Modern Computing Platforms. Viktor K. Decyk and Tajendra V. Singh UCLA Particle-in-Cell Simulations on Modern Computing Platforms Viktor K. Decyk and Tajendra V. Singh UCLA Outline of Presentation Abstraction of future computer hardware PIC on GPUs OpenCL and Cuda Fortran

More information

Code Saturne on POWER8 clusters: First Investigations

Code Saturne on POWER8 clusters: First Investigations Code Saturne on POWER8 clusters: First Investigations C. MOULINEC, V. SZEREMI, D.R. EMERSON (STFC Daresbury Lab., UK) Y. FOURNIER (EDF R&D, FR) P. VEZOLLE, L. ENAULT (IBM Montpellier, FR) B. ANLAUF, M.

More information

AACE: Applications. Director, Application Acceleration Center of Excellence National Institute for Computational Sciences glenn-

AACE: Applications. Director, Application Acceleration Center of Excellence National Institute for Computational Sciences glenn- AACE: Applications R. Glenn Brook Director, Application Acceleration Center of Excellence National Institute for Computational Sciences glenn- brook@tennessee.edu Ryan C. Hulguin Computational Science

More information

Hybrid OpenMP-MPI Turbulent boundary Layer code over 32k cores

Hybrid OpenMP-MPI Turbulent boundary Layer code over 32k cores Hybrid OpenMP-MPI Turbulent boundary Layer code over 32k cores T/NT INTERFACE y/ x/ z/ 99 99 Juan A. Sillero, Guillem Borrell, Javier Jiménez (Universidad Politécnica de Madrid) and Robert D. Moser (U.

More information

simulation framework for piecewise regular grids

simulation framework for piecewise regular grids WALBERLA, an ultra-scalable multiphysics simulation framework for piecewise regular grids ParCo 2015, Edinburgh September 3rd, 2015 Christian Godenschwager, Florian Schornbaum, Martin Bauer, Harald Köstler

More information

Introduction to MPI. EAS 520 High Performance Scientific Computing. University of Massachusetts Dartmouth. Spring 2014

Introduction to MPI. EAS 520 High Performance Scientific Computing. University of Massachusetts Dartmouth. Spring 2014 Introduction to MPI EAS 520 High Performance Scientific Computing University of Massachusetts Dartmouth Spring 2014 References This presentation is almost an exact copy of Dartmouth College's Introduction

More information

Multigrid Pattern. I. Problem. II. Driving Forces. III. Solution

Multigrid Pattern. I. Problem. II. Driving Forces. III. Solution Multigrid Pattern I. Problem Problem domain is decomposed into a set of geometric grids, where each element participates in a local computation followed by data exchanges with adjacent neighbors. The grids

More information

Application / User Viewpoint

Application / User Viewpoint SC 07 Reno, Nevada November 15, 2007 Fortran@50 Application / User Viewpoint Henry Tufo Computer Science Section Head Associate Professor and Director, Computational Science Center Department of Computer

More information

Introduction to Parallel Computing

Introduction to Parallel Computing Portland State University ECE 588/688 Introduction to Parallel Computing Reference: Lawrence Livermore National Lab Tutorial https://computing.llnl.gov/tutorials/parallel_comp/ Copyright by Alaa Alameldeen

More information

Domain Decomposition: Computational Fluid Dynamics

Domain Decomposition: Computational Fluid Dynamics Domain Decomposition: Computational Fluid Dynamics December 0, 0 Introduction and Aims This exercise takes an example from one of the most common applications of HPC resources: Fluid Dynamics. We will

More information

CMSC 714 Lecture 6 MPI vs. OpenMP and OpenACC. Guest Lecturer: Sukhyun Song (original slides by Alan Sussman)

CMSC 714 Lecture 6 MPI vs. OpenMP and OpenACC. Guest Lecturer: Sukhyun Song (original slides by Alan Sussman) CMSC 714 Lecture 6 MPI vs. OpenMP and OpenACC Guest Lecturer: Sukhyun Song (original slides by Alan Sussman) Parallel Programming with Message Passing and Directives 2 MPI + OpenMP Some applications can

More information

Toward portable I/O performance by leveraging system abstractions of deep memory and interconnect hierarchies

Toward portable I/O performance by leveraging system abstractions of deep memory and interconnect hierarchies Toward portable I/O performance by leveraging system abstractions of deep memory and interconnect hierarchies François Tessier, Venkatram Vishwanath, Paul Gressier Argonne National Laboratory, USA Wednesday

More information

Adventures in Load Balancing at Scale: Successes, Fizzles, and Next Steps

Adventures in Load Balancing at Scale: Successes, Fizzles, and Next Steps Adventures in Load Balancing at Scale: Successes, Fizzles, and Next Steps Rusty Lusk Mathematics and Computer Science Division Argonne National Laboratory Outline Introduction Two abstract programming

More information

Performance of the 3D-Combustion Simulation Code RECOM-AIOLOS on IBM POWER8 Architecture. Alexander Berreth. Markus Bühler, Benedikt Anlauf

Performance of the 3D-Combustion Simulation Code RECOM-AIOLOS on IBM POWER8 Architecture. Alexander Berreth. Markus Bühler, Benedikt Anlauf PADC Anual Workshop 20 Performance of the 3D-Combustion Simulation Code RECOM-AIOLOS on IBM POWER8 Architecture Alexander Berreth RECOM Services GmbH, Stuttgart Markus Bühler, Benedikt Anlauf IBM Deutschland

More information

The IBM Blue Gene/Q: Application performance, scalability and optimisation

The IBM Blue Gene/Q: Application performance, scalability and optimisation The IBM Blue Gene/Q: Application performance, scalability and optimisation Mike Ashworth, Andrew Porter Scientific Computing Department & STFC Hartree Centre Manish Modani IBM STFC Daresbury Laboratory,

More information

Cosmology Simulations with Enzo

Cosmology Simulations with Enzo Cosmology Simulations with Enzo John Wise (Georgia Tech) Enzo Workshop 17 Oct 2013 Outline Introduction to unigrid cosmology simulations Introduction to nested grid cosmology simulations Using different

More information

Dynamic Load Distributions for Adaptive Computations on MIMD Machines using Hybrid Genetic Algorithms (a subset)

Dynamic Load Distributions for Adaptive Computations on MIMD Machines using Hybrid Genetic Algorithms (a subset) Dynamic Load Distributions for Adaptive Computations on MIMD Machines using Hybrid Genetic Algorithms (a subset) TIMOTHY H. KAISER PhD, Computer Science - UNM MS, Electrical Engineering - (Applied Physics)

More information

Fourteen years of Cactus Community

Fourteen years of Cactus Community Fourteen years of Cactus Community Frank Löffler Center for Computation and Technology Louisiana State University, Baton Rouge, LA September 6th 2012 Outline Motivation scenario from Astrophysics Cactus

More information

Review of previous examinations TMA4280 Introduction to Supercomputing

Review of previous examinations TMA4280 Introduction to Supercomputing Review of previous examinations TMA4280 Introduction to Supercomputing NTNU, IMF April 24. 2017 1 Examination The examination is usually comprised of: one problem related to linear algebra operations with

More information

SHARCNET Workshop on Parallel Computing. Hugh Merz Laurentian University May 2008

SHARCNET Workshop on Parallel Computing. Hugh Merz Laurentian University May 2008 SHARCNET Workshop on Parallel Computing Hugh Merz Laurentian University May 2008 What is Parallel Computing? A computational method that utilizes multiple processing elements to solve a problem in tandem

More information

Introduction to parallel Computing

Introduction to parallel Computing Introduction to parallel Computing VI-SEEM Training Paschalis Paschalis Korosoglou Korosoglou (pkoro@.gr) (pkoro@.gr) Outline Serial vs Parallel programming Hardware trends Why HPC matters HPC Concepts

More information

Towards Exascale Programming Models HPC Summit, Prague Erwin Laure, KTH

Towards Exascale Programming Models HPC Summit, Prague Erwin Laure, KTH Towards Exascale Programming Models HPC Summit, Prague Erwin Laure, KTH 1 Exascale Programming Models With the evolution of HPC architecture towards exascale, new approaches for programming these machines

More information

Particle-based simulations in Astrophysics

Particle-based simulations in Astrophysics Particle-based simulations in Astrophysics Jun Makino Particle Simulator Research Team, AICS/ Earth-Life Science Institute(ELSI), Tokyo Institute of Technology Feb 28, 2013 3rd AICS International Symposium

More information

PiTP Summer School 2009

PiTP Summer School 2009 PiTP Summer School 2009 Plan for my lectures Volker Springel Lecture 1 Basics of collisionless dynamics and the N-body approach Lecture 2 Gravitational solvers suitable for collisionless dynamics, parallelization

More information

Hybrid MPI + OpenMP Approach to Improve the Scalability of a Phase-Field-Crystal Code

Hybrid MPI + OpenMP Approach to Improve the Scalability of a Phase-Field-Crystal Code Hybrid MPI + OpenMP Approach to Improve the Scalability of a Phase-Field-Crystal Code Reuben D. Budiardja reubendb@utk.edu ECSS Symposium March 19 th, 2013 Project Background Project Team (University of

More information

HACC: Extreme Scaling and Performance Across Diverse Architectures

HACC: Extreme Scaling and Performance Across Diverse Architectures HACC: Extreme Scaling and Performance Across Diverse Architectures Salman Habib HEP and MCS Divisions Argonne National Laboratory Vitali Morozov Nicholas Frontiere Hal Finkel Adrian Pope Katrin Heitmann

More information

Performance Characteristics of Hybrid MPI/OpenMP Scientific Applications on a Largescale Multithreaded BlueGene/Q Supercomputer

Performance Characteristics of Hybrid MPI/OpenMP Scientific Applications on a Largescale Multithreaded BlueGene/Q Supercomputer Performance Characteristics of Hybrid MPI/OpenMP Scientific Applications on a Largescale Multithreaded BlueGene/Q Supercomputer Xingfu Wu Department of Computer Science & Engineering, Texas A&M University

More information

CFD exercise. Regular domain decomposition

CFD exercise. Regular domain decomposition CFD exercise Regular domain decomposition Reusing this material This work is licensed under a Creative Commons Attribution- NonCommercial-ShareAlike 4.0 International License. http://creativecommons.org/licenses/by-nc-sa/4.0/deed.en_us

More information

Designing a Domain-specific Language to Simulate Particles. dan bailey

Designing a Domain-specific Language to Simulate Particles. dan bailey Designing a Domain-specific Language to Simulate Particles dan bailey Double Negative Largest Visual Effects studio in Europe Offices in London and Singapore Large and growing R & D team Squirt Fluid Solver

More information

GAMER : a GPU-accelerated Adaptive-MEsh-Refinement Code for Astrophysics GPU 與自適性網格於天文模擬之應用與效能

GAMER : a GPU-accelerated Adaptive-MEsh-Refinement Code for Astrophysics GPU 與自適性網格於天文模擬之應用與效能 GAMER : a GPU-accelerated Adaptive-MEsh-Refinement Code for Astrophysics GPU 與自適性網格於天文模擬之應用與效能 Hsi-Yu Schive ( 薛熙于 ), Tzihong Chiueh ( 闕志鴻 ), Yu-Chih Tsai ( 蔡御之 ), Ui-Han Zhang ( 張瑋瀚 ) Graduate Institute

More information

Comparison and analysis of parallel tasking performance for an irregular application

Comparison and analysis of parallel tasking performance for an irregular application Comparison and analysis of parallel tasking performance for an irregular application Patrick Atkinson, University of Bristol (p.atkinson@bristol.ac.uk) Simon McIntosh-Smith, University of Bristol Motivation

More information

Domain Decomposition: Computational Fluid Dynamics

Domain Decomposition: Computational Fluid Dynamics Domain Decomposition: Computational Fluid Dynamics May 24, 2015 1 Introduction and Aims This exercise takes an example from one of the most common applications of HPC resources: Fluid Dynamics. We will

More information

What is Cactus? Cactus is a framework for developing portable, modular applications

What is Cactus? Cactus is a framework for developing portable, modular applications What is Cactus? Cactus is a framework for developing portable, modular applications What is Cactus? Cactus is a framework for developing portable, modular applications focusing, although not exclusively,

More information

Optimizing TELEMAC-2D for Large-scale Flood Simulations

Optimizing TELEMAC-2D for Large-scale Flood Simulations Available on-line at www.prace-ri.eu Partnership for Advanced Computing in Europe Optimizing TELEMAC-2D for Large-scale Flood Simulations Charles Moulinec a,, Yoann Audouin a, Andrew Sunderland a a STFC

More information

Analyzing the Performance of IWAVE on a Cluster using HPCToolkit

Analyzing the Performance of IWAVE on a Cluster using HPCToolkit Analyzing the Performance of IWAVE on a Cluster using HPCToolkit John Mellor-Crummey and Laksono Adhianto Department of Computer Science Rice University {johnmc,laksono}@rice.edu TRIP Meeting March 30,

More information

A More Realistic Way of Stressing the End-to-end I/O System

A More Realistic Way of Stressing the End-to-end I/O System A More Realistic Way of Stressing the End-to-end I/O System Verónica G. Vergara Larrea Sarp Oral Dustin Leverman Hai Ah Nam Feiyi Wang James Simmons CUG 2015 April 29, 2015 Chicago, IL ORNL is managed

More information

Demonstration of Legion Runtime Using the PENNANT Mini-App

Demonstration of Legion Runtime Using the PENNANT Mini-App Demonstration of Legion Runtime Using the PENNANT Mini-App Charles Ferenbaugh Los Alamos National Laboratory LA-UR-14-29180 1 A brief overview of PENNANT Implements a small subset of basic physics from

More information

Trends and Challenges in Multicore Programming

Trends and Challenges in Multicore Programming Trends and Challenges in Multicore Programming Eva Burrows Bergen Language Design Laboratory (BLDL) Department of Informatics, University of Bergen Bergen, March 17, 2010 Outline The Roadmap of Multicores

More information

computational Fluid Dynamics - Prof. V. Esfahanian

computational Fluid Dynamics - Prof. V. Esfahanian Three boards categories: Experimental Theoretical Computational Crucial to know all three: Each has their advantages and disadvantages. Require validation and verification. School of Mechanical Engineering

More information

GPU Debugging Made Easy. David Lecomber CTO, Allinea Software

GPU Debugging Made Easy. David Lecomber CTO, Allinea Software GPU Debugging Made Easy David Lecomber CTO, Allinea Software david@allinea.com Allinea Software HPC development tools company Leading in HPC software tools market Wide customer base Blue-chip engineering,

More information

Cactus Framework: Scaling and Lessons Learnt

Cactus Framework: Scaling and Lessons Learnt Cactus Framework: Scaling and Lessons Learnt Gabrielle Allen, Erik Schnetter, Jian Tao Center for Computation & Technology Departments of Computer Science & Physics Louisiana State University Also: Christian

More information

Shape of Things to Come: Next-Gen Physics Deep Dive

Shape of Things to Come: Next-Gen Physics Deep Dive Shape of Things to Come: Next-Gen Physics Deep Dive Jean Pierre Bordes NVIDIA Corporation Free PhysX on CUDA PhysX by NVIDIA since March 2008 PhysX on CUDA available: August 2008 GPU PhysX in Games Physical

More information

Placement de processus (MPI) sur architecture multi-cœur NUMA

Placement de processus (MPI) sur architecture multi-cœur NUMA Placement de processus (MPI) sur architecture multi-cœur NUMA Emmanuel Jeannot, Guillaume Mercier LaBRI/INRIA Bordeaux Sud-Ouest/ENSEIRB Runtime Team Lyon, journées groupe de calcul, november 2010 Emmanuel.Jeannot@inria.fr

More information

A Study of High Performance Computing and the Cray SV1 Supercomputer. Michael Sullivan TJHSST Class of 2004

A Study of High Performance Computing and the Cray SV1 Supercomputer. Michael Sullivan TJHSST Class of 2004 A Study of High Performance Computing and the Cray SV1 Supercomputer Michael Sullivan TJHSST Class of 2004 June 2004 0.1 Introduction A supercomputer is a device for turning compute-bound problems into

More information

Scalable Debugging with TotalView on Blue Gene. John DelSignore, CTO TotalView Technologies

Scalable Debugging with TotalView on Blue Gene. John DelSignore, CTO TotalView Technologies Scalable Debugging with TotalView on Blue Gene John DelSignore, CTO TotalView Technologies Agenda TotalView on Blue Gene A little history Current status Recent TotalView improvements ReplayEngine (reverse

More information

Porting a parallel rotor wake simulation to GPGPU accelerators using OpenACC

Porting a parallel rotor wake simulation to GPGPU accelerators using OpenACC DLR.de Chart 1 Porting a parallel rotor wake simulation to GPGPU accelerators using OpenACC Melven Röhrig-Zöllner DLR, Simulations- und Softwaretechnik DLR.de Chart 2 Outline Hardware-Architecture (CPU+GPU)

More information

SCALASCA parallel performance analyses of SPEC MPI2007 applications

SCALASCA parallel performance analyses of SPEC MPI2007 applications Mitglied der Helmholtz-Gemeinschaft SCALASCA parallel performance analyses of SPEC MPI2007 applications 2008-05-22 Zoltán Szebenyi Jülich Supercomputing Centre, Forschungszentrum Jülich Aachen Institute

More information

Aerodynamic Study of a Realistic Car W. TOUGERON

Aerodynamic Study of a Realistic Car W. TOUGERON Aerodynamic Study of a Realistic Car W. TOUGERON Tougeron CFD Engineer 2016 Abstract This document presents an aerodynamic CFD study of a realistic car geometry. The aim is to demonstrate the efficiency

More information

MPI Casestudy: Parallel Image Processing

MPI Casestudy: Parallel Image Processing MPI Casestudy: Parallel Image Processing David Henty 1 Introduction The aim of this exercise is to write a complete MPI parallel program that does a very basic form of image processing. We will start by

More information

Load Balancing and Data Migration in a Hybrid Computational Fluid Dynamics Application

Load Balancing and Data Migration in a Hybrid Computational Fluid Dynamics Application Load Balancing and Data Migration in a Hybrid Computational Fluid Dynamics Application Esteban Meneses Patrick Pisciuneri Center for Simulation and Modeling (SaM) University of Pittsburgh University of

More information