Enzo-P / Cello. Scalable Adaptive Mesh Refinement for Astrophysics and Cosmology. San Diego Supercomputer Center. Department of Physics and Astronomy

Size: px
Start display at page:

Download "Enzo-P / Cello. Scalable Adaptive Mesh Refinement for Astrophysics and Cosmology. San Diego Supercomputer Center. Department of Physics and Astronomy"

Transcription

1 Enzo-P / Cello Scalable Adaptive Mesh Refinement for Astrophysics and Cosmology James Bordner 1 Michael L. Norman 1 Brian O Shea 2 1 University of California, San Diego San Diego Supercomputer Center 2 Michigan State University Department of Physics and Astronomy Extreme Scaling Workshop 2012 Blue Waters / XSEDE Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

2 Enzo overview Parallel astrophysics and cosmology implemented in C++ /Fortran approximately 150K SLOC parallelized using MPI / OpenMP Vast range of scales astrophysical fluid dynamics hydrodynamic cosmology Adaptive mesh refinement (AMR) Growing development community [ Norman et al ] Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

3 Enzo s physics and algorithms Eulerian hydrodynamics piecewise-parabolic method (PPM) Lagrangian dark matter particle-mesh method (PM) Self-gravity FFT s on root-level grid multigrid on refinement patches Local physics heating, cooling, chemistry, etc. MHD, RHD (ray-tracing or implicit FLD) [JohnWise] Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

4 Enzo s pursuit of scalability Enzo born in early 1990 s Extreme meant 100 processors Continual scalability improvements MPI/OpenMP parallelism neighbor-finding algorithm I/O optimizations Further improvement getting harder increasing scalability requirements easy improvements already made Motivates concurrent rewriting Enzo-P petascale Enzo fork Cello AMR framework [SamSkillman,MattTurk] Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

5 Enzo s AMR data structure Patch-based SAMR Each patch is a C++ object root grid root patch refinement patch Patches assigned to processes N P root patches, grid size 64 3, Refinement patches generally smaller Refinement patches initially local to parent Load balancing relocates refinement patches Patch data (grids, particles) are distributed AMR hierarchy structure is replicated [TomAbel,JohnWise,RalfKaehler] Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

6 Enzo s timestepping Adaptive timestepping by level parallel within level less computation by O(N L ) reduced parallel efficiency EvolveLevel(L,dt L 1 ) 1 refresh level L ghosts 2 compute timestep dt L 3 advance level L by dt L 4 EvolveLevel(L+1,dt L ) 5 correct fluxes t t L=0 L=1 L=2 L= dt dt1 3 dt 2 x x Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

7 Enzo s scaling issues Memory usage AMR structure is non-scalable ghost zone layer three zones deep memory fragmentation Data locality disrupted by load balancing Parallel task definition widely varying patch sizes granularity determined by AMR Parallel task scheduling parallel within a level synchronization between levels [ElizabethTasker] Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

8 Talk outline 1 Enzo 1 overview 2 design 3 scaling issues 2 Talk outline 3 Enzo-P / Cello 1 overview 2 design 3 scaling solutions 4 implementation using Charm++ 5 recursively generated parallel data structures Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

9 Enzo-P / Cello overview Enzo-P intended to be petascale Enzo Cello is a scalable AMR framework Parallelism using Charm++ MPI as a backup 25K SLOC Work in progress PPM HD / PPML MHD on distributed Cartesian grids prototyping Charm++ AMR implementations Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

10 Cello s AMR data structure Forest Tree Patch Block Tree-based SAMR Patches define unit of refinement cubical, varying size Blocks define parallel tasks flexible size and shape One Block per Charm++ chare Tree structure optionally distributed Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

11 Cello s timestepping Optional adaptive timestepping includes by Patch or Block local synchronization parallelism between levels t L=0 L=1 L=2 L= dt dt1 2 1 dt 2 1 x Quantized timesteps avoids sliver timesteps dt 0 but dt s smaller than optimal reduced numerical roundoff t x Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

12 Cello s improvements to scaling Reducing replicated AMR structure Fewer replicated patches patch-merging technique truncates full subtrees 2to3 reduction original patch-merged Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

13 Cello s improvements to scaling Reducing replicated AMR structure Fewer replicated patches patch-merging technique truncates full subtrees 2to3 reduction Smaller replicated patches Enzo: grid = 1544 bytes Cello: Node = 16 bytes 100 reduction! original Enzo grid patch-merged Cello Node Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

14 Cello s improvements to scaling Distributed AMR Structure Forest of Trees each assigned a process range simple indexing load balancing issues Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

15 Cello s improvements to scaling Distributed AMR Structure Forest of Trees each assigned a process range simple indexing load balancing issues Space-filling curve improved load balancing requires global scan greater surface area Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

16 Cello s improvements to scaling Dynamically load balancing AMR block data Dynamic Load Balancing Use Charm++ load balancing Space-filling curves equally distribute load maintain data locality no parent-child communication Use measured performance computation memory usage Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

17 Cello s improvements to scaling Parallel tasks Task definition Flexible choice of size / shape Reduced size variability constant grid Block sizes variable subcycling, # particles Enzo Cello Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

18 Cello s improvements to scaling Parallel tasks Task definition Flexible choice of size / shape Reduced size variability constant grid Block sizes variable subcycling, # particles Task scheduling Charm++: asynchronous, data-driven Blocks advance when ghosts refreshed t Enzo Cello L=0 L=1 L=2 L=3 x Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

19 Cello s implementation using Charm++ Charm++ program structure Main ChareA px() ChareB Main() py() pw() ChareC pz() pv() ACharm++ Program Charm++ program Charm++ objects are chares invoke entry methods communicate via messages Charm++ runtime system maps chares to processors schedules entry methods migrates chares to load balance Additional scalability features fault tol.: checkpoint/restart dynamic load balancing Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

20 Cello s implementation using Charm++ Charm++ collections of chares Chare Arrays distributed array of chares migratable elements flexible indexing Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

21 Cello s implementation using Charm++ Charm++ collections of chares Chare Arrays distributed array of chares migratable elements flexible indexing Chare Groups one chare per processor (non-migratable) Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

22 Cello s implementation using Charm++ Charm++ collections of chares Chare Arrays distributed array of chares migratable elements flexible indexing Chare Groups one chare per processor (non-migratable) Chare Nodegroups one chare per node (non-migratable) Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

23 Cello s implementation using Charm++ Three implementation strategies 1. Single chare array efficient: single access restricted Tree depth H Hierarchy T P B B P B B T P B B P B B T P B B P B B T P Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

24 Cello s implementation using Charm++ Three implementation strategies 1. Single chare array efficient: single access restricted Tree depth 2. Composite chare arrays tree of chare arrays less restricted Tree depth H F Hierarchy T P B B P B B T Forest P B B P Tree B B T P B B P B B T P Patch T P T T P P B B Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

25 Cello s implementation using Charm++ Three implementation strategies 1. Single chare array efficient: single access restricted Tree depth 2. Composite chare arrays tree of chare arrays less restricted Tree depth H F Hierarchy T P B B P B B T Forest P B B P Tree B B T P B B P B B T P Patch T P T T P P B B 3. Singleton chares no depth restrictions possible performance issues how to generate...? T P B B P B B T P B B P B B T P B B P B B T P Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

26 Recursively generated parallel data structures Generating a software network root seed 1 Start with single Seed conceptually complete p.d.s. with processor range 2 grow() spawns remote Seeds conceptually partitioned p.d.s. with processor subranges Seeds interlinked 3 Recurse to individual elements final Seeds complete p.d.s. scaffolding: previous Seeds scaffolding data structure topology Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

27 Recursively generated parallel data structures Using the software network Three types of Seed links parent: reductions neighbor: collaborations child: distributions Scalable links per node: O(1) generation: O(log N) Load balancing space-filling curves hierarchical Seed parent Seed neighbor Seeds child Seeds reduce link neighbor links broadcast links Usable in Cello grids / octrees easy SeedGrid, SeedTree, etc. Usable with MPI send encoded seed creation link: process rank + pointer Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

28 Summary Enzo Enzo-P / Cello Parallelization MPI Charm++ SAMR patch-based tree-based AMR structure replicated optionally distributed Remote patches 1544 bytes 16 bytes Timestepping level-adaptive block-adaptive Block sizes 1000 variation constant Task scheduling level-parallel dependency-driven Load balancing patch migration space-filling curves Data locality LB conflict no LB conflict NSF PHY , AST Extreme Scaling Workshop Enzo-P / Cello 16 July / 21

Enzo-P / Cello. Formation of the First Galaxies. San Diego Supercomputer Center. Department of Physics and Astronomy

Enzo-P / Cello. Formation of the First Galaxies. San Diego Supercomputer Center. Department of Physics and Astronomy Enzo-P / Cello Formation of the First Galaxies James Bordner 1 Michael L. Norman 1 Brian O Shea 2 1 University of California, San Diego San Diego Supercomputer Center 2 Michigan State University Department

More information

A Scalable Adaptive Mesh Refinement Framework For Parallel Astrophysics Applications

A Scalable Adaptive Mesh Refinement Framework For Parallel Astrophysics Applications A Scalable Adaptive Mesh Refinement Framework For Parallel Astrophysics Applications James Bordner, Michael L. Norman San Diego Supercomputer Center University of California, San Diego 15th SIAM Conference

More information

Formation of the First Galaxies Enzo-P / Cello Adaptive Mesh Refinement

Formation of the First Galaxies Enzo-P / Cello Adaptive Mesh Refinement Formation of the First Galaxies Enzo-P / ello Adaptive Mesh Refinement James Bordner, Michael L. Norman, Brian O Shea June 20, 2013 Abstract Enzo [8] is a mature and highly successful parallel adaptive

More information

Experiences with ENZO on the Intel R Many Integrated Core (Intel MIC) Architecture

Experiences with ENZO on the Intel R Many Integrated Core (Intel MIC) Architecture Experiences with ENZO on the Intel R Many Integrated Core (Intel MIC) Architecture 1 Introduction Robert Harkness National Institute for Computational Sciences Oak Ridge National Laboratory The National

More information

Adaptive Mesh Astrophysical Fluid Simulations on GPU. San Jose 10/2/2009 Peng Wang, NVIDIA

Adaptive Mesh Astrophysical Fluid Simulations on GPU. San Jose 10/2/2009 Peng Wang, NVIDIA Adaptive Mesh Astrophysical Fluid Simulations on GPU San Jose 10/2/2009 Peng Wang, NVIDIA Overview Astrophysical motivation & the Enzo code Finite volume method and adaptive mesh refinement (AMR) CUDA

More information

Future of Enzo. Michael L. Norman James Bordner LCA/SDSC/UCSD

Future of Enzo. Michael L. Norman James Bordner LCA/SDSC/UCSD Future of Enzo Michael L. Norman James Bordner LCA/SDSC/UCSD SDSC Resources Data to Discovery Host SDNAP San Diego network access point for multiple 10 Gbs WANs ESNet, NSF TeraGrid, CENIC, Internet2, StarTap

More information

Experiences with ENZO on the Intel Many Integrated Core Architecture

Experiences with ENZO on the Intel Many Integrated Core Architecture Experiences with ENZO on the Intel Many Integrated Core Architecture Dr. Robert Harkness National Institute for Computational Sciences April 10th, 2012 Overview ENZO applications at petascale ENZO and

More information

Dynamic Load Partitioning Strategies for Managing Data of Space and Time Heterogeneity in Parallel SAMR Applications

Dynamic Load Partitioning Strategies for Managing Data of Space and Time Heterogeneity in Parallel SAMR Applications Dynamic Load Partitioning Strategies for Managing Data of Space and Time Heterogeneity in Parallel SAMR Applications Xiaolin Li and Manish Parashar The Applied Software Systems Laboratory Department of

More information

Parallel Algorithms: Adaptive Mesh Refinement (AMR) method and its implementation

Parallel Algorithms: Adaptive Mesh Refinement (AMR) method and its implementation Parallel Algorithms: Adaptive Mesh Refinement (AMR) method and its implementation Massimiliano Guarrasi m.guarrasi@cineca.it Super Computing Applications and Innovation Department AMR - Introduction Solving

More information

Peta-Scale Simulations with the HPC Software Framework walberla:

Peta-Scale Simulations with the HPC Software Framework walberla: Peta-Scale Simulations with the HPC Software Framework walberla: Massively Parallel AMR for the Lattice Boltzmann Method SIAM PP 2016, Paris April 15, 2016 Florian Schornbaum, Christian Godenschwager,

More information

RAMSES on the GPU: An OpenACC-Based Approach

RAMSES on the GPU: An OpenACC-Based Approach RAMSES on the GPU: An OpenACC-Based Approach Claudio Gheller (ETHZ-CSCS) Giacomo Rosilho de Souza (EPFL Lausanne) Romain Teyssier (University of Zurich) Markus Wetzstein (ETHZ-CSCS) PRACE-2IP project EU

More information

HANDLING LOAD IMBALANCE IN DISTRIBUTED & SHARED MEMORY

HANDLING LOAD IMBALANCE IN DISTRIBUTED & SHARED MEMORY HANDLING LOAD IMBALANCE IN DISTRIBUTED & SHARED MEMORY Presenters: Harshitha Menon, Seonmyeong Bak PPL Group Phil Miller, Sam White, Nitin Bhat, Tom Quinn, Jim Phillips, Laxmikant Kale MOTIVATION INTEGRATED

More information

Preliminary Experiences with the Uintah Framework on on Intel Xeon Phi and Stampede

Preliminary Experiences with the Uintah Framework on on Intel Xeon Phi and Stampede Preliminary Experiences with the Uintah Framework on on Intel Xeon Phi and Stampede Qingyu Meng, Alan Humphrey, John Schmidt, Martin Berzins Thanks to: TACC Team for early access to Stampede J. Davison

More information

Computational Astrophysics 5 Higher-order and AMR schemes

Computational Astrophysics 5 Higher-order and AMR schemes Computational Astrophysics 5 Higher-order and AMR schemes Oscar Agertz Outline - The Godunov Method - Second-order scheme with MUSCL - Slope limiters and TVD schemes - Characteristics tracing and 2D slopes.

More information

Forest-of-octrees AMR: algorithms and interfaces

Forest-of-octrees AMR: algorithms and interfaces Forest-of-octrees AMR: algorithms and interfaces Carsten Burstedde joint work with Omar Ghattas, Tobin Isaac, Georg Stadler, Lucas C. Wilcox Institut für Numerische Simulation (INS) Rheinische Friedrich-Wilhelms-Universität

More information

Introducing Overdecomposition to Existing Applications: PlasComCM and AMPI

Introducing Overdecomposition to Existing Applications: PlasComCM and AMPI Introducing Overdecomposition to Existing Applications: PlasComCM and AMPI Sam White Parallel Programming Lab UIUC 1 Introduction How to enable Overdecomposition, Asynchrony, and Migratability in existing

More information

Welcome to the 2017 Charm++ Workshop!

Welcome to the 2017 Charm++ Workshop! Welcome to the 2017 Charm++ Workshop! Laxmikant (Sanjay) Kale http://charm.cs.illinois.edu Parallel Programming Laboratory Department of Computer Science University of Illinois at Urbana Champaign 2017

More information

CHRONO::HPC DISTRIBUTED MEMORY FLUID-SOLID INTERACTION SIMULATIONS. Felipe Gutierrez, Arman Pazouki, and Dan Negrut University of Wisconsin Madison

CHRONO::HPC DISTRIBUTED MEMORY FLUID-SOLID INTERACTION SIMULATIONS. Felipe Gutierrez, Arman Pazouki, and Dan Negrut University of Wisconsin Madison CHRONO::HPC DISTRIBUTED MEMORY FLUID-SOLID INTERACTION SIMULATIONS Felipe Gutierrez, Arman Pazouki, and Dan Negrut University of Wisconsin Madison Support: Rapid Innovation Fund, U.S. Army TARDEC ASME

More information

Towards a Reconfigurable HPC Component Model

Towards a Reconfigurable HPC Component Model C2S@EXA Meeting July 10, 2014 Towards a Reconfigurable HPC Component Model Vincent Lanore1, Christian Pérez2 1 ENS de Lyon, LIP 2 Inria, LIP Avalon team 1 Context 1/4 Adaptive Mesh Refinement 2 Context

More information

Scalable Dynamic Adaptive Simulations with ParFUM

Scalable Dynamic Adaptive Simulations with ParFUM Scalable Dynamic Adaptive Simulations with ParFUM Terry L. Wilmarth Center for Simulation of Advanced Rockets and Parallel Programming Laboratory University of Illinois at Urbana-Champaign The Big Picture

More information

Adaptive-Mesh-Refinement Hydrodynamic GPU Computation in Astrophysics

Adaptive-Mesh-Refinement Hydrodynamic GPU Computation in Astrophysics Adaptive-Mesh-Refinement Hydrodynamic GPU Computation in Astrophysics H. Y. Schive ( 薛熙于 ) Graduate Institute of Physics, National Taiwan University Leung Center for Cosmology and Particle Astrophysics

More information

Advanced Parallel Programming. Is there life beyond MPI?

Advanced Parallel Programming. Is there life beyond MPI? Advanced Parallel Programming Is there life beyond MPI? Outline MPI vs. High Level Languages Declarative Languages Map Reduce and Hadoop Shared Global Address Space Languages Charm++ ChaNGa ChaNGa on GPUs

More information

High performance computing and numerical modeling

High performance computing and numerical modeling High performance computing and numerical modeling Volker Springel Plan for my lectures Lecture 1: Collisional and collisionless N-body dynamics Lecture 2: Gravitational force calculation Lecture 3: Basic

More information

Topology and affinity aware hierarchical and distributed load-balancing in Charm++

Topology and affinity aware hierarchical and distributed load-balancing in Charm++ Topology and affinity aware hierarchical and distributed load-balancing in Charm++ Emmanuel Jeannot, Guillaume Mercier, François Tessier Inria - IPB - LaBRI - University of Bordeaux - Argonne National

More information

Programming Models for Supercomputing in the Era of Multicore

Programming Models for Supercomputing in the Era of Multicore Programming Models for Supercomputing in the Era of Multicore Marc Snir MULTI-CORE CHALLENGES 1 Moore s Law Reinterpreted Number of cores per chip doubles every two years, while clock speed decreases Need

More information

Load Balancing Techniques for Asynchronous Spacetime Discontinuous Galerkin Methods

Load Balancing Techniques for Asynchronous Spacetime Discontinuous Galerkin Methods Load Balancing Techniques for Asynchronous Spacetime Discontinuous Galerkin Methods Aaron K. Becker (abecker3@illinois.edu) Robert B. Haber Laxmikant V. Kalé University of Illinois, Urbana-Champaign Parallel

More information

Polygonization of Implicit Surfaces

Polygonization of Implicit Surfaces Polygonization of Implicit Surfaces Hongxin Zhang and Jieqing Feng 2007-01-11 State Key Lab of CAD&CG Zhejiang University Contents Polygonization of Implicit Surfaces Other Methods for Displaying Implicit

More information

Scientific Computing at Million-way Parallelism - Blue Gene/Q Early Science Program

Scientific Computing at Million-way Parallelism - Blue Gene/Q Early Science Program Scientific Computing at Million-way Parallelism - Blue Gene/Q Early Science Program Implementing Hybrid Parallelism in FLASH Christopher Daley 1 2 Vitali Morozov 1 Dongwook Lee 2 Anshu Dubey 1 2 Jonathon

More information

Development of a Computational Framework for Block-Based AMR Simulations

Development of a Computational Framework for Block-Based AMR Simulations Procedia Computer Science Volume 29, 2014, Pages 2351 2359 ICCS 2014. 14th International Conference on Computational Science Development of a Computational Framework for Block-Based AMR Simulations Hideyuki

More information

Generic finite element capabilities for forest-of-octrees AMR

Generic finite element capabilities for forest-of-octrees AMR Generic finite element capabilities for forest-of-octrees AMR Carsten Burstedde joint work with Omar Ghattas, Tobin Isaac Institut für Numerische Simulation (INS) Rheinische Friedrich-Wilhelms-Universität

More information

Parallelizing Adaptive Triangular Grids with Refinement Trees and Space Filling Curves

Parallelizing Adaptive Triangular Grids with Refinement Trees and Space Filling Curves Parallelizing Adaptive Triangular Grids with Refinement Trees and Space Filling Curves Daniel Butnaru butnaru@in.tum.de Advisor: Michael Bader bader@in.tum.de JASS 08 Computational Science and Engineering

More information

AREPO: a moving-mesh code for cosmological hydrodynamical simulations

AREPO: a moving-mesh code for cosmological hydrodynamical simulations AREPO: a moving-mesh code for cosmological hydrodynamical simulations E pur si muove: Galiliean-invariant cosmological hydrodynamical simulations on a moving mesh Springel, 2010 arxiv:0901.4107 Rubens

More information

simulation framework for piecewise regular grids

simulation framework for piecewise regular grids WALBERLA, an ultra-scalable multiphysics simulation framework for piecewise regular grids ParCo 2015, Edinburgh September 3rd, 2015 Christian Godenschwager, Florian Schornbaum, Martin Bauer, Harald Köstler

More information

Adaptive Mesh Refinement (AMR)

Adaptive Mesh Refinement (AMR) Adaptive Mesh Refinement (AMR) Carsten Burstedde Omar Ghattas, Georg Stadler, Lucas C. Wilcox Institute for Computational Engineering and Sciences (ICES) The University of Texas at Austin Collaboration

More information

Fault tolerant issues in large scale applications

Fault tolerant issues in large scale applications Fault tolerant issues in large scale applications Romain Teyssier George Lake, Ben Moore, Joachim Stadel and the other members of the project «Cosmology at the petascale» SPEEDUP 2010 1 Outline Computational

More information

Contents. Preface xvii Acknowledgments. CHAPTER 1 Introduction to Parallel Computing 1. CHAPTER 2 Parallel Programming Platforms 11

Contents. Preface xvii Acknowledgments. CHAPTER 1 Introduction to Parallel Computing 1. CHAPTER 2 Parallel Programming Platforms 11 Preface xvii Acknowledgments xix CHAPTER 1 Introduction to Parallel Computing 1 1.1 Motivating Parallelism 2 1.1.1 The Computational Power Argument from Transistors to FLOPS 2 1.1.2 The Memory/Disk Speed

More information

Towards Exascale Programming Models HPC Summit, Prague Erwin Laure, KTH

Towards Exascale Programming Models HPC Summit, Prague Erwin Laure, KTH Towards Exascale Programming Models HPC Summit, Prague Erwin Laure, KTH 1 Exascale Programming Models With the evolution of HPC architecture towards exascale, new approaches for programming these machines

More information

Visualization Challenges for Large Scale Astrophysical Simulation Data. Ultrascale Visualization Workshop

Visualization Challenges for Large Scale Astrophysical Simulation Data. Ultrascale Visualization Workshop Visualization Challenges for Large Scale Astrophysical Simulation Data Ralf Kähler (KIPAC/SLAC) Tom Abel (KIPAC/Stanford) Marcelo Alvarez (CITA) Oliver Hahn (Stanford) Hans-Christian Hege (ZIB) Ji-hoon

More information

Lagrangian methods and Smoothed Particle Hydrodynamics (SPH) Computation in Astrophysics Seminar (Spring 2006) L. J. Dursi

Lagrangian methods and Smoothed Particle Hydrodynamics (SPH) Computation in Astrophysics Seminar (Spring 2006) L. J. Dursi Lagrangian methods and Smoothed Particle Hydrodynamics (SPH) Eulerian Grid Methods The methods covered so far in this course use an Eulerian grid: Prescribed coordinates In `lab frame' Fluid elements flow

More information

PREPARING AN AMR LIBRARY FOR SUMMIT. Max Katz March 29, 2018

PREPARING AN AMR LIBRARY FOR SUMMIT. Max Katz March 29, 2018 PREPARING AN AMR LIBRARY FOR SUMMIT Max Katz March 29, 2018 CORAL: SIERRA AND SUMMIT NVIDIA Volta fueling supercomputers IBM Power 9 + NVIDIA Volta V100 Sierra (LLNL): 4 GPUs/node, ~4300 nodes Summit (ORNL):

More information

Large Scale Simulations of the Non-Thermal Universe

Large Scale Simulations of the Non-Thermal Universe Available on-line at www.prace-ri.eu Partnership for Advanced Computing in Europe Large Scale Simulations of the Non-Thermal Universe Claudio Gheller a,, Graziella Ferini a, Maciej Cytowski b, Franco Vazza

More information

C SCI 335 Software Analysis & Design III Lecture Notes Prof. Stewart Weiss Chapter 4: B Trees

C SCI 335 Software Analysis & Design III Lecture Notes Prof. Stewart Weiss Chapter 4: B Trees B-Trees AVL trees and other binary search trees are suitable for organizing data that is entirely contained within computer memory. When the amount of data is too large to fit entirely in memory, i.e.,

More information

Hierarchical Partitioning Techniques for Structured Adaptive Mesh Refinement Applications

Hierarchical Partitioning Techniques for Structured Adaptive Mesh Refinement Applications Hierarchical Partitioning Techniques for Structured Adaptive Mesh Refinement Applications Xiaolin Li (xlli@caip.rutgers.edu) and Manish Parashar (parashar@caip.rutgers.edu) Department of Electrical and

More information

Scalable Interaction with Parallel Applications

Scalable Interaction with Parallel Applications Scalable Interaction with Parallel Applications Filippo Gioachin Chee Wai Lee Laxmikant V. Kalé Department of Computer Science University of Illinois at Urbana-Champaign Outline Overview Case Studies Charm++

More information

Petascale Multiscale Simulations of Biomolecular Systems. John Grime Voth Group Argonne National Laboratory / University of Chicago

Petascale Multiscale Simulations of Biomolecular Systems. John Grime Voth Group Argonne National Laboratory / University of Chicago Petascale Multiscale Simulations of Biomolecular Systems John Grime Voth Group Argonne National Laboratory / University of Chicago About me Background: experimental guy in grad school (LSCM, drug delivery)

More information

Parallel Implementation of 3D FMA using MPI

Parallel Implementation of 3D FMA using MPI Parallel Implementation of 3D FMA using MPI Eric Jui-Lin Lu y and Daniel I. Okunbor z Computer Science Department University of Missouri - Rolla Rolla, MO 65401 Abstract The simulation of N-body system

More information

Kevin J. Barker. Scott Pakin and Darren J. Kerbyson

Kevin J. Barker. Scott Pakin and Darren J. Kerbyson Experiences in Performance Modeling: The Krak Hydrodynamics Application Kevin J. Barker Scott Pakin and Darren J. Kerbyson Performance and Architecture Laboratory (PAL) http://www.c3.lanl.gov/pal/ Computer,

More information

Fourteen years of Cactus Community

Fourteen years of Cactus Community Fourteen years of Cactus Community Frank Löffler Center for Computation and Technology Louisiana State University, Baton Rouge, LA September 6th 2012 Outline Motivation scenario from Astrophysics Cactus

More information

Parallel Programming Concepts. Parallel Algorithms. Peter Tröger

Parallel Programming Concepts. Parallel Algorithms. Peter Tröger Parallel Programming Concepts Parallel Algorithms Peter Tröger Sources: Ian Foster. Designing and Building Parallel Programs. Addison-Wesley. 1995. Mattson, Timothy G.; S, Beverly A.; ers,; Massingill,

More information

Shallow Water Simulations on Graphics Hardware

Shallow Water Simulations on Graphics Hardware Shallow Water Simulations on Graphics Hardware Ph.D. Thesis Presentation 2014-06-27 Martin Lilleeng Sætra Outline Introduction Parallel Computing and the GPU Simulating Shallow Water Flow Topics of Thesis

More information

Adaptive Blocks: A High Performance Data Structure

Adaptive Blocks: A High Performance Data Structure Adaptive Blocks: A High Performance Data Structure Quentin F. Stout Electrical Engineering and Computer Science Center for Parallel Computing University of Michigan Ann Arbor, MI 48109-2122 qstout@umich.edu

More information

GAMER : a GPU-accelerated Adaptive-MEsh-Refinement Code for Astrophysics GPU 與自適性網格於天文模擬之應用與效能

GAMER : a GPU-accelerated Adaptive-MEsh-Refinement Code for Astrophysics GPU 與自適性網格於天文模擬之應用與效能 GAMER : a GPU-accelerated Adaptive-MEsh-Refinement Code for Astrophysics GPU 與自適性網格於天文模擬之應用與效能 Hsi-Yu Schive ( 薛熙于 ), Tzihong Chiueh ( 闕志鴻 ), Yu-Chih Tsai ( 蔡御之 ), Ui-Han Zhang ( 張瑋瀚 ) Graduate Institute

More information

B-Trees. Version of October 2, B-Trees Version of October 2, / 22

B-Trees. Version of October 2, B-Trees Version of October 2, / 22 B-Trees Version of October 2, 2014 B-Trees Version of October 2, 2014 1 / 22 Motivation An AVL tree can be an excellent data structure for implementing dictionary search, insertion and deletion Each operation

More information

ALE and AMR Mesh Refinement Techniques for Multi-material Hydrodynamics Problems

ALE and AMR Mesh Refinement Techniques for Multi-material Hydrodynamics Problems ALE and AMR Mesh Refinement Techniques for Multi-material Hydrodynamics Problems A. J. Barlow, AWE. ICFD Workshop on Mesh Refinement Techniques 7th December 2005 Acknowledgements Thanks to Chris Powell,

More information

Scalable Software Components for Ultrascale Visualization Applications

Scalable Software Components for Ultrascale Visualization Applications Scalable Software Components for Ultrascale Visualization Applications Wes Kendall, Tom Peterka, Jian Huang SC Ultrascale Visualization Workshop 2010 11-15-2010 Primary Collaborators Jian Huang Tom Peterka

More information

The ITAPS Mesh Interface

The ITAPS Mesh Interface The ITAPS Mesh Interface Carl Ollivier-Gooch Advanced Numerical Simulation Laboratory, University of British Columbia Needs and Challenges for Unstructured Mesh Usage Application PDE Discretization Mesh

More information

Introduction to parallel Computing

Introduction to parallel Computing Introduction to parallel Computing VI-SEEM Training Paschalis Paschalis Korosoglou Korosoglou (pkoro@.gr) (pkoro@.gr) Outline Serial vs Parallel programming Hardware trends Why HPC matters HPC Concepts

More information

Communication and Topology-aware Load Balancing in Charm++ with TreeMatch

Communication and Topology-aware Load Balancing in Charm++ with TreeMatch Communication and Topology-aware Load Balancing in Charm++ with TreeMatch Joint lab 10th workshop (IEEE Cluster 2013, Indianapolis, IN) Emmanuel Jeannot Esteban Meneses-Rojas Guillaume Mercier François

More information

Computational Fluid Dynamics with the Lattice Boltzmann Method KTH SCI, Stockholm

Computational Fluid Dynamics with the Lattice Boltzmann Method KTH SCI, Stockholm Computational Fluid Dynamics with the Lattice Boltzmann Method KTH SCI, Stockholm March 17 March 21, 2014 Florian Schornbaum, Martin Bauer, Simon Bogner Chair for System Simulation Friedrich-Alexander-Universität

More information

A Parallel-Object Programming Model for PetaFLOPS Machines and BlueGene/Cyclops Gengbin Zheng, Arun Singla, Joshua Unger, Laxmikant Kalé

A Parallel-Object Programming Model for PetaFLOPS Machines and BlueGene/Cyclops Gengbin Zheng, Arun Singla, Joshua Unger, Laxmikant Kalé A Parallel-Object Programming Model for PetaFLOPS Machines and BlueGene/Cyclops Gengbin Zheng, Arun Singla, Joshua Unger, Laxmikant Kalé Parallel Programming Laboratory Department of Computer Science University

More information

EE382N (20): Computer Architecture - Parallelism and Locality Lecture 13 Parallelism in Software IV

EE382N (20): Computer Architecture - Parallelism and Locality Lecture 13 Parallelism in Software IV EE382 (20): Computer Architecture - Parallelism and Locality Lecture 13 Parallelism in Software IV Mattan Erez The University of Texas at Austin EE382: Parallelilsm and Locality (c) Rodric Rabbah, Mattan

More information

Performance Optimization of a Massively Parallel Phase-Field Method Using the HPC Framework walberla

Performance Optimization of a Massively Parallel Phase-Field Method Using the HPC Framework walberla Performance Optimization of a Massively Parallel Phase-Field Method Using the HPC Framework walberla SIAM PP 2016, April 13 th 2016 Martin Bauer, Florian Schornbaum, Christian Godenschwager, Johannes Hötzer,

More information

What is DARMA? DARMA is a C++ abstraction layer for asynchronous many-task (AMT) runtimes.

What is DARMA? DARMA is a C++ abstraction layer for asynchronous many-task (AMT) runtimes. DARMA Janine C. Bennett, Jonathan Lifflander, David S. Hollman, Jeremiah Wilke, Hemanth Kolla, Aram Markosyan, Nicole Slattengren, Robert L. Clay (PM) PSAAP-WEST February 22, 2017 Sandia National Laboratories

More information

SCALASCA parallel performance analyses of SPEC MPI2007 applications

SCALASCA parallel performance analyses of SPEC MPI2007 applications Mitglied der Helmholtz-Gemeinschaft SCALASCA parallel performance analyses of SPEC MPI2007 applications 2008-05-22 Zoltán Szebenyi Jülich Supercomputing Centre, Forschungszentrum Jülich Aachen Institute

More information

CSE 530A. B+ Trees. Washington University Fall 2013

CSE 530A. B+ Trees. Washington University Fall 2013 CSE 530A B+ Trees Washington University Fall 2013 B Trees A B tree is an ordered (non-binary) tree where the internal nodes can have a varying number of child nodes (within some range) B Trees When a key

More information

Enabling scalable parallel implementations of structured adaptive mesh refinement applications

Enabling scalable parallel implementations of structured adaptive mesh refinement applications J Supercomput (2007) 39: 177 203 DOI 10.1007/s11227-007-0110-z Enabling scalable parallel implementations of structured adaptive mesh refinement applications Sumir Chandra Xiaolin Li Taher Saif Manish

More information

Ray Tracing. Cornell CS4620/5620 Fall 2012 Lecture Kavita Bala 1 (with previous instructors James/Marschner)

Ray Tracing. Cornell CS4620/5620 Fall 2012 Lecture Kavita Bala 1 (with previous instructors James/Marschner) CS4620/5620: Lecture 37 Ray Tracing 1 Announcements Review session Tuesday 7-9, Phillips 101 Posted notes on slerp and perspective-correct texturing Prelim on Thu in B17 at 7:30pm 2 Basic ray tracing Basic

More information

Solid Modeling. Thomas Funkhouser Princeton University C0S 426, Fall Represent solid interiors of objects

Solid Modeling. Thomas Funkhouser Princeton University C0S 426, Fall Represent solid interiors of objects Solid Modeling Thomas Funkhouser Princeton University C0S 426, Fall 2000 Solid Modeling Represent solid interiors of objects Surface may not be described explicitly Visible Human (National Library of Medicine)

More information

VTK-m: Uniting GPU Acceleration Successes. Robert Maynard Kitware Inc.

VTK-m: Uniting GPU Acceleration Successes. Robert Maynard Kitware Inc. VTK-m: Uniting GPU Acceleration Successes Robert Maynard Kitware Inc. VTK-m Project Supercomputer Hardware Advances Everyday More and more parallelism High-Level Parallelism The Free Lunch Is Over (Herb

More information

Adaptive Mesh Refinement

Adaptive Mesh Refinement Aleander Knebe, Universidad Autonoma de Madrid Adaptive Mesh Refinement AMR codes Poisson s equation ΔΦ( ) = 4πGρ( ) Poisson s equation F ( ) = m Φ( ) ΔΦ( ) = 4πGρ( ) particle approach F ( Gm i ) = i m

More information

Data Structures and Algorithms

Data Structures and Algorithms Data Structures and Algorithms CS245-2008S-19 B-Trees David Galles Department of Computer Science University of San Francisco 19-0: Indexing Operations: Add an element Remove an element Find an element,

More information

Scalability of Uintah Past Present and Future

Scalability of Uintah Past Present and Future DOE for funding the CSAFE project (97-10), DOE NETL, DOE NNSA NSF for funding via SDCI and PetaApps, INCITE, XSEDE Scalability of Uintah Past Present and Future Martin Berzins Qingyu Meng John Schmidt,

More information

Cosmology Simulations with Enzo

Cosmology Simulations with Enzo Cosmology Simulations with Enzo John Wise (Georgia Tech) Enzo Workshop 17 Oct 2013 Outline Introduction to unigrid cosmology simulations Introduction to nested grid cosmology simulations Using different

More information

Spatial Data Structures

Spatial Data Structures Spatial Data Structures Hierarchical Bounding Volumes Regular Grids Octrees BSP Trees Constructive Solid Geometry (CSG) [Angel 9.10] Outline Ray tracing review what rays matter? Ray tracing speedup faster

More information

A CASE STUDY OF COMMUNICATION OPTIMIZATIONS ON 3D MESH INTERCONNECTS

A CASE STUDY OF COMMUNICATION OPTIMIZATIONS ON 3D MESH INTERCONNECTS A CASE STUDY OF COMMUNICATION OPTIMIZATIONS ON 3D MESH INTERCONNECTS Abhinav Bhatele, Eric Bohm, Laxmikant V. Kale Parallel Programming Laboratory Euro-Par 2009 University of Illinois at Urbana-Champaign

More information

Scalasca performance properties The metrics tour

Scalasca performance properties The metrics tour Scalasca performance properties The metrics tour Markus Geimer m.geimer@fz-juelich.de Scalasca analysis result Generic metrics Generic metrics Time Total CPU allocation time Execution Overhead Visits Hardware

More information

ELASTIC: Dynamic Tuning for Large-Scale Parallel Applications

ELASTIC: Dynamic Tuning for Large-Scale Parallel Applications Workshop on Extreme-Scale Programming Tools 18th November 2013 Supercomputing 2013 ELASTIC: Dynamic Tuning for Large-Scale Parallel Applications Toni Espinosa Andrea Martínez, Anna Sikora, Eduardo César

More information

Stitched Together: Transitioning CMS to a Hierarchical Threaded Framework

Stitched Together: Transitioning CMS to a Hierarchical Threaded Framework Stitched Together: Transitioning CMS to a Hierarchical Threaded Framework CD Jones and E Sexton-Kennedy Fermilab, P.O.Box 500, Batavia, IL 60510-5011, USA E-mail: cdj@fnal.gov, sexton@fnal.gov Abstract.

More information

CS 475: Parallel Programming Introduction

CS 475: Parallel Programming Introduction CS 475: Parallel Programming Introduction Wim Bohm, Sanjay Rajopadhye Colorado State University Fall 2014 Course Organization n Let s make a tour of the course website. n Main pages Home, front page. Syllabus.

More information

COPYRIGHTED MATERIAL. Introduction: Enabling Large-Scale Computational Science Motivations, Requirements, and Challenges.

COPYRIGHTED MATERIAL. Introduction: Enabling Large-Scale Computational Science Motivations, Requirements, and Challenges. Chapter 1 Introduction: Enabling Large-Scale Computational Science Motivations, Requirements, and Challenges Manish Parashar and Xiaolin Li 1.1 MOTIVATION The exponential growth in computing, networking,

More information

Dynamic Load Balancing for Weather Models via AMPI

Dynamic Load Balancing for Weather Models via AMPI Dynamic Load Balancing for Eduardo R. Rodrigues IBM Research Brazil edrodri@br.ibm.com Celso L. Mendes University of Illinois USA cmendes@ncsa.illinois.edu Laxmikant Kale University of Illinois USA kale@cs.illinois.edu

More information

Scalable Shared Memory Programing

Scalable Shared Memory Programing Scalable Shared Memory Programing Marc Snir www.parallel.illinois.edu What is (my definition of) Shared Memory Global name space (global references) Implicit data movement Caching: User gets good memory

More information

Center Extreme Scale CS Research

Center Extreme Scale CS Research Center Extreme Scale CS Research Center for Compressible Multiphase Turbulence University of Florida Sanjay Ranka Herman Lam Outline 10 6 10 7 10 8 10 9 cores Parallelization and UQ of Rocfun and CMT-Nek

More information

An Efficient CUDA Implementation of a Tree-Based N-Body Algorithm. Martin Burtscher Department of Computer Science Texas State University-San Marcos

An Efficient CUDA Implementation of a Tree-Based N-Body Algorithm. Martin Burtscher Department of Computer Science Texas State University-San Marcos An Efficient CUDA Implementation of a Tree-Based N-Body Algorithm Martin Burtscher Department of Computer Science Texas State University-San Marcos Mapping Regular Code to GPUs Regular codes Operate on

More information

I/O Analysis and Optimization for an AMR Cosmology Application

I/O Analysis and Optimization for an AMR Cosmology Application I/O Analysis and Optimization for an AMR Cosmology Application Jianwei Li Wei-keng Liao Alok Choudhary Valerie Taylor ECE Department, Northwestern University {jianwei, wkliao, choudhar, taylor}@ece.northwestern.edu

More information

Using Charm++ to Support Multiscale Multiphysics

Using Charm++ to Support Multiscale Multiphysics LA-UR-17-23218 Using Charm++ to Support Multiscale Multiphysics On the Trinity Supercomputer Robert Pavel, Christoph Junghans, Susan M. Mniszewski, Timothy C. Germann April, 18 th 2017 Operated by Los

More information

Large Data Visualization

Large Data Visualization Large Data Visualization Seven Lectures 1. Overview (this one) 2. Scalable parallel rendering algorithms 3. Particle data visualization 4. Vector field visualization 5. Visual analytics techniques for

More information

Physical Level of Databases: B+-Trees

Physical Level of Databases: B+-Trees Physical Level of Databases: B+-Trees Adnan YAZICI Computer Engineering Department METU (Fall 2005) 1 B + -Tree Index Files l Disadvantage of indexed-sequential files: performance degrades as file grows,

More information

The Case for Collective Pattern Specification

The Case for Collective Pattern Specification The Case for Collective Pattern Specification Torsten Hoefler, Jeremiah Willcock, ArunChauhan, and Andrew Lumsdaine Advances in Message Passing, Toronto, ON, June 2010 Motivation and Main Theses Message

More information

IOS: A Middleware for Decentralized Distributed Computing

IOS: A Middleware for Decentralized Distributed Computing IOS: A Middleware for Decentralized Distributed Computing Boleslaw Szymanski Kaoutar El Maghraoui, Carlos Varela Department of Computer Science Rensselaer Polytechnic Institute http://www.cs.rpi.edu/wwc

More information

Scalable Fault Tolerance Schemes using Adaptive Runtime Support

Scalable Fault Tolerance Schemes using Adaptive Runtime Support Scalable Fault Tolerance Schemes using Adaptive Runtime Support Laxmikant (Sanjay) Kale http://charm.cs.uiuc.edu Parallel Programming Laboratory Department of Computer Science University of Illinois at

More information

Spatial Data Structures

Spatial Data Structures 15-462 Computer Graphics I Lecture 17 Spatial Data Structures Hierarchical Bounding Volumes Regular Grids Octrees BSP Trees Constructive Solid Geometry (CSG) April 1, 2003 [Angel 9.10] Frank Pfenning Carnegie

More information

Module 4: Index Structures Lecture 13: Index structure. The Lecture Contains: Index structure. Binary search tree (BST) B-tree. B+-tree.

Module 4: Index Structures Lecture 13: Index structure. The Lecture Contains: Index structure. Binary search tree (BST) B-tree. B+-tree. The Lecture Contains: Index structure Binary search tree (BST) B-tree B+-tree Order file:///c /Documents%20and%20Settings/iitkrana1/My%20Documents/Google%20Talk%20Received%20Files/ist_data/lecture13/13_1.htm[6/14/2012

More information

CS350: Data Structures B-Trees

CS350: Data Structures B-Trees B-Trees James Moscola Department of Engineering & Computer Science York College of Pennsylvania James Moscola Introduction All of the data structures that we ve looked at thus far have been memory-based

More information

Interactive Isosurface Ray Tracing of Large Octree Volumes

Interactive Isosurface Ray Tracing of Large Octree Volumes Interactive Isosurface Ray Tracing of Large Octree Volumes Aaron Knoll, Ingo Wald, Steven Parker, and Charles Hansen Scientific Computing and Imaging Institute University of Utah 2006 IEEE Symposium on

More information

15-853:Algorithms in the Real World. Outline. Parallelism: Lecture 1 Nested parallelism Cost model Parallel techniques and algorithms

15-853:Algorithms in the Real World. Outline. Parallelism: Lecture 1 Nested parallelism Cost model Parallel techniques and algorithms :Algorithms in the Real World Parallelism: Lecture 1 Nested parallelism Cost model Parallel techniques and algorithms Page1 Andrew Chien, 2008 2 Outline Concurrency vs. Parallelism Quicksort example Nested

More information

Spatial Data Structures

Spatial Data Structures 15-462 Computer Graphics I Lecture 17 Spatial Data Structures Hierarchical Bounding Volumes Regular Grids Octrees BSP Trees Constructive Solid Geometry (CSG) March 28, 2002 [Angel 8.9] Frank Pfenning Carnegie

More information

Application Example Running on Top of GPI-Space Integrating D/C

Application Example Running on Top of GPI-Space Integrating D/C Application Example Running on Top of GPI-Space Integrating D/C Tiberiu Rotaru Fraunhofer ITWM This project is funded from the European Union s Horizon 2020 Research and Innovation programme under Grant

More information

Trees. Q: Why study trees? A: Many advance ADTs are implemented using tree-based data structures.

Trees. Q: Why study trees? A: Many advance ADTs are implemented using tree-based data structures. Trees Q: Why study trees? : Many advance DTs are implemented using tree-based data structures. Recursive Definition of (Rooted) Tree: Let T be a set with n 0 elements. (i) If n = 0, T is an empty tree,

More information

Solving Petascale Turbulent Combustion Problems with the Uintah Software

Solving Petascale Turbulent Combustion Problems with the Uintah Software Solving Petascale Turbulent Combustion Problems with the Uintah Software Martin Berzins DOE NNSA PSAAP2 Center Thanks to DOE ASCI (97-10), NSF, DOE NETL+NNSA, NSF, INCITE, XSEDE, ALCC, ORNL, ALCF for funding

More information