What is Cactus? Cactus is a framework for developing portable, modular applications

Size: px
Start display at page:

Download "What is Cactus? Cactus is a framework for developing portable, modular applications"

Transcription

1 What is Cactus? Cactus is a framework for developing portable, modular applications

2 What is Cactus? Cactus is a framework for developing portable, modular applications focusing, although not exclusively, on high-performance simulation codes

3 What is Cactus? Cactus is a framework for developing portable, modular applications focusing, although not exclusively, on high-performance simulation codes designed to allow experts in different fields to develop modules based upon their experience and to use modules developed by experts in other fields with minimal knowledge of the internals or operation of the other modules

4 Cactus Goals Portable Different development machines Different production machines

5 Cactus Goals Portable Different development machines Different production machines Modular Standard interfaces for module interaction for easier code interaction, writing and debugging Interchangeable modules with same functionality

6 Cactus Goals Portable Different development machines Different production machines Modular Standard interfaces for module interaction for easier code interaction, writing and debugging Interchangeable modules with same functionality Easy to use Good documentation Try to let users program the way they are used to Support all major (HPC) programming languages

7 Philosophy Open code base to encourage community contributions

8 Philosophy Open code base to encourage community contributions Strict quality control for base framework

9 Philosophy Open code base to encourage community contributions Strict quality control for base framework Development always driven by real user requirements

10 Philosophy Open code base to encourage community contributions Strict quality control for base framework Development always driven by real user requirements Support and develop for a wide range of application domains

11 What is Cactus for? Assume: Computational problem Too large for single machine Distributed development Multiple programming languages

12 What is Cactus for? Assume: Computational problem Too large for single machine OpenMP MPI Distributed development Multiple programming languages

13 What is Cactus for? Assume: Computational problem Too large for single machine OpenMP MPI Distributed development Modularize Problem Versioning system(s) Multiple programming languages

14 What is Cactus for? Assume: Computational problem Too large for single machine OpenMP MPI Distributed development Modularize Problem Versioning system(s) Multiple programming languages Modularize Problem Interfaces for inter-language communication

15 What is Cactus for? Assume: Computational problem Too large for single machine OpenMP MPI Distributed development Modularize Problem Versioning system(s) Multiple programming languages Modularize Problem Interfaces for inter-language communication

16 Modularization Motivation Simulations are complex, may contain several models at once. Example: Long Gamma Ray Burst: General Relativity (black hole) Relativistic Hydrodynamics (star) Microphysics, equation of state (shock wave) Neutrino radiation (cooling, heating) Magnetic Fields (Jet formation) Photon radiation (afterglow)

17 Modularization Motivation Added Problems: Example: Einstein Toolkit (not untypical) Code 12+ years old Grad students leave after 3 productive years Most original authors not available anymore Developers distributed over many places in several continents Most physicists are not good programmers

18 Component Architecture Split program into independent components (as much as possible) Framework provides glue between these Components are developed independently by small groups End user assembles all code: no central control, no authorative version Framework itself does no real work Components don t interact directly with each other

19 Cactus History Direct descendant of many years of code development in Ed Seidel s group of researchers at NCSA 1995, Paul Walker, Joan Masso, Ed Seidel, and John Shalf: Cactus 1.0 Originally for numerical relativity Over the years generalized for use by scientists in other domains

20 Current Users and Developers

21 Covers

22 Cactus Funding Organizations: Max-Planck-Gesellschaft Center for Computation & Technology at LSU National Center for Supercomputing Applications Lawrence Berkeley National Laboratory Washington University University of Tübingen Grants: NSF (PHY , , , , , PIF , , ) Europ. Commission (HPRN-CT , IST ) DFN-Verein (TK 6-2-AN 200) DFG (TiKSL) ONR (COMI) DOE/BOR (OE DE-FG02-04ER46136, BOR DOE/LEQSF)

23 Cactus Awards IEEE SCALE09 Challenge Winner 2009 IEEE Sidney Fernback Award 2006 High-Performance Bandwidth Challenge SC2002 High-Performance Computing Challenge SC2002 Gordon Bell Prize for Supercomputing SC2001 HPC Most Stellar Challenge Award SC1998 Heinz Billing Prize for Scientific Computing 1998

24 The Flesh The flesh is the central component of Cactus. It interfaces with modular components called thorns. The flesh provides: Variables & Data Types Parameters Functions for: Parallelisation Input/Output Coordinates Reduction Interpolation Information Staggering Indexing Ghostzones

25 Thorns Some thorns provide additional functionality, while others serve as applications. Thorns are grouped into arrangements which supply some common functionality. Example thorns: CactusIO CactusIOJpeg CactusConnect HTTPD PUGH PUGH input and output operations JPEG image data compression and writing operations networking starts the HTTP daemon for remote connections unigrid driver + tools; reductions and interpolations unigrid driver handles grid scalars, arrays and functions

26 Application Toolkits The Einstein Toolkit is a collection of arrangements for computational relativity. The toolkit includes a vacuum spacetime solver (McLachlan), a relativistic hydrodynamics solver, along with thorns for initial data, analysis and computational infrastructure. The Cactus Computational Toolkit is a collection of arrangements that provides general computational infrastructure.

27 Typical list of component tasks Evolution systems Boundary conditions Initial conditions Time stepping method Finite Difference methods Simulation grid (distributed arrays) I/O (more on next slide) Simulation domain specification Termination condition Twitter client

28 I/O Capabilities Usual I/O and checkpointing in different formats: Screen output ASCII file output HDF5 file in-/output Online Jpeg rendering Online VisIt visualization

29 More Capabilities: Grids, Boundaries, Symmetries Grids Only structured meshes (at the moment) Unigrid (PUGH) Adaptive Mesh Refinement (Carpet) Boundaries / Symmetries Periodic Static Mirror symmetries Rotational symmetries Problemspecific boundaries

30 The Cactus Computational Toolkit Core modules (thorns) providing many basic utilities: I/O methods Boundary conditions Parallel unigrid driver Reduction and Interpolation operators Interface to external elliptic solvers Web-based interaction and monitoring interface Simple example thorns (wavetoy)

31 Many arrangements with many modules... CactusBase CactusBench CactusConnect CactusElliptic CactusExamples CactusExternal CactusIO CactusNumerical CactusPUGH CactusPUGHIO CactusTest CactusUtils CactusWave Basic utility and interface thorns Benchmark utility thorns Network utility thorns Elliptic solvers / interface thorns Example thorns External library interface thorns General I/O thorns General numerical methods Cactus Unigrid Driver thorn I/O thorns specifix for PUGH driver Thorns for Cactus testing Misc. utility thorns Wave example thorns

Introduction to the Cactus Framework

Introduction to the Cactus Framework Introduction to the Cactus Framework The Cactus team Oct 25 2011 1 Introduction Outline 2 The Cactus Computational Toolkit Overview Capabilities 3 Visualization 4 Cactus Structure Overview The Flesh The

More information

Fourteen years of Cactus Community

Fourteen years of Cactus Community Fourteen years of Cactus Community Frank Löffler Center for Computation and Technology Louisiana State University, Baton Rouge, LA September 6th 2012 Outline Motivation scenario from Astrophysics Cactus

More information

High Performance and Grid Computing Applications with the Cactus Framework. HPCC Program Grand Challenges (1995)

High Performance and Grid Computing Applications with the Cactus Framework. HPCC Program Grand Challenges (1995) High Performance and Grid Computing Applications with the Cactus Framework Gabrielle Allen Department of Computer Science Center for Computation & Technology Louisiana State University HPCC Program Grand

More information

Cactus Tutorial. Introduction to Cactus. Yaakoub El Khamra. Cactus Developer, Frameworks Group CCT 27 March, 2007

Cactus Tutorial. Introduction to Cactus. Yaakoub El Khamra. Cactus Developer, Frameworks Group CCT 27 March, 2007 Cactus Tutorial Introduction to Cactus Yaakoub El Khamra Cactus Developer, Frameworks Group CCT 27 March, 2007 Agenda Introduction to Cactus What is Cactus Flesh and thorns Cactus Computational Toolkit

More information

Cactus Framework: Scaling and Lessons Learnt

Cactus Framework: Scaling and Lessons Learnt Cactus Framework: Scaling and Lessons Learnt Gabrielle Allen, Erik Schnetter, Jian Tao Center for Computation & Technology Departments of Computer Science & Physics Louisiana State University Also: Christian

More information

The Cactus Framework. Erik Schnetter September 2006

The Cactus Framework. Erik Schnetter September 2006 The Cactus Framework Erik Schnetter September 2006 Outline History The Cactus User Community Cactus Usage Patterns Bird s eye view Cactus is a freely available, portable, and manageable environment for

More information

Parallelism. Wolfgang Kastaun. May 9, 2008

Parallelism. Wolfgang Kastaun. May 9, 2008 Parallelism Wolfgang Kastaun May 9, 2008 Outline Parallel computing Frameworks MPI and the batch system Running MPI code at TAT The CACTUS framework Overview Mesh refinement Writing Cactus modules Links

More information

Cactus: Current Status and Future Plans

Cactus: Current Status and Future Plans Cactus: Current Status and Future Plans Tom Goodale goodale@aei-potsdam.mpg.de Penn State Numrel Lunch February 27 h 2003 What Is Cactus Cactus is a framework for developing portable, modular applications,

More information

Integration of Trilinos Into The Cactus Code Framework

Integration of Trilinos Into The Cactus Code Framework Integration of Trilinos Into The Cactus Code Framework Josh Abadie Research programmer Center for Computation & Technology Louisiana State University Summary Motivation Objectives The Cactus Code Trilinos

More information

From Physics Model to Results: An Optimizing Framework for Cross-Architecture Code Generation

From Physics Model to Results: An Optimizing Framework for Cross-Architecture Code Generation From Physics Model to Results: An Optimizing Framework for Cross-Architecture Code Generation Erik Schnetter, Perimeter Institute with M. Blazewicz, I. Hinder, D. Koppelman, S. Brandt, M. Ciznicki, M.

More information

Introduction to the Einstein Toolkit Details

Introduction to the Einstein Toolkit Details Introduction to the Einstein Toolkit Details Roland Haas TAPIR, Caltech July 26 th 2013 Einstein Toolkit main components Toolkit is based on three core components Cactus GetComponents provides the numerical

More information

The Cactus Framework and Toolkit: Design and Applications

The Cactus Framework and Toolkit: Design and Applications The Cactus Framework and Toolkit: Design and Applications Tom Goodale 1, Gabrielle Allen 1, Gerd Lanfermann 1, Joan Massó 2, Thomas Radke 1, Edward Seidel 1, and John Shalf 3 1 Max-Planck-Institut für

More information

Datura The new HPC-Plant at Albert Einstein Institute

Datura The new HPC-Plant at Albert Einstein Institute Datura The new HPC-Plant at Albert Einstein Institute Nico Budewitz Max Planck Institue for Gravitational Physics, Germany Cluster Day, 2011 Outline 1 History HPC-Plants at AEI -2009 Peyote, Lagavulin,

More information

Grid Computing in Numerical Relativity and Astrophysics

Grid Computing in Numerical Relativity and Astrophysics Grid Computing in Numerical Relativity and Astrophysics Gabrielle Allen: gallen@cct.lsu.edu Depts Computer Science & Physics Center for Computation & Technology (CCT) Louisiana State University Challenge

More information

Adaptive Mesh Astrophysical Fluid Simulations on GPU. San Jose 10/2/2009 Peng Wang, NVIDIA

Adaptive Mesh Astrophysical Fluid Simulations on GPU. San Jose 10/2/2009 Peng Wang, NVIDIA Adaptive Mesh Astrophysical Fluid Simulations on GPU San Jose 10/2/2009 Peng Wang, NVIDIA Overview Astrophysical motivation & the Enzo code Finite volume method and adaptive mesh refinement (AMR) CUDA

More information

Performance Profiling with Cactus Benchmarks

Performance Profiling with Cactus Benchmarks Performance Profiling with Cactus Benchmarks Sasanka Madiraju April 6th, 2006 System Science Master s Project Report Department of Computer Science Louisiana State University Acknowledgment This project

More information

PHYSICS TOMORROW. The Grid from an Application Viewpoint. Ed Seidel

PHYSICS TOMORROW. The Grid from an Application Viewpoint. Ed Seidel The Grid from an Application Viewpoint PHYSICS TOMORROW TODAY Ed Seidel Max-Planck-Institut für Gravitationsphysik (Albert Einstein Institute) + Lots of colleagues eseidel@aei.mpg.de Co-Chair, GGF Applications

More information

Parallel Programming with MPI on Clusters

Parallel Programming with MPI on Clusters Parallel Programming with MPI on Clusters Rusty Lusk Mathematics and Computer Science Division Argonne National Laboratory (The rest of our group: Bill Gropp, Rob Ross, David Ashton, Brian Toonen, Anthony

More information

Erik Schnetter Rochester, August Friday, August 27, 2010

Erik Schnetter Rochester, August Friday, August 27, 2010 Erik Schnetter Rochester, August 2010 1 Goal: have state-of-the-art set of tools for NR available as open source Organised by Einstein Consortium, open to everyone See http://einsteintoolkit.org 2 Guiding

More information

A Scalable Adaptive Mesh Refinement Framework For Parallel Astrophysics Applications

A Scalable Adaptive Mesh Refinement Framework For Parallel Astrophysics Applications A Scalable Adaptive Mesh Refinement Framework For Parallel Astrophysics Applications James Bordner, Michael L. Norman San Diego Supercomputer Center University of California, San Diego 15th SIAM Conference

More information

Component Specification in the Cactus Framework: The Cactus Configuration Language

Component Specification in the Cactus Framework: The Cactus Configuration Language Component Specification in the Cactus Framework: The Cactus Configuration Language Gabrielle Allen Center for Computation & Technology Department of Computer Science Louisiana State University Baton Rouge,

More information

The Cactus Framework: Design, Applications and Future Directions. Cactus Code

The Cactus Framework: Design, Applications and Future Directions. Cactus Code The Cactus Framework: Design, Applications and Future Directions Gabrielle Allen gallen@cct.lsu.edu Center for Computation & Technology Departments of Computer Science & Physics Louisiana State University

More information

Adaptive Mesh Refinement in Titanium

Adaptive Mesh Refinement in Titanium Adaptive Mesh Refinement in Titanium http://seesar.lbl.gov/anag Lawrence Berkeley National Laboratory April 7, 2005 19 th IPDPS, April 7, 2005 1 Overview Motivations: Build the infrastructure in Titanium

More information

Application / User Viewpoint

Application / User Viewpoint SC 07 Reno, Nevada November 15, 2007 Fortran@50 Application / User Viewpoint Henry Tufo Computer Science Section Head Associate Professor and Director, Computational Science Center Department of Computer

More information

Performance and Optimization Abstractions for Large Scale Heterogeneous Systems in the Cactus/Chemora Framework

Performance and Optimization Abstractions for Large Scale Heterogeneous Systems in the Cactus/Chemora Framework Performance and Optimization Abstractions for Large Scale Heterogeneous Systems in the Cactus/Chemora Framework Erik Schne+er Perimeter Ins1tute for Theore1cal Physics XSCALE 2013, Boulder, CO, 2013-08-

More information

Integrating Web 2.0 Technologies with Scientific Simulation Codes for Real-Time Collaboration

Integrating Web 2.0 Technologies with Scientific Simulation Codes for Real-Time Collaboration Integrating Web 2.0 Technologies with Scientific Simulation Codes for Real-Time Collaboration Gabrielle Allen 1,2, Frank Löffler 1, Thomas Radke 3, Erik Schnetter 1,4, Edward Seidel 4,5 1 Center for Computation

More information

arxiv: v1 [physics.comp-ph] 24 Jul 2013

arxiv: v1 [physics.comp-ph] 24 Jul 2013 From Physics Model to Results: An Optimizing Framework for Cross-Architecture Code Generation arxiv:1307.6488v1 [physics.comp-ph] 24 Jul 2013 Marek Blazewicz 1,2,, Ian Hinder 3, David M. Koppelman 4,5,

More information

Dynamic Deployment of a Component Framework with the Ubiqis System

Dynamic Deployment of a Component Framework with the Ubiqis System Dynamic Deployment of a Component Framework with the Ubiqis System Steven Brandt (1) Gabrielle Allen (1,2) Matthew Eastman (1) Matthew Kemp (1) Erik Schnetter (1,3) (1) Center for Computation & Technology,

More information

Large-scale Gas Turbine Simulations on GPU clusters

Large-scale Gas Turbine Simulations on GPU clusters Large-scale Gas Turbine Simulations on GPU clusters Tobias Brandvik and Graham Pullan Whittle Laboratory University of Cambridge A large-scale simulation Overview PART I: Turbomachinery PART II: Stencil-based

More information

Optimising the Mantevo benchmark suite for multi- and many-core architectures

Optimising the Mantevo benchmark suite for multi- and many-core architectures Optimising the Mantevo benchmark suite for multi- and many-core architectures Simon McIntosh-Smith Department of Computer Science University of Bristol 1 Bristol's rich heritage in HPC The University of

More information

Introduction to Parallel Programming in OpenMp Dr. Yogish Sabharwal Department of Computer Science & Engineering Indian Institute of Technology, Delhi

Introduction to Parallel Programming in OpenMp Dr. Yogish Sabharwal Department of Computer Science & Engineering Indian Institute of Technology, Delhi Introduction to Parallel Programming in OpenMp Dr. Yogish Sabharwal Department of Computer Science & Engineering Indian Institute of Technology, Delhi Lecture - 01 Introduction to Parallel Computing Architectures

More information

6/20/16. Blue Waters User Monthly Teleconference

6/20/16. Blue Waters User Monthly Teleconference 6/20/16 Blue Waters User Monthly Teleconference Agenda Symposium File System Upgrade Status Maintenance Changes CUDA 7.5 Programming Environment Utilization Recent Events XSEDE Opportunities PUBLICATIONS!

More information

Fluent User Services Center

Fluent User Services Center Solver Settings 5-1 Using the Solver Setting Solver Parameters Convergence Definition Monitoring Stability Accelerating Convergence Accuracy Grid Independence Adaption Appendix: Background Finite Volume

More information

cdo Data Processing (and Production) Luis Kornblueh, Uwe Schulzweida, Deike Kleberg, Thomas Jahns, Irina Fast

cdo Data Processing (and Production) Luis Kornblueh, Uwe Schulzweida, Deike Kleberg, Thomas Jahns, Irina Fast cdo Data Processing (and Production) Luis Kornblueh, Uwe Schulzweida, Deike Kleberg, Thomas Jahns, Irina Fast Max-Planck-Institut für Meteorologie, DKRZ September 24, 2014 MAX-PLANCK-GESELLSCHAFT Data

More information

Day 2 August 06, 2004 (Friday)

Day 2 August 06, 2004 (Friday) An Overview of Grid Computing Day 2 August 06, 2004 (Friday) By CDAC Experts Contact :vcvrao@cdacindia.com; betatest@cdacindia.com URL : http://www.cs.umn.edu/~vcvrao 1 Betatesting Group,NPSF, C-DAC,Pune

More information

Introducing Overdecomposition to Existing Applications: PlasComCM and AMPI

Introducing Overdecomposition to Existing Applications: PlasComCM and AMPI Introducing Overdecomposition to Existing Applications: PlasComCM and AMPI Sam White Parallel Programming Lab UIUC 1 Introduction How to enable Overdecomposition, Asynchrony, and Migratability in existing

More information

Developing the TELEMAC system for HECToR (phase 2b & beyond) Zhi Shang

Developing the TELEMAC system for HECToR (phase 2b & beyond) Zhi Shang Developing the TELEMAC system for HECToR (phase 2b & beyond) Zhi Shang Outline of the Talk Introduction to the TELEMAC System and to TELEMAC-2D Code Developments Data Reordering Strategy Results Conclusions

More information

AstroGrid-D. Advanced Prototype Implementation of Monitoring & Steering Methods. Documentation and Test Report 1. Deliverable D6.5

AstroGrid-D. Advanced Prototype Implementation of Monitoring & Steering Methods. Documentation and Test Report 1. Deliverable D6.5 AstroGrid-D Deliverable D6.5 Advanced Prototype Implementation of Monitoring & Steering Methods Documentation and Test Report 1 Deliverable D6.5 Authors Thomas Radke (AEI) Editors Thomas Radke (AEI) Date

More information

Scientific Computing at Million-way Parallelism - Blue Gene/Q Early Science Program

Scientific Computing at Million-way Parallelism - Blue Gene/Q Early Science Program Scientific Computing at Million-way Parallelism - Blue Gene/Q Early Science Program Implementing Hybrid Parallelism in FLASH Christopher Daley 1 2 Vitali Morozov 1 Dongwook Lee 2 Anshu Dubey 1 2 Jonathon

More information

ScalaIOTrace: Scalable I/O Tracing and Analysis

ScalaIOTrace: Scalable I/O Tracing and Analysis ScalaIOTrace: Scalable I/O Tracing and Analysis Karthik Vijayakumar 1, Frank Mueller 1, Xiaosong Ma 1,2, Philip C. Roth 2 1 Department of Computer Science, NCSU 2 Computer Science and Mathematics Division,

More information

I/O Optimization Strategies in the PLUTO Code

I/O Optimization Strategies in the PLUTO Code Available on-line at www.prace-ri.eu Partnership for Advanced Computing in Europe I/O Optimization Strategies in the PLUTO Code A. Mignone a,, G. Muscianisi b, M. Rivi b, G. Bodo c a Diparnto di Fisica

More information

Community Software Development with the Astrophysics Simulation Collaboratory

Community Software Development with the Astrophysics Simulation Collaboratory CONCURRENCY AND COMPUTATION: PRACTICE AND EXPERIENCE Concurrency Computat.: Pract. Exper. 2001; volume (number): 000 000 Community Software Development with the Astrophysics Simulation Collaboratory 5

More information

Scientific Components and Frameworks

Scientific Components and Frameworks High Performance Computing: Concepts, Methods & Means Scientific Components and Frameworks Prof. Daniel S. Katz Department of Electrical and Computer Engineering Louisiana State University April 24 th,

More information

Algorithms, System and Data Centre Optimisation for Energy Efficient HPC

Algorithms, System and Data Centre Optimisation for Energy Efficient HPC 2015-09-14 Algorithms, System and Data Centre Optimisation for Energy Efficient HPC Vincent Heuveline URZ Computing Centre of Heidelberg University EMCL Engineering Mathematics and Computing Lab 1 Energy

More information

SPH: Why and what for?

SPH: Why and what for? SPH: Why and what for? 4 th SPHERIC training day David Le Touzé, Fluid Mechanics Laboratory, Ecole Centrale de Nantes / CNRS SPH What for and why? How it works? Why not for everything? Duality of SPH SPH

More information

Python, Clawpack, PyClaw, and PetClaw (using PETSc)

Python, Clawpack, PyClaw, and PetClaw (using PETSc) Python, Clawpack, PyClaw, and PetClaw (using PETSc) Randall J. LeVeque Kyle Mandli Applied Mathematics University of Washington Conservation Laws Package www.clawpack.org Supported in part by NSF, ONR,

More information

Overview and Recent Developments of Dynamic Mesh Capabilities

Overview and Recent Developments of Dynamic Mesh Capabilities Overview and Recent Developments of Dynamic Mesh Capabilities Henrik Rusche and Hrvoje Jasak h.rusche@wikki-gmbh.de and h.jasak@wikki.co.uk Wikki Gmbh, Germany Wikki Ltd, United Kingdom 6th OpenFOAM Workshop,

More information

Scientific Visualization Services at RZG

Scientific Visualization Services at RZG Scientific Visualization Services at RZG Klaus Reuter, Markus Rampp klaus.reuter@rzg.mpg.de Garching Computing Centre (RZG) 7th GOTiT High Level Course, Garching, 2010 Outline 1 Introduction 2 Details

More information

SC12 HPC Educators session: Unveiling parallelization strategies at undergraduate level

SC12 HPC Educators session: Unveiling parallelization strategies at undergraduate level SC12 HPC Educators session: Unveiling parallelization strategies at undergraduate level E. Ayguadé, R. M. Badia, D. Jiménez, J. Labarta and V. Subotic August 31, 2012 Index Index 1 1 The infrastructure:

More information

Tuning I/O Performance for Data Intensive Computing. Nicholas J. Wright. lbl.gov

Tuning I/O Performance for Data Intensive Computing. Nicholas J. Wright. lbl.gov Tuning I/O Performance for Data Intensive Computing. Nicholas J. Wright njwright @ lbl.gov NERSC- National Energy Research Scientific Computing Center Mission: Accelerate the pace of scientific discovery

More information

Unstructured Grid Numbering Schemes for GPU Coalescing Requirements

Unstructured Grid Numbering Schemes for GPU Coalescing Requirements Unstructured Grid Numbering Schemes for GPU Coalescing Requirements Andrew Corrigan 1 and Johann Dahm 2 Laboratories for Computational Physics and Fluid Dynamics Naval Research Laboratory 1 Department

More information

ORAP Forum October 10, 2013

ORAP Forum October 10, 2013 Towards Petaflop simulations of core collapse supernovae ORAP Forum October 10, 2013 Andreas Marek 1 together with Markus Rampp 1, Florian Hanke 2, and Thomas Janka 2 1 Rechenzentrum der Max-Planck-Gesellschaft

More information

Cactus-Simulationen on the Grid

Cactus-Simulationen on the Grid ------------------------------- Cactus-Simulationen on the Grid ------------------------------- Contact Person: Thomas Radke, AEI tradke@aei.mpg.de, 0331 / 567 7626 -------------------- 1. Scenario Overview

More information

Pre- and post-processing in Nektar++

Pre- and post-processing in Nektar++ Pre- and post-processing in Nektar++ D. Moxey, C. Cantwell, R. M. Kirby, S. Sherwin Department of Aeronautics, Imperial College London Nektar++ workshop 7 th July 2015 Outline Motivation Pre-processing

More information

Semester Final Report

Semester Final Report CSUMS SemesterFinalReport InLaTex AnnKimball 5/20/2009 ThisreportisageneralsummaryoftheaccumulationofknowledgethatIhavegatheredthroughoutthis semester. I was able to get a birds eye view of many different

More information

Introduction to High Performance Parallel I/O

Introduction to High Performance Parallel I/O Introduction to High Performance Parallel I/O Richard Gerber Deputy Group Lead NERSC User Services August 30, 2013-1- Some slides from Katie Antypas I/O Needs Getting Bigger All the Time I/O needs growing

More information

Turbostream: A CFD solver for manycore

Turbostream: A CFD solver for manycore Turbostream: A CFD solver for manycore processors Tobias Brandvik Whittle Laboratory University of Cambridge Aim To produce an order of magnitude reduction in the run-time of CFD solvers for the same hardware

More information

Development Environments for HPC: The View from NCSA

Development Environments for HPC: The View from NCSA Development Environments for HPC: The View from NCSA Jay Alameda National Center for Supercomputing Applications, University of Illinois at Urbana-Champaign DEHPC 15 San Francisco, CA 18 October 2015 Acknowledgements

More information

A Framework for Visualizing Hierarchical Computations

A Framework for Visualizing Hierarchical Computations A Framework for Visualizing Hierarchical Computations Terry J. Ligocki 1, Brian Van Straalen 2,JohnM.Shalf 1,GuntherH. Weber 3,4, and Bernd Hamann 1,3 1 Lawrence Berkeley National Laboratory, National

More information

Compiler Technology for Problem Solving on Computational Grids

Compiler Technology for Problem Solving on Computational Grids Compiler Technology for Problem Solving on Computational Grids An Overview of Programming Support Research in the GrADS Project Ken Kennedy Rice University http://www.cs.rice.edu/~ken/presentations/gridcompilers.pdf

More information

Improving Uintah s Scalability Through the Use of Portable

Improving Uintah s Scalability Through the Use of Portable Improving Uintah s Scalability Through the Use of Portable Kokkos-Based Data Parallel Tasks John Holmen1, Alan Humphrey1, Daniel Sunderland2, Martin Berzins1 University of Utah1 Sandia National Laboratories2

More information

3D Resistivity Logging Instruments

3D Resistivity Logging Instruments Parallel hp-finite Element Simulations of 3D Resistivity Logging Instruments M.Paszyński (1), D.Pardo (2), L.Demkowicz (3), C.Torres-Verdin (2) (1) Department of Modeling and Information Technology, AGH

More information

The parallelization of a block-tridiagonal matrix system for an electromagnetic wave simulation in TOKAMAK(TORIC) by MPI Fortran

The parallelization of a block-tridiagonal matrix system for an electromagnetic wave simulation in TOKAMAK(TORIC) by MPI Fortran The parallelization of a block-tridiagonal matrix system for an electromagnetic wave simulation in TOKAMAK(TORIC) by MPI Fortran Jungpyo Lee Graduate Student in a department

More information

Communication-Avoiding Optimization of Geometric Multigrid on GPUs

Communication-Avoiding Optimization of Geometric Multigrid on GPUs Communication-Avoiding Optimization of Geometric Multigrid on GPUs Amik Singh James Demmel, Ed. Electrical Engineering and Computer Sciences University of California at Berkeley Technical Report No. UCB/EECS-2012-258

More information

Real Parallel Computers

Real Parallel Computers Real Parallel Computers Modular data centers Overview Short history of parallel machines Cluster computing Blue Gene supercomputer Performance development, top-500 DAS: Distributed supercomputing Short

More information

CFDnet: Computational Fluid Dynamics on the Internet

CFDnet: Computational Fluid Dynamics on the Internet CFDnet: Computational Fluid Dynamics on the Internet F. E. Ham, J. Militzer and A. Bemfica Department of Mechanical Engineering Dalhousie University - DalTech Halifax, Nova Scotia Abstract CFDnet is computational

More information

Pratibha Gopalam, A. N. Cartwright, Bina Ramamurthy. University at Buffalo

Pratibha Gopalam, A. N. Cartwright, Bina Ramamurthy. University at Buffalo Center for Active Learning of Microelectronics and Photonics Java Enabled Opto-Electronic Learning Tools and A Supporting Framework Pratibha Gopalam, A. N. Cartwright, Bina Ramamurthy Computer Science

More information

Multi-Physics Multi-Code Coupling On Supercomputers

Multi-Physics Multi-Code Coupling On Supercomputers Multi-Physics Multi-Code Coupling On Supercomputers J.C. Cajas 1, G. Houzeaux 1, M. Zavala 1, M. Vázquez 1, B. Uekermann 2, B. Gatzhammer 2, M. Mehl 2, Y. Fournier 3, C. Moulinec 4 1) er, Edificio NEXUS

More information

Implementation of an integrated efficient parallel multiblock Flow solver

Implementation of an integrated efficient parallel multiblock Flow solver Implementation of an integrated efficient parallel multiblock Flow solver Thomas Bönisch, Panagiotis Adamidis and Roland Rühle adamidis@hlrs.de Outline Introduction to URANUS Why using Multiblock meshes

More information

HPC Algorithms and Applications

HPC Algorithms and Applications HPC Algorithms and Applications Dwarf #5 Structured Grids Michael Bader Winter 2012/2013 Dwarf #5 Structured Grids, Winter 2012/2013 1 Dwarf #5 Structured Grids 1. dense linear algebra 2. sparse linear

More information

The Icosahedral Nonhydrostatic (ICON) Model

The Icosahedral Nonhydrostatic (ICON) Model The Icosahedral Nonhydrostatic (ICON) Model Scalability on Massively Parallel Computer Architectures Florian Prill, DWD + the ICON team 15th ECMWF Workshop on HPC in Meteorology October 2, 2012 ICON =

More information

Example 13 - Shock Tube

Example 13 - Shock Tube Example 13 - Shock Tube Summary This famous experiment is interesting for observing the shock-wave propagation. Moreover, this case uses the representation of perfect gas and compares the different formulations:

More information

Parallel Languages: Past, Present and Future

Parallel Languages: Past, Present and Future Parallel Languages: Past, Present and Future Katherine Yelick U.C. Berkeley and Lawrence Berkeley National Lab 1 Kathy Yelick Internal Outline Two components: control and data (communication/sharing) One

More information

Energy Efficiency for Exascale. High Performance Computing, Grids and Clouds International Workshop Cetraro Italy June 27, 2012 Natalie Bates

Energy Efficiency for Exascale. High Performance Computing, Grids and Clouds International Workshop Cetraro Italy June 27, 2012 Natalie Bates Energy Efficiency for Exascale High Performance Computing, Grids and Clouds International Workshop Cetraro Italy June 27, 2012 Natalie Bates The Exascale Power Challenge Where do we get a 1000x improvement

More information

MRI Induced Heating of a Pacemaker. Peter Krenz, Application Engineer

MRI Induced Heating of a Pacemaker. Peter Krenz, Application Engineer MRI Induced Heating of a Pacemaker Peter Krenz, Application Engineer 1 Problem Statement Electric fields generated during MRI exposure are dissipated in tissue of the human body resulting in a temperature

More information

Parallel Mesh Partitioning in Alya

Parallel Mesh Partitioning in Alya Available online at www.prace-ri.eu Partnership for Advanced Computing in Europe Parallel Mesh Partitioning in Alya A. Artigues a *** and G. Houzeaux a* a Barcelona Supercomputing Center ***antoni.artigues@bsc.es

More information

Introduction to GT3. Introduction to GT3. What is a Grid? A Story of Evolution. The Globus Project

Introduction to GT3. Introduction to GT3. What is a Grid? A Story of Evolution. The Globus Project Introduction to GT3 The Globus Project Argonne National Laboratory USC Information Sciences Institute Copyright (C) 2003 University of Chicago and The University of Southern California. All Rights Reserved.

More information

Achieving Efficient Strong Scaling with PETSc Using Hybrid MPI/OpenMP Optimisation

Achieving Efficient Strong Scaling with PETSc Using Hybrid MPI/OpenMP Optimisation Achieving Efficient Strong Scaling with PETSc Using Hybrid MPI/OpenMP Optimisation Michael Lange 1 Gerard Gorman 1 Michele Weiland 2 Lawrence Mitchell 2 Xiaohu Guo 3 James Southern 4 1 AMCG, Imperial College

More information

COSC6365. Introduction to HPC. Lecture 21. Lennart Johnsson Department of Computer Science

COSC6365. Introduction to HPC. Lecture 21. Lennart Johnsson Department of Computer Science Introduction to HPC Lecture 21 Department of Computer Science Most slides from UC Berkeley CS 267 Spring 2011, Lecture 12, Dense Linear Algebra (part 2), Parallel Gaussian Elimination. Jim Demmel Dense

More information

Experiences with ENZO on the Intel R Many Integrated Core (Intel MIC) Architecture

Experiences with ENZO on the Intel R Many Integrated Core (Intel MIC) Architecture Experiences with ENZO on the Intel R Many Integrated Core (Intel MIC) Architecture 1 Introduction Robert Harkness National Institute for Computational Sciences Oak Ridge National Laboratory The National

More information

Scalable Computing: Practice and Experience Volume 11, Number 2, pp

Scalable Computing: Practice and Experience Volume 11, Number 2, pp Scalable Computing: Practice and Experience Volume 11, Number 2, pp. 205 220. http://www.scpe.org ISSN 1895-1767 c 2010 SCPE LARGE SCALE PROBLEM SOLVING USING AUTOMATIC CODE GENERATION AND DISTRIBUTED

More information

Parallel I/O Performance Study and Optimizations with HDF5, A Scientific Data Package

Parallel I/O Performance Study and Optimizations with HDF5, A Scientific Data Package Parallel I/O Performance Study and Optimizations with HDF5, A Scientific Data Package MuQun Yang, Christian Chilan, Albert Cheng, Quincey Koziol, Mike Folk, Leon Arber The HDF Group Champaign, IL 61820

More information

First Steps of YALES2 Code Towards GPU Acceleration on Standard and Prototype Cluster

First Steps of YALES2 Code Towards GPU Acceleration on Standard and Prototype Cluster First Steps of YALES2 Code Towards GPU Acceleration on Standard and Prototype Cluster YALES2: Semi-industrial code for turbulent combustion and flows Jean-Matthieu Etancelin, ROMEO, NVIDIA GPU Application

More information

The State and Needs of IO Performance Tools

The State and Needs of IO Performance Tools The State and Needs of IO Performance Tools Scalable Tools Workshop Lake Tahoe, CA August 6 12, 2017 This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National

More information

Development of an Integrated Modeling Framework for Simulations of Coastal Processes in Deltaic Environments Using High-Performance Computing

Development of an Integrated Modeling Framework for Simulations of Coastal Processes in Deltaic Environments Using High-Performance Computing Development of an Integrated Modeling Framework for Simulations of Coastal Processes in Deltaic Environments Using High-Performance Computing Q. Jim Chen Department of Civil and Environmental Engineering

More information

Grid Programming Models: Current Tools, Issues and Directions. Computer Systems Research Department The Aerospace Corporation, P.O.

Grid Programming Models: Current Tools, Issues and Directions. Computer Systems Research Department The Aerospace Corporation, P.O. Grid Programming Models: Current Tools, Issues and Directions Craig Lee Computer Systems Research Department The Aerospace Corporation, P.O. Box 92957 El Segundo, CA USA lee@aero.org Domenico Talia DEIS

More information

General Plasma Physics

General Plasma Physics Present and Future Computational Requirements General Plasma Physics Center for Integrated Computation and Analysis of Reconnection and Turbulence () Kai Germaschewski, Homa Karimabadi Amitava Bhattacharjee,

More information

The Cactus Framework. Tom Goodale LSU Center for Computation and Technology and Cardiff University

The Cactus Framework. Tom Goodale LSU Center for Computation and Technology and Cardiff University The Cactus Framework Tom Goodale LSU Center for Computation and Technology and Cardiff University goodale@cct.lsu.edu Introduction What is Cactus? CACTUS is a generic, freely available, modular, portable

More information

Supercomputing and Science An Introduction to High Performance Computing

Supercomputing and Science An Introduction to High Performance Computing Supercomputing and Science An Introduction to High Performance Computing Part VII: Scientific Computing Henry Neeman, Director OU Supercomputing Center for Education & Research Outline Scientific Computing

More information

ALE and AMR Mesh Refinement Techniques for Multi-material Hydrodynamics Problems

ALE and AMR Mesh Refinement Techniques for Multi-material Hydrodynamics Problems ALE and AMR Mesh Refinement Techniques for Multi-material Hydrodynamics Problems A. J. Barlow, AWE. ICFD Workshop on Mesh Refinement Techniques 7th December 2005 Acknowledgements Thanks to Chris Powell,

More information

LEHMAN COLLEGE OF THE CITY UNIVERSITY OF NEW YORK DEPARTMENT OF PHYSICS AND ASTRONOMY CURRICULUM CHANGE

LEHMAN COLLEGE OF THE CITY UNIVERSITY OF NEW YORK DEPARTMENT OF PHYSICS AND ASTRONOMY CURRICULUM CHANGE Name of Program and Degree Award: Physics, B.S. Hegis Number: 1902.00 Program Code: 34031 Effective Term: Fall 2018 1. Type of Change: Course Requirements for the B.S. Physics Program Physics, B.S. (60

More information

High Performance Computing

High Performance Computing High Performance Computing ADVANCED SCIENTIFIC COMPUTING Dr. Ing. Morris Riedel Adjunct Associated Professor School of Engineering and Natural Sciences, University of Iceland Research Group Leader, Juelich

More information

Enzo-P / Cello. Formation of the First Galaxies. San Diego Supercomputer Center. Department of Physics and Astronomy

Enzo-P / Cello. Formation of the First Galaxies. San Diego Supercomputer Center. Department of Physics and Astronomy Enzo-P / Cello Formation of the First Galaxies James Bordner 1 Michael L. Norman 1 Brian O Shea 2 1 University of California, San Diego San Diego Supercomputer Center 2 Michigan State University Department

More information

The Study of Ship Motions in Regular Waves using a Mesh-Free Numerical Method

The Study of Ship Motions in Regular Waves using a Mesh-Free Numerical Method The Study of Ship Motions in Regular Waves using a Mesh-Free Numerical Method by Bruce Kenneth Cartwright, B. Eng., M. Sc. Submitted in fulfilment of the requirements for the Degree of Master of Philosophy

More information

THE NATIONAL DATA SERVICE(S) & NDS CONSORTIUM A Call to Action for Accelerating Discovery Through Data Services we can Build Ed Seidel

THE NATIONAL DATA SERVICE(S) & NDS CONSORTIUM A Call to Action for Accelerating Discovery Through Data Services we can Build Ed Seidel THE NATIONAL DATA SERVICE(S) & NDS CONSORTIUM A Call to Action for Accelerating Discovery Through Data Services we can Build Ed Seidel National Center for Supercomputing Applications University of Illinois

More information

A Multiscale Non-hydrostatic Atmospheric Model for Regional and Global Applications

A Multiscale Non-hydrostatic Atmospheric Model for Regional and Global Applications A Multiscale Non-hydrostatic Atmospheric Model for Regional and Global Applications James D. Doyle 1, Frank Giraldo 2, Saša Gaberšek 1 1 Naval Research Laboratory, Monterey, CA, USA 2 Naval Postgraduate

More information

HDF5 and STEP/NRF Database for SINDA/G

HDF5 and STEP/NRF Database for SINDA/G HDF5 and STEP/NRF Database for SINDA/G October 22 st, 2003 Ron Behee Network Analysis Inc Why NAI Chose HDF5 Widely used for scientific data Proposed database for STEP/NRF Public domain from NCSA Good

More information

Chapter 1: An Overview of Computers and Programming Languages. Objectives. Objectives (cont d.) Introduction

Chapter 1: An Overview of Computers and Programming Languages. Objectives. Objectives (cont d.) Introduction Chapter 1: An Overview of Computers and Programming Languages Objectives Objectives (cont d.) In this chapter, you will: Learn about different types of computers Explore hardware and software Learn about

More information

Introduction to 3D Scientific Visualization. Training in Visualization for PRACE Summer of HPC 2013 Leon Kos, University of Ljubljana, Slovenia

Introduction to 3D Scientific Visualization. Training in Visualization for PRACE Summer of HPC 2013 Leon Kos, University of Ljubljana, Slovenia Introduction to 3D Scientific Visualization Training in Visualization for PRACE Summer of HPC 2013 Leon Kos, University of Ljubljana, Slovenia Motto Few correctly put words is worth hundreds of images.

More information

Towards Self-Verification in Finite Difference Code Generation

Towards Self-Verification in Finite Difference Code Generation Towards Self-Verification in Finite Difference Code Generation Jan Hückelheim 1, Ziqing Luo 2, Fabio Luporini 1,Navjot Kukreja 1, Michael Lange 1, Gerard Gorman 1, Stephen Siegel 2, Matthew Dwyer 3, Paul

More information