Supercomputing Engine for Mathematica. Supercomputing 2009 Portland, Oregon
|
|
- Edgar Rose
- 5 years ago
- Views:
Transcription
1 Supercomputing Engine for Mathematica Supercomputing 2009 Portland, Oregon
2 Supercomputing Engine for Mathematica Dean E. Dauger, Ph. D. President, Dauger Research, Inc.
3 The First Mac Cluster - established G3/266ʼs Mac OS 8.1 Decyk, Dauger TheKokelaar, Dawson Cluster
4 Dauger Research, Inc. High-Performance, Scientific, and Cluster Computing Software Plug-and-Play Supercomputer-Compatible Clusters Pooch App - Now on Linux! Source-Code Tutorials Visualization & Simulation Consulting Services Optimization Parallelization Vectorization
5 Why Parallel Computing? Problems too large to solve on one computer Takes too much time Requires more memory can outgrow RAM capacity Programming API standardized Message-Passing Interface (MPI) specification established in 1994 dominant software interface at supercomputing centers portable MPI code in Fortran and C
6 Why Parallel Computing? Problems too large to solve on one computer core Computational power per core no longer doubles Doubling cores is how Mooreʼs Law technically holds Multicore utilization is not automatic Software must make up for where hardware leaves off Choose your parallel programming paradigm wisely Multicore Eroding Mooreʼs Law
7 Mathematica How to parallelize it? Excels in high-level and symbolic processing and modeling Used for many different fields of study But how to make it faster!
8 Front End and Kernel Mathematicaʼs Structure Front End Kernel
9 Front End and Kernel Mathematicaʼs Structure Front End Instructions and Data Kernel Processing
10 Front End and Kernel Mathematicaʼs Structure Front End Display Kernel Results
11 Front End and Kernel How to parallelize it? Front End Kernel
12 Parallelizing Front-End/Kernel Working in between Supercomputing Engine for Mathematica
13 Parallelizing Front-End/Kernel Working in between Supercomputing Engine for Mathematica Kernels are unaware they are running in parallel
14 MPI in Mathematica Supercomputing Engine for Mathematica Closely follows MPI standard in Mathematica environment Basic MPI calls (mpisend, mpirecv) Asynchronous MPI calls (mpiisend, mpiirecv, mpitest) Collective MPI calls (mpibcast, mpigather, mpialltoall) High-level parallel calls for common tasks ParallelTable, EdgeCell, ParallelFourier, ElementManage Basic Parallel I/O Automatically locates, launches, configures, and coordinates Mathematica kernels via Pooch from command line or Mathematicaʼs Front End Builds on any licensed gridmathematica
15 Why MPI instead of grid? Typical grid implementation
16 Why MPI instead of grid? MPI implementation
17 Parallelizing Front-End/Kernel Working in between Supercomputing Engine for Mathematica
18 Parallelizing Front-End/Kernel Working in between
19 Parallelizing Front-End/Kernel Working in between
20 Parallelizing Front-End/Kernel Working in between Instructions and Data Processing
21 Parallelizing Front-End/Kernel Working in between Display Results
22 Parallel Code using MPI Code coordinating parallel work using messages N tasks or virtual processors running simultaneously, labeled 0 through N-1 executables often use identification data to determine algorithmically what part of the problem on which to work tasks pass messages amongst themselves to organize data and coordinate work any group of tasks can communicate ( O(N 2 ) connections) simple sends and receives collective calls - broadcast, gather, reduce, transpose, etc. synchronization not required, but often implied by messages
23 Game of Life Cellular Automata - J. Conway, Princeton * * * * * * 59 1 * * *2010 Life of one cell depends on its neighbors
24 Parallel Life Message-Passing for Cellular Automata * * * * * * 59 1 * * *2010 Data to propagate between partitions
25 Parallel Life Message-Passing for Cellular Automata * 2 * * 079 * * 59 * * * *2010 Message exchange maintains guard cells
26 Plasma Simulation Plasma Dynamics * * * * * * * * * ** * * * *** ** * * * * * * ** * * * * ** * * * * * * * * * * * * * * * ** * * ** * ** * * * * ** * * * * * * ** * * * * ** * * ** * *** * * * * * * * * * * * * * * * * * * * * * * * * ** * * * ** ** ** * * * ** * ** * * * ** * * * * * * * ** * ** * * * * ** ** ** * * * *** * ** * ** * * *** * * * * * *** * ***** ** * ** * * * * * * * *** * * * * * * * * ** * ** * * * * * * * * ** * ** * ** * * * * * ** *** * * * * ** * ** * * * * * * * ** * * * * ** * * * * * * * * ** * * * ** ** * * * * * * * * * * * * * V. K. Decyk, C. D. Norton, Comp. Phys. Communications 164 (2004) 80-85
27 Parallel Plasma Simulation Message-Passing for Plasma Dynamics * * * * * * * * * ** * * * *** ** * * * * * * ** * * * * ** * * * * * * * * * * * * * * * ** * * ** * ** * * * * ** * * * * * * ** * * * * ** * * ** * *** * * * * * * * * * * * * * * * * * * * * * * * * ** * * * ** ** ** * * * ** * ** * * * ** * * * * * * * ** * ** * * * * ** ** ** * * * *** * ** * ** * * *** * * * * * *** * ***** ** * ** * * * * * * * *** * * * * * * * * ** * ** * * * * * * * * ** * ** * ** * * * * * ** *** * * * * ** * ** * * * * * * * ** * * * * ** * * * * * * * * ** * * * ** ** * * * * * * * * * * * * * V. K. Decyk, C. D. Norton, Comp. Phys. Communications 164 (2004) 80-85
28 Parallel Plasma Simulation Message-Passing for Plasma Dynamics ** * * * * * * * * * * * * * * ** * * ** * ** ** ** * * * ** * * * * * * *** ** ** * * * * ** * ** * *** * * * * * * * ** * ** * * *** ** * * * ** * * * * * * * * * * * * * * ** * * * * * * * * * * ** * * * * * * * * * * * * * * * * * * * * * * * * * ** * * * * * * * * * * * * * * * * * * ** ** * * * * * ** * ** * * * *** * * * * * * ** ** * * * ** * * * * * * * ** * * * * * * * * * * * * * ** * * * * * * * ** * * * * * * * * * * * * * * * * *** * * * * * * ** * * * * * ** *** * * * * * * * * * * * * * * * * * * * * * Particles propagate from one partition to its neighbor
29 Parallel Plasma Simulation Message-Passing for Plasma Dynamics ** * * * * * * * * * * * * * * ** * * ** * ** ** ** * * * ** * * * * * * *** ** ** * * * * ** * ** * *** * * * * * * * ** * ** * * *** ** * * * ** * * * * * * * * * * * * * * ** * * * * * * * * * * ** * * * * * * * * * * * * * * * * * * * * * * * * * ** * * * * * * * * * * * * * * * * * * ** ** * * * * * ** * ** * * * *** * * * * * * ** ** * * * ** * * * * * * * ** * * * * * * * * * * * * * ** * * * * * * * ** * * * * * * * * * * * * * * * * *** * * * * * * ** * * * * * ** *** * * * * * * * * * * * * * * * * * * * * * Particles propagate from one partition to its neighbor
30 Message-Passing Patterns Supported via MPI Master-Slave All to All Tree Nearest Neighbor Irregular Or Combinations
31 For a Demonstration See Booth 2295 Dean E. Dauger, Ph. D. d@daugerresearch.com
32 Ping-pong Benchmark gridmathematica and SEM based on benchmark by V. K. Decyk sending digits of π as text Message Size
33 Swap Benchmark gridmathematica and SEM based on benchmark by V. K. Decyk sending digits of π as text Message Size
34 We found SEM to be very efficient in terms of stability and use for financial engineers who do not have the time to optimize load balancing issues but want to focus on modeling. - Vineer Bhansali, Managing Director, PIMCO
35 "I CAN endorse SEM with pride. I would emphasize the flexibility and scalability of SEM. And SEM can make writing Mathematica programs even more flexible than before." - Yuko Matsuda, Professor, Tokyo Institute of Technology
36 Parallelizing Front-End/Kernel Working in between
37 Parallelizing Other Applications? Factored into Front-end/Kernel Model Matlab If we know the language Maple MathCAD Supercomputing Engine Data Mining Graphics Renderers Your Factored Code...
38 Who to Contact Dean E. Dauger, Ph. D. Dauger Research, Inc. Zvi Tannenbaum Advanced Cluster Systems, LLC &
39 Roadmap Advanced Cluster Systems Booth Booth #2295 Mon-Thurs Exhibitor Forum Session Room C123-4 Wed 11 am
40 For More Information
41 Reference Library Documentation Supercomputing Engine for Mathematica Site Advanced Cluster Systems Site Tutorials on Writing Parallel Code Mac Clustering on National Television Related Publications Multicore Eroding Mooreʼs Law
42 Q&A Dean E. Dauger, Ph. D. President, Dauger Research, Inc. &
Supercomputing Engine for Mathematica. Machine Evaluation Workshop 19-2 Dec 2008 Runcorn, Daresbury, United Kingdom
Supercomputing Engine for Mathematica Machine Evaluation Workshop 19-2 Dec 2008 Runcorn, Daresbury, United Kingdom Supercomputing Engine for Mathematica Dean E. Dauger, Ph. D. President, Dauger Research,
More informationSupercomputing Engine for Mathematica. Supercomputing 2008 Austin, Texas
Supercomputing Engine for Mathematica Supercomputing 2008 Austin, Texas Supercomputing Engine for Mathematica Dean E. Dauger, Ph. D. President, Dauger Research, Inc. d@daugerresearch.com The First Mac
More informationYet Another Parallel Computations in Mathematica
Yet Another Parallel Computations in Mathematica Yuko Matsuda (matsuda@symbolics.jp) Symbolic Systems LLC Dean Dauger Dauger Research Inc. Zvi Tannenbaum Advanced Cluster Systems LLC 2 Abstract Wolfram
More informationCSE. Parallel Algorithms on a cluster of PCs. Ian Bush. Daresbury Laboratory (With thanks to Lorna Smith and Mark Bull at EPCC)
Parallel Algorithms on a cluster of PCs Ian Bush Daresbury Laboratory I.J.Bush@dl.ac.uk (With thanks to Lorna Smith and Mark Bull at EPCC) Overview This lecture will cover General Message passing concepts
More informationScalasca performance properties The metrics tour
Scalasca performance properties The metrics tour Markus Geimer m.geimer@fz-juelich.de Scalasca analysis result Generic metrics Generic metrics Time Total CPU allocation time Execution Overhead Visits Hardware
More informationOn-Demand Supercomputing Multiplies the Possibilities
Microsoft Windows Compute Cluster Server 2003 Partner Solution Brief Image courtesy of Wolfram Research, Inc. On-Demand Supercomputing Multiplies the Possibilities Microsoft Windows Compute Cluster Server
More informationBasic Communication Operations (Chapter 4)
Basic Communication Operations (Chapter 4) Vivek Sarkar Department of Computer Science Rice University vsarkar@cs.rice.edu COMP 422 Lecture 17 13 March 2008 Review of Midterm Exam Outline MPI Example Program:
More informationParticle-in-Cell Simulations on Modern Computing Platforms. Viktor K. Decyk and Tajendra V. Singh UCLA
Particle-in-Cell Simulations on Modern Computing Platforms Viktor K. Decyk and Tajendra V. Singh UCLA Outline of Presentation Abstraction of future computer hardware PIC on GPUs OpenCL and Cuda Fortran
More informationCS 470 Spring Mike Lam, Professor. Distributed Programming & MPI
CS 470 Spring 2017 Mike Lam, Professor Distributed Programming & MPI MPI paradigm Single program, multiple data (SPMD) One program, multiple processes (ranks) Processes communicate via messages An MPI
More informationBulk Synchronous and SPMD Programming. The Bulk Synchronous Model. CS315B Lecture 2. Bulk Synchronous Model. The Machine. A model
Bulk Synchronous and SPMD Programming The Bulk Synchronous Model CS315B Lecture 2 Prof. Aiken CS 315B Lecture 2 1 Prof. Aiken CS 315B Lecture 2 2 Bulk Synchronous Model The Machine A model An idealized
More informationDeveloping a Thin and High Performance Implementation of Message Passing Interface 1
Developing a Thin and High Performance Implementation of Message Passing Interface 1 Theewara Vorakosit and Putchong Uthayopas Parallel Research Group Computer and Network System Research Laboratory Department
More informationProgramming with MPI on GridRS. Dr. Márcio Castro e Dr. Pedro Velho
Programming with MPI on GridRS Dr. Márcio Castro e Dr. Pedro Velho Science Research Challenges Some applications require tremendous computing power - Stress the limits of computing power and storage -
More informationParallel & Cluster Computing. cs 6260 professor: elise de doncker by: lina hussein
Parallel & Cluster Computing cs 6260 professor: elise de doncker by: lina hussein 1 Topics Covered : Introduction What is cluster computing? Classification of Cluster Computing Technologies: Beowulf cluster
More informationDeveloping PIC Codes for the Next Generation Supercomputer using GPUs. Viktor K. Decyk UCLA
Developing PIC Codes for the Next Generation Supercomputer using GPUs Viktor K. Decyk UCLA Abstract The current generation of supercomputer (petaflops scale) cannot be scaled up to exaflops (1000 petaflops),
More informationCS 470 Spring Mike Lam, Professor. Distributed Programming & MPI
CS 470 Spring 2018 Mike Lam, Professor Distributed Programming & MPI MPI paradigm Single program, multiple data (SPMD) One program, multiple processes (ranks) Processes communicate via messages An MPI
More informationProgramming with MPI. Pedro Velho
Programming with MPI Pedro Velho Science Research Challenges Some applications require tremendous computing power - Stress the limits of computing power and storage - Who might be interested in those applications?
More informationDistributed Memory Parallel Programming
COSC Big Data Analytics Parallel Programming using MPI Edgar Gabriel Spring 201 Distributed Memory Parallel Programming Vast majority of clusters are homogeneous Necessitated by the complexity of maintaining
More informationPractical Scientific Computing: Performanceoptimized
Practical Scientific Computing: Performanceoptimized Programming Programming with MPI November 29, 2006 Dr. Ralf-Peter Mundani Department of Computer Science Chair V Technische Universität München, Germany
More informationAdventures in Load Balancing at Scale: Successes, Fizzles, and Next Steps
Adventures in Load Balancing at Scale: Successes, Fizzles, and Next Steps Rusty Lusk Mathematics and Computer Science Division Argonne National Laboratory Outline Introduction Two abstract programming
More informationL17: Introduction to Irregular Algorithms and MPI, cont.! November 8, 2011!
L17: Introduction to Irregular Algorithms and MPI, cont.! November 8, 2011! Administrative Class cancelled, Tuesday, November 15 Guest Lecture, Thursday, November 17, Ganesh Gopalakrishnan CUDA Project
More informationPerformance properties The metrics tour
Performance properties The metrics tour Markus Geimer & Brian Wylie Jülich Supercomputing Centre scalasca@fz-juelich.de January 2012 Scalasca analysis result Confused? Generic metrics Generic metrics Time
More informationIntroduction to Parallel Programming
Introduction to Parallel Programming Overview Parallel programming allows the user to use multiple cpus concurrently Reasons for parallel execution: shorten execution time by spreading the computational
More informationBasic MPI Communications. Basic MPI Communications (cont d)
Basic MPI Communications MPI provides two non-blocking routines: MPI_Isend(buf,cnt,type,dst,tag,comm,reqHandle) buf: source of data to be sent cnt: number of data elements to be sent type: type of each
More informationThe Use of the MPI Communication Library in the NAS Parallel Benchmarks
The Use of the MPI Communication Library in the NAS Parallel Benchmarks Theodore B. Tabe, Member, IEEE Computer Society, and Quentin F. Stout, Senior Member, IEEE Computer Society 1 Abstract The statistical
More informationScalasca performance properties The metrics tour
Scalasca performance properties The metrics tour Markus Geimer m.geimer@fz-juelich.de Scalasca analysis result Generic metrics Generic metrics Time Total CPU allocation time Execution Overhead Visits Hardware
More informationParallel Simulation of Dendritic Growth On Unstructured Grids
Parallel Simulation of Dendritic Growth On Unstructured Grids, Julian Hammer, Dietmar Fey Friedrich-Alexander-Universität Erlangen-Nürnberg IA 3 Workshop @SC11, Nov. 13th, 2011 Outline 1 What and why?
More informationPerformance properties The metrics tour
Performance properties The metrics tour Markus Geimer & Brian Wylie Jülich Supercomputing Centre scalasca@fz-juelich.de August 2012 Scalasca analysis result Online description Analysis report explorer
More information<Insert Picture Here> Oracle Coherence & Extreme Transaction Processing (XTP)
Oracle Coherence & Extreme Transaction Processing (XTP) Gary Hawks Oracle Coherence Solution Specialist Extreme Transaction Processing What is XTP? Introduction to Oracle Coherence
More informationPerformance properties The metrics tour
Performance properties The metrics tour Markus Geimer & Brian Wylie Jülich Supercomputing Centre scalasca@fz-juelich.de Scalasca analysis result Online description Analysis report explorer GUI provides
More informationUNIT 9C Randomness in Computation: Cellular Automata Principles of Computing, Carnegie Mellon University
UNIT 9C Randomness in Computation: Cellular Automata 1 Exam locations: Announcements 2:30 Exam: Sections A, B, C, D, E go to Rashid (GHC 4401) Sections F, G go to PH 125C. 3:30 Exam: All sections go to
More informationMA471. Lecture 5. Collective MPI Communication
MA471 Lecture 5 Collective MPI Communication Today: When all the processes want to send, receive or both Excellent website for MPI command syntax available at: http://www-unix.mcs.anl.gov/mpi/www/ 9/10/2003
More informationMessage Passing Interface (MPI)
What the course is: An introduction to parallel systems and their implementation using MPI A presentation of all the basic functions and types you are likely to need in MPI A collection of examples What
More informationUCLA Parallel PIC Framework
UCLA Parallel PIC Framework Viktor K. Decyk 1,2, Charles D. Norton 2 1 Department of Physics and Astronomy University of California, Los Angeles Los Angeles, California 90095-1547 2 Jet Propulsion Laboratory
More informationStandard MPI - Message Passing Interface
c Ewa Szynkiewicz, 2007 1 Standard MPI - Message Passing Interface The message-passing paradigm is one of the oldest and most widely used approaches for programming parallel machines, especially those
More informationCS4961 Parallel Programming. Lecture 18: Introduction to Message Passing 11/3/10. Final Project Purpose: Mary Hall November 2, 2010.
Parallel Programming Lecture 18: Introduction to Message Passing Mary Hall November 2, 2010 Final Project Purpose: - A chance to dig in deeper into a parallel programming model and explore concepts. -
More informationClusters of SMP s. Sean Peisert
Clusters of SMP s Sean Peisert What s Being Discussed Today SMP s Cluters of SMP s Programming Models/Languages Relevance to Commodity Computing Relevance to Supercomputing SMP s Symmetric Multiprocessors
More informationCS 6230: High-Performance Computing and Parallelization Introduction to MPI
CS 6230: High-Performance Computing and Parallelization Introduction to MPI Dr. Mike Kirby School of Computing and Scientific Computing and Imaging Institute University of Utah Salt Lake City, UT, USA
More informationPlug-and-Play Cluster Computing : HPC Designed for the Mainstream Scientist
Plug-and-Play Cluster Computing : HPC Designed for the Mainstream Scientist Dean E. Dauger 1 and Viktor K. Decyk 2 1 Dauger Research, Inc., P. O. Box 3074, Huntington Beach, CA 92605 USA http://daugerresearch.com/
More informationBig Orange Bramble. August 09, 2016
Big Orange Bramble August 09, 2016 Overview HPL SPH PiBrot Numeric Integration Parallel Pi Monte Carlo FDS DANNA HPL High Performance Linpack is a benchmark for clusters Created here at the University
More informationNAG at Manchester. Michael Croucher (University of Manchester)
NAG at Manchester Michael Croucher (University of Manchester) Michael.Croucher@manchester.ac.uk www.walkingrandomly.com Twitter: @walkingrandomly My background PhD Computational Physics from Sheffield
More informationIntroduction to MPI. Ricardo Fonseca. https://sites.google.com/view/rafonseca2017/
Introduction to MPI Ricardo Fonseca https://sites.google.com/view/rafonseca2017/ Outline Distributed Memory Programming (MPI) Message Passing Model Initializing and terminating programs Point to point
More informationExercises: April 11. Hermann Härtig, TU Dresden, Distributed OS, Load Balancing
Exercises: April 11 1 PARTITIONING IN MPI COMMUNICATION AND NOISE AS HPC BOTTLENECK LOAD BALANCING DISTRIBUTED OPERATING SYSTEMS, SCALABILITY, SS 2017 Hermann Härtig THIS LECTURE Partitioning: bulk synchronous
More informationMoore s Law. Computer architect goal Software developer assumption
Moore s Law The number of transistors that can be placed inexpensively on an integrated circuit will double approximately every 18 months. Self-fulfilling prophecy Computer architect goal Software developer
More informationChapter 4: Threads. Chapter 4: Threads. Overview Multicore Programming Multithreading Models Thread Libraries Implicit Threading Threading Issues
Chapter 4: Threads Silberschatz, Galvin and Gagne 2013 Chapter 4: Threads Overview Multicore Programming Multithreading Models Thread Libraries Implicit Threading Threading Issues 4.2 Silberschatz, Galvin
More informationPerformance Analysis and Optimal Utilization of Inter-Process Communications on Commodity Clusters
Performance Analysis and Optimal Utilization of Inter-Process Communications on Commodity Yili TSENG Department of Computer Systems Technology North Carolina A & T State University Greensboro, NC 27411,
More informationMPI and comparison of models Lecture 23, cs262a. Ion Stoica & Ali Ghodsi UC Berkeley April 16, 2018
MPI and comparison of models Lecture 23, cs262a Ion Stoica & Ali Ghodsi UC Berkeley April 16, 2018 MPI MPI - Message Passing Interface Library standard defined by a committee of vendors, implementers,
More informationThe MOSIX Scalable Cluster Computing for Linux. mosix.org
The MOSIX Scalable Cluster Computing for Linux Prof. Amnon Barak Computer Science Hebrew University http://www. mosix.org 1 Presentation overview Part I : Why computing clusters (slide 3-7) Part II : What
More informationCISC 879 Software Support for Multicore Architectures Spring Student Presentation 6: April 8. Presenter: Pujan Kafle, Deephan Mohan
CISC 879 Software Support for Multicore Architectures Spring 2008 Student Presentation 6: April 8 Presenter: Pujan Kafle, Deephan Mohan Scribe: Kanik Sem The following two papers were presented: A Synchronous
More informationCINES MPI. Johanne Charpentier & Gabriel Hautreux
Training @ CINES MPI Johanne Charpentier & Gabriel Hautreux charpentier@cines.fr hautreux@cines.fr Clusters Architecture OpenMP MPI Hybrid MPI+OpenMP MPI Message Passing Interface 1. Introduction 2. MPI
More informationLOAD BALANCING DISTRIBUTED OPERATING SYSTEMS, SCALABILITY, SS Hermann Härtig
LOAD BALANCING DISTRIBUTED OPERATING SYSTEMS, SCALABILITY, SS 2016 Hermann Härtig LECTURE OBJECTIVES starting points independent Unix processes and block synchronous execution which component (point in
More informationScreencast: What is [Open] MPI?
Screencast: What is [Open] MPI? Jeff Squyres May 2008 May 2008 Screencast: What is [Open] MPI? 1 What is MPI? Message Passing Interface De facto standard Not an official standard (IEEE, IETF, ) Written
More informationScreencast: What is [Open] MPI? Jeff Squyres May May 2008 Screencast: What is [Open] MPI? 1. What is MPI? Message Passing Interface
Screencast: What is [Open] MPI? Jeff Squyres May 2008 May 2008 Screencast: What is [Open] MPI? 1 What is MPI? Message Passing Interface De facto standard Not an official standard (IEEE, IETF, ) Written
More informationNon-Blocking Collectives for MPI
Non-Blocking Collectives for MPI overlap at the highest level Torsten Höfler Open Systems Lab Indiana University Bloomington, IN, USA Institut für Wissenschaftliches Rechnen Technische Universität Dresden
More informationSzámítogépes modellezés labor (MSc)
Számítogépes modellezés labor (MSc) Running Simulations on Supercomputers Gábor Rácz Physics of Complex Systems Department Eötvös Loránd University, Budapest September 19, 2018, Budapest, Hungary Outline
More informationLecture 4: Locality and parallelism in simulation I
Lecture 4: Locality and parallelism in simulation I David Bindel 6 Sep 2011 Logistics Distributed memory machines Each node has local memory... and no direct access to memory on other nodes Nodes communicate
More informationMPI Programming. Henrik R. Nagel Scientific Computing IT Division
1 MPI Programming Henrik R. Nagel Scientific Computing IT Division 2 Outline Introduction Finite Difference Method Finite Element Method LU Factorization SOR Method Monte Carlo Method Molecular Dynamics
More informationChapter 4: Multithreaded Programming
Chapter 4: Multithreaded Programming Silberschatz, Galvin and Gagne 2013 Chapter 4: Multithreaded Programming Overview Multicore Programming Multithreading Models Thread Libraries Implicit Threading Threading
More informationSPEC MPI2007 Benchmarks for HPC Systems
SPEC MPI2007 Benchmarks for HPC Systems Ron Lieberman Chair, SPEC HPG HP-MPI Performance Hewlett-Packard Company Dr. Tom Elken Manager, Performance Engineering QLogic Corporation Dr William Brantley Manager
More informationParallel Programming Models. Parallel Programming Models. Threads Model. Implementations 3/24/2014. Shared Memory Model (without threads)
Parallel Programming Models Parallel Programming Models Shared Memory (without threads) Threads Distributed Memory / Message Passing Data Parallel Hybrid Single Program Multiple Data (SPMD) Multiple Program
More informationPyConZA High Performance Computing with Python. Kevin Colville Python on large clusters with MPI
PyConZA 2012 High Performance Computing with Python Kevin Colville Python on large clusters with MPI Andy Rabagliati Python to read and store data on CHPC Petabyte data store www.chpc.ac.za High Performance
More informationFurther MPI Programming. Paul Burton April 2015
Further MPI Programming Paul Burton April 2015 Blocking v Non-blocking communication Blocking communication - Call to MPI sending routine does not return until the send buffer (array) is safe to use again
More informationDistributed Memory Programming With MPI (4)
Distributed Memory Programming With MPI (4) 2014 Spring Jinkyu Jeong (jinkyu@skku.edu) 1 Roadmap Hello World in MPI program Basic APIs of MPI Example program The Trapezoidal Rule in MPI. Collective communication.
More informationChapter 4: Threads. Operating System Concepts. Silberschatz, Galvin and Gagne
Chapter 4: Threads Silberschatz, Galvin and Gagne Chapter 4: Threads Overview Multithreading Models Thread Libraries Threading Issues Operating System Examples Linux Threads 4.2 Silberschatz, Galvin and
More informationGPU-based Distributed Behavior Models with CUDA
GPU-based Distributed Behavior Models with CUDA Courtesy: YouTube, ISIS Lab, Universita degli Studi di Salerno Bradly Alicea Introduction Flocking: Reynolds boids algorithm. * models simple local behaviors
More informationLarge Scale Debugging of Parallel Tasks with AutomaDeD!
International Conference for High Performance Computing, Networking, Storage and Analysis (SC) Seattle, Nov, 0 Large Scale Debugging of Parallel Tasks with AutomaDeD Ignacio Laguna, Saurabh Bagchi Todd
More informationExperiencing Cluster Computing Message Passing Interface
Experiencing Cluster Computing Message Passing Interface Class 6 Message Passing Paradigm The Underlying Principle A parallel program consists of p processes with different address spaces. Communication
More informationIntroduction to Parallel Programming with MPI
Introduction to Parallel Programming with MPI PICASso Tutorial October 25-26, 2006 Stéphane Ethier (ethier@pppl.gov) Computational Plasma Physics Group Princeton Plasma Physics Lab Why Parallel Computing?
More informationTolerating Message Latency through the Early Release of Blocked Receives
Tolerating Message Latency through the Early Release of Blocked Receives Jian Ke 1, Martin Burtscher 1, and Evan Speight 2 1 Computer Systems Laboratory, School of Electrical & Computer Engineering, Cornell
More informationAutomatic trace analysis with the Scalasca Trace Tools
Automatic trace analysis with the Scalasca Trace Tools Ilya Zhukov Jülich Supercomputing Centre Property Automatic trace analysis Idea Automatic search for patterns of inefficient behaviour Classification
More informationIntroduction to High-Performance Computing
Introduction to High-Performance Computing Dr. Axel Kohlmeyer Associate Dean for Scientific Computing, CST Associate Director, Institute for Computational Science Assistant Vice President for High-Performance
More informationHow to perform HPL on CPU&GPU clusters. Dr.sc. Draško Tomić
How to perform HPL on CPU&GPU clusters Dr.sc. Draško Tomić email: drasko.tomic@hp.com Forecasting is not so easy, HPL benchmarking could be even more difficult Agenda TOP500 GPU trends Some basics about
More informationMessage-Passing Computing
Chapter 2 Slide 41þþ Message-Passing Computing Slide 42þþ Basics of Message-Passing Programming using userlevel message passing libraries Two primary mechanisms needed: 1. A method of creating separate
More informationF# for Scientists. Jon Harrop Flying Frog Consultancy Ltd. Foreword by Don Syme A JOHN WILEY & SONS, INC., PUBLICATION WILEY
F# for Scientists Jon Harrop Flying Frog Consultancy Ltd. Foreword by Don Syme WILEY A JOHN WILEY & SONS, INC., PUBLICATION Preface Acknowledgments List of Figi ares List of Tables Acronyms 1 Introduction
More informationGhost Cell Pattern. Fredrik Berg Kjolstad. January 26, 2010
Ghost Cell Pattern Fredrik Berg Kjolstad University of Illinois Urbana-Champaign, USA kjolsta1@illinois.edu Marc Snir University of Illinois Urbana-Champaign, USA snir@illinois.edu January 26, 2010 Problem
More informationLS-DYNA Scalability Analysis on Cray Supercomputers
13 th International LS-DYNA Users Conference Session: Computing Technology LS-DYNA Scalability Analysis on Cray Supercomputers Ting-Ting Zhu Cray Inc. Jason Wang LSTC Abstract For the automotive industry,
More informationMM5 Modeling System Performance Research and Profiling. March 2009
MM5 Modeling System Performance Research and Profiling March 2009 Note The following research was performed under the HPC Advisory Council activities AMD, Dell, Mellanox HPC Advisory Council Cluster Center
More informationMetropolitan Road Traffic Simulation on FPGAs
Metropolitan Road Traffic Simulation on FPGAs Justin L. Tripp, Henning S. Mortveit, Anders Å. Hansson, Maya Gokhale Los Alamos National Laboratory Los Alamos, NM 85745 Overview Background Goals Using the
More informationChapter 4: Multi-Threaded Programming
Chapter 4: Multi-Threaded Programming Chapter 4: Threads 4.1 Overview 4.2 Multicore Programming 4.3 Multithreading Models 4.4 Thread Libraries Pthreads Win32 Threads Java Threads 4.5 Implicit Threading
More informationParallel Programming. Jin-Soo Kim Computer Systems Laboratory Sungkyunkwan University
Parallel Programming Jin-Soo Kim (jinsookim@skku.edu) Computer Systems Laboratory Sungkyunkwan University http://csl.skku.edu Challenges Difficult to write parallel programs Most programmers think sequentially
More informationScalable Software Components for Ultrascale Visualization Applications
Scalable Software Components for Ultrascale Visualization Applications Wes Kendall, Tom Peterka, Jian Huang SC Ultrascale Visualization Workshop 2010 11-15-2010 Primary Collaborators Jian Huang Tom Peterka
More informationTopics. Lecture 7. Review. Other MPI collective functions. Collective Communication (cont d) MPI Programming (III)
Topics Lecture 7 MPI Programming (III) Collective communication (cont d) Point-to-point communication Basic point-to-point communication Non-blocking point-to-point communication Four modes of blocking
More informationHigh Scalability of Lattice Boltzmann Simulations with Turbulence Models using Heterogeneous Clusters
SIAM PP 2014 High Scalability of Lattice Boltzmann Simulations with Turbulence Models using Heterogeneous Clusters C. Riesinger, A. Bakhtiari, M. Schreiber Technische Universität München February 20, 2014
More informationParallel Programming Using MPI
Parallel Programming Using MPI Gregory G. Howes Department of Physics and Astronomy University of Iowa Iowa High Performance Computing Summer School University of Iowa Iowa City, Iowa 6-8 June 2012 Thank
More informationParallel Computing and the MPI environment
Parallel Computing and the MPI environment Claudio Chiaruttini Dipartimento di Matematica e Informatica Centro Interdipartimentale per le Scienze Computazionali (CISC) Università di Trieste http://www.dmi.units.it/~chiarutt/didattica/parallela
More informationFall CSE 633 Parallel Algorithms. Cellular Automata. Nils Wisiol 11/13/12
Fall 2012 CSE 633 Parallel Algorithms Cellular Automata Nils Wisiol 11/13/12 Simple Automaton: Conway s Game of Life Simple Automaton: Conway s Game of Life John H. Conway Simple Automaton: Conway s Game
More informationAn MPI failure detector over PMPI 1
An MPI failure detector over PMPI 1 Donghoon Kim Department of Computer Science, North Carolina State University Raleigh, NC, USA Email : {dkim2}@ncsu.edu Abstract Fault Detectors are valuable services
More informationDPHPC Recitation Session 2 Advanced MPI Concepts
TIMO SCHNEIDER DPHPC Recitation Session 2 Advanced MPI Concepts Recap MPI is a widely used API to support message passing for HPC We saw that six functions are enough to write useful
More information30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy
Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy Why serial is not enough Computing architectures Parallel paradigms Message Passing Interface How
More information東京大学情報基盤中心准教授片桐孝洋 Takahiro Katagiri, Associate Professor, Information Technology Center, The University of Tokyo
Overview of MPI 東京大学情報基盤中心准教授片桐孝洋 Takahiro Katagiri, Associate Professor, Information Technology Center, The University of Tokyo 台大数学科学中心科学計算冬季学校 1 Agenda 1. Features of MPI 2. Basic MPI Functions 3. Reduction
More informationCS 426. Building and Running a Parallel Application
CS 426 Building and Running a Parallel Application 1 Task/Channel Model Design Efficient Parallel Programs (or Algorithms) Mainly for distributed memory systems (e.g. Clusters) Break Parallel Computations
More informationIntroduction to High-Performance Computing
Introduction to High-Performance Computing Simon D. Levy BIOL 274 17 November 2010 Chapter 12 12.1: Concurrent Processing High-Performance Computing A fancy term for computers significantly faster than
More informationMPI Case Study. Fabio Affinito. April 24, 2012
MPI Case Study Fabio Affinito April 24, 2012 In this case study you will (hopefully..) learn how to Use a master-slave model Perform a domain decomposition using ghost-zones Implementing a message passing
More informationAnna Morajko.
Performance analysis and tuning of parallel/distributed applications Anna Morajko Anna.Morajko@uab.es 26 05 2008 Introduction Main research projects Develop techniques and tools for application performance
More informationChapter 4: Threads. Chapter 4: Threads
Chapter 4: Threads Silberschatz, Galvin and Gagne 2013 Chapter 4: Threads Overview Multicore Programming Multithreading Models Thread Libraries Implicit Threading Threading Issues Operating System Examples
More informationEfficient Multi-GPU CUDA Linear Solvers for OpenFOAM
Efficient Multi-GPU CUDA Linear Solvers for OpenFOAM Alexander Monakov, amonakov@ispras.ru Institute for System Programming of Russian Academy of Sciences March 20, 2013 1 / 17 Problem Statement In OpenFOAM,
More informationBenchmarking Porting Costs of the SKYNET High Performance Signal Processing Middleware
Benchmarking Porting Costs of the SKYNET High Performance Signal Processing Middleware Michael J. Linnig and Gary R. Suder Engineering Fellows Sept 12, 2014 Linnig@Raytheon.com Gary_R_Suder@Raytheon.com
More informationHPC Parallel Programing Multi-node Computation with MPI - I
HPC Parallel Programing Multi-node Computation with MPI - I Parallelization and Optimization Group TATA Consultancy Services, Sahyadri Park Pune, India TCS all rights reserved April 29, 2013 Copyright
More informationOracle Database Competency Center
Oracle Database Competency Center Suchai Yenruedee Consulting & Customer Support Director Advanced Solutions Application Hosting Services Database Competency Center Space: 167.54 sqm. Location: 7th Floor
More informationMPI 5. CSCI 4850/5850 High-Performance Computing Spring 2018
MPI 5 CSCI 4850/5850 High-Performance Computing Spring 2018 Tae-Hyuk (Ted) Ahn Department of Computer Science Program of Bioinformatics and Computational Biology Saint Louis University Learning Objectives
More informationPerformance database technology for SciDAC applications
Performance database technology for SciDAC applications D Gunter 1, K Huck 2, K Karavanic 3, J May 4, A Malony 2, K Mohror 3, S Moore 5, A Morris 2, S Shende 2, V Taylor 6, X Wu 6, and Y Zhang 7 1 Lawrence
More information