Lightweight MPI Communicators with Applications to
|
|
- Mercy Melton
- 5 years ago
- Views:
Transcription
1 Lightweight MPI Communicators with Applications to Michael Axtmann, Peter Sanders, Armin Wiebigke IPDPS May 22, 2018 KIT The Research University in the Helmholtz Association
2 Overview Communicators and communication Disadvantages of communicator construction Solutions for MPI RBC communicators Case study on sorting 1
3 Communicators in MPI MPI COMM WORLD Subcommunicator 0 3; 1 4; Blocking point-to-point 0 1 Send Nonblocking point-to-point 0 1 ISend Compute and Test 2 Receive 2 IReceive and Test Blocking collective 0 1 Scan Nonblocking collective 0 1 IScan Compute and Test 2 Scan 2 IScan Test 2
4 MPI Examples Communication over rows and columns Divide and conquer Usage of communicators Divide tasks into fine-grained subproblems Elegant algorithms and comfortable programming Communicators make life easier at no cost!? 3
5 Current Implementations OpenMPI and MPICH PE group Mapping from PE ID to process ID required Explicit representation as table Subcommunicator 0 3; 1 4; Context ID Separates communication between communicators part of each message Unique for all PEs of the PE group Blocking Allgather-operation on context ID mask Subcommunicator Context ID 0 Subcommunicator Context ID 1 3/0 2/
6 Current Implementations OpenMPI and MPICH PE group Mapping from PE ID to process ID required Explicit representation as table Subcommunicator 0 3; 1 4; 2 1 Communicator creation takes time linear to the communicator size Context ID Separates communication between communicators part of each message Unique for all PEs of the PE group Blocking Allgather-operation on context ID mask Subcommunicator Context ID 0 Subcommunicator Context ID 1 3/0 2/
7 Current Implementations OpenMPI and MPICH PE group Mapping from PE ID to process ID required Explicit representation as table Subcommunicator 0 3; 1 4; 2 1 Communicator creation takes time linear to the communicator size Context ID Separates communication between communicators part of each message Unique for all PEs of the PE group Blocking Allgather-operation on context ID mask Subcommunicator Context ID 0 Subcommunicator Context ID 1 Communicator creation is a blocking collective operation 3/0 2/
8 Blocking Communicator Creation... nonblocking collective operations can mitigate possible synchronizing effects enabling communication-computation overlap perform collective operations on overlapping communicators, which would lead to deadlocks with blocking operations. MPI Standard A collective operation is invoked by all PEs of a communicator BUT: Communicator creation breaks nonblocking idea Nonblocking collective Nonblocking collective with communicator creation 0 IScan Compute and Test 0 Communicator creation IScanTest IScan Test 2 5
9 Communicator Construction Splitting a communicator into two communicators of half the size Splitting PE Running Time / Comm Size [µs] IBM MPI Comm create group IBM MPI Comm split Intel MPI Comm create group Intel MPI Comm split Comm Size (Cores) Communicator construction time is linear to PE group size 6 SuperMUC
10 Communicator Construction Splitting 2 15 PEs into two communicators of size 2 14 Collective operation on 2 14 cores Splitting Collective operation PE Running Time [ms] MPI_Reduce MPI_Exscan MPI_Comm_split Message Length [B] Communicator construction is expensive compared to collectives 7 SuperMUC cores IBM MPI
11 Communicator Construction Splitting a communicator into overlapping communicators of size four PE group invokest MPI Comm create group Alternating Cascading Time PE PE Running time [ms] Alternating Cascade Comm Size (Cores) Blocking communicator creation causes delays 8 SuperMUC Intel MPI
12 Proposals for MPI PE group Sparse representations E.g. MPI_Group_range_incl Subcommunicator f i r s t =1, l a s t =4, s t r i d e =2 0 1 Context ID User-defined tag Calculate by MPI: Concatenation of counters Subcommunicator Context ID 0 Subcommunicator Context ID 1 3/0 2/ MPI COMM WORLD {0} {1} {2} {3} {0,0} {0,1} {2,0} {2,1} {2,2} 9
13 Our RBC library Range-based communicator in O(1) time Local construction Select MPI or RBC operations Local splitting: Split_RBC_Comm(Comm&, Comm&, tttint first, int last, int stride) Only adjust range Blocking Ops Nonblocking Ops Classes Local Ops rbc::bcast rbc::ibcast rbc::request rbc::create RBC Comm rbc::reduce rbc::ireduce rbc::comm rbc::split RBC Comm rbc::allreduce rbc::iallreduce RBC::Comm rank rbc::scan rbc::iscan rbc::comm size rbc::gather rbc::igather rbc::gatherv rbc::igatherv rbc::barrier rbc::ibarrier rbc::send rbc::isend rbc::recv rbc::irecv rbc::probe rbc::iprobe rbc::wait rbc::test rbc::waitall rbc::comm + parent comm : MPI Comm + first : int + last : int + stride : int Initial MPI communicator f i r s t =1, l a s t =7, s t r i d e =3 10
14 Our RBC library Implementation Details (Non)blocking point-to-point communication Maps rank to rank of MPI communicator Call MPI counterpart (Non)blocking collective operations Calls point-to-point operations of RBC One globally reserved tag Nonblocking details Optional user-defined tag Round-based schedule rbc::ibcast(void *buff, int cnt, MPI_Datatype datatype, int root, tttrbc::comm comm, rbc::request *request, int tag = RBC_IBCAST_TAG) Time Broadcast PE 11
15 RBC vs. MPI Splitting a communicator into two communicators of half the size Splitting PE Running Time / Comm Size [µs] IBM MPI Comm create group IBM MPI Comm split Intel MPI Comm create group Intel MPI Comm split RBC rbc::split_rbc_comm Comm Size (Cores) RBC splitting comes with almost no cost 12 SuperMUC
16 RBC vs. MPI Splitting a communicator with 2 15 PEs into two communicators of size 2 14 and performing a broadcast operation on both communicators PE Splitting Collective operation Running time [ms] Message Length [B] RBC splitting comes with almost no cost IBM MPI Ibcast IBM rbc::ibcast 13 SuperMUC cores IBM MPI
17 Cost of Communicators Splitting a communicator into overlapping communicators of size four PE group invokest MPI Comm create group Alternating Cascading Time PE PE Running time [ms] MPI Alternating RBC Cascade MPI Cascade RBC Alternating Comm Size (Cores) Cascades do not have an effect on RBC 14 SuperMUC Intel MPI
18 Quicksort on Distributed Systems Single ported message passing Sending of l machine words: α + βl Analyze critical path Small inputs Minimal latency O(α log 2 p + β n p log p + n p log n) Sort locally Splitter selection PE 1 PE 2 PE 3 PE 4 Hypercube Quicksort Static communication pattern Precomputable communicators Bad for skewed inputs Only works for p = 2 k Split Redistribution Merge Recurse Subcube 1 Subcube 2 15
19 Janus Sort O(α log 2 p + βp n log p + p n log n) Arbitrary p Calculate communicators on the fly Perfectly balanced data Roman God Janus Twofaced 1 PE 0 PE 1 PE 2 PE 3 PE 4 PE 5 Local sort Recursion step Group 1 Group 2 16 Image taken by Fubar Obfusco
20 Janus Sort group j group j + 1 PE i Janus PE PE i + 2 PE i + 3 PE i + 4 execution 17
21 Janus Sort group j group j + 1 PE i Janus PE PE i + 2 PE i + 3 PE i + 4 execution 17
22 Janus Sort group j group j + 1 PE i Janus PE PE i + 2 PE i + 3 PE i + 4 execution partitioning partitioning 17
23 Janus Sort group j group j + 1 PE i Janus PE PE i + 2 PE i + 3 PE i + 4 execution partitioning data exchange partitioning 17
24 Janus Sort group j group j + 1 PE i Janus PE PE i + 2 PE i + 3 PE i + 4 execution partitioning data exchange partitioning comm creation comm creation Janus PE Janus PE PE i + 2 Janus PE PE i + 4 base case group k group k
25 Janus Sort group j group j + 1 PE i Janus PE PE i + 2 PE i + 3 PE i + 4 execution partitioning data exchange partitioning comm creation comm creation Janus PE Janus PE PE i + 2 Janus PE PE i + 4 base case group k group k + 1 Pivot selection: O(α log p) Regular PE IBcast Wait Left group IBcast Janus PE Waitall Right group IBcast 17
26 Janus Sort group j group j + 1 PE i Janus PE PE i + 2 PE i + 3 PE i + 4 execution partitioning data exchange partitioning comm creation comm creation Janus PE Janus PE PE i + 2 Janus PE PE i + 4 base case group k group k + 1 Pivot selection: O(α log p) Regular PE Left group Janus PE Right group Partitioning: O(log n p ) Binary search Binary search Binary search 17
27 Janus Sort group j group j + 1 PE i Janus PE PE i + 2 PE i + 3 PE i + 4 execution partitioning data exchange partitioning comm creation comm creation Janus PE Janus PE PE i + 2 Janus PE PE i + 4 base case group k group k + 1 Pivot selection: O(α log p) Partitioning: O(log n p ) Data exchange: O(α + β n p ) Regular PE IScan Wait IBcast Wait ISend IRecv Waitall Left group IScan IBcast ISend IRecv Janus PE Waitall Waitall Waitall Right group IScan IBcast ISend IRecv 17
28 Janus Sort group j group j + 1 PE i Janus PE PE i + 2 PE i + 3 PE i + 4 execution partitioning data exchange partitioning comm creation comm creation Janus PE Janus PE PE i + 2 Janus PE PE i + 4 base case group k group k + 1 Pivot selection: O(α log p) Regular PE Left group New Janus PE Right group Partitioning: O(log n p ) Data exchange: O(α + β n p ) Comm creation: O(1) Create comm Creation of even subcommunicator Creation of odd subcommunicator 17
29 Janus Sort Experimental Results Sorting of double values cores Weak scaling on the SuperMUC double values per core Juqueen RBC Split MPI Split SuperMUC RBC Split MPI Split RBC Split MPI Split Run time [s] Run Time [s] Elements per core Cores 18 IBM MPI
30 Conclusion Communicator creation is expensive Blocking communicator creation breaks idea of nonblocking communication Extension for MPI Range-based communicators Case study on sorting Code published at 19
Lecture 3: Sorting 1
Lecture 3: Sorting 1 Sorting Arranging an unordered collection of elements into monotonically increasing (or decreasing) order. S = a sequence of n elements in arbitrary order After sorting:
More informationWe can use a max-heap to sort data.
Sorting 7B N log N Sorts 1 Heap Sort We can use a max-heap to sort data. Convert an array to a max-heap. Remove the root from the heap and store it in its proper position in the same array. Repeat until
More informationParallel Systems Course: Chapter VIII. Sorting Algorithms. Kumar Chapter 9. Jan Lemeire ETRO Dept. November Parallel Sorting
Parallel Systems Course: Chapter VIII Sorting Algorithms Kumar Chapter 9 Jan Lemeire ETRO Dept. November 2014 Overview 1. Parallel sort distributed memory 2. Parallel sort shared memory 3. Sorting Networks
More informationAlgorithms and Applications
Algorithms and Applications 1 Areas done in textbook: Sorting Algorithms Numerical Algorithms Image Processing Searching and Optimization 2 Chapter 10 Sorting Algorithms - rearranging a list of numbers
More informationLecture 16: Sorting. CS178: Programming Parallel and Distributed Systems. April 2, 2001 Steven P. Reiss
Lecture 16: Sorting CS178: Programming Parallel and Distributed Systems April 2, 2001 Steven P. Reiss I. Overview A. Before break we started talking about parallel computing and MPI 1. Basic idea of a
More informationLast Time. Intro to Parallel Algorithms. Parallel Search Parallel Sorting. Merge sort Sample sort
Intro to MPI Last Time Intro to Parallel Algorithms Parallel Search Parallel Sorting Merge sort Sample sort Today Network Topology Communication Primitives Message Passing Interface (MPI) Randomized Algorithms
More informationCopyright The McGraw-Hill Companies, Inc. Permission required for reproduction or display. Chapter 9
Chapter 9 Document Classification Document Classification Problem Search directories, subdirectories for documents (look for.html,.txt,.tex, etc.) Using a dictionary of key words, create a profile vector
More informationParallel Systems Course: Chapter VIII. Sorting Algorithms. Kumar Chapter 9. Jan Lemeire ETRO Dept. Fall Parallel Sorting
Parallel Systems Course: Chapter VIII Sorting Algorithms Kumar Chapter 9 Jan Lemeire ETRO Dept. Fall 2017 Overview 1. Parallel sort distributed memory 2. Parallel sort shared memory 3. Sorting Networks
More informationSorting. Sorting in Arrays. SelectionSort. SelectionSort. Binary search works great, but how do we create a sorted array in the first place?
Sorting Binary search works great, but how do we create a sorted array in the first place? Sorting in Arrays Sorting algorithms: Selection sort: O(n 2 ) time Merge sort: O(nlog 2 (n)) time Quicksort: O(n
More informationParallel Programming with MPI and OpenMP
Parallel Programming with MPI and OpenMP Michael J. Quinn Chapter 6 Floyd s Algorithm Chapter Objectives Creating 2-D arrays Thinking about grain size Introducing point-to-point communications Reading
More informationIntroduction to MPI. May 20, Daniel J. Bodony Department of Aerospace Engineering University of Illinois at Urbana-Champaign
Introduction to MPI May 20, 2013 Daniel J. Bodony Department of Aerospace Engineering University of Illinois at Urbana-Champaign Top500.org PERFORMANCE DEVELOPMENT 1 Eflop/s 162 Pflop/s PROJECTED 100 Pflop/s
More informationMPI Optimisation. Advanced Parallel Programming. David Henty, Iain Bethune, Dan Holmes EPCC, University of Edinburgh
MPI Optimisation Advanced Parallel Programming David Henty, Iain Bethune, Dan Holmes EPCC, University of Edinburgh Overview Can divide overheads up into four main categories: Lack of parallelism Load imbalance
More informationIntroduction to Parallel Computing
Introduction to Parallel Computing George Karypis Sorting Outline Background Sorting Networks Quicksort Bucket-Sort & Sample-Sort Background Input Specification Each processor has n/p elements A ordering
More informationLecture 16: Recapitulations. Lecture 16: Recapitulations p. 1
Lecture 16: Recapitulations Lecture 16: Recapitulations p. 1 Parallel computing and programming in general Parallel computing a form of parallel processing by utilizing multiple computing units concurrently
More informationLecture 8 Parallel Algorithms II
Lecture 8 Parallel Algorithms II Dr. Wilson Rivera ICOM 6025: High Performance Computing Electrical and Computer Engineering Department University of Puerto Rico Original slides from Introduction to Parallel
More informationWith regards to bitonic sorts, everything boils down to doing bitonic sorts efficiently.
Lesson 2 5 Distributed Memory Sorting Bitonic Sort The pseudo code for generating a bitonic sequence: 1. create two bitonic subsequences 2. Make one an increasing sequence 3. Make one into a decreasing
More informationQuickSort
QuickSort 7 4 9 6 2 2 4 6 7 9 4 2 2 4 7 9 7 9 2 2 9 9 1 QuickSort QuickSort on an input sequence S with n elements consists of three steps: n n n Divide: partition S into two sequences S 1 and S 2 of about
More informationSorting Algorithms. Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar
Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text ``Introduction to Parallel Computing'', Addison Wesley, 2003. Topic Overview Issues in Sorting on Parallel
More informationDivide and Conquer Algorithms. Sathish Vadhiyar
Divide and Conquer Algorithms Sathish Vadhiyar Introduction One of the important parallel algorithm models The idea is to decompose the problem into parts solve the problem on smaller parts find the global
More informationSorting and Selection
Sorting and Selection Introduction Divide and Conquer Merge-Sort Quick-Sort Radix-Sort Bucket-Sort 10-1 Introduction Assuming we have a sequence S storing a list of keyelement entries. The key of the element
More informationAll-Pairs Shortest Paths - Floyd s Algorithm
All-Pairs Shortest Paths - Floyd s Algorithm Parallel and Distributed Computing Department of Computer Science and Engineering (DEI) Instituto Superior Técnico October 31, 2011 CPD (DEI / IST) Parallel
More informationThe Message Passing Interface (MPI) TMA4280 Introduction to Supercomputing
The Message Passing Interface (MPI) TMA4280 Introduction to Supercomputing NTNU, IMF January 16. 2017 1 Parallelism Decompose the execution into several tasks according to the work to be done: Function/Task
More informationCSC630/CSC730 Parallel & Distributed Computing
CSC630/CSC730 Parallel & Distributed Computing Parallel Sorting Chapter 9 1 Contents General issues Sorting network Bitonic sort Bubble sort and its variants Odd-even transposition Quicksort Other Sorting
More informationCollective Communication in MPI and Advanced Features
Collective Communication in MPI and Advanced Features Pacheco s book. Chapter 3 T. Yang, CS240A. Part of slides from the text book, CS267 K. Yelick from UC Berkeley and B. Gropp, ANL Outline Collective
More informationCSC 447: Parallel Programming for Multi- Core and Cluster Systems
CSC 447: Parallel Programming for Multi- Core and Cluster Systems Parallel Sorting Algorithms Instructor: Haidar M. Harmanani Spring 2016 Topic Overview Issues in Sorting on Parallel Computers Sorting
More informationCSL 860: Modern Parallel
CSL 860: Modern Parallel Computation PARALLEL ALGORITHM TECHNIQUES: BALANCED BINARY TREE Reduction n operands => log n steps Total work = O(n) How do you map? Balance Binary tree technique Reduction n
More informationParallel Algorithms for (PRAM) Computers & Some Parallel Algorithms. Reference : Horowitz, Sahni and Rajasekaran, Computer Algorithms
Parallel Algorithms for (PRAM) Computers & Some Parallel Algorithms Reference : Horowitz, Sahni and Rajasekaran, Computer Algorithms Part 2 1 3 Maximum Selection Problem : Given n numbers, x 1, x 2,, x
More informationMergesort again. 1. Split the list into two equal parts
Quicksort Mergesort again 1. Split the list into two equal parts 5 3 9 2 8 7 3 2 1 4 5 3 9 2 8 7 3 2 1 4 Mergesort again 2. Recursively mergesort the two parts 5 3 9 2 8 7 3 2 1 4 2 3 5 8 9 1 2 3 4 7 Mergesort
More informationLecture 19 Sorting Goodrich, Tamassia
Lecture 19 Sorting 7 2 9 4 2 4 7 9 7 2 2 7 9 4 4 9 7 7 2 2 9 9 4 4 2004 Goodrich, Tamassia Outline Review 3 simple sorting algorithms: 1. selection Sort (in previous course) 2. insertion Sort (in previous
More informationSorting Goodrich, Tamassia Sorting 1
Sorting Put array A of n numbers in increasing order. A core algorithm with many applications. Simple algorithms are O(n 2 ). Optimal algorithms are O(n log n). We will see O(n) for restricted input in
More informationDocument Classification Problem
Document Classification Problem Search directories, subdirectories for documents (look for.html,.txt,.tex, etc.) Using a dictionary of key words, create a profile vector for each document Store profile
More informationPart - II. Message Passing Interface. Dheeraj Bhardwaj
Part - II Dheeraj Bhardwaj Department of Computer Science & Engineering Indian Institute of Technology, Delhi 110016 India http://www.cse.iitd.ac.in/~dheerajb 1 Outlines Basics of MPI How to compile and
More informationData structures. More sorting. Dr. Alex Gerdes DIT961 - VT 2018
Data structures More sorting Dr. Alex Gerdes DIT961 - VT 2018 Divide and conquer Very general name for a type of recursive algorithm You have a problem to solve: - Split that problem into smaller subproblems
More informationParallel Programming
Parallel Programming for Multicore and Cluster Systems von Thomas Rauber, Gudula Rünger 1. Auflage Parallel Programming Rauber / Rünger schnell und portofrei erhältlich bei beck-shop.de DIE FACHBUCHHANDLUNG
More informationHPC Parallel Programing Multi-node Computation with MPI - I
HPC Parallel Programing Multi-node Computation with MPI - I Parallelization and Optimization Group TATA Consultancy Services, Sahyadri Park Pune, India TCS all rights reserved April 29, 2013 Copyright
More informationLecture 8: Mergesort / Quicksort Steven Skiena
Lecture 8: Mergesort / Quicksort Steven Skiena Department of Computer Science State University of New York Stony Brook, NY 11794 4400 http://www.cs.stonybrook.edu/ skiena Problem of the Day Give an efficient
More informationECE 574 Cluster Computing Lecture 13
ECE 574 Cluster Computing Lecture 13 Vince Weaver http://web.eece.maine.edu/~vweaver vincent.weaver@maine.edu 21 March 2017 Announcements HW#5 Finally Graded Had right idea, but often result not an *exact*
More informationQuick-Sort. Quick-Sort 1
Quick-Sort 7 4 9 6 2 2 4 6 7 9 4 2 2 4 7 9 7 9 2 2 9 9 Quick-Sort 1 Outline and Reading Quick-sort ( 4.3) Algorithm Partition step Quick-sort tree Execution example Analysis of quick-sort (4.3.1) In-place
More informationProgramming II (CS300)
1 Programming II (CS300) Chapter 12: Sorting Algorithms MOUNA KACEM mouna@cs.wisc.edu Spring 2018 Outline 2 Last week Implementation of the three tree depth-traversal algorithms Implementation of the BinarySearchTree
More informationReview of MPI Part 2
Review of MPI Part Russian-German School on High Performance Computer Systems, June, 7 th until July, 6 th 005, Novosibirsk 3. Day, 9 th of June, 005 HLRS, University of Stuttgart Slide Chap. 5 Virtual
More informationInformation Coding / Computer Graphics, ISY, LiTH
Sorting on GPUs Revisiting some algorithms from lecture 6: Some not-so-good sorting approaches Bitonic sort QuickSort Concurrent kernels and recursion Adapt to parallel algorithms Many sorting algorithms
More informationQuick Sort. CSE Data Structures May 15, 2002
Quick Sort CSE 373 - Data Structures May 15, 2002 Readings and References Reading Section 7.7, Data Structures and Algorithm Analysis in C, Weiss Other References C LR 15-May-02 CSE 373 - Data Structures
More informationLecture 6 Sequences II. Parallel and Sequential Data Structures and Algorithms, (Fall 2013) Lectured by Danny Sleator 12 September 2013
Lecture 6 Sequences II Parallel and Sequential Data Structures and Algorithms, 15-210 (Fall 2013) Lectured by Danny Sleator 12 September 2013 Material in this lecture: Today s lecture is about reduction.
More informationCSL 730: Parallel Programming
CSL 73: Parallel Programming General Algorithmic Techniques Balance binary tree Partitioning Divid and conquer Fractional cascading Recursive doubling Symmetry breaking Pipelining 2 PARALLEL ALGORITHM
More informationOptimization of MPI Applications Rolf Rabenseifner
Optimization of MPI Applications Rolf Rabenseifner University of Stuttgart High-Performance Computing-Center Stuttgart (HLRS) www.hlrs.de Optimization of MPI Applications Slide 1 Optimization and Standardization
More informationDocument Classification
Document Classification Introduction Search engine on web Search directories, subdirectories for documents Search for documents with extensions.html,.txt, and.tex Using a dictionary of key words, create
More informationMessage Passing Interface - MPI
Message Passing Interface - MPI Parallel and Distributed Computing Department of Computer Science and Engineering (DEI) Instituto Superior Técnico October 24, 2011 Many slides adapted from lectures by
More informationParallel Programming in C with MPI and OpenMP
Parallel Programming in C with MPI and OpenMP Michael J. Quinn Chapter 9 Document Classification Chapter Objectives Complete introduction of MPI functions Show how to implement manager-worker programs
More informationI/O Model. Cache-Oblivious Algorithms : Algorithms in the Real World. Advantages of Cache-Oblivious Algorithms 4/9/13
I/O Model 15-853: Algorithms in the Real World Locality II: Cache-oblivious algorithms Matrix multiplication Distribution sort Static searching Abstracts a single level of the memory hierarchy Fast memory
More informationMessage Passing Interface - MPI
Message Passing Interface - MPI Parallel and Distributed Computing Department of Computer Science and Engineering (DEI) Instituto Superior Técnico March 31, 2016 Many slides adapted from lectures by Bill
More informationPresentation for use with the textbook, Algorithm Design and Applications, by M. T. Goodrich and R. Tamassia, Wiley, 2015
Presentation for use with the textbook, Algorithm Design and Applications, by M. T. Goodrich and R. Tamassia, Wiley, 2015 Quick-Sort 7 4 9 6 2 2 4 6 7 9 4 2 2 4 7 9 7 9 2 2 9 9 2015 Goodrich and Tamassia
More informationimmediate send and receive query the status of a communication MCS 572 Lecture 8 Introduction to Supercomputing Jan Verschelde, 9 September 2016
Data Partitioning 1 Data Partitioning functional and domain decomposition 2 Parallel Summation applying divide and conquer fanning out an array of data fanning out with MPI fanning in the results 3 An
More informationBetter sorting algorithms (Weiss chapter )
Better sorting algorithms (Weiss chapter 8.5 8.6) Divide and conquer Very general name for a type of recursive algorithm You have a problem to solve. Split that problem into smaller subproblems Recursively
More informationQuick-Sort fi fi fi 7 9. Quick-Sort Goodrich, Tamassia
Quick-Sort 7 4 9 6 2 fi 2 4 6 7 9 4 2 fi 2 4 7 9 fi 7 9 2 fi 2 9 fi 9 Quick-Sort 1 Quick-Sort ( 10.2 text book) Quick-sort is a randomized sorting algorithm based on the divide-and-conquer paradigm: x
More informationCS102 Sorting - Part 2
CS102 Sorting - Part 2 Prof Tejada 1 Types of Sorts Incremental Approach Bubble Sort, Selection Sort, Insertion Sort, etc. Work slowly toward solution one step at a time Generally iterative in nature Divide
More informationLecture 23: Parallelism. To Use Library. Getting Good Results. CS 62 Fall 2016 Kim Bruce & Peter Mawhorter
Lecture 23: Parallelism CS 62 Fall 2016 Kim Bruce & Peter Mawhorter Some slides based on those from Dan Grossman, U. of Washington To Use Library Getting Good Results Create a ForkJoinPool Instead of subclass
More informationCSE 160 Lecture 13. Non-blocking Communication Under the hood of MPI Performance Stencil Methods in MPI
CSE 160 Lecture 13 Non-blocking Communication Under the hood of MPI Performance Stencil Methods in MPI Announcements Scott B. Baden /CSE 260/ Winter 2014 2 Today s lecture Asynchronous non-blocking, point
More informationSorting (Chapter 9) Alexandre David B2-206
Sorting (Chapter 9) Alexandre David B2-206 Sorting Problem Arrange an unordered collection of elements into monotonically increasing (or decreasing) order. Let S = . Sort S into S =
More informationHighly Scalable Parallel Sorting. Edgar Solomonik and Laxmikant Kale University of Illinois at Urbana-Champaign April 20, 2010
Highly Scalable Parallel Sorting Edgar Solomonik and Laxmikant Kale University of Illinois at Urbana-Champaign April 20, 2010 1 Outline Parallel sorting background Histogram Sort overview Histogram Sort
More informationCISC 235: Topic 4. Balanced Binary Search Trees
CISC 235: Topic 4 Balanced Binary Search Trees Outline Rationale and definitions Rotations AVL Trees, Red-Black, and AA-Trees Algorithms for searching, insertion, and deletion Analysis of complexity CISC
More informationUnit-2 Divide and conquer 2016
2 Divide and conquer Overview, Structure of divide-and-conquer algorithms, binary search, quick sort, Strassen multiplication. 13% 05 Divide-and- conquer The Divide and Conquer Paradigm, is a method of
More informationParallel Sorting. Sathish Vadhiyar
Parallel Sorting Sathish Vadhiyar Parallel Sorting Problem The input sequence of size N is distributed across P processors The output is such that elements in each processor P i is sorted elements in P
More informationUnit 6 Chapter 15 EXAMPLES OF COMPLEXITY CALCULATION
DESIGN AND ANALYSIS OF ALGORITHMS Unit 6 Chapter 15 EXAMPLES OF COMPLEXITY CALCULATION http://milanvachhani.blogspot.in EXAMPLES FROM THE SORTING WORLD Sorting provides a good set of examples for analyzing
More informationContents. Preface xvii Acknowledgments. CHAPTER 1 Introduction to Parallel Computing 1. CHAPTER 2 Parallel Programming Platforms 11
Preface xvii Acknowledgments xix CHAPTER 1 Introduction to Parallel Computing 1 1.1 Motivating Parallelism 2 1.1.1 The Computational Power Argument from Transistors to FLOPS 2 1.1.2 The Memory/Disk Speed
More informationCSE 421 Greedy Alg: Union Find/Dijkstra s Alg
CSE 1 Greedy Alg: Union Find/Dijkstra s Alg Shayan Oveis Gharan 1 Dijkstra s Algorithm Dijkstra(G, c, s) { d s 0 foreach (v V) d[v] //This is the key of node v foreach (v V) insert v onto a priority queue
More informationAgenda. MPI Application Example. Praktikum: Verteiltes Rechnen und Parallelprogrammierung Introduction to MPI. 1) Recap: MPI. 2) 2.
Praktikum: Verteiltes Rechnen und Parallelprogrammierung Introduction to MPI Agenda 1) Recap: MPI 2) 2. Übungszettel 3) Projektpräferenzen? 4) Nächste Woche: 3. Übungszettel, Projektauswahl, Konzepte 5)
More informationParallel Programming. Functional Decomposition (Document Classification)
Parallel Programming Functional Decomposition (Document Classification) Document Classification Problem Search directories, subdirectories for text documents (look for.html,.txt,.tex, etc.) Using a dictionary
More informationThrill : High-Performance Algorithmic
Thrill : High-Performance Algorithmic Distributed Batch Data Processing in C++ Timo Bingmann, Michael Axtmann, Peter Sanders, Sebastian Schlag, and 6 Students 2016-12-06 INSTITUTE OF THEORETICAL INFORMATICS
More informationMarch 3, George Mason University Sorting Networks. Indranil Banerjee. Parallel Sorting: Hardware Level Parallelism
Sorting George Mason University ibanerje@gmu.edu March 3, 2016 Sorting GMU March 3, 2016 1 / 19 There are mainly two approaches to sorting in parallel: 1 Non-oblivious: Comparisons are data dependent Example:
More informationParallel programming MPI
Parallel programming MPI Distributed memory Each unit has its own memory space If a unit needs data in some other memory space, explicit communication (often through network) is required Point-to-point
More informationSorting (Chapter 9) Alexandre David B2-206
Sorting (Chapter 9) Alexandre David B2-206 1 Sorting Problem Arrange an unordered collection of elements into monotonically increasing (or decreasing) order. Let S = . Sort S into S =
More informationProblem. Input: An array A = (A[1],..., A[n]) with length n. Output: a permutation A of A, that is sorted: A [i] A [j] for all. 1 i j n.
Problem 5. Sorting Simple Sorting, Quicksort, Mergesort Input: An array A = (A[1],..., A[n]) with length n. Output: a permutation A of A, that is sorted: A [i] A [j] for all 1 i j n. 98 99 Selection Sort
More informationCSE 530A. B+ Trees. Washington University Fall 2013
CSE 530A B+ Trees Washington University Fall 2013 B Trees A B tree is an ordered (non-binary) tree where the internal nodes can have a varying number of child nodes (within some range) B Trees When a key
More informationDistributed Systems + Middleware Advanced Message Passing with MPI
Distributed Systems + Middleware Advanced Message Passing with MPI Gianpaolo Cugola Dipartimento di Elettronica e Informazione Politecnico, Italy cugola@elet.polimi.it http://home.dei.polimi.it/cugola
More informationCME 194 Introduc0on to MPI
CME 194 Introduc0on to MPI Essen0a Callidus h8p://cme194.stanford.edu Recap Last class: Communicators & Derived Datatypes Communica0on between arbitrary subsets of processes Grid style communica0on Communicate
More informationMerge Sort Goodrich, Tamassia Merge Sort 1
Merge Sort 7 2 9 4 2 4 7 9 7 2 2 7 9 4 4 9 7 7 2 2 9 9 4 4 2004 Goodrich, Tamassia Merge Sort 1 Review of Sorting Selection-sort: Search: search through remaining unsorted elements for min Remove: remove
More informationCS 6230: High-Performance Computing and Parallelization Introduction to MPI
CS 6230: High-Performance Computing and Parallelization Introduction to MPI Dr. Mike Kirby School of Computing and Scientific Computing and Imaging Institute University of Utah Salt Lake City, UT, USA
More informationChapter 12: Indexing and Hashing. Basic Concepts
Chapter 12: Indexing and Hashing! Basic Concepts! Ordered Indices! B+-Tree Index Files! B-Tree Index Files! Static Hashing! Dynamic Hashing! Comparison of Ordered Indexing and Hashing! Index Definition
More informationDivide-and-Conquer. Dr. Yingwu Zhu
Divide-and-Conquer Dr. Yingwu Zhu Divide-and-Conquer The most-well known algorithm design technique: 1. Divide instance of problem into two or more smaller instances 2. Solve smaller instances independently
More informationJohn Mellor-Crummey Department of Computer Science Rice University
Parallel Sorting John Mellor-Crummey Department of Computer Science Rice University johnmc@rice.edu COMP 422/534 Lecture 23 6 April 2017 Topics for Today Introduction Sorting networks and Batcher s bitonic
More informationBasic MPI Communications. Basic MPI Communications (cont d)
Basic MPI Communications MPI provides two non-blocking routines: MPI_Isend(buf,cnt,type,dst,tag,comm,reqHandle) buf: source of data to be sent cnt: number of data elements to be sent type: type of each
More informationChapter 12: Indexing and Hashing
Chapter 12: Indexing and Hashing Basic Concepts Ordered Indices B+-Tree Index Files B-Tree Index Files Static Hashing Dynamic Hashing Comparison of Ordered Indexing and Hashing Index Definition in SQL
More informationDistributed Memory Programming with MPI
Distributed Memory Programming with MPI Moreno Marzolla Dip. di Informatica Scienza e Ingegneria (DISI) Università di Bologna moreno.marzolla@unibo.it Algoritmi Avanzati--modulo 2 2 Credits Peter Pacheco,
More informationChapter 12: Indexing and Hashing (Cnt(
Chapter 12: Indexing and Hashing (Cnt( Cnt.) Basic Concepts Ordered Indices B+-Tree Index Files B-Tree Index Files Static Hashing Dynamic Hashing Comparison of Ordered Indexing and Hashing Index Definition
More informationCOMP 322: Fundamentals of Parallel Programming. Lecture 34: Introduction to the Message Passing Interface (MPI), contd
COMP 322: Fundamentals of Parallel Programming Lecture 34: Introduction to the Message Passing Interface (MPI), contd Vivek Sarkar, Eric Allen Department of Computer Science, Rice University Contact email:
More informationLecture 9: MPI continued
Lecture 9: MPI continued David Bindel 27 Sep 2011 Logistics Matrix multiply is done! Still have to run. Small HW 2 will be up before lecture on Thursday, due next Tuesday. Project 2 will be posted next
More informationOutline. Communication modes MPI Message Passing Interface Standard
MPI THOAI NAM Outline Communication modes MPI Message Passing Interface Standard TERMs (1) Blocking If return from the procedure indicates the user is allowed to reuse resources specified in the call Non-blocking
More informationCAD Algorithms. Categorizing Algorithms
CAD Algorithms Categorizing Algorithms Mohammad Tehranipoor ECE Department 2 September 2008 1 Categorizing Algorithms Greedy Algorithms Prim s Algorithm (Minimum Spanning Tree) A subgraph that is a tree
More informationMergesort again. 1. Split the list into two equal parts
Quicksort Mergesort again 1. Split the list into two equal parts 5 3 9 2 8 7 3 2 1 4 5 3 9 2 8 7 3 2 1 4 Mergesort again 2. Recursively mergesort the two parts 5 3 9 2 8 7 3 2 1 4 2 3 5 8 9 1 2 3 4 7 Mergesort
More informationWeek 10. Sorting. 1 Binary heaps. 2 Heapification. 3 Building a heap 4 HEAP-SORT. 5 Priority queues 6 QUICK-SORT. 7 Analysing QUICK-SORT.
Week 10 1 Binary s 2 3 4 5 6 Sorting Binary s 7 8 General remarks Binary s We return to sorting, considering and. Reading from CLRS for week 7 1 Chapter 6, Sections 6.1-6.5. 2 Chapter 7, Sections 7.1,
More informationCopyright The McGraw-Hill Companies, Inc. Permission required for reproduction or display. Chapter 8
Chapter 8 Matrix-vector Multiplication Chapter Objectives Review matrix-vector multiplicaiton Propose replication of vectors Develop three parallel programs, each based on a different data decomposition
More informationChapter 1 Divide and Conquer Algorithm Theory WS 2014/15 Fabian Kuhn
Chapter 1 Divide and Conquer Algorithm Theory WS 2014/15 Fabian Kuhn Divide And Conquer Principle Important algorithm design method Examples from Informatik 2: Sorting: Mergesort, Quicksort Binary search
More informationQuick-Sort. Quick-sort is a randomized sorting algorithm based on the divide-and-conquer paradigm:
Presentation for use with the textbook Data Structures and Algorithms in Java, 6 th edition, by M. T. Goodrich, R. Tamassia, and M. H. Goldwasser, Wiley, 2014 Quick-Sort 7 4 9 6 2 2 4 6 7 9 4 2 2 4 7 9
More informationAlgorithm classification
Types of Algorithms Algorithm classification Algorithms that use a similar problem-solving approach can be grouped together We ll talk about a classification scheme for algorithms This classification scheme
More informationChapter 4. Divide-and-Conquer. Copyright 2007 Pearson Addison-Wesley. All rights reserved.
Chapter 4 Divide-and-Conquer Copyright 2007 Pearson Addison-Wesley. All rights reserved. Divide-and-Conquer The most-well known algorithm design strategy: 2. Divide instance of problem into two or more
More informationCOMP 322: Fundamentals of Parallel Programming
COMP 322: Fundamentals of Parallel Programming https://wiki.rice.edu/confluence/display/parprog/comp322 Lecture 37: Introduction to MPI (contd) Vivek Sarkar Department of Computer Science Rice University
More informationLesson 1 4. Prefix Sum Definitions. Scans. Parallel Scans. A Naive Parallel Scans
Lesson 1 4 Prefix Sum Definitions Prefix sum given an array...the prefix sum is the sum of all the elements in the array from the beginning to the position, including the value at the position. The sequential
More informationPresentation for use with the textbook, Algorithm Design and Applications, by M. T. Goodrich and R. Tamassia, Wiley, 2015
Presentation for use with the textbook, Algorithm Design and Applications, by M. T. Goodrich and R. Tamassia, Wiley, 2015 Merge Sort 2015 Goodrich and Tamassia Merge Sort 1 Application: Internet Search
More informationUsing Lamport s Logical Clocks
Fast Classification of MPI Applications Using Lamport s Logical Clocks Zhou Tong, Scott Pakin, Michael Lang, Xin Yuan Florida State University Los Alamos National Laboratory 1 Motivation Conventional trace-based
More informationSAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 6. Sorting Algorithms
SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 6 6.0 Introduction Sorting algorithms used in computer science are often classified by: Computational complexity (worst, average and best behavior) of element
More information