Co-arrays to be included in the Fortran 2008 Standard
|
|
- Sydney Waters
- 5 years ago
- Views:
Transcription
1 Co-arrays to be included in the Fortran 2008 Standard John Reid, ISO Fortran Convener The ISO Fortran Committee has decided to include co-arrays in the next revision of the Standard. Aim of this talk: introduce co-arrays and explain why we believe that they will lead to easier development of parallel programs, faster execution times, and better maintainability. No knowledge of Fortran 2003 is needed. I will explain the Fortran 95 features used. Kyoto University, 30 October, 2006 Tokyo University, 31 October, 2006
2 Summary of co-array model SPMD Single Program, Multiple Data Replicated to a number of images Images have indices 1, 2,... Number of images fixed during execution Each image has its own set of local variables Images execute asynchronously except when explicitly synchronized: sync all, sync team, notify, query,... Variables declared as co-arrays are accessible on another image through second set of array subscripts, delimited by [ ] and mapped to image indices by the usual rule Intrinsics: this_image, num_images,... collectives such as co_sum Critical construct 2
3 Examples of co-array syntax real :: r[*]! Scalar co-array real :: x(n)[*]! Array co-array! Co-arrays always have assumed! co-size (equal to number of images) real :: t! Local scalar integer :: p! Local scalar t = r[p] x(:) = x(:)[p]! Reference without [] is to local part x(:)[p] = r 3
4 Implementation model Usually each image resides on one processor. However, several images may share a processor (e.g. for debugging) and one image may execute on a multiple processors (e.g. with OpenMP). A co-array has the same set of bounds on all images, so the compiler can arrange that it occupies the same set of addresses within each image. On a shared-memory machine, a co-array can be implemented as a single large array. On any machine, a co-array may be implemented so that each image can calculate the memory address of an element on another image. 4
5 Synchronization With a few exceptions, the images execute asynchronously. If syncs are needed, the user supplies them explicitly.! Barrier on all images sync all! Barrier on the images of a team sync team (team)! Check into a barrier, but do not wait notify (image-set)! Wait for others to check into barrier query (image-set) For example, to read data on image 1 and get it to other images: if(this_image()==1) read(*,*)p sync all p = p[1] 5
6 Critical sections Exceptionally, it may be necessary to limit execution to one image at a time: critical p[6] = p[6] + 1 : end critical 6
7 Collective subroutines Intrinsics and involve synchronization. All have optional argument team. On every image, given the co-arrays real :: x[*], y(n)[*] real :: sum, sums(n) call co_sum(x,sum) returns p x[p] and call co_sum(y(:),sums(:)) returns y(:)[p]. p Others: co_all co_any co_count co_maxloc co_maxval co_product True if all values are true True if any value is true Numbers of true elements Image indices of maximum values Maximum values Products of elements 7
8 Dynamic co-arrays Only dynamic form: the allocatable co-array. All images synchronize at an allocate or deallocate statement so that they can all perform their allocations and deallocations in the same order. The bounds must not vary between images. Automatic co-arrays or co-array-valued functions would require automatic synchronization, so are not allowed. Co-Arrays and SAVE Unless allocatable or a dummy argument, a coarray must be given the SAVE attribute. This is to avoid the need for synchronization when co-arrays go out of scope on return from a procedure. 8
9 Procedures A subobject of a co-array without [ ], may be passed to a co-array. The ordinary rules of Fortran 95 apply to the local part; the co-rank and co-bounds are defined afresh. The interface must be explicit. No copy-in copy-out. The rules for resolving generic procedure references remain unchanged. No co-array syntax is permitted in pure procedures. 9
10 Structure components A co-array may be of a derived type with allocatable or pointer components. Pointers must have targets in their own image: q => z[i]%p! Not allowed allocate(z[i]%p)! Not allowed No automatic synchronization. Each image works independently. Provides a simple but powerful mechanism for cases where the size varies from image to image, avoiding loss of optimization. 10
11 Input/output Syntax to allow teams of images to access a single file. Allows local buffering. To open for a team: OPEN(unit,...,TEAM=team,...) There is an implied sync team (team) and the unit must not be opened on other images. Only cases: sequential write While an image is writing a record, the processor blocks other images. Thus each record comes from a single image. direct access Up to the programmer to synchronize access to a single record by more than one image. 11
12 Optimization Most of the time, the compiler can optimize as if the image is on its own, using its temporary storage such as cache, registers, etc. There is no coherency requirement except on synchronization. It also has scope to optimize communication. 12
13 Comparison with MPI (i) MPI is the de-facto standard but is awkward to program. Here is an example due to Jef Dawson of AHPCRC-NCSI. With co-arrays, to send the first m elements of an array from one image to another: real :: a(n)[*] me=this_image() if ( me.eq.2 ) a(1:m)=a(1:m)[1] sync all and with MPI: real :: a(n) call mpi_comm_rank(mpi_comm_world, & myrank, errcode) if (myrank.eq.0) call mpi_send & (a,m,mpi_float,1,tag1, & mpi_comm_world,errcode) if (myrank.eq.1) call mpi_recv & (a,m,mpi_float,0,tag1, & mpi_comm_world,status,errcode) 13
14 Comparison with MPI (ii) Experience on the Cray vector computers with the Cray compiler suggests that there is a performance advantage as the number of processes increases. For example, Dawson (2004) reports speed-up of 60 on 64 processors of the Cray X1 for a stencil update code, compared with 35 for MPI. Dawson, Jef (2004). Co-array Fortran for productivity and performance. In Army HPC Research Center Bulletin, 14, 4. 14
15 Advantages of co-arrays Easy to write code the compiler looks after the communication References to local data are obvious as such. Easy to maintain code more concise than MPI and easy to see what is happening Integrated with Fortran type checking, type conversion on assignment,... The compiler can optimize communication Local optimizations still available Does not make severe demands on the compiler, e.g. for coherency. 15
16 Further reading Numrich, Robert W. and Reid, John (2005). Co-arrays in the next Fortran Standard. ACM Fortran Forum, 24, 2, Also N1642.pdf in ftp://ftp.nag.co.uk/sc22wg5/n1601-n
Fortran 2008: what s in it for high-performance computing
Fortran 2008: what s in it for high-performance computing John Reid, ISO Fortran Convener, JKR Associates and Rutherford Appleton Laboratory Fortran 2008 has been completed and is about to be published.
More informationFortran Coarrays John Reid, ISO Fortran Convener, JKR Associates and Rutherford Appleton Laboratory
Fortran Coarrays John Reid, ISO Fortran Convener, JKR Associates and Rutherford Appleton Laboratory This talk will explain the objectives of coarrays, give a quick summary of their history, describe the
More informationParallel Programming in Fortran with Coarrays
Parallel Programming in Fortran with Coarrays John Reid, ISO Fortran Convener, JKR Associates and Rutherford Appleton Laboratory Fortran 2008 is now in FDIS ballot: only typos permitted at this stage.
More informationReport from WG5 convener
Report from WG5 convener Content of Fortran 2008 Framework was decided at last years WG5 meeting and was not substantially changed at this year s WG5 meeting. Two large items bits and intelligent macros
More informationProgramming for High Performance Computing in Modern Fortran. Bill Long, Cray Inc. 17-May-2005
Programming for High Performance Computing in Modern Fortran Bill Long, Cray Inc. 17-May-2005 Concepts in HPC Efficient and structured layout of local data - modules and allocatable arrays Efficient operations
More informationLecture V: Introduction to parallel programming with Fortran coarrays
Lecture V: Introduction to parallel programming with Fortran coarrays What is parallel computing? Serial computing Single processing unit (core) is used for solving a problem One task processed at a time
More informationParallel Programming with Coarray Fortran
Parallel Programming with Coarray Fortran SC10 Tutorial, November 15 th 2010 David Henty, Alan Simpson (EPCC) Harvey Richardson, Bill Long, Nathan Wichmann (Cray) Tutorial Overview The Fortran Programming
More informationCoarrays in the next Fortran Standard
ISO/IEC JTC1/SC22/WG5 N1824 Coarrays in the next Fortran Standard John Reid, JKR Associates, UK April 21, 2010 Abstract Coarrays will be included in the next Fortran Standard, known informally as Fortran
More informationAppendix D. Fortran quick reference
Appendix D Fortran quick reference D.1 Fortran syntax... 315 D.2 Coarrays... 318 D.3 Fortran intrisic functions... D.4 History... 322 323 D.5 Further information... 324 Fortran 1 is the oldest high-level
More informationCoarrays in the next Fortran Standard
ISO/IEC JTC1/SC22/WG5 N1724 Coarrays in the next Fortran Standard John Reid, JKR Associates, UK March 18, 2008 Abstract The WG5 committee, at its meeting in Delft, May 2005, decided to include coarrays
More informationProceedings of the GCC Developers Summit
Reprinted from the Proceedings of the GCC Developers Summit June 17th 19th, 2008 Ottawa, Ontario Canada Conference Organizers Andrew J. Hutton, Steamballoon, Inc., Linux Symposium, Thin Lines Mountaineering
More informationISO/IEC : TECHNICAL CORRIGENDUM 2
ISO/IEC 1539-1:2010 - TECHNICAL CORRIGENDUM 2 ISO/IEC/JTC1/SC22/WG5-N1957 Notes for WG5: Edits included in this document from the interpretations in N1932 as amended by 12-193 and 12-194 and in N1949 as
More informationParallel and High Performance Computing CSE 745
Parallel and High Performance Computing CSE 745 1 Outline Introduction to HPC computing Overview Parallel Computer Memory Architectures Parallel Programming Models Designing Parallel Programs Parallel
More informationParallel Programming without MPI Using Coarrays in Fortran SUMMERSCHOOL
Parallel Programming without MPI Using Coarrays in Fortran SUMMERSCHOOL 2007 2015 August 5, 2015 Ge Baolai SHARCNET Western University Outline What is coarray How to write: Terms, syntax How to compile
More informationMore Coarray Features. SC10 Tutorial, November 15 th 2010 Parallel Programming with Coarray Fortran
More Coarray Features SC10 Tutorial, November 15 th 2010 Parallel Programming with Coarray Fortran Overview Multiple Dimensions and Codimensions Allocatable Coarrays and Components of Coarray Structures
More informationInformation technology Programming languages Fortran Part 1: Base language
INTERNATIONAL STANDARD ISO/IEC 1539-1:2010 TECHNICAL CORRIGENDUM 2 Published 2013-06-01 INTERNATIONAL ORGANIZATION FOR STANDARDIZATION МЕЖДУНАРОДНАЯ ОРГАНИЗАЦИЯ ПО СТАНДАРТИЗАЦИИ ORGANISATION INTERNATIONALE
More informationTechnical Report on further interoperability with C
Technical Report on further interoperability with C John Reid, ISO Fortran Convener, JKR Associates and Rutherford Appleton Laboratory Fortran 2003 (or 2008) provides for interoperability of procedures
More informationNew Programming Paradigms: Partitioned Global Address Space Languages
Raul E. Silvera -- IBM Canada Lab rauls@ca.ibm.com ECMWF Briefing - April 2010 New Programming Paradigms: Partitioned Global Address Space Languages 2009 IBM Corporation Outline Overview of the PGAS programming
More informationChapter 4. Fortran Arrays
Chapter 4. Fortran Arrays Fortran arrays are any object with the dimension attribute. In Fortran 90/95, and in HPF, arrays may be very different from arrays in older versions of Fortran. Arrays can have
More informationTechnical Specification on further interoperability with C
Technical Specification on further interoperability with C John Reid, ISO Fortran Convener Fortran 2003 (or 2008) provides for interoperability of procedures with nonoptional arguments that are scalars,
More informationParallel Programming Languages. HPC Fall 2010 Prof. Robert van Engelen
Parallel Programming Languages HPC Fall 2010 Prof. Robert van Engelen Overview Partitioned Global Address Space (PGAS) A selection of PGAS parallel programming languages CAF UPC Further reading HPC Fall
More informationEvolution of Fortran. Presented by: Tauqeer Ahmad. Seminar on Languages for Scientific Computing
Evolution of Fortran Presented by: Seminar on Languages for Scientific Computing Outline (1) History of Fortran Versions FORTRAN I FORTRAN II FORTRAN III FORTRAN IV FORTRAN 66 FORTRAN 77 Evolution of FORTRAN
More informationTowards Exascale Computing with Fortran 2015
Towards Exascale Computing with Fortran 2015 Alessandro Fanfarillo National Center for Atmospheric Research Damian Rouson Sourcery Institute Outline Parallelism in Fortran 2008 SPMD PGAS Exascale challenges
More informationMigrating A Scientific Application from MPI to Coarrays. John Ashby and John Reid HPCx Consortium Rutherford Appleton Laboratory STFC UK
Migrating A Scientific Application from MPI to Coarrays John Ashby and John Reid HPCx Consortium Rutherford Appleton Laboratory STFC UK Why and Why Not? +MPI programming is arcane +New emerging paradigms
More informationAn Open64-based Compiler and Runtime Implementation for Coarray Fortran
An Open64-based Compiler and Runtime Implementation for Coarray Fortran talk by Deepak Eachempati Presented at: Open64 Developer Forum 2010 8/25/2010 1 Outline Motivation Implementation Overview Evaluation
More informationPortable, MPI-Interoperable! Coarray Fortran
Portable, MPI-Interoperable! Coarray Fortran Chaoran Yang, 1 Wesley Bland, 2! John Mellor-Crummey, 1 Pavan Balaji 2 1 Department of Computer Science! Rice University! Houston, TX 2 Mathematics and Computer
More informationNAGWare f95 Recent and Future Developments
NAGWare f95 Recent and Future Developments Malcolm Cohen The Numerical Algorithms Group Ltd., Oxford Nihon Numerical Algorithms Group KK, Tokyo Contents 1. Standard Fortran 2. High Performance 3. NAGWare
More informationCo-array Fortran for parallel programming
ISO/IEC JTC1/SC22/WG5 N1317 Co-array Fortran for parallel programming Robert W. Numrich, Silicon Graphics, Inc. and John Reid, Rutherford Appleton Laboratory Abstract Co-array Fortran, formerly known as
More informationPGI Accelerator Programming Model for Fortran & C
PGI Accelerator Programming Model for Fortran & C The Portland Group Published: v1.3 November 2010 Contents 1. Introduction... 5 1.1 Scope... 5 1.2 Glossary... 5 1.3 Execution Model... 7 1.4 Memory Model...
More informationA Comparison of Co-Array Fortran and OpenMP Fortran for SPMD Programming
The Journal of Supercomputing, 22, 231 250, 2002 2002 Kluwer Academic Publishers. Manufactured in The Netherlands. A Comparison of Co-Array Fortran and OpenMP Fortran for SPMD Programming ALAN J. WALLCRAFT
More informationIntroduction to Parallel Programming
Introduction to Parallel Programming Section 5. Victor Gergel, Professor, D.Sc. Lobachevsky State University of Nizhni Novgorod (UNN) Contents (CAF) Approaches to parallel programs development Parallel
More informationFortran Coding Standards and Style
Fortran Coding Standards and Style The Fortran Company Version 20160112 Copyright 2015-2016, The Fortran Company All rights reserved. Redistribution, with or without modification, is permitted provided
More informationIMPLEMENTATION AND EVALUATION OF ADDITIONAL PARALLEL FEATURES IN COARRAY FORTRAN
IMPLEMENTATION AND EVALUATION OF ADDITIONAL PARALLEL FEATURES IN COARRAY FORTRAN A Thesis Presented to the Faculty of the Department of Computer Science University of Houston In Partial Fulfillment of
More informationCo-Array Fortran for parallel programming
RAL-TR-1998-060 1 Co-Array Fortran for parallel programming by 2 3 R. W. Numrich and J. K. Reid Abstract Co-Array Fortran, formerly known as F, is a small extension of Fortran 95 for parallel processing.
More informationParallel Programming Features in the Fortran Standard. Steve Lionel 12/4/2012
Parallel Programming Features in the Fortran Standard Steve Lionel 12/4/2012 Agenda Overview of popular parallelism methodologies FORALL a look back DO CONCURRENT Coarrays Fortran 2015 Q+A 12/5/2012 2
More informationParallel Programming Models. Parallel Programming Models. Threads Model. Implementations 3/24/2014. Shared Memory Model (without threads)
Parallel Programming Models Parallel Programming Models Shared Memory (without threads) Threads Distributed Memory / Message Passing Data Parallel Hybrid Single Program Multiple Data (SPMD) Multiple Program
More informationCoarray Fortran: Past, Present, and Future. John Mellor-Crummey Department of Computer Science Rice University
Coarray Fortran: Past, Present, and Future John Mellor-Crummey Department of Computer Science Rice University johnmc@cs.rice.edu CScADS Workshop on Leadership Computing July 19-22, 2010 1 Staff Bill Scherer
More informationCo-array Fortran Performance and Potential: an NPB Experimental Study. Department of Computer Science Rice University
Co-array Fortran Performance and Potential: an NPB Experimental Study Cristian Coarfa Jason Lee Eckhardt Yuri Dotsenko John Mellor-Crummey Department of Computer Science Rice University Parallel Programming
More informationAdditional Parallel Features in Fortran An Overview of ISO/IEC TS 18508
Additional Parallel Features in Fortran An Overview of ISO/IEC TS 18508 Dr. Reinhold Bader Leibniz Supercomputing Centre Introductory remarks Technical Specification a Mini-Standard permits implementors
More informationTS Further Interoperability of Fortran with C WG5/N1917
TS 29113 Further Interoperability of Fortran with C WG5/N1917 7th May 2012 12:21 Draft document for DTS Ballot (Blank page) 2012/5/7 TS 29113 Further Interoperability of Fortran with C WG5/N1917 Contents
More informationHolland Computing Center Kickstart MPI Intro
Holland Computing Center Kickstart 2016 MPI Intro Message Passing Interface (MPI) MPI is a specification for message passing library that is standardized by MPI Forum Multiple vendor-specific implementations:
More informationChapter 3. Fortran Statements
Chapter 3 Fortran Statements This chapter describes each of the Fortran statements supported by the PGI Fortran compilers Each description includes a brief summary of the statement, a syntax description,
More informationCS4961 Parallel Programming. Lecture 16: Introduction to Message Passing 11/3/11. Administrative. Mary Hall November 3, 2011.
CS4961 Parallel Programming Lecture 16: Introduction to Message Passing Administrative Next programming assignment due on Monday, Nov. 7 at midnight Need to define teams and have initial conversation with
More informationImplementation and Evaluation of Coarray Fortran Translator Based on OMNI XcalableMP. October 29, 2015 Hidetoshi Iwashita, RIKEN AICS
Implementation and Evaluation of Coarray Fortran Translator Based on OMNI XcalableMP October 29, 2015 Hidetoshi Iwashita, RIKEN AICS Background XMP Contains Coarray Features XcalableMP (XMP) A PGAS language,
More informationMorden Fortran: Concurrency and parallelism
Morden Fortran: Concurrency and parallelism GENERAL SUMMERSCHOOL INTEREST SEMINARS 2007 2017 April 19, 2017 Ge Baolai SHARCNET Western University Outline Highlights of some Fortran 2008 enhancement Array
More informationReview More Arrays Modules Final Review
OUTLINE 1 REVIEW 2 MORE ARRAYS Using Arrays Why do we need dynamic arrays? Using Dynamic Arrays 3 MODULES Global Variables Interface Blocks Modular Programming 4 FINAL REVIEW THE STORY SO FAR... Create
More informationFortran Bill Long, Cray Inc. 21-May Cray Proprietary
Fortran 2003 Bill Long, Cray Inc. 21-May-2004 Cray Proprietary Fortran 2003 Specification for Fortran 2003 (f03) is finished Standard should be official in September 569 pages including Appendices and
More informationSupercomputing in Plain English Exercise #6: MPI Point to Point
Supercomputing in Plain English Exercise #6: MPI Point to Point In this exercise, we ll use the same conventions and commands as in Exercises #1, #2, #3, #4 and #5. You should refer back to the Exercise
More informationAdvanced Features. SC10 Tutorial, November 15 th Parallel Programming with Coarray Fortran
Advanced Features SC10 Tutorial, November 15 th 2010 Parallel Programming with Coarray Fortran Advanced Features: Overview Execution segments and Synchronisation Non-global Synchronisation Critical Sections
More informationProgramming Languages
Programming Languages Tevfik Koşar Lecture - VIII February 9 th, 2006 1 Roadmap Allocation techniques Static Allocation Stack-based Allocation Heap-based Allocation Scope Rules Static Scopes Dynamic Scopes
More informationDangerously Clever X1 Application Tricks
Dangerously Clever X1 Application Tricks CUG 2004 James B. White III (Trey) trey@ornl.gov 1 Acknowledgement Research sponsored by the Mathematical, Information, and Division, Office of Advanced Scientific
More informationCurrent Developments in Fortran Standards. David Muxworthy 15 June 2012
Current Developments in Fortran Standards David Muxworthy d.muxworthy@bcs.org.uk 15 June 2012 Purpose of standardization BSI was founded in 1901, amongst other things to: co-ordinate the efforts of producers
More informationPortable, MPI-Interoperable! Coarray Fortran
Portable, MPI-Interoperable! Coarray Fortran Chaoran Yang, 1 Wesley Bland, 2! John Mellor-Crummey, 1 Pavan Balaji 2 1 Department of Computer Science! Rice University! Houston, TX 2 Mathematics and Computer
More informationControl Abstraction. Hwansoo Han
Control Abstraction Hwansoo Han Review of Static Allocation Static allocation strategies Code Global variables Own variables (live within an encapsulation - static in C) Explicit constants (including strings,
More informationThe new features of Fortran 2003
The new features of Fortran 2003 David Muxworthy BSI Fortran Convenor Pages in Fortran Standards 0 100 200 300 400 500 600 700 1966 1967 1968 1969 1970 1971 1972 1973 1974 1975 1976 1977 1978 1979 1980
More informationA Coarray Fortran Implementation to Support Data-Intensive Application Development
A Coarray Fortran Implementation to Support Data-Intensive Application Development Deepak Eachempati, Alan Richardson, Terrence Liao, Henri Calandra and Barbara Chapman Department of Computer Science,
More informationMPI: A Message-Passing Interface Standard
MPI: A Message-Passing Interface Standard Version 2.1 Message Passing Interface Forum June 23, 2008 Contents Acknowledgments xvl1 1 Introduction to MPI 1 1.1 Overview and Goals 1 1.2 Background of MPI-1.0
More information6.1 Multiprocessor Computing Environment
6 Parallel Computing 6.1 Multiprocessor Computing Environment The high-performance computing environment used in this book for optimization of very large building structures is the Origin 2000 multiprocessor,
More informationMPI and OpenMP (Lecture 25, cs262a) Ion Stoica, UC Berkeley November 19, 2016
MPI and OpenMP (Lecture 25, cs262a) Ion Stoica, UC Berkeley November 19, 2016 Message passing vs. Shared memory Client Client Client Client send(msg) recv(msg) send(msg) recv(msg) MSG MSG MSG IPC Shared
More informationNAGWare f95 and reliable, portable programming.
NAGWare f95 and reliable, portable programming. Malcolm Cohen The Numerical Algorithms Group Ltd., Oxford How to detect errors using NAGWare f95, and how to write portable, reliable programs. Support for
More informationCompiler support for Fortran 2003 and 2008 standards Ian Chivers & Jane Sleightholme Fortranplus
Compiler support for Fortran 2003 and Ian Chivers & Jane Sleightholme Fortranplus ian@fortranplus.co.uk www.fortranplus.co.uk 15 th June 2012 jane@fortranplus.co.uk A bit about us Ian & Jane Worked in
More informationMPI Message Passing Interface
MPI Message Passing Interface Portable Parallel Programs Parallel Computing A problem is broken down into tasks, performed by separate workers or processes Processes interact by exchanging information
More informationUsing Co-Array Fortran to Enhance the Scalability of the EPIC Code
Using Co-Array Fortran to Enhance the Scalability of the EPIC Code Jef Dawson Army High Performance Computing Research Center, Network Computing Services, Inc. ABSTRACT: Supercomputing users continually
More informationIntroduction to Parallel Programming
Introduction to Parallel Programming Overview Parallel programming allows the user to use multiple cpus concurrently Reasons for parallel execution: shorten execution time by spreading the computational
More informationdbx90: Fortran debugger March 9, 2009
dbx90: Fortran debugger March 9, 2009 1 Name dbx90 a Fortran 90/95 debugger for use with the NAG Fortran compiler. 2 Usage dbx90 [option]... executable-file 3 Description dbx90 is a Fortran 90/95 debugger
More informationOpenACC 2.6 Proposed Features
OpenACC 2.6 Proposed Features OpenACC.org June, 2017 1 Introduction This document summarizes features and changes being proposed for the next version of the OpenACC Application Programming Interface, tentatively
More informationThe Complete Compendium on Cooperative Computing using Coarrays. c 2008 Andrew Vaught October 29, 2008
Preface The Complete Compendium on Cooperative Computing using Coarrays. c 2008 Andrew Vaught October 29, 2008 Over the last several decades, the speed of computing has increased exponentially, a phenononom
More informationPGI Fortran & C Accelerator Programming Model. The Portland Group
PGI Fortran & C Accelerator Programming Model The Portland Group Published: v0.72 December 2008 Contents 1. Introduction...3 1.1 Scope...3 1.2 Glossary...3 1.3 Execution Model...4 1.4 Memory Model...5
More informationFortran 90 - A thumbnail sketch
Fortran 90 - A thumbnail sketch Michael Metcalf CERN, Geneva, Switzerland. Abstract The main new features of Fortran 90 are presented. Keywords Fortran 1 New features In this brief paper, we describe in
More informationWelcome. Modern Fortran (F77 to F90 and beyond) Virtual tutorial starts at BST
Welcome Modern Fortran (F77 to F90 and beyond) Virtual tutorial starts at 15.00 BST Modern Fortran: F77 to F90 and beyond Adrian Jackson adrianj@epcc.ed.ac.uk @adrianjhpc Fortran Ancient History (1967)
More informationCompiler Support for the Fortran 2003 and 2008 Standards Revision 17
Introduction Compiler Support for the Fortran 2003 and 2008 Standards Revision 17 Ian D Chivers & Jane Sleightholme Ian Chivers: Rhymney Consulting, London. Jane Sleightholme: FortranPlus, London. ian@rhymneyconsulting.co.uk
More informationA Message Passing Standard for MPP and Workstations
A Message Passing Standard for MPP and Workstations Communications of the ACM, July 1996 J.J. Dongarra, S.W. Otto, M. Snir, and D.W. Walker Message Passing Interface (MPI) Message passing library Can be
More informationProgramming with MPI
Programming with MPI p. 1/?? Programming with MPI Point-to-Point Transfers Nick Maclaren nmm1@cam.ac.uk May 2008 Programming with MPI p. 2/?? Digression Most books and courses teach point--to--point first
More informationPACKAGE SPECIFICATION HSL 2013
PACKAGE SPECIFICATION HSL 2013 1 SUMMARY Given a rank-one or rank-two allocatable array, reallocates the array to have a different size, and can copy all or part of the original array into the new array.
More informationComparing OpenACC 2.5 and OpenMP 4.1 James C Beyer PhD, Sept 29 th 2015
Comparing OpenACC 2.5 and OpenMP 4.1 James C Beyer PhD, Sept 29 th 2015 Abstract As both an OpenMP and OpenACC insider I will present my opinion of the current status of these two directive sets for programming
More informationIndex. classes, 47, 228 coarray examples, 163, 168 copystring, 122 csam, 125 csaxpy, 119 csaxpyval, 120 csyscall, 127 dfetrf,14 dfetrs, 14
Index accessor-mutator routine example in a module, 7 PUBLIC or PRIVATE components, 6 ACM, ix editors of CALGO, ix Adams, Brainerd et al., see books, Fortran reference Airy s equation boundary value problem,
More informationParallel Programming. Libraries and Implementations
Parallel Programming Libraries and Implementations Reusing this material This work is licensed under a Creative Commons Attribution- NonCommercial-ShareAlike 4.0 International License. http://creativecommons.org/licenses/by-nc-sa/4.0/deed.en_us
More informationPerformance Comparison between Two Programming Models of XcalableMP
Performance Comparison between Two Programming Models of XcalableMP H. Sakagami Fund. Phys. Sim. Div., National Institute for Fusion Science XcalableMP specification Working Group (XMP-WG) Dilemma in Parallel
More informationIntroduction to the PGAS (Partitioned Global Address Space) Languages Coarray Fortran (CAF) and Unified Parallel C (UPC) Dr. R. Bader Dr. A.
Introduction to the PGAS (Partitioned Global Address Space) Languages Coarray Fortran (CAF) and Unified Parallel C (UPC) Dr. R. Bader Dr. A. Block January 2011 Applying PGAS to classical HPC languages
More informationComputer Architecture
Jens Teubner Computer Architecture Summer 2016 1 Computer Architecture Jens Teubner, TU Dortmund jens.teubner@cs.tu-dortmund.de Summer 2016 Jens Teubner Computer Architecture Summer 2016 2 Part I Programming
More informationJohn Mellor-Crummey Department of Computer Science Center for High Performance Software Research Rice University
Co-Array Fortran and High Performance Fortran John Mellor-Crummey Department of Computer Science Center for High Performance Software Research Rice University LACSI Symposium October 2006 The Problem Petascale
More informationIntroduction to OpenCL!
Lecture 6! Introduction to OpenCL! John Cavazos! Dept of Computer & Information Sciences! University of Delaware! www.cis.udel.edu/~cavazos/cisc879! OpenCL Architecture Defined in four parts Platform Model
More informationMPI: Message Passing Interface An Introduction. S. Lakshmivarahan School of Computer Science
MPI: Message Passing Interface An Introduction S. Lakshmivarahan School of Computer Science MPI: A specification for message passing libraries designed to be a standard for distributed memory message passing,
More informationElementary Parallel Programming with Examples. Reinhold Bader (LRZ) Georg Hager (RRZE)
Elementary Parallel Programming with Examples Reinhold Bader (LRZ) Georg Hager (RRZE) Two Paradigms for Parallel Programming Hardware Designs Distributed Memory M Message Passing explicit programming required
More informationAcknowledgments. Amdahl s Law. Contents. Programming with MPI Parallel programming. 1 speedup = (1 P )+ P N. Type to enter text
Acknowledgments Programming with MPI Parallel ming Jan Thorbecke Type to enter text This course is partly based on the MPI courses developed by Rolf Rabenseifner at the High-Performance Computing-Center
More informationCan Accelerators Really Accelerate Harmonie?
Can Accelerators Really Accelerate Harmonie? Enda O Brien, Adam Ralph Irish Centre for High-End Computing Motivation There is constant demand for more performance Conventional compute cores not getting
More informationChip Multiprocessors COMP Lecture 9 - OpenMP & MPI
Chip Multiprocessors COMP35112 Lecture 9 - OpenMP & MPI Graham Riley 14 February 2018 1 Today s Lecture Dividing work to be done in parallel between threads in Java (as you are doing in the labs) is rather
More informationIntroduction to MPI. Ekpe Okorafor. School of Parallel Programming & Parallel Architecture for HPC ICTP October, 2014
Introduction to MPI Ekpe Okorafor School of Parallel Programming & Parallel Architecture for HPC ICTP October, 2014 Topics Introduction MPI Model and Basic Calls MPI Communication Summary 2 Topics Introduction
More informationSINGLE-SIDED PGAS COMMUNICATIONS LIBRARIES. Basic usage of OpenSHMEM
SINGLE-SIDED PGAS COMMUNICATIONS LIBRARIES Basic usage of OpenSHMEM 2 Outline Concept and Motivation Remote Read and Write Synchronisation Implementations OpenSHMEM Summary 3 Philosophy of the talks In
More informationParallel Paradigms & Programming Models. Lectured by: Pham Tran Vu Prepared by: Thoai Nam
Parallel Paradigms & Programming Models Lectured by: Pham Tran Vu Prepared by: Thoai Nam Outline Parallel programming paradigms Programmability issues Parallel programming models Implicit parallelism Explicit
More informationA Comparison of Unified Parallel C, Titanium and Co-Array Fortran. The purpose of this paper is to compare Unified Parallel C, Titanium and Co-
Shaun Lindsay CS425 A Comparison of Unified Parallel C, Titanium and Co-Array Fortran The purpose of this paper is to compare Unified Parallel C, Titanium and Co- Array Fortran s methods of parallelism
More informationIntroduction to Parallel Computing. CPS 5401 Fall 2014 Shirley Moore, Instructor October 13, 2014
Introduction to Parallel Computing CPS 5401 Fall 2014 Shirley Moore, Instructor October 13, 2014 1 Definition of Parallel Computing Simultaneous use of multiple compute resources to solve a computational
More informationHigh Performance Fortran http://www-jics.cs.utk.edu jics@cs.utk.edu Kwai Lam Wong 1 Overview HPF : High Performance FORTRAN A language specification standard by High Performance FORTRAN Forum (HPFF), a
More informationFortran 2003 and Beyond. Bill Long, Cray Inc. 17-May-2005
Fortran 2003 and Beyond Bill Long, Cray Inc. 17-May-2005 Fortran 2003 and Fortran 2008 Fortran is now ISO/IEC 1539:1-2004 Common name is Fortran 2003 or f03 This cancels and replaces the previous standard,
More informationProgramming with MPI
Programming with MPI p. 1/?? Programming with MPI One-sided Communication Nick Maclaren nmm1@cam.ac.uk October 2010 Programming with MPI p. 2/?? What Is It? This corresponds to what is often called RDMA
More informationFirst Experiences with Application Development with Fortran Damian Rouson
First Experiences with Application Development with Fortran 2018 Damian Rouson Overview Fortran 2018 in a Nutshell ICAR & Coarray ICAR WRF-Hydro Results Conclusions www.yourwebsite.com Overview Fortran
More informationOpenMP - II. Diego Fabregat-Traver and Prof. Paolo Bientinesi WS15/16. HPAC, RWTH Aachen
OpenMP - II Diego Fabregat-Traver and Prof. Paolo Bientinesi HPAC, RWTH Aachen fabregat@aices.rwth-aachen.de WS15/16 OpenMP References Using OpenMP: Portable Shared Memory Parallel Programming. The MIT
More informationIntroduction to OpenMP. Tasks. N.M. Maclaren September 2017
2 OpenMP Tasks 2.1 Introduction Introduction to OpenMP Tasks N.M. Maclaren nmm1@cam.ac.uk September 2017 These were introduced by OpenMP 3.0 and use a slightly different parallelism model from the previous
More informationExamining the Code. [Reading assignment: Chapter 6, pp ]
Examining the Code [Reading assignment: Chapter 6, pp. 91-104] Static white-box testing Static white-box testing is the process of carefully and methodically reviewing the software design, architecture,
More informationMessage Passing Programming. Designing MPI Applications
Message Passing Programming Designing MPI Applications Reusing this material This work is licensed under a Creative Commons Attribution- NonCommercial-ShareAlike 4.0 International License. http://creativecommons.org/licenses/by-nc-sa/4.0/deed.en_us
More information