HSA foundation! Advanced Topics on Heterogeneous System Architectures. Politecnico di Milano! Seminar Room A. Alario! 23 November, 2015!

Similar documents
HSA Foundation! Advanced Topics on Heterogeneous System Architectures. Politecnico di Milano! Seminar Room (Bld 20)! 15 December, 2017!

HETEROGENEOUS SYSTEM ARCHITECTURE: PLATFORM FOR THE FUTURE

CLICK TO EDIT MASTER TITLE STYLE. Click to edit Master text styles. Second level Third level Fourth level Fifth level

THE PROGRAMMER S GUIDE TO THE APU GALAXY. Phil Rogers, Corporate Fellow AMD

Heterogeneous SoCs. May 28, 2014 COMPUTER SYSTEM COLLOQUIUM 1

THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S BEYOND THE GPU

SIMULATOR AMD RESEARCH JUNE 14, 2015

Heterogeneous Computing

Heterogeneous Architecture. Luca Benini

Copyright Khronos Group Page 1. Vulkan Overview. June 2015

Modern Processor Architectures. L25: Modern Compiler Design

HSA QUEUEING HOT CHIPS TUTORIAL - AUGUST 2013 IAN BRATT PRINCIPAL ENGINEER ARM

Panel Discussion: The Future of I/O From a CPU Architecture Perspective

AMD s Unified CPU & GPU Processor Concept

Modern Processor Architectures (A compiler writer s perspective) L25: Modern Compiler Design

The Role of Standards in Heterogeneous Programming

Handout 3. HSAIL and A SIMT GPU Simulator

HSAIL: PORTABLE COMPILER IR FOR HSA

OpenCL: History & Future. November 20, 2017

SIGGRAPH Briefing August 2014

Khronos Connects Software to Silicon

Parallel Programming on Larrabee. Tim Foley Intel Corp

AMD ACCELERATING TECHNOLOGIES FOR EXASCALE COMPUTING FELLOW 3 OCTOBER 2016

Profiling and Debugging OpenCL Applications with ARM Development Tools. October 2014

OpenCL Overview. Shanghai March Neil Trevett Vice President Mobile Content, NVIDIA President, The Khronos Group

Towards a codelet-based runtime for exascale computing. Chris Lauderdale ET International, Inc.

FPGA Reconfiguration!

Antonio R. Miele Marco D. Santambrogio

HETEROGENEOUS COMPUTING IN THE MODERN AGE. 1 HiPEAC January, 2012 Public

Copyright Khronos Group Page 1

Linux Kernel Driver Support to Heterogeneous System Architecture

x Welcome to the jungle. The free lunch is so over

Chapter 4: Threads. Chapter 4: Threads. Overview Multicore Programming Multithreading Models Thread Libraries Implicit Threading Threading Issues

HSA FOR APPLICATION PROGRAMMING

Compiling for HSA accelerators with GCC

Efficient CPU GPU data transfers CUDA 6.0 Unified Virtual Memory

OPERATING SYSTEM. Chapter 4: Threads

Ultra Large-Scale FFT Processing on Graphics Processor Arrays. Author: J.B. Glenn-Anderson, PhD, CTO enparallel, Inc.

HSAemu A Full System Emulator for HSA Platforms

Take GPU Processing Power Beyond Graphics with Mali GPU Computing

ClearSpeed Visual Profiler

Porting Fabric Engine to NVIDIA Unified Memory: A Case Study. Peter Zion Chief Architect Fabric Engine Inc.

Press Briefing SIGGRAPH 2015 Neil Trevett Khronos President NVIDIA Vice President Mobile Ecosystem. Copyright Khronos Group Page 1

From Application to Technology OpenCL Application Processors Chung-Ho Chen

GPGPU on ARM. Tom Gall, Gil Pitney, 30 th Oct 2013

Copyright Khronos Group Page 1

Press Briefing SIGGRAPH 2015 Neil Trevett Khronos President NVIDIA Vice President Mobile Ecosystem. Copyright Khronos Group Page 1

Chapter 4: Threads. Chapter 4: Threads

HETEROGENEOUS MEMORY MANAGEMENT. Linux Plumbers Conference Jérôme Glisse

Introduction. CS 2210 Compiler Design Wonsun Ahn

General Purpose GPU Programming (1) Advanced Operating Systems Lecture 14

Using Industry Standards to Exploit the Advantages and Resolve the Challenges of Multicore Technology

Accelerating Vision Processing

Next Generation OpenGL Neil Trevett Khronos President NVIDIA VP Mobile Copyright Khronos Group Page 1

! Readings! ! Room-level, on-chip! vs.!

Parallel Programming Principle and Practice. Lecture 9 Introduction to GPGPUs and CUDA Programming Model

HKG OpenCL Support by NNVM & TVM. Jammy Zhou - Linaro

RUNTIME SUPPORT FOR ADAPTIVE SPATIAL PARTITIONING AND INTER-KERNEL COMMUNICATION ON GPUS

Xen and the Art of Virtualization. CSE-291 (Cloud Computing) Fall 2016

Vulkan Launch Webinar 18 th February Copyright Khronos Group Page 1

OPENCL GPU BEST PRACTICES BENJAMIN COQUELLE MAY 2015

CUDA GPGPU Workshop 2012

CSE 120 Principles of Operating Systems

Parallel and Distributed Computing

Introduction to CUDA Algoritmi e Calcolo Parallelo. Daniele Loiacono

MTAPI: Parallel Programming for Embedded Multicore Systems

OpenCL Base Course Ing. Marco Stefano Scroppo, PhD Student at University of Catania

Chapter 4: Multithreaded Programming

ET International HPC Runtime Software. ET International Rishi Khan SC 11. Copyright 2011 ET International, Inc.

Chapter 3 Parallel Software

OpenACC Course. Office Hour #2 Q&A

EI 338: Computer Systems Engineering (Operating Systems & Computer Architecture)

TOOLS FOR IMPROVING CROSS-PLATFORM SOFTWARE DEVELOPMENT

The OpenVX Computer Vision and Neural Network Inference

Copyright Khronos Group Page 1

Copyright Khronos Group Page 1. Introduction to SYCL. SYCL Tutorial IWOCL

Copyright Khronos Group 2012 Page 1. OpenCL 1.2. August 2012

A brief introduction to OpenMP

Operating Systems 2 nd semester 2016/2017. Chapter 4: Threads

A Translation Framework for Automatic Translation of Annotated LLVM IR into OpenCL Kernel Function

Critically Missing Pieces on Accelerators: A Performance Tools Perspective

Motivation. Threads. Multithreaded Server Architecture. Thread of execution. Chapter 4

GpuWrapper: A Portable API for Heterogeneous Programming at CGG

MIGRATION OF LEGACY APPLICATIONS TO HETEROGENEOUS ARCHITECTURES Francois Bodin, CTO, CAPS Entreprise. June 2011

Intel Xeon Phi Coprocessors

Did I Just Do That on a Bunch of FPGAs?

CSE 4/521 Introduction to Operating Systems

HKG : OpenAMP Introduction. Wendy Liang

Distributed & Heterogeneous Programming in C++ for HPC at SC17

Automatic Pruning of Autotuning Parameter Space for OpenCL Applications

CSE 4/521 Introduction to Operating Systems

Domain Specific Languages for Financial Payoffs. Matthew Leslie Bank of America Merrill Lynch

Colin Riddell GPU Compiler Developer Codeplay Visit us at

Introduction to CUDA Algoritmi e Calcolo Parallelo. Daniele Loiacono

OpenACC 2.6 Proposed Features

OpenCL Press Conference

1 HiPEAC January, 2012 Public TASKS, FUTURES AND ASYNCHRONOUS PROGRAMMING

INTRODUCTION TO OPENCL TM A Beginner s Tutorial. Udeepta Bordoloi AMD

Heterogeneous-Race-Free Memory Models

Chapter 4: Threads. Operating System Concepts 9 th Edit9on

Transcription:

Advanced Topics on Heterogeneous System Architectures HSA foundation! Politecnico di Milano! Seminar Room A. Alario! 23 November, 2015! Antonio R. Miele! Marco D. Santambrogio! Politecnico di Milano!

2 References! This presentation is based on the material and slides published on the HSA foundation website:! http://www.hsafoundation.com/!

3 Heterogeneous processors have proliferated make them better! Heterogeneous SOCs have arrived and are! a tremendous advance over previous! platforms! SOCs combine CPU cores, GPU cores and! other accelerators, with high bandwidth! access to memory! How do we make them even better?! Easier to program! Easier to optimize! Easier to load balance! Higher performance! Lower power! HSA unites accelerators architecturally! Early focus on the GPU compute accelerator, but HSA will go well beyond the GPU!

4 HSA foundation! Founded in June 2012! Developing a new platform! for heterogeneous systems! www.hsafoundation.com! Specifications under! development in working! groups to define the platform! Membership consists of 43 companies and 16 universities! Adding 1-2 new members each month!

HSA consortium! 5

6 HSA goals! To enable power-efficient performance! To improve programmability of heterogeneous processors! To increase the portability of code across processors and platforms! To increase the pervasiveness of heterogeneous solutions throughout the industry!

7 Paradigm shift! Inflection in processor design and programming!

8 Key features of HSA! huma Heterogeneous Unified Memory Architecture! hq Heterogeneous Queuing! HSAIL HSA Intermediate Language!

9 Key features of HSA! huma Heterogeneous Unified Memory Architecture! hq Heterogeneous Queuing! HSAIL HSA Intermediate Language!

10 Legacy GPU compute! Multiple memory pools! Multiple address spaces! No pointer-based data structures! Explicit data copying across PCIe! High latency! Low bandwidth! High overhead dispatch! Need lots of compute on GPU to amortize copy overhead! Very limited GPU memory capacity! Dual source development! Proprietary environments! Expert programmers only!

11 Existing APUs and SoCs! Physical integration of GPUs and CPUs! Data copies on an internal bus! Two memory pools remain! Still queue through the OS! Still requires expert programmers! APU = Accelerated Processing Unit (i.e. a SoC containing also a GPU)!

12 Existing APUs and SoCs! CPU and GPU still have separate memories for the programmer (different virtual memory spaces)! 1. CPU explicitly copies data to GPU memory! 2. GPU executes computation! 3. CPU explicitly copies results back to its own memory!

13 An HSA enabled SoC! Unified Coherent Memory enables data sharing across all processors! Enabling the usage of pointers! Not explicit data transfer -> values move on demand! Pageable virtual addresses for GPUs -> no GPU capacity contraints! Processors architected to operate cooperatively! Designed to enable the application to run on different processors at different times!

Unified coherent memory! 14

15 Unified coherent memory! CPU and GPU have a unified virtual memory spaces! 1. CPU simply passes a pointer to GPU! 2. GPU executes computation! 3. CPU can read the results directly no copy need!!

Unified coherent memory! 16

17 Unified coherent memory! Transmission of input data

Unified coherent memory! 18

Unified coherent memory! 19

Unified coherent memory! 20

Unified coherent memory! 21

22 Unified coherent memory! Transmission of results

Unified coherent memory! 23

Unified coherent memory! 24

Unified coherent memory! 25

Unified coherent memory! 26

Unified coherent memory! 27

Unified coherent memory! 28

29 Key features of HSA! huma Heterogeneous Unified Memory Architecture! hq Heterogeneous Queuing! HSAIL HSA Intermediate Language!

30 hq: heterogeneous queuing! Task queuing runtimes! Popular pattern for task and data parallel programming on Symmetric Multiprocessor (SMP) systems! Characterized by:! A work queue per core! Runtime library that divides large loops into tasks and distributes to queues! A work stealing scheduler that keeps system balanced! HSA is designed to extend this pattern to run on heterogeneous systems!

31 hq: heterogeneous queuing! How compute dispatch operates today in the driver model!

32 hq: heterogeneous queuing! How compute dispatch improves under HSA! Application codes to the hardware! User mode queuing! Hardware scheduling! Low dispatch times!! No Soft Queues! No User Mode Drivers! No Kernel Mode Transitions! No Overhead!

33 hq: heterogeneous queuing! AQL (Architected Queueing Language) enables any agent to enqueue tasks!

34 hq: heterogeneous queuing! A work stealing scheduler that keeps system balanced!

35 Advantages of the queuing model! Today s picture:!

36 Advantages of the queuing model! The unified shared memory allows to share pointers among different processing elements thus avoiding explicit memory transfer requests

37 Advantages of the queuing model! Coherent caches remove the necessity to perform explicit synchronizajon operajon

38 Advantages of the queuing model! The supported signaling mechanism enables asynchronous events between agents without involving the OS kernel

39 Advantages of the queuing model! Tasks are directly enqueued by the applicajons without using OS mechanisms

40 Advantages of the queuing model! HSA picture:!

41 Summary on the queuing model! User mode queuing for low latency dispatch! Application dispatches directly! No OS or driver required in the dispatch path! Architected Queuing Layer! Single compute dispatch path for all hardware! No driver translation, direct to hardware! Allows for dispatch to queue from any agent! CPU or GPU! GPU self-enqueue enables lots of solutions! Recursion! Tree traversal! Wavefront reforming!

42 Other necessary HW mechanisms! Task preemption and context switching have to be supported by all computing resources (also GPUs)!

43 Key features of HSA! huma Heterogeneous Unified Memory Architecture! hq Heterogeneous Queuing! HSAIL HSA Intermediate Language!

44 HSA intermediate layer (HSAIL)! A portable virtual ISA for vendor-independent compilation and distribution! Like Java bytecodes for GPUs! Low-level IR, close to machine ISA level! Most optimizations (including register allocation) performed before HSAIL! Generated by a high-level compiler (LLVM, gcc, Java VM, etc.)! Application binaries may ship with embedded HSAIL! Compiled down to target ISA by a vendor-specific finalizer! Finalizer may execute at run time, install time, or build time!

45 HSA intermediate layer (HSAIL)! HSA compilation stack! HSA runtime stack!

46 HSA intermediate layer (HSAIL)! Explicitly parallel! Designed for data parallel programming! Support for exceptions, virtual functions, and other high level language features! Syscall methods! GPU code can call directly to system services, IO, printf, etc!

47 HSA intermediate layer (HSAIL)! Lower level than OpenCL SPIR! Fits naturally in the OpenCL compilation stack! Suitable to support additional high level languages and programming models:! Java, C++, OpenMP, C++, Python, etc!

48 HSA software stack! HSA supports many languages! HSAIL

49 HSA and OpenCL! HSA is an optimized platform architecture for OpenCL! Not an alternative to OpenCL! OpenCL on HSA will benefit from! Avoidance of wasteful copies! Low latency dispatch! Improved memory model! Pointers shared between CPU and GPU! OpenCL 2.0 leverages HSA Features! Shared Virtual Memory! Platform Atomics!

50 HSA and Java! Targeted at Java 9 (2015 release)! Allows developers to efficiently represent data parallel algorithms in Java! Sumatra repurposes Java 8 s multicore Stream/Lambda API s to enable both CPU or GPU computing! At runtime, Sumatra enabled Java Virtual Machine (JVM) will dispatch selected constructs to available HSA enabled devices!

51 HSA and Java! Evolution of the Java acceleration before the Sumatra project!

HSA software stack! 52

53 HSA runtime! A thin, user-mode API that provides the interface necessary for the host to launch compute kernels to the available HSA components! The overall goal is to provide a high-performance dispatch mechanism that is portable across multiple HSA vendor architectures! The dispatch mechanism differentiates the HSA runtime from other language runtimes by architected argument setting and kernel launching at the hardware and specification level!

54 HSA runtime! The HSA core runtime API is standard across all HSA vendors, such that languages which use the HSA runtime can run on different vendor s platforms that support the API! The implementation of the HSA runtime may include kernel-level components (required for some hardware components, ex: AMD Kaveri) or may be entirely userspace (for example, simulators or CPU implementations)!

HSA runtime! 55

56 HSA taking platform to programmers! Balance between CPU and GPU for performance and power efficiency! Make GPUs accessible to wider audience of programmers! Programming models close to today s CPU programming models! Enabling more advanced language features on GPU! Shared virtual memory enables complex pointer-containing data structures (lists, trees, etc) and hence more applications on GPU! Kernel can enqueue work to any other device in the system (e.g. GPU->GPU, GPU->CPU)! Enabling task-graph style algorithms, Ray-Tracing, etc.!

57 HSA taking platform to programmers! Complete tool-chain for programming, debugging and profiling! HSA provides a compatible architecture across a wide range of programming models and HW implementations!

58 HSA programming model! Single source! Host and device code side-by-side in same source file! Written in same programming language! Single unified coherent address space! Freely share pointers between host and device! Similar memory model as multi-core CPU! Parallel regions identified with existing language syntax! Typically same syntax used for multi-core CPU! HSAIL is the compiler IR that supports these programming models!

Specifications and software! 59

60 HSA architecture V1! GPU compute C++ support! User Mode Scheduling! Fully coherent memory between CPU & GPU! GPU uses pageable system memory via CPU pointers! GPU graphics pre-emption! GPU compute context switch!

Partners roadmaps! 61

62 Partners roadmaps! 2015

Partners roadmaps! 63

Partners roadmaps! 64

Partners roadmaps! 65

Partners roadmaps! 66