CLICK TO EDIT MASTER TITLE STYLE. Click to edit Master text styles. Second level Third level Fourth level Fifth level

Similar documents
THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S BEYOND THE GPU

HSA Foundation! Advanced Topics on Heterogeneous System Architectures. Politecnico di Milano! Seminar Room (Bld 20)! 15 December, 2017!

HSA foundation! Advanced Topics on Heterogeneous System Architectures. Politecnico di Milano! Seminar Room A. Alario! 23 November, 2015!

Panel Discussion: The Future of I/O From a CPU Architecture Perspective

HETEROGENEOUS SYSTEM ARCHITECTURE: PLATFORM FOR THE FUTURE

THE PROGRAMMER S GUIDE TO THE APU GALAXY. Phil Rogers, Corporate Fellow AMD

SIMULATOR AMD RESEARCH JUNE 14, 2015

Modern Processor Architectures. L25: Modern Compiler Design

Modern Processor Architectures (A compiler writer s perspective) L25: Modern Compiler Design

Handout 3. HSAIL and A SIMT GPU Simulator

AMD ACCELERATING TECHNOLOGIES FOR EXASCALE COMPUTING FELLOW 3 OCTOBER 2016

Heterogeneous Computing

Copyright Khronos Group Page 1. Vulkan Overview. June 2015

HSAIL: PORTABLE COMPILER IR FOR HSA

OpenCL Overview. Shanghai March Neil Trevett Vice President Mobile Content, NVIDIA President, The Khronos Group

Heterogeneous SoCs. May 28, 2014 COMPUTER SYSTEM COLLOQUIUM 1

HSA QUEUEING HOT CHIPS TUTORIAL - AUGUST 2013 IAN BRATT PRINCIPAL ENGINEER ARM

Khronos Connects Software to Silicon

Simplify Software Integration for FPGA Accelerators with OPAE

From Application to Technology OpenCL Application Processors Chung-Ho Chen

Challenges for GPU Architecture. Michael Doggett Graphics Architecture Group April 2, 2008

Windows 7 Overview. Windows 7. Objectives. The History of Windows. CS140M Fall Lake 1

Sequential Consistency for Heterogeneous-Race-Free

AMD s Unified CPU & GPU Processor Concept

Vulkan Launch Webinar 18 th February Copyright Khronos Group Page 1

Profiling and Debugging OpenCL Applications with ARM Development Tools. October 2014

Introduction to CUDA Algoritmi e Calcolo Parallelo. Daniele Loiacono

Curriculum 2013 Knowledge Units Pertaining to PDC

Introduction. L25: Modern Compiler Design

SIGGRAPH Briefing August 2014

Introduction to CUDA Algoritmi e Calcolo Parallelo. Daniele Loiacono

GPUfs: Integrating a file system with GPUs

Using Industry Standards to Exploit the Advantages and Resolve the Challenges of Multicore Technology

Press Briefing SIGGRAPH 2015 Neil Trevett Khronos President NVIDIA Vice President Mobile Ecosystem. Copyright Khronos Group Page 1

ROCm: An open platform for GPU computing exploration

FUSION PROCESSORS AND HPC

Generic System Calls for GPUs

Copyright Khronos Group Page 1

Xen and the Art of Virtualization. CSE-291 (Cloud Computing) Fall 2016

Compiling for HSA accelerators with GCC

Copyright Khronos Group 2012 Page 1. OpenCL 1.2. August 2012

Systems I: Programming Abstractions

Graph Streaming Processor

HETEROGENEOUS COMPUTING IN THE MODERN AGE. 1 HiPEAC January, 2012 Public

ECE 8823: GPU Architectures. Objectives

Graphics Hardware 2008

Copyright Khronos Group Page 1

Visualization of OpenCL Application Execution on CPU-GPU Systems

Utilizing Linux Kernel Components in K42 K42 Team modified October 2001

Next Generation OpenGL Neil Trevett Khronos President NVIDIA VP Mobile Copyright Khronos Group Page 1

Building High Performance, Power Efficient Cortex and Mali systems with ARM CoreLink. Robert Kaye

Heterogeneous Architecture. Luca Benini

OpenCL Press Conference

The Architecture of Virtual Machines Lecture for the Embedded Systems Course CSD, University of Crete (April 29, 2014)

Multi2sim Kepler: A Detailed Architectural GPU Simulator

Optimizing Cache Coherent Subsystem Architecture for Heterogeneous Multicore SoCs

RapidIO.org Update. Mar RapidIO.org 1

ARM Security Solutions and Numonyx Authenticated Flash

Next Generation Visual Computing

Parallel Programming Principle and Practice. Lecture 9 Introduction to GPGPUs and CUDA Programming Model

O p t i m i z e d U E F I I m p l e m e n t a t i o n o n I n t e l X e o n B a s e d O C P P l a t f o r m

Parallel Programming on Larrabee. Tim Foley Intel Corp

The Next Steps in the Evolution of ARM Cortex-M

High Performance Computing on GPUs using NVIDIA CUDA

Copyright Khronos Group Page 1

IBM s statements regarding its plans, directions, and intent are subject to change or withdrawal without notice at IBM s sole discretion.

The Bifrost GPU architecture and the ARM Mali-G71 GPU

Bringing it all together: The challenge in delivering a complete graphics system architecture. Chris Porthouse

RapidIO.org Update.

Introduction to RISC-V

INTRODUCTION TO OPENCL TM A Beginner s Tutorial. Udeepta Bordoloi AMD

Press Briefing SIGGRAPH 2015 Neil Trevett Khronos President NVIDIA Vice President Mobile Ecosystem. Copyright Khronos Group Page 1

OpenCL Base Course Ing. Marco Stefano Scroppo, PhD Student at University of Catania

ad-heap: an Efficient Heap Data Structure for Asymmetric Multicore Processors

LINUX CONTAINERS. Where Enterprise Meets Embedded Operating Environments WHEN IT MATTERS, IT RUNS ON WIND RIVER

Copyright Khronos Group Page 1. OpenCL BOF SIGGRAPH 2013

Application Programming

CUDA Programming Model

Operating System. Chapter 4. Threads. Lynn Choi School of Electrical Engineering

Predictable Interrupt Management and Scheduling in the Composite Component-based System

GPUfs: Integrating a file system with GPUs

HSAemu A Full System Emulator for HSA Platforms

Update on Khronos Open Standard APIs for Vision Processing Neil Trevett Khronos President NVIDIA Vice President Mobile Ecosystem

10 Steps to Virtualization

Design Patterns for Real-Time Computer Music Systems

Martin Kruliš, v

MTAPI: Parallel Programming for Embedded Multicore Systems

The Challenges of System Design. Raising Performance and Reducing Power Consumption

CUDA GPGPU Workshop 2012

Navigating the Vision API Jungle: Which API Should You Use and Why? Embedded Vision Summit, May 2015

For use by students enrolled in #71251 CSE430 Fall 2012 at Arizona State University. Do not use if not enrolled.

CUDA PROGRAMMING MODEL Chaithanya Gadiyam Swapnil S Jadhav

GPU Fundamentals Jeff Larkin November 14, 2016

ROEVER ENGINEERING COLLEGE DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING

Process Scheduling Queues

Accelerating Vision Processing

Take GPU Processing Power Beyond Graphics with Mali GPU Computing

Step into Green Computing Consortium

Software Driven Verification at SoC Level. Perspec System Verifier Overview

Porting Fabric Engine to NVIDIA Unified Memory: A Case Study. Peter Zion Chief Architect Fabric Engine Inc.

Transcription:

CLICK TO EDIT MASTER TITLE STYLE Second level THE HETEROGENEOUS SYSTEM ARCHITECTURE ITS (NOT) ALL ABOUT THE GPU PAUL BLINZER, FELLOW, HSA SYSTEM SOFTWARE, AMD SYSTEM ARCHITECTURE WORKGROUP CHAIR, HSA FOUNDATION 1 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

Single-thread Performance Throughput Performance Targeted Application Performance THREE CLICK ERAS TO EDIT OF PROCESSOR MASTER TITLE PERFORMANCE STYLE Single-Core Era Era Enabled by: Moore s Observation Voltage Scaling Micro-Architecture Constrained by: Power Complexity Second level Multi-Core Enabled by: Moore s Observation Desire for Throughput 20 years of SMP arch Constrained by: Power Parallel SW availability Scalability Heterogeneous Systems Era Enabled by: Moore s Observation Abundant data parallelism Power efficient data parallel processing (GPUs) Constrained by: Programming models Communication overheads o we are here? o we are here o we are here Time Time (# of Processors) Time (Data-parallel exploitation) 2 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

THE CLICK PROGRESS TO EDIT MASTER OF FLOATING TITLE POINT STYLE PERFORMANCE Second level 3 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

WHY CLICK THE TO DIFFERENCE? EDIT MASTER ITS TITLE ABOUT STYLE THE PURPOSE CPUs: built for general purpose, serial instruction threads, often high data locality, Click to lot s edit of Master conditional text styles execution and dealing with data interdependency CPU cache hierarchy is focused on general purpose data access from/to execution units, feeding back Second level previously computed data to the execution units with very low latency Comparatively few registers (vs GPUs), but large caches keep often used arbitrary data close to the execution units Fourth level Execution model Fifth is synchronous level GPUs: purpose-built for highly data parallel, latency tolerant workloads Apply the same sequence of instructions over and over on data with little variation but high throughput ( streaming data ), passing the data from one processing stage to another (latency tolerance) Compute units have a relatively large register file store Using a lot of specialty caches (constant cache, Texture Cache, etc), data caches optimized for SW data prefetch separate memory spaces (e.g. GPU local memory) Execution model is asynchronous 4 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

THE CLICK HSA TO (= EDIT HETEROGENEOUS MASTER TITLE SYSTEM STYLE ARCHITECTURE) GOALS To bring accelerators forward as a first class processor within the system Unified process address space access across all processors Operates Second in level pageable system memory Full memory coherency between the CPU and HSA components Well-defined Fourth relaxed level consistency memory model suited for many high level languages User mode dispatch/scheduling (eliminates drivers from the dispatch path) QoS through pre-emption and context switching* It is not dictating a specific platform or component microarchitecture! It defines the what needs to be supported, not the how it needs to be supported Focus is on providing a robust data parallel application execution infrastructure for regular languages 5 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

THE CLICK GOALS TO EDIT OF HSA, MASTER PART TITLE 2 STYLE To attract mainstream programmers By making it easier writing data parallel code Support Second broader level set of languages beyond traditional GPGPU Languages Support Third for Task level Parallel & Nested Data Parallel Runtimes Rich debugging Fourth and level performance analysis support Create a platform architecture accessible for all accelerator types, not only GPU Focused on the APU/SOC and compute, but not preventing other accelerator types to participate But one can t do it alone 6 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

WHAT CLICK IS TO THE EDIT HSA MASTER FOUNDATION? TITLE STYLE This is the short version The HSA Foundation is a not-for-profit consortium of SOC and SOC IP Second vendors, level OEMs, academia, OSVs and ISVs defining a consistent heterogeneous platform architecture to make it dramatically easier to program heterogeneous parallel devices It spans multiple Fifth host level platform architectures and programmable data parallel components collaborating within the same HSA system architecture - CPU: x86, ARM, MIPS, - Accelerator types: GPUs, DSPs, Specifications define HW & SW platform requirements for applications to target the feature set from high level languages and APIs The System Architecture specification defines the required component and platform features for HSA compliant components Other HSA specifications leverage these properties for the SW definition Feb 2014 7 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

THE CLICK KEY TO DELIVERABLES EDIT MASTER OF TITLE THE STYLE HSA FOUNDATION IP, Specifications and API s The Click primary to edit specifications Master text styles are The HSA Platform System Architecture Specification Second level Focus on hardware requirements and low level system software Support Small Model (32bit) and Large Model ( 64bit) The HSA Programmer Reference Manual Definition of an HSAIL Virtual ISA Binary format (BRIG) Compiler writers guide and Libraries developer guide The HSA System Runtime Specification HSA Tools LLVM to HSAIL Compiler HSAIL Assembler HSA Conformance Suite System Runtime Specification Platform (Software) System Architecture Specification Open source community and reference implementation of SW stack (Linux) by AMD To jump start the ecosystem Allow a single shared implementation where appropriate Enable university research in all areas Tools Conformance Programmer s Reference Manual 8 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

OPENCL CLICK TO EDIT WITH MASTER HSA, NOT TITLE OPENCL STYLE VS HSA! HSA is an optimized platform architecture, which will run OpenCL very well It is a complementary standard, not a competitor to OpenCL It is Second focused level on the hardware and system platform runtime definition more than an API itself It is ready Third to level support many more languages than C/C++, including managed code languages OpenCL on HSA benefits from a rich and consistent platform infrastructure Pointers shared between CPU and GPU (Shared Virtual Memory), Avoidance of wasteful copies Low latency dispatch Improved and consistent memory model Virtual function calls Flexible control flow Exception generation and handling Device and platform atomics 9 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

THE CLICK SYSTEM TO EDIT ARCHITECTURE MASTER TITLE WORKGROUP STYLE OF THE HSA FOUNDATION The membership spans a wide variety of IP and platform architecture Click to edit owners Master text styles Several Second host level platform architectures (x86, ARM, MIPS, ) The specifications define a common set of platform properties provide a dependable hardware and system foundation for application software, libraries and runtimes Fifth level Eliminate weak points in the system software- and hardware architecture of traditional platforms, causing overhead processing data parallel workloads Focusing on a few key concepts (queues, signals, memory) optimized and used throughout the architecture 10 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

THE CLICK SYSTEM TO EDIT ARCHITECTURE MASTER TITLE WORKGROUP STYLE OF THE HSA FOUNDATION The main deliverables are: Well-defined, consistent and dependable memory model all HSA agents operate in Share Second access level to process virtual memory between HSA agents ( ptr-is-ptr ) Low-latency workload dispatch, contained in user-mode queues Scalability across Fourth a level wide range of platforms, from smartphones to clients and servers These properties Fifth are level leveraged in the HSA Programmer s Reference, HSAIL and HSA Runtime specifications 11 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

DESIGN CLICK TO CRITERIA EDIT MASTER FOR THE TITLE HSA STYLE INFRASTRUCTURE HSA defines platform & HW requirements Software can depend on Comprehensive Memory model (well-defined visibility and ordering rules for transactions) Shared Virtual Memory (same page table mapping, HW access enforcement) Second level Page fault capability for every HSA agent Cache Coherency Domains Architected, memory-based signaling and synchronization User Mode Queues, Architected Queuing Language (AQL), Service Queues for runtime callbacks Preemption / Quality of Service* Error Reporting Hardware Debug and profiling support requirements Architected Topology Discovery Thin System SW Layer enables HW features for use by an application runtime Mainly responsible for hw init, access enforcement and resource management Provides a consistent, dependable feature set for application layer through SW primitives 12 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

HSA CLICK SECURITY TO EDIT AND MASTER EXECUTION TITLE STYLE MODEL HSA components operate in the same security infrastructure as the host CPU User and privileged memory distinction Hardware Second enforced level process space isolation Page attributes Third level (Read, write, execute) protections enforced by HW and apply as defined by system software Internally, the platform partitions functionality by privilege level User mode queues can only run AQL packets within the defined process context HSA defines Quality of Service requirements Requires support for mechanisms to schedule both HSA and non-hsa workloads for devices that support both task types with appropriate priority, latency, throughput and scheduling constraints. Context Switch Preempt Terminate and Context Reset 13 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

HSA CLICK COMMAND TO EDIT MASTER AND DISPATCH TITLE STYLE FLOW Second level Application C Application B Optional Dispatch Buffer C B C Hardware Queue B C B C C Accelerator Hardware (GPU) SW view: User-mode dispatches to HW No Kernel Driver overhead Low dispatch times CPU & GPU dispatch APIs HW view: Application A A Hardware Queue A A HW / microcode controlled HW scheduling Architected Queuing Language (AQL) HW-managed protection Hardware Queue 14 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

WHAT CLICK DEFINES TO EDIT MASTER HSA PLATFORMS TITLE STYLE AND COMPONENTS? An HSA compatible platform consists of HSA components and HSA agents Both adhere to the same various system architecture requirements (memory model, etc) HSA Second agents level can submit AQL packets for execution, may be but are not required to be an HSA component (e.g. host CPU Third cores level can be an HSA agent, but step out of the role) HSA components Fourth can level be a target of AQL, by implication they are also HSA agents (= generate AQL packets) Defined in the System Architecture specification The HSAIL and Programmer s Reference Manual specifications define the software execution model Architected mechanisms to enqueue and dispatch workloads from one HSA agent queue to another eliminate the need to use the host CPU for these purposes for a lot of scenarios Architected infrastructure allows exchanging data with non-hsa compliant components in a platform Fundamental data types are naturally aligned 15 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

WHAT CLICK DEFINES TO EDIT MASTER HSA PLATFORMS TITLE STYLE AND COMPONENTS? There are two different machine models ( small and large ) target different functionality levels It takes into account different feature requirements for different platform environments Second level In all cases, the same HSA application programming model is used to target HSA agents and provides the same power efficient Third level and low-latency dispatch mechanisms, synchronization primitives and SW programming model Applications written to target HSA small model machines will generally work on large model machines If the large model platform and host Operating System provides a 32bit process environment Properties Small Machine Model Large Machine Model Platform targets embedded or personal device space (controllers, smartphones, etc.) PC, workstation, cloud Server, etc running more demanding workloads Native pointer size 32bit 64bit (+ 32bit ptr if 32bit processes are supported) Floating point size Half (FP16*), Single (FP32) precision Half (FP16*), Single (FP32), Double (FP64) precision Atomic ops size 32bit 32bit, 64bit *min. Load and store on memory 16 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

THE CLICK HSA TO MEMORY EDIT MASTER MODEL TITLE REQUIREMENTS STYLE A memory model defines how writes by one work item or agent becomes visible Click to edit other Master work items text styles and agents Rules Second that level need to be adhered to by compilers and application threads It defines Third visibility level and ordering rules of write and read events across work items, HSA agents and interactions Fourth with non-hsa level components in the system Important to define scope for performance optimizations in the compiler and allow reordering in the Finalizer HSA is defined to be compatible with C++11, Java and.net Memory Models Relaxed consistency memory model for parallel compute performance Inherently maps to many CPU and device architectures Efficient sequential consistency mechanisms supported to fit high-level language programming models Scope visibility controlled by Load.Acquire / Store.Release & Barrier semantics 17 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

THE CLICK HSA TO MEMORY EDIT MASTER MODEL TITLE REQUIREMENTS STYLE A consistent, full set of atomic operations is available Naturally aligned on size, small machine model supports 32bit, large machine model supports 32bit and 64bit Second level Cache Coherency between HSA agents (& host CPU) is maintained by default Key feature Fourth of the level HSA system & platform environment Defined transition Fifth points level to non-hsa data domains (e.g. graphics, audio, ) 18 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

THE CLICK HSA TO SIGNALING EDIT MASTER INFRASTRUCTURE TITLE STYLE HSA supports hardware-assisted signaling and synchronization primitives Defines Second consistent, level memory based semantics to synchronize with work items processed by HSA agents e.g. 32bit Fourth or 64bit level value, content update, wait on value by HSA agents and AQL packets Typically hardware-assisted, power-efficient & low-latency way to synchronize execution of work items between threads on HSA agents Allows one-to-one and one-to-many signaling The signaling semantics follow general atomicity requirements defined in the memory model System Software, runtime & application SW can use this infrastructure to efficiently build higher-level synchronization primitives like mutexes, semaphores, Source: wikimedia commons 19 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

WHAT CLICK IS TO HSAIL? EDIT MASTER TITLE STYLE HSAIL is the intermediate language for parallel compute in HSA HSAIL is a low level instruction set designed for parallel compute in a shared virtual memory environment. Generated Second level by a high level compiler (LLVM, gcc, Java VM, etc) Compiled Third down level to GPU ISA or other parallel processor ISA by an IHV Finalizer Finalizer may Fourth execute level at run time, install time or build time, depending on platform type HSAIL is SIMT in form and does not dictate hardware microarchitecture HSAIL is designed for fast compile time, moving most optimizations to HL compiler Limited register set avoids full register allocation in Finalizer HSAIL is at the same level as PTX: an intermediate assembly or Virtual Machine Target Represented as bit-code in in a Brig file format with support late binding of libraries. 20 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

LOOKING CLICK TO BEYOND EDIT MASTER THE TRADITIONAL TITLE STYLE GPU LANGUAGES Dynamic Languages are one of the most interesting areas for exploration of heterogeneous Click to edit Master parallel text runtimes styles and have many applications Second level Dynamic Languages have different compilation foundations targeting HSA / HSAIL Java/Scala, Fourth JavaScript, level Dart, etc See Project Sumatra - http://openjdk.java.net/projects/sumatra/ Formal Project for GPU Acceleration for Java in OpenJDK HSA allows efficient support for other standards based language environments like OpenMP, Fortran, GO, Haskell And domain specific languages like Halide, Julia, and many others 21 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

Shared Memory and Coherency Fabric LOOKING CLICK TO BEYOND EDIT MASTER THE DATA TITLE PARALLEL STYLE COMPUTE APPLICATION The initial release of the HSA specifications focuses on data parallel Click to compute edit Master language text styles and application scenarios Focus Second is on level integrating GPUs into the general software infrastructure But the Third next level generations of the specifications may apply to other domains With their domain-specific HW processor language focus By design the HSA infrastructure is quite easy to extend Initial focus is on data parallel compute tasks But other areas of support are under consideration for the future Architected Topology infrastructure allows to reliably identify and address domain specific accelerator capabilities By design the HSA infrastructure is easy to virtualize Programming model does leverage few, simple hardware & platform paradigms (queues, signals, memory) for its operation Future spec work may put additional requirements to cover such environments CPU Audio Processor Video Decode Encode Security Processor GPU Image Processor DSP Fixed Function Accelerator 22 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

IN CLICK SUMMARY TO EDIT MASTER TITLE STYLE HSA is not about a specific API, feature or runtime It is about a paradigm to efficiently access the various heterogeneous components in a system by software It allows Second application level programmers to use the languages of their choice to efficiently implement their code HSA is not about a specific hardware or vendor or Operating System It defines a few Fifth fundamental level requirements and concepts as building blocks software at all levels can depend on HW vendors can efficiently expose their compute acceleration features to software in an architected way OS, runtimes and application frameworks can build efficient data and task parallel runtimes leveraging these Application software can more easily use the right tool for the job through high level language support HSA is an open and flexible concept Collaborative participation through the HSA Foundation is encouraged for companies and academia The first set of standards by the HSA Foundation is either released or imminent This is a good time to engage 23 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

WHERE CLICK TO TO FIND EDIT FURTHER MASTER TITLE INFORMATION STYLE ON HSA? HSA Click Foundation to edit Master Website: text http://www.hsafoundation.com styles The main Second location level for specs, developer info, tools, publications and many things more Also the location Third level to apply for membership Tools are now available at GitHUB HSA Foundation libhsa Assembler and Disassembler https://github.com/hsafoundation/hsail-tools HSAIL Instruction Set Simulator https://github.com/hsafoundation/hsail-instruction-set-simulator HSA ISS Loader Library for Java and C++ for creation and dspatch HSAIL Kernels https://github.com/hsafoundation/okra-interface-to-hsail-simulator More to come soon 24 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

CLICK TO EDIT MASTER TITLE STYLE With thanks to Phil Rogers, Greg Stoner Sasa Marinkovic and others for some materials and feedback Second level Trademark Attribution HSA Foundation, the HSA Foundation logo and combinations thereof are trademarks of HSA Foundation, Inc. in the United States and/or other jurisdictions. Other names used in this presentation are for identification purposes only and may be trademarks of their respective owners. 2014 HSA Foundation, Inc. All rights reserved. 25 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014

ANY CLICK QUESTIONS? TO EDIT MASTER TITLE STYLE Of course there are, so go ahead Second level 26 THE HETEROGENEOUS SYSTEM ARCHITECTURE IT S (NOT) ALL ABOUT THE GPU MARCH 1, 2014