Symmetric Computing. ISC 2015 July John Cazes Texas Advanced Computing Center
|
|
- Rosamond Gibbs
- 5 years ago
- Views:
Transcription
1 Symmetric Computing ISC 2015 July 2015 John Cazes Texas Advanced Computing Center
2 Symmetric Computing Run MPI tasks on both MIC and host Also called heterogeneous computing Two executables are required: CPU MIC Works with Intel MPI (IMPI) and MVAPICH2
3 Definition of a Node A node contains a host component and a MIC component Host refers to the Sandy Bridge component MIC refers to one or two Intel Xeon Phi co-processor cards NODE Host 2x Intel 2.7 GHz E cores MIC 1-2 Intel Xeon PHI SE10P 61 cores / 244 HW threads
4 Environment variables for MIC By default, environment variables are inherited by all MPI tasks Since the MIC has a different architecture, several environment variables must be modified OMP_NUM_THREADS # of threads on MIC LD_LIBRARY_PATH must point to MIC libraries I_MPI_PIN_MODE controls the placement of tasks KMP_AFFINITY controls thread binding
5 Symmetric run on 1 Node 16 tasks on host mpiexec.hydra \ n 16 host localhost./host.exe \ : env OMP_NUM_THREADS 30 \ env LD_LIBRARY_PATH $MIC_LD_LIBRARY_PATH \ env I_MPI_PIN_MODE mpd \ env KMP_AFFINITY balanced \ n 4 host mic0./mic.exe 4 tasks on mic0 Environment variables for MIC tasks
6 Steps to create a symmetric run 1. Compile a host executable and a MIC executable: mpicc openmp o host.exe my_code.c mpicc openmp mmic o mic.exe my_code.c 2. Determine the appropriate number of tasks and threads for both MIC and host: 16 tasks/host 1 thread/mpi task 4 tasks/mic 30 threads/mpi task
7 Steps to create a symmetric run 3. Create a batch script to distribute the job #!/bin/bash # # symmetric.slurm # Generic symmetric script MPI + OpenMP # #SBATCH J symmetric #Job name #SBATCH -o symmetric.%j.out #stdout; %j expands to jobid #SBATCH e symmetric.%j.err #stderr; skip to combine #SBATCH p development #queue #SBATCH N 2 #Number of nodes #SBATCH n 32 #Total number of MPI tasks #SBATCH t 00:30:00 #max time #SBATCH A TG #necessary if multiple projects export MIC_PPN=4 export MIC_OMP_NUM_THREADS=30 ibrun.symm m./mic.exe c./host.exe
8 Steps to create a symmetric run 1. Compile a host executable and a MIC executable 2. Determine the appropriate number of tasks and threads for both MIC and host 3. Create the batch script 4. Submit the batch script sbatch symmetric.slurm
9 Symmetric launcher ibrun.symm Usage: ibrun.symm m./<mic_executable> -c./<cpu_executable> Analog of ibrun for symmetric execution # of MIC tasks and threads are controlled by environment variables MIC_PPN=<# of MPI tasks/mic card> MIC_OMP_NUM_THREADS=<# of OMP threads/mic MPI task> MIC_MY_NSLOTS=<Total # of MIC MPI tasks>
10 Symmetric launcher # of host tasks determined by batch script (same as regular ibrun) ibrun.symm does not support np and n flags Command line arguments may be passed within quotes ibrun.symm m./my_exe.mic args c./my_exe.cpu args
11 Symmetric launcher If the executables require redirection or complicated command lines, a simple shell script may be used: run_mic.sh #!/bin/sh a.out.mic <args> < inputfile run_cpu.sh #!/bin/bash a.out.host <args> < inputfile ibrun.symm m./run_mic.sh c./run_cpu.sh Note: The csh and tcsh shells are not available on MIC. So, the MIC script must begin with #!/bin/sh or #!/bin/bash
12 Symmetric Launcher Example #SBATCH N 4 n 32 export OMP_NUM_THREADS=2 export MIC_OMP_NUM_THREADS=60 export MIC_PPN=2 The MPI tasks will be allocated in consecutive order by node (CPU tasks first, then MIC tasks). For example, the task allocation described by the above script snippet will be: NODE 1 8 host tasks (0-7) 2 MIC tasks (8-9) NODE 2 8 host tasks (10-17) 2 MIC tasks (18-19) NODE 3 8 host tasks (20-27) 2 MIC tasks (28-29) NODE 4 8 host tasks (30-37) 2 MIC tasks (38-39)
13 IMPI Task Binding When using IMPI, the mode of process binding may be controlled with the following environment variable: I_MPI_PIN_MODE=<pinmode> mpd pm lib mpd daemon pins MPI processes at startup (Best performance for MIC) Hydra launcher pins MPI processes at startup (Doesn t appear to work on MIC) MPI library pins processes BUT this does not guarantee colocation of CPU and memory (Default) I_MPI_PIN_MODE=mpd (default for ibrun.symm)
14 IMPI Task Binding You can also lay out tasks across the local cores Explicitly: I_MPI_PIN_PROCESSOR_LIST=<proclist> export I_MPI_PIN_PROCESSOR_LIST=1-7,9-15 Grouped: I_MPI_PIN_PROCESSOR_LIST=<map> export I_MPI_PIN_PROCESSOR=spread bunch scatter spread The processes are mapped as closely as possible on the socket The processes are mapped as remotely as possible to avoid sharing common resources: caches, cores The processes are mapped consecutively with the possibility to not share common resources
15 IMPI Task Binding Be careful when using MIC and host MIC 244 H/W threads and 1 socket Host 16 cores and 2 sockets To set I_MPI_PROCESSOR_LIST for MIC simply use the MIC prefix, e.g. export MIC_I_MPI_PROCESSOR_LIST=1,61,121,181 Turn on debugging to see where tasks are placed: export I_MPI_DEBUG=4 Recommend using IMPI defaults
16 IMPI Task Binding Using IMPI defaults Example: 4 MPI tasks/mic with 60 threads each export KMP_AFFINITY=balanced
17 MVAPICH2 Task Binding Binding is disabled by default HOWEVER Since the Xeon Phi appears as one processor to MVAPICH2, all MPI tasks will be bound to the same hardware threads without explicit binding Example: 4 MPI tasks/mic with 60 threads each export KMP_AFFINITY=balanced
18 MVAPICH2 Task Binding Enable affinity with a direct mapping to avoid overlap Example: 4 MPI tasks/mic with 60 threads each export KMP_AFFINITY=balanced export MV2_MIC_MAPPING=1-60:61-120: :
19 Thread Placement Thread placement may be controlled with the following environment variable MIC_KMP_AFFINITY=<type> compact pack threads close to each other scatter Round-Robin threads to cores compact balanced keep OMP thread ids consecutive (MIC only) scatter explicit use the proclist modifier to pin threads none does not pin threads balanced
20 MPI Performance MPI performance is highly dependent on the path the data takes Imbalance in perfomance is mitigated by the use of a proxy Proxy runs on the host and controls the movement of data through the host Both Intel MPI and MVAPICH2 provide proxy support Source: Configuration and Benchmarks of Peer-to-Peer Communication over Gigabit Ethernet and InfiniBand in a Cluster with Intel Xeon Phi Coprocessors by Vadim Karpusenko and Andrey Vladimirov (Colfax International)
21 Proxy can help quite a bit for both Intel MPI and MVAPICH2 even on one node Pingpong Host to MIC
22 Doesn t help much for latency bound codes Pingpong Host to MIC off node
23 MIC to MIC bandwidth is almost as good as host to MIC Pingpong MIC to MIC off node
24 Intel Trace Analyzer and Collector (ITAC) ITAC is a graphical tool for understanding MPI application behavior, quickly finding bottlenecks, and achieving high performance. Use with Intel and Intel MPI Load the ITAC module: module load itac Trace files provide a timeline of MPI execution More powerful than an MPI profiler such as mpip or IPM Trace files can become quite large Supports symmetric execution on the Xeon Phi
25 Compile and Run for ITAC To instrument for ITAC, the application must be compiled and linked with the trace option for both MIC and host executable. C/C++ : Fortran: mpicc trace mmic -g o mycode.mic mycode.c mpicc trace xavx g o mycode.host mycode.c mpif90 trace mmic o mycode.mic mycode.f90 mpif90 trace xavx o mycode.host mycode.f90 Run normally: ibrun.symm m./mycode.mic c./mycode.host ITAC will generate multiple trace files for each task and a master file named <executable_name>.stf
26 Analyze ITAC Trace files Once the trace files are created, the traceanalyzer used to analyze the results. $ traceanalyzer mycode.stf tool may be Since this is a graphical tool, the user should be logged in with X11 enabled or via a VNC session. SSH: ssh Y stampede.tacc.utexas.edu VNC: Follow the instructions in the Stampede user guide for Remote Desktop Access ( )
27 Example with 2 nodes and 2 MPI tasks per host and MIC ITAC Summary
28 MPI Pingpong benchmark ITAC Flat Profile
29 Note the long MPI_Finalize time for 4 of the 8 tasks These are the MPI tasks running on the MICs. ITAC Event Timeline
30 Collective Parallel I/O Stampede has the Lustre filesystem mounted on the MIC as an NSF mount from the host Results in very poor performance Lustre client is available for the MIC we haven t tested it But, there is a way to get decent parallel I/O performance for collective I/O operations
31 ROMIO Hints Use ROMIO hints to limit the writers of collective I/O to the hosts Example: Running on 2 nodes 2 hosts and 2 MICs c stampede.tacc.utexas.edu c stampede.tacc.utexas.edu c mic0.stampede.tacc.utexas.edu c mic0.stampede.tacc.utexas.edu Set hints to only write from the host $ echo cb_config_list c stampede.tacc.utexas.edu:1,\ c stampede.tacc.utexas.edu:1 > romio_hints $ export ROMIO_HINTS=romio_hints $ ibrun.symm m./test_io.mic c./test_io.host NOTE: Must use hostname returned by MPI_Get_processor_name
32 Balance How to balance the code? Sandy Bridge Xeon Phi Memory 32 GB 8 GB Cores Clock Speed 2.7 GHz 1.1 GHz Memory Bandwidth 51.2 GB/s(x2) 352 GB/s Vector Length 4 DP words 8 DP words
33 Balance Example: Memory balance Balance memory use and performance by using a different # of tasks/threads on host and MIC Host 16 tasks/1 thread/task 2GB/task Xeon PHI 4 tasks/60 threads/task 2GB/task
34 Balance Example: Performance balance Balance performance by tuning the # of tasks and threads on host and MIC Host? tasks/? threads/task?gb/task Xeon PHI? tasks/? threads/task?gb/task
35 MPI with Offload Sections ADVANTAGES Offload Sections may easily be added to MPI/OpenMP codes with directives Intel compiler will automatically detect and compile offloaded sections CAVEATS However, there may be no MPI calls within offload sections Each host task will spawn an offload section Must explicitly place each offload set of threads if you have more than one MPI task per node offload
36 Exercises Exercise 1 Run in MPI symmetric mode using mpiexec.hydra Exercise 2 Run in a symmetric mode using MIC and host Exercise 3 Compare performance with and without proxy Exercise 4 Compare collective I/O performance with and without hints
37 John Cazes For more information:
Symmetric Computing. SC 14 Jerome VIENNE
Symmetric Computing SC 14 Jerome VIENNE viennej@tacc.utexas.edu Symmetric Computing Run MPI tasks on both MIC and host Also called heterogeneous computing Two executables are required: CPU MIC Currently
More informationSymmetric Computing. John Cazes Texas Advanced Computing Center
Symmetric Computing John Cazes Texas Advanced Computing Center Symmetric Computing Run MPI tasks on both MIC and host and across nodes Also called heterogeneous computing Two executables are required:
More informationSymmetric Computing. Jerome Vienne Texas Advanced Computing Center
Symmetric Computing Jerome Vienne Texas Advanced Computing Center Symmetric Computing Run MPI tasks on both MIC and host Also called heterogeneous computing Two executables are required: CPU MIC Currently
More informationUnderstanding Native Mode, Vectorization, and Symmetric MPI. Bill Barth Director of HPC, TACC June 16, 2013
Understanding Native Mode, Vectorization, and Symmetric MPI Bill Barth Director of HPC, TACC June 16, 2013 What is a native application? It is an application built to run exclusively on the MIC coprocessor.
More informationMIC Lab Parallel Computing on Stampede
MIC Lab Parallel Computing on Stampede Aaron Birkland and Steve Lantz Cornell Center for Advanced Computing June 11 & 18, 2013 1 Interactive Launching This exercise will walk through interactively launching
More informationNative Computing and Optimization. Hang Liu December 4 th, 2013
Native Computing and Optimization Hang Liu December 4 th, 2013 Overview Why run native? What is a native application? Building a native application Running a native application Setting affinity and pinning
More informationNative Computing and Optimization on Intel Xeon Phi
Native Computing and Optimization on Intel Xeon Phi ISC 2015 Carlos Rosales carlos@tacc.utexas.edu Overview Why run native? What is a native application? Building a native application Running a native
More informationParallel Applications on Distributed Memory Systems. Le Yan HPC User LSU
Parallel Applications on Distributed Memory Systems Le Yan HPC User Services @ LSU Outline Distributed memory systems Message Passing Interface (MPI) Parallel applications 6/3/2015 LONI Parallel Programming
More informationSCALABLE HYBRID PROTOTYPE
SCALABLE HYBRID PROTOTYPE Scalable Hybrid Prototype Part of the PRACE Technology Evaluation Objectives Enabling key applications on new architectures Familiarizing users and providing a research platform
More informationIntel Parallel Studio XE Cluster Edition - Intel MPI - Intel Traceanalyzer & Collector
Intel Parallel Studio XE Cluster Edition - Intel MPI - Intel Traceanalyzer & Collector A brief Introduction to MPI 2 What is MPI? Message Passing Interface Explicit parallel model All parallelism is explicit:
More informationLab MIC Experiments 4/25/13 TACC
Lab MIC Experiments 4/25/13 TACC # pg. Subject Purpose directory 1 3 5 Offload, Begin (C) (F90) Compile and Run (CPU, MIC, Offload) offload_hello 2 7 Offload, Data Optimize Offload Data Transfers offload_transfer
More informationScalasca support for Intel Xeon Phi. Brian Wylie & Wolfgang Frings Jülich Supercomputing Centre Forschungszentrum Jülich, Germany
Scalasca support for Intel Xeon Phi Brian Wylie & Wolfgang Frings Jülich Supercomputing Centre Forschungszentrum Jülich, Germany Overview Scalasca performance analysis toolset support for MPI & OpenMP
More informationNative Computing and Optimization on the Intel Xeon Phi Coprocessor. John D. McCalpin
Native Computing and Optimization on the Intel Xeon Phi Coprocessor John D. McCalpin mccalpin@tacc.utexas.edu Intro (very brief) Outline Compiling & Running Native Apps Controlling Execution Tuning Vectorization
More informationOverview of Intel Xeon Phi Coprocessor
Overview of Intel Xeon Phi Coprocessor Sept 20, 2013 Ritu Arora Texas Advanced Computing Center Email: rauta@tacc.utexas.edu This talk is only a trailer A comprehensive training on running and optimizing
More informationRunning HARMONIE on Xeon Phi Coprocessors
Running HARMONIE on Xeon Phi Coprocessors Enda O Brien Irish Centre for High-End Computing Disclosure Intel is funding ICHEC to port & optimize some applications, including HARMONIE, to Xeon Phi coprocessors.
More informationJURECA Tuning for the platform
JURECA Tuning for the platform Usage of ParaStation MPI 2017-11-23 Outline ParaStation MPI Compiling your program Running your program Tuning parameters Resources 2 ParaStation MPI Based on MPICH (3.2)
More informationAdvanced Job Launching. mapping applications to hardware
Advanced Job Launching mapping applications to hardware A Quick Recap - Glossary of terms Hardware This terminology is used to cover hardware from multiple vendors Socket The hardware you can touch and
More informationNative Computing and Optimization on the Intel Xeon Phi Coprocessor. Lars Koesterke John D. McCalpin
Native Computing and Optimization on the Intel Xeon Phi Coprocessor Lars Koesterke John D. McCalpin lars@tacc.utexas.edu mccalpin@tacc.utexas.edu Intro (very brief) Outline Compiling & Running Native Apps
More informationLab MIC Offload Experiments 7/22/13 MIC Advanced Experiments TACC
Lab MIC Offload Experiments 7/22/13 MIC Advanced Experiments TACC # pg. Subject Purpose directory 1 3 5 Offload, Begin (C) (F90) Compile and Run (CPU, MIC, Offload) offload_hello 2 7 Offload, Data Optimize
More informationNative Computing and Optimization on the Intel Xeon Phi Coprocessor. John D. McCalpin
Native Computing and Optimization on the Intel Xeon Phi Coprocessor John D. McCalpin mccalpin@tacc.utexas.edu Outline Overview What is a native application? Why run native? Getting Started: Building a
More informationIntroduction to Intel Xeon Phi programming techniques. Fabio Affinito Vittorio Ruggiero
Introduction to Intel Xeon Phi programming techniques Fabio Affinito Vittorio Ruggiero Outline High level overview of the Intel Xeon Phi hardware and software stack Intel Xeon Phi programming paradigms:
More informationIntel Xeon Phi Coprocessor
Intel Xeon Phi Coprocessor A guide to using it on the Cray XC40 Terminology Warning: may also be referred to as MIC or KNC in what follows! What are Intel Xeon Phi Coprocessors? Hardware designed to accelerate
More informationIntroduction to the Intel Xeon Phi on Stampede
June 10, 2014 Introduction to the Intel Xeon Phi on Stampede John Cazes Texas Advanced Computing Center Stampede - High Level Overview Base Cluster (Dell/Intel/Mellanox): Intel Sandy Bridge processors
More informationIntroduc)on to Xeon Phi
Introduc)on to Xeon Phi MIC Training Event at TACC Lars Koesterke Xeon Phi MIC Xeon Phi = first product of Intel s Many Integrated Core (MIC) architecture Co- processor PCI Express card Stripped down Linux
More informationIntroduction to Xeon Phi. Bill Barth January 11, 2013
Introduction to Xeon Phi Bill Barth January 11, 2013 What is it? Co-processor PCI Express card Stripped down Linux operating system Dense, simplified processor Many power-hungry operations removed Wider
More informationMessage Passing Interface (MPI) on Intel Xeon Phi coprocessor
Message Passing Interface (MPI) on Intel Xeon Phi coprocessor Special considerations for MPI on Intel Xeon Phi and using the Intel Trace Analyzer and Collector Gergana Slavova gergana.s.slavova@intel.com
More informationTools for Intel Xeon Phi: VTune & Advisor Dr. Fabio Baruffa - LRZ,
Tools for Intel Xeon Phi: VTune & Advisor Dr. Fabio Baruffa - fabio.baruffa@lrz.de LRZ, 27.6.- 29.6.2016 Architecture Overview Intel Xeon Processor Intel Xeon Phi Coprocessor, 1st generation Intel Xeon
More informationOur new HPC-Cluster An overview
Our new HPC-Cluster An overview Christian Hagen Universität Regensburg Regensburg, 15.05.2009 Outline 1 Layout 2 Hardware 3 Software 4 Getting an account 5 Compiling 6 Queueing system 7 Parallelization
More informationAn Introduction to the Intel Xeon Phi. Si Liu Feb 6, 2015
Training Agenda Session 1: Introduction 8:00 9:45 Session 2: Native: MIC stand-alone 10:00-11:45 Lunch break Session 3: Offload: MIC as coprocessor 1:00 2:45 Session 4: Symmetric: MPI 3:00 4:45 1 Last
More informationCluster Clonetroop: HowTo 2014
2014/02/25 16:53 1/13 Cluster Clonetroop: HowTo 2014 Cluster Clonetroop: HowTo 2014 This section contains information about how to access, compile and execute jobs on Clonetroop, Laboratori de Càlcul Numeric's
More informationPractical Introduction to
1 2 Outline of the workshop Practical Introduction to What is ScaleMP? When do we need it? How do we run codes on the ScaleMP node on the ScaleMP Guillimin cluster? How to run programs efficiently on ScaleMP?
More informationKlaus-Dieter Oertel, May 28 th 2013 Software and Services Group Intel Corporation
S c i c o m P 2 0 1 3 T u t o r i a l Intel Xeon Phi Product Family Intel MPI Klaus-Dieter Oertel, May 28 th 2013 Software and Services Group Intel Corporation Legal Disclaimer INFORMATION IN THIS DOCUMENT
More informationLAB. Preparing for Stampede: Programming Heterogeneous Many-Core Supercomputers
LAB Preparing for Stampede: Programming Heterogeneous Many-Core Supercomputers Dan Stanzione, Lars Koesterke, Bill Barth, Kent Milfeld dan/lars/bbarth/milfeld@tacc.utexas.edu XSEDE 12 July 16, 2012 1 Discovery
More informationNon-uniform memory access (NUMA)
Non-uniform memory access (NUMA) Memory access between processor core to main memory is not uniform. Memory resides in separate regions called NUMA domains. For highest performance, cores should only access
More informationIntroduction to CINECA Computer Environment
Introduction to CINECA Computer Environment Today you will learn... Basic commands for UNIX environment @ CINECA How to submitt your job to the PBS queueing system on Eurora Tutorial #1: Example: launch
More informationAdvanced MPI: MPI Tuning or How to Save SUs by Optimizing your MPI Library! The lab.
Advanced MPI: MPI Tuning or How to Save SUs by Optimizing your MPI Library! The lab. Jérôme VIENNE viennej@tacc.utexas.edu Texas Advanced Computing Center (TACC). University of Texas at Austin Tuesday
More informationReusing this material
XEON PHI BASICS Reusing this material This work is licensed under a Creative Commons Attribution- NonCommercial-ShareAlike 4.0 International License. http://creativecommons.org/licenses/by-nc-sa/4.0/deed.en_us
More informationNUMA Control for Hybrid Applications. Hang Liu TACC February 7 th, 2011
NUMA Control for Hybrid Applications Hang Liu TACC February 7 th, 2011 Hybrid Applications Typical definition of hybrid application Uses both message passing (MPI) and a form of shared memory algorithm
More informationStampede User Environment
Version DJ2013-01 8 Jan 2012 Stampede User Environment Doug James 10 Jan 2012 Overview Effective users, good citizens Getting Started Access to Stampede Getting Acquainted A Tour of Stampede Getting Work
More informationHands-on. MPI basic exercises
WIFI XSF-UPC: Username: xsf.convidat Password: 1nt3r3st3l4r WIFI EDUROAM: Username: roam06@bsc.es Password: Bsccns.4 MareNostrum III User Guide http://www.bsc.es/support/marenostrum3-ug.pdf Remember to
More informationIntroduc)on to Xeon Phi
Introduc)on to Xeon Phi ACES Aus)n, TX Dec. 04 2013 Kent Milfeld, Luke Wilson, John McCalpin, Lars Koesterke TACC What is it? Co- processor PCI Express card Stripped down Linux opera)ng system Dense, simplified
More informationNUMA Control for Hybrid Applications. Hang Liu & Xiao Zhu TACC September 20, 2013
NUMA Control for Hybrid Applications Hang Liu & Xiao Zhu TACC September 20, 2013 Parallel Paradigms OpenMP MPI Run a bunch of threads in shared memory(spawned by a single a.out). Run a bunch of a.out s
More informationCompiling applications for the Cray XC
Compiling applications for the Cray XC Compiler Driver Wrappers (1) All applications that will run in parallel on the Cray XC should be compiled with the standard language wrappers. The compiler drivers
More informationAmbiente CINECA: moduli, job scripts, PBS. A. Grottesi (CINECA)
Ambiente HPC @ CINECA: moduli, job scripts, PBS A. Grottesi (CINECA) Bologna 2017 In this tutorial you will learn... How to get familiar with UNIX environment @ CINECA How to submit your job to the PBS
More informationMulticore Performance and Tools. Part 1: Topology, affinity, clock speed
Multicore Performance and Tools Part 1: Topology, affinity, clock speed Tools for Node-level Performance Engineering Gather Node Information hwloc, likwid-topology, likwid-powermeter Affinity control and
More informationHands-on with Intel Xeon Phi
Hands-on with Intel Xeon Phi Lab 2: Native Computing and Vector Reports Bill Barth Kent Milfeld Dan Stanzione 1 Lab 2 What you will learn about Evaluating and Analyzing vector performance. Controlling
More informationIntel Knights Landing Hardware
Intel Knights Landing Hardware TACC KNL Tutorial IXPUG Annual Meeting 2016 PRESENTED BY: John Cazes Lars Koesterke 1 Intel s Xeon Phi Architecture Leverages x86 architecture Simpler x86 cores, higher compute
More informationIntel MPI Library for Windows* OS. Developer Guide
Intel MPI Library for Windows* OS Developer Guide Contents Legal Information... 4 1. Introduction... 5 1.1. Introducing Intel MPI Library... 5 1.2. Conventions and Symbols... 5 1.3. Related Information...
More informationDebugging Intel Xeon Phi KNC Tutorial
Debugging Intel Xeon Phi KNC Tutorial Last revised on: 10/7/16 07:37 Overview: The Intel Xeon Phi Coprocessor 2 Debug Library Requirements 2 Debugging Host-Side Applications that Use the Intel Offload
More informationDesigning Optimized MPI Broadcast and Allreduce for Many Integrated Core (MIC) InfiniBand Clusters
Designing Optimized MPI Broadcast and Allreduce for Many Integrated Core (MIC) InfiniBand Clusters K. Kandalla, A. Venkatesh, K. Hamidouche, S. Potluri, D. Bureddy and D. K. Panda Presented by Dr. Xiaoyi
More informationIntroduction to Unix Environment: modules, job scripts, PBS. N. Spallanzani (CINECA)
Introduction to Unix Environment: modules, job scripts, PBS N. Spallanzani (CINECA) Bologna PATC 2016 In this tutorial you will learn... How to get familiar with UNIX environment @ CINECA How to submit
More informationPreliminary Experiences with the Uintah Framework on on Intel Xeon Phi and Stampede
Preliminary Experiences with the Uintah Framework on on Intel Xeon Phi and Stampede Qingyu Meng, Alan Humphrey, John Schmidt, Martin Berzins Thanks to: TACC Team for early access to Stampede J. Davison
More informationNo Time to Read This Book?
Chapter 1 No Time to Read This Book? We know what it feels like to be under pressure. Try out a few quick and proven optimization stunts described below. They may provide a good enough performance gain
More informationChoosing Resources Wisely Plamen Krastev Office: 38 Oxford, Room 117 FAS Research Computing
Choosing Resources Wisely Plamen Krastev Office: 38 Oxford, Room 117 Email:plamenkrastev@fas.harvard.edu Objectives Inform you of available computational resources Help you choose appropriate computational
More informationIntroduction to GALILEO
Introduction to GALILEO Parallel & production environment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Alessandro Grottesi a.grottesi@cineca.it SuperComputing Applications and
More informationAusAn TACC. OpenMP. TACC OpenMP Team
UT AusAn TACC OpenMP Thread Affinity TACC OpenMP Team milfeld/lars/agomez@tacc.utexas.edu Learning ObjecAve Understand why you should care about affinity Basic concepts OMP_PLACES OMP_PROC_BIND Standard
More informationIntroduction to GALILEO
November 27, 2016 Introduction to GALILEO Parallel & production environment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it SuperComputing Applications and Innovation Department
More informationBenchmark results on Knight Landing (KNL) architecture
Benchmark results on Knight Landing (KNL) architecture Domenico Guida, CINECA SCAI (Bologna) Giorgio Amati, CINECA SCAI (Roma) Roma 23/10/2017 KNL, BDW, SKL A1 BDW A2 KNL A3 SKL cores per node 2 x 18 @2.3
More informationBenchmark results on Knight Landing architecture
Benchmark results on Knight Landing architecture Domenico Guida, CINECA SCAI (Bologna) Giorgio Amati, CINECA SCAI (Roma) Milano, 21/04/2017 KNL vs BDW A1 BDW A2 KNL cores per node 2 x 18 @2.3 GHz 1 x 68
More informationBlue Gene/Q User Workshop. User Environment & Job submission
Blue Gene/Q User Workshop User Environment & Job submission Topics Blue Joule User Environment Loadleveler Task Placement & BG/Q Personality 2 Blue Joule User Accounts Home directories organised on a project
More informationOpenMP Exercises. These exercises will introduce you to using OpenMP for parallel programming. There are four exercises:
OpenMP Exercises These exercises will introduce you to using OpenMP for parallel programming. There are four exercises: 1. OMP Hello World 2. Worksharing Loop 3. OMP Functions 4. Hand-coding vs. MKL To
More informationExercises: Abel/Colossus and SLURM
Exercises: Abel/Colossus and SLURM November 08, 2016 Sabry Razick The Research Computing Services Group, USIT Topics Get access Running a simple job Job script Running a simple job -- qlogin Customize
More informationBring your application to a new era:
Bring your application to a new era: learning by example how to parallelize and optimize for Intel Xeon processor and Intel Xeon Phi TM coprocessor Manel Fernández, Roger Philp, Richard Paul Bayncore Ltd.
More informationMVAPICH MPI and Open MPI
CHAPTER 6 The following sections appear in this chapter: Introduction, page 6-1 Initial Setup, page 6-2 Configure SSH, page 6-2 Edit Environment Variables, page 6-5 Perform MPI Bandwidth Test, page 6-8
More informationBefore We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop
Before We Start Sign in hpcxx account slips Windows Users: Download PuTTY Google PuTTY First result Save putty.exe to Desktop Research Computing at Virginia Tech Advanced Research Computing Compute Resources
More informationAdvanced Threading and Optimization
Mikko Byckling, CSC Michael Klemm, Intel Advanced Threading and Optimization February 24-26, 2015 PRACE Advanced Training Centre CSC IT Center for Science Ltd, Finland!$omp parallel do collapse(3) do p4=1,p4d
More informationPRACE PATC Course: Intel MIC Programming Workshop, MKL. Ostrava,
PRACE PATC Course: Intel MIC Programming Workshop, MKL Ostrava, 7-8.2.2017 1 Agenda A quick overview of Intel MKL Usage of MKL on Xeon Phi Compiler Assisted Offload Automatic Offload Native Execution Hands-on
More informationAccelerator Programming Lecture 1
Accelerator Programming Lecture 1 Manfred Liebmann Technische Universität München Chair of Optimal Control Center for Mathematical Sciences, M17 manfred.liebmann@tum.de January 11, 2016 Accelerator Programming
More informationIntel Manycore Testing Lab (MTL) - Linux Getting Started Guide
Intel Manycore Testing Lab (MTL) - Linux Getting Started Guide Introduction What are the intended uses of the MTL? The MTL is prioritized for supporting the Intel Academic Community for the testing, validation
More informationSlurm and Abel job scripts. Katerina Michalickova The Research Computing Services Group SUF/USIT October 23, 2012
Slurm and Abel job scripts Katerina Michalickova The Research Computing Services Group SUF/USIT October 23, 2012 Abel in numbers Nodes - 600+ Cores - 10000+ (1 node->2 processors->16 cores) Total memory
More informationScientific Computing with Intel Xeon Phi Coprocessors
Scientific Computing with Intel Xeon Phi Coprocessors Andrey Vladimirov Colfax International HPC Advisory Council Stanford Conference 2015 Compututing with Xeon Phi Welcome Colfax International, 2014 Contents
More informationIntroduc)on to Hyades
Introduc)on to Hyades Shawfeng Dong Department of Astronomy & Astrophysics, UCSSC Hyades 1 Hardware Architecture 2 Accessing Hyades 3 Compu)ng Environment 4 Compiling Codes 5 Running Jobs 6 Visualiza)on
More informationFor Distributed Performance
For Distributed Performance Intel Parallel Studio XE 2017 development suite Empowering Faster Code Faster Delivering HPC Development Solutions Over 20 years Industry Collaboration on Standards PARALLELISM
More informationEffective Use of CCV Resources
Effective Use of CCV Resources Mark Howison User Services & Support This talk... Assumes you have some familiarity with a Unix shell Provides examples and best practices for typical usage of CCV systems
More informationA Unified Approach to Heterogeneous Architectures Using the Uintah Framework
DOE for funding the CSAFE project (97-10), DOE NETL, DOE NNSA NSF for funding via SDCI and PetaApps A Unified Approach to Heterogeneous Architectures Using the Uintah Framework Qingyu Meng, Alan Humphrey
More informationIntel MPI Cluster Edition on Graham A First Look! Doug Roberts
Intel MPI Cluster Edition on Graham A First Look! Doug Roberts SHARCNET / COMPUTE CANADA Intel Parallel Studio XE 2016 Update 4 Cluster Edition for Linux 1. Intel(R) MPI Library 5.1 Update 3 Cluster Ed
More informationSubmitting and running jobs on PlaFRIM2 Redouane Bouchouirbat
Submitting and running jobs on PlaFRIM2 Redouane Bouchouirbat Summary 1. Submitting Jobs: Batch mode - Interactive mode 2. Partition 3. Jobs: Serial, Parallel 4. Using generic resources Gres : GPUs, MICs.
More informationGetting Performance from OpenMP Programs on NUMA Architectures
Getting Performance from OpenMP Programs on NUMA Architectures Christian Terboven, RWTH Aachen University terboven@itc.rwth-aachen.de EU H2020 Centre of Excellence (CoE) 1 October 2015 31 March 2018 Grant
More informationLeibniz Supercomputer Centre. Movie on YouTube
SuperMUC @ Leibniz Supercomputer Centre Movie on YouTube Peak Performance Peak performance: 3 Peta Flops 3*10 15 Flops Mega 10 6 million Giga 10 9 billion Tera 10 12 trillion Peta 10 15 quadrillion Exa
More informationRunning applications on the Cray XC30
Running applications on the Cray XC30 Running on compute nodes By default, users do not access compute nodes directly. Instead they launch jobs on compute nodes using one of three available modes: 1. Extreme
More informationHow to for compiling and running MPI Programs. Prepared by Kiriti Venkat
How to for compiling and running MPI Programs. Prepared by Kiriti Venkat What is MPI? MPI stands for Message Passing Interface MPI is a library specification of message-passing, proposed as a standard
More informationWindows-HPC Environment at RWTH Aachen University
Windows-HPC Environment at RWTH Aachen University Christian Terboven, Samuel Sarholz {terboven, sarholz}@rz.rwth-aachen.de Center for Computing and Communication RWTH Aachen University PPCES 2009 March
More informationCode optimization in a 3D diffusion model
Code optimization in a 3D diffusion model Roger Philp Intel HPC Software Workshop Series 2016 HPC Code Modernization for Intel Xeon and Xeon Phi February 18 th 2016, Barcelona Agenda Background Diffusion
More informationIntroduction to the Xeon Phi programming model. Fabio AFFINITO, CINECA
Introduction to the Xeon Phi programming model Fabio AFFINITO, CINECA What is a Xeon Phi? MIC = Many Integrated Core architecture by Intel Other names: KNF, KNC, Xeon Phi... Not a CPU (but somewhat similar
More informationHPC Architectures. Types of resource currently in use
HPC Architectures Types of resource currently in use Reusing this material This work is licensed under a Creative Commons Attribution- NonCommercial-ShareAlike 4.0 International License. http://creativecommons.org/licenses/by-nc-sa/4.0/deed.en_us
More informationIntroduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende
Introduction to the NCAR HPC Systems 25 May 2018 Consulting Services Group Brian Vanderwende Topics to cover Overview of the NCAR cluster resources Basic tasks in the HPC environment Accessing pre-built
More informationNew User Seminar: Part 2 (best practices)
New User Seminar: Part 2 (best practices) General Interest Seminar January 2015 Hugh Merz merz@sharcnet.ca Session Outline Submitting Jobs Minimizing queue waits Investigating jobs Checkpointing Efficiency
More informationNUMA-aware OpenMP Programming
NUMA-aware OpenMP Programming Dirk Schmidl IT Center, RWTH Aachen University Member of the HPC Group schmidl@itc.rwth-aachen.de Christian Terboven IT Center, RWTH Aachen University Deputy lead of the HPC
More informationSlurm and Abel job scripts. Katerina Michalickova The Research Computing Services Group SUF/USIT November 13, 2013
Slurm and Abel job scripts Katerina Michalickova The Research Computing Services Group SUF/USIT November 13, 2013 Abel in numbers Nodes - 600+ Cores - 10000+ (1 node->2 processors->16 cores) Total memory
More informationParallel Programming on Ranger and Stampede
Parallel Programming on Ranger and Stampede Steve Lantz Senior Research Associate Cornell CAC Parallel Computing at TACC: Ranger to Stampede Transition December 11, 2012 What is Stampede? NSF-funded XSEDE
More informationThe Stampede is Coming: A New Petascale Resource for the Open Science Community
The Stampede is Coming: A New Petascale Resource for the Open Science Community Jay Boisseau Texas Advanced Computing Center boisseau@tacc.utexas.edu Stampede: Solicitation US National Science Foundation
More informationTech Computer Center Documentation
Tech Computer Center Documentation Release 0 TCC Doc February 17, 2014 Contents 1 TCC s User Documentation 1 1.1 TCC SGI Altix ICE Cluster User s Guide................................ 1 i ii CHAPTER 1
More informationTo connect to the cluster, simply use a SSH or SFTP client to connect to:
RIT Computer Engineering Cluster The RIT Computer Engineering cluster contains 12 computers for parallel programming using MPI. One computer, cluster-head.ce.rit.edu, serves as the master controller or
More informationDmitry Durnov 15 February 2017
Cовременные тенденции разработки высокопроизводительных приложений Dmitry Durnov 15 February 2017 Agenda Modern cluster architecture Node level Cluster level Programming models Tools 2/20/2017 2 Modern
More informationSupercomputing in Plain English Exercise #6: MPI Point to Point
Supercomputing in Plain English Exercise #6: MPI Point to Point In this exercise, we ll use the same conventions and commands as in Exercises #1, #2, #3, #4 and #5. You should refer back to the Exercise
More informationOBTAINING AN ACCOUNT:
HPC Usage Policies The IIA High Performance Computing (HPC) System is managed by the Computer Management Committee. The User Policies here were developed by the Committee. The user policies below aim to
More informationProgress Report on Transparent Checkpointing for Supercomputing
Progress Report on Transparent Checkpointing for Supercomputing Jiajun Cao, Rohan Garg College of Computer and Information Science, Northeastern University {jiajun,rohgarg}@ccs.neu.edu August 21, 2015
More informationIntel MPI Library for Linux* OS. Developer Guide
Intel MPI Library for Linux* OS Developer Guide Contents Legal Information... 4 1. Introduction... 5 1.1. Introducing Intel MPI Library... 5 1.2. Conventions and Symbols... 5 1.3. Related Information...
More informationSimulation using MIC co-processor on Helios
Simulation using MIC co-processor on Helios Serhiy Mochalskyy, Roman Hatzky PRACE PATC Course: Intel MIC Programming Workshop High Level Support Team Max-Planck-Institut für Plasmaphysik Boltzmannstr.
More informationHow to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions
How to run applications on Aziz supercomputer Mohammad Rafi System Administrator Fujitsu Technology Solutions Agenda Overview Compute Nodes Storage Infrastructure Servers Cluster Stack Environment Modules
More information