Using Compute Canada. Masao Fujinaga Information Services and Technology University of Alberta

Similar documents
Graham vs legacy systems

Introduction to HPC2N

HPC Introductory Course - Exercises

Duke Compute Cluster Workshop. 3/28/2018 Tom Milledge rc.duke.edu

For Dr Landau s PHYS8602 course

How to Use a Supercomputer - A Boot Camp

Duke Compute Cluster Workshop. 11/10/2016 Tom Milledge h:ps://rc.duke.edu/

Introduction to SLURM on the High Performance Cluster at the Center for Computational Research

Working with Shell Scripting. Daniel Balagué

Duke Compute Cluster Workshop. 10/04/2018 Tom Milledge rc.duke.edu

Introduction to HPC2N

Introduction to the Cluster

Sherlock for IBIIS. William Law Stanford Research Computing

Introduction to High-Performance Computing (HPC)

Introduction to the SHARCNET Environment May-25 Pre-(summer)school webinar Speaker: Alex Razoumov University of Ontario Institute of Technology

Introduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende

HPC Workshop. Nov. 9, 2018 James Coyle, PhD Dir. Of High Perf. Computing

Computing with the Moore Cluster

Introduction to GACRC Teaching Cluster

Supercomputing environment TMA4280 Introduction to Supercomputing

An Introduction to Gauss. Paul D. Baines University of California, Davis November 20 th 2012

Slurm and Abel job scripts. Katerina Michalickova The Research Computing Services Group SUF/USIT November 13, 2013

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing

Exercises: Abel/Colossus and SLURM

Using Cartesius and Lisa. Zheng Meyer-Zhao - Consultant Clustercomputing

Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop

Using a Linux System 6

Slurm basics. Summer Kickstart June slide 1 of 49

New User Tutorial. OSU High Performance Computing Center

Linux Tutorial. Ken-ichi Nomura. 3 rd Magics Materials Software Workshop. Gaithersburg Marriott Washingtonian Center November 11-13, 2018

Introduction to HPC2N

Introduction to GACRC Teaching Cluster

Slurm and Abel job scripts. Katerina Michalickova The Research Computing Services Group SUF/USIT October 23, 2012

Submitting and running jobs on PlaFRIM2 Redouane Bouchouirbat

Our Workshop Environment

Our new HPC-Cluster An overview

Batch Usage on JURECA Introduction to Slurm. May 2016 Chrysovalantis Paschoulas HPS JSC

Our Workshop Environment

Tech Computer Center Documentation

Introduction to HPC Using zcluster at GACRC

Scientific Computing in practice

Introduction to Joker Cyber Infrastructure Architecture Team CIA.NMSU.EDU

Introduction to PICO Parallel & Production Enviroment

Introduction to GALILEO

High Performance Computing (HPC) Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC

Using Sapelo2 Cluster at the GACRC

ICS-ACI System Basics

XSEDE New User Tutorial

Introduction to GACRC Teaching Cluster PHYS8602

Working on the NewRiver Cluster

Introduction to HPC Using zcluster at GACRC

Introduction to High-Performance Computing (HPC)

Our Workshop Environment

Intel MPI Cluster Edition on Graham A First Look! Doug Roberts

Introduction to Discovery.

Intel Manycore Testing Lab (MTL) - Linux Getting Started Guide

Our Workshop Environment

Choosing Resources Wisely. What is Research Computing?

SCALABLE HYBRID PROTOTYPE

High Performance Computing Cluster Basic course

Our Workshop Environment

Submitting batch jobs

OpenMP Exercises. These exercises will introduce you to using OpenMP for parallel programming. There are four exercises:

XSEDE New User Tutorial

Our Workshop Environment

Crash Course in High Performance Computing

Introduction to GALILEO

Introduction to GALILEO

How to run a job on a Cluster?

Batch Systems & Parallel Application Launchers Running your jobs on an HPC machine

Migrating from Zcluster to Sapelo

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing

The cluster system. Introduction 22th February Jan Saalbach Scientific Computing Group

XSEDE New User Tutorial

OBTAINING AN ACCOUNT:

CSC BioWeek 2016: Using Taito cluster for high throughput data analysis

Introduction to High Performance Computing at Case Western Reserve University. KSL Data Center

Practical Introduction to Message-Passing Interface (MPI)

UoW HPC Quick Start. Information Technology Services University of Wollongong. ( Last updated on October 10, 2011)

Introduction to Discovery.

Introduction to UBELIX

STARTING THE DDT DEBUGGER ON MIO, AUN, & MC2. (Mouse over to the left to see thumbnails of all of the slides)

P a g e 1. HPC Example for C with OpenMPI

Introduction to HPC Resources and Linux

Introduction to HPC Using zcluster at GACRC On-Class GENE 4220

RHRK-Seminar. High Performance Computing with the Cluster Elwetritsch - II. Course instructor : Dr. Josef Schüle, RHRK

Introduction to the PDC environment. PDC Center for High Performance Computing KTH, Sweden

New User Seminar: Part 2 (best practices)

Compilation and Parallel Start

High Performance Computing Cluster Advanced course

Programming Environment

CSC BioWeek 2018: Using Taito cluster for high throughput data analysis

TITANI CLUSTER USER MANUAL V.1.3

Compiling applications for the Cray XC

INTRODUCTION TO THE CLUSTER

Beginner's Guide for UK IBM systems

CSC Supercomputing Environment

Guillimin HPC Users Meeting July 14, 2016

Submitting batch jobs Slurm on ecgate Solutions to the practicals

Transcription:

Using Compute Canada Masao Fujinaga Information Services and Technology University of Alberta

Introduction to cedar batch system jobs are queued priority depends on allocation and past usage

Cedar Nodes "Base" compute nodes: 576 nodes 128 GB of memory, 32 cores "Large" compute nodes: 128 nodes 256 GB of memory, 32 cores "Bigmem500" 24 nodes 0.5 TB (512 GB) of memory, 32 cores "Bigmem1500" nodes 24 nodes 1.5 TB of memory, 32 cores "GPU base" nodes: 114 nodes 128 GB of memory, 24 cores, 4 GPUs "GPU large" nodes. 32 nodes 256 GB of memory, 24 cores, 4 GPUs "Bigmem3000" nodes 4 nodes 3 TB of memory, 32 cores

Software for connecting to Compute Canada Windows MobaXterm http://mobaxterm.mobatek.net Mac XQuartz http://xquartz.macosforge.org/landing/

Install MobaXterm (Windows)

Start MobaXterm

Start local terminal

Connect with MobaXterm

Transfer files with MobaXterm

Connect with Terminal (mac or linux)

Transfer files mac/linux scp myfile mycomputecanadaid@cedar.westgrid.ca: FileZilla windows MobaXterm WinSCP

Connect to cedar Logging in ssh -Y mycomputecanadaid@cedar.computecanada.ca Interactive use compiling testing, debugging use less than 1 hour and 2 cores longer interactive use: salloc

Untar workshop files home directory cp ~masao/workshopsept2017.tar.gz. tar -xf workshopsept2017.tar.gz cd workshopsept2017 current directory

Setting up the environment Modules module avail Organized according to compiler, MPI, cuda.

Finding modules Module spider module spider blast+ You will need to load all module(s) on any one of the lines below before the "blast+/2.6.0" module is available to load. nixpkgs/16.09 gcc/5.4.0 module load gcc blast+

Software Compute Canada website https://docs.computecanada.ca/wiki/available_software

Other software Install own https://docs.computecanada.ca/wiki/installing_software_in_your_home_directory https://docs.computecanada.ca/wiki/python python27-scipy-stack/2017a, python35-scipy-stack/2017a python27-mpi4py/2.0.0, python35-mpi4py/2.0.0 https://docs.computecanada.ca/wiki/r r-bundle-bioconductor/3.4 (gcc) https://docs.computecanada.ca/wiki/perl Request installation support@computecanada.ca

Compiling compile icc -o hello hello.c ifort -o hello hello.f90 run./hello

Man pages man icc space bar to get a page /mysearch to search for mysearch q to quit man -k mytopic get pages related to mytopic

Editing graphical editors nedit, gedit non-graphical editors vi, emacs editing through Winscp

Batch job #!/bin/bash #SBATCH --reservation=masao_10 #SBATCH --account=def-myaccount #SBATCH --ntasks=1 #SBATCH --mem-per-cpu=2000mb #SBATCH --time=00:05:00 #SBATCH --mail-type=all #SBATCH --mail-user=myemail@emailaddress./hello

Job submission sbatch sbatch job-cedar.slm Submitted batch job 1136843 squeue squeue -u masao JOBID USER ACCOUNT 1136843 masao def-masao_cp NAME job-cedar.slm ST START_TIME PD N/A TIME_LEFT NODES CPUS GRES MIN_MEM NODELIST (REASON) 5:00 1 1 (null) 2000M (Priority) NAME job-cedar.slm ST START_TIME R 2017-09-19T08:31 TIME_LEFT NODES CPUS GRES MIN_MEM NODELIST (REASON) 4:55 1 1 (null) 2000M cdr310 (None) running JOBID 1136843 USER ACCOUNT masao def-masao_cp

Job output slurm-<jobnumber>.out Common errors out of walltime out of memory batch script in DOS format No end-of-line character at the end

pi program compile icc -o pi pi.c interactive run./pi input from terminal

pi job batch run create pi.in./pi < pi.in > pi.out no input file./pi << EOF-pi > pi.out 100 1000 0 EOF-pi

MPI compile mpicc -o hello-mpi hello-mpi.c mpif90 -o hello-mpi hello-mpi.f90 Run ulimit -l 2097152 mpiexec -np 2./hello-mpi

Interactive batch job Submitting an interactive batch job salloc --time=6:0:0 --ntasks=2 --account=def-myaccount -reservation=masao_10 no batch script may have to wait a long time depending on how busy the system is

Benchmarking Timing reproducible, without other users time command walltime

Parallel benchmark Time the real time it takes to run the program using various number of processors For any parallel program, it is important to know the parallel efficiency of the program

benchmark exercise compile sample1.f capital oh zero mpif90 -o sample1 sample1.f90 -fixed -O0 submit script1.sbatch. This will run the program five times, changing -n to 1, 2, 4, 8, and 12 look at the real time ( grep real out.time ) plot using www.wolframalpha.com Replace t1,t2,t4,t8,t16 with your values plot time plot{{1,t1},{2,t2},{4,t4},{8,t8},{16,t16}} plot speedup plot{{1,1},{2,t1/t2},{4,t1/t4},{8,t1/t8},{16,t1/t16}}

benchmark results

Optimizing compiler optimization default: -O2 higher optimizations: -O3, -axcore-avx2 Exercise icc -o sample2 sample2.c -O1 time./sample2 time runs of sample2 after having compiled with -O0 -O1 -O2 -O3 -O3 -axcore-avx2 -O3 -axcore-avx2 -fp-model precise Note the checksum

Math Library Intel Math Kernel Library (MKL) http://software.intel.com/en-us/mkl_11.1_ref -mkl=sequential

Filesystems home directory Project Default project Projects Links to other project directories scratch Use for running jobs quotas diskusage_report

Moving data in Compute Canada Globus https://www.computecanada.ca/research-portal/national-services/globus-portal/ web-based fault tolerant can transfer to your desktop

Upcoming ARC Training Sessions October 3 10:00 AM 11:00 AM PDT October 5 11:00 AM 1:00 PM PDT October 24-26 1:00-3:00 PM MDT (each day) November 1 11:00 AM 1:00 PM PDT November 21 11:00 AM 1:00 PM PST Data Visualization on Compute Canada's Supercomputers Intro to Parallel Programming for Shared Memory Machines Scheduling & Job Management: How to Get the Most from a Cluster Introduction to Classical Molecular Dynamics Simulations Exploring Containerization with Singularity

Getting Help https://docs.computecanada.ca use the search support@computecanada.ca specify user ID which system job ID error messages script location