Introduction to Advanced Research Computing (ARC)

Size: px
Start display at page:

Download "Introduction to Advanced Research Computing (ARC)"

Transcription

1 Introduction to Advanced Research Computing (ARC) September 29, 2016 By: Pier-Luc St-Onge 1

2 Financial Partners 2

3 Setup for the workshop 1. Get a user ID and password paper (provided in class): ##: ********** 2. Access to local computer (replace ## and with appropriate values, is provided in class): a. User name: csuser## (ex.: csuser99) b. (ex.: sec@[s99) 3. Slides: 3

4 Outline Introduction to ARC Laptop vs HPC Cluster Moving Data With Globus Software Development Types of Parallelism Scientific Software and Modules Best Practices The Scheduler and Job Submission Compute Canada and Calcul Québec 4

5 Introduction to ARC 5

6 Introduction This workshop is built from the HPCS 2015 Introduction to ARC workshop: HPCS website: Etherpad with extra information: Presentation: Instructors and participants Name and research domain 6

7 Other research domains Genomics Molecular Dynamics (MD) Finite Element Analysis (FEA) Computational Fluid Dynamics (CFD) Astronomy and Astrophysics Geospatial Data Analysis 3D renderings, Image Analysis Digital Humanities (DH) Artificial Intelligence 7

8 Advanced Research Computing High Performance Computing (HPC) Parallel code Running code on accelerators Profiling code Management of massive data Complete pipelines for jobs submission Sharing data Web portals for special projects Cloud computing / storage 8

9 ARC is also about having enough Resources Problem: we want to compute a dense matrix multiplication C = A * B Each matrix is made of * floating point numbers in double precision How much memory do you need for this job? a) 3.2 GB b) 8.6 GB c) 12.9 GB d) 25.8 GB 9

10 Different Tools for Different Needs Portable computer HPC Compute node 2-4 cores 4-16 GB Local storage, USB 128GB to 2TB Wi-Fi and Ethernet General usage Cloud storage Gaming GPU 32 bits Direct access 8-32 cores GB Network storage 150TB to 5PB Ethernet and fast network (Infiniband) Multi-node tasks Accelerators 64 bits, ECC mem. Access by queued jobs 10

11 Zoom on a HPC Compute Node RAM Proc. Proc. RAM Infiniband RAM Local hard-drive (scratch) RAM Proc. Proc. Ethernet RAM Proc. Proc. RAM Accelerator RAM RAM Proc. Proc. 11

12 Why Infiniband? Some tasks need more than one compute node Split workload, memory usage and storage access High bandwidth Ethernet: 1Gbps (~100MB/s) Infiniband: Gbps (2-10GB/s) Best for I/O on large files Reduce latency Ethernet: ~30µs (33k iops = I/O operations per sec.) Infiniband: 2µs (500k iops) Best for several small communications 12

13 Storage Available on Compute Nodes Bandwidth RAMDISK Local scratch Project space Scratch space Capacity (Not to scale) Latency 13

14 Typical Cluster Job Login node Scheduler Job Nodes Ethernet & Infiniband Internet Storage home projet scratch ssh 14

15 Typical Cluster Usage 1. Connect to a login node a. Import data b. Compile code or install applications 2. Submit jobs to the scheduler a. Define needed resources: processors, accelerators, memory, walltime limit b. Wait in queue - monitor your position c. Automatic execution of your jobs on compute nodes 3. Get your results 15

16 Typical Environment on Cluster The operating system is usually Linux (CentOS, Scientific Linux, ) To connect to a cluster, we usually use an encrypted session through SSH (Secure SHell) To transfer data safely: SCP (Secure CoPy) and others tools Some web portals may also provide an access Development environment Multiple text editors Scientific software Version control software Tools for transferring large amount of data 10Gbps, 40Gbps or 100Gbps between sites 16

17 Moving Data With Globus 17

18 Moving Data Over SSH Secure CoPy with SCP or WinSCP: Most clusters only allow SSH connections SSH: Secure SHell (username and password) SCP uses the SSH protocol to copy files Simple tool for small amount of data (<1GB) On Windows: use WinSCP or FileZilla Two panels, where one side shows your files on the cluster, then drag-and-drop files or folders On Linux or Mac OS X: use the scp command: scp file.txt scp -pr 18

19 Globus - Why? Secure CoPy with SCP or WinSCP: One process and one thread Fails on dropped connections BBCP - parallel transfers Need a calibration for the round-trip time Command line with multiple arguments Globus - Web interface and CLI No need for calibration Convenient Web interface Confirmation after the transfer Possible to share data with other users 19

20 Go to Compute Canada Research Portal CC > Research Portal > National Services > Data Movement (Globus) Create an account if you do not have Or Log in to your account if you already have an account 20

21 Account Creation and Login Go to: Select Compute Canada if you have a CCDB account, or Globus ID if not. To create a Globus ID account: 21

22 Compute Canada Login and Globus ID Account Creation 22

23 Exercise: Sign-up! Globus ID credentials are stored on Globus: Can be a different username than the one on CCDB, or it can be the same. But both accounts (on CCDB and Globus ID) are different! For sure, use a different password! Put the green Post-it on the computer screen when done 23

24 Globus Home Interface To Logout Endpoint Viewports Transfer options Globus Endpoint on your computer 24

25 Choosing an Endpoint at Compute Canada 25

26 Authentication to an Endpoint The authentication is done on a server corresponding to the endpoint Globus will never know your credentials for a given cluster/endpoint The authentication server only grants access to Globus. 26

27 Files and Directories 27

28 Exercise: Import Pictures Connect to computecanada#guillimin Use the class account credentials (on your paper) class## ********** Connect to intro-arc_guillimin_plstonge Select both exercises and pictures folders Click on the blue button to initiate the transfer to your (class##) home directory Wait for the report When the transfer is completed, put the green Post-it on top of the computer screen 28

29 Confirmation TASK DETAILS Task ID: 5ef1f2c6-81f5-11e6-afd b92c261 Task Type: TRANSFER Status: SUCCEEDED Source: intro-arc_guillimin_plstonge (...) Destination: computecanada#guillimin (...) Label: n/a 29

30 Transfer Report 30

31 Managing Endpoints 31

32 Create a Local Endpoint 32

33 Create a Local Endpoint 1. Choose a unique Endpoint name a. Generate the Setup Key b. Copy the Setup Key in a text editor 2. Download and Install Globus Connect Personal a. Start it and paste the Setup Key 3. A Globus icon ( g) appears in the taskbar a. [Right-]click -> Preferences -> Access b. Select accessible directories you want to export 4. Use your Endpoint: user#uniquename 33

34 Sharing Option 34

35 Software Development Types of Parallelism 35

36 Vocabulary Serial tasks Any task that cannot be split in two simultaneous sequences of actions Examples: starting a process, reading a file, any communication between two processes Parallel tasks Data parallelism: same action applied on different data. Could be serial tasks done in parallel. Process parallelism: one action on one set of data. Action split in multiple processes or threads. Data partitioning: rectangles or blocks 36

37 Parallel tasks Shared memory Multiple threads share the same memory space in a single process: full read and write access. Distributed memory Each process has its own memory space Information is sent and received by messages Parallel efficiency (scaling) Amdahl s law: how long does it take to compute a task with an infinite number of processors? Gustafson's law: what size of problem can we solve in a given time with N processors? 37

38 Software Development Scientific Software and Modules 38

39 Connecting to Guillimin - SSH Client From Linux, open a Terminal window From Mac OS X, open a Terminal window From Windows, install PuTTY, MobaXterm or any other SSH (2.0 compliant) client /download.html putty.exe is a standalone executable Simply run putty.exe 39

40 Connecting to Guillimin Terminal Server name: guillimin.calculquebec.ca Username: class## ## = Two digits on the first line of the small piece of paper Password: ********** Sequence of 10 characters on small piece of paper Linux/Mac OS X example: ssh class99@guillimin.calculquebec.ca Accept the server s key fingerprint 40

41 Connecting to Guillimin PuTTY 41

42 Scientific Software You will use or create scientific software Applications Libraries Make sure you are using a code language widely used in your research domain Make sure your code is well documented and easy to maintain Keep in mind that future researchers are going to use your software Do not reinvent the wheel: use modules! 42

43 Modules Exercise: module avail Compilers: GNU, Intel, LLVM, PGI, Java Language interpreters: Python, R Libraries and other applications: OpenMPI, MVAPICH2, Boost MKL, FFTW GROMACS, NAMD, Quantum-Espresso, NetCDF Other tools: Paraview, GNU-Parallel, GDB, Valgrind 43

44 Exercises with Modules which gcc module add gcc/4.8.2 module list which gcc module rm gcc/4.8.2 module add ifort_icc openmpi echo $LD_LIBRARY_PATH echo $CC module purge 44

45 New Lmod/EasyBuild-based module system Default set of modules on Colosse Both legacy and Lmod modules are available on Guillimin Exercise on Guillimin: 1. Load the main toolchain Intel + OpenMPI + MKL: module load iomkl/2015b 2. List loaded modules: module list 3. You now have access to much more modules: module avail 45

46 Introduction to GNU-Parallel Automatic online process scheduler for serial tasks User provides a command line template and one or multiple sequences of arguments For N cores on a compute node, GNU-Parallel will start N processes As soon as one process is complete, it will start the next process according to values in sequences Example: module load parallel/ parallel echo {1} ::: $(seq 1 960) less 46

47 Introduction to Threads Threads are used for parallel tasks on a single node One process, one shared memory space, but multiple threads Synchronization issues, like race conditions Two threads reading or writing to the same variable Instead of POSIX threads: OpenMP Threading Building Blocks 47

48 Introduction to MPI Message Passing Interface (MPI) is a standard for passing messages between multiple processes on multiple nodes Each process can know its rank number Same set of functions, but different implementations MPICH2, MVAPICH2, OpenMPI, Intel MPI,... Example: module load iomkl/2015b mpiexec -n 4 hostname 48

49 Software Development Best Practices 49

50 Version Control & Revision Control Bad practice: Keeping multiple versions in sub-directories Sending a specific version by Trying to merge two versions manually Good practice: using a version control system Choose one that fits your project: Git (git) Mercurial (hg) Subversion (svn) Repository: shared on a remote server 50

51 Exercises with git mkdir -p ~/projecta; cd ~/projecta; module add git git init # Local repository touch script.sh; git status git add script.sh # Ready to commit git commit -m "intro-arc" # Do commit Add line env > env.log in script.sh git git git git diff # See differences add script.sh # Ready to commit commit -m "Cleaner job" # Do commit log # Show past commits 51

52 One Last Good Practice Using SSH Keys Having access with a local passphrase: ssh-keygen Generating public/private rsa key pair. Enter file in which to save the key (~/.ssh/id_rsa): Enter passphrase (empty for no passphrase): Enter same passphrase again: Copy the public key to a remote site (id_rsa.pub): scp.ssh/id_rsa.pub <user>@<server>: ssh <user>@<server> \ "cat id_rsa.pub >>.ssh/authorized_keys" 52

53 Exercise - Code Compilation Go to the exercises folder cd ~/exercises Load the following modules: iomkl/2015b Boost/ Python ImageMagick/ Build the application: make 53

54 The Scheduler and Job Submission 54

55 The Scheduler Jobs are not running on login nodes You need to submit your jobs to the scheduler: Resources you need: n processors, memory, accelerators Job duration Account name The script to run on worker nodes Where you want your results 55

56 Schedulers in Calcul Québec Job component Torque + Maui/Moab Submission command Torque: qsub Maui/Moab: msub Account -A xyz-123-aa Processors (small L) Some clusters force full nodes -l nodes=n:ppn=p -l procs=pp Memory (small L) -l pmem=1700m -l mem=20g Job duration (small L) -l walltime=d:hh:mm:ss Job name -N name Where you want your results -o output_path -e error_path your_app > out.log 56

57 Environment Variables During Job Variable $PBS_JOBID $PBS_JOBNAME $PBS_NODEFILE Description Job ID Ex.: server.name Job name set by -N Ex.: introarc The name of the file that contains a list of assigned nodes for the job $PBS_O_WORKDIR The current working directory ($PWD) when the job has been submitted. $PBS_ARRAYID One of the values specified by the job array: -t 1-3,6,

58 Exercise - Manage your Jobs Torque - jobs and job array details qstat -u $USER qstat -t -u $USER Maui / Moab - jobs showq -u $USER (Optional) Torque - cancel job qdel jobid (Optional) Maui / Moab - cancel job mjobctl -c jobid canceljob jobid 58

59 Exercises: Edit and Submit a Job Scripts Go to the exercises folder Check the README.en file for general instructions Start with the exercise 1-base. All the instructions are in the README.en file The solution is in solution.sh Then try exercises 2, 3 and/or 6 For exercise 2, a description of & and wait is in the README.en file of the exercise You need to understand the solution of the exercise 3 to do exercise 4 59

60 Compute Canada and Calcul Québec 60

61 Compute Canada Four consortia (from West to East): WestGrid British Columbia, Alberta, Saskatchewan, Manitoba Compute Ontario (SciNet + SHARCNET + HPCVL) Ontario Calcul Québec Québec ACENET New Brunswick, Newfoundland, Nova Scotia, Prince Edward Island 61

62 Example - Calcul Québec Consortium Compute Canada Calcul Québec - Compute Resources Briarée & Hadès (U. Montréal) Colosse & Helios (U. Laval) Guillimin (McGill) MS2 & MP2 (U. Sherbrooke) Calcul Québec - Human Resources About 40 employees ARC analysts System administrators Scientific and operational directors 62

63 Biggest Clusters - GP1, GP2 and GP3 are coming soon... Name / Institution Cores Summary MP2 / U. Sherbrooke cores/node, 32+ GB/node, QDR IB GPC / U. Toronto cores/node, 16+ GB/node, D/QDR IB Guillimin / McGill U ,16 cores/node, 24+ GB/node, QDR IB Orcinus / U. British Columbia ,12 cores/node, 16+ GB/node, D/QDR IB Orca / SHARCNET ,24 cores/node, 32 GB/node, QDR IB Colosse / U. Laval cores/node, 24+ GB/node, QDR IB Briarée / U. Montréal cores/node, 24+ GB/node, QDR IB Parallel / U. Calgary cores/node, 24 GB/node, QDR IB Example of job requirements: 8 or 16 cores/node, 30 GB/node, QDR IB 63

64 Roles and Allocations CCRI : abc ABCD University Alloc.: abc-123-aa (default alloc.) CCRI : def HIJK University Alloc.: def-456-ab 50 core*years xyz Prof Prof CCRI : abc DEFG University Alloc.: abc-123-ac 200 core*years xyz Student 64

65 About Allocations Amount of compute cycles (only a target): 1 core provides 2 to 3 GHz (cycles per second) Core*seconds, core*hours or core*years (cy) Each site has a default allocation (up to 80 cy) Amount of storage space (hard limit): Private scratch space (default: up to 1TB) Group shared project space (default: up to 1TB) Special allocations Resource Allocation Competition (RAC 2017) ource-allocation-competitions/ 65

66 How to Access Resources The Complete Procedure CCDB: A professor creates an account in CCDB Request role of Principal Investigator (PI) - CCRI (Optional) Request a special allocation (RAC) Special allocation granted on a specific cluster A student creates an account in CCDB Request sponsored role - need to provide the PI s CCRI Apply for a consortium account (ex.: Calcul Québec) In the Calcul Québec portal, activate your access to any cluster Choose clusters based on special allocations 66

67 Usage Policy Each user has his own account; never share your private account! No heavy process (in time and space) on login nodes Resources must be used in an optimal way Large parallel jobs must use tools that scale well Cold data should not stay on disk Maximize the amount of research done in Canada Must acknowledge Compute Canada and any regional partner organization 67

68 Service and Support Training workshops: Future: Past: Documentation: Documentation Support: 68

69 Conclusion 69

70 Conclusion Need help? WestGrid: SciNet: SHARCNET: Calcul Québec: ACENET: Useful information to send us: User name Job ID (if applicable) Path to {job script, output/error files, code} or as attached files 70

Introduction to Advanced Research Computing

Introduction to Advanced Research Computing Introduction to Advanced Research Computing and Best Practices January 28, 2016 By: Pier-Luc St-Onge 1 Partners and sponsors 2 Setup for the workshop 1. Get a user ID and password paper (provided in class):

More information

Guillimin HPC Users Meeting March 17, 2016

Guillimin HPC Users Meeting March 17, 2016 Guillimin HPC Users Meeting March 17, 2016 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Outline Compute Canada News System Status Software Updates Training

More information

Introduction to High Performance Computing

Introduction to High Performance Computing Introduction to High Performance Computing By Pier-Luc St-Onge pier-luc.st-onge@mcgill.ca September 11, 2014 Objectives Familiarize new users with the concepts of High Performance Computing (HPC). Outline

More information

Guillimin HPC Users Meeting October 20, 2016

Guillimin HPC Users Meeting October 20, 2016 Guillimin HPC Users Meeting October 20, 2016 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Please be kind to your fellow user meeting attendees Limit

More information

Guillimin HPC Users Meeting February 11, McGill University / Calcul Québec / Compute Canada Montréal, QC Canada

Guillimin HPC Users Meeting February 11, McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Guillimin HPC Users Meeting February 11, 2016 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Compute Canada News Scheduler Updates Software Updates Training

More information

Outline. March 5, 2012 CIRMMT - McGill University 2

Outline. March 5, 2012 CIRMMT - McGill University 2 Outline CLUMEQ, Calcul Quebec and Compute Canada Research Support Objectives and Focal Points CLUMEQ Site at McGill ETS Key Specifications and Status CLUMEQ HPC Support Staff at McGill Getting Started

More information

Guillimin HPC Users Meeting March 16, 2017

Guillimin HPC Users Meeting March 16, 2017 Guillimin HPC Users Meeting March 16, 2017 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Please be kind to your fellow user meeting attendees Limit to

More information

Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop

Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop Before We Start Sign in hpcxx account slips Windows Users: Download PuTTY Google PuTTY First result Save putty.exe to Desktop Research Computing at Virginia Tech Advanced Research Computing Compute Resources

More information

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing Quick Start Guide by Burak Himmetoglu Supercomputing Consultant Enterprise Technology Services & Center for Scientific Computing E-mail: bhimmetoglu@ucsb.edu Contents User access, logging in Linux/Unix

More information

Supercomputing environment TMA4280 Introduction to Supercomputing

Supercomputing environment TMA4280 Introduction to Supercomputing Supercomputing environment TMA4280 Introduction to Supercomputing NTNU, IMF February 21. 2018 1 Supercomputing environment Supercomputers use UNIX-type operating systems. Predominantly Linux. Using a shell

More information

PARALLEL COMPUTING IN R USING WESTGRID CLUSTERS STATGEN GROUP MEETING 10/30/2017

PARALLEL COMPUTING IN R USING WESTGRID CLUSTERS STATGEN GROUP MEETING 10/30/2017 PARALLEL COMPUTING IN R USING WESTGRID CLUSTERS STATGEN GROUP MEETING 10/30/2017 PARALLEL COMPUTING Dataset 1 Processor Dataset 2 Dataset 3 Dataset 4 R script Processor Processor Processor WHAT IS ADVANCED

More information

Computing with the Moore Cluster

Computing with the Moore Cluster Computing with the Moore Cluster Edward Walter An overview of data management and job processing in the Moore compute cluster. Overview Getting access to the cluster Data management Submitting jobs (MPI

More information

Graham vs legacy systems

Graham vs legacy systems New User Seminar Graham vs legacy systems This webinar only covers topics pertaining to graham. For the introduction to our legacy systems (Orca etc.), please check the following recorded webinar: SHARCNet

More information

Introduction to HPC Using the New Cluster at GACRC

Introduction to HPC Using the New Cluster at GACRC Introduction to HPC Using the New Cluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What is the new cluster

More information

Guillimin HPC Users Meeting June 16, 2016

Guillimin HPC Users Meeting June 16, 2016 Guillimin HPC Users Meeting June 16, 2016 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Compute Canada News System Status Software Updates Training News

More information

Introduction to PICO Parallel & Production Enviroment

Introduction to PICO Parallel & Production Enviroment Introduction to PICO Parallel & Production Enviroment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Domenico Guida d.guida@cineca.it Nicola Spallanzani n.spallanzani@cineca.it

More information

Allinea DDT Debugger. Dan Mazur, McGill HPC March 5,

Allinea DDT Debugger. Dan Mazur, McGill HPC  March 5, Allinea DDT Debugger Dan Mazur, McGill HPC daniel.mazur@mcgill.ca guillimin@calculquebec.ca March 5, 2015 1 Outline Introduction and motivation Guillimin login and DDT configuration Compiling for a debugger

More information

Introduction to the SHARCNET Environment May-25 Pre-(summer)school webinar Speaker: Alex Razoumov University of Ontario Institute of Technology

Introduction to the SHARCNET Environment May-25 Pre-(summer)school webinar Speaker: Alex Razoumov University of Ontario Institute of Technology Introduction to the SHARCNET Environment 2010-May-25 Pre-(summer)school webinar Speaker: Alex Razoumov University of Ontario Institute of Technology available hardware and software resources our web portal

More information

Introduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende

Introduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende Introduction to the NCAR HPC Systems 25 May 2018 Consulting Services Group Brian Vanderwende Topics to cover Overview of the NCAR cluster resources Basic tasks in the HPC environment Accessing pre-built

More information

Introduction to HPC Resources and Linux

Introduction to HPC Resources and Linux Introduction to HPC Resources and Linux Burak Himmetoglu Enterprise Technology Services & Center for Scientific Computing e-mail: bhimmetoglu@ucsb.edu Paul Weakliem California Nanosystems Institute & Center

More information

High Performance Computing (HPC) Club Training Session. Xinsheng (Shawn) Qin

High Performance Computing (HPC) Club Training Session. Xinsheng (Shawn) Qin High Performance Computing (HPC) Club Training Session Xinsheng (Shawn) Qin Outline HPC Club The Hyak Supercomputer Logging in to Hyak Basic Linux Commands Transferring Files Between Your PC and Hyak Submitting

More information

Guillimin HPC Users Meeting. Bryan Caron

Guillimin HPC Users Meeting. Bryan Caron July 17, 2014 Bryan Caron bryan.caron@mcgill.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Outline Compute Canada News Upcoming Maintenance Downtime in August Storage System

More information

The cluster system. Introduction 22th February Jan Saalbach Scientific Computing Group

The cluster system. Introduction 22th February Jan Saalbach Scientific Computing Group The cluster system Introduction 22th February 2018 Jan Saalbach Scientific Computing Group cluster-help@luis.uni-hannover.de Contents 1 General information about the compute cluster 2 Available computing

More information

Our new HPC-Cluster An overview

Our new HPC-Cluster An overview Our new HPC-Cluster An overview Christian Hagen Universität Regensburg Regensburg, 15.05.2009 Outline 1 Layout 2 Hardware 3 Software 4 Getting an account 5 Compiling 6 Queueing system 7 Parallelization

More information

High Performance Computing (HPC) Using zcluster at GACRC

High Performance Computing (HPC) Using zcluster at GACRC High Performance Computing (HPC) Using zcluster at GACRC On-class STAT8060 Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC?

More information

Using a Linux System 6

Using a Linux System 6 Canaan User Guide Connecting to the Cluster 1 SSH (Secure Shell) 1 Starting an ssh session from a Mac or Linux system 1 Starting an ssh session from a Windows PC 1 Once you're connected... 1 Ending an

More information

The JANUS Computing Environment

The JANUS Computing Environment Research Computing UNIVERSITY OF COLORADO The JANUS Computing Environment Monte Lunacek monte.lunacek@colorado.edu rc-help@colorado.edu What is JANUS? November, 2011 1,368 Compute nodes 16,416 processors

More information

New User Tutorial. OSU High Performance Computing Center

New User Tutorial. OSU High Performance Computing Center New User Tutorial OSU High Performance Computing Center TABLE OF CONTENTS Logging In... 3-5 Windows... 3-4 Linux... 4 Mac... 4-5 Changing Password... 5 Using Linux Commands... 6 File Systems... 7 File

More information

UoW HPC Quick Start. Information Technology Services University of Wollongong. ( Last updated on October 10, 2011)

UoW HPC Quick Start. Information Technology Services University of Wollongong. ( Last updated on October 10, 2011) UoW HPC Quick Start Information Technology Services University of Wollongong ( Last updated on October 10, 2011) 1 Contents 1 Logging into the HPC Cluster 3 1.1 From within the UoW campus.......................

More information

Guillimin HPC Users Meeting. Bart Oldeman

Guillimin HPC Users Meeting. Bart Oldeman June 19, 2014 Bart Oldeman bart.oldeman@mcgill.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Outline Compute Canada News Upcoming Maintenance Downtime in August Storage System

More information

PACE. Instructional Cluster Environment (ICE) Orientation. Research Scientist, PACE

PACE. Instructional Cluster Environment (ICE) Orientation. Research Scientist, PACE PACE Instructional Cluster Environment (ICE) Orientation Mehmet (Memo) Belgin, PhD Research Scientist, PACE www.pace.gatech.edu What is PACE A Partnership for an Advanced Computing Environment Provides

More information

A Brief Introduction to The Center for Advanced Computing

A Brief Introduction to The Center for Advanced Computing A Brief Introduction to The Center for Advanced Computing May 1, 2006 Hardware 324 Opteron nodes, over 700 cores 105 Athlon nodes, 210 cores 64 Apple nodes, 128 cores Gigabit networking, Myrinet networking,

More information

Introduction to HPC Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC Introduction to HPC Using zcluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What is HPC Concept? What is

More information

Migrating from Zcluster to Sapelo

Migrating from Zcluster to Sapelo GACRC User Quick Guide: Migrating from Zcluster to Sapelo The GACRC Staff Version 1.0 8/4/17 1 Discussion Points I. Request Sapelo User Account II. III. IV. Systems Transfer Files Configure Software Environment

More information

Introduction to HPC Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC Introduction to HPC Using zcluster at GACRC On-class PBIO/BINF8350 Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What

More information

Introduction to GALILEO

Introduction to GALILEO November 27, 2016 Introduction to GALILEO Parallel & production environment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it SuperComputing Applications and Innovation Department

More information

Introduction to GALILEO

Introduction to GALILEO Introduction to GALILEO Parallel & production environment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Domenico Guida d.guida@cineca.it Maurizio Cremonesi m.cremonesi@cineca.it

More information

PACE. Instructional Cluster Environment (ICE) Orientation. Mehmet (Memo) Belgin, PhD Research Scientist, PACE

PACE. Instructional Cluster Environment (ICE) Orientation. Mehmet (Memo) Belgin, PhD  Research Scientist, PACE PACE Instructional Cluster Environment (ICE) Orientation Mehmet (Memo) Belgin, PhD www.pace.gatech.edu Research Scientist, PACE What is PACE A Partnership for an Advanced Computing Environment Provides

More information

Introduction to Discovery.

Introduction to Discovery. Introduction to Discovery http://discovery.dartmouth.edu The Discovery Cluster 2 Agenda What is a cluster and why use it Overview of computer hardware in cluster Help Available to Discovery Users Logging

More information

Using Sapelo2 Cluster at the GACRC

Using Sapelo2 Cluster at the GACRC Using Sapelo2 Cluster at the GACRC New User Training Workshop Georgia Advanced Computing Resource Center (GACRC) EITS/University of Georgia Zhuofei Hou zhuofei@uga.edu 1 Outline GACRC Sapelo2 Cluster Diagram

More information

Introduction to HPC Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC Introduction to HPC Using zcluster at GACRC On-class STAT8330 Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu Slides courtesy: Zhoufei Hou 1 Outline What

More information

Practical Introduction to Message-Passing Interface (MPI)

Practical Introduction to Message-Passing Interface (MPI) 1 Practical Introduction to Message-Passing Interface (MPI) October 1st, 2015 By: Pier-Luc St-Onge Partners and Sponsors 2 Setup for the workshop 1. Get a user ID and password paper (provided in class):

More information

New User Seminar: Part 2 (best practices)

New User Seminar: Part 2 (best practices) New User Seminar: Part 2 (best practices) General Interest Seminar January 2015 Hugh Merz merz@sharcnet.ca Session Outline Submitting Jobs Minimizing queue waits Investigating jobs Checkpointing Efficiency

More information

Guillimin HPC Users Meeting July 14, 2016

Guillimin HPC Users Meeting July 14, 2016 Guillimin HPC Users Meeting July 14, 2016 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Outline Compute Canada News System Status Software Updates Training

More information

XSEDE New User Tutorial

XSEDE New User Tutorial April 2, 2014 XSEDE New User Tutorial Jay Alameda National Center for Supercomputing Applications XSEDE Training Survey Make sure you sign the sign in sheet! At the end of the module, I will ask you to

More information

Guillimin HPC Users Meeting April 13, 2017

Guillimin HPC Users Meeting April 13, 2017 Guillimin HPC Users Meeting April 13, 2017 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Please be kind to your fellow user meeting attendees Limit to

More information

ICS-ACI System Basics

ICS-ACI System Basics ICS-ACI System Basics Adam W. Lavely, Ph.D. Fall 2017 Slides available: goo.gl/ss9itf awl5173 ICS@PSU 1 Contents 1 Overview 2 HPC Overview 3 Getting Started on ACI 4 Moving On awl5173 ICS@PSU 2 Contents

More information

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved. Minnesota Supercomputing Institute Introduction to Job Submission and Scheduling Andrew Gustafson Interacting with MSI Systems Connecting to MSI SSH is the most reliable connection method Linux and Mac

More information

Introduc)on to Hyades

Introduc)on to Hyades Introduc)on to Hyades Shawfeng Dong Department of Astronomy & Astrophysics, UCSSC Hyades 1 Hardware Architecture 2 Accessing Hyades 3 Compu)ng Environment 4 Compiling Codes 5 Running Jobs 6 Visualiza)on

More information

ACEnet for CS6702 Ross Dickson, Computational Research Consultant 29 Sep 2009

ACEnet for CS6702 Ross Dickson, Computational Research Consultant 29 Sep 2009 ACEnet for CS6702 Ross Dickson, Computational Research Consultant 29 Sep 2009 What is ACEnet? Shared resource......for research computing... physics, chemistry, oceanography, biology, math, engineering,

More information

Working on the NewRiver Cluster

Working on the NewRiver Cluster Working on the NewRiver Cluster CMDA3634: Computer Science Foundations for Computational Modeling and Data Analytics 22 February 2018 NewRiver is a computing cluster provided by Virginia Tech s Advanced

More information

OBTAINING AN ACCOUNT:

OBTAINING AN ACCOUNT: HPC Usage Policies The IIA High Performance Computing (HPC) System is managed by the Computer Management Committee. The User Policies here were developed by the Committee. The user policies below aim to

More information

Introduction to High Performance Computing at UEA. Chris Collins Head of Research and Specialist Computing ITCS

Introduction to High Performance Computing at UEA. Chris Collins Head of Research and Specialist Computing ITCS Introduction to High Performance Computing at UEA. Chris Collins Head of Research and Specialist Computing ITCS Introduction to High Performance Computing High Performance Computing at UEA http://rscs.uea.ac.uk/hpc/

More information

Introduction to Discovery.

Introduction to Discovery. Introduction to Discovery http://discovery.dartmouth.edu The Discovery Cluster 2 Agenda What is a cluster and why use it Overview of computer hardware in cluster Help Available to Discovery Users Logging

More information

A Brief Introduction to The Center for Advanced Computing

A Brief Introduction to The Center for Advanced Computing A Brief Introduction to The Center for Advanced Computing February 8, 2007 Hardware 376 Opteron nodes, over 890 cores Gigabit networking, Myrinet networking, Infiniband networking soon Hardware: nyx nyx

More information

Using Compute Canada. Masao Fujinaga Information Services and Technology University of Alberta

Using Compute Canada. Masao Fujinaga Information Services and Technology University of Alberta Using Compute Canada Masao Fujinaga Information Services and Technology University of Alberta Introduction to cedar batch system jobs are queued priority depends on allocation and past usage Cedar Nodes

More information

Introduction to CINECA Computer Environment

Introduction to CINECA Computer Environment Introduction to CINECA Computer Environment Today you will learn... Basic commands for UNIX environment @ CINECA How to submitt your job to the PBS queueing system on Eurora Tutorial #1: Example: launch

More information

A Hands-On Tutorial: RNA Sequencing Using High-Performance Computing

A Hands-On Tutorial: RNA Sequencing Using High-Performance Computing A Hands-On Tutorial: RNA Sequencing Using Computing February 11th and 12th, 2016 1st session (Thursday) Preliminaries: Linux, HPC, command line interface Using HPC: modules, queuing system Presented by:

More information

KISTI TACHYON2 SYSTEM Quick User Guide

KISTI TACHYON2 SYSTEM Quick User Guide KISTI TACHYON2 SYSTEM Quick User Guide Ver. 2.4 2017. Feb. SupercomputingCenter 1. TACHYON 2 System Overview Section Specs Model SUN Blade 6275 CPU Intel Xeon X5570 2.93GHz(Nehalem) Nodes 3,200 total Cores

More information

Introduction to HPC Using zcluster at GACRC On-Class GENE 4220

Introduction to HPC Using zcluster at GACRC On-Class GENE 4220 Introduction to HPC Using zcluster at GACRC On-Class GENE 4220 Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu Slides courtesy: Zhoufei Hou 1 OVERVIEW GACRC

More information

HPC Introductory Course - Exercises

HPC Introductory Course - Exercises HPC Introductory Course - Exercises The exercises in the following sections will guide you understand and become more familiar with how to use the Balena HPC service. Lines which start with $ are commands

More information

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing Quick Start Guide by Burak Himmetoglu Supercomputing Consultant Enterprise Technology Services & Center for Scientific Computing E-mail: bhimmetoglu@ucsb.edu Linux/Unix basic commands Basic command structure:

More information

Introduction to GALILEO

Introduction to GALILEO Introduction to GALILEO Parallel & production environment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Alessandro Grottesi a.grottesi@cineca.it SuperComputing Applications and

More information

Introduction to HPC Using the New Cluster at GACRC

Introduction to HPC Using the New Cluster at GACRC Introduction to HPC Using the New Cluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What is the new cluster

More information

Introduction to HPC Using the New Cluster at GACRC

Introduction to HPC Using the New Cluster at GACRC Introduction to HPC Using the New Cluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What is the new cluster

More information

How to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions

How to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions How to run applications on Aziz supercomputer Mohammad Rafi System Administrator Fujitsu Technology Solutions Agenda Overview Compute Nodes Storage Infrastructure Servers Cluster Stack Environment Modules

More information

Effective Use of CCV Resources

Effective Use of CCV Resources Effective Use of CCV Resources Mark Howison User Services & Support This talk... Assumes you have some familiarity with a Unix shell Provides examples and best practices for typical usage of CCV systems

More information

Using the computational resources at the GACRC

Using the computational resources at the GACRC An introduction to zcluster Georgia Advanced Computing Resource Center (GACRC) University of Georgia Dr. Landau s PHYS4601/6601 course - Spring 2017 What is GACRC? Georgia Advanced Computing Resource Center

More information

Guillimin HPC Users Meeting January 13, 2017

Guillimin HPC Users Meeting January 13, 2017 Guillimin HPC Users Meeting January 13, 2017 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Please be kind to your fellow user meeting attendees Limit

More information

Using Cartesius and Lisa. Zheng Meyer-Zhao - Consultant Clustercomputing

Using Cartesius and Lisa. Zheng Meyer-Zhao - Consultant Clustercomputing Zheng Meyer-Zhao - zheng.meyer-zhao@surfsara.nl Consultant Clustercomputing Outline SURFsara About us What we do Cartesius and Lisa Architectures and Specifications File systems Funding Hands-on Logging

More information

Introduction to High-Performance Computing (HPC)

Introduction to High-Performance Computing (HPC) Introduction to High-Performance Computing (HPC) Computer components CPU : Central Processing Unit cores : individual processing units within a CPU Storage : Disk drives HDD : Hard Disk Drive SSD : Solid

More information

GACRC User Training: Migrating from Zcluster to Sapelo

GACRC User Training: Migrating from Zcluster to Sapelo GACRC User Training: Migrating from Zcluster to Sapelo The GACRC Staff Version 1.0 8/28/2017 GACRC Zcluster-Sapelo Migrating Training 1 Discussion Points I. Request Sapelo User Account II. III. IV. Systems

More information

Crash Course in High Performance Computing

Crash Course in High Performance Computing Crash Course in High Performance Computing Cyber-Infrastructure Days October 24, 2013 Dirk Colbry colbrydi@msu.edu Research Specialist Institute for Cyber-Enabled Research https://wiki.hpcc.msu.edu/x/qamraq

More information

PACE Orientation. Research Scientist, PACE

PACE Orientation. Research Scientist, PACE PACE Orientation Mehmet (Memo) Belgin, PhD Research Scientist, PACE www.pace.gatech.edu What is PACE A Partnership for an Advanced Computing Environment Provides faculty and researchers vital tools to

More information

Please include the following sentence in any works using center resources.

Please include the following sentence in any works using center resources. The TCU High-Performance Computing Center The TCU HPCC currently maintains a cluster environment hpcl1.chm.tcu.edu. Work on a second cluster environment is underway. This document details using hpcl1.

More information

Running Jobs, Submission Scripts, Modules

Running Jobs, Submission Scripts, Modules 9/17/15 Running Jobs, Submission Scripts, Modules 16,384 cores total of about 21,000 cores today Infiniband interconnect >3PB fast, high-availability, storage GPGPUs Large memory nodes (512GB to 1TB of

More information

XSEDE New User Tutorial

XSEDE New User Tutorial May 13, 2016 XSEDE New User Tutorial Jay Alameda National Center for Supercomputing Applications XSEDE Training Survey Please complete a short on-line survey about this module at http://bit.ly/hamptonxsede.

More information

Practical Introduction to

Practical Introduction to 1 2 Outline of the workshop Practical Introduction to What is ScaleMP? When do we need it? How do we run codes on the ScaleMP node on the ScaleMP Guillimin cluster? How to run programs efficiently on ScaleMP?

More information

A Brief Introduction to The Center for Advanced Computing

A Brief Introduction to The Center for Advanced Computing A Brief Introduction to The Center for Advanced Computing November 10, 2009 Outline 1 Resources Hardware Software 2 Mechanics: Access Transferring files and data to and from the clusters Logging into the

More information

User Guide of High Performance Computing Cluster in School of Physics

User Guide of High Performance Computing Cluster in School of Physics User Guide of High Performance Computing Cluster in School of Physics Prepared by Sue Yang (xue.yang@sydney.edu.au) This document aims at helping users to quickly log into the cluster, set up the software

More information

XSEDE New User Tutorial

XSEDE New User Tutorial June 12, 2015 XSEDE New User Tutorial Jay Alameda National Center for Supercomputing Applications XSEDE Training Survey Please remember to sign in for today s event: http://bit.ly/1fashvo Also, please

More information

Introduction to Unix Environment: modules, job scripts, PBS. N. Spallanzani (CINECA)

Introduction to Unix Environment: modules, job scripts, PBS. N. Spallanzani (CINECA) Introduction to Unix Environment: modules, job scripts, PBS N. Spallanzani (CINECA) Bologna PATC 2016 In this tutorial you will learn... How to get familiar with UNIX environment @ CINECA How to submit

More information

Choosing Resources Wisely Plamen Krastev Office: 38 Oxford, Room 117 FAS Research Computing

Choosing Resources Wisely Plamen Krastev Office: 38 Oxford, Room 117 FAS Research Computing Choosing Resources Wisely Plamen Krastev Office: 38 Oxford, Room 117 Email:plamenkrastev@fas.harvard.edu Objectives Inform you of available computational resources Help you choose appropriate computational

More information

Introduction to CINECA HPC Environment

Introduction to CINECA HPC Environment Introduction to CINECA HPC Environment 23nd Summer School on Parallel Computing 19-30 May 2014 m.cestari@cineca.it, i.baccarelli@cineca.it Goals You will learn: The basic overview of CINECA HPC systems

More information

UF Research Computing: Overview and Running STATA

UF Research Computing: Overview and Running STATA UF : Overview and Running STATA www.rc.ufl.edu Mission Improve opportunities for research and scholarship Improve competitiveness in securing external funding Matt Gitzendanner magitz@ufl.edu Provide high-performance

More information

Introduction to High Performance Computing (HPC) Resources at GACRC

Introduction to High Performance Computing (HPC) Resources at GACRC Introduction to High Performance Computing (HPC) Resources at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? Concept

More information

XSEDE New User Tutorial

XSEDE New User Tutorial October 20, 2017 XSEDE New User Tutorial Jay Alameda National Center for Supercomputing Applications XSEDE Training Survey Please complete a short on line survey about this module at http://bit.ly/xsedesurvey.

More information

Cryptography Application : SSH. Cyber Security & Network Security March, 2017 Dhaka, Bangladesh

Cryptography Application : SSH. Cyber Security & Network Security March, 2017 Dhaka, Bangladesh Cryptography Application : SSH Cyber Security & Network Security 20-22 March, 2017 Dhaka, Bangladesh Issue Date: [31-12-2015] Revision: [v.1] What is Safely Authentication I am Assured of Which Host I

More information

GPU Cluster Usage Tutorial

GPU Cluster Usage Tutorial GPU Cluster Usage Tutorial How to make caffe and enjoy tensorflow on Torque 2016 11 12 Yunfeng Wang 1 PBS and Torque PBS: Portable Batch System, computer software that performs job scheduling versions

More information

Introduction to HPCC at MSU

Introduction to HPCC at MSU Introduction to HPCC at MSU Chun-Min Chang Research Consultant Institute for Cyber-Enabled Research Download this presentation: https://wiki.hpcc.msu.edu/display/teac/2016-03-17+introduction+to+hpcc How

More information

Cerebro Quick Start Guide

Cerebro Quick Start Guide Cerebro Quick Start Guide Overview of the system Cerebro consists of a total of 64 Ivy Bridge processors E5-4650 v2 with 10 cores each, 14 TB of memory and 24 TB of local disk. Table 1 shows the hardware

More information

Getting started with the CEES Grid

Getting started with the CEES Grid Getting started with the CEES Grid October, 2013 CEES HPC Manager: Dennis Michael, dennis@stanford.edu, 723-2014, Mitchell Building room 415. Please see our web site at http://cees.stanford.edu. Account

More information

Introduction to HPC2N

Introduction to HPC2N Introduction to HPC2N Birgitte Brydsø HPC2N, Umeå University 4 May 2017 1 / 24 Overview Kebnekaise and Abisko Using our systems The File System The Module System Overview Compiler Tool Chains Examples

More information

For Dr Landau s PHYS8602 course

For Dr Landau s PHYS8602 course For Dr Landau s PHYS8602 course Shan-Ho Tsai (shtsai@uga.edu) Georgia Advanced Computing Resource Center - GACRC January 7, 2019 You will be given a student account on the GACRC s Teaching cluster. Your

More information

Introduction to the ITA computer system

Introduction to the ITA computer system Introduction to the ITA computer system Tiago M. D. Pereira Slides: https://folk.uio.no/tiago/teaching/unix2017 Institute of Theoretical Astrophysics Today s lecture in a nutshell 1. Network and users,

More information

Introduction to HPC Using the New Cluster at GACRC

Introduction to HPC Using the New Cluster at GACRC Introduction to HPC Using the New Cluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu 1 Outline What is GACRC? What is the new cluster

More information

Kohinoor queuing document

Kohinoor queuing document List of SGE Commands: qsub : Submit a job to SGE Kohinoor queuing document qstat : Determine the status of a job qdel : Delete a job qhost : Display Node information Some useful commands $qstat f -- Specifies

More information

Introduction to High Performance Computing Using Sapelo2 at GACRC

Introduction to High Performance Computing Using Sapelo2 at GACRC Introduction to High Performance Computing Using Sapelo2 at GACRC Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu 1 Outline High Performance Computing (HPC)

More information

Practical Introduction to Message-Passing Interface (MPI)

Practical Introduction to Message-Passing Interface (MPI) 1 Outline of the workshop 2 Practical Introduction to Message-Passing Interface (MPI) Bart Oldeman, Calcul Québec McGill HPC Bart.Oldeman@mcgill.ca Theoretical / practical introduction Parallelizing your

More information

SuperMike-II Launch Workshop. System Overview and Allocations

SuperMike-II Launch Workshop. System Overview and Allocations : System Overview and Allocations Dr Jim Lupo CCT Computational Enablement jalupo@cct.lsu.edu SuperMike-II: Serious Heterogeneous Computing Power System Hardware SuperMike provides 442 nodes, 221TB of

More information