Introduction to Advanced Research Computing (ARC)
|
|
- Estella Osborne
- 5 years ago
- Views:
Transcription
1 Introduction to Advanced Research Computing (ARC) September 29, 2016 By: Pier-Luc St-Onge 1
2 Financial Partners 2
3 Setup for the workshop 1. Get a user ID and password paper (provided in class): ##: ********** 2. Access to local computer (replace ## and with appropriate values, is provided in class): a. User name: csuser## (ex.: csuser99) b. (ex.: sec@[s99) 3. Slides: 3
4 Outline Introduction to ARC Laptop vs HPC Cluster Moving Data With Globus Software Development Types of Parallelism Scientific Software and Modules Best Practices The Scheduler and Job Submission Compute Canada and Calcul Québec 4
5 Introduction to ARC 5
6 Introduction This workshop is built from the HPCS 2015 Introduction to ARC workshop: HPCS website: Etherpad with extra information: Presentation: Instructors and participants Name and research domain 6
7 Other research domains Genomics Molecular Dynamics (MD) Finite Element Analysis (FEA) Computational Fluid Dynamics (CFD) Astronomy and Astrophysics Geospatial Data Analysis 3D renderings, Image Analysis Digital Humanities (DH) Artificial Intelligence 7
8 Advanced Research Computing High Performance Computing (HPC) Parallel code Running code on accelerators Profiling code Management of massive data Complete pipelines for jobs submission Sharing data Web portals for special projects Cloud computing / storage 8
9 ARC is also about having enough Resources Problem: we want to compute a dense matrix multiplication C = A * B Each matrix is made of * floating point numbers in double precision How much memory do you need for this job? a) 3.2 GB b) 8.6 GB c) 12.9 GB d) 25.8 GB 9
10 Different Tools for Different Needs Portable computer HPC Compute node 2-4 cores 4-16 GB Local storage, USB 128GB to 2TB Wi-Fi and Ethernet General usage Cloud storage Gaming GPU 32 bits Direct access 8-32 cores GB Network storage 150TB to 5PB Ethernet and fast network (Infiniband) Multi-node tasks Accelerators 64 bits, ECC mem. Access by queued jobs 10
11 Zoom on a HPC Compute Node RAM Proc. Proc. RAM Infiniband RAM Local hard-drive (scratch) RAM Proc. Proc. Ethernet RAM Proc. Proc. RAM Accelerator RAM RAM Proc. Proc. 11
12 Why Infiniband? Some tasks need more than one compute node Split workload, memory usage and storage access High bandwidth Ethernet: 1Gbps (~100MB/s) Infiniband: Gbps (2-10GB/s) Best for I/O on large files Reduce latency Ethernet: ~30µs (33k iops = I/O operations per sec.) Infiniband: 2µs (500k iops) Best for several small communications 12
13 Storage Available on Compute Nodes Bandwidth RAMDISK Local scratch Project space Scratch space Capacity (Not to scale) Latency 13
14 Typical Cluster Job Login node Scheduler Job Nodes Ethernet & Infiniband Internet Storage home projet scratch ssh 14
15 Typical Cluster Usage 1. Connect to a login node a. Import data b. Compile code or install applications 2. Submit jobs to the scheduler a. Define needed resources: processors, accelerators, memory, walltime limit b. Wait in queue - monitor your position c. Automatic execution of your jobs on compute nodes 3. Get your results 15
16 Typical Environment on Cluster The operating system is usually Linux (CentOS, Scientific Linux, ) To connect to a cluster, we usually use an encrypted session through SSH (Secure SHell) To transfer data safely: SCP (Secure CoPy) and others tools Some web portals may also provide an access Development environment Multiple text editors Scientific software Version control software Tools for transferring large amount of data 10Gbps, 40Gbps or 100Gbps between sites 16
17 Moving Data With Globus 17
18 Moving Data Over SSH Secure CoPy with SCP or WinSCP: Most clusters only allow SSH connections SSH: Secure SHell (username and password) SCP uses the SSH protocol to copy files Simple tool for small amount of data (<1GB) On Windows: use WinSCP or FileZilla Two panels, where one side shows your files on the cluster, then drag-and-drop files or folders On Linux or Mac OS X: use the scp command: scp file.txt scp -pr 18
19 Globus - Why? Secure CoPy with SCP or WinSCP: One process and one thread Fails on dropped connections BBCP - parallel transfers Need a calibration for the round-trip time Command line with multiple arguments Globus - Web interface and CLI No need for calibration Convenient Web interface Confirmation after the transfer Possible to share data with other users 19
20 Go to Compute Canada Research Portal CC > Research Portal > National Services > Data Movement (Globus) Create an account if you do not have Or Log in to your account if you already have an account 20
21 Account Creation and Login Go to: Select Compute Canada if you have a CCDB account, or Globus ID if not. To create a Globus ID account: 21
22 Compute Canada Login and Globus ID Account Creation 22
23 Exercise: Sign-up! Globus ID credentials are stored on Globus: Can be a different username than the one on CCDB, or it can be the same. But both accounts (on CCDB and Globus ID) are different! For sure, use a different password! Put the green Post-it on the computer screen when done 23
24 Globus Home Interface To Logout Endpoint Viewports Transfer options Globus Endpoint on your computer 24
25 Choosing an Endpoint at Compute Canada 25
26 Authentication to an Endpoint The authentication is done on a server corresponding to the endpoint Globus will never know your credentials for a given cluster/endpoint The authentication server only grants access to Globus. 26
27 Files and Directories 27
28 Exercise: Import Pictures Connect to computecanada#guillimin Use the class account credentials (on your paper) class## ********** Connect to intro-arc_guillimin_plstonge Select both exercises and pictures folders Click on the blue button to initiate the transfer to your (class##) home directory Wait for the report When the transfer is completed, put the green Post-it on top of the computer screen 28
29 Confirmation TASK DETAILS Task ID: 5ef1f2c6-81f5-11e6-afd b92c261 Task Type: TRANSFER Status: SUCCEEDED Source: intro-arc_guillimin_plstonge (...) Destination: computecanada#guillimin (...) Label: n/a 29
30 Transfer Report 30
31 Managing Endpoints 31
32 Create a Local Endpoint 32
33 Create a Local Endpoint 1. Choose a unique Endpoint name a. Generate the Setup Key b. Copy the Setup Key in a text editor 2. Download and Install Globus Connect Personal a. Start it and paste the Setup Key 3. A Globus icon ( g) appears in the taskbar a. [Right-]click -> Preferences -> Access b. Select accessible directories you want to export 4. Use your Endpoint: user#uniquename 33
34 Sharing Option 34
35 Software Development Types of Parallelism 35
36 Vocabulary Serial tasks Any task that cannot be split in two simultaneous sequences of actions Examples: starting a process, reading a file, any communication between two processes Parallel tasks Data parallelism: same action applied on different data. Could be serial tasks done in parallel. Process parallelism: one action on one set of data. Action split in multiple processes or threads. Data partitioning: rectangles or blocks 36
37 Parallel tasks Shared memory Multiple threads share the same memory space in a single process: full read and write access. Distributed memory Each process has its own memory space Information is sent and received by messages Parallel efficiency (scaling) Amdahl s law: how long does it take to compute a task with an infinite number of processors? Gustafson's law: what size of problem can we solve in a given time with N processors? 37
38 Software Development Scientific Software and Modules 38
39 Connecting to Guillimin - SSH Client From Linux, open a Terminal window From Mac OS X, open a Terminal window From Windows, install PuTTY, MobaXterm or any other SSH (2.0 compliant) client /download.html putty.exe is a standalone executable Simply run putty.exe 39
40 Connecting to Guillimin Terminal Server name: guillimin.calculquebec.ca Username: class## ## = Two digits on the first line of the small piece of paper Password: ********** Sequence of 10 characters on small piece of paper Linux/Mac OS X example: ssh class99@guillimin.calculquebec.ca Accept the server s key fingerprint 40
41 Connecting to Guillimin PuTTY 41
42 Scientific Software You will use or create scientific software Applications Libraries Make sure you are using a code language widely used in your research domain Make sure your code is well documented and easy to maintain Keep in mind that future researchers are going to use your software Do not reinvent the wheel: use modules! 42
43 Modules Exercise: module avail Compilers: GNU, Intel, LLVM, PGI, Java Language interpreters: Python, R Libraries and other applications: OpenMPI, MVAPICH2, Boost MKL, FFTW GROMACS, NAMD, Quantum-Espresso, NetCDF Other tools: Paraview, GNU-Parallel, GDB, Valgrind 43
44 Exercises with Modules which gcc module add gcc/4.8.2 module list which gcc module rm gcc/4.8.2 module add ifort_icc openmpi echo $LD_LIBRARY_PATH echo $CC module purge 44
45 New Lmod/EasyBuild-based module system Default set of modules on Colosse Both legacy and Lmod modules are available on Guillimin Exercise on Guillimin: 1. Load the main toolchain Intel + OpenMPI + MKL: module load iomkl/2015b 2. List loaded modules: module list 3. You now have access to much more modules: module avail 45
46 Introduction to GNU-Parallel Automatic online process scheduler for serial tasks User provides a command line template and one or multiple sequences of arguments For N cores on a compute node, GNU-Parallel will start N processes As soon as one process is complete, it will start the next process according to values in sequences Example: module load parallel/ parallel echo {1} ::: $(seq 1 960) less 46
47 Introduction to Threads Threads are used for parallel tasks on a single node One process, one shared memory space, but multiple threads Synchronization issues, like race conditions Two threads reading or writing to the same variable Instead of POSIX threads: OpenMP Threading Building Blocks 47
48 Introduction to MPI Message Passing Interface (MPI) is a standard for passing messages between multiple processes on multiple nodes Each process can know its rank number Same set of functions, but different implementations MPICH2, MVAPICH2, OpenMPI, Intel MPI,... Example: module load iomkl/2015b mpiexec -n 4 hostname 48
49 Software Development Best Practices 49
50 Version Control & Revision Control Bad practice: Keeping multiple versions in sub-directories Sending a specific version by Trying to merge two versions manually Good practice: using a version control system Choose one that fits your project: Git (git) Mercurial (hg) Subversion (svn) Repository: shared on a remote server 50
51 Exercises with git mkdir -p ~/projecta; cd ~/projecta; module add git git init # Local repository touch script.sh; git status git add script.sh # Ready to commit git commit -m "intro-arc" # Do commit Add line env > env.log in script.sh git git git git diff # See differences add script.sh # Ready to commit commit -m "Cleaner job" # Do commit log # Show past commits 51
52 One Last Good Practice Using SSH Keys Having access with a local passphrase: ssh-keygen Generating public/private rsa key pair. Enter file in which to save the key (~/.ssh/id_rsa): Enter passphrase (empty for no passphrase): Enter same passphrase again: Copy the public key to a remote site (id_rsa.pub): scp.ssh/id_rsa.pub <user>@<server>: ssh <user>@<server> \ "cat id_rsa.pub >>.ssh/authorized_keys" 52
53 Exercise - Code Compilation Go to the exercises folder cd ~/exercises Load the following modules: iomkl/2015b Boost/ Python ImageMagick/ Build the application: make 53
54 The Scheduler and Job Submission 54
55 The Scheduler Jobs are not running on login nodes You need to submit your jobs to the scheduler: Resources you need: n processors, memory, accelerators Job duration Account name The script to run on worker nodes Where you want your results 55
56 Schedulers in Calcul Québec Job component Torque + Maui/Moab Submission command Torque: qsub Maui/Moab: msub Account -A xyz-123-aa Processors (small L) Some clusters force full nodes -l nodes=n:ppn=p -l procs=pp Memory (small L) -l pmem=1700m -l mem=20g Job duration (small L) -l walltime=d:hh:mm:ss Job name -N name Where you want your results -o output_path -e error_path your_app > out.log 56
57 Environment Variables During Job Variable $PBS_JOBID $PBS_JOBNAME $PBS_NODEFILE Description Job ID Ex.: server.name Job name set by -N Ex.: introarc The name of the file that contains a list of assigned nodes for the job $PBS_O_WORKDIR The current working directory ($PWD) when the job has been submitted. $PBS_ARRAYID One of the values specified by the job array: -t 1-3,6,
58 Exercise - Manage your Jobs Torque - jobs and job array details qstat -u $USER qstat -t -u $USER Maui / Moab - jobs showq -u $USER (Optional) Torque - cancel job qdel jobid (Optional) Maui / Moab - cancel job mjobctl -c jobid canceljob jobid 58
59 Exercises: Edit and Submit a Job Scripts Go to the exercises folder Check the README.en file for general instructions Start with the exercise 1-base. All the instructions are in the README.en file The solution is in solution.sh Then try exercises 2, 3 and/or 6 For exercise 2, a description of & and wait is in the README.en file of the exercise You need to understand the solution of the exercise 3 to do exercise 4 59
60 Compute Canada and Calcul Québec 60
61 Compute Canada Four consortia (from West to East): WestGrid British Columbia, Alberta, Saskatchewan, Manitoba Compute Ontario (SciNet + SHARCNET + HPCVL) Ontario Calcul Québec Québec ACENET New Brunswick, Newfoundland, Nova Scotia, Prince Edward Island 61
62 Example - Calcul Québec Consortium Compute Canada Calcul Québec - Compute Resources Briarée & Hadès (U. Montréal) Colosse & Helios (U. Laval) Guillimin (McGill) MS2 & MP2 (U. Sherbrooke) Calcul Québec - Human Resources About 40 employees ARC analysts System administrators Scientific and operational directors 62
63 Biggest Clusters - GP1, GP2 and GP3 are coming soon... Name / Institution Cores Summary MP2 / U. Sherbrooke cores/node, 32+ GB/node, QDR IB GPC / U. Toronto cores/node, 16+ GB/node, D/QDR IB Guillimin / McGill U ,16 cores/node, 24+ GB/node, QDR IB Orcinus / U. British Columbia ,12 cores/node, 16+ GB/node, D/QDR IB Orca / SHARCNET ,24 cores/node, 32 GB/node, QDR IB Colosse / U. Laval cores/node, 24+ GB/node, QDR IB Briarée / U. Montréal cores/node, 24+ GB/node, QDR IB Parallel / U. Calgary cores/node, 24 GB/node, QDR IB Example of job requirements: 8 or 16 cores/node, 30 GB/node, QDR IB 63
64 Roles and Allocations CCRI : abc ABCD University Alloc.: abc-123-aa (default alloc.) CCRI : def HIJK University Alloc.: def-456-ab 50 core*years xyz Prof Prof CCRI : abc DEFG University Alloc.: abc-123-ac 200 core*years xyz Student 64
65 About Allocations Amount of compute cycles (only a target): 1 core provides 2 to 3 GHz (cycles per second) Core*seconds, core*hours or core*years (cy) Each site has a default allocation (up to 80 cy) Amount of storage space (hard limit): Private scratch space (default: up to 1TB) Group shared project space (default: up to 1TB) Special allocations Resource Allocation Competition (RAC 2017) ource-allocation-competitions/ 65
66 How to Access Resources The Complete Procedure CCDB: A professor creates an account in CCDB Request role of Principal Investigator (PI) - CCRI (Optional) Request a special allocation (RAC) Special allocation granted on a specific cluster A student creates an account in CCDB Request sponsored role - need to provide the PI s CCRI Apply for a consortium account (ex.: Calcul Québec) In the Calcul Québec portal, activate your access to any cluster Choose clusters based on special allocations 66
67 Usage Policy Each user has his own account; never share your private account! No heavy process (in time and space) on login nodes Resources must be used in an optimal way Large parallel jobs must use tools that scale well Cold data should not stay on disk Maximize the amount of research done in Canada Must acknowledge Compute Canada and any regional partner organization 67
68 Service and Support Training workshops: Future: Past: Documentation: Documentation Support: 68
69 Conclusion 69
70 Conclusion Need help? WestGrid: SciNet: SHARCNET: Calcul Québec: ACENET: Useful information to send us: User name Job ID (if applicable) Path to {job script, output/error files, code} or as attached files 70
Introduction to Advanced Research Computing
Introduction to Advanced Research Computing and Best Practices January 28, 2016 By: Pier-Luc St-Onge 1 Partners and sponsors 2 Setup for the workshop 1. Get a user ID and password paper (provided in class):
More informationGuillimin HPC Users Meeting March 17, 2016
Guillimin HPC Users Meeting March 17, 2016 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Outline Compute Canada News System Status Software Updates Training
More informationIntroduction to High Performance Computing
Introduction to High Performance Computing By Pier-Luc St-Onge pier-luc.st-onge@mcgill.ca September 11, 2014 Objectives Familiarize new users with the concepts of High Performance Computing (HPC). Outline
More informationGuillimin HPC Users Meeting October 20, 2016
Guillimin HPC Users Meeting October 20, 2016 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Please be kind to your fellow user meeting attendees Limit
More informationGuillimin HPC Users Meeting February 11, McGill University / Calcul Québec / Compute Canada Montréal, QC Canada
Guillimin HPC Users Meeting February 11, 2016 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Compute Canada News Scheduler Updates Software Updates Training
More informationOutline. March 5, 2012 CIRMMT - McGill University 2
Outline CLUMEQ, Calcul Quebec and Compute Canada Research Support Objectives and Focal Points CLUMEQ Site at McGill ETS Key Specifications and Status CLUMEQ HPC Support Staff at McGill Getting Started
More informationGuillimin HPC Users Meeting March 16, 2017
Guillimin HPC Users Meeting March 16, 2017 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Please be kind to your fellow user meeting attendees Limit to
More informationBefore We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop
Before We Start Sign in hpcxx account slips Windows Users: Download PuTTY Google PuTTY First result Save putty.exe to Desktop Research Computing at Virginia Tech Advanced Research Computing Compute Resources
More informationQuick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing
Quick Start Guide by Burak Himmetoglu Supercomputing Consultant Enterprise Technology Services & Center for Scientific Computing E-mail: bhimmetoglu@ucsb.edu Contents User access, logging in Linux/Unix
More informationSupercomputing environment TMA4280 Introduction to Supercomputing
Supercomputing environment TMA4280 Introduction to Supercomputing NTNU, IMF February 21. 2018 1 Supercomputing environment Supercomputers use UNIX-type operating systems. Predominantly Linux. Using a shell
More informationPARALLEL COMPUTING IN R USING WESTGRID CLUSTERS STATGEN GROUP MEETING 10/30/2017
PARALLEL COMPUTING IN R USING WESTGRID CLUSTERS STATGEN GROUP MEETING 10/30/2017 PARALLEL COMPUTING Dataset 1 Processor Dataset 2 Dataset 3 Dataset 4 R script Processor Processor Processor WHAT IS ADVANCED
More informationComputing with the Moore Cluster
Computing with the Moore Cluster Edward Walter An overview of data management and job processing in the Moore compute cluster. Overview Getting access to the cluster Data management Submitting jobs (MPI
More informationGraham vs legacy systems
New User Seminar Graham vs legacy systems This webinar only covers topics pertaining to graham. For the introduction to our legacy systems (Orca etc.), please check the following recorded webinar: SHARCNet
More informationIntroduction to HPC Using the New Cluster at GACRC
Introduction to HPC Using the New Cluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What is the new cluster
More informationGuillimin HPC Users Meeting June 16, 2016
Guillimin HPC Users Meeting June 16, 2016 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Compute Canada News System Status Software Updates Training News
More informationIntroduction to PICO Parallel & Production Enviroment
Introduction to PICO Parallel & Production Enviroment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Domenico Guida d.guida@cineca.it Nicola Spallanzani n.spallanzani@cineca.it
More informationAllinea DDT Debugger. Dan Mazur, McGill HPC March 5,
Allinea DDT Debugger Dan Mazur, McGill HPC daniel.mazur@mcgill.ca guillimin@calculquebec.ca March 5, 2015 1 Outline Introduction and motivation Guillimin login and DDT configuration Compiling for a debugger
More informationIntroduction to the SHARCNET Environment May-25 Pre-(summer)school webinar Speaker: Alex Razoumov University of Ontario Institute of Technology
Introduction to the SHARCNET Environment 2010-May-25 Pre-(summer)school webinar Speaker: Alex Razoumov University of Ontario Institute of Technology available hardware and software resources our web portal
More informationIntroduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende
Introduction to the NCAR HPC Systems 25 May 2018 Consulting Services Group Brian Vanderwende Topics to cover Overview of the NCAR cluster resources Basic tasks in the HPC environment Accessing pre-built
More informationIntroduction to HPC Resources and Linux
Introduction to HPC Resources and Linux Burak Himmetoglu Enterprise Technology Services & Center for Scientific Computing e-mail: bhimmetoglu@ucsb.edu Paul Weakliem California Nanosystems Institute & Center
More informationHigh Performance Computing (HPC) Club Training Session. Xinsheng (Shawn) Qin
High Performance Computing (HPC) Club Training Session Xinsheng (Shawn) Qin Outline HPC Club The Hyak Supercomputer Logging in to Hyak Basic Linux Commands Transferring Files Between Your PC and Hyak Submitting
More informationGuillimin HPC Users Meeting. Bryan Caron
July 17, 2014 Bryan Caron bryan.caron@mcgill.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Outline Compute Canada News Upcoming Maintenance Downtime in August Storage System
More informationThe cluster system. Introduction 22th February Jan Saalbach Scientific Computing Group
The cluster system Introduction 22th February 2018 Jan Saalbach Scientific Computing Group cluster-help@luis.uni-hannover.de Contents 1 General information about the compute cluster 2 Available computing
More informationOur new HPC-Cluster An overview
Our new HPC-Cluster An overview Christian Hagen Universität Regensburg Regensburg, 15.05.2009 Outline 1 Layout 2 Hardware 3 Software 4 Getting an account 5 Compiling 6 Queueing system 7 Parallelization
More informationHigh Performance Computing (HPC) Using zcluster at GACRC
High Performance Computing (HPC) Using zcluster at GACRC On-class STAT8060 Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC?
More informationUsing a Linux System 6
Canaan User Guide Connecting to the Cluster 1 SSH (Secure Shell) 1 Starting an ssh session from a Mac or Linux system 1 Starting an ssh session from a Windows PC 1 Once you're connected... 1 Ending an
More informationThe JANUS Computing Environment
Research Computing UNIVERSITY OF COLORADO The JANUS Computing Environment Monte Lunacek monte.lunacek@colorado.edu rc-help@colorado.edu What is JANUS? November, 2011 1,368 Compute nodes 16,416 processors
More informationNew User Tutorial. OSU High Performance Computing Center
New User Tutorial OSU High Performance Computing Center TABLE OF CONTENTS Logging In... 3-5 Windows... 3-4 Linux... 4 Mac... 4-5 Changing Password... 5 Using Linux Commands... 6 File Systems... 7 File
More informationUoW HPC Quick Start. Information Technology Services University of Wollongong. ( Last updated on October 10, 2011)
UoW HPC Quick Start Information Technology Services University of Wollongong ( Last updated on October 10, 2011) 1 Contents 1 Logging into the HPC Cluster 3 1.1 From within the UoW campus.......................
More informationGuillimin HPC Users Meeting. Bart Oldeman
June 19, 2014 Bart Oldeman bart.oldeman@mcgill.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Outline Compute Canada News Upcoming Maintenance Downtime in August Storage System
More informationPACE. Instructional Cluster Environment (ICE) Orientation. Research Scientist, PACE
PACE Instructional Cluster Environment (ICE) Orientation Mehmet (Memo) Belgin, PhD Research Scientist, PACE www.pace.gatech.edu What is PACE A Partnership for an Advanced Computing Environment Provides
More informationA Brief Introduction to The Center for Advanced Computing
A Brief Introduction to The Center for Advanced Computing May 1, 2006 Hardware 324 Opteron nodes, over 700 cores 105 Athlon nodes, 210 cores 64 Apple nodes, 128 cores Gigabit networking, Myrinet networking,
More informationIntroduction to HPC Using zcluster at GACRC
Introduction to HPC Using zcluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What is HPC Concept? What is
More informationMigrating from Zcluster to Sapelo
GACRC User Quick Guide: Migrating from Zcluster to Sapelo The GACRC Staff Version 1.0 8/4/17 1 Discussion Points I. Request Sapelo User Account II. III. IV. Systems Transfer Files Configure Software Environment
More informationIntroduction to HPC Using zcluster at GACRC
Introduction to HPC Using zcluster at GACRC On-class PBIO/BINF8350 Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What
More informationIntroduction to GALILEO
November 27, 2016 Introduction to GALILEO Parallel & production environment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it SuperComputing Applications and Innovation Department
More informationIntroduction to GALILEO
Introduction to GALILEO Parallel & production environment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Domenico Guida d.guida@cineca.it Maurizio Cremonesi m.cremonesi@cineca.it
More informationPACE. Instructional Cluster Environment (ICE) Orientation. Mehmet (Memo) Belgin, PhD Research Scientist, PACE
PACE Instructional Cluster Environment (ICE) Orientation Mehmet (Memo) Belgin, PhD www.pace.gatech.edu Research Scientist, PACE What is PACE A Partnership for an Advanced Computing Environment Provides
More informationIntroduction to Discovery.
Introduction to Discovery http://discovery.dartmouth.edu The Discovery Cluster 2 Agenda What is a cluster and why use it Overview of computer hardware in cluster Help Available to Discovery Users Logging
More informationUsing Sapelo2 Cluster at the GACRC
Using Sapelo2 Cluster at the GACRC New User Training Workshop Georgia Advanced Computing Resource Center (GACRC) EITS/University of Georgia Zhuofei Hou zhuofei@uga.edu 1 Outline GACRC Sapelo2 Cluster Diagram
More informationIntroduction to HPC Using zcluster at GACRC
Introduction to HPC Using zcluster at GACRC On-class STAT8330 Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu Slides courtesy: Zhoufei Hou 1 Outline What
More informationPractical Introduction to Message-Passing Interface (MPI)
1 Practical Introduction to Message-Passing Interface (MPI) October 1st, 2015 By: Pier-Luc St-Onge Partners and Sponsors 2 Setup for the workshop 1. Get a user ID and password paper (provided in class):
More informationNew User Seminar: Part 2 (best practices)
New User Seminar: Part 2 (best practices) General Interest Seminar January 2015 Hugh Merz merz@sharcnet.ca Session Outline Submitting Jobs Minimizing queue waits Investigating jobs Checkpointing Efficiency
More informationGuillimin HPC Users Meeting July 14, 2016
Guillimin HPC Users Meeting July 14, 2016 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Outline Compute Canada News System Status Software Updates Training
More informationXSEDE New User Tutorial
April 2, 2014 XSEDE New User Tutorial Jay Alameda National Center for Supercomputing Applications XSEDE Training Survey Make sure you sign the sign in sheet! At the end of the module, I will ask you to
More informationGuillimin HPC Users Meeting April 13, 2017
Guillimin HPC Users Meeting April 13, 2017 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Please be kind to your fellow user meeting attendees Limit to
More informationICS-ACI System Basics
ICS-ACI System Basics Adam W. Lavely, Ph.D. Fall 2017 Slides available: goo.gl/ss9itf awl5173 ICS@PSU 1 Contents 1 Overview 2 HPC Overview 3 Getting Started on ACI 4 Moving On awl5173 ICS@PSU 2 Contents
More informationMinnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.
Minnesota Supercomputing Institute Introduction to Job Submission and Scheduling Andrew Gustafson Interacting with MSI Systems Connecting to MSI SSH is the most reliable connection method Linux and Mac
More informationIntroduc)on to Hyades
Introduc)on to Hyades Shawfeng Dong Department of Astronomy & Astrophysics, UCSSC Hyades 1 Hardware Architecture 2 Accessing Hyades 3 Compu)ng Environment 4 Compiling Codes 5 Running Jobs 6 Visualiza)on
More informationACEnet for CS6702 Ross Dickson, Computational Research Consultant 29 Sep 2009
ACEnet for CS6702 Ross Dickson, Computational Research Consultant 29 Sep 2009 What is ACEnet? Shared resource......for research computing... physics, chemistry, oceanography, biology, math, engineering,
More informationWorking on the NewRiver Cluster
Working on the NewRiver Cluster CMDA3634: Computer Science Foundations for Computational Modeling and Data Analytics 22 February 2018 NewRiver is a computing cluster provided by Virginia Tech s Advanced
More informationOBTAINING AN ACCOUNT:
HPC Usage Policies The IIA High Performance Computing (HPC) System is managed by the Computer Management Committee. The User Policies here were developed by the Committee. The user policies below aim to
More informationIntroduction to High Performance Computing at UEA. Chris Collins Head of Research and Specialist Computing ITCS
Introduction to High Performance Computing at UEA. Chris Collins Head of Research and Specialist Computing ITCS Introduction to High Performance Computing High Performance Computing at UEA http://rscs.uea.ac.uk/hpc/
More informationIntroduction to Discovery.
Introduction to Discovery http://discovery.dartmouth.edu The Discovery Cluster 2 Agenda What is a cluster and why use it Overview of computer hardware in cluster Help Available to Discovery Users Logging
More informationA Brief Introduction to The Center for Advanced Computing
A Brief Introduction to The Center for Advanced Computing February 8, 2007 Hardware 376 Opteron nodes, over 890 cores Gigabit networking, Myrinet networking, Infiniband networking soon Hardware: nyx nyx
More informationUsing Compute Canada. Masao Fujinaga Information Services and Technology University of Alberta
Using Compute Canada Masao Fujinaga Information Services and Technology University of Alberta Introduction to cedar batch system jobs are queued priority depends on allocation and past usage Cedar Nodes
More informationIntroduction to CINECA Computer Environment
Introduction to CINECA Computer Environment Today you will learn... Basic commands for UNIX environment @ CINECA How to submitt your job to the PBS queueing system on Eurora Tutorial #1: Example: launch
More informationA Hands-On Tutorial: RNA Sequencing Using High-Performance Computing
A Hands-On Tutorial: RNA Sequencing Using Computing February 11th and 12th, 2016 1st session (Thursday) Preliminaries: Linux, HPC, command line interface Using HPC: modules, queuing system Presented by:
More informationKISTI TACHYON2 SYSTEM Quick User Guide
KISTI TACHYON2 SYSTEM Quick User Guide Ver. 2.4 2017. Feb. SupercomputingCenter 1. TACHYON 2 System Overview Section Specs Model SUN Blade 6275 CPU Intel Xeon X5570 2.93GHz(Nehalem) Nodes 3,200 total Cores
More informationIntroduction to HPC Using zcluster at GACRC On-Class GENE 4220
Introduction to HPC Using zcluster at GACRC On-Class GENE 4220 Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu Slides courtesy: Zhoufei Hou 1 OVERVIEW GACRC
More informationHPC Introductory Course - Exercises
HPC Introductory Course - Exercises The exercises in the following sections will guide you understand and become more familiar with how to use the Balena HPC service. Lines which start with $ are commands
More informationQuick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing
Quick Start Guide by Burak Himmetoglu Supercomputing Consultant Enterprise Technology Services & Center for Scientific Computing E-mail: bhimmetoglu@ucsb.edu Linux/Unix basic commands Basic command structure:
More informationIntroduction to GALILEO
Introduction to GALILEO Parallel & production environment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Alessandro Grottesi a.grottesi@cineca.it SuperComputing Applications and
More informationIntroduction to HPC Using the New Cluster at GACRC
Introduction to HPC Using the New Cluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What is the new cluster
More informationIntroduction to HPC Using the New Cluster at GACRC
Introduction to HPC Using the New Cluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What is the new cluster
More informationHow to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions
How to run applications on Aziz supercomputer Mohammad Rafi System Administrator Fujitsu Technology Solutions Agenda Overview Compute Nodes Storage Infrastructure Servers Cluster Stack Environment Modules
More informationEffective Use of CCV Resources
Effective Use of CCV Resources Mark Howison User Services & Support This talk... Assumes you have some familiarity with a Unix shell Provides examples and best practices for typical usage of CCV systems
More informationUsing the computational resources at the GACRC
An introduction to zcluster Georgia Advanced Computing Resource Center (GACRC) University of Georgia Dr. Landau s PHYS4601/6601 course - Spring 2017 What is GACRC? Georgia Advanced Computing Resource Center
More informationGuillimin HPC Users Meeting January 13, 2017
Guillimin HPC Users Meeting January 13, 2017 guillimin@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Please be kind to your fellow user meeting attendees Limit
More informationUsing Cartesius and Lisa. Zheng Meyer-Zhao - Consultant Clustercomputing
Zheng Meyer-Zhao - zheng.meyer-zhao@surfsara.nl Consultant Clustercomputing Outline SURFsara About us What we do Cartesius and Lisa Architectures and Specifications File systems Funding Hands-on Logging
More informationIntroduction to High-Performance Computing (HPC)
Introduction to High-Performance Computing (HPC) Computer components CPU : Central Processing Unit cores : individual processing units within a CPU Storage : Disk drives HDD : Hard Disk Drive SSD : Solid
More informationGACRC User Training: Migrating from Zcluster to Sapelo
GACRC User Training: Migrating from Zcluster to Sapelo The GACRC Staff Version 1.0 8/28/2017 GACRC Zcluster-Sapelo Migrating Training 1 Discussion Points I. Request Sapelo User Account II. III. IV. Systems
More informationCrash Course in High Performance Computing
Crash Course in High Performance Computing Cyber-Infrastructure Days October 24, 2013 Dirk Colbry colbrydi@msu.edu Research Specialist Institute for Cyber-Enabled Research https://wiki.hpcc.msu.edu/x/qamraq
More informationPACE Orientation. Research Scientist, PACE
PACE Orientation Mehmet (Memo) Belgin, PhD Research Scientist, PACE www.pace.gatech.edu What is PACE A Partnership for an Advanced Computing Environment Provides faculty and researchers vital tools to
More informationPlease include the following sentence in any works using center resources.
The TCU High-Performance Computing Center The TCU HPCC currently maintains a cluster environment hpcl1.chm.tcu.edu. Work on a second cluster environment is underway. This document details using hpcl1.
More informationRunning Jobs, Submission Scripts, Modules
9/17/15 Running Jobs, Submission Scripts, Modules 16,384 cores total of about 21,000 cores today Infiniband interconnect >3PB fast, high-availability, storage GPGPUs Large memory nodes (512GB to 1TB of
More informationXSEDE New User Tutorial
May 13, 2016 XSEDE New User Tutorial Jay Alameda National Center for Supercomputing Applications XSEDE Training Survey Please complete a short on-line survey about this module at http://bit.ly/hamptonxsede.
More informationPractical Introduction to
1 2 Outline of the workshop Practical Introduction to What is ScaleMP? When do we need it? How do we run codes on the ScaleMP node on the ScaleMP Guillimin cluster? How to run programs efficiently on ScaleMP?
More informationA Brief Introduction to The Center for Advanced Computing
A Brief Introduction to The Center for Advanced Computing November 10, 2009 Outline 1 Resources Hardware Software 2 Mechanics: Access Transferring files and data to and from the clusters Logging into the
More informationUser Guide of High Performance Computing Cluster in School of Physics
User Guide of High Performance Computing Cluster in School of Physics Prepared by Sue Yang (xue.yang@sydney.edu.au) This document aims at helping users to quickly log into the cluster, set up the software
More informationXSEDE New User Tutorial
June 12, 2015 XSEDE New User Tutorial Jay Alameda National Center for Supercomputing Applications XSEDE Training Survey Please remember to sign in for today s event: http://bit.ly/1fashvo Also, please
More informationIntroduction to Unix Environment: modules, job scripts, PBS. N. Spallanzani (CINECA)
Introduction to Unix Environment: modules, job scripts, PBS N. Spallanzani (CINECA) Bologna PATC 2016 In this tutorial you will learn... How to get familiar with UNIX environment @ CINECA How to submit
More informationChoosing Resources Wisely Plamen Krastev Office: 38 Oxford, Room 117 FAS Research Computing
Choosing Resources Wisely Plamen Krastev Office: 38 Oxford, Room 117 Email:plamenkrastev@fas.harvard.edu Objectives Inform you of available computational resources Help you choose appropriate computational
More informationIntroduction to CINECA HPC Environment
Introduction to CINECA HPC Environment 23nd Summer School on Parallel Computing 19-30 May 2014 m.cestari@cineca.it, i.baccarelli@cineca.it Goals You will learn: The basic overview of CINECA HPC systems
More informationUF Research Computing: Overview and Running STATA
UF : Overview and Running STATA www.rc.ufl.edu Mission Improve opportunities for research and scholarship Improve competitiveness in securing external funding Matt Gitzendanner magitz@ufl.edu Provide high-performance
More informationIntroduction to High Performance Computing (HPC) Resources at GACRC
Introduction to High Performance Computing (HPC) Resources at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? Concept
More informationXSEDE New User Tutorial
October 20, 2017 XSEDE New User Tutorial Jay Alameda National Center for Supercomputing Applications XSEDE Training Survey Please complete a short on line survey about this module at http://bit.ly/xsedesurvey.
More informationCryptography Application : SSH. Cyber Security & Network Security March, 2017 Dhaka, Bangladesh
Cryptography Application : SSH Cyber Security & Network Security 20-22 March, 2017 Dhaka, Bangladesh Issue Date: [31-12-2015] Revision: [v.1] What is Safely Authentication I am Assured of Which Host I
More informationGPU Cluster Usage Tutorial
GPU Cluster Usage Tutorial How to make caffe and enjoy tensorflow on Torque 2016 11 12 Yunfeng Wang 1 PBS and Torque PBS: Portable Batch System, computer software that performs job scheduling versions
More informationIntroduction to HPCC at MSU
Introduction to HPCC at MSU Chun-Min Chang Research Consultant Institute for Cyber-Enabled Research Download this presentation: https://wiki.hpcc.msu.edu/display/teac/2016-03-17+introduction+to+hpcc How
More informationCerebro Quick Start Guide
Cerebro Quick Start Guide Overview of the system Cerebro consists of a total of 64 Ivy Bridge processors E5-4650 v2 with 10 cores each, 14 TB of memory and 24 TB of local disk. Table 1 shows the hardware
More informationGetting started with the CEES Grid
Getting started with the CEES Grid October, 2013 CEES HPC Manager: Dennis Michael, dennis@stanford.edu, 723-2014, Mitchell Building room 415. Please see our web site at http://cees.stanford.edu. Account
More informationIntroduction to HPC2N
Introduction to HPC2N Birgitte Brydsø HPC2N, Umeå University 4 May 2017 1 / 24 Overview Kebnekaise and Abisko Using our systems The File System The Module System Overview Compiler Tool Chains Examples
More informationFor Dr Landau s PHYS8602 course
For Dr Landau s PHYS8602 course Shan-Ho Tsai (shtsai@uga.edu) Georgia Advanced Computing Resource Center - GACRC January 7, 2019 You will be given a student account on the GACRC s Teaching cluster. Your
More informationIntroduction to the ITA computer system
Introduction to the ITA computer system Tiago M. D. Pereira Slides: https://folk.uio.no/tiago/teaching/unix2017 Institute of Theoretical Astrophysics Today s lecture in a nutshell 1. Network and users,
More informationIntroduction to HPC Using the New Cluster at GACRC
Introduction to HPC Using the New Cluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu 1 Outline What is GACRC? What is the new cluster
More informationKohinoor queuing document
List of SGE Commands: qsub : Submit a job to SGE Kohinoor queuing document qstat : Determine the status of a job qdel : Delete a job qhost : Display Node information Some useful commands $qstat f -- Specifies
More informationIntroduction to High Performance Computing Using Sapelo2 at GACRC
Introduction to High Performance Computing Using Sapelo2 at GACRC Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu 1 Outline High Performance Computing (HPC)
More informationPractical Introduction to Message-Passing Interface (MPI)
1 Outline of the workshop 2 Practical Introduction to Message-Passing Interface (MPI) Bart Oldeman, Calcul Québec McGill HPC Bart.Oldeman@mcgill.ca Theoretical / practical introduction Parallelizing your
More informationSuperMike-II Launch Workshop. System Overview and Allocations
: System Overview and Allocations Dr Jim Lupo CCT Computational Enablement jalupo@cct.lsu.edu SuperMike-II: Serious Heterogeneous Computing Power System Hardware SuperMike provides 442 nodes, 221TB of
More information