June 26, Explanatory meeting for users of supercomputer system -- Overview of UGE --

Size: px
Start display at page:

Download "June 26, Explanatory meeting for users of supercomputer system -- Overview of UGE --"

Transcription

1 June 26, 2012 Explanatory meeting for users of supercomputer system -- Overview of UGE --

2 What is Univa Grid Engine (UGE)? It is software that is used to construct a grid computing system. It functions as a batch job system. It is a commercial product derived from Sun Grid Engine 6.2U5 (the last opensource version). The main developers of SGE participated in the development of UGE. The commands, etc., for entering jobs in the UGE are the same as that in SGE. 1

3 Advantages of using UGE Multiple jobs can be smoothly executed in a sequential manner. When more than one user enter several jobs simultaneously, UGE carries out the scheduling. Effective scheduling is carried out according to the memory, CPU, etc., required by a job. Precautions for using UGE Parallelization of jobs and other such functions cannot be performed. If the resource demand when a job is entered is not declared properly, large-scale hang-up of computer may occur. 2

4 How to use a supercomputer 1 Log into the gateway node (gw.ddbj.nig.ac.jp) 2Execute qlogin and log into the interactive node 3 Enter a job from the host for which qlogin is executed into UGE 4 In the UGE, the job is executed at a node with small load 5 The job execution result is output to the home directory, lustre 6 The job execution result is checked 1 Job Job Job gateway 1 unit (activestandby) dedicated to communications with external 2 8 units for professional use, 8 units for research dedicated to interactive 6 3 Result Result Result Result Result 4 81 units for professional use, 208 units for research dedicated to batch 5 3

5 Basic terms (concept) (1) Host (node) Physically existing computer qmaster Job Job execd execd execd execd Master host Host in which master daemon (qmaster) of UGE operates The master daemon is the daemon that controls the UGE, which carries out acceptance of job, scheduling, delivery to execution host, collection, etc. Execution host Host in which the execution daemon (execd) of UGE operates The execution daemon accepts the instruction of job execution from the master daemon and executes the job. 4

6 Basic terms (concept) (2) submit host Job Job qmaster Submit host Host that can enter a job into UGE The execution host for which login is enabled using the qlogin command is a submit host execd execd execd execd execd execd Queue Target of job entry It is configured using more than one execution host. There are several types of queues depending on their application. execd JobSlot JobSlot JobSlot JobSlot JobSlot JobSlot JobSlot JobSlot Job slot Container for executing a job that is set for each execution host. A job is entered into a queue and finally stored in the slot. 5

7 Two UGE environments The supercomputer system has two UGE environments as shown below. The usable environmental setting is configured at login, so a user does not need to configure the settings. UGE environment for DDBJ operation UGE environment that can be used with account for DDBJ operation SGE_ROOT=/home/geadmin/UGES SGE_CELL=uges UGE environment for research UGE environment that can be used with account for general research SGE_ROOT=/home/geadmin/UGER SGE_CELL=uger 6

8 Types of queues (for research, as of June 26) Queue name Number of job slots Upper limit of execution time (days) Application week_hdd.q If queue/resource is not specified, a job is entered in this queue week_ssd.q Used when a job that may take short execution time is executed using ssd month_hdd.q Used when a job that may require long execution time is executed month_ssd.q Used when a job that may take long execution time is executed using ssd month_gpu.q Used when a job that uses gpu is executed month_medium.q Used when a job that uses a medium node is executed month_fat.q Used when a job that uses a fat node is executed debug.q 64 1 Used for checking the operation of a job login.q Used for entering a job 7

9 Types of queues (for commercial use, as of June 10) Queue name Number of job slots Upper limit of execution time (days) Application month_hdd.q If queue/resource is not specified, a job is entered in this queue debug.q 32 1 Used for checking the operation of a job login.q 64 - Used for entering a job 8

10 Upper limit of execution time The upper limit of execution time is set for resolving jamming of jobs that are waiting for execution when a jam is generated. A job that exceeds the upper limit of the execution time is killed. The execution time is computed as the actual time after a job is executed. (It is not the CPU use time. The time of waiting in a queue is not included.) Before entering a job, it is necessary to know the execution time using the environment for checking the operation. 9

11 qlogin When a job is entered, log into the host that has sufficient resource of login.q using the qlogin command from the gateway host (common to environment for both research and DDBJ operation). $ qlogin Your job 329 ("QLOGIN") has been submitted waiting for interactive job to be scheduled... Your interactive job 329 has been successfully scheduled. Establishing builtin session to host t217i... $ uname -n t217 When logging into the execution host, ensure that you use the qlogin command to do so. Do not directly log into the system to execute a job, as the load dispersion mechanism would not function properly. (*A user who directly logs into the system is recorded.) 10

12 Job entry (1) A job is entered by creating a shell script that is described for UGE. An example is shown below (file name: test.sh). #!/bin/sh #$ -S /bin/sh pwd hostname date sleep 20 date echo "to stderr" 1>&2 #$ at the second line is the prefix for specifying the UGE option. The interpreter that is used when this shell script is operated on UGE is specified with #$ -S (in this example, the interpreter is /bin/sh). If this line is omitted, it is necessary to specify -S path of interpreter to be used with the command option when a job is entered. 11

13 Job entry (2) Enter a job using the qsub command $ qsub test.sh When a job is entered, it is inserted in the queue that waits for execution. The status of the entered job can be checked using the qstat command (to be described). After execution, check the output of the job. A file in which the standard output of job and the standard error output are recorded is output to the home directory. $ cat ~/test.sh.o325 /lustre1/home/ddbjuser t 年 3 月 21 日水曜日 11:15:01 JST 2012 年 3 月 21 日水曜日 11:15:21 JST $ cat ~/test.sh.e325 to stderr 12

14 Main options of qsub (1) -S <path of interpreter> -cwd Specify the path of the interpreter when the script file is executed. The interpreter of the script language, such as Perl or Ruby, can be specified in addition to the shell, e.g., (sh is specified): -S /bin/sh (Perl is specified): -S /usr/local/bin/perl A job is executed not in the home directory but in the directory where the qsub command is executed. If this option is specified, the file containing the standard output and standard error output are output to the directory where qsub command is executed. -o <output destination of standard output> -e <output destination of standard error output> Specify the output destinations of the standard output of job and the standard error output. If the standard output or standard error output is not output as a file, specify /dev/null as the output destination, e.g., -o /dev/null -e /dev/null 13

15 Main options of qsub (2) -N <alias name of job> The name of job that can be checked using qstat is changed to the specified name. If it is not specified, the job name is the same as the script name. -l resource demand 1,resource demand 2, -l resource demand 1 l resource demand 2 l They are mainly used for selecting a queue or changing the upper limit of memory usage capacity. Details are described later. 14

16 Checking job status The status of an entered job is checked using the qstat command If a job exists in the waiting queue, qw is displayed as the state $ qstat job-id prior name user state submit/start at test.sh ddbjuser qw 03/19/ :11:56 While a job is executed, "r" is displayed for state $ qstat job-id prior name user state submit/start at test.sh ddbjuser r 03/19/ :11:56 The main states are shown below. States can also be displayed in combinations of more than one. r qw t E d A job is being executed on the execution host A job is waiting in a queue A job is being transferred to the execution host An error is generated in a job A job is being deleted 15

17 Main options of qstat -f -u [uid] The queue usage status is displayed additionally. e.g., qstat f The specified job of [uid] is also displayed. If it is set to " * ", the jobs of all users are displayed. e.g., qstat u * -j [jobid] The detailed information of specified job of [jobid] is checked. The reason of error status Eqw can be checked. e.g., qstat j

18 Deletion of a job The qdel command is used to delete a job. A job is deleted by specifying the job ID or UID. When the job ID is specified (specify only job ID) $ qsub test.sh Your job 326 ("test.sh") has been submitted $ qdel 326 ddbjuser has deleted job 326 When UID is specified (UID is specified using -u option) $ qsub test.sh Your job 327 ("test.sh") has been submitted $ qsub test.sh Your job 328 ("test.sh") has been submitted $ qdel -u ddbjuser ddbjuser has registered the job 327 for deletion ddbjuser has registered the job 328 for deletion 17

19 Checking job execution result The details of a job that has been executed are checked using the qacct command. The resource that was actually consumed by the job, etc., can be checked. $ qacct -j 325 ============================================================== qname week_hdd.q hostname t165i group se owner ddbjuser project NONE (*An omission*) cpu mem io iow maxvmem M arid undefined 18

20 Precautions before entering a job * Before entering multiple jobs, ensure that a test is conducted. >> If the memory is insufficient, several hosts can cause a hang-up >> A lot of error jobs can cause overload of UGE Do not allocate or output an input file or a file of final output to a directory of each local host such, as /tmp and /ssd >> The input cannot be read by a host in which a job is being executed >> The result cannot be referenced after execution *1 The number of jobs that are executed by a job at the same time should be 1. (Do not fork multiple processes with one job without using def_slot (to be described)) *1 The number of threads of a process should be 1. (Do not execute a process with multiple threads without using def_slot (to be described)) >> Load dispersion cannot occur and the host causes a hang-up 19

21 Switching usage of queue (for research) (1) The usage of a queue can be switched by specifying a resource with the "-l" option Do not specify a resource $ qsub test.sh week_hdd.q,week_ssd.q is used The priority order is week_hdd.q > week_ssd.q Specify "month" (* Specify it when a long calculation time is estimated) $ qsub l month test.sh month_hdd.q,month_ssd.q,month_gpu.q is used The priority order is month_hdd.q > month_ssd.q > month_gpu.q Specify "ssd" (* Specify it when a job that uses SSD is entered) $ qsub l ssd test.sh Only week_ssd.q is used 20

22 Switching usage of queue (for research) (2) Speicfy "month" and " ssd" (* Specify them when a long calculation time is estimated for a job that uses SSD) $ qsub l month l ssd test.sh month_ssd.q,month_gpu.q is used The priority order is month_ssd.q > month_gpu.q Specify "month" and " gpu" (* Specify them when a job that uses GPU is entered) $ qsub l month l gpu test.sh Only month_gpu.q is used * When a host on which GPU is mounted is used, ensure that "-l month is specified * Only one job that requires GPU can operate on one host on which GPU is mounted Specify "month" and " medium" (* Specify them when a job that uses a medium node is entered) $ qsub l month l medium test.sh Only month_medium.q is used * When a medium node is used, ensure that "-l month is specified 21

23 Switching usage of queue (for research) (3) Specify "month" and " fat" (* Specify them when a job that uses a fat node is entered) $ qsub l month l fat test.sh Only month_fat.q is used * When a fat node is used, ensure that "-l month is specified Specify "debug" (* Specify it when checking the operation of a job) $ qsub l debug test.sh debug.q is used Specify "debug" and "gpu" (* Specify them when checking the operation of a job that uses GPU) $ qsub l debug l gpu test.sh A host on which GPU is mounted in debug.q is used 22

24 Switching usage of queue (for research) (4) *Caution* To use a GPU, medium, or fat node, ensure that you specify "month" in addition to the resources ("gpu, "medium," "fat, respectively). With the current queue configuration, all of the GPU, medium, and fat nodes are allocated only for a queue for a time-consuming calculation, so it is necessary to specify the resource "month" for using those queues. If "month" is not specified, the current queues have no resource that satisfies the resource specification condition, and so the submission is normally carried out, but the job is not executed. 23

25 Switching usage of queue (for professional use) Do not specify a resource $ qsub test.sh month_hdd.q is used Specify "debug" (* Specify it when checking the operation of a job) $ qsub l debug test.sh debug.q is used 24

26 When large amount of memory is used (1) The memory available for a UGE job is restricted to 4 GB by default. When a large amount of memory is used, declare the memory usage using the "-l" option When 8 GB of memory is used for 1 job $ qsub l s_vmem=8g l mem_req=8g test.sh When 128 GB of memory is used for 1 job on a medium node $ qsub l s_vmem=128g l mem_req=128g l month l medium test.sh 25

27 When large amount of memory is used (2) s_vmem: Declares the upper limit value of virtual memory that can be used by a job. The job cannot use the memory exceeding the amount specified here. mem_req: Declares the memory to be used. For the execution host, the value of "mem_req" is set as an index that indicates the remaining amount of memory, and the value increases or decreases depending on the job execution status. It is used as an index of load dispersion. Job mem_req= 8G Job mem_req= 128G qmaster execd mem_req=64g execd mem_req=64g execd mem_req=2048g Job mem_req= execd 8G mem_req=56g Job execd mem_req= 128G mem_req=1920g 1. While a job is being executed, mem_req value of the host decreases for the portion declared by the job 2. If the job is terminated, mem_req for the portion declared by the job is reset to the original value 3. If mem_req of the host is smaller than mem_req declared by a job, the job is not executed for the host 26

28 Array job (1) If a job is entered as an array job, it is possible to give a different parameter for the same job and execute it repeatedly. If the "-t" option of qsub is used, the array job can be entered. $ cat arraytest.sh #!/bin/sh #$ -S /bin/sh echo --- echo JOB_ID: ${JOB_ID} echo SGE_TASK_ID: ${SGE_TASK_ID} echo SGE_TASK_FIRST: ${SGE_TASK_FIRST} echo SGE_TASK_LAST: ${SGE_TASK_LAST} echo SGE_TASK_STEPSIZE: ${SGE_TASK_STEPSIZE} echo --- $ qsub t 1-6:2 arraytest.sh Your job-array :2 ("arraytest.sh") has been submitted $ qstat job-id prior name user state submit/start at queue slots jatask-id arraytest. ddbjuser r 03/19/ :43:13 week_hdd.q@t168i arraytest. ddbjuser r 03/19/ :43:13 week_hdd.q@t168i arraytest. ddbjuser r 03/19/ :43:13 week_hdd.q@t178i

29 Array job (2) $ ls arraytest.sh.o1031.* arraytest.sh.o arraytest.sh.o arraytest.sh.o $ cat arraytest.sh.o JOB_ID: 1031 SGE_TASK_ID: 1 SGE_TASK_FIRST: 1 SGE_TASK_LAST: 6 SGE_TASK_STEPSIZE: $ cat arraytest.sh.o JOB_ID: 1031 SGE_TASK_ID: 5 SGE_TASK_FIRST: 1 SGE_TASK_LAST: 10 SGE_TASK_STEPSIZE: To prevent UGE overload, the number of jobs that can be entered by one user is limited. If you try to enter jobs exceeding this limit, an error is generated and you cannot enter the jobs. The upper limit of jobs to be entered per user for this system is If a job is entered as an array job, the load on UGE can be reduced. If array jobs of 5000 jobs are entered, the jobs for 5000 * SGE_TASK_ID can be executed. The upper limit of SGE_TASK_ID is

30 MPI job (1) An example of shell script that is used to enter an MPI job is shown below. $ cat mpitest.sh #!/bin/sh #$ -S /bin/sh #$ -pe mpi 2-24 #$ -cwd /usr/local/bin/mpirun -np $NSLOTS -machinefile $TMPDIR/machines./mpitest -pe <MPI execution environment name> <minimum parallel count>-<maximum parallel count> $NSLOTS Specify MPI execution environment (to be described), minimum parallel count, and maximum parallel count Values that are automatically decided from [minimum parallel count] to [maximum parallel count] shown above are set according to the free space of queue -machinefile $TMPDIR/machines The file $TMPDIR/machines is automatically generated by UGE 29

31 MPI job (2) Enter an MPI job in UGE $ qsub mpitest.sh Your job 1292 ("mpitest.sh") has been submitted $ qstat job-id prior name user state submit/start at queue slots ja-task-id mpitest.sh ddbjuser r 03/19/ :55:24 week_hdd.q@t303i 24 $ cat mpitest.sh.o1292 Hellow World from Process 0 of 24 running on t303 Hellow World from Process 1 of 24 running on t290 (* 中略 *) Hellow World from Process 19 of 24 running on t311 Main MPI execution environment mpi: executes parallel jobs using as many hosts as possible mpi-fillup: executes parallel jobs using the same host whenever possible 30

32 Use of parallel environment def_slot (1) Use it when you enter a job that may cause overload if it is entered directly, such as a job that forks multiple processes or a job that executes a multi-thread process. $ qsub pe def_slot 2 test.sh The number of job slots consumed by this job is redefined by a value following "def_slot. In this example, this job consumes two job slots. It is used as an index of a value that specifies the maximum number of processes that are simultaneously initialized in the target job, and the maximum number of threads used by a process initialized in the job. 31

33 Use of parallel environment def_slot (2) *Cautions* If def_slot is specified, the amount of resource demand becomes "resource amount specified by -l" "number of slots specified by def_slot" Note that the excessive resource may be demanded unintentionally. If one of the following options is specified, the amount of resource demand becomes 32 GB. $ qsub pe def_slot 4 l max_vmem=8g l mem_req=8g test.sh If the resource demand is not clarified, the default value is applied. In the following case, the amount of resource demand becomes 16 GB. $ qsub pe def_slot 4 test.sh In the following case, the amount of resource demand becomes 80 GB. However, a thin node does not have a node that meets the condition, so the job is not executed if it is submitted. $ qsub pe def_slot 10 l max_vmem=8g l mem_req=8g test.sh 32

34 Contact information If you have any enquiries or opinions, please contact Supercomputer SE team, National Institute of Genetics Mail: Room: w202 Extension: upercom-intro.html 33

35 Revision history Date of revision Revised content March 21, 2012 May 10, 2012 Newly created The number of job slots for "type of queue" is corrected according to the current situation. Precautions on use of def_slot are added. June 18, 2012 Upper limit of execution time for month_*.q is changed to 62 days from 31 days, type of queue is corrected. June 26, 2012 Configuration of the queues is changed, type of queue is corrected. The types of queues for professional use are reduced, Switching usage of queue (for professional use) is modified. 34

Feb 22, Explanatory meeting for users of supercomputer system -- Knowhow for entering jobs in UGE --

Feb 22, Explanatory meeting for users of supercomputer system -- Knowhow for entering jobs in UGE -- Feb 22, 2013 Explanatory meeting for users of supercomputer system -- Knowhow for entering jobs in UGE -- Purpose of this lecture To understand the method for smoothly executing a few hundreds or thousands

More information

Introduction to HPC Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC Introduction to HPC Using zcluster at GACRC On-class STAT8330 Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu Slides courtesy: Zhoufei Hou 1 Outline What

More information

Shark Cluster Overview

Shark Cluster Overview Shark Cluster Overview 51 Execution Nodes 1 Head Node (shark) 2 Graphical login nodes 800 Cores = slots 714 TB Storage RAW Slide 1/17 Introduction What is a High Performance Compute (HPC) cluster? A HPC

More information

Shark Cluster Overview

Shark Cluster Overview Shark Cluster Overview 51 Execution Nodes 1 Head Node (shark) 1 Graphical login node (rivershark) 800 Cores = slots 714 TB Storage RAW Slide 1/14 Introduction What is a cluster? A cluster is a group of

More information

High Performance Computing (HPC) Using zcluster at GACRC

High Performance Computing (HPC) Using zcluster at GACRC High Performance Computing (HPC) Using zcluster at GACRC On-class STAT8060 Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC?

More information

A Hands-On Tutorial: RNA Sequencing Using High-Performance Computing

A Hands-On Tutorial: RNA Sequencing Using High-Performance Computing A Hands-On Tutorial: RNA Sequencing Using Computing February 11th and 12th, 2016 1st session (Thursday) Preliminaries: Linux, HPC, command line interface Using HPC: modules, queuing system Presented by:

More information

Using the computational resources at the GACRC

Using the computational resources at the GACRC An introduction to zcluster Georgia Advanced Computing Resource Center (GACRC) University of Georgia Dr. Landau s PHYS4601/6601 course - Spring 2017 What is GACRC? Georgia Advanced Computing Resource Center

More information

Cluster User Training

Cluster User Training Cluster User Training From Bash to parallel jobs under SGE in one terrifying hour Christopher Dwan, Bioteam First delivered at IICB, Kolkata, India December 14, 2009 UNIX ESSENTIALS Unix command line essentials

More information

Batch system usage arm euthen F azo he Z J. B T

Batch system usage arm euthen F azo he Z J. B T Batch system usage 10.11.2010 General stuff Computing wikipage: http://dvinfo.ifh.de Central email address for questions & requests: uco-zn@desy.de Data storage: AFS ( /afs/ifh.de/group/amanda/scratch/

More information

X Grid Engine. Where X stands for Oracle Univa Open Son of more to come...?!?

X Grid Engine. Where X stands for Oracle Univa Open Son of more to come...?!? X Grid Engine Where X stands for Oracle Univa Open Son of more to come...?!? Carsten Preuss on behalf of Scientific Computing High Performance Computing Scheduler candidates LSF too expensive PBS / Torque

More information

Introduction to HPC Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC Introduction to HPC Using zcluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu 1 Outline What is GACRC? What is HPC Concept? What

More information

SGE Roll: Users Guide. Version Edition

SGE Roll: Users Guide. Version Edition SGE Roll: Users Guide Version 4.2.1 Edition SGE Roll: Users Guide : Version 4.2.1 Edition Published Sep 2006 Copyright 2006 University of California and Scalable Systems This document is subject to the

More information

National Biochemical Computational Research https://nbcr.net/accounts/apply.php. Familiarize yourself with the account policy

National Biochemical Computational Research  https://nbcr.net/accounts/apply.php. Familiarize yourself with the account policy Track 3: Molecular Visualization and Virtual Screening NBCR Summer Institute Session: NBCR clusters introduction August 11, 2006 Nadya Williams nadya@sdsc.edu Where to start National Biochemical Computational

More information

Introduction to HPC Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC Introduction to HPC Using zcluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What is HPC Concept? What is

More information

Introduction to HPC Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC Introduction to HPC Using zcluster at GACRC On-class PBIO/BINF8350 Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What

More information

HPCC New User Training

HPCC New User Training High Performance Computing Center HPCC New User Training Getting Started on HPCC Resources Eric Rees, Ph.D. High Performance Computing Center Fall 2018 HPCC User Training Agenda HPCC User Training Agenda

More information

An Introduction to Cluster Computing Using Newton

An Introduction to Cluster Computing Using Newton An Introduction to Cluster Computing Using Newton Jason Harris and Dylan Storey March 25th, 2014 Jason Harris and Dylan Storey Introduction to Cluster Computing March 25th, 2014 1 / 26 Workshop design.

More information

Kohinoor queuing document

Kohinoor queuing document List of SGE Commands: qsub : Submit a job to SGE Kohinoor queuing document qstat : Determine the status of a job qdel : Delete a job qhost : Display Node information Some useful commands $qstat f -- Specifies

More information

To connect to the cluster, simply use a SSH or SFTP client to connect to:

To connect to the cluster, simply use a SSH or SFTP client to connect to: RIT Computer Engineering Cluster The RIT Computer Engineering cluster contains 12 computers for parallel programming using MPI. One computer, phoenix.ce.rit.edu, serves as the master controller or head

More information

Joint High Performance Computing Exchange (JHPCE) Cluster Orientation.

Joint High Performance Computing Exchange (JHPCE) Cluster Orientation. Joint High Performance Computing Exchange (JHPCE) Cluster Orientation http://www.jhpce.jhu.edu/ Schedule - Introductions who are we, who are you? - Terminology - Logging in and account setup - Basics of

More information

Quick Guide for the Torque Cluster Manager

Quick Guide for the Torque Cluster Manager Quick Guide for the Torque Cluster Manager Introduction: One of the main purposes of the Aries Cluster is to accommodate especially long-running programs. Users who run long jobs (which take hours or days

More information

Grid Engine Users Guide. 5.5 Edition

Grid Engine Users Guide. 5.5 Edition Grid Engine Users Guide 5.5 Edition Grid Engine Users Guide : 5.5 Edition Published May 08 2012 Copyright 2012 University of California and Scalable Systems This document is subject to the Rocks License

More information

Cluster Clonetroop: HowTo 2014

Cluster Clonetroop: HowTo 2014 2014/02/25 16:53 1/13 Cluster Clonetroop: HowTo 2014 Cluster Clonetroop: HowTo 2014 This section contains information about how to access, compile and execute jobs on Clonetroop, Laboratori de Càlcul Numeric's

More information

Batch Systems. Running calculations on HPC resources

Batch Systems. Running calculations on HPC resources Batch Systems Running calculations on HPC resources Outline What is a batch system? How do I interact with the batch system Job submission scripts Interactive jobs Common batch systems Converting between

More information

Answers to Federal Reserve Questions. Administrator Training for University of Richmond

Answers to Federal Reserve Questions. Administrator Training for University of Richmond Answers to Federal Reserve Questions Administrator Training for University of Richmond 2 Agenda Cluster overview Physics hardware Chemistry hardware Software Modules, ACT Utils, Cloner GridEngine overview

More information

Introduction to HPC Using zcluster at GACRC On-Class GENE 4220

Introduction to HPC Using zcluster at GACRC On-Class GENE 4220 Introduction to HPC Using zcluster at GACRC On-Class GENE 4220 Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu Slides courtesy: Zhoufei Hou 1 OVERVIEW GACRC

More information

SGE Roll: Users Guide. Version 5.3 Edition

SGE Roll: Users Guide. Version 5.3 Edition SGE Roll: Users Guide Version 5.3 Edition SGE Roll: Users Guide : Version 5.3 Edition Published Dec 2009 Copyright 2009 University of California and Scalable Systems This document is subject to the Rocks

More information

Gridengine. Contents. Aim. Configuration of gridengine. From reading group / nlp lunch

Gridengine. Contents. Aim. Configuration of gridengine. From reading group / nlp lunch Gridengine From reading group / nlp lunch Contents 1 Aim 2 Configuration of gridengine 3 Preparation (Login info) 4 How to use gridengine (Hello World Example) 5 Useful Commands 6 Other environmental variables

More information

CGRB Computational Infrastructure

CGRB Computational Infrastructure CGRB Computational Infrastructure shell.cgrb.oregonstate.edu (ssh) waterman CGRB Computational Infrastructure (compute nodes)... shell.cgrb.oregonstate.edu (ssh) waterman CGRB Computational Infrastructure

More information

Grid Engine - A Batch System for DESY. Andreas Haupt, Peter Wegner DESY Zeuthen

Grid Engine - A Batch System for DESY. Andreas Haupt, Peter Wegner DESY Zeuthen Grid Engine - A Batch System for DESY Andreas Haupt, Peter Wegner 15.6.2005 DESY Zeuthen Introduction Motivations for using a batch system more effective usage of available computers (e.g. reduce idle

More information

SINGAPORE-MIT ALLIANCE GETTING STARTED ON PARALLEL PROGRAMMING USING MPI AND ESTIMATING PARALLEL PERFORMANCE METRICS

SINGAPORE-MIT ALLIANCE GETTING STARTED ON PARALLEL PROGRAMMING USING MPI AND ESTIMATING PARALLEL PERFORMANCE METRICS SINGAPORE-MIT ALLIANCE Computational Engineering CME5232: Cluster and Grid Computing Technologies for Science and Computing COMPUTATIONAL LAB NO.2 10 th July 2009 GETTING STARTED ON PARALLEL PROGRAMMING

More information

Batch Systems & Parallel Application Launchers Running your jobs on an HPC machine

Batch Systems & Parallel Application Launchers Running your jobs on an HPC machine Batch Systems & Parallel Application Launchers Running your jobs on an HPC machine Partners Funding Reusing this material This work is licensed under a Creative Commons Attribution- NonCommercial-ShareAlike

More information

Grid Engine Users Guide. 7.0 Edition

Grid Engine Users Guide. 7.0 Edition Grid Engine Users Guide 7.0 Edition Grid Engine Users Guide : 7.0 Edition Published Dec 01 2017 Copyright 2017 University of California and Scalable Systems This document is subject to the Rocks License

More information

HPC DOCUMENTATION. 3. Node Names and IP addresses:- Node details with respect to their individual IP addresses are given below:-

HPC DOCUMENTATION. 3. Node Names and IP addresses:- Node details with respect to their individual IP addresses are given below:- HPC DOCUMENTATION 1. Hardware Resource :- Our HPC consists of Blade chassis with 5 blade servers and one GPU rack server. a.total available cores for computing: - 96 cores. b.cores reserved and dedicated

More information

The cluster system. Introduction 22th February Jan Saalbach Scientific Computing Group

The cluster system. Introduction 22th February Jan Saalbach Scientific Computing Group The cluster system Introduction 22th February 2018 Jan Saalbach Scientific Computing Group cluster-help@luis.uni-hannover.de Contents 1 General information about the compute cluster 2 Available computing

More information

Univa Grid Engine Troubleshooting Quick Reference

Univa Grid Engine Troubleshooting Quick Reference Univa Corporation Grid Engine Documentation Univa Grid Engine Troubleshooting Quick Reference Author: Univa Engineering Version: 8.4.4 October 31, 2016 Copyright 2012 2016 Univa Corporation. All rights

More information

Submit a Job. Want to run a batch script: #!/bin/sh echo Starting job date /usr/bin/time./hello date echo Ending job. qsub A HPC job.

Submit a Job. Want to run a batch script: #!/bin/sh echo Starting job date /usr/bin/time./hello date echo Ending job. qsub A HPC job. Submit a Job Want to run a batch script: #!/bin/sh echo Starting job date /usr/bin/time./hello date echo Ending job Have to ask scheduler to do it. qsub A 20090528HPC job.sge #!/bin/sh #$ -N ht3d-hyb #$

More information

Answers to Federal Reserve Questions. Training for University of Richmond

Answers to Federal Reserve Questions. Training for University of Richmond Answers to Federal Reserve Questions Training for University of Richmond 2 Agenda Cluster Overview Software Modules PBS/Torque Ganglia ACT Utils 3 Cluster overview Systems switch ipmi switch 1x head node

More information

Sun Grid Engine - A Batch System for DESY

Sun Grid Engine - A Batch System for DESY Sun Grid Engine - A Batch System for DESY Wolfgang Friebel, Peter Wegner 28.8.2001 DESY Zeuthen Introduction Motivations for using a batch system more effective usage of available computers (e.g. more

More information

MERCED CLUSTER BASICS Multi-Environment Research Computer for Exploration and Discovery A Centerpiece for Computational Science at UC Merced

MERCED CLUSTER BASICS Multi-Environment Research Computer for Exploration and Discovery A Centerpiece for Computational Science at UC Merced MERCED CLUSTER BASICS Multi-Environment Research Computer for Exploration and Discovery A Centerpiece for Computational Science at UC Merced Sarvani Chadalapaka HPC Administrator University of California

More information

Resource Management Systems

Resource Management Systems Resource Management Systems RMS DCC/FCUP Grid Computing 1 NQE (Network Queue Environment) DCC/FCUP Grid Computing 2 NQE #QSUB eo #QSUB J m #QSUB o %fred@gale/nppa_latte:/home/gale/fred/mary.jjob.output

More information

Installing and running COMSOL 4.3a on a Linux cluster COMSOL. All rights reserved.

Installing and running COMSOL 4.3a on a Linux cluster COMSOL. All rights reserved. Installing and running COMSOL 4.3a on a Linux cluster 2012 COMSOL. All rights reserved. Introduction This quick guide explains how to install and operate COMSOL Multiphysics 4.3a on a Linux cluster. It

More information

Intel Manycore Testing Lab (MTL) - Linux Getting Started Guide

Intel Manycore Testing Lab (MTL) - Linux Getting Started Guide Intel Manycore Testing Lab (MTL) - Linux Getting Started Guide Introduction What are the intended uses of the MTL? The MTL is prioritized for supporting the Intel Academic Community for the testing, validation

More information

Why You Should Consider Grid Computing

Why You Should Consider Grid Computing Why You Should Consider Grid Computing Kenny Daily BIT Presentation 8 January 2007 Outline Motivational Story Electric Fish Grid Computing Overview N1 Sun Grid Engine Software Use of UCI's cluster My Research

More information

Introduction to GALILEO

Introduction to GALILEO Introduction to GALILEO Parallel & production environment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Domenico Guida d.guida@cineca.it Maurizio Cremonesi m.cremonesi@cineca.it

More information

OBTAINING AN ACCOUNT:

OBTAINING AN ACCOUNT: HPC Usage Policies The IIA High Performance Computing (HPC) System is managed by the Computer Management Committee. The User Policies here were developed by the Committee. The user policies below aim to

More information

Name Department/Research Area Have you used the Linux command line?

Name Department/Research Area Have you used the Linux command line? Please log in with HawkID (IOWA domain) Macs are available at stations as marked To switch between the Windows and the Mac systems, press scroll lock twice 9/27/2018 1 Ben Rogers ITS-Research Services

More information

User Guide of High Performance Computing Cluster in School of Physics

User Guide of High Performance Computing Cluster in School of Physics User Guide of High Performance Computing Cluster in School of Physics Prepared by Sue Yang (xue.yang@sydney.edu.au) This document aims at helping users to quickly log into the cluster, set up the software

More information

Using ISMLL Cluster. Tutorial Lec 5. Mohsan Jameel, Information Systems and Machine Learning Lab, University of Hildesheim

Using ISMLL Cluster. Tutorial Lec 5. Mohsan Jameel, Information Systems and Machine Learning Lab, University of Hildesheim Using ISMLL Cluster Tutorial Lec 5 1 Agenda Hardware Useful command Submitting job 2 Computing Cluster http://www.admin-magazine.com/hpc/articles/building-an-hpc-cluster Any problem or query regarding

More information

Job Management on LONI and LSU HPC clusters

Job Management on LONI and LSU HPC clusters Job Management on LONI and LSU HPC clusters Le Yan HPC Consultant User Services @ LONI Outline Overview Batch queuing system Job queues on LONI clusters Basic commands The Cluster Environment Multiple

More information

NBIC TechTrack PBS Tutorial. by Marcel Kempenaar, NBIC Bioinformatics Research Support group, University Medical Center Groningen

NBIC TechTrack PBS Tutorial. by Marcel Kempenaar, NBIC Bioinformatics Research Support group, University Medical Center Groningen NBIC TechTrack PBS Tutorial by Marcel Kempenaar, NBIC Bioinformatics Research Support group, University Medical Center Groningen 1 NBIC PBS Tutorial This part is an introduction to clusters and the PBS

More information

NBIC TechTrack PBS Tutorial

NBIC TechTrack PBS Tutorial NBIC TechTrack PBS Tutorial by Marcel Kempenaar, NBIC Bioinformatics Research Support group, University Medical Center Groningen Visit our webpage at: http://www.nbic.nl/support/brs 1 NBIC PBS Tutorial

More information

Introduction to HPC Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC Introduction to HPC Using zcluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu Slides courtesy: Zhoufei Hou OVERVIEW GACRC High Performance

More information

Advanced Topics in High Performance Scientific Computing [MA5327] Exercise 1

Advanced Topics in High Performance Scientific Computing [MA5327] Exercise 1 Advanced Topics in High Performance Scientific Computing [MA5327] Exercise 1 Manfred Liebmann Technische Universität München Chair of Optimal Control Center for Mathematical Sciences, M17 manfred.liebmann@tum.de

More information

How to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions

How to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions How to run applications on Aziz supercomputer Mohammad Rafi System Administrator Fujitsu Technology Solutions Agenda Overview Compute Nodes Storage Infrastructure Servers Cluster Stack Environment Modules

More information

Assignment 2 Using Paraguin to Create Parallel Programs

Assignment 2 Using Paraguin to Create Parallel Programs Overview Assignment 2 Using Paraguin to Create Parallel Programs C. Ferner and B. Wilkinson Minor clarification Oct 11, 2013 The goal of this assignment is to use the Paraguin compiler to create parallel

More information

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing Quick Start Guide by Burak Himmetoglu Supercomputing Consultant Enterprise Technology Services & Center for Scientific Computing E-mail: bhimmetoglu@ucsb.edu Contents User access, logging in Linux/Unix

More information

GPU Cluster Usage Tutorial

GPU Cluster Usage Tutorial GPU Cluster Usage Tutorial How to make caffe and enjoy tensorflow on Torque 2016 11 12 Yunfeng Wang 1 PBS and Torque PBS: Portable Batch System, computer software that performs job scheduling versions

More information

Migrating from Zcluster to Sapelo

Migrating from Zcluster to Sapelo GACRC User Quick Guide: Migrating from Zcluster to Sapelo The GACRC Staff Version 1.0 8/4/17 1 Discussion Points I. Request Sapelo User Account II. III. IV. Systems Transfer Files Configure Software Environment

More information

Submitting and running jobs on PlaFRIM2 Redouane Bouchouirbat

Submitting and running jobs on PlaFRIM2 Redouane Bouchouirbat Submitting and running jobs on PlaFRIM2 Redouane Bouchouirbat Summary 1. Submitting Jobs: Batch mode - Interactive mode 2. Partition 3. Jobs: Serial, Parallel 4. Using generic resources Gres : GPUs, MICs.

More information

Grid Computing. Department of Statistics Staff Retreat February 2008

Grid Computing. Department of Statistics Staff Retreat February 2008 Slide 1 Grid Computing Department of Statistics Staff Retreat February 2008 This is a talk I gave to the Department in February 2008. I followed from a talk by Werner regarding the Department s future

More information

Introduction to HPC Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC Introduction to HPC Using zcluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu Slides courtesy: Zhoufei Hou 1 OVERVIEW GACRC High Performance

More information

Batch Systems. Running your jobs on an HPC machine

Batch Systems. Running your jobs on an HPC machine Batch Systems Running your jobs on an HPC machine Reusing this material This work is licensed under a Creative Commons Attribution- NonCommercial-ShareAlike 4.0 International License. http://creativecommons.org/licenses/by-nc-sa/4.0/deed.en_us

More information

OpenPBS Users Manual

OpenPBS Users Manual How to Write a PBS Batch Script OpenPBS Users Manual PBS scripts are rather simple. An MPI example for user your-user-name: Example: MPI Code PBS -N a_name_for_my_parallel_job PBS -l nodes=7,walltime=1:00:00

More information

Parallelism. Wolfgang Kastaun. May 9, 2008

Parallelism. Wolfgang Kastaun. May 9, 2008 Parallelism Wolfgang Kastaun May 9, 2008 Outline Parallel computing Frameworks MPI and the batch system Running MPI code at TAT The CACTUS framework Overview Mesh refinement Writing Cactus modules Links

More information

Unix Processes. What is a Process?

Unix Processes. What is a Process? Unix Processes Process -- program in execution shell spawns a process for each command and terminates it when the command completes Many processes all multiplexed to a single processor (or a small number

More information

ITCS 4145/5145 Assignment 2

ITCS 4145/5145 Assignment 2 ITCS 4145/5145 Assignment 2 Compiling and running MPI programs Author: B. Wilkinson and Clayton S. Ferner. Modification date: September 10, 2012 In this assignment, the workpool computations done in Assignment

More information

How to Run NCBI BLAST on zcluster at GACRC

How to Run NCBI BLAST on zcluster at GACRC How to Run NCBI BLAST on zcluster at GACRC BLAST: Basic Local Alignment Search Tool Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu 1 OVERVIEW What is BLAST?

More information

Intermediate Usage of Shared Computing Cluster (SCC) Charles Jahnke Research Computing Services Information Services & Technology

Intermediate Usage of Shared Computing Cluster (SCC) Charles Jahnke Research Computing Services Information Services & Technology Intermediate Usage of Shared Computing Cluster (SCC) Charles Jahnke Research Computing Services Information Services & Technology Topics for Today Introduction The Cluster Running Jobs Not Hands-on Monitoring

More information

HPC Resources at Lehigh. Steve Anthony March 22, 2012

HPC Resources at Lehigh. Steve Anthony March 22, 2012 HPC Resources at Lehigh Steve Anthony March 22, 2012 HPC at Lehigh: Resources What's Available? Service Level Basic Service Level E-1 Service Level E-2 Leaf and Condor Pool Altair Trits, Cuda0, Inferno,

More information

Logging in to the CRAY

Logging in to the CRAY Logging in to the CRAY 1. Open Terminal Cray Hostname: cray2.colostate.edu Cray IP address: 129.82.103.183 On a Mac 2. type ssh username@cray2.colostate.edu where username is your account name 3. enter

More information

Advanced Research Computing. ARC3 and GPUs. Mark Dixon

Advanced Research Computing. ARC3 and GPUs. Mark Dixon Advanced Research Computing Mark Dixon m.c.dixon@leeds.ac.uk ARC3 (1st March 217) Included 2 GPU nodes, each with: 24 Intel CPU cores & 128G RAM (same as standard compute node) 2 NVIDIA Tesla K8 24G RAM

More information

Introduction to HPC Resources and Linux

Introduction to HPC Resources and Linux Introduction to HPC Resources and Linux Burak Himmetoglu Enterprise Technology Services & Center for Scientific Computing e-mail: bhimmetoglu@ucsb.edu Paul Weakliem California Nanosystems Institute & Center

More information

KISTI TACHYON2 SYSTEM Quick User Guide

KISTI TACHYON2 SYSTEM Quick User Guide KISTI TACHYON2 SYSTEM Quick User Guide Ver. 2.4 2017. Feb. SupercomputingCenter 1. TACHYON 2 System Overview Section Specs Model SUN Blade 6275 CPU Intel Xeon X5570 2.93GHz(Nehalem) Nodes 3,200 total Cores

More information

Programming Environment on Ranger Cluster

Programming Environment on Ranger Cluster Programming Environment on Ranger Cluster Cornell Center for Advanced Computing December 8, 2010 12/8/2010 www.cac.cornell.edu 1 User Guides TACC Ranger (http://services.tacc.utexas.edu/index.php/ranger-user-guide)

More information

ACEnet for CS6702 Ross Dickson, Computational Research Consultant 29 Sep 2009

ACEnet for CS6702 Ross Dickson, Computational Research Consultant 29 Sep 2009 ACEnet for CS6702 Ross Dickson, Computational Research Consultant 29 Sep 2009 What is ACEnet? Shared resource......for research computing... physics, chemistry, oceanography, biology, math, engineering,

More information

Computational Skills Primer. Lecture 2

Computational Skills Primer. Lecture 2 Computational Skills Primer Lecture 2 Your Background Who has used SCC before? Who has worked on any other cluster? Do you have previous experience working with basic linux and command line usage (CLI)?

More information

How to for compiling and running MPI Programs. Prepared by Kiriti Venkat

How to for compiling and running MPI Programs. Prepared by Kiriti Venkat How to for compiling and running MPI Programs. Prepared by Kiriti Venkat What is MPI? MPI stands for Message Passing Interface MPI is a library specification of message-passing, proposed as a standard

More information

Computing with the Moore Cluster

Computing with the Moore Cluster Computing with the Moore Cluster Edward Walter An overview of data management and job processing in the Moore compute cluster. Overview Getting access to the cluster Data management Submitting jobs (MPI

More information

HPCC User Group Meeting

HPCC User Group Meeting High Performance Computing Center HPCC User Group Meeting Planned Update of Quanah and Hrothgar Eric Rees Research Associate - HPCC June 6, 2018 HPCC User Group Meeting Agenda 6/6/2018 HPCC User Group

More information

Duke Compute Cluster Workshop. 10/04/2018 Tom Milledge rc.duke.edu

Duke Compute Cluster Workshop. 10/04/2018 Tom Milledge rc.duke.edu Duke Compute Cluster Workshop 10/04/2018 Tom Milledge rc.duke.edu rescomputing@duke.edu Outline of talk Overview of Research Computing resources Duke Compute Cluster overview Running interactive and batch

More information

Getting started with the CEES Grid

Getting started with the CEES Grid Getting started with the CEES Grid October, 2013 CEES HPC Manager: Dennis Michael, dennis@stanford.edu, 723-2014, Mitchell Building room 415. Please see our web site at http://cees.stanford.edu. Account

More information

la a quota v lfs quota /lustre/ df /tmp $ dd if=/dev/zero of=/lustre/$user/4g.dat bs=1m count=4096 4096+0 records in 4096+0 records out 4294967296 bytes (4.3 GB) copied, 41.0291 seconds, 105 MB/s $

More information

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing Quick Start Guide by Burak Himmetoglu Supercomputing Consultant Enterprise Technology Services & Center for Scientific Computing E-mail: bhimmetoglu@ucsb.edu Linux/Unix basic commands Basic command structure:

More information

CycleServer Grid Engine Support Install Guide. version

CycleServer Grid Engine Support Install Guide. version CycleServer Grid Engine Support Install Guide version 1.34.4 Contents CycleServer Grid Engine Guide 1 Administration 1 Requirements 1 Installation 1 Monitoring Additional Grid Engine Clusters 3 Monitoring

More information

Parallel Programming Assignment 3 Compiling and running MPI programs

Parallel Programming Assignment 3 Compiling and running MPI programs Parallel Programming Assignment 3 Compiling and running MPI programs Author: Clayton S. Ferner and B. Wilkinson Modification date: October 11a, 2013 This assignment uses the UNC-Wilmington cluster babbage.cis.uncw.edu.

More information

Introduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende

Introduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende Introduction to the NCAR HPC Systems 25 May 2018 Consulting Services Group Brian Vanderwende Topics to cover Overview of the NCAR cluster resources Basic tasks in the HPC environment Accessing pre-built

More information

The Team. Using Sun Grid Engine and Globus to Schedule Jobs Across a Combination of Local and Remote Machines

The Team. Using Sun Grid Engine and Globus to Schedule Jobs Across a Combination of Local and Remote Machines The Team Using Sun Grid Engine and Globus to Schedule Jobs Across a Combination of Local and Remote Machines By Geoff Cawood and Paul Graham Edinburgh Parallel Computing Centre (EPCC) Telephone: +44 131

More information

Duke Compute Cluster Workshop. 3/28/2018 Tom Milledge rc.duke.edu

Duke Compute Cluster Workshop. 3/28/2018 Tom Milledge rc.duke.edu Duke Compute Cluster Workshop 3/28/2018 Tom Milledge rc.duke.edu rescomputing@duke.edu Outline of talk Overview of Research Computing resources Duke Compute Cluster overview Running interactive and batch

More information

UoW HPC Quick Start. Information Technology Services University of Wollongong. ( Last updated on October 10, 2011)

UoW HPC Quick Start. Information Technology Services University of Wollongong. ( Last updated on October 10, 2011) UoW HPC Quick Start Information Technology Services University of Wollongong ( Last updated on October 10, 2011) 1 Contents 1 Logging into the HPC Cluster 3 1.1 From within the UoW campus.......................

More information

High Performance Beowulf Cluster Environment User Manual

High Performance Beowulf Cluster Environment User Manual High Performance Beowulf Cluster Environment User Manual Version 3.1c 2 This guide is intended for cluster users who want a quick introduction to the Compusys Beowulf Cluster Environment. It explains how

More information

PBS Pro Documentation

PBS Pro Documentation Introduction Most jobs will require greater resources than are available on individual nodes. All jobs must be scheduled via the batch job system. The batch job system in use is PBS Pro. Jobs are submitted

More information

HPCC - Hrothgar Getting Started User Guide Gromacs

HPCC - Hrothgar Getting Started User Guide Gromacs HPCC - Hrothgar Getting Started User Guide Gromacs High Performance Computing Center Texas Tech University HPCC - Hrothgar 2 Table of Contents 1. Introduction... 3 2. Setting up the environment... 3 For

More information

New User Seminar: Part 2 (best practices)

New User Seminar: Part 2 (best practices) New User Seminar: Part 2 (best practices) General Interest Seminar January 2015 Hugh Merz merz@sharcnet.ca Session Outline Submitting Jobs Minimizing queue waits Investigating jobs Checkpointing Efficiency

More information

Exercises: Abel/Colossus and SLURM

Exercises: Abel/Colossus and SLURM Exercises: Abel/Colossus and SLURM November 08, 2016 Sabry Razick The Research Computing Services Group, USIT Topics Get access Running a simple job Job script Running a simple job -- qlogin Customize

More information

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved. Minnesota Supercomputing Institute Introduction to Job Submission and Scheduling Andrew Gustafson Interacting with MSI Systems Connecting to MSI SSH is the most reliable connection method Linux and Mac

More information

UF Research Computing: Overview and Running STATA

UF Research Computing: Overview and Running STATA UF : Overview and Running STATA www.rc.ufl.edu Mission Improve opportunities for research and scholarship Improve competitiveness in securing external funding Matt Gitzendanner magitz@ufl.edu Provide high-performance

More information

The Supercomputing Facility for Bioinformatics & Computational Biology, IIT Delhi

The Supercomputing Facility for Bioinformatics & Computational Biology, IIT Delhi Supercomputing Facility for Bioinformatics & Computational Biology, IIT Delhi The Supercomputing Facility for Bioinformatics & Computational Biology, IIT Delhi User Manual Dr. B. Jayaram (Professor of

More information

Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop

Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop Before We Start Sign in hpcxx account slips Windows Users: Download PuTTY Google PuTTY First result Save putty.exe to Desktop Research Computing at Virginia Tech Advanced Research Computing Compute Resources

More information

To connect to the cluster, simply use a SSH or SFTP client to connect to:

To connect to the cluster, simply use a SSH or SFTP client to connect to: RIT Computer Engineering Cluster The RIT Computer Engineering cluster contains 12 computers for parallel programming using MPI. One computer, cluster-head.ce.rit.edu, serves as the master controller or

More information