Log into Jasper.westgrid.ca: ssh -X use pucy if you are working in windows Log into the workshop cluster from jasper:

Size: px
Start display at page:

Download "Log into Jasper.westgrid.ca: ssh -X use pucy if you are working in windows Log into the workshop cluster from jasper:"

Transcription

1 Using a cluster effec/vely Scheduling and Job Management Log into Jasper.westgrid.ca: ssh -X yourusername@jasper.westgrid.ca use pucy if you are working in windows Log into the workshop cluster from jasper: ssh X cl2n230 Copy the working directory to your own and go into it. cp -r /global/solware/workshop/scheduling-wg cd scheduling-wg-2015 You can find a copy of the slides and materials for this workshop in the following link hcps://goo.gl/m1qmcq

2 Scheduling and Job Management 2 Using a cluster effec/vely

3 Presenta/on contents Job submission part 2 Understanding Jobs

4 PBS Jobs and memory It is very important to specify memory correctly If you don t ask for enough and your job uses more,your job will be killed. If you ask for too much, it will take a much longer /me to schedule a job, and you will be was/ng resources. If you ask for more memory than is available on the cluster your job will never run. The scheduling system will not stop you from submiyng such a job or even warn you. If you don t know how much memory your jobs will need ask for a large amount in your first job and run checkjob v v <jobid> or qstat f <jobid>. Along other informa/on, you should see how much memory your job used. If you don t specify any memory then your job will get a very small default maximum memory (256MB on Jasper).

5 PBS Jobs and memory Always ask for slightly less than total memory on node as some memory is used for OS, and your job will not start un/l enough memory is available. You may specify the maximum memory available to your job in one of 2 ways. Ask for a total memory used by your jobs #PBS l mem=24000mb Ask for memory used per process/core in your job #PBS l pmem=2000mb

6 Features and Par//ons Some/mes nodes have certain proper/es: fast processor, bigger disk, SSD, Fast connec/on or they belong to certain research group. Such nodes are given a feature name by the sysadmin so you can ask for the nodes by feature name in your pbs job script. If you would like to specify that your job will run on nodes with feature ssd: #PBS l feature=ssd Par//ons are set of nodes that you can specify and your job must run with all of its components in a par//on to run in the newer node par//on on orcinus: #PBS -l par//on=qdr

7 PBS jobs and GPUS To request GPU use the nodes nota/on and add :gpu=x for #PBS l nodes=2:gpus=3:ppn=4 Modern torque scheduling programs recognize GPUs as well as the state of the GPU.

8 SoLware licenses and generic resources Some/mes not only cluster hardware is required to be scheduled for a job but other resources as well, such as solware licenses, telescope or other instrument /me. To request generic resources or licenses: #PBS -W x=gres:matlab=2 #PBS -l other=matlab=2 You can see the list of solware licenses and generic resources available on the cluster with the jobinfo n command.

9 PBS script commands PBS script command #PBS -l mem=4gb #PBS l pmem=4gb #PBS l feature=ssd #PBS -l par//on=qdr #PBS l nodes=2:blue:ppn=2 #PBS l nodes=2:gpus=3:ppn=4 #PBS l nodes=cl2n002+cl2n003 #PBS l host=cl2n002 #PBS -I Descrip2on Requests 4 GB of memory in total Requests 4GB of memory per process Requests 1 procesor on node with a feature ssd Requests to run in the QDR par//on Request 2 cores on each of 2 nodes with blue feature. Request 4 cores and 2 gpus on each of 2 nodes Requests 2 nodes cl2n002 and cl2n003 Requests host or node cl2n002 Request an interac/ve Job

10 Memory, Features, SoLware licenses BREAK FOR PRACTICE

11 Job Submission Requiring Full nodes Some/mes there is a need for exclusive access to guarantee that no other job will be running on the same nodes as your job To guarantee that the job will only run on nodes with other jobs you own use: #PBS -l naccesspolicy=singleuser To guarantee that the job will only run on nodes with no other Job use: #PBS -n #PBS -l naccesspolicy=singlejob To guarantee that the each part of the job will only run on a separate node without anything else running on that node use: #PBS -l naccesspolicy=singletask Your group may get charged for using the whole node and not just the resources requested, and it may take a long /me to gather resources needed for these special jobs.

12 Job submission mul/ple projects If you are part of two different WestGrid projects and are running jobs for both, you need to specify the accoun/ng group for each project so that the correct priority of the job can be determined and so that the usage is charged to the correct group. In order to specify an accoun/ng group for a Job use: #PBS A <accoun/ng group> You can find more informa/on about your accoun/ng groups (RAPI) on the WestGrid s accounts portal: hcps://portal.westgrid.ca/user/my_account.php You can see your accoun/ng group informa/on with the jobinfo a command.

13 Job dependencies If you want one job to start one aler another finishes use the qsub W depend=alerok:<jobid1> job2.pbs If one can break apart a long job into several shorter jobs then the shorter jobs will olen be able to be ran faster. This is also the technique to use if the required job run/me is longer than the maximum wall/me allowed on the cluster. jobn1= $( qsub job1.pbs ) qsub -W depend=alerok:$jobn1 job2.pbs

14 Prologue, Epilogue and Data staging Prologue script runs before your job starts for a maximum of 5 minutes. #PBS -l prologue=/home/fujinaga/prologue.script Epilogue script runs aler your job is finished for a maximum of 5 minutes. #PBS -l epilogue=/home/fujinaga/epilogue.script These scripts are nice if you need to document some more informa/on about the state of your job in the scheduling system. Jobs can resubmit themselves with an appropriate script in the epilogue on some systems.

15 sample epilogue.script #!/bin/sh export MYOUTPUT="$HOME/$1-epilogue.out echo "Epilogue Args:" echo "Job ID: $1" echo "User ID: $2" echo "Group ID: $3" echo "Job Name: $4" echo "Session ID: $5" echo "Resource List: $6" echo "Resources Used: $7" echo "Queue Name: $8" echo "Account String: $9" echo exit 0

16 Temporary available local storage Some solware like Gaussian needs to write and read many small files to disk. The cluster (lustre) file system cannot do this well and this becomes a performance problem for the job and the cluster its running on. Each node has local disk, that is shared by all jobs running on the node. One specifies the requests the local storage via #PBS l file=1000mb. There is a directory created for each job when it is run. When the job finished this directory is automa/cally erased. The directory name is $TMPDIR. A example of using the temporary local storage: #PBS l file=1000mb cd $TMPDIR <run my job > mkdir $HOME/$PBS_JOBID/ cp <file I wish to save> $HOME/$PBS_JOBID/

17 PBS script commands PBS script command #PBS -l naccesspolicy=singleuser #PBS -l naccesspolicy=singlejob #PBS -n #PBS -l naccesspolicy=singletask #PBS A <accoun/ng group> #PBS -W x=gres:matlab=2 #PBS -l other=matlab=2 qsub W depend=alerok:<job1id> j2.pbs #PBS -l epilogue=/home/fujinaga/ epilogue.script #PBS l prologue=/home/ fujinaga/prologue.script #PBS l nodes=5:ppn=12+nodes=1:ppn=1 Descrip2on Requests to only run on nodes with other jobs of same user Requests to only run on nodes with no other jobs Requests that the each part of the job will only run on a separate node without anything else running on that node. Requests that a specific accoun/ng group be used for this job Requests 2 units of a generic resource or solware license MATLAB Job 2 that depends on job1 and will not start un/l job1 completes successfully. Runs epilogue script for maximum of 5 minutes aler job is complete. Runs prologue script for maximum of 5 minutes before job is complete. Requests 5 nodes with 12 processors each and a single node with 1 core.

18 PBS Environment Variables Environment Variable PBS_JOBNAME PBS_ARRAYID PBS_GPUFILE PBS_O_WORKDIR PBS_TASKNUM PBS_O_HOME PBS_JOBID PBS_NUM_NODES PBS_NUM_PPN PBS_O_HOST PBS_QUEUE PBS_NODEFILE PBS_O_PATH Descrip2on User specified job name Job array index for this job list of GPUs allocated to the job located 1 per line: <host>-gpu<number> Job's submission directory Number of tasks requested Home directory of submiyng user Unique pbs job id Number of nodes allocated to the job Number of procs per node allocated to the job Host on which job script is currently running Job queue File containing line delimited list on nodes allocated to the job Path variable used to locate executables within job script

19 Job submission prac/ce BREAK FOR PRACTICE

20 GeYng informa/on on your Job Command jobinfo -j qstat t u $USER qstat a qstat r showq showq i showq b qstat f <Jobid> checkjob <Jobid> checkjob -v -v <Jobid> What its used for List all your jobs and their state List all your array jobs and the subcomponents and their state. List all jobs on the system and their state. List all running jobs on the system. List all jobs on the system and their state. List all jobs being considered for scheduling and their priority Lists all blocked (unable to be run) jobs List detailed informa/on on Job List detailed informa/on on Job List detailed informa/on on Job, including history and why it is not running now on each node.

21 jobinfo -j ~]$ jobinfo -j JobID State Proc WCLimit User Opsys Class Features Running 1 3:00:00:00 kamil - batch Idle 12 8:00:00 kamil - batch Idle 12 8:00:00 kamil - batch Idle 12 8:00:00 kamil - batch Idle 12 8:00:00 kamil - batch Idle 12 8:00:00 kamil - batch Idle 12 8:00:00 kamil - batch Idle 12 8:00:00 kamil - batch Idle 12 8:00:00 kamil - batch -

22 qstat -t u $USER [kamil@jasper ~]$ qstat -t -u kamil jasper-usradm.westgrid.ca: Req'd Req'd Elap Job ID Username Queue Jobname SessID NDS TSK Memory Time S Time [349].jasper-us kamil batch "sumrate" :59:00 Q [350].jasper-us kamil batch "sumrate" :59:00 Q [351].jasper-us kamil batch "sumrate" :59:00 Q [352].jasper-us kamil batch "sumrate" :59:00 R 09:40: [353].jasper-us kamil batch "sumrate" :59:00 R 09:40: [354].jasper-us kamil batch "sumrate" :59:00 R 09:40: [355].jasper-us kamil batch "sumrate" :59:00 R 09:40: [356].jasper-us kamil batch "sumrate" :59:00 R 09:40:12

23 qstat -a hungabee:~ # qstat -a hungabee: Req'd Req'd Elap Job ID Username Queue Jobname SessID NDS TSK Memory Time S Time hungabee fujinaga hall Alliaria.RunAllP :00 R 32: hungabee fujinaga hall Lythrum.RunAllPa :00 Q hungabee tmah hall Nektar_job_3D :00 R 31: hungabee tmah hall Nektar_job_3D :00 Q hungabee tmah hiru cakile.abyssalt :00 R 17: hungabee jyang hiru runscript.hungab :00 R 14: hungabee jyang hiru runscript.hungab :00 R 11: hungabee kamil hiru f_rpx10_c64_f :00 R 06: hungabee kamil hiru f_rpx10_c128_f :00 R 06: hungabee kamil iru f_rpx10_c256_f :00 Q hungabee tmcguire hiru E1e4eta70N :00 R 01:04

24 qstat -r hungabee:~ # qstat -r hungabee: Req'd Req'd Job ID Username Queue Jobname SessID NDS TSK Memory Time S Time hungabee fujinaga hall Alliaria.RunAllP :00 R 32: hungabee tmah hall Nektar_job_3D :00 R 31: hungabee tmah hiru cakile.abyssalt :00 R 17: hungabee jyang hiru runscript.hungab :00 R 14: hungabee jyang hiru runscript.hungab :00 R 11: hungabee kamil hiru f_rpx10_c64_f :00 R 06: hungabee kamil hiru f_rpx10_c128_f :00 R 06: hungabee tmcguire hiru E1e4eta70N :00 R 01:04 Elap

25 showq hungabee:~ # showq active jobs JOBID USERNAME STATE PROCS REMAINING STARTTIME fujinaga Running 64 5:11:32 Thu Apr 10 03:51: kamil Running 64 6:29:27 Thu Apr 10 09:09: tmcguire Running 512 1:15:03:42 Wed Apr 9 01:43:26 4 active jobs 640 of 2048 processors in use by local jobs (31.25%) 80 of 256 nodes active (31.25%) eligible jobs JOBID USERNAME STATE PROCS WCLIMIT QUEUETIME fujinaga Idle :00:00 Thu Apr 10 03:51:27 1 eligible jobs blocked jobs JOBID USERNAME STATE PROCS WCLIMIT QUEUETIME jyang Deferred 1 3:00:00:00 Thu Apr 10 10:35:37 1 blocked jobs Total jobs: 5

26 showq -b etc]# showq b blocked jobs JOBID USERNAME STATE PROCS WCLIMIT QUEUETIME fujinaga BatchHold 1 1:12:00:00 Fri Apr 4 07:15: fujinaga BatchHold 1 1:12:00:00 Fri Apr 4 07:15: fujinaga BatchHold 1 1:12:00:00 Fri Apr 4 07:15: fujinaga BatchHold 1 1:12:00:00 Fri Apr 4 07:15: fujinaga BatchHold 1 1:12:00:00 Fri Apr 4 07:15: fujinaga BatchHold 1 1:12:00:00 Fri Apr 4 07:15: [74] tmcguire Idle 5 3:00:00:00 Sat Apr 5 12:27: jyang Deferred 12 00:01:00 Mon Apr 7 11:52: jyang Deferred 12 00:01:00 Mon Apr 7 11:58: tmah Deferred 4 3:00:00:00 Mon Apr 7 15:07: blocked jobs Total jobs: 3426

27 jobinfo -i or showq -i etc]# showq -i eligible jobs JOBID PRIORITY XFACTOR Q USERNAME GROUP PROCS WCLIMIT CLASS SYSTEMQUEUETIME * fujinaga fujinaga 16 10:30:00 batch Thu Apr 10 10:11: * fujinaga fujinaga 16 10:30:00 batch Thu Apr 10 10:11: [482]* kamil kamil 1 1:00:59:00 batch Thu Apr 10 00:25: [404]* kamil kamil 1 1:00:59:00 batch Thu Apr 10 00:25: [405]* kamil kamil 1 1:00:59:00 batch Thu Apr 10 00:25: * jyang jyang 12 3:00:00:00 batch Wed Apr 9 15:31: * tmcguire tmcguire 8 2:00:00 batch Thu Apr 10 10:27: * jyang jyang 12 3:00:00:00 batch Wed Apr 9 15:31: [539]* tmah tmah 5 3:00:00:00 batch Wed Apr 9 15:36:01 9 eligible jobs Total jobs: 9

28 qstat f [kamil@cl2n234 testwrapper]$ qstat -f 508.cl2n234 Job Id: 508.cl2n234 Job_Name = partest-lq.pbs Job_Owner = kamil@cl2n234 job_state = Q queue = parallel server = cl2n234 Checkpoint = u ctime = Thu Apr 10 13:15: Error_Path = cl2n234:/lustre/home/kamil/test/pbs/jasper/testwrapper/partes t-lq.pbs.e508 Hold_Types = n Join_Path = n Keep_Files = n Mail_Points = abe Mail_Users = kamil@ualberta.ca mtime = Thu Apr 10 13:15: Output_Path = cl2n234:/lustre/home/kamil/test/pbs/jasper/testwrapper/parte st-lq.pbs.o508 Priority = 0 qtime = Thu Apr 10 13:15:

29 qstat f (con/nued) Rerunable = True Resource_List.nodect = 1 Resource_List.nodes = 1:ppn=12 Resource_List.pmem = 256mb Resource_List.walltime = 03:00:00 Shell_Path_List = /bin/sh Variable_List = PBS_O_QUEUE=parallel,PBS_O_HOME=/home/kamil, PBS_O_LOGNAME=kamil, PBS_O_PATH=/usr/kerberos/bin:/usr/local/bin:/bin:/usr/bin:/lustre/jas per/software/jobinfo/jobinfo//bin:/opt/sgi/sgimc/bin:/opt/moab/moab-ve rsion/bin:/opt/moab/moab-version/sbin:/var/spool/torque/torque-version /bin,pbs_o_mail=/var/spool/mail/kamil,pbs_o_shell=/bin/bash, PBS_O_LANG=en_US.UTF-8, PBS_O_WORKDIR=/lustre/home/kamil/test/pbs/jasper/testwrapper, PBS_O_HOST=cl2n234,PBS_O_SERVER=cl2n234 etime = Thu Apr 10 13:15: submit_args = partest-lq.pbs fault_tolerant = False job_radix = 0 submit_host = cl2n234

30 checkjob <jobid> torque-setup]# checkjob 508 job 508 AName: partest-lq.pbs State: Idle Creds: user:kamil group:kamil account:ndz-983-aa class:parallel WallTime: 00:00:00 of 3:00:00 BecameEligible: Thu Apr 10 13:18:14 SubmitTime: Thu Apr 10 13:15:43 (Time Queued Total: 00:20:16 Eligible: 00:20:08) TemplateSets: DEFAULT NodeMatchPolicy: EXACTNODE Total Requested Tasks: 12

31 checkjob <jobid> (con/nued) Req[0] TaskCount: 12 Partition: ALL Memory >= 256M Disk >= 0 Swap >= 0 Dedicated Resources Per Task: PROCS: 1 MEM: 256M SystemID: Moab SystemJID: 508 Notification Events: JobStart,JobEnd,JobFail Notification Address: kamil@ualberta.ca Flags: RESTARTABLE Attr: checkpoint StartPriority: cl2n236 available: 12 tasks supported cl2n235 available: 12 tasks supported NOTE: job can run in partition torque (24 procs available 12 procs required)

32 Checkjob -v -v <jobid> ~]$ checkjob -v -v job (RM job ' jasper-usradm.westgrid.ca ) AName: jrc.egsinp_w193 State: Idle Creds: user:kamil group:kamil account:cas-124-aa class:batch WallTime: 00:00:00 of 1:06:00:00 SubmitTime: Thu Apr 10 13:31:02 (Time Queued Total: 00:08:04 Eligible: 00:00:10) TemplateSets: DEFAULT NodeMatchPolicy: EXACTNODE Total Requested Tasks: 1 Total Requested Nodes: 1 Req[0] TaskCount: 1 Partition: ALL Memory >= 2048M Disk >= 0 Swap >= 0 Available Memory >= 0 Available Swap >= 0 Dedicated Resources Per Task: PROCS: 1 MEM: 2048M SWAP: 2048M NodeSet=ONEOF:FEATURE:X5675-QDR:X5675-DDR:L5420-DDR NodeCount: 1 SystemID: Moab SystemJID: Notification Events: JobFail

33 Checkjob -v -v <jobid> (con/nued) UMask: 0000 OutputFile: jasper.westgrid.ca:/home/kamil/egsnrc/codes/dosxyznrc_nob/ ProfilePhantom02IC10_10x10_Emean5_9MeVMonoAltPrimColModFFMT50Ang08KMNRC.egsinp_w193.eo ErrorFile: jasper.westgrid.ca:/home/kamil/egsnrc/codes/dosxyznrc_nob/ ProfilePhantom02IC10_10x10_Emean5_9MeVMonoAltPrimColModFFMT50Ang08KMNRC.egsinp_w193.eo EnvVariables: PBS_O_QUEUE=batch,PBS_O_HOME=/home/ kamil,pbs_o_logname=kamil,pbs_o_path=/home/kamil/egsnrc/codes/bin/x86_64-unknown-linuxgnu-f95:/home/kamil/egsnrc/bin/x86_64-unknown-linux-gnu-f95:/usr/kerberos/bin:/usr/ local/bin:/bin:/usr/bin:/usr/bin:/lustre/jasper/software/jobinfo/jobinfo//bin:/opt/sgi/ sgimc/bin:/opt/moab/moab-version/bin:/opt/moab/moab-version/sbin:/var/spool/torque/ torque-version/bin,pbs_o_mail=/var/spool/mail/kamil,pbs_o_shell=/bin/ bash,pbs_o_lang=en_us.utf-8,pbs_o_workdir=/lustre/home/kamil/egsnrc/codes/ dosxyznrc_nob,pbs_o_host=jasper.westgrid.ca,pbs_o_server=jasper-usradm.westgrid.ca Partition List: [ALL] SrcRM: jasper-usradm DstRM: jasper-usradm DstRMJID: jasperusradm.westgrid.ca Submit Args: -j eo -l pmem=2gb,vmem=2gb -e ProfilePhantom02IC10_10x10_Emean5_9MeVMonoAltPrimColModFFMT50Ang08KMNRC.egsinp_w193.eo - N jrc.egsinp_w193 -l walltime=30:00:00 Flags: RESTARTABLE Attr: checkpoint StartPriority: Priority Analysis: Job PRIORITY* Cred(Class) FS(Accnt) Serv(QTime) Weights ( 1) 1000( 1) 1( 1) ( 0.0) 100.0(-11.1) 0.0( 0.0) PE: 1.00 Node Availability for Partition jasper-usradm

34 Checkjob -v -v <jobid> (con/nued) Node Availability for Partition jasper-usradm cl1n001 cl1n002 cl2n002 cl2n003 cl2n028 cl2n029 cl2n030 cl2n031 rejected: Reserved (wlcg_ops ) allocationpriority=0.00 rejected: Reserved (wlcg_ops ) allocationpriority=0.00 rejected: Memory allocationpriority=0.00 rejected: Memory allocationpriority=0.00 rejected: State (Busy) allocationpriority=0.00 rejected: State (Busy) allocationpriority=0.00 rejected: Memory allocationpriority=0.00 rejected: Memory allocationpriority=0.00 NOTE: job req cannot run in partition jasper-usradm (available procs do not meet requirements : 0 of 1 procs found) idle procs: 354 feasible procs: 0 Node Rejection Summary: [Memory: 128][State: 284][Reserved: 4] BLOCK MSG: job violates idle HARD MAXIJOB limit of 5 for user kamil partition ALL (Req: 1 InUse: 5) (recorded at last scheduling iteration)

35 Demonstra/on on cluster SSH cluster and show all the following commands and how to interpret them jobinfo -j qstat -t -u $USER qstat -a qstat -r showq showq -i showq -b qstat -f <jobid> Checkjob <jobid> Checkjob -v -v <jobid>

36 Job informa/on prac/ce BREAK FOR PRACTICE

37 QUESTIONS?

38 The End

SGI Altix Running Batch Jobs With PBSPro Reiner Vogelsang SGI GmbH

SGI Altix Running Batch Jobs With PBSPro Reiner Vogelsang SGI GmbH SGI Altix Running Batch Jobs With PBSPro Reiner Vogelsang SGI GmbH reiner@sgi.com Module Objectives After completion of this module you should be able to Submit batch jobs Create job chains Monitor your

More information

Introduction to HPC Using the New Cluster at GACRC

Introduction to HPC Using the New Cluster at GACRC Introduction to HPC Using the New Cluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What is the new cluster

More information

Introduction to HPC Using the New Cluster at GACRC

Introduction to HPC Using the New Cluster at GACRC Introduction to HPC Using the New Cluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What is the new cluster

More information

Introduction to HPC Using the New Cluster at GACRC

Introduction to HPC Using the New Cluster at GACRC Introduction to HPC Using the New Cluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu Outline What is GACRC? What is the new cluster

More information

PBS Pro Documentation

PBS Pro Documentation Introduction Most jobs will require greater resources than are available on individual nodes. All jobs must be scheduled via the batch job system. The batch job system in use is PBS Pro. Jobs are submitted

More information

pypbs Documentation Release dev Tyghe Vallard, Michael panciera

pypbs Documentation Release dev Tyghe Vallard, Michael panciera pypbs Documentation Release 0.2.0-dev Tyghe Vallard, Michael panciera May 12, 2015 Contents 1 Additional Commands 3 1.1 pbsstatus................................................. 3 1.2 qpeek...................................................

More information

Introduction to HPC Using the New Cluster at GACRC

Introduction to HPC Using the New Cluster at GACRC Introduction to HPC Using the New Cluster at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu 1 Outline What is GACRC? What is the new cluster

More information

Cluster Tools Batch Queue Job Control W. Trevor King May 20, 2008

Cluster Tools Batch Queue Job Control W. Trevor King May 20, 2008 Cluster Tools Batch Queue Job Control W. Trevor King May 20, 2008 1 Submitting jobs You can submit jobs to the batch queue for later proccessing with qsub. Batch queueing can get pretty fancy, so qsub

More information

Introduction to HPC Using Sapelo Cluster at GACRC

Introduction to HPC Using Sapelo Cluster at GACRC Introduction to HPC Using Sapelo Cluster at GACRC Level 1/2 Georgia Advanced Computing Resource Center (GACRC) EITS/University of Georgia Zhuofei Hou zhuofei@uga.edu 1 Outline GACRC Sapelo Cluster Overview

More information

User Guide of High Performance Computing Cluster in School of Physics

User Guide of High Performance Computing Cluster in School of Physics User Guide of High Performance Computing Cluster in School of Physics Prepared by Sue Yang (xue.yang@sydney.edu.au) This document aims at helping users to quickly log into the cluster, set up the software

More information

Advanced Scripting Using PBS Environment Variables

Advanced Scripting Using PBS Environment Variables Advanced Scripting Using PBS Environment Variables Your job submission script has a number of environment variables that can be used to help you write some more advanced scripts. These variables can make

More information

and how to use TORQUE & Maui Piero Calucci

and how to use TORQUE & Maui Piero Calucci Queue and how to use & Maui Scuola Internazionale Superiore di Studi Avanzati Trieste November 2008 Advanced School in High Performance and Grid Computing Outline 1 We Are Trying to Solve 2 Using the Manager

More information

Queue systems. and how to use Torque/Maui. Piero Calucci. Scuola Internazionale Superiore di Studi Avanzati Trieste

Queue systems. and how to use Torque/Maui. Piero Calucci. Scuola Internazionale Superiore di Studi Avanzati Trieste Queue systems and how to use Torque/Maui Piero Calucci Scuola Internazionale Superiore di Studi Avanzati Trieste March 9th 2007 Advanced School in High Performance Computing Tools for e-science Outline

More information

Running Jobs, Submission Scripts, Modules

Running Jobs, Submission Scripts, Modules 9/17/15 Running Jobs, Submission Scripts, Modules 16,384 cores total of about 21,000 cores today Infiniband interconnect >3PB fast, high-availability, storage GPGPUs Large memory nodes (512GB to 1TB of

More information

Intel Manycore Testing Lab (MTL) - Linux Getting Started Guide

Intel Manycore Testing Lab (MTL) - Linux Getting Started Guide Intel Manycore Testing Lab (MTL) - Linux Getting Started Guide Introduction What are the intended uses of the MTL? The MTL is prioritized for supporting the Intel Academic Community for the testing, validation

More information

Computing on Mio Data & Useful Commands

Computing on Mio Data & Useful Commands Computing on Mio Data & Useful Commands Timothy H. Kaiser, Ph.D. tkaiser@mines.edu Director - CSM High Performance Computing Director - Golden Energy Computing Organization http://inside.mines.edu/mio/tutorial/

More information

Introduction to Discovery.

Introduction to Discovery. Introduction to Discovery http://discovery.dartmouth.edu The Discovery Cluster 2 Agenda What is a cluster and why use it Overview of computer hardware in cluster Help Available to Discovery Users Logging

More information

PACE Orientation OIT-ART

PACE Orientation OIT-ART PACE Orientation OIT-ART Mehmet (Memo) Belgin, PhD Research Scientist, OIT-ART www.pace.gatech.edu What is PACE A Partnership for an Advanced Computing Environment Provides faculty and researchers vital

More information

Introduction to Discovery.

Introduction to Discovery. Introduction to Discovery http://discovery.dartmouth.edu March 2014 The Discovery Cluster 2 Agenda Resource overview Logging on to the cluster with ssh Transferring files to and from the cluster The Environment

More information

Introduction to Discovery.

Introduction to Discovery. Introduction to Discovery http://discovery.dartmouth.edu The Discovery Cluster 2 Agenda What is a cluster and why use it Overview of computer hardware in cluster Help Available to Discovery Users Logging

More information

GACRC User Training: Migrating from Zcluster to Sapelo

GACRC User Training: Migrating from Zcluster to Sapelo GACRC User Training: Migrating from Zcluster to Sapelo The GACRC Staff Version 1.0 8/28/2017 GACRC Zcluster-Sapelo Migrating Training 1 Discussion Points I. Request Sapelo User Account II. III. IV. Systems

More information

Introduction to HPC Using the New Cluster (Sapelo) at GACRC

Introduction to HPC Using the New Cluster (Sapelo) at GACRC Introduction to HPC Using the New Cluster (Sapelo) at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu 1 Outline What is GACRC? What is the

More information

Martinos Center Compute Cluster

Martinos Center Compute Cluster Why-N-How: Intro to Launchpad 8 September 2016 Lee Tirrell Laboratory for Computational Neuroimaging Adapted from slides by Jon Kaiser 1. Intro 2. Using launchpad 3. Summary 4. Appendix: Miscellaneous

More information

OpenPBS Users Manual

OpenPBS Users Manual How to Write a PBS Batch Script OpenPBS Users Manual PBS scripts are rather simple. An MPI example for user your-user-name: Example: MPI Code PBS -N a_name_for_my_parallel_job PBS -l nodes=7,walltime=1:00:00

More information

Introduction to HPC Using the New Cluster (Sapelo) at GACRC

Introduction to HPC Using the New Cluster (Sapelo) at GACRC Introduction to HPC Using the New Cluster (Sapelo) at GACRC Georgia Advanced Computing Resource Center University of Georgia Zhuofei Hou, HPC Trainer zhuofei@uga.edu 1 Outline What is GACRC? What is the

More information

Introduction to the HPC Resources at the University of Pittsburgh

Introduction to the HPC Resources at the University of Pittsburgh Introduction to the HPC Resources at the University of Pittsburgh Kim F. Wong and Albert DeFusco Center for Simulation and Modeling July 18, 2013 Agenda This hands-on introductory workshop will cover the

More information

Introduction to HPC Using Sapelo Cluster at GACRC

Introduction to HPC Using Sapelo Cluster at GACRC Introduction to HPC Using Sapelo Cluster at GACRC Level 1/2 Georgia Advanced Computing Resource Center (GACRC) EITS/University of Georgia Zhuofei Hou zhuofei@uga.edu 1 Outline GACRC What is Sapelo cluster?

More information

Viglen NPACI Rocks. Getting Started and FAQ

Viglen NPACI Rocks. Getting Started and FAQ Viglen NPACI Rocks Getting Started and FAQ Table of Contents Viglen NPACI Rocks...1 Getting Started...3 Powering up the machines:...3 Checking node status...4 Through web interface:...4 Adding users:...7

More information

A Brief Introduction to The Center for Advanced Computing

A Brief Introduction to The Center for Advanced Computing A Brief Introduction to The Center for Advanced Computing May 1, 2006 Hardware 324 Opteron nodes, over 700 cores 105 Athlon nodes, 210 cores 64 Apple nodes, 128 cores Gigabit networking, Myrinet networking,

More information

PACE. Instructional Cluster Environment (ICE) Orientation. Research Scientist, PACE

PACE. Instructional Cluster Environment (ICE) Orientation. Research Scientist, PACE PACE Instructional Cluster Environment (ICE) Orientation Mehmet (Memo) Belgin, PhD Research Scientist, PACE www.pace.gatech.edu What is PACE A Partnership for an Advanced Computing Environment Provides

More information

Quick Guide for the Torque Cluster Manager

Quick Guide for the Torque Cluster Manager Quick Guide for the Torque Cluster Manager Introduction: One of the main purposes of the Aries Cluster is to accommodate especially long-running programs. Users who run long jobs (which take hours or days

More information

PACE. Instructional Cluster Environment (ICE) Orientation. Mehmet (Memo) Belgin, PhD Research Scientist, PACE

PACE. Instructional Cluster Environment (ICE) Orientation. Mehmet (Memo) Belgin, PhD  Research Scientist, PACE PACE Instructional Cluster Environment (ICE) Orientation Mehmet (Memo) Belgin, PhD www.pace.gatech.edu Research Scientist, PACE What is PACE A Partnership for an Advanced Computing Environment Provides

More information

Please include the following sentence in any works using center resources.

Please include the following sentence in any works using center resources. The TCU High-Performance Computing Center The TCU HPCC currently maintains a cluster environment hpcl1.chm.tcu.edu. Work on a second cluster environment is underway. This document details using hpcl1.

More information

A Brief Introduction to The Center for Advanced Computing

A Brief Introduction to The Center for Advanced Computing A Brief Introduction to The Center for Advanced Computing February 8, 2007 Hardware 376 Opteron nodes, over 890 cores Gigabit networking, Myrinet networking, Infiniband networking soon Hardware: nyx nyx

More information

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved. Minnesota Supercomputing Institute Introduction to Job Submission and Scheduling Andrew Gustafson Interacting with MSI Systems Connecting to MSI SSH is the most reliable connection method Linux and Mac

More information

High Performance Beowulf Cluster Environment User Manual

High Performance Beowulf Cluster Environment User Manual High Performance Beowulf Cluster Environment User Manual Version 3.1c 2 This guide is intended for cluster users who want a quick introduction to the Compusys Beowulf Cluster Environment. It explains how

More information

Introduction to HPC Using Sapelo Cluster at GACRC

Introduction to HPC Using Sapelo Cluster at GACRC Introduction to HPC Using Sapelo Cluster at GACRC Georgia Advanced Computing Resource Center EITS/University of Georgia Zhuofei Hou, zhuofei@uga.edu 1 Outline GACRC Sapelo Cluster Job Submission Workflow

More information

Using the IBM Opteron 1350 at OSC Batch Processing. October 19-20, 2010

Using the IBM Opteron 1350 at OSC Batch Processing. October 19-20, 2010 Using the IBM Opteron 1350 at OSC Batch Processing October 19-20, 2010 Online Information Technical information: http://www.osc.edu/supercomputing/ Hardware Software Environment Training Notices http://www.osc.edu/supercomputing/computing/#batch

More information

Introduction to GALILEO

Introduction to GALILEO Introduction to GALILEO Parallel & production environment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Domenico Guida d.guida@cineca.it Maurizio Cremonesi m.cremonesi@cineca.it

More information

Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop

Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop Before We Start Sign in hpcxx account slips Windows Users: Download PuTTY Google PuTTY First result Save putty.exe to Desktop Research Computing at Virginia Tech Advanced Research Computing Compute Resources

More information

UF Research Computing: Overview and Running STATA

UF Research Computing: Overview and Running STATA UF : Overview and Running STATA www.rc.ufl.edu Mission Improve opportunities for research and scholarship Improve competitiveness in securing external funding Matt Gitzendanner magitz@ufl.edu Provide high-performance

More information

OBTAINING AN ACCOUNT:

OBTAINING AN ACCOUNT: HPC Usage Policies The IIA High Performance Computing (HPC) System is managed by the Computer Management Committee. The User Policies here were developed by the Committee. The user policies below aim to

More information

UBDA Platform User Gudie. 16 July P a g e 1

UBDA Platform User Gudie. 16 July P a g e 1 16 July 2018 P a g e 1 Revision History Version Date Prepared By Summary of Changes 1.0 Jul 16, 2018 Initial release P a g e 2 Table of Contents 1. Introduction... 4 2. Perform the test... 5 3 Job submission...

More information

Using Sapelo2 Cluster at the GACRC

Using Sapelo2 Cluster at the GACRC Using Sapelo2 Cluster at the GACRC New User Training Workshop Georgia Advanced Computing Resource Center (GACRC) EITS/University of Georgia Zhuofei Hou zhuofei@uga.edu 1 Outline GACRC Sapelo2 Cluster Diagram

More information

A Brief Introduction to The Center for Advanced Computing

A Brief Introduction to The Center for Advanced Computing A Brief Introduction to The Center for Advanced Computing November 10, 2009 Outline 1 Resources Hardware Software 2 Mechanics: Access Transferring files and data to and from the clusters Logging into the

More information

Introduction to Molecular Dynamics on ARCHER: Instructions for running parallel jobs on ARCHER

Introduction to Molecular Dynamics on ARCHER: Instructions for running parallel jobs on ARCHER Introduction to Molecular Dynamics on ARCHER: Instructions for running parallel jobs on ARCHER 1 Introduction This handout contains basic instructions for how to login in to ARCHER and submit jobs to the

More information

Effective Use of CCV Resources

Effective Use of CCV Resources Effective Use of CCV Resources Mark Howison User Services & Support This talk... Assumes you have some familiarity with a Unix shell Provides examples and best practices for typical usage of CCV systems

More information

NBIC TechTrack PBS Tutorial

NBIC TechTrack PBS Tutorial NBIC TechTrack PBS Tutorial by Marcel Kempenaar, NBIC Bioinformatics Research Support group, University Medical Center Groningen Visit our webpage at: http://www.nbic.nl/support/brs 1 NBIC PBS Tutorial

More information

Introduc)on to Pacman

Introduc)on to Pacman Introduc)on to Pacman Don Bahls User Consultant dmbahls@alaska.edu (Significant Slide Content from Tom Logan) Overview Connec)ng to Pacman Hardware Programming Environment Compilers Queuing System Interac)ve

More information

High Performance Computing (HPC) Club Training Session. Xinsheng (Shawn) Qin

High Performance Computing (HPC) Club Training Session. Xinsheng (Shawn) Qin High Performance Computing (HPC) Club Training Session Xinsheng (Shawn) Qin Outline HPC Club The Hyak Supercomputer Logging in to Hyak Basic Linux Commands Transferring Files Between Your PC and Hyak Submitting

More information

Introduction to HPC Using Sapelo Cluster at GACRC

Introduction to HPC Using Sapelo Cluster at GACRC Introduction to HPC Using Sapelo Cluster at GACRC Georgia Advanced Computing Resource Center (GACRC) EITS/University of Georgia Zhuofei Hou zhuofei@uga.edu 1 Outline GACRC Sapelo Cluster Job Submission

More information

High Performance Compu2ng Using Sapelo Cluster

High Performance Compu2ng Using Sapelo Cluster High Performance Compu2ng Using Sapelo Cluster Georgia Advanced Compu2ng Resource Center EITS/UGA Zhuofei Hou, Training Advisor zhuofei@uga.edu 1 Outline GACRC What is High Performance Compu2ng (HPC) Sapelo

More information

Batch Systems. Running calculations on HPC resources

Batch Systems. Running calculations on HPC resources Batch Systems Running calculations on HPC resources Outline What is a batch system? How do I interact with the batch system Job submission scripts Interactive jobs Common batch systems Converting between

More information

Introduction to PICO Parallel & Production Enviroment

Introduction to PICO Parallel & Production Enviroment Introduction to PICO Parallel & Production Enviroment Mirko Cestari m.cestari@cineca.it Alessandro Marani a.marani@cineca.it Domenico Guida d.guida@cineca.it Nicola Spallanzani n.spallanzani@cineca.it

More information

The cluster system. Introduction 22th February Jan Saalbach Scientific Computing Group

The cluster system. Introduction 22th February Jan Saalbach Scientific Computing Group The cluster system Introduction 22th February 2018 Jan Saalbach Scientific Computing Group cluster-help@luis.uni-hannover.de Contents 1 General information about the compute cluster 2 Available computing

More information

Purdue University - ITaP. March 6-10, 2017

Purdue University - ITaP. March 6-10, 2017 Clusters 101 Purdue University - ITaP Gladys Andino Dan Dietz Jieyu Gao Lev Gorenstein Erik Gough Stephen Harrell Randy Herban Steve Kelley Boyu Zhang Xiao Zhu rcac-help@purdue.edu March 6-10, 2017 Slides

More information

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing Quick Start Guide by Burak Himmetoglu Supercomputing Consultant Enterprise Technology Services & Center for Scientific Computing E-mail: bhimmetoglu@ucsb.edu Contents User access, logging in Linux/Unix

More information

UoW HPC Quick Start. Information Technology Services University of Wollongong. ( Last updated on October 10, 2011)

UoW HPC Quick Start. Information Technology Services University of Wollongong. ( Last updated on October 10, 2011) UoW HPC Quick Start Information Technology Services University of Wollongong ( Last updated on October 10, 2011) 1 Contents 1 Logging into the HPC Cluster 3 1.1 From within the UoW campus.......................

More information

CENTER FOR HIGH PERFORMANCE COMPUTING. Overview of CHPC. Martin Čuma, PhD. Center for High Performance Computing

CENTER FOR HIGH PERFORMANCE COMPUTING. Overview of CHPC. Martin Čuma, PhD. Center for High Performance Computing Overview of CHPC Martin Čuma, PhD Center for High Performance Computing m.cuma@utah.edu Spring 2014 Overview CHPC Services HPC Clusters Specialized computing resources Access and Security Batch (PBS and

More information

Batch Systems & Parallel Application Launchers Running your jobs on an HPC machine

Batch Systems & Parallel Application Launchers Running your jobs on an HPC machine Batch Systems & Parallel Application Launchers Running your jobs on an HPC machine Partners Funding Reusing this material This work is licensed under a Creative Commons Attribution- NonCommercial-ShareAlike

More information

Batch Systems. Running your jobs on an HPC machine

Batch Systems. Running your jobs on an HPC machine Batch Systems Running your jobs on an HPC machine Reusing this material This work is licensed under a Creative Commons Attribution- NonCommercial-ShareAlike 4.0 International License. http://creativecommons.org/licenses/by-nc-sa/4.0/deed.en_us

More information

Introduction to HPC Using Sapelo at GACRC

Introduction to HPC Using Sapelo at GACRC Introduction to HPC Using Sapelo at GACRC Georgia Advanced Computing Resource Center University of Georgia Suchitra Pakala pakala@uga.edu Slides courtesy: Zhuofei Hou 1 Outline GACRC Sapelo Cluster at

More information

Introduc)on to Hyades

Introduc)on to Hyades Introduc)on to Hyades Shawfeng Dong Department of Astronomy & Astrophysics, UCSSC Hyades 1 Hardware Architecture 2 Accessing Hyades 3 Compu)ng Environment 4 Compiling Codes 5 Running Jobs 6 Visualiza)on

More information

Image Sharpening. Practical Introduction to HPC Exercise. Instructions for Cirrus Tier-2 System

Image Sharpening. Practical Introduction to HPC Exercise. Instructions for Cirrus Tier-2 System Image Sharpening Practical Introduction to HPC Exercise Instructions for Cirrus Tier-2 System 2 1. Aims The aim of this exercise is to get you used to logging into an HPC resource, using the command line

More information

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing Quick Start Guide by Burak Himmetoglu Supercomputing Consultant Enterprise Technology Services & Center for Scientific Computing E-mail: bhimmetoglu@ucsb.edu Linux/Unix basic commands Basic command structure:

More information

Workshop Set up. Workshop website: Workshop project set up account at my.osc.edu PZS0724 Nq7sRoNrWnFuLtBm

Workshop Set up. Workshop website:   Workshop project set up account at my.osc.edu PZS0724 Nq7sRoNrWnFuLtBm Workshop Set up Workshop website: https://khill42.github.io/osc_introhpc/ Workshop project set up account at my.osc.edu PZS0724 Nq7sRoNrWnFuLtBm If you already have an OSC account, sign in to my.osc.edu

More information

HPC Resources at Lehigh. Steve Anthony March 22, 2012

HPC Resources at Lehigh. Steve Anthony March 22, 2012 HPC Resources at Lehigh Steve Anthony March 22, 2012 HPC at Lehigh: Resources What's Available? Service Level Basic Service Level E-1 Service Level E-2 Leaf and Condor Pool Altair Trits, Cuda0, Inferno,

More information

bwunicluster Tutorial Access, Data Transfer, Compiling, Modulefiles, Batch Jobs

bwunicluster Tutorial Access, Data Transfer, Compiling, Modulefiles, Batch Jobs bwunicluster Tutorial Access, Data Transfer, Compiling, Modulefiles, Batch Jobs Frauke Bösert, SCC, KIT 1 Material: Slides & Scripts https://indico.scc.kit.edu/indico/event/263/ @bwunicluster/forhlr I/ForHLR

More information

Logging in to the CRAY

Logging in to the CRAY Logging in to the CRAY 1. Open Terminal Cray Hostname: cray2.colostate.edu Cray IP address: 129.82.103.183 On a Mac 2. type ssh username@cray2.colostate.edu where username is your account name 3. enter

More information

The JANUS Computing Environment

The JANUS Computing Environment Research Computing UNIVERSITY OF COLORADO The JANUS Computing Environment Monte Lunacek monte.lunacek@colorado.edu rc-help@colorado.edu What is JANUS? November, 2011 1,368 Compute nodes 16,416 processors

More information

Migrating from Zcluster to Sapelo

Migrating from Zcluster to Sapelo GACRC User Quick Guide: Migrating from Zcluster to Sapelo The GACRC Staff Version 1.0 8/4/17 1 Discussion Points I. Request Sapelo User Account II. III. IV. Systems Transfer Files Configure Software Environment

More information

Introduction to HPCC at MSU

Introduction to HPCC at MSU Introduction to HPCC at MSU Chun-Min Chang Research Consultant Institute for Cyber-Enabled Research Download this presentation: https://wiki.hpcc.msu.edu/display/teac/2016-03-17+introduction+to+hpcc How

More information

Sharpen Exercise: Using HPC resources and running parallel applications

Sharpen Exercise: Using HPC resources and running parallel applications Sharpen Exercise: Using HPC resources and running parallel applications Andrew Turner, Dominic Sloan-Murphy, David Henty, Adrian Jackson Contents 1 Aims 2 2 Introduction 2 3 Instructions 3 3.1 Log into

More information

Shell Scripting. With Applications to HPC. Edmund Sumbar Copyright 2007 University of Alberta. All rights reserved

Shell Scripting. With Applications to HPC. Edmund Sumbar Copyright 2007 University of Alberta. All rights reserved AICT High Performance Computing Workshop With Applications to HPC Edmund Sumbar research.support@ualberta.ca Copyright 2007 University of Alberta. All rights reserved High performance computing environment

More information

Running Jobs on Blue Waters. Greg Bauer

Running Jobs on Blue Waters. Greg Bauer Running Jobs on Blue Waters Greg Bauer Policies and Practices Placement Checkpointing Monitoring a job Getting a nodelist Viewing the torus 2 Resource and Job Scheduling Policies Runtime limits expected

More information

bwunicluster Tutorial Access, Data Transfer, Compiling, Modulefiles, Batch Jobs

bwunicluster Tutorial Access, Data Transfer, Compiling, Modulefiles, Batch Jobs bwunicluster Tutorial Access, Data Transfer, Compiling, Modulefiles, Batch Jobs Frauke Bösert, SCC, KIT 1 Material: Slides & Scripts https://indico.scc.kit.edu/indico/event/263/ @bwunicluster/forhlr I/ForHLR

More information

Parameter searches and the batch system

Parameter searches and the batch system Parameter searches and the batch system Scientific Computing Group css@rrzn.uni-hannover.de Parameter searches and the batch system Scientific Computing Group 1st of October 2012 1 Contents 1 Parameter

More information

Moab Workload Manager on Cray XT3

Moab Workload Manager on Cray XT3 Moab Workload Manager on Cray XT3 presented by Don Maxwell (ORNL) Michael Jackson (Cluster Resources, Inc.) MOAB Workload Manager on Cray XT3 Why MOAB? Requirements Features Support/Futures 2 Why Moab?

More information

Sharpen Exercise: Using HPC resources and running parallel applications

Sharpen Exercise: Using HPC resources and running parallel applications Sharpen Exercise: Using HPC resources and running parallel applications Contents 1 Aims 2 2 Introduction 2 3 Instructions 3 3.1 Log into ARCHER frontend nodes and run commands.... 3 3.2 Download and extract

More information

PARALLEL COMPUTING IN R USING WESTGRID CLUSTERS STATGEN GROUP MEETING 10/30/2017

PARALLEL COMPUTING IN R USING WESTGRID CLUSTERS STATGEN GROUP MEETING 10/30/2017 PARALLEL COMPUTING IN R USING WESTGRID CLUSTERS STATGEN GROUP MEETING 10/30/2017 PARALLEL COMPUTING Dataset 1 Processor Dataset 2 Dataset 3 Dataset 4 R script Processor Processor Processor WHAT IS ADVANCED

More information

Getting started with the CEES Grid

Getting started with the CEES Grid Getting started with the CEES Grid October, 2013 CEES HPC Manager: Dennis Michael, dennis@stanford.edu, 723-2014, Mitchell Building room 415. Please see our web site at http://cees.stanford.edu. Account

More information

PACE Orientation. Research Scientist, PACE

PACE Orientation. Research Scientist, PACE PACE Orientation Mehmet (Memo) Belgin, PhD Research Scientist, PACE www.pace.gatech.edu What is PACE A Partnership for an Advanced Computing Environment Provides faculty and researchers vital tools to

More information

TORQUE Resource Manager5.0.2 release notes

TORQUE Resource Manager5.0.2 release notes TORQUE Resource Manager release notes The release notes file contains the following sections: New Features on page 1 Differences on page 2 Known Issues on page 4 Resolved issues on page 4 New Features

More information

Introduction to HPC Resources and Linux

Introduction to HPC Resources and Linux Introduction to HPC Resources and Linux Burak Himmetoglu Enterprise Technology Services & Center for Scientific Computing e-mail: bhimmetoglu@ucsb.edu Paul Weakliem California Nanosystems Institute & Center

More information

Shark Cluster Overview

Shark Cluster Overview Shark Cluster Overview 51 Execution Nodes 1 Head Node (shark) 1 Graphical login node (rivershark) 800 Cores = slots 714 TB Storage RAW Slide 1/14 Introduction What is a cluster? A cluster is a group of

More information

Introduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende

Introduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende Introduction to the NCAR HPC Systems 25 May 2018 Consulting Services Group Brian Vanderwende Topics to cover Overview of the NCAR cluster resources Basic tasks in the HPC environment Accessing pre-built

More information

Introduction to High-Performance Computing (HPC) at USC

Introduction to High-Performance Computing (HPC) at USC 6/2/17 Introduction to High-Performance Computing (HPC) at USC Erin Shaw and Cesar Sul And Avalon Johnson (Slides) Advanced Cyberinfrastructure Research and Education Facilitation USC Center for High-Performance

More information

Shark Cluster Overview

Shark Cluster Overview Shark Cluster Overview 51 Execution Nodes 1 Head Node (shark) 2 Graphical login nodes 800 Cores = slots 714 TB Storage RAW Slide 1/17 Introduction What is a High Performance Compute (HPC) cluster? A HPC

More information

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved. Minnesota Supercomputing Institute Introduction to MSI Systems Andrew Gustafson The Machines at MSI Machine Type: Cluster Source: http://en.wikipedia.org/wiki/cluster_%28computing%29 Machine Type: Cluster

More information

Compiling applications for the Cray XC

Compiling applications for the Cray XC Compiling applications for the Cray XC Compiler Driver Wrappers (1) All applications that will run in parallel on the Cray XC should be compiled with the standard language wrappers. The compiler drivers

More information

Introduction to High-Performance Computing (HPC)

Introduction to High-Performance Computing (HPC) Introduction to High-Performance Computing (HPC) Computer components CPU : Central Processing Unit cores : individual processing units within a CPU Storage : Disk drives HDD : Hard Disk Drive SSD : Solid

More information

Submit a Job. Want to run a batch script: #!/bin/sh echo Starting job date /usr/bin/time./hello date echo Ending job. qsub A HPC job.

Submit a Job. Want to run a batch script: #!/bin/sh echo Starting job date /usr/bin/time./hello date echo Ending job. qsub A HPC job. Submit a Job Want to run a batch script: #!/bin/sh echo Starting job date /usr/bin/time./hello date echo Ending job Have to ask scheduler to do it. qsub A 20090528HPC job.sge #!/bin/sh #$ -N ht3d-hyb #$

More information

SLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education

SLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education SLURM: Resource Management and Job Scheduling Software Advanced Computing Center for Research and Education www.accre.vanderbilt.edu Simple Linux Utility for Resource Management But it s also a job scheduler!

More information

NBIC TechTrack PBS Tutorial. by Marcel Kempenaar, NBIC Bioinformatics Research Support group, University Medical Center Groningen

NBIC TechTrack PBS Tutorial. by Marcel Kempenaar, NBIC Bioinformatics Research Support group, University Medical Center Groningen NBIC TechTrack PBS Tutorial by Marcel Kempenaar, NBIC Bioinformatics Research Support group, University Medical Center Groningen 1 NBIC PBS Tutorial This part is an introduction to clusters and the PBS

More information

New User Tutorial. OSU High Performance Computing Center

New User Tutorial. OSU High Performance Computing Center New User Tutorial OSU High Performance Computing Center TABLE OF CONTENTS Logging In... 3-5 Windows... 3-4 Linux... 4 Mac... 4-5 Changing Password... 5 Using Linux Commands... 6 File Systems... 7 File

More information

Workflow Optimization for Large Scale Bioinformatics

Workflow Optimization for Large Scale Bioinformatics Workflow Optimization for Large Scale Bioinformatics Ramses van Zon SciNet/Compute Canada September 25, 2012 Ramses van Zon (SciNet) Optimizing Large Scale Bioinformatics September 25, 2012 1 / 87 Outline

More information

PBS Pro and Ansys Examples

PBS Pro and Ansys Examples PBS Pro and Ansys Examples Introduction This document contains a number of different types of examples of using Ansys on the HPC, listed below. 1. Single-node Ansys Job 2. Single-node CFX Job 3. Single-node

More information

Grid Examples. Steve Gallo Center for Computational Research University at Buffalo

Grid Examples. Steve Gallo Center for Computational Research University at Buffalo Grid Examples Steve Gallo Center for Computational Research University at Buffalo Examples COBALT (Computational Fluid Dynamics) Ercan Dumlupinar, Syracyse University Aerodynamic loads on helicopter rotors

More information

Installing and running COMSOL 4.3a on a Linux cluster COMSOL. All rights reserved.

Installing and running COMSOL 4.3a on a Linux cluster COMSOL. All rights reserved. Installing and running COMSOL 4.3a on a Linux cluster 2012 COMSOL. All rights reserved. Introduction This quick guide explains how to install and operate COMSOL Multiphysics 4.3a on a Linux cluster. It

More information

XSEDE New User Tutorial

XSEDE New User Tutorial April 2, 2014 XSEDE New User Tutorial Jay Alameda National Center for Supercomputing Applications XSEDE Training Survey Make sure you sign the sign in sheet! At the end of the module, I will ask you to

More information