X Grid Engine. Where X stands for Oracle Univa Open Son of more to come...?!?

Similar documents
Batch Systems. Running calculations on HPC resources

Grid Engine - A Batch System for DESY. Andreas Haupt, Peter Wegner DESY Zeuthen

Batch system usage arm euthen F azo he Z J. B T

Shark Cluster Overview

Why You Should Consider Grid Computing

SGE Roll: Users Guide. Version Edition

Shark Cluster Overview

Resource Management Systems

National Biochemical Computational Research Familiarize yourself with the account policy

A Hands-On Tutorial: RNA Sequencing Using High-Performance Computing

Joint High Performance Computing Exchange (JHPCE) Cluster Orientation.

Quick Guide for the Torque Cluster Manager

June 26, Explanatory meeting for users of supercomputer system -- Overview of UGE --

Cluster User Training

Grid Engine Users Guide. 5.5 Edition

Introduction to HPC Using zcluster at GACRC

SGE Roll: Users Guide. Version 5.3 Edition

OBTAINING AN ACCOUNT:

User interface for a computational cluster: resource description approach

Introduction to PICO Parallel & Production Enviroment

Sun Grid Engine - A Batch System for DESY

Introduction to GALILEO

Submit a Job. Want to run a batch script: #!/bin/sh echo Starting job date /usr/bin/time./hello date echo Ending job. qsub A HPC job.

Getting started with the CEES Grid

ACEnet for CS6702 Ross Dickson, Computational Research Consultant 29 Sep 2009

Answers to Federal Reserve Questions. Administrator Training for University of Richmond

Center for Mathematical Modeling University of Chile HPC 101. HPC systems basics and concepts. Juan Carlos Maureira B.

The cluster system. Introduction 22th February Jan Saalbach Scientific Computing Group

Answers to Federal Reserve Questions. Training for University of Richmond

Grid Engine Users Guide. 7.0 Edition

NBIC TechTrack PBS Tutorial. by Marcel Kempenaar, NBIC Bioinformatics Research Support group, University Medical Center Groningen

Using ISMLL Cluster. Tutorial Lec 5. Mohsan Jameel, Information Systems and Machine Learning Lab, University of Hildesheim

NBIC TechTrack PBS Tutorial

MERCED CLUSTER BASICS Multi-Environment Research Computer for Exploration and Discovery A Centerpiece for Computational Science at UC Merced

HPCC New User Training

High Performance Computing (HPC) Using zcluster at GACRC

Name Department/Research Area Have you used the Linux command line?

Grid Computing. Department of Statistics Staff Retreat February 2008

Slurm basics. Summer Kickstart June slide 1 of 49

Batch Systems. Running your jobs on an HPC machine

Introduction to HPC Using zcluster at GACRC

Computing with the Moore Cluster

Kohinoor queuing document

Batch Systems & Parallel Application Launchers Running your jobs on an HPC machine

Feb 22, Explanatory meeting for users of supercomputer system -- Knowhow for entering jobs in UGE --

Introduction to HPC Using zcluster at GACRC

Our new HPC-Cluster An overview

An Introduction to Cluster Computing Using Newton

Intel Manycore Testing Lab (MTL) - Linux Getting Started Guide

The Supercomputing Facility for Bioinformatics & Computational Biology, IIT Delhi

N1GE6 Checkpointing and Berkeley Lab Checkpoint/Restart. Liang PENG Lip Kian NG

Submitting and running jobs on PlaFRIM2 Redouane Bouchouirbat

An Advance Reservation-Based Computation Resource Manager for Global Scheduling

SGE 6.0 configuration guide, version 1.1

Parallel Computing at DESY Zeuthen. Introduction to Parallel Computing at DESY Zeuthen and the new cluster machines

ITCS 4145/5145 Assignment 2

Allowing Users to Run Services at the OLCF with Kubernetes

Univa Grid Engine Troubleshooting Quick Reference

Introducing the HTCondor-CE

PBS Pro Documentation

User Guide of High Performance Computing Cluster in School of Physics

Using the computational resources at the GACRC

High Performance Computing (HPC) Club Training Session. Xinsheng (Shawn) Qin

Requesting Resources on an HPC Facility

UF Research Computing: Overview and Running STATA

MIGRATING TO THE SHARED COMPUTING CLUSTER (SCC) SCV Staff Boston University Scientific Computing and Visualization

Exercise Companion Guide

Cluster Clonetroop: HowTo 2014

Martinos Center Compute Cluster

Introduction to HPC Using zcluster at GACRC

Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop

Knights Landing production environment on MARCONI

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

OpenPBS Users Manual

Grid Compute Resources and Job Management

RHRK-Seminar. High Performance Computing with the Cluster Elwetritsch - II. Course instructor : Dr. Josef Schüle, RHRK

Introduction to HPC Using zcluster at GACRC On-Class GENE 4220

How to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions

Grid Compute Resources and Grid Job Management

XSEDE High Throughput Computing Use Cases

New User Seminar: Part 2 (best practices)

Content. MPIRUN Command Environment Variables LoadLeveler SUBMIT Command IBM Simple Scheduler. IBM PSSC Montpellier Customer Center

Introduction to High-Performance Computing (HPC)

Introduction to Discovery.

Shopping List. resource manager: TORQUE scheduler: Maui (optional) account manager: Gold

Hadoop On Demand User Guide

SINGAPORE-MIT ALLIANCE GETTING STARTED ON PARALLEL PROGRAMMING USING MPI AND ESTIMATING PARALLEL PERFORMANCE METRICS

Introduction to Discovery.

Using Cartesius and Lisa. Zheng Meyer-Zhao - Consultant Clustercomputing

Introduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende

HOD User Guide. Table of contents

Programming Environment on Ranger Cluster

Introduction to the Marc2 Cluster

Workload management at KEK/CRC -- status and plan

Gridengine. Contents. Aim. Configuration of gridengine. From reading group / nlp lunch

Automatic Dependency Management for Scientific Applications on Clusters. Ben Tovar*, Nicholas Hazekamp, Nathaniel Kremer-Herman, Douglas Thain

GUI Installation Manual

Introduction to High-Performance Computing (HPC)

Read mapping with BWA and BOWTIE

Working on the NewRiver Cluster

Transcription:

X Grid Engine Where X stands for Oracle Univa Open Son of more to come...?!? Carsten Preuss on behalf of Scientific Computing High Performance Computing

Scheduler candidates LSF too expensive PBS / Torque not really stable Condor too complicated HTC scheduler (HPC one needed) (S)GE future unknown (initial developer Sun was bought by Oracle during the test phase) Slurm 23.04.12 2

Clusters LSF cluster 240 nodes, 8 cores, 16/24/32 GB 2200 cores at all Slot based scheduling SGE cluster for Alice Grid jobs on top of it Load based scheduling /u, /d, /misc and /user/local 23.04.12 3

Clusters II GE cluster 100 nodes, 16 cores each, 32GB Resource based scheduling new GE cluster 400 nodes, 24 cores, 64 GB 9600 cores at all Infiniband instead of Ethernet resource based scheduling both cluster don't have any NFS system (not even /u) mounted only Lustre for data Software comes from CVMFS (CERN Virtual Maschine File System) HTTP based read only file system 23.04.12 software will be installed via build servers 4 will be administrated by the experiments

New cluster MiniCube build up inside the Testinghalle new cooling concept will have it's own Lustre cluster everything connected via Infiniband will be operational in April (this year ;-) ) 23.04.12 5

Queue concept the new cluster(s) do not have experiment specific queues anymore all queues are resource based ones old queue system was sometimes reason for underutilization of the farm even if jobs were waiting for scheduling the slots per experiment are now done via Fair Share 23.04.12 6

Cluster overview qstat -g c will give an overview over the available queues and the associated slots qhost shows all available batch nodes with load indices qhost -j -q (-h lxbxxx) shows all hosts with their queue instances and jobs running on them 23.04.12 7

User groups Users are differed into 3 groups Default can use up to 100 slots at a time Advanced 400 slots Power can use more or less all slots and resources for the new cluster the rules have to be adapted 23.04.12 8

CVMFS software is installed on so called build servers by the IT department /cvmfs/it.gsi.de/oracle/.. experiments /cvmfs/alice.gsi.de/aliroot/... /cvmfs/hades.gsi.de/... /cvmfs/fairroot.gsi.de/.... 23.04.12 9

Submitting a job login to lxsqueezelust64 (pool) qsub /bin/hostname Unable to run job: Script length does not match declared length. Exiting. What's this? GE expects you to submit a script not a binary qsub -b(inary) y(es) /bin/hostname But wait! What about the output? It will be written to /u/cpreuss which does not exists on the cluster! The job will fail... qsub -wd /lustre/rz/cpreuss/ -b y /bin/hostname Or go to your target directory and use the -cwd option to qsub 23.04.12 10

Submitting a job II differences in submitting scripts and binaries GE will copy the submitted script to the batch node but only the submitted script, be carefull if your script calls another script! a submitted binary will not be copied, so you have to use the path/name on the target batch host if you have a script which is already on the cluster you can use it via the -b y option, GE will not mind. 23.04.12 11

Submitting jobs III in LSF you had to define a queue which your job should go to in GE you can't define queues (they exist but only for administrative usage) what you should define are the resource requirements 23.04.12 12

Submitting jobs IV h_cpu CPU time h_rt Runtime h_rss real memory h_vmem virt. Memory and many more, one can use (qconf -sc will show all variables) be carefull these requirements are needed to sort your jobs to queues which fits to the job needs these requirements are limits, too! exceeding them and your jobs will be killed by GE qsub -l h_rt=1:0:0,h_vmem=2g. 23.04.12 13 1 hour runtime and 2 GB of virt. memory

qsub V -wd /tmp OR /lustre/... -cwd if you are already in the target directorie -terse will return only the jobid nice for automated job submitting and controlling -w e w v will verify your jobs and report Errors Warnings 23.04.12 14 just verify, don't submit anything

qsub VI - Job Arrays qsub -t 1-5:2.. more or less the same as in LSF (but has to start with 1 or greater, not 0) starts one job with 3 tasks task IDs will be 1,3,5 complete job is selected by the jobid tasks by jobid.taskid (not jobid[taskid] like in LSF) qsub -t 1-5 -tc 2. will run only 2 tasks at a time for job throttling 23.04.12 15

qsub VII Dependency Jobs dependency jobs are jobs which are waiting for others to finish in GE the dependency jobs can only wait till the defined job is finished complete no possibility to define the job status (Exit/Done) also not possible to define a number of tasks to be finished if the job is waiting for an array qsub -hold_jid jobid. will pend in state 'hqw' untill job jobid has finished 23.04.12 16

qsub VIII Parallel environments in GE we support parallel environments qsub -pe openmpi 10-20 will submit a job with min. 10 and max. 20 tasks, depending on how full the cluster is the tasks will be clustered on single nodes according to the allocating strategy used by the PE openmpi 23.04.12 17

qstat -j jobid Controlling jobs will show details for the job if the job is pending GE will show you all reasons why a very, very longish output... if the job is running or qstat -j jobid grep usage shows the current used resources can be used to monitor the jobs in real time qstat -j jobid -ext Done jobs can be shown only via qstat -s z 23.04.12 it is NOT Zombie as stated in the manual! 18 GE does not know the states Zombie and UNKWN like LSF

Modifying jobs if you see that the given parameters for your jobs don't make sense you can just change them, no need to kill and submit everything again qalter -j h_rss=4g jobid will change the memory requriement all parameters can be altered, but some only on pending jobs 23.04.12 19

Suspending/deleting/resubmiting jobs qmod -sj jobid job will go to state 's' qmod -usj jobid and the job will run qdel (-f) jobid will delete a job, using '-f' will force this action qresub jobid 23.04.12 20 this will resubmit an exact copy of a job

Job lifecycle The jobs can (and will) have several states : qmod -h jobid qsub qw hqw qmod -cj jobid Eqw s T t z r dr 23.04.12 21 qdel (-f) jobid

Prioritys and Fair Share as written some pages before the experiments don't have a dedicated amount of slots for their usage instead of this the complete cluster is now allocated to the users via Fair Share advantages users can use big parts of the cluster (if it is empty) large consumers are penalized 23.04.12 22

Prioritys and Fair Share qsub -P highprio. will account your resource usage to the Alice project 'highprio' without giving a project the resource usage is accountet to the default project jobs are sorted in order to their priority qstat -u '*' shows in the second column the job priority the spreading of the shares over the different departments will/should be discussed by the department haeds in the future... 23.04.12 23

Infos about done jobs qstat -s z shows only a list of done jobs no additional infos qacct -j jobid should be used to retrieve more detailed infos like exit status and used resources also a resource accumulation is available unfortunately qacct can't be used by users at the moment 23.04.12 24

Some remarks... the help for a command is always 'qsub -help' not -h nor help it is 'qstat -u.' but 'qacct -o.' using qstat shows the submit time/date on pending jobs but the start time/date on running jobs qacct can't be used at the moment do not use queus give fair resource requirements to your jobs, not too much but also not too less there are no user preexecution scripts anymore selecting all jobs/users/hosts is done via '*' not 0 (like in LSF) 23.04.12 25

...and some Do's and Dont's submit test jobs before starting large production checking if the job will run at all checking the resource requirements if submitting several equal jobs, use job arrays instead of single jobs don't get output via mail use the local disk /tmp with care renember that there will be 24 jobs using a single disk, no RAID5 have an eye on your production don't assume that everything will go smooth if there is something going wrong tell us! 23.04.12 26 don't wait until everything is going down the drain