Name Department/Research Area Have you used the Linux command line?

Similar documents
A Hands-On Tutorial: RNA Sequencing Using High-Performance Computing

Introduction to Joker Cyber Infrastructure Architecture Team CIA.NMSU.EDU

Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop

A Brief Introduction to The Center for Advanced Computing

High Performance Computing (HPC) Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

HPC DOCUMENTATION. 3. Node Names and IP addresses:- Node details with respect to their individual IP addresses are given below:-

UoW HPC Quick Start. Information Technology Services University of Wollongong. ( Last updated on October 10, 2011)

Shark Cluster Overview

A Brief Introduction to The Center for Advanced Computing

A Brief Introduction to The Center for Advanced Computing

Introduction to HPC Using zcluster at GACRC

Using a Linux System 6

Getting started with the CEES Grid

Joint High Performance Computing Exchange (JHPCE) Cluster Orientation.

Introduction to HPC Using zcluster at GACRC On-Class GENE 4220

Shark Cluster Overview

An Introduction to Cluster Computing Using Newton

Introduction to High-Performance Computing (HPC)

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing

UF Research Computing: Overview and Running STATA

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

Supercomputing environment TMA4280 Introduction to Supercomputing

New User Tutorial. OSU High Performance Computing Center

Intel Manycore Testing Lab (MTL) - Linux Getting Started Guide

Using ISMLL Cluster. Tutorial Lec 5. Mohsan Jameel, Information Systems and Machine Learning Lab, University of Hildesheim

Working on the NewRiver Cluster

Sherlock for IBIIS. William Law Stanford Research Computing

OBTAINING AN ACCOUNT:

Introduction to HPC Using zcluster at GACRC

Effective Use of CCV Resources

Using CLC Genomics Workbench on Turing

Introduction to PICO Parallel & Production Enviroment

Using the computational resources at the GACRC

Computing with the Moore Cluster

High Performance Computing (HPC) Club Training Session. Xinsheng (Shawn) Qin

PARALLEL COMPUTING IN R USING WESTGRID CLUSTERS STATGEN GROUP MEETING 10/30/2017

The cluster system. Introduction 22th February Jan Saalbach Scientific Computing Group

NBIC TechTrack PBS Tutorial

Sharpen Exercise: Using HPC resources and running parallel applications

Introduction to High-Performance Computing (HPC)

Our new HPC-Cluster An overview

Introduction to HPC Using zcluster at GACRC

Introduction to BioHPC

Introduction to GALILEO

Using the IBM Opteron 1350 at OSC. October 19-20, 2010

Kohinoor queuing document

Introduction to HPC Using the New Cluster at GACRC

Batch Systems & Parallel Application Launchers Running your jobs on an HPC machine

Introduction to HPC Using zcluster at GACRC

Using Sapelo2 Cluster at the GACRC

Introduction to High Performance Computing (HPC) Resources at GACRC

Working with Shell Scripting. Daniel Balagué

MERCED CLUSTER BASICS Multi-Environment Research Computer for Exploration and Discovery A Centerpiece for Computational Science at UC Merced

Part I. UNIX Workshop Series: Quick-Start

Batch Systems. Running calculations on HPC resources

How to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions

Introduction to Linux for BlueBEAR. January

User Guide of High Performance Computing Cluster in School of Physics

Image Sharpening. Practical Introduction to HPC Exercise. Instructions for Cirrus Tier-2 System

Using Cartesius and Lisa. Zheng Meyer-Zhao - Consultant Clustercomputing

Grid Engine Users Guide. 5.5 Edition

Anthill User Group Meeting, 2015

Introduction: What is Unix?

NBIC TechTrack PBS Tutorial. by Marcel Kempenaar, NBIC Bioinformatics Research Support group, University Medical Center Groningen

Martinos Center Compute Cluster

Introduction to High Performance Computing Using Sapelo2 at GACRC

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing

Graham vs legacy systems

Introduction to Linux Environment. Yun-Wen Chen

Migrating from Zcluster to Sapelo

Review of Fundamentals. Todd Kelley CST8207 Todd Kelley 1

MIGRATING TO THE SHARED COMPUTING CLUSTER (SCC) SCV Staff Boston University Scientific Computing and Visualization

Answers to Federal Reserve Questions. Training for University of Richmond

HPC system startup manual (version 1.20)

Slurm and Abel job scripts. Katerina Michalickova The Research Computing Services Group SUF/USIT November 13, 2013

HPC Workshop. Nov. 9, 2018 James Coyle, PhD Dir. Of High Perf. Computing

High Performance Compu2ng Using Sapelo Cluster

Review of Fundamentals

Our Workshop Environment

Exercises: Abel/Colossus and SLURM

For Dr Landau s PHYS8602 course

CS/Math 471: Intro. to Scientific Computing

Bioinformatics Facility at the Biotechnology/Bioservices Center

An Introduction to Gauss. Paul D. Baines University of California, Davis November 20 th 2012

Installing and running COMSOL 4.3a on a Linux cluster COMSOL. All rights reserved.

Introduction to the SHARCNET Environment May-25 Pre-(summer)school webinar Speaker: Alex Razoumov University of Ontario Institute of Technology

Our Workshop Environment

Introduction to HPC Resources and Linux

Introduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende

Our Workshop Environment

Introduction to High Performance Computing (HPC) Resources at GACRC

Cluster Clonetroop: HowTo 2014

Guillimin HPC Users Meeting April 13, 2017

CSE Linux VM. For Microsoft Windows. Based on opensuse Leap 42.2

SGE Roll: Users Guide. Version Edition

HPC Course Session 3 Running Applications

Basic UNIX commands. HORT Lab 2 Instructor: Kranthi Varala

Transcription:

Please log in with HawkID (IOWA domain) Macs are available at stations as marked To switch between the Windows and the Mac systems, press scroll lock twice 9/27/2018 1 Ben Rogers ITS-Research Services August 29, 2018 9/27/2018 2 1

Name Department/Research Area Have you used the Linux command line? Resources Wiki Documentation - https://wiki.uiowa.edu/display/hpcdocs/argon+cl uster Hpc-sysadmins@iowa.uiowa.edu 9/27/2018 3 What is a Compute Cluster? High Performance Computing High Throughput Computing The Argon Cluster Mapping Your Problem to a Cluster 9/27/2018 4 2

Large number of computers Software that allows them to work together A tool for solving computational problems that require more memory or cpu than is available on a single system http://www.flickr.com/photos/fullaperture/5435786866/sizes/l/in/photostream/ 9/27/2018 5 Using multiple computers in a coordinated way to solve a single problem Provides the ability to: Use 10s-1000s of cores to solve a single problem Allows access to 10s-1000s of GB of Ram Likely to require substantial code modification to use a library such as MPI Common Examples: Computational Fluid Dynamics Molecular Dynamics 9/27/2018 6 3

Using multiple computers in a coordinated way to solve many individual problems Provides the ability to: Analyze many data sets simultaneously Efficiently perform a parameter sweep Requires minimal code modifications Common Examples: Image Analysis Genomics 9/27/2018 7 Collaborative Cluster CentOS 7 (Linux) ~370 compute nodes ~10,000 processor cores ~20,000 threads (slots) 96-512GB of Ram/node 100Gb Omnipath Network ~110 GPU Accelerators 9/27/2018 8 4

More capabilities & heterogeneity coming to the Argon cluster. Plan to keep the Argon name and refresh hardware periodically. Added more GPU capacity in August 2018. Bringing Neon into Argon January 2019. 9/27/2018 9 Home Account Storage NFS 1TB per User /home/$hawkid Shared Scratch Storage NFS: 150TB - /nfsscratch Deleted after 60 days Local SSD Scratch 900GB/Compute Node /localscratch No Backups! 9/27/2018 10 5

Your Desktop Login Node Login Node Head (Scheduler) Node Storage Node Compute Node Compute Node Compute Node Compute Node Compute Node Compute Node Compute Node Compute Node 9/27/2018 11 Home & Scratch Storage Not Shared Between Clusters Today In future there will be a single logical cluster so this will no longer be relevant Paid Storage Service - $40/TB/Year/Copy of Data May be shared between clusters Available outside cluster Backups available Found at /Shared/$LAB & /Dedicated/$LAB 12 6

CIFS (Windows) Network File Share Available for home accounts & paid storage https://wiki.uiowa.edu/display/hpcdocs/home+ac counts Argon Home Server Information \\argon-home.hpc.uiowa.edu\myhawkid sftp Available for all file shares https://wiki.uiowa.edu/display/hpcdocs/using+sc P+or+SFTP 9/27/2018 13 Not quite! Just running on Argon won t necessarily make your program faster. http://basementgeographer.blogspot.com/2012/03/international-racing-colours.html 9/27/2018 14 7

Questions Does your problem require more resources (CPU, GPU, memory) than available on your desktop system? Does your job run on Linux? Can your job run in batch mode? Is your job HPC or HTC? Next Steps Develop Strategy for Running Jobs Install Software Home Account Install Singularity Central Install (Existing via modules) Develop Job Submission Scripts Run Your Job 9/27/2018 15 Run Freesurfer on 1000 MRIs Takes 20 Hours per MRI Requires 2GB of Memory/analysis Desktop Analysis Time 20 Hours x 1000 MRIs = 20,000 Hours 2.3 Years! But I have a Quad Core Desktop with 8GB That s still over six months! http://surfer.nmr.mgh.harvard.edu/ 9/27/2018 16 8

Good fit for cluster? Yes Type of problem HTC Software Runs on Linux in batch mode Time to Analyze On Argon As little as 20 hours Time dependent on cores available, likely complete within a week. Possible to run all analyses simultaneously 1000 processor cores Total on Argon > 10,000 2TB of memory Total on Argon > 89TB 9/27/2018 17 Argon uses the Son of Grid Engine (SGE) scheduler to allocate resources. To request specific compute resources you need to know about: Queues Slots Resources 9/27/2018 18 9

Investor queues Investor owned, guaranteed access Access restricted to specified users UI queue Treated like an investor queue but everyone has access all.q No job limits Subject to eviction 9/27/2018 19 Slots are equivalent to a share of a machines resources (1 slot on a 56 core machine = 1/56 th of the system resources) Processor Cores + Hyperthreads Memory The number of slots you specify can determine how soon your job will be able to run When using mpi (HPC jobs) we recommend using full machines (eg. Request a number of slots evenly divisible by the number of slots on a machine.) 9/27/2018 20 10

There are many resources that can be requested. Review wiki for details. Most common to be aware of are: Memory Example: -l mem_256g=true GPU Example: -l gpu=true # of GPUs Example: -l ngpus=2 CPU Architecture -l cpu_arch=broadwell 9/27/2018 21 Getting logged in to Argon Need to use your HawkID (Iowa domain) password Windows will use Putty Mac will use ssh from Terminal Make sure you do not check save password You will be prompted to accept a key. Say yes to accepting the key ssh p 40 argon.hpc.uiowa.edu 9/27/2018 22 11

Why? Significant increase in security What? Requires password plus something you have What does it look like? 9/27/2018 23 I m connected; how do I start working? Argon is a batch, queued system Jobs must be submitted to a queue, wait their turn; then they are processed When running in a batch system you do not interact with your program in real time. Have to specify options in your job script 9/27/2018 24 12

Similar to regular shell scripts First line must contain path to a valid shell Ex. #!/bin/bash Comments start # Directives start #$ Specify options to Grid Engine These are not comments Explanation of sleeper script in editor 9/27/2018 25 Copy job file from /RS/hpc_training/sleeper.sh cp /RS/hpc_training/sleeper.sh. Edit sleeper.sh Vi or emacs if you have a favorite editor If new, try nano Change to your email address Launch job qsub sleeper.sh Your job 5588776 ("sleeper") has been submitted 9/27/2018 26 13

Check the status of your job qstat u [username] Shows all your jobs Check for output files [Scriptname].o[jobnumber] Example: Sleeper.o3101075 9/27/2018 27 What if you want to cancel a job? qdel Let s submit and try it Edit sleep 60 to sleep 180 (sleeps for 180 seconds) Submit job Check status qdel [jobnumber] Wait a few seconds and then check status 9/27/2018 28 14

What are environment modules? System for changing your shell environment Can be loaded and unloaded Find modules via module avail Load via module load Unload via module unload Try to run matlab Load the Matlab module load matlab/r2016b Now try to running matlab again If a program is not working, you may need to load a module 9/27/2018 29 I/O streams Standard input (stdin) Standard output (stdout) (1) Standard error (stderr) (2) Both stdout & stderr (&) Only some shells (bash) support this Redirect using > for output and < for input 2> redirects stderr and &> redirects both What s an example of how you ve used them? Redirection happens when your job runs on the cluster 9/27/2018 30 15

Newline characters I m getting weird errors and I don t see anything wrong with my file. 9/27/2018 31 9/27/2018 32 16

Use a windows editor that has a UNIX line ending setting such as notepad++ (free software) Edit your text files exclusively on Argon Use dos2unix on Argon to convert the files 9/27/2018 33 Individual Consulting By appointment Contact Us: HPC-Sysadmins@iowa.uiowa.edu For additional details visit http://www.hpc.uiowa.edu 9/27/2018 34 17

hpc-sysadmins@iowa.uiowa.edu ben-rogers@uiowa.edu Suggestions? Additional Information http://www.hpc.uiowa.edu http://its.uiowa.edu/hpc https://wiki.uiowa.edu/display/hpcdocs/argon+cluste r https://wiki.uiowa.edu/display/hpcdocs/twofactor+authentication 9/27/2018 35 18