EIC system user manual

Similar documents
Introduction to NCAR HPC. 25 May 2017 Consulting Services Group Brian Vanderwende

Introduction to Cheyenne. 12 January, 2017 Consulting Services Group Brian Vanderwende

The RWTH Compute Cluster Environment

Introduction Workshop 11th 12th November 2013

Introduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende

Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop

Exercise: Calling LAPACK

Cluster Clonetroop: HowTo 2014

UMass High Performance Computing Center

LSF at SLAC. Using the SIMES Batch Cluster. Neal Adams. Stanford Linear Accelerator Center

Introduction to PICO Parallel & Production Enviroment

Introduction to High-Performance Computing (HPC)

Introduction to HPC Using zcluster at GACRC

Introduction to CINECA Computer Environment

Tech Computer Center Documentation

CAM Tutorial. configure, build & run. Dani Coleman July

Duke Compute Cluster Workshop. 3/28/2018 Tom Milledge rc.duke.edu

Introduction to GALILEO

High Performance Computing (HPC) Using zcluster at GACRC

SCALABLE HYBRID PROTOTYPE

Cerebro Quick Start Guide

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

Computing with the Moore Cluster

Image Sharpening. Practical Introduction to HPC Exercise. Instructions for Cirrus Tier-2 System

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing

Introduction to HPC Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC

Introduction to Unix Environment: modules, job scripts, PBS. N. Spallanzani (CINECA)

LSF HPC :: getting most out of your NUMA machine

Intel Manycore Testing Lab (MTL) - Linux Getting Started Guide

Meteorology 5344, Fall 2017 Computational Fluid Dynamics Dr. M. Xue. Computer Problem #l: Optimization Exercises

Our new HPC-Cluster An overview

RWTH GPU-Cluster. Sandra Wienke March Rechen- und Kommunikationszentrum (RZ) Fotos: Christian Iwainsky

User Guide of High Performance Computing Cluster in School of Physics

HPC Workshop. Nov. 9, 2018 James Coyle, PhD Dir. Of High Perf. Computing

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing

Introduction to HPC Numerical libraries on FERMI and PLX

UAntwerpen, 24 June 2016

XSEDE New User Tutorial

The GPU-Cluster. Sandra Wienke Rechen- und Kommunikationszentrum (RZ) Fotos: Christian Iwainsky

Compiling applications for the Cray XC

Introduction to GALILEO

Xeon Phi Native Mode - Sharpen Exercise

Duke Compute Cluster Workshop. 10/04/2018 Tom Milledge rc.duke.edu

KISTI TACHYON2 SYSTEM Quick User Guide

Using the Yale HPC Clusters

Introduction to HPC Using zcluster at GACRC On-Class GENE 4220

Ambiente CINECA: moduli, job scripts, PBS. A. Grottesi (CINECA)

CSC Supercomputing Environment

Exercise 1: Connecting to BW using ssh: NOTE: $ = command starts here, =means one space between words/characters.

Useful Unix Topics. Jeremy Sanders. October 2011

Introduction to Using OSCER s Linux Cluster Supercomputer This exercise will help you learn to use Sooner, the

The JANUS Computing Environment

Sharpen Exercise: Using HPC resources and running parallel applications

Xeon Phi Native Mode - Sharpen Exercise

Introduction to GALILEO

Duke Compute Cluster Workshop. 11/10/2016 Tom Milledge h:ps://rc.duke.edu/

Introduction to HPC Using the New Cluster at GACRC

Using the computational resources at the GACRC

PACE. Instructional Cluster Environment (ICE) Orientation. Mehmet (Memo) Belgin, PhD Research Scientist, PACE

High Performance Computing How-To Joseph Paul Cohen

SGI OpenFOAM TM Quick Start Guide

UoW HPC Quick Start. Information Technology Services University of Wollongong. ( Last updated on October 10, 2011)

Research. We make it happen. Unix Basics. User Support Group help-line: personal:

Beginner's Guide for UK IBM systems

Basic Shell Commands

Minerva Scientific Computing Environment. An Introduction to Minerva

Windows-HPC Environment at RWTH Aachen University

MIGRATING TO THE SHARED COMPUTING CLUSTER (SCC) SCV Staff Boston University Scientific Computing and Visualization

Graham vs legacy systems

NCAR Computation and Information Systems Laboratory (CISL) Facilities and Support Overview

Genius Quick Start Guide

Using the Yale HPC Clusters

Batch environment PBS (Running applications on the Cray XC30) 1/18/2016

ARMINIUS Brief Instructions

High Performance Beowulf Cluster Environment User Manual

HPC Resources at Lehigh. Steve Anthony March 22, 2012

Introduction to High-Performance Computing (HPC)

Submitting and running jobs on PlaFRIM2 Redouane Bouchouirbat

XSEDE New User Tutorial

Introduction to HPC Using zcluster at GACRC

XSEDE New User Tutorial

Introduction to High Performance Computing at UEA. Chris Collins Head of Research and Specialist Computing ITCS

Slurm and Abel job scripts. Katerina Michalickova The Research Computing Services Group SUF/USIT November 13, 2013

An Introduction to Cluster Computing Using Newton

OBTAINING AN ACCOUNT:

TITANI CLUSTER USER MANUAL V.1.3

Using Cartesius and Lisa. Zheng Meyer-Zhao - Consultant Clustercomputing

Sharpen Exercise: Using HPC resources and running parallel applications

Introduc)on to Hyades

PACE. Instructional Cluster Environment (ICE) Orientation. Research Scientist, PACE

XSEDE New User Tutorial

TotalView. Debugging Tool Presentation. Josip Jakić

Advanced Topics in High Performance Scientific Computing [MA5327] Exercise 1

Introduction to HPC Using zcluster at GACRC

Using the IBM Opteron 1350 at OSC. October 19-20, 2010

Effective Use of CCV Resources

COMP Superscalar. COMPSs at BSC. Supercomputers Manual

How to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions

MIC Lab Parallel Computing on Stampede

Transcription:

EIC system user manual how to use system Feb 28 th 2013 SGI Japan Ltd.

Index EIC system overview File system, Network User environment job script Submitting job Displaying status of job Canceling,deleting job How to use peripheral devices Application software 2

EIC system overview IB-SW Voltaire 4x QDR 36 ports frontend server 1 node high spec server 2 nodes parallel server 2 nodes Altix UV 100 Xeon X7542 2.66GHz 8CPU/128GB Altix UV 1000 Xeon X7542 2.66GHz 128CPU/4TB Altix UV 1000 Xeon X7542 2.66GHz 256CPU/4TB FC-SW Brocade 300 8Gbps 24 ports x2 Disk storage 218TB backup storage 163TB InfiniteStorage 5000 218TB Altix XE 500 Xeon E5520 2.26GHz 2CPU/48GB InfiniteStorage 5000 163TB 3 Altix XE 500 Xeon E5520 2.26GHz 2CPU/48GB CXFS server

File system, network LAN NFS Frontend server UV 100 High spec server UV 1000 High spec server UV 1000 Parallel serever UV 1000 Parallel server UV 1000 CXFS server Altix XE User workstation User workstation 6 nodes User workstation 6 nodes 6 nodes 8Gbps FC SW /home 40TB backup area 160TB /work 80TB 4

User environment EIC user can login and use the following servers. hostname Hardware IP address notice eic SGI UV 100 133.11.57.80 Frontend server eic00 Dell Precision T3500 133.11.57.84 Workstation with DAT drive eic01 Dell Precision T3500 133.11.57.85 Workstation with DAT drive eic02 Dell Precision T3500 133.11.57.86 Workstation with Blu-ray drive eic03 Dell Precision T3500 133.11.57.87 Workstation with Blu-ray drive eic04 Dell Precision T3500 133.11.57.88 Workstation eic05 Dell Precision T3500 133.11.57.89 Workstation 5

User environment File system /home(home area) is total 40TB, you can use 150GB as default quota limit. /work(temporary area) is total 80TB, you can use 2000GB as default quota limit. files which has never been accessed for 30 days are deleted. 6

User environment TSS(interactive job) limitation limitation CPU TIME 1 hour MEMORY SIZE 1GB STACK SIZE 4GB CORE SIZE 0GB Number of CPU 1 Please use LSF batch software if you will run a job over TSS limitation. 7

User environment environment variable On EIC system your environment variable was already set to use. You don t have to set environment variable by yourself. There might be a trouble if you move environment variable files (ex.cshrc) in EIC system from other system, please pay your attention. When you face any problem (ex can t submit batch job, can t check output file) after you migrated any environment variable files, delete your.cshrc on your home directory. 8

How to login How to login Please login a frontend server eic for making a program, compiling, interactive debugging, submitting a batch job, frontend sever s hostname is eic.eri.u-tokyo.ac.jp telnet,rsh,rlogin are not permitted on frontend server, please use ssh(secureshell). Login from Linux workstation. You can login to use SSH from your Linux workstation. $ ssh l username eic.eri.u-tokyo.ac.jp 9

How to login(windows) how to login Please use Windows SSH software (TeraTerm,Putty) HOST:eic.eri.u-tokyo.ac.jp username:username password:password following is TeraTerm sample. 10

job script Creating job script file Create job script file to submit a batch job, sample is in a right square. You must define.. #BSUB-q queue name #BSUB-n number of cpu cores #BSUB -o output file name dplace insert before command or program to improve performance. Attention job s standard output or error are temporary saved on /home, finally written into a file which you define as o. If you don t define o, outputs are sent to eic as email. Email size is limited in 1MB, please define o filename or re-direct filename on command line. #!/usr/bin/csh #BSUB -q A #BSUB -n 1 #BSUB -o sample.out dplace./sample 4000 11

Submitting job Use bsub to submit a batch job. Re-direct a job script to bsub command. $ cat sample.csh #!/usr/bin/csh #BSUB -q A #BSUB -n 1 #BSUB -o sample.out dplace./sample 4000 $ bsub < sample.csh set LSB_SUB_MAX_NUM_PROCESSORS is 6 Job <958> is submitted to queue <A>. job ID is printed. 12

Displaying job status bstatus bstatus displays status of jobs you submitted. $ bstatus JOBID USER STAT QUEUE FROM_HOST EXEC_HOST JOB_NAME SUBMIT_TIME 959 sgi RUN C eic 24*eicp1 *para2.csh Feb 3 16:26 961 sgi PEND C eic *para3.csh Feb 3 16:32 row STAT displays status of jobs. RUN ----- job is running PEND---- job is pending bjobs command displays only your jobs. 13

Canceling, Deleting job bkill bkill can cancel or delete jobs. define job ID. $ bjobs JOBID USER STAT QUEUE FROM_HOST EXEC_HOST JOB_NAME SUBMIT_TIME 957 sgi RUN C eic 24*eicp1./para.csh Feb 3 15:44 $ bkill 957 Job <957> is being terminated $ bjobs No unfinished job found 14

Queue configuration Queue name Runtime Memory limit Maximum memory limit Parallel limit (cores) Queue name: Queue name Runtime: Wallclock time limitation per job(only queue A limits cputime) Memory limit: Memory limitation per job(default) Maximum memory limit:memory limitation per job(if you define M when you submit) Parallel limit: number of CPU(cores) per job Job limit: number of running job per user Job limit (cores) A 2h(cputime) 8GB 16GB 1(6) 1(6) B 100h 32GB 32GB 1(6) 4(24) C 80h 128GB 128GB 4(24) 3(72) D 70h 256GB 256GB 8(48) 3(144) E 50h 256GB 512GB 16(96) 2(192) F 40h 512GB 1024GB 32(192) 1(192) M 12h 8GB 8GB MATLAB 15

MPI job script sample $ cat go.24 #!/usr/bin/csh #BSUB -q C #BSUB -n 24 #BSUB -o test.out mpirun -np 24 dplace -s1./xhpl < /dev/null >& out.mpi 24 mpi parallel job(sample) Define 24(number of parallel) on BSUB n and mpirun -np Insert dplace -s1 before running module name. 16

Submitting MPI job $ bsub <./go.24 set LSB_SUB_MAX_NUM_PROCESSORS is 24 Job <751> is submitted to queue <C>. $ bjobs JOBID USER STAT QUEUE FROM_HOST EXEC_HOST JOB_NAME SUBMIT_TIME 751 sgi RUN C eic 24*eicp1./go.36 Feb 3 10:13 24 parallel job runs on 4CPU(24cores) EIC servers have 12cores on local memory, all jobs are automatically set to multiple of 12 if number of cores is not multiple of 12. 17

OpenMP job script sample $ cat para.csh #!/usr/bin/csh #BSUB -q D #BSUB -n 48 #BSUB -o test.out setenv OMP_NUM_THREADS 48 dplace -x2./para < /dev/null >& out.para 48 OpenMP parallel job(sample) Define 48(number of parallel) on BSUB n and environment variable OMP_NUM_THREADS Insert dplace -x2 before module name(-x2 is not required if build by GNU compiler.) 18

Submitting OpenMP job $ bsub <./para.csh set LSB_SUB_MAX_NUM_PROCESSORS is 48 Job <957> is submitted to queue <D>. $ bjobs JOBID USER STAT QUEUE FROM_HOST EXEC_HOST JOB_NAME SUBMIT_TIME 957 sgi RUN D eic 48*eicp1./para.csh Feb 3 10:13 48 parallel job runs on 8CPU(48cores) EIC servers have 12cores on local memory, all jobs are automatically set to multiple of 12 if number of cores is not multiple of 12. 19

MPI+OpenMP Hybrid Parallel job What is hybrid parallel? MPI processes boot OpenMP threads. recommends 2,3, or 6 as number of OpenMP threads because OpenMP theads in same MPI process should use same local memory. 4mpi x 3 thread sample MPI process OpenMP thread CPU0 CPU1 20

MPI+OpenMP Hybrid job script(sample) $ cat go.csh #!/bin/csh -x #BSUB -q C #BSUB -n 24 #BSUB -o hy1.out limit stacksize unlimited set np=8 set th=3 setenv OMP_NUM_THREADS ${th} mpirun -np ${np} omplace -nt ${th} -c 0-23:bs=${th}+st=3./a.out 24 hybrid parallel job(8mpi x 3 threads) Define number of MPI for np, number of threads per MPI for th. Use omplace instead of dplace. insert following before command name. omplace -nt ${th} -c 0-23:bs=${th}+st=3 -c means using 3cores from core 0 to core 23. 21

Core Hopping What is core hopping? Normally job occupies all 6cores and local memory on same CPU socket. You can reduce used cores per CPU if you would like to use wider memory band width per thread, it s called core hopping. Normal process allocation: occupies 2CPU(12cores) MPI process (or OpenMP threads) CPU0 CPU1 Core hopping allocation : occupies 3CPU(18cores) idle core CPU0 CPU1 CPU2 22

Queue option for core hopping Define not only normal -n, but also -P how many cores you use per CPU. (define P from 1 to 6) You have to select larger queue because you occupy more cores than number of n. see following table. Number of parallel cores per CPU Queue name Queue Option Number of occupied CPU (cores) 8 4 C #BSUB- q C #BSUB -n 8 #BSUB -P 4 32 4 D #BSUB- q D #BSUB -n 32 #BSUB -P 4 64 4 E #BSUB- q E #BSUB -n 64 #BSUB -P 4 2 (12) 8 (48) 16 (96) 23

Core hopping MPI job script #!/usr/bin/csh #BSUB -q D #BSUB -n 32 #BSUB -P 4 #BSUB -o mpi4x8.out source /opt/lsf/local/mpienv.csh 32 4 mpirun -np 32./xhpl < /dev/null >& out 32 parallel MPI job(4 cores per CPU) #BSUB -n number of parallel #BSUB -P cores per cpu(1~6) source /opt/lsf/local/mpienv.csh [number of parallel] [cores per cpu] (if you use sh(bash). /opt/lsf/local/mpienv.sh [number of parallel] [cores per cpu] ) mpirun -np number of parallel command name Delete dplace 24

Core hopping OpenMP job script #!/usr/bin/csh #BSUB -q D #BSUB -n 32 #BSUB -P 4 #BSUB -o out set th=32 setenv OMP_NUM_THREADS ${th} dplace -x2 0-3,6-9,12-15,18-21,24-27,30-33,36-39,42-45./para >& out.para または omplace -nt ${th} -c 0-:bs=4+st=6./para >& out.para 32 parallel OpenMP job(4 cores per CPU) #BSUB -n number of parallel #BSUB -P cores per cpu(1~6) omplace -nt [numberof parallel] -c 0-:bs=[cores per cpu] +st=6 [command name]. 25

Core hopping hybrid job script #!/bin/csh #BSUB -q C #BSUB -n 32 #BSUB -P 4 #BSUB -o hy32-4.out set np=8 set th=4 setenv OMP_NUM_THREADS ${th} mpirun -np ${np} omplace -nt ${th} -c 0-:bs=${th}+st=6./a.out BSUB -n number of parallel #BSUB -P cores per cpu(1~6) setenv OMP_NUM_THREADS [number of OpenMP threads] mpirun -np [number of MPI] omplace -nt [number of OpenMP] -c 0-:bs=(core per cpu)+st=6 command. 26

Displaying core hopping job qstatus qstatus displays core hopping job or normal job. $ bjobs JOBID USER STAT QUEUE FROM_HOST EXEC_HOST JOB_NAME SUBMIT_TIME 33359 sgi RUN E eic 48*eicp1 *t.sample5 Apr 27 11:28 $ qstatus c/p CPUTIME/ CPUTIME WALLTIME MEMORY JOB_ID USER_NAME STAT Q HOST PROC (-P) WALLTIME (hh:mm:ss) (hh:mm:ss) (GB) ------------------------------------------------------------------------------------- 33359 sgi RUN E eicp1 48 4/6 29.4 00:50:08 00:01:42 17.0 27 cores per CPU

How to use printer how to print from eic, eicxx Use lpr eic%lpr -Pprinter_name PSfile_name ex) eic%lpr Pxdp1-6f test.ps how to print text file Use a2ps to print text file eic% a2ps -Pprinter_name ascii.txt ex) eic%a2ps Pxdp1-6f /home/sgi/ascii.txt Displaying print status Use lpq eic%lpq -Pprinter_name ex) eic%lpq Pxdp1-6f Rank Owner Job Files 1st root 2 /home/sgi/test.f Cancel printing Use lprm eic%lprm -Pprinter_name request_id confirm request_id lpq -Pprinter_name ex)eic%lpq Pxdp1-6f Rank Owner Job Files 1st root 2 /home/sgi/test.f eic%lprm Pxdp6-1f 2 28

How to use DAT drive DAT drive connect to eic00 and eic01. /dev/st0------rewinding /dev/nst0----no rewinding Use tar or cpio mt command to rewind or forward. mt change uncompress or compress. When you use DAT tape, please confirm compression mode. 29

How to use DAT Writing tape $ mt -f /dev/st0 rewind rewinding tape media $ mt -f /dev/st0 compression 0 define 0 for uncompression, 1for compression. $ cd /home/sgi/test moving backup directory $ tar cvf /dev/st0. writing current directory to tape, and rewind when it finishes Reading tape $ mt -f /dev/st0 rewind rewinding tape media $ cd /home/sgi/test moving writing directory $tar xvf /dev/st0 Confirming tape media writing data to current directory, and rewind when it finishes $ mt -f /dev/st0 rewind rewinding tape media $tar tvf /dev/st0 confirming tape media See online manual man mt or man tar 30

How to use Blu-ray drive Blu-ray drive connect to eic02 and eic03 bdr command boots GUI writing software. $ bdr confirm target drive as PIONEER BD-RW BDR-205 Rev1.08(p:1 t:0) Select cursor menu on right side of drive name. See User Manual Chapter4, manual is available from http://wwweic.eri.u-tokyo.ac.jp/computer/manual/altixuv/doc/misc/bdrgui.pdf 31

Application software AVS is available on workstations(eic00~eic05). login to workstation, use express command. eic00$ express See manual http://kgt.cybernet.co.jp/article/2497/index.html IMSL Fortran Library IMSL Fortran library Ver7.0 is available on EIC. TSS, OpenMP ifort o [module name] $FFLAGS [source name] $LINK_FNL MPI ifort o [module name] $FFLAGS [source name] $LINK_MPI 32

MATLAB MATLAB you can use matlab on eic Login to eic % ssh -X username@eic.eri.u-tokyo.ac.jp run matlab % matlab You have to use LSF batch if you will run matlab over TSS limitation. See next page for matlab via LSF limitation CPU TIME 1 hour MEMORY SIZE 1GB STACK SIZE 4GB CORE SIZE 0GB Number of CPU 1 33

MATLAB MATLAB via LSF how to use matlab via LSF(batch) Login to eic % ssh -X username@eic.eri.u-tokyo.ac.jp submit a job % bsub -q M n 1 -Is /bin/tcsh or % bsub -q M n 1 -Is /bin/bash Job <1519> is submitted to queue <M>. <<Waiting for dispatch...>> <<Starting on eic>> Confirm DISPLAY variable % env grep DISPLAY eic:xx.0 Change DISPLAY variable. % setenv DISPLAY localhost:xx.0 or % export DISPLAY=localhost:xx.0 % xhost + Run MATLAB % matlab 34

Attention When you finish MATLAB, you have to % exit if you don t exit, MATLAB license will be still used, other user will not be able to use it. MATLAB licenses are 10, you can t run when all licenses are used. when you bsub, MATLAB License is over now is displayed. You can use MATLAB on workstations (eic00~eic05). % matlab (you can t run when all licenses are used.) 35