How to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions

Similar documents
Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop

Introduction to GALILEO

Introduction to PICO Parallel & Production Enviroment

Introduction to GALILEO

Introduction to GALILEO

Batch Systems. Running calculations on HPC resources

Introduction to Joker Cyber Infrastructure Architecture Team CIA.NMSU.EDU

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

Our new HPC-Cluster An overview

Computing with the Moore Cluster

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

User Guide of High Performance Computing Cluster in School of Physics

Introduction to Cheyenne. 12 January, 2017 Consulting Services Group Brian Vanderwende

Introduction to HPC Using the New Cluster at GACRC

Introduction to CINECA HPC Environment

Introduction to High Performance Computing (HPC) Resources at GACRC

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing

Quick Guide for the Torque Cluster Manager

Introduction to HPC Resources and Linux

Knights Landing production environment on MARCONI

Advanced Topics in High Performance Scientific Computing [MA5327] Exercise 1

OBTAINING AN ACCOUNT:

Introduction to High Performance Computing (HPC) Resources at GACRC

Ambiente CINECA: moduli, job scripts, PBS. A. Grottesi (CINECA)

Using Cartesius and Lisa. Zheng Meyer-Zhao - Consultant Clustercomputing

Answers to Federal Reserve Questions. Training for University of Richmond

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

Introduction to HPC Numerical libraries on FERMI and PLX

Using ITaP clusters for large scale statistical analysis with R. Doug Crabill Purdue University

Working on the NewRiver Cluster

Introduction to Unix Environment: modules, job scripts, PBS. N. Spallanzani (CINECA)

Reduces latency and buffer overhead. Messaging occurs at a speed close to the processors being directly connected. Less error detection

Introduction to High Performance Computing Using Sapelo2 at GACRC

Effective Use of CCV Resources

Using Sapelo2 Cluster at the GACRC

Introduction to HPC Using the New Cluster at GACRC

UF Research Computing: Overview and Running STATA

Introduction to NCAR HPC. 25 May 2017 Consulting Services Group Brian Vanderwende

Intel Manycore Testing Lab (MTL) - Linux Getting Started Guide

The cluster system. Introduction 22th February Jan Saalbach Scientific Computing Group

Introduction to CINECA Computer Environment

MIGRATING TO THE SHARED COMPUTING CLUSTER (SCC) SCV Staff Boston University Scientific Computing and Visualization

Cluster Network Products

Batch Systems & Parallel Application Launchers Running your jobs on an HPC machine

Running Jobs, Submission Scripts, Modules

DDT: A visual, parallel debugger on Ra

The GPU-Cluster. Sandra Wienke Rechen- und Kommunikationszentrum (RZ) Fotos: Christian Iwainsky

Introduction to HPCC at MSU

XSEDE New User Tutorial

High Performance Computing (HPC) Club Training Session. Xinsheng (Shawn) Qin

RWTH GPU-Cluster. Sandra Wienke March Rechen- und Kommunikationszentrum (RZ) Fotos: Christian Iwainsky

HPC Middle East. KFUPM HPC Workshop April Mohamed Mekias HPC Solutions Consultant. Agenda

Tech Computer Center Documentation

Introduction to HPC Using the New Cluster at GACRC

Software Ecosystem for Arm-based HPC

Getting started with the CEES Grid

MERCED CLUSTER BASICS Multi-Environment Research Computer for Exploration and Discovery A Centerpiece for Computational Science at UC Merced

Introduction to HPC Using the New Cluster at GACRC

Genius Quick Start Guide

Introduction to Discovery.

Introduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende

Name Department/Research Area Have you used the Linux command line?

How to Use a Supercomputer - A Boot Camp

Sharpen Exercise: Using HPC resources and running parallel applications

Taiwania HPC System User Operation Manual

Transitioning to Leibniz and CentOS 7

PBS Pro Documentation

Batch Systems. Running your jobs on an HPC machine

Introduction to running C based MPI jobs on COGNAC. Paul Bourke November 2006

First steps on using an HPC service ARCHER

Introduc)on to Pacman

Migrating from Zcluster to Sapelo

RHRK-Seminar. High Performance Computing with the Cluster Elwetritsch - II. Course instructor : Dr. Josef Schüle, RHRK

Introduction to parallel computing

Introduction to HPC2N

Introduction to Discovery.

XSEDE New User Tutorial

NBIC TechTrack PBS Tutorial. by Marcel Kempenaar, NBIC Bioinformatics Research Support group, University Medical Center Groningen

Cerebro Quick Start Guide

CENTER FOR HIGH PERFORMANCE COMPUTING. Overview of CHPC. Martin Čuma, PhD. Center for High Performance Computing

HPC Resources at Lehigh. Steve Anthony March 22, 2012

PACE. Instructional Cluster Environment (ICE) Orientation. Mehmet (Memo) Belgin, PhD Research Scientist, PACE

Cluster Clonetroop: HowTo 2014

STARTING THE DDT DEBUGGER ON MIO, AUN, & MC2. (Mouse over to the left to see thumbnails of all of the slides)

Introduction to Discovery.

Installing and running COMSOL 4.3a on a Linux cluster COMSOL. All rights reserved.

NBIC TechTrack PBS Tutorial

Designed for Maximum Accelerator Performance

UAntwerpen, 24 June 2016

FUJITSU PHI Turnkey Solution

HPC DOCUMENTATION. 3. Node Names and IP addresses:- Node details with respect to their individual IP addresses are given below:-

Viglen NPACI Rocks. Getting Started and FAQ

Graham vs legacy systems

The RWTH Compute Cluster Environment

Introduction to Parallel Programming. Martin Čuma Center for High Performance Computing University of Utah

Allinea Unified Environment

Bright Cluster Manager Advanced HPC cluster management made easy. Martijn de Vries CTO Bright Computing

A Brief Introduction to The Center for Advanced Computing

Using ISMLL Cluster. Tutorial Lec 5. Mohsan Jameel, Information Systems and Machine Learning Lab, University of Hildesheim

Sharpen Exercise: Using HPC resources and running parallel applications

Transcription:

How to run applications on Aziz supercomputer Mohammad Rafi System Administrator Fujitsu Technology Solutions

Agenda Overview Compute Nodes Storage Infrastructure Servers Cluster Stack Environment Modules Interactive vs Batch Jobs PBS Scripts 2

Overview KAU VPN Login/Data Nodes PBS Scheduler Compute Cluster PBS script /home/$user /project2/<group>/$user /fefs1/<group>/$user Ethernet Infiniband For account request and queries, please write to hpc@kau.edu.sa 3

Compute Nodes Thin Nodes (thin queue) Fat Nodes (fat queue) 380 Compute Nodes 96 GB memory on each node Total of 9120 cores 112 Compute Nodes 256 GB memory on each node Total of 2688 compute cores GPU Nodes (k20 queue) Intel Xeon Phi Nodes (phi queue) 2 GPGPU Nodes 96 GB memory on each node Total of 4992 cuda cores 2 Intel Xeon Phi Nodes 96 GB memory on each node Total of 120 cores

Storage /home/$user Holds user specific files, scripts and applications Default quota limit is 5GB Accessible from login nodes and all compute nodes /fefs1/<group>/$user High-speed parallel file system with an aggregate of 2 PB of storage space Users infiniband interconnect to provide high-speed low-latency data transfer Accessible from login nodes and all compute nodes /project2/<group>/$user A total of 6 PB storage with archival system Accessible only from kdata1 and kdata2 Completed projects and related data can be moved to this storage area for long term storage

Infrastructure Servers Login Nodes Access Aziz supercomputing facility through these nodes using ssh Should not be used to run user applications Data Servers For data transfer purposes Can be used to transfer data from user system to Aziz or between different file systems in Aziz Project storage is accessible only from data servers Head Nodes Provides management and job scheduling (PBS Scheduler) to the cluster Management Nodes Provides licenses to applications running on Aziz supercomputer Monitoring tools and management software run on these servers

Cluster Stack Applications Climate Aeronautics Genome Chemistry Physics Debugging / performance tuning Allinea DDT GDB Intel Vtune Amplifier Compilers Intel Cluster Studio PGI GNU CUDA Message Passing OpenMPI Intel MPI (Cluster Studio) Workload Manager Altair PBS Professional File System NFS FEFS (Fujitsu Exabyte File System) Cluster Manager / Monitoring HCS Ganglia Server View IPMI Connectivity Infiniband Ethernet Operating System Centos 6.4 Redhat Linux 6.4

Environment Modules Environment modules provides dynamic modification to user s environment Loading and unloading an environment module is analogues to installing and uninstalling an application Examples To load Intel MPI module on Aziz $ module load impi/5.0.3.048 unload Intel MPI $ module unload impi/5.0.3.048 Load different version of Intel MPI $ module load impi/5.1.1.109

Interactive vs Batch Jobs Batch jobs do not require user intervention during course of execution. Once job is submitted to the scheduler, resources are allocated to the job as requested in PBS script and the job runs on compute node. A batch job may contain one or more instructions which will be executed in sequence. Interactivate jobs are usually used so that a person can become familiar with the computing environment and test or debug their code before attempting a long production run as a batch job. Applications like R is usually launched in interactive mode. To run an interactive job, use -I switch of qsub command and shown in example below. $ qsub I l select=1:ncpus=1 q thin

PBS Script Example 01: PBS script to run an application #!/bin/bash #PBS N example01 #PBS l select=1:ncpus=1:mpiprocs=1 #PBS q thin cd /home/user/examples./example01 Command to submit pbs script to scheduler $qsub <Name of the script> Command to view jobs $qstat $qstat u <user name> Command to delete job $qdel <job id>

PBS Script Example 02: PBS script with error and output directories #!/bin/bash #PBS N example02 #PBS l select=1:ncpus=1:mpiprocs=1 #PBS q thin #PBS e /home/user/examples/logs #PBS o /home/user/examples/logs cd /home/user/examples./example02

PBS Script, cont Example 03: PBS script to run an MPI application #!/bin/bash #PBS N example03 #PBS l select=1:ncpus=2:mpiprocs=2 #PBS q thin #PBS e /home/user/examples/logs #PBS o /home/user/examples/logs #PBS l walltime=00:20:00 cd $PBS_O_WORKDIR module load impi/5.1.1.109 mpirun./example03 12

PBS Script, cont Example 04: PBS script to run an MPI application with more number of CPUS #!/bin/bash #PBS N example04 #PBS l select=1:ncpus=24:mpiprocs=24 #PBS q thin #PBS e /home/user/examples/logs #PBS o /home/user/examples/logs #PBS l walltime=00:20:00 cd $PBS_O_WORKDIR module load impi/5.1.1.109 mpirun./example04 13

PBS Script, cont Example 05: PBS script to run an MPI application with multiple nodes #!/bin/bash #PBS N mpi_example #PBS l select=2:ncpus=24:mpiprocs=24 #PBS q thin #PBS e /home/user/examples/logs #PBS o /home/user/examples/logs #PBS l walltime=00:20:00 cd $PBS_O_WORKDIR module load impi/5.1.1.109 echo Job start time: $(date) mpirun./example05 echo Job end time: $(date)./cleanup.sh 14

Thankyou, Merci, Danke schön, Gracias, ありがとう, 谢谢您,شكرا