INTRODUCTION TO THE CLUSTER

Similar documents
Introduction to the Cluster

Introduction to the Cluster

Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop

Introduction to Discovery.

Introduction to High-Performance Computing (HPC)

Introduction to High Performance Computing at UEA. Chris Collins Head of Research and Specialist Computing ITCS

Genius Quick Start Guide

Introduction to Discovery.

Introduction to PICO Parallel & Production Enviroment

Introduction to the SHARCNET Environment May-25 Pre-(summer)school webinar Speaker: Alex Razoumov University of Ontario Institute of Technology

Introduction to HPC Using zcluster at GACRC

Introduction to HPC2N

Introduction to Modules at CHPC

The cluster system. Introduction 22th February Jan Saalbach Scientific Computing Group

Introduction to Joker Cyber Infrastructure Architecture Team CIA.NMSU.EDU

Introduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende

XSEDE New User Training. Ritu Arora November 14, 2014

Introduction to High Performance Computing at UEA. Chris Collins Head of Research and Specialist Computing ITCS

New User Seminar: Part 2 (best practices)

Introduction to GALILEO

Introduction to GALILEO

HPCC New User Training

HPC Introductory Course - Exercises

Frequently Asked Questions

Introduction to Discovery.

Using Sapelo2 Cluster at the GACRC

Introduction to CINECA HPC Environment

Introduction to HPC Using zcluster at GACRC

Introduction to High-Performance Computing (HPC)

Introduction to HPC Using zcluster at GACRC

Parallel Programming Pre-Assignment. Setting up the Software Environment

Illinois Proposal Considerations Greg Bauer

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

High Performance Computing (HPC) Using zcluster at GACRC

HPC DOCUMENTATION. 3. Node Names and IP addresses:- Node details with respect to their individual IP addresses are given below:-

Knights Landing production environment on MARCONI

Supercomputing environment TMA4280 Introduction to Supercomputing

Introduction to GALILEO

Sherlock for IBIIS. William Law Stanford Research Computing

Name Department/Research Area Have you used the Linux command line?

Introduction to High Performance Computing Using Sapelo2 at GACRC

Introduction to Cuda Visualization. Graphical Application Tunnelling on Palmetto

Introduction to HPC Using zcluster at GACRC On-Class GENE 4220

LBRN - HPC systems : CCT, LSU

University at Buffalo Center for Computational Research

Workstations & Thin Clients

Elastic Compute Service. Quick Start for Windows

Effective Use of CCV Resources

HPC Workshop. Nov. 9, 2018 James Coyle, PhD Dir. Of High Perf. Computing

HPC Architectures. Types of resource currently in use

Intel Manycore Testing Lab (MTL) - Linux Getting Started Guide

HPC and IT Issues Session Agenda. Deployment of Simulation (Trends and Issues Impacting IT) Mapping HPC to Performance (Scaling, Technology Advances)

The Why and How of HPC-Cloud Hybrids with OpenStack

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 13 th CALL (T ier-0)

Introduction to BioHPC New User Training

HPC Aaditya User Policies & Support

Duke Compute Cluster Workshop. 3/28/2018 Tom Milledge rc.duke.edu

Application Guide. Connection Broker. Advanced Connection and Capacity Management For Hybrid Clouds

Using a Linux System 6

Introduction to BioHPC

Choosing Resources Wisely Plamen Krastev Office: 38 Oxford, Room 117 FAS Research Computing

AN INTRODUCTION TO CLUSTER COMPUTING

How to Use a Supercomputer - A Boot Camp

Introduction to HPC2N

Parallel Programming Pre-Assignment. Setting up the Software Environment

Übung zur Vorlesung Architektur paralleler Rechnersysteme

Graham vs legacy systems

User Instruction UBC Department of Botany insync Backup Service. Prepared by: Botany IT

Crash Course in High Performance Computing

How to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions

Introduction to Cheyenne. 12 January, 2017 Consulting Services Group Brian Vanderwende

Introduction to NCAR HPC. 25 May 2017 Consulting Services Group Brian Vanderwende

CS/Math 471: Intro. to Scientific Computing

Introduction to High-Performance Computing (HPC)

Wednesday, August 10, 11. The Texas Advanced Computing Center Michael B. Gonzales, Ph.D. Program Director, Computational Biology

Our Workshop Environment

JHU Economics August 24, Galaxy How To SSH and RDP

Using Compute Canada. Masao Fujinaga Information Services and Technology University of Alberta

The BioHPC Nucleus Cluster & Future Developments

CS 246 Winter Tutorial 1

For Dr Landau s PHYS8602 course

Getting Started with ESX Server 3i Installable Update 2 and later for ESX Server 3i version 3.5 Installable and VirtualCenter 2.5

Introduction to Linux Workshop 2. The George Washington University SEAS Computing Facility

HPC Resources at Lehigh. Steve Anthony March 22, 2012

BRC HPC Services/Savio

Introduction to BioHPC

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

The RWTH Compute Cluster Environment

Introduction to High Performance Computing and an Statistical Genetics Application on the Janus Supercomputer. Purpose

SuperMike-II Launch Workshop. System Overview and Allocations

Introduction to HPC Using zcluster at GACRC

Introduction to HPC Using zcluster at GACRC

Our Workshop Environment

Lmod: A Modern Environment Module System XALT: Understanding HPC Usage via Job Level Collection. June 14, 2017

Our Workshop Environment

Tech Computer Center Documentation

Introduction to BioHPC

Using WestGrid from the desktop Oct on Access Grid

Introduction. Kevin Miles. Paul Henderson. Rick Stillings. Essex Scales. Director of Research Support. Systems Engineer.

Working on the NewRiver Cluster

Transcription:

INTRODUCTION TO THE CLUSTER

WHAT IS A CLUSTER? A computer cluster consists of a group of interconnected servers (nodes) that work together to form a single logical system. COMPUTE NODES GATEWAYS SCHEDULER USER auth L-STORE GPFS DORS CUSTOM GATEWAYS

WHY YOU SHOULD USE A CLUSTER Single user On-demand resources Few simultaneous processes Limited memory Small data storage capability Many users Resources upon availability Many simultaneous jobs Large memory Big data storage capability

THE GATEWAYS The cluster gateways are the main entry point to the cluster. Multiple gateways in round-robin rotation to guarantee access. GATEWAYS Manage and/or edit files Code development Jobs submission Lightweight debugging Run resources intensive processes Custom gateways are gateways paid by and dedicated to specific research groups. They differ from cluster gateways for: CUSTOM GATEWAYS Restricted group access Users can run resources intensive processes

ACCESS TO THE CLUSTER Users can log into the cluster with a Secure Shell (ssh) client. From a terminal: ssh vunetid@login.accre.vanderbilt.edu PuTTY Bash on Windows Full Ubuntu-based Bash shell To install: https://goo.gl/tasj8u Windows 10 only

ACCOUNT CREDENTIALS I forgot my password or My password is expired How can I change my password? Log into the cluster Open a helpdesk ticket: From the terminal: ssh auth www.accre.vanderbilt.edu/help From auth: passwd When prompted, enter your new password Do not use your VU E-password! Type exit to disconnect from auth Within 20 minutes your new password will be active

Older Newer THE COMPUTE NODES Regular nodes Dual multicore CPUs Random Access Memory Family No. of cores RAM / GB No. of nodes 12 128 41 Haswell 128 73 16 256 38 64 31 Sandy Bridge 12 128 190 256 7 24 56 8 128 21 Westmere 48 35 12 64 24 24 24 Nehalem 8 48 1 96 13 Total 6,632 61,536 554

THE COMPUTE NODES Regular nodes Accelerated nodes Dual multicore CPUs 4 x Nvidia GPU 2 x Xeon Phi (KNC) Random Access Memory

THE SCHEDULER 1 Execute user s workloads in the right priority order 2 3 Provide requested resources on compute nodes Optimize cluster utilization A B C D Users do not access compute nodes directly! A B C D

CLUSTER STORAGE GPFS and DORS are distributed parallel filesystems that allow users to get access to the same set of directories on all nodes and all gateways on the cluster. Nightly backup Included with account For purchase GPFS /home /scratch /data DORS /dors Managed by Center for Structural Biology, supported by ACCRE. Provides easy access to data from both desktops and cluster.

GPFS STORAGE QUOTA QUOTA: LIMIT: When exceeded the user receives a warning message. Usage has to return below the quota within the GRACE PERIOD. Cannot be exceeded. Automatically set to the actual quota usage when grace period expires. Data size Number of files QUOTA LIMIT QUOTA LIMIT GRACE PERIOD GPFS /home /scratch /data 15 GB 20 GB 200,000 300,000 7 days 50 GB 200 GB 200,000 1,000,000 14 days Can be purchased at 1 TB increments.

GPFS STORAGE QUOTA QUOTA: LIMIT: When exceeded the user receives a warning message. Usage has to return below the quota within the GRACE PERIOD. Cannot be exceeded. Automatically set to the actual quota usage when grace period expires. Quota usage on GPFS is accounted in different ways. GPFS /home /scratch USER: USER/GROUP: User ownership User or group ownership /data FILESET: Group s data directory content

GPFS STORAGE QUOTA How can I check my current quota usage? accre_storage Shows the current usage for all quotas associated with the user.

GPFS STORAGE QUOTA How can I check my current quota usage? accre_storage Shows the current usage for all quotas associated with the user.

DATA TRANSFER scp source destination Copy data from source to destination. Remote source or destination must be preceded by vunetid@login.accre.vandebilt.edu: scp scp local_path vunetid@login.accre.vanderbilt.edu:remote_path scp scp vunetid@login.accre.vanderbilt.edu:remote_path local_path

DATA TRANSFER WinSCP https://winscp.net scp

THE SOFTWARE STACK Researchers should focus on the science behind the software they use. Easily search available software Effortless access to selected software Seamless shell initialization LMOD Lua-based framework that provides a convenient way to customize user s environment through software modules. Supports all shells available on the cluster. Incompatible software conflict resolution

SOFTWARE HIERARCHY Software is organized in a tree structure and displayed accordingly to the loaded dependencies. GCC/5.4.0 Intel/2016.3.210 OpenMPI/1.10.3 OpenMPI/1.10.3-RoCE IntelMPI/5.1.3.181 Amber/16 Amber/16 Amber/16

LMOD - THE ESSENTIALS module avail <mod> If no module is passed, print a list of all modules that are available to be loaded. If a module is specified, show all available modules with that name. module load mod1 mod2 Load the specified modules. module unload mod1 mod2 Unload the specified modules. module list Show all modules loaded in the current environment. module purge Remove all loaded modules from the environment.

LMOD - A FEW SHORTCUTS module list ml module load mod1 ml mod1 module unload mod1 ml -mod1 module avail ml av Traditional science is all about finding shortcuts Rudy Rucker

LMOD - RULES OF THE GAME 1 Only one version of a given software can be loaded at any time. LMOD will automatically unload previously loaded modules on name clash. 2 Incompatible programs are prevented to be loaded at the same time. E.g. Anaconda2 / Anaconda3 / Python X.Y.Z 3 If the version is not specified, the default module is loaded for that software. The default module corresponds to the most recent version available. 4 When unloading a module, Lmod does not automatically unload its dependencies. In this way unloading a module will not break other loaded modules dependencies.

LMOD - SEARCHING FOR MODULES How can I search among the visible modules? module avail pattern Show only the visible modules that contain the desired pattern. How can I search through all the modules, even the non visible ones? module spider pattern Search all the modules that contain the desired pattern.

LMOD - SAVE LOADED MODULES I always need the same set of modules. How can I have them loaded automatically? OPTION 1: Add module load statements in your ~/.bashrc file. This is the primary cause of software errors for our cluster users! OPTION 2: Save loaded modules in named collections. module save collection_name Save the list of current loaded modules in ~/.lmod.d/collection_name. module restore collection_name Restore the desired named collection in the current environment.

LMOD - HIDDEN MODULES All the dependencies are built from source with the available compilers. The whole software stack will be (mostly) independent from OS libraries. All non-essential dependencies are hidden for user clarity. module --show-hidden avail Show all the visible modules, including hidden ones. module --show-hidden spider pattern Search across all modules, including hidden ones. To load a hidden module, the version must be specified.

SOFTWARE POLICIES New compilers/mpi with relative software stacks are available every 12 months. Software stacks older than 3 years will be removed. What if the software I need is not available via Lmod? ACCRE uses EasyBuild to build the software stack. Open a ticket to request the installation. If not available via EasyBuild, we will discuss the alternatives.

NEED MORE HELP? Check our Frequently Asked Question webpage: 1 www.accre.vanderbilt.edu/faq Submit a ticket from the helpdesk: 2 www.accre.vanderbilt.edu/help 3 Open a ticket to request an appointment with an ACCRE specialist. DO NOT submit tickets in Rush cluster! Rush tickets are for cluster-wide issues only.

IT S CLUSTER TOUR TIME!

CONNECT WITH REMOTE DISPLAY SUPPORT From a terminal: ssh -X vunetid@login.accre.vanderbilt.edu Install XQuartz and connect as for Linux. www.xquartz.org 1. Install and launch Xming 2. Configure PuTTY https://sourceforge.net/projects/xming