Using the IAC Chimera Cluster

Similar documents
Before We Start. Sign in hpcxx account slips Windows Users: Download PuTTY. Google PuTTY First result Save putty.exe to Desktop

Getting started with the CEES Grid

A Brief Introduction to The Center for Advanced Computing

HPC Resources at Lehigh. Steve Anthony March 22, 2012

A Brief Introduction to The Center for Advanced Computing

UBDA Platform User Gudie. 16 July P a g e 1

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

A Brief Introduction to The Center for Advanced Computing

Supercomputing resources at the IAC

The JANUS Computing Environment

CENTER FOR HIGH PERFORMANCE COMPUTING. Overview of CHPC. Martin Čuma, PhD. Center for High Performance Computing

Introduction to Discovery.

OBTAINING AN ACCOUNT:

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing

Our new HPC-Cluster An overview

Introduction to Discovery.

OpenPBS Users Manual

P a g e 1. HPC Example for C with OpenMPI

Introduction to Discovery.

Introduction to GALILEO

Computing with the Moore Cluster

Cluster Network Products

Introduction to PICO Parallel & Production Enviroment

Reduces latency and buffer overhead. Messaging occurs at a speed close to the processors being directly connected. Less error detection

Answers to Federal Reserve Questions. Training for University of Richmond

High Performance Beowulf Cluster Environment User Manual

Introduction to HPC Using the New Cluster at GACRC

The cluster system. Introduction 22th February Jan Saalbach Scientific Computing Group

Installing and running COMSOL 4.3a on a Linux cluster COMSOL. All rights reserved.

User Guide of High Performance Computing Cluster in School of Physics

Introduc)on to Hyades

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

Quick Guide for the Torque Cluster Manager

Interpolation calculation using cloud computing

Using ITaP clusters for large scale statistical analysis with R. Doug Crabill Purdue University

How to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions

Quick Start Guide. by Burak Himmetoglu. Supercomputing Consultant. Enterprise Technology Services & Center for Scientific Computing

Running Jobs, Submission Scripts, Modules

Cerebro Quick Start Guide

Outline. March 5, 2012 CIRMMT - McGill University 2

Working on the NewRiver Cluster

ACEnet for CS6702 Ross Dickson, Computational Research Consultant 29 Sep 2009

Answers to Federal Reserve Questions. Administrator Training for University of Richmond

Moab Workload Manager on Cray XT3

Introduction to High Performance Computing (HPC) Resources at GACRC

Tech Computer Center Documentation

Simple examples how to run MPI program via PBS on Taurus HPC

High Performance Computing (HPC) Using zcluster at GACRC

Cluster Clonetroop: HowTo 2014

Introduc)on*to*ARC*Resources* and*user*environment*

Introduction to HPC Resources and Linux

Using Sapelo2 Cluster at the GACRC

Knights Landing production environment on MARCONI

Introduction to HPC Using zcluster at GACRC

XSEDE New User Tutorial

New User Tutorial. OSU High Performance Computing Center

Advanced Topics in High Performance Scientific Computing [MA5327] Exercise 1

Guillimin HPC Users Meeting February 11, McGill University / Calcul Québec / Compute Canada Montréal, QC Canada

Introduction to HPC Using zcluster at GACRC

Your Microservice Layout

Name Department/Research Area Have you used the Linux command line?

Guillimin HPC Users Meeting. Bart Oldeman

SGE Roll: Users Guide. Version Edition

and how to use TORQUE & Maui Piero Calucci

Introduction to High Performance Computing (HPC) Resources at GACRC

Introduction to HPC Using zcluster at GACRC

Workflow Optimization for Large Scale Bioinformatics

Introduction to SLURM on the High Performance Cluster at the Center for Computational Research

Effective Use of CCV Resources

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

New User Seminar: Part 2 (best practices)

Introduction to HPC Using the New Cluster at GACRC

UF Research Computing: Overview and Running STATA

Cheese Cluster Training

Guillimin HPC Users Meeting March 17, 2016

Queue systems. and how to use Torque/Maui. Piero Calucci. Scuola Internazionale Superiore di Studi Avanzati Trieste

Shell Scripting. With Applications to HPC. Edmund Sumbar Copyright 2007 University of Alberta. All rights reserved

Genius Quick Start Guide

Using the IBM Opteron 1350 at OSC. October 19-20, 2010

Introduction to HPCC at MSU

Programming Techniques for Supercomputers. HPC RRZE University Erlangen-Nürnberg Sommersemester 2018

Advanced Scripting Using PBS Environment Variables

ITCS 4145/5145 Assignment 2

Introduction to CINECA HPC Environment

Introduction to the HPC Resources at the University of Pittsburgh

Spartan '08 for Linux Installation Instructions

Hybrid MPI+OpenMP Parallel MD

Sherlock for IBIIS. William Law Stanford Research Computing

Computer System: An Overview

ACCRE High Performance Compute Cluster

XSEDE New User Tutorial

The Why and How of HPC-Cloud Hybrids with OpenStack

HPC DOCUMENTATION. 3. Node Names and IP addresses:- Node details with respect to their individual IP addresses are given below:-

RHRK-Seminar. High Performance Computing with the Cluster Elwetritsch - II. Course instructor : Dr. Josef Schüle, RHRK

UAntwerpen, 24 June 2016

An Advance Reservation-Based Computation Resource Manager for Global Scheduling

DATARMOR: Comment s'y préparer? Tina Odaka

Introduction to High-Performance Computing (HPC)

SGE Roll: Users Guide. Version 5.3 Edition

Grid Engine Users Guide. 5.5 Edition

Transcription:

Using the IAC Chimera Cluster Ángel de Vicente (Tel.: x5387) SIE de Investigación y Enseñanza

Chimera overview Beowulf type cluster Chimera: a monstrous creature made of the parts of multiple animals. Mailing list: beowulf@iac.es Web page: http://chimera Course on Adv. Prog. and Parallel Comp. (June 11 25)

Schematic View

Hardware Details Nodes: 1 master node (EM64T) 16 old i686 nodes: 32 Xeon 2.80 Ghz (chi32) 16 new EM64T nodes: 32 Xeon 3.20 Ghz (chi64) RAM: 98 GB (master: 2 + chi32: 32 + chi64: 64) Disk: ~ 5TB (master: 280 + chi32: 480 + chi64: 4.5TB) Network: two independent Gigabit networks (user applications and admin, nfs, etc.)

Disk space User available space: (all) /home (NFS master): 50 GB /scratch (NFS master): 195 GB (chi32) /local_scratch (local): (per node) 20 GB (chi64) /mnt/pvfs2 (PVFS2 chi64): 3.9 TB /home quotas to be implemented automatic deletion in the other partitions to be implemented as well.

PVFS2 Introduction Stripes data across disks (chi64 in Chimera) Larger files can be created, and potential bandwith is increased. Multiple user interfaces: MPI IO support Traditional Linux file system

PVFS2 Example With MPI IO: /scratch (NFS) /mnt/pvfs2 (PVFS2) Processors: 60 Write bandwith: 24MB/s 892MB/s Read bandwith: 116MB/s 482MB/s Traditional Linux file system: local disk /scratch (NFS) /mnt/pvfs2(pvfs2) Processors: 1 Write 900 MB 14.77s 43.942s 11.779s Read 900 MB (wc) 6.401s 10.007s 45.942s

Modules package Dynamic modification of a user's environment: PATH, MANPATH, etc. Shared and/or private modulefiles. Useful in managing different versions of applications. Very simple to use: module help avail list load unload Use module commands is.bashrc for common environment. Useful for dealing with chi32 vs. chi64

Compiling code Code compiled in 64 bits can only run in chi64. Code compiled in 32 bits can run in chi32, chi64 or chimera (chi32 + chi64). By default you login into a 64bits environment. (see this by running uname a) Modules are by default 64 bits. 32 bits versions end with _32 Environment and modules' bitness should match.

Compiling code (2) Compiling example for 64 bits: [angelv@chimera sieminar]$ mpicc o cpi_64 cpi.c [angelv@chimera sieminar]$ file cpi_64 cpi_64: ELF 64 bit LSB executable, AMD x86 64, version 1 (SYSV), for GNU/Linux 2.4.0, dynamically linked (uses shared libs), not stripped Compiling example for 32 bits: env32 puts us into a 32 bits environment [angelv@chimera sieminar]$ module list (verify 32 bits versions) [angelv@chimera sieminar]$ mpicc o cpi_32 cpi.c [angelv@chimera sieminar]$ file cpi_32 cpi_32: ELF 32 bit LSB executable, Intel 80386, version 1 (SYSV), for GNU/Linux 2.2.5, dynamically linked (uses shared libs), not stripped

Submitting jobs to the cluster Chimera's queueing system: Torque: Resource Manager Maui: Scheduler Maui/Torque basic commands: showq, qsub, checkjob, canceljob qsub needs a submitting file: [angelv@chimera sieminar]$ cat submit cpi #!/bin/sh NP=$(wc l $PBS_NODEFILE awk '{print $1}') cd $PBS_O_WORKDIR mpirun np $NP machinefile $PBS_NODEFILE./cpi

Submitting jobs to the cluster (2) With qsub you specify: the number of nodes required, the time required, the bitness of nodes required, etc. Example submissions: To chi64 (default): qsub l nodes=4:ppn=2,walltime=03:00:00 submit cpi To chi32: qsub l nodes=4:ppn=2 q chi32 submit cpi To chimera: qsub l nodes=4:ppn=2 q chimera submit cpi

Scheduling policies Current policies NOT FIFO (/usr/local/maui/maui.cfg): Time in queue Expansion factor Backfilling Number of requested processors Fairshare Max time for a job: 3.5 days for 128 processors. Usage of Beoiac (old cluster): 54.18% (last 2 years) The early bird catches the worm!

Monitorization Graphical view of scheduling status (same output as showq, but perhaps easier to interpret) http://chimera/cgi bin/mauistatus.pl Graphical view of different metrics of the cluster (are your allocated nodes really doing something?) http://chimera/ganglia/

Other resources at the IAC Condor system (~ 180 machines, ideal for parameter studies). Future CALP node (512 nodes, 20% exclusive to IAC)

References Beowulf.org (http://www.beowulf.org) Chimera@wikipedia (http://en.wikipedia.org/wiki/chimera_%28mythology%29) IAC mailing list (http://listas.iac.es/mailman/listinfo/beowulf) Chimera IAC web page (http://chimera/) IAC Course on Parallel Comp. (http://goya/sie/forum/viewtopic.php?t=141) PVFS2 (http://www.pvfs.org) Modules package (http://modules.sourceforge.net) Maui (http://www.clusterresources.com/pages/products/maui cluster scheduler.php) Torque (http://www.clusterresources.com/pages/products/torque resource manager.php) Condor IAC web page (http://www.iac.es/sieinvens/sinfin/condor/)