Bright Cluster Manager

Similar documents
Bright Cluster Manager Advanced HPC cluster management made easy. Martijn de Vries CTO Bright Computing

Optimizing Cluster Utilisation with Bright Cluster Manager

Bright Cluster Manager

Introduction to High-Performance Computing (HPC)

Introduction to Abel/Colossus and the queuing system

How to Use a Supercomputer - A Boot Camp

Introduction to Joker Cyber Infrastructure Architecture Team CIA.NMSU.EDU

Introduction. Kevin Miles. Paul Henderson. Rick Stillings. Essex Scales. Director of Research Support. Systems Engineer.

HySecure Quick Start Guide. HySecure 5.0

Sherlock for IBIIS. William Law Stanford Research Computing

Managing complex cluster architectures with Bright Cluster Manager

The Why and How of HPC-Cloud Hybrids with OpenStack

Cloud Control Panel (CCP) User Guide

Graham vs legacy systems

Introduction to Operating Systems (Part II)

Slurm basics. Summer Kickstart June slide 1 of 49

Grid Computing Competence Center Large Scale Computing Infrastructures (MINF 4526 HS2011)

High Performance Computing Cluster Advanced course

November 7, DAN WILSON Global Operations Architecture, Concur. OpenStack Summit Hong Kong JOE ARNOLD

Bright Cluster Manager 7.1. Cloudbursting Manual. Revision: Date: Thu, 10 Dec 2015

Allowing Users to Run Services at the OLCF with Kubernetes

Bright Cluster Manager 7.0. Cloudbursting Manual. Revision: 1fabda4. Date: Fri Mar

What's New in SUSE LINUX Enterprise Server 9

Paperspace. Architecture Overview. 20 Jay St. Suite 312 Brooklyn, NY Technical Whitepaper

2013 AWS Worldwide Public Sector Summit Washington, D.C.

What s new in HTCondor? What s coming? HTCondor Week 2018 Madison, WI -- May 22, 2018

SUN CUSTOMER READY HPC CLUSTER: REFERENCE CONFIGURATIONS WITH SUN FIRE X4100, X4200, AND X4600 SERVERS Jeff Lu, Systems Group Sun BluePrints OnLine

Introduction to High-Performance Computing (HPC)

Bright Cluster Manager 7.1. Installation Manual. Revision: Date: Thu, 10 Dec 2015

Bright Cluster Manager 7.2. Installation Manual. Revision: e9f97c1. Date: Tue May

RHRK-Seminar. High Performance Computing with the Cluster Elwetritsch - II. Course instructor : Dr. Josef Schüle, RHRK

Linux Administration

Installation Prerequisites

ACCRE High Performance Compute Cluster

Working with Shell Scripting. Daniel Balagué

MERCED CLUSTER BASICS Multi-Environment Research Computer for Exploration and Discovery A Centerpiece for Computational Science at UC Merced

Research Project 1. Failover research for Bright Cluster Manager

Red Hat HPC Solution Overview. Platform Computing

PBS PROFESSIONAL VS. MICROSOFT HPC PACK

Introduction to the NCAR HPC Systems. 25 May 2018 Consulting Services Group Brian Vanderwende

Evolving HPC Solutions Using Open Source Software & Industry-Standard Hardware

How to run applications on Aziz supercomputer. Mohammad Rafi System Administrator Fujitsu Technology Solutions

Introduction to the Cluster

Bright Cluster Manager 5.2. Administrator Manual. Revision: Date: Fri, 27 Nov 2015

Bright Cluster Manager 7.3. Installation Manual. Revision: Date: Thu, 25 Jan 2018

Opportunities for container environments on Cray XC30 with GPU devices

Habanero Operating Committee. January

Installation Prerequisites

ECE 574 Cluster Computing Lecture 4

STARTING THE DDT DEBUGGER ON MIO, AUN, & MC2. (Mouse over to the left to see thumbnails of all of the slides)

Write a technical report Present your results Write a workshop/conference paper (optional) Could be a real system, simulation and/or theoretical

Elastic Efficient Execution of Varied Containers. Sharma Podila Nov 7th 2016, QCon San Francisco

Running Galaxy in an HPC environment requirements, challenges and some solutions : the LIFEPORTAL

Introduction to Visualization on Stampede

Resiliency Replication Appliance Installation Guide Version 7.2

Amazon AWS-Solution-Architect-Associate Exam

Experiences with OracleVM 3.3

Choosing Resources Wisely Plamen Krastev Office: 38 Oxford, Room 117 FAS Research Computing

OBTAINING AN ACCOUNT:

SuperMike-II Launch Workshop. System Overview and Allocations

Bright Cluster Manager 8.0. Cloudbursting Manual. Revision: e Date: Mon Apr

Porting SLURM to the Cray XT and XE. Neil Stringfellow and Gerrit Renker

Slurm and Abel job scripts. Katerina Michalickova The Research Computing Services Group SUF/USIT November 13, 2013

Bright Cluster Manager 7.0. Installation Manual. Revision: 1fabda4. Date: Fri Mar

Cornell Red Cloud: Campus-based Hybrid Cloud. Steven Lee Cornell University Center for Advanced Computing

Workstations & Thin Clients

Introduction to BioHPC

Deploy the ExtraHop Trace Appliance with VMware

PODShell: Simplifying HPC in the Cloud Workflow

Introduction to High Performance Computing and an Statistical Genetics Application on the Janus Supercomputer. Purpose

High Performance Computing Cluster Basic course

Intellicus Cluster and Load Balancing- Linux. Version: 18.1

The BioHPC Nucleus Cluster & Future Developments

Bright Cluster Manager 7.3. Cloudbursting Manual. Revision: 079bc45. Date: Mon Sep

Zadara Enterprise Storage in

Answers to Federal Reserve Questions. Training for University of Richmond

How то Use HPC Resources Efficiently by a Message Oriented Framework.

Power Systems with POWER8 Scale-out Technical Sales Skills V1

Overview of a virtual cluster

Getting started with the CEES Grid

Unit 2: Manage Files Graphically with Nautilus Objective: Manage files graphically and access remote systems with Nautilus

Using a Linux System 6

BeeGFS. Parallel Cluster File System. Container Workshop ISC July Marco Merkel VP ww Sales, Consulting

Introduction to SLURM on the High Performance Cluster at the Center for Computational Research

TELSTRA CLOUD SERVICES CLOUD INFRASTRUCTURE PRICING GUIDE UNITED KINGDOM

Amazon Web Services Training. Training Topics:

Reduces latency and buffer overhead. Messaging occurs at a speed close to the processors being directly connected. Less error detection

AWS Storage Gateway. Amazon S3. Amazon EFS. Amazon Glacier. Amazon EBS. Amazon EC2 Instance. storage. File Block Object. Hybrid integrated.

LBRN - HPC systems : CCT, LSU

Genomics on Cisco Metacloud + SwiftStack

Batch Usage on JURECA Introduction to Slurm. May 2016 Chrysovalantis Paschoulas HPS JSC

HPC File Systems and Storage. Irena Johnson University of Notre Dame Center for Research Computing

THE ZADARA CLOUD. An overview of the Zadara Storage Cloud and VPSA Storage Array technology WHITE PAPER

VMware vsphere 6.0 / 6.5 Infrastructure Deployment Boot Camp

OpenNebula on VMware: Cloud Reference Architecture

High Performance Computing Cloud - a PaaS Perspective

HPC Architectures. Types of resource currently in use

On-demand provisioning of HEP compute resources on cloud sites and shared HPC centers

GNU/Linux advanced administration

Vendor: IBM. Exam Code: C Exam Name: Power Systems with POWER8 Scale-out Technical Sales Skills V1. Version: Demo

Transcription:

Bright Cluster Manager Using Slurm for Data Aware Scheduling in the Cloud Martijn de Vries CTO

About Bright Computing Bright Computing 1. Develops and supports Bright Cluster Manager for HPC systems, server farms, grids and clouds 2. Incorporated in USA and The Netherlands (Offices in San Jose and Amsterdam)

Architecture Cluster Management GUI Cluster Management Shell SOAP+SSL CMDaemon Bright Cluster SOAP+SSL node001 Web-Based User Portal head node node002 Third-Party Applications node003

Management Interfaces Graphical User Interface (GUI) Offers administrator full cluster control Standalone desktop application Manages multiple clusters simultaneously Runs on Linux & Windows Built on top of Mozilla XUL engine Cluster Management Shell (CMSH) All GUI functionality also available through Cluster Management Shell Interactive and scriptable in batch mode 4

Bright Cluster Manager Elements Cluster Management GUI User Portal SSL / SOAP / X509 / IPtables Cluster Management Daemon Cluster Management Shell Provisioning Slurm Monitoring Automation Health Checks Management Compilers Libraries Debuggers Profilers SLES / RHEL / CentOS / SL SLES / RHEL / CentOS / SL ScaleMP vsmp CPU GPU Memory Disk Ethernet Interconnect IPMI / ilo PDU

6

7

8

Integration with Slurm Slurm default choice for workload management system Slurm up and running at first boot Node & partition configuration Topology configuration HA configuration Workload management metrics Health checking Job monitoring and control Integrated in Cluster Management API 9

10

Cloud Bursting Cluster On Demand Scenario node001 head node node002 node003

Mixing Local and Cloud Resources Cloud does not work well for all HPC workloads Sensitive data/computations Problems getting huge amounts of data in/out Workload may depend on low latency / high bandwidth Workload may depend on non-standard compute resources Workload may depend on advanced shared storage (e.g. Lustre) Not everyone will replace HPC cluster with EC2 account Allow local cluster to be extended with cloud resources to give best of both worlds Allow workload suitable for cloud to be off-loaded Allow traditional HPC users to try out and migrate to cloud 12

Cloud Bursting Cluster Extension Scenario node006 node004 node007 node005 head node node001 node002 node003

14

Cloud Network Map 15

Cloud Director Cloud Director acts as a head node in the cloud Provides gateway between local and cloud nodes Provisions software image to cloud nodes Serves shared storage for cloud nodes Mirrors network services for the cloud nodes (e.g. LDAP, DNS) Cloud node booting process Instances are created with 1GB EBS and ngb ephemeral/ebs disk Bright Node Installer AMI goes on EBS disk Node Installer continues with normal procedure to bring up node Software image gets provisioned onto second disk 16

Uniformity Cloud nodes behave the same way as local nodes Same method of provisioning Same software image and user environment Same workload management set-up Same management interface that allows to control cluster Same monitoring & health checking Everything can talk to everything Accomplished using VPN, routing, network mapping VPN set-up automated and does not require firewall set-up (requires just outgoing access on 1194/udp) Single global DNS namespace 17

Workload Driven Cloud Resizing Nodes are created in the cloud: Manually by administrator using CMGUI/CMSH Automatically based on workload by cloud-resize utility Cloud-resize called periodically from crond Three inputs to cloud-resize: Current workload Current number of cloud nodes Policy (Python module) When more cloud nodes are needed (as determined by policy), more nodes are created in the cloud based on configured node properties When more nodes come online (~2-5m), Slurm will schedule jobs onto nodes 18

Workload Management in the Cloud Typical setup: one workload management queue per region Jobs that may run in the cloud should be submitted to one of the cloud queues Alternatively, cloud nodes and regular nodes can be combined in same queue. Cloud nodes also have workload management features which can be used as job-constraints Example: #!/bin/sh #SBATCH -J TestJob #SBATCH --ntasks=16 #SBATCH --constraint=us-east-1 Workload management system will schedule jobs onto cloud nodes the same way as on local nodes Nodes NFS mount /home and /cm/shared: Local nodes mount from head node Cloud nodes mount from a cloud director 19

Data Locality Problem Problem: Jobs usually require input data and produce output data Input and/or output data may require significant transfer time Resources charged by the hour, so input/output data should be transferred while resources are not yet allocated Data moving mechanics should be hidden from users as much as possible Solution: Bright introduces job submission utility cmsub which allows data dependencies of jobs to be made explicit in Slurm Useful for cloud, but can also be useful for e.g. Fetching data from tape archive Staging data to local compute nodes to overcome throughput limitations of parallel filesystem (needed for exascale) 20

Cloud Bursting Data-Aware Scheduling to the Cloud node006 node004 job node007 node005 data head node node001 node002 node003 1. User submits job to queue 2. Bright creates data-transfer job 3. Slurm runs compute job when data-transfer job is complete 4. Bright transfers output data back after completion

Slurm Job with Data Dependencies Example #!/bin/sh #SBATCH -J Data-Transfer-Test #SBATCH --ntasks=1 #CMSUB --input=/home/martijn/data-transfer-test/inputfile.txt #CMSUB --regions=eu-west-1 # Do the heavy work of reversing the lines tac inputfile.txt >outputfile-$slurm_job_id.txt # Schedule output file to be transferred back CM_SCHEDULE_TRANSFER(/home/martijn/data-transfertest/outputfile-$SLURM_JOB_ID.txt) echo Processed data on `hostname` 22

Data Aware Workload Management User submits job to workload management system using cmsub The cmsub utility will: Submit input data transfer job to Slurm Submit compute job with dependency on input transfer job Submit output data transfer job with dependency user job Data transfer jobs run on head node, so compute nodes need not be allocated while data is being transferred in/out of cloud Option to remove or keep data in the cloud after job completed Cmsub prevents multiple transfers of same data Partial data transfers are handled elegantly Users may also take responsibility for transferring data outside of cmsub 23

Future Directions Scheduling priorities of data transfers and compute jobs should be interdependent Order in which data should be transferred depends on: Estimated transfer time (data size, target location) Estimated job run time Job priority Resources requested by job Simple example: Job 1: run time: 1h input data: 10GB (10h) Job 2: run time: 10h input data: 1GB (1h) Naïve scheduling: 10h + 1h + 10h = 21h Optimal scheduling: 1h + 10h + 1h = 12h Making things worse: what about priority for output data? 24

Questions? Martijn de Vries martijn.devries@brightcomputing.com