Slurm Roadmap. Morris Jette, Danny Auble (SchedMD) Yiannis Georgiou (Bull)

Similar documents
Versions and 14.11

Directions in Workload Management

1 Bull, 2011 Bull Extreme Computing

Slurm Overview. Brian Christiansen, Marshall Garey, Isaac Hartung SchedMD SC17. Copyright 2017 SchedMD LLC

Scheduling By Trackable Resources

SLURM Operation on Cray XT and XE

Slurm Roadmap. Danny Auble, Morris Jette, Tim Wickberg SchedMD. Slurm User Group Meeting Copyright 2017 SchedMD LLC

Introduction to Slurm

Slurm Workload Manager Overview SC15

Slurm BOF SC13 Bull s Slurm roadmap

Resource Management at LLNL SLURM Version 1.2

High Throughput Computing with SLURM. SLURM User Group Meeting October 9-10, 2012 Barcelona, Spain

Slurm Birds of a Feather

Exascale Process Management Interface

Heterogeneous Job Support

Federated Cluster Support

High Scalability Resource Management with SLURM Supercomputing 2008 November 2008

Tutorial 3: Cgroups Support On SLURM

Hosts & Partitions. Slurm Training 15. Jordi Blasco & Alfred Gil (HPCNow!)

Native SLURM on Cray XC30. SLURM Birds of a Feather SC 13

Scheduler Optimization for Current Generation Cray Systems

Towards Exascale: Leveraging InfiniBand to accelerate the performance and scalability of Slurm jobstart.

Slurm Burst Buffer Support

Slurm Workload Manager Introductory User Training

Extending SLURM with Support for GPU Ranges

Workload Managers. A Flexible Approach

Slurm Version Overview

Resource Management with Linux Control Groups in HPC Clusters. Yiannis Georgiou

Executing dynamic heterogeneous workloads on Blue Waters with RADICAL-Pilot

EReinit: Scalable and Efficient Fault-Tolerance for Bulk-Synchronous MPI Applications

NERSC Site Report One year of Slurm Douglas Jacobsen NERSC. SLURM User Group 2016

Spark Over RDMA: Accelerate Big Data SC Asia 2018 Ido Shamay Mellanox Technologies

MDHIM: A Parallel Key/Value Store Framework for HPC

Habanero Operating Committee. January

Slurm at CEA. status and evolutions. 13 septembre 2013 CEA 10 AVRIL 2012 PAGE 1. SLURM User Group - September 2013 F. Belot, F. Diakhaté, M.

CEA Site Report. SLURM User Group Meeting 2012 Matthieu Hautreux 26 septembre 2012 CEA 10 AVRIL 2012 PAGE 1

Batch Usage on JURECA Introduction to Slurm. May 2016 Chrysovalantis Paschoulas HPS JSC

Big Data 7. Resource Management

Introduction to SLURM on the High Performance Cluster at the Center for Computational Research

Paving the Road to Exascale

Slurm Support for Linux Control Groups

Using GPUaaS in Cloud Foundry

Enhancing an Open Source Resource Manager with Multi-Core/Multi-threaded Support

Cray Operating System Plans and Status. Charlie Carroll May 2012

Slurm Configuration Impact on Benchmarking

cli_filter command line filtration, manipulation, and introspection of job submissions

The Road to ExaScale. Advances in High-Performance Interconnect Infrastructure. September 2011

MediaTek CorePilot 2.0. Delivering extreme compute performance with maximum power efficiency

OVERHEADS ENHANCEMENT IN MUTIPLE PROCESSING SYSTEMS BY ANURAG REDDY GANKAT KARTHIK REDDY AKKATI

Deployment Planning Guide

Brigham Young University

Big Data for Engineers Spring Resource Management

Write a technical report Present your results Write a workshop/conference paper (optional) Could be a real system, simulation and/or theoretical

Brigham Young University Fulton Supercomputing Lab. Ryan Cox

TensorFlow: A System for Learning-Scale Machine Learning. Google Brain

Distributed Computing: PVM, MPI, and MOSIX. Multiple Processor Systems. Dr. Shaaban. Judd E.N. Jenne

Quobyte The Data Center File System QUOBYTE INC.

Accelerator programming with OpenACC

Construct a sharable GPU farm for Data Scientist. Layne Peng DellEMC OCTO TRIGr

Lecture 11 Hadoop & Spark

Adaptive Cluster Computing using JavaSpaces

Chapter 5. The MapReduce Programming Model and Implementation

Philip C. Roth. Computer Science and Mathematics Division Oak Ridge National Laboratory

Managing and Deploying GPU Accelerators. ADAC17 - Resource Management Stephane Thiell and Kilian Cavalotti Stanford Research Computing Center

Testing SLURM open source batch system for a Tierl/Tier2 HEP computing facility

Users and utilization of CERIT-SC infrastructure

SEMem: deployment of MPI-based in-memory storage for Hadoop on supercomputers

Deploying SLURM on XT, XE, and Future Cray Systems

Power Constrained HPC

CPU-GPU Heterogeneous Computing

In-Network Computing. Paving the Road to Exascale. 5th Annual MVAPICH User Group (MUG) Meeting, August 2017

TotalView Release Notes

Slurm Inter-Cluster Project. Stephen Trofinoff CSCS Via Trevano 131 CH-6900 Lugano 24-September-2014

ECE 574 Cluster Computing Lecture 23

Future Routing Schemes in Petascale clusters

How to Boost the Performance of Your MPI and PGAS Applications with MVAPICH2 Libraries

MOHA: Many-Task Computing Framework on Hadoop

Cray XC Scalability and the Aries Network Tony Ford

Best Practices for Setting BIOS Parameters for Performance

Leveraging Flash in HPC Systems

A Platform for Fine-Grained Resource Sharing in the Data Center

Birds of a Feather Presentation

IO virtualization. Michael Kagan Mellanox Technologies

SLURM Event Handling. Bill Brophy

Introduction & Motivation Problem Statement Proposed Work Evaluation Conclusions Future Work

TITLE: PRE-REQUISITE THEORY. 1. Introduction to Hadoop. 2. Cluster. Implement sort algorithm and run it using HADOOP

Porting SLURM to the Cray XT and XE. Neil Stringfellow and Gerrit Renker

Designing and Building Efficient HPC Cloud with Modern Networking Technologies on Heterogeneous HPC Clusters

Assignment 5. Georgia Koloniari

Cycle Sharing Systems

Interconnect Your Future

SLURM Simulator improvements and evaluation

FUSION PROCESSORS AND HPC

Apache Hadoop 3. Balazs Gaspar Sales Engineer CEE & CIS Cloudera, Inc. All rights reserved.

Chelsio Communications. Meeting Today s Datacenter Challenges. Produced by Tabor Custom Publishing in conjunction with: CUSTOM PUBLISHING

Using MySQL in a Virtualized Environment. Scott Seighman Systems Engineer Sun Microsystems

GlusterFS Current Features & Roadmap

IRIX Resource Management Plans & Status

Paving the Road to Exascale Computing. Yossi Avni

What's New in SUSE LINUX Enterprise Server 9

Transcription:

Slurm Roadmap Morris Jette, Danny Auble (SchedMD) Yiannis Georgiou (Bull)

Exascale Focus Heterogeneous Environment Scalability Reliability Energy Efficiency New models (Cloud/Virtualization/Hadoop)

Following Releases Agile development with two major releases per year plus periodic bug fixes Switching to Ubuntu version number scheme <year>.<month> 13.12 == December 2013 14.06 == June 2014 14.12 == December 2014

Upcoming Release 13.12

Scalability Optimize MPI Network Use New switch/generic plugin Captures network interface configuration on each node (name and IP address) Provides job with network interface details about allocated nodes using PMI2 Step/Job Epilog Hierarchical Communications

Scalability Scheduling New partition configuration parameters: AllowAccounts, AllowQOS, DenyAccounts, DenyQOS Added load-based scheduling Allocate resources on least loaded node first Can improve serial job performance, but maximizes fragmentation

Reliability Failure Management Configurable hot-spare nodes Running jobs can replace or remove failed or failing nodes Jobs can extend time limit based upon failures Jobs can drain nodes they perceive to be failing Configurable access control lists and limits

Energy Efficiency Energy Use Optimizations Integration of Layout framework, which will be the basis for optimizing resource management, job placement and scheduling based on resources characteristics Power cap enforcement and considerations in job placement Reservations of Power cap Job energy consumption as a new factor in fair-share scheduling

New Models (Cloud/Virtualization/Hadoop) Hadoop Integration Work being performed by Intel Eliminates need for dedicated Hadoop cluster Better scalability Launch: Hadoop/YARN (~N), Slurm (~log N) Wireup: Hadoop/YARN (~N2 ), Slurm (~log N) No modifications to Hadoop Completely transparent to existing applications

New Models (Cloud/Virtualization/Hadoop) Slurm-Hadoop Architecture Java Daemon Intercepts calls from client and translates for Slurm Slurmctld plugin Accepts and processes RM requests Web interface for job/system status Slurmd plugin Spawn local processes, support shuffle and other ops

New Models (Cloud/Virtualization/Hadoop) Slurm-Hadoop Status Under development Tentative schedule Demo at SC13 Release 4Q 2013 or 1Q 2014 Slurm plugins in version 13.12 Translator daemon will be available from Intel

Other Features Licenses Management in Accounting and support of FlexLM (Flexnet Publisher) Stable version of Jobacct_gather/cgroup plugin Support of PAM with cgroups Improved sview scalability Added job_container plugin infrastructure Improved integration with Cray systems

Future Release 14.06

Heterogeneous Environment Heterogeneous resources Distributed architecture to support the management of resources with MIC Lightweight slurmd upon MIC Router capabilities on slurmd based on static and dynamic trees Support of I/O as new resource along with data locality Extension on the Job requirements description to support heterogeneous resources

Heterogeneous Environment Improved GPU and MIC Support Support for heterogeneous GPUs salloc --gres=gpu:1 salloc --gres=gpu:tesla:1,gpu... Use MIC to offload work from CPUs or as independent compute node

Scalability Scalability and scheduling Scalability optimizations for MPI initialization Improved scheduling support for job dependencies (e.g. pre-processing, post-processing, co-processing on I/O nodes, etc.) to optimize overall system utilization

Reliability Failure Management Distribute hot spare resources through system Optimize job s replacement resources with respect to network latency

Energy Efficiency Energy Use Optimizations Work likely to continue for several years Finer grain Power Management Optimize system throughput with respect to varying power caps Consider DVFS or other models in conjunction with power caps Limit rate of change in system power consumption

Release 14.12 and beyond

Directions Multi-parameter scheduling based on the layout framework Fault-tolerance and jobs dynamic adaptation through communication protocol between Slurm, MPI libraries and the application Network Communication Scalability optimizations