Introduction to Distributed HTC and overlay systems

Similar documents
One Pool To Rule Them All The CMS HTCondor/glideinWMS Global Pool. D. Mason for CMS Software & Computing

High Throughput WAN Data Transfer with Hadoop-based Storage

CMS Grid Computing at TAMU Performance, Monitoring and Current Status of the Brazos Cluster

Shooting for the sky: Testing the limits of condor. HTCondor Week May 2015 Edgar Fajardo On behalf of OSG Software and Technology

glideinwms UCSD Condor tunning by Igor Sfiligoi (UCSD) UCSD Jan 18th 2012 Condor Tunning 1

HTCondor overview. by Igor Sfiligoi, Jeff Dost (UCSD)

(Refer Slide Time: 1:26)

Day 9: Introduction to CHTC

CMS experience of running glideinwms in High Availability mode

CLIENT ONBOARDING PLAN & SCRIPT

glideinwms architecture by Igor Sfiligoi, Jeff Dost (UCSD)

CLIENT ONBOARDING PLAN & SCRIPT

Who am I? I m a python developer who has been working on OpenStack since I currently work for Aptira, who do OpenStack, SDN, and orchestration

Deep Learning for Visual Computing Prof. Debdoot Sheet Department of Electrical Engineering Indian Institute of Technology, Kharagpur

SIMPLE FREEDOM: Follow Up Training Plan, Scripts, and Booklet

Lecture 1: January 22

Lesson 3 Transcript: Part 1 of 2 - Tools & Scripting

Singularity in CMS. Over a million containers served

Changing landscape of computing at BNL

Grid Computing Competence Center Large Scale Computing Infrastructures (MINF 4526 HS2011)

Fundamentals of Database Systems Prof. Arnab Bhattacharya Department of Computer Science and Engineering Indian Institute of Technology, Kanpur

Best practices to achieve optimal memory allocation and remote desktop user experience

Scientific Computing on Emerging Infrastructures. using HTCondor

HIGH ENERGY PHYSICS ON THE OSG. Brian Bockelman CCL Workshop, 2016

Lecture 1: January 23

Office 365 provided by Hugh Baird College

Getting Started Self Serve User Guide. For Getting Started Trainers

Application of Virtualization Technologies & CernVM. Benedikt Hegner CERN

Promoting Component Architectures in a Dysfunctional Organization

Backpacking with Code: Software Portability for DHTC Wednesday morning, 9:00 am Christina Koch Research Computing Facilitator

So, coming back to this picture where three levels of memory are shown namely cache, primary memory or main memory and back up memory.

Outline. ASP 2012 Grid School

Having said this, there are some workarounds you can use to integrate video into your course.

In our first lecture on sets and set theory, we introduced a bunch of new symbols and terminology.

/ Cloud Computing. Recitation 3 Sep 13 & 15, 2016

Hi everyone. I hope everyone had a good Fourth of July. Today we're going to be covering graph search. Now, whenever we bring up graph algorithms, we

OPERATING SYSTEMS. After A.S.Tanenbaum, Modern Operating Systems, 3rd edition. Uses content with permission from Assoc. Prof. Florin Fortis, PhD

Win-Back Campaign- Re-Engagement Series

Writing Cognitive Swift Apps developerworks Open Tech Talk March 8, 2017

Keeping Sane - Managing your

glideinwms Training Glidein Internals How they work and why by Igor Sfiligoi, Jeff Dost (UCSD) glideinwms Training Glidein internals 1

I'm Andy Glover and this is the Java Technical Series of. the developerworks podcasts. My guest is Brian Jakovich. He is the

VIMS Individual Users Guide. How to access your VIMS site, login and use all the features!

Using Remember the Milk with the Get Organized! System. Frank Buck. Consulting, Inc.

Out for Shopping-Understanding Linear Data Structures English

Building Campus HTC Sharing Infrastructures. Derek Weitzel University of Nebraska Lincoln (Open Science Grid Hat)

Scaling Without Sharding. Baron Schwartz Percona Inc Surge 2010

STAUNING Credit Application Internet Sales Process with /Voic Templates to Non-Responsive Prospects 2018 Edition

Data Structures and Algorithms Dr. Naveen Garg Department of Computer Science and Engineering Indian Institute of Technology, Delhi.

CMS Tier-2 Program for user Analysis Computing on the Open Science Grid Frank Würthwein UCSD Goals & Status

Embedded Systems Dr. Santanu Chaudhury Department of Electrical Engineering Indian Institute of Technology, Delhi

A Virtual Comet. HTCondor Week 2017 May Edgar Fajardo On behalf of OSG Software and Technology

Linux Local Security about Passwords and Data NZPAPER.BLOGSPOT.COM. Nz Paper Linux and Web Application Security. Zeeshan Khan 4/15/2013

T0-T1-T2 networking. Vancouver, 31 August 2009 LHCOPN T0-T1-T2 Working Group

Lecture 1: Overview

Four Steps to Unleashing The Full Potential of Your Database

MITOCW watch?v=w_-sx4vr53m

MITOCW watch?v=0jljzrnhwoi

glideinwms: Quick Facts

File System Performance (and Abstractions) Kevin Webb Swarthmore College April 5, 2018

UTILIZING THE NEW ALDA WEBSITE (CHAPTER LEADERS GROUP) PRESENTER: BRIAN JENSEN SEPTEMBER 16, 2016

Introducing: main Function, Comments, Statements

Interfacing HTCondor-CE with OpenStack: technical questions

4.1 Interval Scheduling

Persistence & State. SWE 432, Fall 2016 Design and Implementation of Software for the Web

Queuing and Scheduling on Compute Clusters

Q &A on Entity Relationship Diagrams. What is the Point? 1 Q&A

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback

In the following slides we will see some critical issues related to migration to BizTalk 2013.

For more info on Cloud9 see their documentation:

X Grid Engine. Where X stands for Oracle Univa Open Son of more to come...?!?

Sixth lecture; classes, objects, reference operator.

Weebly 101. Make an Affordable, Professional Website in Less than an Hour

Lesson 2. Introducing Apps. In this lesson, you ll unlock the true power of your computer by learning to use apps!

Embedded Linux Day 2

Data Intensive processing with irods and the middleware CiGri for the Whisper project Xavier Briand

Configuring a glideinwms factory

Large Scale Computing Infrastructures

Lecture 3: O/S Organization. plan: O/S organization processes isolation

FindResearch.org: How to Encourage Sharing of Research Artifacts

It s possible to get your inbox to zero and keep it there, even if you get hundreds of s a day.

For Volunteers An Elvanto Guide

Rsyslog: going up from 40K messages per second to 250K. Rainer Gerhards

An update on the scalability limits of the Condor batch system

The LHC Computing Grid

10601 Machine Learning. Model and feature selection

BBC LEARNING ENGLISH 6 Minute English Wireless furniture for phones

Audience: Info Workers, Dev

Optimizing Parallel Access to the BaBar Database System Using CORBA Servers

Using GPUaaS in Cloud Foundry

Servers & Developers. Julian Nadeau Production Engineer

How Do I Sync My Iphone To Another Computer Without Losing Everything

APPENDIX B: Questionnaire

Week - 01 Lecture - 04 Downloading and installing Python

How to Get Your Inbox to Zero Every Day

How to Get a Website Up and Running

How To Use My Alternative High

CS61A Lecture 28 Distributed Computing. Jom Magrotker UC Berkeley EECS August 6, 2012

Digital Marketing Manager, Marketing Manager, Agency Owner. Bachelors in Marketing, Advertising, Communications, or equivalent experience

Early X1 Experiences at Boeing. Jim Glidewell Information Technology Services Boeing Shared Services Group

Transcription:

Introduction to Distributed HTC and overlay systems Tuesday morning session Igor Sfiligoi <isfiligoi@ucsd.edu> University of California San Diego

Logistical reminder It is OK to ask questions - During the lecture - During the demos - During the exercises - During the breaks If I don't know the answer, I will find someone who likely does

High Throughput Computing Yesterday you were introduced to HTC - Often called batch system computing - A paradigm that emphasizes maximizing the amount of useful computing Batch slots over long periods of time Jobs Scheduler

Local HTC What you have really experienced so far is local HTC i.e. computing on dedicated cluster of dedicated resources - Managed by a single admin group - Co-located in a single location

Is there anything else? As you might expect, there are several insulated local HTC clusters installed around the world And there are non-htc systems out there, too

Is there anything else? As you might expect, there are several insulated local HTC clusters installed around the world And there are non-htc systems out there, too And you point is???

Just local HTC You moved from a single PC - O(1) cores To a local HTC cluster - Say, O(100) cores daily avg Great! I can have my results back in 1/100 th of the time

Just local HTC You moved from a single PC - O(1) cores To a local HTC cluster - Say, O(100) cores daily avg But is it fast enough? The result of the 10 body simulation is very promising! I want to run a 100 body one! It will still take me over a month to get the results back!

How to get more? If you find out that you are resource constrained - What do you do?

How to get more? If you find out that you are resource constrained - What do you do? Beg for a larger share of the local pool - i.e. better priority compared to the other users of the same pool There are likely 100s of users But you better be doing something really important

How to get more? If you find out that you are resource constrained - What do you do? Beg for a larger share of the local pool Pay to get more resources bought into the pool - Great for long term needs If you can afford it - But will not help you in the short term Getting new machines installed can take months!

How to get more? If you find out that you are resource constrained - What do you do? Beg for a larger share of the local pool Pay to get more resources bought Get the needed resources somewhere else - i.e. not locally

How to get more? If you find out that you are resource constrained - What do you do? Beg for a larger share of the local pool Pay to get more resources bought Get the needed resources somewhere else - i.e. not locally This is what this lecture is all about!

Distributed HTC A computing paradigm that aims at maximizing useful computation using any available resource located anywhere on the planet This place is huge! As a corollary - Compute resources are owned and operated by several independent groups

Implications of distributed computing We will be dealing with multiple independent compute systems - That do not know about each other They are owned by several independent groups, remember?

Implications of distributed computing We will be dealing with multiple independent compute systems - That do not know about each other No global HTC scheduler out of the box Scheduler

Implications of distributed computing We will be dealing with multiple independent compute systems - That do not know about each other No global HTC scheduler out of the box - Will have to stitch them all together

The naïve way The simplest way is to partition your jobs and submit a subset to each cluster Scheduler Scheduler Scheduler

The naïve way The simplest way is to partition your jobs and submit a subset to each cluster The drawbacks - You may need multiple user accounts - You may need several different tools - Doing a good partitioning is hard Want proportional to the resources you will get - And what if those CPUs are not in a HTC setup at all?

The naïve way The simplest way is to partition your jobs and submit a subset to each cluster The drawbacks - You may What's need multiple the user accounts - You may alternative??? need several different tools - Doing a good partitioning hard Want proportional to the resources you will get - And what if those CPUs are not in a HTC setup at all?

Use an overlay system Use a systems that looks and feels like a regular HTC to users - But has compute nodes all over the world I like the sound of this!

Use an overlay system Use a systems that looks and feels like a regular HTC to users - But has compute nodes all over the world But... didn't I like you the just say there was sound no such of this! thing???

Use an overlay system Use a systems that looks and feels like a regular HTC to users - But has compute nodes all over the world I said out of the box But one can create one. But... didn't you just say there was no such thing???

Creating a dynamic overlay sys No single person cannot manage all the existing resources Scheduler

Creating a dynamic overlay sys But we can lease a subset of them - We discuss the how later

Creating a dynamic overlay sys But we can lease a subset of them And instantiate a HTC system on them Scheduler

Creating a dynamic overlay sys But we can lease a subset of them And instantiate a HTC system on them - Now we can schedule user jobs on them - Only our resources are considered Jobs Scheduler

DHTC through an overlay sys Just another HTC system - Well, almost - More details in a few slides Cool! Jobs Scheduler

Overlay system ownership Setting up an overlay a major task - Comparable with installing a dedicated cluster Long term maintenance is also costly Not something a final user would want to do

Typical overlay sys operators Existing HTC admins, e.g. - The UW HTC cluster can overflow into OSG - The UCSD operates one for local users Scientific communities, e.g. - The CMS LHC experiment The Open Science Grid itself - With the OSG Connect More on this later today

Is DHTC really just HTC? Even with overlays, there are some differences between DHTC and HTC With or without overlays, the core reasons are: - Multiple independent HW operators - Not all resources are co-located

The multiple owners problem In the Grid world, the resource owner decides which Operating System, which OS services and which libraries to install - A way smaller problem in the Cloud world - But most of the current DHTC landscape is based on the Grid paradigm Different clusters likely configured differently

The multiple owners problem In the Grid world, the resource owner decides which Operating System, which OS services and which libraries to install - A way smaller problem in the Cloud world - But most of the current DHTC landscape is based on the Grid paradigm Different clusters likely configured differently Even if you could get your pet library/service installed on some clusters, you cannot expect to get it installed everywhere

Heterogeneity DHTC systems are way more heterogeneous than local HTC ones Two ways to approach this: - Minimize external dependencies in your compute jobs Make them self-contained Adapt to the running environment (e.g. multiple binaries, one per platform) Do not use licensed software

Heterogeneity DHTC systems are way more heterogeneous than local HTC ones Two ways to approach this: - Minimize external dependencies in your compute jobs Make them self-contained Sounds like a lot of work! Adapt to the running environment (e.g. multiple binaries, one per platform) Do not use licensed software

Heterogeneity DHTC systems are way more heterogeneous than local HTC ones Two ways to approach this: - Minimize external dependencies - Use only a subset of the resources Restrict where your jobs can run Your job will of course take longer to finish May still get you the result sooner

Heterogeneity DHTC systems are way more heterogeneous than local HTC ones Two ways to approach this: - Minimize external dependencies - Use only a subset of the resources Restrict where your jobs can run Your job will of course take longer to finish So long for May still get you the result DHTC! sooner

Heterogeneity DHTC systems are way more heterogeneous than local HTC ones Two ways to approach this: - Minimize external dependencies - Use only a subset of the resources Restrict where your jobs can run Unfortunately, Your those job will are of the course only two take longer to finish So long for May still alternatives. get you the result DHTC! sooner

No shared file system As a side effect, you cannot expect a globally shared file system - It's just yet another user requested service You will have to deal with data explicitly - Either using the HTC scheduler capabilities (remember yesterday's lecture) - Or, embed file transfer to and from permanent storage into your own jobs More details tomorrow

The location problem Nodes in different locations need a way to talk to each other - This is what networks are for If your computation is mostly about CPU cycles, with little input and output data - Node location is not an issue at all If you have lots of data - Remember, throughput is typically inversely proportional with the distance

The data problem Transferring large amounts of data over Wide Area Network can take a lot of time You should try to compute close to the data source and/or sink - Network-wise - More about this tomorrow

Bottom line For simple computation, DHTC is very similar to HTC As soon as you require any complex setup for your jobs, you are in for a rough ride - This includes large datasets

Infrastructure considerations DHTC is likely to give you access to many more compute slots Which is mostly a good thing - You get your results faster But could crash your HTC system, e.g. - Can your storage system handle more data traffic? - Can the job scheduling system handle an order of magnitude more nodes?

Infrastructure considerations DHTC is likely to give you access to many more compute slots Which is mostly a good thing - You get your results faster But could crash your HTC system, e.g. - Can your Hopefully storage not something system handle more data traffic? final users should deal with but it is good to keep in mind. - Can the job scheduling system handle an order of magnitude more nodes?

Questions? Questions? Comments? - Feel free to ask me questions later: Igor Sfiligoi <isfiligoi@ucsd.edu> Upcoming sessions - Where to get the needed resources - glideinwms the OSG overlay software - Hands-on exercises - Tour

Beware the power Courtesy of fanpop.com

Copyright statement Some images contained in this presentation are the copyrighted property of ToonClipart. As such, these images are being used under a license from said entities, and may not be copied or downloaded without explicit permission from ToonClipart.