STORK: Making Data Placement a First Class Citizen in the Grid

Similar documents
Stork: State of the Art

A Fully Automated Faulttolerant. Distributed Video Processing and Off site Replication

Condor-G and DAGMan An Introduction

A Framework for Reliable and Efficient Data Placement in Distributed Computing Systems

Condor-G Stork and DAGMan An Introduction

George Kola. A position in industry or research lab that involves systems design, performance analysis and systems development.

Profiling Grid Data Transfer Protocols and Servers. George Kola, Tevfik Kosar and Miron Livny University of Wisconsin-Madison USA

DiskRouter: A Flexible Infrastructure for High Performance Large Scale Data Transfers

Data Management 1. Grid data management. Different sources of data. Sensors Analytic equipment Measurement tools and devices

Grid Compute Resources and Grid Job Management

Pegasus WMS Automated Data Management in Shared and Nonshared Environments

DATA PLACEMENT IN WIDELY DISTRIBUTED SYSTEMS. by Tevfik Kosar. A dissertation submitted in partial fulfillment of the requirements for the degree of

Pegasus Workflow Management System. Gideon Juve. USC Informa3on Sciences Ins3tute

Grid Compute Resources and Job Management

The Problem of Grid Scheduling

What makes workflows work in an opportunistic environment?

Phoenix: Making Data-intensive Grid Applications Fault-tolerant

THE GLOBUS PROJECT. White Paper. GridFTP. Universal Data Transfer for the Grid

Profiling Grid Data Transfer Protocols and Servers

High Throughput Urgent Computing

Condor and Workflows: An Introduction. Condor Week 2011

The Google File System

The glite File Transfer Service

Part IV. Workflow Mapping and Execution in Pegasus. (Thanks to Ewa Deelman)

The Google File System

Stork Data Scheduler: Mitigating the Data Bottleneck in e-science

Distributed and Adaptive Execution of Condor DAGMan Workflows

A New Paradigm in Data Intensive Computing: Stork and the Data-Aware Schedulers

Introduction to Grid Computing

Grid Programming: Concepts and Challenges. Michael Rokitka CSE510B 10/2007

Sphinx: A Scheduling Middleware for Data Intensive Applications on a Grid

PROOF-Condor integration for ATLAS

Grid Data Management

Production Grids. Outline

Map-Reduce. Marco Mura 2010 March, 31th

Data Staging: Moving large amounts of data around, and moving it close to compute resources

GFS: The Google File System. Dr. Yingwu Zhu

Li Yu. University of Notre Dame

Day 1 : August (Thursday) An overview of Globus Toolkit 2.4

Grid Computing. Lectured by: Dr. Pham Tran Vu Faculty of Computer and Engineering HCMC University of Technology

Outline. ASP 2012 Grid School

The Google File System. Alexandru Costan

High Performance Computing Course Notes Grid Computing I

Parallel Computing: MapReduce Jin, Hai

Reduction of Workflow Resource Consumption Using a Density-based Clustering Model

Outline. INF3190:Distributed Systems - Examples. Last week: Definitions Transparencies Challenges&pitfalls Architecturalstyles

UW-ATLAS Experiences with Condor

SPHINX: A Fault-Tolerant System for Scheduling in Dynamic Grid Environments

GFS: The Google File System

Data Staging: Moving large amounts of data around, and moving it close to compute resources

Chapter 4:- Introduction to Grid and its Evolution. Prepared By:- NITIN PANDYA Assistant Professor SVBIT.

Distributed Systems CS6421

User Tools and Languages for Graph-based Grid Workflows

Heterogeneous Grid Computing: Issues and Early Benchmarks

Increasing Performance for PowerCenter Sessions that Use Partitions

WMS overview and Proposal for Job Status

Automating Real-time Seismic Analysis

Managing large-scale workflows with Pegasus

An update on the scalability limits of the Condor batch system

High Throughput Computing. Miron Livny Computer Sciences Department University of Wisconsin-Madison

Supporting Application- Tailored Grid File System Sessions with WSRF-Based Services

Error detection and error classification: failure awareness in data transfer scheduling

XSEDE High Throughput Computing Use Cases

Announcements (March 1) Query Processing: A Systems View. Physical (execution) plan. Announcements (March 3) Physical plan execution

Pegasus. Automate, recover, and debug scientific computations. Rafael Ferreira da Silva.

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

Midterm Exam. October 20th, Thursday NSC

EFFICIENT ALLOCATION OF DYNAMIC RESOURCES IN A CLOUD

GLADE: A Scalable Framework for Efficient Analytics. Florin Rusu (University of California, Merced) Alin Dobra (University of Florida)

Aggregation and Degradation in JetStream: Streaming analytics in the wide area

CIS Operating Systems Contiguous Memory Allocation. Professor Qiang Zeng Spring 2018

Effective Data Transfers through Parallelism in Cloud Networks

Ian Foster, CS554: Data-Intensive Computing

TreeSearch User Guide

Workflow, Planning and Performance Information, information, information Dr Andrew Stephen M c Gough

Deploying virtualisation in a production grid

Current Topics in OS Research. So, what s hot?

Flat Datacenter Storage. Edmund B. Nightingale, Jeremy Elson, et al. 6.S897

Day 9: Introduction to CHTC

CSE Traditional Operating Systems deal with typical system software designed to be:

A Federated Grid Environment with Replication Services

The Google File System

Process Migration via Remote Fork: a Viable Programming Model? Branden J. Moor! cse 598z: Distributed Systems December 02, 2004

EFFICIENT SCHEDULING TECHNIQUES AND SYSTEMS FOR GRID COMPUTING

Paxos Replicated State Machines as the Basis of a High- Performance Data Store

Cloud Computing. Up until now

Overview of ATLAS PanDA Workload Management

A BigData Tour HDFS, Ceph and MapReduce

<Insert Picture Here> QCon: London 2009 Data Grid Design Patterns

BOSCO Architecture. Derek Weitzel University of Nebraska Lincoln

OpenSFS Test Cluster Donation. Dr. Nihat Altiparmak, Assistant Professor Computer Engineering & Computer Science Department University of Louisville

Grid Scheduling Architectures with Globus

CLOUD-SCALE FILE SYSTEMS

Join Processing for Flash SSDs: Remembering Past Lessons

ECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective

CCB The Condor Connection Broker. Dan Bradley Condor Project CS and Physics Departments University of Wisconsin-Madison

Cloud Computing. Summary

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

Understanding StoRM: from introduction to internals

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

Transcription:

STORK: Making Data Placement a First Class Citizen in the Grid Tevfik Kosar University of Wisconsin-Madison May 25 th, 2004 CERN

Need to move data around.. TB PB TB PB

While doing this.. Locate the data Access heterogeneous resources Face with all kinds of failures Allocate and de-allocate storage Move the data Clean-up everything All of these need to be done reliably and efficiently!

Stork A scheduler for data placement activities in the Grid What Condor is for computational jobs, Stork is for data placement Stork comes with a new concept: Make data placement a first class citizen in the Grid.

Outline Introduction The Concept Stork Features Big Picture Case Studies Conclusions

The Concept Stage-in Execute the Job Stage-out Individual Jobs

The Concept Allocate space for input & output data Stage-in Stage-in Execute the Job Stage-out Execute the job Release input space Stage-out Individual Jobs Release output space

The Concept Allocate space for input & output data Stage-in Stage-in Execute the Job Stage-out Execute the job Release input space Stage-out Data Placement Jobs Computational Jobs Release output space

The Concept DAG specification DaP A A.submit DaP B B.submit Job C C.submit.. Parent A child B Parent B child C Parent C child D, E.. A B DAGMan C D E F C E Condor Job Queue Stork Job Queue

Why Stork? Stork understands the characteristics and semantics of data placement jobs. Can make smart scheduling decisions, for reliable and efficient data placement.

Understanding Job Characteristics & Semantics Job_type = transfer, reserve, release? Source and destination hosts, files, protocols to use? hdetermine concurrency level hcan select alternate protocols hcan select alternate routes hcan tune network parameters (tcp buffer size, I/O block size, # of parallel streams) h

Support for Heterogeneity Protocol translation using Stork memory buffer.

Support for Heterogeneity Protocol translation using Stork Disk Cache.

Flexible Job Representation and Multilevel Policy Support [ ] Type = Transfer ; Src_Url = srb://ghidorac.sdsc.edu/kosart.condor/x.dat ; Dest_Url = nest://turkey.cs.wisc.edu/kosart/x.dat ; Max_Retry = 10; Restart_in = 2 hours ;

Failure Recovery and Efficient Resource Utilization Fault tolerance hjust submit a bunch of data placement jobs, and then go away.. Control number of concurrent transfers from/to any storage system hprevents overloading Space allocation and De-allocations hmake sure space is available

Run-time Adaptation Dynamic protocol selection [ dap_type = transfer ; src_url = drouter://slic04.sdsc.edu/tmp/test.dat ; dest_url = drouter://quest2.ncsa.uiuc.edu/tmp/test.dat ; alt_protocols = nest-nest, gsiftp-gsiftp ; ] [ ] dap_type = transfer ; src_url = any://slic04.sdsc.edu/tmp/test.dat ; dest_url = any://quest2.ncsa.uiuc.edu/tmp/test.dat ;

Run-time Adaptation Run-time Protocol Auto-tuning [ link = slic04.sdsc.edu quest2.ncsa.uiuc.edu ; protocol = gsiftp ; ] bs = 1024KB; //block size tcp_bs = 1024KB; //TCP buffer size p = 4;

Outline Introduction The Concept Stork Features Big Picture Case Studies Conclusions

PLANNER Abstract DAG JOB DESCRIPTIONS USER

RLS PLANNER Abstract DAG JOB DESCRIPTIONS USER Concrete DAG WORKFLOW MANAGER

RLS PLANNER Abstract DAG JOB DESCRIPTIONS USER Concrete DAG WORKFLOW MANAGER COMPUTATION SCHEDULER DATA PLACEMENT SCHEDULER STORAGE SYSTEMS COMPUTE NODES

RLS PLANNER Abstract DAG JOB DESCRIPTIONS USER Concrete DAG WORKFLOW MANAGER COMPUTATION SCHEDULER DATA PLACEMENT SCHEDULER POLICY ENFORCER COMPUTE NODES C. JOB LOG FILES D. JOB LOG FILES STORAGE SYSTEMS

RLS PLANNER Abstract DAG JOB DESCRIPTIONS USER Concrete DAG WORKFLOW MANAGER COMPUTATION SCHEDULER DATA PLACEMENT SCHEDULER POLICY ENFORCER COMPUTE NODES C. JOB LOG FILES D. JOB LOG FILES STORAGE SYSTEMS NETWORK MONITORING TOOLS FEEDBACK MECHANISM DATA MINER

RLS PEGASUS Abstract DAG JOB DESCRIPTIONS USER Concrete DAG DAGMAN CONDOR/ CONDOR-G STORK MATCHMAKER COMPUTE NODES C. JOB LOG FILES D. JOB LOG FILES STORAGE SYSTEMS NETWORK MONITORING TOOLS FEEDBACK MECHANISM DATA MINER

Outline Introduction The Concept Stork Features Big Picture Case Studies Conclusions

Case Study I: SRB-UniTree Data Pipeline Transfer ~3 TB of DPOSS data from SRB @SDSC to UniTree @NCSA SRB Server Submit Site UniTree Server A data transfer pipeline created with Stork SDSC Cache NCSA Cache

Failure Recovery UniTree not responding Diskrouter reconfigured and restarted SDSC cache reboot & UW CS Network outage Software problem

Case Study -II

Dynamic Protocol Selection

Runtime Adaptation Before Tuning: parallelism = 1 block_size = 1 MB tcp_bs = 64 KB After Tuning: parallelism = 4 block_size = 1 MB tcp_bs = 256 KB

Case Study -III 4 3 Split files 2 1 DiskRouter/ Globus-url-copy Condor Pool @UW 5 Condor File Transfer Mechanism Staging Site @UW 7 6 User submits a DAG at management site Management Site @UW Merge files WCER SRB put 8 SRB Server @SDSC Control flow Input Data flow Output Data flow Processing DiskRouter/ Globus-url-copy Other Condor Stork: Making Data Pools Placement a First Class Citizen in the Grid Other Replicas

Conclusions Regard data placement as individual jobs. Treat computational and data placement jobs differently. Introduce a specialized scheduler for data placement. Provide end-to-end automation, fault tolerance, run-time adaptation, multilevel policy support, reliable and efficient transfers.

Future work Enhanced interaction between Stork and higher level planners hbetter coordination of CPU and I/O Interaction between multiple Stork servers and job delegation Enhanced authentication mechanisms More run-time adaptation

Related Publications Tevfik Kosar and Miron Livny. Stork: Making Data Placement a First Class Citizen in the Grid. In Proceedings of 24 th IEEE Int. Conference on Distributed Computing Systems (ICDCS 2004), Tokyo, Japan, March 2004. George Kola, Tevfik Kosar and Miron Livny. A Fully Automated Faulttolerant System for Distributed Video Processing and Off-site Replication. To appear in Proceedings of 14 th ACM Int. Workshop on etwork and Operating Systems Support for Digital Audio and Video (Nossdav 2004), Kinsale, Ireland, June 2004. Tevfik Kosar, George Kola and Miron Livny. A Framework for Selfoptimizing, Fault-tolerant, High Performance Bulk Data Transfers in a Heterogeneous Grid Environment. In Proceedings of 2 nd Int. Symposium on Parallel and Distributed Computing (ISPDC 2003), Ljubljana, Slovenia, October 2003. George Kola, Tevfik Kosar and Miron Livny. Run-time Adaptation of Grid Data Placement Jobs. In Proceedings of Int. Workshop on Adaptive Grid Middleware (AGridM 2003), New Orleans, LA, September 2003.

You don t have to FedEx your data anymore.. Stork delivers it for you! For more information: Email: kosart@cs.wisc.edu http://www.cs.wisc.edu/condor/stork