WMS overview and Proposal for Job Status

Similar documents
Grid services. Enabling Grids for E-sciencE. Dusan Vudragovic Scientific Computing Laboratory Institute of Physics Belgrade, Serbia

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

AliEn Resource Brokers

DataGrid. Document identifier: Date: 24/11/2003. Work package: Partner: Document status. Deliverable identifier:

LCG-2 and glite Architecture and components

Ganga - a job management and optimisation tool. A. Maier CERN

Advanced Job Submission on the Grid

glite Grid Services Overview

Workload Management. Stefano Lacaprara. CMS Physics Week, FNAL, 12/16 April Department of Physics INFN and University of Padova

Ganga The Job Submission Tool. WeiLong Ueng

DIRAC pilot framework and the DIRAC Workload Management System

International Collaboration to Extend and Advance Grid Education. glite WMS Workload Management System

Architecture of the WMS

Distributed Computing Framework. A. Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille

HEP replica management

Architecture Proposal

DIRAC Distributed Infrastructure with Remote Agent Control

CMS Tier-2 Program for user Analysis Computing on the Open Science Grid Frank Würthwein UCSD Goals & Status

Grid Infrastructure For Collaborative High Performance Scientific Computing

DataGrid D EFINITION OF ARCHITECTURE, TECHNICAL PLAN AND EVALUATION CRITERIA FOR SCHEDULING, RESOURCE MANAGEMENT, SECURITY AND JOB DESCRIPTION

Grid Compute Resources and Job Management

Cloud Computing. Up until now

Monitoring the Usage of the ZEUS Analysis Grid

Introduction to Grid Infrastructures

DIRAC: A Scalable Lightweight Architecture for High Throughput Computing

Gergely Sipos MTA SZTAKI

The glite middleware. Ariel Garcia KIT

DIRAC Distributed Infrastructure with Remote Agent Control

NUSGRID a computational grid at NUS

Problemi di schedulazione distribuita su Grid

DataGrid. Document identifier: Date: 16/06/2003. Work package: Partner: Document status. Deliverable identifier:

Improved 3G Bridge scalability to support desktop grid executions

OSGMM and ReSS Matchmaking on OSG

DIRAC Documentation. Release integration. DIRAC Project. 09:29 20/05/2016 UTC

Grid Compute Resources and Grid Job Management

The University of Oxford campus grid, expansion and integrating new partners. Dr. David Wallom Technical Manager

A Practical Approach for a Workflow Management System

Overview of HEP software & LCG from the openlab perspective

PoS(ACAT2010)039. First sights on a non-grid end-user analysis model on Grid Infrastructure. Roberto Santinelli. Fabrizio Furano.

Chapter 3. Design of Grid Scheduler. 3.1 Introduction

EUROPEAN MIDDLEWARE INITIATIVE

The Problem of Grid Scheduling

RELEASE OF GANGA. Basics and organisation What Ganga should do tomorrow Ganga design What Ganga will do today Next steps

DIRAC File Replica and Metadata Catalog

Resource Allocation in computational Grids

GRID COMPANION GUIDE

g-eclipse A Framework for Accessing Grid Infrastructures Nicholas Loulloudes Trainer, University of Cyprus (loulloudes.n_at_cs.ucy.ac.

ALICE Grid/Analysis Tutorial Exercise-Solutions

Logic Networks on the Grid: Handling 15 Million Jobs

2/26/2017. For instance, consider running Word Count across 20 splits

Grid Scheduling Architectures with Globus

DataGrid TECHNICAL PLAN AND EVALUATION CRITERIA FOR THE RESOURCE CO- ALLOCATION FRAMEWORK AND MECHANISMS FOR PARALLEL JOB PARTITIONING

ATLAS Production System in ATLAS Data Challenge 2 Luc Goossens (CERN/EP/ATC) Kaushik De (UTA)

LHCb Computing Strategy

AGATA Analysis on the GRID

Look What I Can Do: Unorthodox Uses of HTCondor in the Open Science Grid

Grid Architectural Models

Status of the DIRAC Project

Distributed production managers meeting. Armando Fella on behalf of Italian distributed computing group

DIRAC: a community grid solution

Tutorial for CMS Users: Data Analysis on the Grid with CRAB

Parallel Computing in EGI

CERN: LSF and HTCondor Batch Services

On the employment of LCG GRID middleware

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

Analisi Tier2 e Tier3 Esperienze ai Tier-2 Giacinto Donvito INFN-BARI

Sphinx: A Scheduling Middleware for Data Intensive Applications on a Grid

Grid Interoperation and Regional Collaboration

A Simulation Model for Large Scale Distributed Systems

DataGrid. Document identifier: Date: 28/10/2003. Work package: Partner: Document status. Deliverable identifier:

Bookkeeping and submission tools prototype. L. Tomassetti on behalf of distributed computing group

Tutorial 4: Condor. John Watt, National e-science Centre

Philippe Charpentier PH Department CERN, Geneva

The glite middleware. Presented by John White EGEE-II JRA1 Dep. Manager On behalf of JRA1 Enabling Grids for E-sciencE

Tier2 Centre in Prague

Edinburgh (ECDF) Update

CREAM-WMS Integration

Michigan Grid Research and Infrastructure Development (MGRID)

DTM a lightweight computing virtualization system based on irods. Yonny Cardenas, Pascal Calvat, Jean-Yves Nief, Thomas Kachelhoffer

GRID COMPUTING APPLIED TO OFF-LINE AGATA DATA PROCESSING. 2nd EGAN School, December 2012, GSI Darmstadt, Germany

Grid Programming: Concepts and Challenges. Michael Rokitka CSE510B 10/2007

Dynamic fault tolerant grid workflow in the water threat management project

AutoPyFactory: A Scalable Flexible Pilot Factory Implementation

OSG Lessons Learned and Best Practices. Steven Timm, Fermilab OSG Consortium August 21, 2006 Site and Fabric Parallel Session

Distributed Monte Carlo Production for

M. Roehrig, Sandia National Laboratories. Philipp Wieder, Research Centre Jülich Nov 2002

Troubleshooting Grid authentication from the client side

The ARDA project: Grid analysis prototypes of the LHC experiments

Jozef Cernak, Marek Kocan, Eva Cernakova (P. J. Safarik University in Kosice, Kosice, Slovak Republic)

Client tools know everything

Deploying virtualisation in a production grid

CMS HLT production using Grid tools

Monitoring System for the GRID Monte Carlo Mass Production in the H1 Experiment at DESY

LHCbDirac: distributed computing in LHCb

glite Middleware Usage

Dr. Giuliano Taffoni INAF - OATS

Implementing GRID interoperability

Setup Desktop Grids and Bridges. Tutorial. Robert Lovas, MTA SZTAKI

New Directions and BNL

Grid Computing. Olivier Dadoun LAL, Orsay. Introduction & Parachute method. Socle 2006 Clermont-Ferrand Orsay)

Transcription:

WMS overview and Proposal for Job Status Author: V.Garonne, I.Stokes-Rees, A. Tsaregorodtsev. Centre de physiques des Particules de Marseille Date: 15/12/2003 Abstract In this paper, we describe briefly the workload Management System (WMS) of DIRAC (Distributed Infrastructure with Remote Agent Control). Dirac will be the infrastructure used for this year s DC 04 production. This document also discusses the different states of a Job during its lifecycle and the job parameters required by the system. WMS Architecture Overview Figure 2. WMS Architecture The WMS consists of the following key components: User where jobs originally come from and where results are returned. The User submits jobs in JDL. JDL is used to maintain interoperability with others grid projects such EDG, 1

Alien and LCG. Job receiver Job DB Optimizer Match Maker Agent Computing Element accepts job submissions from the User and registers them into the Job DB. stores information pertaining to each job. sorts jobs into Grid queues with a priority computed by optimizer. decides what jobs to give to a particular computing resource Monitors a Computing Element (CE) and fetches jobs from the Match Maker provides computational resources to the user (see also Computing Element chapter) Computing Element The Computing Element (CE) is an abstraction which provides an interface to local computing resources. Although single node and grid Computing Elements are possible, in general we model each Computing Element as a single Gatekeeper which manages/accesses a cluster of computing Worker Nodes. We assume such a system consists of its own local scheduler and queues (as illustrated in Figure 2). At present, DIRAC provides interfaces to LSF, PBS, BQS and Globus. Figure 1: Computing Element A CE is classified as Available if it is immediately (un-utilised CPUs) or imminently (CPUs with no queued jobs) able to execute a job. This criterion of availability depends on the nature of the CE, so we define different criteria of availability: For Batch System type Total Queuing Jobs=0 or Total Queuing Jobs/Total CPUs < 0.3 For a local PC Total Running Jobs == 0 2

Job status During its lifetime a job may go through a number of states. These states and there meaning are listed below. Ready Waiting Queued Running Done Outputready Rescheduled Failed The user has submitted the Job to the User Interface The Job is queued in Grid queues The Job is queued in a Cluster on a local queue (for example PBS) of a Computing Element The Job is running on a worker node The execution of the Job has completed The Output of the Job is ready To be defined To be defined Job Information & Parameters: Example Values Job Information Job Parameters JobID 209 Owner Nobody Site ccali.in2p3.fr/bqs-a Type Test SubmissionDate 2003-12-19 SubmissionTime 09:41:07 Status outputready JobID Name Value 209 LocalBatchId I3191743 209 Worker Node ccwali64 209 Memory 027904kB 209 Model PentiumIII(Coppermine) 3

209 CPU 996.894 209 CPU comsumed 36.97 209 Execution Time 37.3256549835 Job Life in the WMS When the job is done it is not necessary to keep all its information in the system. We plan to use an agent called the Job Monitor in order to clear out information from old jobs. This agent will transfer information concerning a job from the Job DB to a Provenance DB when it discovers jobs to be cleared. The criteria for clearing a job is still to be fixed. The Provenance DB will also be used as an accounting service. For the moment we propose to use the Bookkeeping DB as Provenance DB. Implementation details The WMS is implemented as a set of classes written in Python. This provides a clean object-oriented design together with a rapid development environment. It uses the XML- RPC protocol and Jabber to communicate with different services. The Job DB uses MySQL. DIRAC jobs are described with the Condor ClassAds Job Description Language (JDL). Job splitting and Merging proposal With the WMS architecture, we propose different approaches to do the job splitting and merging. We firstly considered jobs with no input data. Splitting and Merging approach for jobs with no input data The job is submitted to DIRAC and then divided into a set of sub-jobs, all of which are targeted for the same destination CE. A Merge job is also submitted to the same CE, which allows all sub-job outputs to be locally available, thus facilitating the merge, whose only purpose is to combine subjob outputs. The JDL of the parent job must contain these options in order for splitting to be possible: - JobType - NbEvents - SizePerEvent - CPUPerEvent - Splittable (true or false) - CEName 4

The Job Receiver stores the parent job in the DB and returns a JobId. Then it notifies a Splitter Service which is a specialized Optimiser. This splits the job into the n subjobs and one merging job. The jobs are queued and will eventually match to the specified CE when its Agent requests a job. The Splitter will also keep track of the relation between the parent s JobId and the SubJob s JobId in a DB. This information will be available via the Job Monitoring Service. The merging job will be executed only if all inputs are ready. Parent Job JobType = MinBias; NbEvents = 10000 ; SizePEreevnt = 114; CPUPerEvent = 60; Splittable = true; Site = X; Splitter Optimizer Splits job in n subjobs and Creates a merging job Theses job are then stores in Job DB and sorted in a queue by the Splitter. The relation between JobId and SubJobId are stored in a DB. Merging Job JobType = Merger; Executable = /bin/cat; Site = X; InputData = {mc_01.rawh; mc_02.raw.h, } SubJob 1 SubJob 2 JobType = MinBias; NbEvents JobType = 200 SubJob = ; MinBias; n. NbEvents = 200 ;. JobType = MinBias; NbEvents = 200 ; Site = X;. Job Queue rank 1 2 4 JobId 5 4 6 Implication: There are various implications of this scheme: - A new optimiser is required which is able to split jobs and manage the sub job dependencies. 5

- A way to report the status of the parent job from the Job Monitoring Service. For example: 23 % in done status, 2% in running status. The access to this information could be done by a special process which monitors subjobs and computes some statistics. - A Dependency Checker Agent will also check if data is available for a job. This builds on the idea of virtual data where a job can be submitted before the data is available. Specifically, this will be required to hold the merge job until all the outputs are ready. This agent will be interfaced with the file catalogue. Splitting and Merging approach for jobs with input data This is the same as the scenario without input data, except that fragmentation of the input data set must be considered. Here the Dependency Checker Agent will be make sure that a job is split into groups such that all the files in a sub job are accessible from a single CE. To facilitate this the Dependency Checker can act as or in conjunction with a Replica Manager and initiate data replication to insure that groups of input files for sub jobs are accessible from a single CE which has available computing resources. Terminology Glossary class-ad CE job-ad JDL LRMS PID SE UI VO WMS Classified advertisement Computing Element Class-ad describing a job Job Description Language Local Resource Management System Process Identifier Storage Element User Interface Virtual Organisation Workload Management System 6