LHCb Computing Strategy

Similar documents
LHCb Computing Resource usage in 2017

Andrea Sciabà CERN, Switzerland

DIRAC pilot framework and the DIRAC Workload Management System

LHCb Computing Status. Andrei Tsaregorodtsev CPPM

Overview of LHCb applications and software environment. Bologna Tutorial, June 2006

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Ganga - a job management and optimisation tool. A. Maier CERN

Monte Carlo Production on the Grid by the H1 Collaboration

CHIPP Phoenix Cluster Inauguration

DIRAC Distributed Infrastructure with Remote Agent Control

Computing at Belle II

Workload Management. Stefano Lacaprara. CMS Physics Week, FNAL, 12/16 April Department of Physics INFN and University of Padova

Status of KISTI Tier2 Center for ALICE

LHCb Computing Resources: 2018 requests and preview of 2019 requests

DIRAC Distributed Infrastructure with Remote Agent Control

Data and Analysis preservation in LHCb

PoS(ACAT2010)039. First sights on a non-grid end-user analysis model on Grid Infrastructure. Roberto Santinelli. Fabrizio Furano.

DESY. Andreas Gellrich DESY DESY,

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

LHCb Computing Technical Design Report

LHCb Computing Resources: 2019 requests and reassessment of 2018 requests

150 million sensors deliver data. 40 million times per second

The LHC Computing Grid. Slides mostly by: Dr Ian Bird LCG Project Leader 18 March 2008

LHCb Distributed Conditions Database

Computing. DOE Program Review SLAC. Rainer Bartoldus. Breakout Session 3 June BaBar Deputy Computing Coordinator

The CMS Computing Model

HammerCloud: A Stress Testing System for Distributed Analysis

The grid for LHC Data Analysis

The LHC Computing Grid

Computing / The DESY Grid Center

Batch Services at CERN: Status and Future Evolution

Data handling and processing at the LHC experiments

UW-ATLAS Experiences with Condor

How the Monte Carlo production of a wide variety of different samples is centrally handled in the LHCb experiment

DESY at the LHC. Klaus Mőnig. On behalf of the ATLAS, CMS and the Grid/Tier2 communities

Challenges of the LHC Computing Grid by the CMS experiment

Summary of the LHC Computing Review

Distributed Computing Framework. A. Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille

LHC Computing Models

ATLAS operations in the GridKa T1/T2 Cloud

Ganga The Job Submission Tool. WeiLong Ueng

CMS Belgian T2. G. Bruno UCL, Louvain, Belgium on behalf of the CMS Belgian T2 community. GridKa T1/2 meeting, Karlsruhe Germany February

Storage and I/O requirements of the LHC experiments

DIRAC data management: consistency, integrity and coherence of data

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

The Grid: Processing the Data from the World s Largest Scientific Machine

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

Virtualizing a Batch. University Grid Center

Overview of ATLAS PanDA Workload Management

Austrian Federated WLCG Tier-2

Philippe Charpentier PH Department CERN, Geneva

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

A Prototype of the CMS Object Oriented Reconstruction and Analysis Framework for the Beam Test Data

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

HEP Grid Activities in China

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Evolution of Database Replication Technologies for WLCG

Data services for LHC computing

INTRODUCTION TO THE ANAPHE/LHC++ SOFTWARE SUITE

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

Storage Resource Sharing with CASTOR.

RUSSIAN DATA INTENSIVE GRID (RDIG): CURRENT STATUS AND PERSPECTIVES TOWARD NATIONAL GRID INITIATIVE

Grid Computing Activities at KIT

CMS Tier-2 Program for user Analysis Computing on the Open Science Grid Frank Würthwein UCSD Goals & Status

1. Introduction. Outline

Considerations for a grid-based Physics Analysis Facility. Dietrich Liko

Preparing for High-Luminosity LHC. Bob Jones CERN Bob.Jones <at> cern.ch

Benchmarking the ATLAS software through the Kit Validation engine

Data Management for the World s Largest Machine

The CMS data quality monitoring software: experience and future prospects

Support for multiple virtual organizations in the Romanian LCG Federation

Data Transfers Between LHC Grid Sites Dorian Kcira

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback

The High-Level Dataset-based Data Transfer System in BESDIRAC

Monitoring System for the GRID Monte Carlo Mass Production in the H1 Experiment at DESY

Evolution of Database Replication Technologies for WLCG

The Grid. Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop

The ATLAS Production System

The ARDA project: Grid analysis prototypes of the LHC experiments

Monitoring of Computing Resource Use of Active Software Releases at ATLAS

Popularity Prediction Tool for ATLAS Distributed Data Management

ALHAD G. APTE, BARC 2nd GARUDA PARTNERS MEET ON 15th & 16th SEPT. 2006

ATLAS Distributed Computing Experience and Performance During the LHC Run-2

The INFN Tier1. 1. INFN-CNAF, Italy

ilcdirac and Continuous Integration: Automated Testing for Distributed Computing

DØ Southern Analysis Region Workshop Goals and Organization

A Simulation Model for Large Scale Distributed Systems

Clouds in High Energy Physics

CMS Computing Model with Focus on German Tier1 Activities

Multi-threaded, discrete event simulation of distributed computing systems

RELEASE OF GANGA. Basics and organisation What Ganga should do tomorrow Ganga design What Ganga will do today Next steps

Reprocessing DØ data with SAMGrid

STATUS OF PLANS TO USE CONTAINERS IN THE WORLDWIDE LHC COMPUTING GRID

IEPSAS-Kosice: experiences in running LCG site

Belle & Belle II. Takanori Hara (KEK) 9 June, 2015 DPHEP Collaboration CERN

The ATLAS EventIndex: an event catalogue for experiments collecting large amounts of data

Beob Kyun KIM, Christophe BONNAUD {kyun, NSDC / KISTI

The European DataGRID Production Testbed

WHEN the Large Hadron Collider (LHC) begins operation

LHC Computing Grid. Technical Design Report. Version: June The LCG TDR Editorial Board Chair: J. Knobloch Project Leader: L.

Transcription:

LHCb Computing Strategy Nick Brook Computing Model 2008 needs Physics software Harnessing the Grid DIRC GNG Experience & Readiness HCP, Elba May 07 1

Dataflow RW data is reconstructed: e.g. Calo. Energy clusters Particle ID Tracks t recons time only enough info is stored to allow physics pre-selection to run at a later stage - reduced DST (rdst) - stored separately from RW Reconstruction performed twice a year: quasi real time after LHC shutdown HCP, Elba May 07 2

Dataflow - Stripping Stripping performed 4 times per year rdst: 20kB/evt RW: 35kB/evt DST: 110kB/evt rdst is analysed in production-mode event streams for further analysis; 20-30 streams lgorithm developed by physics working groups - use as i/p rdst & RW Event to be output will have additional reconstructed info added: (full) DST+ RW data Event Tag Collection - created to allow quick access to data; contain metadata HCP, Elba May 07 3

Dataflow - nalysis User physics analysis will be primarily performed on the output of the stripping Output from stripping is selfcontained i.e. no need to navigate between files nalysis generates quasiprivate data e.g. Ntuple and/or personal DSTs Data publicly accessible - enable remote collaboration HCP, Elba May 07 4

Use of computing centres Main user analysis supported at CERN+6 Tier-1 centres Tier-2 centres essentially Monte Carlo production facilities Plan to make use of LHCb online farm for re-processing HCP, Elba May 07 5

2008 Resource Summary Estimated 4 10 6 secs physics (including machine efficiency) ssume 8 10 9 events from event filter farm to CERN computer centre CPU (2.8GHz P4 years) Disk (TB) Tape (TB) CERN 360 350 631 Tier-1 s 1770 1025 860 Tier-2 s 4550 - - HCP, Elba May 07 6

LHCb software Event model / Physics event model GenParts MCParts Simul. Gauss MCHits Detector Description Digit. Boole RawData Digits Recons. & HLT Brunel Conditions Database DST nalysis DaVinci MiniDST OD Gaudi LHCb data processing applications and data flow HCP, Elba May 07 7

LHCb software framework Object diagram of the Gaudi architecture HCP, Elba May 07 8

LHCb software framework Gaudi is architecture-centric, customisable framework dopted by TLS; used by GLST & HRP Same framework used both online & offline lgorithmic part of data processing as a set of OO objects decoupling between the objects describing the data and the algorithms allows programmers to concentrate separately on both. allows a longer stability for the data objects (the LHCb event model) as algorithms evolve much more rapidly n important design choice has been to distinguish between a transient and a persistent representation of the data objects changed from persistency solutions without the algorithms being affected. Event Model classes only contain enough basic internal functionality for giving algorithms access to their content and derived information lgorithms and tools perform the actual data transformations HCP, Elba May 07 9

DIRC - community Grid solution The DIRC Workload & Data Management System (WMS & DMS) is made up of Central Services and Distributed gents The main aims of DIRC are: To integrate all of the heterogeneous compute resources available to LHCb Minimize the human intervention at sites Use worldwide LHC Computing Grid services wherever possible DIRC realizes these goals via: Pilot gent paradigm Overlay Network paradigm HCP, Elba May 07 10

DIRC Overlay Network Paradigm DIRC gents are deployed close to resources Forms an overlay network of gents masking the underlying diversity of the available compute resource Services interact with gents Computing Resources PCs Site Clusters Grid HCP, Elba May 07 11

DIRC Overlay Network Paradigm DIRC gents are deployed close to resources Forms an overlay network of gents masking the underlying diversity of the available compute resource Services interact with gents Computing Resources PCs Site Clusters Grid HCP, Elba May 07 12

DIRC Overlay Network Paradigm DIRC gents are deployed close to resources Forms an overlay network of gents masking the underlying diversity of the available compute resource Services interact with gents Computing Resources PCs gents Site Clusters Grid HCP, Elba May 07 13

DIRC Overlay Network Paradigm DIRC gents are deployed close to resources Forms an overlay network of gents masking the underlying diversity of the available compute resource Services interact with gents Computing Resources PCs Service 1 gents Service 3 Site Clusters Service 2 Grid HCP, Elba May 07 14

DIRC Pilot gent Paradigm DIRC is a PULL scheduling system gents first occupy a resource and then request jobs from a central task queue This late binding allows execution environment to be checked in advance Pilot gents are sent to the glite Resource Broker as normal jobs Facilitate PULL approach on PUSH system LCG jobs are Pilot jobs in the context of the DIRC WMS ctual workload management performed by DIRC HCP, Elba May 07 15

DIRC Workload Management System Heterogeneous groupings of resources such as clusters / Grids become homogeneous via DIRC DIRC can therefore be viewed as a (very) large batch system ccounting Priority Mechanism Fair share HCP, Elba May 07 16

GNG - user interface to the Grid Goal Simplify the management of analysis for end-user physicists by developing a tool for accessing Grid services with built-in knowledge of how Gaudi works Required user functionality Job preparation and configuration Job submission, monitoring and control Bookkeeping browsing, etc. Done in collaboration with TLS Use Grid middleware services interface to the Grid via Dirac and create synergy between the two projects HCP, Elba May 07 17

Ganga jobs job in Ganga is constructed from a set of building blocks, not all required for every job Job pplication Backend Input Dataset Output Dataset Splitter Merger What to run Where to run Data read by application Data written by application Rule for dividing into subjobs Rule for combining outputs HCP, Elba May 07 18

Ganga: how the pieces fit together Ganga has built-in support for TLS and LHCb pplications Component architecture allows customisation for other user groups TLS applications LHCb applications Other applications User interface for job definition and management Metadata catalogues File catalogues Tools for data management Data storage and retrieval Remote repository Local repository Ganga job archives Ganga monitoring loop Local batch systems Experiment-specific workload-management systems Distributed (Grid) systems Processing systems (backends) HCP, Elba May 07 19

LHCb Simulation Production Typical MC Production job lasts 24hrs Recently achieved 10K concurrent production jobs Throughput only limited by available capacity of LCG ~80 distinct sites accessed via Grid or directly Sustained resource usage over extended periods of time System is stable for simulation HCP, Elba May 07 20

Breakdown of Production HCP, Elba May 07 21

LHCb Reconstruction Results Reconstruction challenge is ongoing Current issues include:site service instability; tape failures pril 2007, reconstruction jobs successfully running at all LHCb Tier-1 sites CERN (Switzerland) IN2P3 (France) GridKa (Germany) CNF (Italy) NIKHEF (Netherlands) PIC (Spain) RL (U.K.) HCP, Elba May 07 22

LHCb nalysis ll Nos of users LHCb 596 unique GNG Users 99 users from LHCb ~41k GNG sessions since start of year ~10k LHCb sessions Nos of sessions Month Month HCP, Elba May 07 23

LHCb nalysis 393k jobs passed through DIRC analysis system since start of year Users happy with efficiency ccess to large amount of resources HCP, Elba May 07 24

Summary Computing model being finalised currently under stress test particularly access to data Software framework robust & mature final version of reconstruction s/w available 3 different persistency solutions without major upheaval LHCb DIRC system allows efficient use of Grid resources Monte Carlo production now routine Reconstruction under stress test User analysis on the Grid seeing increase in use GNG interface between s/w framework & DIRC HCP, Elba May 07 25

Summary Computing model being finalised currently under stress test particularly access to data Software framework robust & mature final version of reconstruction s/w available 3 different persistency solutions without major upheaval LHCb DIRC system allows efficient use of Grid resources Monte Carlo production now routine Reconstruction under stress test User analysis on the Grid seeing increase in use GNG interface between s/w framework & DIRC Confident LHCb computing will be ready for data taking HCP, Elba May 07 26