The grid for LHC Data Analysis

Similar documents
The LHC Computing Grid. Slides mostly by: Dr Ian Bird LCG Project Leader 18 March 2008

CHIPP Phoenix Cluster Inauguration

The Grid: Processing the Data from the World s Largest Scientific Machine

The LHC Computing Grid

The Grid. Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Travelling securely on the Grid to the origin of the Universe

First Experience with LCG. Board of Sponsors 3 rd April 2009

The EGEE-III Project Towards Sustainable e-infrastructures

RUSSIAN DATA INTENSIVE GRID (RDIG): CURRENT STATUS AND PERSPECTIVES TOWARD NATIONAL GRID INITIATIVE

CERN and Scientific Computing

Andrea Sciabà CERN, Switzerland

A short introduction to the Worldwide LHC Computing Grid. Maarten Litmaath (CERN)

LHC Computing Grid today Did it work?

The CMS Computing Model

Grid Computing a new tool for science

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

150 million sensors deliver data. 40 million times per second

Storage and I/O requirements of the LHC experiments

Grid Computing Activities at KIT

I Service Challenge e l'implementazione dell'architettura a Tier in WLCG per il calcolo nell'era LHC

Computing for LHC in Germany

Moving e-infrastructure into a new era the FP7 challenge

e-science Infrastructure and Applications in Taiwan Eric Yen and Simon C. Lin ASGC, Taiwan Apr. 2008

Challenges of the LHC Computing Grid by the CMS experiment

IEPSAS-Kosice: experiences in running LCG site

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

Tier-2 DESY Volker Gülzow, Peter Wegner

Computing grids, a tool for international collaboration and against digital divide Guy Wormser Director of CNRS Institut des Grilles (CNRS, France)

LHCb Computing Resource usage in 2017

The European DataGRID Production Testbed

LHCb Computing Strategy

Grid and Cloud Activities in KISTI

High Performance Computing from an EU perspective

A European Vision and Plan for a Common Grid Infrastructure

Constant monitoring of multi-site network connectivity at the Tokyo Tier2 center

Summary of the LHC Computing Review

The LHC Computing Grid

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 27 July 2010

Status of KISTI Tier2 Center for ALICE

Workload Management. Stefano Lacaprara. CMS Physics Week, FNAL, 12/16 April Department of Physics INFN and University of Padova

Grid Challenges and Experience

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback

Distributed e-infrastructures for data intensive science

The INFN Tier1. 1. INFN-CNAF, Italy

Experience of the WLCG data management system from the first two years of the LHC data taking

Preparing for High-Luminosity LHC. Bob Jones CERN Bob.Jones <at> cern.ch

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

e-infrastructure: objectives and strategy in FP7

Experience of Data Grid simulation packages using.

DESY. Andreas Gellrich DESY DESY,

Austrian Federated WLCG Tier-2

Outline. Infrastructure and operations architecture. Operations. Services Monitoring and management tools

CernVM-FS beyond LHC computing

CC-IN2P3: A High Performance Data Center for Research

Spanish Tier-2. Francisco Matorras (IFCA) Nicanor Colino (CIEMAT) F. Matorras N.Colino, Spain CMS T2,.6 March 2008"

Features and Future. Frédéric Hemmer - CERN Deputy Head of IT Department. Enabling Grids for E-sciencE. BEGrid seminar Brussels, October 27, 2006

Computing Model Tier-2 Plans for Germany Relations to GridKa/Tier-1

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 21 July 2011

PoS(EGICF12-EMITC2)106

GRIDS INTRODUCTION TO GRID INFRASTRUCTURES. Fabrizio Gagliardi

The LHC computing model and its evolution. Dr Bob Jones CERN

High Energy Physics data analysis

Batch Services at CERN: Status and Future Evolution

where the Web was born Experience of Adding New Architectures to the LCG Production Environment

e-science for High-Energy Physics in Korea

LHCb Distributed Conditions Database

Data Transfers Between LHC Grid Sites Dorian Kcira

Philippe Charpentier PH Department CERN, Geneva

Computing / The DESY Grid Center

High Throughput WAN Data Transfer with Hadoop-based Storage

Reprocessing DØ data with SAMGrid

The EU DataGrid Testbed

Grid Computing: dealing with GB/s dataflows

Operating the Distributed NDGF Tier-1

UW-ATLAS Experiences with Condor

Giovanni Lamanna LAPP - Laboratoire d'annecy-le-vieux de Physique des Particules, Université de Savoie, CNRS/IN2P3, Annecy-le-Vieux, France

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

Pan-European Grid einfrastructure for LHC Experiments at CERN - SCL's Activities in EGEE

Large scale commissioning and operational experience with tier-2 to tier-2 data transfer links in CMS

The ARDA project: Grid analysis prototypes of the LHC experiments

EGEE (JRA4) Loukik Kudarimoti DANTE. RIPE 51, Amsterdam, October 12 th, 2005 Enabling Grids for E-sciencE.

Support for multiple virtual organizations in the Romanian LCG Federation

Grid Computing: dealing with GB/s dataflows

De BiG Grid e-infrastructuur digitaal onderzoek verbonden

HEP Grid Activities in China

Grid Interoperation and Regional Collaboration

DESY at the LHC. Klaus Mőnig. On behalf of the ATLAS, CMS and the Grid/Tier2 communities

Bob Jones. EGEE and glite are registered trademarks. egee EGEE-III INFSO-RI

Storage Resource Sharing with CASTOR.

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

EGEE and Interoperation

Pushing the Limits. ADSM Symposium Sheelagh Treweek September 1999 Oxford University Computing Services 1

Grid Security Policy

HammerCloud: A Stress Testing System for Distributed Analysis

Grids and Security. Ian Neilson Grid Deployment Group CERN. TF-CSIRT London 27 Jan

Accelerating Throughput from the LHC to the World

arxiv: v1 [physics.ins-det] 1 Oct 2009

Application of Virtualization Technologies & CernVM. Benedikt Hegner CERN

Transcription:

The grid for LHC Data Analysis ICAP 2006 Conference Chamonix 5 October 2006 Les Robertson - CERN LHC Computing Grid Project Leader

The LHC Computing Challenges 1. Data After reduction by triggers and data acquisition filter the raw data will be recorded at 100 1,000 MBytes/sec With processed datasets, across the 4 experiments, 15 PetaBytes of new data each year 2. Computation A few thousand users and about 100K of today s fastest PC processors 3. Funding Most of the computing resources will not be at CERN distributed across ~100 scientific computing centres around the world les.robertson@cern,ch 2

LCG - The LHC Computing Grid Purpose Develop, build and maintain a distributed computing environment for the storage and analysis of data from the four LHC experiments Ensure the computing service and common application libraries and tools Collaboration experiments and ~100 regional computing centres Phase I 2002-05 - Development & planning Phase II 2006-2008 Deployment & commissioning of the initial services les.robertson@cern,ch 3

Distribution of Computing Services Summary of Computing Resource Requirements All experiments - 2008 First full year From LCG TDR - June 2005 CERN All Tier-1s All Tier-2s Total CPU (MSPECint2000s) 25 56 61 142 Disk (PetaBytes) 7 31 19 57 Tape (PetaBytes) 18 35 53 All Tier-2s 43% CPU Disk Tape CERN 18% All Tier-2s 33% CERN 12% CERN 34% All Tier-1s 39% All Tier-1s 55% All Tier-1s 66% les.robertson@cern,ch 4

LCG Service Hierarchy Tier-0 the accelerator centre Data acquisition & initial processing Long-term data curation Distribution of data Tier-1 centres Canada Triumf (Vancouver) France IN2P3 (Lyon) Germany Forschunszentrum Karlsruhe Italy CNAF (Bologna) Netherlands NIKHEF/SARA (Amsterdam) Nordic countries distributed Tier-1 Spain PIC (Barcelona) Taiwan Academia SInica (Taipei) UK CLRC (Oxford) US FermiLab (Illinois) Brookhaven (NY) Tier-1 online to the data acquisition process high availability Managed Mass Storage grid-enabled data service Data-heavy analysis National, regional support Tier-2 ~100 centres in ~40 countries Simulation End-user analysis batch and interactive les.robertson@cern,ch 5

Using the grid Grid Computing Centre User job JDL List of files grid grid job JDL grid job job JDL files JDL grid job files files JDL grid job grid job files JDL grid job JDL files JDL grid job files files JDL grid job files JDL files job generator Grid File Catalog Grid Resource Broker Grid Information Service Grid Computing Centre Grid Computing Centre Grid Computing Centre Grid Computing Centre Bookkeeping Data collection Grid Computing Centre Application les.robertson@cern,ch 6 Grid

LCG Baseline Grid Services Information system Security framework Security model Certificate infrastructure Policy and processes Virtual Organisation management services User registration Authentication & authorisation Compute Resource Services Interface to site Job submission, monitoring & control Storage management services Standard interface (SRM) Data transfer services Basic file transfer Scheduling and recovery service Grid operations Duty operator responsibility cycles around 5 large sites Monitoring tools Availability/reliability test framework Accounting User support Resource Broker Grid Catalog Applications software installation Information publisher Authentication and authorisation services Grid catalogue services Database services Data management toolkit POSIX-like I/O service g r i d s i t e les.robertson@cern,ch 7

HEP analysis using the grid Quite different from large simulation But: Emphasis on low latency, variable work-load Multi-user concurrent activities Lots of data Still an activity for power users Lots of end-users (physicists in the collaborations) Ultimately most analysis will need the grid User environment for the grid: All collaborations are building their own interfaces/environments Purpose: Facilitate: job preparation, job generation, job monitor/control, bookkeeping and data collection Hide: error recovery and details of the computing environments (local batch system, grid) Most of the experiments use at least 2 grid infrastructures les.robertson@cern,ch 8

Example: ATLAS and LHCb Ganga Ganga GUI GUI based on dockable windows Job details Logical Folders Scriptor Job Monitor Splitter 15 February 2006 19 les.robertson@cern,ch 9

WLCG depends on two major science grid infrastructures. EGEE OSG - Enabling Grids for E-Science - US Open Science Grid les.robertson@cern,ch 10

Wide Area Network Tier-2s and Tier-1s are inter-connected by the general purpose research networks Any Tier-2 may access data at any Tier-1 Brookhaven IN2P3 GridKa Dedicated 10 Gbit optical network TRIUMF ASCC Nordic CNAF RAL Fermilab SARA PIC LCG

Production Grids for LHC EGEE Grid Last month, running jobs for the whole Grid 14K 60 50 Jobs/Day - EGEE Grid Jobs/day EGEE Grid lhcb cms atlas alice 40 K jobs/day 30 20 10 0 Jun- 05 Jul- 05 Aug- 05 Sep- 05 Oct- 05 Nov- 05 Dec- 05 Jan- 06 month les.robertson@cern,ch 12 Feb- 06 Mar- 06 Apr- 06 May- 06 Jun- 06 Jul- 06 Aug- 06 alice atlas cms lhcb geant4 dteam non-lhc EGEE Grid ~50K jobs/day ~14K simultaneous jobs during prolonged periods ~20% of the 2008 requirement

Data Distribution CERN T1s April 06 test period nominal rate when LHC is operating - 1.6 GB/s reached but only for one day Sustained data rate 80% of the target 1.6 GBytes/sec 1.3 GBytes/sec August experiment-driven transfers (ATLAS and CMS) sustained 50% of the SC4 target under much more realistic conditions 0.8 GBytes/sec CMS transferred a steady 1 PByte/month between Tier-1s & Tier-2s during a 90 day period ATLAS distributed 1.25 PBytes from CERN during a 6-week period les.robertson@cern,ch 13

2006 Commissioning Schedule SC4 becomes initial service when reliability and performance goals met Continued testing of computing models, basic services 2007 Testing DAQ Tier-0 (??) & integrating into DAQ Tier-0 Tier-1 data flow Building up end-user analysis support Exercising the computing systems, ramping up job rates, data management performance,. Introduce residual services Full FTS services; 3D; SRM v2.2; VOMS roles Initial service commissioning increase reliability, performance, capacity to target levels, experience in monitoring, 24 X 7 operation,. 01jul07 - service commissioned - full 2007 capacity, performance first physics 2008 Experiments les.robertson@cern,ch 14 Sites & Services

EGEE-II INFSO-RI-031688 Enabling Grids for E-sciencE More than 20 applications from 7 domains High Energy Physics (Pilot domain) 4 LHC experiments Other HEP (DESY, Fermilab, etc.) Biomedicine (Pilot domain) Bioinformatics Medical imaging Earth Sciences Earth Observation Solid Earth Physics Hydrology Climate Computational Chemistry Fusion Astronomy Cosmic microwave background Gamma ray astronomy Geophysics Industrial applications Applications Running

Example from Biomedical applications: Avian Flu Millions of chemical compounds available in laboratories High Throughput Screening 2$/compound, nearly impossible 300,000 Chemical compounds: ZINC Chemical combinatorial library Molecular docking (Autodock) ~100 CPU years, 600 GB data Data challenge on EGEE grid ~6 weeks on ~2000 computers Target (PDB) : Neuraminidase (8 structures) Hits sorting and refining In-vitro screening of 100 hits

Another example: the gridcc project (www.gridcc.org) Using Grid technology to operate distributed instruments And couple them with grid resources for simulation and analysis Pilot applications: Remote operation of an accelerator Meteorology (Ensemble Limited Area Forecasting) Device Farm for the Support of Cooperative Distributed Measurements in Telecommunications and Networking Laboratories Geo-hazards: Remote Operation of Geophysical Monitoring Network Main GridCC Pilot Applications: Remote Operation of an Accelerator Elettra Synchrotron G. Maron, EGEE06 Conference,CERN, 25-29 September, 2006

Different applications LHC Monte Carlo simulation Fusion Drug discovery (like avian flu and malaria) Similar characteristics Jobs are CPU-intensive Large number of independent jobs Run by few (expert) users Small input; large output & needs Massive CPU needs Data collections for further analyses ATLAS ITER Drug discovery

Towards General Science Infrastructures EGEE Funded 50% by the EU - 90 partners - 34 countries Funding for people - middleware development, operations, pilot applications, outreach No funding for hardware Many sites outside Europe Current (second) phase ends March 2008! Expectation that there will be a third phase of 2-3 years Post EGEE OSG Early discussions on a long term science grid infrastructure (cf. the GEANT research and education network backbone) Organised as a consortium with some funding ($6M/year for next 5 years) from DoE and NSF for facility development and operations No middleware activity, no funding for hardware les.robertson@cern,ch 19

How to get going NA4 website (http://egeena4.lal.in2p3.fr/) EGEE Conferences and Users Forums Share your expertise, learn from other users. Be open to collaboration with others Next event: Manchester May 2007 EGEE website www.eu-egee.org OSG website www.opensciencegrid.org les.robertson@cern,ch 20

LCG http://gridportal.hep.ph.ic.ac.uk/rtm/ les.robertson@cern,ch 21