Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

Similar documents
Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Data Transfers Between LHC Grid Sites Dorian Kcira

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

The CMS Computing Model

New strategies of the LHC experiments to meet the computing requirements of the HL-LHC era

Summary of the LHC Computing Review

High-Energy Physics Data-Storage Challenges

Andrea Sciabà CERN, Switzerland

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

DESY at the LHC. Klaus Mőnig. On behalf of the ATLAS, CMS and the Grid/Tier2 communities

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

UW-ATLAS Experiences with Condor

CMS Grid Computing at TAMU Performance, Monitoring and Current Status of the Brazos Cluster

Towards Network Awareness in LHC Computing

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

HammerCloud: A Stress Testing System for Distributed Analysis

Computing for LHC in Germany

The INFN Tier1. 1. INFN-CNAF, Italy

Experience of the WLCG data management system from the first two years of the LHC data taking

First Experience with LCG. Board of Sponsors 3 rd April 2009

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

WLCG Transfers Dashboard: a Unified Monitoring Tool for Heterogeneous Data Transfers.

Accelerating Throughput from the LHC to the World

CouchDB-based system for data management in a Grid environment Implementation and Experience

Computing. DOE Program Review SLAC. Rainer Bartoldus. Breakout Session 3 June BaBar Deputy Computing Coordinator

LHCb Computing Status. Andrei Tsaregorodtsev CPPM

Overview of ATLAS PanDA Workload Management

The ATLAS Distributed Analysis System

Batch Services at CERN: Status and Future Evolution

RUSSIAN DATA INTENSIVE GRID (RDIG): CURRENT STATUS AND PERSPECTIVES TOWARD NATIONAL GRID INITIATIVE

The LHC Computing Grid

UK Tier-2 site evolution for ATLAS. Alastair Dewhurst

Storage Resource Sharing with CASTOR.

Update of the Computing Models of the WLCG and the LHC Experiments

Austrian Federated WLCG Tier-2

SPINOSO Vincenzo. Optimization of the job submission and data access in a LHC Tier2

Travelling securely on the Grid to the origin of the Universe

Computing Model Tier-2 Plans for Germany Relations to GridKa/Tier-1

An Analysis of Storage Interface Usages at a Large, MultiExperiment Tier 1

Operating the Distributed NDGF Tier-1

CERN and Scientific Computing

LHCb Computing Resources: 2019 requests and reassessment of 2018 requests

150 million sensors deliver data. 40 million times per second

Data services for LHC computing

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

arxiv: v1 [physics.ins-det] 1 Oct 2009

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback

Belle & Belle II. Takanori Hara (KEK) 9 June, 2015 DPHEP Collaboration CERN

Data Management for the World s Largest Machine

Stephen J. Gowdy (CERN) 12 th September 2012 XLDB Conference FINDING THE HIGGS IN THE HAYSTACK(S)

Understanding the T2 traffic in CMS during Run-1

Preparing for High-Luminosity LHC. Bob Jones CERN Bob.Jones <at> cern.ch

T0-T1-T2 networking. Vancouver, 31 August 2009 LHCOPN T0-T1-T2 Working Group

Status of KISTI Tier2 Center for ALICE

Grid and Cloud Activities in KISTI

Influence of Distributing a Tier-2 Data Storage on Physics Analysis

ATLAS Distributed Computing Experience and Performance During the LHC Run-2

Storage on the Lunatic Fringe. Thomas M. Ruwart University of Minnesota Digital Technology Center Intelligent Storage Consortium

Constant monitoring of multi-site network connectivity at the Tokyo Tier2 center

LHCb Computing Resources: 2018 requests and preview of 2019 requests

CernVM-FS beyond LHC computing

Distributed Monte Carlo Production for

CHIPP Phoenix Cluster Inauguration

Philippe Charpentier PH Department CERN, Geneva

LHCb Computing Resource usage in 2017

HTC/HPC Russia-EC. V. Ilyin NRC Kurchatov Institite Moscow State University

Grid Computing Activities at KIT

Big Data Analytics and the LHC

The LHC Computing Grid. Slides mostly by: Dr Ian Bird LCG Project Leader 18 March 2008

The Compact Muon Solenoid Experiment. Conference Report. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

Lessons Learned in the NorduGrid Federation

ANSE: Advanced Network Services for [LHC] Experiments

where the Web was born Experience of Adding New Architectures to the LCG Production Environment

C3PO - A Dynamic Data Placement Agent for ATLAS Distributed Data Management

The Grid: Processing the Data from the World s Largest Scientific Machine

CERN openlab II. CERN openlab and. Sverre Jarp CERN openlab CTO 16 September 2008

CMS Computing Model with Focus on German Tier1 Activities

ISTITUTO NAZIONALE DI FISICA NUCLEARE

The ATLAS PanDA Pilot in Operation

The Software Defined Online Storage System at the GridKa WLCG Tier-1 Center

System upgrade and future perspective for the operation of Tokyo Tier2 center. T. Nakamura, T. Mashimo, N. Matsui, H. Sakamoto and I.

arxiv: v1 [cs.dc] 20 Jul 2015

Popularity Prediction Tool for ATLAS Distributed Data Management

PoS(EGICF12-EMITC2)106

Reprocessing DØ data with SAMGrid

The ATLAS EventIndex: Full chain deployment and first operation

Edinburgh (ECDF) Update

Grid Data Management

Experiences with the new ATLAS Distributed Data Management System

1. Introduction. Outline

The PanDA System in the ATLAS Experiment

Storage and I/O requirements of the LHC experiments

Considerations for a grid-based Physics Analysis Facility. Dietrich Liko

LHCb Computing Strategy

The grid for LHC Data Analysis

One Pool To Rule Them All The CMS HTCondor/glideinWMS Global Pool. D. Mason for CMS Software & Computing

Online data storage service strategy for the CERN computer Centre G. Cancio, D. Duellmann, M. Lamanna, A. Pace CERN, Geneva, Switzerland

A short introduction to the Worldwide LHC Computing Grid. Maarten Litmaath (CERN)

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

and the GridKa mass storage system Jos van Wezel / GridKa

Transcription:

Challenges and Evolution of the LHC Production Grid April 13, 2011 Ian Fisk 1

Evolution Uni x ALICE Remote Access PD2P/ Popularity Tier-2 Tier-2 Uni u Open Lab m Tier-2 Science Uni x Grid Uni z USA Tier-2 Brookhaven Lab a UK USA FermiLab France Physics Tier 1 CERN Department Tier2 Italy. Uni a Uni n CMS Full Mesh Desktop # Lab b NL Germany Lab c "! Uni y Uni b Over the development the evolution of the WLCG Production grid has oscillated between structure and flexibility - Driven by capabilities of the infrastructure and the needs of the experiments 2

Evolution LHC Computing has grown up with Grid development - Many previous experiments have achieved distributed computing - LHC experiments started with a fully distributed LHC Computing Grid was approved by CERN Council Sept. 20 2001 First Grid Deployment Board was Oct. 2002 LCG was built on services developed in Europe and the US. LCG has collaborated with a number of Grid Projects It evolved into the Worldwide LCG (WLCG) EGEE, EGI, NorduGrid, and Open Science Grid Services Support the 4 LHC Experiments NORDUGRID Grid Solution for Wide Area Computing and Data Handling 3

WLCG Today Today >140 sites ~150k CPU cores Hit 1M jobs per day >50 PB disk 4

To greater and lesser extents LHC Computing model are based on the MONARC model - Developed more than a decade ago - Foresaw Tiered Computing Facilities to meet the needs of the LHC Experiments Architectures Tier3 Univ WG 2 Tier3 Univ WG M Tier3 Univ WG 1 Tier2 Center 20k Si95 20 Tbytes Disk, Robot 622 Mbits /s 622Mbits/s Model Circa 2005 622 Mbits/s CERN/CMS 350k Si95 350 Tbytes Disk; Robot 622 Mbits/s FNAL/BNL 70k Si95 70 Tbytes Disk; Robot Fig. 4-1 Computing for an LHC Experiment Based on a Hierarchy of Computing Centers. Capacities for CPU and disk are representative and are provided to give an approximate scale). ALICE ATLAS CMS LHCb - 16 - N X 622 Mbits/s 622 Mbits /s Tier-0 Tier-0 Tier-0 Cloud Tier-1 Tier-1 Tier-1 Tier-1 Tier-1 Tier-1 Tier-1 Tier-1 Mesh Tier-2 Tier-2 Tier-2 Tier-2 Tier-2 Tier-2 Tier-2 5

Working Today At the LHC most analysis work is conducted far away from the data archives and storage is widely distributed Organized Processing Storage Data Serving Chaotic Analysis Prompt Processing Archival Storage Tape CERN Tape Tier-1 Tier-1 Tier-1 Tape Tier-2 Tier-2 Tier-2 Ti 6

Processing Scale 350M hours at Tier-1 2010 was the first full year of running - Adding Tier-1 and Tier-2 computing time LHC used roughly 80 CPU millennia in 2010 Percentage of Utilized Hours 200% 180% 160% 140% 120% 100% 80% 60% 40% 20% Usage Tier-1 Only ALICE Percentage Use ATLAS Percentage Use CMS Percentage Use LHCb Percentage Use Total Percentage Use 0% 1 2 3 4 5 6 7 8 9 10 11 12 Month 2010 7

Scale of Storage Decreases in the cost of disk and technology to run big disk farms - LHC is no longer talking about 10% disk caches ALICE ATLAS CMS LHCb DZero CDF T0 Disk (TB) 6100 7000 4500 1500 ~500 ~500 T0 Tape (TB) 6800 12200 21600 2500 5900 6600 T1 Disk (TB) 7900 24800 19500 3500 T1 Tape (TB) 13100 30100 52400 3470 T2 Disk (TB) 6600 37600 19900 20 Disk Total (TB) 20600 69400 43900 5020 Tape Total (TB) 19900 42300 74000 5970 - In 2011 majority of the currently accessed data could be disk resident 8

Scale of Archival Storage Challenge is growing volume of data that is produced With the current technology evolution CERN will have robotic capacity for half an exabyte Experiment Data in CERN Castor Experiment Data in FNAL Enstore per day (26PB total) 9

Analysis Disk Storage Percentage of Hours Used! 200%! 180%! 160%! 140%! 120%! 100%! 80%! 60%! 40%! 20%! 0%! Total Percentage of Tier-2 Usage! ALICE ATLAS CMS LHCb January! February! March! April! May! June! July! August! September!October! November!December! Tier-2 Tier-2 Tier-2 ALICE Percentage Use! ATLAS Percentage Use! CMS Percentage Use! Example from LHC - Tier-2s are very heavily utilized - Many of the challenging IO Applications are conducted at centers with exclusively disk Tier-2s vary from 10s of TB at the smallest site to 1PB of disk at the larger sites - There have been many more options to manage this much space In 2011 there are more than 60PB of T2 Disk in LHC GFS GPFS 10

Evolving Challenge - Data Management Data is backed up on tape. Organized processing centers have substantial disk caches Analysis centers have large disk resources Good options in technology for virtualizing disk storage What s the problem? - There are almost 100 Tier-2 sites that make up WLCG Managing the space accessed by users efficiently is an interesting problem 11

Placement Computing models for LHC were based on to greater and lesser extents on the MONARC computing model of 2000 and relied heavily on data placement - Jobs were sent to datasets already resident on sites - Multiple copies of the data would be hosted on the distributed infrastructure - General concern that the network would be insufficient or unreliable Tape Tier-2 Tier-1 Tier-2 Richard Mount 12

Distribution - Change from Tier-1 Tier-2 Tier-1 Tier-2 Transfers West Tier-1 Tier-2 - To Tier-1 Tier-2 Transfers East 13

Data Management 2000 Space (TB) CMS Experiments have chosen a variety of philosophy 1500 1000 - ATLAS started with replication of nearly all data out to regions - CMS divided into Central background samples, physics groups, and the local community 500 0 Analysis Ops B-Physics b-tag E-gamma EWK Exotica Forward HI Higgs Jet-MET Local Muon QCD SUSY Tau-pFlow Top Tracker-DPG Tracker-OPG Trigger 14

For a system intended to protect against weak networking, we re using a lot of network - LHC Experiments reprocessed a lot of data in 2010 - Refreshing large disk caches requires a lot of networking Access Number of Accesses 700000 600000 500000 400000 300000 200000 100000 0 ATLAS In CMS 30 % of samples subscribed by physicists not used for 3 months during 2010 Traffic on OPN up to 70 Gb/s! - ATLAS reprocessing campaigns CMS 100-1000 CPU Hours 15

Tape Tier-1 Placement In an environment that discounts the network the sites are treated independently - On the time scale of a job submitted and running on a site it is assumed the local environment cannot be changed From a data access perspective in 2011 data available over the network from a disk at a remote site may be closer than data on the local tape installation Tape Tier-1 Tier-1 Tier-1 Tape Tier-2 16

Dynamic Replication ATLAS introduced Panda Dynamic Data Placement (PD2P) Tier-1 Tier-2 Jobs are sent to Tier-1 and data replicated to a Tier-2 at submission time 17

Data Placement and ReUse Dynamic placement now accounts for a lot of the networking Re-brokering jobs is increasing the reuse of samples and the efficiency 18

If you want to understand how better to manage storage space, important to know how it s used Popularity Interesting challenge to track the utilization of 30PB worth of files spread over more than 50 sites - Equally important to know what s not accessed 19

Clean-Up and Replication Once popularity is understood - Popular data can be replicated multiple times - Unused data replicas can be cleaned up Data Popularity will be tracked at the file level - Improves granularity and should improve reuse of the service 20 -

Analysis Data We like to think of high energy data as series of embarrassing parallel events 1 2 3 4 5 6 7 8 In reality it s not how we either write or read the files - More like Big gains in how storage is used by optimizing how events are read and streamed to an application - Big improvements from the Root team and application teams in this area 21

Wide Area Access With properly optimized IO other methods of managing the data and the storage are available - Sending data directly to applications over the WAN Not immediately obvious that this increases the wide area network transfers - If a sample is only accessed once, then transferring it before hand or in real time are the same number of bytes sent - If we only read a portion of the file, then it might be fewer bytes 22

xrootd Caching Demonstrator Case Global Xrootd Redirector User Analysis Tier 3 Site Xrootd Xrootd Cache Cache Xrootd Cache Xrootd Local Data Xrootd Local Redirector Remote Site Xrootd Remote Site Xrootd Current Xrootd demonstrator in CMS is intended to support the university computing - Facility in Nebraska and Bari with data served from a variety of locations - Tier-3 receiving data runs essentially diskless Similar installation being prepared in ATLAS 23

This Tier-3 has a 10Gb/s network CPU Efficiency competitive Data Served Performance 8TB/day peak about 1.5TB average,000 wall hours in the last day. Best USCMS T2 efficiency about 80% CPU efficiency about 60% in Omaha 24

Networking ALICE Distributes Data in this way - Rate from the ALICE Xrootd servers is comparable in peaks to other LHC experiments ALICE 1GB/s 25

Future? Once you have streams of objects and optimized IO, the analysis application an application like skimming does not look so different from video streaming - Read in incoming stream of objects. Once in a while read the entire event Web delivery of content in a distributed system is an interesting problem, but one with lots of existing tools - Early interest in Content Delivery Networks and other technologies capable of delivering a stream of data to lots of applications 26

Outlook First year of running on LHC went well - We are learning rapidly how to operate the infrastructure more efficiently - Making more dynamic use of the storage and making better use of the networking - 2011 and 2012 are long runs at the LHC Data volumes and user activities are both increasing 27