Evaluation of the computing resources required for a Nordic research exploitation of the LHC

Similar documents
ATLAS NOTE. December 4, ATLAS offline reconstruction timing improvements for run-2. The ATLAS Collaboration. Abstract

Prompt data reconstruction at the ATLAS experiment

Multi-threaded, discrete event simulation of distributed computing systems

Storage and I/O requirements of the LHC experiments

Reliability Engineering Analysis of ATLAS Data Reprocessing Campaigns

PoS(High-pT physics09)036

Data Management for the World s Largest Machine

First LHCb measurement with data from the LHC Run 2

IEPSAS-Kosice: experiences in running LCG site

Tracking and flavour tagging selection in the ATLAS High Level Trigger

The ATLAS Conditions Database Model for the Muon Spectrometer

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

The CMS Computing Model

Data Quality Monitoring Display for ATLAS experiment

Clustering and Reclustering HEP Data in Object Databases

Verification and Diagnostics Framework in ATLAS Trigger/DAQ

CMS Alignement and Calibration workflows: lesson learned and future plans

An SQL-based approach to physics analysis

The JINR Tier1 Site Simulation for Research and Development Purposes

A New Segment Building Algorithm for the Cathode Strip Chambers in the CMS Experiment

New strategies of the LHC experiments to meet the computing requirements of the HL-LHC era

Fast pattern recognition with the ATLAS L1Track trigger for the HL-LHC

Physics Analysis Tools for Beauty Physics in ATLAS

Grid Computing: dealing with GB/s dataflows

Andrea Sciabà CERN, Switzerland

ATLAS, CMS and LHCb Trigger systems for flavour physics

The GAP project: GPU applications for High Level Trigger and Medical Imaging

Virtualizing a Batch. University Grid Center

CLOUDS OF JINR, UNIVERSITY OF SOFIA AND INRNE JOIN TOGETHER

ATLAS Experiment and GCE

CMS High Level Trigger Timing Measurements

Summary of the LHC Computing Review

Grid Computing a new tool for science

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

UW-ATLAS Experiences with Condor

Monitoring of Computing Resource Use of Active Software Releases at ATLAS

PROOF-Condor integration for ATLAS

LHCb Computing Resources: 2018 requests and preview of 2019 requests

Modules and Front-End Electronics Developments for the ATLAS ITk Strips Upgrade

The LHC Computing Grid. Slides mostly by: Dr Ian Bird LCG Project Leader 18 March 2008

CMS Conference Report

The ATLAS EventIndex: an event catalogue for experiments collecting large amounts of data

Adding timing to the VELO

CMS - HLT Configuration Management System

Reprocessing DØ data with SAMGrid

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

Usage statistics and usage patterns on the NorduGrid: Analyzing the logging information collected on one of the largest production Grids of the world

Performance of Tracking, b-tagging and Jet/MET reconstruction at the CMS High Level Trigger

The BABAR Database: Challenges, Trends and Projections

The ATLAS Trigger Simulation with Legacy Software

Overview. About CERN 2 / 11

Data Reconstruction in Modern Particle Physics

AGIS: The ATLAS Grid Information System

CERN openlab II. CERN openlab and. Sverre Jarp CERN openlab CTO 16 September 2008

Alignment of the ATLAS Inner Detector tracking system

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN

First results from the LHCb Vertex Locator

The BaBar Computing Model *

A LVL2 Zero Suppression Algorithm for TRT Data

Future trends in distributed infrastructures the Nordic Tier-1 example

b-jet identification at High Level Trigger in CMS

CERN s Business Computing

International Cooperation in High Energy Physics. Barry Barish Caltech 30-Oct-06

arxiv:hep-ph/ v1 11 Mar 2002

The Database Driven ATLAS Trigger Configuration System

ISTITUTO NAZIONALE DI FISICA NUCLEARE

PoS(Baldin ISHEPP XXII)134

Atlas event data model optimization studies based on the use of segmented VArray in Objectivity/DB.

150 million sensors deliver data. 40 million times per second

Storage Resource Sharing with CASTOR.

CSCS CERN videoconference CFD applications

PoS(EPS-HEP2017)523. The CMS trigger in Run 2. Mia Tosi CERN

Detector Control LHC

An ATCA framework for the upgraded ATLAS read out electronics at the LHC

Atlantis: Visualization Tool in Particle Physics

arxiv: v1 [physics.ins-det] 1 Oct 2009

The ATLAS Tier-3 in Geneva and the Trigger Development Facility

Early experience with the Run 2 ATLAS analysis model

HEP replica management

CERN and Scientific Computing

Physics Analysis Software Framework for Belle II

Stephen J. Gowdy (CERN) 12 th September 2012 XLDB Conference FINDING THE HIGGS IN THE HAYSTACK(S)

The Fourth Level Trigger Online Reconstruction Farm of HERA-B 1

PARALLEL PROCESSING OF LARGE DATA SETS IN PARTICLE PHYSICS

ATLAS Tracking Detector Upgrade studies using the Fast Simulation Engine

Data handling and processing at the LHC experiments

Grid Computing: dealing with GB/s dataflows

CernVM-FS beyond LHC computing

Computing at the Large Hadron Collider. Frank Würthwein. Professor of Physics University of California San Diego November 15th, 2013

Tuning Pipelined Scientific Data Analyses for Efficient Multicore Execution

The CMS data quality monitoring software: experience and future prospects

DESY at the LHC. Klaus Mőnig. On behalf of the ATLAS, CMS and the Grid/Tier2 communities

A Simulation Model for Large Scale Distributed Systems

Batch Services at CERN: Status and Future Evolution

Unified System for Processing Real and Simulated Data in the ATLAS Experiment

Monte Carlo programs

Performance of the MRPC based Time Of Flight detector of ALICE at LHC

PoS(TIPP2014)204. Tracking at High Level Trigger in CMS. Mia TOSI Universitá degli Studi di Padova e INFN (IT)

Disentangling P ANDA s time-based data stream

Transcription:

PROCEEDINGS Evaluation of the computing resources required for a Nordic research exploitation of the LHC and Sverker Almehed, Chafik Driouichi, Paula Eerola, Ulf Mjörnmark, Oxana Smirnova,TorstenÅkesson Dept. of Elementary Particle Physics, Lund University, Box 8, 22 Lund, Sweden E-mail: christina.jarlskog@quark.lu.se, sverker.almehed@quark.lu.se, chafik.driouichi@quark.lu.se, paula.eerola@quark.lu.se, ulf.mjornmark@quark.lu.se, oxana.smirnova@quark.lu.se, torsten.akesson@quark.lu.se Abstract: A simulation study to evaluate the required computing resources for a research exploitation of the Large Hadron Collider (LHC) has been performed. The evaluation was done as a case study, assuming existence of a Nordic regional centre and using the requirements for performing a specific physics analysis as a yard-stick. Other input parameters were: assumption for the distribution of researchers at the institutions involved, an analysis model, and two different functional structures of the computing resources.. A case-study for the LHC data processing model The Large Hadron Collider (LHC) is the world s biggest accelerator, being built at the European Particle Physics Laboratory, CERN. By the time of its completion in 26, it will be capable of accelerating and colliding beams of protons at centre of mass energies of 4 TeV. LHC experiments expect to have a recorded raw data rate of about PetaByte per year at the beginning of the LHC operation []. The geographical spread of the collaboration increases the complexity of the access and analysis of the data. One of the important measurements at LHC will be that of the parameter sin(2β), where the angle β is an angle of the CKM unitarity triangle describing quark mixing. It is a central parameter to demonstrate CP violation in the B-meson system [2]. In order to measure this parameter one needs to reconstruct and tag decays B d J/ψK S. In addition to the signal events, one also needs to analyze control samples (J/ψK and J/ψK + events). The ATLAS detector [] will trigger such events at the low-luminosity running of LHC. PrHEP hep2 Speaker. On leave from JINR, 498 Dubna, Russia.

The high-level trigger [3], the event filter, consists of full event processing with off-line type software. The final event filter output rate of J/ψ decaying to µ + µ or e + e was estimated to give, in one day of data taking, a total of 3. 6 di-lepton events, approximately []. In the following, it is assumed that these events will be written out in a separate raw data stream. In order to cope with the LHC data analysis and storage requirements, a tiered hierarchy of distributed regional centres was proposed by the MONARC project (Models of Networked Analysis at Regional Centres for LHC Experiments) [4]. In this scheme, the main centre is CERN (Tier), where the data reconstruction is expected to take place. The Tier centres have a capacity next largest to CERN. Among the possible activities of the Tier s, the production and reconstruction of fully-simulated data requires significant resources. The data analysis and fast simulation will be mainly the responsibility of the Tier2 centres of a smaller capacity. Computer farms at institutions and workstations constitute lower tiers. Data recorded directly from the online stream, including the signals from the detector elements and the on-line reconstruction results (called raw data or RAW), are expected to reside at CERN, being stored on a mass storage. During the processing at this Tier centre, the raw data will first be run through a reconstruction program, which calculates charged particle trajectories and energy depositions in the calorimeters. The reconstruction results are called ESD (Event Summary Data). Further processing algorithms are used at the Tier to prepare AOD (Analysis Object Data), which contain reduced information from ESD, and tag data or TAG, which are a small set of variables describing the event. The information in the TAG data set is meant to be used for initial selection of the AOD data to be analyzed. The size of these data types per event is expected to be MB for RAW,. MB for ESD,. MB for AOD and. MB for TAG. The MONARC project developed a simulation tool [5] to model various configurations of regional centres. It allows to determine optimal resources and strategies needed to achieve the highest efficiency of tasks performed by users. In this paper, a MONARC simulation study to evaluate the required computing resources in the Nordic countries for a research exploitation of the LHC has been performed, using the measurement of sin(2β) as one of the several physics cases. The simulations addressed the processing and analysis required for one day of data-taking, which includes (a) reconstruction of RAW data (production of ESD, AOD and TAG data) at Tier (CERN), (b) analysis of AOD data at a Nordic Tier2 (or Tier), (c) fast simulation of AOD data at a Nordic Tier2 (or Tier) and (d) full simulation and reconstruction of RAW data at a Nordic Tier. The amount of data was assumed to correspond to one day of data-taking, i.e. about 3 million real data events and 6 million simulated events. It was assumed that this specific analysis will be conducted by four experimental groups in the Nordic countries: at the Niels Bohr Institute (NBI) in Copenhagen, at the University of Oslo, at the University of Bergen and at the University of Lund. Each experimental group was assumed to consist of five researchers. All the CPU power was placed in NBI, which was considered both in a Tier and in a Tier2 configuration. The other three institutes represent users of the computing power of NBI. When the NBI was considered to be a Tier2 centre, the full simulation was assumed to be performed in three Tier centres, producing two million events each. It was PrHEP hep2 2

assumed here that those Tier centres would be located in UK, France and CERN. 2. Reconstruction at the Tier (CERN) To evaluate the time needed to reconstruct 3. million of RAW events at CERN, 2 the whole batch was split into sub-jobs (a subjob being a task running on a single node),.5.5 and simulation runs were performed by varying the number of sub-jobs for the reconstruc- 2 4 6 8 2 4 6 tion chain. The jobs for the creation of ESD, AOD and TAG were made sequential in the simulation. There were 3 nodes assumed Figure : Total execution time for the reconstruction of the data at CERN. of 2 SPECint95 each. Due to the division into sub-jobs, the total number of events processed in the simulation was not always equal to 9.3 million events (where all types of data are taken into account). For this reason, the equivalent execution time was calculated by dividing the 9.3 million events by the processing rate given by each simulation run. This time is shown in Fig.. As Fig. suggests, the reconstruction task for the given channel at CERN can be performed well within one day. It is clear from the same plot that the number of sub-jobs can be optimized. 3. Tier2 analysis and simulation Total execution time (days) The analysis of AOD data was assumed to be performed by twenty researchers (five per institute), each analyzing a) b) the complete sample once in a number of sub-jobs of equal number of 2 4 2 4 /person /person events. The number of sub-jobs per c) d) person was varied as follows:, 5,, 5, 2 and 4. Nodes were assumed to 2 4 2 4 be single-processor of 2 SPECint95 /person /person each. It was assumed that the output e) of the analysis for each event was fifteen real numbers and five integer num- 2 4 /person bers characterizing the event (invariant mass, decay time etc), the output size Figure 2: Total execution time for the analysis of thus being estimated at 4 bytes per AOD data at Tier2, configured with different amount event. The time to analyze one event of nodes: a) nodes, b) 2 nodes, c) 3 nodes, was assumed to be 3 SPECint95 s. The d) 4 nodes and e) 8 nodes. execution time as a function of the number of sub-jobs submitted by each analyzer and for different numbers of nodes in the Tier2 isshowninfig.2. PrHEP hep2 3

The fast simulation of six million events was assumed to be performed by one operator once. The size of the simulated.5 a).5 b) data was assumed to be the same as that of the real data. It was calculated that the extra time to write events.5 c).5 d) 2 4 6 8 2 4 6 8 of twice this size to the databases would be of the order of a few minutes and 2 4 6 8 2 4 6 8 could therefore be neglected. The time to generate one event was estimated to.5 e) be 7 SPECint95 s. The execution time as a function of the number of sub-jobs 2 4 6 8 and for different numbers of nodes is given in Fig. 3. The overall conclusion Figure 3: Total execution time for the fast simulation of AOD data at Tier2, configured with different from Figs. 2 and 3 is that the optimum combination is to have as many amount of nodes: a) nodes, b) 2 nodes, c) 3 jobs as nodes and that a centre with nodes, d) 4 nodes and e) 8 nodes. 2 nodes would seem to be well suited for the tasks of analysis and fast simulation. It was shown in the study that the time required to transfer files by ftp to and from the Tier2 can be neglected for a WAN speed of 25 MB/s or more. 4. Tier study Full simulation of RAW data is the most demanding task as far as CPU time is concerned: 5 SPECint95 s per event 25 2 SI95/node were required. The CPU per node at the Tier was assumed 5 SI95/node 2 to be either 2 SPECint95 or 5 SPECint95. Generation of 5 6 million events was simulated. The execution time as a function of the number of jobs (equal to the number of nodes) is given in 5 Fig. 4. The best estimate for the execution time was 2 days and 2 4 6 8 (nodes) 9 hours for a Tier centre with 9 nodes of 5 SPECint95 per node. The execution time for the reconstruction of the fullysimulated RAW data is given in Table. Figure 4: Execution time for full simulation at the Other activities at the Tier can be analysis of AOD data Tier. and fast simulation, as assumed in the Tier2 case. The execution times for Tier nodes of 2 SPECint95 and of 5 SPECint95 (per node) are given in Table. The overall conclusion for the Tier study is that all activities can be performed within one day, except the full simulation of RAW data. PrHEP hep2 5. Conclusions The aim of this study was to evaluate the amount of computing resources needed to per- The time estimate is based on the present performance of the ATLAS full simulation program. 4

analysis, fast simulation, reconstruction, 3mln.events, 6mln.events, 6mln.events, 2 jobs & nodes 2 jobs & nodes 5 jobs & nodes Tier (I) 2h 36m 2h 57m 6h 23m Tier (II) h 8m h 2m 2h 52m Table : Execution times at Tier in different configurations: case (I) corresponds to 2 SPECint95 per node, and case (II) to 5 SPECint95 per node. form a particular physics analysis task at a future LHC experiment by several groups of researchers in the Nordic countries. The task in question was the measurement of CP violation in decays B d J/ψK S. The objective was to perform all the analysis on-fly, which implied that the data acquired in one day should be immediately processed and analyzed. The required capacities of the Tier centre (at CERN) and a Nordic regional centre were investigated. By assuming the CERN capacity as having 3 single-processor nodes with 2 SPECint95 per node, it was found that it is not only sufficient for performing the task in question, but can accommodate many more jobs. The Nordic regional centre was considered in Tier2 and Tier configurations. While it was shown that a Tier2 centre can perform data analysis and fast simulation of events with a rate faster than the data production rate at the LHC, this is not the case for the full-scale detector simulation at the Nordic Tier centre. A solution would be either to increase the size of the Nordic regional centre to several thousands of nodes, or to share the full simulation task with other Tier centres worldwide. The present analysis concerns only one particular high-energy physics task, while a regional centre will serve many other research groups not only in physics but also in other sciences. Therefore, the results have to be considered as a single typical use-case, one of many at a future Nordic Regional Computing Centre. Acknowledgements We would like to thank Krzysztof Sliwa, Laura Perini and Iosif Legrand for their assistance. PrHEP hep2 References [] ATLAS Collaboration, ATLAS Detector and Physics Performance Technical Design Report, CERN/LHCC/99-4. ATLAS TDR 4 (May 999), Vol. I. [2] For a review, see Y. Nir and H. Quinn in B Decays (ed. S. Stone), World Scientific 994, p. 362, or Ann. Rev. Nucl. and Part. Sci. 42, 2 (992). [3] ATLAS Collaboration, ATLAS High-Level Triggers. DAQ and DCS Technical Proposal, CERN/LHCC/2-7 (March 2). [4] MONARC Collaboration, Models of Networked Analysis at Regional Centres for LHC experiments (MONARC), Mid-project progress report, LCB 99-5. [5] MONARC Collaboration, Multi-threaded, discrete event simulation of distributed computing systems, presented by I. Legrand in CHEP2, to be published in CPC Journal special edition CHEP2. 5