High Energy Physics data analysis

Similar documents
IllustraCve Example of Distributed Analysis in ATLAS Spanish Tier2 and Tier3

Considerations for a grid-based Physics Analysis Facility. Dietrich Liko

Andrea Sciabà CERN, Switzerland

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

150 million sensors deliver data. 40 million times per second

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

The LHC Computing Grid

Status of KISTI Tier2 Center for ALICE

Spanish Tier-2. Francisco Matorras (IFCA) Nicanor Colino (CIEMAT) F. Matorras N.Colino, Spain CMS T2,.6 March 2008"

The Grid: Processing the Data from the World s Largest Scientific Machine

HEP Grid Activities in China

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

The LHC Computing Grid. Slides mostly by: Dr Ian Bird LCG Project Leader 18 March 2008

The CMS Computing Model

The INFN Tier1. 1. INFN-CNAF, Italy

UW-ATLAS Experiences with Condor

SPINOSO Vincenzo. Optimization of the job submission and data access in a LHC Tier2

Austrian Federated WLCG Tier-2

ATLAS operations in the GridKa T1/T2 Cloud

CHIPP Phoenix Cluster Inauguration

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

PoS(EGICF12-EMITC2)106

Grid Computing Activities at KIT

Operating the Distributed NDGF Tier-1

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

where the Web was born Experience of Adding New Architectures to the LCG Production Environment

The ATLAS EventIndex: Full chain deployment and first operation

The ATLAS Tier-3 in Geneva and the Trigger Development Facility

First Experience with LCG. Board of Sponsors 3 rd April 2009

Computing Model Tier-2 Plans for Germany Relations to GridKa/Tier-1

Analisi Tier2 e Tier3 Esperienze ai Tier-2 Giacinto Donvito INFN-BARI

The National Analysis DESY

The ATLAS Distributed Analysis System

The grid for LHC Data Analysis

GRID COMPUTING APPLIED TO OFF-LINE AGATA DATA PROCESSING. 2nd EGAN School, December 2012, GSI Darmstadt, Germany

Philippe Charpentier PH Department CERN, Geneva

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 27 July 2010

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Tier2 Centre in Prague

The ATLAS EventIndex: an event catalogue for experiments collecting large amounts of data

PROOF-Condor integration for ATLAS

The LHC Computing Grid

Virtualizing a Batch. University Grid Center

System upgrade and future perspective for the operation of Tokyo Tier2 center. T. Nakamura, T. Mashimo, N. Matsui, H. Sakamoto and I.

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Presented by Manfred Alef Contributions of Jos van Wezel, Andreas Heiss

Distributed Data Management on the Grid. Mario Lassnig

CouchDB-based system for data management in a Grid environment Implementation and Experience

IEPSAS-Kosice: experiences in running LCG site

RUSSIAN DATA INTENSIVE GRID (RDIG): CURRENT STATUS AND PERSPECTIVES TOWARD NATIONAL GRID INITIATIVE

The Grid. Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop

Storage and I/O requirements of the LHC experiments

The Global Grid and the Local Analysis

The CMS experiment workflows on StoRM based storage at Tier-1 and Tier-2 centers

Constant monitoring of multi-site network connectivity at the Tokyo Tier2 center

Computing for LHC in Germany

ATLAS Data Challenge 2: A Massive Monte Carlo Production on the Grid

Storage Resource Sharing with CASTOR.

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

TAG Based Skimming In ATLAS

Tier-2 structure in Poland. R. Gokieli Institute for Nuclear Studies, Warsaw M. Witek Institute of Nuclear Physics, Cracow

Workload Management. Stefano Lacaprara. CMS Physics Week, FNAL, 12/16 April Department of Physics INFN and University of Padova

ATLAS 実験コンピューティングの現状と将来 - エクサバイトへの挑戦 坂本宏 東大 ICEPP

PIC Tier-1 Report. 1st Board for the follow-up of GRID Spain activities Madrid 5/03/2008. G. Merino and M. Delfino, PIC

PoS(ACAT2010)039. First sights on a non-grid end-user analysis model on Grid Infrastructure. Roberto Santinelli. Fabrizio Furano.

CESGA: General description of the institution. FINIS TERRAE: New HPC Supercomputer for RECETGA: Galicia s Science & Technology Network

UK Tier-2 site evolution for ATLAS. Alastair Dewhurst

ATLAS COMPUTING AT OU

Cluster Setup and Distributed File System

Support for multiple virtual organizations in the Romanian LCG Federation

Tier-2 DESY Volker Gülzow, Peter Wegner

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 21 July 2011

( PROPOSAL ) THE AGATA GRID COMPUTING MODEL FOR DATA MANAGEMENT AND DATA PROCESSING. version 0.6. July 2010 Revised January 2011

Summary of the LHC Computing Review

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

The ATLAS Production System

CMS Computing Model with Focus on German Tier1 Activities

The LHC Computing Grid Project in Spain (LCG-ES) Presentation to RECFA

Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU

Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008

Edinburgh (ECDF) Update

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN

CernVM-FS beyond LHC computing

Grid Operation at Tokyo Tier-2 Centre for ATLAS

Service Availability Monitor tests for ATLAS

Unified storage systems for distributed Tier-2 centres

High Throughput WAN Data Transfer with Hadoop-based Storage

RDMS CMS Computing Activities before the LHC start

The EU DataGrid Testbed

ATLAS Experiment and GCE

XRAY Grid TO BE OR NOT TO BE?

Monte Carlo Production on the Grid by the H1 Collaboration

Improved ATLAS HammerCloud Monitoring for Local Site Administration

New strategies of the LHC experiments to meet the computing requirements of the HL-LHC era

The Portuguese National Grid Initiative (INGRID)

Opportunities A Realistic Study of Costs Associated

ISTITUTO NAZIONALE DI FISICA NUCLEARE

Influence of Distributing a Tier-2 Data Storage on Physics Analysis

FREE SCIENTIFIC COMPUTING

Grid Computing: dealing with GB/s dataflows

Transcription:

escience Intrastructure T2-T3 T3 for High Energy Physics data analysis Presented by: Álvaro Fernandez Casani (Alvaro.Fernandez@ific.uv.es) IFIC Valencia (Spain) Santiago González de la Hoz, Gabriel Amorós, Álvaro Fernández, Mohamed Kaci, Alejandro Lamas, Luis March, Elena Oliver, José Salt, Javier Sánchez, Miguel Villaplana, Roger Vives

Outline Introduction to the LHC and ATLAS Computing Model LHC and ATLAS experiment The Event Data Model A hierarchical Computing Model ATLAS Spanish Tier-2 Distributed Tier-2 Resources MC production Data Movement and Network Storage Element Analysis Use cases User Support Tier-3 prototype at IFIC-Valencia Extension of our Tier-2 A PC farm outside grid for interactive analysis: Proof Typical use of a Tier-3 Spanish escience Initiative Conclusions Alvaro Fernández Casaní, CGW2008 14-October-2008 2

Large Hadron Collider (LHC) The LHC is p-p collider: s = 14 TeV and ʆ =10 34 cm -2 s -1 (10 35 in high lumi phase) There are 4 detectors: - 2 for general purposes: ATLAS and CMS - 1 for B physics: LHCb - 1 for heavy ions: ALICE Alvaro Fernández Casaní, CGW2008 14-October-2008 3

ATLAS Computing The offline computing: - Output event rate: 200 Hz ~ 10 9 events/year - Average event size (raw data): 1.6 MB/event Processing: - 40,000 of today s fastest PCs Storage: - Raw data recording rate 320 MB/sec - Accumulating at 5-8 PB/year A solution: Grid technologies Worldwide LHC Computing Grid (WLCG) ATLAS Data Challenge (DC) ATLAS Production System (ProdSys) Alvaro Fernández Casaní, CGW2008 14-October-2008 4

Introduction: Alvaro Fernández Casaní, CGW2008 14-October-2008 5

Introduction: Analysis Data Format Derived Physics Dataset (DPD) after many discussions last year in the context of the Analysis Forum will consist (for most analysis) of skimmed/slimmed/thinned AODs plus relevant blocks of computed quantities (such as invariant masses). Produced at Tier-1s and Tier-2s Stored in the same format as ESD and AOD at Tier-3s Therefore readable both from Athena and from ROOT Alvaro Fernández Casaní, CGW2008 14-October-2008 6

ATLAS Spanish Distributed Tier2 UAM SWE Cloud: Spain-Portugal Tier-1: PIC-Barcelona Tier-2: UAM, IFAE & IFIC LIP & Coimbra IFIC IFAE Enable Physics Analysis by Spanish ATLAS Users Tier-1s send AOD data to Tier-2s Continuous production of ATLAS MC events Tier-2s produce simulated data and send them to Tier-1s To contribute to ATLAS + LCG Computing Common Tasks Sustainable growth of infrastructure according to the scheduled ATLAS ramp-up and stable operation T1/T2 Relationship FTS (File Transfer System) channels are installed for these data for production use All other data transfers go through normal network routes In this model, a number of data management services are installed only at Tier-1s and act also on their associated Tier- 2s: VO Box, FTS channel server, Local file catalogue (part of Distributed Data Management) Alvaro Fernández Casaní, CGW2008 14-October-2008 7

Spanish Distributed Tier2: Resources Ramp-up of Tier-2 Resources (after LHC rescheduling) numbers are cumulative Evolution of ALL ATLAS T-2 resources according to the estimations made by ATLAS CB (October 2006) Spanish ATLAS T-2 assuming a contribution of a 5% to the whole effort Strong increase of resources Present resources of the Spanish ATLAS T-2 (October 08) New acquisitions in progress to get the pledged resources IFAE UAM IFIC TOTAL CPU (ksi2k) 201 275 132 608 Disk (TB) 104 100 36 240 Accounting values are normalized according to WLCG recommendations Alvaro Fernández Casaní, CGW2008 14-October-2008 8

Spanish Distributed Tier2: Monte Carlo production 2006 & 2007 -The production in T2-ES follows the same trend as LCG/EGEE (good performance of the ES-ATLAS-T2) -The ES-ATLAS-T2 average contribution to the total Data Production in LCG/EGEE was 2.8% in 2006-2007. Taking into account that 250 centers/institutes are participating, 10 of them are Tier-1 2008 - Since January-2008, ATLAS has migrated to PANDA executor -The production efficiency has been positively affected; the average efficiency was 50% and now is 75%-80% @ T2-ES -T2-ES contribution to jobs production is being 1.6% up to now Alvaro Fernández Casaní, CGW2008 14-October-2008 9

Spanish Distributed Tier2: Network and Data Transfer It is provided by the Spanish NREN RedIRIS - (Aug 08) Connection at 10 Gbps to University backbone - 10 Gbps among RedIRIS POP in Valencia, Madrid and Catalunya Atlas collaboration: - More than 9 PetaBytes (> 10 million of files) transferred in the last 6 months among Tiers - The ATLAS link requirement between Tier- 1 and Tier-2s has to be 50 MBytes/s (400 Mbps) in a real data taken scenario. Data transfer between Spanish Tier1 and Tier2. We reached 250 Mbps using gridftp between T1 -> T2 (CCRC 08) Data transfer from CASTOR (IFIC) for a TICAL private production. We reached 720 Mbps (plateau) In about 20 hours (4th March 08) High rate is possible Alvaro Fernández Casaní, CGW2008 14-October-2008 10

Spanish distributed Tier2: Storage Element System Distributed Tier2: UAM(25%), IFAE(25%) and IFIC(50%) IFIC IFAE UAM SE (Disk Storage) Lustre+StoRM dcache/disk+srm posix dcache Inside our Tier2 two SE options are used. In case that Storm/Lustre won t work as expected we will switch to dcache Alvaro Fernández Casaní, CGW2008 14-October-2008 11

IFIC- Storage Network Sun X4500 disk servers: 2 servers with 48x500Gb disks ~= 36 Tb (installing) Tier-2 infrastructure: 9 servers with 48x500Gb disks ~= 216 Tb (installing) Grid-CSIC infrastructure: 5 servers with 48x1Tb disks ~= 200 Tb Configuration: RAID 5 (5x8 + 1x6). Usage ration 80% 1 raid per disk controller Performance (Bonie++ tests) Write: 444,585 Kb/s Read: 1,777,488 Kb/s Network performance: 4 Gb ports. OSS tested with channel bonding configuration ocupies link over 4 clients Alvaro Fernández Casaní, CGW2008 14-October-2008 12

StoRM Storage Element system at IFIC-Valencia Posix SRM v2 (server on Lustre) Being used in our IFIC-Tier2. Endpoint: srm://srmv2.ific.uv.es:8443:/srm/mana gerv2 Lustre in production in our Tier2 High performance file system Standard file system, easy to use Higher IO capacity due to the cluster file system Used in supercomputer centers Free version available Direct access from WN www.lustre.org https://twiki.ific.uv.es/twiki/bin/view/atlas/lustrestorm Alvaro Fernández Casaní, CGW2008 14-October-2008 13

Tests: RFIO without Castor Athena analysis 12.0.6 AOD's 4 MB/s CPU limited and Athena Alvaro Fernández Casaní, CGW2008 14-October-2008 14

Aplication: Atlas Distributed Analysis using the Grid Heterogeneous grid environment based on 3 grid infrastructures: OSG, EGEE, Nordugrid Grids have different middleware, replica catalogs and tools to submit jobs. Naive assumption: Grid ~large batch system Provide complicated job configuration jdl file (Job Description Language) Find suitable ATLAS (Athena) software, installed as distribution kits in the Grid Locate the data on different storage elements Job splitting, monitoring and book-keeping Etc.. NEED FOR AUTOMATION AND INTEGRATION OF VARIOUS DIFFERENT COMPONENTS We have for that Two Frontends: Panda & Ganga Alvaro Fernández Casaní, CGW2008 14-October-2008 15

Aplication: Atlas Distributed Analysis using the Grid Current Situation A user-friendly job definition and management tool Allows simple switching between testing on a local batch system and large-scale data processing on distributed resources (Grid) Developed in the context of ATLAS and LHCb Python framework Support for development work from UK (PPARCG/GridPP), Germany (D-Grid) and EU (EGEE/ARDA) Ganga is based on a simple, but flexible, job abstraction A job is constructed from a set of building blocks, not all required for every job Ganga offers three ways of user interaction: Shell command line Interactive IPython shell Graphical User Interface Alvaro Fernández Casaní, CGW2008 14-October-2008 16

Spanish Distributed Tier2: User Support A twiki web page has been provided in order to guide new users on how to do for using the grid and local ressources for the atlas data analysis https://twiki.ific.uv.es/twiki/bin/view/atlas/atlastier2 It is envisaging the introduction of a Ticketing-System-like Alvaro Fernández Casaní, CGW2008 14-October-2008 17

What is an ATLAS Tier3? Summary of ATLAS Tier3 workshop in January 2008 (https://twiki.cern.ch/twiki/bin/view/atlas/tier3taskforce) These have many forms: No single answer Size Very small ones (one professor and two students) Very large ones (regional/national centers) Connectivity and Access to data A Tier3 next to a Tier1/2 looks different than a Tier3 in the middle of nowhere Basically represent resources not for general ATLAS usage Some fraction of T1/T2 resources Local University clusters Desktop/laptop machines Tier-3 task force provided recommended solutions (plural) http://indico.cern.ch/getfile.py/access?contribid=30&sessionid=14&resid=0&m aterialid=slides&confid=22132 Alvaro Fernández Casaní, CGW2008 14-October-2008 18

Tier3 IFIC prototype: user access - Discussed in ATLAS Tier3 task force and currently taken: (https://twiki.cern.ch/twiki/bin/view/atlas/atlascomputing?topic=tier3taskforce) a) To install some User Interfaces and at least one CE dedicated to the Tier3: To have the ATLAS software (production releases & DDM tools) installed automatically The user has to login in the UI s and they can send jobs to the Grid It is possible to ask for development releases installation In our case, every UI can see Lustre (/lustre/ific.uv.es/grid) as a local file system (Useful to read files). b) SEVIEW developed for https access c) The Ganga client is installed in AFS Alvaro Fernández Casaní, CGW2008 14-October-2008 19

Tier3 IFIC prototype: (PC farm outside Grid) - Interactive analysis on DPD using ROOT-PROOF a) Outside the Grid b) ~20-30 nodes 4 machines to install PROOF and make some tests with Lustre: DELL PE1950 III 2 Quad-Core Xeon E5430 2.66GHz/2x6MB 1333FSb 16 GB RAM (2GB per core; 8x2GB) 2 HD 146 GB (15000 rpm) TESTS (Johannes Elmsheuser munich): The Lustre filesystem shows a nearly equivalent behaviour as the local storage. dcache performed in this test not as good as the others. dcache performance was not optimised, since many files were stored in the same pool node We could observe anearly linear speed up to 40 nodes. As expected, the average total time for one analysis is proportional to the number of users. Johannes Elmsheuser Alvaro Fernández Casaní, CGW2008 14-October-2008 20

AFS / Lustre Desktop/Laptop prototype at IFIC RB/W MS CE WN WN WN WN WN Tier-2 Extra Tier-2 Resources coupled to Tier2) -Ways to submit our jobs to other grid sites -Tools to transfer data Tier-3 dispatcher Workers -Work with ATLAS software -Use of final analysis tools (i.e. root) -User disk space Alvaro Fernández Casaní, CGW2008 14-October-2008 21

Spanish escience Initiative See Talk by J. Marco on this conference NGI initiative to be part of EGI IFIC provides its T2-T3 infrastructure for HEP. Additionally more resources from GRID-CSIC to the NGI. IFIC Members in the Infrastructure and Middleware panels Study and support of aplications to be ported and supported on this einfrastructure (i.e- Medical Physics.HadronTherapy) Alvaro Fernández Casaní, CGW2008 14-October-2008 22

Conclusions escience Infrastructure for Tier2 and Tier3 was presented Distributed Tier-2 among 3 sites in Spain. October 08: 608 KSI2000 240 TB Tier-3 at IFIC configuration and software setup that matches the requirements according to the DPD analysis needs as formulated by the ATLAS analysis model group. Some local resources, beyond Tier-1s and Tier-2s, are required to do physics analysis in ATLAS.. IFIC is active member of the Spanish escience NGI. Commitee members at middleware and infrastructure panels. Lessons learned in high energy physics to be applied to new applications Alvaro Fernández Casaní, CGW2008 14-October-2008 23