I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

Similar documents
Analisi Tier2 e Tier3 Esperienze ai Tier-2 Giacinto Donvito INFN-BARI

SPINOSO Vincenzo. Optimization of the job submission and data access in a LHC Tier2

Cluster Setup and Distributed File System

Using Hadoop File System and MapReduce in a small/medium Grid site

The INFN Tier1. 1. INFN-CNAF, Italy

The Legnaro-Padova distributed Tier-2: challenges and results

ARC integration for CMS

High Throughput WAN Data Transfer with Hadoop-based Storage

Testing SLURM open source batch system for a Tierl/Tier2 HEP computing facility

Understanding StoRM: from introduction to internals

A scalable storage element and its usage in HEP

Edinburgh (ECDF) Update

Andrea Sciabà CERN, Switzerland

Service Availability Monitor tests for ATLAS

Status of KISTI Tier2 Center for ALICE

Austrian Federated WLCG Tier-2

The CMS experiment workflows on StoRM based storage at Tier-1 and Tier-2 centers

The glite middleware. Presented by John White EGEE-II JRA1 Dep. Manager On behalf of JRA1 Enabling Grids for E-sciencE

A High Availability Solution for GRID Services

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Presented by Manfred Alef Contributions of Jos van Wezel, Andreas Heiss

glite Grid Services Overview

Workload Management. Stefano Lacaprara. CMS Physics Week, FNAL, 12/16 April Department of Physics INFN and University of Padova

EGEE and Interoperation

CS60021: Scalable Data Mining. Sourangshu Bhattacharya

Distributed production managers meeting. Armando Fella on behalf of Italian distributed computing group

Spanish Tier-2. Francisco Matorras (IFCA) Nicanor Colino (CIEMAT) F. Matorras N.Colino, Spain CMS T2,.6 March 2008"

Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008

Data Management. Parallel Filesystems. Dr David Henty HPC Training and Support

Lecture 11 Hadoop & Spark

CMS Belgian T2. G. Bruno UCL, Louvain, Belgium on behalf of the CMS Belgian T2 community. GridKa T1/2 meeting, Karlsruhe Germany February

dcache Introduction Course

UW-ATLAS Experiences with Condor

Lessons Learned in the NorduGrid Federation

LCG data management at IN2P3 CC FTS SRM dcache HPSS

Beob Kyun KIM, Christophe BONNAUD {kyun, NSDC / KISTI

Parallel Storage Systems for Large-Scale Machines

INDIGO-DataCloud Architectural Overview

The Global Grid and the Local Analysis

ALICE Grid Activities in US

Bookkeeping and submission tools prototype. L. Tomassetti on behalf of distributed computing group

and the GridKa mass storage system Jos van Wezel / GridKa

The LHC Computing Grid

TITLE: PRE-REQUISITE THEORY. 1. Introduction to Hadoop. 2. Cluster. Implement sort algorithm and run it using HADOOP

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

A BigData Tour HDFS, Ceph and MapReduce

CMS Tier-2 Program for user Analysis Computing on the Open Science Grid Frank Würthwein UCSD Goals & Status

XRAY Grid TO BE OR NOT TO BE?

CMS users data management service integration and first experiences with its NoSQL data storage

August Li Qiang, Huang Qiulan, Sun Gongxing IHEP-CC. Supported by the National Natural Science Fund

Outline. ASP 2012 Grid School

Analytics in the cloud

Distributed Filesystem

Distributed Systems 16. Distributed File Systems II

Tier2 Centre in Prague

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

Data Access and Data Management

Challenges of the LHC Computing Grid by the CMS experiment

Clouds at other sites T2-type computing

GRID COMPUTING APPLIED TO OFF-LINE AGATA DATA PROCESSING. 2nd EGAN School, December 2012, GSI Darmstadt, Germany

Grid Architectural Models

Philippe Charpentier PH Department CERN, Geneva

Application of Virtualization Technologies & CernVM. Benedikt Hegner CERN

ISTITUTO NAZIONALE DI FISICA NUCLEARE

Data storage services at KEK/CRC -- status and plan

The Software Defined Online Storage System at the GridKa WLCG Tier-1 Center

Introduction to SRM. Riccardo Zappi 1

The National Analysis DESY

Monte Carlo Production on the Grid by the H1 Collaboration

Monitoring system for geographically distributed datacenters based on Openstack. Gioacchino Vino

Deploying virtualisation in a production grid

BIG DATA AND HADOOP ON THE ZFS STORAGE APPLIANCE

DESY. Andreas Gellrich DESY DESY,

Grid Infrastructure For Collaborative High Performance Scientific Computing

MI-PDB, MIE-PDB: Advanced Database Systems

PROOF-Condor integration for ATLAS

Data Transfers Between LHC Grid Sites Dorian Kcira

Integration of Cloud and Grid Middleware at DGRZR

Long Term Data Preservation for CDF at INFN-CNAF

The glite middleware. Ariel Garcia KIT

Virtualization. A very short summary by Owen Synge

The LHC Computing Grid

Optimization of Italian CMS Computing Centers via MIUR funded Research Projects

Storage for HPC, HPDA and Machine Learning (ML)

Data Grid Infrastructure for YBJ-ARGO Cosmic-Ray Project

g-eclipse A Framework for Accessing Grid Infrastructures Nicholas Loulloudes Trainer, University of Cyprus (loulloudes.n_at_cs.ucy.ac.

An Introduction to GPFS

Interconnect EGEE and CNGRID e-infrastructures

VMs at a Tier-1 site. EGEE 09, Sander Klous, Nikhef

CA485 Ray Walshe Google File System

Operating the Distributed NDGF Tier-1

EMC Forum EMC ViPR and ECS: A Lap Around Software-Defined Services

Clouds in High Energy Physics

Monitoring tools in EGEE

Implementing GRID interoperability

AGATA Analysis on the GRID

ATLAS Tier-3 UniGe

CHIPP Phoenix Cluster Inauguration

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

Distributed File Systems II

OPEN STORAGE IN THE ENTERPRISE with GlusterFS and Ceph

Transcription:

I Tier-3 di CMS-Italia: stato e prospettive Claudio Grandi Workshop CCR GRID 2011

Outline INFN Perugia Tier-3 R&D Computing centre: activities, storage and batch system CMS services: bottlenecks and workarounds INFN Bologna Tier-3 Computing centre: configuration, activities and resources Functionalities Conclusions 2

INFN Perugia Tier-3

INFN Perugia computing centre activities Computing centre was setup since 2004 Main experiments making use of the computing farm are CMS, Babar, SuperB, NA48/62, Theo, Erna, Glast, Ise, Virgo and Grid. ~ 50 active users for the last 6 months 4

Storage dcache based disk-only storage managing ~ 35 TB 14 pools of ~ 2TB: equal size pools for better load balancing 1 pool of 5 G dedicated for storage certification jobs 2.5 TB NFS for POSIX-like storage 5

Batch system Scheduler accepting local and Grid jobs Support of the local submission using CRAB ~ 80 % of CMS analysis jobs are submitted locally using CRAB Torque/Maui scheduling: 3 internal queues for each VO (VO-short, VO-medium and VO-long) 1 Grid queue for each VO 1 queue for certification jobs Maui fair share policy of ~ 260 cores in a private network (25% CMS and 75% others) 6

CMS services: bottlenecks and workarounds

Services Xen-based virtualization of services: SE-dCache admin node, LCG-CE (CREAM-CE on-going), UI and SLC5 nodes Phedex more than 30 TB of data transferred since 2007 dcache is used as SE since 2007 and has allowed to use the entire bandwidth available Squid server to access conditions data, NoSQL database (couchdb) instance for CRAB development and tests. 8

Bottlenecks Very small external bandwidth (100Mb) Low data transfer rate using Phedex Users are penalized by the use of the entire bandwidth Large number of CRAB jobs failing during the stage-out step 9

Workarounds Phedex Limit the farm traffic to 40-50 Mb during the working hours to avoid penalizing other users 2 Phedex configurations CRAB Limiting the GridFTP queue in dcache has reduced the failure rate by ~ 20% CRAB 3 will reduce much more the failure rate (stage-out asincrono) Tests the stage-out of users outputs in T2_IT_Bari and access them using xrootd from there (see Giacinto s talks): Promising results: From users point of view, the performances in accessing data residing in Bari are the same as accessing data residing in Perugia No known problems 10

R&D in Tier-3 Perugia

Storage optimization Looking for a storage system which can provide (compared to dcache): better read performance higher fault tolerance In collaboration with CMS Tier-2 Bari (Giacinto Donvito) Data replica in Lustre is not easy to configure (see Giacinto s talk in CCR workshop 2010) Hadoop can provide needed performance and scalability by means of commodity HW Storm can work only with filesystems which support ACL and it is not the case of HDFS BeStMan can be configured to work with HDFS 12

Why Hadoop? Small/medium sites use commodity hardware more than large sites HDFS provides: exposed to hardware failure. Native data replication (block replication) DataNode failure ~ transparent Rack awareness Splitting files in different DataNode when storing them in HDFS when reading them back. can improve the performance 13

HDFS-based SE schema - Tier-3 Perugia setup Remote user Site BDII BeStMan Gridmap-file Xrootd HDFS GridFTP HDFS-based SE Local user WN WN WN 14

Setup 1. GridFtp: Compile HDFS plugin for GridFTP used in OSG sites: http:// t2.unl.edu:8094/browser/gridftp_hdfs/src/ Start glite GridFTP server to read/write from/to HDFS: globusgridftp-server -p 2811 -dsi hdfs -debug 2. SRM v2.2: Load fuse module in order to be able to mount HDFS (hadoopfuse-0.19.1-15.el5) Install BeStMan (2.2.1.3) and configure it using --with-gridmap-pathlocal and --with-tokens-list (to manage experiments storage area in HDFS) options 3. Install xrootd/hdfs: rpm -ivh xrootd-1.3.2-6.el5.x86_64.rpm --nodeps (rpms downloaded from here http://vdt.cs.wisc.edu/hadoop/testing/1.0/ rhel5.5//x86_64/) 4. Provider script: publishes dynamic information by calling srm-ping command. 15

Future directions System deployment in INFN-Grid production: Develop scripts for system configuration using INFN-Grid profile Maintain the system Monitor the system measuring its performance in the production environment System tuning: Tune the GridFTP plugin Tune the system configuration 16

Other R&D WN on demand: dynamic virtual machine manager for heterogeneous execution environment Optimize the use of computing resources for specific purposes applications Performance optimization of Maui cluster scheduler Development of a tool to predict the system behavior when a new set of maui configuration parameters is set 17

INFN Bologna Tier-3

INFN-Bologna Tier-3 Joint project of INFN and Università di Bologna Supports local research groups including LHC experiments Integrated with the INFN-CNAF Tier-1 infrastructure ~ 50 dual quad-core boxes accessible through CNAF-LSF batch system and managed through the WNoDeS system 150 TB on CNAF-GPFS storage system home directories and software areas mounted through CNFS The boundaries of the Tier-1 and Tier-3 will be virtual Each site can expand into the other according to policies in LSF The profile of the virtual machine defines to which site it belongs Computing and Storage Elements are independent Differently from the Tier-1 also offers interactive services to local users The site is certified and currently is being commissioned for the experiments 19

Functionalities Standard Grid site for the supported experiments and for the local researchers of the Università di Bologna E.g. for CMS: PhEDEx service, standard software installations, CMS SAM/Nagios tests,... Local users support: Direct access to a dedicated LSF queue Read/Write access to a portion of the GPFS storage Read access to the SE area (Write through SRM) No access to the Tier-1 queues and storage! Interactive access to a few nodes R&D: Virtual Interactive Pools (VIP) UI on demand managed by the WNoDeS system 20

Conclusions We have described 2 different configurations of Tier-3 sites, each one has its own issues and development directions: 1. INFN Perugia Tier-3: Computing centre was setup since 2004 Support of CMS services: CRAB development Local support of CRAB Data transfer using Phedex Small external bandwidth CRAB jobs failing during the stage-out step Tuning dcache configuration has reduced the failure rate CRAB 3 will reduce much more the failure rate First tests to access outputs stored at Tier-2 Bari using xrootd seem promising The functionalities tests of HDFS based SE in INFN-Grid environment are done with success. 21

Conclusions 2. INFN Bologna Tier-3: There is a clear benefit in term of hardware provisioning in being attached to a big site High quality hardware and infrastructure All basic services managed at the Tier-1 level by CNAF people: GPFS, LSF, Squids, monitoring,... The complexity increases and a good communication is needed Strong division of responsibilities: often an intervention requires people both from CNAF and Bologna Unforeseen interactions between components: need to be careful not to impact the Tier-1 operations Perfect playground for R&D activities High quality site without explicit duties to the collaboration 22

Backup 23

CMS jobs submission For the last 6 months Local jobs submission using PBS plugin of CRAB and Grid jobs submission are both supported ~ 80 % of analysis jobs are submitted locally using PBS plugin of CRAB Only ~ 50 % of CMS jobs submitted locally uses PBS plugin of CRAB