Computing / The DESY Grid Center

Similar documents
The National Analysis DESY

DESY. Andreas Gellrich DESY DESY,

Computing for LHC in Germany

NAF & NUC reports. Y. Kemp for NAF admin team H. Stadie for NUC 4 th annual Alliance Workshop Dresden,

DESY at the LHC. Klaus Mőnig. On behalf of the ATLAS, CMS and the Grid/Tier2 communities

Status of KISTI Tier2 Center for ALICE

The Global Grid and the Local Analysis

Site Report. Stephan Wiesand DESY -DV

The LHC Computing Grid

Grid Computing Activities at KIT

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Operating the Distributed NDGF Tier-1

CMS Computing Model with Focus on German Tier1 Activities

Grid and Cloud Activities in KISTI

Austrian Federated WLCG Tier-2

dcache, activities Patrick Fuhrmann 14 April 2010 Wuppertal, DE 4. dcache Workshop dcache.org

LHCb Computing Strategy

Distributing storage of LHC data - in the nordic countries

Tier-2 DESY Volker Gülzow, Peter Wegner

Spanish Tier-2. Francisco Matorras (IFCA) Nicanor Colino (CIEMAT) F. Matorras N.Colino, Spain CMS T2,.6 March 2008"

Summary of the LHC Computing Review

150 million sensors deliver data. 40 million times per second

First Experience with LCG. Board of Sponsors 3 rd April 2009

Monte Carlo Production on the Grid by the H1 Collaboration

Distributed Computing Framework. A. Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

ARC integration for CMS

IEPSAS-Kosice: experiences in running LCG site

RUSSIAN DATA INTENSIVE GRID (RDIG): CURRENT STATUS AND PERSPECTIVES TOWARD NATIONAL GRID INITIATIVE

Virtualizing a Batch. University Grid Center

UW-ATLAS Experiences with Condor

CHIPP Phoenix Cluster Inauguration

Lessons Learned in the NorduGrid Federation

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

The LHC Computing Grid. Slides mostly by: Dr Ian Bird LCG Project Leader 18 March 2008

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

The German National Analysis Facility What it is and how to use it efficiently

Challenges of the LHC Computing Grid by the CMS experiment

Batch Services at CERN: Status and Future Evolution

The Grid. Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop

The ATLAS Tier-3 in Geneva and the Trigger Development Facility

Monitoring System for the GRID Monte Carlo Mass Production in the H1 Experiment at DESY

Batch system usage arm euthen F azo he Z J. B T

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Presented by Manfred Alef Contributions of Jos van Wezel, Andreas Heiss

The Grid: Processing the Data from the World s Largest Scientific Machine

Edinburgh (ECDF) Update

Virtualization. A very short summary by Owen Synge

Installation of CMSSW in the Grid DESY Computing Seminar May 17th, 2010 Wolf Behrenhoff, Christoph Wissing

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

Preparing for High-Luminosity LHC. Bob Jones CERN Bob.Jones <at> cern.ch

ATLAS operations in the GridKa T1/T2 Cloud

A short introduction to the Worldwide LHC Computing Grid. Maarten Litmaath (CERN)

and the GridKa mass storage system Jos van Wezel / GridKa

CMS Belgian T2. G. Bruno UCL, Louvain, Belgium on behalf of the CMS Belgian T2 community. GridKa T1/2 meeting, Karlsruhe Germany February

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

CernVM-FS beyond LHC computing

The INFN Tier1. 1. INFN-CNAF, Italy

A distributed tier-1. International Conference on Computing in High Energy and Nuclear Physics (CHEP 07) IOP Publishing. c 2008 IOP Publishing Ltd 1

Visita delegazione ditte italiane

The grid for LHC Data Analysis

A scalable storage element and its usage in HEP

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback

ATLAS COMPUTING AT OU

Giovanni Lamanna LAPP - Laboratoire d'annecy-le-vieux de Physique des Particules, Université de Savoie, CNRS/IN2P3, Annecy-le-Vieux, France

Data services for LHC computing

HTCondor with KRB/AFS Setup and first experiences on the DESY interactive batch farm

CMS Tier-2 Program for user Analysis Computing on the Open Science Grid Frank Würthwein UCSD Goals & Status

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

HammerCloud: A Stress Testing System for Distributed Analysis

Pan-European Grid einfrastructure for LHC Experiments at CERN - SCL's Activities in EGEE

ATLAS Experiment and GCE

PROOF-Condor integration for ATLAS

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

Pre-Commercial Procurement project - HNSciCloud. 20 January 2015 Bob Jones, CERN

Data Analysis in ATLAS. Graeme Stewart with thanks to Attila Krasznahorkay and Johannes Elmsheuser

The ATLAS Production System

EMI Deployment Planning. C. Aiftimiei D. Dongiovanni INFN

Support for multiple virtual organizations in the Romanian LCG Federation

Long term data preservation and virtualization

Overview of ATLAS PanDA Workload Management

Monitoring the Usage of the ZEUS Analysis Grid

ATLAS Tier-3 UniGe

New strategies of the LHC experiments to meet the computing requirements of the HL-LHC era

dcache: challenges and opportunities when growing into new communities Paul Millar on behalf of the dcache team

Virtualization of the ATLAS Tier-2/3 environment on the HPC cluster NEMO

STATUS OF PLANS TO USE CONTAINERS IN THE WORLDWIDE LHC COMPUTING GRID

Patrick Fuhrmann (DESY)

DESY site report. HEPiX Spring 2016 at DESY. Yves Kemp, Peter van der Reest. Zeuthen,

dcache Introduction Course

EUDET Kickoff. Brief Overview of the Meeting at DESY E.Elsen

High Performance Computing from an EU perspective

Clouds at other sites T2-type computing

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN

Towards a Strategy for Data Sciences at UW

ALICE Grid Activities in US

High-Energy Physics Data-Storage Challenges

A validation system for Data Preservation in HEP. Marco Strutz

Clouds in High Energy Physics

Grid-related related Activities around KISTI Tier2 Center for ALICE

Transcription:

Computing / The DESY Grid Center Developing software for HEP - dcache - ILC software development The DESY Grid Center - NAF, DESY-HH and DESY-ZN Grid overview - Usage and outcome Yves Kemp for DESY IT & DESY DV 70. DESY PRC DESY Zeuthen 14.10.2010

dcache: Status dcache as a project: > Eight people at DESY Management, infrastructure, testing Development: NFS 4.1, Webinterface, main part > Two people at Fermilab Development: SRM, gplazma, Res. Manager > Two people at NDGF Development: Specific NDGF needs FTE Funding: The current situation: > Originally, developed at DESY for HERA > Now massively used in WLCG More than ½ of all LHC data on dcache SE In about 40 Tier 2 and 8/11 Tier 1 Total data stored in dcache (WLCG) ~20 PByte > One of the four middlewares in EMI EMI: European Middleware Initiative dcache project leader also EMI Data Area Leader DESY pays two developers and gets 10 in return Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 2

dcache, Future Future of dcache 1) New customers CFEL, European XFEL, LOFAR: LOw Frequency ARray in the Netherlands and Europe SNIC: Swedish National Infrastructure for Computing, many branches of Science will use a dcache system distributed over many Swedish CC 2) New developments Focusing on standard access protocols: Be competitive with industry solutions and benefit from their developments (e.g. WebDAV & NFS 4.1, also of interest in WLCG) EMI + HGF imposing extended code quality control 3) Sustained funding for next three years through EMI + HGF is guaranteed Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 3

ILC software development > Software support of ILC detector development: core software tools reconstruction software Grid production data storage software releases > People at DESY IT: two scientists, one engineer Working in close collaboration with DESY FLC group and KEK, SLAC, CERN, LLR,... > Recent activities: improve realism in detector simulation develop new tracking package develop Grid production system improvement of core tools (persistency, EDM, Conditions data,...) > Projects involved: EUDET - (ends 2010): task leader ANALYSIS AIDA starts 02/2011: coordination of WP2 Common Software & task leader Reconstruction http://ilcsoft.desy.de/ Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 4

ILC detector studies: Core Tools User community diverse: > ILD/CLIC detector studies > Calice calo testbeam > LC-TPC testbeam > EUDET - Pixel Telescope Overview of the different components and their interactions: Synergies between testbeam & global detector optimization Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 5

DESY GridCenter DESY GridCenter combining Grid / LHC Tier-2 and the National Analysis Facility + + Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 6

Which communities? > HERA experiments: H1, Zeus, Hermes, HERA-B > LHC experiments: ATLAS, CMS, LHCb > Future HEP experiments: ILC, CALICE > Astro: IceCube, CTA > Theory: ILDG > DESY photon science (CFEL, XFEL, ) > Monitoring and training VOs > DESY Grid Center platform for many diverse DESY activities. > Many different services needed see next slide Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 7

Which services? > Providing Data storage > Providing Computing power > Providing additional services to DESY, national and global community: Grid User interface on AFS: Used at different institutes in Germany, no need for them to deploy and maintain an own version of it. Grid Workload Management Systems (WMS): Largest WMS resource in Germany with ~6000 jobs/day, approx. 1/3 DESY, 1/3 Germany, 1/3 World > Central VO services for (among others) HERA, IceCube & ILDG VOMS registry server, LFC file catalogue, > Additional VO-specific site services, like SQUID server, PHEDEX- Server, Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 8

Special NAF services > Access to experiment data on the Grid Storage Element > CPU cycles for analysis: interactive and local batch Complement the Grid resources New techniques like PROOF > Additional storage: Lustre parallel file system > Home directories on AFS, accessible from everywhere Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 9

NAF CPU > Different contributions: Core contribution (DESY, BMBF, D-Grid, Alliance ) DESY ATLAS group LHCb contribution CMS Uni-Hamburg each contributing to the same infrastructure. > Concept: Sharing infrastructure works NAF CPU usage 1.4.-30.9.2010 All participants within ~10% of contribution > Different expectations Financers want good occupation Users want many resources and (probably more important): no/small waiting time Experiment and users want fast access to data > Meet different expectations 1.4.2010 30.9.2010: 66% Occupancy Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 10

NAF CPU usage by institutes > The NAF is well accepted among the German user community > DESY plays a big part, but not the biggest one > UniHH have invested hardware in NAF, and using it extensively > Some non-german researches invited by experiments to fully complete their analysis chains Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 11

The best metric for NAF success: > First CMS analysis explicitly mentioning the NAF: > Constraining SUSY models with Fittino using measurements before, with and beyond the LHC (arxiv:0907.2589v1 [hep-ph]) also mentions NAF. > Many more analyses going on on the NAF: > ATLAS Minimum bias > Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 12

DESY Grid Occupancy > DESY CPU Grid: Shared by many VOs > and many different job profiles CPU intensive Network/Storage intensive Local RAM or local HD intensive April-now DESY-HH > Plots below are job-centric not taking into account other used resources > Sharing of resources works > Both DESY-HH and DESY-ZN fulfilling 2010 T2 MoU pledges. April-now DESY-ZN Running jobs Queued jobs Running jobs All jobs Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 13

dcache storage: > Central data store for all experiments ATLAS (HH+ZN) ATLAS (HH+ZN) > Used from the Grid CPU and NAF CPU cluster > LHC VOs have enormously raised volumes of data available CMS > Data availability key aspect of the NAF Focus on storage capacity and available bandwidth > T2 MoU Pledges: ATLAS: 740 TB @ 1.4.2010 CMS: 400 TB @ 1.4.2010 > Additional space for NAF usage! Used space Free space > ToDo next month: (Experiments and providers) (re)evaluate needs for NAF additions (re)evaluate needed bandwidth Develop tools to prevent cold data Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 14

ATLAS TAG DB > Idea: Have a database with a very condensed event format (~1kB/evt) > a very fast pre-selection of events as input for data analysis > DB at CERN, PIC, BNL, TRIUMPF and DESY (not Tier-2 duty!) > Based on Oracle, knowledge at DESY > Currently: ~9 TB used, 18 TB available CPU&DB resources sufficient for the moment > DESY service for global ATLAS! DE users all directed to DESY MC only at DESY, serving all requests Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 15

Reliability & Availability > Users expect to be able to always work on all subsystems > As a Tier-2 DESY has committed to: Note: This is responding to operational problems, not solving them. > From 1.4.-31.9: Both DESY-HH and DESY-ZN are above Tier-2 requirements Only includes Grid-CPU and dcache storage, for VO OPS (WLCG official accounting) > What about NAF? More difficult to measure: More systems, more complex system E.g. also waiting time in queue is subjectively folded into perceived availability and reliability more complex metric Trying to evaluate using simple user-like test jobs NAF successful 97.5% of the time NAF test jobs July 2010, ATLAS Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 16

Experiment s view on availability might vary: Example CMS > Additional to official WLCG tests: Check availability of CMS SW Checking availability of CMS site services like Squid DB > Also around ~97% Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 17

Grid: Readiness for LHC data analysis Global readiness: Example: ATLAS (slide from Guenter Duckeck) Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 18

Grid: Readiness for LHC data analysis Tier-2 Readiness: Example: CMS analysis jobs (last month) DESY Tier-2 #1 for analysis in CMS DESY RWTH Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 19

The multiple faces of support > Tier-2 / Grid Mainly contacted by expert users, usually via GGUS Direct contact with appropriate provider 2 nd level support team Also consulting experts from (mostly) DESY VOs for their work on the Grid > NAF facility problems User knowledge level more diverse and not local to DESY. Contact naf-helpdesk@desy.de: General 1 st level DESY support Either answers themselves or asks appropriate 2 nd level team Expert users ( VO admins ) know 2 nd level channels and can contact these directly Majority of first-level contacts outside DESY > NAF VO related problems Is organized by the experiments themselves, usually via mailing lists Very valuable work done by the German experiment groups themselves Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 20

GridCenter Review Task Force > FH director set up a task force to answer three questions: Critically assess the existing resources: Are they optimally set up and used? Which monitoring tools do we have for users, decision-makers and non-experts? What are the requirements and plans for 2011-2014 > The realm of the task force does not cover Tier-2 MoU numbers! > Timeline: August 2010-December 2010 > Participants from ATLAS, CMS and IT/DV providers 7 members in total Liaison to Grid project board and NAF users committee > Current status: bi-weekly meetings, helping preparation of this talk Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 21

Summary and outlook > With dcache, DESY provides important pillar for HEP computing and the Grid. > DESY IT software development plays important role for ILC > Grid masters LHC analysis storm > DESY GridCenter in good shape and well used DESY offering many services for DESY, German and Global communities NAF important element in German HEP analysis landscape. Most important metric: First LHC analyses performed on the NAF! Yves Kemp Computing / The DESY Grid Center 14.10.2010 Page 22