The National Analysis DESY

Similar documents
The Global Grid and the Local Analysis

NAF & NUC reports. Y. Kemp for NAF admin team H. Stadie for NUC 4 th annual Alliance Workshop Dresden,

Computing / The DESY Grid Center

The German National Analysis Facility What it is and how to use it efficiently

Batch system usage arm euthen F azo he Z J. B T

DESY. Andreas Gellrich DESY DESY,

Computing for LHC in Germany

Site Report. Stephan Wiesand DESY -DV

Grid Computing Activities at KIT

DESY at the LHC. Klaus Mőnig. On behalf of the ATLAS, CMS and the Grid/Tier2 communities

HTCondor with KRB/AFS Setup and first experiences on the DESY interactive batch farm

Installation of CMSSW in the Grid DESY Computing Seminar May 17th, 2010 Wolf Behrenhoff, Christoph Wissing

ATLAS operations in the GridKa T1/T2 Cloud

Analisi Tier2 e Tier3 Esperienze ai Tier-2 Giacinto Donvito INFN-BARI

Tier-2 DESY Volker Gülzow, Peter Wegner

PROOF-Condor integration for ATLAS

SPINOSO Vincenzo. Optimization of the job submission and data access in a LHC Tier2

UW-ATLAS Experiences with Condor

Computing at DESY Zeuthen. Wolfgang Friebel

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

The Software Defined Online Storage System at the GridKa WLCG Tier-1 Center

and the GridKa mass storage system Jos van Wezel / GridKa

High Energy Physics data analysis

Status of KISTI Tier2 Center for ALICE

Edinburgh (ECDF) Update

Virtualizing a Batch. University Grid Center

CMS Computing Model with Focus on German Tier1 Activities

Considerations for a grid-based Physics Analysis Facility. Dietrich Liko

Grid Computing. Olivier Dadoun LAL, Orsay. Introduction & Parachute method. Socle 2006 Clermont-Ferrand Orsay)

The ATLAS Tier-3 in Geneva and the Trigger Development Facility

Austrian Federated WLCG Tier-2

Parallel Computing at DESY Zeuthen. Introduction to Parallel Computing at DESY Zeuthen and the new cluster machines

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Presented by Manfred Alef Contributions of Jos van Wezel, Andreas Heiss

Monitoring the Usage of the ZEUS Analysis Grid

A scalable storage element and its usage in HEP

DESY site report. HEPiX Spring 2016 at DESY. Yves Kemp, Peter van der Reest. Zeuthen,

Computing Model Tier-2 Plans for Germany Relations to GridKa/Tier-1

The Grid. Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop

OBTAINING AN ACCOUNT:

LCG data management at IN2P3 CC FTS SRM dcache HPSS

glite Grid Services Overview

Grid Computing: dealing with GB/s dataflows

Cluster Setup and Distributed File System

The Grid: Processing the Data from the World s Largest Scientific Machine

IEPSAS-Kosice: experiences in running LCG site

Challenges of the LHC Computing Grid by the CMS experiment

HEP Grid Activities in China

Service Availability Monitor tests for ATLAS

Virtualization of the ATLAS Tier-2/3 environment on the HPC cluster NEMO

AGATA Analysis on the GRID

Bookkeeping and submission tools prototype. L. Tomassetti on behalf of distributed computing group

dcache Introduction Course

150 million sensors deliver data. 40 million times per second

A validation system for Data Preservation in HEP. Marco Strutz

Tutorial for CMS Users: Data Analysis on the Grid with CRAB

Application of Virtualization Technologies & CernVM. Benedikt Hegner CERN

The Wuppertal Tier-2 Center and recent software developments on Job Monitoring for ATLAS

Data Challenges in Photon Science. Manuela Kuhn GridKa School 2016 Karlsruhe, 29th August 2016

XRAY Grid TO BE OR NOT TO BE?

Summary of the LHC Computing Review

Introduction to High Performance Computing and an Statistical Genetics Application on the Janus Supercomputer. Purpose

Outline. ASP 2012 Grid School

The LHC Computing Grid

Data Access and Data Management

Monitoring System for the GRID Monte Carlo Mass Production in the H1 Experiment at DESY

Scientific data management

dcache Introduction Course

Introduction to HPC Using zcluster at GACRC

Grid Engine - A Batch System for DESY. Andreas Haupt, Peter Wegner DESY Zeuthen

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

Long term data preservation and virtualization

DQ2 - Data distribution with DQ2 in Atlas

Operating two InfiniBand grid clusters over 28 km distance

The LHC Computing Grid

The ATLAS Production System

Computing. DOE Program Review SLAC. Rainer Bartoldus. Breakout Session 3 June BaBar Deputy Computing Coordinator

Virtualization. A very short summary by Owen Synge

Editorial Manager(tm) for Journal of Physics: Conference Series Manuscript Draft

Grid Computing: dealing with GB/s dataflows

Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

HPC system startup manual (version 1.20)

Computing at Belle II

DIRAC pilot framework and the DIRAC Workload Management System

Computing at the Large Hadron Collider. Frank Würthwein. Professor of Physics University of California San Diego November 15th, 2013

VMs at a Tier-1 site. EGEE 09, Sander Klous, Nikhef

Grid Computing. Olivier Dadoun LAL, Orsay Introduction & Parachute method. APC-Grid February 2007

Integration of Cloud and Grid Middleware at DGRZR

High Performance Computing (HPC) Using zcluster at GACRC

Database Services at CERN with Oracle 10g RAC and ASM on Commodity HW

Tier-2 structure in Poland. R. Gokieli Institute for Nuclear Studies, Warsaw M. Witek Institute of Nuclear Physics, Cracow

A Long-distance InfiniBand Interconnection between two Clusters in Production Use

Monte Carlo Production on the Grid by the H1 Collaboration

Using a dynamic data federation for running Belle-II simulation applications in a distributed cloud environment

CMS Tier-2 Program for user Analysis Computing on the Open Science Grid Frank Würthwein UCSD Goals & Status

Spanish Tier-2. Francisco Matorras (IFCA) Nicanor Colino (CIEMAT) F. Matorras N.Colino, Spain CMS T2,.6 March 2008"

Introduction to HPC Using zcluster at GACRC

Storage and I/O requirements of the LHC experiments

Slurm basics. Summer Kickstart June slide 1 of 49

Transcription:

The National Analysis Facility @ DESY Yves Kemp for the NAF team DESY IT Hamburg & DV Zeuthen 10.9.2008 GridKA School

NAF: National Analysis Facility Why a talk about an Analysis Facility at a Grid School? This talk will show you what kind of requirements for computing resources Analysis has where the Grid can meet them and where it cannot planning details of the NAF GridKA 10.9.08 NAF @ DESY Yves Kemp 2

HEP Computing: Data centric Collisions at 40 MHz Recorded at ~100 Hz Total RAW data by all LHC experiments: 15 PB/year Balloon (30 Km) CD stack with 1 year LHC data! (~ 20 Km) Concorde (15 Km) Mt. Blanc (4.8 Km) Atlas pp collision GridKA 10.9.08 NAF @ DESY Yves Kemp 3

Different tasks: Different requirements Coordinated & global tasks MC Production Event Generation: no I; small O; little CPU Detector Simulation: small I; large O & CPU Event Reconstruction/Reprocessing Reprocessing: full I; full O; large CPU Selections: large I; large O; large CPU Uncoordinated, unstructured & local tasks Analysis Usually: large I; small O; little CPU Performed by many users, many times! LHC StartUp phase: Short turn-around GridKA 10.9.08 NAF @ DESY Yves Kemp 4

The Grid: Distribute Data over distributed computing centers GridKA 10.9.08 NAF @ DESY Yves Kemp 5

GridKA DESY Each layer: Specialized for certain tasks e.g. T2: -Analysis -User access -AOD storage DESY Grid: Serves Atlas&CMS ILC & Calice HERA-Experiments

Do we need something in addition? Grid and the Tier model well suited for Global & coordinated tasks Analysis Local & uncoordinated, unstructured Provide best possible infrastructure and tools for German researchers In addition to global Grid resources Join forces and create synergies among German scientists The NAF: National Analysis Facility Located at DESY: Data is there GridKA 10.9.08 NAF @ DESY Yves Kemp 7

The frame for the NAF: The NAF is part of the Strategic Helmholtz Alliance More: http://terascale.desy.de/ Only accessible by German research groups for LHC and ILC tasks Planned for a size of about 1.5 av. Tier 2, but with more data Starting as joint activity @ DESY Requirements papers from German Atlas and CMS groups GridKA 10.9.08 NAF @ DESY Yves Kemp 8

Starting with Atlas & CMS Requirement papers. Some points: Interactive login Code development & testing, Experiment SW and tools Uniform access Central registry Batch-like resources: Local access: short queue, for testing purpose Large part (only) available via Gridmechanisms Fast response wanted for local&grid Hosted Data: Personal/group storage AFS home directories (and access to other AFS cells) High-capacity /High-bandwidth storage Local part (potentially with backup) Grid part: Enlargement of the T2 part AODs (Full set in case for Atlas, maybe trade some for ESD?) TAG database User/Group data Additional services PROOF farm, with connection to highbandwidth storage Flexible setup Allows reassignment of hosts between different types of services GridKA 10.9.08 NAF @ DESY Yves Kemp 9

Infrastructure building blocks NAF gsissh qsub Interactive Local Batch Proof NAF Grid I n t e r c o n n e c t AFS Parallel Cluster FS Dedicated space AFS/Kerberos scp SRM? grid-submit Grid Cluster DESY Grid dcache Storage Grid-ftp, SRM GridKA 10.9.08 NAF @ DESY Yves Kemp 10

Grid Part of NAF Use VOMS!! voms-proxy-init --voms atlas cms atlas:/atlas/de cms:/cms/dcms NAF Grid ressources integrated into DESY Grid Cluster Separate Fairshare and Priority for German users Access to storage based on VOMS groups/roles to come! GridKA 10.9.08 NAF @ DESY Yves Kemp 11

NAF login, interactive gsissh proxy proxy certificate based login mapping VO1 login nodes afs account VO1 (e.g. schmidt) mapping VO2 afs account VO2 (e.g. schmidtilc) registry Compute Nodes: - Workgroup servers - Local batch nodes registry administrator VO member accounts shell access rights quotas resources user home directories AFS naf.desy.de group directories back-up GridKA 10.9.08 NAF @ DESY Yves Kemp 12

IO and Storage New AFS cell: naf.desy.de User & Working group directories Special software area Safe and distributed storage Cluster File System High Bandwidth (O(GB/s)) to large Storage (O(10TB) Copy data from Grid, process data, save results to AFS or Grid Scratch-like space, lifetime t.b.d., but longer than typical job Locally connected via InfiniBand, remote access via TCP/IP dcache Well-known product and access methods Central entry point for data import and exchange Special space for German users GridKA 10.9.08 NAF @ DESY Yves Kemp 13

Storage organization ATLAS DESY has 100% of the AODs : Distributed between HH and Zn More than the nominal T2 pledge: Additions are the NAF part RDO/ESD at a smaller level, if requested and if space CMS Concept of T2 hosting an analysis DESY-HH try to host as many analysis as possible Have all interesting data for physics ILC/CALICE Already have MC data (ILC) and real data (CALICE) But at a smaller scale dcache SE to host these data! GridKA 10.9.08 NAF @ DESY Yves Kemp 14

Access to storage AFS well known product, access clear Lustre: is a cluster filesystem Use as a normal filesystem (OK: some limitations concerning locking and handling of many small files ) dcache: different access methods: Via Grid methods: LFC /pnfs mount: BUT: Security and performance problems! DESY summer student Malte Nuhn: Development of secure and low-resource consuming tools for replacing /pnfs mount GridKA 10.9.08 NAF @ DESY Yves Kemp 15

Software Experiment specific software: Grid and Interactive world: DESY provides space and tools: Experiments install their software themselves Because of current nature of Grid and Interactive parts: Two different areas Common software: Grid world: Standard worker node installation Interactive world: Compilers, debuggers ROOT, CERNLIB Operation System: Currently all Grid WNs on SL4 (64 bit) InteractiveSL4 (64bit) (some SL5 testing machines). No SL3 GridKA 10.9.08 NAF @ DESY Yves Kemp 16

PROOF: Experience from CMS Courtesy of Wolf Behrenhoff GridKA 10.9.08 NAF @ DESY Yves Kemp 17

PROOF from CMS, cntd. running on SGE batch farm access to Lustre, dcache,... every user starts his/her own PROOF cluster crashes, segfaults,... never affect others no need to deal with authentication, permissions,... simple setup, scripted start/stop no version/compatibility problems. One user can run 5.14 (CMSSW 1.x), others can use 5.18 at the same time First user doing real analysis since July 4th Courtesy of Wolf Behrenhoff More information: See PROOF/ROOT tutorial for general PROOF or attend the CMS course this afternoon GridKA 10.9.08 NAF @ DESY Yves Kemp 18

Support, Documentation, NUC Docu Main entry point: http://naf.desy.de/ Links to experiment-specific pages linked from here Support General entry point: naf-helpdesk@desy.de Experiment-specific support: See their docu NAF Users Committee NUC Atlas: Wolfgang Ehrenfeld, DESY Jan-Erik Sundermann, Freiburg CMS: Hartmut Stadie, Uni-HH Carsten Hof, Aachen GridKA 10.9.08 NAF @ DESY Yves Kemp 19

Summary & Outlook NAF already has many active users All building blocks in place Still tuning needed for some Additional services wanted e.g. TAG-DB for ATLAS: to come YOU should get an account (if you not already have one:-)) CMS tutorial this afternoon on NAF ATLAS tutorial next week @ Munich on NAF GridKA 10.9.08 NAF @ DESY Yves Kemp 20

Backup: Current and prospected hardware NAF-Batch: currently 264 cores (HH+Zn) 2008: HH +256 ; Zn +128 NAF-Grid: German groups have each: 10% of 1262 cores fairshare. Lustre: ~60 TB (in HH) dcache: (T2 & NAF!!) Enlargement of HH 480 TB / ZN 90TB in 2008 + other backbone systems GridKA 10.9.08 NAF @ DESY Yves Kemp 21