The LHC Computing Grid Project in Spain (LCG-ES) Presentation to RECFA

Similar documents
Summary of the LHC Computing Review

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Andrea Sciabà CERN, Switzerland

Computing for LHC in Germany

The Grid: Processing the Data from the World s Largest Scientific Machine

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

Spanish Tier-2. Francisco Matorras (IFCA) Nicanor Colino (CIEMAT) F. Matorras N.Colino, Spain CMS T2,.6 March 2008"

Data Analysis in Experimental Particle Physics

LHCb Computing Strategy

The LHC Computing Grid. Slides mostly by: Dr Ian Bird LCG Project Leader 18 March 2008

The LHC Computing Grid

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 27 July 2010

Preparing for High-Luminosity LHC. Bob Jones CERN Bob.Jones <at> cern.ch

CHIPP Phoenix Cluster Inauguration

The INFN Tier1. 1. INFN-CNAF, Italy

IEPSAS-Kosice: experiences in running LCG site

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

A Simulation Model for Large Scale Distributed Systems

CERN and Scientific Computing

The European DataGRID Production Testbed

Storage and I/O requirements of the LHC experiments

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Storage Resource Sharing with CASTOR.

CESGA: General description of the institution. FINIS TERRAE: New HPC Supercomputer for RECETGA: Galicia s Science & Technology Network

The grid for LHC Data Analysis

ELFms industrialisation plans

The Grid. Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop

CERN openlab II. CERN openlab and. Sverre Jarp CERN openlab CTO 16 September 2008

High Energy Physics data analysis

HEP Grid Activities in China

Virtualizing a Batch. University Grid Center

CC-IN2P3: A High Performance Data Center for Research

where the Web was born Experience of Adding New Architectures to the LCG Production Environment

The CMS Computing Model

Storage on the Lunatic Fringe. Thomas M. Ruwart University of Minnesota Digital Technology Center Intelligent Storage Consortium

DESY at the LHC. Klaus Mőnig. On behalf of the ATLAS, CMS and the Grid/Tier2 communities

ISTITUTO NAZIONALE DI FISICA NUCLEARE

Travelling securely on the Grid to the origin of the Universe

Giovanni Lamanna LAPP - Laboratoire d'annecy-le-vieux de Physique des Particules, Université de Savoie, CNRS/IN2P3, Annecy-le-Vieux, France

RUSSIAN DATA INTENSIVE GRID (RDIG): CURRENT STATUS AND PERSPECTIVES TOWARD NATIONAL GRID INITIATIVE

e-infrastructures in FP7 INFO DAY - Paris

MONTE CARLO SIMULATION FOR RADIOTHERAPY IN A DISTRIBUTED COMPUTING ENVIRONMENT

High Performance Computing from an EU perspective

High Performance Computing Course Notes Grid Computing I

150 million sensors deliver data. 40 million times per second

Computing / The DESY Grid Center

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

GRIDS INTRODUCTION TO GRID INFRASTRUCTURES. Fabrizio Gagliardi

Status of KISTI Tier2 Center for ALICE

Batch Services at CERN: Status and Future Evolution

Insight: that s for NSA Decision making: that s for Google, Facebook. so they find the best way to push out adds and products

Support for multiple virtual organizations in the Romanian LCG Federation

The LHC Computing Grid

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

UW-ATLAS Experiences with Condor

Grid Challenges and Experience

DESY. Andreas Gellrich DESY DESY,

First Experience with LCG. Board of Sponsors 3 rd April 2009

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

High-Energy Physics Data-Storage Challenges

The EU DataGrid Testbed

Study of the viability of a Green Storage for the ALICE-T1. Eduardo Murrieta Técnico Académico: ICN - UNAM

Tier-2 structure in Poland. R. Gokieli Institute for Nuclear Studies, Warsaw M. Witek Institute of Nuclear Physics, Cracow

The European Data Grid Project

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 21 July 2011

Tier-2 DESY Volker Gülzow, Peter Wegner

The JINR Tier1 Site Simulation for Research and Development Purposes

Getting prepared for the LHC Run2: the PIC Tier-1 case

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

CSCS CERN videoconference CFD applications

Distributed Data Management with Storage Resource Broker in the UK

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

N. Marusov, I. Semenov

New strategies of the LHC experiments to meet the computing requirements of the HL-LHC era

Online data storage service strategy for the CERN computer Centre G. Cancio, D. Duellmann, M. Lamanna, A. Pace CERN, Geneva, Switzerland

Grid and Cloud Activities in KISTI

Big Data Analytics and the LHC

Grid Computing: dealing with GB/s dataflows

e-infrastructure: objectives and strategy in FP7

High Throughput WAN Data Transfer with Hadoop-based Storage

Distributed Monte Carlo Production for

Experience of the WLCG data management system from the first two years of the LHC data taking

Top Trends in DBMS & DW

CERN s Business Computing

High Performance Computing Data Management. Philippe Trautmann BDM High Performance Computing Global Research

Towards Network Awareness in LHC Computing

CMS Belgian T2. G. Bruno UCL, Louvain, Belgium on behalf of the CMS Belgian T2 community. GridKa T1/2 meeting, Karlsruhe Germany February

Metro Ethernet for Government Enhanced Connectivity Drives the Business Transformation of Government

Interconnect EGEE and CNGRID e-infrastructures

Experimental Computing. Frank Porter System Manager: Juan Barayoga

Computing Model Tier-2 Plans for Germany Relations to GridKa/Tier-1

Towards a Strategy for Data Sciences at UW

( PROPOSAL ) THE AGATA GRID COMPUTING MODEL FOR DATA MANAGEMENT AND DATA PROCESSING. version 0.6. July 2010 Revised January 2011

Data services for LHC computing

Programmable Information Highway (with no Traffic Jams)

ACCI Recommendations on Long Term Cyberinfrastructure Issues: Building Future Development

Tier2 Centre in Prague

ALHAD G. APTE, BARC 2nd GARUDA PARTNERS MEET ON 15th & 16th SEPT. 2006

A DAQ Architecture for Agata experiment

Transcription:

* Collaboration of the Generalitat de Catalunya, IFAE and Universitat Autònoma de Barcelona Port d The LHC Computing Grid Project in Spain (LCG-ES) Presentation to RECFA Prof. Manuel Delfino Coordinating Principal Investigator, LCG-ES Physics Department, Universitat Autònoma de Barcelona Director, Port d (PIC)*

LHC is multi-petabyte, Tera-object data-intensive scientific computing TeraBytes 14'000 12'000 10'000 8'000 6'000 4'000 2'000 0 Long Term Tape Storage Estimates Current Experiments Accumulation: 10 PB/year Signal/Background up to 1:10 12 COMPASS LHC 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 Year

Complex LHC data = more CPU/byte K SI95 5,000 4,000 3,000 2,000 1,000 Estimated CPU Capacity required at CERN LHC Moore s law some measure of the capacity technology advances provide for a constant number of processors or investment 0 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 Jan 2000: 3.5K SI95

1-100 GB/sec Detector 0.1 to 1 GB/sec Raw data LHC analysis will be driven by the demands of ~6000 physicists worldwide 35K SI95 Event Filter (selection & reconstruction) 1 PB / year 500 TB One Experiment Event Summary Data ~200 MB/sec 350K SI95 64 GB/sec Batch Physics Analysis 200 TB / year Processed Data ~100 MB/sec 250K SI95 Event Reconstruction analysis objects Event Simulation Interactive Data Analysis Thousands of scientists worldwide

Technology Domains for developing solutions for LHC Data Analysis DEVELOPER VIEW FABRIC GRID APPLICATION GRIDIFICATION USER VIEW

From physical to virtual communities User point of view: Virtual analysis communities Transparent Analysis Region Physical setup H γγ H µµ b-physics orldwide multi-cultural llaboration Transparent Analysis Region LHC Computing Grid Infrastructure Institutional, National, Geographical Setups Transparent Analysis Region

Benefits of Grid Infrastructure: User-centric view Grid Infrastructure Field-specific middleware Scientists Collaborating General middleware Collaborating Network and Computing Infrastructures Interoperating Security, Registration, Authentication and Authorization

The Internet connects Computers The Grid enables Virtual Communities Grid Services Architecture Application pecialized services : user- or pln-specific distributed services User anaging multiple resources : iquitous infrastructure services haring single resources : gotiating access, controlling use alking to things : communication nternet protocols) & security ontrolling things locally : Access, & control of, resources Collective Resource Connectivity Fabric Adapted by permission from Foster and Kesselman Application Transport Internet Link

From farms to computing fabrics 12 10 Thousand dual-cpu boxes 0.8 0.8 5 8 Hundreds of tape drives 1.5 6 * 250 24 * Farm Network * Data Rate in Gbps 960 * LAN-WAN Routers Multi-Gigabit Ethernet switche Data Network 0.8 Real-time detector data Grid Interface 10 Thousand disk units Computing fabric at CERN (2006) [per experiment!]

For a fixed Quality of Service for the Users: - Costs will be dominated by the level of Automation and the reliability of Certification procedures. - This in turn is related to the Architecture of the solution. (Hardware) Trend: These two are increasing

LCG-ES is a Coordinated Project of 7 Groups involved in 3 LHC experiments. It is a development and deployment project, not an R&D project. It is managed similar to an EU project. You can think of it as the overlapped pre-production, pre-assembly and development of Engineering Design Report of the Computing Subdetector in 36 months. USC (LHCb) IFCA (CMS) CIEMAT (CMS) UAM (ATLAS) IFAE (ATLAS) UB (LHCb) IFIC (ATLAS)

Background leading to LCG-ES Overall situation of Spanish groups in computing was analyzed and found to be too weak to make just-in-time LHC ramp-up Analysis for LEP mostly done at CERN towards the end General lack of technicians, system admins and sw professionals Work on LHC dispersed, not fully recognized by LHC collaborations (in the shadows of US ATLAS/CMS, France, etc ) On the other hand, we had a small string of high quality activities Various contributions to online and slow controls. FALCON quasi-online farm for ALEPH (+ Florida reprocess farm) CERN RD-47 (HEP Data Processing using Commodity Components) R&D in Object Databases for HEP Data Participation in EU DataGrid and EU CrossGrid Initial boosts from ad-hoc mechanisms, lobbying and FPA 2 multi-disciplinary farms (~200 nodes) at IFIC and IFCA Special Action for 1 year of 500 K (about 50% materials, 50% personnel

LCG-ES: High-Level Objectives Coordinated within the global LCG project plan and the Data Challenges of the experiments Develop and test under realistic conditions a series of prototypes, increasingly more powerful and complex, of data processing systems for the LHC utilizing Grid techniques. Research the possibilities of sharing complex data processing infrastructures between several communities (those of ATLAS, CMS and LHCb in our case).

Finer-grain objectives (1 of 3) 1. Establish a solid base of knowledge and experience in Grid distributed computing in order to ensure participation of Spanish groups in LHC data analysis through a shared infrastructure. 2. Establish a solid base of knowledge and experience in management and processing of the large data volumes (Petabyte level) that will be generated by the LHC experiments. 3. Actively participate in Grid computing projects in order to ensure the deployment in Spain of a Grid and DataGrid infrastructure which is integrated internationally.

Finer-grain objectives (2 of 3) 4. Take active part in development of Grid computing within the LHC experiments and contribute to the LHC Computing Grid Technical Design Report. 5. Create Monte Carlo production centers such that groups can contribute to the official productions of the LHC experiments. In particular: Generate official MC consisted with Spanish participation in the LHC program and the objectives of groups within their experiment. Explore efficient and affordable methods to deliver the data of these official productions to the international collaborators in the experiments.

Finer-grain objectives (3 of 3) 6. Deliver to the groups a local software development environment (for simulation, reconstruction and analysis) appropriate for their experiment. 7. Facilitate the transfer of the experience acquired, fostering the possible re-use of this knowledge in other HEP projects or in other disciplines.

Work Methodology Major agreements by groups Commit to give services to each other Allow jobs from any experiment to run at institute s facilities (with some safeguards) Opened the possibility to weave deliverables together into a Spanish Grid Economy linked to the global LCG Grid Forced to use build-to-cost, as cost estimates are difficult with fastmoving technologies and demand seems to far exceed reasonable funding expectations in 2003-2005 Allow a reasonable amount of personnel scaling in order to study farm and data management automation to minimize personnel needs in the production system

Risk Analysis Major agreements by groups (R1: Management support) Commit to give services to each other Allow jobs from any experiment to run at institute s facilities (with some safeguards) Opened the possibility to weave deliverables together into a Spanish Grid Economy linked to the global LCG Grid (R2: Will Grid techniques deliver. Alternatives?) Forced to use build-to-cost, as cost estimates are difficult with fastmoving technologies and demand seems to far exceed reasonable funding expectations in 2003-2005 (R3: Fulfilling official MC. R4: Production scale-up) Allow a reasonable amount of personnel scaling in order to study farm and data management automation to minimize personnel needs in the production system (R5: Ratchet effect, local vs national personnel policy)

LCG-ES Budget summary Notes: 1. Spanish contribution to LCG-CERN (300 K ) handled administratively through IFAE (Additional 100 K for LCG-CERN Phase 1 to be funded in 2005) 2. Project budget is 50% human resources. If you include human resources funded by institutes and universities, it is approximately 2/3 people. Consistent with Gartner Group????

LCG-ES Personnel details (estimates) Personnel per Work Package Personnel per Group and Funding Source

Deliverables to fulfill the objectives D = Analysis Farm S = SW Dev Platform G = SW repository W = SW Gridification CF = MC Fabric M = Virtual MC farm D = Data Transform C = Gridified Data Store G = Security Architect S = Tech MC Support C = Data Chal. Coord. EAD, GSW, CDC USC EAD, MCF, GSW, CDC IFCA EAD, MCF, EDS, RSG, SEG, CTS CIEMAT UAM EAD, GVM EAD, ETD, PIC, EDS, GSW IFAE IFIC UB EAD,EDS, CTS EAD, MCF, CTS, CDC Stay away from glitz. Concentrate on deployment, MC & analysis Use local Univ. for TT to other disciplines 600 KCHF materials contribution to LCG-CERN

UI JDL Use Grid technology to generate a common infrastructure Input Sandbox Replica Catalogue Information Service ob Submit Event Output Sandbox Resource Broker Input Sandbox gging & ok-keeping keeping Job Submission Service Brokerinfo Storag Eleme Job Status Output Sandbox Compute

Functionality developed by EDG, Globus, etc. brought to high performance level by LCG + Data Challenges EU DataGrid Monitor http://ccwp7.in2p3.fr/mapcenter/

Tier3 physics department sktop γ Tier2 Lab a grid for a regional group The LHC Computing Centre β Lab b α The LHC Computing Model: 1999-2002 and still evolving Uni x Tier 1 Uni y USA Italy CERN Tier 1 Lab m CERN Tier 0. Germany Uni b UK France. Uni a Lab c grid for a physics study group Uni n

The Tier-1 Dilemma Classic T1: Serve 10% of a global experiment Global LHC Community Spanish Grid Economy Zone Spain: ~7% of CERN ~3% of LHC

LCG Requirements Ctte (SC2) revisited Tier 1/2/3 issue. New document focuses more on services than on capacity Global LHC Community PIC Spanish Grid Economy Zone Prototype a Grid Services Exchange focused on LHC data PIC = Port d Legacy-free center for data-intensive scientific computing Exploit commodity hardware and automation to lower costs Store and offer produced datasets to global LCG securely

Objectives of the PIC Coordination centre for data-intensive scientific computing for users in Spain and open to other users in the EU and other nations. Coordination centre for LHC computing of Spanish HEP, maintaining contacts particularly with CERN. Reference centre in data-intensive scientific computation techniques, including transfer of these techniques to scientific disciplines which may benefit from them. Establishment of collaborations with other institutes involved in scientific computation, in particular supercomputing and massively parallel computing. Development of techniques and general services in the context of the future Global Information Grid. Development of a Centre of Excellence to enable participation in European projects for the development of a future International Grid Infrastructure.

PIC Base Funding Plan (Note: CIEMAT agreement not signed yet) Port d' (PIC) Base Funding Plan (Meuros) 1.2 1 0.8 0.6 0.4 0.2 CIEMAT-I CIEMAT-P GENCAT-I GENCAT-P UAB-I UAB-M&O 0 2002 2003 2004 2005 2006 2007 Year

PIC infrastructure at UAB UAB is turning Edifici D into a shared infrastructure to house computationally oriented scientific research activities. The largest user of the building will be the PIC. PIC will have ~150 m 2 office space there (mostly open-plant type, we already have 90 m 2 ). Some space reserved for visitors. The PIC tape robots and data servers are housed in a partition of the computer room of Edifici D. 150 m 2 of high quality machine room with false floor 200 KVA electricity with battery backup capilarized to about 70 16-Amp circuits each with smoke detector 300 KW of triply redundant air conditioning 500 KVA diesel generator (Rolls-Royce!!) The Catalan and Spanish A&R Network are being boosted to the Gbps range. GÈANT is already multi-gbps.

A few pictures of where we are with PIC Overall view of machine room Space for: 2 STK cylindrical robots 35 19 racks Future: Double tape space by using additional room

A few pictures of where we are with PIC Latest Generation L5500 Storagetek Tape Robot (6000 slots, >1.4 PBytes capacity) 2 Latest Generation 9940B drives (200 GB cartridge, 30 MB/sec)

One Grid Infrastructure makes many Grids possible Grid Infrastructure General middleware Field-specific middleware Field-specific middleware Field-specific middleware Distinct, Collaborating Network and Computing Infrastructures Communities of geographically dispersed, collaborating scientists from various fields

Grids and Grid Infrastructure: Provider-centric view Grid Infrastructure Multiple Scientific Communities collaborating in subsets Grid-enabled Computing and Storage Resources Network Infrastructure General middleware

Questions of Terminology e-science Enhanced Science possible by continuous collaboration o mobile scientists supported by digital resources Grids The set of all instances of a Grid a Grid The set of digital resources that a Community is using at any given time to support their e-activity Grid Infrastructure The set of persistent Services and Processes which perm construction and reconfiguration of Grids Research Network Infrastructure and/or Internet The net that we know and which will grow influenced by and influencing Grid Infrastructure and Grids

An illustration: GÉANT and GRIDs (as seen by EU IST programme) GRIDs use GÉANT infrastructure Enhanced Science or e-sciencee GÉANT profits from technological innovation GRIDs empowered GÉANT International dimension Grids in support of various subjects Grid Infrastructure Prototypes GÉANT network Adapted by permission from A. Karlson, EU DG-INFSO R&D on GRIDs

Lead-times and momentum in R+D Internet Web Web Services Grid DNS, Akamai, SSL, Yahoo, co-location, Web Hosting, etc.

Evolution of data-processing environments Decade 2020 2010 2000 1990 1980 1970 1960 1950 Ambient Computing Ubiquitous wireless self-configuring devices Virtual Communities sharing Digitized Data Digital ID infrastructure, Grid Infrastructures The Network Society Fiber optics, Ethernet, TCP/IP, Internet Personal Productivity Microprocessors, PC, Mac Process Automation Mini-computers, Relational Databases Time-sharing computing Operating systems Numerical methods Mainframes Programming Computers 1 10 100 1000 10000 100000 1000000 10000000 Complexity and Functionality

Evolution of data-processing environments 2020 Ambient Computing Ubiquitous wireless self-configuring devices 2010 Virtual Communities sharing Digitized Data Digital ID infrastructure, Grid Infrastructures We are here 2000 The Network Society Fiber optics, Ethernet, TCP/IP, Internet Decade 1990 1980 Personal Productivity Microprocessors, PC, Mac Process Automation Mini-computers, Relational Databases WWW invented here Internet born here 1970 Time-sharing computing Operating systems First e-mail ever sent 1960 Numerical methods Mainframes 1950 Programming Computers 1 10 100 1000 10000 100000 1000000 10000000 Complexity and Functionality

Deliverables to fulfill the objectives D = Analysis Farm S = SW Dev Platform G = SW repository W = SW Gridification CF = MC Fabric M = Virtual MC farm D = Data Transform C = Gridified Data Store G = Security Architect S = Tech MC Support C = Data Chal. Coord. EAD, GSW, CDC USC EAD, MCF, GSW, CDC IFCA EAD, MCF, EDS, RSG, SEG, CTS CIEMAT UAM EAD, GVM EAD, ETD, PIC, EDS, GSW IFAE IFIC UB EAD,EDS, CTS EAD, MCF, CTS, CDC Stay away from glitz. Concentrate on deployment, MC & analysis Use local Univ. for TT to other disciplines 600 KCHF materials contribution to LCG-CERN