The LHC Computing Grid. Slides mostly by: Dr Ian Bird LCG Project Leader 18 March 2008

Similar documents
The LHC Computing Grid

The grid for LHC Data Analysis

The Grid: Processing the Data from the World s Largest Scientific Machine

The Grid. Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop

Travelling securely on the Grid to the origin of the Universe

CHIPP Phoenix Cluster Inauguration

A short introduction to the Worldwide LHC Computing Grid. Maarten Litmaath (CERN)

First Experience with LCG. Board of Sponsors 3 rd April 2009

The EGEE-III Project Towards Sustainable e-infrastructures

Grid Computing a new tool for science

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Distributed e-infrastructures for data intensive science

RUSSIAN DATA INTENSIVE GRID (RDIG): CURRENT STATUS AND PERSPECTIVES TOWARD NATIONAL GRID INITIATIVE

Computing grids, a tool for international collaboration and against digital divide Guy Wormser Director of CNRS Institut des Grilles (CNRS, France)

The LHC computing model and its evolution. Dr Bob Jones CERN

CERN and Scientific Computing

Storage and I/O requirements of the LHC experiments

150 million sensors deliver data. 40 million times per second

IEPSAS-Kosice: experiences in running LCG site

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

High Performance Computing from an EU perspective

Pan-European Grid einfrastructure for LHC Experiments at CERN - SCL's Activities in EGEE

Moving e-infrastructure into a new era the FP7 challenge

Preparing for High-Luminosity LHC. Bob Jones CERN Bob.Jones <at> cern.ch

Grid Computing September 2010 Marian Babik CERN. The LHC Computing Grid Marian Babik (orig. by Marian Babik (orig. by Rafal Otto, GridCafe),

e-science Infrastructure and Applications in Taiwan Eric Yen and Simon C. Lin ASGC, Taiwan Apr. 2008

LHC Computing Grid today Did it work?

A European Vision and Plan for a Common Grid Infrastructure

Andrea Sciabà CERN, Switzerland

Introduction to Grid Computing May 2013 Marian Babik CERN. The LHC Computing Grid Marian Babik (orig. by Rafal Otto, GridCafe), CERN

CERN openlab II. CERN openlab and. Sverre Jarp CERN openlab CTO 16 September 2008

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

e-science for High-Energy Physics in Korea

The LHC Computing Grid

Status of KISTI Tier2 Center for ALICE

Computing for LHC in Germany

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback

Batch Services at CERN: Status and Future Evolution

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

Constant monitoring of multi-site network connectivity at the Tokyo Tier2 center

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 27 July 2010

The Portuguese National Grid Initiative (INGRID)

Experience of the WLCG data management system from the first two years of the LHC data taking

I Service Challenge e l'implementazione dell'architettura a Tier in WLCG per il calcolo nell'era LHC

Austrian Federated WLCG Tier-2

Grid and Cloud Activities in KISTI

Grid Computing Activities at KIT

Summary of the LHC Computing Review

PoS(EGICF12-EMITC2)106

Operating the Distributed NDGF Tier-1

Visita delegazione ditte italiane

Bob Jones. EGEE and glite are registered trademarks. egee EGEE-III INFSO-RI

Challenges of the LHC Computing Grid by the CMS experiment

where the Web was born Experience of Adding New Architectures to the LCG Production Environment

Virtualizing a Batch. University Grid Center

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

Grid Challenges and Experience

CC-IN2P3: A High Performance Data Center for Research

GRIDS INTRODUCTION TO GRID INFRASTRUCTURES. Fabrizio Gagliardi

The CMS Computing Model

CERN s Business Computing

CHEP Report. Jeff Templon

HealthGrids: In Search for Sustainable Solutions

Features and Future. Frédéric Hemmer - CERN Deputy Head of IT Department. Enabling Grids for E-sciencE. BEGrid seminar Brussels, October 27, 2006

The INFN Tier1. 1. INFN-CNAF, Italy

HTC/HPC Russia-EC. V. Ilyin NRC Kurchatov Institite Moscow State University

Computing / The DESY Grid Center

LHCb Computing Strategy

Data Management for the World s Largest Machine

e-infrastructure: objectives and strategy in FP7

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 21 July 2011

Storage Resource Sharing with CASTOR.

Interconnection of Armenian e- Infrastructures with the pan- Euroepan Integrated Environments

EGEE and Interoperation

Accelerating Throughput from the LHC to the World

Stephen J. Gowdy (CERN) 12 th September 2012 XLDB Conference FINDING THE HIGGS IN THE HAYSTACK(S)

Spanish Tier-2. Francisco Matorras (IFCA) Nicanor Colino (CIEMAT) F. Matorras N.Colino, Spain CMS T2,.6 March 2008"

DESY. Andreas Gellrich DESY DESY,

Integration of an Asian NGI with European counterparts

Data Handling for LHC: Plans and Reality

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

The European DataGRID Production Testbed

Grid: data delen op wereldschaal

N. Marusov, I. Semenov

Computing Model Tier-2 Plans for Germany Relations to GridKa/Tier-1

Grids and Security. Ian Neilson Grid Deployment Group CERN. TF-CSIRT London 27 Jan

De BiG Grid e-infrastructuur digitaal onderzoek verbonden

WLCG and Grid Computing Summer 2011 Part1: WLCG Markus Schulz. IT Grid Technology Group, CERN WLCG

Grid Computing: dealing with GB/s dataflows

High Energy Physics data analysis

Tier-2 DESY Volker Gülzow, Peter Wegner

Big Data Analytics and the LHC

Grid Computing at the IIHE

CernVM-FS beyond LHC computing

Computing at Belle II

LHCb Computing Resource usage in 2017

Evaluation of the computing resources required for a Nordic research exploitation of the LHC

Tier-2 structure in Poland. R. Gokieli Institute for Nuclear Studies, Warsaw M. Witek Institute of Nuclear Physics, Cracow

DESY at the LHC. Klaus Mőnig. On behalf of the ATLAS, CMS and the Grid/Tier2 communities

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

Transcription:

The LHC Computing Grid Slides mostly by: Dr Ian Bird LCG Project Leader 18 March 2008 The LHC Computing Grid February 2008

Some precursors Computing for HEP means data handling Fixed-target experiments are also at the forefront COMPASS >300 TB per year starting 2001 (still running) Is used to investigating new computing technologies One of the first reconstruction programmes entirely written using C++ and modern techniques Raw data recorded at 35-70 MB/s (CMS expect ~200 MB/s) Test of very-large database technologies First user of CASTOR (transparent access of tape data)

Clusters of Inexpensive Processors Requirements driven We started this phase with a simple architecture that enables sharing of storage across CPU servers, that proved stable and has survived from RISC to Quad-core Parallel, high throughput Sustained price/perf improvement ~60% /yr Ferranti Mercury IBM 709 IBM 7090 CDC 7600 <--- IBM Mainframes---> Apollo DN10.000s First Supercomputer 1989 20 MIPS/proc CDC 6600 1990 - SUN, SGI, IBM, HP, DEC,... each with its own flavour of Unix 1996 the first PC service 1998 COMPASS Computing farm 2008 dual quad core systems 50K MIPS/chip ~20k cores available == ~20 MSI2K 5 orders of magnitude in 18 years First PC systems COMPASS Computing Farm First RISC systems Last Supercomputer Cray X-MP LHC (14 TeV) LEP II (200GeV) LEP (100GeV) COMPASS

The LHC Data Challenge CERN IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it The accelerator will be completed in 2008 and run for 10-15 years Experiments will produce about 15 Million Gigabytes of data each year (about 20 million CDs!) LHC data analysis requires a computing power equivalent to ~100,000 of today's fastest PC processors Requires many cooperating computer centres, as CERN can only provide ~20% of the capacity

Summary of Computing Resource Requirements All experiments - 2008 From LCG TDR - June 2005 CERN All Tier-1s All Tier-2s Total CPU (MSPECint2000s) 25 56 61 142 Disk (PetaBytes) 7 31 19 57 Tape (PetaBytes) 18 35 53 CPU Disk Tape All Tier-2s 43% CERN 18% All Tier-2s 33% CERN 12% CERN 34% All Tier-1s 39% All Tier-1s 55% All Tier-1s 66%

Solution: the Grid Use the Grid to unite computing resources of particle physics institutes around the world The World Wide Web provides seamless access to information that is stored in many millions of different geographical locations The Grid is an infrastructure that provides seamless access to computing power and data storage capacity distributed over the globe CERN IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it

How does the Grid work? It makes multiple computer centres look like a single system to the end-user Advanced software, called middleware, automatically finds the data the scientist needs, and the computing power to analyse it. Middleware balances the load on different resources. It also handles security, accounting, monitoring and much more. CERN IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it

View of the ATLAS detector (under construction) 150 million sensors deliver data 40 million times per second

CERN IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it Frédéric Hemmer, CERN, IT Department

LHC Computing Grid project (LCG) More than 140 computing centres 12 large centres for primary data management: CERN (Tier-0) and eleven Tier-1s 38 federations of smaller Tier-2 centres 35 countries involved CERN IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it

LCG Service Hierarchy Tier-0: the accelerator centre Data acquisition & initial processing Long-term data safekeeping Distribution of data Tier-1 centres Canada Triumf (Vancouver) Spain PIC (Barcelona) France IN2P3 (Lyon) Taiwan Academia SInica (Taipei) Germany Forschungszentrum KarlsruheUK CLRC (Oxford) Italy CNAF (Bologna) US FermiLab (Illinois) Netherlands NIKHEF/SARA (Amsterdam) Brookhaven (NY) Nordic countries distributed Tier-1 Tier-1: online to the data acquisition process high availability Managed Mass Storage grid-enabled data service Data-heavy analysis National, regional support Tier-2: ~140 centres in ~35 countries Simulation End-user analysis batch and interactive

WLCG Collaboration CERN IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it The Collaboration 4 LHC experiments ~140 computing centres 12 large centres (Tier-0, Tier-1) 38 federations of smaller Tier-2 centres ~35 countries Memorandum of Understanding Agreed in October 2005, now being signed Resources Focuses on the needs of the four LHC experiments Commits resources each October for the coming year 5-year forward look Agrees on standards and procedures Relies on EGEE and OSG (and other regional efforts)

Data Transfer Data distribution from CERN to Tier-1 sites The target rate was achieved in 2006 under test conditions Autumn 2007 & CCRC 08 under more realistic experiment testing, reaching & sustaining target rate with ATLAS and CMS active Target 2008

Jan-07 Feb-07 Mar-07 Apr-07 May-07 Jun-07 Jul-07 Aug-07 Sep-07 Oct-07 Nov-07 Dec-07 6000000 5000000 4000000 3000000 2000000 1000000 0 LCG Jobs run 2007 LHCb CMS 165k/day ATLAS ALICE Grid activity WLCG ran ~ 44 M jobs in 2007 workload has continued to increase now at ~ 165k jobs/day Distribution of work across Tier0/Tier1/Tier 2 really illustrates the importance of the grid system Tier 2 contribution is around 50%; > 85% is external to CERN Tier 2 sites

Impact of the LHC Computing Grid in Europe LCG has been the driving force for the European multiscience Grid EGEE (Enabling Grids for E-sciencE) EGEE is now a global effort, and the largest Grid infrastructure worldwide Co-funded by the European Commission (Cost: ~130 M over 4 years, funded by EU ~70M ) EGEE already used for >100 applications, including Bio-informatics Education, Training Medical Imaging

EGEE The EGEE project Started in April 2004, now in second phase with 91 partners in 32 countries 3 rd phrase (2008-2010) in preparation Objectives Large-scale, production-quality grid infrastructure for e-science Attracting new resources and users from industry as well as science Maintain and further improve glite Grid middleware

Registered Collaborating Projects 25 projects have registered as of September 2007: web page Infrastructures geographical or thematic coverage Applications improved services for academia, industry and the public Support Actions key complementary functions

Collaborating infrastructures

Enabling Grids for E-sciencE Archeology Astronomy Astrophysics Civil Protection Comp. Chemistry Earth Sciences Finance Fusion Geophysics High Energy Physics Life Sciences Multimedia Material Sciences >250 sites 48 countries >50,000 CPUs >20 PetaBytes >10,000 users >150 VOs >150,000 jobs/day EGEE-II INFSO-RI-031688

In silico drug discovery Enabling Grids for E-sciencE Diseases such as HIV/AIDS, SARS, Bird Flu etc. are a threat to public health due to world wide exchanges and circulation of persons Grids open new perspectives to in silico drug discovery Reduced cost and adding an accelerating factor in the search for new drugs International collaboration is required for: Early detection Epidemiological watch Prevention Search for new drugs Search for vaccines Avian influenza: bird casualties EGEE-II INFSO-RI-031688

Enabling Grids for E-sciencE http://wisdom.healthgrid.org/ WISDOM EGEE-II INFSO-RI-031688

For more information: www.cern.ch/lcg www.eu-egee.org www.eu-egi.org/ www.gridcafe.org CERN IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it Thank you for your kind attention!