CERN and Scientific Computing

Similar documents
The LHC Computing Grid. Slides mostly by: Dr Ian Bird LCG Project Leader 18 March 2008

The LHC Computing Grid

CC-IN2P3: A High Performance Data Center for Research

The CMS Computing Model

A short introduction to the Worldwide LHC Computing Grid. Maarten Litmaath (CERN)

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Grid Computing a new tool for science

CERN openlab II. CERN openlab and. Sverre Jarp CERN openlab CTO 16 September 2008

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Travelling securely on the Grid to the origin of the Universe

The grid for LHC Data Analysis

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 27 July 2010

Storage and I/O requirements of the LHC experiments

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

150 million sensors deliver data. 40 million times per second

HOME GOLE Status. David Foster CERN Hawaii GLIF January CERN - IT Department CH-1211 Genève 23 Switzerland

Stephen J. Gowdy (CERN) 12 th September 2012 XLDB Conference FINDING THE HIGGS IN THE HAYSTACK(S)

Experience of the WLCG data management system from the first two years of the LHC data taking

First Experience with LCG. Board of Sponsors 3 rd April 2009

The LHC computing model and its evolution. Dr Bob Jones CERN

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

CSCS CERN videoconference CFD applications

Batch Services at CERN: Status and Future Evolution

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 21 July 2011

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

DESY at the LHC. Klaus Mőnig. On behalf of the ATLAS, CMS and the Grid/Tier2 communities

High-Energy Physics Data-Storage Challenges

Belle & Belle II. Takanori Hara (KEK) 9 June, 2015 DPHEP Collaboration CERN

Virtualizing a Batch. University Grid Center

Constant monitoring of multi-site network connectivity at the Tokyo Tier2 center

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback

Visita delegazione ditte italiane

The Grid: Processing the Data from the World s Largest Scientific Machine

Storage Resource Sharing with CASTOR.

Andrea Sciabà CERN, Switzerland

ATLAS Experiment and GCE

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

The INFN Tier1. 1. INFN-CNAF, Italy

LHCb Computing Resource usage in 2017

Summary of the LHC Computing Review

Overview. About CERN 2 / 11

Computing at Belle II

Moving e-infrastructure into a new era the FP7 challenge

Distributed e-infrastructures for data intensive science

Austrian Federated WLCG Tier-2

Data Transfers Between LHC Grid Sites Dorian Kcira

Computing grids, a tool for international collaboration and against digital divide Guy Wormser Director of CNRS Institut des Grilles (CNRS, France)

International Cooperation in High Energy Physics. Barry Barish Caltech 30-Oct-06

UW-ATLAS Experiences with Condor

Spanish Tier-2. Francisco Matorras (IFCA) Nicanor Colino (CIEMAT) F. Matorras N.Colino, Spain CMS T2,.6 March 2008"

Grid Computing at the IIHE

Towards Network Awareness in LHC Computing

Grid: data delen op wereldschaal

Preparing for High-Luminosity LHC. Bob Jones CERN Bob.Jones <at> cern.ch

LHCb Computing Resources: 2019 requests and reassessment of 2018 requests

Giovanni Lamanna LAPP - Laboratoire d'annecy-le-vieux de Physique des Particules, Université de Savoie, CNRS/IN2P3, Annecy-le-Vieux, France

Online data storage service strategy for the CERN computer Centre G. Cancio, D. Duellmann, M. Lamanna, A. Pace CERN, Geneva, Switzerland

Multi-threaded, discrete event simulation of distributed computing systems

IEPSAS-Kosice: experiences in running LCG site

The Grid. Processing the Data from the World s Largest Scientific Machine II Brazilian LHC Computing Workshop

CHIPP Phoenix Cluster Inauguration

Benchmarking the ATLAS software through the Kit Validation engine

CERN Lustre Evaluation

Grid Computing September 2010 Marian Babik CERN. The LHC Computing Grid Marian Babik (orig. by Marian Babik (orig. by Rafal Otto, GridCafe),

ISTITUTO NAZIONALE DI FISICA NUCLEARE

The European Data Grid Project

CouchDB-based system for data management in a Grid environment Implementation and Experience

Grid Computing Activities at KIT

De BiG Grid e-infrastructuur digitaal onderzoek verbonden

CERN network overview. Ryszard Jurga

CERN s Business Computing

arxiv: v1 [physics.ins-det] 1 Oct 2009

The EU DataGrid Testbed

Application of Virtualization Technologies & CernVM. Benedikt Hegner CERN

Insight: that s for NSA Decision making: that s for Google, Facebook. so they find the best way to push out adds and products

Evaluation of the computing resources required for a Nordic research exploitation of the LHC

Grid and Cloud Activities in KISTI

Storage on the Lunatic Fringe. Thomas M. Ruwart University of Minnesota Digital Technology Center Intelligent Storage Consortium

Tier-2 DESY Volker Gülzow, Peter Wegner

CLOUDS OF JINR, UNIVERSITY OF SOFIA AND INRNE JOIN TOGETHER

Clouds in High Energy Physics

Introduction to Grid Computing May 2013 Marian Babik CERN. The LHC Computing Grid Marian Babik (orig. by Rafal Otto, GridCafe), CERN

Philippe Laurens, Michigan State University, for USATLAS. Atlas Great Lakes Tier 2 collocated at MSU and the University of Michigan

Big Data Analytics and the LHC

Data handling and processing at the LHC experiments

Computing. DOE Program Review SLAC. Rainer Bartoldus. Breakout Session 3 June BaBar Deputy Computing Coordinator

Distributing storage of LHC data - in the nordic countries

Data Issues for next generation HPC

CMS Computing Model with Focus on German Tier1 Activities

Data Management for the World s Largest Machine

Challenges of the LHC Computing Grid by the CMS experiment

Status of KISTI Tier2 Center for ALICE

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN

SoLID GEM Detectors in US

LHCb Computing Strategy

Tackling tomorrow s computing challenges today at CERN. Maria Girone CERN openlab CTO

CERN Network activities update

I Service Challenge e l'implementazione dell'architettura a Tier in WLCG per il calcolo nell'era LHC

Using the In-Memory Columnar Store to Perform Real-Time Analysis of CERN Data. Maaike Limper Emil Pilecki Manuel Martín Márquez

Transcription:

CERN and Scientific Computing Massimo Lamanna CERN Information Technology Department Experiment Support Group

1960: 26 GeV proton in the 32 cm CERN hydrogen bubble chamber 1960: IBM 709 at the Geneva airport 1972: Experiment at the CERN PS (East hall) 1972: Installation of the CDC 7600 2

1983: Computer centre mainframes (IBM, Siemens, CDC) (Bd 513) 1983: UA1 discovers Ws and Z0 at the CERN SPPS 1990: The first web server: this machine was used by Tim Berners-Lee in 1990 to develop and run the first WWW server, multi-media browser and web editor 1989-2000: LEP experiments 3

Early days of COMPASS (1998) A new experiment being built with new computing challenges DAQ = 35 MB/s (later over 60 MB/s) 0.5 PB / year Massive use of PC technology The SM2 magnet The COMPASS Computing Farm 4

COMPASS Event Display 5

20 th of November 2009 27 km tunnel (2000+ superconductive dipoles) 450 GeV injection energy 7 TeV max beam energy 3.5+3.5 TeV (March the 30 th world record) 6

Very recent LHC pp collisions! 7

Interesting facts (CERN Computer centre) Number of machines About 4,500 batch (18,000 CPUs) About 3,000 disk servers (50,000 hard drives) Several hundred tape servers, console head-nodes, database and Grid servers etc. Storage Capacity 5+ PB disk 25+ PB tape (IBM and Sun/StorageTek) There will be an additional 15 PB each year needed for the LHC data (3*10^6 DVDs!) Network Capacity Connection at 10 gigabits/second to each Tier 1, plus backup, plus regular (firewalled etc) internet Speed record: 1.1 TB in less than half hour (CERN-CalTech) Number of staff CERN: ~2700 ; IT department (computing) ~250 and ~200 on shorter-term Grid projects 25-NOV-09 M. Lamanna Early 2009 data 8 Fast ramp up going on for the LHC start

9

What makes High-Energy Physics so special (also in computing)? High rates 100-1,000 MB/s over long periods of time (10+ years, ~10 7 s/year i.e several months a year) High data volumes Rates + derived types + replications o(10) PB/year/experiment Lot of files, relatively small DB (compared to the file storage) Several 10 5 PC cores to reconstruct, simulate and analyse the data Reconstruction 10-100 s range on one PC Large distributed collaborations O(1000) physicists Everyone ( everywhere ) gets in touch with data, in general with custom applications (based on common framework). High concurrency Human activity, interactive : need to minimise the latency 10

LHC Computing Grid Worldwide infrastructure (EGEE + OSG + NDGF) 11

CPU Number crunching boxes No resident scientific data Shared facility for all CERN users (basically every physicists participating in an experiment at CERN) Faster PC means more SPECs per box SPEC_INT_2k still very much used 1,000 SI2K ~ Pentium4 @ 3GHz (~3GFlops) 12

Disks Staging area: they keep hot data Access to these disks is managed by dedicated PCs (serving/receiving data to the PC crunching numbers) Moore s law at work here! Gb per goes up Disk capacity goes up GB/cm 3 goes up Mass-market items! 13

Tapes Data custodial We build accelerator and experiments to collect scientific data Write-once / Read-many At variance with backups (Write-once / Read-never) Evolving (a la Moore s law) but more gently than PCs No surprise, none of us has a tape library at home, I guess Expect to store 40 PB (40,000 TB) of data per year Scientific data, corresponding derived data (reconstruction, analysis), simulation data 14

CA-TRIUMF US-T1-BNL NDGF UK-T1-RAL NL-T1 DE-KIT FR-CCIN2P3 IT-INFN-CNAF ES-PIC 10 clouds Only the Tier1 centres are shown About 100 Tier2s in total TW-ASGC 15

EXAMPLES OF OTHER APPLICATION (CERN COLLABORATIONS) 16

ITU conference (2006) The problem: Assign frequencies for digital radio and television (international treaty) Critical point: Need on dependability: verify (iteratively) the compatibility between radio stations Solution: Use the EGEE grid + a system used in ATLAS and LHCb to increase the reliability of the Grid 17

Bird Flu Basic idea: Compute how a given chemical interacts with a protein (e.g. belonging to a virus) High affinity means the chemical is a potential drug against the virus In silico (i.e. use your PC): Scan millions of chemicals (~10 3 s per chemical-protein pair) With 1,000 PCs, 1 docking per second Good candidate given to biologist (verification longer and more expensive- than in silico docking) In practice, you enrich the initial sample saving time Molecular docking (and money) Essential to fight to pandemia (H5N1) or to fight neglected diseases (like Malaria) WISDOM collaboration Malaria H5N1 (Bird Flu) 18

19

Hydrological modeling Observations SWAT Scenarios1 Observations Climate SWAT Observations Land cover SWAT Scenarios2 Scenarios3 CERN contribution: gridification of SWAT Using Ganga/Diane 20

Questions? 21