The CMS Computing Model

Similar documents
Data Transfers Between LHC Grid Sites Dorian Kcira

CMS Computing Model with Focus on German Tier1 Activities

Stephen J. Gowdy (CERN) 12 th September 2012 XLDB Conference FINDING THE HIGGS IN THE HAYSTACK(S)

Data handling and processing at the LHC experiments

1. Introduction. Outline

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN

LHCb Computing Resources: 2018 requests and preview of 2019 requests

LHCb Computing Resources: 2019 requests and reassessment of 2018 requests

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

arxiv: v1 [physics.ins-det] 1 Oct 2009

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

CERN and Scientific Computing

LHCb Computing Resource usage in 2017

Virtualizing a Batch. University Grid Center

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Summary of the LHC Computing Review

Computing. DOE Program Review SLAC. Rainer Bartoldus. Breakout Session 3 June BaBar Deputy Computing Coordinator

Detector Control LHC

Storage and I/O requirements of the LHC experiments

LHC Computing Models

Big Data Analytics and the LHC

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

Real-time dataflow and workflow with the CMS tracker data

High-Energy Physics Data-Storage Challenges

Computing at the Large Hadron Collider. Frank Würthwein. Professor of Physics University of California San Diego November 15th, 2013

Andrea Sciabà CERN, Switzerland

Storage on the Lunatic Fringe. Thomas M. Ruwart University of Minnesota Digital Technology Center Intelligent Storage Consortium

Experience of the WLCG data management system from the first two years of the LHC data taking

Storage Resource Sharing with CASTOR.

First LHCb measurement with data from the LHC Run 2

Challenges of the LHC Computing Grid by the CMS experiment

CERN openlab II. CERN openlab and. Sverre Jarp CERN openlab CTO 16 September 2008

CouchDB-based system for data management in a Grid environment Implementation and Experience

UW-ATLAS Experiences with Condor

Deferred High Level Trigger in LHCb: A Boost to CPU Resource Utilization

New strategies of the LHC experiments to meet the computing requirements of the HL-LHC era

PoS(High-pT physics09)036

CMS Belgian T2. G. Bruno UCL, Louvain, Belgium on behalf of the CMS Belgian T2 community. GridKa T1/2 meeting, Karlsruhe Germany February

DESY at the LHC. Klaus Mőnig. On behalf of the ATLAS, CMS and the Grid/Tier2 communities

Experience with Data-flow, DQM and Analysis of TIF Data

Reprocessing DØ data with SAMGrid

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback

Monte Carlo Production Management at CMS

Status of KISTI Tier2 Center for ALICE

LHCb Computing Status. Andrei Tsaregorodtsev CPPM

CC-IN2P3: A High Performance Data Center for Research

Grid Computing: dealing with GB/s dataflows

Tackling tomorrow s computing challenges today at CERN. Maria Girone CERN openlab CTO

Data Reconstruction in Modern Particle Physics

August 31, 2009 Bologna Workshop Rehearsal I

Update of the Computing Models of the WLCG and the LHC Experiments

Computing at Belle II

CSCS CERN videoconference CFD applications

The CMS L1 Global Trigger Offline Software

LHC Computing Grid. Technical Design Report. Version: June The LCG TDR Editorial Board Chair: J. Knobloch Project Leader: L.

CMS data quality monitoring: Systems and experiences

Review of the Compact Muon Solenoid (CMS) Collaboration Heavy Ion Computing Proposal

Towards Network Awareness in LHC Computing

Large scale commissioning and operational experience with tier-2 to tier-2 data transfer links in CMS

I Service Challenge e l'implementazione dell'architettura a Tier in WLCG per il calcolo nell'era LHC

Data Quality Monitoring Display for ATLAS experiment

The LHC Computing Grid

Grid Computing Activities at KIT

IEPSAS-Kosice: experiences in running LCG site

Electronics, Trigger and Data Acquisition part 3

HEP Grid Activities in China

PROOF-Condor integration for ATLAS

The GAP project: GPU applications for High Level Trigger and Medical Imaging

2008 JINST 3 S Online System. Chapter System decomposition and architecture. 8.2 Data Acquisition System

Continuation Report/Proposal: Vanderbilt CMS Tier 2 Computing Project Reporting Period: November 1, 2010 to March 31, 2012

A L I C E Computing Model

Evaluation of the computing resources required for a Nordic research exploitation of the LHC

The CMS data quality monitoring software: experience and future prospects

Improving Packet Processing Performance of a Memory- Bounded Application

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 21 July 2011

Philippe Laurens, Michigan State University, for USATLAS. Atlas Great Lakes Tier 2 collocated at MSU and the University of Michigan

Muon Reconstruction and Identification in CMS

Data Processing and Analysis Requirements for CMS-HI Computing

ATLAS PILE-UP AND OVERLAY SIMULATION

Monitoring system for geographically distributed datacenters based on Openstack. Gioacchino Vino

Open data and scientific reproducibility

Data oriented job submission scheme for the PHENIX user analysis in CCJ

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 27 July 2010

ATLAS Offline Data Quality Monitoring

The grid for LHC Data Analysis

Trigger and Data Acquisition: an ATLAS case study

ANSE: Advanced Network Services for [LHC] Experiments

CMS-HI-US Computing Proposal Update to the U.S.D.O.E 1

Overview. About CERN 2 / 11

Batch Services at CERN: Status and Future Evolution

LHCb Computing Strategy

Prompt data reconstruction at the ATLAS experiment

Trigger and Data Acquisition at the Large Hadron Collider

PoS(EGICF12-EMITC2)106

Modeling and Validating Time, Buffering, and Utilization of a Large-Scale, Real-Time Data Acquisition System

High Energy Physics data analysis

Transcription:

The CMS Computing Model Dorian Kcira California Institute of Technology SuperComputing 2009 November 14-20 2009, Portland, OR

CERN s Large Hadron Collider 5000+ Physicists/Engineers 300+ Institutes 70+ Countries TOTEM CMS pp, general purpose Atlas pp, general purpose ALICE : Heavy Ions LHCb: B-physics 2/35 27 km Tunnel under Switzerland & France

3/35 Computing for Particle Physics, J.J.Bunn, JHU escience, October 2006

Closing the CMS Detector

Central Silicon Strip Tracker

Detectors, Trigger, Data Acquisition TRIGGER and DAQ 40 MHz 100 khz 300 Hz Two event selection systems HLT is a farm of commercial processors Event selection based on offline software using full event data 7

Physics Events In the Large Hadron Collider protons at very high energies will collide with each-other The particles and energies produced in one of these collisions as well as the respective signatures in the detectors are constitute a data Event Large scale Monte Carlo (MC) simulation of the physics models and detector response is also performed, that results in MC events. Computing in HEP has to deal with storage, processing and accessing of these events

Complexity of LHC Events Simulated Higgs decay 9/35 Computing for Particle Physics, J.J.Bunn, JHU escience, October 2006

07 Nov, 2009 20:28 Beam Splash Event

Cosmic Ray Event

Events: Rates, Sizes, Processing CMS data events will be recorded at a rate of 300Hz Event Content: o RAW: Detector + Trigger information: 1.5MB/event (2MB when SIM included) o RECO: Reconstructed physicsobjects: 0.5 MB/event o AOD: Analysis Object Data : 0.1 MB/event Processing power needed per event: o1 data event reconstruction: 10s on a 3GHz core o1 MC event simulation and reconstruction: 10 times more

2009/2010 Data Start of collisions in LHC expected again within days Expected amounts of recorded data o 2009 run: November 2009 March 2010 => 726M events o 2010 run: April 2010 September 2010 => 1555M events In total this would result in 3.42 PB RAW data + 1.14PB RECO (x 3 for re-reconstruction) Only to reconstruct the data on a single 3GHz core it would take ~ 100 years.

LHC Data Grid Hierarchy Experiment Tier 1 ~PByte/sec 10-40 Gbps Online: custom fast electronics, pipelined, buffered, FPGAs Online System Tier 0 +1 ~150-1500 MB/s CERN Center PBs of Disk; Tape Robot >10 Tier1 and ~100 Tier2 Centers Tier 3 IN2P3 Center Physics data cache ~1-10 Gbps Institute Workstations/Laptops RAL Center Institute Institute 1 to 10 Gbps Tier 4 Tier 2 Institute INFN Center Caltech Tier2 Center FNAL Center ~10 Gbps Tier2 Center Tier2 Center Tier2 Center Tier2 Center Physicists work on analysis channels Each institute has ~10 physicists working on one or more channels Few Petabytes by 2009 An Exabyte ~5-7 Years later 100 Gbps+ Data Networks

Tier Architecture of Computing Resources 20% of resources at CERN, 40% at T1s and 40% at T2s Moving from strict hierarchy towards mesh model T3 T3 T3

Tier-0 Center On CERN site Repacking, prompt reconstruction, storage of RAW+RECO, data archiving and distribution CERN Analysis Facility, latency critical taks: calibration, alignment, monitoring

Tier-1 Centers Share RAW data for custodial storage Data reprocessing and selection Extraction of AOD Archiving of simulation samples

Tier-2 + Tier-3 Centers Monte Carlo Simulation Tier2 Monte Carlo generation Physics Group analysis General User Analysis User analysis Tier3 Small part dedicated to CMS

CMS data flow and Data on(off) Flow line computing To regional centers 10 Gbit/s Tier 0 50 TeraFlops Remote control rooms Readout Syst em s Comput ing Services High Level Trigger 10 TeraFlops Controls: 1 Gbit/s Events Data: 10 Gbit/s Run Cont r ol Builder Net wor ks Event M anager Filt er Syst em s Det ect or Fr ont end Level 1 Tr igger Controls: 1 Gbit/s Raw Data: 2000 Gbit/s CMS SS09 tsv

LHC Data Transfers

Data: CMS Data Management Discover, transfer and access data in a distributed computing environment o Events grouped into files : avoid too small files merging to some GB. Millions of files per year o Fileblocks: 1-10TB, thousands per year. Smallest unit o Datasets: correspond logically to datataking periods/conditions and to simulation samples. Small or up to 10TB Software Components: o DBS (Dataset Bookkeeping system) : provides the means to define, discover and use CMS event data o DLS (Dataset Location Service) : DLS provides the means to locate replicas of data in the distributed system ophedex (Physics Experiment Data Export) : Movement of data, integration with OSG/EGEE transfer services o local file catalogue solutions : a trivial file catalogue as a baseline solution

DBS

PhEDEx

CMS Remote Analysis Builder: CRAB

Summary The Compact Muon Soleid at the Large Hadron Collider has significant computing needs Data driven computing model Distributed tier computing infrastructure 2009/2010 expect 6 PetaByte of collision data These data must be processed, accessed and analyzed CMS computing is ready for LHC collision data expected in the next weeks.