Computing at Belle II

Similar documents

Belle & Belle II. Takanori Hara (KEK) 9 June, 2015 DPHEP Collaboration CERN

Storage and I/O requirements of the LHC experiments

1. Introduction. Outline

LHCb Computing Strategy

The CMS Computing Model

Monte Carlo Production on the Grid by the H1 Collaboration

The software library of the Belle II experiment

The High-Level Dataset-based Data Transfer System in BESDIRAC

Belle II Software and Computing

Andrea Sciabà CERN, Switzerland

Physics Analysis Software Framework for Belle II

Summary of the LHC Computing Review

Preparing for High-Luminosity LHC. Bob Jones CERN Bob.Jones <at> cern.ch

LHCb Distributed Conditions Database

DESY at the LHC. Klaus Mőnig. On behalf of the ATLAS, CMS and the Grid/Tier2 communities

ALICE ANALYSIS PRESERVATION. Mihaela Gheata DASPOS/DPHEP7 workshop

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

Data handling and processing at the LHC experiments

Beam test measurements of the Belle II vertex detector modules

The Belle II Software From Detector Signals to Physics Results

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Tier-2 structure in Poland. R. Gokieli Institute for Nuclear Studies, Warsaw M. Witek Institute of Nuclear Physics, Cracow

Data oriented job submission scheme for the PHENIX user analysis in CCJ

Hall D and IT. at Internal Review of IT in the 12 GeV Era. Mark M. Ito. May 20, Hall D. Hall D and IT. M. Ito. Introduction.

Batch Services at CERN: Status and Future Evolution

The LHC Computing Grid. Slides mostly by: Dr Ian Bird LCG Project Leader 18 March 2008

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

LHCb Computing Status. Andrei Tsaregorodtsev CPPM

CMS Computing Model with Focus on German Tier1 Activities

CERN and Scientific Computing

The Grid: Processing the Data from the World s Largest Scientific Machine

First Experience with LCG. Board of Sponsors 3 rd April 2009

The ATLAS Production System

Virtualizing a Batch. University Grid Center

Data handling with SAM and art at the NOνA experiment

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN

arxiv:cs/ v2 [cs.dc] 12 Mar 2004

The LHC Computing Grid

CERN openlab II. CERN openlab and. Sverre Jarp CERN openlab CTO 16 September 2008

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Distributed e-infrastructures for data intensive science

New strategies of the LHC experiments to meet the computing requirements of the HL-LHC era

HEP Grid Activities in China

Grid Computing Activities at KIT

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

Reprocessing DØ data with SAMGrid

Clouds at other sites T2-type computing

arxiv: v1 [physics.ins-det] 13 Dec 2018

Clouds in High Energy Physics

IEPSAS-Kosice: experiences in running LCG site

ATLAS Analysis Workshop Summary

Computing for ILC experiments

Belle Monte-Carlo production on the Amazon EC2 cloud Martin Sevior, Tom Fifield (University of Melbourne)

ATLAS 実験コンピューティングの現状と将来 - エクサバイトへの挑戦 坂本宏 東大 ICEPP

DESY. Andreas Gellrich DESY DESY,

Overview of ATLAS PanDA Workload Management

Considerations for a grid-based Physics Analysis Facility. Dietrich Liko

A L I C E Computing Model

Grid and Cloud Activities in KISTI

How the Monte Carlo production of a wide variety of different samples is centrally handled in the LHCb experiment

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

SiD Simulation & Reconstruction. Norman Graf (for the sim/reco team) LCWS 2010, Beijing March 28, 2010

LHCb Computing Resources: 2018 requests and preview of 2019 requests

The LHC computing model and its evolution. Dr Bob Jones CERN

PoS(EGICF12-EMITC2)106

Spanish Tier-2. Francisco Matorras (IFCA) Nicanor Colino (CIEMAT) F. Matorras N.Colino, Spain CMS T2,.6 March 2008"

Installation of CMSSW in the Grid DESY Computing Seminar May 17th, 2010 Wolf Behrenhoff, Christoph Wissing

Computing for LHC in Germany

Data Analysis in ATLAS. Graeme Stewart with thanks to Attila Krasznahorkay and Johannes Elmsheuser

The LHC Computing Grid

Distributed Computing Framework. A. Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille

DIRAC pilot framework and the DIRAC Workload Management System

Data Transfers Between LHC Grid Sites Dorian Kcira

BOSS and LHC computing using CernVM and BOINC

Using a dynamic data federation for running Belle-II simulation applications in a distributed cloud environment

Computing. DOE Program Review SLAC. Rainer Bartoldus. Breakout Session 3 June BaBar Deputy Computing Coordinator

Data Processing and Analysis Requirements for CMS-HI Computing

Update of the Computing Models of the WLCG and the LHC Experiments

LHC Computing Models

Austrian Federated WLCG Tier-2

Status of KISTI Tier2 Center for ALICE

Big Data Analytics and the LHC

Data Management for the World s Largest Machine

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

HIGH ENERGY PHYSICS ON THE OSG. Brian Bockelman CCL Workshop, 2016

RUSSIAN DATA INTENSIVE GRID (RDIG): CURRENT STATUS AND PERSPECTIVES TOWARD NATIONAL GRID INITIATIVE

ATLAS ITk Layout Design and Optimisation

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 27 July 2010

ATLAS Experiment and GCE

ATLAS PILE-UP AND OVERLAY SIMULATION

Challenges of the LHC Computing Grid by the CMS experiment

Data and Analysis preservation in LHCb

Grid Computing at Ljubljana and Nova Gorica

N. Marusov, I. Semenov

ATLAS operations in the GridKa T1/T2 Cloud

The ATLAS Distributed Analysis System

Long term data preservation and virtualization

Beob Kyun KIM, Christophe BONNAUD {kyun, NSDC / KISTI

Transcription:

Computing at Belle II CHEP 22.05.2012 Takanori Hara for the Belle II Computing Group

Physics Objective of Belle and Belle II Confirmation of KM mechanism of CP in the Standard Model CP in the SM too small (by many orders of magnitude) to generate observed baryon asymmetry in the universe Need sources of CP beyond the SM Super B factory Complementary to LHCb Page 2

KEK Site Tsukuba Tokyo e+ e Narita Mt. Fuji Page 3

Projection of Luminosity at SuperKEKB Page 4

Estimated Data Rates Experiment Event Size [kb] Rate [Hz] Rate [MB/s] High rate scenario for Belle II DAQ: Belle II 300 6,000 1,800 ALICE (HI) 12,500 100 1,250 ALICE (pp) 1,000 100 100 ATLAS 1,600 200 320 CMS 1,500 150 225 2,000 50 LCG TDR (2005): High data rate is a challenge! LHCb 25 Page 5

Belle II Collaboration ~400 members 65 institutes from 19 countries Distributed collaboration Page 6

Considerations for Belle II Computing Belle: Computing centralized at KEK Belle II: 50 times more data, distributed collaboration Go for a distributed computing system? More complicated, requires more effort to set up and maintain It allows the Belle II members to contribute via computing facilities in their country It provides redundancy The distributed infrastructure already exists Use existing grid sites and services (glite, EMI, DIRAC) Page 7

Grid Sites Country Sites Belle VO Comment Australia Tier2/3 Supported Cloud system planned Austria Tier2 China Tier2 Czech Republic Tier2 Supported Germany Tier1/2 Supported India Tier2 Japan KEK Supported Korea Tier2 Supported Poland Tier2/3 Supported Russia Tier2 Slovenia Tier2 Taiwan Tier1/2 USA OSG Poster Session 1, #159, M.Sevior DIRAC server New data center planned Cloud system developed Supported Supported Site at PNNL is set up Page 8

Computing Tasks Raw data processing Tape as storage medium Store and process at KEK, replication to just one remote site Simpler than LCG model Monte Carlo Production 6 times the real data size Produced in managed way, (almost) no input data needed Well suited for a distributed environment, including cloud Physics Analysis Random, uncoordinated access Store input data on disk Ntuple analysis on local resources for fast turn-around Page 9

Computing Model Raw Data Storage and Processing MC Production (optional) MC Production and Ntuple Production Ntuple Analysis Page 10

Network Connections Page 11

Resource Estimates Estimates depend on several unknown parameters: Accelerator performance Event data size Simulation/reconstruction code performance Analysis requirements Page 12

Organized Analysis Problem: inefficient resource usage by many users Limit resources per user, but maintain free access to data Offer high-performance organized analysis as a service User User User User Group Convener User User User Develop analysis/skim modules User User Group Convener Group Convener Production Team Integrate code and run jobs Collect and check code Page 13

Distributed Computing System DIRAC (developed by LHCb) Pilot jobs VO centric Extendible Belle MC production on grid sites, local clusters, and Amazon EC2 AMGA Metadata of files and datasets Poster Session 2, #58, K.Cho Page 14

Distributed Computing System Happyface (developed by CMS) Monitoring of grid sites CVMFS Software distribution Page 15

Workflow Abstraction Input data files (analysis) jobs FileA Output data files FileX Job1 FileB FileY Job2 FileC... Input dataset FileZ... Project... Output dataset Don't deal with single files and jobs, but with datasets and projects Page 16

Analysis Projects Analysis projects provide high level user interface Bookkeeping of jobs Dataset: output files created by a project Tools to create, list, replicate, remove, download datasets Page 17

User Interface: gbasf2 Same python steering file as for offline basf2 job, but with additional parameters for the grid job Page 18

Analysis Code Submission with gbasf2 Three options planned: Project Locally compiled code User Requires SL5 compatible system DIRAC Compiled Code Job Job Job Page 19

Analysis Code Submission with gbasf2 Three options planned: Project Locally compiled code Requires SL5 compatible system DIRAC Code Prepr. Job Local source code User Automatic compilation in preprocessing job Compiled Code Job Job Job Page 20

Analysis Code Submission with gbasf2 Three options planned: Requires SL5 compatible system Code DIRAC SVN Prepr. Job Local source code Project Locally compiled code User Automatic compilation in preprocessing job Compiled Code Source code committed to svn repository Job Job Job Automatic commit of local code Automatic compilation in preprocessing job Code version stored in metadata of output dataset Documentation which code was used to produce a file Page 21

Software Framework: basf2 Inspired by frameworks of Belle (basf) + other experiments Used for simulation, reconstruction, analysis, and DAQ ROOT I/O as data format Software bus with dynamically loaded modules Python steering Parallel processing Poster Session 2, #155, R.Itoh Simulation: Geant4 Tracking: GenFit Event Processing, Thu 15:10, M.Nadler Poster Session 2, #74, J.Lettenbichler Page 22

Code Management Developer with different level of experience, distributed around the world Need reliable, user-friendly, well-maintainable code Tools: Central code repository (svn), code browser, doxygen, style formatting tool, issue tracker, twiki, continuous integration system (buildbot), mailing lists Page 23

Summary Belle II at the SuperKEKB accelerator will search for New Physics with O(50) times more data than current B factories Huge data volume is a challenge for the computing Formation of grid sites federation in progress Distributed computing user interface: gbasf2 Distributed computing system based on existing technologies and infrastructures Workflow abstraction with projects and datasets Easy transition from offline software framework: basf2 Upgrade of accelerator and detector to be complemented by upgrade of software and computing system Page 24

Backup Page 25

KEKB Performance Design World record luminosity: 2.1 x 1034 cm-2s-1 Twice design 1 ab-1 of integrated luminosity Page 26

SuperKEKB Upgrade: Nano Beam Scheme Belle II New IR e- 2.6 A New Superconducting / permanent final focusing quads near the IP e+ 3.6 A Replace long TRISTAN dipoles with shorter ones (HER). New damping ring Larger crossing angle 2φ = 22 mrad 83 mrad TiN coated beam pipe with antechambers New low emmitance electron source Smaller asymmetry 3.5 / 8 GeV 4 / 7 Gev Page 27

Belle II Detector Page 28

Tasks of Computing Facilities Page 29

Prompt Reconstruction DQM SVD CDC HLT EVB2 Express Reco selected events Constants Making Disk Calibration Constants... KLM PXD KEK Grid Farm Prompt Reco mdst Analysis Job Grid File Catalog Page 30

Python Steering Example import os from basf2 import * #Register modules gearbox = register_module("gearbox") #Set parameters gearbox.param("inputfilexml",os.path.join(basf2dir,"belle2.xml")) gearbox.param("savetoroot", True) gearbox.param("outputfileroot", "Belle2.root") #Create paths main = create_path() #Add modules to paths main.add_module(gearbox) #Process events process(main,1) Page 31

Database Distribution via Frontier Export/ Import xml DB Transparent to the user Page 32

gbasf2 Page 33

(Commercial) Cloud Computing Resource demands vary with time Fair-share can solve this issue only to some extent Cloud computing allows to buy resources on demand Well suited to absorb peaks in varying resource demand Page 34

Cloud Computing in Belle II Risk: vendor lock-in No permanent data storage on the cloud Much less critical for CPU resources Large data transfer / storage not cost efficient (now) Use cloud primarily for MC production No data processing Maybe physics analysis Accounting issues Baseline of computing resources provided by the grid Cloud computing is option for peak demands Page 35