Data Reconstruction in Modern Particle Physics

Similar documents
Fast pattern recognition with the ATLAS L1Track trigger for the HL-LHC

PoS(EPS-HEP2017)492. Performance and recent developments of the real-time track reconstruction and alignment of the LHCb detector.

Adding timing to the VELO

CMS Conference Report

Tracking and Vertex reconstruction at LHCb for Run II

PoS(High-pT physics09)036

Tracking and flavour tagging selection in the ATLAS High Level Trigger

Muon Reconstruction and Identification in CMS

Track reconstruction with the CMS tracking detector

b-jet identification at High Level Trigger in CMS

ATLAS PILE-UP AND OVERLAY SIMULATION

MIP Reconstruction Techniques and Minimum Spanning Tree Clustering

PoS(EPS-HEP2017)523. The CMS trigger in Run 2. Mia Tosi CERN

Track pattern-recognition on GPGPUs in the LHCb experiment

CMS FPGA Based Tracklet Approach for L1 Track Finding

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

First LHCb measurement with data from the LHC Run 2

8.882 LHC Physics. Track Reconstruction and Fitting. [Lecture 8, March 2, 2009] Experimental Methods and Measurements

CSCS CERN videoconference CFD applications

ATLAS NOTE. December 4, ATLAS offline reconstruction timing improvements for run-2. The ATLAS Collaboration. Abstract

Performance of Tracking, b-tagging and Jet/MET reconstruction at the CMS High Level Trigger

ALICE tracking system

Deep Learning Photon Identification in a SuperGranular Calorimeter

A New Segment Building Algorithm for the Cathode Strip Chambers in the CMS Experiment

ATLAS ITk Layout Design and Optimisation

Physics CMS Muon High Level Trigger: Level 3 reconstruction algorithm development and optimization

Modules and Front-End Electronics Developments for the ATLAS ITk Strips Upgrade

PoS(TIPP2014)204. Tracking at High Level Trigger in CMS. Mia TOSI Universitá degli Studi di Padova e INFN (IT)

The CMS Computing Model

The GAP project: GPU applications for High Level Trigger and Medical Imaging

Robustness Studies of the CMS Tracker for the LHC Upgrade Phase I

HPS Data Analysis Group Summary. Matt Graham HPS Collaboration Meeting June 6, 2013

ATLAS, CMS and LHCb Trigger systems for flavour physics

Performance of the ATLAS Inner Detector at the LHC

Primary Vertex Reconstruction at LHCb

Stephen J. Gowdy (CERN) 12 th September 2012 XLDB Conference FINDING THE HIGGS IN THE HAYSTACK(S)

Integrated CMOS sensor technologies for the CLIC tracker

π ± Charge Exchange Cross Section on Liquid Argon

arxiv:hep-ph/ v1 11 Mar 2002

The LHCb Upgrade. LHCC open session 17 February Large Hadron Collider Physics (LHCP) Conference New York, 2-7 June 2014

LHCb Computing Resources: 2018 requests and preview of 2019 requests

PoS(IHEP-LHC-2011)002

The Compact Muon Solenoid Experiment. Conference Report. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

Charged Particle Reconstruction in HIC Detectors

GLAST tracking reconstruction Status Report

Prompt data reconstruction at the ATLAS experiment

Real-time Analysis with the ALICE High Level Trigger.

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN

HLT Hadronic L0 Confirmation Matching VeLo tracks to L0 HCAL objects

Full Offline Reconstruction in Real Time with the LHCb Detector

Tracking and compression techniques

Determination of the aperture of the LHCb VELO RF foil

The performance of the ATLAS Inner Detector Trigger Algorithms in pp collisions at the LHC

Data Analysis in ATLAS. Graeme Stewart with thanks to Attila Krasznahorkay and Johannes Elmsheuser

Electron and Photon Reconstruction and Identification with the ATLAS Detector

Event reconstruction in STAR

Precision Timing in High Pile-Up and Time-Based Vertex Reconstruction

Charged Particle Tracking at Cornell: Gas Detectors and Event Reconstruction

Alignment of the ATLAS Inner Detector tracking system

Particle Track Reconstruction with Deep Learning

LHCb Computing Resources: 2019 requests and reassessment of 2018 requests

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

Simulating the RF Shield for the VELO Upgrade

PoS(Baldin ISHEPP XXII)134

Tracking and Vertexing performance in CMS

1. INTRODUCTION 2. MUON RECONSTRUCTION IN ATLAS. A. Formica DAPNIA/SEDI, CEA/Saclay, Gif-sur-Yvette CEDEX, France

THE ATLAS INNER DETECTOR OPERATION, DATA QUALITY AND TRACKING PERFORMANCE.

HEP Experiments: Fixed-Target and Collider

Beam test measurements of the Belle II vertex detector modules

LHC Detector Upgrades

The HEP.TrkX Project: Deep Learning for Particle Tracking

First results from the LHCb Vertex Locator

TORCH: A large-area detector for precision time-of-flight measurements at LHCb

Track reconstruction of real cosmic muon events with CMS tracker detector

Evaluation of the computing resources required for a Nordic research exploitation of the LHC

Design of the new ATLAS Inner Tracker (ITk) for the High Luminosity LHC

The LHCb upgrade. Outline: Present LHCb detector and trigger LHCb upgrade main drivers Overview of the sub-detector modifications Conclusions

Atlantis: Visualization Tool in Particle Physics

IEPSAS-Kosice: experiences in running LCG site

New strategies of the LHC experiments to meet the computing requirements of the HL-LHC era

TPC Detector Response Simulation and Track Reconstruction

Trigger and Data Acquisition at the Large Hadron Collider

3D-Triplet Tracking for LHC and Future High Rate Experiments

Modelling of non-gaussian tails of multiple Coulomb scattering in track fitting with a Gaussian-sum filter

Data handling and processing at the LHC experiments

Virtualizing a Batch. University Grid Center

The AMS-02 Anticoincidence Counter. Philip von Doetinchem I. Phys. Inst. B, RWTH Aachen for the AMS-02 Collaboration DPG, Freiburg March 2008

Upgraded Swimmer for Computationally Efficient Particle Tracking for Jefferson Lab s CLAS12 Spectrometer

CMS High Level Trigger Timing Measurements

Study of the Higgs boson coupling to the top quark and of the b jet identification with the ATLAS experiment at the Large Hadron Collider.

Data Quality Monitoring at CMS with Machine Learning

PoS(ACAT08)101. An Overview of the b-tagging Algorithms in the CMS Offline Software. Christophe Saout

Summary of the LHC Computing Review

Charged Particle Tracking at Cornell: Gas Detectors and Event Reconstruction

Overview. About CERN 2 / 11

CMS Alignement and Calibration workflows: lesson learned and future plans

SoLID GEM Detectors in US

Clustering and Reclustering HEP Data in Object Databases

Status of the TORCH time-of-flight detector

Transcription:

Data Reconstruction in Modern Particle Physics Daniel Saunders, University of Bristol 1

About me Particle Physics student, final year. CSC 2014, tcsc 2015, icsc 2016 Main research interests. Detector upgrades for LHC - pixel detectors for LHCb. Neutrino experiments at nuclear reactors - data reconstruction and analysis. But day to day Professional ROOT and git complainer. Developing C++ and python projects to perform above. 2

But it s not all about me I assume you know a bit about: The LHC. LHC-like particle physics detectors, and that they give information like: Particle energies. Particle paths. The Tardis prompts questions for the audience. 3

Aim of these lectures What just happened? Example collision event from CMS. 4

Aim of these lectures What just happened? LHC detectors produces O(10) petabytes of data per year [1]. Data is processed to the stage of physics papers measurements and discoveries. Example collision event from CMS. Higgs discovery at CMS. Many steps involved. Each step has computing costs, varying inefficiencies, often in large backgrounds. We ll consider some steps in detail, looking at tradeoffs between these three factors. 5

Lecture Outline - Scope Aim of these lectures: How to take detector output and make physics measurements. Lecture 1 Introduction and context: Elements of LHC detectors (CMS & LHCb). Data rates and formats. Data taking strategies: Triggers. Lecture 2 Event reconstruction principles: Tracking. Particle identification. Vertexing. Optimisations: Parallelism. The Grid. 6

Lecture Outline - Beyond Scope Particle Physics detectors exist: Able to measure the position of a crossing particle and energy. Measurements are to a limited resolution. Detectors aren t perfect: can be inefficient, impure and have bad resolution. The exact workings of detectors is not considered. Lecture concepts are general and apply to all particle physics experiments. For detector concepts, see: Many physics measurements are made by comparing reconstructed data to simulation. A large topic! Many sophisticated data analysis techniques used. We will stop at this point. Many of these topics are discussed in other lectures of this school. 7

Lecture 1 Introduction and Data Taking 8

Introduction and Motivation LHC computing is an enormous task: Data rate of typical LHC experiments in run two (i.e. current) ~1GB/s [2]. In extreme cases (CMS [3] & ATLAS), this is post 1 in ~10,000 filter by electronics. Still large background of uninteresting events. Collaborations of thousands of scientists: Many kinds of analysis requiring different data selection. Competitive spirit (e.g. ATLAS vs CMS). Many kinds of people (Computer Scientists, Engineers, Physicists etc.) with varying computing skills. 9

LHC Reminder LHC is a proton - proton (or Pb - Pb) collider. Each collision is called an event. LHC examples used in these lectures, but many concepts apply to other big data experiments in particle physics and beyond. 10

LHC Collisions Particles are grouped into bunches of ~ 100 billion protons. LHC collides two proton beams, circling in opposite directions. A collision between two bunches happens once every 25ns (so 40MHz). In each collision, ~25 pairs of proton collisions. About 1 per million is used for physics. The rest is well understood background of un-interesting events. 11

LHC Detectors Large dedicated detectors built surrounding four collision points: ATLAS & CMS: general purpose detectors, discovery machines. ALICE: high energy plasma studies, conditions approx 10-6 seconds post big bang. LHCb: precision antimatter studies. Many other smaller experiments. Detectors generally either perform: Direct searches (e.g. CMS) - looking for new phenomena, such as the Higgs Boson. Indirect searches (e.g. LHCb) - compare precision measurements to theoretical predictions. Differences can be a sign of new physics that has been unaccounted. Choose CMS and LHCb to look at in more detail. 12

LHC Detectors - CMS Calorimeters - measures energy. Some particles stop. Muon chambers - detect particles that leave the detector (most likely muons). Silicon tracker - tracks particles near the collision region. Magnet - bends tracks, allowing for momentum measurement. 13

LHC Detectors - CMS CMS slice, with different particle examples. 14

LHC Detectors - LHCb Calorimeters - measures energy. Some particles stop. RICH (Ring Imaging Cherenkov Detector) - particle ID from radiation induced by crossing particles. Silicon tracker - tracks particles near the collision region. Magnet - bends charged tracks, allowing for momentum measurement. Muon chambers - detect particles that leave the detector (most likely muons). 15

All LHC computing is connected to data processing. The Task What just happened? Need to take detector output (~1GB/s of electrical signals) and perform: Needed live. Processed similar rate to data taking. Detector related studies: Online data quality monitoring. Calibrations. Physics analysis: Discovery of new particles. Comparisons with simulation. Example collision event from CMS. 16

Data Flow Data reconstruction generally involves several steps of processing and reduction: Detector output Physics measurements 17

Data Flow Data reconstruction generally involves several steps of processing and reduction: 0x01e84c10: 0x01e8 0x01e84c10: 0x0000 0x01e84c10: 0x0000 0x01e84c10: 0x01e8 0x01e84c10: 0x0000 0x01e84c10: 0x0000 0x01e84c10: 0x01e8 0x01e84c10: 0x01e8 Stage Trigger Event Reconstruction Stripping (AKA Skimming) Description Hardware Implemented Timescale Data reduction factor Initial selection for finding interesting events. Local electronics or CPU/ GPU processing farm. Live. Reconstruct triggered data into list of particles. Inside trigger and/or the Grid (see later). Almost live (requires detector calibration). Repeated ~yearly. Signature selection trained by prior physics knowledge. The Grid. Any point, ~monthly turn around. 10 6 * (permanent loss). 10x (used for Physics). Analysis dependant. *CMS example. Lecture 1 Lecture 2 18

Triggers 19

Triggers The trigger decides what data to save in real time. Typically just 1 in a million events considered interesting by trigger. Can be formed of several levels, data reduction and increasing complexity at each step. Executed on FPGAs and/or local CPU/GPU processing farms. Use simple algorithms that perform quickly. Perform partial/full event reconstruction (see later). Efficient (aim >80%) to maximise storage and reconstruction resources. All other events permanently discarded. Many O(50) configurations used (a few examples shown later): Tuned to physics analysis of the experiment. Try to be least biased whilst recording interesting events. 20

Triggers The trigger decides what data to save in real time. Typically just 1 in a million events considered interesting by trigger. Can be formed of several levels, data reduction and increasing complexity at each step. Executed on FPGAs and/or local CPU/GPU processing farms. Use simple algorithms that perform quickly. Perform partial/full event reconstruction (see later). Efficient (aim >80%) to maximise storage and reconstruction resources. All other events permanently discarded. Many O(50) configurations used (a few examples shown later): Tuned to physics analysis of the experiment. Try to be least biased whilst recording interesting events. A theorist has just predicted a new physics particle, produced as frequently as the Higgs Bosons in LHC, but with significantly different topology and energy distributions: what are the chances it s discovered? Almost zero! 21

Triggers Worked Example - LHCb Data rate of entire detector too high for all to be used in trigger: Use a subset of information. Introduce multiple levels of triggers use more information in higher levels. May deliberately reduce resolution of detectors to reduce data size. E.g. combine cells in tracker, or use a less precise data type. LHCb model is very efficient, allowing for physics analysis immediately after trigger - not always possible. Particles cross each other every 25ns. First trigger selects 1 in 40 events. Performed in hardware, using subset of detector data. Software trigger selects 1 in 100 events. Since called less frequently, can use full detector information for full event reconstruction. Combined trigger selects 1 in 40k events for physics analysis. 22

Triggers Worked Example - LHCb Data rate of entire detector too high for all to be used in trigger: Use a subset of information. Introduce multiple levels of triggers use more information in higher levels. May deliberately reduce resolution of detectors to reduce data size. E.g. combine cells in tracker, or use a less precise data type. LHCb model is very efficient, allowing for physics analysis immediately after trigger - not always possible. CMS selects 1 in O(10 6 ), where as LHCb selects 1 in O(10 4 ) for the same output data rate - what causes the difference? LHCb opening angle is smaller! Less coverage. LHCb events are very common - can reduce luminosity (i.e. number of collisions per crossing). Particles cross each other every 25ns. First trigger selects 1 in 40 events. Performed in hardware, using subset of detector data. Software trigger selects 1 in 100 events. Since called less frequently, can use full detector information for full event reconstruction. Combined trigger selects 1 in 40k events for physics analysis. 23

Trigger Example 1 - ECal at CMS (L0) Filtering for events we don t know about. Typical algorithms identify data variables that distinguish signals from backgrounds. Difficult to define signal and background for discoveries - try to keep general: ECal Particles Missing Energy Vector Known processes are approx symmetrically distributed in energy. Missing energy could be a sign of something new produced - dark matter. Separation between background and W eν events as missing energy in the traversal direction. 24

Trigger Example 2 - Vertexing at LHCb Filtering for events we do know about. LHCb is designed for precision studies of collisions involving b quarks: Complimentary to direct searches (e.g. discovering new particles). Perform indirect searches by comparing experimental rates to theoretical predictions. Trigger can be designed for specific physics: typical signature involves a displaced vertex. Not easy to reconstruct - performed in software farm at higher level trigger. (Reconstruction of these objects discussed in lecture 2). p p 1cm Detector Vertex particle Tracked particle Displaced vertex in LHCb - signature of interesting event. LHCb VELO HLT ROC. 25

Trigger Bias Data sets from triggers inevitably biased by trigger. E.g. experiment finds deficit Higgs candidates with ET < 5 GeV (unsurprising if ET Trig = 5 GeV). Can be accounted for: Comparisons with simulation, although many factors (detector performance, collider conditions). Comparison with non-triggered data: Far lower rate! Have to extrapolate. 26

Trigger Hardware Example 1 - LHCb Frequent small events. Level Level0* High Level Trigger* Input rate 40 MHz 1MHz, 70GB/s Hardware FPGAs. Local cluster using 20k [7] CPUs. Output rate 1MHz 10kHz, 700MB/s [2] Event filter factor 40x 100x Notes Uses subset of detector data (ECal and muons only). Full reconstruction performed. *Values accurate to ±50%. 27

Lecture 1 Summary LHC detectors output a large amount of data. Four major experiments, each outputting O(1) GB/s physics data (post trigger). Many results required live (detector monitoring, calibrations etc.). Reconstruction performed on the GRID (Lecture 2). Triggers are used to filter for interesting events in real time. Can be multilayered, using electronics, CPUs and GPUs. Complexity ranges from simple algorithms through to full event reconstruction. Designs heavily influenced by the physics aims of the experiment. 0x01e84c10: 0x01 0x01e84c10: 0x00 0x01e84c10: 0x00 0x01e84c10: 0x01 0x01e84c10: 0x00 0x01e84c10: 0x00 0x01e84c10: 0x01 0x01e84c10: 0x01 28

Lecture 2 Event Reconstruction 29

Event Reconstruction Triggered detector collision data particle interactions. Seek the following information as input for physics analysis: What particles were created? Where were they produced? What were the parent particles? To find this, perform (at least): Tracking: Reconstruct particle trajectories into tracks. Vertexing: Group particles into vertices. Particle ID: Find the particle identification of each track (e.g. a muon, electron etc.). Requirements: Fast. Good quality for physics analysis. Usually anti correlated - a fast algorithm often leads to inefficiency and impurities (see later). 30

Tracking Algorithms Aim: to play a game join the dots at 1kHz with many fake dots. Tracking particles through detectors involves two step. Pattern recognition: identifying which detector hits for a track. Track fit: approximate the path of the particle with an equation. No one size fits all solution. Many detectors use different combinations of algorithms (e.g. LHCb uses 4 different algorithms for difference combinations of sub detectors, but basic ideas are the same). Usually a trade off between: Efficiency: fraction of real tracks found Purity: fraction of tracks that are real Typically these two are anti correlated: a good efficiency typically has a bad purity, and vice versa. Both good efficiency and purity is usually computationally expensive - see later. Computational speed. 31

One of the hardest cases - Pb 32 collisions in ALICE, a real event.

Tracking - Pattern Recognition Example Consider tracking in the Vertex Locator (VELO) of LHCb. The VELO is a silicon strip detector, consisting of 42 layers. Gives an x, y and z position. LHCb VELO 33

Tracking - Pattern Recognition Example Looking side on: Particle tracks clearly visible to eye. Extra hits present, typically electrical noise or secondary short tracks. Recall data points in the format: (x, y, z, time) Time resolution only accurate to which collision the particles come from (25ns, sometimes worse ). x z (beam) Have to find an algorithm to track using this information and in these conditions. Many choices - consider the following (LHC) examples LHCb VELO data event (2d projection) 34

Tracking - Pattern Recognition Example Name Description Scalability Combinatorial Form every track from each possible combination of hits. Access each track by quality (e.g. χ 2 ) and tag. ntracks! LHCb VELO data event (2d projection, top half) 35

Tracking - Pattern Recognition Example Name Description Scalability Combinatorial Form every track from each possible combination of hits. Access each track by quality (e.g. χ 2 ) and tag. ntracks! LHCb VELO data event (2d projection, top half) 36

Tracking - Pattern Recognition Example Name Description Scalability Combinatorial Form every track from each possible combination of hits. Access each track by quality (e.g. χ 2 ) and tag. ntracks! LHCb VELO data event (2d projection, top half) 37

Tracking - Pattern Recognition Example Name Description Scalability Combinatorial Form every track from each possible combination of hits. Access each track by quality (e.g. χ 2 ) and tag. ntracks! LHCb VELO data event (2d projection, top half) 38

Tracking - Pattern Recognition Example Name Description Scalability Combinatorial Hough Transform Form every track from each possible combination of hits. Access each track by quality (e.g. χ 2 ) and tag. Transform points into a system where clusters form. If straight tracks, take the difference between consecutive hits. Group (e.g. in a histogram) and tag peaks. ntracks! x LHCb VELO data event (2d projection, top half) 39

Tracking - Pattern Recognition Example Name Description Scalability Combinatorial Hough Transform Form every track from each possible combination of hits. Access each track by quality (e.g. χ 2 ) and tag. Transform points into a system where clusters form. If straight tracks, take the difference between consecutive hits. Group (e.g. in a histogram) and tag peaks. ntracks! x LHCb VELO data event (2d projection, top half) 40

Tracking - Pattern Recognition Example Name Description Scalability Combinatorial Hough Transform Seeding Form every track from each possible combination. Access each track by quality (e.g. χ 2 ) and tag. Transform points into a system where clusters form. E.g. for straight tracks, take the difference between consecutive hits. Group (e.g. in a histogram) and tag peaks. Form seeds from pairs of hits on a sub set of the detector. Extrapolate the seed and count hits intercepted. Tag if sufficient number of hits. ntracks! x nlog(n) LHCb VELO data event (2d projection, top half) 41

Tracking - Pattern Recognition Example Name Description Scalability Combinatorial Hough Transform Seeding Form every track from each possible combination. Access each track by quality (e.g. χ 2 ) and tag. Transform points into a system where clusters form. E.g. for straight tracks, take the difference between consecutive hits. Group (e.g. in a histogram) and tag peaks. Form seeds from pairs of hits on a sub set of the detector. Extrapolate the seed and count hits intercepted. Tag if sufficient number of hits. ntracks! x nlog(n) Question: which algorithm to pick? Even in the case where efficiency and purity are constant? Still not enough information 42

Tracking - Pattern Recognition Algorithms Recall three main factors in choosing such algorithms: Efficiency: fraction of real tracks found Purity: fraction of tracks that are real Computational speed. Toy simulation for LHCb VELO: LHCb VELO toy event (2d projection) 43

Tracking - Pattern Recognition Algorithms Recall three main factors in choosing such algorithms: Efficiency: fraction of real tracks found Purity: fraction of tracks that are real Computational speed. Any use case for green? Curves! Toy simulation for LHCb VELO: 44

Tracking - Pattern Recognition Algorithms Typically use a combination of these algorithms. Each example taken from LHC activities: Combinatorial often used at testbeams: Low occupancy, so fast. Efficient and pure. Timepix3 Tracking Telescope Testbeam Data Hough transforms used for more complicated shapes (e.g. rings in LHCb RICH*). LHCb RICH Subdetector RICH Data All LHC experiments use seeding extensively (highest occupancy). ATLAS Tracker *Often not needed to actually reconstruct rings. ATLAS Inner Layers 45

Track Fitting Tracking particles through detectors involves two step. Pattern recognition: identifying which detector hits for a track. Track fit: approximate the path of the particle with an equation. Typically use a Kalman filter. Basic steps: Track is approximated as a zig-zag (fewer free parameters than co-ordinates!). Start with seed or estimate of track parameters (e.g. straight line fit). Propagate to the next plane (approximating B field, account for scattering in material). Predict position of next particle, weighting by closest hits (needs too be tuned). Kalman Filter Example 46

Track Refinement Common to tune pattern recognition to be efficient and impure refine selection later using full particle information. Can use χ 2 to find well fitting tracks. Can also use/combine with other parameters: Number of hits (complimentary information to χ 2 ). Fits from different sub detectors LHCb Typically build an MVA out of different quality parameters - LHCb uses a neutral net. Caution: if fake/ghost tracks are formed from parts of real tracks, they may be lost. 47

Tracking Notes Kalman filter and pattern recognition can be merged into a single step computationally more efficient (see later). Detector hits can sometimes form part of multiple tracks. Detector spatial resolution too low to separate near tracks. Secondary tracks produced by interactions with the detector material. 48

Vertexing (Briefly) Vertexing involves clustering tracks that originate from the same point. Easy in cases where vertex location is known - extrapolate all tracks and apply selection criteria. Else, Physics input can narrow search region significantly. Can use analytic methods (e.g. distance of closest approach) to seed search. Common to seed by projecting into 2D plane and searching for point of high track density (essentially a peak finding/clustering problem). p p x p 1cm Vertex example in LHCb Velo. Detector Vertex particle Tracked particle End on projection. 49

Particle ID (Briefly) Classify each track as a type of particle event by event: Needed to refine selections for offline analysis (remove background). Many kinds of particle: Not just fundamental particles, also composite hadrons (e.g. Pion, Kaon). Some easy cases: 50

Particle ID (Briefly) Many techniques needed for other cases. RICH detector at LHCb uses Cherenkov radiation: Light emitted when a particle slows passing through a material. Emission is isotropic, and forms rings on detectors. Often not required to reconstruct the ring itself - instead, test different hypothesis. Sometimes multiple solutions (e.g. high momentum) - can still assign probabilities. 51

Particle ID (Briefly) Many techniques needed for other cases. RICH detector at LHCb uses Cherenkov radiation: Light emitted when a particle slows passing through a material. Emission is isotropic, and forms rings on detectors. Often not required to reconstruct the ring itself - instead, test different hypothesis. Sometimes multiple solutions (e.g. high momentum) - can still assign probabilities. 52

Event Reconstruction Implementation 53

Event Reconstruction Implementation Each reconstruction stage typically (sometimes by necessity) follows sequentially, e.g: Input Histogram plotting Track finding Track fitting Vertexing Particle ID Histogram plotting Output Such a chain can be performed for a single event, or large set of events. Reminder: each event is (usually) statistically independent of each-other. Strategy for single core is obvious, but for multi core, not so much. Nowadays, reconstruction involves tens of thousands of CPUs worldwide - need efficient strategy. Currently limited by memory: E.g. CMS end of 2011 could only 6 out of 8 cores on average. 54

Event Reconstruction Optimisations Some advantages: Data is essentially a list of collision data. Each collision is (mostly) separate from the next. Some disadvantages: Many packages needed, not all thread safe. Large overhead of memory to run even a small job, can quickly reach O(1)GB: (e.g. calibration and alignment values). External libraries. 55

Event Reconstruction Optimisations - Parallelism Many possible optimisation strategies: Parallelism - many possible options: Run multiple jobs on the same multi core machine. Very common strategy currently. Split input file (easy), and merge output file (easy). Potential pitfalls? Memory! 56

Event Reconstruction Optimisations - Parallelism Many possible optimisation strategies: Parallelism - many possible options: Many segments can be performed in parallel. Many segments (e.g. tracking) depend on those previous, which can give long serial sections. Potential pitfalls? Amdahl s Law! Typically limited at factor of 5 [9] speedup. 57

Event Reconstruction Optimisations - Parallelism Large segments of sequential code limit theoretical speedup: E.g. tracking, expected to get harder. Typically limited to factor of 5 speedup (~80% parallel portion). Successfully implemented at CMS (see later). 58

Event Reconstruction Optimisations - Parallelism Many possible optimisation strategies: Parallelism - many possible options: Process multiple events concurrently in one big job. Large overhead of memory (e.g. calibration constants) can be shared between threads. Time 59

Event Reconstruction Optimisations - Parallelism Current limited by memory and not CPU power. Manageable at run 1, but not for higher luminosity. Parallelism strategy aims for similar performance whilst saving memory. 60

The Grid Majority of reconstruction jobs performed on the Grid (excluding HLTs): ~2 million jobs processed each day. 170 computing centres in 42 different countries. Managed by each institute: Large variation of technology and software used. 61

Centres are organised in a tier system. Tier 0 (CERN): Receives data from experiments. First copy of data to tape. Distribute reconstruction in Tier1. Re-reconsuction when LHC is not on. Tier 1 (distributed, 10GB/s connection): Offline event reconstruction performed here (&storage). Distribute jobs on Tier 2. Tier 2 (distributed): University sites. Ideal for analysis jobs performed on reconsulted data (e.g Higgs Searches). The Grid 62

Lecture 2 Summary Event reconstruction typically involves a set of algorithms executed in a particular order (often by necessity): Each algorithm offers varying efficiency, purity and speed. Tracking in particular often a large sequential portion of event reconstruction. Parallelism required to utilise available computing resources: No obvious methods to parallelise. Until recently, ok to have multiple processes on multi core machines: Now memory issues - multi threading required to reduce large overhead. In process of multi threading reconstruction packages. Many computing resources available, not just relying on CERN: The Grid is used for processing many reconstruction and analysis jobs. Spread worldwide, using a variety of different technologies. With flat computing budgets and increased luminosities, reconstruction is challenging: Clever ideas welcome! 63

References [1] Introduction to Physics Computing, CSC 2014. [2] CERN computing website. [3] CMS upgrade trigger TDR. [4] LHCb trigger TDR. [6] LHCb Starter Kit. [7] Sascha Stahl, LHCb High Level Trigger, EPS-HEP 2015. [8] ALICE HLT High Speed Tracking on GPU, IEEE Transactions on Nuclear Science, Vol. 58, No. 4, August 2011. [9] Software Design in the Many-Cores era, CSC 2014. [10] Study of a Fine Grained Threaded Framework Design, Christopher Jones, http://indico.cern.ch/event/149557/session/3/contribution/194/attachments/150160/212729/threaded_framework_talk.pdf [11] Using the CMS Threaded Application in a Production Environment, Christopher Jones, CHEP 2015, http://indico.cern.ch/event/304944/session/2/contribution/120/attachments/578690/796858/threading_production_chep2015.pdf + many more [12] LHCb Topological Trigger Reoptimization, https://cdsweb.cern.ch/record/2019813/files/lhcb-proc-2015-018.pdf 64