CMS FPGA Based Tracklet Approach for L1 Track Finding

Similar documents
The Compact Muon Solenoid Experiment. Conference Report. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

PoS(EPS-HEP2017)523. The CMS trigger in Run 2. Mia Tosi CERN

ATLAS ITk Layout Design and Optimisation

Tracking and Vertexing performance in CMS

Track reconstruction with the CMS tracking detector

LHC Detector Upgrades

Fast pattern recognition with the ATLAS L1Track trigger for the HL-LHC

The LHCb upgrade. Outline: Present LHCb detector and trigger LHCb upgrade main drivers Overview of the sub-detector modifications Conclusions

The performance of the ATLAS Inner Detector Trigger Algorithms in pp collisions at the LHC

Tracking and flavour tagging selection in the ATLAS High Level Trigger

PoS(TIPP2014)204. Tracking at High Level Trigger in CMS. Mia TOSI Universitá degli Studi di Padova e INFN (IT)

CMS Conference Report

Design of the new ATLAS Inner Tracker (ITk) for the High Luminosity LHC

ATLAS PILE-UP AND OVERLAY SIMULATION

Muon Reconstruction and Identification in CMS

Performance of Tracking, b-tagging and Jet/MET reconstruction at the CMS High Level Trigger

ATLAS, CMS and LHCb Trigger systems for flavour physics

CMS R&D for HL-LHC L1 Tracking

arxiv:hep-ph/ v1 11 Mar 2002

The ATLAS Trigger System: Past, Present and Future

Modules and Front-End Electronics Developments for the ATLAS ITk Strips Upgrade

Physics CMS Muon High Level Trigger: Level 3 reconstruction algorithm development and optimization

b-jet identification at High Level Trigger in CMS

Performance studies of the Roman Pot timing detectors in the forward region of the IP5 at LHC

Adding timing to the VELO

The Compact Muon Solenoid Experiment. Conference Report. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

Electron and Photon Reconstruction and Identification with the ATLAS Detector

The LHCb Upgrade. LHCC open session 17 February Large Hadron Collider Physics (LHCP) Conference New York, 2-7 June 2014

First Operational Experience from the LHCb Silicon Tracker

3D-Triplet Tracking for LHC and Future High Rate Experiments

Direct photon measurements in ALICE. Alexis Mas for the ALICE collaboration

Track pattern-recognition on GPGPUs in the LHCb experiment

The Track-Finding Processor for the Level-1 Trigger of the CMS Endcap Muon System

First results from the LHCb Vertex Locator

Associative Memory Pattern Matching for the L1 Track Trigger of CMS at the HL-LHC

Tracking POG Update. Tracking POG Meeting March 17, 2009

The Phase-2 ATLAS ITk Pixel Upgrade

Construction of the Phase I upgrade of the CMS pixel detector

THE ATLAS INNER DETECTOR OPERATION, DATA QUALITY AND TRACKING PERFORMANCE.

Integrated CMOS sensor technologies for the CLIC tracker

Performance of the ATLAS Inner Detector at the LHC

Charged Particle Reconstruction in HIC Detectors

Time of CDF (II)

ATLAS NOTE. December 4, ATLAS offline reconstruction timing improvements for run-2. The ATLAS Collaboration. Abstract

PoS(IHEP-LHC-2011)002

Real-time Analysis with the ALICE High Level Trigger.

Burn-In Software Development. Taras Fedorchuk, Taras Shevchenko National University of Kyiv 3rd September 2018

Robustness Studies of the CMS Tracker for the LHC Upgrade Phase I

A LVL2 Zero Suppression Algorithm for TRT Data

HPS Data Analysis Group Summary. Matt Graham HPS Collaboration Meeting June 6, 2013

Description and performance of track and primaryvertex reconstruction with the CMS tracker

TPC Detector Response Simulation and Track Reconstruction

Expected feedback from 3D for SLHC Introduction. LHC 14 TeV pp collider at CERN start summer 2008

Velo readout board RB3. Common L1 board (ROB)

L1 track trigger for the CMS HL-LHC upgrade using AM chips and FPGAs

Software for implementing trigger algorithms on the upgraded CMS Global Trigger System

HLT Hadronic L0 Confirmation Matching VeLo tracks to L0 HCAL objects

A 3-D Track-Finding Processor for the CMS Level-1 Muon Trigger

Straw Detectors for the Large Hadron Collider. Dirk Wiedner

Endcap Modules for the ATLAS SemiConductor Tracker

The Compact Muon Solenoid Experiment. Conference Report. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

Stefania Beolè (Università di Torino e INFN) for the ALICE Collaboration. TIPP Chicago, June 9-14

MIP Reconstruction Techniques and Minimum Spanning Tree Clustering

CSC Trigger Motherboard

PoS(High-pT physics09)036

Upgrading the ATLAS Tile Calorimeter electronics

TPC Detector Response Simulation and Track Reconstruction

ATLAS Simulation Computing Performance and Pile-Up Simulation in ATLAS

Tracking and compression techniques

Data Reconstruction in Modern Particle Physics

Alignment of the CMS silicon tracker using Millepede II

ATLAS Tracking Detector Upgrade studies using the Fast Simulation Engine

The new detector readout system for the ATLAS experiment

Precision Timing in High Pile-Up and Time-Based Vertex Reconstruction

Evaluation of GPU-based track-triggering for the CMS detector at CERN s HL-LHC

First LHCb measurement with data from the LHC Run 2

A New Variable-Resolution Associative Memory for High Energy Physics

Trigger and Data Acquisition: an ATLAS case study

Challenges and performance of the frontier technology applied to an ATLAS Phase-I calorimeter trigger board dedicated to the jet identification

TORCH: A large-area detector for precision time-of-flight measurements at LHCb

Track-Finder Test Results and VME Backplane R&D. D.Acosta University of Florida

The CLICdp Optimization Process

TPC Detector Response Simulation and Track Reconstruction

Level-1 Regional Calorimeter Trigger System for CMS

Alignment of the ATLAS Inner Detector tracking system

Tracking and Vertex reconstruction at LHCb for Run II

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN

A New Segment Building Algorithm for the Cathode Strip Chambers in the CMS Experiment

Track reconstruction of real cosmic muon events with CMS tracker detector

ALICE tracking system

The LHCb VERTEX LOCATOR performance and VERTEX LOCATOR upgrade

Full Offline Reconstruction in Real Time with the LHCb Detector

The ALICE electromagnetic calorimeter high level triggers

Level 0 trigger decision unit for the LHCb experiment

Investigating the Micron Automata Processor for HEP Pattern Recognition Applications

PSEC-4: Review of Architecture, etc. Eric Oberla 27-oct-2012

The CMS alignment challenge

SLAC Testbeam Data Analysis: High Occupancy Tracking & FE-I4 Cluster Study

Beam test measurements of the Belle II vertex detector modules

The GAP project: GPU applications for High Level Trigger and Medical Imaging

Transcription:

CMS FPGA Based Tracklet Approach for L1 Track Finding Anders Ryd (Cornell University) On behalf of the CMS Tracklet Group Presented at AWLC June 29, 2017 Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:1

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:2 Outline/Introduction Motivation for tracking in L1 Trigger Challenge and CMS HL-LHC tracker Tracklet algorithm Simulation and challenges Tracklet demonstrator Performance and results Next steps and outlook This project has involved developing tracking algorithms suitable for execution on FPGAs Student contributions to this effort has been critical Firmware written by students with engineering 'guidance' The goal of this project was to demonstrate the feasibility of tracking at 40 MHz by the fall of 2016 for the CMS tracker TDR We are now proceeding to defining the upgrade project based on the demonstration

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:3 HL-LHC Trigger Challenge The LHC is planned to be upgraded during Long Shutdown 3 (LS3) in 2024-2026 Leveled luminosity up to 7.5 10 34 cm -2 s -1 Pileup of up to 200 200 proton-proton interactions on average in each proton-proton bunch collision every 25 ns (40 MHz) About 8 10 9 proton-proton interactions per second For the HL-LHC operations CMS plans to record up to 7.5 khz of collisions Need to reduce rate by factor of ~5,000. This is done in two stages: L1 (hardware trigger) from 40 MHz 750 khz HLT (High Level Trigger in software) from 750 khz 7.5 khz. The current L1 trigger in CMS does not have access to tracking information. To meet the HL-LHC requirements the CMS L1 trigger will have access to tracking information (p T >2 GeV) for each bunch crossing

Example of tracking in L1: Muon Triggers Track matching to muon candidates has high efficiency Muons+L1Tracks provide much sharper threshold Sharp threshold allows a significant rate reduction: Factor ~10 reduction @20 GeV Tracking at L1 is also powerful tool for electrons, τ s, jets, photons as detailed in the CMS HL-LHC TP: https://cds.cern.ch/record/2020886 Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:4

HL-LHC Environment: Minbias 14 TeV minbias: 6.5 charged particles, mostly pions, per unit of rapidity or 33 charged particles in the tracking volume η <2.5 With 200 PU 6600 charged particles per bunch crossing. Soft spectrum peaks at p T of about 200 MeV. The average min bias event has 33 2.7%=0.89 tracks with p T >2.0 For PU=200 we expect ~180 tracks with p T >2.0. p T distribution in 14 TeV minbias 2.7% above 2.0 GeV 0.8% above 3.0 GeV Data volume: 200 PU average of about 13,000 stubs/bunch crossing Each stub is ~36 bits At 40 MHz BX rate we have 20 Tbits/s Only ~5% of stubs are from tracks with p T >2GeV Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:5

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:6 Correlating hits in closely spaced sensors give p T discrimination in B=3.8 T Correlations formed on module data reduction for trigger readout p T Modules Strip-strip (2S) Modules 2 5 cm strips 90 um pitch Pixel-strip (PS) Modules 2 2.5 cm strips 100 um pitch 1.5 mm macro pixels

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:7 CMS HL-LHC Tracker Layout ¼ of Outer Tracker 6 Layers 5 Disks Layer 1 Stub Finding Efficiency Stub Finding Efficiency per Layer

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:8 L1 Tracking Goals Highest possible efficiency for high p T isolated tracks Lepton triggers: e, µ, and τ Good z resolution, ~1 mm, for pileup suppression Veto objects from different pp interactions Good efficiency for tracking in jets Need to be able to identify jet vertex for pileup suppression Efficiency down to p T =2 GeV Low momentum tracks used for track based isolation Good track purity Important e.g. for τ identification More powerfull, but complicated, L1 algorithms such as particle flow are under study All this delivered in ~5 µs to meet overall L1 trigger latency requirements

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:9 Tracklet an all FPGA Approach The Tracklet approach described in this talk is an all FPGA solution that takes advantage of the rapid development of commercial FPGAs Algorithm implements a conventional road search technique Seeds are formed from pairs of stubs from neighboring layers (or disk) Projected to other layers (disks) to add stubs Combinatorics addressed by aggressive partitioning Implementation massively parallel The implementation makes use of the high precision of the hits in the silicon detector to project the seeds to neighboring layers Calculations are performed in the FPGAs using the powerful Digitial Signal Processing (DSP) units Data routing capability important for the hardware implementation of the algorithm

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:10 Tracklet Based Track Finding Form track seeds, tracklets, from pairs of stubs in neighboring layers y x

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:11 Tracklet Based Track Finding Form track seeds, tracklets, from pairs of stubs in neighboring layers Match stubs on road defined by tracklet and IP constraint y Matching resolutions for tracklets in L1+L2 projected to L3 φ resolution 0.5 mm window x rδφ residual (cm) Residual from matching used in final trackfit

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:12 Tracklet Based Track Finding Form track seeds, tracklets, from pairs of stubs in neighboring layers Match stubs on road defined by tracklet Fit the hits matched to the tracklet using a linearized χ 2 fit Tracklet parameters good linear fit works very will y x

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:13 Tracklet Based Track Finding Form track seeds, tracklets, from pairs of stubs in neighboring layers Match stubs on road defined by tracklet Fit the hits matched to the tracklet using a linearized χ 2 fit Seeding is done in parallel in different layers Duplicate tracks are removed if they share 2 or more stubs y x

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:14 Implementation Approach Make use of multiplexing: Spatial: Divide detector into 28 φ sectors. Seeding (tracklet finding) local within sector - ~10% stub duplication Time: Factor of 6 time multiplexing. Each sector processor receives a new event every 150 ns. From simulations (at 200 PU) we have per sector: ~500 stubs Form ~70 tracklets Find ~15 track (including duplicates) What computational resources are need to implement this tracking in an FPGA?

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:15 Resource Estimates Approximate number of DSP operations required for the different steps in the algorithm Task # objects DSP operations per object Tracklet Parameters 70 20 1400 Tracklet Projections (4 per tracklet) 70 4 10 2800 Matching 200 4 800 Track fit 15 20 300 Total 5300 Total DSP operations Assuming a factor of 6 time multiplexing (150 ns between events) and a 240 MHz project clock each DSP in the FPGA can perform 36 operations. A vertex 7 (690T) has about 3600 DSPs and can perform ~130,000 operations per event We need about 4% of the resources The Ultrascale+ FPGAs will have up to 12,000 DSPs and O(1%) would be needed for the L1 tracking However: Challenge is to handle combinatorics and tails

Baseline Implementation We have established a 'baseline project for demonstration: Factor of 6 time MUX: Each board receives a new event every 150 ns (LHC bunch crossing frequency is 40 MHz) 28 sectors - Tracking performed for p T >2 GeV Tracklets formed locally in sector Projections sent to neighboring sectors A sector covers the full η range Seeding (tracklet finding) done in multiple layer (disk) combinations: Fit tracks with 4 or more stubs I.e. tracklet + at least two matched stubs With tracklet seed a linearized χ 2 fit works very well Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:16

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:17 Combinatorics The obvious challenge for implementing a tracking algorithm with a fixed latency is how to handle the combinatorics. The tracklet approach described here has two key steps were combinatorics is an issue: Forming tracklets from pairs of stubs Matching tracklets projected to other layers (or disks) to stubs Both cases are addressed in a similar manner: The detector is divided into smaller regions and the combinatorics problem is solved by massive parallelism To illustrate this we consider the tracklet finding in some detail Remember that we work here in one sector, so 1/28 of the full detector _ The plots of the occupancy is taken from a sampler of tt with 200 minbias events superimposed

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:18 Tracklet Formation Average number of stubs in a sector: Layer 1: ~80 stubs Layer 2: ~50 stubs Total of ~4000 possible pairs Most of these don't satisfy: z 0 <15 cm p T >2 GeV Divide each layer into sub regions ('Virtual Modules' VMs) Layer 1: 8 z 3 φ = 24 VMs Layer 2: 8 z 4 φ = 32 VMs Total of 24 32=768 pairs of VMs 96 of these can form tracklets Search for tracklets in parallel in these 96 pairs of VMs. Stubs in L1 Sector Stubs in L2 Sector # stubs # stubs Layer 2 Layer 1 Two of the 8 z bins

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:19 Virtual Module Occupancy Stubs per VM in Layer1 Stubs per VM in Layer1 Layer 1: <# stubs>=3.4 Layer 2: <# stubs>=2.0 Linear scale # stubs per VM # stubs per VM Log scale # stubs per VM # stubs per VM

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:20 Tracklet Formation Combinatorics Stub pairs per virtual module combination Tracklets found in L1+L2: z 0 <15 cm p T >2 GeV Average of ~6 stub pairs Average of ~30 tracklets # stubs pairs

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:21 Truncating Tracklet Finding There will be a limit to how many stub pairs we can try in the tracklet engine (TE) The effect of the truncation for finding isolated muons in events with pileup is illustrated below Currently our implementation can comfortably process 36 stub pairs per 150 ns (factor of 6 time multiplexing) Redundancy from multiple seeding layers gives good tracking efficiency Algorithm improvements will essentially eliminate truncation

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:22 Track Fit Performance The integer performance and full floating point calculation gives the same track parameter resolution

Track Finding Efficiency Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:23

Overview of Firmware Project Eight processing steps + two transmission (red) implements the algorithm Stub organization Forming tracklets Anders Ryd Cornell University Projection transmission to neighbors Organize Match tracklet Match projections to transtracklet projections stubs mission FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Track fit Page:24

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:25 Tracklet Demonstrator Stub Source Emulation ~100 Gbit/s* 200 Gbit/s ~100 Gbit/s* * Only need to receive stubs that needed to demonstrate functionality of the central sector Same board used for stub source emulation and the L1Track receiver Sector n+1 80 Gbit/s 80 Gbit/s 80 Gbit/s Sector n Sector n-1 80 Gbit/s Sector board Input: 360 Gbit/s Output: 180 Gbit/s 20 Gbit/s L1Track Receiver (Trigger) Full Sector Demonstration: Covers η <2.4 Uses factor 4 time MUX p T >2 GeV Allow measurement of latency from stub input to tracks available

Latency Model Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:26

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:27 Latency Model Measured Latency: 3333 ns Agrees with model to 0.38% or three 240 MHz clocks

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:28 FPGA Resources From the resource usage in the half-sector project we can extrapolate to the full sector The current Virtex-7 FPGAs don t have sufficient resources Next generation Ultrascale+ FPGAs provide the needed resources The cheaper Kintex FPGAs are also an interesting possibility Half-Sector Resource Usage Extrapolate

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:29 Future Improvements We are exploring a better detector segmentation (Virtual Module) configuration Using narrow VMs in f that covers the full length of the detector allows a reduction in the combinatorics since jets are not contained in VM. Dynamic load balancing Average resource utilization is low - resources assigned to handle tails. Dynamic load balance at the most parallel stages (Tracklet Engines and Match Engines) allow better resource utilization. Improved duplicate removal using χ 2 information Select best track and optimize resource usage in duplicate removal Use of HLS (High-Level Synthesis) instead of Verilog Simplify the development

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:30 Summary/Outlook The tracklet approach a road-search technique implemented on FPGAs have been demonstrated The approach meets the CMS requirements for latency and track performance The implementation provides robustness by seeding in parallel in different seeding layers Further improvements are under study to implement load balancing to reduce truncation in tails and reduce FPGA resources needed New hardware is being developed for the next generation boards for the track trigger implementation Will use Xilinx Ultrascale+ FPGAs and optical links at speeds up to 25 Gbits/s CMS is forming a common L1 tracking effort to define the reference track trigger project based on an all FPGA solution

Backup Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:31

Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:32 Development We have a C++ simulation that reproduces bitwise the firmware. Useful for algorithm development and physics performance studies. Implementation in Verilog (compared with C++) FPGA firmware simulation Execution on test stand _

Seeding Redundancy Anders Ryd Cornell University FPGA Based L1 Tracking at HL-LHC AWLC June 26-30, 2017 Page:33