PERFORMING TRACK RECONSTRUCTION AT THE ALICE TPC USING A FAST HOUGH TRANSFORM METHOD

Similar documents
PoS(High-pT physics09)036

Fast pattern recognition with the ATLAS L1Track trigger for the HL-LHC

A New Segment Building Algorithm for the Cathode Strip Chambers in the CMS Experiment

ALICE tracking system

PoS(Baldin ISHEPP XXII)134

Tracking and compression techniques

Real-time Analysis with the ALICE High Level Trigger.

The Compact Muon Solenoid Experiment. Conference Report. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

High Level Trigger System for the LHC ALICE Experiment

Performance of the MRPC based Time Of Flight detector of ALICE at LHC

PATHFINDER A track finding package based on Hough transformation

Muon Reconstruction and Identification in CMS

Tracking and flavour tagging selection in the ATLAS High Level Trigger

Klaus Dehmelt EIC Detector R&D Weekly Meeting November 28, 2011 GEM SIMULATION FRAMEWORK

CMS Conference Report

arxiv: v1 [physics.ins-det] 11 Jul 2015

ATLAS, CMS and LHCb Trigger systems for flavour physics

PoS(EPS-HEP2017)492. Performance and recent developments of the real-time track reconstruction and alignment of the LHCb detector.

Charged Particle Reconstruction in HIC Detectors

PoS(TIPP2014)204. Tracking at High Level Trigger in CMS. Mia TOSI Universitá degli Studi di Padova e INFN (IT)

First results from the LHCb Vertex Locator

Performance of Tracking, b-tagging and Jet/MET reconstruction at the CMS High Level Trigger

First LHCb measurement with data from the LHC Run 2

THE ATLAS INNER DETECTOR OPERATION, DATA QUALITY AND TRACKING PERFORMANCE.

Detector Control LHC

ATLAS ITk Layout Design and Optimisation

SoLID GEM Detectors in US

The ATLAS Conditions Database Model for the Muon Spectrometer

for the DESY/ ECFA study detector

b-jet identification at High Level Trigger in CMS

Performance of the ATLAS Inner Detector at the LHC

Event reconstruction in STAR

SoLID GEM Detectors in US

The Track-Finding Processor for the Level-1 Trigger of the CMS Endcap Muon System

Control slice prototypes for the ALICE TPC detector

arxiv:hep-ph/ v1 11 Mar 2002

The ALICE High Level Trigger

Implementing Online Calibration Feed Back Loops in the Alice High Level Trigger

Open charm measurements with new Vertex Detector at NA61/SHINE experiment at CERN SPS

arxiv: v1 [physics.ins-det] 26 Dec 2017

ATLAS NOTE. December 4, ATLAS offline reconstruction timing improvements for run-2. The ATLAS Collaboration. Abstract

Charged Particle Tracking at Cornell: Gas Detectors and Event Reconstruction

Stefania Beolè (Università di Torino e INFN) for the ALICE Collaboration. TIPP Chicago, June 9-14

Atlantis: Visualization Tool in Particle Physics

3D-Triplet Tracking for LHC and Future High Rate Experiments

Simulating the RF Shield for the VELO Upgrade

Adding timing to the VELO

CMS reconstruction improvements for the tracking in large pile-up events

Determination of the aperture of the LHCb VELO RF foil

ALICE Run3/Run4 Computing Model simulation software

Inside-out tracking at CDF

TORCH: A large-area detector for precision time-of-flight measurements at LHCb

TPC tracking and particle identification in high-density environment

Status of the TORCH time-of-flight detector

Charged Particle Tracking at Cornell: Gas Detectors and Event Reconstruction

Direct photon measurements in ALICE. Alexis Mas for the ALICE collaboration

1. Introduction. Outline

Beam test measurements of the Belle II vertex detector modules

Tracking and Vertex reconstruction at LHCb for Run II

ALICE is one of the four large-scale experiments at the

arxiv: v1 [physics.ins-det] 16 Oct 2017

ATLAS Tracking Detector Upgrade studies using the Fast Simulation Engine

PoS(IHEP-LHC-2011)002

Alignment of the ATLAS Inner Detector tracking system

Pattern recognition systems Lab 3 Hough Transform for line detection

8.882 LHC Physics. Track Reconstruction and Fitting. [Lecture 8, March 2, 2009] Experimental Methods and Measurements

CMS Alignement and Calibration workflows: lesson learned and future plans

Robustness Studies of the CMS Tracker for the LHC Upgrade Phase I

Lawrence Berkeley National Laboratory Lawrence Berkeley National Laboratory

CRU Readout Simulation for the Upgrade of the ALICE Time Projection Chamber at CERN

1. INTRODUCTION 2. MUON RECONSTRUCTION IN ATLAS. A. Formica DAPNIA/SEDI, CEA/Saclay, Gif-sur-Yvette CEDEX, France

The CMS alignment challenge

arxiv: v1 [hep-ex] 7 Jul 2011

PoS(EPS-HEP2017)523. The CMS trigger in Run 2. Mia Tosi CERN

LHC-B. 60 silicon vertex detector elements. (strips not to scale) [cm] [cm] = 1265 strips

Upgraded Swimmer for Computationally Efficient Particle Tracking for Jefferson Lab s CLAS12 Spectrometer

arxiv:physics/ v1 [physics.ins-det] 18 Dec 1998

Performance of the GlueX Detector Systems

arxiv: v1 [physics.ins-det] 26 Dec 2017

Track reconstruction with the CMS tracking detector

SLAC PUB 8389 Mar 2000 TRACKING IN FULL MONTE CARLO DETECTOR SIMULATIONS OF 500 GeV e + e COLLISIONS a M.T. RONAN Lawrence Berkeley National Laborator

CMS FPGA Based Tracklet Approach for L1 Track Finding

Overview. About CERN 2 / 11

Full Offline Reconstruction in Real Time with the LHCb Detector

Overview of the American Detector Models

TEC single wire resolution

The full detector simulation for the ATLAS experiment: status and outlook

Benchmarking message queue libraries and network technologies to transport large data volume in

The GAP project: GPU applications for High Level Trigger and Medical Imaging

Alignment of the CMS Silicon Tracker

HLT Hadronic L0 Confirmation Matching VeLo tracks to L0 HCAL objects

Update of the BESIII Event Display System

ONLINE MONITORING SYSTEM FOR THE EXPERIMENT

Locating the neutrino interaction vertex with the help of electronic detectors in the OPERA experiment

SoLID GEM Detectors in US

Track reconstruction for the Mu3e experiment based on a novel Multiple Scattering fit Alexandr Kozlinskiy (Mainz, KPH) for the Mu3e collaboration

V. Karimäki, T. Lampén, F.-P. Schilling, The HIP algorithm for Track Based Alignment and its Application to the CMS Pixel Detector, CMS Note

The LHCb upgrade. Outline: Present LHCb detector and trigger LHCb upgrade main drivers Overview of the sub-detector modifications Conclusions

π ± Charge Exchange Cross Section on Liquid Argon

Alignment of the CMS silicon tracker using Millepede II

Transcription:

Ó³ Ÿ. 2016.. 13, º 5(203).. 1020Ä1027 Š Œ œ ƒˆˆ ˆ ˆŠ PERFORMING TRACK RECONSTRUCTION AT THE ALICE TPC USING A FAST HOUGH TRANSFORM METHOD C. S. Kouzinopoulos 1, P. Hristov 2 CERN, European Laboratory for Particle Physics, Geneva The Hough Transform algorithm is a popular image analysis method that is widely used to perform global pattern recognition in images through the identiˇcation of local patterns in a suitably chosen parameter space. The algorithm can also be used to perform track reconstruction and to estimate the trajectory of individual particles passed through the active elements of a detector volume. This paper presents a fast reconstruction method for the Time Projection Chamber (TPC) of the ALICE experiment at LHC. The method, that combines a linear Hough Transform algorithm with a fast ˇlling of the Hough Transform parameter space, is developed within AliceO 2, the new computing framework of ALICE for Run3. PACS: 29.85.-c; 07.05.Pj INTRODUCTION ALICE (A Large Ion Collider Experiment) is a general-purpose, heavy-ion detector at the CERN Large Hadron Collider (LHC) which focuses on QCD, the strong-interaction sector of the Standard Model. It is designed to address the physics of strongly interacting matter and the quarkägluon plasma at extreme values of energy density and temperature in nucleusänucleus collisions [1]. A signiˇcant increase of the LHC luminosity for heavy ions is expected in Run3 after the Long Shutdown 2 (LS2) period, leading to PbÄPb collision rates of about 50 khz and an integrated luminosity L int =10nb 1. The main physics topics addressed by the ALICE upgrade are precise measurements of heavy avour hadrons, low-momentum quarkonia, and low mass di-leptons [2]. After the upgrade, continuous read-out will be implemented instead of the classic trigger system for the main detectors, including ITS and TPC, to deal with event pile-up and avoid trigger-generated dead time. The integration of online and ofine data processing requires a common onlineäofine software framework, AliceO 2 and a common computing facility, the O 2 computing infrastructure, dedicated to both data collection and processing. AliceO 2, that is currently under development for Run3, is built on top of ALFA, the new ALICEÄFAIR concurrency framework for high-quality parallel data processing and reconstruction on heterogeneous computing systems [3]. 1 E-mail: charalampos.kouzinopoulos@cern.ch 2 E-mail: peter.hristov@cern.ch

Performing Track Reconstruction at the ALICE TPC Using a Fast Hough Transform Method 1021 Since approximately 92.5% of the data rate for PbÄPb collisions at 50 khz will be generated by the ALICE TPC, one of the main processing challenges of the O 2 computing infrastructure will be to achieve an efˇcient TPC track reconstruction. This paper presents one possible algorithm to perform track reconstruction for the TPC, a fast Hough Transform method. The method is based on prior work [4] where it showed good and stable track reconstruction performance over the entire range of expected event multiplicities in central PbÄPb collisions; it is currently being implemented as part of AliceO 2. 1. ALICE TPC DETECTOR FOR Run3 TPC is the main tracking detector in the central barrel of ALICE. Together with the ITS, TRD and TOF detectors, it provides charged particle momentum measurement, particle identiˇcation, and vertex determination with sufˇcient momentum resolution, two-track separation and de/dx resolution for studies of hadronic and leptonic signals in the region p t < 10 GeV/c and pseudorapidities η < 0.9 [5, 6]. TPC is the largest detector of this type, with an overall active volume of about 90 m 3. It employs a cylindrical gas volume with an inner radius of approximately 84 cm, an outer radius of about 250 cm, and an overall length of about 5 m in the beam axis. It is divided into two half volumes of equal size, separated by a central electrode at high voltage to generate the drift ˇeld. The readout plane is azimuthal segmented; the chambers are mounted into 18 trapezoidal sectors at each end plate, each sector covering 20 in azimuth. The readout at each sector is divided into two chambers, the Inner Readout Chamber (IROC) and Outer Readout Chamber (OROC), resulting in a radial segmentation of the readout plane, as can be seen in Fig. 1. On each side of the central electrode, the primary ionization electrons drift towards the end plate, where their signal is measured by approximately 550,000 readout pads (5504 pads per IROC and 9856 pads per OROC on each of the 36 TPC sectors). Fig. 1. A schematic view of the current layout of the TPC sectors and readout chambers

1022 Kouzinopoulos C. S., Hristov P. Currently, the signal measurement is performed via conventional Multi-Wire Proportional Chambers (MWPCs) with cathode pad readout, while the readout chambers operate with an active bipolar gating grid. This schema though has a principal rate limitation to about 3.5 khz and thus is not suitable for Run3. For this reason, it will be replaced by a multistage Gas Electron Multiplier (GEM) detectors system as charge ampliˇer. GEMs operate reliably in high-rate applications and provide intrinsic ion blocking capabilities that avoid massive charge accumulation in the drift volume from back-drifting ions and prevent excessive space-charge distortions, enabling therefore the TPC to operate in a continuous, un-gated readout mode at collision rates of 50 khz [6]. The upgraded readout chambers will employ stacks of four GEM foils for gas ampliˇcation and anode pad readout, with three GEM foils per layer in the OROCs and a single large GEM foil per layer for the IROCs. The signals coming from the readout pads on the readout plane are passed to the frontend cards (FEC), digitized and processed by a DSP chip and are subsequently transferred to a Common Readout Unit (CRU). From the CRU, the individual data fragments are ordered by their geometrical position on the pad rows and are forwarded to the O 2 computing infrastructure on a pad-by-pad basis. 2. THE O 2 COMPUTING INFRASTRUCTURE The O 2 computing infrastructure will support both online synchronous data reduction and asynchronous and iterative data processing [2]. An overview of the infrastructure, from the generation of the raw data on the detector level till the ˇnal data storage on the Storage Servers, is depicted in Fig. 2. The computing infrastructure consists of two different categories of computing nodes, the First Level Processors (FLPs) and the Event Processing Nodes (EPNs). Approximately 250 FLPs constitute the front-end of the O 2 infrastructure; their role will be to read out the raw data samples as produced by the ALICE detectors over the front-end links, to perform low-level standalone processing (clusterization, masking, calibration), and to compress the data by up to a factor of 2.5 by converting raw to clusterized data and applying lossless compression algorithms. The resulting data will subsequently be merged into data packages called Sub-Time Frames (STFs), that are temporary stored locally before being transmitted to the EPNs over a switching network. Each FLP will collect data from a speciˇc subdetector only, the produced STFs therefore will contain only partial information from the physics event that was buffered in the subdetector electronics. Fig. 2. High-level overview of the data aggregation, from the fragments read-out from the detectors to the full Time Frames stored on the Storage Servers [2]

Performing Track Reconstruction at the ALICE TPC Using a Fast Hough Transform Method 1023 Around 1500 EPNs will aggregate from the FLPs all the STFs corresponding to the same event in order to form complete Time Frames. The EPNs will also perform additional processing on the data, including track and event-level reconstruction, as well as further compression of the data volume up to an estimated factor of 8 with lossy data reduction by discarding noninteresting clusters. To reduce data trafˇc contention when several FLPs are sending to the same EPN, a trafˇc shaping mechanism will be used, which will buffer the STFs at the FLPs and use a delayed transmission schema to the EPNs [7]. The data dispatch and building components needed by both FLPs and EPNs are implemented using ALFA, as a part of the AliceO 2 computing framework. ALFA is a modular set of packages including a data transport layer and the capability to coordinate multiple data processing components. Processing tasks are organized into topologies, consisting of independent processes called devices that communicate on either inter-process or distributed fashion using asynchronous message queues. The devices are grouped into three categories: sources Å devices without inputs that feed the processing pipeline with data; message-based processors Å devices that operate on messages without interpreting their content; content-based processors Å devices that access and process the data, employing user algorithms. One such task that will be part of the AliceO 2 computing framework as a content-based processor device is a track reconstruction. 3. USING THE HOUGH TRANSFORM METHOD FOR TRACK RECONSTRUCTION The task of track reconstruction is to reconstruct the path which a charged particle took through the active components of a detector by combining the hits as registered by the readout pads to tracks. From the parameters describing the tracks, the transverse momentum can be obtained. By extrapolating tracks to the inner part of the detector, secondary vertices can be reconstructed; by extrapolating tracks to the outer part of the detector, the entering point into the calorimeter can be calculated so that measurements in the calorimeter can be matched to the particle which produced the track [8]. One method that can be used for track reconstruction on the TPC is Hough Transform [9] (HT), an algorithm commonly used in computer vision applications to determine all the possible lines in an image based on a given set of image pixels. When used for track reconstruction, HT performs global track recognition by the identiˇcation of local patterns in a parameter space and mapping image space points to parameter space curves. Since the solenoid magnetic ˇeld of ALICE is almost constant, the trajectories of the particles originating from a primary vertex on the beam axis are helices, if multiple scattering and energy losses in the detector volume are neglected. The particles follow a circular trajectory in the transverse plane; reconstructing a track is then equivalent to ˇnding the parameters that describe their trajectory. Different methods can be used for parameterization, but an efˇcient way is to transform the particle trajectories to straight lines by using the conformal mapping [10] transformation. This leads to uncorrelated variables that are closely related to the geometry of the TPC and to simple mathematical calculations. As described in [4], two circles can be deˇned inside a TPC sector, the ˇrst lying near the middle edge of the sector, while the second one lying near the outer edge. Then, as shown

1024 Kouzinopoulos C. S., Hristov P. Fig. 3. a) A schematic view of a TPC sector on the Cartesian cluster system. The dashed line represents a track crossing the sector. The two deˇned circles are also visible. b) The same view in the conformal mapping space. The track and the two circles are visible as straight lines [4] in Fig. 3, a, the track crossing the TPC sector can be represented by two points (x 1, y 1 and x 2, y 2 ) on these circles. The circles are given by the constants α 1 and α 2 : α 1 = x 1 r 2 1 = x 1 x 2 1 + α 2 = x 2 y2 1 r2 2 The track can be parameterized using the variables β 1 and β 2 : β 1 = y 1 r 2 1 = x 2 x 2 2 +. (1) y2 2 = y 1 x 2 1 + β 2 = y 2 y2 1 r2 2 = y 2 x 2 2 +. (2) y2 2 Then, for each space point along the track trajectory it is true that y x 2 + y 2 β 1 x x 2 + y 2 α = β 2 β 1. (3) α 1 2 α 1

Performing Track Reconstruction at the ALICE TPC Using a Fast Hough Transform Method 1025 If A and B are deˇned as A = y x 2 + y 2 (α x 1 α 2 ) x x x 2 + y 2 α, B = 2 + y 2 α 1 x 2 x 2 + y 2 α, (4) 2 then Eq. (3) transforms into a straight line in the conformal mapping parameter space β 1 = A + Bβ 2. (5) A track contained fully inside the TPC sector is represented in the parameter space by a series of straight lines covering the whole azimuthal angle around the track parameters position and thus producing a round peak. This allows one to achieve good track reconstruction efˇciency, to minimize the amount of fake track candidates, and to extract uncorrelated track candidate parameters [4]. Since the offset A and the slope B of the line depend only on the Cartesian coordinates (x, y) of the digit and the constants α 1 and α 2, they can be precalculated for each of the TPC pads and stored to look-up tables to speed-up the execution of the method. 4. OVERVIEW OF TRACK RECONSTRUCTION ON THE O 2 INFRASTRUCTURE The HT method presented in this paper will be executed on the O 2 computing infrastructure to perform track reconstruction using a series of individual processing steps; a high-level overview of the procedure can be seen in Fig. 4. The HT device will receive as an input clusterized data. For this reason, it will be executed on the FLPs after the TPC raw data (pad, arrival time and signal amplitude information) are converted to clusters by a one-dimensional FPGA algorithm. On the ˇrst run, HT will undergo a preprocessing phase to allocate and setup all the necessary memory structures of the method, as well as to calculate the A and B parameters and to store them to look-up tables, as described in Sec. 3 above. The track reconstruction will be performed on the FLPs on a per TPC sector basis. The FLPs will not have information for a full TPC sector, since the STFs will contain only partial Fig. 4. Overview of the main steps in the HT based reconstruction for the TPC

1026 Kouzinopoulos C. S., Hristov P. information from the physics event but instead each FLP will have access to only parts of the TPC sector data. For this reason, a partial reconstruction will be performed at this stage with a ˇnal merging being performed at a later stage on the EPNs. Due to the nature of the continuous readout, the FLPs will not have information on the event timestamps. The event extraction phase, where tracks and secondary vertices attribute to primary vertices and their association to trigger timestamps from the FIT detector, will be executed as the ˇnal step of the asynchronous Time Frame reconstruction on the EPNs. Instead, a time window will be used on the FLPs to separate clusters from different events based on arbitrary timestamps. The candidate tracks produced by the HT device at this stage will have to pass through a veriˇcation phase on the EPNs. The HT method will then perform the conformal mapping transformation on the cluster data; each cluster with a charge greater than a speciˇed threshold will be transformed to a line in the parameter space. Two different parameter space storage methods will have to be considered for the implementation of the device: A series of two-dimensional histograms. The histogram bin size and limits must be chosen to obtain the best compromise between the efˇciency of the track reconstruction, the resolution and the time performance. A key-value pair map container, a solution used in [8]. The use of a map container can reduce the necessary memory space needed for the algorithm but can be slower in terms of CPU utilization. The value of all the histogram bins/map entries that are crossed by the line are increased by a certain weight. In the original implementation of the method [4], a weight corresponding to the count of the occupied TPC pad rows along the track was preferred over a constant weight or a weight equal to the charge count of the point; all three methods will be reevaluated for the current implementation. The bins with the highest weight correspond to the point where all parameters intersect. Since there can exist bins with the same weight, the adjacent bins to the one with the local maxima have also to be inspected. The bins found in this way give an approximation of the track parameters. The resulting data structures will be passed to a peak ˇnder on the FLPs to locate all neighbour ˇlled bins/entries and form track candidate peaks. The resulting tracks are candidates and will be veriˇed or rejected on the EPNs, after the Time Frames are aggregated. The partial peaks/tracks of the individual peak ˇnders will be transmitted to the EPNs as a part of the STFs. Each EPN will have full information on a speciˇc event from all TPC sectors and in addition the time reference from the ITS and TRD detectors. The output of the partial HT executions on the FLPs will be merged while processing aggregated Time Frames. Track candidates situated too close to the entrance or exit of the data chunk of a given FLP will have to be merged. Track candidates situated too close to the entrance or exit of a given TPC sector will also have to be merged. Track candidates formed from clusters of different events will have to be carefully reevaluated and possibly rejected as fakes. The track parameters obtained with this method are only an estimate of the original ones, due to the binning of the Hough space. To improve the estimation of the track parameters, a simple track ˇtting algorithm will be implemented at the EPNs. As the last step, the tracks from all TPC sectors are collected on the EPNs and stored in an event-summary data-block.

Performing Track Reconstruction at the ALICE TPC Using a Fast Hough Transform Method 1027 CONCLUSIONS This paper presented the HT method that can be used to perform track reconstruction for the ALICE TPC for Run3. The method is based on prior work for the ALICE experiment, showing good and stable track reconstruction performance. HT is currently developed as a content-based processor device, as part of the AliceO 2 software framework and will be executed on the O 2 computing infrastructure. REFERENCES 1. The ALICE Collab. The ALICE Experiment at the CERN LHC // J. Instr. 2008. V. 3, No. 08. P. S08002. 2. Buncic P., Krzewicki M., Vande Vyvre P. Technical Design Report for the Upgrade of the OnlineÄ Ofine Computing System // Tech. Rep. CERN-LHCC-2015-006. ALICE-TDR-019. Geneva: CERN, 2015. 3. Al-Turany M. et al. The New ALICEÄFAIR Software Framework // J. Phys.: Conf. Ser. 2015. V. 664, No. 7. 4. Cheshkov C. Fast Hough Transform Track Reconstruction for the ALICE TPC // Nucl. Instr. Meth. A. 2005. P. 35Ä39. ALICE-INT-2005-024. 5. Dellacasa G. et al. (ALICE Collab.). ALICE: Technical Design Report of the Time Projection Chamber. CERN / LHCC 2000-001 ALICE TDR 7. 2000. 6. Upgrade of the ALICE Time Projection Chamber. Tech. Rep. CERN-LHCC-2013-020. ALICE-TDR- 016. Geneva: CERN, 2013. 7. Rybalchenko A. et al. Efˇcient Time Frame Building for Online Data Reconstruction in ALICE Experiment // J. Phys.: Conf. Ser. 2015. 8. Heinze I. Development of a Hough Transformation Track Finder for Time Projection Chambers. Ph. D. Dissertation. Hamburg Univ., 2013. 9. Duda R. O., Hart P. E. Use of the Hough Transformation to Detect Lines and Curves in Pictures // Commun. ACM. 1972. V. 15, No. 1. P. 11Ä15. 10. Yepes P. A Fast Track Pattern Recognition // Nucl. Instr. Meth. A. 1996. V. 380, No. 3. P. 582Ä585.