A LVL2 Zero Suppression Algorithm for TRT Data

Similar documents
ATLAS NOTE. December 4, ATLAS offline reconstruction timing improvements for run-2. The ATLAS Collaboration. Abstract

Performance of the ATLAS Inner Detector at the LHC

Modules and Front-End Electronics Developments for the ATLAS ITk Strips Upgrade

ATLAS TDAQ RoI Builder and the Level 2 Supervisor system

CMS FPGA Based Tracklet Approach for L1 Track Finding

Fast pattern recognition with the ATLAS L1Track trigger for the HL-LHC

A New Segment Building Algorithm for the Cathode Strip Chambers in the CMS Experiment

arxiv:hep-ph/ v1 11 Mar 2002

Disentangling P ANDA s time-based data stream

ATLANTIS - a modular, hybrid FPGA/CPU processor for the ATLAS. University of Mannheim, B6, 26, Mannheim, Germany

Scenarios for a ROB system built with SHARC processors. Jos Vermeulen, 7 September 1999 Paper model results updated on 7 October 1999

First Operational Experience from the LHCb Silicon Tracker

THE ATLAS INNER DETECTOR OPERATION, DATA QUALITY AND TRACKING PERFORMANCE.

DRAFT. Options for the ROB Complex

Performance of the GlueX Detector Systems

The Compact Muon Solenoid Experiment. CMS Note. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

Atlantis MultiRob Scenario

The MROD. The MDT Precision Chambers ROD. Adriaan König University of Nijmegen. 5 October nd ATLAS ROD Workshop 1

ATLAS ITk Layout Design and Optimisation

Updated impact parameter resolutions of the ATLAS Inner Detector

ATLAS PILE-UP AND OVERLAY SIMULATION

Straw Detectors for the Large Hadron Collider. Dirk Wiedner

The GAP project: GPU applications for High Level Trigger and Medical Imaging

The raw event format in the ATLAS Trigger & DAQ

S-LINK: A Prototype of the ATLAS Read-out Link

Evaluation of network performance for triggering using a large switch

The MROD. The Read Out Driver for the ATLAS MDT Muon Precision Chambers

The ATLAS Level-1 Muon to Central Trigger Processor Interface

Real-time Analysis with the ALICE High Level Trigger.

PoS(High-pT physics09)036

Tracking and flavour tagging selection in the ATLAS High Level Trigger

Electron and Photon Reconstruction and Identification with the ATLAS Detector

Endcap Modules for the ATLAS SemiConductor Tracker

ATLAS, CMS and LHCb Trigger systems for flavour physics

PoS(ACAT)049. Alignment of the ATLAS Inner Detector. Roland Haertel Max-Planck-Institut für Physik, Munich, Germany

The Phase-2 ATLAS ITk Pixel Upgrade

First LHCb measurement with data from the LHC Run 2

Beam test measurements of the Belle II vertex detector modules

Adding timing to the VELO

PU : General View. From FEB S2P DSP FIFO SLINK

LHC Detector Upgrades

The FTK to Level-2 Interface Card (FLIC)

ATLAS Simulation Computing Performance and Pile-Up Simulation in ATLAS

Simulating the RF Shield for the VELO Upgrade

The performance of the ATLAS Inner Detector Trigger Algorithms in pp collisions at the LHC

Alignment of the ATLAS Inner Detector tracking system

Scintillator-strip Plane Electronics

Alignment of the CMS Silicon Tracker

Velo readout board RB3. Common L1 board (ROB)

arxiv:physics/ v1 [physics.ins-det] 18 Dec 1998

The First Integration Test of the ATLAS End-cap Muon Level 1 Trigger System

Physics CMS Muon High Level Trigger: Level 3 reconstruction algorithm development and optimization

Event reconstruction in STAR

I/O Choices for the ATLAS. Insertable B Layer (IBL) Abstract. Contact Person: A. Grillo

The Track-Finding Processor for the Level-1 Trigger of the CMS Endcap Muon System

CMS Conference Report

TORCH: A large-area detector for precision time-of-flight measurements at LHCb

Improving Packet Processing Performance of a Memory- Bounded Application

Results of a Sliced System Test for the ATLAS End-cap Muon Level-1 Trigger

Time of CDF (II)

Evaluation of the computing resources required for a Nordic research exploitation of the LHC

The First Integration Test of the ATLAS End-Cap Muon Level 1 Trigger System

The Compact Muon Solenoid Experiment. Conference Report. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

The GTPC Package: Tracking and Analysis Software for GEM TPCs

First results from the LHCb Vertex Locator

LHC-B. 60 silicon vertex detector elements. (strips not to scale) [cm] [cm] = 1265 strips

Muon Reconstruction and Identification in CMS

The ATLAS Trigger System: Past, Present and Future

ROB IN Performance Measurements

The full detector simulation for the ATLAS experiment: status and outlook

The LHCb upgrade. Outline: Present LHCb detector and trigger LHCb upgrade main drivers Overview of the sub-detector modifications Conclusions

Clustering and Reclustering HEP Data in Object Databases

Update on PRad GEMs, Readout Electronics & DAQ

Simulation of Beam-Beam Background at CLIC. André Sailer (CERN-PH-LCD, HU Berlin) LCWS2010: BDS+MDI Joint Session 29 March, 2010, Beijing

Construction of the Phase I upgrade of the CMS pixel detector

Level-1 Data Driver Card of the ATLAS New Small Wheel Upgrade Compatible with the Phase II 1 MHz Readout

ROB-IN Functional demonstrator of the ATLAS Trigger / DAQ Read-Out Buffer O.Gachelin, M.Huet, P.Le Dû, M.Mur C.E.A.

PoS(IHEP-LHC-2011)002

b-jet identification at High Level Trigger in CMS

Alignment of the CMS silicon tracker using Millepede II

Track reconstruction with the CMS tracking detector

Determination of the aperture of the LHCb VELO RF foil

The Compact Muon Solenoid Experiment. CMS Note. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

The LHCb Upgrade. LHCC open session 17 February Large Hadron Collider Physics (LHCP) Conference New York, 2-7 June 2014

Tracking and Vertexing performance in CMS

Direct photon measurements in ALICE. Alexis Mas for the ALICE collaboration

ROBIN Functional demonstrator of the ATLAS Trigger / DAQ Read-Out Buffer O.Gachelin, M.Huet, P.Le Dû, M.Mur C.E.A. Saclay - DAPNIA

SoLID GEM Detectors in US

1. INTRODUCTION 2. MUON RECONSTRUCTION IN ATLAS. A. Formica DAPNIA/SEDI, CEA/Saclay, Gif-sur-Yvette CEDEX, France

PoS(EPS-HEP2017)523. The CMS trigger in Run 2. Mia Tosi CERN

Level-1 Regional Calorimeter Trigger System for CMS

Track-Finder Test Results and VME Backplane R&D. D.Acosta University of Florida

PoS(TIPP2014)204. Tracking at High Level Trigger in CMS. Mia TOSI Universitá degli Studi di Padova e INFN (IT)

Trigger and Data Acquisition: an ATLAS case study

CSC Trigger Motherboard

Robustness Studies of the CMS Tracker for the LHC Upgrade Phase I

Optical Data Interface ODI-2 Transport Layer Preliminary Specification. Revision Date

PoS(EPS-HEP2017)492. Performance and recent developments of the real-time track reconstruction and alignment of the LHCb detector.

The CMS Event Builder

CMS Alignement and Calibration workflows: lesson learned and future plans

Transcription:

A LVL2 Zero Suppression Algorithm for TRT Data R. Scholte,R.Slopsema,B.vanEijk, N. Ellis, J. Vermeulen May 5, 22 Abstract In the ATLAS experiment B-physics studies will be conducted at low and intermediate luminosity (resp. 33 cm 2 s and 2 33 cm 2 s ). When a µ with p T > 6GeV/c(or in new scenarios a p T > 8 GeV/c) is detected, the first step of the LVL2 trigger is to confirm the muon. For the events that are retained a full scan of the Inner detector is foreseen. In this note a compression algorithm is discussed, that reduces the data volume for the LVL2 network in case of a full TRT scan. Measurements are performed on a prototype ROBIn and a number of Pentium-based machines. Results on data reduction and processing time are presented. ATL-DAQ-23-6 25 July 23 Introduction The dataset used for this study is based on the B-physics trigger selection as it is assumed in []. The following statements, as well as the rest of the note, assume this old trigger selection scenario. In this dataset, B-physics events are selected by searching for events with a muon with p T > 6 GeV/c. This results in an acceptance rate of about 23 khz [2]. The LVL2 trigger starts with a confirmation of the LVL trigger using the muon spectrometer and inner detector. This reduces the rate to about 5 khz if the track is validated in the inner detector. Next the TRT may be searched for charged tracks. The LVL muon can come from one of two B-mesons produced in the proton-proton collision. With an unguided track search charged decay products are looked for. For this purpose the complete TRT detector needs to be read out. The full TRT scan results in a heavy load for the LVL2 network. For every event with one or more validated muons, all TRT ROBs (256) have to be read out. Using the ROD data format [3], the total data volume produced by the scan in the output of the ROBs is typically 256 4 Bytes = kbytes for low luminosity. The resulting 5 MBytes/s required bandwidth can be reduced by a zero suppression algorithm, optimised for a relatively high percentage of non-hit straws, which is the case at low luminosity running. In this note, such a zero suppression algorithm is studied. It has been investigated whether it is feasible to run the preferred algorithm in the ROBs. Since the rate with which the full scan of the TRT is performed is 5 khz, the algorithm itself should take less than.2 ms to run. It has recently been decided that the compression algorithm described in this note will be implemented in the RODs instead of the ROBs. The numbers given in this note for the data reduction should still University of Twente CERN NIKHEF Alternative algorithms in which the complete pixel detector and/or SCT is read out are also of interest.

hold, since the ROB adds only a header and a trailer to the data coming from the ROD, and these are not used by the algorithm. The timing of the algorithm as measured on the Pentium processors, could still serve as a benchmark. 2 Data volumes and format The TRT barrel contains a total of 588 straws distributed over two half barrels [3]. Each half-barrel is built of three concentric layers. For the read out, the barrel is segmented in 32 projective wedges in φ per half barrel. Data from one segment is read out into a single ROD. The 8 wheels of an endcap are divided into 96 readout segments. Each segment receives data from all 8 rings and covers.65 rad in φ. The data going into the RODs has a size of 27 bits per channel. It contains information from three consecutive bunch-crossings (BXs). Per BX, one bit indicates the state of the high threshold discriminator. A signal exceeding the high threshold (5 kev) indicates a transition radiation candidate (electron). The low threshold (.2 kev) is used to identify all minimum ionising particles. The state of the low threshold is given by 8 bits per BX, each bit indicating the state in a 3.25 (=25/8) ns time window. The total input data volume per ROD including headers, trailers and byte alignment is about 5 kbits/event. The data per ROD is transferred to a ROB. At the moment it is envisioned that there will be a : correspondence between ROBs and RODs for the TRT, and thus this is also assumed in this note. Before transmission to the ROB the data is compressed. Only data of validated straws is sent (see table ). A straw is validated when the low threshold is exceeded between 42. ns and 54.5 ns after the bunch crossing. Each particle crossing the straw usually causes ionization near the straw wall. The maximum drift time it takes for an electron to reach the wire is 42. ns. This signal is delayed by the electronics by about 7-8 ns. By applying a narrow window between 42 ns and 54.5 ns, hits caused by particles from neighbouring bunch crossings are suppressed. The window size and position is chosen to allow efficient identification of the hits from the beam-crossing under investigation, while the number of accepted hits from neighbouring beam-crossings is minimized [4]. A straw is declared VALID if any of the last two low threshold bits of the second BX or the first two of the third BX are set. All other straws are defined as NOT VALID. For VALID straws the number of leading edges (transitions from to ) in the low threshold bits of the first two BXs are determined. Depending on the number of leading edges the following data is transferred: <><ttt><ttt><h><t> for two leading edges <><tttt><h><t> for one leading edge <><><L&H><T> for zero leading edges where: <tttt> or <ttt> are a 4-bit or 3-bit 2 encoding location of the transition (from to ) bit. <H> / <L> are one bit ORs of all of the high / low threshold bits, <L&H> is a one bit encoding of all low and high threshold bits, <T> is one bit to indicate that a trailing edge occurred in the third BX. If two leading edges are encountered in the same time-slice (BX) the first will be dropped and only the second edge will be encoded. In Table 5 in [5] it is shown that having a maximum of one edge per time-slice results in a negligible difference compared to the situation where all 2 the first 3 bits encode the first BX interval, the second 3 bits encode the second interval 2

Data type <> <><> <><> <> <> Validation NOT VALID NOT VALID VALID VALID VALID Size (bits) 2 2+2 2+3 2+6 2+8 Table : Different types of data in the ROBs for the TRT. edges within one time-slice are counted. For NOT VALID straws all low threshold bits are checked. If no bit is set two bits are transferred, else four bits are transferred: <> if no bit is set <><><L&H> if at least one bit is set In table an overview of the five possible formats of the TRT data sent to the ROBs is given. The ROB adds to each event fragment a header and trailer. The header is 32 Bytes long, the trailer 2 Bytes. In a recent note [5] a slightly different scenario for data compression in the ROD (the Istanbul scenario) was described. An estimate of the deviation between using the Istanbul scenario and the old scenario was made. Using the mean values of percentage of straws with, and 2 leading edges and the percentage of non-hit straws, it was found that the differences are small: the Istanbul scenario results in about 5% smaller event fragment sizes. In this note only the old scenario has been studied. 3 Zero Suppression Algorithm In [3] an algorithm for data reduction in the ROB is given. In this algorithm, only data for hit straws is sent to the LVL2 trigger. Each data item has an address header indicating the straw position. For both the barrel and endcap modules eleven address header bits are needed to give each straw a specific header (2 <664< 2 ). When this zero suppression algorithm is used, the data sent to the LVL2 trigger consists mainly of address header bits. An alternative algorithm, first suggested by N. Ellis [6], is not to give each hit straw a unique address header, but an offset header indicating the number of straws (offset) to the previous straw that has been hit. This offset header can be much smaller than eleven bits. The optimum offset header size N depends on the percentage of straws hit and their spread among the straws in a readout segment. If the offset between two hit straws is larger than 2 N straws, it is not possible to encode it in N bits. In that case, an extra data item has to be added to specify the offset to the header of the next hit straw. In general, the zero suppression algorithm with offset produces a smaller output volume than the algorithm without offset. The former has the disadvantage that a larger number of processing steps per event is needed: for each non-hit straw the value of the offset counter has to be checked, and extra data items have to be copied to the output. 4 Generation of input Data The TRT digitized detector data sample was created using a signal sample piled up on top of a minimum bias sample. The samples used are the same as in [5]. For low, intermediate and high luminosity, on average 2.3, 4.6 and 23 events, respectively, were piled up on the signal sample. The piled up signal was then digitized by the digitization routine for the TRT [5]. 3

barrel endcap luminosity: low int high low int high %<>: 92.68 85. 6.5 92.65 85.25 62.4 %<><>: 4.73 9.7 9.7 4.75 9.47 8.44 %<><>:.4.5.68.7.24 2.9 %<>: 2.34 4.5 4.35 2.42 4.76 5.3 %<>:.3.49 2.76..29.75 Table 2: Overview of the data used as input for the zero suppression algorithm. The numbers are in percentages. Due to rounding errors they do not exactly sum up to %. The digitization routine already wrote information on leading edges and validness of straws that were hit. The routine was modified to also write x,y,z coordinates of the center of the straw as well as the phi segment (a number from -92) and the layer (a number from -3 for the barrel and -22 for the endcap) in which the straw is located. A class was created which contains as data members the phi segment number, the layer numbers and a list of x,y and z coordinates. By looping over all events and all straws a database was created, consisting of objects (instantiations of the aforementioned class) uniquely defined by a phi segment. Each object contained the x,y,z coordinates of all the straws. Within one phi segment the straws are ordered layer-by-layer. Subsequently the digitized data sample is now read in per event per ROD. A ROD spans 6 phi segments in the barrel and 2 in the endcap. The straws in the data sample are compared with the straws in the database. If a straw is missing in the data sample (it is not hit) <> is written, if a straw is present the correct ROD output is written. The output stream thus created is ordered in a layer-by-layer fashion. This may not reflect the real readout sequence, but the differences are thought to be small. The output stream data agrees nicely with the numbers found in [5], where a subset was used of the same datasample that is used in this note. The output stream can now be used as input for the zero-suppression algorithm. In table 2 the average percentage of different straw data possibilities are shown. 5 Implementation An implementation of the zero suppression algorithm has been written in C. The basic operation per event consists of copying the ROD header and trailer to the output, checking for each straw the type of data and copying data and header for hit straws to the output. If a straw has not been hit, a counter indicating the offset to the previous hit is incremented. The header for hit straws is formed by the N bit encoding of this offset counter. In addition a check is made to see if the offset counter equals 2 N. If this is the case a N bit header with all bits set to followed by two zero s is copied to the output. Time measurements are made per event. A time stamp is set when the copying of the event header starts, another one when the copying of the event trailer is done. So only the time spent in the algorithm is measured, no file operations or printout statements are included in the time measurement. In the implementation for the Pentium platform, the time stamps are obtained from assembler code with which the number of clock cycles since the computer booted is read out. Information on the CPU speed is then used to convert this number into ms. The difference in time, together with the event size, is copied to the event header to be read out later. In addition, an implementation for the CRUSH [7] has been written. The algorithms are 4

implemented on the SHARC processor. Also here the number of clock cycles is counted and is later converted into ms. In the tests, only the data compression program was running on the SHARC. In reality also other tasks have to be executed, such as handling incoming RoI requests and deleting events [7]. Assembler code specific to the CRUSH has been used to execute operations such as bit shifts and logical comparisons of bytes (the most time consuming parts of the algorithms). The rest of the code is identical to the standard C implementation. For both implementations compiler optimisation has been used. 6 Results Distributions were made of the ratio of output and input size on an event-by-event basis while looping over all ROBs. This was done for different header sizes N. In fig. the result is shown for the barrel (fig. ) and the endcap (fig. ), in the case of low luminosity running. The plots for intermediate luminosity running are presented in fig. 2. The input sizes are standard ROD output format including headers and trailers. The output sizes include ROB headers and trailers. The solid line (in all four plots) gives the results in case of zero suppression without offset, i.e. using an eleven-bit address header for all hit straws. outputsize / inputsize [-].2.8.6 outputsize / inputsize [-].2.8.6.4.4.2.2 2 4 6 8 2 nr. of headerbits [-] 2 4 6 8 2 nr. of headerbits [-] Figure :. Output size/input size, averaged over all barrel data channels within one ROB, averaged over all ROBs, as a function of the number of header bits used in the zero suppression algorithm with offset. The plot is made with low luminosity data.. The same plot for the endcap data channels within one ROB. The solid line in both plots represents the result in case of zero suppression without offset. The plot is made with low luminosity data. The error bars indicate the spread of the ratio instead of the error on the average value. It is chosen here to show the spread in ratio instead of the error on the average value, since the ratio depends on the input (and output) size which is not a fixed number (at one particular bit header size) but varies for every event and for every ROB. This explains the rather large error bars in figs and 2. For smaller header sizes than the optimal header size the ratio is smaller due to an increase in the number of extra data items (<..><>). This results in a larger output. Above the optimal header size, the extra header bits that have to be sent for each hit straw result in a larger output. For the barrel the optimal header size for low luminosity is 6 bits, which gives a 5

outputsize / inputsize [-].4.2.8.6 outputsize / inputsize [-].4.2.8.6.4.2.4.2 2 4 6 8 2 nr. of headerbits [-] 2 4 6 8 2 nr. of headerbits [-] Figure 2:. Output size/input size, averaged over all barrel data channels within one ROB, averaged over all ROBs, as a function of the number of header bits used in the zero suppression algorithm with offset. The plot is made with intermediate luminosity data.. The same plot for the endcap data channels within one ROB. The solid line in both plots represents the result in case of zero suppression without offset. The plot is made with intermediate luminosity data. The error bars indicate the spread of the ratio instead of the error on the average value. ratio of output/input size of 46.2%. The ratio with a 5 bit header size is very close (46.%). For intermediate luminosity the minimum ratio for the barrel (7.%) is observed when a 4 bit header size is used. This ratio is larger than the typical ratios in the low luminosity case, since relatively more straws are hit, and thus the average number of sequential non-hits is smaller. For the endcap the optimal header sizes are 5 bits for low luminosity (ratio of 5.64%) and 4 bits for intermediate luminosity (ratio of 73.44%). In fig. 3 the distribution of input data size for both barrel ROBs (fig. 3) and endcap ROBs (fig. 3) are shown, for low and intermediate luminosity. The distributions for the input sizes are relatively narrow, the distributions for the output sizes have a considerably larger width, see fig. 4. In fig. 4 the barrel output is shown for the optimal header size for low luminosity (N=6) and for intermediate luminosity (N=4). In fig. 4 the output size in bytes is shown for the endcap for low (N=5) and intermediate luminosity (N=4). The long tail in the output size distribution is a result of the large spread in percentages of hit straws. Large output sizes stem from relatively large values of non-empty straws. For example the entry with 788 bytes in the barrel low luminosity distribution corresponds to an event with 27.9% non-empty straws within one ROB. 6

25 Mean low luminosity : 389.9 RMS low luminosity : 42.4 2 Mean low luminosity : 385.2 RMS low luminosity : 38.7 Mean intermediate luminosity : 43.5 Mean intermediate luminosity : 426. 2 RMS intermediate luminosity : 56.7 RMS intermediate luminosity : 45.4 8 5 6 4 5 2 2 4 6 8 bytes/rob 2 4 6 8 bytes/rob Figure 3:. Distribution of input event fragment sizes for one barrel ROB. The solid line gives the distribution for low luminosity, the dashed line for intermediate luminosity.. Distribution of input event fragment sizes for one endcap ROB. The solid line gives the distribution for low luminosity, the dashed line for intermediate luminosity. 4 2 Mean low luminosity : 23.6 RMS low luminosity : 8.4 Mean intermediate luminosity : 3.8 RMS intermediate luminosity : 2.5 6 Mean low luminosity : 23.9 RMS low luminosity : 88.6 5 Mean intermediate luminosity : 39.6 RMS intermediate luminosity : 7.9 4 8 6 4 2 3 2 2 4 6 8 bytes/rob 2 4 6 8 bytes/rob Figure 4:. Distribution of output event fragment sizes for barrel ROBs where the zero suppression algorithm with offset has been used, with the optimal number of 6 header bits for low luminosity (solid line) and 4 header bits for intermediate luminosity (dashed line).. The same plot for endcap ROBs now with the optimal number of 5 header bits for low luminosity (solid line) and 4 header bits for intermediate luminosity (dashed line). 7

6. Processing times The processing time of the algorithm was measured on the SHARC processor (clock frequency 4 MHz) and on various Pentium based Linux systems. For the optimal header sizes for barrel and endcap, the distribution of processing time on a GHz machine is shown in fig. 5. Note the large tail resulting from the spread in input size. The distributions of processing times for the optimal header sizes, running on a SHARC processor are shown in fig. 6. 4 Mean low luminosity :.37 Mean low luminosity :.36 RMS low luminosity :.5 RMS low luminosity :.5 2 Mean intermediate luminosity :.42 5 Mean intermediate luminosity :.42 RMS intermediate luminosity :.6 RMS intermediate luminosity :.5 4 8 3 6 4 2 2.2.3.4.5.6.7.8.2.3.4.5.6.7.8 Figure 5:. Distribution of processing times on a GHz Pentium platform, for the optimal header size of (N=6) bits for low luminosity (solid line) and (N=4) bits for intermediate luminosity (dashed line), for barrel data.. Distribution of processing times on a GHz Pentium platform, for the optimal header size of (N=5) bits for low luminosity (solid line) and (N=4) bits for intermediate luminosity (dashed line), for endcap data. 35 Mean low luminosity :.898 RMS low luminosity :.97 7 Mean low luminosity :.89 RMS low luminosity :.78 3 Mean intermediate luminosity :.965 RMS intermediate luminosity :. 6 Mean intermediate luminosity :.892 RMS intermediate luminosity :.2 25 5 2 4 5 3 2 5.4.6.8 2 2.2 2.4 2.6 2.8.4.6.8 2 2.2 2.4 2.6 2.8 Figure 6:. Distribution of processing times on the SHARC, for the optimal header size of (N=6) bits for low luminosity (solid line) and (N=4) bits for intermediate luminosity (dashed line), for barrel data.. Distribution of processing times for the SHARC, for the optimal header size of (N=5) bits for low luminosity (solid line) and (N=4) bits for intermediate luminosity (dashed line), for endcap data. 8

.6.5.4.3.6.5.4.3.2.2.. 2 4 6 8 2 nr. of headerbits [-] 2 4 6 8 2 nr. of headerbits [-] Figure 7:. Average processing time and the spread, measured on a GHz Pentium platform, vs. the number of header bits, for endcap data, low luminosity.. Average processing time and the spread, measured on a GHz Pentium platform, vs. the number of header bits, for endcap data, intermediate luminosity. The distributions in fig. 6 look more gaussian than the distributions in fig. 5, probably due to the fact that no other processes were running on the SHARC processor, while the algorithm measured on Pentium platforms had to share its time with other processes of the operating system. It was found that there is no significant dependence of the processing time on the number of header bits. This is shown in fig. 7 for the measurement done on a GHz Pentium platform for endcap data. The same plot for measurements done on the SHARC processor looks similar, with processing times scaled accordingly. outputsize / inputsize [-] 2.8.6.4.2 outputsize / inputsize [-] 2.8.6.4.2.8.6.4.2.2.3.4.5.6.7.8.8.6.4.2.2.3.4.5.6.7.8 Figure 8:. Scatterplot of the processing time on a GHz Pentium platform vs. the ratio of output and input size. The data used was for the endcap ROBs, at low luminosity running. The optimal header size (N=5) was used.. The same plot but now for endcap ROBs, intermediate luminosity, with an optimal header size of (N=4). 9

..8.6..8.6.4.4.2.2 5 5 2 25 3 35 4 nr of straws hit [-] 2 3 4 5 nr of straws hit [-] Figure 9:. Scatterplot of the processing time on a GHz Pentium platform vs. the number of straws hit (seen by one ROB). The data used was for endcap ROBs, low luminosity. The optimal header size (N=5) was used.. The same plot but now for endcap ROBs, intermediate luminosity, with an optimal header size of (N=4). For the measurements done on a GHz Pentium platform, the ratio of input size and output size versus the processing time, for the endcap ROBs, is shown in fig. 8 for low and intermediate luminosity. The plots show a weak linear dependence of the ratio of input and output size on the processing time, but with a large spread in the data. The average processing times on the different platforms are shown in table 3. A scatterplot was also made of the number of straws hit vs. the processing time. A clear linear dependence can be seen. A straight line fit through the data shows that of order 5 to 6 ns are spent per straw, i.e. on the GHz Pentium platform about 5-6 cycles per straw hit. The same data for the SHARC shows that approximately 2-25 cycles are needed per straw. The difference is due to different architecture for the processors as well as different instruction sets and different speeds for transferring data to and from the memory. Differences in system and processor architecture cause the number of clock cycles for the various Pentium platforms to differ from each other (see table 3). The rather large offset from zero in fig. 9 indicates that the algorithm spends a large time, evaluating caseand if-then-else statements. Processor Frequency Processing Time #ClockCycles MHz ms - SHARC 4..82 6.69 4 PentiumII 45.82 3.62 4 PentiumIII 733.5 3.56 4 PentiumIII.38 3.62 4 PentiumIV 4.38 5. 4 PentiumIV 7.3 4.95 4 Table 3: Average processing time per event for zero suppression with offset (N=5) for endcap ROBs for low luminosity, on different processors. In the last column the number of clock cycles per event is given.

7 Summary and Conclusions The most relevant parameters in this note are summarized in table 4. Also included is the reduction factor when the output size for the zero suppression algorithm is compared with the output size one gets for an bit address header. barrel endcap luminosity low int low int header size [bits] 6 4 5 4 output size [bytes] 24 3 24 32 ratio (output size/input size) [%] 46. 7. 5.7 73.4 reduction [%] 2.9 3.5 23. 3.6 Proc. time (SHARC) [ms].9.97.82.89 Proc. time ( GHz Pentium) [ms].37.42.38.42 Table 4: summary of the average of the most relevant parameters for the zero suppression algorithm for barrel and endcap, for low and intermediate luminosity. ( w.r.t. bit header, no offset) We have conducted several measurements with zero suppression algorithms for TRT data with and without offset on data sizes and processing times. The measurements were made on a prototype ROBIn (the CRUSH) as well as on Pentium-based platforms. We can draw the following conclusions : With the data-sets used in this note, the zero suppression with a 5-bit offset header for the endcap ROBs and a 6-bit header for barrel ROBs, gives a maximum reduction in data size for the full TRT scan for low luminosity. For intermediate luminosity the optimal header size is 4 bits for both barrel and endcap ROBs. The zero-suppression processing time hardly varies with the header size. Zero suppression on the CRUSH is too slow. On average, the processing time needs to be smaller than.2 ms. The processing times measured are.82 ms for endcap ROBs and.9 ms for barrel ROBs. Zero suppression on Pentium-processors can be done easily. Less than 5 µsecond per event, can be achieved already by a GHz processor. Since the algorithm consists mainly of evaluating simple if-then-else statements and case statements, an implementation of the algorithm in FPGAs could be advantageous. It is probably advantageous to move the zero suppression up to the RODs. At ROD-level the data is already compressed into the ROB format (see section 2). Performing the zero suppression algorithm at this level can be done fast if it is possible to implement it in FPGAs. An additional advantage is the size reduction of all event fragments sent from ROD to ROB. The results of this note depend on the order of reading out the straws. A layer-by-layer order (within one phi-segment, i.e. one ROB) has been assumed. The differences are thought to be small for other read out sequences. 8 Acknowledgement The authors of this note would like to thank Mogens Dam for making available the data set and subsequent help with the TRT digitization routine.

REFERENCES 2 References [] ATLAS High-Level Triggers, DAQ and DCS, Technical Proposal, CERN/LHCC 2-7, March 2 [2] ATLAS detector and physics performance, Technical Design Report (TDR), CERN/LHCC 99-5, 999 [3] Detector and read-out specification, and buffer-roi relations, for level-2 studies, P. Clarke et al., ATL-DAQ-99-4, October 999 [4] ATLAS Inner Detector, Technical Design Report (TDR), CERN/LHCC 97-6, April 997 [5] A Study of ROD compression schemes for the TRT, Mogens Dam, ATLAS DAQ note, ATL-INDET-2-9 [6] N. Ellis, report to TRT electronics meeting, CERN, 3 november 999 [7] A SHARC based ROB complex: design and measurement results, J.Vermeulen et al., ATLAS DAQ note, ATL-DAQ-2-2, 7 March 2