Atlantis MultiRob Scenario

Similar documents
ATLANTIS - a modular, hybrid FPGA/CPU processor for the ATLAS. University of Mannheim, B6, 26, Mannheim, Germany

Scenarios for a ROB system built with SHARC processors. Jos Vermeulen, 7 September 1999 Paper model results updated on 7 October 1999

ROB-IN Functional demonstrator of the ATLAS Trigger / DAQ Read-Out Buffer O.Gachelin, M.Huet, P.Le Dû, M.Mur C.E.A.

The FTK to Level-2 Interface Card (FLIC)

Readout-Nodes. Master-Node S-LINK. Crate Controller VME ROD. Read out data (PipelineBus) VME. PipelineBus Controller PPM VME. To DAQ (S-Link) PPM

ROBIN Functional demonstrator of the ATLAS Trigger / DAQ Read-Out Buffer O.Gachelin, M.Huet, P.Le Dû, M.Mur C.E.A. Saclay - DAPNIA

Evaluation of network performance for triggering using a large switch

S-LINK: A Prototype of the ATLAS Read-out Link

A LVL2 Zero Suppression Algorithm for TRT Data

Read-out of High Speed S-LINK Data Via a Buffered PCI Card

Trigger and Data Acquisition at the Large Hadron Collider

ATLAS TDAQ RoI Builder and the Level 2 Supervisor system

The raw event format in the ATLAS Trigger & DAQ

DRAFT. Options for the ROB Complex

A Fast Ethernet Tester Using FPGAs and Handel-C

The MROD. The MDT Precision Chambers ROD. Adriaan König University of Nijmegen. 5 October nd ATLAS ROD Workshop 1

LHC Detector Upgrades

2008 JINST 3 T The ATLAS ROBIN TECHNICAL REPORT

PROTOTYPING HARDWARE FOR THE ATLAS READOUT BUFFERS

The ATLAS Data Acquisition System: from Run 1 to Run 2

CMX (Common Merger extension module) Y. Ermoline for CMX collaboration Preliminary Design Review, Stockholm, 29 June 2011

ATLANTIS A Hybrid FPGA/RISC Based Re-configurable System

Using SCI to Implement the Local-Global Architecture for the ATLAS Level 2 Trigger

Ethernet Networks for the ATLAS Data Collection System: Emulation and Testing

RPC Trigger Overview

Alternative Ideas for the CALICE Back-End System

An Upgraded ATLAS Central Trigger for 2015 LHC Luminosities

THE ATLAS DATA ACQUISITION SYSTEM IN LHC RUN 2

Front End Electronics. Level 1 Trigger

The new detector readout system for the ATLAS experiment

ROB IN Performance Measurements

The ATLAS Level-1 Muon to Central Trigger Processor Interface

The CMS Event Builder

2008 JINST 3 S Online System. Chapter System decomposition and architecture. 8.2 Data Acquisition System

The Baseline DataFlow System of the ATLAS Trigger & DAQ

Velo readout board RB3. Common L1 board (ROB)

The MROD. The Read Out Driver for the ATLAS MDT Muon Precision Chambers

LVL1 e/γ RoI Builder Prototype

FELI. : the detector readout upgrade of the ATLAS experiment. Soo Ryu. Argonne National Laboratory, (on behalf of the FELIX group)

Trigger and Data Acquisition: an ATLAS case study

RT2016 Phase-I Trigger Readout Electronics Upgrade for the ATLAS Liquid-Argon Calorimeters

Dominique Gigi CMS/DAQ. Siena 4th October 2006

The ATLAS Data Flow System for LHC Run 2

An ATCA framework for the upgraded ATLAS read out electronics at the LHC

The electron/photon and tau/hadron Cluster Processor for the ATLAS First-Level Trigger - a Flexible Test System

Detector Control LHC

NetFPGA Hardware Architecture

Upgrade of the ATLAS Level-1 Trigger with event topology information

L1 and Subsequent Triggers

The GAP project: GPU applications for High Level Trigger and Medical Imaging

Module Performance Report. ATLAS Calorimeter Level-1 Trigger- Common Merger Module. Version February-2005

arxiv: v1 [nucl-ex] 26 Oct 2008

LHCb Online System BEAUTY-2002

Fast pattern recognition with the ATLAS L1Track trigger for the HL-LHC

Read-Out Driver (ROD) and Data Challenges

The First Integration Test of the ATLAS End-Cap Muon Level 1 Trigger System

ROM Status Update. U. Marconi, INFN Bologna

M100 GigE Series. Multi-Camera Vision Controller. Easy cabling with PoE. Multiple inspections available thanks to 6 GigE Vision ports and 4 USB3 ports

Tracking and flavour tagging selection in the ATLAS High Level Trigger

The ATLAS High Level Trigger Region of Interest Builder

The WaveDAQ system: Picosecond measurements with channels

EMU FED. --- Crate and Electronics. ESR, CERN, November B. Bylsma, S. Durkin, Jason Gilmore, Jianhui Gu, T.Y. Ling. The Ohio State University

Improving Packet Processing Performance of a Memory- Bounded Application

Modeling and Validating Time, Buffering, and Utilization of a Large-Scale, Real-Time Data Acquisition System

SoLID GEM Detectors in US

The First Integration Test of the ATLAS End-cap Muon Level 1 Trigger System

The Read-Out Driver (ROD) for the ATLAS Liquid Argon Calorimeters

The CMS Computing Model

10 Gbit/s Challenge inside the Openlab framework

The Global Trigger Emulator System for the CMS experiment

Stefan Koestner on behalf of the LHCb Online Group ( IEEE - Nuclear Science Symposium San Diego, Oct.

1 MHz Readout. LHCb Technical Note. Artur Barczyk, Guido Haefeli, Richard Jacobsson, Beat Jost, and Niko Neufeld. Revision: 1.0

How to Choose the Right Bus for Your Measurement System

Presentation Outline. Data Concentrator Card for ECAL. ECAL Data Volume and Raw Data generation. DCC Conceptual Design

Data acquisition system of COMPASS experiment - progress and future plans

arxiv: v2 [nucl-ex] 6 Nov 2008

M100 GigE Series. Multi-Camera Vision Controller. Easy cabling with PoE. Multiple inspections available thanks to 6 GigE Vision ports and 4 USB3 ports

Chapter 12: Multiprocessor Architectures

2008 JINST 3 S Data Acquisition. Chapter 9

J. Castelo. IFIC, University of Valencia. SPAIN DRAFT. V1.0 previous to 5/6/2002 phone meeting

A generic firmware core to drive the Front-End GBT-SCAs for the LHCb upgrade

THE ALFA TRIGGER SIMULATOR

High Performance Simultaneous Data Acquisition

Development and test of a versatile DAQ system based on the ATCA standard

S100 Series. Compact Smart Camera. High Performance: Dual Core Cortex-A9 processor and Xilinx FPGA. acquisition and preprocessing

ATLAS PILE-UP AND OVERLAY SIMULATION

Field Program mable Gate Arrays

The Intelligent FPGA Data Acquisition

Supra-linear Packet Processing Performance with Intel Multi-core Processors

Trigger Layout and Responsibilities

arxiv: v1 [physics.ins-det] 11 Jul 2015

The CMS Event Builder

A Data Readout Approach for Physics Experiment*

PXI EXPRESS SPECIFICATION TUTORIAL

RUPRECHT-KARLS-UNIVERSITÄT HEIDELBERG

NUCLEAIRE EUROPEAN ORGANIZATION FOR NUCLEAR RESEARCH CERN LIBRARIES, GENEVA CERN-ECP

50GeV KEK IPNS. J-PARC Target R&D sub gr. KEK Electronics/Online gr. Contents. Read-out module Front-end

Development and test of the DAQ system for a Micromegas prototype to be installed in the ATLAS experiment

IBM Network Processor, Development Environment and LHCb Software

CMS Conference Report

Transcription:

Atlantis MultiRob Scenario --------------------------------- The ATLANTIS system is described in the attached document robscenario.pdf and can be viewed as a standard PC with a number of FPGA co-processors. However due to the compactpci board format a larger number of co-processors is supported and each of these is capable of handling up to 4 electrical links @60MB/s (M-Links, compatible to one specific S-Link implementation). The three basic functions of a Rob Complex (as seen by Level-2) RobIn, RSI, and RobC(ontroller) are usually (in the present PP studies) distributed in a way that RSI and RobC are handled by a single host cpu while the RobIn function is handled by a dedicated co-processor (of any kind). By adding more than one RobIn module per cpu one creates what we call a MultiRob. The benfit one can gain from a MultiRob-Complex can be limited by a number of factors:. Bandwidth of the host bus serving the input links (from RobIn's) 2. Bandwidth of the host bus serving the output link (to RSI) [if shared with input bus] 3. Bandwidth of the network link attached to the RSI 4. Realtime response characteristics of the host CPU (probably dominated by OS issues) 5. Processing power for managing input links (buffer management, preprocessing etc.) Most of the items above (additionally) suffer from the high message rates we expect for the RobComplex. The bandwidth of the input links is not considered critical in the scope of the RobComplex. The basic goals of the Atlantis MultiRob project are to Demonstrate a full MultiRob Complex, with real input and output, Measure the benefit of local (on the co-processors, and eventually via a private bus) managing and preprocessing (=> multi RobIn), compared to a single RobIn or a collection of single RobIn s on a single bus, Investigate the effect of the reduced message rates allowed by a RoI based rather than RoI-Fragment based traffic (for both RobC and RSI).

ROB Scenario on Atlantis Introduction This document describes a scenario for a ROB complex on the FPGA processorsystem ATLANTIS. The system consists of ROB-IN, ROB-Controller an ROB-OUT (RSI). Section 2 gives a summary about the requirements of the trigger system to a ROB Complex. Section 3 describes the setup of a ATLANTIS based ROB. Section 4 presents the ROB-IN based on the ATLANTIS I/O Board, Section 5 the ROB-Controler and Section 6 the ROB to switch interface. 2 Requirements This section is a summary of requirements for a ROB complex regarding the dierent subdetectors, the Level 2 trigger and the Event Filter. 2. Data from the ROD's The fragment size coming from the ROD's for the dierent subdetectors ist shown in table. Assuming a LVL acceptance rate of 75 khz the datarates between ROD and ROB-IN can be computed. Subdetector Typical Fragment Size Data Rate per ROB-IN Number of Links /event TRT 8kbits 75MB/s 256 SCT 2.5kbits 7MB/s 96 Pixel max 6.kbits 57.2MB/s 84 EM Calorimeter 4kbits 34MB/s 760 Hadronic Calorimeter 7kbit 67.5MB/s 64 Precision Muon 2-6 kbit 8,8MB/s-56.3MB/s 92 Trigger Muon 3kbit 28.7MB/s 48 Table : Eventsize and data rates tranfered from ROD to the one ROB-IN ant high luminosity [2] 2.2 Data to LVL2 and EF The data rates at the ROB output to the LVL2 trigger are dierent for the several subdetectors. This depends on the trigger menu which is used to examine the event. To get an overview about the data rates between ROB and LVL2 refer to the Modelling parameters of Jos Vermeulen[3]: For the dierent subdetectors the data rates of this model are summarized in Table 2 and 3. The data rate to the Event Filter depends on the LVL2 acceptance rate. The full data of every accepted event is transmited. Assuming a LVL2 acceptance rate of.4 khz [4] leads to the data rates in table 4.

Subdetector TRT SCT Pixel EM Calorimeter Hadronic Calorimeter Precision Muon Trigger Muon Maximal data rate per ROB-IN.5MB/s.MB/s 7.9MB/s 3MB/s Table 2: Maximal data rates at high luminosity from modell[3] Subdetector TRT SCT Pixel EM Calorimeter Hadronic Calorimeter Precision Muon Trigger Muon Maximal data rate per ROB-IN 6.5MB/s 8.6MB/s 2MB/s 5.9MB/s 3.9MB/s Table 3: Maximal data rates at low luminosity from modell[3] Subdetector TRT SCT Pixel EM Calorimeter Hadronic Calorimeter Precision Muon Trigger Muon Data rate to the EF per ROB-IN.4 MB/s 2.2 MB/s. MB/s 2.45 MB/s.2MB/s 0.4 - MB/s 0.5 MB/s Table 4: Data rates to the Event Filter 3 Setup for a ATLANTIS based ROB Complex The backbone of the ATLANTIS system is a Compact PCI crate with up to 4 PCI slots. One of these slots is always used for the host CPU which isanormal PC. At this time we usea8slot PCI backplane. So 7 slots are left to take Atlantis I/O boards. These act as four port ROB-IN's. A privat bus system connects all I/O boards independent of the PCI Bus. Over this second bus they communicate with a data rate up to 0.5GB/s with each other. The Host CPU acts as a ROB-Controller for all ROB-IN's. The ROB ouput port (RSI) is a network interface card (Ethernet). The whole setup can be seen in gure 4 The ROB-IN The ATLANTIS I/O board acts as a four port ROB-IN in the presented ROB Complex. The main task of the I/O board is the buering and preprocessing of data recieved from the ROD's. It acts as ROB-IN. A schematic structure can be seen in gure 2. In the actual conguration a 200MHz Pentium System with 28MB memory, harddisk, video adapter and fast ethernet. 2

2 H E L = J A * 7 5 * =?F=AM EJD &"+ F=?J2+ IJI 2 + * 7 5 BH 4 BH 4 BH 4 BH 4 KFJ%" J IM EJ? D B=IJAJD AH AJ Figure : The ATLANTIS based ROB Complex A H O A H O. 2 / ) 2 4 8 ) 6 - * 7 5 A H O A H O. 2 / ) 2 + * 7 5 ) J=JEI >=H@ Figure 2: The ATLANTIS I/O board 4. Input links The board is equipped with up to four MLink input ports. MLink is a LVDS based interface card which is compatible to the electrical SLink interface (SCSI). The dierence to SLink is a mechanicaly smaller interface card. Optional two standard SLink interface cards can 3

be used with the ATLANTIS I/O Board instead of the MLink cards. With 40 MHz SLink clock the maximal speed per input interface is 60MB/s which is the upper limit of the specication at this time. 4.2 FPGA's Two FPGA's on the I/O board are processing the incoming and outgoing data. Two input channels are dedicated to one FPGA. The design for one FPGA contains two input processes with memory management. So two FPGA's have four processes for storing data coming from 4 input links. The second task for the FPGA's is the preprocessing and the output to the PCI bus. One output task is used for the whole I/O board. 4.3 Memory The memory consists of four banks dual port SRAM with a datawidth of 32 bit, two per FPGA, which are independent accessable for each input process. The memory size depends on the event fragmet size coming from the ROD's, the event rate, the LVL2 processing time and the LVL2 acceptance rate with the EF readout time[]: MEM = EFSZ (ER LV L2 pt + LV L2 ar EF) SAV E EFSZ : Event fragment size ER:Eventrate from ROD LVL2 pt : Level 2 processing time LVL2 ar : Level 2 accept rate EF : Time between LVL2 accept and arrivel at EF SAVE : Safety factor This leads to a memory size of MB. For the I/O board are 4-8 MB per input link provided. The maximum transfer rate to memory is 60 MHz (40MHz clock speed). 4.4 Output to PCI The output process reads the requested event out of the buer memory and preprocesses it. This data package is send over the PCI Bus to the Host CPU. If a requested ROI requiers the data of several input ports, the output process merges the data to one big package and passes it over the PCI bus to the Host CPU. This reduces the number of small packages on the PCI bus and increases the PCI data rate. It reduces also the number of messages from the ROB Controler to request data, because only one ROB-IN board must be accessed. Measurements with the microenable ROB-IN, whose PCI interface is the same as on ATLANTIS Boards, show that at low data package sizes the PCI bandwidth increases dramatically with the package size. With a standard PCI bus (32 bit, 33 MHz) each of the seven I/O boards and output processes get a maximal bandwidth of 9 MB/s. This is not sucient for the most subdetectors particulary with low luminosity (table 4 and 3). So the number of I/O Boards should be matched to the required output datarates. The PCI bus will be one bottlenck of the ATLANTIS based ROB Complex. Using a 64 bit bus with 66 MHz increases the bandwidth of the PCI bus to 528 MB in future systems and eliminate the PCI bottleneck. 4

Figure 3: PCI output bandwidth vers. data size measured with microenable 4.5 Pre-processing Pre-processing is done in the FPGA's before passing the data to the PCI bus. This reduces the data volume and decreases the requierments for the following instances. For the TRT Rob a suggestion for a data format can be found in [5]. Also measurements have been done in [5] with a ROB-IN based on microenable. An other suggestion for a data format can be found in [2] which reduces the data volume with factor 0.7. A preprocessor algorithem for the SCT has been done in [6]. 4.6 The private bus The private bus is a connection between the ATLANTIS boards. It is not connected to the Host CPU. The bus has no specic protocol or protocol engine. It is a connection between two neighboured I/O Boards which gives the FPGA's the possibility to exchange signals. The privat bus is used for collecting data from neighboured I/O boards. This is useful if data from more then one input port is requested and the data can be found on two or more I/O boards. In that case one I/O board has the possibility to collect all data and pass one big data package over the PCI bus instead of several small packages. Also the ROB Controller passes only one message to one I/O board. Both increases the bandwidth of the PCI Bus and reduces the messages between ROB Controller an ROB-IN's. 5 The ROB-Controller The host CPU of the ATLANTIS system acts as ROB-Controler. The host CPU answers requests from outside to the ROB and dispatches them to the dierent I/O boards. It analyses the requested ROI's and requests the matching event data from the dierent boards. 5

6 ROB to Switch Interface The Rob to Switch Interface connects the ROB Complex with the LVL2 Trigger, the Eventlter and the Run Control. The actual ATLANTIS Host CPU has a build in 00 Mbit ethernet adapter. This is sucient for the rst studies of the system. But with four I/O Boards the bandwidth of that adapter is too low. In future systems a Gigabit ethernet adapter or a 622Mbit ATM adapter is more reasonable. References [] O. Boyle, R. McLaren. ROB-IN Input Module For The Demonstrator ROB, ECP Division, CERN, 996. [2] P.Clarke, F. J. Wickens. Detector and Read-Out Specication, and Buer-ROI Relations, for Level-2 Studies. Atlas DAQ Note Final DRAFT 9/999. [3] Jos Vermeulen. Modelling slides, http://www.nikhef.nl/pub/experiments/atlas/daq/modelling- 2-6-99/Presentation-June-99-update.pdf. [4] Atlas Technical Proposal Chapter 5. CERN, 994. [5] Ruediger Rissmann. Implementierungen von Vorverarbeitungsalgorithmen fur den AT- LAS Level 2 Trigger auf dem FPGA-Prozessor microenable. Universitiy of Heidelberg, 999. [6] R. Dankers, J.Bains. A Data Preparation Algorithm for the Precision Tracker LVL2 FEX. ATL-DAQ-99-00, CERN, 999. 6