Fermi National Accelerator Laboratory

Similar documents
REAL-TIME MULTIPLE NETWORKED VIEWER CAPABILITY OF THE DIII D EC DATA ACQUISITION SYSTEM

ESNET Requirements for Physics Reseirch at the SSCL

In-Field Programming of Smart Meter and Meter Firmware Upgrade

Cell Switching (ATM) Commonly transmitted over SONET other physical layers possible. Variable vs Fixed-Length Packets

Advanced Synchrophasor Protocol DE-OE-859. Project Overview. Russell Robertson March 22, 2017

FY97 ICCS Prototype Specification

Go SOLAR Online Permitting System A Guide for Applicants November 2012

ENDF/B-VII.1 versus ENDFB/-VII.0: What s Different?

SmartSacramento Distribution Automation

COMPUTATIONAL FLUID DYNAMICS (CFD) ANALYSIS AND DEVELOPMENT OF HALON- REPLACEMENT FIRE EXTINGUISHING SYSTEMS (PHASE II)

Portable Data Acquisition System

The APS Control System-Network

Development of Web Applications for Savannah River Site

Graphical Programming of Telerobotic Tasks

Large Scale Test Simulations using the Virtual Environment for Test Optimization

REAL-TIME CONTROL OF DIII D PLASMA DISCHARGES USING A LINUX ALPHA COMPUTING CLUSTER

Fermi National Accelerator Laboratory

Packet Switching - Asynchronous Transfer Mode. Introduction. Areas for Discussion. 3.3 Cell Switching (ATM) ATM - Introduction

Intelligent Grid and Lessons Learned. April 26, 2011 SWEDE Conference

Real Time Price HAN Device Provisioning

On Demand Meter Reading from CIS

Bridging The Gap Between Industry And Academia

Washington DC October Consumer Engagement. October 4, Gail Allen, Sr. Manager, Customer Solutions

Design of a Weighted Fair Queueing Cell Scheduler for ATM Networks

Adding a System Call to Plan 9

GA A22720 THE DIII D ECH MULTIPLE GYROTRON CONTROL SYSTEM

Alignment and Micro-Inspection System

Optimizing Bandwidth Utilization in Packet Based Telemetry Systems. Jeffrey R Kalibjian

Integrated Volt VAR Control Centralized

PJM Interconnection Smart Grid Investment Grant Update

RD-31 Status report 95

OPTIMIZING CHEMICAL SENSOR ARRAY SIZES

INTERF'ACING TO ACCELERATOR INSTRUMENTATION*

HWP2 Application level query routing HWP1 Each peer knows about every other beacon B1 B3

Entergy Phasor Project Phasor Gateway Implementation

zorder-lib: Library API for Z-Order Memory Layout

ACCELERATOR OPERATION MANAGEMENT USING OBJECTS*

High Scalability Resource Management with SLURM Supercomputing 2008 November 2008

DatasheetDirect.com. Visit to get your free datasheets. This datasheet has been downloaded by

NATIONAL GEOSCIENCE DATA REPOSITORY SYSTEM

CHANGING THE WAY WE LOOK AT NUCLEAR

Chapter 1 Introduction

NIF ICCS Test Controller for Automated & Manual Testing

A VERSATILE DIGITAL VIDEO ENGINE FOR SAFEGUARDS AND SECURITY APPLICATIONS

Performance of Multihop Communications Using Logical Topologies on Optical Torus Networks

Chapter 3 Packet Switching

MAS. &lliedsignal. Design of Intertransition Digitizing Decomutator KCP Federal Manufacturing & Technologies. K. L.

EECS 122: Introduction to Computer Networks Switch and Router Architectures. Today s Lecture

Research paper Measured Capacity of an Ethernet: Myths and Reality

MULTIPLE HIGH VOLTAGE MODULATORS OPERATING INDEPENDENTLY FROM A SINGLE COMMON 100 kv dc POWER SUPPLY

Trigger and Data Acquisition at the Large Hadron Collider

Increasing Performance for PowerCenter Sessions that Use Partitions

Testing PL/SQL with Ounit UCRL-PRES

PJM Interconnection Smart Grid Investment Grant Update

Forging a Future in Memory: New Technologies, New Markets, New Applications. Ed Doller Chief Technology Officer

Virtual University of Pakistan. Describe the Hidden Node and Exposed Node problems in Standard? VUSR. [Larry L. Peterson]

Generic Architecture. EECS 122: Introduction to Computer Networks Switch and Router Architectures. Shared Memory (1 st Generation) Today s Lecture

Understanding Packet Counters in show policy map interface Output

A COMBINED PLC AND CPU APPROACH TO MULTIPROCESSOR CONTROL

What Is Congestion? Effects of Congestion. Interaction of Queues. Chapter 12 Congestion in Data Networks. Effect of Congestion Control

Architecture of Computers and Parallel Systems Part 2: Communication with Devices

The CMS Event Builder

,,, Presented at 1983 Conf. on Real-Time Computer Applications i n Particle and Nuclear Physics,

Module 10 Frame Relay and ATM

Congestion in Data Networks. Congestion in Data Networks

LosAlamos National Laboratory LosAlamos New Mexico HEXAHEDRON, WEDGE, TETRAHEDRON, AND PYRAMID DIFFUSION OPERATOR DISCRETIZATION

Problem 7. Problem 8. Problem 9

National Accelerator Laboratory

Contributors: Surabhi Jain, Gengbin Zheng, Maria Garzaran, Jim Cownie, Taru Doodi, and Terry L. Wilmarth

different problems from other networks ITU-T specified restricted initial set Limited number of overhead bits ATM forum Traffic Management

METADATA REGISTRY, ISO/IEC 11179

Technical Committee. Interoperability Abstract Test Suites for the Physical Layer. af-test

Java Based Open Architecture Controller

Electronic Weight-and-Dimensional-Data Entry in a Computer Database

FPGA Implementation of RDMA-Based Data Acquisition System Over 100 GbE

TUCKER WIRELINE OPEN HOLE WIRELINE LOGGING

and opinions of authors expressed herein do not necessarily state or reflect those of the United States Government or any agency thereof.

The NMT-5 Criticality Database

Master Course Computer Networks IN2097

Asynchronous Transfer Mode

DESIGN AND IMPLEMENTATION OF AN AVIONICS FULL DUPLEX ETHERNET (A664) DATA ACQUISITION SYSTEM

Eng 3553 Lab #5 TCP Throughput

CS/EE 577 Final Exam December 17, 1996

What Is Congestion? Computer Networks. Ideal Network Utilization. Interaction of Queues

6.9. Communicating to the Outside World: Cluster Networking

Integrated Device Technology, Inc Stender Way, Santa Clara, CA Phone #: (408) Fax #: (408) Errata Notification

@ST1. JUt EVALUATION OF A PROTOTYPE INFRASOUND SYSTEM ABSTRACT. Tom Sandoval (Contractor) Los Alamos National Laboratory Contract # W7405-ENG-36

Stereo Vision Based Automated Grasp Planning

Final Report for LDRD Project Learning Efficient Hypermedia N avi g a ti o n

Superconducting Super Collider Laboratory. Bridging Fortran to Object Oriented Paradigm for HEP Data Modeling Task. J. Huang

CMS Conference Report

1/29/2008. From Signals to Packets. Lecture 6 Datalink Framing, Switching. Datalink Functions. Datalink Lectures. Character and Bit Stuffing.

An Empirical Study of Reliable Multicast Protocols over Ethernet Connected Networks

BWXT Y-12 Y-12. A BWXT/Bechtel Enterprise COMPUTER GENERATED INPUTS FOR NMIS PROCESSOR VERIFICATION. Y-12 National Security Complex

Architectural Differences nc. DRAM devices are accessed with a multiplexed address scheme. Each unit of data is accessed by first selecting its row ad

MOLECULAR DYNAMICS ON DISTRIBUTED-MEMORY MIMD COMPUTERS WITH LOAD BALANCING 1. INTRODUCTION

440GX Application Note

ATM. Asynchronous Transfer Mode. (and some SDH) (Synchronous Digital Hierarchy)

The Network Processor Revolution

Big Data Computing for GIS Data Discovery

Transcription:

Fermi National Accelerator Laboratory FERMILAB-Conf-96/439-E CDF CDF DAQ Upgrade and CMS DAQ R&D: Event Builder Tests Using an ATM Switch G. Bauer, T. Daniels, K. Kelley, P. Sphicas, K. Sumorok, S. Tether, J. Tseng and D. Vucinic Massachusetts Institute of Technology Cambridge, Massachusetts 02139 E. Barsotti, M. Bowden, J. Patrick Fermi National Accelerator Laboratory P.0. Box 500, Batavia, Illinois 60510 December 1996 Published Proceedings of the 2nd International Data Acquisition Workshop, Osaka, Japan, November 13-15, 1996 Operated by Universities Research Association Inc. under Contract No. DE-AC02-76CH03000 with the United States Department of Energy

Disclaimer This report was prepared as an account of work sponsored by an agency of the United States Government. Neither the United States Government nor any agency thereof, nor any of their employees, makes any warranty, expressed OT implied, or assumes any legal liability or responsibility GOT the accuracy, completeness, or usefulness of any information, apparatus, product, or process disclosed, or represents that its use would not infringe ptivately owned rights. Reference herein to any specific commercial product, process, or service by trade name, trademark, manufacturer, or otherwise, does not necessarily constitute or imply its endorsement, recommendation, OT favoring by the United States Government or any agency thereof. The views and opinions of authors expressed herein do not necessarily state or reflect those of the United States Government OT any agency thereof. Distribution Approved for public release; further dissemination unlimited.

CDF DAQ UPGRADE AND CMS DAQ R&D: EVENT BUILDER TESTS USING AN ATM SWITCH G. Bauer, T. Daniels! K. Kelley, P. Sphicas. K. Sumorok, S. Tether, J. Tseng, D. Vucinid Massachusetts Institute of Technology Cambridge, Massachusetts 02139, USA E. Barsotti, M. Bowden, J. Patrick Fermi National Accelerator Laboratory Batavia, Illinois 60510, USA Presented by J. Tseng E-mail: jtsengqfnal.gov ABSTRACT The present data acquisition system of the CDF experiment has to be upgraded for the higher luminosities expected during the Run II (1999+) data-taking period. The core of the system. consisting of a control network based on reflective memories will remain the same. The network used for data transfers, however, will have to be changed. We have investigated ATM as a possible replacement technology for the current Ultranet switch. \Ve present preliminary results on this new ATM-based event builder system. 1. Event Building at CDF The Collider Detector at Fermilab i (CDF) is a general purpose particle detector which has taken over 100 pb- of data at the Fermilab Tevatron since 1987 and is scheduled to take data again in 1999, accumulating well over 10 pb- per week. To take advantage of the high luminosity of the upgraded Tevatron, the three-level trigger heirarchy will be preserved, where the first two levels, implemented in hardware, will reduce the event rate from 7.6 million events/s to about 300 events/s (up to 1000 events/s) which are then assembled and analyzed by the Level 3 trigger. The average event size will be about 150 kb, assembled from about a dozen sources, the fragments ranging from several kilobytes to about 16 kb. The largest fragment size is expected to be 32 kb; to transfer such a fragment at 300 events/s requires that an individual link sustain traffic of 10 MB/s. 2 A promising technology for building such events is ATM, which is also being investigated for use in the CMS experiment. 3 This article reports on preliminary studies conducted at CDF with an ATM-based event builder test system. 2. The Event Builder Test System Figure la shows the conceptual architecture of the CDF DAQ system with its separate command and event networks. The Run Ib (1994-96) event network was a commercial network called Ultranet; it is this network for which an ATM switch

~L3 programs j receive; j \1 I rscanner Event ; Manager %txd, Network ATM Switch / / G-+ commands + event data Scanners Front ends I x ig 5 -Ring Fig. 1. (a) Conceptual architecture of the CDF data acquisition system. Data flows up from the front ends under the control of the Scanner Manager. (b) Th e event builder test system. The eight computers connected to the ATM switch are four Motorola MVME1603 s and four Radstone Tech- nology RS603 s. The Scanner Manager (SM) computer is a Motorola MVME1604. The SCRAMSet ring connects the VME reflective memory (RM) modules. is being investigated as a replacement. The test system for the upgrade is built around a FORE Systems ASX-1000 non-blocking ATM switch with 13 K cell output buffers. * The switch is currently equipped with eight 155 Mbps input/output ports but is expandable to up to 64 such ports. Non-blocking refers to the fact that the switch s internal bandwidth accomodates the maximum input bandwidth, even when the switch is expanded. Each port is connected to a PowerPC-based VME single- board computer running VxWorks 5.2; these computers can act as either Scanners, which in the real DAQ s y s t em read the detector front ends and send the event frag- ments through the event network, or Receivers, which combine the fragments into events for the Level 3 system. A 4 + 4 system is shown in Figure lb. The ATM in- terfaces are Interphase 4515 PCI-ATM adapters with 128 kb on-board RAM ( packet RAM ). A ninth PowerPC computer is used as a Scanner Manager and is connected to the other computers via the command network, in this case a Systran SCRAMNet ring of VME reflective memories. An ATM address consists of a virtual path identifier (VPI) and a virtual circuit identifier (VCI). In the test system, the VP1 f unctions as a physical address, uniquely assigned to individual computers, while the VCI addresses one of several event build- ing buffers on a given Receiver. A %irtual connection connects each Scanner with each event building buffer. The CDF event building software from Run I has been ported to this system and adapted for test purposes. This setup makes possible realistic performance measure- ments such as data and event throughput and investigations of cell losses, as well as

a) DMA 1 SEND 1 DMA 1 SEND 1 DMA 1 SEND 1 DMA 1 SEND b) DMA SEND 1 DMA 1 SEND I time t Fig. 2. XT11 driver send operations, slow (a) and fast (b) versions. refinements as necessary or advantageous. 3. ATM Interface The ATM interface driver software used in the present system uses the bare AAL protocol with a hardware-calculated CRC in order to investigate system performance without the overheads of lossless protocols such as TCP/IP. The driver software s in-house development also allows for flexible configuration and eases optimization. One optimization concerns how data is sent using the interface card. A send operation consists of two steps: first, the packet to be sent is loaded using DMA from the CPU s local memory into the packet RAM. The board is then instructed to send the packet, whereby the hardware takes over the operation, fragmenting the packet into ATM cells, packaging the cells in SONET frames, and sending the frames over the optical link. If the two steps are performed separately, as shown in Figure 2a, the output link is not used during the DMA step, which then contributes to transmission overhead. The DMA throughput has been measured to be 37.3 MB/s in the absence of other DMA operations, whereas the theoretical maximum payload throughput over the optical link, counting ATM and SONET overheads, is 17 MB/s; the theoretical maximum payload throughput of the slow send operation is therefore 11.7 MB/s. The marginal payload throughput is measured to be 11.2 MB/s, with the actual throughput using 32000-byte packets at 10.5 MB/s. It is possible to hide the DMA overhead by loading one packet while sending another as shown in Figure 2b. This transmission mode requires multiple output buffers in the packet RAM. The output hardware then interleaves cells from buffers being sent to different ATM addresses; sending two buffers to the same address causes the second buffer to wait on the first. With this fast transmission mode, which is used in all the subsequent tests, the payload throughput is measured to be 16.2 MB/s with very little overhead. 4. Direct Driver Tests The most basic tests involving the ATM components are those in which Naend computers perform uncoordinated rapid-fire packet transmissions to each of N,,, re- ceiving computers. The number of packets moving through the switch at a given

z20 ;- 218 : 516 r $14 r a) N + 1 Test. System 1 - n ; 16 b) N+N Relative Throughput &12 r *El0 i 8: r i, (16.2 MB/$/N 5 5.E 4 = X - -+ c- *....-*-.... *...--.- 0 1 1 I 1 1 5 0 I I I I I I 0 2 4 6 8 0 5 10 15 N = Number of senders R,, = 1-9 1 throughput (MB/s) Fig. 3. (a) Maximum sending rate for N slnd equivalent senders for one to seven senders to one receiver. The curve is the theoretical maximum, above which cells will be lost. The points occasion- ally lie below the curve because of the coarse-grained rate limit control. (b) Relative throughput (to Rrr, the 1 -+ 1 data throughput) vs. R11: for Nscnd = N,,,. time is therefore Ndend x N,,,. The driver is called directly, and the control network is ignored. These tests therefore reflect the best possible throughput performance. One obvious issue in this setup with multiple senders and receivers is that if several senders send data to a single receiver faster than it can be received, the ATM switch will simply drop the overflowing cells. However, the interface can be instructed to restrict its own sending rate on any given virtual connection by setting a hardware prescale counter. If the maximum reception rate is uma5, one naively expects that the maximum sending rate from equivalent senders will be u,,,/n~~~, above which cells will be lost, and indeed this is seen to be the case in Figure 3a. Since the rate limit is implemented per virtual connection, a sender s unused bandwidth can be used to send to other receivers in an Naend -+ N,,, system. The total throughput of the system should therefore scale as Ndend x N,,,. This scaling behavior is shown for Naend = N,,, by the plateaus in Figure 3b. The falling relative throughput for Rii > z,,,,/n,,, is due to having saturated the senders ATM links. Furthermore, no cells were lost at any set rate limit. It was also confirmed that the same data was received as was sent. 5. Traffic Shaping In the direct driver tests, there is only one virtual connection between each sender and receiver. However, as noted previously, the event builder system allocates a different VCI to each event building buffer so that one event may be built while

host 1 VPI 3 DMA / 50% SEND host 1 VPI 3 /DMA 1 50% SEND host 1 VPI 3 IDMA 1 50% SEND host 1 VPI 3 14 50% SEND I host 3 overload host 2 VPI 3 [ DMA j 50% SEND host 2 VPI 3 [ DMA 1 50% SEND host 2 VPI 3 1 DMA, 50% SEND host 2 VPI 3 w 50% SEND I time time 1 t w Fig. 4. Left: overloading a receiver ( VP1 3 ) with multiple transmit buffers on different senders in spite of rate division. Right: the solution implemented in the driver. another is processed by the Level 3 program. Hence a situation such as the one in Figure 4 (left) may arise where two transmit buffers on the same computer send to different event building buffers on the same receiving computer, in this case with VP1 3: the receiver is overloaded in spite of the rate limitation. One possible solution is to set the rate limit at 25% rather than 50%; however, this would leave some unused bandwidth as well as require more sophisticated control by the Scanner Manager to detect such clashes. It is simpler to restrict senders from transmitting multiple buffers to the same VPI, much as the hardware restricts sending to the same ATM address. Now the second buffer waits for the first to be sent, as shown in Figure 4 (right). DMA overhead remains hidden. With this restriction in place, the rate limit can be set to Vmaz/Nsend without consideration of the multiple sending and receiving buffers. To build an event using these divided sending rates, the Scanner Manager broad- casts one SEND-EVENT message to all the Scanners; each Scanner then sends at its allocated rate, after which it sends its acknowledgement back to the Scanner Manager. Multiple events are built concurrently as in the direct driver tests. This rate divi- sion algorithm is in contrast with the barrel shifter algorithm, which in all forms requires each Scanner to be informed one at a time via the command network when it is to send its data at the full rate. At CDF, this algorithm has been implemented with the Scanner Manager sending the individual SEND-EVENT commands, interleav- ing events being sent to different Receivers. Thus, the barrel shifter method incurs substantial control overhead from generating and passing these messages. The rate division and barrel shifter algorithms can be compared by running the event builder system without actually passing any data through the event network. In this case, an event is simply a complete round of control messages. These tests therefore measure the best possible (non-empty) event throughputs for the two algorithms. The results for a 4 + 4 system are shown in Figure 5a. The barrel shifter plateaus around 450 events/s; the target for Run II is 300 events/s, but 1000 events/s is desired. The rate division method, on the other hand, reaches 1000 events/s, albeit without sending any actual data. However, direct measurements show that the CPU is quickly saturated in the rate division test; a computer upgrade will likely further increase the event throughput. The

2 Y s $ 0.8 5 L5 0.6 G 60 c =oo 0 4-4 0 Barrel Shifter 4+4! ok1 ilitaliiil lli 0 0.5 1 1.5 2 Scanner Manager polling rate (khz) vnwx/d (MB/s) Fig. 5. (a) Event (messages only) throughput as a function of the rate at which the Scanner Manager polls for messages, for barrel shifter and rate division traffic shaping algorithms. The decreasing performance of the rate division method above 500 Hz is a result of CPU saturation. (b) A first look at the event builder test system data throughput, with each Scanner sending fixed-size 32000-byte fragments at the set rate limit v,,,/d. 6. Event Builder System Test The concern in using rate division for traffic shaping is that cells might be lost. It is clear from the direct. driver tests, however, that cell loss can be made very rare or even nonexistent. A first look at sending fixed-size Ndmd x 32000 byte events through the full event builder test system, with all the control software, is shown in Figure 5b. No cell loss was observed. It is encouraging to note that in the 4 -+ 4 system, each link is carrying more than 13 MB/s; the Run II target is 10 MB/s. The event throughput is 450 events/s with event fragments twice the size of the average largest fragment size for Run II. 7. Conclusion This article has reported on results from an event builder test system utilizing an ATM switch and in-house driver and control software. Tests using the ATM driver directly show the expected behavior regarding rate limitations and scaling, all without cell loss. Tests h ave also begun with the full test system utilizing all the control software. Again, event loss is not observed. Future work will more fully characterize the performance of the system, not only in terms of packet size and rate limits, but also with variable event sizes, in order to better simulate a real event

builder system. A software simulation of the switch and event builder system is also planned. Upgrades are already in progress to take advantage of faster computers and optical links as well as to expand to an 8 + 8 test system. The current tests suggest scalability to larger systems such as that projected at CMS, and in the nearer future, with only relatively modest upgrades foreseeable well within the next year, that an ATM-based event builder can meet or exceed CDF Run II performance targets. 8. References 1. F. Abe, et az.(cdf Collaboration), Nud. Instrum. Meth. A271, 387 (1988); F. Abe, et az.(cdf Collaboration), Phys. Rev. D50, 2966 (1994). 2. The CDF II Collaboration, The CDF II Detector: Technical Design Report, FERMILAB-PUB-96/390-E, October, 1996. 3. The CMS Collaboration, Technical Proposal, CERN/LHCC 94-38 (LHCC/Pl), D ecember 15, 1994. 4. FORE Systems, ForeRunner ATM Switch Architecture, April, 1996.