THE CMS apparatus [1] is a multi-purpose detector designed

Similar documents
CMS Note Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

Modules and Front-End Electronics Developments for the ATLAS ITk Strips Upgrade

The CCU25: a network oriented Communication and Control Unit integrated circuit in a 0.25 µm CMOS technology.

Muon Reconstruction and Identification in CMS

Fast pattern recognition with the ATLAS L1Track trigger for the HL-LHC

CMS Tracker PLL Reference Manual

APV-25 based readout electronics for the SBS front GEM Tracker

Quad Module Hybrid Development for the ATLAS Pixel Layer Upgrade

Control and Monitoring of the Front-End Electronics in ALICE

The ATLAS Conditions Database Model for the Muon Spectrometer

THE ATLAS INNER DETECTOR OPERATION, DATA QUALITY AND TRACKING PERFORMANCE.

Physics CMS Muon High Level Trigger: Level 3 reconstruction algorithm development and optimization

Results of Radiation Test of the Cathode Front-end Board for CMS Endcap Muon Chambers

Detector Control LHC

Track reconstruction of real cosmic muon events with CMS tracker detector

Control slice prototypes for the ALICE TPC detector

b-jet identification at High Level Trigger in CMS

arxiv:hep-ph/ v1 11 Mar 2002

The Phase-2 ATLAS ITk Pixel Upgrade

Trigger Server: TSS, TSM, Server Board

Deployment of the CMS Tracker AMC as backend for the CMS pixel detector

A real time electronics emulator with realistic data generation for reception tests of the CMS ECAL front-end boards

Radiation test and application of FPGAs in the Atlas Level 1 Trigger.

ATLAS, CMS and LHCb Trigger systems for flavour physics

RT2016 Phase-I Trigger Readout Electronics Upgrade for the ATLAS Liquid-Argon Calorimeters

CMS Conference Report

Investigation of Proton Induced Radiation Effects in 0.15 µm Antifuse FPGA

Production and Quality Assurance of Detector Modules for the LHCb Silicon Tracker

Level-1 Regional Calorimeter Trigger System for CMS

PoS(High-pT physics09)036

The Compact Muon Solenoid Experiment. Conference Report. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

First experiences with the ATLAS pixel detector control system at the combined test beam 2004

A flexible stand-alone testbench for facilitating system tests of the CMS Preshower

Construction of the Phase I upgrade of the CMS pixel detector

The Compact Muon Solenoid Experiment. Conference Report. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

Ignacy Kudla, Radomir Kupczak, Krzysztof Pozniak, Antonio Ranieri

ATLAS ITk Layout Design and Optimisation

Alignment of the ATLAS Inner Detector tracking system

Integrated CMOS sensor technologies for the CLIC tracker

Oliver Holme Diogo Di Calafiori Günther Dissertori. For the CMS collaboration

Level-1 Data Driver Card of the ATLAS New Small Wheel Upgrade Compatible with the Phase II 1 MHz Readout

Endcap Modules for the ATLAS SemiConductor Tracker

Prototyping of large structures for the Phase-II upgrade of the pixel detector of the ATLAS experiment

Performance of the ATLAS Inner Detector at the LHC

First Operational Experience from the LHCb Silicon Tracker

Adding timing to the VELO

Physics Procedia 00 (2011) A high speed serializer ASIC for ATLAS Liquid Argon calorimeter upgrade

The Compact Muon Solenoid Experiment. Conference Report. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

Tracking and Vertexing performance in CMS

Radiation tests of key components of the ALICE TOF TDC Readout Module

DCUF User Guide. G. Magazzu * A. Marchioro and P. Moreira CERN - EP/MIC, Geneva Switzerland. November 14, 2003 V.3.0 1

The LHCb upgrade. Outline: Present LHCb detector and trigger LHCb upgrade main drivers Overview of the sub-detector modifications Conclusions

The Front-End Driver Card for the CMS Silicon Strip Tracker Readout.

Detector Control System for Endcap Resistive Plate Chambers

PoS(EPS-HEP2017)492. Performance and recent developments of the real-time track reconstruction and alignment of the LHCb detector.

Atlantis: Visualization Tool in Particle Physics

The GBT-SCA, a radiation tolerant ASIC for detector control and monitoring applications in HEP experiments

The Sorting Processor Project

CMS Alignement and Calibration workflows: lesson learned and future plans

2003 TEC beam test : a quick overview

A Configurable Radiation Tolerant Dual-Ported Static RAM macro, designed in a 0.25 µm CMOS technology for applications in the LHC environment.

Muon Port Card Upgrade Status May 2013

Validation of the front-end electronics and firmware for LHCb vertex locator.

PoS(ACAT08)100. FROG: The Fast & Realistic OPENGL Event Displayer

ATLAS NOTE. December 4, ATLAS offline reconstruction timing improvements for run-2. The ATLAS Collaboration. Abstract

IEEE Nuclear Science Symposium San Diego, CA USA Nov. 3, 2015

Performance of Tracking, b-tagging and Jet/MET reconstruction at the CMS High Level Trigger

PoS(TIPP2014)204. Tracking at High Level Trigger in CMS. Mia TOSI Universitá degli Studi di Padova e INFN (IT)

FEC. Front End Control unit for Embedded Slow Control D R A F T. C. Ljuslin C. Paillard

Vertex Detector Electronics: ODE to ECS Interface

The performance of the ATLAS Inner Detector Trigger Algorithms in pp collisions at the LHC

Electron and Photon Reconstruction and Identification with the ATLAS Detector

Short Introduction to DCS, JCOP Framework, PVSS. PVSS Architecture and Concept. JCOP Framework concepts and tools.

The CMS Tracker: Front End Readout and Control Systems & DAQ

Readout Systems. Liquid Argon TPC Analog multiplexed ASICs SiPM arrays. CAEN 2016 / 2017 Product Catalog

The Belle Silicon Vertex Detector. T. Tsuboyama (KEK) 6 Dec Workshop New Hadrons with Various Flavors 6 7 Dec Nagoya Univ.

The CMS L1 Global Trigger Offline Software

The LHCb Upgrade. LHCC open session 17 February Large Hadron Collider Physics (LHCP) Conference New York, 2-7 June 2014

GLAST Silicon Microstrip Tracker Status

Front-End Electronics Configuration System for CMS. Philippe Gras CERN - University of Karlsruhe

Trigger Report. W. H. Smith U. Wisconsin. Calorimeter & Muon Trigger: Highlights Milestones Concerns Near-term Activities CMS

CTF3 BPM acquisition system

DT TPG STATUS. Trigger meeting, September INFN Bologna; INFN Padova; CIEMAT Madrid. Outline: most updates on Sector Collector system

Tracking and flavour tagging selection in the ATLAS High Level Trigger

Project Specification Project Name: Atlas SCT HV Power Supply (ASH) Version: V2.04. Project History

A New Segment Building Algorithm for the Cathode Strip Chambers in the CMS Experiment

CMS FPGA Based Tracklet Approach for L1 Track Finding

Summary of Optical Link R&D

SoLID GEM Detectors in US

Tracker Optical Link Upgrade Options and Plans

ATMEL ATF280E Rad Hard SRAM Based FPGA. Bernard BANCELIN ATMEL Nantes SAS, Aerospace Business Unit

IEEE Proof Web Version

Burn-In Test Software Development

Tracking and Vertex reconstruction at LHCb for Run II

On-chip Phase Locked Loop (PLL) design for clock multiplier in CMOS Monolithic Active Pixel Sensors (MAPS)

PoS(EPS-HEP2017)523. The CMS trigger in Run 2. Mia Tosi CERN

Using XDAQ in Application Scenarios of the CMS Experiment

The ATLAS Level-1 Muon to Central Trigger Processor Interface

Alignment of the CMS Silicon Tracker

The Compact Muon Solenoid Experiment. CMS Note. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

Transcription:

The CERN CMS Tracker Control System F. Drouhin, L. Gross, D. Vintache, A. Marchioro, C. Paillard, C. Ljuslin, L. Mirabito, P.G. Verdini Abstract Due to the high integration level of the experiments planned for the Large Hadron Collider (LHC) of the European Organization for Nuclear Research (CERN), the data acquisition and the control systems need complex developments both in hardware and software. The purpose of this paper is to describe the control system of a sub-detector of one of the CERN experiments, the Tracker of the Compact Muon Solenoid (CMS). The CMS Tracker control system is based on dedicated hardware and software. The hardware is based on the Front End Controller (), an interface board that hosts token rings for the communication with the Control and Communication Unit () modules. These in turn contain dedicated I/O channels for the front end readout and control chips. The software is built in layers: one device driver, a C++ dedicated application program interface (API) plus a database for the storage of all the information needed for the front end electronics. This system will also be adopted in some other CMS sub-detectors. Index Terms Control System, Software, Oracle Database, ASIC I. INTRODUCTION THE CMS apparatus [1] is a multi-purpose detector designed for operation at the CERN LHC [2] and optimized for the identification of the Higgs boson and the measurement of its properties over a wide mass range, as well as for the search of new physics phenomena. A high-precision muon detection and tracking system is installed around a four Tesla superconduction solenoid. Inside the magnet are located the hadronic and electromagnetic calorimeters, and the Central Tracker. The CMS Tracker [3] consists of an inner Pixel system and of an outer Silicon Strip Tracker (SST); the control system of the latter is described in this paper. The SST pseudorapidity coverage ranges up to η 2.5. It will provide at least ten points per track, resulting in a transverse momentum resolution for isolated tracks in the central region better than δp (15 p +0.5)%, where p is measured in TeV/c. This in turn will result in a reconstruction efficiency better than 98 % for muon tracks over the full F. Drouhin is with Université de Haute-Alsace, Mulhouse, France. Corresponding author: Frederic.Drouhin@cern.ch, Institut de Recherches Subatomiques de Strasbourg, Batiment 21, 23 rue du Loess, BP28, F67037 Strasbourg Cedex L. Gross is with Institut de Recherches Subatomiques de Strasbourg, France D. Vintache is with Institut de Recherches Subatomiques de Strasbourg, France A. Marchioro is with European Organization for Nuclear Research, Geneva, C. Paillard is with European Organization for Nuclear Research, Geneva, C. Ljuslin is with European Organization for Nuclear Research, Geneva, L. Mirabito is with Institut de Physique Nucléaire, Lyon, France P.G. Verdini is with Istituto Nazionale di Fisica Nucleare, Pisa, Italy pseudorapidity range, and better than 90 % for high energy electrons, while keeping the occupancy and the amount of ghost tracks at the percent level. These features require a single-hit resolution of 20µm, well within the possibilities of the technology chosen. The SST has been designed to guarantee stable operating conditions for ten years of LHC running, nominally. One of the critical issues for the viability of the system is the longterm survivability after heavy irradiation. The CMS SST will operate in a hostile environment, where the integrated particle fluence, expressed in 1 MeV neutron equivalent, is expected to reach 1.6 10 14 cm 2. In order to minimize the negative effects of annealing of radiation-induced damage to the Silicon bulk, and to prevent thermal runaway in the later stages of operation, the ambient temperature in the SST will be kept around 20 o C, when the saturation vapour pressure of water over ice corresponds to 3000 ppm. The need to compensate for accrued radiation damage in the sensors and in the front-end electronics has led to the development of complex, programmable readout electronics; the extreme environmental conditions have prompted the development of a radiation-hard ASIC which can be deployed on every individual module and which can be used to monitor the temperature of the sensors and of the front-end hybrid, the low voltage levels and the bias currents drawn by the sensors. The control system for the CMS SST must be capable of monitoring the temperature, humidity, voltage and current measurements coming from 16,000 modules, and at the same time ensure the download of correct parameters to the front-end electronics while checking for the occurrence of Single Event Upset phenomena in the chip registers. A dedicated control system, both in hardware and in software, has been developed to face this challenge. With increasing radiation exposure, the response of the silicon sensors as well as that the readout electronics is bound to change. Therefore, for the CMS Tracker special electronics has been developed whose mode of operation can be modified by programming specific registers. The Single Event Upset (SEU) phenomenon can change register values (bit changes). There are five kinds of front end chips shown in figure 1: - The APV [4]: front end chip for the multiplexed readout of 128 strips; - The APV Multiplexer (APV Mux): chip that multiplexes the outputs of two APV on a single optical link; - The PLL: programmable delay and clock rephaser; - The Laserdriver: used for the transfer of data between the front end electronics and the Front End Driver (analogue to digital converter with memory) over an optical fiber;

Fig. 1. A Tracker module with two silicon sensors and the electronic hybrid that hosts the front end electronics - The DCU (Detector Control Unit): this chip is very important in the CMS Tracker control system because it gives condition parameters for voltage, current and temperature. It contains an eight channel analog to digital converter, two constant current sources and a temperature sensor. It will monitor two sets of thermistors, one on the sensor and one on the hybrid, its own internal temperature, the silicon detector leakage current and the two (1.25 V and 2.5 V) low voltages. So there are three temperatures, two voltages and one current measurement for each hybrid. Moreover, each DCU has a unique hardware identification number that can be read by software. This identification number will be used in order to identify in the final Tracker construction the position of each module and to act as a link between the construction database that stores the module information (pedestal, calibration,...), the condition database and the online parameter database. The APV chips (4 or 6 per module), APV Mux, PLL and DCU (in the following text referenced to generically as devices ) each have several registers to configure. All registers are accessed over an Inter Integrated Circuit (I 2 C) bus with the standard modes described in the specification [5] and a specific addressing mode for the APV with a non standard extension of the I 2 C extended mode, the RAL mode. Before starting data taking through the data acquisition system [6], the front end electronics must be configured properly and each device register must be set through the I 2 Cbus. For the Silicon Strip Tracker, about 16,000 modules will be produced and installed. Thus the potential problems of the system (both in hardware and software) come from its complexity and from its dimensions. In fact, the parameter set and therefore the number of registers to access is about 1,680,000. Most of the device registers must be tuned in order to find the best conditions for taking physics data. In parallel to this electronic design, a control system was developed to set and retrieve all the device parameters. The next sections describe the hardware and software architecture. Note that this hardware and software control architecture is also used in other parts of the CMS experiment with the same / rings but different front end electronic chips. II. HARDWARE ARCHITECTURE A. Hardware description Token Ring Fig. 2. rings. 16 i2c channels 1 Memory channel 1 JTAG channel : Control and Communication Unit : Front End Controller 4 PIO channel i2c channels Hybrids / Devices 1 Trigger channel Hardware architecture. Note that one can manage up to 8 token Figure 2 shows the hardware architecture of the control system. This architecture is based on a specific board, the Front End Controller () [7], shown in figure 3, which hosts a token ring with modules, Communication and Control Units () [8], shown in figure 4. Each manages several kinds of channels:

- 16 I2 C channels for the access to the front end chips. Each channel can be configured for different speeds (100, 200, 400 khz and 1 MHz). - 4 Input/Output-like parallel bus controllers (Parallel input output (PIO)) such as the ones used in the Motorola PIA, etc. These channels are used in the Tracker in order to reset the front end readout electronics; - 1 memory channel, memory-like bus controller to access devices such as static memories, analog digital converters, etc. - 1 JTAG1 master controller ; - 1 trigger distribution controller ; - 1 node controller (the control itself is seen as a special channel capable, for instance, to report the status of the other channels). Fig. 3. Current prototype of the PMC Front End Controller Board for PCI support The memory, JTAG and the trigger channels are not used in the Tracker control system, but some other parts of the CMS experiment use the memory channel and check the possibility to use the trigger channel. The is built to be radiation resistant but SEU can cause bit changes, so a Cyclic Redundancy Checksum (CRC) and an auto-correction algorithm are implemented in case of errors. For SEU and noise reasons, the connexion between the and the ring is optical with laserdrivers. In the Tracker, the is completed with a DCU for the voltage and temperature measurements which must also be read. In order to prevent any problem due to the loss of a that would break the ring and consequently lead to lose all the on the ring, an additional set of connection, are used 1 IEEE Standard 1149.1 (JTAG), IEEE Standard Test Access Port and Boundary-Scan Architecture Fig. 4. Current prototype of the Control and Command Unit module with 6 I2 C channels and the two rings A/B for reliability to implement a second, redundant ring. This redundant ring makes it possible to bypass one by configuring the input and output of each and. Once the ring is built, for each access, whatever the channel is, a frame is built with a number, a address, a channel, a transaction and a command number. The channel number specifies one of the channels. The transaction number allows to recover the action which initiates the frame, for example, a read command. The command number gives the command to be executed (read, write, read modify write). For the I2 C accesses, the frame must also contain an address and possibly some data for write or read-modify-write operations. Note that the ring number is not given; it is configured via a specific register into the and. For research and development reasons, two versions of the have been developed, one with an electrical and one with an optical connection to the ring. Both are developed under the Peripheral Component Interconnect (PCI) bus but the final version of the will be on the VME bus with eight optical rings. An intermediate version will be available in mid-2004. For the same reasons two versions of exist (6 and 25). III. S OFTWARE ARCHITECTURE The software architecture presented in figure III is built upon the hardware architecture presented in the previous section. The software is split up into several parts: - Run Control [9]: controls the data acquisition and control system with the help of the Detector Control System. - Information Message service (IMS) integrated in the run control: dedicated task that receives all the messages (information, error,...) belonging to the run control. - device driver: a Linux device driver for the. - supervisor (, the FecSoftware API): central part of the control, it manages several s in order to read and write values into the front end electronics.

Supervisor PCI bus i2c access Oracle OCCI SOAP (Finite State Machine) Database Supervisor CMS General Run Control Supervisor IMS subscribe DCU Analyser Diagnosis system Database access through Oracle Call C++ Interface (OCCI) Finite state machine: SOAP message Hardware access Device driver access (software: open, read, write, ioctl, close) Fig. 5. FecSoftware Architecture - Database: Oracle database used to store all the values needed for the hardware. It manages all the parameters to set the hardware with the help of a versioning system. - DCU analyzer: dedicated process to analyze all the values provided by the DCU. - diagnostic system: system that diagnoses the source of errors and possibly takes a decision and signals it to the run control system. The software control is built around the device driver, database, (FecSoftware API), DCU analyzer and the diagnostic system. The control software issue is divided in two actions: - Download and upload parameters in the front end electronics: these two operations require that all / / channels (PIO and I 2 C channels) needed for the access, are enabled and configured. Once the control hardware is configured, the software sends the value needed to each register of the front end electronics (download). At the end of the download operation, the parameters set are read back from each device and compared to the values downloaded. If a difference appears, the faulty device register is stored in the database and an error is reported to the run control through the IMS. A similar operation can be undertaken directly between the event builder and the control processes (FecSupervisor) to make calibration or a delay curve. During this kind of operation, the event builder asks the control system to set PLL or APV values. The download/upload system must be done in a relatively short time (about 2 minutes for the current measurements) - Control the environmental parameters: this operation consists in reading all DCU on the hybrid and and to send all this information to the IMS. By a mechanism of web subscription, the information is sent to the DCU analyzer to check the correctness of the parameters (voltages and temperatures). These readings are performed every thirty seconds but this rate can be changed dynamically depending on the needs of the system. This system must detect a failure in any system in terms of temperatures and power. A second system exists but it is controlled through hardware locks to switch off all the Tracker to prevent, burning for example. Note that the DCU system is important to avoid the switching off of the complete system. If a problem occurs in temperature or humidity, the DCU analyzer must detect the divergence of the system and must send a message or take the decision to switch off the system correctly one part at a time. Each part of the software control system will be described with its requirements, the mode of operation and performances. The components are: device driver, Database, C++ Fec- Software API, DCU analyzer and finally the diagnostic system. IV. CONCLUSION A. FecSoftware summary Throughout this paper, all the interfaces needed to make system accesses and configurations will be presented: - /Ring/ hardware: physical layer; - device driver: low level software layer; - FecDevice class: interface to the device driver; - FecAccess class: concurrent access management; - DeviceAccess class and its sub-classes: specific device access; - FecSupervisor and FecAccessManager classes: software application.

All the tools and the API are completely documented with the Doxygen tool [11] and this documentation is available on the web page [10]. It gives developers the ability to develop and use the software for new front end electronic chip accesses and new channels. So the tools developed (device driver, C++ API) are used in some other parts of the CMS experiment: the Preshower, the µ chambers and the Electromagnetic Calorimeter for the time being. All software and tools described here are used in several Tracker production centers (Aachen, Karlsruhe, Louvain la Neuve, IPN Lyon, INFN Pisa, IReS Strasbourg, etc.) for the tests and the production of the detectors (silicon sensors for the Tracker). B. Performance This system has been fully tested in beam tests (2002, 2003) and has proved to fulfill all the requirements needed by the final experiment in terms of performance and functionalities for the Tracker. The download performance is good, for the whole Tracker in the worst case (only one supervisor and one database machine) the complete Tracker parameters settings are downloaded in 205 s. Concerning the upload performance, the time needed is not so good: about 2000 s is required (always in the worst case) due essentially to the database upload time. But this time is only valid if the complete Tracker front end electronics is in a bad state, in case of a power cut for example. In the best case (no errors), the upload time will only cost about 5 s to retrieve the parameters from the hardware and do a comparison, knowing that only the devices with errors in settings will be stored in the database. Note that these numbers come from the different curves that will be shown in the paper. Note that the performances described in this paper are given for hardware currently used. Due to choices of the Tracker and CMS collaboration, the Tracker will use the VME bus and personal computers for crate controllers with a PCI/VME links (see next paragraph on the evolution). We are obviously planning to repeat the measurements for the final hardware. is to manage the system through the same API, and the interfaces must be the same when using PCI bus or VME bus. Another upgrade is to develop the diagnostic system for the Tracker control. The use of an expert system is justified by the complex interconnections of the system. For example if the detector s data are not correct (detected by the data acquisition system), the diagnostic system must ask the control system if all the parameters are correct (power, values correctness in the front end chips) and an expert system can take the decision to recover the system by switching one of the or to the redundancy channel, resetting the or simply to re-downloading the values. The current plans are to develop this system and see how far we can automatize the solution of problems, i.e. whether this will be just and aid to the expert on chift or an automatic recovery tool. REFERENCES [1] CMS Collaboration, Technical Proposal, CERN/LHCC 94-38, December 1994 [2] http://lhc-new-homepage.web.cern.ch/lhc-new-homepage/ The Large Hadron Collider (LHC) project of CERN, [3] CMS Collaboration, The Tracker Projet - Technical Design Report, CERN/LHCC 98-6, April 1998 [4] P. Placidi, A. Marchioro, P. Moreira, K. Kloukinas, A 40 MHz Clock and Trigger Recovery Circuit for the CMS Tracker Fabricated in a 0.25µm CMOS Technology and Using a Self calibration Technique, presented at the 5th Workshop on Electronics for LHC Experiments, Snowmass, CO, USA, September 1999. [5] Philips Semiconductors, The I 2 C-Bus Specification, Version 2.1, document order number: 9398 39340011, January 2000 [6] L. Mirabito et al., Tracker data acquisition for beamtest and integration, Internal note, CMS IN 2003-021, May 2003 [7] A. Marchioro, C. LJustin, C. Paillard, Optical, Front End Control Unit for Embedded Slow Control, Draft 0.16, not pulished, February 2003 [8] A. Marchioro, C. LJustin, C. Paillard, 25, Communication and Control Unit for Embedded Slow Control, Draft 2.0, not pulished, May 2002 [9] M. Bellato et al. Run Control and Monitor System for the CMS Experiment, Submitted to Computing in High-Energy and Nuclear Physics, La Jolla CA, USA, March 24-28, 2003, [10] F. Drouhin, L. Gross, D. Vintache The FecSoftware documentation, http://cmsdoc.cern.ch/cms/cmt/system aspects/fe [11] Doxygen web site, http://www.doxygen.org/index.html C. Evolution The next upgrade of the system will be of the hardware. In fact, the final format of the board will be the VME and the software must be changed in order to take care of this bus. The CMS collaboration has chosen to use PCI/VME link, essentially for two reasons: - A Personal Computer with standard PCI interface is cheaper than a VME-board computer and has a more modern, powerfull processor - All CMS online software is developed on Linux PC that takes advantage of recent Linux kernel developments and all available Linux tools. The tests will be performed on a PCI/VME link, but the final link has not been chosen yet. The goal of this upgrade