Overview. About CERN 2 / 11

Similar documents
Monitoring system for geographically distributed datacenters based on Openstack. Gioacchino Vino

CERN openlab II. CERN openlab and. Sverre Jarp CERN openlab CTO 16 September 2008

Benchmarking message queue libraries and network technologies to transport large data volume in

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Virtualizing a Batch. University Grid Center

Using the In-Memory Columnar Store to Perform Real-Time Analysis of CERN Data. Maaike Limper Emil Pilecki Manuel Martín Márquez

1. Introduction. Outline

CERN s Business Computing

Grid Computing a new tool for science

Tackling tomorrow s computing challenges today at CERN. Maria Girone CERN openlab CTO

CSCS CERN videoconference CFD applications

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

IEPSAS-Kosice: experiences in running LCG site

Storage and I/O requirements of the LHC experiments

CouchDB-based system for data management in a Grid environment Implementation and Experience

arxiv: v1 [physics.ins-det] 16 Oct 2017

ATLAS Experiment and GCE

Fault Detection using Advanced Analytics at CERN's Large Hadron Collider

Philippe Laurens, Michigan State University, for USATLAS. Atlas Great Lakes Tier 2 collocated at MSU and the University of Michigan

PoS(High-pT physics09)036

Travelling securely on the Grid to the origin of the Universe

Grid Computing: dealing with GB/s dataflows

The coolest place on earth

Batch Services at CERN: Status and Future Evolution

The LHC Computing Grid

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Stephen J. Gowdy (CERN) 12 th September 2012 XLDB Conference FINDING THE HIGGS IN THE HAYSTACK(S)

The CMS Computing Model

CC-IN2P3: A High Performance Data Center for Research

Technical Case Study CERN the European Organization for Nuclear Research

Table 1 The Elastic Stack use cases Use case Industry or vertical market Operational log analytics: Gain real-time operational insight, reduce Mean Ti

HPE SimpliVity. The new powerhouse in hyperconvergence. Boštjan Dolinar HPE. Maribor Lancom

Scalable global grid catalogue for Run3 and beyond

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

Database Services at CERN with Oracle 10g RAC and ASM on Commodity HW

Apache Spark 2.0 Performance Improvements Investigated With Flame Graphs. Luca Canali CERN, Geneva (CH)

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

Improving Packet Processing Performance of a Memory- Bounded Application

Preparing for High-Luminosity LHC. Bob Jones CERN Bob.Jones <at> cern.ch

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback

Computing at the Large Hadron Collider. Frank Würthwein. Professor of Physics University of California San Diego November 15th, 2013

High-Energy Physics Data-Storage Challenges

International Cooperation in High Energy Physics. Barry Barish Caltech 30-Oct-06

Fast packet processing in the cloud. Dániel Géhberger Ericsson Research

CERN and Scientific Computing

Geant4 Computing Performance Benchmarking and Monitoring

Big Data Analytics and the LHC

Storage Resource Sharing with CASTOR.

Accelerating Hadoop Applications with the MapR Distribution Using Flash Storage and High-Speed Ethernet

Grid Computing: dealing with GB/s dataflows

AWS & Intel: A Partnership Dedicated to fueling your Innovations. Thomas Kellerer BDM CSP, Intel Central Europe

Deploy a High-Performance Database Solution: Cisco UCS B420 M4 Blade Server with Fusion iomemory PX600 Using Oracle Database 12c

BlueGene/L. Computer Science, University of Warwick. Source: IBM

CMS High Level Trigger Timing Measurements

Data transfer over the wide area network with a large round trip time

Extremely Fast Distributed Storage for Cloud Service Providers

CLOUDS OF JINR, UNIVERSITY OF SOFIA AND INRNE JOIN TOGETHER

Accelerating Big Data: Using SanDisk SSDs for Apache HBase Workloads

Datasheet FUJITSU Software ServerView Cloud Monitoring Manager V1.1

Volunteer Computing at CERN

for the Physics Community

An ATCA framework for the upgraded ATLAS read out electronics at the LHC

ISTITUTO NAZIONALE DI FISICA NUCLEARE

Scaling Internet TV Content Delivery ALEX GUTARIN DIRECTOR OF ENGINEERING, NETFLIX

NCP Computing Infrastructure & T2-PK-NCP Site Update. Saqib Haleem National Centre for Physics (NCP), Pakistan

DataON and Intel Select Hyper-Converged Infrastructure (HCI) Maximizes IOPS Performance for Windows Server Software-Defined Storage

A MIMD Multi Threaded Processor

SPDK China Summit Ziye Yang. Senior Software Engineer. Network Platforms Group, Intel Corporation

Upgrade to Microsoft SQL Server 2016 with Dell EMC Infrastructure

SUPERMICRO, VEXATA AND INTEL ENABLING NEW LEVELS PERFORMANCE AND EFFICIENCY FOR REAL-TIME DATA ANALYTICS FOR SQL DATA WAREHOUSE DEPLOYMENTS

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN

Experience of the WLCG data management system from the first two years of the LHC data taking

Accelerating Microsoft SQL Server 2016 Performance With Dell EMC PowerEdge R740

HOW TO BUILD A MODERN AI

ATLAS Distributed Computing Experience and Performance During the LHC Run-2

Communication has significant impact on application performance. Interconnection networks therefore have a vital role in cluster systems.

Evaluation of the computing resources required for a Nordic research exploitation of the LHC

The ALICE Glance Shift Accounting Management System (SAMS)

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

Data Transfers Between LHC Grid Sites Dorian Kcira

Modeling and Validating Time, Buffering, and Utilization of a Large-Scale, Real-Time Data Acquisition System

Apache Spark Graph Performance with Memory1. February Page 1 of 13

e-science for High-Energy Physics in Korea

An Oracle White Paper September Oracle Utilities Meter Data Management Demonstrates Extreme Performance on Oracle Exadata/Exalogic

W H I T E P A P E R. Comparison of Storage Protocol Performance in VMware vsphere 4

Micron and Hortonworks Power Advanced Big Data Solutions

Learning with Purpose

RADU POPESCU IMPROVING THE WRITE SCALABILITY OF THE CERNVM FILE SYSTEM WITH ERLANG/OTP

STATUS OF PLANS TO USE CONTAINERS IN THE WORLDWIDE LHC COMPUTING GRID

Monitoring for IT Services and WLCG. Alberto AIMAR CERN-IT for the MONIT Team

Interconnect Your Future Enabling the Best Datacenter Return on Investment. TOP500 Supercomputers, November 2017

Insight: that s for NSA Decision making: that s for Google, Facebook. so they find the best way to push out adds and products

Cavium FastLinQ 25GbE Intelligent Ethernet Adapters vs. Mellanox Adapters

Intel Solid State Drive Data Center Family for PCIe* in Baidu s Data Center Environment

The Compact Muon Solenoid Experiment. Conference Report. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

Grid Computing at the IIHE

Low-Power Server Combines Efficiency with Durable, Mobile Performance

Cloud Analytics and Business Intelligence on AWS

Storage Optimization with Oracle Database 11g

VEXATA FOR ORACLE. Digital Business Demands Performance and Scale. Solution Brief

Transcription:

Overview CERN wanted to upgrade the data monitoring system of one of its Large Hadron Collider experiments called ALICE (A La rge Ion Collider Experiment) to ensure the experiment s high efficiency. They needed to constantly monitor their 2000 nodes processing data at 3.4 TB/s which leads to an incredible 600 khz metric rate. These metrics are collected and aggregated by Flume and Spark, and CERN chose InfluxDB to store these metrics. About CERN CERN, the world's largest particle physics center, consists of 16,868 members and collaborators, covers over 6,250,000 m2 of land, spans 700 buildings, and conducts various experiments. Founded in 1954, the CERN laboratory sits astride the Franco-Swiss border near Geneva. It was one of Europe's first joint ventures and now has 22 member states. At CERN, physicists and engineers are probing the fundamental structure of the universe. The instruments used at CERN are particle accelerators and detectors. Accelerators boost beams of particles to high energies before the beams are made to collide with each other or with stationary targets. Detectors observe and record the results of these collisions. The particles are made to collide together at close to the speed of light. The process gives physicists clues about how particles interact, and provides insights into the fundamental laws of nature. InfluxDB used for critical monitoring of accelerator systems, experiments and data centers at CERN The latest addition to CERN s accelerator complex, the Large Hadron Collider is the world s largest and most powerful particle accelerator. ALICE (A Large Ion Collider Experiment) is a heavy-ion detector on the LHC ring. It is designed to study the physics of strongly interacting matter at extreme energy densities, where a phase of matter called quark-gluon plasma forms. 2 / 11

There are many projects at CERN that are using InfluxDB. The largest installation is with the monitoring of one of our data centers where we are using 31 instances of InfluxDB, ingesting over 1.6 TB of metrics a day. Adam Wegrzynek, Senior Engineer LHC and ALICE at CERN The CERN Accelerator Complex hosts a number of projects, among them the Large Hadron Collider (LHC): Large due to its size (a ring approximately 27 km in circumference), superconducting magnets with a number of accelerating structures to boost the energy of the particles along the way Hadron because it accelerates protons or ions, which are hadron Collider because these particles form two high-energy particle beams travelling, at close to the speed of light, in opposite directions, and which collide in the four interaction points where the major experiments are located 3 / 11

Located 100 meters underground, LHC is the largest cryogenic system in the world and one of the coldest places on Earth, operating at -271.3 C. Such a cold temperature is required to operate the magnets that keep the protons on course. One of the LHC experiments, ALICE is a detector specialized in measuring and analyzing lead-ion collisions. It studies quark-gluon plasma ( a state of matter thought to have formed just after the big bang ) as it expands and cools, observing how it progressively gives rise to the particles that constitute the matter of our universe today. Collision Data Processing When the particles are about to collide, a special hardware, called Trigger, tells all the other subsystems to start collecting data. During the collision, the particle shower is created, and detectors generate the signal, which is pushed via special electronic devices into the computing firm. This data is then processed and compressed. Many different algorithms are applied and it is eventually sent to storage. 4 / 11

The reconstructed data is stored and analyzed in the Worldwide LHC Computing Grid (WLCG), which spans over 170 computing centers in 42 different countries, linking up national and international grid infrastructures. It runs around 2 million tasks a day using 750,000 cores. It already stores around 800 petabytes of data. InfluxDB Use Cases in CERN IT The IT department at CERN is responsible for Tier 0, the core of the WLCG. The Tier 0 core consists of two interconnected data centers, providing 20% of the computing power. InfluxDB is used to monitor this Tier 0, which is quite a large center. For this reason, 31 instances of InfluxDB are used, and 1.6 terabytes of metrics a day are written to InfluxDB. 5 / 11

WLCG Tier 0: InfluxDB The IT department at CERN also has a service system called DB On Demand, a custom-made deployment and cluster management system. It has a simple web self-service interface through which users can request any database they want, among them InfluxDB. CERN already has 90 projects using this service, collectively writing an impressive 1.5 million points per second. InfluxDB in CERN s DB On Demand Service System InfluxDB in the ALICE Experiment We write everything through InfluxDB, which is our main hardware s system. Why InfluxDB? The new monitoring system of ALICE consists of 19 detectors specialized in heavy ions. The heavy ions allow to form quark-gluon plasma. These deep collisions generate a temperature 100,000 times hotter than the Sun. 6 / 11

The upgraded monitoring system of ALICE will continuously read out about around 7.2 terabytes a day per second, over 9,000 fibers. Then it applies first-level compressing, decreasing it to 500 GB per second and then, eventually, GPU-accelerated reconstruction that brings down the data to 100 GB. This is written for storage and eventually to the WLCG. Such a system requires extensive monitoring. The team estimated around 600,000 metrics per second arriving asynchronously from 100,000 different sources, because this is the number of processes they expect to have in their firm. They also didn't want to introduce any high latency into the system because they have many critical values they want to ship to the shift crew in the control center. Monitoring System Upgrade To achieve their monitoring system upgrade, they considered tools such as MonALISA, Zabbix, StatsD, collectd, Prometheus, InfluxDB, PostreSQL, Apache Flume, Apache Kafka, Apache Spark, Riemann, and Grafana. After evaluating many tools, they chose InfluxDB as their Time Series Database because it met their monitoring system requirements in terms of latency, scalability, and write throughput. Technical Architecture Monitoring: Flow 7 / 11

The monitoring flow for ALICE s new data monitoring system is as follows: Collectd is used to gather performance metrics and monitor CERN s custom hardware. This data is shipped to Flume over UDP protocol. Flume routes data to Spark for aggregation and data is re-injected back to Flume. Some data is shipped directly to Grafana without going through InfluxDB, and other data goes to Riemann to trigger alarms and notifications for the operations team. Everything is written through InfluxDB, their main hardware s system. The Custom Monitoring Library In ALICE s new monitoring system, processing devices implement the compression and reconstruction logic and are linked against ALICE s custom C++ monitoring library. Written by Adam Wegrzynek, purely in C++, this library allows writing user-defined metrics, to various backends. It supports many monitoring systems and can monitor the process it's running within. It governs how much CPU, memory and network each process is using, for the 100,000 processes it's running. It also performs initial aggregation to calculate some direct metrics like rate, average, deltas, etc. So central monitoring can be offloaded from basic processing. The library can also monitor custom metrics. It is open source and free to use. Test Configuration for InfluxDB To find out if their system is able to handle the data load, CERN have a test configuration node with one InfluxDB database and the following hardware specs: CPU: 2x Intel Xeon E5-2670 v3 Memory: 64 GB DDR4 8 / 11

Network: Mellanox MT27520 40 GbE1 Storage: QLogic FastLinQ QL4121 25 GbE2 2x Intel S3610 800GB SSD (RAID 0) OS: CERN CentOS 7.4 3.10.0-693.21.1 InfluxDB: 1.5.2 Internals of the Receive Stack Some optimizations were performed, for testing purposes, to get the best values possible. InfluxDB: Optimization These optimizations involved the following: Linux Scaling Governor (a power-saving mode that scales CPU frequency) Receive Side Scaling - RSS ( which allows using multiple cues) RX Flow Hash (to add more variables like port numbers to be able to use all the cues) Socket receive queue memory Irqbalance NIC, IRQs and InfluxDB at same NUMA node LRO/GRO (Large Receive Offloads/Generic Receive Offloads) SO_REUSEPORT? (an option that allows having the same port number for multiple sockets) Visualizing Performance The below plot shows the percentage of successfully stored metrics in a function of the metric rate in kilohertz. On the yellow curve, at around 500 kilohertz, metrics begin to drop. From this graph, the CERN team can find out up to which point they can use the system, what the drop will 9 / 11

be, and learn the consequences of going higher. Upon visualizing performance, the team found that InfluxDB behaves quite well. InfluxDB: Writing They also executed a simple query to estimate how much time it takes to read the data when there is no major tasks running in the machine. The request completes at around 4 milliseconds, which was sufficient since that meant it could perform 550 requests per second. InfluxDB: Reading To improve the reading time when the system was performing reading and writing simultaneously, they are in the process of partitioning the system and using multiple InfluxDB instances. This would enable them to read and write all the data they want. Since they use Flume, it can choose where to write and push data. What's Next for ALICE? Future steps planned for the system include: 10 / 11

Partitioning Alarming Grafana real-time data source Custom hardware sensors Summary At CERN, several major projects use InfluxDB, and CERN s DB On Demand service promotes the usage of InfluxDB among CERN users. InfluxDB was chosen for ALICE s new monitoring system due to being a purpose-built Time Series Database that allows for high throughput ingest, compression and real-time querying of that same data in line with the system s monitoring requirements. 11 / 11