The ATLAS EventIndex: Full chain deployment and first operation

Similar documents
The ATLAS EventIndex: an event catalogue for experiments collecting large amounts of data

Event cataloguing and other database applications in ATLAS

The ATLAS EventIndex: an event catalogue for experiments collecting large amounts of data

THE ATLAS DISTRIBUTED DATA MANAGEMENT SYSTEM & DATABASES

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

PROOF-Condor integration for ATLAS

ATLAS Oracle database applications and plans for use of the Oracle 11g enhancements

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

Streamlining CASTOR to manage the LHC data torrent

Big Data Analytics Tools. Applied to ATLAS Event Data

Big Data Tools as Applied to ATLAS Event Data

Backup and Recovery for Hadoop: Plans, Survey and User Inputs

The ATLAS Distributed Analysis System

DQ2 - Data distribution with DQ2 in Atlas

UW-ATLAS Experiences with Condor

ATLAS Data Management Accounting with Hadoop Pig and HBase

A Fast and High Throughput SQL Query System for Big Data

Experiences with the new ATLAS Distributed Data Management System

MSG: An Overview of a Messaging System for the Grid

Distributed Systems 16. Distributed File Systems II

Chapter 5. The MapReduce Programming Model and Implementation

Storage for HPC, HPDA and Machine Learning (ML)

Introduction to Hadoop and MapReduce

ATLAS Analysis Workshop Summary

Computing Model Tier-2 Plans for Germany Relations to GridKa/Tier-1

TAG Based Skimming In ATLAS

Considerations for a grid-based Physics Analysis Facility. Dietrich Liko

CouchDB-based system for data management in a Grid environment Implementation and Experience

10 Million Smart Meter Data with Apache HBase

Tuning Intelligent Data Lake Performance

Copyright 2012, Oracle and/or its affiliates. All rights reserved.

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

Database Services at CERN with Oracle 10g RAC and ASM on Commodity HW

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

EsgynDB Enterprise 2.0 Platform Reference Architecture

Big Data Programming: an Introduction. Spring 2015, X. Zhang Fordham Univ.

Bring Context To Your Machine Data With Hadoop, RDBMS & Splunk

Data Clustering on the Parallel Hadoop MapReduce Model. Dimitrios Verraros

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN

Lightweight Streaming-based Runtime for Cloud Computing. Shrideep Pallickara. Community Grids Lab, Indiana University

Summary of the LHC Computing Review

Workload management at KEK/CRC -- status and plan

Virtualizing a Batch. University Grid Center

150 million sensors deliver data. 40 million times per second

Tuning Intelligent Data Lake Performance

Computing. DOE Program Review SLAC. Rainer Bartoldus. Breakout Session 3 June BaBar Deputy Computing Coordinator

Topics. Big Data Analytics What is and Why Hadoop? Comparison to other technologies Hadoop architecture Hadoop ecosystem Hadoop usage examples

BigTable: A Distributed Storage System for Structured Data

LHCb Computing Status. Andrei Tsaregorodtsev CPPM

A Glimpse of the Hadoop Echosystem

Where We Are. Review: Parallel DBMS. Parallel DBMS. Introduction to Data Management CSE 344

The ATLAS Production System

MOHA: Many-Task Computing Framework on Hadoop

Oracle at CERN CERN openlab summer students programme 2011

3-2. Index construction. Most slides were adapted from Stanford CS 276 course and University of Munich IR course.

Improving overall Robinhood performance for use on large-scale deployments Colin Faber

MATE-EC2: A Middleware for Processing Data with Amazon Web Services

Index Construction. Dictionary, postings, scalable indexing, dynamic indexing. Web Search

August Li Qiang, Huang Qiulan, Sun Gongxing IHEP-CC. Supported by the National Natural Science Fund

WLCG Transfers Dashboard: a Unified Monitoring Tool for Heterogeneous Data Transfers.

Hadoop An Overview. - Socrates CCDH

LCG data management at IN2P3 CC FTS SRM dcache HPSS

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

Index construction CE-324: Modern Information Retrieval Sharif University of Technology

A Review Paper on Big data & Hadoop

Andrea Sciabà CERN, Switzerland

Index construction CE-324: Modern Information Retrieval Sharif University of Technology

Memory: Page Table Structure. CSSE 332 Operating Systems Rose-Hulman Institute of Technology

Technological Challenges in the GAIA Archive

What is the maximum file size you have dealt so far? Movies/Files/Streaming video that you have used? What have you observed?

Data Management. Parallel Filesystems. Dr David Henty HPC Training and Support

Stephen J. Gowdy (CERN) 12 th September 2012 XLDB Conference FINDING THE HIGGS IN THE HAYSTACK(S)

EventStore A Data Management System

Using space-filling curves for multidimensional

Dept. Of Computer Science, Colorado State University

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

Monitoring system for geographically distributed datacenters based on Openstack. Gioacchino Vino

Monitoring of Computing Resource Use of Active Software Releases at ATLAS

Next-Generation Cloud Platform

Embedded Technosolutions

Early experience with the Run 2 ATLAS analysis model

Bigtable. A Distributed Storage System for Structured Data. Presenter: Yunming Zhang Conglong Li. Saturday, September 21, 13

CSCI 5417 Information Retrieval Systems Jim Martin!

Delving Deep into Hadoop Course Contents Introduction to Hadoop and Architecture

Cloud Computing CS

To Shard or Not to Shard That is the question! Peter Zaitsev April 21, 2016

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

CS November 2018

Monitoring WLCG with lambda-architecture: a new scalable data store and analytics platform for monitoring at petabyte scale.

CS November 2017

Storage Resource Sharing with CASTOR.

Page 1. Goals for Today" Background of Cloud Computing" Sources Driving Big Data" CS162 Operating Systems and Systems Programming Lecture 24

CIS 601 Graduate Seminar. Dr. Sunnie S. Chung Dhruv Patel ( ) Kalpesh Sharma ( )

Data handling and processing at the LHC experiments

CS / Cloud Computing. Recitation 11 November 5 th and Nov 8 th, 2013

The amount of data increases every day Some numbers ( 2012):

2013 AWS Worldwide Public Sector Summit Washington, D.C.

CS / Cloud Computing. Recitation 3 September 9 th & 11 th, 2014

Bigtable: A Distributed Storage System for Structured Data By Fay Chang, et al. OSDI Presented by Xiang Gao

Information Retrieval

Transcription:

The ATLAS EventIndex: Full chain deployment and first operation Álvaro Fernández Casaní Instituto de Física Corpuscular () Universitat de València CSIC On behalf of the ATLAS Collaboration 1

Outline ATLAS Experiment Motivation for an event index Prototype and first operation Distributed data collection Hadoop Core Query Services Summary and Conclusions 2

Computing Challenges 2 0 1 2 Experiments like ATLAS produce large amounts of data: 2 billion real events and 4 billion simulated events were produced in 2011 and the same in 2012. These numbers increase if we take into account all the reprocessing. After current Long Shutdown 1, in 2015: Increasing rate to 1 khz. Event size 0.8/1MB. RAW DATA 800/1000 MB/s Translates globally in CPU +20%, DISK +15%, TAPE +15% each year. 3

Event Cataloguing All this information needs cataloguing according to different points of view to meet multiple use cases and search criteria. A system that contains the reference to the file that includes every event at every stage of processing is necessary to recall selected events from data storage systems. ATLAS already has a database called EventTag, with some issues: Using traditional DBMS Oracle, expensive to maintain Scaling up in hardware is expensive. Not all use cases implemented. Not completely automatized: asynchronous, error prone Consistency issues Performance not satisfactory EventTag is potentially very useful but has to improve. 4

EventIndex We need another approach: EVENTINDEX PROJECT, using Structured Storage ( ie: NoSQL) technologies: Appropriate to scale for increased data rates (Commodity HW, Fixed cost/unit), scalable. Cheaper, easier to use and faster for sparse searches over large amounts of data. A complete catalogue of all ATLAS events: All events, real and simulated data. All processing stages Event Info Guid Guid Needed EventIndex information ~= 300bytes to 1Kbyte per event: Event identifiers Online trigger pattern and hit counts References (pointers) to the events at each processing stage (RAW, ESD, AOD, NTUP) in all permanent files on storage 5

Use Cases Event picking: Give me the reference (pointer) to "this" event in "that" format for a given processing cycle. Event skimming: Give me the list of events passing "this" selection and their references. Production consistency checks: Technical checks that processing cycles are complete. Panda event service (see talk in this session): Give me the references for the events on this GUID file (to be distributed to HPC or cloud clusters for processing) 6

Prototype Overview 1 ) DATA COLLECTION 3) QUERY SERVICES 2) HADOOP CORE 7

Data Collection - Producer EventIndex Information is produced in Athena (Atlas software framework) with python scripts, running in Tier0 and GRID sites. Current tools include POOL, RAW, TAG formats, and produce the information in a intermediate file to be sent afterwards. Production jobs measured last May 2013 give us the daily data to be loaded as 265 K gridjobs/day -> 300 Million events/day -> 1Kb/event. Means around 300 GB/day. Rate of the Producers is therefore measured as 20Hz of file records containing ~3.4 khz of event records, to be increased the following years up to 80Hz and 30kHz of event records. 8

Data Collection - Messaging Collection of the produced Event Info is done with a messaging system based on the STOMP protocol, and ActiveMQ brokers. Message size is set small 1-10kB to keep broker queues agile. Atomic transactions on Producers: if connections breaks no partial processing occurs. 2 brokers for HA: @CERN and future @Wigner Stomp performance in our benchmarks reaches enough 350 msg/s and 10Mb/s per Producer. Measured performance for sending real events reached 200K event/s and 60Mb/s (1Broker, 6 Prod/s, 4 Cons, 50K events/job) 9

Data Collection - Consumer Consumers are in charge of retrieving the information from the brokers and insert it into Hadoop. Messages tagged by group (JMSXGroupID) All messages from 1 Producer will go to 1 Consumer (while is alive) Files stored on Mapfile format usable by Hadoop Core Services 10

Hadoop Core (1) Data uploaded from Consumers is stored in MapFiles files in HDFS. Catalog is stored in Hbase. Data can be indexed (e.g. Using inverted indices for trigger info). Index files just have key + references to data files. Some index files created at upload, others added. Searches can be performed using keys (immediate results), or on data files. Searches can be done with full scans or using MapReduce jobs. 11

Files in HDFS are logically grouped in collections of filesets, representing event collections. Each Fileset contains: a master record, pointers to before /after filesets (vertical partitions), slave filesets ( horizontal partitions), and indexes. Each file is represented by one entry in a HBASE table. The Catalog checks presence and consistency of all components. Hadoop Core (2) 12

Hadoop Core (3) Important Use case: query/count the events that satisfy a given trigger condition, or a combination. Trigger chains need to be first decoded by knowing the trigger menu via the run number and finding in the Conditions Metadata (COMA) database the corresponding trigger info for each bit in the trigger words. Trigger decoding (get chain names from trigger bits and run number) is also cached to improve the future searches. 13

Query Services A Typical query from the Panda Event Service for all events from a GUID takes ~3-10 secs Query services are the interface for users to the Core Hadoop: Command line interface, Web Services with graphical and text/plain output suitable for the Panda Event Service. Simple data scan operations on the test dataset (1 TB of 2011 data) using the available Hadoop cluster of 20 nodes take about 15 minutes. worst case, reads all data Queries that give the run number and event number (event picking use case) and use the so-far unoptimised accessor method return their result in 3.7 seconds 14

Summary and Conclusions The ATLAS experiment needs an EventIndex to catalogue billions of events taken yearly. Several TBs of catalogue data are needed to index several PBs of event data. Full chain tested successfully with first prototype: EventIndex production worldwide at different Grid Sites. Messaging system can amply cope with the required rates for the following years. Hadoop core tested with previously loaded 1TB of 2011 data, and data coming from the data collection. Query services to satisfy the use cases. Results are encouraging, and performance suitable to improve. Next steps are: Automate procedures. Load all Run1 data. Tune the system. According to the original plan and the current experience the system will be fully working at the start of 2015, ready for the new LHC data. 15