Data Analysis in ATLAS. Graeme Stewart with thanks to Attila Krasznahorkay and Johannes Elmsheuser

Similar documents
Early experience with the Run 2 ATLAS analysis model

ATLAS Analysis Workshop Summary

Expressing Parallelism with ROOT

Monitoring of Computing Resource Use of Active Software Releases at ATLAS

CMS Analysis Workflow

A new petabyte-scale data derivation framework for ATLAS

Prompt data reconstruction at the ATLAS experiment

Event cataloguing and other database applications in ATLAS

Large Scale Software Building with CMake in ATLAS

Big Data Tools as Applied to ATLAS Event Data

The CMS data quality monitoring software: experience and future prospects

Big Data Analytics Tools. Applied to ATLAS Event Data

Experience with Data-flow, DQM and Analysis of TIF Data

Machine Learning in Data Quality Monitoring

Monte Carlo programs

The Run 2 ATLAS Analysis Event Data Model

The ATLAS Distributed Analysis System

Data handling and processing at the LHC experiments

PARALLEL PROCESSING OF LARGE DATA SETS IN PARTICLE PHYSICS

Stephen J. Gowdy (CERN) 12 th September 2012 XLDB Conference FINDING THE HIGGS IN THE HAYSTACK(S)

Considerations for a grid-based Physics Analysis Facility. Dietrich Liko

Machine Learning Software ROOT/TMVA

TAG Based Skimming In ATLAS

Belle & Belle II. Takanori Hara (KEK) 9 June, 2015 DPHEP Collaboration CERN

Practical Statistics for Particle Physics Analyses: Introduction to Computing Examples

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Clustering and Reclustering HEP Data in Object Databases

Framework design experience from art. Marc Paterno EIC software workshop 25 September 2015

Data Analysis in Experimental Particle Physics

Event Displays and LArg

Analysis & Tier 3s. Amir Farbin University of Texas at Arlington

CERN Open Data and Data Analysis Knowledge Preservation

LHCb Computing Resources: 2018 requests and preview of 2019 requests

Validation of Missing ET reconstruction in xaod

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN

ATLAS PILE-UP AND OVERLAY SIMULATION

New data access with HTTP/WebDAV in the ATLAS experiment

Reducing CPU Consumption of Geant4 Simulation in ATLAS

LHCb Computing Status. Andrei Tsaregorodtsev CPPM

Toward real-time data query systems in HEP

Parallelism. Parallel Hardware. Introduction to Computer Systems

Three OPTIMIZING. Your System for Photoshop. Tuning for Performance

VISPA: Visual Physics Analysis Environment

CMS data quality monitoring: Systems and experiences

Popularity Prediction Tool for ATLAS Distributed Data Management

Computing at Belle II

Distributed Data Management on the Grid. Mario Lassnig

The Attraction of Complexity

BIG Data How to handle it. Mark Holton, College of Engineering, Swansea University,

ATLAS Nightly Build System Upgrade

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

New strategies of the LHC experiments to meet the computing requirements of the HL-LHC era

The ATLAS EventIndex: Full chain deployment and first operation

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

pyframe Ryan Reece A light-weight Python framework for analyzing ROOT ntuples in ATLAS University of Pennsylvania

CMS - HLT Configuration Management System

Computing. DOE Program Review SLAC. Rainer Bartoldus. Breakout Session 3 June BaBar Deputy Computing Coordinator

Physics Analysis Software Framework for Belle II

ATLAS NOTE. December 4, ATLAS offline reconstruction timing improvements for run-2. The ATLAS Collaboration. Abstract

The ATLAS Trigger Simulation with Legacy Software

INTRODUCTION TO THE ANAPHE/LHC++ SOFTWARE SUITE

Data Reconstruction in Modern Particle Physics

HEP replica management

The ATLAS EventIndex: an event catalogue for experiments collecting large amounts of data

A PROPOSAL FOR MODELING THE CONTROL SYSTEM FOR THE SPANISH LIGHT SOURCE IN UML

HammerCloud: A Stress Testing System for Distributed Analysis

Collider analysis recasting with Rivet & Contur. Andy Buckley, University of Glasgow Jon Butterworth, University College London MC4BSM, 20 April 2018

Introduction to Concurrent Software Systems. CSCI 5828: Foundations of Software Engineering Lecture 12 09/29/2016

Hall D and IT. at Internal Review of IT in the 12 GeV Era. Mark M. Ito. May 20, Hall D. Hall D and IT. M. Ito. Introduction.

Data services for LHC computing

ATLAS operations in the GridKa T1/T2 Cloud

Automated Testing of Tableau Dashboards

Interactive 3D graphics in HEP

Martin Dove s RMC Workflow Diagram

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

The ATLAS Conditions Database Model for the Muon Spectrometer

Update of the Computing Models of the WLCG and the LHC Experiments

ONS Beta website. 7 December 2015

LHCb Computing Resource usage in 2017

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback

Software. CPU implements "machine code" instructions. --Each machine code instruction is extremely simple. --To run, expanded to about 10 machine code

Data and Analysis preservation in LHCb

Introduction to Concurrent Software Systems. CSCI 5828: Foundations of Software Engineering Lecture 08 09/17/2015

Construction Change Order analysis CPSC 533C Analysis Project

An SQL-based approach to physics analysis

Computing / The DESY Grid Center

DESY at the LHC. Klaus Mőnig. On behalf of the ATLAS, CMS and the Grid/Tier2 communities

ALICE ANALYSIS PRESERVATION. Mihaela Gheata DASPOS/DPHEP7 workshop

Using the In-Memory Columnar Store to Perform Real-Time Analysis of CERN Data. Maaike Limper Emil Pilecki Manuel Martín Márquez

LHCb Computing Resources: 2019 requests and reassessment of 2018 requests

Existing Tools in HEP and Particle Astrophysics

Monitoring of large-scale federated data storage: XRootD and beyond.

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 27 July 2010

How to discover the Higgs Boson in an Oracle database. Maaike Limper

TDDC88 Lab 4 Software Configuration Management

Influence of Distributing a Tier-2 Data Storage on Physics Analysis

Insight: that s for NSA Decision making: that s for Google, Facebook. so they find the best way to push out adds and products

A Tool for Conditions Tag Management in ATLAS

2012 Enterprise Strategy Group. Enterprise Strategy Group Getting to the bigger truth. TM

Transcription:

Data Analysis in ATLAS Graeme Stewart with thanks to Attila Krasznahorkay and Johannes Elmsheuser 1

ATLAS Data Flow into Analysis RAW detector data and simulated RDO data are reconstructed into our xaod format Called xaod to make it clear that it has our Run 2 EDM From these xaod events a number of derived formats are made, called DxAODs There are about 100 of these derived formats, each targeting a particular group of analyses, each targeting 1% of the total size of the primary xaods 2

xaod to DxAOD Production of DxAODs is done via a train model This is to amortise the costs of staging in the xaod input, and repeating analysis tasks About 14 trains are run, each producing ~7 outputs Trains are organised based on the output size of the DxAODs (merging is easier that way) Size is strongly correlated to production time; optimise delivery time to physics groups Each DxAOD is produced from the xaod applying any/all of Skimming - remove events that don t pass some selection criteria Slimming & Thinning - remove of whole/partial containers from events that are uninteresting Enhancing - add additional content to the DxAOD If multiple derivations all require the same enhancements they are done only once In fact for 2017 ATLAS is moving most jet finding and flavour tagging into the derivations step - to save space in the primary xaod and to catch final code improvements 3

Conditions and Calibration First pass calibration is done in our calibration loop of 48 hours Usual calibrations for beam spot, alignment, noisy channels, etc. derived from express stream Then bulk reconstruction proceeds to produce xaods at the Tier-0 These are lower level calibrations, stored in the database, used within Athena Refined calibrations from combined performance groups are prepared only once the bulk reconstruction is ready This is rather a long process, taking several months Serialised down the reconstruction chain (tracking, jet finding, MET, flavour tagging, ) This process is fine tuned to maximise the physics performance E.g., b-tagging requires several weeks to train their neural networks So it is out of the question to do that in any calibration loop Generally these calibrations may be used in DxAOD production itself (even applying on the fly fixes to xaod) 4

CP Calibration Storage and Access Once CP calibrations are prepared they are stored on CVMFS in a group data area Individual groups are responsible for the mechanics of versioning and ensuring that their software packages will find the correct version of the calibrations In general a set of final calibrations are provided for each major software release, with a set of pre-recommendations coming from MC and then revised once detector data arrives These calibrations are accessed both in the xaod DxAOD step, and during the DxAOD analysis The vast majority of groups use these calibrations as is In some special bleeding edge cases the CP group and the analysis group may work together on refinements, e.g, the H ZZ* 4l and egamma CP 5

Analysis Software Provision ATLAS provides two base releases to support analysis AthAnalysisBase - a stripped down build of Gaudi+Athena framework, containing only the packages needed by analysts AnalysisBase - a ROOT based analysis toolkit, with a much smaller footprint Perceived lower barrier to entry; does lack some features Of the two, AnalysisBase is by far the most popular Both releases can read the xaod format (big improvement from Run 1) 6

Analysis Software Frameworks Many analysis groups built on the AnalysisBase software Adding extra layers to apply the correct calibrations, deal with systematics and luminosity, etc. Over time these additions were generalised to more analyses and became more generic There are now a lot of these and their use is widespread Management frequently ask if these could be rationalised Despite the fact that effort is being duplicated, these are very successful and well established frameworks - the effort to rationalise them would be substantial 7

Code Management and Development Git is the future, SVN is the past Most groups do use CERN hosted services SSO integration is probably the killer feature A surprising amount of development is still lxplus based Everyone s reference platform lxplus Tier 1 / 2 Tier 3 Institute PC Personal machine other 8

Software Validation ATLAS software is well validated up to the xaod level This is physics validation, comparing new outputs with reference samples Centrally organized effort in CP and detector SW groups Every small or major change in reconstruction or simulation SW is validated with hundreds of histogram comparisons in each sub-domain Happens almost on a weekly or bi-weekly basis As the xaod content is largely what is in the DxAODs then these are implicitly validated DxAOD production is constantly monitored for performance also E.g., sudden degradation in CPU time would be noticed Analysis group validation is performed on a more ad-hoc basis Internal validations of recommendations in the analysis of recent data/mc performance Usually driven by major SW releases or conferences 9

Analysis Data Access and Tools ATLAS analysis is using C++ and Python, very little else No widespread use of non-hep tools However, there is some trend towards looking at newer machine learning techniques (i.e., beyond what is in TMVA or other traditional HEP machine learning) Such tools are usually used to prepare something that goes back into ROOT, e.g., a trained network Here there is some back and forth format translation: xaod ROOT Ntuple JSON, txt or CSV Compiled C / C++ Python / PyROOT ROOT C++ Macros other 10

Conditions and Calibration Lower level conditions needs the full Athena framework to be accessed Just because the tools that interpret the database entries only exist in Athena and rely on quite fat libraries, like COOL ATLAS are considering simplifying this for Run 3 (as a common project) Analysis calibration files are mostly plain ROOT Ntuples Accessing this data is mainly done via tools written by the CP groups themselves There is special support for these tools to ensure they can be compiled into an Athena release and into a ROOT release Insofar as they are usually quite simple data structures, reading them on your own is possible But highly discouraged 11

Analysis Bifurcation Points Almost all analysis groups pass through at least one additional format between DxAOD and final plot data Flat ntuples are still quite widely used in later stages Almost all analyses use one of the common framework layers previously shown These are common Beyond that, code usually becomes more specialised Intermediate file formats used in an analysis 12

Grid Use Almost everyone starts on the grid, or at least in some substantial private cluster That s just the reality of LHC datasets! Metadata: ATLAS Metadata Interface (AMI) used for discovery of appropriate data Container = Many Datasets Dataset = many files ATLAS tools for good run list and luminosity built into the system Really substantial amounts of data are downloaded as well ATLAS Grid Tier 2 with special rights ATLAS Tier 3 Private Tier 3 Personal machine other 13

For the Future - worries and mitigations Dramatic increases in LHC luminosity are not going to be matched by resources For analysis the time to complete one run over increasingly large datasets is only going up N.B. here we suffer more than other workflows as the IO rate per TB is decreasing (larger disks, no increase in read rate) Analysis could lose a lot of flexibility, less scope to try something different Less copies of data available mean that remote reading of data will only grow Work to do that efficiently Can we provide concurrency plugins that support analysis on modern CPUs easily? We can t expect analysis to become experts in parallel programing 14