LCG Conditions Database Project

Similar documents
LHCb Distributed Conditions Database

The ATLAS Conditions Database Model for the Muon Spectrometer

COOL performance optimization using Oracle hints

Access to ATLAS Geometry and Conditions Databases

A Tool for Conditions Tag Management in ATLAS

The CORAL Project. Dirk Düllmann for the CORAL team Open Grid Forum, Database Workshop Barcelona, 4 June 2008

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

Evolution of Database Replication Technologies for WLCG

Evolution of Database Replication Technologies for WLCG

Ideas for evolution of replication CERN

Event cataloguing and other database applications in ATLAS

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

Persistent storage of non-event data in the CMS databases

Persistent storage of non-event data in the CMS databases

Database monitoring and service validation. Dirk Duellmann CERN IT/PSS and 3D

ATLAS Oracle database applications and plans for use of the Oracle 11g enhancements

Summary of the LHC Computing Review

LHCb Conditions Database Graphical User Interface

Implementing a modular framework in a conditions database explorer for ATLAS

DIRAC File Replica and Metadata Catalog

LCG Persistency Framework (CORAL, COOL, POOL): Status and Outlook in 2012

Andrea Sciabà CERN, Switzerland

Deferred High Level Trigger in LHCb: A Boost to CPU Resource Utilization

PoS(EGICF12-EMITC2)106

DQ2 - Data distribution with DQ2 in Atlas

Streamlining CASTOR to manage the LHC data torrent

ATLAS DCS Overview SCADA Front-End I/O Applications

Evolution of ATLAS conditions data and its management for LHC Run-2

Oracle at CERN CERN openlab summer students programme 2011

DIRAC data management: consistency, integrity and coherence of data

First Experience with LCG. Board of Sponsors 3 rd April 2009

Large Scale Software Building with CMake in ATLAS

LHCb Computing Status. Andrei Tsaregorodtsev CPPM

ELFms industrialisation plans

Prompt data reconstruction at the ATLAS experiment

Optimizing Parallel Access to the BaBar Database System Using CORBA Servers

The ATLAS Distributed Analysis System

First experiences with the ATLAS pixel detector control system at the combined test beam 2004

Virtualizing a Batch. University Grid Center

Servicing HEP experiments with a complete set of ready integreated and configured common software components

The Run 2 ATLAS Analysis Event Data Model

Distributing BaBar Data using the Storage Resource Broker (SRB)

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

Data Warehousing & Big Data at OpenWorld for your smartphone

CERN s Business Computing

CSE 444: Database Internals. Lectures 26 NoSQL: Extensible Record Stores

CouchDB-based system for data management in a Grid environment Implementation and Experience

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

AMGA metadata catalogue system

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

References. What is Bigtable? Bigtable Data Model. Outline. Key Features. CSE 444: Database Internals

Evolution of Cloud Computing in ATLAS

CHEP 2013 October Amsterdam K De D Golubkov A Klimentov M Potekhin A Vaniachine

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

GoldenGate Zbigniew Baranowski

Spread the Database Love with Heterogeneous Replication. MC Brown, VP, Products

Frank Gaede, DESY, LCWS 2005, Stanford, CA March Overview of Simulation and Reconstruction Tools in Europe

The LHC Computing Grid

Experience with Data-flow, DQM and Analysis of TIF Data

The CMS Computing Model

Manual Trigger Sql Server 2008 Insert Multiple Rows At Once

The ATLAS Production System

The BABAR Database: Challenges, Trends and Projections

Experiences with the new ATLAS Distributed Data Management System

The Google File System

150 million sensors deliver data. 40 million times per second

Understanding StoRM: from introduction to internals

AGIS: The ATLAS Grid Information System

Experience with the Open Source based implementation for ATLAS Conditions Data Management System

LHCb Computing Technical Design Report

Event visualisation for the ATLAS experiment - the technologies involved

Front-End Electronics Configuration System for CMS. Philippe Gras CERN - University of Karlsruhe

Overview of ATLAS PanDA Workload Management

A never-ending database migration

GAUDI - The Software Architecture and Framework for building LHCb data processing applications. Marco Cattaneo, CERN February 2000

Data services for LHC computing

Manual Trigger Sql Server 2008 Insert Update Delete

Tests of PROOF-on-Demand with ATLAS Prodsys2 and first experience with HTTP federation

INSPIRE & Environment Data in the EU

Course Syllabus: In-Depth unipaas Programming Techniques

Newly-Created, Work-in-Progress (WIP), Approval Cycle, Approved or Copied-from-Previously-Approved, Work-in-Progress (WIP), Approval Cycle, Approved

LHCb Computing Strategy

ANALYSIS OF THE FEATURES OF THE OPTIMAL LOGICAL STRUCTURE OF DISTRIBUTED DATABASES

Data Management for the World s Largest Machine

Sql Server 2005 Create Script To Copy Database Schema And All The Objects

Batch Services at CERN: Status and Future Evolution

CernVM-FS beyond LHC computing

The ATLAS Trigger Simulation with Legacy Software

The LHCb Alignment Framework

How the Monte Carlo production of a wide variety of different samples is centrally handled in the LHCb experiment

Heterogeneous Relational Databases for a Grid-enabled Analysis Environment

Oracle 11g Partitioning new features and ILM

Oracle and.net Introduction and What s New. Copyright 2017, Oracle and/or its affiliates. All rights reserved.

CMS - HLT Configuration Management System

CMS conditions database web application service

Grid Computing Activities at KIT

RADU POPESCU IMPROVING THE WRITE SCALABILITY OF THE CERNVM FILE SYSTEM WITH ERLANG/OTP

Computing at Belle II

THE ATLAS DISTRIBUTED DATA MANAGEMENT SYSTEM & DATABASES

Software installation and condition data distribution via CernVM File System in ATLAS

Transcription:

Computing in High Energy and Nuclear Physics (CHEP 2006) TIFR, Mumbai, 13 Feb 2006 LCG Conditions Database Project COOL Development and Deployment: Status and Plans On behalf of the COOL team (A.V., D.Front, M.Clemencic, S.A.Schmidt, U.Moosbrugger)

What are conditions data? Outline Background to COOL COOL software overview and status Development and deployment perspectives 2 COOL Status and Plans

What are conditions data? Non event detector data that vary with time And may also exist in different versions Data producers from both online and offline Geometry, readout, slow control, alignment, calibration... Data used for event processing and more Detector experts Alignment and calibration Event reconstruction and analysis 3 COOL Status and Plans

Prehistory Background of common activities in 2000 2003 Collaboration of IT DB, LHCb, Atlas, COMPASS, Harp Borrowing a few design ideas from the BaBar experience C++ API definition and Objectivity implementation Oracle implementation of original API ( BLOB payload) MySQL implementation of extended API (flexible payload) LCG Conditions DB project launched in 2003 Subproject of LCG Persistency Framework (POOL) 4 COOL Status and Plans

2003 until CHEP 2004 Two parallel activities (common project mandate) Integrate existing Oracle/MySQL packages into LCG AA Review old software and API to plan new developments Project faced two main problems in this phase Lack of manpower for new developments Divergence of the two packages Decision to develop new software just after CHEP Following public AA Meeting discussion in October Development of COOL started in November 2004 5 COOL Status and Plans

Sub detector #1 Experiment framework Sub detector #2 Sub detector #N Experiment conditions data common software and conventions Common project scope NOT the problems specific to one experiment or one data type (handled by each experiment) Conditions DB Access (COOL) Time varying single version data (~online) Time varying multi version data (~offline) Software for time varying and versioned data: a common component with a well defined task (RDBMS implementation of technology neutral API ) C++ Relational Access (CORAL) Oracle Access MySQL Access SQLite Access NOT the generic C++ access to relational data (handled by CORAL) Oracle OCI MyODBC API SQLite API Oracle DB MySQL DB SQLite DB Relational Database Deployment and Data Distribution (3D) NOT the generic deployment of relational services and distribution of relational data (handled by 3D at CERN by IT PSS) 6 COOL Status and Plans

Conditions data: metadata model Designed to handle data objects that Can be classified into independent data items VARY WITH TIME May have different versions (for given time and data item) A CondDBObject has Metadata: Data item identifier Interval of validity [since, until] Version information Data payload : Actual data variables (temperatures, calibration parameters ) Separate fields or encoded as a LOB This 3 D metadata model is still valid! A CondDBObject Pere Mato (Feb 2000) Main use case: fetch data valid at given time and tag Inverse lookup (from temperature to time or version) is not a priority 7 COOL Status and Plans

COOL software overview Merge best ideas of two previous packages New API enhances both original and extended APIs Single implementation for many relational backends Encapsulated behind C++ API (no direct SQL user access) Support for Oracle, MySQL and SQLite via CORAL Attention to Oracle performance (bulk operations, bind variables ) Maximize integration with other LCG projects Reuse infrastructure and software (SPI, SEAL, POOL ) Mutually beneficial collaboration with CORAL team 8 COOL Status and Plans

COOL relational implementation Modeling of condition data objects System managed common metadata Data items: many tables ( folders ), each with many channels Interval of validity IOV: since, until Versioning information with handling of interval overlaps User defined schema for data payload Support for simple C++ types as CORAL AttributeList objectid channelid since until pressure temperature Metadata System controlled (versioning metadata not shown) Data payload User defined schema (different tables for different schemas) 9 COOL Status and Plans

Milestones Nov 2004: start of COOL software development Brand new code, merge ideas of two pre existing packages Initially ~1.3 FTE for development (A.V. and S.A.S.) Apr 2005: first COOL production release 1.0.0 Support for Oracle and MySQL through POOL RAL Basic insertion/retrieval (single/multi version, single/bulk) Oct 2005: Atlas use case performance validation One job every 5s, each retrieving 100 MB in 100k rows Latest of many releases: COOL 1.2.8 (Jan 2006) SQLite (July), CLOB, PyCool, multi channel bulk ops (Aug.), performance tests (Oct.), data copy (Nov.), CORAL (Jan.) Team has grown to ~3.5 FTE (IT/LCG, Atlas and LHCb) 10 COOL Status and Plans

Future perspectives Software consolidation and enhancements Major API and schema changes in 1.3.0 (~Apr 2006) Implement new features following experiment priorities Support real life deployment before LHC startup Continue to test and improve Tier0 software performance Support Atlas and LHCb in setting up distributed services Collaboration with 3D and IT PSS service teams is crucial 11 COOL Status and Plans

12 COOL Status and Plans Software enhancement plans Next on the list (COOL 1.3.0 and later) AMD64 port and storage precision in API (int32 vs. int64) Add table with channel metadata ; schema evolution tools Improve versioning: user tags (later: HVS hierarchical tags) Later: add CORAL monitoring/authentication/indirection More requests for new functionalities are pending Often received at weekly phone meetings Handled according to experiment priorities and available manpower Formal review later on when API is more stable

13 COOL Status and Plans Atlas (offline) deployment COOL fully integrated into Athena since mid 2005 Small payload stored inline, complex payload as POOL refs Development priorities: CORAL API, schema evolution, HVS Most data still in Lisbon MySQL implementation Transition phase: complete migration to COOL by mid 2006 Deployment priorities Commissioning (now); simulation (Apr); reconstruction (Oct) Static replication now (Oracle >SQLite) Explore T0 T1 dynamic replication via Oracle streams in 3D Interest in distributed Oracle access via Frontier too

14 COOL Status and Plans LHCb deployment Conditions DB (COOL) one of many databases COOL holds conditions data for reconstruction/analysis Other data in PVSS, LFC, Bookkeeping, Configuration DBs Deployment model (online and offline) Masters (r/w) at the pit and CERN T0 Replicas (r/o) at T1 Oracle replication via Oracle streams Data challenge plans in 2006 Alignment/calibration challenge in Oct (with all T1 sites)

15 COOL Status and Plans Summary COOL software is recent but of production quality Software developments started in fall 2004 after CHEP04 Manpower has increased from ~1.3 to ~3.5 FTE Single implementation for Oracle, MySQL and SQLite Tight integration with other LCG projects Mutually beneficial collaboration with CORAL project Service integration with IT PSS at CERN and 3D project Focus moving from development to deployment Development of new functionalities is far from finished Deployment is progressing fast in Atlas and LHCb

For more information LCG Conditions Database Project Web page http://lcgapp.cern.ch/project/conddb Related talks and posters at this conference COOL performance and distribution tests (A. Valassi) CORAL relational database access software (I. Papadopoulos) POOL object persistency into relational databases (G. Govi) Software for a variable Atlas detector description (V. Tsulaia) LHCb conditions database framework (M. Clemencic) Database access in Atlas computing model (S. Vaniachine) 16 COOL Status and Plans