Data Flow & Leve1 1 Pipeline

Similar documents
Gamma-ray Large Area Space Telescope. Work Breakdown Structure

GLAST Large Area Telescope:

I&T&C Organization Chart

GLAST Large Area Telescope:

Subsystem Development. T&DF Development. LAT Flight Software Testbed. LAT Integration & Test. Mission Systems Integration & Test EGSE TKR EM1 LOF

Electrical Ground Support Equipment Overview

GLAST Large Area Telescope Instrument Science Operations Center

GLAST Large Area Telescope:

Experiences From The Fermi Data Archive. Dr. Thomas Stephens Wyle IS/Fermi Science Support Center

GLAST Large Area Telescope

Electrical Ground Support Equipment Verification Test Support. EGSE Verification Test Support

Planning Production Monitoring

Transient Data Store and Persistent Data

GLAST Large Area Telescope Instrument Science Operations Center

The GLAST Event Reconstruction: What s in it for DC-1?

Topics for the TKR Software Review Tracy Usher, Leon Rochester

LAT Flight Software. End-To-End Testing

Gamma-ray Large Area Space Telescope (GLAST) Large Area Telescope (LAT) Integration and Test Subsystem Electrical Performance Tests

Lunar Reconnaissance Orbiter (LRO)

The use of HepRep in GLAST

CRaTER Scence Operation Center Requirements Document. Dwg. No

Science Analysis Software Overview

Existing Tools in HEP and Particle Astrophysics

Simulation of GLAST. Alessandro de Angelis. SLAC, February 21, University of Udine and INFN Trieste

GLAST tracking reconstruction Status Report

GLAST Large Area Telescope:

Preliminary results in an ongoing study.

GLAST Large Area Telescope Instrument Science Operations Center

Wide-field Infrared Survey Explorer (WISE) WISE Mission Operations System Preliminary Design Review. WSDC Architecture

GLAST Large Area Telescope Instrument Science Operations Center

The CMS Computing Model

WISE Science Data Center

Test Beam PSF Analysis. PSF Analysis. Brian Baughman, UCSC Jose-Angel Hernando, UCSC

The Use of HepRep in GLAST

AIA Data processing and Distribution: from Telemetry to Science data

HiRISE Operations Center Overview. Eric Eliason Co-Investigator and Ground System Manager October 13, 2004

Hall D and IT. at Internal Review of IT in the 12 GeV Era. Mark M. Ito. May 20, Hall D. Hall D and IT. M. Ito. Introduction.

Performance quality monitoring system (PQM) for the Daya Bay experiment

LHCb Computing Status. Andrei Tsaregorodtsev CPPM

WSDC Hardware Architecture

A Prototype of the CMS Object Oriented Reconstruction and Analysis Framework for the Beam Test Data

ALICE ANALYSIS PRESERVATION. Mihaela Gheata DASPOS/DPHEP7 workshop

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

PRISMA - Ground Segment. Roma, TPZ-Roma-DG, 02/03/2017

PN treatment in ESAS - SAS validation - Calibration files and Configuration Control

Distributed Archive System for the Cherenkov Telescope Array

EventStore A Data Management System

S&OC System Requirements Review: GSRD Traceability Matrix

Event Displays and LArg

SDO EVE Data Processing Overview & Data Access

Computing. DOE Program Review SLAC. Rainer Bartoldus. Breakout Session 3 June BaBar Deputy Computing Coordinator

DVCS software and analysis tutorial

CMS Simulation Software

CLAS12 DAQ, Trigger and Online Computing Requirements. Sergey Boyarinov Sep 25, 2017

Optimizing Parallel Access to the BaBar Database System Using CORBA Servers

Polarization Ray Trace, Radiometric Analysis, and Calibration

Physics and Detector Simulations. Norman Graf (SLAC) 2nd ECFA/DESY Workshop September 24, 2000

Databases in the Cloud

CRaTER SOC Operations

Track pattern-recognition on GPGPUs in the LHCb experiment

GLAST Large Area Telescope

WISE Data Processing Overview

Processing System Hardware

Power Distribution Unit (PDU) Specification - Level V Specification

The Belle II Software From Detector Signals to Physics Results

I&T&C Organization Chart

GLAST Large Area Telescope: I & T

Data and Analysis preservation in LHCb

The ATLAS EventIndex: Full chain deployment and first operation

Performance quality monitoring system for the Daya Bay reactor neutrino experiment

Progress on G4 FDIRC Simulation. Doug Roberts University of Maryland

4-3 Telemetry and Command Processing System for Experiments

Science Data Processing and Data Products

Performance of the GlueX Detector Systems

Security, Monitoring, and Control of the Re-engineered Hubble Space Telescope Control Center System

A DTN-Based Multiple Access Fast Forward Serivce for the NASA Space Network

LHCb Computing Strategy

GLAST Silicon Microstrip Tracker Status

Geant4 activities at DESY

LHCb Computing Resources: 2018 requests and preview of 2019 requests

Implementing Onboard Data Compression for the Van Allen Probes

Data Analysis in ATLAS. Graeme Stewart with thanks to Attila Krasznahorkay and Johannes Elmsheuser

Monte Carlo programs

Current Joint Polar Satellite System (JPSS) Common Ground System (CGS) Technical Performance Measures (TPMs)

Dynamics 365. for Finance and Operations, Enterprise edition (onpremises) system requirements

CT Systems and their standards

Performance of the ATLAS Inner Detector at the LHC

FACT-Tools. Processing High- Volume Telescope Data. Jens Buß, Christian Bockermann, Kai Brügge, Maximilian Nöthe. for the FACT Collaboration

The BaBar Computing Model *

National Aeronautics and Space Administration Jet Propulsion Laboratory California Institute of Technology WISE Archive.

Belle II Software and Computing

Netherlands Institute for Radio Astronomy. May 18th, 2009 Hanno Holties

Implementation of a Middleware Based Ground System March 2, 2005, GSAW2005 Conference

ATLAS Tracking Detector Upgrade studies using the Fast Simulation Engine

SharePoint 2010 Technical Case Study: Microsoft SharePoint Server 2010 Enterprise Intranet Collaboration Environment

FAMOS: A Dynamically Configurable System for Fast Simulation and Reconstruction for CMS

Space Radiation and Plasma Environment Monitoring Workshop 13 and 14 May ESTEC. Efacec Space Radiation Monitors MFS & BERM

Time Domain Alerts from LSST & ZTF

SCHEDULE AND TIMELINE

CLAS12 DAQ & Trigger Status and Timeline. Sergey Boyarinov Oct 3, 2017

Transcription:

Data Flow & Leve1 1 Pipeline High level specs in L3 & L4 documents: Level 4 - LAT-SS-00505-01 Level 3 - LAT-SS-00020-01 Pipeline Server Implementation Plan - LAT-TD-00773-01 (draft in review) Database Specification - LAT-TD-00553-01 Quick summary of Pipeline scope Process real data EM?, CU, Flight Generate bulk MC Facilitate calibrations (monitoring and specialty datasets) Near real time diagnostics from Level 1 chain Fully automated Quick turnaround R.Dubois 1

Mission Operations Architecture TDRS Alerts GLAST LAT Instrument Operations Center Source: Scott Williams Two Contacts Per day Routine Data LAT data handling LAT performance Standard product processing Alerts Large loads Target of Opportunity Commands Malindi Spacecraft, IOC and GBM data Mission Ops Center Observatory safety Spacecraft health Commanding Mission scheduling Instrument data handling Schedules Level 0 data for archiving LAT Data Command Loads Status Science Support Center Science scheduling Archiving Guest Observer Support Standard product processing Standard products Schedule requests Standard products Schedule requests GRB Coordinates Network Burst and transient Alerts GBM Data Status Command Loads GBM Instrument Operations Center GBM data handling GBM performance Standard product processing R.Dubois 2

LAT Operations Facility Functions The LOF will consist of about 1/2 the resource of the FUSE Control Center at Johns Hopkins University (as shown here). Mission Operations Center (MOC) S/C LAT, and GBM Data, Commands Space and Ground Segments LAT Instrument Operations & Science Data Processing LAT Operations Facility WBS 4.1.B S. Williams, SU-HEPL LAT Data Verification LAT Health & Safety Monitoring LAT Commanding Test & Calibration Data Acquisition Validating & Maintaining Flight Software Alert Processing Level 0 Data, LAT Procs & Uploads Science Plan, Schedules, Level 0 Data Level 0 Data, GBM Procs & Uploads Level 0 Science & Hsk Data Performance & Cal Data Science Plan, LAT Schedules Science Support Center (SSC) Level 1 Data, High Level Products, GBM IOC Data Products, GBM Schedules, Science Plan GBM Instrument Operations Center R.Dubois 3 Data Processing Facility WBS 4.1.D R. Dubois, SU-SLAC Science Data Processing Optimizing Analysis & Processing Algorithms LAT Calibration LAT Performance Assessment Data Distribution Analysis Software Mirror Sites Level 1 Data, High Level Products, LAT IOC Data Products Source: Scott Williams

LAT Data Path Realtime data: 32 kbps S-band SSR dump: 10 to 150 Mbps X-band Source: Scott Williams Malindi Prime - 2 Mbps Intelsat Backup - 500 kbps ASINet Central Node Fucino, Italy 32 Mbps JSC Houston Malindi gets ~35 Gbits per day from GLAST and 26 Gbits From Swift, and AGILE Equivalent to ~400 kbps continuous Realtime S-Band Data and SSR Data MOC (TBD) NISN or ASINet NISN or Internet 2 via JPL LAT IOC Stanford NISN should support whatever rates we deem necessary for the desired latency Another option is Internet 2 NISN or Internet 2 via ARC Realtime S/C and LAT HSK Data Followed by processed Level 0 Data R.Dubois 4

Data Flow Data recon + MC on disk. Abstract fullrecon output into L1 DB for analysis MOC MOC DPF DPF MC Recon Calibs Italian mirror French mirror IOC IOC L1 DB Fully automated server, with RDB for data catalogue + processing state. Uses SLAC batch CPU and disk farms. L2 DB Parts of L2 processing also automated SSC SSC R.Dubois 5

Instrument Simulations and Reconstruction 3 GeV gamma interaction Source Fluxes Particle Transport Instrument data Gleam Raw Data Recon 3 GeV gamma recon Geometry Background Rejection - Particle ID CAL Detail Instrumental Calibrations R.Dubois 6

What is GLEAM? Already described by Richard, see his talk: this talk will fill in a few details. The name (originally an acronym by L. Rochester) for an application that does everything: Manage an (optional) GUI interface, with a 3-D display Manage the detector description, including full geometry Manage a catalog of particle sources, ranging from simple test-beams to GRBs; handle geometry of the orbit. Run the Geant4 simulation engine to trace particle trajectories, simulating interactions, scattering, decays, etc, and record deposited energy in sensitive detectors ( hits ) Manage conversion of hits to digis, corresponding exactly to raw data collected by the detector Construct a trigger from the deposited information Run reconstruction algorithms to analyze strips, CsI xtal info, and ACD tiles. Write out the hits, digis, and recon to ROOT files for further analysis Read digis at least to avoid re-running simulation Summarize the track fit results as a PSF and Aeff Source: Toby Burnett R.Dubois 7

Proof! incoming photon LAT Spacecraft Solar panels Source: Toby Burnett R.Dubois 8

Proof, cont. Incoming 100- MeV photon Source: Toby Burnett reconstructed photon direction G4 pair converts in in a thick W layer Cal response, recon W strip energy Si R.Dubois 9

Gleam Data Products Current Persistent Store definition will evolve with more experience MC Particle tree, Truth Hits in ACD, CAL, TKR Digitization raw data from TKR, CAL, ACD Level 0.25 Trigger word Recon ACD Relate tracks to hit tile(s) CAL Output of clustering Position, direction from the one cluster Level 0.5 Total observed and corrected energies leakage TKR Clusters of strips Pattern recognition, Fit track candidates vertices R.Dubois 10

Upcoming Gleam Data Products Modifying event chain to include Event Interpretation To be initiated this summer Take subsystem recons as input standard analysis combine TKR vertices, ACD hit tiles, CAL energy/direction to synthesize event Apply background and quality criteria to classify event Full output will go into Level 0.5 structure Extract summary info for Level 1 DB summary ntuple Probably not the same as the Level 0.5 ntuple used for instrument studies Housekeeping to accompany science data has yet to be defined. Still need to deliver header info for each dataset R.Dubois 11

Numerology from pdrapp predecessor to Gleam used for PDR performance estimate. will have revised numbers from Gleam this summer with iterated structures Downlink = 3 GB/day @ 2.2M events : raw data ~ 1.5 kb/event MC simulation takes 0.16 sec/event for all_gamma Recon takes 0.08 sec/event MC sim generates 1.5 kb/event Recon generates 4 kb/event Assume equal MC generation to downlinked data rate during flight operations A year s re-reconstruction would add 4 TB and 100 CPU-weeks to the load negotiate with SLAC Computing Center for additional resources when needed an advantage of tapping into a large batch farm! R.Dubois 12

Processing Requirements - Disk 10000 9000 8000 flight Disk Usage (GB) 8.5 TB/year; < 60 TB total Raw+Recon+MC GB 7000 6000 5000 Mock Data Challenges, Calib, I&T 4000 3000 2000 balloon Perf studies 1000 0 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011 R.Dubois 13 Year (note: BABAR temp space is 25 TB)

Processing Requirements - CPU 450 400 350 300 Mock Data Challenges, Calib, I&T CPU weeks vs time flight 8 CPUs dedicated Raw+Recon+MC CPU (weeks) 250 200 150 100 balloon Perf studies 50 0 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011 Time (SLAC will have ~2000 CPUs for BABAR) R.Dubois 14

Additional Pipeline Products Strip files (or tagged events in Level 0.5 DB) for low level calibration (peds, gains etc) Select CNO etc for subsystem calibrations Tag random-trigger events Pipeline could run calibration algorithms if useful; humans will have to check anyhow before results put back into the system. Near real-time monitoring Run system test diagnostics and feed back to IOC operations Use full power of correlated analysis to identify problems Will take form of plots and statistics, tracked in pipeline database Keep track of Livetime Correlate livetime counters with operational modes and generate history Follow up onboard alerts; look for new ones R.Dubois 15

Processing Pipeline IOC IOC Level 0 Batch system HSM Automated Tape Archive Level 0 Level 1, diagnostics ~50 ~50 CPUs; CPUs; ~50 ~50 TB TB disk disk by by 2010 2010 Plenty of margin on CPUs R.Dubois 16

LAT Data Processing Database Heart of data processing facility is a database to handle state of processing, as well as an automated server. relational database tracks state of file based datasets throughout lifetime in the system, from arrival at IOC or MC generation, through Level 1 output. Automated server will poll IOC generated database entries for new Level 0 datasets and take immediate action, as well as generate MC data, and log all actions to the database. R.Dubois 17

Prototype DB Table Layout LAT-TD-00553-01 if you re interested read this R.Dubois 18

Data Manager Prototype Existing Data Manager Prototype is set of perl scripts that: Performs automated MC batch processing using LSF and SLAC batch farm (e.g. produced 50 M background events, 10 M gammas for PDR studies) Provides utilities for processing, filtering, and displaying results of MC runs Provides very preliminary scripts for entering results of MC runs into Oracle tables Will evolve into Data Manager for Data Processing Facility (DPF) by being split into a server and set of utility packages as described in Data Manager spec R.Dubois 19

Key Pipeline Concepts Media Everything on disk Input to the Pipeline FITS files appear on disk; tagged in DB The Processing Chain Elements configured in the DB Customize to task (real data, MC, etc) Communication with Processes Interface hiding batch system details so that a non-slac implementation can use a different mechanism Communication with DB Server does all communicating with DB, which is standard SQL Communication with Server CLI, gui Authenticate users LAT-TD-00773-01 (in prep) Archive all files to tape R.Dubois 20

Additional Details Current plan is to export Level 1 db to SSC Several ways to mirror Haven t started talking to LAT mirrors yet still early in the game Pipeline database knows about all datasets, so any mirroring should be easy DB keeps track of all details of code used to do processing + parameters used Can also be used for Level 2 automated processing Planning & Implementation DB spec and Implementation plan to be reviewed by LAT-SSC working group Alex Schlessinger targeted to implement pipeline Starting to think of implementation details already Try to have something ready for EM and for CDR MC generation On the TODO list: handle burst alerts & tracking livetime R.Dubois 21