Data services for LHC computing

Similar documents
Analytics Platform for ATLAS Computing Services

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

CERN Tape Archive (CTA) :

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

CERN European Organization for Nuclear Research, 1211 Geneva, CH

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

Distributing storage of LHC data - in the nordic countries

Streamlining CASTOR to manage the LHC data torrent

Storage Virtualization. Eric Yen Academia Sinica Grid Computing Centre (ASGC) Taiwan

Visita delegazione ditte italiane

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Data Movement & Tiering with DMF 7

Past. Inputs: Various ATLAS ADC weekly s Next: TIM wkshop in Glasgow, 6-10/06

Overview. Jakub T. Mościcki, IT/DSS. Meeting with Palestinian fellows

ATLAS Distributed Computing Experience and Performance During the LHC Run-2

CernVM-FS beyond LHC computing

Big Data Analytics and the LHC

The CMS Computing Model

Data Transfers Between LHC Grid Sites Dorian Kcira

WLCG Transfers Dashboard: a Unified Monitoring Tool for Heterogeneous Data Transfers.

Summary of the LHC Computing Review

Database Services at CERN with Oracle 10g RAC and ASM on Commodity HW

New strategies of the LHC experiments to meet the computing requirements of the HL-LHC era

UW-ATLAS Experiences with Condor

Andrea Sciabà CERN, Switzerland

ANSE: Advanced Network Services for [LHC] Experiments

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

LHCb Computing Strategy

Preparing for High-Luminosity LHC. Bob Jones CERN Bob.Jones <at> cern.ch

Considerations for a grid-based Physics Analysis Facility. Dietrich Liko

Towards Network Awareness in LHC Computing

Introducing SUSE Enterprise Storage 5

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN

CouchDB-based system for data management in a Grid environment Implementation and Experience

THE ATLAS DISTRIBUTED DATA MANAGEMENT SYSTEM & DATABASES

Storage for HPC, HPDA and Machine Learning (ML)

Constant monitoring of multi-site network connectivity at the Tokyo Tier2 center

Storage on the Lunatic Fringe. Thomas M. Ruwart University of Minnesota Digital Technology Center Intelligent Storage Consortium

LHCb Computing Resources: 2018 requests and preview of 2019 requests

Evolution of Cloud Computing in ATLAS

High-Energy Physics Data-Storage Challenges

Online data storage service strategy for the CERN computer Centre G. Cancio, D. Duellmann, M. Lamanna, A. Pace CERN, Geneva, Switzerland

Reliability Engineering Analysis of ATLAS Data Reprocessing Campaigns

Insight: that s for NSA Decision making: that s for Google, Facebook. so they find the best way to push out adds and products

LHCb Computing Resource usage in 2017

150 million sensors deliver data. 40 million times per second

LHC and LSST Use Cases

THE EMC ISILON STORY. Big Data In The Enterprise. Deya Bassiouni Isilon Regional Sales Manager Emerging Africa, Egypt & Lebanon.

LHCb Computing Resources: 2019 requests and reassessment of 2018 requests

1. Introduction. Outline

The INFN Tier1. 1. INFN-CNAF, Italy

Singularity in CMS. Over a million containers served

Evolution of Rack Scale Architecture Storage

Evolution of the HEP Content Distribution Network. Dave Dykstra CernVM Workshop 6 June 2016

Computing / The DESY Grid Center

Experiences with the new ATLAS Distributed Data Management System

Grid Computing Activities at KIT

Status of KISTI Tier2 Center for ALICE

RADU POPESCU IMPROVING THE WRITE SCALABILITY OF THE CERNVM FILE SYSTEM WITH ERLANG/OTP

PoS(EGICF12-EMITC2)106

ALHAD G. APTE, BARC 2nd GARUDA PARTNERS MEET ON 15th & 16th SEPT. 2006

DESY at the LHC. Klaus Mőnig. On behalf of the ATLAS, CMS and the Grid/Tier2 communities

ATLAS Experiment and GCE

STATUS OF PLANS TO USE CONTAINERS IN THE WORLDWIDE LHC COMPUTING GRID

The ATLAS Tier-3 in Geneva and the Trigger Development Facility

Federated data storage system prototype for LHC experiments and data intensive science

Federated Data Storage System Prototype based on dcache

Data Management. Parallel Filesystems. Dr David Henty HPC Training and Support

RESEARCH DATA DEPOT AT PURDUE UNIVERSITY

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

Big Data Tools as Applied to ATLAS Event Data

Striped Data Server for Scalable Parallel Data Analysis

Efficient HTTP based I/O on very large datasets for high performance computing with the Libdavix library

Opportunities for container environments on Cray XC30 with GPU devices

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 21 July 2011

Using CernVM-FS to deploy Euclid processing S/W on Science Data Centres

Leveraging Software-Defined Storage to Meet Today and Tomorrow s Infrastructure Demands

Data Movement & Storage Using the Data Capacitor Filesystem

Data Access and Data Management

The Software Defined Online Storage System at the GridKa WLCG Tier-1 Center

Isilon: Raising The Bar On Performance & Archive Use Cases. John Har Solutions Product Manager Unstructured Data Storage Team

The JINR Tier1 Site Simulation for Research and Development Purposes

Monitoring system for geographically distributed datacenters based on Openstack. Gioacchino Vino

Storage and Storage Access

Recent Developments in the CernVM-FS Server Backend

Security in the CernVM File System and the Frontier Distributed Database Caching System

CERN Open Data and Data Analysis Knowledge Preservation

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

Data Analysis in ATLAS. Graeme Stewart with thanks to Attila Krasznahorkay and Johannes Elmsheuser

Costefficient Storage with Dataprotection

SLATE. Services Layer at the Edge. First Meeting of the National Research Platform Montana State University August 7-8, 2017

Experience with Data-flow, DQM and Analysis of TIF Data

Data transfer over the wide area network with a large round trip time

THOUGHTS ON SDN IN DATA INTENSIVE SCIENCE APPLICATIONS

Cluster Setup and Distributed File System

Storage Resource Sharing with CASTOR.

Exam : Title : Storage Sales V2. Version : Demo

C3PO - A Dynamic Data Placement Agent for ATLAS Distributed Data Management

Transcription:

Data services for LHC computing SLAC 1 Xavier Espinal on behalf of IT/ST

DAQ to CC 8GB/s+4xReco Hot files Reliable Fast Processing DAQ Feedback loop WAN aware Tier-1/2 replica, multi-site High throughout to tape 350+MB/s/drive - 12GB/s Pb-Pb back-up Filesystem feeling $HOME, SW-dist, Data disk and gc? Few fast streams CDR 2x40Gbps Endpoint Mounts ie. /atlas in the WNs Non-LHC and Local Less structured, small communities Unexpected usage Catalogue=Namespace 2 Consistent Many slow clients Repro, reco, analysis constant >20k

cernbox AFS cvmfs NFS RBD S3 3

cernbox AFS cvmfs NFS RBD S3 4

Evolved to Tape oriented system Key feature Per stream speed Biggest physics-repo worldwide 175PB and +500M files Towards a pluggable tape backend (EOS) Cold by definition: hight throughput, high latency Tape best technology for data repositories: TCO media power density and resilient/reliable very large disk caches nowadays Total LHC data:130 PB 11PB in July LHC Raw Data Recording (1bin/1month) #tapes 5

CERN Tape Archive Biggest physics-repo worldwide 175PB and +500M files Towards a pluggable tape backend (EOS) Cold by definition: hight throughput, high latency Tape best technology for data repositories: TCO media power density and resilient/reliable very large disk caches nowadays Total LHC data:130 PB 11PB in July LHC Raw Data Recording (1bin/1month) #tapes 6

EOS now October 2016 +1200 +45000 850M 150PB files EB era Easily scalable (#disk #servers) Performant and manageable LHC Main storage platform 7

EOS now October 2016 +1200 +45000 850M 150PB files EB era Easily scalable (#disk #servers) Performant and manageable LHC Main storage platform 8

EOS now October 2016 +1200 +45000 850M 150PB files EB era Easily scalable (#disk #servers) Performant and manageable LHC Main storage platform 9

EOS now October 2016 +1200 +45000 850M 150PB files EB era Easily scalable (#disk #servers) Performant and manageable LHC Main storage platform 10

EOS now October 2016 +1200 +45000 850M 150PB files EB era Easily scalable (#disk #servers) Performant and manageable LHC Main storage platform 11

EOS now October 2016 +1200 +45000 850M 150PB files EB era Easily scalable (#disk #servers) Performant and manageable LHC Main storage platform 12

CERN made for LHC experiments needs, but Adaptable Catering with different uses Data processing User Analysis LHC Data Recording Sync&Share Community storage CERNBOX Collaborate Share Offline work Sync SWAN (Jupyter) Users 6000 (Δ 7d =60) #files 110M (Δ 7d =250K) #dirs 14M Quota 1TB/user Used Space 240TB Deployed Space 1.5PB 13 lmascetti

Goals Make data access easy Make analysis simple Facilitate Science My Laptop Small scale analysis Test jobs AFS $home batch/interactive services Large scale experiment processing User extensive analysis protocols (xrdcp,rfio,*) /cvmfs Data Access Main experiment data repositories 14

Goals Make data access easy Make analysis simple Facilitate Science batch/interactive services Large scale experiment processing User extensive analysis My Laptop Small scale analysis Test jobs Mounts squids /cvmfs/athena fuse /mycernbox AFS $home Data Access Main experiment data repositories fuse /eos/atlas 15

Goals Make data access easy Make analysis simple Facilitate Science My Laptop Small scale analysis Test jobs Data Access Main experiment data repositories batch/interactive services Large scale experiment processing User extensive analysis Mounts squids /cvmfs/athena fuse /mycernbox fuse /eos/atlas EOS CERNBOX does your files /cernbox/jdoe EOS experiment does big data /eos/lhcb Different QoS, different patterns, overlaps 16

Goals Make data access easy Make analysis simple Facilitate Science Physicist code: topmass.kumac on his laptop on /mycernbox and sync d via cernbox client Physicist identify an interesting dataset /eos/atlas/phys-top goldenrun052014 Submit jobs to lxbatch/wlcg to process the data EOS Fuse: /eos/atlas/phys-top EOS Fuse: /mycernbox/topmass.kumac Experiment SW: /cvmfs/athena Results (ntuples) aggregated on /mycernbox/topmass are synced on his laptop as the if desired jobs are being completed Working on final plots on his laptop and Latex-ing the paper directly on /mycernnbox/topmass/paper Share on-the-fly: n-tuples Final plots Publication via /mycernbox is the enabling technology binding all this Multi QoS Access patterns Protocols Redundancy 17

Goals Make data access easy Make analysis simple Facilitate Science Interfaced CERN storage services with Blue Waters NCSA using WLCG s FTS3 to manage the data workflow Open door for HPC environments to link with our HTC and Distributed Computing expertise 18

Goals Make data access easy Make analysis simple Facilitate Science 19

Pushing boundaries Raw data recording for Jan to mid-august 20

Pushing boundaries LHC running at full speed before ICHEP 2016, unprecedent amounts of data 10PB 5PB Raw data recording per month in 2016 500 250 TB TB Raw data recording for June and July, per day 21

Pushing boundaries LHC running at full speed before ICHEP 2016, unprecedent amounts of data Systems reaching limits and not exporting data fast enough from Tier-0 to WLCG 2PB 22

Raw data recording for Jan to mid-august 23

Raw data recording for Jan to mid-august 24

Raw data recording for Jan to mid-august 25

Pushing boundaries CMS accumulated backlog 2PB CMS Tier-0 to T* data export speed 7GB/s Issue solved after one week. Many investigations, many actions and many experts from Storage, Network, Experiments and FTS worked together to identify the issues and solve them. 26

Goals summary Ensure a coherent development and operation of storage services at CERN for all aspects of physics data Keep developing and operating Storage Services for Physics at the highest level Communicating Understanding Delivering Keep the ability to adapt and react fast Evaluate and investigate evolutions in technologies for better service/$ Envision new models on data mananagement and analysis Problem/solution Ask/Implement In-house knowhow More for less Operational costs New applications Sync&Share SWAN LHC@myPC 27

More on CERNBOX: CERNBox: the data hub for data analysis (J.Moscicki) - Poster session Total LHC data:130 More PBon SWAN: SWAN: a Service for Web-Based Data Analysis in the Cloud(D.Piparo/E.Tejedor)12th/Oct-11:45 (SierraB) More on distributed EOS distributed: Global EOS: exploring the 300-ms-latency region (L.Mascetti) - Poster session More on Cern Tape Archive (CTA): An efficient, modular and simple tape archiving solution for LHC Run-3 (S. Murray) - Poster Session From Physics to industry: EOS Outside HEP (XE) 29