Using a dynamic data federation for running Belle-II simulation applications in a distributed cloud environment

Similar documents
Running HEP Workloads on Distributed Clouds

Clouds at other sites T2-type computing

Evolution of Cloud Computing in ATLAS

Clouds in High Energy Physics

UK Tier-2 site evolution for ATLAS. Alastair Dewhurst

Bootstrapping a (New?) LHC Data Transfer Ecosystem

Computing activities in Napoli. Dr. Silvio Pardi (INFN-Napoli) Belle II Italian collaboration meeting 21 November 2017 Pisa - Italy

Introduction to SciTokens

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Data Storage. Paul Millar dcache

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

Efficient HTTP based I/O on very large datasets for high performance computing with the Libdavix library

New data access with HTTP/WebDAV in the ATLAS experiment

Tests of PROOF-on-Demand with ATLAS Prodsys2 and first experience with HTTP federation

INDIGO-DataCloud Architectural Overview

Using Puppet to contextualize computing resources for ATLAS analysis on Google Compute Engine

SPINOSO Vincenzo. Optimization of the job submission and data access in a LHC Tier2

Computing at Belle II

Storage Virtualization. Eric Yen Academia Sinica Grid Computing Centre (ASGC) Taiwan

Outline. ASP 2012 Grid School

The PanDA System in the ATLAS Experiment

On-demand provisioning of HEP compute resources on cloud sites and shared HPC centers

Analisi Tier2 e Tier3 Esperienze ai Tier-2 Giacinto Donvito INFN-BARI

HIGH ENERGY PHYSICS ON THE OSG. Brian Bockelman CCL Workshop, 2016

Promoting Open Standards for Digital Repository. case study examples and challenges

Scientific data management

ATLAS Experiment and GCE

PoS(ISGC 2011 & OGF 31)052

COP Cloud Computing. Presented by: Sanketh Beerabbi University of Central Florida

What s new in HTCondor? What s coming? HTCondor Week 2018 Madison, WI -- May 22, 2018

Understanding StoRM: from introduction to internals

LHCb Computing Status. Andrei Tsaregorodtsev CPPM

HTCondor Week 2015: Implementing an HTCondor service at CERN

The Cirrus Research Computing Cloud

Visita delegazione ditte italiane

Considerations for a grid-based Physics Analysis Facility. Dietrich Liko

2017 Resource Allocations Competition Results

Dynamic Federations. Seamless aggregation of standard-protocol-based storage endpoints

The Echo Project An Update. Alastair Dewhurst, Alison Packer, George Vasilakakos, Bruno Canning, Ian Johnson, Tom Byrne

Data Access and Data Management

The National Analysis DESY

UW-ATLAS Experiences with Condor

Federated data storage system prototype for LHC experiments and data intensive science

Parallel Storage Systems for Large-Scale Machines

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 27 July 2010

High Availability & Disaster Recovery. Witt Mathot

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

Exploiting Virtualization and Cloud Computing in ATLAS

WLCG Transfers Dashboard: a Unified Monitoring Tool for Heterogeneous Data Transfers.

Singularity tests at CC-IN2P3 for Atlas

glite Grid Services Overview

The ATLAS Software Installation System v2 Alessandro De Salvo Mayuko Kataoka, Arturo Sanchez Pineda,Yuri Smirnov CHEP 2015

HOW TO PLAN & EXECUTE A SUCCESSFUL CLOUD MIGRATION

Exploring cloud storage for scien3fic research

PoS(EGICF12-EMITC2)106

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

UVA HPC & BIG DATA COURSE. Cloud Computing. Adam Belloum

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

THE ATLAS DISTRIBUTED DATA MANAGEMENT SYSTEM & DATABASES

Benchmarking third-party-transfer protocols with the FTS

dcache Introduction Course

ATLAS Tier-3 UniGe

Part2: Let s pick one cloud IaaS middleware: OpenStack. Sergio Maffioletti


Federated Data Storage System Prototype based on dcache

File Transfer: Basics and Best Practices. Joon Kim. Ph.D. PICSciE. Research Computing 09/07/2018

Introducing SUSE Enterprise Storage 5

Application of Virtualization Technologies & CernVM. Benedikt Hegner CERN

Current Status of the Ceph Based Storage Systems at the RACF

The Software Defined Online Storage System at the GridKa WLCG Tier-1 Center

Data Transfers Between LHC Grid Sites Dorian Kcira

FTS3 a file transfer service for Grids, HPCs and Clouds

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

DESY site report. HEPiX Spring 2016 at DESY. Yves Kemp, Peter van der Reest. Zeuthen,

Spanish Tier-2. Francisco Matorras (IFCA) Nicanor Colino (CIEMAT) F. Matorras N.Colino, Spain CMS T2,.6 March 2008"

Containerized Cloud Scheduling Environment

Distributed Computing Framework. A. Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille

SCS Distributed File System Service Proposal

Azure Everywhere. Brandon Murray, Cami Williams, David Haver, Kevin Carter, Russ Henderson

13th International Workshop on Advanced Computing and Analysis Techniques in Physics Research ACAT 2010 Jaipur, India February

The SciTokens Authorization Model: JSON Web Tokens & OAuth

Data Movement & Tiering with DMF 7

High Availability and Disaster Recovery. Cherry Lin, Jonathan Quinn

COMPUTE CANADA GLOBUS PORTAL

The Legnaro-Padova distributed Tier-2: challenges and results

Paperspace. Deployment Guide. Cloud VDI. 20 Jay St. Suite 312 Brooklyn, NY Technical Whitepaper

Use of containerisation as an alternative to full virtualisation in grid environments.

Choosing an Interface

High Energy Physics data analysis

Cloud Storage with AWS: EFS vs EBS vs S3 AHMAD KARAWASH

ARC integration for CMS

VMware AirWatch Content Gateway Guide For Linux

A scalable storage element and its usage in HEP

CIT 668: System Architecture. Amazon Web Services

Influence of Distributing a Tier-2 Data Storage on Physics Analysis

CERN: LSF and HTCondor Batch Services

A VOSpace deployment: interoperability and integration in big infrastructure

Status of KISTI Tier2 Center for ALICE

Cloud Computing with Nimbus

CMS Belgian T2. G. Bruno UCL, Louvain, Belgium on behalf of the CMS Belgian T2 community. GridKa T1/2 meeting, Karlsruhe Germany February

Transcription:

Using a dynamic data federation for running Belle-II simulation applications in a distributed cloud environment Marcus Ebert mebert@uvic.ca on behalf of the HEP-RC UVic group: Frank Berghaus, Kevin Casteels, Colson Driemel, Colin Leavett-Brown, Michael Paterson, Rolf Seuster, Randall Sobie, Ryan Taylor (University of Victoria) Fernando Fernandez Galindo, Reda Tafirout (TRIUMF) Why we use a dynamic data federation (Dynafed) What is Dynafed Dynafed for Belle-II jobs at UVic Performance comparisons

Traditional GRID site User Central Service data storage Compute user job Grid Site 07/12/2018 CHEP 2018, Sofia 2

Cloud computing for the GRID User Central Service data storage Compute Cloud user job Grid Site 07/12/2018 CHEP 2018, Sofia 3

Cloud computing for the GRID User Central Service Problem: Traditional storage systems are not flexible enough! not efficient in that situation overloading of single SE Grid Site (UVic) data storage? Job Scheduler Canadian West Coast Clouds Canadian East Coast Clouds Clouds on other continents 07/12/2018 CHEP 2018, Sofia 4

User Distributed cloud-storage and cloud-compute for the GRID Central Service Storage cloud not only useful for cloud jobs but also for compute-only sites and experiments as a whole! Grid Site Storage Cloud DYNAFED Any storage close to the job Job Scheduler UVic: CloudScheduler Any cloud in the world More about the compute and accounting in: Quasi-online accounting and monitoring system for distributed clouds (Wednesday, 11:30am T7) Sim@P1: Using Cloudscheduler for offline processing on the ATLAS HLT farm (Tuesday 11:30 T7) 07/12/2018 CHEP 2018, Sofia 5

Dynafed redirector for a dynamic data federation, developed by CERN IT for data transfers, client is redirected to a storage element with the data access through http(s)/dav(s) protocols can federate existing sites without configuration changes at sites site needs to be accessible by http(s)/dav(s) world wide distributed data can be made accessible under common name space and from single endpoint in addition, can also directly access S3 and Azure based storage no credentials visible to the client preauthorized URL with limited lifetime is used to access files on the storage no large Grid storage installation needed (DPM, dcache,...) X509/VOMS based authentication/access authorization can be used with dynafed http://heprc.blogspot.com for grid-mapfile based authentication/authorization different posts have also links to dynafed installation instructions in our TWiki more details in poster #69 by Fabrizio Furano and Oliver Keeble 07/12/2018 CHEP 2018, Sofia 6

Some features using Dynafed redirecting client to nearest site that has data in the future other characteristics could be added, like latency, bandwidth, storage space, cost,... client tools can get new redirect to another site if anything happens with an already established connection site outage, network problems at a site,... root based tools can speak webdav and access data over network using dynafed TFile *f=tfile::open("davs://dynafed.server:port/belle/path/to/file/file.root") uses external davix libraries writing into Dynafed also goes to one of the connected sites uses also location based redirect 07/12/2018 CHEP 2018, Sofia 7

Dynafed for Belle-II at UVic Dynafed uses http(s)/dav(s) protocol which is supported by the gfal2 tools currently Belle-II only supports srm and locally available data gfalfs can be used to mount endpoint use Dynafed as endpoint for gfalfs everything behind Dynafed appears as local data still benefits from all Dynafed features location based redirect fail-over when endpoint goes down workflow: copy data from SE or local directory to workdir of the job no streaming over network ~3000 cores used in parallel for Belle-II single core jobs each job needs input file ~30TB data per day transferred to jobs very inefficient if using long-distance transfers (site SE -> cloud) can also easily overload a single site SE 07/12/2018 CHEP 2018, Sofia 8

Dynafed for Belle-II at UVic Dynafed uses http(s)/dav(s) protocol which is supported by the gfal2 tools currently Belle-II only supports srm and locally available data gfalfs can be used to mount endpoint use Dynafed as endpoint for gfalfs everything behind Dynafed appears as local data still benefits from all Dynafed features location based redirect fail-over when endpoint goes down workflow: copy data from SE or local directory to workdir of the job no streaming over network ~3000 cores used in parallel for Belle-II single core jobs each job needs input file ~30TB data per day transferred to jobs very inefficient if using long-distance transfers (site SE -> cloud) can also easily overload a single site SE see Frank s Poster (#161) about Dynafed for Atlas 07/12/2018 CHEP 2018, Sofia 9

Dynafed monitoring general request timeline endpoint status timeline requests served per endpoint information which cloud/machine requested how many files 07/12/2018 CHEP 2018, Sofia 10

Testing Dynafed usage to make test comparable: using gfal-copy dynafed instance at Victoria (Canadian West Coast) Grid and cloud storage endpoints behind Dynafed Grid storage: UVic SE, BNL SE cloud storage: Ceph at UVic, minio at Compute Canada Westcloud and Eastcloud VMs using volumes, minio at group s own cloud using large local root disk 500 3GB files copied to a worker node VM copied to /dev/shm to not rely on virtual disk access gfal2 usage with: srm access to UVic SE (site SE) http access to UVic SE dynafed access using grid sites behind it dynafed access using cloud storage endpoints behind it VM on Compute Canada Westcloud (Victoria,BC), Eastcloud (Sherbrooke,QC), group s own development cloud (Victoria, BC) do all copy processes on all clouds and compare results cloud shared with other users endpoint used by production jobs and other VOs 07/12/2018 CHEP 2018, Sofia 11

Testing gfalfs/dynafed Victoria - Sherbrooke: 3853km Sherbrooke - BNL: 515km Victoria : 1xminio, 1xown Ceph, 1 Belle-II SE, 1xshared Openstack, 1xown Openstack Sherbrooke : 1xminio, 1xshared Openstack BNL : 1x Belle-II SE 07/12/2018 CHEP 2018, Sofia 12

Direct access to site SE (srm) direct access to site SE using srm protocol Westcloud nodes have direct access to our site SE other clouds need to go through cloud router first 13

Direct access to site SE (http) direct access to site SE using http protocol performs better than srm access consistent across all clouds 14

Access to grid sites (dynafed) using dynafed to access grid endpoints only all 3 VMs access the same dynafed URL using dynafed not slower than direct http access consistent between Westcloud and own cloud Eastcloud data access benefits from dynafed dynafed chooses nearest endpoint can use BNL grid site instead of site SE 15

Access to cloud sites (dynafed) using dynafed to access cloud endpoints only all 3 VMs access the same dynafed URL cloud storage on Westcloud performs not as good as grid storage data on volume mounted in minio VM (network based) Ceph runs outside of cloud many more layers than accessing site SE needs to go through cloud network to access data on minio/ceph on own cloud, minio performs best data access within cloud data stored on local disk of minio VM on Eastcloud setup like on Westcloud smaller cloud on ComputeCanada clouds other users run VMs/data transfers too 16

Summary Dynafed works very well for our distributed cloud computing when reading input data instant fail-over to other sites kept jobs running relieves pressure on site SE much more efficient to use than just site SE gfalfs is a great way to use Dynafed when only local data access tools are supported all features of dynafed can be used no performance issues when using dynafed compared to direct http access to site SE http access performs better than srm access dynafed makes it easy to add storage to the federation using dynafed as site SE will make the setup of grid storage much simpler can natively use object storage which eliminates the whole grid storage infrastructure setup Very positive experience in using it over the last year! 07/12/2018 CHEP 2018, Sofia 17

Backup

Cloud compute at UVic Scheduler Status Communication job submission University cluster VM VM VM VM VM VM Cloudscheduler (starts/terminates VMs as needed) Other Universities VM VM VM VM VM VM HTCondor run jobs for Atlas and Belle-II on ~15 different clouds Canada, USA, CERN, UK, Germany, Australia ~6000 cores Public Clouds VM VM VM VM VM VM Image distribution between clouds: Glint Cloud types: Openstack, OpenNebula, Amazon, Microsoft Azure, Google Cloud 07/12/2018 CHEP 2018, Sofia 19

gfalfs for Belle-II at UVic mount: gfalfs -s ${HOME}/b2data/belle davs://dynafed02.heprc.uvic.ca:8443/belle usage: export VO_BELLE_DATA_DIR=$HOME/b2data VOMS proxy of the job is used for the authentication and access authorization Worked very well for about a year so far. Problem occured recently: segfaults of gfalfs happen ~since beginning of year in libcrypto.so as part of openssl ticket with CERN IT open hard to debug since not reproducible manually cronjob makes sure dynafed gets remounted when that happens 07/12/2018 CHEP 2018, Sofia 20

Advantages of using S3 based storage easy to manage no extra servers needed, no need for the whole Grid infrastructure on site (DPM, mysql, apache, gridftp, xrootd, VOMS information, grid-mapfile, accounting,...) just use private/public access key in central Dynafed installation no need for extra manpower to manage a grid storage site small group with budget to provide storage but no manpower for it: Just buy S3 based xtb for y years and put the information into dynafed ---> instantly available to the Grid, no need to buy/manage/update extra hardware if university/lab has already large Ceph installation --> just ask for/create a bucket, and put credentials in dynafed industry standard adopted from Amazon by Open Source and commercial cloud and storage solutions HPC, Openstack, Ceph, Google, Rackspace cloud storage, NetApp, IBM,... scalable traditional local file storage servers based on traditional filesystems will become harder to manage/use with growing capacity needs, same for other bundle solutions (DPM,...) raid5 dead, raid6 basically dead too, ZFS will get problems with network performance 07/12/2018 CHEP 2018, Sofia 21