Computing activities in Napoli. Dr. Silvio Pardi (INFN-Napoli) Belle II Italian collaboration meeting 21 November 2017 Pisa - Italy

Similar documents
Tests of PROOF-on-Demand with ATLAS Prodsys2 and first experience with HTTP federation

Clouds at other sites T2-type computing


Europe and its Open Science Cloud: the Italian perspective. Luciano Gaido Plan-E meeting, Poznan, April

Using a dynamic data federation for running Belle-II simulation applications in a distributed cloud environment

The Legnaro-Padova distributed Tier-2: challenges and results

Clouds in High Energy Physics

Evolution of Cloud Computing in ATLAS

CERN Network activities update

Status of KISTI Tier2 Center for ALICE

Understanding StoRM: from introduction to internals

LHCb experience running jobs in virtual machines

Helix Nebula Science Cloud Pre-Commercial Procurement pilot. 5 April 2016 Bob Jones, CERN

Helix Nebula Science Cloud Pre-Commercial Procurement pilot. 9 March 2016 Bob Jones, CERN

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

Pre-Commercial Procurement project - HNSciCloud. 20 January 2015 Bob Jones, CERN

Exploring cloud storage for scien3fic research

Grid Computing Activities at KIT

Parallel Storage Systems for Large-Scale Machines

Running HEP Workloads on Distributed Clouds

Storage Virtualization. Eric Yen Academia Sinica Grid Computing Centre (ASGC) Taiwan

The Echo Project An Update. Alastair Dewhurst, Alison Packer, George Vasilakakos, Bruno Canning, Ian Johnson, Tom Byrne

Dynamic Federations. Seamless aggregation of standard-protocol-based storage endpoints

Spanish Tier-2. Francisco Matorras (IFCA) Nicanor Colino (CIEMAT) F. Matorras N.Colino, Spain CMS T2,.6 March 2008"

LHCb Computing Resource usage in 2017

Promoting Open Standards for Digital Repository. case study examples and challenges

Introduction to SRM. Riccardo Zappi 1

Federated Data Storage System Prototype based on dcache

LHCb Distributed Conditions Database

Distributed production managers meeting. Armando Fella on behalf of Italian distributed computing group

Cluster Setup and Distributed File System

Data Storage. Paul Millar dcache

Distributing storage of LHC data - in the nordic countries

Computing for LHC in Germany

ARC integration for CMS

On-demand provisioning of HEP compute resources on cloud sites and shared HPC centers

Bridging cloud computing innovation & Open Science

LHCb Computing Status. Andrei Tsaregorodtsev CPPM

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

BSA Best Practices Webinars Distributed Installations Sean Berry Customer Engineering

Edinburgh (ECDF) Update

Long Term Data Preservation for CDF at INFN-CNAF

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

ATLAS computing activities and developments in the Italian Grid cloud

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Efficient HTTP based I/O on very large datasets for high performance computing with the Libdavix library

Disaster recovery of the INFN Tier-1 data center: lesson learned. Luca dell Agnello INFN - CNAF CHEP 2018 conference Sofia, July

Optimization of Italian CMS Computing Centers via MIUR funded Research Projects

Rethink Storage: The Next Generation Of Scale- Out NAS

Constant monitoring of multi-site network connectivity at the Tokyo Tier2 center

Application of Virtualization Technologies & CernVM. Benedikt Hegner CERN

Analisi Tier2 e Tier3 Esperienze ai Tier-2 Giacinto Donvito INFN-BARI

Computing / The DESY Grid Center

XCache plans and studies at LMU Munich

Challenges of Big Data Movement in support of the ESA Copernicus program and global research collaborations

The creation of a Tier-1 Data Center for the ALICE experiment in the UNAM. Lukas Nellen ICN-UNAM

System upgrade and future perspective for the operation of Tokyo Tier2 center. T. Nakamura, T. Mashimo, N. Matsui, H. Sakamoto and I.

Andrea Sciabà CERN, Switzerland

Data storage services at KEK/CRC -- status and plan

Benchmarking third-party-transfer protocols with the FTS

Benchmarking the ATLAS software through the Kit Validation engine

The INFN Tier1. 1. INFN-CNAF, Italy

INTEGRATING HPFS IN A CLOUD COMPUTING ENVIRONMENT

Data transfer over the wide area network with a large round trip time

DESY. Andreas Gellrich DESY DESY,

EUROPEAN MIDDLEWARE INITIATIVE

hybrid cloud for science Kickoff Phase 3 Pilot FeBRUARY, 6 th / 7 th 2018 Team T-Systems/Huawei/Cyfronet/Divia

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Presented by Manfred Alef Contributions of Jos van Wezel, Andreas Heiss

Beob Kyun KIM, Christophe BONNAUD {kyun, NSDC / KISTI

Cloud interoperability and elasticity with COMPSs

New data access with HTTP/WebDAV in the ATLAS experiment

ATLAS Experiment and GCE

Austrian Federated WLCG Tier-2

Federated data storage system prototype for LHC experiments and data intensive science

OpenNebula on VMware: Cloud Reference Architecture

Federated access to e-infrastructures worldwide

Influence of Distributing a Tier-2 Data Storage on Physics Analysis

Database monitoring and service validation. Dirk Duellmann CERN IT/PSS and 3D

ISTITUTO NAZIONALE DI FISICA NUCLEARE

Prototype DIRAC portal for EISCAT data Short instruction

271 Waverley Oaks Rd. Telephone: Suite 206 Waltham, MA USA

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

Workload management at KEK/CRC -- status and plan

COP Cloud Computing. Presented by: Sanketh Beerabbi University of Central Florida

Building Service Platforms using OpenStack and CEPH: A University Cloud at Humboldt University

Visita delegazione ditte italiane

Helix Nebula The Science Cloud

Developing Microsoft Azure Solutions (70-532) Syllabus

UK Tier-2 site evolution for ATLAS. Alastair Dewhurst

CernVM-FS beyond LHC computing

13th International Workshop on Advanced Computing and Analysis Techniques in Physics Research ACAT 2010 Jaipur, India February

The ATLAS Software Installation System v2 Alessandro De Salvo Mayuko Kataoka, Arturo Sanchez Pineda,Yuri Smirnov CHEP 2015

Setting up Microsoft Exchange Server 2016 with Avi

SPINOSO Vincenzo. Optimization of the job submission and data access in a LHC Tier2

glite Grid Services Overview

Hyper-V Innovations for the SMB. IT Pro Camp, Northwest Florida State College, Niceville, FL, October 5, 2013

Developing Microsoft Azure Solutions (70-532) Syllabus

Developing Microsoft Azure Solutions (70-532) Syllabus

A European Cloud federation

Helix Nebula The Science Cloud

Transcription:

Computing activities in Napoli Dr. Silvio Pardi (INFN-Napoli) Belle II Italian collaboration meeting 21 November 2017 Pisa - Italy

Activities in Napoli Grid Services Networking Http/Webdav and Dynamic Federation SCOReS-GARR Project Helix Nebula Science Cloud VCYCLE

Grid Services: Computing GRID Cluster updated with the possibility to use opportunistic resource from local project (Prisma, SCOReS) LCG.Napoli.it has been one of the most used Site in the last 4 Quarters

Grid Services: Storage Storage totally Full Since February 2017 NAPOLI-SE has been one of the most used storage in the last 4 Quarters. The

Grid Services: Site Update Recent update A new Chiller has been implemented in order to improve the data centre stability Short term plan (2017-2018) A part of a storage for testing will be integrated in the production DPM by to the end of the year (about 140 TB) Medium/Long term plan (2018-2020) University of Napoli Federico II and INFN-Napoli are working to apply for the next National Call for Infrastructure empowering in the contest of Innovation ad Big-Data Foreseen update are: Install new Racks (the current site can be expanded with up to 11) New storage and computing resource will be acquired Hardware Renewal

Networking: Network Rappresentative Silvio Pardi and Bruno Hoeft Network Representative for Europe. Participation at the WLCG-GDB focus on Network representing Belle II (January 2017) Participation at LHCONE KEK JP (October 2017) Participation at Super Computing 17 Denver USA (November 2017) Network estimation update at Super Computing.

Networking: Data Challenge

Networking: Data Challenge

Http/Webdav and Dynamic Federation # STORGE DIRAC NAME HOSTNAME TYPE 1 DESY-DE dcache-belle-webdav.desy.de DCACHE 2 GRIDKA-SE f01-075-140-e.gridka.de DCACHE 3 NTU-SE bgrid3.phys.ntu.edu.tw DCACHE 4 SIGNET-SE dcache.ijs.si DCACHE 5 UVic-SE charon01.westgrid.ca DCACHE 6 Adelaide-SE coepp-dpm-01.ersa.edu.au DPM 7 CESNET-SE dpm1.egee.cesnet.cz DPM 8 CYFRONNET-SE dpm.cyf-kr.edu.pl DPM 9 Frascati-SE atlasse.lnf.infn.it DPM 10 HEPHY-SE hephyse.oeaw.ac.at DPM 11 Melbourne-SE b2se.mel.coepp.org.au DPM 12 Napoli-SE belle-dpm-01.na.infn.it DPM 13 ULAKBIM-SE torik1.ulakbim.gov.tr DPM 14 IPHC-SE sbgse1.in2p3.fr DPM 15 CNAF-SE ds-202-11-01.cr.cnaf.infn.it STORM 16 ROMA3-SE storm-01.roma3.infn.it STORM 17 KEK-SE Kek-se03.cc.kek.jp STORM 18 McGill-SE gridftp02.clumeq.mcgill.ca STORM Testing Dynafed server in Napoli updated at the last version 2 weeks ago. The testbed is expanded Now included 18 of the 23 SRM in production. https://dynafed01.na.infn.it/myfed/

Federation Views Dynafed server in Napoli https://dynafed01.na.infn.it/myfed myfed/persite/ Shows the file systems of each storage separately (without aggregation) myfed/belle/ Aggregation of all the directory /DATA/belle and /TMP/belle/ myfed/belle-cache-path/ Testing area for caching myfed/dav-napoli/ Testing area exported by a webdav-only storage in Napoli myfed/s3-rhea/ S3 area for HNSciCloud myfed/s3-tsystem/ S3 area for HNSciCloud myfed/s3-ibm/ S3 area for HNSciCloud

Http/Webdav and Dynamic Federation Ongoing Activities: Test of usage via DIRAC (Ueda) Test S3 storage areas in Dynafed Test standard WedDav Storage in Dynafed using a local synology Testing ACL in Dynafed The Dynafed Server in Napoli is part of the WLCG demonstrator for data federation as representative of the Belle II use-case.

SCOReS Project Italian Acronym for: Study of a Caching system for optimize the usage of Opportunistic Resources and site without pledged storage, for e-science application (SCOReS). Project funded by GARR within a National call consisting in a 1+1Year fellowship. Davide Michelino - project fellowship Silvio Pardi Project Tutor for INFN-Napoli

Cache Use-Cases Cache can affects performance in many scenarios: At site level: Cache increase Analysis performance for those job run on the same data-set. Cache can help all sites close to the one hosting cache Storage-Less Site Paradigm Cloud Storage Access to Cloud storage with limited bandwidth vs the client Limit the number of GET requests on Cloud Storage

Caching laboratory with DPM CHEP 2016

Concept of Volatile Pool A Pool in DPM is a collection of File System managed as a single storage area. A Volatile Pool is a special pool that can pull files from external sources, which can be used to deploy simple data caches. When User stat a file in a Volatile pool, the DPM head node runs a script and then reply with file information even if the file is not there. When User get a file from the Volatile pool, the Disk Node providing the file system of the pool, send the file to the client if ready, otherwise a script is locally runned in order to retrieve the file from some external source.

File Get Data GriD File download HEAD NODE DISK NODE Get a File File is there No Script To retrieve file File ready File Sender Yes Sent File to the Client

Dynafed + Volatile Pool Cache Real File What happen if we aggregate a Webdav endpoint with a DPM Volatile Pool? When Dynafed stat files inside the real webdav endpoint, it receive always an reply even from the Volatile Pool. So that the metalink representing a file in Dynafed, included always at the real URL and the corresponding virtual copy in the cache (even if does not exist yet) Moreover Thanks to the GeoPlugin, Dynafed prioritize the cache copy if the Volatile Pool is local to the Client or close to it. This combination allow to create a cache system

OUR IMPLEMENTATION DPM DOME User Interface 20Gb/s WAN 10Gb/s LAN DYNAFED 10Gb/s LAN SOFTWARE VERSION dpm-1.9.0-1 dynafed-1.3.1-2 davix-0.6.6-1 DPMHEAD NODE KEK Storage CESNET Storage DPMDISK1 DPMDISK2 pool-recas pool-cache Private Cloud storage S3

Preliminary Tests Schema CESNET-SE USER UI 2. Redirect to the Cache 1. Request Dynafed Server /myfed/belle-cache-path KEK-SE 5. CACHE WARMING UViv-SE Volatile Pool

Mbit/s Test Results 49MB Mbit/s (Higher is better) 2000 1800 1600 1400 1200 1000 800 600 400 200 0 1GB Test 1684 1739 1739 32 16 40 20 152 47 KEK Uvic CESNET Direct Empty Cache Warm Cache

Test Results 1GB Mbit/s (Higher is better) Suppose to be identical Storage Tuning should fix it

To-do and Ongoing works Bug fixing Study the cache dynamic (makespace function) Study system reliability Define some test use-case.

HNSciCloud Project Dr. Silvio Pardi 23 Ref: HNSciCloud Introduction - Bob Jones - Prototype Phase kickoff meeting

HNSciCloud Providers The 3 April the winners of the bids to build prototypes have been announced at CERN during the Launching the Helix Nebula Science Cloud Prototype Phase webcast event. http://www.hnscicloud.eu/events/launching-the-helix-nebula-science-cloudprototype-phase-webcast-3-april-2017-cern Contractor 1: T-Systems, Huawei, Cyfronet, Divia Contractor 2: IBM Contractor 3: RHEA Group, T-Systems, exoscale, SixSq Dr. Silvio Pardi 24

HNSciCloud Update Part of the jobs used HNSciCloud resources: The current status is the follow: 2 Cloud Infrastructures work properly but there is a persistent random issue with cloud-init that affect the VCYCLE server activity. Cloud providers are trying to solve the issue. 1 Cloud is still not directly usable with VCYCLE because the lack of a set of compatible API. Two option: Create a specific VCYCLE plugin Wait that the provider will produce the full set of API compatible

Cloud: What is VCYCLE VCYLCE is VM lifecycle manager developed by GRIDPP, it is designed to create VMs on Cloud endpoints offering EC2, Openstack or Azure interface. VCYCLE can be easily integrated in DIRAC and the accounting system is compliant with APEL. VCYCLE is currently uses in production by LHCb. https://www.gridpp.ac.uk/vcycle/

VCYCLE for Belle II Testing activity started during the last B2GM26 (suggested by Ueda) In February we meet Andrew McNab (developer of vcycle) at CERN and we had a for a working session. We configured a VCYCLE endpoint attached to the ReCaS/PRISMA Cloud and we integrated it in the DIRAC VALIDATION server of PNNL. The work required a specific customization for the Belle II usecase.

VCYCLE Testbed setup in Napoli recas-vcycle01.na.infn.it VCYCLE VM FACTORY SERVICE https://recas-vcycle01.na.infn.it:8443 HTTP CONTEXTUALIZATION ENDPOINT Virtual Server for all the services 4 cores 8GB Ram 50GB HD VCYCLE Tenant on a Local Openstack Cloud VM1 VM2 VM3 VM4 https://recas-vcycle01.na.infn.it:443 HTTP ENDPOINT FOR LOGGING KEK DIRAC SERVER SITE: VCYLE.Napoli.it

VCYCLE for HNSciCloud and Prisma Napoli RHEA belle-vcycle01 VM2 VM4 T-SYSTEM belle-squid01 VM1 VM3 IBM belle-vcycle01 VM2 VM4 belle-vcycle01 VM2 VM4 belle-squid01 VM1 VM3 belle-squid01 VM1 VM3 Napoli PRISMA belle-vcycle01 belle-squid01 VM2 VM4 VM1 VM3 Belle-KEK Site:VCYCLE.Napoli.it CLOUD WORKS UNDER THE SAME DIRAC Site VCYCLE.Napoli.it BelleDIRAC-PNNL-VALIDATION Site:VCYCLE.Napoli.it

DIRAC Jobs Among the Cloud site in Belle II VCYCLE.Napoli.it was the most used. French askes for use our VCYCLE server to integrate their Openstack resources in DIRAC.

People Prof. Guido Russo Dr. Silvio Pardi Dr. Davide Michelino (GARR fellowship)