Exploring cloud storage for scien3fic research

Similar documents
Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Evaluation of the Huawei UDS cloud storage system for CERN specific data

Storage Virtualization. Eric Yen Academia Sinica Grid Computing Centre (ASGC) Taiwan

Towards Transparent Integration of Heterogeneous Cloud Storage Platforms

Application of Virtualization Technologies & CernVM. Benedikt Hegner CERN

VMs at a Tier-1 site. EGEE 09, Sander Klous, Nikhef

Understanding StoRM: from introduction to internals

Andrea Sciabà CERN, Switzerland

Visita delegazione ditte italiane

-Presented By : Rajeshwari Chatterjee Professor-Andrey Shevel Course: Computing Clusters Grid and Clouds ITMO University, St.

Next Generation Storage for The Software-Defned World

CernVM-FS beyond LHC computing

Storage for HPC, HPDA and Machine Learning (ML)

The CMS data quality monitoring software: experience and future prospects

Ceph Intro & Architectural Overview. Abbas Bangash Intercloud Systems

ATLAS Experiment and GCE

Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU

CouchDB-based system for data management in a Grid environment Implementation and Experience

Big Data in OpenStack Storage

High Performance Computing on MapReduce Programming Framework

Cloud Storage. Patrick Osborne Director of Product Management. Sam Fineberg Distinguished Technologist.

South African Science Gateways

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

The ATLAS EventIndex: Full chain deployment and first operation

IBM Aspera Direct-to- Cloud Storage

LHCb Distributed Conditions Database

Data Movement & Tiering with DMF 7

Grid Computing Activities at KIT

Elastic Cloud Storage (ECS)

irods usage at CC-IN2P3 Jean-Yves Nief

Grid Computing: dealing with GB/s dataflows

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

Cloud Computing and Hadoop Distributed File System. UCSB CS170, Spring 2018

Data Grid Infrastructure for YBJ-ARGO Cosmic-Ray Project

Object Storage Service. Product Introduction. Issue 04 Date HUAWEI TECHNOLOGIES CO., LTD.

Computing activities in Napoli. Dr. Silvio Pardi (INFN-Napoli) Belle II Italian collaboration meeting 21 November 2017 Pisa - Italy

OPENSTACK PRIVATE CLOUD WITH GITHUB

How To Guide: Long Term Archive for Rubrik. Using SwiftStack Storage as a Long Term Archive for Rubrik

Aspera Direct-to-Cloud Storage WHITE PAPER

Basics of Cloud Computing Lecture 2. Cloud Providers. Satish Srirama

New research on Key Technologies of unstructured data cloud storage

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

Federated Data Storage System Prototype based on dcache

LHCb Computing Status. Andrei Tsaregorodtsev CPPM

IBM Spectrum NAS, IBM Spectrum Scale and IBM Cloud Object Storage

IBM Aspera Direct-to-Cloud Storage

Benchmarking third-party-transfer protocols with the FTS

EGEE and Interoperation

Storage Resource Sharing with CASTOR.

Developing Enterprise Cloud Solutions with Azure

Enosis: Bridging the Semantic Gap between

REFERENCE ARCHITECTURE Quantum StorNext and Cloudian HyperStore

Virtualizing a Batch. University Grid Center

Monitoring system for geographically distributed datacenters based on Openstack. Gioacchino Vino

Centre de Calcul de l Institut National de Physique Nucléaire et de Physique des Particules. Singularity overview. Vanessa HAMAR

Ansible Tower Quick Setup Guide

Xcellis Technical Overview: A deep dive into the latest hardware designed for StorNext 5

ROCK INK PAPER COMPUTER

A New Key-value Data Store For Heterogeneous Storage Architecture Intel APAC R&D Ltd.

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

Lightweight Streaming-based Runtime for Cloud Computing. Shrideep Pallickara. Community Grids Lab, Indiana University

THE ATLAS DISTRIBUTED DATA MANAGEMENT SYSTEM & DATABASES

Learning Management System (LMS) + Content Management System (CMS)

The ATLAS Tier-3 in Geneva and the Trigger Development Facility

Using S3 cloud storage with ROOT and CvmFS

Large Scale Computing Infrastructures

SLI Learning Search Connect For Magento 2

Scale-out Object Store for PB/hr Backups and Long Term Archive April 24, 2014

Best Practices in Designing Cloud Storage based Archival solution Sreenidhi Iyangar & Jim Rice EMC Corporation

Windows Azure Services - At Different Levels

IEPSAS-Kosice: experiences in running LCG site

Data Transfers Between LHC Grid Sites Dorian Kcira

HPC File Systems and Storage. Irena Johnson University of Notre Dame Center for Research Computing

powered by Cloudian and Veritas

HPC learning using Cloud infrastructure

Chapter 2 Introduction to the WS-PGRADE/gUSE Science Gateway Framework

CERN s Business Computing

CSE6331: Cloud Computing

Summary of the LHC Computing Review

Using a dynamic data federation for running Belle-II simulation applications in a distributed cloud environment

OpenStack SwiftOnFile: User Identity for Cross Protocol Access Demystified Dean Hildebrand, Sasikanth Eda Sandeep Patil, Bill Owen IBM

MOHA: Many-Task Computing Framework on Hadoop

glite Grid Services Overview

Build Cloud like Rackspace with OpenStack Ansible

Genomics on Cisco Metacloud + SwiftStack

Rio-2 Hybrid Backup Server

SPINOSO Vincenzo. Optimization of the job submission and data access in a LHC Tier2

Dell EMC CIFS-ECS Tool

Filesystems on SSCK's HP XC6000

AMAZON S3 FOR SCIENCE GRIDS: A VIABLE SOLUTION?

CC-IN2P3 activity. irods in production: irods developpements in Lyon: SRB to irods migration. Hardware setup. Usage. Prospects.

Developing Microsoft Azure Solutions (70-532) Syllabus

SwiftStack Object Storage

Cloud Open Source Innovation on Software Defined Storage

Software Defined Storage. Mark Carlson, Alan Yoder, Leah Schoeb, Don Deel, Carlos Pratt, Chris Lionetti, Doug Voigt

Improve Web Application Performance with Zend Platform

Building Storage-as-a-Service Businesses

DIRAC pilot framework and the DIRAC Workload Management System

HEP Grid Activities in China

Transcription:

Exploring cloud storage for scien3fic research Fabio Hernandez fabio@in2p3.fr Lu Wang Lu.Wang@ihep.ac.cn 第十六届全国科学计算与信息化会议暨科研大数据论坛 h"p://indico.ihep.ac.cn/conferencedisplay.py?confid=3138 Dalian, July 8th 2013

2 Who I am Computing science background Working in the field of computing for high-energy physics for more than 20 years software development for scientific data management (data transfer over high-throughput networks, mass storage & retrieval, cataloguing,...) and operations of IT services for research Deeply involved in planning, prototyping, deploying and operating the computing infrastructure for CERN s Large Hadron Collider technical leadership of the French contribution to the LHC computing grid served in the management board and grid deployment board of the WLCG collaboration Served as deputy director of IN2P3 computing centre, which hosts and operates the French tier-1, a major center for the LHC computing grid Visiting professor at the Institute of High Energy Physics, Chinese Academy of Sciences, in Prof. Gang CHEN s team working on cloud-based storage systems and unstructured data stores

3 Summary We have been exploring what it takes to use cloud storage for storing data in a scientific research context Here we report on our progress so far and the prospects

4 Contents A word on cloud storage The big picture: what we would like to do Current status Perspectives Conclusions

5 Cloud storage Data storage as a service (DaaS) Storage capacity delivered on demand billing based on metered usage File-based cloud storage unit of storage is a file (sequence of bytes) files are stored in containers reduced set of basic operations: list, store, retrieve, delete files; create, delete containers large-scale repositories of objects (images, video, data files, ) REST-based APIs for interacting with the service for uploading, downloading and removing files for listing the contents of the containers and retrieving file metadata

6 Cloud storage (cont.) Benefits non-stop access from any connected computer eliminate physical storage boundaries: sharing of data is made easier massively scalable: Amazon S3 stores 2 trillion objects, 1.1M requests per second as of April 2013* Protocols S3: introduced by Amazon, also supported by Google and other providers OpenStack Swift & RackSpace SNIA s Cloud Data Management Interface (CDMI) Implementors and providers Amazon, Google, Rackspace, Microsoft Azure, OpenStack Swift: middleware for deploying cloud storage services storage appliances: Huawei, NetApp, * Source: http://aws.typepad.com/aws/2013/04/amazon-s3-two-trillion-objects-11-million-requests-second.html

7 S3 protocol Amazon s S3: Simple Storage Service De facto standard protocol for object storage well documented, works on top of HTTP(S) several client-side tools support it, both CLI- and GUI-based several service providers and implementors (Amazon, Google, Eucalyptus,...) S3 storage model objects (i.e. files) stored within containers called buckets limits: 100 buckets per user, infinite number of files per bucket, 5TB per file, bucket namespace shared among all users of a single provider flat object namespace within buckets: no hierarchy, no directories no POSIX semantics: partial reading, no partial writing, no locks, file versioning per-bucket and per-file access control based on ACLs authentication based on secret shared between the user and the server user is provided with credentials of type: access key and secret key file sharing mechanism via standard URL Bucket with objects

8 Objective

9 Objective How can we exploit cloud storage in the context of data processing for high energy physics experiments? targets: experiments as a whole and individuals How can we make this as transparent as possible from the experiment/individual point of view? users should not be concerned by the specifics of the storage provider nor the protocol used

10 Current status

11 Building blocks Storage servers which expose cloud protocols either operated in-house or by a external (commercial) service provider Tools for interacting with the storage backend both GUI-based and CLI-based (scriptable) Seamless integration of cloud protocols to the experiment s software stack

12 Use case 1: experimental data storage Use cases storage of experimental data to be processed by batch jobs executing in the local farm storage of experimental data to be shared with remote sites participating in a research collaboration central storage repository for collecting simulation data produced by jobs executed at remote sites Required tools command line tools for downloading and uploading files: to be used by batch jobs (local or grid) data access libraries that understand cloud protocols, e.g. CERN s ROOT framework

13 Use case 2: personal file storage Goal: provide a storage area for individual users controlled and managed by the end user significant but limited capacity usage convenience is more important than raw performance kind of personal storage element: accessible from the user s desktop computer and readable and writable by the user s (local and grid) batch jobs individual user can share his own files with other individuals What tools can we provide or recommend so that individuals can interact with their own storage area? both from their desktop environment and from their jobs GUI and command-line file system emulation

14 Cloud storage overview

15 Experimental data storage Improvements of S3 support in CERN s ROOT data analysis framework experiments can efficiently read remote files using the S3 protocol as if the files were stored in a local file system no experiment-specific software needs to be modified TFile* f = TFile::Open( s3://fsc01.ihep.ac.cn/mybucket/path/to/mydatafile.root ) features: partial reads, vector read (provided the S3 server supports it), web proxy handling, HTTP and HTTPS works with Amazon, Google, Huawei, OpenStack Swift, available since ROOT v5.34.05 (Feb. 2013)

16 Experimental data storage (cont.) RELATIVE PERFORMANCE FOR SCANNING SAMPLE ATLAS ROOT FILE [FILESIZE: 768MB, 11.918 EVENTS] NOTES ROOT v5.34.05 ext3 Lustre S3 http 1.0 1.1 1.2 Reading of all entries in the physics tree of the file. When using ext3, the process takes 37 seconds on a worker node running SL5. S3 https 1.5 Downloading the whole file to the local disk takes 5 secs. s3fs * 1.5 Storage back-end: Huawei UDS 0.8 1.0 1.2 1.4 1.6 * s3fs: slightly modified version to remove virtual host notation (not supported by Huawei UDS) WALLCLOCK TIME RELATIVE TO EXT3 (shorter is better)

17 CLI-based interface Command-line interface we developed a CLI-based S3 client compatible with Amazon, Google, OpenStack Swift, Huawei, GO programming language exploit GO s built-in concurrency stand-alone executable small size, so to avoid pre-installation and be downloadable $ mucura --help Usage: mucura <command> [opts...] args... mucura --help mucura --version Accepted commands: lb list existing buckets mb make a new bucket ls list bucket contents rb remove a bucket up upload a file to a bucket dl download a file rm remove files Use 'mucura help <command>' for getting more information on a specific command $ mucura dl http://s3.amazonaws.com/mucura/path/to/myfile /tmp works on MacOS, Linux and (soon) Windows

18 GUI-based interface CloudBerry Explorer CloudBerry Lab Windows 7 Cyberduck (MacOS X) Expandrive (MacOS X)

19 Extensions to ROOT framework We developed an extension of ROOT: a set of C++ classes for supporting cloud storage protocols delivered as a shared library usable both within ROOT macros and from interpreter (CINT) compatible with all versions of ROOT since v5.24.00b (Oct. 2009) no need to modify ROOT source code Benefits read-only access to remote files using cloud protocols is now possible with legacy ROOT versions no modification to existing ROOT macros transparent access to your ROOT data files: currently supports S3 and Swift protocols (Amazon, Google, OpenStack, ) you can also install it and access the same remote files via ROOT from the comfort of your personal computer, wherever you are connected to the network

20 Extensions to ROOT framework (cont.) Load ROOT C++ macro Draw the histogram contained in the remote Swift file

21 Ongoing work

22 Ongoing work Testing with BES III experimental data BES III jobs running at IHEP batch farm can read files in ROOT format hosted by IHEP s OpenStack Swift testbed larger scale tests to be performed in the coming weeks File system emulation layer in development FUSE-based file system layer for accessing remote files through S3 and Swift protocols planned features: multi-provider integration, client-side encryption, managed local cache for offline access,

23 Ongoing work (cont.)

24 Ongoing work (cont.) Understanding I/O patterns work conducted by Lu WANG goal: analyze the I/O patterns exhibited by BES III jobs when accessing data files, as seen by the file system currently collecting data using Lustre file system expected benefits: problem detection, system tuning, automation e.g. this information should be useful to tune the storage system (client-side, network, data and metadata servers) to match the usage made by the experiment, and conversely use collected data for classification (clustering) purposes

25 BES III I/O patterns: simulation job height of the bar is proportional to the amount of bytes written file position Jobs writes its output file I/O operation sequence number PRELIMINARY

26 BES III I/O patterns: reconstruction job Job reads input raw data Job reads random trigger raw data

27 BES III I/O patterns: reconstruction job (cont.) Job writes the output file

28 Perspectives

29 Perspectives Use the OpenStack Swift testbed platform deployed at IHEP for performing larger scale tests 6 file servers + 2 head nodes, 24TB aggregated capacity Performance tests accessing data from remote sites uploading and downloading entire files, remote access of BES data via ROOT from long distance client Add support to ROOT for Rackspace files and integrate with Chinese cloud storage providers (Aliyun, China Telecom Cloud, )

30 Conclusions We have made progress understanding what is needed to integrate cloud-based storage for supporting high energy physics experiments both for experimental data and for individual s files Cloud storage paradigm and tools look promising for our data processing needs More work needs to be done to reach production readiness larger scale testing involve end-users, both experiments and individuals

31 Questions & Comments