Ceph-based storage services for Run2 and beyond

Size: px
Start display at page:

Download "Ceph-based storage services for Run2 and beyond"

Transcription

1 Journal of Physics: Conference Series PAPER OPEN ACCESS Ceph-based storage services for Run2 and beyond To cite this article: Daniel C. van der Ster et al 2015 J. Phys.: Conf. Ser View the article online for updates and enhancements. This content was downloaded from IP address on 23/01/2019 at 11:22

2 Ceph-based storage services for Run2 and beyond Daniel C. van der Ster, Massimo Lamanna, Luca Mascetti, Andreas J. Peters, Hervé Rousseau CERN, Geneva, Switzerland Abstract. In 2013, CERN IT evaluated then deployed a petabyte-scale Ceph cluster to support OpenStack use-cases in production. With now more than a year of smooth operations, we will present our experience and tuning best-practices. Beyond the cloud storage use-cases, we have been exploring Ceph-based services to satisfy the growing storage requirements during and after Run2. First, we have developed a Ceph back-end for CASTOR, allowing this service to deploy thin disk server nodes which act as gateways to Ceph; this feature marries the strong data archival and cataloging features of CASTOR with the resilient and high performance Ceph subsystem for disk. Second, we have developed RADOSFS, a lightweight storage API which builds a POSIX-like filesystem on top of the Ceph object layer. When combined with Xrootd, RADOSFS can offer a scalable object interface compatible with our HEP data processing applications. Lastly the same object layer is being used to build a scalable and inexpensive NFS service for several user communities. 1. Introduction The CERN IT Department has operated a 3 petabyte Ceph [1] cluster in production for roughly 18 months. With a primary use-case of hosting OpenStack [2] Cinder volumes and Glance images, the cluster presently hosts around 176 terabytes of user data, replicated 3 times. This data is exposed to end-user virtual machines via close to 3000 RADOS block devices, specifically as 1200 Cinder volumes and 1800 Glance images. The cluster is currently hosted in Meyrin, Switzerland and a second cluster is planned for CERN s data centre in Budapest. Over the course of these 18 months, our team has gathered a variety of operations experience, much of which will be summarized in this short article. Section 2 presents the OpenStack usecases in more detail, and highlights our conclusions regarding tuning Ceph for IOPS capacity. Section 3 summarizes ongoing work to exploit Ceph RADOS in particular to store physics data. Finally, we conclude in section Block Storage for OpenStack The model offered by OpenStack Cinder volumes, whereby users and developers can attach an arbitrarily large block device to a virtual machine, simplifies the operations of core IT and LHC experiment services. Notable examples include ATLAS PanDA, CMS GlideinWMS, and the CMS Webtools, all of which have replaced dedicated hardware with virtual machines and attached storage. In addition to most core IT services, the Data and Storage Services Group is in the process of migrating NFS users out of a proprietary filer appliance and onto a solution composed of ZFS-formatted volumes on virtual NFS servers. Nearly 100 terabytes of these mini- Filer servers have already been deployed. The CERN central CVMFS service also employs Content from this work may be used under the terms of the Creative Commons Attribution 3.0 licence. Any further distribution of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI. Published under licence by IOP Publishing Ltd 1

3 attached volumes in the Squid Proxy services, and work is currently underway to migrate the stratum zero data from a filer appliance onto attached volumes. Indeed, with close to 1000 attached volumes, it is clear that the virtual block device model is proving essential to the virtual CERN IT infrastructure IOPS and SSDs One key feature that makes Cinder volumes operable is QoS types, which allows operators to throttle the IOPS and bandwidth available to each attached volume. Paying attention to available and consumed IOPS is critical to a successful Ceph service. Using a write-ahead journal to guarantee write durability, Ceph acknowledges a write only after it has been persisted on all replicas. Therefore, the latency of the Ceph journal is central to the available IOPS on a given Ceph object storage daemon (ceph-osd). In our tests, we have concluded that the two core Ceph use-cases (namely, object and block storage), can be treated differently with regard to journal requirements. For object storage, overall throughput is normally more important than latency; co-locating the journal on the cephosd disk usually offers adequate performance, even after considering the double-write penalty coming from journalled writes. On the other hand, block storage has been observed to rely on a low small write latency in order to offer a responsive experience; we have concluded that a solid-state journal is essential for ceph-osd being used for block storage. In particular, we have observed a 5-10x increase in the 4kB write IOPS capacity after moving ceph-osd journals from a co-located spinning disk onto an SSD [3]. Additionally, SSD journals have been measured to decrease the small write latency from 50ms to around 5ms. Given that SSD journal devices are usually partitioned for many ceph-osd s, it is important to select a drive with adequate random write performance, and of course being a write-only journal the endurance offered by the drive should be evaluated carefully. The Ceph user community has concluded that the Intel DC S3700 is a good choice for this use-case; we use the 200GB version and store five ceph-osd journals per SSD Tuning Ceph and Linux One issue that our cluster suffered during the Dumpling release cycle was that the ceph-mon LevelDBs were ever increasing in size, sometimes by many gigabytes per day. As a workaround we scheduled a manual compaction on each ceph-mon every two days. Thankfully this issue seems to have been resolved after upgrading to the Firefly. Early in the life of the cluster we found that the ceph-mon reliability was highly dependent on the latency of the LevelDB filesystem; we currently store ceph-mon data directories on SSDs and no longer observe monitor elections that can plague a non-responsive cluster. For the ceph-osd processes, we have paid particular attention to the scrub and deep-scrub processes. By default, each placement group (PG) is scrubbed weekly, and this often happens at the same time each week. After observing that this thundering herd of scrub processes was hurting client latencies, we developed two strategies to mitigate the problem. First, we wrote a process to preemptively schedule PG scrubs in order to smear them out over the week and decrease the number of concurrent scrubs. Second, we requested changes to the ceph-osd process to allow the ceph-osd disk thread to be ionice d to the idle priority. With these two mechanisms applied, scrubbing is now generally transparent to our end-users. Handling many small files on a ceph-osd also exposes a few weaknesses in Linux, or at least in RHEL 6 systems. First, we have found the NUMA zone reclaim feature can lead to whole servers freezing for several minutes while the kernel drops the inode caches as a result of a NUMA zone running out of memory. We now have a general policy of disabling this feature with the sysctl vm.zone reclaim mode = 0. UpdateDB can be similarly disruptive if it is configured to index the ceph-osd FileStores; we now add /var/lib/ceph to the PRUNEPATHS in 2

4 /etc/updatedb.conf. Finally, Ceph is widely known to use an excessive number of threads and sockets; servers are recommended to increase kernel.pid max to its maximum possible value, , and clients are required to increase the allowed number of open files with ulimit -n to a value larger than Significant Incidents Our Ceph cluster has been exposed to several disk and host failures during 18 months of operations. Disks fail monthly on average, and this type of failure has been found to be completely transparent to our end-users. The SSD journals currently have a perfect reliability record, though the authors acknowledge that they are tempting fate by mentioning this statistic. On two occasions an entire ceph-osd host went offline. In both of these cases the host and data could be recovered within a few hours by an operator. After this experience we now disable automatic backfilling for host failures by setting the mon osd down out subtree limit option to host. In October 2014 we suffered a power cut and UPS failure which took down 3 out of 5 cephmon daemons. The cluster was down for 18 minutes as a result. In the postmortem, it was found that during this outage block device IOs were paused in flight and resumed once the cluster quorum was restored. No filesystem corruptions were reported, and we have therefore concluded that virtio-blk is superior to virtio-scsi, which would have timed out the IOs. In March 2015 a central router issue led to significant packet loss on the Ceph networks. This led to OSDs being marked up and down repeated over many hours, and around 20 OSDs reached a heartbeat timeout and committed suicide (a data safeguard feature). This outage required manual intervention to restart the down OSDs and backfilling was ongoing for around 12 hours. Again, no filesystem corruptions were reported after this incident. 3. Ceph for Physics Data In addition to block storage, we are also investigating how to exploit RADOS within our existing physics data stores [4]. CASTOR version added support for the RadosStriper; with this feature CASTOR disk servers would act as thin gateways to objects stored in RADOS [4]. Additionally, various strategies for using RADOS are being explored for the EOS Diamond release, included use-cases for both metadata and data [5]. For other corner cases, we have been evaluating an S3 service built with the Ceph RADOSGW daemon. Our deployment uses civetweb and haproxy for horizontal scalability and the most visible user is currently BOINC. We presently store more than 40 million objects behind the Ceph S3 service A 30 Petabyte Test These physics data use-cases demonstrate that larger Ceph clusters may soon be needed. However, we are not aware of production Ceph clusters above the 3-5 PB size. In order to evaluate Ceph s scalability, we provisioned 150 servers each with 48x 4TB drives and only 64GB of RAM 1 for a short term test [6]. After optimizing our Puppet configuration templates, we succeeded to deploy the cluster within roughly 2 days and then performed various performance tests which demonstrated up to 52 GB/s write performance and little performance degradation during backfilling. During the testing we exposed one scalability limitation in which ceph-osd processes consume an amount of memory proportional to the number of OSDs in the cluster. Ceph uses an osdmap to represent the cluster s structure, including the directory and statuses of OSDs, the CRUSH map, etc... Having 7200 OSDs, the osdmap was 4MB in size, and since each OSD caches up 1 This is acknowledged to be below the recommended 2GB RAM per OSD. 3

5 to 500 previous versions of the osdmap, each OSD consumed 3-4GB RAM. With 48 OSDs per server, this implied that at least 200GB of RAM is needed per server just to cache the osdmaps. We feel this is not an efficient usage of expensive RAM, especially given old osdmaps are rarely used in a healthy Ceph cluster. On the positive side, we found that by decrease several options related to the osdmap cache we were able to decrease memory usage to under 500MB per cephosd, thereby achieving a stable cluster. However, we are looking forward to any improved design which would remove this limitation. 4. Summary This short paper presented the operational status of Ceph in the CERN IT Department. The production cluster is currently 3PB in size and is primarily used for OpenStack Cinder and Glance block devices. This block storage service is proving to be quite popular and indispensable, with close to 1000 virtual machines relying on Ceph to provide attached volumes. Looking forward, we have ongoing developments in CASTOR and EOS to exploit Ceph object storage for better reliability and scalability. These projects motivated a 30 petabyte Ceph test in order to evaluate the solution at a scale relevant to physics data storage. This test demonstrated that operating Ceph at that scale is feasible with workarounds, however we identified limitations which prevent further scaling in terms of more ceph-osd processes. References [1] Weil, Sage et al. Ceph: A scalable, high-performance distributed file system. Proceedings of the 7th symposium on Operating systems design and implementation. USENIX Association, [2] Bell, Tim et al. Scaling the CERN OpenStack cloud. CHEP Okinawa, Japan. [3] Van der Ster, Daniel. Block Storage Service: Status and Performance. CERN-IT-Note June [4] Mascetti, Luca et al. Disk storage at CERN. CHEP Okinawa, Japan. [5] Peters, Andreas Joachim et al. Integrating CEPH in EOS. CHEP Okinawa, Japan. [6] Van der Ster, Daniel and Rousseau, Hervé. Ceph 30PB Test Report. CERN-IT-Note May

Block Storage Service: Status and Performance

Block Storage Service: Status and Performance Block Storage Service: Status and Performance Dan van der Ster, IT-DSS, 6 June 2014 Summary This memo summarizes the current status of the Ceph block storage service as it is used for OpenStack Cinder

More information

CERN European Organization for Nuclear Research, 1211 Geneva, CH

CERN European Organization for Nuclear Research, 1211 Geneva, CH Disk storage at CERN L Mascetti, E Cano, B Chan, X Espinal, A Fiorot, H González Labrador, J Iven, M Lamanna, G Lo Presti, JT Mościcki, AJ Peters, S Ponce, H Rousseau and D van der Ster CERN European Organization

More information

Ceph Intro & Architectural Overview. Abbas Bangash Intercloud Systems

Ceph Intro & Architectural Overview. Abbas Bangash Intercloud Systems Ceph Intro & Architectural Overview Abbas Bangash Intercloud Systems About Me Abbas Bangash Systems Team Lead, Intercloud Systems abangash@intercloudsys.com intercloudsys.com 2 CLOUD SERVICES COMPUTE NETWORK

More information

A fields' Introduction to SUSE Enterprise Storage TUT91098

A fields' Introduction to SUSE Enterprise Storage TUT91098 A fields' Introduction to SUSE Enterprise Storage TUT91098 Robert Grosschopff Senior Systems Engineer robert.grosschopff@suse.com Martin Weiss Senior Consultant martin.weiss@suse.com Joao Luis Senior Software

More information

Current Status of the Ceph Based Storage Systems at the RACF

Current Status of the Ceph Based Storage Systems at the RACF Journal of Physics: Conference Series PAPER OPEN ACCESS Current Status of the Ceph Based Storage Systems at the RACF To cite this article: A. Zaytsev et al 2015 J. Phys.: Conf. Ser. 664 042027 View the

More information

The Comparison of Ceph and Commercial Server SAN. Yuting Wu AWcloud

The Comparison of Ceph and Commercial Server SAN. Yuting Wu AWcloud The Comparison of Ceph and Commercial Server SAN Yuting Wu wuyuting@awcloud.com AWcloud Agenda Introduction to AWcloud Introduction to Ceph Storage Introduction to ScaleIO and SolidFire Comparison of Ceph

More information

Introduction to Ceph Speaker : Thor

Introduction to Ceph Speaker : Thor Introduction to Ceph Speaker : Thor Outline What s Ceph? Ceph Architecture Ceph Functions Ceph at CERN Ceph UI Ceph Solution Architectures What is Ceph?. Distributed storage system - Fault tolerant, no

More information

What's new in Jewel for RADOS? SAMUEL JUST 2015 VAULT

What's new in Jewel for RADOS? SAMUEL JUST 2015 VAULT What's new in Jewel for RADOS? SAMUEL JUST 2015 VAULT QUICK PRIMER ON CEPH AND RADOS CEPH MOTIVATING PRINCIPLES All components must scale horizontally There can be no single point of failure The solution

More information

ROCK INK PAPER COMPUTER

ROCK INK PAPER COMPUTER Introduction to Ceph and Architectural Overview Federico Lucifredi Product Management Director, Ceph Storage Boston, December 16th, 2015 CLOUD SERVICES COMPUTE NETWORK STORAGE the future of storage 2 ROCK

More information

Data preservation for the HERA experiments at DESY using dcache technology

Data preservation for the HERA experiments at DESY using dcache technology Journal of Physics: Conference Series PAPER OPEN ACCESS Data preservation for the HERA experiments at DESY using dcache technology To cite this article: Dirk Krücker et al 2015 J. Phys.: Conf. Ser. 66

More information

RED HAT CEPH STORAGE ROADMAP. Cesar Pinto Account Manager, Red Hat Norway

RED HAT CEPH STORAGE ROADMAP. Cesar Pinto Account Manager, Red Hat Norway RED HAT CEPH STORAGE ROADMAP Cesar Pinto Account Manager, Red Hat Norway cpinto@redhat.com THE RED HAT STORAGE MISSION To offer a unified, open software-defined storage portfolio that delivers a range

More information

Evaluation of the Huawei UDS cloud storage system for CERN specific data

Evaluation of the Huawei UDS cloud storage system for CERN specific data th International Conference on Computing in High Energy and Nuclear Physics (CHEP3) IOP Publishing Journal of Physics: Conference Series 53 (4) 44 doi:.88/74-6596/53/4/44 Evaluation of the Huawei UDS cloud

More information

클라우드스토리지구축을 위한 ceph 설치및설정

클라우드스토리지구축을 위한 ceph 설치및설정 클라우드스토리지구축을 위한 ceph 설치및설정 Ph.D. Sun Park GIST, NetCS Lab. 2015. 07. 15 1 목차 Cloud Storage Services? Open Source Cloud Storage Softwares Introducing Ceph Storage Ceph Installation & Configuration Automatic

More information

SolidFire and Ceph Architectural Comparison

SolidFire and Ceph Architectural Comparison The All-Flash Array Built for the Next Generation Data Center SolidFire and Ceph Architectural Comparison July 2014 Overview When comparing the architecture for Ceph and SolidFire, it is clear that both

More information

A Gentle Introduction to Ceph

A Gentle Introduction to Ceph A Gentle Introduction to Ceph Narrated by Tim Serong tserong@suse.com Adapted from a longer work by Lars Marowsky-Brée lmb@suse.com Once upon a time there was a Free and Open Source distributed storage

More information

INTRODUCTION TO CEPH. Orit Wasserman Red Hat August Penguin 2017

INTRODUCTION TO CEPH. Orit Wasserman Red Hat August Penguin 2017 INTRODUCTION TO CEPH Orit Wasserman Red Hat August Penguin 2017 CEPHALOPOD A cephalopod is any member of the molluscan class Cephalopoda. These exclusively marine animals are characterized by bilateral

More information

Visita delegazione ditte italiane

Visita delegazione ditte italiane Visita delegazione ditte italiane CERN IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it Massimo Lamanna/CERN IT department - Data Storage Services group Innovation in Computing in High-Energy

More information

Benchmark of a Cubieboard cluster

Benchmark of a Cubieboard cluster Benchmark of a Cubieboard cluster M J Schnepf, D Gudu, B Rische, M Fischer, C Jung and M Hardt Steinbuch Centre for Computing, Karlsruhe Institute of Technology, Karlsruhe, Germany E-mail: matthias.schnepf@student.kit.edu,

More information

CMS users data management service integration and first experiences with its NoSQL data storage

CMS users data management service integration and first experiences with its NoSQL data storage Journal of Physics: Conference Series OPEN ACCESS CMS users data management service integration and first experiences with its NoSQL data storage To cite this article: H Riahi et al 2014 J. Phys.: Conf.

More information

The DMLite Rucio Plugin: ATLAS data in a filesystem

The DMLite Rucio Plugin: ATLAS data in a filesystem Journal of Physics: Conference Series OPEN ACCESS The DMLite Rucio Plugin: ATLAS data in a filesystem To cite this article: M Lassnig et al 2014 J. Phys.: Conf. Ser. 513 042030 View the article online

More information

Streamlining CASTOR to manage the LHC data torrent

Streamlining CASTOR to manage the LHC data torrent Streamlining CASTOR to manage the LHC data torrent G. Lo Presti, X. Espinal Curull, E. Cano, B. Fiorini, A. Ieri, S. Murray, S. Ponce and E. Sindrilaru CERN, 1211 Geneva 23, Switzerland E-mail: giuseppe.lopresti@cern.ch

More information

Online data storage service strategy for the CERN computer Centre G. Cancio, D. Duellmann, M. Lamanna, A. Pace CERN, Geneva, Switzerland

Online data storage service strategy for the CERN computer Centre G. Cancio, D. Duellmann, M. Lamanna, A. Pace CERN, Geneva, Switzerland Online data storage service strategy for the CERN computer Centre G. Cancio, D. Duellmann, M. Lamanna, A. Pace CERN, Geneva, Switzerland Abstract. The Data and Storage Services group at CERN is conducting

More information

Provisioning with SUSE Enterprise Storage. Nyers Gábor Trainer &

Provisioning with SUSE Enterprise Storage. Nyers Gábor Trainer & Provisioning with SUSE Enterprise Storage Nyers Gábor Trainer & Consultant @Trebut gnyers@trebut.com Managing storage growth and costs of the software-defined datacenter PRESENT Easily scale and manage

More information

Distributed File Storage in Multi-Tenant Clouds using CephFS

Distributed File Storage in Multi-Tenant Clouds using CephFS Distributed File Storage in Multi-Tenant Clouds using CephFS FOSDEM 2018 John Spray Software Engineer Ceph Christian Schwede Software Engineer OpenStack Storage In this presentation Brief overview of key

More information

Evaluating Cloud Storage Strategies. James Bottomley; CTO, Server Virtualization

Evaluating Cloud Storage Strategies. James Bottomley; CTO, Server Virtualization Evaluating Cloud Storage Strategies James Bottomley; CTO, Server Virtualization Introduction to Storage Attachments: - Local (Direct cheap) SAS, SATA - Remote (SAN, NAS expensive) FC net Types - Block

More information

Introducing SUSE Enterprise Storage 5

Introducing SUSE Enterprise Storage 5 Introducing SUSE Enterprise Storage 5 1 SUSE Enterprise Storage 5 SUSE Enterprise Storage 5 is the ideal solution for Compliance, Archive, Backup and Large Data. Customers can simplify and scale the storage

More information

Discover CephFS TECHNICAL REPORT SPONSORED BY. image vlastas, 123RF.com

Discover CephFS TECHNICAL REPORT SPONSORED BY. image vlastas, 123RF.com Discover CephFS TECHNICAL REPORT SPONSORED BY image vlastas, 123RF.com Discover CephFS TECHNICAL REPORT The CephFS filesystem combines the power of object storage with the simplicity of an ordinary Linux

More information

CS-580K/480K Advanced Topics in Cloud Computing. Object Storage

CS-580K/480K Advanced Topics in Cloud Computing. Object Storage CS-580K/480K Advanced Topics in Cloud Computing Object Storage 1 When we use object storage When we check Facebook, twitter Gmail Docs on DropBox Check share point Take pictures with Instagram 2 Object

More information

virtual machine block storage with the ceph distributed storage system sage weil xensummit august 28, 2012

virtual machine block storage with the ceph distributed storage system sage weil xensummit august 28, 2012 virtual machine block storage with the ceph distributed storage system sage weil xensummit august 28, 2012 outline why you should care what is it, what it does how it works, how you can use it architecture

More information

CernVM-FS beyond LHC computing

CernVM-FS beyond LHC computing CernVM-FS beyond LHC computing C Condurache, I Collier STFC Rutherford Appleton Laboratory, Harwell Oxford, Didcot, OX11 0QX, UK E-mail: catalin.condurache@stfc.ac.uk Abstract. In the last three years

More information

Performance of popular open source databases for HEP related computing problems

Performance of popular open source databases for HEP related computing problems Journal of Physics: Conference Series OPEN ACCESS Performance of popular open source databases for HEP related computing problems To cite this article: D Kovalskyi et al 2014 J. Phys.: Conf. Ser. 513 042027

More information

A New Key-value Data Store For Heterogeneous Storage Architecture Intel APAC R&D Ltd.

A New Key-value Data Store For Heterogeneous Storage Architecture Intel APAC R&D Ltd. A New Key-value Data Store For Heterogeneous Storage Architecture Intel APAC R&D Ltd. 1 Agenda Introduction Background and Motivation Hybrid Key-Value Data Store Architecture Overview Design details Performance

More information

ATLAS Nightly Build System Upgrade

ATLAS Nightly Build System Upgrade Journal of Physics: Conference Series OPEN ACCESS ATLAS Nightly Build System Upgrade To cite this article: G Dimitrov et al 2014 J. Phys.: Conf. Ser. 513 052034 Recent citations - A Roadmap to Continuous

More information

Open vstorage RedHat Ceph Architectural Comparison

Open vstorage RedHat Ceph Architectural Comparison Open vstorage RedHat Ceph Architectural Comparison Open vstorage is the World s fastest Distributed Block Store that spans across different Datacenter. It combines ultrahigh performance and low latency

More information

Catalogic DPX TM 4.3. ECX 2.0 Best Practices for Deployment and Cataloging

Catalogic DPX TM 4.3. ECX 2.0 Best Practices for Deployment and Cataloging Catalogic DPX TM 4.3 ECX 2.0 Best Practices for Deployment and Cataloging 1 Catalogic Software, Inc TM, 2015. All rights reserved. This publication contains proprietary and confidential material, and is

More information

Samba and Ceph. Release the Kraken! David Disseldorp

Samba and Ceph. Release the Kraken! David Disseldorp Samba and Ceph Release the Kraken! David Disseldorp ddiss@samba.org Agenda Ceph Overview State of Samba Integration Performance Outlook Ceph Distributed storage system Scalable Fault tolerant Performant

More information

Use of containerisation as an alternative to full virtualisation in grid environments.

Use of containerisation as an alternative to full virtualisation in grid environments. Journal of Physics: Conference Series PAPER OPEN ACCESS Use of containerisation as an alternative to full virtualisation in grid environments. Related content - Use of containerisation as an alternative

More information

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model Journal of Physics: Conference Series The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model To cite this article: S González de la Hoz 2012 J. Phys.: Conf. Ser. 396 032050

More information

Finding a Needle in a Haystack. Facebook s Photo Storage Jack Hartner

Finding a Needle in a Haystack. Facebook s Photo Storage Jack Hartner Finding a Needle in a Haystack Facebook s Photo Storage Jack Hartner Paper Outline Introduction Background & Previous Design Design & Implementation Evaluation Related Work Conclusion Facebook Photo Storage

More information

Using S3 cloud storage with ROOT and CvmFS

Using S3 cloud storage with ROOT and CvmFS Journal of Physics: Conference Series PAPER OPEN ACCESS Using S cloud storage with ROOT and CvmFS To cite this article: María Arsuaga-Ríos et al 05 J. Phys.: Conf. Ser. 66 000 View the article online for

More information

The Legnaro-Padova distributed Tier-2: challenges and results

The Legnaro-Padova distributed Tier-2: challenges and results The Legnaro-Padova distributed Tier-2: challenges and results Simone Badoer a, Massimo Biasotto a,fulviacosta b, Alberto Crescente b, Sergio Fantinel a, Roberto Ferrari b, Michele Gulmini a, Gaetano Maron

More information

di-eos "distributed EOS": Initial experience with split-site persistency in a production service

di-eos distributed EOS: Initial experience with split-site persistency in a production service Journal of Physics: Conference Series OPEN ACCESS di-eos "distributed EOS": Initial experience with split-site persistency in a production service To cite this article: A J Peters et al 2014 J. Phys.:

More information

Striped Data Server for Scalable Parallel Data Analysis

Striped Data Server for Scalable Parallel Data Analysis Journal of Physics: Conference Series PAPER OPEN ACCESS Striped Data Server for Scalable Parallel Data Analysis To cite this article: Jin Chang et al 2018 J. Phys.: Conf. Ser. 1085 042035 View the article

More information

Data transfer over the wide area network with a large round trip time

Data transfer over the wide area network with a large round trip time Journal of Physics: Conference Series Data transfer over the wide area network with a large round trip time To cite this article: H Matsunaga et al 1 J. Phys.: Conf. Ser. 219 656 Recent citations - A two

More information

Dell Fluid Data solutions. Powerful self-optimized enterprise storage. Dell Compellent Storage Center: Designed for business results

Dell Fluid Data solutions. Powerful self-optimized enterprise storage. Dell Compellent Storage Center: Designed for business results Dell Fluid Data solutions Powerful self-optimized enterprise storage Dell Compellent Storage Center: Designed for business results The Dell difference: Efficiency designed to drive down your total cost

More information

Tests of PROOF-on-Demand with ATLAS Prodsys2 and first experience with HTTP federation

Tests of PROOF-on-Demand with ATLAS Prodsys2 and first experience with HTTP federation Journal of Physics: Conference Series PAPER OPEN ACCESS Tests of PROOF-on-Demand with ATLAS Prodsys2 and first experience with HTTP federation To cite this article: R. Di Nardo et al 2015 J. Phys.: Conf.

More information

SUSE OpenStack Cloud Production Deployment Architecture. Guide. Solution Guide Cloud Computing.

SUSE OpenStack Cloud Production Deployment Architecture. Guide. Solution Guide Cloud Computing. SUSE OpenStack Cloud Production Deployment Architecture Guide Solution Guide Cloud Computing Table of Contents page Introduction... 2 High Availability Configuration...6 Network Topography...8 Services

More information

The Echo Project An Update. Alastair Dewhurst, Alison Packer, George Vasilakakos, Bruno Canning, Ian Johnson, Tom Byrne

The Echo Project An Update. Alastair Dewhurst, Alison Packer, George Vasilakakos, Bruno Canning, Ian Johnson, Tom Byrne The Echo Project An Update Alastair Dewhurst, Alison Packer, George Vasilakakos, Bruno Canning, Ian Johnson, Tom Byrne What is Echo? Service S3 / Swift XrootD / GridFTP Ceph cluster with Erasure Coded

More information

WLCG Transfers Dashboard: a Unified Monitoring Tool for Heterogeneous Data Transfers.

WLCG Transfers Dashboard: a Unified Monitoring Tool for Heterogeneous Data Transfers. WLCG Transfers Dashboard: a Unified Monitoring Tool for Heterogeneous Data Transfers. J Andreeva 1, A Beche 1, S Belov 2, I Kadochnikov 2, P Saiz 1 and D Tuckett 1 1 CERN (European Organization for Nuclear

More information

System level traffic shaping in disk servers with heterogeneous protocols

System level traffic shaping in disk servers with heterogeneous protocols Journal of Physics: Conference Series OPEN ACCESS System level traffic shaping in disk servers with heterogeneous protocols To cite this article: Eric Cano and Daniele Francesco Kruse 14 J. Phys.: Conf.

More information

Security in the CernVM File System and the Frontier Distributed Database Caching System

Security in the CernVM File System and the Frontier Distributed Database Caching System Security in the CernVM File System and the Frontier Distributed Database Caching System D Dykstra 1 and J Blomer 2 1 Scientific Computing Division, Fermilab, Batavia, IL 60510, USA 2 PH-SFT Department,

More information

White Paper Features and Benefits of Fujitsu All-Flash Arrays for Virtualization and Consolidation ETERNUS AF S2 series

White Paper Features and Benefits of Fujitsu All-Flash Arrays for Virtualization and Consolidation ETERNUS AF S2 series White Paper Features and Benefits of Fujitsu All-Flash Arrays for Virtualization and Consolidation Fujitsu All-Flash Arrays are extremely effective tools when virtualization is used for server consolidation.

More information

Optimizing Ceph Object Storage For Production in Multisite Clouds

Optimizing Ceph Object Storage For Production in Multisite Clouds Optimizing Ceph Object Storage For Production in Multisite Clouds Do s and don ts of multisite object storage John Wilkins Michael Hackett May 9, 2018 Objectives High level understanding of Ceph Object

More information

Archive Solutions at the Center for High Performance Computing by Sam Liston (University of Utah)

Archive Solutions at the Center for High Performance Computing by Sam Liston (University of Utah) Archive Solutions at the Center for High Performance Computing by Sam Liston (University of Utah) The scale of the data housed at the Center for High Performance Computing (CHPC) has dramatically increased

More information

IBM Spectrum NAS, IBM Spectrum Scale and IBM Cloud Object Storage

IBM Spectrum NAS, IBM Spectrum Scale and IBM Cloud Object Storage IBM Spectrum NAS, IBM Spectrum Scale and IBM Cloud Object Storage Silverton Consulting, Inc. StorInt Briefing 2017 SILVERTON CONSULTING, INC. ALL RIGHTS RESERVED Page 2 Introduction Unstructured data has

More information

Rio-2 Hybrid Backup Server

Rio-2 Hybrid Backup Server A Revolution in Data Storage for Today s Enterprise March 2018 Notices This white paper provides information about the as of the date of issue of the white paper. Processes and general practices are subject

More information

A High-Availability Cloud for Research Computing

A High-Availability Cloud for Research Computing This is a post-print version of the following article: J. Riley, J. Noss, J. Cuff, I. M. Llorente, A High- Availability Cloud for Research Computing, IEEE Computer, pp: 91-95, Issue No. 06 - June (2017

More information

Evolution of Database Replication Technologies for WLCG

Evolution of Database Replication Technologies for WLCG Journal of Physics: Conference Series PAPER OPEN ACCESS Evolution of Database Replication Technologies for WLCG To cite this article: Zbigniew Baranowski et al 2015 J. Phys.: Conf. Ser. 664 042032 View

More information

Datacenter Storage with Ceph

Datacenter Storage with Ceph Datacenter Storage with Ceph John Spray john.spray@redhat.com jcsp on #ceph-devel Agenda What is Ceph? How does Ceph store your data? Interfaces to Ceph: RBD, RGW, CephFS Latest development updates Datacenter

More information

-Presented By : Rajeshwari Chatterjee Professor-Andrey Shevel Course: Computing Clusters Grid and Clouds ITMO University, St.

-Presented By : Rajeshwari Chatterjee Professor-Andrey Shevel Course: Computing Clusters Grid and Clouds ITMO University, St. -Presented By : Rajeshwari Chatterjee Professor-Andrey Shevel Course: Computing Clusters Grid and Clouds ITMO University, St. Petersburg Introduction File System Enterprise Needs Gluster Revisited Ceph

More information

ATLAS Tracking Detector Upgrade studies using the Fast Simulation Engine

ATLAS Tracking Detector Upgrade studies using the Fast Simulation Engine Journal of Physics: Conference Series PAPER OPEN ACCESS ATLAS Tracking Detector Upgrade studies using the Fast Simulation Engine To cite this article: Noemi Calace et al 2015 J. Phys.: Conf. Ser. 664 072005

More information

THE CEPH POWER SHOW. Episode 2 : The Jewel Story. Daniel Messer Technical Marketing Red Hat Storage. Karan Singh Sr. Storage Architect Red Hat Storage

THE CEPH POWER SHOW. Episode 2 : The Jewel Story. Daniel Messer Technical Marketing Red Hat Storage. Karan Singh Sr. Storage Architect Red Hat Storage THE CEPH POWER SHOW Episode 2 : The Jewel Story Karan Singh Sr. Storage Architect Red Hat Storage Daniel Messer Technical Marketing Red Hat Storage Kyle Bader Sr. Storage Architect Red Hat Storage AGENDA

More information

Why software defined storage matters? Sergey Goncharov Solution Architect, Red Hat

Why software defined storage matters? Sergey Goncharov Solution Architect, Red Hat Why software defined storage matters? Sergey Goncharov Solution Architect, Red Hat sgonchar@redhat.com AGENDA Storage and Datacenter evolution Red Hat Storage portfolio Red Hat Gluster Storage Red Hat

More information

Nutanix Tech Note. Virtualizing Microsoft Applications on Web-Scale Infrastructure

Nutanix Tech Note. Virtualizing Microsoft Applications on Web-Scale Infrastructure Nutanix Tech Note Virtualizing Microsoft Applications on Web-Scale Infrastructure The increase in virtualization of critical applications has brought significant attention to compute and storage infrastructure.

More information

Ceph: A Scalable, High-Performance Distributed File System PRESENTED BY, NITHIN NAGARAJ KASHYAP

Ceph: A Scalable, High-Performance Distributed File System PRESENTED BY, NITHIN NAGARAJ KASHYAP Ceph: A Scalable, High-Performance Distributed File System PRESENTED BY, NITHIN NAGARAJ KASHYAP Outline Introduction. System Overview. Distributed Object Storage. Problem Statements. What is Ceph? Unified

More information

The Fastest And Most Efficient Block Storage Software (SDS)

The Fastest And Most Efficient Block Storage Software (SDS) The Fastest And Most Efficient Block Storage Software (SDS) StorPool: Product Summary 1. Advanced Block-level Software Defined Storage, SDS (SDS 2.0) Fully distributed, scale-out, online changes of everything,

More information

The Google File System

The Google File System The Google File System Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung December 2003 ACM symposium on Operating systems principles Publisher: ACM Nov. 26, 2008 OUTLINE INTRODUCTION DESIGN OVERVIEW

More information

New Approach to Unstructured Data

New Approach to Unstructured Data Innovations in All-Flash Storage Deliver a New Approach to Unstructured Data Table of Contents Developing a new approach to unstructured data...2 Designing a new storage architecture...2 Understanding

More information

ECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective

ECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective ECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective Part II: Software Infrastructure in Data Centers: Distributed File Systems 1 Permanently stores data Filesystems

More information

The Google File System

The Google File System October 13, 2010 Based on: S. Ghemawat, H. Gobioff, and S.-T. Leung: The Google file system, in Proceedings ACM SOSP 2003, Lake George, NY, USA, October 2003. 1 Assumptions Interface Architecture Single

More information

Experiences with the new ATLAS Distributed Data Management System

Experiences with the new ATLAS Distributed Data Management System Experiences with the new ATLAS Distributed Data Management System V. Garonne 1, M. Barisits 2, T. Beermann 2, M. Lassnig 2, C. Serfon 1, W. Guan 3 on behalf of the ATLAS Collaboration 1 University of Oslo,

More information

MarkLogic Server. Scalability, Availability, and Failover Guide. MarkLogic 9 May, Copyright 2018 MarkLogic Corporation. All rights reserved.

MarkLogic Server. Scalability, Availability, and Failover Guide. MarkLogic 9 May, Copyright 2018 MarkLogic Corporation. All rights reserved. Scalability, Availability, and Failover Guide 1 MarkLogic 9 May, 2017 Last Revised: 9.0-4, January, 2018 Copyright 2018 MarkLogic Corporation. All rights reserved. Table of Contents Table of Contents Scalability,

More information

Ceph Block Devices: A Deep Dive. Josh Durgin RBD Lead June 24, 2015

Ceph Block Devices: A Deep Dive. Josh Durgin RBD Lead June 24, 2015 Ceph Block Devices: A Deep Dive Josh Durgin RBD Lead June 24, 2015 Ceph Motivating Principles All components must scale horizontally There can be no single point of failure The solution must be hardware

More information

Storage Resource Sharing with CASTOR.

Storage Resource Sharing with CASTOR. Storage Resource Sharing with CASTOR Olof Barring, Benjamin Couturier, Jean-Damien Durand, Emil Knezo, Sebastien Ponce (CERN) Vitali Motyakov (IHEP) ben.couturier@cern.ch 16/4/2004 Storage Resource Sharing

More information

Ceph at the DRI. Peter Tiernan Systems and Storage Engineer Digital Repository of Ireland TCHPC

Ceph at the DRI. Peter Tiernan Systems and Storage Engineer Digital Repository of Ireland TCHPC Ceph at the DRI Peter Tiernan Systems and Storage Engineer Digital Repository of Ireland TCHPC DRI: The Digital Repository Of Ireland (DRI) is an interactive, national trusted digital repository for contemporary

More information

All-NVMe Performance Deep Dive Into Ceph + Sneak Preview of QLC + NVMe Ceph

All-NVMe Performance Deep Dive Into Ceph + Sneak Preview of QLC + NVMe Ceph All-NVMe Performance Deep Dive Into Ceph + Sneak Preview of QLC + NVMe Ceph Ryan Meredith Sr. Manager, Storage Solutions Engineering 2018 Micron Technology, Inc. All rights reserved. Information, products,

More information

StorageCraft OneXafe and Veeam 9.5

StorageCraft OneXafe and Veeam 9.5 TECHNICAL DEPLOYMENT GUIDE NOV 2018 StorageCraft OneXafe and Veeam 9.5 Expert Deployment Guide Overview StorageCraft, with its scale-out storage solution OneXafe, compliments Veeam to create a differentiated

More information

Table of Contents GEEK GUIDE CEPH: OPEN-SOURCE SDS

Table of Contents GEEK GUIDE CEPH: OPEN-SOURCE SDS Table of Contents About the Sponsor... 4 Introduction... 5 Overview: What Is Software-Defined Storage?... 6 Benefits of SDS... 9 Single Management Interface...9 Reduced CAPEX...10 Scalable...11 No Single

More information

Understanding Write Behaviors of Storage Backends in Ceph Object Store

Understanding Write Behaviors of Storage Backends in Ceph Object Store Understanding Write Behaviors of Storage Backends in Object Store Dong-Yun Lee, Kisik Jeong, Sang-Hoon Han, Jin-Soo Kim, Joo-Young Hwang and Sangyeun Cho How Amplifies Writes client Data Store, please

More information

HammerCloud: A Stress Testing System for Distributed Analysis

HammerCloud: A Stress Testing System for Distributed Analysis HammerCloud: A Stress Testing System for Distributed Analysis Daniel C. van der Ster 1, Johannes Elmsheuser 2, Mario Úbeda García 1, Massimo Paladin 1 1: CERN, Geneva, Switzerland 2: Ludwig-Maximilians-Universität

More information

Scaling Without Sharding. Baron Schwartz Percona Inc Surge 2010

Scaling Without Sharding. Baron Schwartz Percona Inc Surge 2010 Scaling Without Sharding Baron Schwartz Percona Inc Surge 2010 Web Scale!!!! http://www.xtranormal.com/watch/6995033/ A Sharding Thought Experiment 64 shards per proxy [1] 1 TB of data storage per node

More information

MySQL and Ceph. A tale of two friends

MySQL and Ceph. A tale of two friends ysql and Ceph A tale of two friends Karan Singh Sr. Storage Architect Red Hat Taco Scargo Sr. Solution Architect Red Hat Agenda Ceph Introduction and Architecture Why ysql on Ceph ysql and Ceph Performance

More information

StorMagic SvSAN: A virtual SAN made simple

StorMagic SvSAN: A virtual SAN made simple Data Sheet StorMagic SvSAN: A virtual SAN made simple StorMagic SvSAN SvSAN is a software-defined storage solution designed to run on two or more servers. It is uniquely architected with the combination

More information

StorageCraft OneBlox and Veeam 9.5 Expert Deployment Guide

StorageCraft OneBlox and Veeam 9.5 Expert Deployment Guide TECHNICAL DEPLOYMENT GUIDE StorageCraft OneBlox and Veeam 9.5 Expert Deployment Guide Overview StorageCraft, with its scale-out storage solution OneBlox, compliments Veeam to create a differentiated diskbased

More information

TCC, so your business continues

TCC, so your business continues TCC, so your business continues 1 Whitepaper Executive summary The TCC (Tuxis Cloud Concept) is a smart way to build hosting infrastructures for your private cloud and virtual datacenter. It natively integrates

More information

Ambry: LinkedIn s Scalable Geo- Distributed Object Store

Ambry: LinkedIn s Scalable Geo- Distributed Object Store Ambry: LinkedIn s Scalable Geo- Distributed Object Store Shadi A. Noghabi *, Sriram Subramanian +, Priyesh Narayanan +, Sivabalan Narayanan +, Gopalakrishna Holla +, Mammad Zadeh +, Tianwei Li +, Indranil

More information

Distributed File Storage in Multi-Tenant Clouds using CephFS

Distributed File Storage in Multi-Tenant Clouds using CephFS Distributed File Storage in Multi-Tenant Clouds using CephFS Openstack Vancouver 2018 May 23 Patrick Donnelly CephFS Engineer Red Hat, Inc. Tom Barron Manila Engineer Red Hat, Inc. Ramana Raja CephFS Engineer

More information

Storage on the Lunatic Fringe. Thomas M. Ruwart University of Minnesota Digital Technology Center Intelligent Storage Consortium

Storage on the Lunatic Fringe. Thomas M. Ruwart University of Minnesota Digital Technology Center Intelligent Storage Consortium Storage on the Lunatic Fringe Thomas M. Ruwart University of Minnesota Digital Technology Center Intelligent Storage Consortium tmruwart@dtc.umn.edu Orientation Who are the lunatics? What are their requirements?

More information

The CMS data quality monitoring software: experience and future prospects

The CMS data quality monitoring software: experience and future prospects The CMS data quality monitoring software: experience and future prospects Federico De Guio on behalf of the CMS Collaboration CERN, Geneva, Switzerland E-mail: federico.de.guio@cern.ch Abstract. The Data

More information

Deploying Software Defined Storage for the Enterprise with Ceph. PRESENTATION TITLE GOES HERE Paul von Stamwitz Fujitsu

Deploying Software Defined Storage for the Enterprise with Ceph. PRESENTATION TITLE GOES HERE Paul von Stamwitz Fujitsu Deploying Software Defined Storage for the Enterprise with Ceph PRESENTATION TITLE GOES HERE Paul von Stamwitz Fujitsu Agenda Yet another attempt to define SDS Quick Overview of Ceph from a SDS perspective

More information

System upgrade and future perspective for the operation of Tokyo Tier2 center. T. Nakamura, T. Mashimo, N. Matsui, H. Sakamoto and I.

System upgrade and future perspective for the operation of Tokyo Tier2 center. T. Nakamura, T. Mashimo, N. Matsui, H. Sakamoto and I. System upgrade and future perspective for the operation of Tokyo Tier2 center, T. Mashimo, N. Matsui, H. Sakamoto and I. Ueda International Center for Elementary Particle Physics, The University of Tokyo

More information

Red Hat Gluster Storage performance. Manoj Pillai and Ben England Performance Engineering June 25, 2015

Red Hat Gluster Storage performance. Manoj Pillai and Ben England Performance Engineering June 25, 2015 Red Hat Gluster Storage performance Manoj Pillai and Ben England Performance Engineering June 25, 2015 RDMA Erasure Coding NFS-Ganesha New or improved features (in last year) Snapshots SSD support Erasure

More information

vsan Disaster Recovery November 19, 2017

vsan Disaster Recovery November 19, 2017 November 19, 2017 1 Table of Contents 1. Disaster Recovery 1.1.Overview 1.2.vSAN Stretched Clusters and Site Recovery Manager 1.3.vSAN Performance 1.4.Summary 2 1. Disaster Recovery According to the United

More information

HCI: Hyper-Converged Infrastructure

HCI: Hyper-Converged Infrastructure Key Benefits: Innovative IT solution for high performance, simplicity and low cost Complete solution for IT workloads: compute, storage and networking in a single appliance High performance enabled by

More information

Geographically Distributed Software Defined Storage (the proposal)

Geographically Distributed Software Defined Storage (the proposal) Geographically Distributed Software Defined Storage (the proposal) A. Y. Shevel 1,2,a, S. E. Khoruzhnikov 1, V.A. Grudinin 1, O. L. Sadov 1, A. B. Kairkanov 1 1 ITMO University, 49, Kronverksky prospect,

More information

The Software Defined Online Storage System at the GridKa WLCG Tier-1 Center

The Software Defined Online Storage System at the GridKa WLCG Tier-1 Center The Software Defined Online Storage System at the GridKa WLCG Tier-1 Center CHEP 2018, Sofia Jan Erik Sundermann, Jolanta Bubeliene, Ludmilla Obholz, Andreas Petzold STEINBUCH CENTRE FOR COMPUTING (SCC)

More information

Delivering unprecedented performance, efficiency and flexibility to modernize your IT

Delivering unprecedented performance, efficiency and flexibility to modernize your IT SvSAN 6 Data Sheet Delivering unprecedented performance, efficiency and flexibility to modernize your IT StorMagic SvSAN SvSAN is a software-defined storage solution designed to run on two or more servers.

More information

Ceph. The link between file systems and octopuses. Udo Seidel. Linuxtag 2012

Ceph. The link between file systems and octopuses. Udo Seidel. Linuxtag 2012 Ceph OR The link between file systems and octopuses Udo Seidel Agenda Background CephFS CephStorage Summary Ceph what? So-called parallel distributed cluster file system Started as part of PhD studies

More information

Red Hat Ceph Storage and Samsung NVMe SSDs for intensive workloads

Red Hat Ceph Storage and Samsung NVMe SSDs for intensive workloads Red Hat Ceph Storage and Samsung NVMe SSDs for intensive workloads Power emerging OpenStack use cases with high-performance Samsung/ Red Hat Ceph reference architecture Optimize storage cluster performance

More information

Geant4 Computing Performance Benchmarking and Monitoring

Geant4 Computing Performance Benchmarking and Monitoring Journal of Physics: Conference Series PAPER OPEN ACCESS Geant4 Computing Performance Benchmarking and Monitoring To cite this article: Andrea Dotti et al 2015 J. Phys.: Conf. Ser. 664 062021 View the article

More information