Open vstorage RedHat Ceph Architectural Comparison
|
|
- Preston Greer
- 5 years ago
- Views:
Transcription
1 Open vstorage RedHat Ceph Architectural Comparison Open vstorage is the World s fastest Distributed Block Store that spans across different Datacenter. It combines ultrahigh performance and low latency connections with a data integrity that has no comparison. Data is distributed across datacenters using both Replication and Erasure Coding. Joining Performance and Integrity is not a simple bolt-on solution and requires a from-the-ground-up approach. Disk Failures, Node Failures and even Datacenter Failures do not present data loss and hence do not threaten any of your Data Integrity. You have been lead to believe that in order to have a 100% Data Loss Protection you have to compromise on Performance. While this might sound logical and acceptable, in is time to step out of the box and demand a noncompromise Storage Platform. With Open vstorage you can have your cake and eat it too! This document provides an overview of the Nexenta architecture and highlights the difference with the Open vstorage architecture. The comparison is not intended to be exhaustive, but covers the most relevant items where both solutions differ as seen from a customer perspective. Antwerpse Steenweg 19, 9080 Lochristi Belgium Phone: Mail: Info@openvstorage.com
2 Introduction Ceph, now part of RedHat, is an open-source, distributed object store designed to provide excellent performance, reliability and scalability. Next to scalable object storage interface Ceph offers a file system and block storage interface. All these interfaces make Ceph a popular storage solution. It has even been labeled the de facto storage backend for OpenStack 1. This should not come as a surprise as 57% of the OpenStack users and 45% of the cores deployments are using Ceph RBD as block storage solution 2. But although it is the most used OpenStack block storage solution, there are other solutions on the market which might be able to push Ceph from its OpenStack throne. Open vstorage takes a different approach which allows it to beat the performance of Ceph and while offering more functionality. Architectural Design RedHat Ceph At the core of Ceph is the RADOS object-based storage system. It provides the foundation for some of Ceph s advanced features such as RADOS Block Device (RBD), the RADOS S3 Gateway and the Ceph File System. The RADOS architecture consists out of a cluster map, a diagram explaining how nodes and disks are distributed, a set of OSDs (Object Storage Daemons) and monitors. The cluster map is built upon the CRUSH (Controlled Replication Under Scalable Hashing) algorithm to efficiently compute information about data location, instead of having to depend on a central lookup table. This is a critical part as the CRUSH algorithm utilizes a hash function that makes the mapping completely deterministic and calculable. The OSDs are responsible for storing the actual data on disks. The monitors contain the cluster map and clients can retrieve the map from the monitors. Ceph clients such as RBD run on the client side and write objects to the OSDs based upon the cluster map received from the monitors
3 In many cases RBD, the block interface, is used to provide storage for Virtual Machines by offering a virtual block device to the hypervisor. Underlying RBD splits the virtual block device into 4MB chunks that are stored as objects on the RADOS object store. This means a 4K write ends up as a 4MB write on the backend. Data on a Ceph backend is typically stored 3 times as this is the default policy. But let s have a look at the write flow in detail. The client writes to the primary OSD, a disk of the Ceph cluster. As soon as the data is written to the first OSD, it is written to 2 additional OSDs. Once both the secondary and tertiary OSD acknowledge the write to the primary OSD, the primary OSD will acknowledge to the client that the write is safely stored on the backend. Now let s have a closer look as how data is stored on the OSDs. As soon as IO goes into the OSD, it gets written twice as the first IO is done to the Ceph journal and later it is written a second time to the backend file system for safeguarding. If the data is safely stored on the backend file system, the journal gets trimmed to reclaim space. The same dual write also applies to the secondary and tertiary copy so a single write IO is written 6 times, twice per OSD multiplied by 3 different OSDs, before it can be acknowledged to the application.
4 Open vstorage The basic components of Open vstorage are the Open vstorage Edge, the Open vstorage Volume Driver and ALBA. The Open vstorage Edge exposes block devices to applications that need a volume. The Edge component communicates across RDMA, a low-latency, high-throughput networking protocol, directly with the memory of the server running the Volume Driver. The Volume Driver is the technology that converts block storage into objects (Storage Container Objects) that can be stored on the ALBA backend. This ALBA backend is a special purpose object storage solution and is made up out of storage nodes that run ALBA daemons. These daemons manage the local storage devices (HDDs, SSDs, PCIe flash cards...) and contribute these devices to the storage pools. The Volume Driver is a combination of a locationbased approach (performance) and a log structured approach (unlimited history, unlimited snapshots, thin cloning). Each incoming 4K write is appended to the write buffer. This write buffer can actually be seen as a Transaction Log storing one or more Storage Container Objects, a consecutive group of incoming writes. The incoming write is also dispatched to a Transaction Log onto a second node in the cluster to prevent data loss. This principle of dispatching incoming writes to an additional write buffer is referred to as the Open vstorage Distributed Transaction Log (DTL). It is important to note that both the write buffer and the DTL are very small in size, as they only need to hold data that is not yet protected by the ALBA backend. They are typically limited to 256MB per volume.
5 Once a SCO is full, it gets split into chunks, compressed, encrypted and spread across the ALBA backend for redundancy. This ALBA backend is typically built on top of a pool of large capacity drives with an allflash performance tier as acceleration layer. On reads the Volume Driver fetches the correct block directly from the right chunk on the ALBA backend across RDMA. Similarities Software-defined storage Both Ceph and Open vstorage are software-defined storage solutions. They are infrastructure agnostic and can run on any x86 hardware. The actual specifications of the hardware have of course a huge impact on the performance numbers you can achieve. Both solutions use SSDs to speed up the performance and use SATA drives to store the data in the backend. Scalability Both solutions are built to scale capacity wise from a couple of TB up to a few petabyte. Each node that gets added to the cluster puts its capacity at the disposal of the storage pool. For example CERN is running a single production cluster of 3PB based upon Ceph and has tested a cluster of 30PB3. The same applies to Open vstorage. A cluster can grow as big as 1024 nodes and up to 30 PB. Performance of both solutions also scales with every node added as more drives are added to distribute the IO load. 3
6 Key Differences Deterministic vs Explicit Meta Data With Ceph the placement of data on the backend is computation based (CRUSH), this makes the whole process repeatable and deterministic. So no lookup on a hash table, only formulas and calculations are used to determine which disks contain the object. Open vstorage on the other hand uses explicit metadata as it stores the exact location of the objects in the backend in a distributed database. Note that based upon the data in the backend, the distributed database can always be reconstructed in case there would be an issue with the database. Open vstorage has optimized the footprint of the metadata set and manages to keep the whole set in RAM for performance reasons. While having explicit metadata consumes more storage space, the benefits of explicit metadata are substantial: In case of a broken disk Open vstorage can easily regenerate the data from the broken disks, as it knows exactly what was on the disk. It even knows which bits of data are most at risk due to many missing chunks and can repair the most vulnerable data first. As Ceph doesn t keep track of what is on the disks, it needs to wait until scrubbing picks up that objects are missing. As Open vstorage stores backend metadata it can offer APE (Asynchronous Policy Enforcement) which allows to spread data according to the best available policy. In case data can t be written optimally, it will be first stored sub-optimal and later, when for example nodes are replaced or the network link between the datacenters is restored, data will be re-written with the most optimal policy. Ceph doesn t have this type of features and will just stop all writes. Huge Performance Difference A lot has been said about the performance of Ceph or better the lack thereof. Execute a Google Search for Ceph Performance Tuning and you will find 25,000 suggestions to improve the performance. The truth is that Ceph was designed for scalability and wasn t developed to be the fastest distributed block storage solution. Just to give you some numbers; running a cluster of around k SATA HDD drives will sadly only give you roughly IOPs. Even with all-flash hardware, the random write performance isn t impressive. On an InfiniFlash Sandisk server with 64 SSDs, Ceph can only reach 180,000 IOPS for 4K random writes4. On top, since Ceph only acknowledges the write after 3 disks have acknowledged the write, a single slow disk will impact all VMs as clients spread data across all disks in the cluster. 4 ance_optimization.pptx
7 Open vstorage performance of a single node. Internal benchmarks show that Open vstorage can be 10x faster than Ceph on the same hardware. This huge performance difference can be explained by the fact that both solutions have a radical different design: Ceph uses a location based approach where each volume is divided into 4MB fragments and writes are location based hence happen randomly. Open vstorage uses a log structured approach using immutable objects that creates a lock less design and leverages multiple cores to get very high performance. This approach leads to lower latency and hence better performance. Erasure Coding Since the Hammer release, April 2015, Ceph supports erasure coding on the backend. But, as Ceph works location based and uses 4MB extents, the update of a 4k block means that the whole 4MB extent to which the block belongs needs to be erasure coded again. This is a very high penalty. One could of course lower the size of an extent but that will make the metadata grow out of control. Basically erasure coding isn t usable with RBD so replication must be used. Open vstorage on the other hand uses a log-structured approach where consecutive writes are appended to an immutable Storage Container Object (SCO). This SCO can be erasure coded by means of forward error correction techniques and the resulting fragments can be stored across multiple hosts or even datacenters. Space Efficiency As discussed earlier, Ceph can only work with an X-way replication strategy in case RBD is used. By default Ceph proposes a 3-way replication strategy. This isn t very space efficient, as every 4MB will lead to 12MB being stored on the physical disks. This means three times the amount of nodes, three times the amount of networking equipment, three times the amount of power and cooling, basically tripling the TCO. Open vstorage is much more space efficient as it uses forward error correction which provides 100x better reliability as compared to 3 copies although storing only 1.25x more data. This means that for every 1MB, 1.25MB will be saved on the backend. Secondly, Open vstorage allows for the creation of different error correction policies for flash and HDD and hence the customer can select a lower redundancy factor on flash and a higher redundancy factor on HDD. In addition, contrary to Ceph, Open vstorage can further reduce its storage footprint by compressing the fragments before storing them on
8 SSDs and HDDs. Ceph has compression functionality but since updating a 4K block means the whole 4MB extent needs to be compressed again, it makes no sense to enable it for RBD. Multi datacenter Ceph is designed to be used on local networks, as it requires low latency links between hosts to be performant. Since the writes need to be acknowledged by 3 storage nodes, the latency for writes would be too high in case the secondary or tertiary write would have to go to other datacenters. Open vstorage on the other hand is designed to store data safely and efficiently across multiple datacenters. To ensure local performance it uses an all-flash tier per datacenter while for disaster recovery reasons data can be stored on a capacity tier across multiple datacenters. The fact that Open vstorage is designed to be used in a multi datacenter approach is also visible in that Open vstorage supports 2 levels of erasure coding: a global and a local policy. The local policy makes sure that a single disk or host failure doesn t require data from other datacenters to be sent over the network in order to rebuild the missing chunks of data. As Ceph relies on replication the chunks will need to come from a different datacenter to replace the missing replicas. This puts additional stress on the network. Encryption of the backend Data security is an essential aspect of a large-scale storage implementation. Encryption, encoding the data with a secret key, is the most effective way to achieve data security. Open vstorage encrypts the data before the SCOs are transferred to the backend. This means that data being transferred over the network to the backend is already encrypted. Ceph on the other hand only encrypts data when it is being stored on the OSD. This means data is in a clear format when it is transferred over the network. This approach poses a higher risk as data might get wiretapped while in transit to the backend. Snapshots and Clones For Open vstorage taking a snapshot is as simple as placing a marker in the metadata. Therefore it takes less than a millisecond to complete a snapshot. Due to the design, taking a snapshot in Ceph involves a lot more work as Ceph works location based. When a snapshot is created and data is being overwritten, the old data needs to be copied, a copy-on-write process, to safeguard the overwritten data. This adds additional latency to the write path. Both solutions offer thin cloning but both implement this feature quite differently. Ceph uses an approach where there is no tracking of which objects exist in a clone. A read that hits a non-existent object will attempt to read from the parent snapshot, and this will continue recursively until an object exists or an image with no parent is found. It goes without saying that each lookup attempt adds additional latency to the read path. Open vstorage on the other hand keeps track of the metadata of the volume. Finding the data only requires to look up the right SCO in the metadata. Even if the block being read is in the root snapshot and the volume has been cloned multiple times, the block can be fetched immediately from the backend with a single read operation. This single-hop-read allows Open vstorage to minimize the read latency even for the most complex clone trees.
9 Flash Friendly As Ceph is location based it requires high endurance flash technology as random writes, generate small updates across the device. Many updates of the same location within the volume will also cause the flash chips to wear out faster than normal as you are constantly updating the same cells of the flash memory. Open vstorage is much more flash friendly as the write buffer sequentializes the random writes in large fragments that can be written sequentially. Due to random write amplification, sequential write endurance for SSDs and PCIe flash cards is typically higher so the flash drives are less likely to fail with Open vstorage compared to Ceph. As every update to a volume also gets appended to a new SCO, overwriting the same LBA of a volume many times will hammer the same cells of the flash memory over and over. Since Open vstorage is less intensive on flash as compared to Ceph, cheaper, lower endurance SSDs can be used for the performance tier. The bottom line Open vstorage and Ceph are software based distributed block storage solutions and they both scale very well. Although both can be installed on the same hardware, the different design approaches result in a remarkable performance difference. Based upon tests it has been proven that Open vstorage can be 10 times faster than Ceph on the same hardware. Next to the difference in performance, Open vstorage is more space efficient. Ceph uses replication to protect the data while Open vstorage can use both erasure coding and replication. This allows Open vstorage to store data across datacenters and survive even a complete datacenter going offline.
Open vstorage EMC SCALEIO Architectural Comparison
Open vstorage EMC SCALEIO Architectural Comparison Open vstorage is the World s fastest Distributed Block Store that spans across different Datacenter. It combines ultrahigh performance and low latency
More informationTurning Object. Storage into Virtual Machine Storage. White Papers
Turning Object Open vstorage is the World s fastest Distributed Block Store that spans across different Datacenter. It combines ultrahigh performance and low latency connections with a data integrity that
More informationSolidFire and Ceph Architectural Comparison
The All-Flash Array Built for the Next Generation Data Center SolidFire and Ceph Architectural Comparison July 2014 Overview When comparing the architecture for Ceph and SolidFire, it is clear that both
More informationA Gentle Introduction to Ceph
A Gentle Introduction to Ceph Narrated by Tim Serong tserong@suse.com Adapted from a longer work by Lars Marowsky-Brée lmb@suse.com Once upon a time there was a Free and Open Source distributed storage
More informationA product by CloudFounders. Wim Provoost Open vstorage
A product by CloudFounders Wim Provoost (@wimpers_be) Open vstorage (@openvstorage) http://www.openvstorage.com CloudFounders vrun Converged infrastructure that combines the benefits of the hyperconverged
More informationVMware Virtual SAN Technology
VMware Virtual SAN Technology Today s Agenda 1 Hyper-Converged Infrastructure Architecture & Vmware Virtual SAN Overview 2 Why VMware Hyper-Converged Software? 3 VMware Virtual SAN Advantage Today s Agenda
More informationDisclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme
STO1926BU A Day in the Life of a VSAN I/O Diving in to the I/O Flow of vsan John Nicholson (@lost_signal) Pete Koehler (@vmpete) VMworld 2017 Content: Not for publication #VMworld #STO1926BU Disclaimer
More informationNimble Storage Adaptive Flash
Nimble Storage Adaptive Flash Read more Nimble solutions Contact Us 800-544-8877 solutions@microage.com MicroAge.com TECHNOLOGY OVERVIEW Nimble Storage Adaptive Flash Nimble Storage s Adaptive Flash platform
More informationThe Google File System
October 13, 2010 Based on: S. Ghemawat, H. Gobioff, and S.-T. Leung: The Google file system, in Proceedings ACM SOSP 2003, Lake George, NY, USA, October 2003. 1 Assumptions Interface Architecture Single
More informationRED HAT CEPH STORAGE ROADMAP. Cesar Pinto Account Manager, Red Hat Norway
RED HAT CEPH STORAGE ROADMAP Cesar Pinto Account Manager, Red Hat Norway cpinto@redhat.com THE RED HAT STORAGE MISSION To offer a unified, open software-defined storage portfolio that delivers a range
More informationThe Fastest And Most Efficient Block Storage Software (SDS)
The Fastest And Most Efficient Block Storage Software (SDS) StorPool: Product Summary 1. Advanced Block-level Software Defined Storage, SDS (SDS 2.0) Fully distributed, scale-out, online changes of everything,
More informationDeploying Software Defined Storage for the Enterprise with Ceph. PRESENTATION TITLE GOES HERE Paul von Stamwitz Fujitsu
Deploying Software Defined Storage for the Enterprise with Ceph PRESENTATION TITLE GOES HERE Paul von Stamwitz Fujitsu Agenda Yet another attempt to define SDS Quick Overview of Ceph from a SDS perspective
More informationUsing persistent memory and RDMA for Ceph client write-back caching Scott Peterson, Senior Software Engineer Intel
Using persistent memory and RDMA for Ceph client write-back caching Scott Peterson, Senior Software Engineer Intel 2018 Storage Developer Conference. Intel Corporation. All Rights Reserved. 1 Ceph Concepts
More informationThe Google File System
The Google File System Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung Google SOSP 03, October 19 22, 2003, New York, USA Hyeon-Gyu Lee, and Yeong-Jae Woo Memory & Storage Architecture Lab. School
More informationRed Hat Ceph Storage and Samsung NVMe SSDs for intensive workloads
Red Hat Ceph Storage and Samsung NVMe SSDs for intensive workloads Power emerging OpenStack use cases with high-performance Samsung/ Red Hat Ceph reference architecture Optimize storage cluster performance
More informationWhite Paper. Nexenta Replicast
White Paper Nexenta Replicast By Caitlin Bestler, September 2013 Table of Contents Overview... 3 Nexenta Replicast Description... 3 Send Once, Receive Many... 4 Distributed Storage Basics... 7 Nexenta
More informationDell EMC Isilon All-Flash
Enterprise Strategy Group Getting to the bigger truth. ESG Lab Validation Dell EMC Isilon All-Flash Scale-out All-flash Storage for Demanding Unstructured Data Workloads By Tony Palmer, Senior Lab Analyst
More informationNutanix Tech Note. Virtualizing Microsoft Applications on Web-Scale Infrastructure
Nutanix Tech Note Virtualizing Microsoft Applications on Web-Scale Infrastructure The increase in virtualization of critical applications has brought significant attention to compute and storage infrastructure.
More informationCold Storage: The Road to Enterprise Ilya Kuznetsov YADRO
Cold Storage: The Road to Enterprise Ilya Kuznetsov YADRO Agenda Technical challenge Custom product Growth of aspirations Enterprise requirements Making an enterprise cold storage product 2 Technical Challenge
More informationMethod to Establish a High Availability and High Performance Storage Array in a Green Environment
Method to Establish a High Availability and High Performance Storage Array in a Green Environment Dr. M. K. Jibbe Director of Quality Architect Team, NetApp APG mahmoudj@netapp.com Marlin Gwaltney Quality
More informationRED HAT CEPH STORAGE ON THE INFINIFLASH ALL-FLASH STORAGE SYSTEM FROM SANDISK
REFERENCE ARCHITECTURE RED HAT CEPH STORAGE ON THE INFINIFLASH ALL-FLASH STORAGE SYSTEM FROM SANDISK ABSTRACT Combining Red Hat Ceph Storage with the InfiniFlash system from SanDisk yields software-defined
More informationThe Google File System
The Google File System Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung December 2003 ACM symposium on Operating systems principles Publisher: ACM Nov. 26, 2008 OUTLINE INTRODUCTION DESIGN OVERVIEW
More informationStaggeringly Large Filesystems
Staggeringly Large Filesystems Evan Danaher CS 6410 - October 27, 2009 Outline 1 Large Filesystems 2 GFS 3 Pond Outline 1 Large Filesystems 2 GFS 3 Pond Internet Scale Web 2.0 GFS Thousands of machines
More informationHCI: Hyper-Converged Infrastructure
Key Benefits: Innovative IT solution for high performance, simplicity and low cost Complete solution for IT workloads: compute, storage and networking in a single appliance High performance enabled by
More information2014 VMware Inc. All rights reserved.
2014 VMware Inc. All rights reserved. Agenda Virtual SAN 1 Why VSAN Software Defined Storage 2 Introducing Virtual SAN 3 Hardware Requirements 4 DEMO 5 Questions 2 The Software-Defined Data Center Expand
More informationKubernetes Integration with Virtuozzo Storage
Kubernetes Integration with Virtuozzo Storage A Technical OCTOBER, 2017 2017 Virtuozzo. All rights reserved. 1 Application Container Storage Application containers appear to be the perfect tool for supporting
More informationModern hyperconverged infrastructure. Karel Rudišar Systems Engineer, Vmware Inc.
Modern hyperconverged infrastructure Karel Rudišar Systems Engineer, Vmware Inc. 2 What Is Hyper-Converged Infrastructure? - The Ideal Architecture for SDDC Management SDDC Compute Networking Storage Simplicity
More informationA New Key-value Data Store For Heterogeneous Storage Architecture Intel APAC R&D Ltd.
A New Key-value Data Store For Heterogeneous Storage Architecture Intel APAC R&D Ltd. 1 Agenda Introduction Background and Motivation Hybrid Key-Value Data Store Architecture Overview Design details Performance
More informationTake Back Lost Revenue by Activating Virtuozzo Storage Today
Take Back Lost Revenue by Activating Virtuozzo Storage Today JUNE, 2017 2017 Virtuozzo. All rights reserved. 1 Introduction New software-defined storage (SDS) solutions are enabling hosting companies to
More informationCeph Block Devices: A Deep Dive. Josh Durgin RBD Lead June 24, 2015
Ceph Block Devices: A Deep Dive Josh Durgin RBD Lead June 24, 2015 Ceph Motivating Principles All components must scale horizontally There can be no single point of failure The solution must be hardware
More informationAccelerating Ceph with Flash and High Speed Networks
Accelerating Ceph with Flash and High Speed Networks Dror Goldenberg VP Software Architecture Santa Clara, CA 1 The New Open Cloud Era Compute Software Defined Network Object, Block Software Defined Storage
More informationStorage Designed to Support an Oracle Database. White Paper
Storage Designed to Support an Oracle Database White Paper Abstract Databases represent the backbone of most organizations. And Oracle databases in particular have become the mainstream data repository
More informationPerformance Testing December 16, 2017
December 16, 2017 1 1. vsan Performance Testing 1.1.Performance Testing Overview Table of Contents 2 1. vsan Performance Testing Performance Testing 3 1.1 Performance Testing Overview Performance Testing
More informationTHE CEPH POWER SHOW. Episode 2 : The Jewel Story. Daniel Messer Technical Marketing Red Hat Storage. Karan Singh Sr. Storage Architect Red Hat Storage
THE CEPH POWER SHOW Episode 2 : The Jewel Story Karan Singh Sr. Storage Architect Red Hat Storage Daniel Messer Technical Marketing Red Hat Storage Kyle Bader Sr. Storage Architect Red Hat Storage AGENDA
More informationSolidFire and Pure Storage Architectural Comparison
The All-Flash Array Built for the Next Generation Data Center SolidFire and Pure Storage Architectural Comparison June 2014 This document includes general information about Pure Storage architecture as
More informationSoftware-defined Storage: Fast, Safe and Efficient
Software-defined Storage: Fast, Safe and Efficient TRY NOW Thanks to Blockchain and Intel Intelligent Storage Acceleration Library Every piece of data is required to be stored somewhere. We all know about
More informationDistributed Filesystem
Distributed Filesystem 1 How do we get data to the workers? NAS Compute Nodes SAN 2 Distributing Code! Don t move data to workers move workers to the data! - Store data on the local disks of nodes in the
More informationCloud object storage in Ceph. Orit Wasserman Fosdem 2017
Cloud object storage in Ceph Orit Wasserman owasserm@redhat.com Fosdem 2017 AGENDA What is cloud object storage? Ceph overview Rados Gateway architecture Questions Cloud object storage Block storage Data
More informationUCS Invicta: A New Generation of Storage Performance. Mazen Abou Najm DC Consulting Systems Engineer
UCS Invicta: A New Generation of Storage Performance Mazen Abou Najm DC Consulting Systems Engineer HDDs Aren t Designed For High Performance Disk 101 Can t spin faster (200 IOPS/Drive) Can t seek faster
More informationThe Google File System
The Google File System Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung SOSP 2003 presented by Kun Suo Outline GFS Background, Concepts and Key words Example of GFS Operations Some optimizations in
More informationIBM Spectrum NAS. Easy-to-manage software-defined file storage for the enterprise. Overview. Highlights
IBM Spectrum NAS Easy-to-manage software-defined file storage for the enterprise Highlights Reduce capital expenditures with storage software on commodity servers Improve efficiency by consolidating all
More informationVMware Virtual SAN. High Performance Scalable Storage Architecture VMware Inc. All rights reserved.
VMware Virtual SAN High Performance Scalable Storage Architecture 2014 VMware Inc. All rights reserved. Agenda Importance of Software Defined approach for Storage Introduction VMware Virtual SAN Key Properties
More informationDatacenter Storage with Ceph
Datacenter Storage with Ceph John Spray john.spray@redhat.com jcsp on #ceph-devel Agenda What is Ceph? How does Ceph store your data? Interfaces to Ceph: RBD, RGW, CephFS Latest development updates Datacenter
More informationUNLEASH YOUR APPLICATIONS
UNLEASH YOUR APPLICATIONS Meet the 100% Flash Scale-Out Enterprise Storage Array from XtremIO Opportunities to truly innovate are rare. Yet today, flash technology has created the opportunity to not only
More informationHyperscaler Storage. September 12, 2016
Storage Networking Industry Association Technical White Paper Hyperscaler Storage Abstract: Hyperscaler storage customers typically build their own storage systems from commodity components. They have
More informationNovember 7, DAN WILSON Global Operations Architecture, Concur. OpenStack Summit Hong Kong JOE ARNOLD
November 7, 2013 DAN WILSON Global Operations Architecture, Concur dan.wilson@concur.com @tweetdanwilson OpenStack Summit Hong Kong JOE ARNOLD CEO, SwiftStack joe@swiftstack.com @joearnold Introduction
More informationDeep Dive on SimpliVity s OmniStack A Technical Whitepaper
Deep Dive on SimpliVity s OmniStack A Technical Whitepaper By Hans De Leenheer and Stephen Foskett August 2013 1 Introduction This paper is an in-depth look at OmniStack, the technology that powers SimpliVity
More informationEMC Virtual Infrastructure for Microsoft Exchange 2010 Enabled by EMC Symmetrix VMAX, VMware vsphere 4, and Replication Manager
EMC Virtual Infrastructure for Microsoft Exchange 2010 Enabled by EMC Symmetrix VMAX, VMware vsphere 4, and Replication Manager Reference Architecture Copyright 2010 EMC Corporation. All rights reserved.
More informationEnterprise Ceph: Everyway, your way! Amit Dell Kyle Red Hat Red Hat Summit June 2016
Enterprise Ceph: Everyway, your way! Amit Bhutani @ Dell Kyle Bader @ Red Hat Red Hat Summit June 2016 Agenda Overview of Ceph Components and Architecture Evolution of Ceph in Dell-Red Hat Joint OpenStack
More informationCeph Software Defined Storage Appliance
Ceph Software Defined Storage Appliance Unified distributed data storage cluster with self-healing, auto-balancing and no single point of failure Lowest power consumption in the industry: 70% power saving
More information9/26/2017 Sangmi Lee Pallickara Week 6- A. CS535 Big Data Fall 2017 Colorado State University
CS535 Big Data - Fall 2017 Week 6-A-1 CS535 BIG DATA FAQs PA1: Use only one word query Deadends {{Dead end}} Hub value will be?? PART 1. BATCH COMPUTING MODEL FOR BIG DATA ANALYTICS 4. GOOGLE FILE SYSTEM
More informationECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective
ECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective Part II: Software Infrastructure in Data Centers: Distributed File Systems 1 Permanently stores data Filesystems
More informationNetApp SolidFire and Pure Storage Architectural Comparison A SOLIDFIRE COMPETITIVE COMPARISON
A SOLIDFIRE COMPETITIVE COMPARISON NetApp SolidFire and Pure Storage Architectural Comparison This document includes general information about Pure Storage architecture as it compares to NetApp SolidFire.
More informationLife In The Flash Director - EMC Flash Strategy (Cross BU)
1 Life In The Flash Lane @SamMarraccini, Director - EMC Flash Strategy (Cross BU) CONSTANT 2 Performance = Moore s Law, Or Does It? MOORE S LAW: 100X PER DECADE FLASH Closes The CPU To Storage Gap FLASH
More informationNVMFS: A New File System Designed Specifically to Take Advantage of Nonvolatile Memory
NVMFS: A New File System Designed Specifically to Take Advantage of Nonvolatile Memory Dhananjoy Das, Sr. Systems Architect SanDisk Corp. 1 Agenda: Applications are KING! Storage landscape (Flash / NVM)
More informationINTRODUCTION TO CEPH. Orit Wasserman Red Hat August Penguin 2017
INTRODUCTION TO CEPH Orit Wasserman Red Hat August Penguin 2017 CEPHALOPOD A cephalopod is any member of the molluscan class Cephalopoda. These exclusively marine animals are characterized by bilateral
More informationOperating Systems. Lecture File system implementation. Master of Computer Science PUF - Hồ Chí Minh 2016/2017
Operating Systems Lecture 7.2 - File system implementation Adrien Krähenbühl Master of Computer Science PUF - Hồ Chí Minh 2016/2017 Design FAT or indexed allocation? UFS, FFS & Ext2 Journaling with Ext3
More informationUnderstanding System Characteristics of Online Erasure Coding on Scalable, Distributed and Large-Scale SSD Array Systems
Understanding System Characteristics of Online Erasure Coding on Scalable, Distributed and Large-Scale SSD Array Systems arxiv:179.5365v2 [cs.dc] 19 Sep 217 Sungjoon Koh, Jie Zhang, Miryeong Kwon, Jungyeon
More informationBenefits of 25, 40, and 50GbE Networks for Ceph and Hyper- Converged Infrastructure John F. Kim Mellanox Technologies
Benefits of 25, 40, and 50GbE Networks for Ceph and Hyper- Converged Infrastructure John F. Kim Mellanox Technologies Storage Transitions Change Network Needs Software Defined Storage Flash Storage Storage
More informationCeph: A Scalable, High-Performance Distributed File System
Ceph: A Scalable, High-Performance Distributed File System S. A. Weil, S. A. Brandt, E. L. Miller, D. D. E. Long Presented by Philip Snowberger Department of Computer Science and Engineering University
More information! Design constraints. " Component failures are the norm. " Files are huge by traditional standards. ! POSIX-like
Cloud background Google File System! Warehouse scale systems " 10K-100K nodes " 50MW (1 MW = 1,000 houses) " Power efficient! Located near cheap power! Passive cooling! Power Usage Effectiveness = Total
More informationImproving Ceph Performance while Reducing Costs
Improving Ceph Performance while Reducing Costs Applications and Ecosystem Solutions Development Rick Stehno Santa Clara, CA 1 Flash Application Acceleration Three ways to accelerate application performance
More informationINFINIDAT Storage Architecture. White Paper
INFINIDAT Storage Architecture White Paper Abstract The INFINIDAT enterprise storage solution is based upon the unique and patented INFINIDAT Storage Architecture (ISA). The INFINIDAT Storage Architecture
More informationA fields' Introduction to SUSE Enterprise Storage TUT91098
A fields' Introduction to SUSE Enterprise Storage TUT91098 Robert Grosschopff Senior Systems Engineer robert.grosschopff@suse.com Martin Weiss Senior Consultant martin.weiss@suse.com Joao Luis Senior Software
More informationTCC, so your business continues
TCC, so your business continues 1 Whitepaper Executive summary The TCC (Tuxis Cloud Concept) is a smart way to build hosting infrastructures for your private cloud and virtual datacenter. It natively integrates
More informationAsynchronous Logging and Fast Recovery for a Large-Scale Distributed In-Memory Storage
Asynchronous Logging and Fast Recovery for a Large-Scale Distributed In-Memory Storage Kevin Beineke, Florian Klein, Michael Schöttner Institut für Informatik, Heinrich-Heine-Universität Düsseldorf Outline
More informationThe What, Why and How of the Pure Storage Enterprise Flash Array. Ethan L. Miller (and a cast of dozens at Pure Storage)
The What, Why and How of the Pure Storage Enterprise Flash Array Ethan L. Miller (and a cast of dozens at Pure Storage) Enterprise storage: $30B market built on disk Key players: EMC, NetApp, HP, etc.
More informationGoogle File System. Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung Google fall DIP Heerak lim, Donghun Koo
Google File System Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung Google 2017 fall DIP Heerak lim, Donghun Koo 1 Agenda Introduction Design overview Systems interactions Master operation Fault tolerance
More informationECONOMICAL, STORAGE PURPOSE-BUILT FOR THE EMERGING DATA CENTERS. By George Crump
ECONOMICAL, STORAGE PURPOSE-BUILT FOR THE EMERGING DATA CENTERS By George Crump Economical, Storage Purpose-Built for the Emerging Data Centers Most small, growing businesses start as a collection of laptops
More information"Software-defined storage Crossing the right bridge"
Navigating the software-defined storage world shaping tomorrow with you René Huebel "Software-defined storage Crossing the right bridge" SDS the model and the promises Control Abstraction The promises
More informationDiscover the all-flash storage company for the on-demand world
Discover the all-flash storage company for the on-demand world STORAGE FOR WHAT S NEXT The applications we use in our personal lives have raised the level of expectations for the user experience in enterprise
More information클라우드스토리지구축을 위한 ceph 설치및설정
클라우드스토리지구축을 위한 ceph 설치및설정 Ph.D. Sun Park GIST, NetCS Lab. 2015. 07. 15 1 목차 Cloud Storage Services? Open Source Cloud Storage Softwares Introducing Ceph Storage Ceph Installation & Configuration Automatic
More informationPRESENTATION TITLE GOES HERE
Enterprise Storage PRESENTATION TITLE GOES HERE Leah Schoeb, Member of SNIA Technical Council SNIA EmeraldTM Training SNIA Emerald Power Efficiency Measurement Specification, for use in EPA ENERGY STAR
More informationEvaluating Cloud Storage Strategies. James Bottomley; CTO, Server Virtualization
Evaluating Cloud Storage Strategies James Bottomley; CTO, Server Virtualization Introduction to Storage Attachments: - Local (Direct cheap) SAS, SATA - Remote (SAN, NAS expensive) FC net Types - Block
More informationBlock Storage Service: Status and Performance
Block Storage Service: Status and Performance Dan van der Ster, IT-DSS, 6 June 2014 Summary This memo summarizes the current status of the Ceph block storage service as it is used for OpenStack Cinder
More informationCeph at DTU Risø Frank Schilder
Ceph at DTU Risø Frank Schilder Ceph at DTU Risø Ceph at DTU Risø Design goals 1) High failure tolerance (long-term) single-disk blue store OSDs, no journal aggregation high replication value for 24/7
More informationVirtual SAN and vsphere w/ Operations Management
Welcome! The webinar will start shortly For audio, dial 877-668-4490 / Code 664 120 829 or Listen on Your Computer Simplify Virtual Storage and Management with VMware Virtual SAN and vsphere w/ Operations
More informationPivot3 Acuity with Microsoft SQL Server Reference Architecture
Pivot3 Acuity with Microsoft SQL Server 2014 Reference Architecture How to Contact Pivot3 Pivot3, Inc. General Information: info@pivot3.com 221 West 6 th St., Suite 750 Sales: sales@pivot3.com Austin,
More informationArchive Solutions at the Center for High Performance Computing by Sam Liston (University of Utah)
Archive Solutions at the Center for High Performance Computing by Sam Liston (University of Utah) The scale of the data housed at the Center for High Performance Computing (CHPC) has dramatically increased
More informationGoogle File System. Arun Sundaram Operating Systems
Arun Sundaram Operating Systems 1 Assumptions GFS built with commodity hardware GFS stores a modest number of large files A few million files, each typically 100MB or larger (Multi-GB files are common)
More informationMicrosoft SQL Server HA and DR with DVX
Microsoft SQL Server HA and DR with DVX 385 Moffett Park Dr. Sunnyvale, CA 94089 844-478-8349 www.datrium.com Technical Report Introduction A Datrium DVX solution allows you to start small and scale out.
More informationECS High Availability Design
ECS High Availability Design March 2018 A Dell EMC white paper Revisions Date Mar 2018 Aug 2017 July 2017 Description Version 1.2 - Updated to include ECS version 3.2 content Version 1.1 - Updated to include
More informationCeph Intro & Architectural Overview. Abbas Bangash Intercloud Systems
Ceph Intro & Architectural Overview Abbas Bangash Intercloud Systems About Me Abbas Bangash Systems Team Lead, Intercloud Systems abangash@intercloudsys.com intercloudsys.com 2 CLOUD SERVICES COMPUTE NETWORK
More informationFLASHARRAY//M Smart Storage for Cloud IT
FLASHARRAY//M Smart Storage for Cloud IT //M AT A GLANCE PURPOSE-BUILT to power your business: Transactional and analytic databases Virtualization and private cloud Business critical applications Virtual
More informationSoftware Defined Storage at the Speed of Flash. PRESENTATION TITLE GOES HERE Carlos Carrero Rajagopal Vaideeswaran Symantec
Software Defined Storage at the Speed of Flash PRESENTATION TITLE GOES HERE Carlos Carrero Rajagopal Vaideeswaran Symantec Agenda Introduction Software Technology Architecture Review Oracle Configuration
More informationThe Google File System
The Google File System Sanjay Ghemawat, Howard Gobioff and Shun Tak Leung Google* Shivesh Kumar Sharma fl4164@wayne.edu Fall 2015 004395771 Overview Google file system is a scalable distributed file system
More informationCluster-Level Google How we use Colossus to improve storage efficiency
Cluster-Level Storage @ Google How we use Colossus to improve storage efficiency Denis Serenyi Senior Staff Software Engineer dserenyi@google.com November 13, 2017 Keynote at the 2nd Joint International
More informationOpenIO SDS on ARM A practical and cost-effective object storage infrastructure based on SoYouStart dedicated ARM servers.
OpenIO SDS on ARM A practical and cost-effective object storage infrastructure based on SoYouStart dedicated ARM servers. Copyright 217 OpenIO SAS All Rights Reserved. Restriction on Disclosure and Use
More informationStorPool Distributed Storage Software Technical Overview
StorPool Distributed Storage Software Technical Overview StorPool 2018 Page 1 of 9 StorPool Overview StorPool is distributed storage software. It pools the attached storage (hard disks or SSDs) of standard
More informationHGST: Market Creator to Market Leader
HGST: Market Creator to Market Leader Gaetano Pastore Enterprise Sales EMEA gaetano.pastore@hgst.com +4915122674411 HGST's Transformation: http://www.youtube.com/watch?v=ehiyhn0jlie Growth of the Digital
More informationVMware vsphere Clusters in Security Zones
SOLUTION OVERVIEW VMware vsan VMware vsphere Clusters in Security Zones A security zone, also referred to as a DMZ," is a sub-network that is designed to provide tightly controlled connectivity to an organization
More informationCeph in a Flash. Micron s Adventures in All-Flash Ceph Storage. Ryan Meredith & Brad Spiers, Micron Principal Solutions Engineer and Architect
Ceph in a Flash Micron s Adventures in All-Flash Ceph Storage Ryan Meredith & Brad Spiers, Micron Principal Solutions Engineer and Architect 217 Micron Technology, Inc. All rights reserved. Information,
More informationCS435 Introduction to Big Data FALL 2018 Colorado State University. 11/7/2018 Week 12-B Sangmi Lee Pallickara. FAQs
11/7/2018 CS435 Introduction to Big Data - FALL 2018 W12.B.0.0 CS435 Introduction to Big Data 11/7/2018 CS435 Introduction to Big Data - FALL 2018 W12.B.1 FAQs Deadline of the Programming Assignment 3
More informationKinetic drive. Bingzhe Li
Kinetic drive Bingzhe Li Consumption has changed It s an object storage world, unprecedented growth and scale In total, a complete redefinition of the storage stack https://www.openstack.org/summit/openstack-summit-atlanta-2014/session-videos/presentation/casestudy-seagate-kinetic-platform-in-action
More informationAurora, RDS, or On-Prem, Which is right for you
Aurora, RDS, or On-Prem, Which is right for you Kathy Gibbs Database Specialist TAM Katgibbs@amazon.com Santa Clara, California April 23th 25th, 2018 Agenda RDS Aurora EC2 On-Premise Wrap-up/Recommendation
More informationOptimizing the Data Center with an End to End Solutions Approach
Optimizing the Data Center with an End to End Solutions Approach Adam Roberts Chief Solutions Architect, Director of Technical Marketing ESS SanDisk Corporation Flash Memory Summit 11-13 August 2015 August
More informationCohesity Architecture White Paper. Building a Modern, Web-Scale Architecture for Consolidating Secondary Storage
Architecture White Paper Building a Modern, Web-Scale Architecture for Consolidating Secondary Storage The Band-Aid Effect : The Hidden Cost of Trying to Make Legacy Solutions Work The combination of explosive
More informationWHITE PAPER Software-Defined Storage IzumoFS with Cisco UCS and Cisco UCS Director Solutions
WHITE PAPER Software-Defined Storage IzumoFS with Cisco UCS and Cisco UCS Director Solutions Introduction While the data handled by companies has an average growth rate of over 50% per annum, growth of
More informationMySQL and Ceph. A tale of two friends
ysql and Ceph A tale of two friends Karan Singh Sr. Storage Architect Red Hat Taco Scargo Sr. Solution Architect Red Hat Agenda Ceph Introduction and Architecture Why ysql on Ceph ysql and Ceph Performance
More informationLeveraging Flash in Scalable Environments: A Systems Perspective on How FLASH Storage is Displacing Disk Storage
Leveraging Flash in Scalable Environments: A Systems Perspective on How FLASH Storage is Displacing Disk Storage Roark Hilomen, Engineering Fellow Systems & Software Solutions May 3, 2016 Forward-Looking
More information