Open vstorage RedHat Ceph Architectural Comparison

Size: px
Start display at page:

Download "Open vstorage RedHat Ceph Architectural Comparison"

Transcription

1 Open vstorage RedHat Ceph Architectural Comparison Open vstorage is the World s fastest Distributed Block Store that spans across different Datacenter. It combines ultrahigh performance and low latency connections with a data integrity that has no comparison. Data is distributed across datacenters using both Replication and Erasure Coding. Joining Performance and Integrity is not a simple bolt-on solution and requires a from-the-ground-up approach. Disk Failures, Node Failures and even Datacenter Failures do not present data loss and hence do not threaten any of your Data Integrity. You have been lead to believe that in order to have a 100% Data Loss Protection you have to compromise on Performance. While this might sound logical and acceptable, in is time to step out of the box and demand a noncompromise Storage Platform. With Open vstorage you can have your cake and eat it too! This document provides an overview of the Nexenta architecture and highlights the difference with the Open vstorage architecture. The comparison is not intended to be exhaustive, but covers the most relevant items where both solutions differ as seen from a customer perspective. Antwerpse Steenweg 19, 9080 Lochristi Belgium Phone: Mail: Info@openvstorage.com

2 Introduction Ceph, now part of RedHat, is an open-source, distributed object store designed to provide excellent performance, reliability and scalability. Next to scalable object storage interface Ceph offers a file system and block storage interface. All these interfaces make Ceph a popular storage solution. It has even been labeled the de facto storage backend for OpenStack 1. This should not come as a surprise as 57% of the OpenStack users and 45% of the cores deployments are using Ceph RBD as block storage solution 2. But although it is the most used OpenStack block storage solution, there are other solutions on the market which might be able to push Ceph from its OpenStack throne. Open vstorage takes a different approach which allows it to beat the performance of Ceph and while offering more functionality. Architectural Design RedHat Ceph At the core of Ceph is the RADOS object-based storage system. It provides the foundation for some of Ceph s advanced features such as RADOS Block Device (RBD), the RADOS S3 Gateway and the Ceph File System. The RADOS architecture consists out of a cluster map, a diagram explaining how nodes and disks are distributed, a set of OSDs (Object Storage Daemons) and monitors. The cluster map is built upon the CRUSH (Controlled Replication Under Scalable Hashing) algorithm to efficiently compute information about data location, instead of having to depend on a central lookup table. This is a critical part as the CRUSH algorithm utilizes a hash function that makes the mapping completely deterministic and calculable. The OSDs are responsible for storing the actual data on disks. The monitors contain the cluster map and clients can retrieve the map from the monitors. Ceph clients such as RBD run on the client side and write objects to the OSDs based upon the cluster map received from the monitors

3 In many cases RBD, the block interface, is used to provide storage for Virtual Machines by offering a virtual block device to the hypervisor. Underlying RBD splits the virtual block device into 4MB chunks that are stored as objects on the RADOS object store. This means a 4K write ends up as a 4MB write on the backend. Data on a Ceph backend is typically stored 3 times as this is the default policy. But let s have a look at the write flow in detail. The client writes to the primary OSD, a disk of the Ceph cluster. As soon as the data is written to the first OSD, it is written to 2 additional OSDs. Once both the secondary and tertiary OSD acknowledge the write to the primary OSD, the primary OSD will acknowledge to the client that the write is safely stored on the backend. Now let s have a closer look as how data is stored on the OSDs. As soon as IO goes into the OSD, it gets written twice as the first IO is done to the Ceph journal and later it is written a second time to the backend file system for safeguarding. If the data is safely stored on the backend file system, the journal gets trimmed to reclaim space. The same dual write also applies to the secondary and tertiary copy so a single write IO is written 6 times, twice per OSD multiplied by 3 different OSDs, before it can be acknowledged to the application.

4 Open vstorage The basic components of Open vstorage are the Open vstorage Edge, the Open vstorage Volume Driver and ALBA. The Open vstorage Edge exposes block devices to applications that need a volume. The Edge component communicates across RDMA, a low-latency, high-throughput networking protocol, directly with the memory of the server running the Volume Driver. The Volume Driver is the technology that converts block storage into objects (Storage Container Objects) that can be stored on the ALBA backend. This ALBA backend is a special purpose object storage solution and is made up out of storage nodes that run ALBA daemons. These daemons manage the local storage devices (HDDs, SSDs, PCIe flash cards...) and contribute these devices to the storage pools. The Volume Driver is a combination of a locationbased approach (performance) and a log structured approach (unlimited history, unlimited snapshots, thin cloning). Each incoming 4K write is appended to the write buffer. This write buffer can actually be seen as a Transaction Log storing one or more Storage Container Objects, a consecutive group of incoming writes. The incoming write is also dispatched to a Transaction Log onto a second node in the cluster to prevent data loss. This principle of dispatching incoming writes to an additional write buffer is referred to as the Open vstorage Distributed Transaction Log (DTL). It is important to note that both the write buffer and the DTL are very small in size, as they only need to hold data that is not yet protected by the ALBA backend. They are typically limited to 256MB per volume.

5 Once a SCO is full, it gets split into chunks, compressed, encrypted and spread across the ALBA backend for redundancy. This ALBA backend is typically built on top of a pool of large capacity drives with an allflash performance tier as acceleration layer. On reads the Volume Driver fetches the correct block directly from the right chunk on the ALBA backend across RDMA. Similarities Software-defined storage Both Ceph and Open vstorage are software-defined storage solutions. They are infrastructure agnostic and can run on any x86 hardware. The actual specifications of the hardware have of course a huge impact on the performance numbers you can achieve. Both solutions use SSDs to speed up the performance and use SATA drives to store the data in the backend. Scalability Both solutions are built to scale capacity wise from a couple of TB up to a few petabyte. Each node that gets added to the cluster puts its capacity at the disposal of the storage pool. For example CERN is running a single production cluster of 3PB based upon Ceph and has tested a cluster of 30PB3. The same applies to Open vstorage. A cluster can grow as big as 1024 nodes and up to 30 PB. Performance of both solutions also scales with every node added as more drives are added to distribute the IO load. 3

6 Key Differences Deterministic vs Explicit Meta Data With Ceph the placement of data on the backend is computation based (CRUSH), this makes the whole process repeatable and deterministic. So no lookup on a hash table, only formulas and calculations are used to determine which disks contain the object. Open vstorage on the other hand uses explicit metadata as it stores the exact location of the objects in the backend in a distributed database. Note that based upon the data in the backend, the distributed database can always be reconstructed in case there would be an issue with the database. Open vstorage has optimized the footprint of the metadata set and manages to keep the whole set in RAM for performance reasons. While having explicit metadata consumes more storage space, the benefits of explicit metadata are substantial: In case of a broken disk Open vstorage can easily regenerate the data from the broken disks, as it knows exactly what was on the disk. It even knows which bits of data are most at risk due to many missing chunks and can repair the most vulnerable data first. As Ceph doesn t keep track of what is on the disks, it needs to wait until scrubbing picks up that objects are missing. As Open vstorage stores backend metadata it can offer APE (Asynchronous Policy Enforcement) which allows to spread data according to the best available policy. In case data can t be written optimally, it will be first stored sub-optimal and later, when for example nodes are replaced or the network link between the datacenters is restored, data will be re-written with the most optimal policy. Ceph doesn t have this type of features and will just stop all writes. Huge Performance Difference A lot has been said about the performance of Ceph or better the lack thereof. Execute a Google Search for Ceph Performance Tuning and you will find 25,000 suggestions to improve the performance. The truth is that Ceph was designed for scalability and wasn t developed to be the fastest distributed block storage solution. Just to give you some numbers; running a cluster of around k SATA HDD drives will sadly only give you roughly IOPs. Even with all-flash hardware, the random write performance isn t impressive. On an InfiniFlash Sandisk server with 64 SSDs, Ceph can only reach 180,000 IOPS for 4K random writes4. On top, since Ceph only acknowledges the write after 3 disks have acknowledged the write, a single slow disk will impact all VMs as clients spread data across all disks in the cluster. 4 ance_optimization.pptx

7 Open vstorage performance of a single node. Internal benchmarks show that Open vstorage can be 10x faster than Ceph on the same hardware. This huge performance difference can be explained by the fact that both solutions have a radical different design: Ceph uses a location based approach where each volume is divided into 4MB fragments and writes are location based hence happen randomly. Open vstorage uses a log structured approach using immutable objects that creates a lock less design and leverages multiple cores to get very high performance. This approach leads to lower latency and hence better performance. Erasure Coding Since the Hammer release, April 2015, Ceph supports erasure coding on the backend. But, as Ceph works location based and uses 4MB extents, the update of a 4k block means that the whole 4MB extent to which the block belongs needs to be erasure coded again. This is a very high penalty. One could of course lower the size of an extent but that will make the metadata grow out of control. Basically erasure coding isn t usable with RBD so replication must be used. Open vstorage on the other hand uses a log-structured approach where consecutive writes are appended to an immutable Storage Container Object (SCO). This SCO can be erasure coded by means of forward error correction techniques and the resulting fragments can be stored across multiple hosts or even datacenters. Space Efficiency As discussed earlier, Ceph can only work with an X-way replication strategy in case RBD is used. By default Ceph proposes a 3-way replication strategy. This isn t very space efficient, as every 4MB will lead to 12MB being stored on the physical disks. This means three times the amount of nodes, three times the amount of networking equipment, three times the amount of power and cooling, basically tripling the TCO. Open vstorage is much more space efficient as it uses forward error correction which provides 100x better reliability as compared to 3 copies although storing only 1.25x more data. This means that for every 1MB, 1.25MB will be saved on the backend. Secondly, Open vstorage allows for the creation of different error correction policies for flash and HDD and hence the customer can select a lower redundancy factor on flash and a higher redundancy factor on HDD. In addition, contrary to Ceph, Open vstorage can further reduce its storage footprint by compressing the fragments before storing them on

8 SSDs and HDDs. Ceph has compression functionality but since updating a 4K block means the whole 4MB extent needs to be compressed again, it makes no sense to enable it for RBD. Multi datacenter Ceph is designed to be used on local networks, as it requires low latency links between hosts to be performant. Since the writes need to be acknowledged by 3 storage nodes, the latency for writes would be too high in case the secondary or tertiary write would have to go to other datacenters. Open vstorage on the other hand is designed to store data safely and efficiently across multiple datacenters. To ensure local performance it uses an all-flash tier per datacenter while for disaster recovery reasons data can be stored on a capacity tier across multiple datacenters. The fact that Open vstorage is designed to be used in a multi datacenter approach is also visible in that Open vstorage supports 2 levels of erasure coding: a global and a local policy. The local policy makes sure that a single disk or host failure doesn t require data from other datacenters to be sent over the network in order to rebuild the missing chunks of data. As Ceph relies on replication the chunks will need to come from a different datacenter to replace the missing replicas. This puts additional stress on the network. Encryption of the backend Data security is an essential aspect of a large-scale storage implementation. Encryption, encoding the data with a secret key, is the most effective way to achieve data security. Open vstorage encrypts the data before the SCOs are transferred to the backend. This means that data being transferred over the network to the backend is already encrypted. Ceph on the other hand only encrypts data when it is being stored on the OSD. This means data is in a clear format when it is transferred over the network. This approach poses a higher risk as data might get wiretapped while in transit to the backend. Snapshots and Clones For Open vstorage taking a snapshot is as simple as placing a marker in the metadata. Therefore it takes less than a millisecond to complete a snapshot. Due to the design, taking a snapshot in Ceph involves a lot more work as Ceph works location based. When a snapshot is created and data is being overwritten, the old data needs to be copied, a copy-on-write process, to safeguard the overwritten data. This adds additional latency to the write path. Both solutions offer thin cloning but both implement this feature quite differently. Ceph uses an approach where there is no tracking of which objects exist in a clone. A read that hits a non-existent object will attempt to read from the parent snapshot, and this will continue recursively until an object exists or an image with no parent is found. It goes without saying that each lookup attempt adds additional latency to the read path. Open vstorage on the other hand keeps track of the metadata of the volume. Finding the data only requires to look up the right SCO in the metadata. Even if the block being read is in the root snapshot and the volume has been cloned multiple times, the block can be fetched immediately from the backend with a single read operation. This single-hop-read allows Open vstorage to minimize the read latency even for the most complex clone trees.

9 Flash Friendly As Ceph is location based it requires high endurance flash technology as random writes, generate small updates across the device. Many updates of the same location within the volume will also cause the flash chips to wear out faster than normal as you are constantly updating the same cells of the flash memory. Open vstorage is much more flash friendly as the write buffer sequentializes the random writes in large fragments that can be written sequentially. Due to random write amplification, sequential write endurance for SSDs and PCIe flash cards is typically higher so the flash drives are less likely to fail with Open vstorage compared to Ceph. As every update to a volume also gets appended to a new SCO, overwriting the same LBA of a volume many times will hammer the same cells of the flash memory over and over. Since Open vstorage is less intensive on flash as compared to Ceph, cheaper, lower endurance SSDs can be used for the performance tier. The bottom line Open vstorage and Ceph are software based distributed block storage solutions and they both scale very well. Although both can be installed on the same hardware, the different design approaches result in a remarkable performance difference. Based upon tests it has been proven that Open vstorage can be 10 times faster than Ceph on the same hardware. Next to the difference in performance, Open vstorage is more space efficient. Ceph uses replication to protect the data while Open vstorage can use both erasure coding and replication. This allows Open vstorage to store data across datacenters and survive even a complete datacenter going offline.

Open vstorage EMC SCALEIO Architectural Comparison

Open vstorage EMC SCALEIO Architectural Comparison Open vstorage EMC SCALEIO Architectural Comparison Open vstorage is the World s fastest Distributed Block Store that spans across different Datacenter. It combines ultrahigh performance and low latency

More information

Turning Object. Storage into Virtual Machine Storage. White Papers

Turning Object. Storage into Virtual Machine Storage. White Papers Turning Object Open vstorage is the World s fastest Distributed Block Store that spans across different Datacenter. It combines ultrahigh performance and low latency connections with a data integrity that

More information

SolidFire and Ceph Architectural Comparison

SolidFire and Ceph Architectural Comparison The All-Flash Array Built for the Next Generation Data Center SolidFire and Ceph Architectural Comparison July 2014 Overview When comparing the architecture for Ceph and SolidFire, it is clear that both

More information

A Gentle Introduction to Ceph

A Gentle Introduction to Ceph A Gentle Introduction to Ceph Narrated by Tim Serong tserong@suse.com Adapted from a longer work by Lars Marowsky-Brée lmb@suse.com Once upon a time there was a Free and Open Source distributed storage

More information

A product by CloudFounders. Wim Provoost Open vstorage

A product by CloudFounders. Wim Provoost Open vstorage A product by CloudFounders Wim Provoost (@wimpers_be) Open vstorage (@openvstorage) http://www.openvstorage.com CloudFounders vrun Converged infrastructure that combines the benefits of the hyperconverged

More information

VMware Virtual SAN Technology

VMware Virtual SAN Technology VMware Virtual SAN Technology Today s Agenda 1 Hyper-Converged Infrastructure Architecture & Vmware Virtual SAN Overview 2 Why VMware Hyper-Converged Software? 3 VMware Virtual SAN Advantage Today s Agenda

More information

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme STO1926BU A Day in the Life of a VSAN I/O Diving in to the I/O Flow of vsan John Nicholson (@lost_signal) Pete Koehler (@vmpete) VMworld 2017 Content: Not for publication #VMworld #STO1926BU Disclaimer

More information

Nimble Storage Adaptive Flash

Nimble Storage Adaptive Flash Nimble Storage Adaptive Flash Read more Nimble solutions Contact Us 800-544-8877 solutions@microage.com MicroAge.com TECHNOLOGY OVERVIEW Nimble Storage Adaptive Flash Nimble Storage s Adaptive Flash platform

More information

The Google File System

The Google File System October 13, 2010 Based on: S. Ghemawat, H. Gobioff, and S.-T. Leung: The Google file system, in Proceedings ACM SOSP 2003, Lake George, NY, USA, October 2003. 1 Assumptions Interface Architecture Single

More information

RED HAT CEPH STORAGE ROADMAP. Cesar Pinto Account Manager, Red Hat Norway

RED HAT CEPH STORAGE ROADMAP. Cesar Pinto Account Manager, Red Hat Norway RED HAT CEPH STORAGE ROADMAP Cesar Pinto Account Manager, Red Hat Norway cpinto@redhat.com THE RED HAT STORAGE MISSION To offer a unified, open software-defined storage portfolio that delivers a range

More information

The Fastest And Most Efficient Block Storage Software (SDS)

The Fastest And Most Efficient Block Storage Software (SDS) The Fastest And Most Efficient Block Storage Software (SDS) StorPool: Product Summary 1. Advanced Block-level Software Defined Storage, SDS (SDS 2.0) Fully distributed, scale-out, online changes of everything,

More information

Deploying Software Defined Storage for the Enterprise with Ceph. PRESENTATION TITLE GOES HERE Paul von Stamwitz Fujitsu

Deploying Software Defined Storage for the Enterprise with Ceph. PRESENTATION TITLE GOES HERE Paul von Stamwitz Fujitsu Deploying Software Defined Storage for the Enterprise with Ceph PRESENTATION TITLE GOES HERE Paul von Stamwitz Fujitsu Agenda Yet another attempt to define SDS Quick Overview of Ceph from a SDS perspective

More information

Using persistent memory and RDMA for Ceph client write-back caching Scott Peterson, Senior Software Engineer Intel

Using persistent memory and RDMA for Ceph client write-back caching Scott Peterson, Senior Software Engineer Intel Using persistent memory and RDMA for Ceph client write-back caching Scott Peterson, Senior Software Engineer Intel 2018 Storage Developer Conference. Intel Corporation. All Rights Reserved. 1 Ceph Concepts

More information

The Google File System

The Google File System The Google File System Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung Google SOSP 03, October 19 22, 2003, New York, USA Hyeon-Gyu Lee, and Yeong-Jae Woo Memory & Storage Architecture Lab. School

More information

Red Hat Ceph Storage and Samsung NVMe SSDs for intensive workloads

Red Hat Ceph Storage and Samsung NVMe SSDs for intensive workloads Red Hat Ceph Storage and Samsung NVMe SSDs for intensive workloads Power emerging OpenStack use cases with high-performance Samsung/ Red Hat Ceph reference architecture Optimize storage cluster performance

More information

White Paper. Nexenta Replicast

White Paper. Nexenta Replicast White Paper Nexenta Replicast By Caitlin Bestler, September 2013 Table of Contents Overview... 3 Nexenta Replicast Description... 3 Send Once, Receive Many... 4 Distributed Storage Basics... 7 Nexenta

More information

Dell EMC Isilon All-Flash

Dell EMC Isilon All-Flash Enterprise Strategy Group Getting to the bigger truth. ESG Lab Validation Dell EMC Isilon All-Flash Scale-out All-flash Storage for Demanding Unstructured Data Workloads By Tony Palmer, Senior Lab Analyst

More information

Nutanix Tech Note. Virtualizing Microsoft Applications on Web-Scale Infrastructure

Nutanix Tech Note. Virtualizing Microsoft Applications on Web-Scale Infrastructure Nutanix Tech Note Virtualizing Microsoft Applications on Web-Scale Infrastructure The increase in virtualization of critical applications has brought significant attention to compute and storage infrastructure.

More information

Cold Storage: The Road to Enterprise Ilya Kuznetsov YADRO

Cold Storage: The Road to Enterprise Ilya Kuznetsov YADRO Cold Storage: The Road to Enterprise Ilya Kuznetsov YADRO Agenda Technical challenge Custom product Growth of aspirations Enterprise requirements Making an enterprise cold storage product 2 Technical Challenge

More information

Method to Establish a High Availability and High Performance Storage Array in a Green Environment

Method to Establish a High Availability and High Performance Storage Array in a Green Environment Method to Establish a High Availability and High Performance Storage Array in a Green Environment Dr. M. K. Jibbe Director of Quality Architect Team, NetApp APG mahmoudj@netapp.com Marlin Gwaltney Quality

More information

RED HAT CEPH STORAGE ON THE INFINIFLASH ALL-FLASH STORAGE SYSTEM FROM SANDISK

RED HAT CEPH STORAGE ON THE INFINIFLASH ALL-FLASH STORAGE SYSTEM FROM SANDISK REFERENCE ARCHITECTURE RED HAT CEPH STORAGE ON THE INFINIFLASH ALL-FLASH STORAGE SYSTEM FROM SANDISK ABSTRACT Combining Red Hat Ceph Storage with the InfiniFlash system from SanDisk yields software-defined

More information

The Google File System

The Google File System The Google File System Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung December 2003 ACM symposium on Operating systems principles Publisher: ACM Nov. 26, 2008 OUTLINE INTRODUCTION DESIGN OVERVIEW

More information

Staggeringly Large Filesystems

Staggeringly Large Filesystems Staggeringly Large Filesystems Evan Danaher CS 6410 - October 27, 2009 Outline 1 Large Filesystems 2 GFS 3 Pond Outline 1 Large Filesystems 2 GFS 3 Pond Internet Scale Web 2.0 GFS Thousands of machines

More information

HCI: Hyper-Converged Infrastructure

HCI: Hyper-Converged Infrastructure Key Benefits: Innovative IT solution for high performance, simplicity and low cost Complete solution for IT workloads: compute, storage and networking in a single appliance High performance enabled by

More information

2014 VMware Inc. All rights reserved.

2014 VMware Inc. All rights reserved. 2014 VMware Inc. All rights reserved. Agenda Virtual SAN 1 Why VSAN Software Defined Storage 2 Introducing Virtual SAN 3 Hardware Requirements 4 DEMO 5 Questions 2 The Software-Defined Data Center Expand

More information

Kubernetes Integration with Virtuozzo Storage

Kubernetes Integration with Virtuozzo Storage Kubernetes Integration with Virtuozzo Storage A Technical OCTOBER, 2017 2017 Virtuozzo. All rights reserved. 1 Application Container Storage Application containers appear to be the perfect tool for supporting

More information

Modern hyperconverged infrastructure. Karel Rudišar Systems Engineer, Vmware Inc.

Modern hyperconverged infrastructure. Karel Rudišar Systems Engineer, Vmware Inc. Modern hyperconverged infrastructure Karel Rudišar Systems Engineer, Vmware Inc. 2 What Is Hyper-Converged Infrastructure? - The Ideal Architecture for SDDC Management SDDC Compute Networking Storage Simplicity

More information

A New Key-value Data Store For Heterogeneous Storage Architecture Intel APAC R&D Ltd.

A New Key-value Data Store For Heterogeneous Storage Architecture Intel APAC R&D Ltd. A New Key-value Data Store For Heterogeneous Storage Architecture Intel APAC R&D Ltd. 1 Agenda Introduction Background and Motivation Hybrid Key-Value Data Store Architecture Overview Design details Performance

More information

Take Back Lost Revenue by Activating Virtuozzo Storage Today

Take Back Lost Revenue by Activating Virtuozzo Storage Today Take Back Lost Revenue by Activating Virtuozzo Storage Today JUNE, 2017 2017 Virtuozzo. All rights reserved. 1 Introduction New software-defined storage (SDS) solutions are enabling hosting companies to

More information

Ceph Block Devices: A Deep Dive. Josh Durgin RBD Lead June 24, 2015

Ceph Block Devices: A Deep Dive. Josh Durgin RBD Lead June 24, 2015 Ceph Block Devices: A Deep Dive Josh Durgin RBD Lead June 24, 2015 Ceph Motivating Principles All components must scale horizontally There can be no single point of failure The solution must be hardware

More information

Accelerating Ceph with Flash and High Speed Networks

Accelerating Ceph with Flash and High Speed Networks Accelerating Ceph with Flash and High Speed Networks Dror Goldenberg VP Software Architecture Santa Clara, CA 1 The New Open Cloud Era Compute Software Defined Network Object, Block Software Defined Storage

More information

Storage Designed to Support an Oracle Database. White Paper

Storage Designed to Support an Oracle Database. White Paper Storage Designed to Support an Oracle Database White Paper Abstract Databases represent the backbone of most organizations. And Oracle databases in particular have become the mainstream data repository

More information

Performance Testing December 16, 2017

Performance Testing December 16, 2017 December 16, 2017 1 1. vsan Performance Testing 1.1.Performance Testing Overview Table of Contents 2 1. vsan Performance Testing Performance Testing 3 1.1 Performance Testing Overview Performance Testing

More information

THE CEPH POWER SHOW. Episode 2 : The Jewel Story. Daniel Messer Technical Marketing Red Hat Storage. Karan Singh Sr. Storage Architect Red Hat Storage

THE CEPH POWER SHOW. Episode 2 : The Jewel Story. Daniel Messer Technical Marketing Red Hat Storage. Karan Singh Sr. Storage Architect Red Hat Storage THE CEPH POWER SHOW Episode 2 : The Jewel Story Karan Singh Sr. Storage Architect Red Hat Storage Daniel Messer Technical Marketing Red Hat Storage Kyle Bader Sr. Storage Architect Red Hat Storage AGENDA

More information

SolidFire and Pure Storage Architectural Comparison

SolidFire and Pure Storage Architectural Comparison The All-Flash Array Built for the Next Generation Data Center SolidFire and Pure Storage Architectural Comparison June 2014 This document includes general information about Pure Storage architecture as

More information

Software-defined Storage: Fast, Safe and Efficient

Software-defined Storage: Fast, Safe and Efficient Software-defined Storage: Fast, Safe and Efficient TRY NOW Thanks to Blockchain and Intel Intelligent Storage Acceleration Library Every piece of data is required to be stored somewhere. We all know about

More information

Distributed Filesystem

Distributed Filesystem Distributed Filesystem 1 How do we get data to the workers? NAS Compute Nodes SAN 2 Distributing Code! Don t move data to workers move workers to the data! - Store data on the local disks of nodes in the

More information

Cloud object storage in Ceph. Orit Wasserman Fosdem 2017

Cloud object storage in Ceph. Orit Wasserman Fosdem 2017 Cloud object storage in Ceph Orit Wasserman owasserm@redhat.com Fosdem 2017 AGENDA What is cloud object storage? Ceph overview Rados Gateway architecture Questions Cloud object storage Block storage Data

More information

UCS Invicta: A New Generation of Storage Performance. Mazen Abou Najm DC Consulting Systems Engineer

UCS Invicta: A New Generation of Storage Performance. Mazen Abou Najm DC Consulting Systems Engineer UCS Invicta: A New Generation of Storage Performance Mazen Abou Najm DC Consulting Systems Engineer HDDs Aren t Designed For High Performance Disk 101 Can t spin faster (200 IOPS/Drive) Can t seek faster

More information

The Google File System

The Google File System The Google File System Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung SOSP 2003 presented by Kun Suo Outline GFS Background, Concepts and Key words Example of GFS Operations Some optimizations in

More information

IBM Spectrum NAS. Easy-to-manage software-defined file storage for the enterprise. Overview. Highlights

IBM Spectrum NAS. Easy-to-manage software-defined file storage for the enterprise. Overview. Highlights IBM Spectrum NAS Easy-to-manage software-defined file storage for the enterprise Highlights Reduce capital expenditures with storage software on commodity servers Improve efficiency by consolidating all

More information

VMware Virtual SAN. High Performance Scalable Storage Architecture VMware Inc. All rights reserved.

VMware Virtual SAN. High Performance Scalable Storage Architecture VMware Inc. All rights reserved. VMware Virtual SAN High Performance Scalable Storage Architecture 2014 VMware Inc. All rights reserved. Agenda Importance of Software Defined approach for Storage Introduction VMware Virtual SAN Key Properties

More information

Datacenter Storage with Ceph

Datacenter Storage with Ceph Datacenter Storage with Ceph John Spray john.spray@redhat.com jcsp on #ceph-devel Agenda What is Ceph? How does Ceph store your data? Interfaces to Ceph: RBD, RGW, CephFS Latest development updates Datacenter

More information

UNLEASH YOUR APPLICATIONS

UNLEASH YOUR APPLICATIONS UNLEASH YOUR APPLICATIONS Meet the 100% Flash Scale-Out Enterprise Storage Array from XtremIO Opportunities to truly innovate are rare. Yet today, flash technology has created the opportunity to not only

More information

Hyperscaler Storage. September 12, 2016

Hyperscaler Storage. September 12, 2016 Storage Networking Industry Association Technical White Paper Hyperscaler Storage Abstract: Hyperscaler storage customers typically build their own storage systems from commodity components. They have

More information

November 7, DAN WILSON Global Operations Architecture, Concur. OpenStack Summit Hong Kong JOE ARNOLD

November 7, DAN WILSON Global Operations Architecture, Concur. OpenStack Summit Hong Kong JOE ARNOLD November 7, 2013 DAN WILSON Global Operations Architecture, Concur dan.wilson@concur.com @tweetdanwilson OpenStack Summit Hong Kong JOE ARNOLD CEO, SwiftStack joe@swiftstack.com @joearnold Introduction

More information

Deep Dive on SimpliVity s OmniStack A Technical Whitepaper

Deep Dive on SimpliVity s OmniStack A Technical Whitepaper Deep Dive on SimpliVity s OmniStack A Technical Whitepaper By Hans De Leenheer and Stephen Foskett August 2013 1 Introduction This paper is an in-depth look at OmniStack, the technology that powers SimpliVity

More information

EMC Virtual Infrastructure for Microsoft Exchange 2010 Enabled by EMC Symmetrix VMAX, VMware vsphere 4, and Replication Manager

EMC Virtual Infrastructure for Microsoft Exchange 2010 Enabled by EMC Symmetrix VMAX, VMware vsphere 4, and Replication Manager EMC Virtual Infrastructure for Microsoft Exchange 2010 Enabled by EMC Symmetrix VMAX, VMware vsphere 4, and Replication Manager Reference Architecture Copyright 2010 EMC Corporation. All rights reserved.

More information

Enterprise Ceph: Everyway, your way! Amit Dell Kyle Red Hat Red Hat Summit June 2016

Enterprise Ceph: Everyway, your way! Amit Dell Kyle Red Hat Red Hat Summit June 2016 Enterprise Ceph: Everyway, your way! Amit Bhutani @ Dell Kyle Bader @ Red Hat Red Hat Summit June 2016 Agenda Overview of Ceph Components and Architecture Evolution of Ceph in Dell-Red Hat Joint OpenStack

More information

Ceph Software Defined Storage Appliance

Ceph Software Defined Storage Appliance Ceph Software Defined Storage Appliance Unified distributed data storage cluster with self-healing, auto-balancing and no single point of failure Lowest power consumption in the industry: 70% power saving

More information

9/26/2017 Sangmi Lee Pallickara Week 6- A. CS535 Big Data Fall 2017 Colorado State University

9/26/2017 Sangmi Lee Pallickara Week 6- A. CS535 Big Data Fall 2017 Colorado State University CS535 Big Data - Fall 2017 Week 6-A-1 CS535 BIG DATA FAQs PA1: Use only one word query Deadends {{Dead end}} Hub value will be?? PART 1. BATCH COMPUTING MODEL FOR BIG DATA ANALYTICS 4. GOOGLE FILE SYSTEM

More information

ECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective

ECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective ECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective Part II: Software Infrastructure in Data Centers: Distributed File Systems 1 Permanently stores data Filesystems

More information

NetApp SolidFire and Pure Storage Architectural Comparison A SOLIDFIRE COMPETITIVE COMPARISON

NetApp SolidFire and Pure Storage Architectural Comparison A SOLIDFIRE COMPETITIVE COMPARISON A SOLIDFIRE COMPETITIVE COMPARISON NetApp SolidFire and Pure Storage Architectural Comparison This document includes general information about Pure Storage architecture as it compares to NetApp SolidFire.

More information

Life In The Flash Director - EMC Flash Strategy (Cross BU)

Life In The Flash Director - EMC Flash Strategy (Cross BU) 1 Life In The Flash Lane @SamMarraccini, Director - EMC Flash Strategy (Cross BU) CONSTANT 2 Performance = Moore s Law, Or Does It? MOORE S LAW: 100X PER DECADE FLASH Closes The CPU To Storage Gap FLASH

More information

NVMFS: A New File System Designed Specifically to Take Advantage of Nonvolatile Memory

NVMFS: A New File System Designed Specifically to Take Advantage of Nonvolatile Memory NVMFS: A New File System Designed Specifically to Take Advantage of Nonvolatile Memory Dhananjoy Das, Sr. Systems Architect SanDisk Corp. 1 Agenda: Applications are KING! Storage landscape (Flash / NVM)

More information

INTRODUCTION TO CEPH. Orit Wasserman Red Hat August Penguin 2017

INTRODUCTION TO CEPH. Orit Wasserman Red Hat August Penguin 2017 INTRODUCTION TO CEPH Orit Wasserman Red Hat August Penguin 2017 CEPHALOPOD A cephalopod is any member of the molluscan class Cephalopoda. These exclusively marine animals are characterized by bilateral

More information

Operating Systems. Lecture File system implementation. Master of Computer Science PUF - Hồ Chí Minh 2016/2017

Operating Systems. Lecture File system implementation. Master of Computer Science PUF - Hồ Chí Minh 2016/2017 Operating Systems Lecture 7.2 - File system implementation Adrien Krähenbühl Master of Computer Science PUF - Hồ Chí Minh 2016/2017 Design FAT or indexed allocation? UFS, FFS & Ext2 Journaling with Ext3

More information

Understanding System Characteristics of Online Erasure Coding on Scalable, Distributed and Large-Scale SSD Array Systems

Understanding System Characteristics of Online Erasure Coding on Scalable, Distributed and Large-Scale SSD Array Systems Understanding System Characteristics of Online Erasure Coding on Scalable, Distributed and Large-Scale SSD Array Systems arxiv:179.5365v2 [cs.dc] 19 Sep 217 Sungjoon Koh, Jie Zhang, Miryeong Kwon, Jungyeon

More information

Benefits of 25, 40, and 50GbE Networks for Ceph and Hyper- Converged Infrastructure John F. Kim Mellanox Technologies

Benefits of 25, 40, and 50GbE Networks for Ceph and Hyper- Converged Infrastructure John F. Kim Mellanox Technologies Benefits of 25, 40, and 50GbE Networks for Ceph and Hyper- Converged Infrastructure John F. Kim Mellanox Technologies Storage Transitions Change Network Needs Software Defined Storage Flash Storage Storage

More information

Ceph: A Scalable, High-Performance Distributed File System

Ceph: A Scalable, High-Performance Distributed File System Ceph: A Scalable, High-Performance Distributed File System S. A. Weil, S. A. Brandt, E. L. Miller, D. D. E. Long Presented by Philip Snowberger Department of Computer Science and Engineering University

More information

! Design constraints. " Component failures are the norm. " Files are huge by traditional standards. ! POSIX-like

! Design constraints.  Component failures are the norm.  Files are huge by traditional standards. ! POSIX-like Cloud background Google File System! Warehouse scale systems " 10K-100K nodes " 50MW (1 MW = 1,000 houses) " Power efficient! Located near cheap power! Passive cooling! Power Usage Effectiveness = Total

More information

Improving Ceph Performance while Reducing Costs

Improving Ceph Performance while Reducing Costs Improving Ceph Performance while Reducing Costs Applications and Ecosystem Solutions Development Rick Stehno Santa Clara, CA 1 Flash Application Acceleration Three ways to accelerate application performance

More information

INFINIDAT Storage Architecture. White Paper

INFINIDAT Storage Architecture. White Paper INFINIDAT Storage Architecture White Paper Abstract The INFINIDAT enterprise storage solution is based upon the unique and patented INFINIDAT Storage Architecture (ISA). The INFINIDAT Storage Architecture

More information

A fields' Introduction to SUSE Enterprise Storage TUT91098

A fields' Introduction to SUSE Enterprise Storage TUT91098 A fields' Introduction to SUSE Enterprise Storage TUT91098 Robert Grosschopff Senior Systems Engineer robert.grosschopff@suse.com Martin Weiss Senior Consultant martin.weiss@suse.com Joao Luis Senior Software

More information

TCC, so your business continues

TCC, so your business continues TCC, so your business continues 1 Whitepaper Executive summary The TCC (Tuxis Cloud Concept) is a smart way to build hosting infrastructures for your private cloud and virtual datacenter. It natively integrates

More information

Asynchronous Logging and Fast Recovery for a Large-Scale Distributed In-Memory Storage

Asynchronous Logging and Fast Recovery for a Large-Scale Distributed In-Memory Storage Asynchronous Logging and Fast Recovery for a Large-Scale Distributed In-Memory Storage Kevin Beineke, Florian Klein, Michael Schöttner Institut für Informatik, Heinrich-Heine-Universität Düsseldorf Outline

More information

The What, Why and How of the Pure Storage Enterprise Flash Array. Ethan L. Miller (and a cast of dozens at Pure Storage)

The What, Why and How of the Pure Storage Enterprise Flash Array. Ethan L. Miller (and a cast of dozens at Pure Storage) The What, Why and How of the Pure Storage Enterprise Flash Array Ethan L. Miller (and a cast of dozens at Pure Storage) Enterprise storage: $30B market built on disk Key players: EMC, NetApp, HP, etc.

More information

Google File System. Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung Google fall DIP Heerak lim, Donghun Koo

Google File System. Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung Google fall DIP Heerak lim, Donghun Koo Google File System Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung Google 2017 fall DIP Heerak lim, Donghun Koo 1 Agenda Introduction Design overview Systems interactions Master operation Fault tolerance

More information

ECONOMICAL, STORAGE PURPOSE-BUILT FOR THE EMERGING DATA CENTERS. By George Crump

ECONOMICAL, STORAGE PURPOSE-BUILT FOR THE EMERGING DATA CENTERS. By George Crump ECONOMICAL, STORAGE PURPOSE-BUILT FOR THE EMERGING DATA CENTERS By George Crump Economical, Storage Purpose-Built for the Emerging Data Centers Most small, growing businesses start as a collection of laptops

More information

"Software-defined storage Crossing the right bridge"

Software-defined storage Crossing the right bridge Navigating the software-defined storage world shaping tomorrow with you René Huebel "Software-defined storage Crossing the right bridge" SDS the model and the promises Control Abstraction The promises

More information

Discover the all-flash storage company for the on-demand world

Discover the all-flash storage company for the on-demand world Discover the all-flash storage company for the on-demand world STORAGE FOR WHAT S NEXT The applications we use in our personal lives have raised the level of expectations for the user experience in enterprise

More information

클라우드스토리지구축을 위한 ceph 설치및설정

클라우드스토리지구축을 위한 ceph 설치및설정 클라우드스토리지구축을 위한 ceph 설치및설정 Ph.D. Sun Park GIST, NetCS Lab. 2015. 07. 15 1 목차 Cloud Storage Services? Open Source Cloud Storage Softwares Introducing Ceph Storage Ceph Installation & Configuration Automatic

More information

PRESENTATION TITLE GOES HERE

PRESENTATION TITLE GOES HERE Enterprise Storage PRESENTATION TITLE GOES HERE Leah Schoeb, Member of SNIA Technical Council SNIA EmeraldTM Training SNIA Emerald Power Efficiency Measurement Specification, for use in EPA ENERGY STAR

More information

Evaluating Cloud Storage Strategies. James Bottomley; CTO, Server Virtualization

Evaluating Cloud Storage Strategies. James Bottomley; CTO, Server Virtualization Evaluating Cloud Storage Strategies James Bottomley; CTO, Server Virtualization Introduction to Storage Attachments: - Local (Direct cheap) SAS, SATA - Remote (SAN, NAS expensive) FC net Types - Block

More information

Block Storage Service: Status and Performance

Block Storage Service: Status and Performance Block Storage Service: Status and Performance Dan van der Ster, IT-DSS, 6 June 2014 Summary This memo summarizes the current status of the Ceph block storage service as it is used for OpenStack Cinder

More information

Ceph at DTU Risø Frank Schilder

Ceph at DTU Risø Frank Schilder Ceph at DTU Risø Frank Schilder Ceph at DTU Risø Ceph at DTU Risø Design goals 1) High failure tolerance (long-term) single-disk blue store OSDs, no journal aggregation high replication value for 24/7

More information

Virtual SAN and vsphere w/ Operations Management

Virtual SAN and vsphere w/ Operations Management Welcome! The webinar will start shortly For audio, dial 877-668-4490 / Code 664 120 829 or Listen on Your Computer Simplify Virtual Storage and Management with VMware Virtual SAN and vsphere w/ Operations

More information

Pivot3 Acuity with Microsoft SQL Server Reference Architecture

Pivot3 Acuity with Microsoft SQL Server Reference Architecture Pivot3 Acuity with Microsoft SQL Server 2014 Reference Architecture How to Contact Pivot3 Pivot3, Inc. General Information: info@pivot3.com 221 West 6 th St., Suite 750 Sales: sales@pivot3.com Austin,

More information

Archive Solutions at the Center for High Performance Computing by Sam Liston (University of Utah)

Archive Solutions at the Center for High Performance Computing by Sam Liston (University of Utah) Archive Solutions at the Center for High Performance Computing by Sam Liston (University of Utah) The scale of the data housed at the Center for High Performance Computing (CHPC) has dramatically increased

More information

Google File System. Arun Sundaram Operating Systems

Google File System. Arun Sundaram Operating Systems Arun Sundaram Operating Systems 1 Assumptions GFS built with commodity hardware GFS stores a modest number of large files A few million files, each typically 100MB or larger (Multi-GB files are common)

More information

Microsoft SQL Server HA and DR with DVX

Microsoft SQL Server HA and DR with DVX Microsoft SQL Server HA and DR with DVX 385 Moffett Park Dr. Sunnyvale, CA 94089 844-478-8349 www.datrium.com Technical Report Introduction A Datrium DVX solution allows you to start small and scale out.

More information

ECS High Availability Design

ECS High Availability Design ECS High Availability Design March 2018 A Dell EMC white paper Revisions Date Mar 2018 Aug 2017 July 2017 Description Version 1.2 - Updated to include ECS version 3.2 content Version 1.1 - Updated to include

More information

Ceph Intro & Architectural Overview. Abbas Bangash Intercloud Systems

Ceph Intro & Architectural Overview. Abbas Bangash Intercloud Systems Ceph Intro & Architectural Overview Abbas Bangash Intercloud Systems About Me Abbas Bangash Systems Team Lead, Intercloud Systems abangash@intercloudsys.com intercloudsys.com 2 CLOUD SERVICES COMPUTE NETWORK

More information

FLASHARRAY//M Smart Storage for Cloud IT

FLASHARRAY//M Smart Storage for Cloud IT FLASHARRAY//M Smart Storage for Cloud IT //M AT A GLANCE PURPOSE-BUILT to power your business: Transactional and analytic databases Virtualization and private cloud Business critical applications Virtual

More information

Software Defined Storage at the Speed of Flash. PRESENTATION TITLE GOES HERE Carlos Carrero Rajagopal Vaideeswaran Symantec

Software Defined Storage at the Speed of Flash. PRESENTATION TITLE GOES HERE Carlos Carrero Rajagopal Vaideeswaran Symantec Software Defined Storage at the Speed of Flash PRESENTATION TITLE GOES HERE Carlos Carrero Rajagopal Vaideeswaran Symantec Agenda Introduction Software Technology Architecture Review Oracle Configuration

More information

The Google File System

The Google File System The Google File System Sanjay Ghemawat, Howard Gobioff and Shun Tak Leung Google* Shivesh Kumar Sharma fl4164@wayne.edu Fall 2015 004395771 Overview Google file system is a scalable distributed file system

More information

Cluster-Level Google How we use Colossus to improve storage efficiency

Cluster-Level Google How we use Colossus to improve storage efficiency Cluster-Level Storage @ Google How we use Colossus to improve storage efficiency Denis Serenyi Senior Staff Software Engineer dserenyi@google.com November 13, 2017 Keynote at the 2nd Joint International

More information

OpenIO SDS on ARM A practical and cost-effective object storage infrastructure based on SoYouStart dedicated ARM servers.

OpenIO SDS on ARM A practical and cost-effective object storage infrastructure based on SoYouStart dedicated ARM servers. OpenIO SDS on ARM A practical and cost-effective object storage infrastructure based on SoYouStart dedicated ARM servers. Copyright 217 OpenIO SAS All Rights Reserved. Restriction on Disclosure and Use

More information

StorPool Distributed Storage Software Technical Overview

StorPool Distributed Storage Software Technical Overview StorPool Distributed Storage Software Technical Overview StorPool 2018 Page 1 of 9 StorPool Overview StorPool is distributed storage software. It pools the attached storage (hard disks or SSDs) of standard

More information

HGST: Market Creator to Market Leader

HGST: Market Creator to Market Leader HGST: Market Creator to Market Leader Gaetano Pastore Enterprise Sales EMEA gaetano.pastore@hgst.com +4915122674411 HGST's Transformation: http://www.youtube.com/watch?v=ehiyhn0jlie Growth of the Digital

More information

VMware vsphere Clusters in Security Zones

VMware vsphere Clusters in Security Zones SOLUTION OVERVIEW VMware vsan VMware vsphere Clusters in Security Zones A security zone, also referred to as a DMZ," is a sub-network that is designed to provide tightly controlled connectivity to an organization

More information

Ceph in a Flash. Micron s Adventures in All-Flash Ceph Storage. Ryan Meredith & Brad Spiers, Micron Principal Solutions Engineer and Architect

Ceph in a Flash. Micron s Adventures in All-Flash Ceph Storage. Ryan Meredith & Brad Spiers, Micron Principal Solutions Engineer and Architect Ceph in a Flash Micron s Adventures in All-Flash Ceph Storage Ryan Meredith & Brad Spiers, Micron Principal Solutions Engineer and Architect 217 Micron Technology, Inc. All rights reserved. Information,

More information

CS435 Introduction to Big Data FALL 2018 Colorado State University. 11/7/2018 Week 12-B Sangmi Lee Pallickara. FAQs

CS435 Introduction to Big Data FALL 2018 Colorado State University. 11/7/2018 Week 12-B Sangmi Lee Pallickara. FAQs 11/7/2018 CS435 Introduction to Big Data - FALL 2018 W12.B.0.0 CS435 Introduction to Big Data 11/7/2018 CS435 Introduction to Big Data - FALL 2018 W12.B.1 FAQs Deadline of the Programming Assignment 3

More information

Kinetic drive. Bingzhe Li

Kinetic drive. Bingzhe Li Kinetic drive Bingzhe Li Consumption has changed It s an object storage world, unprecedented growth and scale In total, a complete redefinition of the storage stack https://www.openstack.org/summit/openstack-summit-atlanta-2014/session-videos/presentation/casestudy-seagate-kinetic-platform-in-action

More information

Aurora, RDS, or On-Prem, Which is right for you

Aurora, RDS, or On-Prem, Which is right for you Aurora, RDS, or On-Prem, Which is right for you Kathy Gibbs Database Specialist TAM Katgibbs@amazon.com Santa Clara, California April 23th 25th, 2018 Agenda RDS Aurora EC2 On-Premise Wrap-up/Recommendation

More information

Optimizing the Data Center with an End to End Solutions Approach

Optimizing the Data Center with an End to End Solutions Approach Optimizing the Data Center with an End to End Solutions Approach Adam Roberts Chief Solutions Architect, Director of Technical Marketing ESS SanDisk Corporation Flash Memory Summit 11-13 August 2015 August

More information

Cohesity Architecture White Paper. Building a Modern, Web-Scale Architecture for Consolidating Secondary Storage

Cohesity Architecture White Paper. Building a Modern, Web-Scale Architecture for Consolidating Secondary Storage Architecture White Paper Building a Modern, Web-Scale Architecture for Consolidating Secondary Storage The Band-Aid Effect : The Hidden Cost of Trying to Make Legacy Solutions Work The combination of explosive

More information

WHITE PAPER Software-Defined Storage IzumoFS with Cisco UCS and Cisco UCS Director Solutions

WHITE PAPER Software-Defined Storage IzumoFS with Cisco UCS and Cisco UCS Director Solutions WHITE PAPER Software-Defined Storage IzumoFS with Cisco UCS and Cisco UCS Director Solutions Introduction While the data handled by companies has an average growth rate of over 50% per annum, growth of

More information

MySQL and Ceph. A tale of two friends

MySQL and Ceph. A tale of two friends ysql and Ceph A tale of two friends Karan Singh Sr. Storage Architect Red Hat Taco Scargo Sr. Solution Architect Red Hat Agenda Ceph Introduction and Architecture Why ysql on Ceph ysql and Ceph Performance

More information

Leveraging Flash in Scalable Environments: A Systems Perspective on How FLASH Storage is Displacing Disk Storage

Leveraging Flash in Scalable Environments: A Systems Perspective on How FLASH Storage is Displacing Disk Storage Leveraging Flash in Scalable Environments: A Systems Perspective on How FLASH Storage is Displacing Disk Storage Roark Hilomen, Engineering Fellow Systems & Software Solutions May 3, 2016 Forward-Looking

More information