Open vstorage EMC SCALEIO Architectural Comparison

Similar documents
Open vstorage RedHat Ceph Architectural Comparison

Turning Object. Storage into Virtual Machine Storage. White Papers

Modern hyperconverged infrastructure. Karel Rudišar Systems Engineer, Vmware Inc.

SOFTWARE DEFINED STORAGE

HCI: Hyper-Converged Infrastructure

The Fastest And Most Efficient Block Storage Software (SDS)

UCS Invicta: A New Generation of Storage Performance. Mazen Abou Najm DC Consulting Systems Engineer

Life In The Flash Director - EMC Flash Strategy (Cross BU)

SolidFire and Ceph Architectural Comparison

VMware Virtual SAN Technology

Take Back Lost Revenue by Activating Virtuozzo Storage Today

2014 VMware Inc. All rights reserved.

Nimble Storage Adaptive Flash

Red Hat Ceph Storage and Samsung NVMe SSDs for intensive workloads

UNLEASH YOUR APPLICATIONS

Software Defined Storage at the Speed of Flash. PRESENTATION TITLE GOES HERE Carlos Carrero Rajagopal Vaideeswaran Symantec

Dell EMC ScaleIO Ready Node

Software Defined Storage

INFINIDAT Storage Architecture. White Paper

NetApp SolidFire and Pure Storage Architectural Comparison A SOLIDFIRE COMPETITIVE COMPARISON

DDN. DDN Updates. DataDirect Neworks Japan, Inc Nobu Hashizume. DDN Storage 2018 DDN Storage 1

Storage Solutions for VMware: InfiniBox. White Paper

VMware vsphere Clusters in Security Zones

THE SUMMARY. CLUSTER SERIES - pg. 3. ULTRA SERIES - pg. 5. EXTREME SERIES - pg. 9

Got Isilon? Need IOPS? Get Avere.

vsan Security Zone Deployment First Published On: Last Updated On:

Copyright 2013 EMC Corporation. All rights reserved. FLASH REDEFINING THE POSSIBLE

THE EMC ISILON STORY. Big Data In The Enterprise. Deya Bassiouni Isilon Regional Sales Manager Emerging Africa, Egypt & Lebanon.

SolidFire and Pure Storage Architectural Comparison

EMC XTREMCACHE ACCELERATES ORACLE

Hyperscaler Storage. September 12, 2016

EMC VMAX 400K SPC-2 Proven Performance. Silverton Consulting, Inc. StorInt Briefing

ECONOMICAL, STORAGE PURPOSE-BUILT FOR THE EMERGING DATA CENTERS. By George Crump

HPC Growing Pains. IT Lessons Learned from the Biomedical Data Deluge

Discover the all-flash storage company for the on-demand world

WHITE PAPER PernixData FVP

Cold Storage: The Road to Enterprise Ilya Kuznetsov YADRO

Dell EMC All-Flash solutions are powered by Intel Xeon processors. Learn more at DellEMC.com/All-Flash

THESUMMARY. ARKSERIES - pg. 3. ULTRASERIES - pg. 5. EXTREMESERIES - pg. 9

DDN. DDN Updates. Data DirectNeworks Japan, Inc Shuichi Ihara. DDN Storage 2017 DDN Storage

A product by CloudFounders. Wim Provoost Open vstorage

Nimble Storage vs HPE 3PAR: A Comparison Snapshot

Storage Designed to Support an Oracle Database. White Paper

Warsaw. 11 th September 2018

NEXT GENERATION DATACENTER XTREMIO and SCALEIO

Four Steps to Unleashing The Full Potential of Your Database

Benefits of 25, 40, and 50GbE Networks for Ceph and Hyper- Converged Infrastructure John F. Kim Mellanox Technologies

Secure Block Storage (SBS) FAQ

New HPE 3PAR StoreServ 8000 and series Optimized for Flash

HGST: Market Creator to Market Leader

Dell EMC Isilon All-Flash

VMWARE EBOOK. Easily Deployed Software-Defined Storage: A Customer Love Story

MODERNISE WITH ALL-FLASH. Intel Inside. Powerful Data Centre Outside.

Next-Generation NVMe-Native Parallel Filesystem for Accelerating HPC Workloads

Operating Systems. Lecture File system implementation. Master of Computer Science PUF - Hồ Chí Minh 2016/2017

EMC XtremIO All-Flash Applications. Sonny Aulakh VP, Sales Engineering November 2014

Architecting For Availability, Performance & Networking With ScaleIO

Modernize with all-flash

Method to Establish a High Availability and High Performance Storage Array in a Green Environment

XTREMIO: TRANSFORMING APPLICATIONS, ENABLING THE AGILE DATA CENTER

EMC Virtual Infrastructure for Microsoft Exchange 2010 Enabled by EMC Symmetrix VMAX, VMware vsphere 4, and Replication Manager

INFINIDAT Data Protection. White Paper

StarWind Storage Appliance

WHITE PAPER Software-Defined Storage IzumoFS with Cisco UCS and Cisco UCS Director Solutions

VMware Virtual SAN. Technical Walkthrough. Massimiliano Moschini Brand Specialist VCI - vexpert VMware Inc. All rights reserved.

Deploy a High-Performance Database Solution: Cisco UCS B420 M4 Blade Server with Fusion iomemory PX600 Using Oracle Database 12c

New Approach to Unstructured Data

How To Get The Most Out Of Flash Deployments

Running Splunk on VxRack FLEX

Why Datrium DVX is Best for VDI

SimpliVity OmniStack with the HyTrust Platform

Reconstruyendo una Nube Privada con la Innovadora Hiper-Convergencia Infraestructura Huawei FusionCube Hiper-Convergente

Cohesity Flash Protect for Pure FlashBlade: Simple, Scalable Data Protection

Technical White Paper: IntelliFlash Architecture

Hyper-converged Secondary Storage for Backup with Deduplication Q & A. The impact of data deduplication on the backup process

IBM Spectrum NAS, IBM Spectrum Scale and IBM Cloud Object Storage

The next step in Software-Defined Storage with Virtual SAN

FLAT DATACENTER STORAGE CHANDNI MODI (FN8692)

Copyright 2018 Dell Inc.

Pivot3 Acuity with Microsoft SQL Server Reference Architecture

It s Time to Move Your Critical Data to SSDs Introduction

Hitachi Virtual Storage Platform Family

VMware Virtual SAN. High Performance Scalable Storage Architecture VMware Inc. All rights reserved.

Modernizing Virtual Infrastructures Using VxRack FLEX with ScaleIO

Modernize Without. Compromise. Modernize Without Compromise- All Flash. All-Flash Portfolio. Haider Aziz. System Engineering Manger- Primary Storage

Veritas NetBackup on Cisco UCS S3260 Storage Server

ECS High Availability Design

Flashed-Optimized VPSA. Always Aligned with your Changing World

The storage challenges of virtualized environments

Trends in Data Protection and Restoration Technologies. Mike Fishman, EMC 2 Corporation

Native vsphere Storage for Remote and Branch Offices

ARCHITECTURE WHITEPAPER

Dell EMC Isilon with Cohesity DataProtect

IBM FlashSystem. IBM FLiP Tool Wie viel schneller kann Ihr IBM i Power Server mit IBM FlashSystem 900 / V9000 Storage sein?

Deep Dive on SimpliVity s OmniStack A Technical Whitepaper

IBM Spectrum NAS. Easy-to-manage software-defined file storage for the enterprise. Overview. Highlights

RAMCloud and the Low- Latency Datacenter. John Ousterhout Stanford University

Dell PowerEdge R730xd Servers with Samsung SM1715 NVMe Drives Powers the Aerospike Fraud Prevention Benchmark

Flash In the Data Center

Cohesity Architecture White Paper. Building a Modern, Web-Scale Architecture for Consolidating Secondary Storage

Transcription:

Open vstorage EMC SCALEIO Architectural Comparison Open vstorage is the World s fastest Distributed Block Store that spans across different Datacenter. It combines ultrahigh performance and low latency connections with a data integrity that has no comparison. Data is distributed across datacenters using both Replication and Erasure Coding. Joining Performance and Integrity is not a simple bolt-on solution and requires a from-the-ground-up approach. Disk Failures, Node Failures and even Datacenter Failures do not present data loss and hence do not threaten any of your Data Integrity. You have been lead to believe that in order to have a 100% Data Loss Protection you have to compromise on Performance. While this might sound logical and acceptable, in is time to step out of the box and demand a noncompromise Storage Platform. With Open vstorage you can have your cake and eat it too! This document provides an overview of the Nexenta architecture and highlights the difference with the Open vstorage architecture. The comparison is not intended to be exhaustive, but covers the most relevant items where both solutions differ as seen from a customer perspective. Antwerpse Steenweg 19, 9080 Lochristi Belgium Phone: +32 9 324 25 74 Mail: Info@openvstorage.com

Introduction ScaleIO is EMC s software-defined, scale-out, block storage solution and is designed for large-scale datacenters. It combines multiple x86 storage nodes into a storage cluster targeted at running high bandwidth, low latency IO workloads. ScaleIO presents top-class performance results. However, to reach these performance numbers, some trade-offs had to be made in the design. This has led to certain limitations with regards to functionality and reliability. Open vstorage takes a different approach which not only results in superior performance vs. ScaleIO, but also offers more functionality. Architectural Design ScaleIO The basic components of ScaleIO1 are the ScaleIO Data Client (SDC) and the ScaleIO Data Server (SDS). The SDC is a lightweight block device driver that exposes local block volumes to applications running on the same server. The actual data is stored on storage nodes that run the SDS. The SDS manages the local storage devices (HDDs, SSDs, PCIe flash cards,...) and contributes these devices to the global storage pool. The role of the SDS is to actually perform the backend IO operations as requested by an SDC. Each ScaleIO volume is divided into 1 MB chunks. These chunks are distributed (striped) across physical disks throughout the cluster. Each chunk has 2 copies for redundancy reasons. Although chunks are 1MB in size, ScaleIO allows to read or write for example 4K instead of the full 1MB. 1 http://www.cisco.com/c/en/us/solutions/collateral/data-center-virtualization/solutions-vspex/whitepaperc11-733544.html

Source EMC 2 On a write, the SDC that exposes the ScaleIO volume, sends the IO to the primary SDS where the chunk is located. The primary SDS sends the IO to the local drive and in parallel to the secondary SDC which holds the second copy of the chunk. Only after an acknowledgment is received from the secondary SDS, the primary SDS acknowledges the write to the SDC. On reads, the SDC connects to the primary SDS to fetch the data it needs. 2 https://www.emc.com/collateral/white-papers/h14344-emc-scaleio-basic-architecture.pdf

Open vstorage The basic components of Open vstorage are the Open vstorage Edge, the Open vstorage Volume Driver and ALBA. The Open vstorage Edge exposes block devices to applications that need a volume. The Edge component communicates via RDMA, a low-latency, high-throughput networking protocol, directly with the memory of the server running the Volume Driver. The Volume Driver is the technology that converts block storage into objects (Storage Container Objects, SCO), which can be stored on the ALBA backend. This ALBA backend is a special purpose object storage solution and is made up out of storage nodes running ALBA daemons. These daemons manage the local storage devices (HDDs, SSDs, PCIe flash cards,...) and contribute these devices to the storage pools. The Volume Driver is a combination of a location-based approach (delivering performance) and a log structured approach (delivering unlimited history, unlimited snapshots, thin cloning). Each incoming 4K write is appended to the write buffer. This write buffer can actually be seen as a Transaction Log storing one or more Storage Container Objects, a consecutive group of incoming writes. The incoming write is also dispatched to a Transaction Log onto a second node in the cluster to prevent data loss. This principle of dispatching incoming writes to an additional write buffer is referred to as the Open vstorage Distributed Transaction Log (DTL). It is important to note that both the write buffer and the DTL are very small in size as they only need to hold data which is not yet protected by the ALBA backend. They are typically limited to 256MB per volume.

Once a SCO is full, it gets split into chunks, compressed, encrypted and spread across the ALBA backend for redundancy. This ALBA backend is typically built on top of a pool of large capacity drives with an allflash performance tier as acceleration layer. On reads the Volume Driver fetches the correct block directly from the right chunk on the ALBA backend via RDMA. Similarities Software-defined storage Both ScaleIO and Open vstorage are software-defined storage solutions which mean they are infrastructure agnostic and can run on any x86 hardware. The actual hardware on which the software runs has of course a huge impact on the performance numbers you can achieve. Scalability ScaleIO is designed to massively scale from 3 nodes up to thousands of nodes. Unlike most traditional storage systems (SANs), performance and throughput scales linearly with every node added. Every storage node added is used to process I/O operations as IO requests are dispersed across the nodes. The same applies to Open vstorage. It can scale to grow as big as 1024 nodes in a cluster and performance and capacity scale with every node added. Next to performance both solutions are built to scale capacity wise from a couple of TB up to a few petabyte. For example Open vstorage can scale to 30 PB in one cluster. Each node with storage which gets added to the cluster puts the capacity of the physical storage devices it controls at the disposal of the cluster wide pool of storage.

Key Differences Huge Performance Difference The performance of ScaleIO is well documented and can reach 200-250k IOPS per node3 for 4k random reads and writes. While this is impressive, Open vstorage typically offers around 500K IOPS for random reads per node, for example a Cisco UCS server with 2 Intel NVMe drives. This means Open vstorage is 3x faster than what typically is already considered as impressive storage performance. This huge performance difference can be explained by the fact that both solutions have radically different designs. ScaleIO uses a location-based approach where each volume is divided into 1MB fragments. Open vstorage uses a log-structured approach and uses RDMA to bypass the kernel and file system as much as possible. This approach leads to lower latency and hence better performance. Data safety ScaleIO uses a 2 copies replication strategy for every bit of data to safeguard against node failures. This 2-way replication strategy protects the user against a single disk failure. With growing disk drive capacity, this strategy will lead inevitably to data loss in large clusters. When a large capacity disk fails, it takes quite some time to rebuild the data from the dead device on other devices in the cluster. During this time the data, which was on the dead device is extremely vulnerable to data loss as there is only a single copy remaining in the cluster. Having a second disk failure might already result in data loss but also rebooting the wrong node at the wrong time leads to data unavailability. Also with 10TB drives and Unrecoverable Bit Error Rates (BER) of 1x10-14, it is almost assured that some data of the disk can t be read. Basically a 2-copy strategy is not a safe approach when storing large amounts of data and will lead to data loss. Open vstorage uses a different approach, which can be compared to solving a Sudoku puzzle. Each SCO, a collection of consecutive writes, is chopped up into chunks and some additional chunks are adjoined. All these chunks are distributed across all the nodes and datacenters in the cluster. The total amount of chunks can be configured but allows for example to recover from a multi node 3 https://www.cloudscaling.com/assets/pdf/h14196-esg-lab-spotlight-proven-performance-and-scalabilitywp.pdf

failure or a complete data center loss. A failure, whether it is a disk, node or data center will cross out some numbers from the complete Sudoku puzzle but as long as you have enough numbers left, you can still solve the puzzle. The same goes for data stored with Open vstorage: as long as you have enough chunks (disk, nodes or data centers) left, you can always recover the data. Basically with Open vstorage storage is truly safe. Encryption Another area where Open vstorage is a step ahead on data safety is the security of the data on the physical disks. While ScaleIO stores the data in the clear, Open vstorage uses AES 256 bit encryption when storing data. Open vstorage even allows to use a different encryption key per volume and can be integrated with different third party key management tools. Snapshots and Clones ScaleIO has limited the amount of snapshots and clones, which can be taken from a volume to 31 instances only. As ScaleIO uses a location-based approach, it needs to safeguard the old data before overwriting data at a certain location. Keeping track of this old data for a single snapshot is complex and slows down performance when snapshotting and as snapshots increase in number. Hence the decision to limit the amount of possible snapshots and clones for a single volume. Being able to create 31 clones severely limits ScaleIO as a backend for e.g. VDI implementations. Open vstorage on the other hand uses a log-structured approach on the backend where data is never overwritten but always appended. This means a snapshot for Open vstorage is a quick and low cost operation, just placing a marker behind the latest write. As snapshots are lightweight, you can create an unlimited amount of snapshots per volume. Snapshot can also be used to create clones. The clones are zero-copy and share the original data with the parent. As an unlimited amount of clones can be made, Open vstorage provides copy data virtualization (aka Actifio, Cohesity) out of the box without the need for separate backup and copy data virtualization software. Space efficiency As discussed earlier, ScaleIO uses a 2-way replication strategy. This isn t very space efficient, as every 1MB will lead to 2MB being stored on the physical disks. This means double the amount of nodes, double the amount of networking equipment, double the amount of power and cooling, basically doubling the TCO. Open vstorage is much more space efficient as it uses forward error correction, which provides 100x better reliability as compared to 2 copies, through writing only 1.25x more data. This means that for every 1MB, 1.25MB will be stored on the backend. Secondly, Open vstorage allows for the creation of different error correction policies for flash and HDD and hence the customer can select a lower redundancy factor on flash and a higher redundancy factor on HDD. In addition, contrary to ScaleIO Open vstorage can further reduce its storage footprint by compressing the fragments before storing them on SSDs and HDDs.

Multi datacenter ScaleIO is designed to be used on local networks, as it requires low latency links both on the read and the write path for performance reasons. Since the writes need to be acknowledged by 2 storage nodes, the latency for writes would be too high in case the second was to go to a second datacenter. Open vstorage is designed to store data safely and efficiently across multiple datacenters. To ensure local performance it uses an all-flash tier while for disaster recovery reasons data can be stored on a capacity tier across multiple datacenters. Open vstorage uses APE (Asynchronous Policy Enforcement) to ensure data gets spread according to the best available policy. In case data can t be written optimally, it will be first stored sub-optimally and later, when for example the network link between the datacenters is restored, data will be re-written with the optimal policy. Flash Friendly As ScaleIO is location based, it requires high endurance flash technology because random writes generate small updates across the device. Many updates of the same location within the volume also cause the flash chips to wear out faster than normal as you are constantly updating the same cells of the flash memory. Open vstorage is much more flash friendly as the write buffer sequentializes the random writes in large fragments, which can be written sequentially. Due to random write amplification, sequential write endurance for SSDs and PCIe flash cards is typically higher so the flash drives are less likely to fail with Open vstorage compared to ScaleIO. As every update to a volume also gets appended to a new SCO, overwriting the same LBA of a volume many times will not hammer the same cells of the flash memory over and over. Since Open vstorage is less intensive on flash compared to ScaleIO, cheaper, lower endurance SSDs can be used for the performance tier. Complete History and Integrated Backup As we move to an era of petabyte scale data sets that change often, the methods around backup and replication need to dramatically change to deal with the fast ingest of data. Secondly, there is a trend towards copy data virtualization whereby backup sets are used for test/dev and analytic workloads. In case of ScaleIO it is purely a primary storage system that would need separate products and tools for backup, replication and copy data virtualization. This not only adds cost and complexity but also affects storage performance as a significant amount of IOPS is wasted to make copies. Open vstorage combines the fact that it can take unlimited snapshots, create as many clones, flash acceleration and its multidatacenter spread using forward error correction to integrate backup, replication and copy data virtualization right into its architecture.

The bottom line Open vstorage and ScaleIO are software based distributed block storage solutions and they both scale very well. Although both can be installed on the same hardware, the different design approaches result in a remarkable performance difference. Open vstorage delivers about 3 times the performance of ScaleIO. Next to the difference in performance, Open vstorage is more space efficient and supports unlimited snapshots. With a large ScaleIO environment data is also at risk as a 2-disk failure might already cause data loss. Open vstorage stores data across datacenters and can survive even a complete datacenter going offline. Lastly, ScaleIO purely addresses the primary storage problem while Open vstorage addresses the entire storage lifecycle.