StorPool Distributed Storage Software Technical Overview

Similar documents
The Fastest And Most Efficient Block Storage Software (SDS)

Scalable, High-Performance Software Defined Storage Solution

Extremely Fast Distributed Storage for Cloud Service Providers

2014 VMware Inc. All rights reserved.

HCI: Hyper-Converged Infrastructure

Hyperconverged Cloud Architecture with OpenNebula and StorPool

SOFTWARE DEFINED STORAGE

IBM Spectrum NAS. Easy-to-manage software-defined file storage for the enterprise. Overview. Highlights

SurFS Product Description

VMware Virtual SAN. Technical Walkthrough. Massimiliano Moschini Brand Specialist VCI - vexpert VMware Inc. All rights reserved.

The OnApp Cloud Platform

Hedvig as backup target for Veeam

Build Cloud like Rackspace with OpenStack Ansible

Modern hyperconverged infrastructure. Karel Rudišar Systems Engineer, Vmware Inc.

THE ZADARA CLOUD. An overview of the Zadara Storage Cloud and VPSA Storage Array technology WHITE PAPER

Delivering Nexenta Software-Based File Services to Cisco HyperFlex

Dell EMC ScaleIO Ready Node

Quobyte The Data Center File System QUOBYTE INC.

Paperspace. Architecture Overview. 20 Jay St. Suite 312 Brooklyn, NY Technical Whitepaper

SolidFire and Ceph Architectural Comparison

THE SUMMARY. CLUSTER SERIES - pg. 3. ULTRA SERIES - pg. 5. EXTREME SERIES - pg. 9

Copyright 2012 EMC Corporation. All rights reserved.

Evaluating Cloud Storage Strategies. James Bottomley; CTO, Server Virtualization

Free up rack space by replacing old servers and storage

SOFTWARE-DEFINED BLOCK STORAGE FOR HYPERSCALE APPLICATIONS

SUSE OpenStack Cloud Production Deployment Architecture. Guide. Solution Guide Cloud Computing.

High performance and functionality

Red Hat Ceph Storage and Samsung NVMe SSDs for intensive workloads

QNAP OpenStack Ready NAS For a Robust and Reliable Cloud Platform

Deploying Software Defined Storage for the Enterprise with Ceph. PRESENTATION TITLE GOES HERE Paul von Stamwitz Fujitsu

Software Defined Storage for the Evolving Data Center

SvSAN Data Sheet - StorMagic

Introduction to Virtualization. From NDG In partnership with VMware IT Academy

VMware Virtual SAN Technology

ECONOMICAL, STORAGE PURPOSE-BUILT FOR THE EMERGING DATA CENTERS. By George Crump

StarWind Storage Appliance

THE OPEN DATA CENTER FABRIC FOR THE CLOUD

EMC Backup and Recovery for Microsoft Exchange 2007

Nutanix Tech Note. Virtualizing Microsoft Applications on Web-Scale Infrastructure

Fast and Easy Persistent Storage for Docker* Containers with Storidge and Intel

How CloudEndure Works

vsan 6.6 Performance Improvements First Published On: Last Updated On:

How CloudEndure Works

Citrix XenServer 7.1 Feature Matrix

Bacula Systems Virtual Machine Performance Backup Suite

Design a Remote-Office or Branch-Office Data Center with Cisco UCS Mini

Benefits of 25, 40, and 50GbE Networks for Ceph and Hyper- Converged Infrastructure John F. Kim Mellanox Technologies

Highly Scalable, Non-RDMA NVMe Fabric. Bob Hansen,, VP System Architecture

VxRack FLEX Technical Deep Dive: Building Hyper-converged Solutions at Rackscale. Kiewiet Kritzinger DELL EMC CPSD Snr varchitect

Turning Object. Storage into Virtual Machine Storage. White Papers

vsan Mixed Workloads First Published On: Last Updated On:

TCC, so your business continues

A product by CloudFounders. Wim Provoost Open vstorage

Database Services at CERN with Oracle 10g RAC and ASM on Commodity HW

Introducing Tegile. Company Overview. Product Overview. Solutions & Use Cases. Partnering with Tegile

How CloudEndure Disaster Recovery Works

Nimble Storage Adaptive Flash

Virtualization of the MS Exchange Server Environment

The storage challenges of virtualized environments

November 7, DAN WILSON Global Operations Architecture, Concur. OpenStack Summit Hong Kong JOE ARNOLD

An Oracle White Paper June Enterprise Database Cloud Deployment with Oracle SuperCluster T5-8

Assessing performance in HP LeftHand SANs

IBM System Storage DCS3700

VMware Virtual SAN. High Performance Scalable Storage Architecture VMware Inc. All rights reserved.

Take control of storage performance

How CloudEndure Disaster Recovery Works

HPE Strategy for VMware Cloud Foundation

DRBD SDS. Open Source Software defined Storage for Block IO - Appliances and Cloud Philipp Reisner. Flash Memory Summit 2016 Santa Clara, CA 1

EMC Backup and Recovery for Microsoft Exchange 2007 SP1. Enabled by EMC CLARiiON CX4-120, Replication Manager, and VMware ESX Server 3.

What is QES 2.1? Agenda. Supported Model. Live demo

VMware vstorage APIs FOR ARRAY INTEGRATION WITH EMC VNX SERIES FOR SAN

Bringing OpenStack to the Enterprise. An enterprise-class solution ensures you get the required performance, reliability, and security

SoftNAS Cloud Performance Evaluation on Microsoft Azure

StorPool System Requirements

HPE MSA 2042 Storage. Data sheet

VMWARE EBOOK. Easily Deployed Software-Defined Storage: A Customer Love Story

Reliable Storage for HA, DR, Clouds and Containers Philipp Reisner, CEO LINBIT

Intel Solid State Drive Data Center Family for PCIe* in Baidu s Data Center Environment

Vendor: EMC. Exam Code: E Exam Name: Cloud Infrastructure and Services Exam. Version: Demo

Forget IOPS: A Proper Way to Characterize & Test Storage Performance Peter Murray SwiftTest

Stellar performance for a virtualized world

StorMagic SvSAN 6.1. Product Announcement Webinar and Live Demonstration. Mark Christie Senior Systems Engineer

Design a Remote-Office or Branch-Office Data Center with Cisco UCS Mini

DELL EMC VXRACK FLEX FOR HIGH PERFORMANCE DATABASES AND APPLICATIONS, MULTI-HYPERVISOR AND TWO-LAYER ENVIRONMENTS

Virtual SAN and vsphere w/ Operations Management

Eliminate the Complexity of Multiple Infrastructure Silos

Deep Dive on SimpliVity s OmniStack A Technical Whitepaper

Data Protection for Cisco HyperFlex with Veeam Availability Suite. Solution Overview Cisco Public

Delivering unprecedented performance, efficiency and flexibility to modernize your IT

OpenIO SDS on ARM A practical and cost-effective object storage infrastructure based on SoYouStart dedicated ARM servers.

HyperFlex. Simplifying your Data Center. Steffen Hellwig Data Center Systems Engineer June 2016

StorMagic SvSAN: A virtual SAN made simple

INTRODUCTION TO CEPH. Orit Wasserman Red Hat August Penguin 2017

HYBRID STORAGE TM. WITH FASTier ACCELERATION TECHNOLOGY

Webinar Series: Triangulate your Storage Architecture with SvSAN Caching. Luke Pruen Technical Services Director

New HPE 3PAR StoreServ 8000 and series Optimized for Flash

Tiered IOPS Storage for Service Providers Dell Platform and Fibre Channel protocol. CloudByte Reference Architecture

Copyright 2012, Oracle and/or its affiliates. All rights reserved.

DataON and Intel Select Hyper-Converged Infrastructure (HCI) Maximizes IOPS Performance for Windows Server Software-Defined Storage

EMC ScaleIO In The Enterprise: The Citi Experience. Tamir Segal Head of ScaleIO Marketing Eliot A. Wilson Senior Engineer, Citi

Transcription:

StorPool Distributed Storage Software Technical Overview StorPool 2018 Page 1 of 9

StorPool Overview StorPool is distributed storage software. It pools the attached storage (hard disks or SSDs) of standard servers to create a single pool of shared block storage. The StorPool software is installed on each server in the cluster and combines the performance and capacity of all drives attached to the servers into one global namespace. StorPool provides standard block devices. You can create one or more volumes through its sophisticated volume manager. StorPool is compatible with ext4 and XFS file systems and with any system designed to work with a block device, e.g. databases and cluster file systems (like OCFS and GFS). StorPool can also be used with no file system, for example when using volumes to store VM images directly. Fig 1. StorPool - Logical diagram Redundancy is provided by multiple copies (replicas) of the data written synchronously across the cluster. Users set the number of replication copies. We recommend 3 copies as a standard and 2 copies for data that is less critical. This technology is superior to RAID in both reliability and performance: Unlike RAID, StorPool replication provides copies on different servers (not copies on one server). So even in the case of a server or component failure you do not lose the data that sits on this server. Rebuilding a failed drive/node is done from many drives (and not from one particular drive as in RAID). This means rebuild times are significantly shorter and they do not impact the overall system performance StorPool also protects data and guarantees data integrity by a 64-bit checksum and a version number for each sector maintained by StorPool. StorPool 2018 Page 2 of 9

StorPool provides a very high degree of flexibility in volume management. Unlike other storage technologies, such as RAID or ZFS, StorPool does not rely on device mirroring (pairing drives for redundancy). So every disk that is added to a StorPool cluster adds capacity to the cluster, not just for new data but also for existing data. Provided that there are sufficient copies of blocks, drives can be added or taken away with no impact to the storage service. Unlike rigid systems like RAID, StorPool does not impose any strict hierarchical storage structure that links and reflects onto the underlying disks. StorPool simply creates a single pool of storage that utilizes the full capacity and performance of a set of commodity drives. Architecture StorPool works on a cluster of servers in a distributed shared-nothing architecture. All functions are performed by all the servers on an equal peer basis. It works on standard off-theshelf servers running GNU/Linux or is attached via iscsi to hosts running other hypervisors/os. The software consists of two parts - a storage server and a storage client that are installed on each physical server (host, node). Each host can be a storage server, a storage client, or both (i.e. converged set-up, converged infrastructure). To storage clients, StorPool volumes appear as block devices under /dev/storpool/* and behave identically to the dedicated physical alternatives. Data on volumes can be read and written by all clients simultaneously and consistency is guaranteed through a synchronous replication protocol. Volumes can be used by clients as they would use a local hard drive or disk array. Fig. 2. Logical network diagram, segregated. In practice we recommend converged deployments a StorPool Client and a StorPool Server on the same physical server. StorPool 2018 Page 3 of 9

Each storage node is responsible for data stored on its local drives. Storage nodes collaborate to provide the storage service. StorPool provides a shared storage pool combining all the available storage capacity. StorPool uses synchronous replication across servers. The StorPool client communicates in parallel with all StorPool servers. In a typical deployment, each virtual machine (or container) is given its own volume in StorPool. The volume is attached to the hypervisor and the virtual machine is configured to use it directly. There is no filesystem, or VM image format (such as QCOW) in between. This gives you the capability to control each VMs storage requirements independently. For example you could have different placement (on SSDs, on HDDs) configuration, replication (2 copies, 3 copies) and different Quality of Service, QoS (IOPS, MB/s) configuration. This architecture also enables you to perform VM snapshots independently for each VM. When used with VMware or Windows/Hyper-V, StorPool is attached through iscsi. We have extended the iscsi target functionality to add High-Availability (HA) and Scale-out functionality of the storage backend. The system is accessed through the standard (build in) iscsi initiators of VMware and Hyper-V (without support for CVS /cluster shared volumes/, yet). Fig. 3. High-level diagram of a Hyper-Converged StorPool system (on the bottom), which also can provide shared storage to VMware or Hyper-V through iscsi. Using StorPool StorPool can be thought of as a SAN replacement or a virtual SAN. StorPool is accessed with the StorPool driver, installed on the client (initiator) machine. The driver provides block-layer volumes which are visible as block devices under /dev/storpool/* on StorPool clients. StorPool 2018 Page 4 of 9

Currently, StorPool supports Linux with KVM, Xen, LXC, Docker and any other technology compatible with the Linux storage stack. StorPool is integrated with OpenStack, OpenNebula, OnApp, CloudStack, lipvirt and Proxmox, as well as custom cloud management solutions. User applications are installed on top of the Linux block device provided by the StorPool storage solution. StorPool is compatible with any system designed to work with a block device. For example, in a virtualized environment virtual volumes are created from the StorPool device and assigned to a single virtual machine directly mapped upon it. Alternatively, StorPool block devices can be used with shared-disk file systems (like GFS2) or Clustered LVM. It is also possible to format volumes with standard file systems such as ext4 or XFS, which enables you to mount the file system on one Linux host at a time. StorPool combines the space, bandwidth and IOPS of all your drives and leaves plenty of CPU and RAM for compute loads. This means virtual machines, applications, databases or any other compute load can run on the same server as StorPool. When using StorPool customers can run both storage and compute loads on the same servers to achieve higher levels of utilization. The StorPool volume management is very flexible and allows users to specify which drives should be used for which copies (e.g. the 1 st copy on SSDs and the 2 nd and 3 rd copies on hard disks). It also allows users to specify which exact drives a specific volume should be stored on. These Data Placement Policies can also be used to ensure high availability and performance, including data locality, e.g. one copy stored on a particular server for faster local access. StorPool has many advantages over traditional storage architectures: It scales in capacity and performance with every added drive or server It works on standard servers alongside applications and virtual machines It delivers very high performance while at the same time it takes minimal CPU and RAM from participating servers It provides better reliability and protects against silent data corruption It comes at a fraction of the cost of other storage solutions Our key benefits, compared to other distributed storage systems, are: Higher Performance (IOPS, MB/s) Higher Efficiency (low CPU usage, low RAM usage) Simplicity and Compatibility Typical Use Cases Public / Private Clouds StorPool is designed to combine storage and compute on the same node to deliver a high reliability, high performance cloud infrastructure while dramatically reducing the Total Cost of Ownership. StorPool can be used on the same servers as virtual machines, in conjunction with the hypervisor (e.g. KVM). StorPool can also be used without a hypervisor, in the so called "bare-metal" cloud architecture, for example to provide resilience and high availability for a StorPool 2018 Page 5 of 9

stand-alone application. This can be achieved by keeping all the data in the StorPool cluster, so if the application server dies it is easy to spin it up on another machine. Virtual Desktops (VDI) StorPool is well suited for virtual desktop environments. StorPool s clustered and non-raid approach leverages the throughput of the entire network and the available performance of the underlying hardware. This means it delivers incredible IOPS performance. In addition the finegrained volume control and thin provisioning makes StorPool an ideal storage solution for VDI. Dev Ops As with the VDI use case StorPool s fine-grained volume control makes it easy and quick for new instances to be assigned storage from the pool. StorPool simply creates a standard storage (block) device from the cluster of servers, so no specialist skills are required to manage it. A junior System Administrator will be able to create (resize and delete) volumes and assign them to virtual machines. How is StorPool Better? Higher Performance StorPool is a genuine scale-out storage system that leverages the entire cluster of servers to give massively superior IOPS performance compared to other software storage solutions (e.g. ZFS, Ceph, etc.). StorPool works much closer to the hardware, so it is much closer to the raw IO speeds of the underlying hardware. StorPool also achieves unmatched performance through its superior network protocol and its copy-on-write on-disk format, while also accelerating random writes and mitigating the "I/O blender effect". Scalability & Flexibility Individual drives in a StorPool cluster can be added and taken away with no impact on the service/data availability, provided sufficient copies have been made. The StorPool cluster is a flat hierarchy of drives and new drives can be added or removed at any time. When new drives are added to a StorPool cluster the capacity of these drives is simply added to the pool. StorPool automatically ensures that copies are placed in separate fault domains (typically separate servers), so there is no need to plan and configure RAIDs. StorPool volume management also offers a higher flexibility than most storage systems. StorPool allows for the creation of multiple volumes each with a different placement policy out of any subset of drives from the pool. For instance you might want some volumes on SATA drives, while other volumes on SSDs. Each volume can also be configured for data locality, keeping data close to where it is used. Simplicity & Ease of Use StorPool creates block storage volumes from a cluster of servers. Volumes can be created and assigned using the StorPool CLI. After creation, volumes are regular block devices and can be monitored with standard tools such as iostat. Unlike many other storage solutions StorPool does not require special skills to operate (i.e. it is not a new file system). А junior Sys Admin is usually able to operate StorPool. StorPool 2018 Page 6 of 9

Lower Cost Unlike big SAN boxes from major storage vendors, StorPool uses commodity hardware. This is why StorPool could cost up to 10 times less than a comparable dedicated storage hardware solution. In addition it is cheaper to set-up, maintain and support. It is much easier to procure as well. Feature Highlights Scale-out, not Scale-Up The StorPool solution is fundamentally about scaling-out (scaling by adding more drives or nodes) rather than scaling-up (adding capacity by replacing a storage box with a larger storage box). This means StorPool can scale independently by IOPS, storage space and bandwidth. There is no bottleneck or single point of failure. StorPool can grow without interruption and in small steps - one hard drive, one server and one network interface at a time. High Performance StorPool combines the IOPS performance of all drives in the cluster and optimizes drive access patterns to provide low latency and handling of storage traffic bursts. Load is distributed equally between all servers through striping and sharding. High Availability and Reliability StorPool uses a replication mechanism that slices and stores copies of the data on different servers. For primary, high performance storage this solution has many advantages compared to RAID systems and provides considerably higher levels of reliability and availability. In case of drive, server or other component failure, StorPool uses another copy of the data located on another server (or rack) and none of your data is lost or even temporarily unavailable. Commodity Hardware StorPool supports drives and servers in a vendor-agnostic manner, allowing you to avoid vendor lock-in. This allows the use of commodity hardware, while preserving reliability and performance requirements. Moreover, unlike RAID, StorPool is drive agnostic - you can mix drives of various types, make, speed or size in a StorPool cluster. Shared Block Device StorPool provides shared block devices with semantics identical to a shared iscsi or FC disk array. Co-existence with hypervisor software StorPool can utilize re-purposed existing servers and can co-exist with hypervisor software on the same server. This means that there is no dedicated hardware for storage, and growing an IaaS cloud solution is achieved by simply adding more servers to the cluster. Compatibility StorPool is compatible with 64-bit Intel and AMD based servers. We support all the Linux-based hypervisors and hypervisor management software. Any Linux software designed to work on a shared storage solution such as an iscsi or FC disk array will work on StorPool. StorPool guarantees the functionality and availability of the storage solution at the Linux block device interface. StorPool 2018 Page 7 of 9

CLI interface and API StorPool provides an easy yet powerful Command Line Interface (CLI) for administration of the data storage solution. It is simple and user-friendly making configuration, provisioning and monitoring fast and efficient. StorPool also provides a RESTful JSON API, exposing all the available functionality, so you can integrate it with any existing management system. We are also working on adding a web-based user interface (GUI). Reliable Support StorPool comes with reliable dedicated support: Remote installation and initial configuration by StorPool's specialists 24x7 support Software updates Compatibility List & Features Software: Servers: We support all the Linux distributions, typical ones are CentOS, RHEL, Debian, Ubuntu LTS KVM, Xen, LXC and other Linux-based virtualization technologies Other operating systems / hypervisors such as Microsoft Windows Server, Hyper-V, VMware vsphere/esx/esxi - through highly available, scale-out iscsi All major file systems and logical volume managers ext2/3/4, XFS, GFS2, LVM, LIO, etc. Servers with 64-bit x86 CPUs from Intel and AMD Recommended minimal configuration 3 cores, 32 GB RAM Recommended Sandy Bridge microarchitecture or newer CPU (e.g. Xeon E5) Network: StorPool server resource utilization 1 core, 3 GB RAM StorPool client resource utilization 1 core, 1 GB RAM 10-Gigabit Ethernet or faster Infiniband QDR/FDR Enterprise-level Feature Set: Distributed, scale-out, shared-nothing cluster architecture high performance, scalability, availability, reliability End-to-end data integrity our protection mechanisms prevent silent data corruption, phantom/partial/misplaced writes, common for systems handling large amounts of data StorPool 2018 Page 8 of 9

Online configuration changes volumes are not brought down during reconfiguration. In-service software upgrades the storage service continues operating while the cluster is being upgraded on a rolling basis. Snapshots and clones - instantaneous, copy-on-write (CoW) snapshots and clones Storage Quality of Service (QoS) per volume fairness between volumes and userconfigurable per-volume limit of IOPS and MB/s Thin provisioning and thin reclaim (TRIM/DISCARD) Write-back cache (WBC) Converged Infrastructure - Storage+Compute use the same servers for storage and applications / virtual machines. StorPool uses only 5-15% of CPU and RAM. Storage pools each pool is a set of drives. There may be a single pool or multiple pools per cluster. Supports hard-drive pools for high capacity use-cases and hybrid SSD+HDD pools for high performance use-cases. Data locality each volume can be configured independently for high performance access from a particular node Data tiering data on HDD pool or on SSD pool. Live migration between pools. Synchronous replication - replication level set per volume RESTful JSON API with API failover Self-healing & automatic data balancing Shared block devices available on all the client nodes, simultaneously accessible and fully consistent. RDMA and NVMe SSD support extremely fast and low latency storage system with high-end SAN feature set. About StorPool StorPool Storage is the leader in new-age software-defined storage. It is unmatched in terms of reliability, performance, hardware resource utilization and scalability. Leading global Enterprises like NASA, CERN, SIEMENS, Deutsche Bourse, NASDAQ Dubai, European Space Agency, Samsung, SiteGround and many others have data running on StorPool. The company has presence in North America, Europe, Middle East and Asia & Oceania regions. With global list of customers, 15 major releases and solid growth, StorPool is the reliable, yet flexible partner for any public or private cloud builder. Contact us to learn more or schedule a free test StorPool Storage info@storpool.com +1 415 670 9320 www.storpool.com StorPool 2018 Page 9 of 9