Understanding Data Locality in VMware vsan First Published On: Last Updated On:

Similar documents
Understanding Data Locality in VMware Virtual SAN

VMware Virtual SAN Data Management Operations

What s New in VMware vsphere Flash Read Cache TECHNICAL MARKETING DOCUMENTATION

VMware Virtual SAN Design and Sizing Guide for Horizon View Virtual Desktop Infrastructures TECHNICAL MARKETING DOCUMENTATION REV A /JULY 2014

vsan 6.6 Performance Improvements First Published On: Last Updated On:

VMware Virtual SAN. Technical Walkthrough. Massimiliano Moschini Brand Specialist VCI - vexpert VMware Inc. All rights reserved.

VMware vsphere APIs for I/O Filtering (VAIO) November 14, 2017

Modern hyperconverged infrastructure. Karel Rudišar Systems Engineer, Vmware Inc.

2014 VMware Inc. All rights reserved.

VMware Virtual SAN Technology

Extremely Fast Distributed Storage for Cloud Service Providers

Nutanix Tech Note. Virtualizing Microsoft Applications on Web-Scale Infrastructure

Why Datrium DVX is Best for VDI

How it can help your organisation

What's New in vsan 6.2 First Published On: Last Updated On:

Performance Testing December 16, 2017

Deep Dive on SimpliVity s OmniStack A Technical Whitepaper

Stellar performance for a virtualized world

VMware Virtual SAN. High Performance Scalable Storage Architecture VMware Inc. All rights reserved.

VMWARE VIRTUAL SAN: ENTERPRISE-GRADE STORAGE FOR HYPER- CONVERGED INFRASTRUCTURES CHRISTOS KARAMANOLIS RAWLINSON RIVERA

SAN Acceleration Using Nexenta Connect View Edition with Third- Party SAN Storage

Expert Reference Series of White Papers. Five Reasons Why You Should Pair VSAN and View

Virtual SAN and vsphere w/ Operations Management

VMWARE VSAN LICENSING GUIDE - MARCH 2018 VMWARE VSAN 6.6. Licensing Guide

Virtualization of the MS Exchange Server Environment

Running VMware vsan Witness Appliance in VMware vcloudair First Published On: April 26, 2017 Last Updated On: April 26, 2017

VMworld 2013 Overview

Dell EMC ScaleIO Ready Node

How Architecture Design Can Lower Hyperconverged Infrastructure (HCI) Total Cost of Ownership (TCO)

VSAN Virtual Desktop Infrastructure Workload First Published On: Last Updated On:

Introduction to Virtualization. From NDG In partnership with VMware IT Academy

Building Your Own Robust and Powerful Software Defined Storage with VMware vsan. Tips on Choosing Hardware for vsan Deployment

VMWARE EBOOK. Easily Deployed Software-Defined Storage: A Customer Love Story

"Software-defined storage Crossing the right bridge"

Nutanix White Paper. Hyper-Converged Infrastructure for Enterprise Applications. Version 1.0 March Enterprise Applications on Nutanix

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme

VMware + OpenStack. Dan Wendlandt Director of Product Management VMware VMware Inc. All rights reserved.

ProphetStor DiskProphet Ensures SLA for VMware vsan

What s New in VMware Virtual SAN (VSAN) v 0.1c/AUGUST 2013

IOmark-VM. VMware VSAN Intel Servers + VMware VSAN Storage SW Test Report: VM-HC a Test Report Date: 16, August

Delivering HCI with VMware vsan and Cisco UCS

COPYRIGHTED MATERIAL. Introducing VMware Infrastructure 3. Chapter 1

HCI: Hyper-Converged Infrastructure

StarWind Virtual SAN Free

What s New in VMware vsphere 4.1 Performance. VMware vsphere 4.1

Building an application-centric " storage platform

The next step in Software-Defined Storage with Virtual SAN

Free up rack space by replacing old servers and storage

Best Practices for Deploying a Mixed 1Gb/10Gb Ethernet SAN using Dell EqualLogic Storage Arrays

Dell Technologies IoT Solution Surveillance with Genetec Security Center

The Impact of Hyper- converged Infrastructure on the IT Landscape

Condusiv s V-locity VM Accelerates Exchange 2010 over 60% on Virtual Machines without Additional Hardware

SAN Virtuosity Fibre Channel over Ethernet

Microsoft SQL Server 2014 on vsan 6.2 All-Flash December 15, 2017

iocontrol Reference Architecture for VMware Horizon View 1 W W W. F U S I O N I O. C O M

Managing Performance Variance of Applications Using Storage I/O Control

vsan Space Efficiency Technologies First Published On: Last Updated On:

Performance & Scalability Testing in Virtual Environment Hemant Gaidhani, Senior Technical Marketing Manager, VMware

Software-Defined Data Infrastructure Essentials

VMware Join the Virtual Revolution! Brian McNeil VMware National Partner Business Manager

INTEL NEXT GENERATION TECHNOLOGY - POWERING NEW PERFORMANCE LEVELS

New Approach to Unstructured Data

Converged Platforms and Solutions. Business Update and Portfolio Overview

TITLE. the IT Landscape

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme

NetApp Clustered Data ONTAP 8.2 Storage QoS Date: June 2013 Author: Tony Palmer, Senior Lab Analyst

Nimble Storage Adaptive Flash

Fast and Easy Persistent Storage for Docker* Containers with Storidge and Intel

Architecting For Availability, Performance & Networking With ScaleIO

Hyper-V Innovations for the SMB. IT Pro Camp, Northwest Florida State College, Niceville, FL, October 5, 2013

NAS for Server Virtualization Dennis Chapman Senior Technical Director NetApp

vsan Mixed Workloads First Published On: Last Updated On:

TECHNICAL WHITE PAPER - JANUARY VMware Horizon 7 on VMware vsan Best Practices TECHNICAL WHITE PAPER

Evaluation Report: Improving SQL Server Database Performance with Dot Hill AssuredSAN 4824 Flash Upgrades

DataON and Intel Select Hyper-Converged Infrastructure (HCI) Maximizes IOPS Performance for Windows Server Software-Defined Storage

HPE Strategy for VMware Cloud Foundation

Identifying Performance Bottlenecks with Real- World Applications and Flash-Based Storage

Infinio Accelerator Product Overview White Paper

Table of Contents HOL HCI

VxRack FLEX Technical Deep Dive: Building Hyper-converged Solutions at Rackscale. Kiewiet Kritzinger DELL EMC CPSD Snr varchitect

Adaptive Resync in vsan 6.7 First Published On: Last Updated On:

VMware vsan 6.6. Licensing Guide. Revised May 2017

NetApp HCI. Ready For Next. Enterprise-Scale Hyper Converged Infrastructure VMworld 2017 Content: Not for publication Gabriel Chapman: Sr. Mgr. - NetA

VMWARE HORIZON 6 ON HYPER-CONVERGED INFRASTRUCTURES. Horizon 6 version 6.2 VMware vsphere 6U1 / VMware Virtual SAN 6U1 Supermicro TwinPro 2 4 Nodes

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme

Performance Sentry VM Provider Objects April 11, 2012

Microsoft SQL Server 2014 on VMware vsan 6.2 All-Flash October 31, 2017

VMware vsan Ready Nodes

Developing a Hyper- Converged Storage Strategy for VMware vcloud Director with VMware vsan

FLASHARRAY//M Smart Storage for Cloud IT

Take control of storage performance

Enterprise Architectures The Pace Accelerates Camberley Bates Managing Partner & Analyst

W H I T E P A P E R. What s New in VMware vsphere 4: Performance Enhancements

VMware vsphere Beginner s Guide

Dell EMC Hyper-Converged Infrastructure

On-Premises Cloud Platform. Bringing the public cloud, on-premises

The vsphere 6.0 Advantages Over Hyper- V

vsan Remote Office Deployment January 09, 2018

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme

Hedvig as backup target for Veeam

Transcription:

Understanding Data Locality in VMware vsan First Published On: 07-20-2016 Last Updated On: 09-30-2016 1

Table of Contents 1. Understanding Data Locality in VMware vsan 1.1.Introduction 1.2.vSAN Design Goals 1.3.Data Locality 1.4.vSAN Design Considerations 1.5.Data Locality and Latency 1.6.Data Placement and Resource Utilization 1.7.Effective Data Locality 1.8.Conclusion 1.9.Acknowledgments 1.10.About the Author 2

1. Understanding Data Locality in VMware vsan This document discusses how vsan design choices exploit data locality to deliver superior aggregate performance while retaining simplicity and efficiency. 3

1.1 Introduction VMware vsan is a new hypervisor-converged, software-defined storage platform that is fully integrated with VMware vsphere. vsan aggregates locally attached disks of hosts that are members of a vsphere cluster, to create a distributed shared storage solution. VMware vsan is a new hypervisor-converged, software-defined storage platform that is fully integrated with VMware vsphere. vsan aggregates locally attached disks of hosts that are members of a vsphere cluster, to create a distributed shared storage solution. The distributed datastore of vsan is an object storage system that leverages the vsphere Storage Policy Based Management (SPBM) framework to deliver applicationcentric storage services and capabilities that are centrally managed through vsphere virtual machine storage policies. This document discusses how vsan design choices exploit data locality to deliver superior aggregate performance while retaining simplicity and efficiency. 1.2 vsan Design Goals One of the primary design goals for vsan is to deliver a new level of storage management simplicity and ease-of-use to the VMware administrator. vsan achieves this by creating an abstracted pool of storage resources that can be used easily and efficiently across the entire vsphere cluster. While capabilities exist to tune performance for individual workloads, vsan strives to eliminate most of the manual effort associated with traditional per-workload storage management. Many of the design choices found in vsan are a direct result of observing thousands of production vsphere clusters. 1.3 Data Locality 4

In computer science, data locality, also known, as locality of reference is the behavior of computer programs according to which a workload accesses a set of data entities or storage locations within some period of time with a predictable access pattern. There are two main types of data locality: Temporal locality-the probability that if some data (or a storage location) is accessed at one point in time, then it will be accessed again soon afterwards. Spatial locality-the probability of accessing some data (or a storage location) soon after some nearby data (or a storage location) on the same medium has been accessed. Sequential locality is a special case of spatial locality, where data (or storage locations) are accessed linearly and according to their physical locations. Data locality is particularly relevant when designing storage caches. For example, flash devices offer impressive performance improvements at a cost so efficient use of these resources becomes important. 1.4 vsan Design Considerations Like any storage system, VMware vsan makes use of data locality. vsan uses a combination of algorithms that take advantage of both temporal and spatial locality of reference to populate the flash-based read caches across a cluster and provide high performance from available flash resources. Examples include: Every time application data is read by a virtual machine, vsan saves a copy of the data in the Read Cache portion of the flash device associated with the disk group where the copy of the data resides. Temporal locality implies that there is high probability that said data will be accessed again before long. In addition, vsan predictively caches disk blocks in the vicinity of the accessed data (in 1MB chunk at a time) to take advantage of spatial locality as well. vsan uses an adaptive replacement algorithm to evict data from the Read Cache when it is deemed unlikely that the data will be accessed again soon, and uses the space for new data more likely to be accessed repeatedly. vsan makes replicas of storage objects across multiple servers for protection purposes. Reads are distributed across the replicas of an object for better load balancing. However, a certain range of logical addresses of an object is always 5

read from the same replica. This approach has two important benefits: 1. Increases the chances that the data accessed is already in the Read Cache. 2. A data block is never cached in more than one flash device. To be clear, a fundamental design decision for vsan is to not implement a persistent client-side local read cache. The decision was based on the following observations regarding local read caching on flash: Local read caching results in very poor balancing of flash utilization (both capacity and performance) across the cluster. Local read caching requires transferring hundreds of gigabytes of data and cache re-warming when virtual machines are vmotioned between hosts to keep compute resources balanced. Local read caching offers negligible practical benefits in terms of performance metrics, such as latency. Let s see in more detail the rationale behind this architectural approach of vsan. 1.5 Data Locality and Latency One theoretical argument in favor of caching VMDK data on the same host as the virtual machine is that of improved read access latencies by avoiding the network latency overhead. If true, this argument would only apply to reads, and not protected writes. As with all clustered storage systems, protected data must be safely written on more than one server, meaning that all writes must be network writes. Thus, any local caching approach will only be theoretically effective for reads, and not for protected writes. The majority of customers with production vsan deployments (and for that matter any hyper-converged storageproduct) are using 10Gigabit Ethernet (10GbE). 10GbE networks have observed latencies in the range of 5-50 microseconds. (Ref: Qlogic s Introduction to Ethernet Latency) Those latencies are achieved with sustained workloads and they include packet processing in the network stack (TCP/IP). Those numbers are consistent with what we observe with the ESX network stack. In the case of vsan this happens on the VMkernel network adapter. 6

One may argue that given the low access latencies of flash devices, network latencies may become apparent when data is accessed remotely. However, a deeper look shows this not to be the case. A typical enterprise grade flash device (SSD or PCIe card) has advertised optimal latencies in the range of 20-100 microseconds (usec). However, this is applicable only for operations issued to the device one at a time. Most flash storage devices have limited internal parallelism; they can typically process only a few operations in parallel. In real-world use cases and certainly in the case of virtualized environments, storage systems use flash devices to maximize the I/O Operations per Second (IOPS) provided to the workloads. Because of the limited internal parallelism of the devices, the storage system needs to maintain a queue of I/O operations on the device to achieve a high number of IOPS (by keeping the pipeline full). Depending on the device, the queue depths required to maximize IOPS may range from 32 to 256 operations. While high queue depths result in better IOPS and throughput, they also increase the observed latency per operation; each operation spends more time in the device queue waiting to be processed. For most flash devices, maximum IOPS are achieved with latencies that are close to and some times exceed 1 millisecond (msec). As a real-world example, consider the recent Intel DC S3700 enterprise SSD, a device with very dependable performance characteristics. The device is rated to deliver up to 75,000 4K reads with a queue size of 32 operations. At the same time, its quality of service with that queue depth is rated to 4K reads in 1 msec, 99.9% of the time (Ref: Solid-State Drive DC S3700 Series page 7). Another typical SSD used in enterprise solutions is the Intel DC S3500. The device is rated to deliver up to 75,000 4K reads with a queue size of 32 operations. In this case, the latency range is even higher. The device is rated to deliver 4K reads in 2 msec, 99.9% of the time, using a queue size of 32. (Ref: Solid-State Drive DC S3700 Series) Given the above data, it becomes clear that the latency introduced by the network is negligible compared to the flash device latencies. The network adds 10-100 usec toa storage device latency that is 1-2 msec range. This has been confirmed in lab experiments: it ends up being extremely difficult to measure the additional latency solely introduced by the network. For the vast majority of workloads, vsan s approach delivers very good read performance from read cache, faster than many traditi onal storage arrays. vsan delivers this high level of read performance while efficiently using SSD resources (only one copy of cached data) as well as preserving an extremely simple experience for vsphere administrators. 1.6 Data Placement and Resource Utilization Given that network access of cached reads doesn t impose a noticeable performance penalty, what additional advantages result from the vsan design, which distributes the cache across the cluster? 7

There are several: Improved Load Balancing and Resource Utilization By being able to distribute data on any host and any device in the cluster, vsan can achieve superior resource utilization. All resources: capacity, IOPs and throughput are available to all the VMs in the cluster, regardless of which server they reside on. Better overall resource utilization in the cluster results in better aggregate performance by eliminating hot spots on individual hosts or devices. It also results in superior economics maximize the bang-for-buck achieved from expensive flash devices. Less Data Migration With read caches local to the client, one has to pay the penalty of bulk data movement and/or cache re-warming every time a virtual machine is vmotioned across hosts in the cluster. These bulk data moves may take longer than expected, and/or consume an inordinate amount of network resources. As a result, there is a considerable performance impact on a VM, for a while after vmotion is completed. Load balancing the compute resources (CPU and memory) in a cluster should not mean that one needs to also do expensive data moves. vsphere DRS (Distributed Resource cheduler), for example, does not need to calculate the cost of moving cache contents when optimizing VM placement in the cluster. For this reason, existing hyper-converged storage products that use client-local caches recommend to their customers that vsphere DRS is disabled in order to reduce vmotion migrations. That eliminates one of the big benefits of virtualization. vsan is not architected with data locality dependencies, and thus delivers more consistent performance. The results from an example test illustrate this. The workload that was run is a typical transactional application, and the orders per second (on a five-minute moving average) were measure throughout the duration of the test. In this test, a virtual machine started running on a host that did not contain any disk replicas (i.e. all disk replicas were running on other hosts). After five minutes the VM was migrated to another host with no disk replicas (marked as the first vmotion event in Figure 3). Five minutes after that it was migrated to a host with one local replica. The application performance remained consistent during and 8

after the migration. 1.7 Effective Data Locality The above arguments are not meant to dismiss the usefulness of read caches local to the client. Such caches can be effective under the following conditions: The latency of accessing a local cache is orders of magnitudes faster (lower latency) than going over the network. With today s technologies, that is possible when the cache is in RAM (latencies measured in nanoseconds). When the medium where the local cache resides is not constrained by the narrow parallelism of storage devices. Again, RAM is the obvious choice here. When a small local cache can go a long way. A good example is when you can share a lot of read-mostly data among many similar virtual machines on a host. Having a small cache is also warranted for a constrained and relatively expensive resource such as RAM. An obvious use case that meets the above conditions is Virtual Desktops. These deployments usually involve: High consolidation ratios (many virtual machines per host) Large ratio of common data between similar OS images, which can be effectively de-duplicated in the local cache. These types of use cases can benefit by using some form of in-memory caching solution that uses de-duplication, such as VMware s Content-Based Read Cache (CBRC). Content-Based Read Cache (CBRC) is a vsphere hypervisor feature that is currently utilized exclusively with Horizon View and is called View Storage Accelerator. It is a very fast in-memory (volatile RAM) dedup ed Read Cache residing on every host where the data is accessed. This feature allows for a read cache layer to be constructed in memory that is optimized for recognizing, handling, and de-duplicating virtual desktop client 9

images. View Storage Accelerator is configured and enabled in the vsphere hypervisor and managed by the Horizon View Composer. View Storage Accelerator delivers a significant reduction in read IOPS, as high as 90% for certain operating system images and workloads. The amount of reduction in IOPS enables large scaling capabilities for the virtual desktops in the case of I/O storm workloads. That s typically found in large virtual desktop deployment scenarios, when many VMs go through a boot sequence at the same time. vsan can be combined with View Storage Accelerator and thus use local read caching based on RAM, which is the more effective application of data locality. 1.8 Conclusion In conclusion, the benefits of better balancing and improved overall resource utilization in a cluster using hypervisor-converged storage with flexible data placement consistently outweigh any practical performance benefits of local caching on flash devices. Local read caches may add additional expense due to bulk data transfers during VM migration (vmotion), they do not apply to protected writes, and they can introduce additional complexity if not fully integrated with other management tools in use. Local caching is best used only where the cache medium is much faster than network and storage device latencies, and where a small cache can go a long way. The View Storage Accelerator feature, 10

based on Content-Based Read Cache, is designed to work exactly within these constraints and is an effective use of local caching for virtual desktop deployments. The VMware engineering team considered all potential options when designing vsan. Their choices reflect the need to deliver superior performance, great economics as well as an extremely simple and consistent management experience for vsphere administrators. 1.9 Acknowledgments I would like to thank Christos Karamanolis, Principal Architect of VMware R&D, and lead engineer of vsan. This paper would not have been possible without him, his contributions. Christos deep knowledge and understanding of the product was leveraged throughout this paper. I would also like to thank Chuck Hollis, Chief Strategist of the Storage and Application Services Business Unit for his contributions and Charu Chaubal, group manager of the Storage and Availability Technical Marketing team for his contributions to this paper. 1.10 About the Author Rawlinson Rivera is a senior architect in the Cloud Infrastructure Technical Marketing group at VMware focused on Software-Defined Storage technologies such as vsan, Virtual Volumes, as well as the integration of VMware storage products with the OpenStack framework. As a previous architect in the VMware Cloud Infrastructure and Management Professional Services organization, he specialized in vsphere and cloud enterprise architectures for VMware Fortune 100 and 500 customers. Rawlinson is among the few VMware Certified Design Experts (VCDX#86) in the world and is the author of multiple books based on VMware and other technologies. Follow Rawlinson s blogs: http://blogs.vmware.com/vsphere/storage http://www.punchingclouds.com Follow Rawlinson on Twitter: @PunchingClouds 11