Expert Reference Series of White Papers. VSAN: Reimagining Storage in vsphere

Similar documents
Expert Reference Series of White Papers. Five Reasons Why You Should Pair VSAN and View

2014 VMware Inc. All rights reserved.

VMware Virtual SAN. Technical Walkthrough. Massimiliano Moschini Brand Specialist VCI - vexpert VMware Inc. All rights reserved.

Introduction to Virtualization. From NDG In partnership with VMware IT Academy

vsan Remote Office Deployment January 09, 2018

Native vsphere Storage for Remote and Branch Offices

Running VMware vsan Witness Appliance in VMware vcloudair First Published On: April 26, 2017 Last Updated On: April 26, 2017

What s New in VMware Virtual SAN (VSAN) v 0.1c/AUGUST 2013

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme

DumpExam. The best exam dump, valid dumps PDF, accurate exam materials provider

StarWind Virtual SAN Free

vsan Disaster Recovery November 19, 2017

VMware vsphere: Taking Virtualization to the Next Level

VMware vsphere Clusters in Security Zones

Installation and Cluster Deployment Guide for VMware

Configuring and Managing Virtual Storage

Installation and Cluster Deployment Guide

Nutanix Tech Note. Virtualizing Microsoft Applications on Web-Scale Infrastructure

vsan Security Zone Deployment First Published On: Last Updated On:

vsan Mixed Workloads First Published On: Last Updated On:

Changes in VCP6.5-DCV exam blueprint vs VCP6

Installation and Cluster Deployment Guide for VMware

By the end of the class, attendees will have learned the skills, and best practices of virtualization. Attendees

VMware vsphere with ESX 4.1 and vcenter 4.1

Deep Dive on SimpliVity s OmniStack A Technical Whitepaper

VMWARE VSAN LICENSING GUIDE - MARCH 2018 VMWARE VSAN 6.6. Licensing Guide

Overview. Prerequisites. VMware vsphere 6.5 Optimize, Upgrade, Troubleshoot

Expert Reference Series of White Papers. Five Reasons VMware vsphere 6.0 is a Game Changer

VMware vsphere VMFS First Published On: Last Updated On:

VMware vsphere 6.5 Boot Camp

VMware vsphere 5.5 Professional Bootcamp

The storage challenges of virtualized environments

Executive Summary. The Need for Shared Storage. The Shared Storage Dilemma for the SMB. The SMB Answer - DroboElite. Enhancing your VMware Environment

SQL Saturday Jacksonville Aug 12, 2017

Eliminate the Complexity of Multiple Infrastructure Silos

Modern hyperconverged infrastructure. Karel Rudišar Systems Engineer, Vmware Inc.

Intelligent Rebuilds in vsan 6.6 January 08, 2018

VMware vsphere 5.5 Advanced Administration

Why Datrium DVX is Best for VDI

VMware Virtual SAN Technology

EMC Integrated Infrastructure for VMware. Business Continuity

ECONOMICAL, STORAGE PURPOSE-BUILT FOR THE EMERGING DATA CENTERS. By George Crump

VMware Virtual SAN. High Performance Scalable Storage Architecture VMware Inc. All rights reserved.

STARWIND VIRTUAL SAN: DIFFERENTIATION [FROM COMPETITORS]

Branch offices and SMBs: choosing the right hyperconverged solution

VMware vsan 6.6. Licensing Guide. Revised May 2017

Optimizing the Data Center with an End to End Solutions Approach

Understanding Data Locality in VMware vsan First Published On: Last Updated On:

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme

What's New in vsan 6.2 First Published On: Last Updated On:

VMware vsphere Customized Corporate Agenda

New Features in VMware vsphere (ESX 4)

"Charting the Course... VMware vsphere 6.7 Boot Camp. Course Summary

IBM Spectrum NAS. Easy-to-manage software-defined file storage for the enterprise. Overview. Highlights

DATA PROTECTION IN A ROBO ENVIRONMENT

Storage Strategies for vsphere 5.5 users

Turning Object. Storage into Virtual Machine Storage. White Papers

VI3 to vsphere 4.0 Upgrade and New Technology Ultimate Bootcamp

VMware - VMware vsphere: Install, Configure, Manage [V6.7]

StorMagic SvSAN: A virtual SAN made simple

Evaluating Cloud Storage Strategies. James Bottomley; CTO, Server Virtualization

StarWind Storage Appliance

VMWARE VIRTUAL SAN: ENTERPRISE-GRADE STORAGE FOR HYPER- CONVERGED INFRASTRUCTURES CHRISTOS KARAMANOLIS RAWLINSON RIVERA

Microsoft Virtualization Delivers More Capabilities, Better Value than VMware

vsan Planning and Deployment Update 1 16 OCT 2018 VMware vsphere 6.7 VMware vsan 6.7

vsan Management Cluster First Published On: Last Updated On:

VMware vsphere Stretched Cluster with X-IO Technologies ISE using Active-Active Mirroring. Whitepaper

SYSTEM UPGRADE, INC Making Good Computers Better. System Upgrade Teaches RAID

RAID SEMINAR REPORT /09/2004 Asha.P.M NO: 612 S7 ECE

Microsoft SQL Server 2014 on VMware vsan 6.2 All-Flash October 31, 2017

How to Increase VMware Application Availability with Shared Storage

Administering VMware Virtual SAN. Modified on October 4, 2017 VMware vsphere 6.0 VMware vsan 6.2

VxRail: Level Up with New Capabilities and Powers GLOBAL SPONSORS

1V Number: 1V0-621 Passing Score: 800 Time Limit: 120 min. 1V0-621

How it can help your organisation

VMware vsphere with ESX 6 and vcenter 6

WHITE PAPER PernixData FVP

Delivering unprecedented performance, efficiency and flexibility to modernize your IT

VMWARE EBOOK. Easily Deployed Software-Defined Storage: A Customer Love Story

Virtual SAN and vsphere w/ Operations Management

VMworld 2013 Overview

Take Back Lost Revenue by Activating Virtuozzo Storage Today

The next step in Software-Defined Storage with Virtual SAN

HCI mit VMware vsan Radikal einfach und vollständig in die SDDC Strategie integriert

Vmware VCP410. VMware Certified Professional on vsphere 4. Download Full Version :

Virtualization of the MS Exchange Server Environment

vsphere Availability Update 1 ESXi 5.0 vcenter Server 5.0 EN

Hitachi Adaptable Modular Storage and Hitachi Workgroup Modular Storage

VMware vsphere with ESX 4 and vcenter

Benefits of Multi-Node Scale-out Clusters running NetApp Clustered Data ONTAP. Silverton Consulting, Inc. StorInt Briefing

Next Gen Storage StoreVirtual Alex Wilson Solutions Architect

Hedvig as backup target for Veeam

Administering VMware vsan. 17 APR 2018 VMware vsphere 6.7 VMware vsan 6.7

Hitachi Adaptable Modular Storage and Workgroup Modular Storage

What s New in VMware vsphere 5.1 Platform

Cisco UCS-Mini Solution with StorMagic SvSAN with VMware vsphere 5.5

Business Continuity and Disaster Recovery. Ed Crowley Ch 12

"Software-defined storage Crossing the right bridge"

How to Protect Your Small or Midsized Business with Proven, Simple, and Affordable VMware Virtualization

Logical Operations Certified Virtualization Professional (CVP) VMware vsphere 6.0 Level 2 Exam CVP2-110

Transcription:

Expert Reference Series of White Papers VSAN: Reimagining Storage in vsphere 1-800-COURSES www.globalknowledge.com

VSAN: Reimagining Storage in vsphere John Hales, Global Knowledge VMware Instructor, A+, Network+, CTT+, MCSE, MCDBA, MOUS, MCT, VCA-DCV, VCA-Cloud, VCA-Workforce Mobility, VCP, VCP-DT, VCAP-DCA, VCI, EMCSA Introduction At VMworld in August of 2013, VMware announced VMware Virtual SAN (VSAN). It was in public beta until early March and went GA (General Availability) on March 10th. VSAN is VMware s native version of Software Defined Storage (SDS). It is simple and easy to setup and managed by user-defined policies that are then applied to VMs as needed. It is this policy-based control that makes VSAN so powerful. This white paper will look at VSAN, including what it is, basic requirements, how it works, and how various types of failures are handled. Some common uses will also be discussed. What is VSAN? Vendor lock-in on the storage side is a big problem in many environments today due to the expense of getting the storage area network (SAN) or Network Attached Storage (NAS) array (often in the hundreds of thousands to millions of dollars per array), not counting the expertise required to operate the array, tune it, and do the necessary provisioning, monitoring, and optimization tasks. In addition, with the cost per gigabyte dropping rapidly, purchasing storage well in advance of when it is actually needed is expensive (it would be cheaper if purchased just before it was needed), but that can lead to a death by a thousand cuts syndrome of constantly having to go back to management and ask for another few disks (which are relatively inexpensive), a new shelf (somewhat more expensive), or even an entire new array (a very expensive proposition). Up until now, the cost was deemed unavoidable and worth the cost to ensure high availability, shared access across hosts, low latencies, etc. These features will still probably be required for large, complex companies (and for core data center functions, etc.) for years to come, but in many other cases, they may not be required (see the Use Cases section for some ideas on where this technology may make sense). VSAN is implemented at the kernel level, and thus doesn t suffer from the performance disadvantages of the Virtual Storage Appliance (VSA), which was (and is) implemented as a virtual appliance (VA). While the VSA was designed as a small-to-medium business (SMB) or remote office / branch office (ROBO) solution where a SAN or NAS array was too expensive, VSAN is designed for use in the enterprise in addition to the VSA use cases. Both the VSA and VSAN have the same basic purpose: take local storage located in individual servers and turn it into shared storage that can be used by HA, vmotion, DRS, etc. VSAN is implemented at the cluster level, similarly to HA and DRS today; in fact, it is just another property of a cluster. It can be enabled in just two clicks, though there are many advanced options that can be set, along with storage policies to provide the needed availability to VMs at the best cost and performance (see the Storage Policies section on page 4 for more detail on this). The nice thing about this product is that you can scale up by adding additional storage within an ESXi host (up to 42 disks); you scale out by simply adding another ESXi host into the cluster (up to the vsphere maximum of 32 nodes). Copyright 2014 Global Knowledge Training LLC. All rights reserved. 2

Requirements VSAN has the following requirements: 3 32 hosts per vsphere cluster HA must be enabled for the cluster (DRS often will be as well) 1 SSD and 1 7 magnetic (spinning) disks, which create a disk group 1 GbE minimum, with 10 GbE recommended vsphere and vcenter (5.5 or higher) VSAN license key A few quick notes on disks and disk groups before we move on... First, SSD space is used for caching (both read and write) only, and thus any discussion of usable space ignores all of the SSD space in every host. Second, VMware s best practice is that 10% of the space in each disk group be SSD to ensure there is enough space for caching (it will work if there is less, but performance may be impacted). Third, each host can have zero to five disk groups located on it. Any host with zero disk groups on it can run VMs like any other host, but storage requests will go to the other nodes. Note: Just because a VM runs on a given host, there is no guarantee the storage needed by that VM is local to the same host. While some may question performance and/or scalability, both of the VSAN solution as well as the CPU performance cost, VMware has tested and shown nearly two million input/output operations per second (IOPS) in a single cluster (read only; roughly half that in a mixed read/write scenario), and at that level, only a 10% hit to CPU performance. While the 10% may sound like a lot, most ESXi servers today are running closer to 50% CPU utilization, so the extra 10% hit will not likely affect VM performance. Each cluster supports up to 4.4 petabytes of space as well, allowing for large amounts of data per cluster. Note that this space is given directly to VSAN to use; if any RAID is used (and usually it is not just the raw disks are given to VSAN to use as it sees fit), only RAID 0 is supported. In fact, in many ways, it acts like Storage Spaces in Windows Server 2012 in this regard. How VSAN Works VSAN is designed to perform well and maximized available space. Previously it was mentioned that it is implemented at the vmkernel level to maximize performance from the compute side. This requires SSD drives for caching to maximize performance from the disk perspective, and requires at least 1 GbE (with 10 GbE recommended) on the network side. To maximize space, all vmdk files are thin provisioned and no parity or mirroring RAID is employed on the hosts. Additional capacity can be added by simply adding disks to a host and giving VSAN access to the new disks. Enabling Enabling VSAN is a very simple process. Once VSAN has been properly licensed, simply go to the desired cluster and check the box for VSAN, like you would for DRS or HA. Note that if HA is already enabled, you will need to temporarily disable it so that VSAN can be configured, then you can re-enable it. Once you check the box, the only major question is how VSAN should get the disks it needs to work: automatically or manually? If you choose automatically, VSAN will automatically use all the SSD and magnetic disks that it can find that are not used elsewhere in the system (for example to boot the host) and will create disk groups automatically. If you want more control over which disks are to be used and/or which disks belong in which disk groups, choose manual and configure the disks as desired. Copyright 2014 Global Knowledge Training LLC. All rights reserved. 3

Storage Policies The power of VSAN is not so much that it turns local storage into shared storage, though that is very impressive, but rather that policies can be setup and applied to VMs and that the system will automatically enforce those policies. There are several policies that can be set in VSAN. They include: Failures to Tolerate: Defines the number of copies of the disk(s) that should be created by specifying the number of concurrent failures that need to be withstood (from zero to three). If zero is specified, any failure of a disk or host will cause the VM to be inaccessible until the failure is corrected (and possibly data restored from backup). The number of copies will be one more than the number specified (i.e., one means two copies). This is sometimes known as RAIN (redundant array of independent nodes) and is always a mirroring-style replication (i.e., RAIN 1, even if more than one failure is tolerated). The purpose of this parameter is availability. Stripe Width: Specifies how many physical disks the data for a single vmdk is spread across. This is always done in RAIN 0 fashion (although it can be combined with the Failures to Tolerate to create a RAIN 0+1 design if desired). This setting is most important when greater performance is necessary and the read caching provided by the SSD drive is insufficient, requiring data to be read from the spinning disk(s), instead of the SSD read cache. If a value greater than one is specified, the data is split into 1 MB chunks across the disks. Note that enabling this feature will consume additional system resources. Object Space Reservation: Reserves the specified percentage of the disk space for the VM (assuming it is thin provisioned); it does not thick-provision the space, but rather uses the calculated space as if it were already provisioned in much the same way that a CPU or memory reservation is used in compute calculations used by vsphere, HA, and DRS. If a vmdk has been configured as either eager or lazy zero thick, this parameter is ignored (i.e., it is already at 100%). By default, a single policy is created and used by everything that uses VSAN, and that policy is not visible in the Web Client. It is simply configured to tolerate the loss of a host, disk, or disk group by setting Failures to Tolerate to one. Note: Don t confuse Storage Policies (used by VSAN only) with Storage Profiles (usable with any datastore type). Storage policies determine the performance and availability of a VM located on a VSAN datastore and are fully automatic once assigned, while Storage Profiles define a preferred storage type (typically based on the speed of the underlying disks) for a VM. The profiles are manually created and manually assigned and do not automatically move VMs to other disks if the profile-configured type is not the actual location of the VM (Storage vmotion would typically be manually invoked to fix the issue). What VSAN Does to VMDKs VSAN will look at the storage policy assigned to each VM and then automatically apply it, placing each.vmdk file on disks it chooses (or the other VM files, collectively known as VM Home In this white paper, the term vmdk has been used throughout for brevity, but it applies equally to the VM Home folders) For example, if you asked to tolerate one failure, VSAN would create two copies of the disk, each on a separate host. While you can determine where individual pieces of a VM are stored via the Web Client, the beauty of VSAN (and SDS in general) is that it really doesn t matter the policy is what matters and the system will automatically enforce the policy (assuming enough hosts and capacity is available). Copyright 2014 Global Knowledge Training LLC. All rights reserved. 4

How Failures are Handled You may be wondering how failures are handled. VSAN is very fault tolerant and can continue to operate in the event of a disk, network, or server failure. The ability to handle failures depends on the storage policies previously described. No data loss will occur in any case as writes are not acknowledged to the VM until all copies on all hosts have acknowledged the write as complete. Let s look at each of the scenarios and review how VSAN responds. Disk Failure In traditional environments, RAID solves the problem and thus the loss of a disk is transparent to vsphere; simply replace the disk and the RAID controller rebuilds the LUN. The VSA works this way as well, but due to this need for redundancy at both the LUN and server levels, if RAID 10 is used (and often it is for performance reasons), only 25% of the space purchased is actually usable (half lost to RAID 10 and half to RAIN 1). VSAN cuts that in half due the use of individual disks (or at worst RAID 0, which is the same from a redundancy as well as disk overhead perspective). With VSAN, your data is replicated by policy to multiple locations (unless you choose to have no redundancy for a non-critical VM), and thus when a disk is lost, VSAN will see that it is not in compliance with the defined storage policy and immediately begin copying the data to another disk to automatically come into compliance again. When speaking of disk loss, it is important to note that the loss of a SSD will cause the disk group to go offline and copying of the data to a new location to automatically begin. Note that no administrator intervention is required in this process at all (except for physically replacing the failed disk of course). This is probably the most common scenario (from a failure perspective the planned loss of a server [temporarily] for maintenance, patching, etc. will probably be the most common scenario in most environments). Network Failure In the event of complete network failure, only VMs that are running on local storage (relative to where the VM is located) will continue to run; HA can attempt to restart the VMs on other nodes if capacity exists where the disk files are located, assuming that HA still has a valid network path (unless a storage policy that striped a single VMDK across multiple nodes was defined). This is why redundant network paths (ideally to redundant switches) are always recommended. A few notes are important when using VSAN in a HA-enabled cluster. First, to handle network partition scenarios, a witness is assigned on an additional host in the cluster to make an odd number of nodes so that a majority (a quorum) could be on line if a partition were to occur. In the event of a network partition, VSAN will always restart any affected VMs in the partition that has quorum. Second, the normal heartbeat communication that takes place across the management network is changed to the VSAN network instead (except for checking for host node isolation, which will still use the default gateway (by default) of the management network). Third, datastore heartbeats are disabled if the cluster only has VSAN datastores as there is no additional availability gained. The failure of the switch is a rare event, however, and thus the foregoing is unlikely to occur, while the loss of a single network port, cable, or NIC is far more probable. In those cases, HA will simply restart any affected VMs on nodes that still have access to storage (locally or across the network) and the VM will be back online quickly. This is also a fairly rare event in most environments. Server Failure The final failure scenario is the failure of a server. When the server fails, HA will restart the VMs elsewhere in the cluster like normal. VSAN does not start rebuilding the lost data on another server right away however, unlike in the disk failure scenario. The reason for this is that the server may be rebooting after a patch for example and Copyright 2014 Global Knowledge Training LLC. All rights reserved. 5

will be back online soon. To prevent lots of extra space being used unnecessarily in situations like these, when the server goes down, a 60-minute timer begins; if the server is back online within that time frame, VSAN will simply synchronize the copies of the data. On the other hand, after the timer runs out, VSAN will automatically create a new copy of the missing data, much like in the disk failure scenario previously described. This brings up an interesting question: What happens if the host is placed in maintenance mode? As with most things in IT, the answer is: It depends. When a host is placed in maintenance mode, there are three things VSAN can do with the data on that host: 1. Full Data Migration: Migrate all data off the host to other nodes in the cluster. This could take the longest to complete and probably would not be used for short term outages, such as patching, but rather for longer term changes, most likely when the node is being removed from the cluster. 2. Ensure Accessibility: Migrate data to other hosts in the cluster if needed to ensure that at least one copy of the data remains accessible. In this case, most data will probably not be migrated, making it faster to get the node in maintenance mode, but accessibility policies may be violated. This is the default and most often used option in most environments. 3. No Data Migration: No data is migrated, even if it is the only copy available; in that case the VM would be unavailable under the node came back online. This method offers the fastest entry into maintenance mode, but again, VM availability may be impacted. Use Cases Common use cases include: VDI, where storage can easily be 40 60% of the cost of implementation ROBO sites, where a SAN or NAS device is not feasible and/or cost justified Tier 2/3 storage needs Disaster recovery (DR), usually at the backup site, at least initially DMZ servers, especially where an air gap is required between internal and DMZ storage Management clusters, which need the usual SAN capabilities, but often don t justify the costs of those capabilities These use cases will be used to prove the capability and resilience of VSAN and as it becomes well tested and proven will probably move into the mainstream for storage needs in vsphere. Conclusion VSAN will fundamentally reshape the role of storage in many organizations over the coming years. In much the same way that virtualization was looked at somewhat skeptically in the early years and is now considered standard practice for most workloads in all environments (including mission critical production) so too storage and network virtualization are in the early phases of adoption, but will soon become mainstream for many use cases. While VMware s stated goal is to coexist with existing SAN and NAS environments, it will likely replace many of them in the coming years, being much less costly, much simpler, and far easier to manage via policies. Copyright 2014 Global Knowledge Training LLC. All rights reserved. 6

Learn More Learn more about how you can improve productivity, enhance efficiency, and sharpen your competitive edge through training. VMware training courses Visit www.globalknowledge.com or call 1-800-COURSES (1-800-268-7737) to speak with a Global Knowledge training advisor. About the Author John Hales (A+, Network+, CTT+, MCSE, MCDBA, MOUS, MCT, VCA-DCV, VCA-Cloud, VCA-Workforce Mobility, VCP, VCP-DT, VCAP-DCA, VCI, EMCSA) is a VMware instructor at Global Knowledge, teaching all of the vsphere and View classes that Global Knowledge offers. John has written a book called Administering vsphere 5: Planning, Implementing, and Troubleshooting published by Cengage, as well as other technical books from exam-preparation books to quick-reference guides, as well as custom courseware for individual customers. John lives with his wife and children in Sunrise, Florida. Copyright 2014 Global Knowledge Training LLC. All rights reserved. 7