W H I T E P A P E R. Comparison of Storage Protocol Performance in VMware vsphere 4

Similar documents
Comparison of Storage Protocol Performance ESX Server 3.5

Managing Performance Variance of Applications Using Storage I/O Control

Performance Report: Multiprotocol Performance Test of VMware ESX 3.5 on NetApp Storage Systems

W H I T E P A P E R. What s New in VMware vsphere 4: Performance Enhancements

Configuration Maximums VMware vsphere 5.0

Configuration Maximums VMware Infrastructure 3: ESX Server 3.5 Update 2, ESX Server 3i version 3.5 Update 2, VirtualCenter 2.

What s New in VMware vsphere 4.1 Performance. VMware vsphere 4.1

PERFORMANCE STUDY OCTOBER 2017 ORACLE MONSTER VIRTUAL MACHINE PERFORMANCE. VMware vsphere 6.5

Maintaining End-to-End Service Levels for VMware Virtual Machines Using VMware DRS and EMC Navisphere QoS

Performance Implications of Storage I/O Control Enabled NFS Datastores in VMware vsphere 5.0

Microsoft Exchange Server 2010 Performance on VMware vsphere 5

Recommendations for Aligning VMFS Partitions

NAS for Server Virtualization Dennis Chapman Senior Technical Director NetApp

Performance of Virtual Desktops in a VMware Infrastructure 3 Environment VMware ESX 3.5 Update 2

Emulex LPe16000B 16Gb Fibre Channel HBA Evaluation

Configuring iscsi in a VMware ESX Server 3 Environment B E S T P R A C T I C E S

EMC Backup and Recovery for Microsoft SQL Server

VMware VMFS Volume Management VMware Infrastructure 3

Maintaining End-to-End Service Levels for VMware Virtual Machines Using VMware DRS and EMC Navisphere QoS

CLOUD PROVIDER POD RELEASE NOTES

Eliminate the Complexity of Multiple Infrastructure Silos

EMC Virtual Infrastructure for Microsoft Exchange 2007

VMware VMmark V1.1 Results

VMware VMmark V1.1 Results

Setup for Microsoft Cluster Service Update 1 Release for ESX Server 3.5, ESX Server 3i version 3.5, VirtualCenter 2.5

10Gb iscsi Initiators

8Gb Fibre Channel Adapter of Choice in Microsoft Hyper-V Environments

SAN Virtuosity Fibre Channel over Ethernet

VMware VMmark V1.1 Results

CLOUD PROVIDER POD RELEASE NOTES

Configuration Maximums

VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014

CLOUD PROVIDER POD. for VMware. Release Notes. VMware Cloud Provider Pod January 2019 Check for additions and updates to these release notes

How to Use a Tomcat Stack on vcloud to Develop Optimized Web Applications. A VMware Cloud Evaluation Reference Document

VMware VMmark V1.1 Results

PAC094 Performance Tips for New Features in Workstation 5. Anne Holler Irfan Ahmad Aravind Pavuluri

VMware VMmark V1.1 Results

VMware vstorage APIs FOR ARRAY INTEGRATION WITH EMC VNX SERIES FOR SAN

Fujitsu PRIMEFLEX for VMware vsan 20,000 User Mailbox Exchange 2016 Mailbox Resiliency Storage Solution

PERFORMANCE CHARACTERIZATION OF MICROSOFT SQL SERVER USING VMWARE CLOUD ON AWS PERFORMANCE STUDY JULY 2018

PRESENTATION TITLE GOES HERE

Virtualizing SQL Server 2008 Using EMC VNX Series and VMware vsphere 4.1. Reference Architecture

Configuration Maximums. Update 1 VMware vsphere 6.5 VMware ESXi 6.5 vcenter Server 6.5

EMC Business Continuity for Microsoft Applications

ARISTA: Improving Application Performance While Reducing Complexity

VMware VMmark V1.1 Results

Surveillance Dell EMC Storage with Cisco Video Surveillance Manager

vstart 50 VMware vsphere Solution Specification

A Performance Characterization of Microsoft SQL Server 2005 Virtual Machines on Dell PowerEdge Servers Running VMware ESX Server 3.

Microsoft SQL Server in a VMware Environment on Dell PowerEdge R810 Servers and Dell EqualLogic Storage

Getting Started with ESX

Mobile Secure Desktop Implementation with Pivot3 HOW-TO GUIDE

WHITE PAPER. Optimizing Virtual Platform Disk Performance

IBM Emulex 16Gb Fibre Channel HBA Evaluation

IOmark- VDI. IBM IBM FlashSystem V9000 Test Report: VDI a Test Report Date: 5, December

Getting Started with ESX

Reference Architecture

Storage Compatibility Guide for ESX Server 3.0

针对 Windows 7 的 服务器和存储大小规划指南 技 术 说 明

VMware VMmark V1.1 Results

Setup for Failover Clustering and Microsoft Cluster Service

Dell Reference Configuration for Large Oracle Database Deployments on Dell EqualLogic Storage

The QLogic 8200 Series is the Adapter of Choice for Converged Data Centers

Setup for Failover Clustering and Microsoft Cluster Service

VMware VMmark V1.1 Results

Deploying Microsoft Exchange Server 2007 mailbox roles on VMware Infrastructure 3 using HP ProLiant servers and HP StorageWorks

vcloud Automation Center Reference Architecture vcloud Automation Center 5.2

iscsi Design Considerations and Deployment Guide VMware Infrastructure 3

Active System Manager Release 8.2 Compatibility Matrix

Storage Considerations for VMware vcloud Director. VMware vcloud Director Version 1.0

AMD: WebBench Virtualization Performance Study

VMware VMmark V1.1 Results

Performance Scaling. When deciding how to implement a virtualized environment. with Dell PowerEdge 2950 Servers and VMware Virtual Infrastructure 3

VMware VMmark V1.1.1 Results

Use of the Internet SCSI (iscsi) protocol

VMware VMmark V1.1 Results

EMC Unified Storage for Oracle Database 11g/10g Virtualized Solution. Enabled by EMC Celerra and Linux using FCP and NFS. Reference Architecture

Creating the Fastest Possible Backups Using VMware Consolidated Backup. A Design Blueprint

Version 1.0 October Reduce CPU Utilization by 10GbE CNA with Hardware iscsi Offload

Configuring and Managing Virtual Storage

EMC Performance Optimization for VMware Enabled by EMC PowerPath/VE

Number of Hosts: 2 Uniform Hosts [yes/no]: yes Total sockets/cores/threads in test: 4/64/64

JMR ELECTRONICS INC. WHITE PAPER

VMware vsphere Storage Appliance Installation and Configuration

EMC Virtual Infrastructure for Microsoft Exchange 2010 Enabled by EMC Symmetrix VMAX, VMware vsphere 4, and Replication Manager

IOmark- VM. HP MSA P2000 Test Report: VM a Test Report Date: 4, March

VMware VMmark V1.1 Results

Vendor and Hardware Platform: HP ProLiant BL620c G7 Virtualization Platform: VMware ESX 4.1 U1 Build VMware vcenter Server 4.

BlackBerry Enterprise Server for Microsoft Exchange Version: 5.0. Performance Benchmarking Guide

BlackBerry AtHoc Networked Crisis Communication Capacity Planning Guidelines. AtHoc SMS Codes

VMware Infrastructure Update 1 for Dell PowerEdge Systems. Deployment Guide. support.dell.com

VMware VMmark V1.1.1 Results

Webinar Series: Triangulate your Storage Architecture with SvSAN Caching. Luke Pruen Technical Services Director

Notes Section Notes for Workload. Configuration Section Configuration

Tested By: Hewlett-Packard Test Date: Configuration Section Configuration

VMware Infrastructure 3 Primer Update 2 and later for ESX Server 3.5, ESX Server 3i version 3.5, VirtualCenter 2.5

Getting Started with ESXi Embedded

Setup for Failover Clustering and Microsoft Cluster Service. Update 1 16 OCT 2018 VMware vsphere 6.7 VMware ESXi 6.7 vcenter Server 6.

IOmark- VM. HP HP ConvergedSystem 242- HC StoreVirtual Test Report: VM- HC b Test Report Date: 27, April

Q500-P20 performance report

Transcription:

W H I T E P A P E R Comparison of Storage Protocol Performance in VMware vsphere 4

Table of Contents Introduction................................................................... 3 Executive Summary............................................................ 3 Experimental Setup............................................................ 3 I/O Workload.................................................................. 4 Experiment: Single VM: Throughput (Read and Write)............................ 5 Experiment: Single VM: CPU Cost Per I/O (Read).................................. 6 Experiment: Multiple VMs: Aggregate Throughput (Read)........................ 7 Conclusion.................................................................... 7 2

Introduction This paper compares the performance of various storage protocols available on VMware vsphere 4. The protocols Fibre Channel, Hardware iscsi, Software iscsi, and NFS are tested using virtual machines on an ESX 4.0 host. Iometer is used to generate the I/O workload. The Fibre Channel experiments were conducted over a 4Gb Fibre Channel network. The Hardware iscsi, Software iscsi, and NFS experiments were conducted over a Gigabit Ethernet connection. Experiments over 10Gb Ethernet and 8Gb Fibre Channel will be included in a future update to this paper. This paper will start by describing key aspects of the test environment: ESX host, storage array, virtual machines, and Iometer workload. Next, performance results for throughput and CPU cost are presented from experiments involving one or more virtual machines. Finally, the key findings of the experiments are summarized. The terms storage server and storage array will be used interchangeably in this paper. Executive Summary The experiments in this paper show that each of the four storage protocols (Fibre Channel, Hardware iscsi, Software iscsi, and NFS) can achieve line-rate throughput for both single virtual machine and multiple virtual machines on an ESX host. These experiments also show that Fibre Channel and Hardware iscsi have substantially lower CPU cost than Software iscsi and NFS. Experimental Setup The table below shows key aspects of the test environment for the ESX host, the storage array, and the virtual machine. ESX Host Component Details Hypervisor VMware ESX 4.0 Processors Four Intel Xeon E7340 Quad-Core 2.4GHz processors Memory 32GB Fibre Channel HBA QLogic QLA2432 4Gb Fibre Channel network 4Gb FC switch NIC for NFS and SW iscsi 1Gb (Intel 82571EB) MTU for NFS, SW iscsi, HW iscsi 1500 bytes iscsi HBA QLogic QL4062c 1Gb (Firmware: 3.0.1.49) IP network for NFS and SW/HW iscsi 1Gb Ethernet with dedicated switch and VLAN (Extreme Summit 400-48t) File system for NFS Native file system on NFS server File system for FC and SW/HW iscsi None (RDM-physical was used) Storage Array Component Details Storage server One server supporting FC, iscsi, and NFS Disk Drives: Number per data LUN 9 Disk Drives: Size 300Gb Disk Drives: Speed 15K RPM Disk Drives: Type Fibre Channel 3

Virtual Machine Component Details Guest OS Windows Server 2008 Enterprise SP1 Virtual processors 1 Memory 512MB Virtual disk for data 100MB Mapped Raw LUN (RDM-physical) File system None (Physical drives were used) SCSI controller LSI Logic Parallel VMware s VMFS file system is recommended for production deployments of virtual machines on iscsi and Fibre Channel arrays. Because NFS storage presents files and not blocks, VMFS is not needed or possible. VMFS was therefore not used in the Fibre Channel and iscsi experiments to attempt to produce results that could be compared across all protocols. I/O Workload Iometer (http://sourceforge.net/projects/iometer) was used to generate the I/O workload for these experiments. Iometer is a free storage performance testing tool that can be configured to measure throughput and latency under a wide variety of access profiles. Iometer Workload Component Details Number of outstanding I/Os 16 Run time 2 min Ramp-up time 2 min Number of workers 1 (per VM) Each virtual (data) disk of the virtual machines used in these experiments is 100MB in size. The small size of these virtual disks ensures that the I/O working set will fit into the cache of the storage array. An experiment with a working set size that fits into the cache of the storage array is commonly referred to as a cached run. For read operations in a cached run experiment, the data is served from the storage array s cache, and read performance is independent of disk latencies. For write operations in a cached run experiment, the rate of write requests at the storage array may exceed the storage array s rate of writing the dirty blocks from the write cache to disk. If this happens, the write cache will eventually fill up. Once the write cache is full, write performance is limited by the rate at which dirty blocks in the write cache are written to disk. This rate is limited by the latency of the disks in the storage array, the RAID configuration, and the number of disk spindles used for the LUN. For these reasons, read performance for cached runs is a better indication of the true performance of a storage protocol on the ESX host, irrespective of the storage array used. 4

Experiment: Single VM: Throughput (Read and Write) Figure 1 shows the sequential read throughput (in MB/sec) of running a single virtual machine in the standard workload configuration for different I/O block sizes, for each of the storage protocols. Figure 1: Read throughput for different I/O block sizes 450 400 Throughput (MB/sec) 350 300 250 200 150 100 50 0 1KB 4KB 8KB 16KB 32KB 64KB 128KB 256KB 512KB I/O Block Sizes NFS SW iscsi HW iscsi FC For Fibre Channel, read throughput is limited by the bandwidth of the 4Gb Fibre Channel link for I/O sizes at or above 64KB. For IP-based protocols, read throughput is limited by the bandwidth of the 1Gb Ethernet link for I/O sizes at or above 32KB. Figure 2 shows the sequential write throughput (in MB/sec) of running a single virtual machine in the standard workload configuration for different I/O block sizes, for each of the storage protocols. Figure 2: Write throughput for different I/O block sizes 450 400 Throughput (MB/sec) 350 300 250 200 150 100 50 0 1KB 4KB 8KB 16KB 32KB 64KB 128KB 256KB 512KB I/O Block Sizes NFS SW iscsi HW iscsi FC 5

For Fibre Channel, the maximum write throughput for any I/O block size is consistently lower than read throughput of the same I/O block size. This is the result of disk write bandwidth limitations on the storage array. For the IP-based protocols, write throughput for block sizes at or above 16KB is limited by the bandwidth of the 1Gb Ethernet link. To summarize, a single Iometer thread running in a virtual machine can saturate the bandwidth of the respective networks for all four storage protocols, for both read and write. Fibre Channel throughput performance is higher because of the higher bandwidth of the Fibre Channel link. For the IP-based protocols, there is no significant throughput difference for most block sizes. Experiment: Single VM: CPU Cost Per I/O (Read) CPU cost is a measure of the amount of CPU resources used by ESX to perform a given amount of I/O. In this paper, the CPU cost of each storage protocol is measured in units of CPU cycles per I/O operation. The cost for different storage protocols is normalized with respect to the cost of software iscsi on ESX 3.5. Figure 3 shows the relative CPU cost of sequential reads in a single virtual machine in the standard workload configuration for a block size of 64 KB for each of the storage protocols. Results on ESX 4.0 are shown next to ESX 3.5 to highlight efficiency improvements on all protocols. The CPU cost of write operations for different storage protocols was not compared as write performance is strongly dependent on the choice of the storage array. Figure 3: Relative CPU cost of 64 KB sequential reads in a single virtual machine 1.2 Relative CPU Cost Per I/O 1.0 0.8 0.6 0.4 0.2 0 NFS SW iscsi HW iscsi FC ESX 3.5 ESX 4.0 For Fibre Channel and Hardware iscsi, a major part of the protocol processing is offloaded to the HBA, and consequently the cost of each I/O is very low. For Software iscsi and NFS, host CPUs are used for protocol processing which increases cost. Furthermore, the cost of NFS and Software iscsi is higher with larger block sizes, such as 64 KB. This is due to the additional CPU cycles needed for each block for check summing, blocking, etc. Software iscsi and NFS are more efficient at smaller blocks and are both capable of delivering high throughput performance when CPU resource is not a bottleneck, as will be shown in the next section. The cost per I/O is dependent on a variety of test parameters, such as platform architecture, block size, and other factors. However, these tests demonstrate improved efficiency in vsphere 4 s storage stack over the previous version. 6

Experiment: Multiple VMs: Aggregate Throughput (Read) Figure 4 shows the aggregate sequential read throughput (in MB/sec) of running 2, 4, 8, 16, and 32 virtual machines in the standard workload configuration for a block size of 64KB. Each virtual machine performs I/O to its dedicated 100MB LUN. Figure 4: Throughput of running multiple VMs in the standard workload configuration 450 400 Throughput (MB/sec) 350 300 250 200 150 100 50 0 2 4 8 16 32 Number of VMs NFS SW iscsi HW iscsi FC For each of the storage protocols, the maximum aggregate throughput is limited by the network bandwidth. There is no degradation in aggregate throughput for a large number of virtual machines. Conclusion In this paper, the performance of four storage protocols was compared for accessing shared storage available on VMware ESX 4.0: Fibre Channel, Hardware iscsi, Software iscsi, and NFS. All four storage protocols for shared storage on ESX are shown to be capable of achieving throughput levels that are only limited by the capabilities of the storage array and the connection between it and the ESX server. ESX shows excellent scalability by maintaining these performance levels in cases of heavy consolidation. For CPU cost, Fibre Channel and Hardware iscsi are more efficient than Software iscsi and NFS. However, when CPU resources are not a bottleneck, Software iscsi and NFS can also be part of a high-performance solution. Earlier versions of ESX were also able to achieve throughput levels that are only limited by the array and bandwidth to it. VMware vsphere 4 continues to support this maximized throughput but can do so with greater efficiency. Improved efficiency on vsphere 4 means the same high levels of performance with more virtual machines. 7

VMware, Inc. 3401 Hillview Ave Palo Alto CA 94304 USA Tel 877-486-9273 Fax 650-427-5001 www.vmware.com Copyright 2009 VMware, Inc. All rights reserved. This product is protected by U.S. and international copyright and intellectual property laws. VMware products are covered by one or more patents listed at http://www.vmware.com/go/patents. VMware is a registered trademark or trademark of VMware, Inc. in the United States and/or other jurisdictions. All other marks and names mentioned herein may be trademarks of their respective companies. VMW_09Q2_WP_VSPHERE_StorageProtocols_P8_R1