Dell EMC Ready Architectures for VDI

Similar documents
Dell EMC Ready Architectures for VDI

Dell EMC Ready Architectures for VDI

Dell EMC Ready Architectures for VDI

Dell EMC Ready Architectures for VDI

Dell EMC XC Series Appliances A Winning VDI Solution with Scalable Infrastructure

Dell EMC vsan Ready Nodes for VDI

VMWare Horizon View 6 VDI Scalability Testing on Cisco 240c M4 HyperFlex Cluster System

Dell EMC Ready Architectures for VDI

Virtualized SQL Server Performance and Scaling on Dell EMC XC Series Web-Scale Hyper-converged Appliances Powered by Nutanix Software

Citrix VDI Scalability Testing on Cisco UCS B200 M3 server with Storage Accelerator

Scalability Testing with Login VSI v16.2. White Paper Parallels Remote Application Server 2018

Dell EMC Ready System for VDI on VxRail

TPC-E testing of Microsoft SQL Server 2016 on Dell EMC PowerEdge R830 Server and Dell EMC SC9000 Storage

Upgrade to Microsoft SQL Server 2016 with Dell EMC Infrastructure

Dell Precision Appliance for Wyse - VMware Horizon

DELL EMC READY BUNDLE FOR VIRTUALIZATION WITH VMWARE AND FIBRE CHANNEL INFRASTRUCTURE

Microsoft SQL Server in a VMware Environment on Dell PowerEdge R810 Servers and Dell EqualLogic Storage

Dell EMC Ready System for VDI on XC Series

The scalability and economics of delivering Citrix Virtual App and Desktop services from Microsoft Azure

AccelStor NeoSapphire VDI Reference Architecture for 500 users with VMware Horizon 7

Dell Precision Appliance for Wyse - Citrix XenDesktop

Reference Architecture: Lenovo Client Virtualization with VMware Horizon and System x Servers

Scale out a 13th Generation XC Series Cluster Using 14th Generation XC Series Appliance

Parallels Remote Application Server. Scalability Testing with Login VSI

DELL EMC READY BUNDLE FOR VIRTUALIZATION WITH VMWARE AND ISCSI INFRASTRUCTURE

Surveillance Dell EMC Storage with Bosch Video Recording Manager

Adobe Acrobat Connect Pro 7.5 and VMware ESX Server

Implementing SQL Server 2016 with Microsoft Storage Spaces Direct on Dell EMC PowerEdge R730xd

VMWARE HORIZON 6 ON HYPER-CONVERGED INFRASTRUCTURES. Horizon 6 version 6.2 VMware vsphere 6U1 / VMware Virtual SAN 6U1 Supermicro TwinPro 2 4 Nodes

Dell EMC. Vblock System 340 with VMware Horizon 6.0 with View

Reference Architecture - Microsoft SharePoint Server 2013 on Dell PowerEdge R630

Dell XC Web-Scale Converged Appliance for VMware Horizon

2000 Persistent VMware View VDI Users on Dell EMC SCv3020 Storage

Windows Server 2016 Impact on VDI: Benchmark Results. By Mark Plettenberg, Ryan Bijkerk and Omar Bouhaj

XenApp and XenDesktop 7.12 on vsan 6.5 All-Flash January 08, 2018

IBM Emulex 16Gb Fibre Channel HBA Evaluation

Performance Lab Report & Architecture Overview Summary of SnapVDI Features and Performance Testing Using Login VSI

Accelerating Microsoft SQL Server 2016 Performance With Dell EMC PowerEdge R740

Login VSI Multimedia Workload

5,000 Persistent VMware View VDI Users on Dell EMC SC9000 Storage

NVIDIA GRID APPLICATION SIZING FOR AUTODESK REVIT 2016

Evaluation Report: HP StoreFabric SN1000E 16Gb Fibre Channel HBA

Nutanix Complete Cluster Reference Architecture for Virtual Desktop Infrastructure

INTEGRATED INFRASTRUCTURE FOR VIRTUAL DESKTOPS ENABLED BY EMC VNXE3300, VMWARE VSPHERE 4.1, AND VMWARE VIEW 4.5

Pivot3 Acuity with Microsoft SQL Server Reference Architecture

Dell EMC PowerEdge R740xd as a Dedicated Milestone Server, Using Nvidia GPU Hardware Acceleration

Reference Architecture: Lenovo Client Virtualization with Citrix XenDesktop and System x Servers

Dell EMC Microsoft Storage Spaces Direct Ready Nodes for VDI

Dell Appliance for Wyse - vworkspace

EMC VSPEX END-USER COMPUTING

Surveillance Dell EMC Storage with Cisco Video Surveillance Manager

All rights reserved. This product is protected by U.S. and international copyright and intellectual property laws.

EVALUATING WINDOWS 10 LEARN WHY YOUR USERS NEED GPU ACCELERATION

Citrix XenDesktop 5.5 on VMware 5 with Hitachi Virtual Storage Platform

Microsoft SharePoint Server 2010 Implementation on Dell Active System 800v

Virtual Desktop Infrastructure with Dell Fluid Cache for SAN

Dell EMC. VxRack System FLEX Architecture Overview

REAL PERFORMANCE RESULTS WITH VMWARE HORIZON AND VIEWPLANNER

Dell Appliance for Wyse - vworkspace

Consolidating Microsoft SQL Server databases on PowerEdge R930 server

NVIDIA GRID. Ralph Stocker, GRID Sales Specialist, Central Europe

Citrix XenDesktop with Provisioning Server for VDI on Dell Compellent SC8000 All Flash Arrays for 3,000 Users

Image Management for View Desktops using Mirage

EVALUATING WINDOWS 10: LEARN WHY YOUR USERS NEED GPU ACCELERATION

Dell EMC Ready Architectures for VDI

Surveillance Dell EMC Storage with FLIR Latitude

VMware Horizon View. VMware Horizon View with Tintri VMstore. TECHNICAL SOLUTION OVERVIEW, Revision 1.1, January 2013

iocontrol Reference Architecture for VMware Horizon View 1 W W W. F U S I O N I O. C O M

EMC VSPEX END-USER COMPUTING

DELL EMC READY BUNDLE FOR VIRTUALIZATION WITH VMWARE VSAN INFRASTRUCTURE

SolidFire Reference Architecture for VMware Horizon View

Dell EMC Storage with Panasonic Video Insight

The Dell Precision T3620 tower as a Smart Client leveraging GPU hardware acceleration

EMC XTREMCACHE ACCELERATES VIRTUALIZED ORACLE

Nutanix Reference Architecture Version 1.1 July 2016 RA-2022

VIRTUAL APPLIANCES. Frequently Asked Questions (FAQ)

Surveillance Dell EMC Storage with Milestone XProtect Corporate

What s New in VMware vsphere 4.1 Performance. VMware vsphere 4.1

DELL EMC VMAX UNISPHERE 360

VMware vrealize Operations for Horizon Administration. 20 SEP 2018 VMware vrealize Operations for Horizon 6.6

Dell EMC Best Practices for Running VMware ESXi 6.5 or Later Clusters on XC Series Appliances and XC Core Systems

Horizon Console Administration. 13 DEC 2018 VMware Horizon 7 7.7

Dell EMC. VxBlock Systems for VMware NSX 6.3 Architecture Overview

Reference Architecture: Lenovo Client Virtualization with Citrix XenDesktop and ThinkSystem Servers

White Paper. The impact of virtualization security on your VDI environment

Citrix XenApp / Microsoft RDSH. How to get the Best User Experience and Performance with NVIDIA vgpu Technology

VMWare Horizon View Solution Guide

LATEST INTEL TECHNOLOGIES POWER NEW PERFORMANCE LEVELS ON VMWARE VSAN

VMware vrealize Operations for Horizon Administration

Microsoft SharePoint Server 2010 on Dell Systems

Technical Overview. Jack Smith Sr. Solutions Architect

Cisco UCS C240 M4 Rack Server with VMware Virtual SAN 6.0 and Horizon 6

VDI with VMware Horizon View and VMware vsphere on Dell Storage PS4210XS Hybrid Storage Arrays

High-Performance, High-Density VDI with Pivot3 Acuity and VMware Horizon 7. Reference Architecture

NLVMUG 16 maart Display protocols in Horizon

Virtual Security Server

High-End Graphics Everywhere! UCS C240 and NVIDIA Grid

Surveillance Dell EMC Storage with Digifort Enterprise

The impact of virtualization security on your VDI environment

SQL Server 2016 on Dell EMC XC Series Reference Architecture

Transcription:

Dell EMC Ready Architectures for VDI Designs for VMware Horizon on Dell EMC XC Family September 2018 H17386 Validation Guide Abstract This validation guide describes the architecture and performance of the integration of VMware Horizon components for virtual desktop infrastructure (VDI) on Dell EMC XC Family devices. Dell EMC Solutions

Copyright 2018 Dell Inc. or its subsidiaries. All rights reserved. Published September 2018 Dell believes the information in this publication is accurate as of its publication date. The information is subject to change without notice. THE INFORMATION IN THIS PUBLICATION IS PROVIDED AS-IS. DELL MAKES NO REPRESENTATIONS OR WARRANTIES OF ANY KIND WITH RESPECT TO THE INFORMATION IN THIS PUBLICATION, AND SPECIFICALLY DISCLAIMS IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. USE, COPYING, AND DISTRIBUTION OF ANY DELL SOFTWARE DESCRIBED IN THIS PUBLICATION REQUIRES AN APPLICABLE SOFTWARE LICENSE. Dell, EMC, and other trademarks are trademarks of Dell Inc. or its subsidiaries. Other trademarks may be the property of their respective owners. Published in the USA. Dell EMC Hopkinton, Massachusetts 01748-9103 1-508-435-1000 In North America 1-866-464-7381 www.dellemc.com 2 Dell EMC Ready Architectures for VDI

CONTENTS Chapter 1 Executive Summary 5 Document purpose... 6 Audience... 6 We value your feedback... 6 Chapter 2 Test Environment Configuration and Best Practices 7 Validated hardware resources... 8 Enterprise hardware... 8 Graphics hardware...8 Network hardware...8 Validated software resources... 9 Validated system versions... 9 Virtual networking configuration... 9 Management server infrastructure...10 NVIDIA GRID License Server...10 SQL Server databases... 11 DNS... 11 High availability...11 VMware Horizon architecture... 12 Chapter 3 Solution Performance and Testing 13 Testing process...14 Resource monitoring... 14 Load generation... 15 Profiles and workloads... 15 Linked vs. Instant Clones...16 Desktop VM test configurations... 17 Test results and analysis... 18 B5 configuration...19 C7 configuration... 26 Chapter 4 Conclusion 41 Density recommendations... 42 Summary... 42 Chapter 5 References 43 Dell EMC documentation... 44 VMware documentation... 44 NVIDIA documentation... 44 Dell EMC Ready Architectures for VDI 3

CONTENTS 4 Dell EMC Ready Architectures for VDI

CHAPTER 1 Executive Summary This chapter presents the following topics: Document purpose...6 Audience... 6 We value your feedback... 6 Executive Summary 5

Executive Summary Document purpose Audience We value your feedback This validation guide details the architecture, components, testing methods, and test results for Dell EMC XC Family devices with VMware Horizon 7. It includes the test environment configuration and best practices for systems that have undergone testing. This guide is intended for architects, developers, and technical administrators of IT environments. It provides an in-depth explanation of the testing methodology and basis for VDI densities. It also validates the value of the Dell EMC Ready Architectures for VDI that deliver Microsoft Windows virtual desktops to users of VMware Horizon 7 VDI components on XC Family devices. Dell EMC and the authors of this document welcome your feedback on the solution and the solution documentation. Contact Dell EMC Solutions team with your comments. Authors: Dell EMC Ready Architectures for VDI Engineering Team, Donna Renfro 6 Dell EMC Ready Architectures for VDI

CHAPTER 2 Test Environment Configuration and Best Practices This chapter presents the following topics: Validated hardware resources... 8 Validated software resources...9 Validated system versions...9 Virtual networking configuration... 9 Management server infrastructure... 10 High availability... 11 VMware Horizon architecture...12 Test Environment Configuration and Best Practices 7

Test Environment Configuration and Best Practices Validated hardware resources Enterprise hardware Dell EMC validated the solution with the specific hardware resources listed in this section. We used the Dell EMC XC Family XC740xd-24 device with the components listed in the following table. We have designated the configurations as B5 and C7, which are referenced throughout the document. Table 1 Validated hardware configurations Configur ation Enterprise platform CPU Memory RAID controller HD configuration Network B5 XC740xd-24 5120 Gold (14- core 2.2 GHz) 384 GB @ 2,400 MT/s HBA 330 2 x 240 GB M.2 2 x 960 GB SSD 4 x 1.8 TB HDD 2 x Mellanox Connect X-4 LX 25 GbE SFP Rack NDC C7 XC740xd-24 6138 Gold (20- core 2.0 GHz) 768 GB @ 2,666 MT/s HBA 330 2 x 240 GB M.2 2 x 960 GB SSD Intel X710 rndc 6 x 1.8 TB HDD Graphics hardware We used the following NVIDIA GPU hardware in our tests for graphics-intensive workloads: NVIDIA Tesla M10 A dual-slot 10.5-inch PCI Express Gen3 graphics card featuring four mid-range 8 GB NVIDIA Maxwell GPUs and a total of 32 GB GDDR5 memory per card NVIDIA Pascal P40 A dual-slot 10.5-inch PCI Express Gen3 graphics card featuring a single high-end NVIDIA Pascal GPU and a total of 24 GB GDDR5 memory per card Network hardware We used the following network hardware in our test environment: Dell Networking S3048 (1 GbE ToR switch) A low-latency top-of-rack (ToR) switch that features 48 x 1 GbE and 4 x 10 GbE ports, a dense 1U design, and up to 260 Gbps switch fabric capacity Dell Networking S4048 (10 GbE ToR switch) A high-density, ultra-low-latency ToR switch that features 48 x 10 GbE SFP+ and 6 x 40 GbE ports and up to 720 Gbps switch fabric capacity 8 Dell EMC Ready Architectures for VDI

Test Environment Configuration and Best Practices Validated software resources Dell EMC validated this solution with the software components listed in the following table. Component Description/Version Hypervisor ESXi 6.5 Update 2 Broker technology VMware Horizon 7 version 7.5.1 Broker database Microsoft SQL Server 2016 Management VM OS Virtual desktop OS Microsoft Windows Server 2016 (Connection Server & DB) Microsoft Windows 10 Enterprise Office application suite Microsoft Office Professional 2016 Login VSI test suite Version 4.1 Platform Nutanix AOS version 5.5.3.1 NVIDIA GRID software (for graphics testing) 6.2 Validated system versions Dell EMC validated this solution using the system versions that are listed in the following table. Table 2 Version matrix for tested system Server configuration Hypervisor Hypervisor version Hypervisor build Hypervisor patches Bios AOS version Winsows 10 version Windows 10 patches Nvidia vgpu version B5 + 3 x P40 ESXi 6.5 U2 8294253 -- 1.3.7 5.5.3.1 1709 KB5969303 6.2 C7 ESXi 6.5 U1 5969303 -- 1.0.7 5.1.3 1703 -- -- C7 + 3 x M60 C7 + 3 x P40 ESXi 6.5 U1 5969303 -- 1.0.7 5.1.3 1703 -- 5.0 ESXi 6.5 U2 8294253 -- 1.3.7 5.5.3.1 1709 KB5969303 6.2 Virtual networking configuration The network configuration for the Dell EMC XC Family devices uses a 10 Gb converged infrastructure model. All required VLANs traverse two 10 Gb NICs configured in an active/active team. For larger scaling, we recommend that you separate the infrastructure management Validated software resources 9

Test Environment Configuration and Best Practices virtual machines (VMs) from the compute VMs to aid in predictable compute host scaling. We used the following VLAN configurations for the compute hosts, management hosts, and idrac in this solution model: Compute hosts Management VLAN: Configured for hypervisor infrastructure traffic L3 routed by using the spine layer Live Migration VLAN: Configured for Live Migration traffic L2 switched by using the leaf layer VDI VLAN: Configured for VDI session traffic L3 routed by using the spine layer Management hosts Management VLAN: Configured for hypervisor management traffic L3 routed by using the spine layer Live Migration VLAN: Configured for Live Migration traffic L2 switched by using the leaf layer VDI Management VLAN: Configured for VDI infrastructure traffic L3 routed by using the spine layer VLAN idrac: Configured for all hardware management traffic L3 routed by using the spine layer Management server infrastructure The following table lists the sizing requirements for the management server components. Table 3 Sizing for XC Family devices, RDSH, and NVIDIA GRID license server (optional) Component vcpu RAM (GB) NIC OS + data vdisk (GB) Tier 2 volume (GB) VMware vcenter Appliance 2 16 1 290 Platform Services Controller 2 2 1 30 Horizon Connection Server 4 16 1 40 SQL Server 4 8 1 40 210 (VMDK) File server 1 4 1 40 2,048 (VMDK) Nutanix CVM 12 32 3 0 RDSH VM 8 32 1 80 NVIDIA GRID License Server 2 4 1 40 + 5 NVIDIA GRID License Server 10 Dell EMC Ready Architectures for VDI When using NVIDIA vgpu cards, graphics-enabled VMs must obtain a license from a GRID License Server on your network to be entitled for vgpu. We installed the GRID License Server software on a system running a Windows 2012 R2 operating system to test vgpu configurations.

Test Environment Configuration and Best Practices SQL Server databases We made the following changes to the GRID License Server to address licensing requirements: Used a reserved fixed IP address Configured a single MAC address Applied time synchronization to all hosts on the same network During validation, a single dedicated SQL Server 2016 VM hosted the VMware databases in the management layer. We separated SQL data, logs, and tempdb into their respective volumes, and created a single database for Horizon Connection Server. We adhered to VMware best practices for this testing, including alignment of disks to be used by SQL Server with a 1,024 KB offset and formatted with a 64 KB file allocation unit size (data, logs, and tempdb). DNS DNS is the basis for Microsoft Active Directory and also controls access to various software components for VMware services. All hosts, VMs, and consumable software components must have a presence in DNS. We used a dynamic namespace integrated with Active Directory and adhered to Microsoft best practices. High availability Although we did not enable high availability (HA) during the validation that is documented in this guide, we strongly recommend that HA be factored into any VDI design and deployment. This process involves following the N+1 model with redundancy at both the hardware and software layers. The design guide for this architecture provides additional recommendations for HA. SQL Server databases 11

Test Environment Configuration and Best Practices VMware Horizon architecture The following figure shows the Horizon communication flow. Figure 1 VMware Horizon architecture 12 Dell EMC Ready Architectures for VDI

CHAPTER 3 Solution Performance and Testing This chapter presents the following topics: Testing process... 14 Test results and analysis... 18 Solution Performance and Testing 13

Testing process Resource monitoring To ensure the optimal combination of end-user experience (EUE) and cost-per-user, we conducted performance analysis and characterization testing on this solution using the Login VSI load-generation tool. Login VSI is a carefully designed, holistic methodology that monitors both hardware resource utilization parameters and EUE during load-testing. We tested each user load against four runs: a pilot run to validate that the infrastructure was functioning and valid data could be captured, and three subsequent runs to enable data correlation. During testing, while the environment was under load, we logged in to a session and completed tasks that correspond to the user workload. While this test is subjective, it helps to provide a better understanding of the EUE in the desktop sessions, particularly under high load. It also helps to ensure reliable data gathering. To ensure that the user experience was not compromised, we monitored the following important resources: Compute host servers VMware vcenter (for VMware vsphere-based solutions) or Microsoft Performance Monitor (for Hyper-V-based solutions) gathers key data (CPU, memory, disk and network usage) from each of the compute hosts during each test run. This data is exported to.csv files for single hosts, and then consolidated to show data from all hosts. While the report does not include specific performance metrics for the management host servers, these servers are monitored during testing to ensure that they are performing at an expected level with no bottlenecks. Hardware resources Resource contention, which occurs when hardware resources have been exhausted, can cause poor EUE. We monitored the relevant resource utilization parameters and applied relatively conservative thresholds, as shown in the following table. Thresholds are carefully selected to deliver an optimal combination of good EUE and cost-per-user while also providing burst capacity for seasonal or intermittent spikes in usage. Table 4 Parameter pass/fail thresholds Parameter Physical host CPU utilization Pass/fail threshold 85% a Physical host memory utilization 85% Network throughput 85% Storage I/O latency 20 ms a. The Ready Solutions for VDI team recommends that average CPU utilization not exceed 85% in a production environment. A 5% margin of error was allocated for this validation effort. Therefore, you will see CPU utilization that exceeds our recommended percentage. Given the nature of LoginVSI testing, this is a reasonable exception for determining our sizing guidance. GPU resources vsphere Client monitoring collects data about the GPU resource use from a script that is run on ESXi 6.5 and later hosts. The script runs for the 14 Dell EMC Ready Architectures for VDI

duration of the test and contains NVIDIA System Management Interface commands. The commands query each GPU and log the GPU processor, temperature, and memory use to a.csv file. Load generation Login VSI from Login VSI, Inc. is the industry-standard tool for testing VDI environments and Remote Desktop Session Host (RDSH) environments. Login VSI installs a standard collection of desktop application software (for example, Microsoft Office, Adobe Acrobat Reader) on each VDI desktop. It then uses launcher systems to connect a specified number of users to available desktops within the environment. When the user is connected, a logon script starts the workload, configures the user environment, and starts the test script. Each launcher system can launch connections to a number of VDI desktops (target machines). A centralized management console configures and manages the launchers and the Login VSI environment. In addition, we used the following login and boot paradigm: Users were logged in within a login timeframe of 1 hour, except when testing lowdensity solutions such as GPU/graphic-based configurations, in which users were logged in every 10 to 15 seconds. All desktops were started before users logged in. All desktops ran an industry-standard anti-virus solution. Windows 10 machines used Windows Defender. Profiles and workloads Machine profiles and user workloads determine the density numbers that the solution can support. Each profile and workload is bound by specific metrics and capabilities, with two targeted at graphics-intensive use cases. Profiles and workloads are defined as follows: Profile The configuration of the virtual desktop; the number of vcpus and the amount of RAM that is configured on the desktop and available to the user Workload The set of applications that is used We load-tested each profile by using a workload that is representative of the profile. The following table describes each use case. Table 5 Virtual desktop profiles and workloads Profile name/ workload Task worker Knowledge worker Workload description The least intensive of the standard workloads. This workload primarily runs Microsoft Excel and Microsoft Internet Explorer, with some minimal Microsoft Word activity, as well as Microsoft Outlook, Adobe, and copy and zip actions. The applications are started and stopped infrequently, which results in lower CPU, memory, and disk I/O usage. Designed for virtual machines with 2 vcpus. This workload includes the following activities: Outlook: Browse messages. Internet Explorer: Browse websites and open a YouTube style video (480p movie trailer) three times in every loop. Word: Initiate one instance to measure response time and another to review and edit a document. Load generation 15

Table 5 Virtual desktop profiles and workloads (continued) Profile name/ workload Workload description Doro PDF Printer and Acrobat Reader: Print a Word document and export it to PDF. Excel: Open a large randomized sheet. PowerPoint: Review and edit a presentation. FreeMind: Run a Java-based Mind Mapping application. Other: Perform various copy and zip actions. Power worker The most intensive of the standard workloads. The following activities are performed with this workload: Begin by opening four instances of Internet Explorer and two instances of Adobe Reader, which remain open throughout the workload. Perform more PDF printer actions than in the other workloads. Watch a 720p and a 1080p video. Reduce the idle time to two minutes. Perform various copy and zip actions. Graphics performance configuration/ multimedia A workload that is designed to heavily stress the CPU when using software graphics acceleration. GPU-accelerated computing offloads the most compute-intensive sections of an application to the GPU while the CPU processes the remaining code. This modified workload uses the following applications for its GPU/CPU-intensive operations: Adobe Acrobat Google Chrome Google Earth Microsoft Excel HTML5 3D spinning balls Internet Explorer MP3 Microsoft Outlook Microsoft PowerPoint Microsoft Word Streaming video Linked vs. Instant Clones Horizon supports two provisioning methods that deliver space-optimized virtual desktop pools: Linked Clones and Instant Clones. The user density per host is not impacted by using one over the other. The differences in the test graphs for these two methods are a result of the following processes: For Linked Clones, all the VMs are rebooted before the test starts to simulate a boot storm. 16 Dell EMC Ready Architectures for VDI

For Instant Clones, the VMs are rebooted after the session logs off, because when a user logs out of the instant clone, the clone is destroyed and recreated for the next user. The following figure shows the differences. Desktop VM test configurations The following table summarizes the compute VM configurations for the profiles and workloads that we tested. Table 6 Desktop VM specifications User profile vcpus ESXi configured memory ESXi reserved memory Screen resolution Operating system Task worker 2 a 2 GB 1 GB 1280 x 720 Windows 10 Enterprise 64-bit Knowledge worker 2 3 GB 1.5 GB 1920 x 1080 Windows 10 Enterprise 64-bit Power worker 2 4 GB 2 GB 1920 x 1080 Windows 10 Enterprise 64-bit Multimedia 4 8 GB 8 GB 1920 x 1080 Windows 10 Enterprise 64-bit a. Dell EMC has validated the LoginVSI Task worker workload with two vcpus assigned per VM, although LoginVSI lists the typical VM vcpu profile for this workload as being a single vcpu. Dell EMC diverges from this definition to deliver virtual desktops with great user experience. Increasing the vcpu count to 2 in the vcpu profile associated with the Task worker workload does have a minor impact on densities but generates improved user experience in return. Desktop VM test configurations 17

Test results and analysis We used the Login VSI test suite to simulate the user experience for several profile types under the typical workload for that type. The following table summarizes the test results for the compute hosts using the various workloads and configurations. Note Dell EMC is aware of the vulnerabilities, known as Meltdown, Spectre, and Foreshadow/L1TF, which affect many modern microprocessors. Ensure that you read the information in the following links: http://www.dell.com/support/article/sln308588 http://www.dell.com/support/article/sln308587 http://dell.com/support/l1-terminal-fault Consider this information in combination with the version information in Table 2 on page 9 to understand the vulnerability mitigation status of the environment we used to derive the test results shown in the following table. Table 7 Test results summary Server Configuration Workload User Density Avg CPU Avg Mem Active Avg IOPS / User B5 + 3 x P40 Multimedia (Virtual PC: P40-2B) 36 82% 328 11.2 C7 Knowledge Worker 150 76% 162 GB 2.11 C7 Power Worker 125 69% 154 GB 2.26 C7 + 3 x M60 Power Worker (Virtual Workstation: M60-1Q) 48 41% 224 GB 4.8 C7 + 3 x P40 Multimedia (Virtual PC: P40-1B) 72 96.5% a 616 2.5 a. The CPU utilization threshold of 85% is relaxed when testing with graphics cards. This test represents maximum utilization of the graphical resources available to the system as well as full user concurrency. Ideally, in a production environment, you would decrease the user density slightly or use higher bin processors to bring the CPU utilization closer to the 85% threshold. All LoginVSI tests completed successfully without reaching VSI maximum, indicating that user experience was good. 18 Dell EMC Ready Architectures for VDI The table headings are defined as follows: User density The number of users per compute host that successfully completed the workload test within the acceptable resource limits for the host. For clusters, this reflects the average of the density achieved for all compute hosts in the cluster. Avg CPU The average CPU usage over the steady-state period. For clusters, this represents the combined average CPU usage of all compute hosts. On the latest Intel processors, the ESXi host CPU metrics exceed the rated 100 percent for the host if Turbo Boost is enabled, which is the default setting. An additional 35 percent of CPU is available from the Turbo Boost feature, but this additional CPU headroom is not reflected in the VMware vsphere metrics where the performance data is gathered. Therefore, CPU usage for ESXi hosts is adjusted and each CPU graph includes a line indicating the potential performance headroom that is provided by Turbo boost. Avg active memory For ESXi hosts, the amount of memory that is actively used, as estimated by the VMkernel based on recently touched memory pages. For

clusters, this is the average amount of guest physical memory that is actively used across all compute hosts over the steady-state period. Avg IOPS per user IOPS calculated from the average disk IOPS over the steady state period divided by the number of users. B5 configuration We performed the following Task Worker, Knowledge Worker, Power Worker, and Multimedia performance testing on the XC740xd system in the B5 configuration described in Validated hardware resources on page 8. Multimedia (P40-2B), 36 vgpu users, ESXi 6.5 U2, Horizon 7.5.1 We ran the following tests on this workload. CPU We populated the GPU-enabled compute host with 36 vgpu-enabled virtual machines and used the NVIDIA P40-2B profile. With all user virtual machines powered on and before starting the test, the CPU usage was approximately 9 percent. The following figure shows the performance data for 36 user sessions on the management and GPU-enabled compute hosts. The CPU reached a steady state average of 82 percent during the test cycle when all users were logged on to the GPU-enabled compute host. GPU usage The GPU Metrics came form the vsphere Client. The GPU usage during the steady state period averaged approximately 38.1 percent and reached a peak usage of 61.66 percent with the multimedia workload. B5 configuration 19

Memory With regard to memory consumption, for this test run there were no constraints on the management or GPU-enabled compute hosts. Of a total of 384 GB available memory per node, the GPU compute host reached a maximum memory consumption of 336 GB with active memory usage reaching a max of 328 GB during the steady state phase. Each user session consumed 9.3 GB of memory and 9.1 GB of active memory after accounting for CVM and management memory usage. There were no variations in memory usage throughout the test, because all vgpu-enabled VM memory was reserved. No memory ballooning or swapping occurred on either host. 20 Dell EMC Ready Architectures for VDI

Network usage Network bandwidth was not an issue on this test run with a steady state peak of approximately 130 Mbps. The busiest period for network traffic was during the logoff phase. The compute and GPU hosts reached a peak of 820 Mbps during the logoff state IOPS The IOPS graphs and IOPS numbers came from the Nutanix Prism web console. They clearly display the initial logon of the desktops, the steady state, the logoff phase, and finally the recreation of the instant clones. The graphs show IOPS data for the cluster and the GPU host. The cluster IOPS reached a maximum of 1,741 disk IOPS during the recreation of the clones after the test finished and averaged 403 disk IOPS during steady state. Based on these numbers each user session generated 11 disk IOPS during steady state. B5 configuration 21

After the test finished, the cluster controller IOPS reached a maximum of 12,600 disk IOPS during the recreation of the clones. The cluster controller IOPS averaged 1,908 disk IOPS during steady state. Based on these numbers each user session generated 53 IOPS in steady state. GPU host disk IOPS After the test finished, the GPU host reached a maximum of 1,342 disk IOPS during the recreation of the clones. Disk IOPS averaged 71 during steady state. Based on these numbers each user session generated 2 IOPS in steady state. 22 Dell EMC Ready Architectures for VDI

Latency The latency graphs and latency numbers came from the Nutanix Prism web console. The graphs clearly display the initial logon of the desktops, the steady state, the logoff phase, and finally the recreation of the instant clones. The graphs show latency data for the cluster and the GPU host. The cluster latency reached a maximum latency of 0.86 ms during the recreation of the clones after the test finished and averaged 0.5 ms during steady state. The GPU host latency reached a maximum latency of 1.96 ms during the recreation of the clones after the test finished and averaged 0.57 during steady state. B5 configuration 23

User experience The user experience score did not reach the Login VSI maximum. When we manually interacted with the sessions during steady state, the mouse and window movement was responsive and video playback was good. The baseline performance of 987 indicated that the user experience for this test run was good. The Index average reached 1,289, which was well below the threshold of 1,987. 24 Dell EMC Ready Architectures for VDI

Notes During the recreation of the virtual machines, there was little network traffic on the compute /GPU host because it created the VMs on the other two hosts, and then migrated them onto the correct host with GPU cards. No disk latency issues occurred during testing. B5 configuration 25

C7 configuration We performed the following multimedia performance testing on the XC740xd system in the C7 configuration described in Validated hardware resources on page 8. Notes: We deployed Horizon and vsphere management roles within the cluster on a single host that also hosted desktops. This optional configuration is beneficial for POCs or small deployments looking to maximize user density. We allocated 12 vcpus and 32 GB of memory to the Nutanix Controller VM CVM) when we configured the Nutanix cluster. We set the Nvidia Pascal P40 GPU accelerator card with vgpu scheduling policy to Fixed Share Scheduler. Knowledge Worker, 445 Users, ESXi 6.5, Horizon 7.3.2 We ran the following tests on this workload. CPU In this workload test, the compute hosts hosted 150 desktop VMs each, while the management host hosted 145 sessions in addition to the Horizon management VMs. The peak CPU Usage was 84 percent on one host during logon phase, while the steady state average was 76 percent across all hosts. The relatively low steady state CPU usage was essential to avoid CPU scheduling problems, which would have placed many of the desktops into the Ready-waiting state, causing VSI errors to a large number of sessions. The average CVM CPU usage during steady state was 7.9 percent, while the management VMs used 1.2 percent altogether. 26 Dell EMC Ready Architectures for VDI Consumed memory The memory consumption averaged 483 GB in steady state across all hosts, and the peak usage on any host was 499 GB during the steady state phase. The peak usage was 65 percent, well below the 85 percent threshold, while the steady state average usage for all hosts was 63 percent. There was no swapping or ballooning during the

test run and each desktop consumed 3.0 GB, after accounting for CVM and management VM memory consumption. The CVM on each host consumed its full 32GB of reserved memory throughout the test run. Active memory Active memory during the test run peaked at 330 GB on the management host during the boot storm, and averaged 162 GB during steady state. Each desktop accounted for 0.86 GB of active memory usage after deducting CVM and management VM active memory. The CVM on each host used a full 32 GB of active memory throughout the test run. C7 configuration 27

Network usage Network usage peaked at 874 Mbps during boot storm on one host, and the average network usage for all hosts was 786 Mbps during steady state. Each desktop produced network throughput of 5.30 Mbps in steady state. IOPS The peak Cluster IOPS for the test run was 9,090 IOPS during the boot storm phase, while the average in steady state was 939 IOPS. Based on these numbers each user session generated 2.11 IOPS in steady state. 28 Dell EMC Ready Architectures for VDI

I/O latency The peak cluster I/O latency was 0.6 ms during the steady state. The average cluster I/O latency during steady state was 0.5 ms. The highest IO latency on any host was 0.7 ms during the boot storm. The chart clearly shows a very steady and very low level of I/O Latency throughout the test run. User experience The baseline performance of 731 indicated that the user experience for this test run was very good. The index average reached 1,122, well below the VSI maximum threshold of 1,732. Although there is considerable space left over for additional sessions according to the VSI graph, adding more session would have resulted in excessive session errors. The irregular shape of the VSI index average curve in the following figure shows the effect of the excess session latency that occurred on a subset of user sessions. C7 configuration 29

Power worker, 375 users, ESXi 6.5, Horizon 7.3.2 We ran the following tests on this workload. CPU usage In this workload test run, each compute host had 125 user sessions, while the designated management host had the full set of management VMs plus 125 desktops. The peak CPU usage was 86 percent on one host during steady state phase, while the steady state average was 69 percent across all hosts. The relatively low steady state CPU usage was essential to avoid session latency problems that would have caused excessive VSI errors. The CVMs on each host averaged 7.3 percent CPU usage during steady state. The management VMs used only 1.2 percent CPU on the management host in steady state. 30 Dell EMC Ready Architectures for VDI

Consumed memory The memory consumption averaged 483 GB in steady state across all hosts, and the peak usage on any host was 554 GB during the logoff phase. The peak usage was 72 percent, well below the 85 percent threshold, while the steady state average usage for all hosts was 63 percent. There was no swapping or ballooning during the test run and each desktop consumed 3.54 GB after accounting for CVM and management VM memory consumption. Active memory Active memory usage reached a maximum of 416 GB on the compute B host during the boot storm, and the average steady state usage for all hosts was 154 GB. Each C7 configuration 31

desktop used 0.96 GB of active memory after deducting for CVM and management VM usage. The CVM used its full 32 GB of active memory throughout the test run. Network usage Network usage peaked at 1,041 Mbps on one host during the steady state phase, and the average network usage for all hosts during steady state was 839 Mbps. Each desktop accounted for 6.71 Mbps in steady state. 32 Dell EMC Ready Architectures for VDI

IOPS The peak cluster IOPS for the test run was 6,585 IOPS during the boot storm phase, while the average in steady state was 848 IOPS. Based on these numbers, each user session generated 2.26 IOPS during steady state I/O latency The peak cluster I/O latency was 0.6 ms during the boot storm, while the peak on any host was 0.6 ms. The average cluster I/O latency during steady state was 0.5 ms. The chart clearly shows a very steady and very low level of I/O latency throughout the test run. C7 configuration 33

User experience The baseline performance of 725 indicated that the user experience for this test run was very good. The index average reached 988, well below the threshold of 1,725. Although the difference between the VSI maximum and the average would seem to indicate that more desktops could be used, the CPU limitations described above would have reduced the user experience dramatically. The irregular shape of the VSI index average curve shown in the following figure shows the effect of the excess session latency that occurred on a subset of user sessions. Multimedia workload (P40-1B), 72 vgpu users, ESXi 6.5 U2, Horizon 7.5.1 We ran the following tests on this workload. CPU usage The GPU-enabled compute host was populated with 72 vgpu-enabled virtual machines and used the NVIDIA P40-1B profile. With all user virtual machines powered on and before starting the test, the CPU usage was approximately 10 percent on the GPU-enabled compute host. The following figure shows the performance data for 72 user sessions on the management and GPU-enabled compute hosts. The CPU reached a steady state average of 96.5 percent during the test cycle when all 48 users were logged on to the GPU-enabled compute host. The management host reached a CPU maximum of 11 percent during steady state. 34 Dell EMC Ready Architectures for VDI

GPU usage We gathered the GPU metrics from the vsphere Web Client. The GPU usage during the steady state period averaged approximately 32.2 percent and reached a peak usage of 40.3 percent with the multimedia workload. Memory With regard to memory consumption for this test run, there were no constraints on the management or GPU-enabled compute hosts. Of a total of 768 GB available memory per node, the GPU compute host reached a maximum memory consumption of 635 GB with active memory usage reaching a maximum of 616 GB during the steady state phase. Each user session consumed 8.8 GB of memory and 8.5 GB of active memory. There were no variations in memory usage throughout the test as all vgpuenabled VM memory was reserved. There was no memory ballooning or swapping on either host. C7 configuration 35

Network usage Network bandwidth was not an issue on this test run, with a steady state peak of approximately 1,059 Mbps. The busiest period for network traffic was during the logoff phase. The compute / GPU host reached a peak of 1,477 Mbps at the start of logoff. 36 Dell EMC Ready Architectures for VDI

IOPS The IOPS graphs and numbers came from the Nutanix Prism web console. They clearly display the boot storm, the initial logon of the desktops, the steady state, and finally the logoff phase. The graphs show IOPS data the cluster the GPU host. The cluster reached a maximum of 1,595 disk IOPS during the boot storm and averaged 183 IOPS during steady state. Based on these numbers, each user session generated 2.5 disk IOPS during steady state. The cluster controller reached a maximum of 3,999 disk IOPS during the boot storm and averaged 1,666 disk IOPS during steady state. Based on these numbers, each user session generated 23.14 IOPS in steady state. C7 configuration 37

GPU host disk IPOS The GPU host reached a maximum of 1,117 disk IOPS during the boot storm and averaged 127 disk IOPS during steady state. Based on these numbers, each user session generated 1.77 IOPS in steady state. Latency The latency graphs and latency numbers came from the Nutanix Prism web console. They clearly display the initial logon of the desktops, the steady state, and then the logoff phase The graphs show latency data for the cluster and the GPU host. The cluster latency reached a maximum of 2.12 ms during the boot storm and averaged 0.63 ms during steady state. 38 Dell EMC Ready Architectures for VDI

The GPU host latency reached a maximum of 3.6 ms during the boot storm and averaged 0.68 ms during steady state. User experience The following figure shows that the user experience score did not reach the Login VSI maximum for this test. When manually interacting with the sessions during steady state the mouse and window movement was responsive and video playback was good. C7 configuration 39

Notes No disk latency issues occurred during testing. 40 Dell EMC Ready Architectures for VDI

CHAPTER 4 Conclusion Density recommendations...42 Summary...42 Conclusion 41

Conclusion Density recommendations We tested all configurations with Microsoft Windows 10 and Microsoft Office 2016. Test results provide recommended user densities, as shown in the following table. Table 8 User density recommendations Server configuration Workload User density B5 + 3 x P40 Multimedia (virtual PC: P40-2B) 36 C7 Knowledge worker 150 C7 Power worker 125 C7 + 3 x M60 Power worker (virtual workstation: M60-1Q) 48 C7 + 3 x P40 Multimedia (virtual PC: P40-1B) 72 Summary The configurations for the XC Family devices have been optimized for VDI. We selected the memory and CPU configurations that provide optimal performance. You can change these configurations to meet your own requirements, but keep in mind that changing the memory and CPU configurations from those that have been validated in this document will affect the user density per host. With the introduction of the six-channels-per-cpu requirement for Skylake, the C7 memory configuration recommendation has increased from the previous guidance of 512 GB to 768 GB. This change was necessary to ensure a balanced memory configuration and optimized performance for your VDI solution. The additional memory is advantageous, considering the resulting increase in operating system resource utilization and the enhanced experience for users when they have access to additional memory allocations. 42 Dell EMC Ready Architectures for VDI

CHAPTER 5 References This chapter presents the following topics: Dell EMC documentation...44 VMware documentation... 44 NVIDIA documentation...44 References 43

References Dell EMC documentation VMware documentation NVIDIA documentation The following Dell EMC documentation provides additional and relevant information. Access to these documents depends on your login credentials. If you do not have access to a document, contact your Dell EMC representative. Dell EMC Virtual Desktop Infrastructure Dell EMC XC Series and XC Core Technical Resource Center This document is part of the documentation set for this architecture, which includes the following: Dell EMC Ready Architectures for VDI: Designs for VMware Horizon on XC Family Design Guide Dell EMC Ready Architectures for VDI: Designs for VMware Horizon on XC Family Deployment Guide Dell EMC Ready Architectures for VDI: Designs for VMware Horizon on XC Family Validation Guide The following VMware documentation provides additional and relevant information: VMware vsphere documentation VMware Horizon 7 documentation Best Practices for Published Application and Desktops in VMware Horizon Apps and VMware Horizon 7 VMware Compatibility Guide Horizon 7 Enterprise Edition Reference Architecture Horizon 7 Enterprise Edition Multi-Site Reference Architecture The following NVIDIA documentation provides additional and relevant information: NVIDIA Virtual GPU Software Quick Start Guide 44 Dell EMC Ready Architectures for VDI