DELL EMC READY BUNDLE FOR VIRTUALIZATION WITH VMWARE AND ISCSI INFRASTRUCTURE

Similar documents
DELL EMC READY BUNDLE FOR VIRTUALIZATION WITH VMWARE AND FIBRE CHANNEL INFRASTRUCTURE

DELL EMC READY BUNDLE FOR VIRTUALIZATION WITH VMWARE VSAN INFRASTRUCTURE

Microsoft SharePoint Server 2010 on Dell Systems

vstart 1000v for Enterprise Virtualization using VMware vsphere: Reference Architecture

Dell EMC Ready Stack: Microsoft Hyper-V on PowerEdge 14G Servers and Unity Storage

Dell EMC SC Series Storage with SAS Front-end Support for VMware vsphere

TPC-E testing of Microsoft SQL Server 2016 on Dell EMC PowerEdge R830 Server and Dell EMC SC9000 Storage

Active System Manager Release 8.2 Compatibility Matrix

Dell EMC. VxRack System FLEX Architecture Overview

Dell EMC Ready Solution for VMware vcloud NFV 3.0 OpenStack Edition Platform

Dell EMC. VxBlock Systems for VMware NSX 6.3 Architecture Overview

TITLE. the IT Landscape

Implementing SharePoint Server 2010 on Dell vstart Solution

Accelerating Microsoft SQL Server 2016 Performance With Dell EMC PowerEdge R740

Microsoft SQL Server in a VMware Environment on Dell PowerEdge R810 Servers and Dell EqualLogic Storage

The Impact of Hyper- converged Infrastructure on the IT Landscape

VxRack System SDDC Enabling External Services

Implementing SQL Server 2016 with Microsoft Storage Spaces Direct on Dell EMC PowerEdge R730xd

Dell Technologies IoT Solution Surveillance with Genetec Security Center

Dell EMC Ready Solutions for Oracle with Dell EMC XtremIO X2 and Data Domain

vstart 50 VMware vsphere Solution Specification

A Dell Technical White Paper Dell Virtualization Solutions Engineering

Upgrade to Microsoft SQL Server 2016 with Dell EMC Infrastructure

vstart 50 VMware vsphere Solution Overview

Dell EMC Ready Stack Design Guide for VMware vsphere on Unity

Reference Architecture - Microsoft SharePoint Server 2013 on Dell PowerEdge R630

Dell EMC. VxBlock Systems for VMware NSX 6.2 Architecture Overview

Consolidating Microsoft SQL Server databases on PowerEdge R930 server

Dell PowerEdge M1000e Blade Enclosure and Dell PS Series SAN Design Best Practices Using Dell S-Series and M-Series Networking

Introducing VMware Validated Designs for Software-Defined Data Center

Remote and Branch Office Reference Architecture for VMware vsphere with Dell PowerEdge VRTX

Introducing VMware Validated Designs for Software-Defined Data Center

DELL EMC TECHNICAL SOLUTION BRIEF. ARCHITECTING A DELL EMC HYPERCONVERGED SOLUTION WITH VMware vsan. Version 1.0. Author: VICTOR LAMA

Introducing VMware Validated Designs for Software-Defined Data Center

Dell Fluid Data solutions. Powerful self-optimized enterprise storage. Dell Compellent Storage Center: Designed for business results

EMC VSPEX END-USER COMPUTING

Dell EMC Ready Bundle for HPC Digital Manufacturing Dassault Systѐmes Simulia Abaqus Performance

NetApp HCI Network Setup Guide

PowerEdge FX2 - Upgrading from 10GbE Pass-through Modules to FN410S I/O Modules

Microsoft SharePoint Server 2010 Implementation on Dell Active System 800v

Vblock Architecture. Andrew Smallridge DC Technology Solutions Architect

2000 Persistent VMware View VDI Users on Dell EMC SCv3020 Storage

EMC CLARiiON CX3-40. Reference Architecture. Enterprise Solutions for Microsoft Exchange Enabled by MirrorView/S

Dell DCPPE-200. Dell PowerEdge Professional. Download Full version :

VxRack SDDC Deep Dive: Inside VxRack SDDC Powered by VMware Cloud Foundation. Harry Meier GLOBAL SPONSORS

VMware vsphere with ESX 4 and vcenter

Dell EMC Validated System for Virtualization Design Guide for VMware with Virtual SAN Infrastructure. A Dell Architecture Guide

Best Practices for Deploying a Mixed 1Gb/10Gb Ethernet SAN using Dell Storage PS Series Arrays

VMware Validated Design for NetApp HCI

VxRail: Level Up with New Capabilities and Powers

Best Practices for Sharing an iscsi SAN Infrastructure with Dell PS Series and SC Series Storage using VMware vsphere Hosts

Dell EMC vsan Ready Nodes for VDI

Reference Architecture for Active System 1000 with Microsoft Hyper-V

Dell EMC Ready Architectures for VDI

Reference Architectures for designing and deploying Microsoft SQL Server Databases in Active System800 Platform

Scale out a 13th Generation XC Series Cluster Using 14th Generation XC Series Appliance

Cisco HyperFlex HX220c M4 Node

Introduction to Virtualization. From NDG In partnership with VMware IT Academy

Dell EMC Ready Architectures for VDI

Paul Hodge 25 Oct 2016 RUN FAST. NEVER STOP. The Next Generation Virtual Platform

The PowerEdge FX Architecture Portfolio Overview Re-inventing the rack server for the data center of the future

Dell Technologies IoT Solution Surveillance with Genetec Security Center

Dell EMC Microsoft Storage Spaces Direct Ready Node

Jake Howering. Director, Product Management

Reference Architecture for Dell VIS Self-Service Creator and VMware vsphere 4

Wyse Datacenter for Citrix XenDesktop Reference Architecture

Cisco HyperFlex HX220c M4 and HX220c M4 All Flash Nodes

VMware vsphere with ESX 4.1 and vcenter 4.1

VxRack FLEX Technical Deep Dive: Building Hyper-converged Solutions at Rackscale. Kiewiet Kritzinger DELL EMC CPSD Snr varchitect

A Reference Architecture for SAP Solutions on Dell and SUSE

Dell EMC. Vblock System 340 with VMware Horizon 6.0 with View

Deployment of VMware Infrastructure 3 on Dell PowerEdge Blade Servers

EMC Integrated Infrastructure for VMware. Business Continuity

EMC Performance Optimization for VMware Enabled by EMC PowerPath/VE

Cisco HyperFlex Systems

CLOUD PROVIDER POD RELEASE NOTES

Dell PowerEdge R910 SQL OLTP Virtualization Study Measuring Performance and Power Improvements of New Intel Xeon E7 Processors and Low-Voltage Memory

ENTERPRISE HYBRID CLOUD 4.1.2

EMC Business Continuity for Microsoft Applications

Dell EMC Big Cloud Fabric Deployment and Best Practices Guide with VMware vsan. Dell EMC Networking Infrastructure Solutions November 2017

V.I.B.E. Virtual. Integrated. Blade. Environment. Harveenpal Singh. System-x PLM

Cisco Nexus 4000 Series Switches for IBM BladeCenter

Improving Blade Economics with Virtualization

EMC VSPEX END-USER COMPUTING

Dell EMC Ready System for VDI on VxRail

Business-Ready Configurations for VMware vsphere 4 with PowerEdge R810 Servers and EqualLogic PS6010 Storage Array Utilizing iscsi Boot Technology

DATA PROTECTION IN A ROBO ENVIRONMENT

SAN Design Best Practices for the Dell PowerEdge M1000e Blade Enclosure and EqualLogic PS Series Storage (1GbE) A Dell Technical Whitepaper

Overview. Cisco UCS Manager User Documentation

EMC STORAGE FOR MILESTONE XPROTECT CORPORATE

VMware Infrastructure Update 1 for Dell PowerEdge Systems. Deployment Guide. support.dell.com

5,000 Persistent VMware View VDI Users on Dell EMC SC9000 Storage

Creating a VMware Software-Defined Data Center REFERENCE ARCHITECTURE VERSION 1.5

Reference Architecture for Microsoft Hyper-V R2 on Dell PowerEdge M- Series Servers with EqualLogic Storage

CLOUD PROVIDER POD. for VMware. Release Notes. VMware Cloud Provider Pod January 2019 Check for additions and updates to these release notes

PBO1064BU VxRack SDDC Deep Dive: Inside VxRack SDDC Powered by VMware Cloud Foundation Jason Marques, Dell EMC Georg Edelmann, VMware VMworld 2017 Con

Cisco UCS-Mini Solution with StorMagic SvSAN with VMware vsphere 5.5

Cisco HyperFlex HX220c M4 and HX220c M4 All Flash Nodes

ENTERPRISE HYBRID CLOUD 4.1.1

Free up rack space by replacing old servers and storage

Transcription:

DELL EMC READY BUNDLE FOR VIRTUALIZATION WITH VMWARE AND ISCSI INFRASTRUCTURE Design Guide APRIL 2017 1

The information in this publication is provided as is. Dell Inc. makes no representations or warranties of any kind with respect to the information in this publication, and specifically disclaims implied warranties of merchantability or fitness for a particular purpose. Use, copying, and distribution of any software described in this publication requires an applicable software license. Copyright 2016 Dell Inc. or its subsidiaries. All Rights Reserved. Dell, EMC, and other trademarks are trademarks of Dell Inc. or its subsidiaries. Other trademarks may be the property of their respective owners. Published in the USA May 2017. Dell EMC believes the information in this document is accurate as of its publication date. The information is subject to change without notice.

TABLE OF CONTENTS EXECUTIVE SUMAMRY... 4 INTRODUCTION... 5 AUDIENCE AND SCOPE... 5 SUPPORTED CONFIGURATION... 5 Compute server... 5 Management server... 6 DESIGN PRINCIPLES... 7 DESIGN OVERVIEW... 7 COMPUTE DESIGN... 8 NETWORK DESIGN... 8... 8 Network connectivity of PowerEdge FX2 Servers... 9 Network configuration of PowerEdge rack servers... 9 Network configuration for LAN traffic with VMware vsphere Distributed Switch (vds)... 10 STORAGE DESIGN... 10 Storage fabric configuration... 10 Storage connectivity for compute and management servers... 11 Storage configuration... 12 MANAGEMENT DESIGN... 12 Management Infrastructure... 12 Management Components... 13 VMWARE VREALIZE AUTOMATION DESIGN... 13 vrealize Automation components... 13 VMware vrealize Automation deployment model sizing... 14 SCALING THE READY BUNDLE... 16 REFERENCES... 17

Executive Sumamry This Design Guide outlines the architecture of the Dell EMC Ready Bundle for Virtualization with VMware and iscsi infrastructure. The Dell EMC Ready Bundle for Virtualization is the most flexible Ready Bundle in the industry. This converged iscsi solution can be built to include Dell PowerEdge rack servers, PowerEdge FX2 converged platforms, Dell SC Storage SC4020 for mid-tier enterprise storage and Dell SC Storage SCv2020 for entry-level storage. Virtualization with VMware vsphere 6 completes the solution to provide a robust platform for enterprise workloads.

Introduction The Dell EMC Ready Bundle for Virtualization with VMware is a family of converged and hyper-converged systems that combine servers, storage, networking, and infrastructure management into an integrated and optimized system that provides a platform for general-purpose virtualized workloads. The components undergo testing and validation to ensure full compatibility with VMware vsphere. The Ready Bundle uses infrastructure best practices, and converged server architecture. This document provides an overview of the design and implementation of the Dell EMC Ready Bundle for Virtualization with VMware and iscsi storage. Audience and scope This document is intended for stakeholders in IT infrastructure and service delivery who have purchased or are considering purchase of Dell EMC Ready Bundle for Virtualization with VMware and iscsi infrastructure. This document provides information and diagrams to familiarize you with the components that make up the bundle. Supported Configuration The following table provides an overview of the supported components: Table 1. Supported Configuration Components Entry level configuration Mid-tier configuration Server platforms Choice of: Dell EMC PowerEdge R630, Dell EMC PowerEdge R730 Choice of: Dell EMC PowerEdge R630, Dell EMC PowerEdge R730, Dell EMC PowerEdge FX2 with Dell EMC PowerEdge FC630 or Dell EMC PowerEdge FC430 LAN connectivity (2) Dell Networking S4048-T (2) SAN connectivity (2) Dell Networking S4048T-T (2) Out-of-band (OOB) connectivity None. LAN switch is used for OOB (1) Dell Networking S3048-ON Storage array Dell Storage SCv2020 Dell Storage SC4020 Management server platform (2) Dell EMC PowerEdge R630 (2) Dell EMC PowerEdge R630 Management software components Dell Storage Manager VMware vcenter Server Appliance VMware vrealize Automation VMware vrealize Business VMware vrealize Log Insight VMware vrealize Operations Manager with optional Dell Storage Management Pack Dell Storage Manager VMware vcenter Server Appliance VMware vrealize Automation VMware vrealize Business VMware vrealize Log Insight VMware vrealize Operations Manager with optional Dell Storage Management Pack Compute server Ready bundle for Virtualization offers ccustomers a choice of rack servers or modular enclosures for their compute infrastructure. The following table details the supported compute servers:

Table 2. Platform model Compute server configurations PowerEdge R630 and PowerEdge R730 PowerEdge FX2 with PowerEdge FC630 PowerEdge FX2 with PowerEdge FC430 Processor (2) Intel Xeon E5 v4 Broadwell processors (2) Intel Xeon E5 v4 Broadwell processors (2) Intel Xeon E5 v4 Broadwell processors Memory 2400 MT/s RDIMMs 2400 MT/s RDIMMs 2400 MT/s RDIMMs Network adapter Intel X710 quad port 10Gb network adapter or QLogic 57800 10 Gb quad port network adapter Intel X710 dual port 10Gb network adapter or QLogic 57810 10 Gb dual port network adapter QLogic 57810 10 Gb dual port network adapter Host bus adapter Intel X710 dual port 10Gb network adapter or QLogic 57810 10 Gb dual port network adapter Intel X710 dual port 10Gb network adapter or QLogic 57810 10 Gb dual port network adapter QLogic 57810 10 Gb dual port network adapter Boot device Internal Dual SD Module with 16 GB SD card for ESXi Internal Dual SD Module with 16 GB SD card for ESXi Internal Dual SD Module with 16 GB SD card for ESXi OOB management idrac8 Enterprise idrac8 Enterprise idrac8 Enterprise Hypervisor VMware ESXi 6.5 VMware ESXi 6.5 VMware ESXi 6.5 FX2 chassis configuration Not applicable CMC and (2) Dell Networking FN IOAs CMC and (2) Dell Networking FN IOAs Management server The management cluster consist of three PowerEdge R630 servers will the following configuration: Table 3. Management server configuration Components Details Number of servers Processor Memory Network adapter Host bus adapter Boot device OOB management (2) Dell EMC PowerEdge R630 (2) Intel Xeon E5 v4 Broadwell processors 2400MT/s RDIMMs Intel X710 quad port 10Gbnetwork adapter or QLogic 57800 10 Gb quad port network adapter Intel X710 dual port 10Gb network adapter or QLogic 57810 10 Gb dual port network adapter Internal Dual SD Module with 16 GB SD card for ESXi idrac8 Enterprise

49 50 51 52 LNK LNK 4 8 12 16 4 8 12 16 4 8 12 16 LNK LNK Design Principles The following principles are central to the design and architecture of the Dell EMC Ready Bundle for Virtualization. The Ready Bundle for Virtualization is built and validated using these design principles. No single point-of-failure: Redundancy is incorporated in the critical aspects 1 of the solution, including server high availability features, networking, and storage. Integrated Management: Provide integrated management using vrealize Operations Manager with associated plugins. Hardware configuration for virtualization: The system is designed for general use case virtualization. Each server is equipped with appropriate processor, memory, host bus, and network adapters as required for virtualization. Best practices adherence: Storage, networking and vsphere best practices of the corresponding components are incorporated into the design to ensure availability, serviceability and optimal performance. vrealize Suite Enabled: The system supports VMware vrealize Suite for managing heterogeneous and hybrid cloud. Flexible configurations: Ready Bundle for Virtualization can be configured to suit most customer needs for a virtualized infrastructure. The solution supports options, such as configuring server model, server processors, server memory, type of storage (Fibre Channel, iscsi and SDS), and network switches based on customer needs. Design Overview This section provides an overview of the architecture including computer, storage, network, and management architecture. The following figure provides a high-level overview of the architecture, including compute servers (showing flexible compute nodes), management servers, LAN switches, SAN switches, storage arrays and out-of-band switches: Out-of-Band Networking Dell Networking S3048-ON Management Cluster PowerEdge R630 Data Center Network LAN 2x 2x SAN Compute Cluster PowerEdge FX2 with FC630/FC430 Dell SC Series PowerEdge R630 PowerEdge R730 Dell Storage Expansion Enclosures Figure 1. Ready Bundle for Virtualization with VMware and iscsi infrastructure Design Overview 1 Out of band management is not considered critical to user workload and does not have redundancy

LNK L N K 1 2 3 4 5 6 7 8 9 1 0 1 1 1 2 1 3 1 4 1 5 1 6 1 7 1 8 1 9 2 0 2 1 2 2 2 3 2 4 2 5 2 6 2 7 2 8 2 9 3 0 3 1 3 2 3 3 3 4 3 5 3 6 3 7 3 8 3 9 4 0 4 1 4 2 4 3 4 4 4 5 4 6 4 7 4 8 A C T 4 9 5 1 5 3 5 0 5 2 5 4 S t a c k - I D Compute Design The latest Intel Xeon v4 Broadwell generation processors power the Ready Bundle for Virtualization with VMware. With up to 22 cores per CPU and clock speeds of up to 3.5Ghz in the Dell PowerEdge rack servers you can reduce CPU socket-based licensing costs and achieve greater VM density. The Dell EMC converged FX2 platform provides a powerful balance of resource density and power consumption in a dense form factor. These server nodes are configured with processors that have a lower thermal design power (TDP) value resulting in lower cooling and electrical costs. The Dell PowerEdge rack platforms support the RDIMM and LRDIMM memory types. Load Reduced DIMM (LRDIMM) uses an imb buffer to isolate electrical loading from the host memory controller. This buffer and isolation allows for the use of quad ranked DIMM to increase overall memory capacity. For general-purpose virtualization solutions, 2400 MT/s RDIMMs are recommended. Memory can be configured in various modes from within the BIOS. Optimizer mode is the default mode and is recommended for most virtualization use cases to provide the optimized memory performance. For improved reliability and resiliency, other modes such as mirror mode and Dell fault-resilient mode are available. The Intel X710 or QLogic network adapters are supported to provide 10 Gb network connectivity to the top of rack LAN switches. Additional Intel X710 or QLogic network adapters are supported to provide 10 Gb network iscsi connectivity to the SAN switches Dell PowerEdge servers support various BIOS configuration profiles that control the processor, memory, and other configuration options. Dell recommends the default profile of performance for the Dell EMC Ready Bundle for Virtualization with VMware. Network Design This section provides an overview of the network architecture including compute and management server connectivity. Details around the Top-of-Rack (ToR) and virtual switch configuration are provided. The network architecture employs Virtual Link Trunking (VLT) connection between the two Top-of-Rack (ToR) switches. In a non- VLT environment, redundancy requires idle equipment which drives up infrastructure costs and increases risks. In a VLT environment, all paths are active, adding immediate value and throughput while still protecting against hardware failures. VLT technology allows a server or bridge to uplink a physical trunk into more than one switch by treating the uplink as one logical trunk. A VLT connected pair of switches acts as a single switch to a connecting bridge or server. Both links from the bridge network can actively forward and receive traffic. VLT provides a replacement for Spanning Tree Protocol (STP) based networks by providing both redundancy and full bandwidth utilization using multiple active paths. Major benefits of VLT technology are: o o o Dual control plane for highly available, resilient network services Full utilization of the active LAG interfaces Active / Active design for seamless operations during maintenance events VLTi Figure 2. Virtual Link Trunk Interconnect (VLTi) Configuration The switches each provide six 40 GbE uplink ports. The Virtual Link Trunk Interconnect (VLTi) configuration in this architecture uses two 40 GbE ports from each Top-of-Rack (ToR) switch to provide an 80 Gb data path between the switches. The remaining four 40Gb ports allow for high speed connectivity to spine switches or directly to the data center core network infrastructure. They can also be used to extend connectivity to other racks.

LNK LNK idrac idrac LNK LNK Gb1 STK/Gb2 1 9 5 210 6 311 7 4 12 8 1 9 5 2 10 6 3 11 7 4 12 8 L N K 1 2 3 4 5 6 7 8 9 1 0 1 1 1 2 1 3 1 4 1 5 1 6 1 7 1 8 1 9 2 0 2 1 2 2 2 3 2 4 2 5 2 6 2 7 2 8 2 9 3 0 3 1 3 2 L N K 1 2 3 4 5 6 7 8 9 1 0 1 1 1 2 1 3 1 4 1 5 1 6 1 7 1 8 1 9 2 0 2 1 2 2 2 3 2 4 2 5 2 6 2 7 2 8 2 9 3 0 3 1 3 2 3 3 3 4 3 5 3 6 3 7 3 8 3 9 4 0 4 1 4 2 4 3 4 4 4 5 4 6 4 7 4 8 3 3 3 4 3 5 3 6 3 7 3 8 3 9 4 0 4 1 4 2 4 3 4 4 4 5 4 6 4 7 4 8 A C T 4 9 5 1 5 3 5 0 5 2 5 4 A C T S t a c k - I D 4 9 5 1 5 3 5 0 5 2 5 4 S t a c k - I D Network connectivity of PowerEdge FX2 Servers This section describes the network connectivity when using PowerEdge FX2 modular infrastructure blade servers. The compute cluster Dell EMC PowerEdge FC630 and the Dell EMC PowerEdge FC430 server nodes connect to the switches through the Dell EMC PowerEdge FN410S modules shown in the following PowerEdge FX2 blade chassis figure: Connectivity between the PowerEdge FX server nodes and the PowerEdge FN410S: The internal architecture of PowerEdge FX2 chassis provides connectivity between an Intel or QLogic 10 GbE Network Daughter Card (NDC) in each PowerEdge FX2 server node and the internal ports of the PowerEdge FN410S module. The FN410 has eight 10 GbE internal ports. Four Dell EMC PowerEdge FC630 server nodes connect to four of the internal ports in each FN410. Eight Dell EMC PowerEdge FC430 server nodes connect to all eight of the internal ports in each PowerEdge FN410S. Connectivity between the Dell EMC PowerEdge FN401S and S4048-ON switches: Two Dell EMC PowerEdge FN410S I/O Aggregators (IOA) in the PowerEdge FX2 architecture provides the Top-of-Rack (ToR) connectivity for the PowerEdge FX2 server nodes. Each IOA provides four external ports. The FN IOA provides a low-touch configuration mode called VLT mode in which Port 9 from each IOA in the chassis forms a Virtual Link Trunk Interconnect (VLTi). Ports 10 and 11 from both the FN IOA form a single port channel which in turn connects to the Top-of-Rack (ToR) switches. PowerEdge FX2 1100W 1100W Port Channel 1 Port Channel 1 VLTi Figure 3. PowerEdge FX2 connectivity Network configuration of PowerEdge rack servers The compute cluster can either be Dell EMC PowerEdge FX2 servers or the Dell PowerEdge rack servers. This section describes the network connectivity if rack servers are used for compute servers, and the management servers. The following image is an example of the connectivity between the compute and management Dell PowerEdge rack servers and switches: The compute and management rack servers have two 10 GbE connections to S4048-ON switches through one Intel or QLogic dual port 10 GbE network daughter card (NDC). VLTi PowerEdge Rack Servers for Compute and Management 1 1 2 750W 750W 1 1 2 750W 750W Figure 4. PowerEdge rack server connectivity

Network configuration for LAN traffic with VMware vsphere Distributed Switch (vds) Customers can achieve bandwidth prioritization for different traffic classes such as host management, vmotion, and VM network using VMware Distributed Virtual Switches. The VMware vsphere Distributed Switch (vds) can be configured, managed, and monitored from a central interface and provides: o o o Simplified virtual machine network configuration Enhanced network monitoring and troubleshooting capabilities Support for network bandwidth partitioning when NPAR is not available The following figures show the virtual distributed switch configuration for the management and compute servers. Figure 5. Distributed Virtual Switch for dual port configuration For configuring iscsi vswitch, configure two vmkernal ports in the iscsi vswitch and bind the corresponding adapter. Each vmkernal port is bound to a physical adapter and then attached to the software or hardware iscsi adapter. Note: When using software iscsi adapter it is recommended to not to bind port. For more information refer: https://kb.vmware.com/selfservice/microsites/search.do?language=en_us&cmd=displaykc&externalid=2038869 Multipathing to an ESXi host is automatic when the server object has more than one HBA or iscsi initiator posts assigned to it. To configure iscsi initiator for multipathing consider following key points: Verify that there is one VMKernal interface for each physical NIC to be used for storage traffic, and that they follow the virtual switch port binding recommendations mentioned in above link. Adjust the failover order on each VMKernal interface for 1:1 VMKernal to physical NIC ratio. Add both VMKernal interfaces to the iscsi adapter network port binding when using hardware iscsi offload. Rescan the iscsi adapter for the new volumes. From within the DSM (Dell Storage Manager) client, create the server object for the host. For more details refer: https://pubs.vmware.com/vsphere-65/topic/com.vmware.icbase/pdf/vsphere-esxi-vcenter-server-65-storage-guide.pdf Storage Design The Ready Bundle uses the Dell Storage Center connected to a pair of 10 Gb Ethernet switches for storage traffic. The Dell SC Storage SC4020 provides a shared storage mechanism to enable a highly available datastore for midtier enterprise storage workloads. Fault domains ensure that data paths are distributed in a way that ensures a component failure does not impact virtual machine operations. The Dell SC Storage SCv2020 series provides the same high availability and redundancy to workloads for entry level enterprise storage. Two or S4048T network switches connect the iscsi fabric. Storage fabric configuration

LNK L N K L N K Gb1 STK/Gb2 A C T A C T G R N = 1 0 G A C T / L N K A G R N = 1 0 G A C T / L N K B G R N = 1 0 G A C T / L N K A G R N = 1 0 G A C T / L N K B G R N = 1 0 G A C T / L N K A G R N = 1 0 G A C T / L N K B L N K G R N = 1 0 G A C T / L N K A G R N = 1 0 G A C T / L N K B 1 2 3 4 5 6 7 8 9 1 0 1 1 1 2 1 3 1 4 1 5 1 6 1 7 1 8 1 9 2 0 2 1 2 2 2 3 2 4 2 5 2 6 2 7 2 8 2 9 3 0 3 1 3 2 3 3 3 4 3 5 3 6 3 7 3 8 3 9 4 0 4 1 4 2 4 3 4 4 4 5 4 6 4 7 4 8 A C T 4 9 5 1 5 3 5 0 5 2 5 4 S t a c k - I D This solution is configured with two fault domains for high availability. For the Dell SC Storage arrays, Port 1 from each controller connects to iscsi switch 1, while Port 2 connects to iscsi switch 2, as shown in the following figure: Each iscsi switch represents a fault domain for the iscsi storage and enables port failover. Figure 6. Storage fabric configuration with Dell SC Storage SC4020 Storage connectivity for compute and management servers Each server has a 10 Gb Ethernet network adapter for connecting to the iscsi storage fabrics. For Dell FX2 platform servers the PCIe network adapter is inserted in to the rear of the chassis and mapped to the appropriate server node using the chassis controller software. When using the PowerEdge FC630 servers, the PCIE slots 2, 4, 6 and 8 are populated with the 10 Gb network adapter as shown in the diagram below. The PowerEdge FC430 server requires the population of all eight slots. PCIe Slots 1 2 3 4 5 6 7 8 1100W 1 9 5 21 0 6 31 1 7 4 1 2 8 PowerEdge FX2 1100W 1 9 5 21 0 6 31 1 7 4 1 2 8 Figure 7. PowerEdge FX2 SAN connectivity For management servers and rack compute servers, each server is configured with a dual port 10 Gb network adapters with each port connecting to the SAN switch as shown.

LNK idrac idrac L N K 1 2 3 4 5 6 7 8 9 1 0 1 1 1 2 1 3 1 4 1 5 1 6 1 7 1 8 1 9 2 0 2 1 2 2 2 3 2 4 2 5 2 6 2 7 2 8 2 9 3 0 3 1 3 2 3 3 3 4 3 5 3 6 3 7 3 8 3 9 4 0 4 1 4 2 4 3 4 4 4 5 4 6 4 7 4 8 A C T 4 9 5 1 5 3 5 0 5 2 5 4 S t a c k - I D VLTi PowerEdge Rack Servers for Compute and Management GRN=10G /LNK A GRN=10G /LNK B 1 1 2 750W 750W GRN=10G /LNK A GRN=10G /LNK B 1 1 2 750W 750W Figure 8. PowerEdge Rack Server SAN connectivity Storage configuration Multiple datastores within the VMware vsphere Cluster enable vsphere HA Datastore Heartbeating. This ensures that partitioning, or isolated host networks do not trigger VM movement within the cluster. By default, the vsphere cluster selects up to five datastores for the Datastore Heartbeating. Management Design Management Infrastructure The management infrastructure consists for two R630 servers as part of a management cluster. Management components are virtualized to provide high availability. The bundle further protects these components using the dedicated management cluster. Redundant 10Gb Ethernet uplinks to the network infrastructure and storage fabric, combined with vsphere High Availability ensure that management components stay online. A Dell Networking S3048 switch is used for OOB connectivity. idrac ports in each management and compute cluster connect to this switch. vrealize Automation IaaS for vra vrealize Business for Cloud vcenter Server vrealize LogInsight vrealize Operations Manager Dell Storage Manager Solution Services NTP VMware vsphere Cluster VMware Management Cluster vmotion Enabled VMware HA Enabled VMware DRS Enabled Datacenter Network DNS Active Directory Two PowerEdge R630 rack servers Figure 9. Management Design

Management Components The following are management components: VMware vcenter Server Appliance VMware vrealize Automation vrealize IaaS VMware vrealize Business VMware vrealize Log Insight VMware vrealize Operations Manager Dell Storage Manager The management software components run on virtual machines contained in your management cluster. The following table lists the management components in the bundle and the VM sizing of those components: Table 4. Component VMs CPU RAM OS (GB) Data NIC cores (GB) (GB) VMware vcenter Server 1 4 12 12 256 1 vrealize Operations Manager 1 4 16 20 254 1 vrealize Automation (vra) 1 4 18 50 90 1 vrealize Business 1 4 8 50 0 1 vrealize Log Insight 1 4 8 20 511 1 vrealize IaaS (for vra) 1 4 6 80 0 1 Dell Storage Manager 1 2 8 80 0 1 VMware vrealize Automation Design The VMware vrealize Automation architecture can be deployed using several models to provide flexibility in resource consumption, availability, and fabric management. The small, medium, and large deployment models all scale up as needed. Note: Do not use the Minimal Deployment model described in the vrealize Automation documentation. This model is for proof of concept deployments only and cannot scale with the environment. vrealize Automation uses DNS entries for all components. When scaling, vrealize Automation uses load balancers to distribute the workload across multiple automation appliances, web servers, and infrastructure managers. These components can be scaled as needed and distributed in the same data center or geographically disperse. In addition to scaling for workload, vrealize Automation with load balancers is highly available. This feature ensures that users who consume vrealize Automation XaaS blueprints are not affected by an outage. vrealize Automation components The following are the components for vrealize Automation: VMware vrealize Automation Appliance - Contains core VRA services and the vrealize Orchestrator service, AD sync connectors, and an internal appliance database. Clusters A/A except DB failover is manual. Infrastructure web server - Runs IIS for user consumption of vrealize services and is fully active-active. Infrastructure Manager Service - Manages the infrastructure deployed with VRA. For small/medium deployment, this service runs on the Infrastructure Web Server and is active-passive with manual failover. Agents - Agents are used for integration with external systems such as Citrix, Hyper-V, and ESXi. Distributed Execution Manager - Performs the orchestration tasks necessary. When using multiple DEM worker servers, Distributed Execution Orchestrators are used to assign, track, and manage the tasks give to the workers.

MSSQL - Tracks infrastructure components, supports Always ON for SQL Server 2016, otherwise a SQL failover cluster is necessary for HA. VMware vrealize Automation deployment model sizing VMware vrealize Automation deployment can be small, medium, or large based on the requirements. The following table provides a summary of the three deployment options: Table 5. vrealize Automation deployment options Small Medium Large Managed Machines 10,000 30,000 50,000 Catalog Items 500 1,000 2,500 Concurrent Provisions 10 50 100 vra Appliances 1 2 2 Windows VMs 1 6 8 MSSQL DB Single server Failover cluster Failover cluster vrealize Business Appliance 1 1 1 Load Balancers 0 3 3

The following images illustrate a large deployment: User Desktop vrealize Automation Appliance DNS Entry Load Balancer vrealize Automation Appliance vrealize Automation Appliance Load Balancer Infrastructure Web DNS Entry Infrastructure Management DNS Entry Load Balancer Infrastructure Web Virtual Machines Infrastructure Management Virtual Machines SQL Database Infrastructure Fabric Distributed Execution Managers Figure 10. vrealize Automation Large Deployment

Scaling the Ready Bundle The solution can be scaled by adding multiple compute nodes (pods) in the customer data center. The Dell Networking Z9100 switch can be used to create a simple yet scalable network. The Z9100 switches serve as the spine switches in the leaf-spine architecture. The Z9100 is a multiline rate switch supporting 10 GbE, 25 GbE, 40 GbE, 50 GbE, and 100 GbE connectivity and can aggregate multiple racks with little or no oversubscription. When connecting multiple racks, using the 40 GbE uplinks from the rack, you can build a large fabric that supports multi-terabit clusters. The density of the Z9100 allows flattening the network tiers and creating an equal-cost fabric from any point to any other point in the network. Figure 11. Multiple Compute PODs scaled out using leaf spine architecture For large domain layer-2 requirements the Extended Virtual Link Trunking (evlt) can be used on the Z9100, as shown in the following figure. The VLT pair formed can scale in terms of hundreds of servers inside multiple racks. Each rack has four 40 GbE links to the core network providing enough bandwidth for all the traffic between each rack. Figure 12. Multiple Compute PODs scaled out using evlt

References Dell EMC SC Series Best Practices with VMware vsphere 5.x 6.x vsphere storage for ESXi 6.0 vrealize Automation Reference Architecture Dell VLT Reference Architecture Dell Configuration Guide for S4048-ON