Dell EMC HPC System for Life Sciences v1.4

Similar documents
Dell EMC Ready Bundle for HPC Digital Manufacturing Dassault Systѐmes Simulia Abaqus Performance

Dell EMC Ready Bundle for HPC Digital Manufacturing ANSYS Performance

GROMACS (GPU) Performance Benchmark and Profiling. February 2016

Emerging Technologies for HPC Storage

LAMMPS-KOKKOS Performance Benchmark and Profiling. September 2015

Altair OptiStruct 13.0 Performance Benchmark and Profiling. May 2015

Design a Remote-Office or Branch-Office Data Center with Cisco UCS Mini

Design a Remote-Office or Branch-Office Data Center with Cisco UCS Mini

Sugon TC6600 blade server

vstart 50 VMware vsphere Solution Specification

Implementing SQL Server 2016 with Microsoft Storage Spaces Direct on Dell EMC PowerEdge R730xd

Dell PowerEdge R720xd with PERC H710P: A Balanced Configuration for Microsoft Exchange 2010 Solutions

Achieve Optimal Network Throughput on the Cisco UCS S3260 Storage Server

NAMD Performance Benchmark and Profiling. January 2015

STAR-CCM+ Performance Benchmark and Profiling. July 2014

Density Optimized System Enabling Next-Gen Performance

A Dell Technical White Paper Dell Virtualization Solutions Engineering

Dell EMC PowerEdge server portfolio: platforms and solutions

HPC Innovation Lab Update. Dell EMC HPC Community Meeting 3/28/2017

HPC and AI Solution Overview. Garima Kochhar HPC and AI Innovation Lab

Intel Select Solutions for Professional Visualization with Advantech Servers & Appliances

Accelerating Microsoft SQL Server 2016 Performance With Dell EMC PowerEdge R740

Microsoft SQL Server in a VMware Environment on Dell PowerEdge R810 Servers and Dell EqualLogic Storage

DESCRIPTION GHz, 1.536TB shared memory RAM, and 20.48TB RAW internal storage teraflops About ScaleMP

Reference Architecture Microsoft Exchange 2013 on Dell PowerEdge R730xd 2500 Mailboxes

Dell EMC PowerEdge and BlueData

Intel Xeon E v4, Windows Server 2016 Standard, 16GB Memory, 1TB SAS Hard Drive and a 3 Year Warranty

Dell PowerVault NX Windows NAS Series Configuration Guide

Microsoft SharePoint Server 2010 on Dell Systems

DELL STORAGE NX WINDOWS NAS SERIES CONFIGURATION GUIDE

Dell PowerEdge Servers Portfolio Guide

DELL EMC READY BUNDLE FOR VIRTUALIZATION WITH VMWARE AND FIBRE CHANNEL INFRASTRUCTURE

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments

OpenFOAM Performance Testing and Profiling. October 2017

Consolidating Microsoft SQL Server databases on PowerEdge R930 server

DELL EMC READY BUNDLE FOR VIRTUALIZATION WITH VMWARE VSAN INFRASTRUCTURE

DELL EMC NX WINDOWS NAS SERIES CONFIGURATION GUIDE

POWEREDGE RACK SERVERS

Architecting Storage for Semiconductor Design: Manufacturing Preparation

Lenovo Database Configuration Guide

REFERENCE ARCHITECTURES OF DELL EMC READY BUNDLE FOR HPC LIFE SCIENCES

John Fragalla TACC 'RANGER' INFINIBAND ARCHITECTURE WITH SUN TECHNOLOGY. Presenter s Name Title and Division Sun Microsystems

Dell EMC Microsoft Exchange 2016 Solution

Intel Xeon E v4, Optional Operating System, 8GB Memory, 2TB SAS H330 Hard Drive and a 3 Year Warranty

Dell Storage NX Windows NAS Series Configuration Guide

Highest Levels of Scalability Simplified Network Manageability Maximum System Productivity

UCS M-Series + Citrix XenApp Optimizing high density XenApp deployment at Scale

The PowerEdge FX Architecture Portfolio Overview Re-inventing the rack server for the data center of the future

Full Featured with Maximum Flexibility for Expansion

Dell PowerVault MD Family. Modular Storage. The Dell PowerVault MD Storage Family

Microsoft SQL Server 2012 Fast Track Reference Architecture Using PowerEdge R720 and Compellent SC8000

IBM Power Advanced Compute (AC) AC922 Server

NetApp High-Performance Storage Solution for Lustre

Accelerating storage performance in the PowerEdge FX2 converged architecture modular chassis

Best Practices for Deploying a Mixed 1Gb/10Gb Ethernet SAN using Dell EqualLogic Storage Arrays

TPC-E testing of Microsoft SQL Server 2016 on Dell EMC PowerEdge R830 Server and Dell EMC SC9000 Storage

SNAP Performance Benchmark and Profiling. April 2014

Enterprise Ceph: Everyway, your way! Amit Dell Kyle Red Hat Red Hat Summit June 2016

Cisco and Cloudera Deliver WorldClass Solutions for Powering the Enterprise Data Hub alerts, etc. Organizations need the right technology and infrastr

Architecting High Performance Computing Systems for Fault Tolerance and Reliability

Dell PowerVault MD Family. Modular storage. The Dell PowerVault MD storage family

SGI Overview. HPC User Forum Dearborn, Michigan September 17 th, 2012

All-Flash High-Performance SAN/NAS Solutions for Virtualization & OLTP

ABySS Performance Benchmark and Profiling. May 2010

LS-DYNA Performance Benchmark and Profiling. April 2015

Dell PowerEdge R730xd Servers with Samsung SM1715 NVMe Drives Powers the Aerospike Fraud Prevention Benchmark

DELL EMC READY BUNDLE FOR VIRTUALIZATION WITH VMWARE AND ISCSI INFRASTRUCTURE

Microsoft SQL Server 2012 Fast Track Reference Configuration Using PowerEdge R720 and EqualLogic PS6110XV Arrays

Cisco UCS C210 M2 General-Purpose Rack-Mount Server

WiRack19 - Computing Server. Wiwynn SV324G2. Highlights. Specification.

Accelerating Hadoop Applications with the MapR Distribution Using Flash Storage and High-Speed Ethernet

Dell EMC PowerEdge R740xd as a Dedicated Milestone Server, Using Nvidia GPU Hardware Acceleration

The PowerEdge M830 blade server

INCREASING DENSITY AND SIMPLIFYING SETUP WITH INTEL PROCESSOR-POWERED DELL POWEREDGE FX2 ARCHITECTURE

An Oracle White Paper December Accelerating Deployment of Virtualized Infrastructures with the Oracle VM Blade Cluster Reference Configuration

DELL EMC READY BUNDLE FOR MICROSOFT EXCHANGE

Altos R320 F3 Specifications. Product overview. Product views. Internal view

SUN CUSTOMER READY HPC CLUSTER: REFERENCE CONFIGURATIONS WITH SUN FIRE X4100, X4200, AND X4600 SERVERS Jeff Lu, Systems Group Sun BluePrints OnLine

VSTOR Vault Mass Storage at its Best Reliable Mass Storage Solutions Easy to Use, Modular, Scalable, and Affordable

CPMD Performance Benchmark and Profiling. February 2014

ISILON ETHERNET BACKEND NETWORK OVERVIEW

Cisco UCS C210 M1 General-Purpose Rack-Mount Server

Upgrade to Microsoft SQL Server 2016 with Dell EMC Infrastructure

GW2000h w/gw175h/q F1 specifications

HPE Scalable Storage with Intel Enterprise Edition for Lustre*

MAKING CLOUD DEPLOYMENTS A REALITY WITH HYPERCONVERGED INFRASTRUCTURE

New Approach to Unstructured Data

DataON and Intel Select Hyper-Converged Infrastructure (HCI) Maximizes IOPS Performance for Windows Server Software-Defined Storage

Cisco HyperFlex HX220c M4 Node

Isilon: Raising The Bar On Performance & Archive Use Cases. John Har Solutions Product Manager Unstructured Data Storage Team

IBM Power AC922 Server

Data Sheet Fujitsu Server PRIMERGY CX250 S2 Dual Socket Server Node

GROMACS Performance Benchmark and Profiling. September 2012

MILC Performance Benchmark and Profiling. April 2013

Running Milestone XProtect with the Dell FS8600 Scale-out File System

CP2K Performance Benchmark and Profiling. April 2011

Cisco UCS C200 M2 High-Density Rack-Mount Server

Lot # 10 - Servers. 1. Rack Server. Rack Server Server

SAN Design Best Practices for the Dell PowerEdge M1000e Blade Enclosure and EqualLogic PS Series Storage (1GbE) A Dell Technical Whitepaper

Deep Learning Performance and Cost Evaluation

Transcription:

Dell EMC HPC System for Life Sciences v1.4 Designed for genomics sequencing analysis, bioinformatics and computational biology Dell EMC Engineering April 2017 A Dell EMC Reference Architecture

Revisions Date April 2017 June 2017 Description Initial release Updated with Isilon THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL ERRORS AND TECHNICAL INACCURACIES. THE CONTENT IS PROVIDED AS IS, WITHOUT EXPRESS OR IMPLIED WARRANTIES OF ANY KIND. Copyright <2016> - <2017> Dell Inc. All rights reserved. Dell and the Dell EMC logo are trademarks of Dell Inc. in the United States and/or other jurisdictions. All other marks and names mentioned herein may be trademarks of their respective companies. 2 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

Table of contents Revisions... 2 Executive summary... 4 Audience... 4 1 Introduction... 5 2 System Design... 6 2.1 Hardware Configuration... 7 2.1.1 Master Node Configuration... 7 2.1.2 Login Node Configuration... 7 2.1.3 Compute Node Configuration... 8 2.1.4 Common Internet File System (CIFS) gateway... 8 2.1.5 Fat Node... 9 2.1.6 Accelerator Node Configuration... 9 2.2 Network Configuration... 9 2.2.1 Management Network... 10 2.2.2 High-Speed Interconnects... 10 2.3 Storage... 12 2.3.1 Dell EMC Ready Bundle for HPC NFS Storage... 12 2.3.2 Dell EMC Ready Bundle for HPC Lustre Storage... 13 2.3.3 Dell EMC Isilon... 14 2.4 Software Configuration... 17 2.4.1 Bright Cluster Manager... 17 2.4.2 BioBuilds... 17 3 Sample Architectures... 18 3.1 Case 1: PowerEdge C6320 compute subsystem with Intel OPA fabric... 18 3.1.1 Solution summary... 18 3.1.2 Cabling information... 19 3.2 Case 2: PowerEdge FC430 compute subsystem with IB FDR fabric... 19 3.2.1 Solution summary... 19 3.2.2 Cabling information... 21 4 Conclusion... 22 5 References... 23 3 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

Executive summary In October 2015, Dell Technologies introduced Genomic Data Analysis Platform (GDAP) v2.0 to answer the growing necessities of rapid genomic analysis due to the availability of next-generation sequencing technologies. Upon the successful implementation of GDAP v2.0, which is capable of processing up to 133 genomes per day while consuming 2 kilowatt-hour (kwh) per genome, we started to explore the life science domains beyond genomics. In addition to Next Generation Sequencing (NGS), other lab equipment used in the life sciences also produce much richer data than before, such as mass spectrometers, electron microscopes and so on. Often the data from various pieces of lab equipment are image based and need heavy computational work to put the raw images into a useful format. However, the data can be highly variable in workloads, types of analysis and tools, and make it difficult to optimize systems. Hence, Dell EMC offers the Dell EMC HPC System for Life Sciences v1.4, a high performance computing environment architected for flexibility and optimized to address various analysis workflows. Audience This document is intended for organizations interested in accelerating genomic research with advanced computing and data management solutions. System administrators, solution architects, and others within those organizations constitute the target audience. 4 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

1 Introduction The Dell EMC HPC System for Life Sciences is a tested, tuned and purpose-built platform, leveraging the most relevant of Dell EMC s high performance computing (HPC) line of products and best-in-class partner products. It encompasses all the hardware resources required for various life sciences data analysis, while providing an optimal balance of compute density, energy efficiency, and performance from Dell EMC s Enterprise server line-up. The Dell EMC HPC System for Life Sciences provides higher flexibility for the solutions. A platform is available in five variants, depending on the cluster interconnects selected, which can be either 10 Gigabit Ethernet (GbE), Intel Omni-Path (OPA), InfiniBand (IB) EDR or IB FDR. In this version, the following options are available: PowerEdge C6320 compute subsystem with Intel OPA fabric (1) PowerEdge C6320 compute subsystem with IB EDR fabric PowerEdge C6320 compute subsystem with 10 GigE fabric PowerEdge FX2 compute subsystem with IB FDR fabric PowerEdge FX2 compute subsystem with 10 GigE fabric The solutions are similar for IB, Intel OPA and 10 GigE versions, with just a few changes in the switching infrastructure and network adapters. These differences are outlined in the Network Components section. The solution ships in a deep 48U rack enclosure, which was chosen because of its ease of mounting PDUs and for a better cable management. This rack houses the management, compute, storage and networking modules of the solution. Also, software modules installed on management hardware will make deploy, manage, and maintain the cluster simple and easier. Not only can each sub-field of life sciences benefit immensely from the massive growth in HPC power, but also life sciences as whole can use HPC as an essential tool to integrate all the necessary data, biochemistry, genomics, proteomics, biophysics, and anatomy of cells, tissues, and organs into a single frame of work. Due to the highly diverse nature of life sciences research, scientists have different needs and face a difficult challenge picking a system that will work for everyone, one system does not fit all. Hence, Dell EMC HPC System for Life Sciences was introduced as a solution to the challenge in life sciences with a modular approach. It was designed to be a plug-and-play turnkey solution so that researchers can spend more time working on matters in their domain, rather than concerning themselves with the computer science aspect of getting the system to function, which deals with cluster deployment and maintenance. This reference architecture describes working solutions for diverse life sciences applications, including molecular dynamics simulation, with a flexible architecture, as well as improvement to the performance of the genomics data analysis platform. 5 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

2 System Design The first step in designing the system is to decide upon the following four basic considerations: Type of workload o Genomics/NGS data analysis only o General purpose and Genomics/NGS data analysis o Adding molecular dynamics simulation capacity o Adding a large memory machine to handle memory intensive workloads such as De Novo assembly and statistical analysis on epigenomics data or metagenomics data Parameter for sizing o Number of compute nodes o Genomes per day to be analyzed o NGS data size from an individual sample Form factor of servers o 2U shared infrastructure of high density that can host 4 compute nodes in one chassis (C6320) Suitable for NGS data processing as well as some statistical analysis, machine learning and simulations o 2U shared infrastructure of very high density that can host 8 compute nodes in one chassis (FC430) Suitable for NGS data processing such as whole genome sequencing and RNA sequencing data pipeline The limitations on the choice of CPUs and the maximum memory capacity compare to C6320 o 4U rack mount servers (R930) Up to 9TB memory capacity makes this server suitable for De Novo assembly with deep sequencing data o 1U rack mount servers (R430) Suitable for management modules such as master node, login node and CIFS gateway o 2U rack mount servers that can host up to 4 accelerators per node (C4130) Suitable for molecular dynamics simulations by using Amber, NAMD and HOOMDblue Types of interconnect: All four options are available for any server except FC430. Mellanox ConnectX-3 (IB FDR) is only high speed interconnect option for FC430. The choice of interconnects is directly related to the number of samples and the size of samples. The bottom line is that we recommend high speed interconnect when there are roughly more than 30 samples (10x whole human genomes) need to be processed in a day. o Intel Omni-Path Host Fabric Interface (HFI) 100 series card o Mellanox ConnectX-4, Single Port, VPI EDR, QSFP28 Adapter o Mellanox ConnectX-3, Single Port, VPI FDR, QSFP+ Adapter, Low Profile o Generic 10GbE Following are the technical specifications of the servers that are considered for the Dell EMC HPC System for Life Sciences. 6 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

2.1 Hardware Configuration There are several considerations when selecting the servers for compute node, login node, fat node and accelerator node, which are the components of the Dell EMC HPC System for Life Sciences. The 1U form factor PowerEdge R430 is the recommended server for master node, login node and CIFS gateway since the number of users expected is a lot less than the users for a generic purpose system. Also, it is possible to add more servers to handle growing number of users. The 4U form factor PowerEdge R930 is recommended as the appropriate server for a fat node. The memory capacity on this server, up to 9TB can cover the most peak memory consumptions of De Novo assembly applications. And, the 1U form factor PowerEdge C4130 is recommended to host accelerators due to the high density of GPUs. For the selection of compute nodes, there are two options available; PowerEdge FC430 or PowerEdge C6320. These servers are not offered as standard off-the-shelf components. All the servers are customizable based on the needs, and the configurations of servers mentioned below are the best fit for life sciences applications. 2.1.1 Master Node Configuration The master node is responsible for managing the compute nodes and optimizing the overall compute capacity. Hence it is also known as a head node. Usually, the master and login nodes are the only nodes that communicate with the outside world and act as a middle point between the actual cluster and the outside network. A master node is also referred to as the front end node because it provides the point of access and testing of the programs you want to run on a cluster system. The master node controls the OS imaging and administration of the cluster. One master node is required, and at least two master nodes are recommended for a high availability configuration. The configuration of Power Edge R430, which is the recommended server for a master node, is provided below: PowerEdge R430, 2-Socket server with Intel Xeon E5-2680 v4 processors 8 x 16GB RDIMM, 2400MT/s, Dual Rank o DDR4 DIMMs at up to 2400MT/s o 12 DIMM slots: 2GB/4GB/8GB/16GB/32GB RAID 1 for H330/H730/H730P (2 HDDs or SSDs) PERC H730 Integrated RAID Controller, 1GB Cache 2 x 500GB 7.2K RPM NLSAS 6Gbps 2.5in Hot-plug Hard Drive,3.5inHYB CARR,13G Interconnect options idrac8 Enterprise with OpenManage Essentials DVD+/-RW SATA Internal for 4HD Chassis 2.1.2 Login Node Configuration Login nodes are where you login for access to the compute nodes. These nodes will be used for all access, compilation and job submission needs. Login node(s) and high availability are optional. The PowerEdge R430 is the recommended server; for its configuration please refer to section 2.1.1. 7 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

2.1.3 Compute Node Configuration A higher number of cores is more important than CPU speed for the applications in NGS data processing. The PowerEdge FC430 offers the ultimate density and is the choice for NGS data processing. Another option for customers who also use other applications in addition to NGS applications is the PowerEdge C6320. PowerEdge C6320 1U, Half-width PowerEdge C6300 Enclosure, 2.5 in Hard Drives with Backplane: houses four PowerEdge C6320 servers PowerEdge C6320, 2-Socket server with Intel Xeon E5-2697 v4 processors 1TB 7.2K RPM SATA 6Gbps 2.5in Cabled Hard Drive 8 x 32GB RDIMM, 2400MT/s, Dual Rank o DDR4 DIMMs at up to 2400MT/s o 16 DIMM slots: 4GB/8GB/16GB/32GB/64GB idrac8 Enterprise PowerEdge FC430 1U, Quarter-width PowerEdge FX2 chassis: houses 8 FC430 servers PowerEdge FC430, 2-Socket server with Intel Xeon E5-2680 v4 processors 1.8" Backplane with 1 Hard Drive and 1 IB Port PowerEdge FC PCIE Mezzanine Adapter 8 x 32GB RDIMM, 2133MT/s, Dual Rank 200GB Solid State Drive usata Mix Use Slim MLC 6Gbps 1.8inHot-plug Drive Interconnect: Mellanox ConnectX-3 FDR Mezzanine adapter idrac8 Enterprise PowerEdge C6320p - 1U, Half-width Intel Xeon Phi 7230. 64 cores, 1.3 GHz 96 GB at 2400 MT/s Internal 1.8 240GB Solid State Drive 6Gbps 2.1.4 Common Internet File System (CIFS) gateway A Dell EMC PowerEdge R430 is used as the CIFS gateway for transferring data generated by the nextgeneration sequencing machines into the storage. The configuration for CIFS gateway is provided below: PowerEdge R430 Server, 2-Socket server with Intel Xeon E5-2680 v4 processors 6 x 8GB RDIMM, 2400 MT/s, Dual Rank o DDR4 DIMMs at up to 2400MT/s o 12 DIMM slots: 2GB/4GB/8GB/16GB/32GB RAID 1 for H330/H730/H730P (2 HDDs or SSDs) PERC H730 Integrated RAID Controller, 1GB Cache 2 x 500GB 7.2K RPM NLSAS 6Gbps 2.5in Hot-plug Hard Drive,3.5inHYB CARR,13G Intel Ethernet X540 DP 10GBASE-T Server Adapter, Low Profile 8 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

idrac8, Basic DVD+/-RW SATA Internal for 4HD Chassis 2.1.5 Fat Node A fat node is a node with a large memory. The PowerEdge R930 is offered as an optional fat node with up to 3TB of memory. All the features benefit de novo genome sequence assemblies like Velvet. The size of the data on which these applications operate is massive. Hosting this data in memory with 64 cores eliminates the overhead caused by interconnects, disk look-ups and swapping, resulting in faster time-to-results. Power Edge R930, 4-Socket server with Intel Xeon E7-8890 v4 processors 96 DIMM slots, DDR4 memory 8GB/16GB/32/64GB DDR4 up to 2400MT/s, maximum 6 TB Up to 24 x 2.5 hot-plug 12Gb/6Gb SAS HDD or SAS/SATA SSD Up to 8 front-accessible Express Flash NVMe PCIe SSD PCI Express Generation 3 slots idrac8, Enterprise with OpenManage Essentials 2.1.6 Accelerator Node Configuration Accelerators are used to speed up computationally intensive applications, such as Amber, HOOMD-blue and NAMD for molecular dynamics simulations. PowerEdge C4130 offers five different NVIDIA GPU configurations. PowerEdge C4130-1U - Up to 4 accelerators per node PowerEdge C4130, 2-socket server with Intel Xeon E5-2690 v4 processors 8 x 16GB RDIMM, 2400MT/s, Dual Rank o 16 DIMM slots, DDR4 Memory o 4GB/8GB/16GB/32GB DDR4 up to 2400MT/s Up to 2 x 1.8 SATA SSD boot drives Optional 96-lane PCIe 3.0 switch for certain accelerator configurations idrac8, Dell EMC OpenManage Essentials Available GPUs 4 x NVIDIA K80 4 x NVIDIA P100 - two versions are available SXM2 module for NVLink-optimized servers provide the best performance and strong-scaling for hyperscale and HPC data centers running applications that scale to multiple GPUs, such as deep learning. PCIe allows HPC data centers to deploy the most advanced GPUs within PCIe-based nodes to support a mix of CPU and GPU workloads. 2.2 Network Configuration The Dell EMC HPC System for Life Sciences is available in Intel OPA and two IB variants. There is also a Dell EMC Network S3048-ON GbE switch used in both configurations, whose purpose is described here. In 9 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

one of the IB variants, the Dell EMC PowerEdge FC430 sleds have 2:1 blocking FDR connectivity to the top of rack FDR switch. The other IB variant is 1:1 non-blocking EDR network for C6320. 2.2.1 Management Network Management traffic typically communicates with the Baseboard Management Controller (BMC) on the compute nodes using IPMI. The management network is used to push images or packages to the compute nodes from the master nodes and for reporting data from client to the master node. Dell EMC Networking S3048-ON and PowerConnect 2848 Switch are considered for management network Dell EMC Network S3048-ON High density, 1U 48-port 1000BASE-T + with four 10GbE uplinks, non-blocking rate-line performance, featuring the Open Networking Install Environment (ONIE). The port assignment of the Dell EMC Networking S3048-ON switch for the Intel OPA or IB versions of the solution is as follows. Ports 01-04 and 27 52 are assigned to the cluster s private management network to be used by Bright Cluster Manager connecting master, login, CIFS gateway, and compute nodes. The PowerEdge C6320 server s Ethernet and idrac constitute a majority of these ports. Ports 06 09 are used for the private network associated with NSS7.0-HA. The rest of the ports, 05 and 12 26 are allocated to the Lustre solution for its private management network Port 10 and 11 are used for the PDUs. Figure 1: Dell EMC Networking S3048-ON switch Note: Installation of four SFP+ to RJ45 transceivers in Dell EMC Networking S3048-ON switch ports 49-52 is required. 2.2.2 High-Speed Interconnects In high performance computing, application performance depends on the number of CPU/GPU cores, memory, interconnect, storage performance, and so on. For a server to perform better, lower latency and higher bandwidth is needed for these systems to communicate with each other. The type of network chosen for computational traffic depends upon the latency, bandwidth, packet size at peak bandwidth, and the message rate. The types of computational network and the switch through which it is offered are listed below: 10 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

Table 1: Interconnect switches Type of Interconnect Fourteen Data Rate (FDR) Extended Data Rate (EDR) 10Gb Ethernet Intel Omni-Path (OPA) Switch Mellanox SX6036 Mellanox SB7790 S4820-T, N4032-F, S6000, Z9500 Dell EMC Networking H1048-OPF Mellanox SX6036 36 FDR (56Gb/s) ports in a switch Port-to-port latency 200ns Up to 4Tb/s aggregate switching capacity 19 rack mountable chassis, 1U with optional redundant power supplies and fan units On-board SM for fabrics up to 648 nodes. Mellanox SB7790 36 EDR (100Gb/s) ports in a switch Port-to-port latency 90ns Up to 7Tb/s aggregate switching capacity 19 rack mountable chassis, 1U with optional redundant power supplies and fan units On-board SM for fabrics up to 2k nodes. Dell EMC Networking S6000 1U high-density 10/40GbE ToR switch with 32 ports of 40GbE (QSFP+) or 96 ports of 10GbE and eight ports of 40GbE or 104 ports of 10GbE. Up to 2.56Tbps of switching capacity. Dell EMC Networking Z9500 High-density 3U with 132 ports of 40GbE (528 ports of 1/10GbE using breakout cables) Dell EMC Networking S4820T 1U high performance ToR switch provides (48) 1/10G BASE-T ports that support 100Mb/1 Gb/10Gb and four 40GbE QSFP+ uplinks. Each 40GbE QSFP+ uplink can be broken out into four 10GbE ports using breakout cables Dell EMC Networking N4032F SFP Switch 24x 10GbE SFP + auto-sensing (10Gb/1Gb) fixed ports 11 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

Up to 32 10GbE ports using breakout cables and optional QSFP+ module One hot swap expansion module bay Dual hot-swappable redundant power supplies (460W) 2.3 Storage The performance requirements of HPC environments with ever-larger compute clusters have placed unprecedented demands on the storage infrastructure. The storage infrastructure consists of the following components: In general, we recommend Dell EMC Ready Bundle for HPC Lustre Storage with HPC NFS Storage as a default storage when a high speed scratch space is required. For Dell EMC Isilon can serve as a default storage as well as a high speed scratch space. Dell EMC Ready Bundle for HPC NFS Storage (NSS 7.0-HA) Dell EMC Ready Bundle for HPC Lustre Storage Dell EMC Isilon 2.3.1 Dell EMC Ready Bundle for HPC NFS Storage NSS7.0 HA is designed to enhance the availability of storage services to the HPC cluster by using a pair of Dell EMC PowerEdge servers and PowerVault storage arrays along with Red Hat HA software stack. The HA cluster consists of a pair of Dell EMC PowerEdge servers and a network switch. The two PowerEdge servers have shared access to disk-based Dell EMC PowerVault storage in a variety of capacities, and both are directly connected to the HPC cluster by using OPA, IB or 10GbE. The two servers are equipped with two fence devices: idrac8 Enterprise, and an APC Power Distribution Unit (PDU). If failures such as storage disconnection, network disconnection, and system stopping from functioning, etc., occur on one server, the HA cluster will failover the storage service to the healthy server with the assistance of the two fence devices; and also ensure that the failed server does not return to life without the administrator s knowledge or control. We previously tested the NSS7.0-HA functionality and performance, and those details are published as a whitepaper (2). Figure 2 shows the configuration of the test bed, and a summary of the test bed is listed below. A 32-node HPC compute cluster (also known as the clients ) was used to provide I/O network traffic for the test bed. A pair of Dell EMC PowerEdge R730 servers were configured as an active-passive HA pair and function as a NFS server for the HPC compute cluster. Both NFS servers were connected to a shared Dell EMC PowerVault MD3460 storage enclosure extended with one Dell EMC PowerVault MD3060e storage enclosure (a 480 TB solution with the two PowerVault MD storage arrays is shown in Figure 2) at the back-end. The user data is stored on an XFS file system created on this storage. The XFS file system was exported to the clients by using NFS. The NFS servers were connected to the clients by using the public network. This network was Intel OPA. For the HA functionality of the NFS servers, a private 1 Gigabit Ethernet network was configured to monitor server health and heartbeat, and to provide a route for the fencing operations by using a Dell EMC Networking 3048-ON Gigabit Ethernet switch. 12 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

Power to the NFS servers was provided by two APC switched PDUs on two separate power buses. Figure 2 NSS7.0-HA configuration 2.3.2 Dell EMC Ready Bundle for HPC Lustre Storage The Dell EMC Ready Bundle for HPC Lustre Storage, referred to as Dell EMC HPC Lustre Storage, is designed for academic and industry users who need to deploy a fully-supported, easy-to-use, highthroughput, scale-out, and cost-effective parallel file system storage solution. The Ready Bundle uses the Intel Enterprise Edition (EE) for Lustre software v.3.1 (3). It is a scale-out storage bundle capable of providing a high performance and high availability storage system. Utilizing an intelligent, extensive and intuitive management interface, the Intel Manager for Lustre (IML), the solution greatly simplifies deploying, managing and monitoring all of the hardware and storage system components. It is easy to scale in capacity, performance, or both, thereby providing a convenient path to grow in the future. Figure 3 shows the relationship of the MDS, MDT, MGS, OSS and OST components of a typical Lustre configuration. Clients in the figure are the HPC cluster s compute nodes. The solution is 120 drive system configuration (60 drives per PowerVault MD3460) with 12 OSTs. Each OST consists of 10x 4TB drives. Total raw storage size is 480TB. Figure 3 shows an overview of a Dell HPC Lustre Storage with Intel EE for Lustre system in a 240-drive configuration. The detailed information including configuration, performance, and other technical help can be found in a whitepaper we published (3). Note the three major components: Management Server (IML) Metadata Server pair (MDS) 13 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

Object Storage Server pair (OSS) Figure 3 Lustre-based storage solution components Figure 4 Dell EMC Ready Bundle for HPC Lustre Storage Components Overview 2.3.3 Dell EMC Isilon For Dell EMC HPC System for Life Sciences, we recommend three grades of Isilon storage nodes, the X410, F800, and H600 SAS. Dell EMC has combined the extreme performance of flash technology with the proven scalability, true multi-protocol access and security of Isilon OneFS operating system simplifying management process significantly by providing the single file system, single-volume architecture, and automation features. The Dell EMC Isilon X410 allows you to easily scale from a 3-node cluster with 108 TB of capacity to a 144- node cluster with over 20 PB, all in a single file system. Management stays simple at any scale. Take 14 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

advantage of high concurrency and streaming performance with the Isilon X410. Support a wide range of workloads, including Hadoop analytics, large-scale home directories, medical imaging, life sciences research, digital media, and electronic design. Figure 5 shows a typical network configuration for Isilon X410. Figure 5 Network configuration of x410 Formerly known as Project Nitro, the highly dense, bladed-node architecture of F800 provides four nodes within a single 4U chassis with capacity options ranging from 92TB to 924TB per chassis (Figure 6 illustrates network configuration for the test). Available in several configurations, F800 units can be combined into a single cluster that provides up to 92.4PB of capacity and over 1.5TB/s of aggregate performance. It is designed for a wide range of next-generation applications and unstructured workloads that require extreme NAS performance including, 4K streaming of data, genomic sequencing, electronic design automation, and near real-time analytics. It can also be deployed as a new cluster, or can seamlessly integrate with existing Isilon clusters to accelerate the performance of an enterprise data lake and lower the overall total cost of ownership (TCO) of a multi-tiered all-flash and high capacity SATA solution. Powered by the OneFS operating system, this new offering from Dell EMC is designed to provide the essential capabilities that enterprises require to modernize IT: extreme performance, massive scalability, operational flexibility, increased efficiency, and enterprise-grade data protection and security. Harness the power of Dell EMC Isilon H600 hybrid scale-out NAS to enable a wide range of high-performance computing (HPC) file workloads on a single platform (a typical network configuration shown in Figure 7). It is equipped with 120 SAS drives, SSD caching, and built-in multiprotocol capabilities. 15 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

Figure 6 Network configuration of F800 Figure 7 Network configuration of H600 16 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

2.4 Software Configuration Along with the hardware components, the solution includes the following software components: Bright Cluster Manager BioBuilds (4) 2.4.1 Bright Cluster Manager Bright Computing is a commercial software that provides comprehensive software solutions for deploying and managing HPC clusters, big data clusters and OpenStack in the data center and in the cloud. Bright cluster Manager can be used to deploy complete clusters over bare metal and manage them effectively. Once the cluster is up and running, the graphical user interface monitors every single node and reports if it detects any software or hardware events. 2.4.2 BioBuilds BioBuilds is a well maintained, versioned, and continuously growing collection of open-source bio-informatics tools from Lab7 (5). They are prebuilt and optimized for a variety of platforms and environments. BioBuilds solves most software challenges faced by the life sciences domain. Imagine a newer version of a tool being released. Updating it may not be straight forward and would probably involve updating all the dependencies the software has as well. BioBuilds includes the software and its supporting dependencies for ease of deployment. Using BioBuilds among all the collaborators can ensure reproducibility since everyone is running the same version of the software. In short, it is a turnkey application package. 17 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

3 Sample Architectures 3.1 Case 1: PowerEdge C6320 compute subsystem with Intel OPA fabric 3.1.1 Solution summary This solution is nearly identical to the solutions with IB EDR and 10 GbE versions, except for a few changes in the switching infrastructure and network adapters. As shown in Figure 8, this solution uses one 48U rack and requires extra deep enclosure. Bright Cluster Manager is the default tool and a proprietary software solution Figure 8 Dell EMC HPC System for Life Sciences with PowerEdge C6320 rack servers and Intel OPA fabric 18 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

stack from Bright Computing. Along with automated deployment and provisioning features for the HPC cluster, it incorporates powerful and intelligent management and monitoring capabilities. The graphical user interface provides advanced configuration options. Table 2: Hardware summary Rack Master nodes Login nodes CIFS gateway Compute nodes High speed interconnect switch Management switch Primary File System Parallel File System 1x 48U 600mm wide x 1200mm Deep 2x PowerEdge R430 2x PowerEdge R430 1x PowerEdge R430 24x PowerEdge C6320 on 6x PowerEdge C6300 Enclosure 1x Dell EMC Networking H 1048-OPF 1x Dell EMC Networking S3048-ON Dell EMC NSS7.0-HA configuration (240TB) Dell EMC Ready Bundle for HPC Lustre Storage (480TB raw space) 3.1.2 Cabling information Management interconnect and high speed interconnect require 32 cables each. Dell EMC NSS7.0-HA needs two cables for both management and high speed interconnect due to the HA configuration. Table 3 Required cables for Case 1, singe rack design Cable types Length Quantity High speed Copper cables 3M 24 interconnect cables 2M 8 Management cables 1GbE RJ-45 CAT6 10ft 24 5ft 8 3.2 Case 2: PowerEdge FC430 compute subsystem with IB FDR fabric 3.2.1 Solution summary The FC430 solution with IB FDR interconnect is nearly identical to the 10 GbE version, except for a few changes in the switching infrastructure, and network adapters and have 2:1 blocking FDR connectivity to the top of rack FDR switch (Figure 9). The port assignment of the Dell EMC Networking S3048-ON switch for the Intel OPA or IB versions of the solution is as follows: Ports 01-04 and 27 52 are assigned to the cluster s private management network to be used by Bright Cluster Manager connecting master, login, CIFS gateway and compute nodes. The PowerEdge C6320 server s Ethernet and idrac constitute a majority of these ports. 19 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

Ports 06 09 are used for the private network associated with NSS7.0-HA. Figure 9 Dell EMC HPC System for Life Sciences with PowerEdge FC430 rack servers with IB FDR fabric The rest of the port 05 and ports 12 26 are allocated to the Dell EMC Ready Bundle for HPC Lustre Storage for its private management network Port 10 and 11 are used for the PDUs. For the 10GbE configuration, the deployment and management of the cluster is done over the 10 Gigabit Ethernet network by using the Dell EMC Networking S4820T switch. So, the first virtual LAN on the S3048- ON, from ports 0 16, is not used. The other two virtual LANs are still used for the same purpose as in the IB FDR configuration. 20 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

Table 4 Hardware summary Rack Master nodes Login nodes CIFS gateway Compute nodes High speed interconnect switch Management switch Primary File System Parallel File System 1x 48U 750mm wide x 1200mm Deep 2x PowerEdge R430 2x PowerEdge R430 1x PowerEdge R430 40x PowerEdge FC430 on 5x PowerEdge FX2 chassis 3x MLNX SX6036 1x Dell EMC Networking S3048-ON Dell EMC NSS7.0-HA configuration (240TB) Dell EMC Ready Bundle for HPC Lustre Storage (480TB raw space) 3.2.2 Cabling information Table 5 Required cables for Case 2, single rack design High speed interconnect cables Cable types Length Quantity Fiber optic cables 3 M 40 Copper cables 2 M 8 Management cables 1GbE RJ-45 CAT6 10 ft 15 5 ft 10 1ft 5 21 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

4 Conclusion The Dell EMC HPC System for Life Sciences provides tested working solutions for diverse life science applications, especially targeted on NGS workload, molecule dynamics simulations, 3-D protein structure reconstruction, and so on. It is a cost-effective solution that is flexible enough to be used for various life science applications. The configurations presented in this guide, which are derived from the Dell EMC HPC System for Life Sciences, are just a starting point. This paper is intended to provide the minimum architecture that could achieve the targeted NGS workload, informed decision making and increased efficiency. It is recommended to reach out to your Dell EMC technical sales representative to develop a tailored HPC plan including Dell EMC Ready Bundles for HPC. 22 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)

5 References 1. Dell EMC HPC System for Life Science V1.1. [Online] http://en.community.dell.com/techcenter/blueprints/blueprint_for_hpc/m/mediagallery/20443473. 2. Dell HPC NFS Storage Solution - High Availability Configuration White Paper (NSS 7.0-HA). [Online] http://en.community.dell.com/techcenter/blueprints/blueprint_for_hpc/m/mediagallery/20442675. 3. Dell HPC Lustre Storage solution with Intel Omni-Path. [Online] http://en.community.dell.com/techcenter/blueprints/blueprint_for_hpc/m/mediagallery/20442903. 4. BioBuilds. [Online] https://biobuilds.org/. 5. Lab 7. [Online] https://www.lab7.io/. 23 Dell EMC HPC System for Life Sciences v1.4 Document ID version (optional)