Benefits of 25, 40, and 50GbE Networks for Ceph and Hyper- Converged Infrastructure John F. Kim Mellanox Technologies

Similar documents
N V M e o v e r F a b r i c s -

16GFC Sets The Pace For Storage Networks

SNIA Developers Conference - Growth of the iscsi RDMA (iser) Ecosystem

Accelerating Ceph with Flash and High Speed Networks

InfiniBand Networked Flash Storage

VM Migration Acceleration over 40GigE Meet SLA & Maximize ROI

The Best Ethernet Storage Fabric

By John Kim, Chair SNIA Ethernet Storage Forum. Several technology changes are collectively driving the need for faster networking speeds.

VMware Virtual SAN Technology

Accelerating Hadoop Applications with the MapR Distribution Using Flash Storage and High-Speed Ethernet

How to Network Flash Storage Efficiently at Hyperscale. Flash Memory Summit 2017 Santa Clara, CA 1

QNAP 25GbE PCIe Expansion Card. Mellanox SN GbE/100GbE Management Switch. QNAP NAS X 25GbE card X 25GbE/100GbE switch

NVMe over Fabrics. High Performance SSDs networked over Ethernet. Rob Davis Vice President Storage Technology, Mellanox

Delivering HCI with VMware vsan and Cisco UCS

WHITE PAPER THE CASE FOR 25 AND 50 GIGABIT ETHERNET WHITE PAPER

Hyper-converged infrastructure with Proxmox VE virtualization platform and integrated Ceph Storage.

TITLE. the IT Landscape

Cavium FastLinQ 25GbE Intelligent Ethernet Adapters vs. Mellanox Adapters

The Impact of Hyper- converged Infrastructure on the IT Landscape

Low-Overhead Flash Disaggregation via NVMe-over-Fabrics

SurFS Product Description

Annual Update on Flash Memory for Non-Technologists

Low-Overhead Flash Disaggregation via NVMe-over-Fabrics Vijay Balakrishnan Memory Solutions Lab. Samsung Semiconductor, Inc.

Implementing SQL Server 2016 with Microsoft Storage Spaces Direct on Dell EMC PowerEdge R730xd

Ceph in a Flash. Micron s Adventures in All-Flash Ceph Storage. Ryan Meredith & Brad Spiers, Micron Principal Solutions Engineer and Architect

Open storage architecture for private Oracle database clouds

Learn Your Alphabet - SRIOV, NPIV, RoCE, iwarp to Pump Up Virtual Infrastructure Performance

Storage Protocol Offload for Virtualized Environments Session 301-F

The Fastest And Most Efficient Block Storage Software (SDS)

Design a Remote-Office or Branch-Office Data Center with Cisco UCS Mini

Design a Remote-Office or Branch-Office Data Center with Cisco UCS Mini

All-Flash High-Performance SAN/NAS Solutions for Virtualization & OLTP

The NE010 iwarp Adapter

Cisco HyperFlex HX220c M4 and HX220c M4 All Flash Nodes

Cisco HyperFlex HX220c M4 Node

2014 VMware Inc. All rights reserved.

Hyper-converged storage for Oracle RAC based on NVMe SSDs and standard x86 servers

Application Acceleration Beyond Flash Storage

Emulex LPe16000B 16Gb Fibre Channel HBA Evaluation

NVMe Over Fabrics (NVMe-oF)

DataON and Intel Select Hyper-Converged Infrastructure (HCI) Maximizes IOPS Performance for Windows Server Software-Defined Storage

Converged Platforms and Solutions. Business Update and Portfolio Overview

Meltdown and Spectre Interconnect Performance Evaluation Jan Mellanox Technologies

G2M Research Fall 2017 NVMe Market Sizing Webinar

The Future of High Performance Interconnects

ETHERNET OPTICS TODAY: 25G NRZ

Audience This paper is targeted for IT managers and architects. It showcases how to utilize your network efficiently and gain higher performance using

Red Hat Ceph Storage and Samsung NVMe SSDs for intensive workloads

VxRail: Level Up with New Capabilities and Powers GLOBAL SPONSORS

IBM Emulex 16Gb Fibre Channel HBA Evaluation

The next step in Software-Defined Storage with Virtual SAN

Flash Storage with 24G SAS Leads the Way in Crunching Big Data

All-Flash High-Performance SAN/NAS Solutions for Virtualization & OLTP

Сетевые технологии для систем хранения данных

Pivot3 Acuity with Microsoft SQL Server Reference Architecture

Next Generation Storage Networking for Next Generation Data Centers. PRESENTATION TITLE GOES HERE Dennis Martin President, Demartek

HyperScalers JetStor appliance with Raidix storage software

Cisco HyperFlex HX220c M4 and HX220c M4 All Flash Nodes

iser as accelerator for Software Defined Storage Rahul Fiske, Subhojit Roy IBM (India)

Dell EMC Hyper-Converged Infrastructure

Active System Manager Release 8.2 Compatibility Matrix

FlashGrid Software Enables Converged and Hyper-Converged Appliances for Oracle* RAC

2017 Storage Developer Conference. Mellanox Technologies. All Rights Reserved.

Silicon Photonics and the Future of Optical Connectivity in the Data Center

Ethernet. High-Performance Ethernet Adapter Cards

GUIDE. Optimal Network Designs with Cohesity

Session 201-B: Accelerating Enterprise Applications with Flash Memory

Dell PowerEdge R730xd Servers with Samsung SM1715 NVMe Drives Powers the Aerospike Fraud Prevention Benchmark

Dell EMC. VxBlock and Vblock Systems 540 Architecture Overview

Enterprise Ceph: Everyway, your way! Amit Dell Kyle Red Hat Red Hat Summit June 2016

Interface Trends for the Enterprise I/O Highway

Modern hyperconverged infrastructure. Karel Rudišar Systems Engineer, Vmware Inc.

Lot # 10 - Servers. 1. Rack Server. Rack Server Server

All-NVMe Performance Deep Dive Into Ceph + Sneak Preview of QLC + NVMe Ceph

RoCE vs. iwarp Competitive Analysis

ARISTA: Improving Application Performance While Reducing Complexity

EXPERIENCES WITH NVME OVER FABRICS

Choosing the Best Network Interface Card for Cloud Mellanox ConnectX -3 Pro EN vs. Intel XL710

Trends in Worldwide Media and Entertainment Storage

SOFTWARE-DEFINED BLOCK STORAGE FOR HYPERSCALE APPLICATIONS

Accelerating Workload Performance with Cisco 16Gb Fibre Channel Deployments

Analyst Perspective: Test Lab Report 16 Gb Fibre Channel Performance and Recommendations

HP BladeSystem c-class Ethernet network adapters

NVM Express over Fabrics Storage Solutions for Real-time Analytics

IBM Hortonworks Design Guide 14-Sep-17 v1

Dell EMC Hyper-Converged Infrastructure

DELL EMC VXRACK FLEX FOR HIGH PERFORMANCE DATABASES AND APPLICATIONS, MULTI-HYPERVISOR AND TWO-LAYER ENVIRONMENTS

"Software-defined storage Crossing the right bridge"

Vblock Architecture. Andrew Smallridge DC Technology Solutions Architect

128GFC. Fiber Channel Leading Charge for Higher Speeds Skip Jones Date: Tuesday, October 15 Time: 2:10pm 2:50pm Location: Room 102B

The Impact of Hyper- converged Infrastructure on the IT Landscape

The New Scale-Out Storage

Software Defined Storage at the Speed of Flash. PRESENTATION TITLE GOES HERE Carlos Carrero Rajagopal Vaideeswaran Symantec

HGST: Market Creator to Market Leader

Creating an agile infrastructure with Virtualized I/O

NVMe over Universal RDMA Fabrics

Cisco UCS Virtual Interface Card 1225

THE ZADARA CLOUD. An overview of the Zadara Storage Cloud and VPSA Storage Array technology WHITE PAPER

Introduction to Infiniband

NVM Express Awakening a New Storage and Networking Titan Shaun Walsh G2M Research

Transcription:

Benefits of 25, 40, and 50GbE Networks for Ceph and Hyper- Converged Infrastructure John F. Kim Mellanox Technologies

Storage Transitions Change Network Needs Software Defined Storage Flash Storage Storage Arrays Server-SAN Hard Disk Drive Flash SSDs Major storage technology transitions Software Defined Storage SSDs replacing hard disk drives Hyper-converged and Cloud All require faster networks, and efficient transport More east-west traffic Faster storage needs faster networks Higher bandwidth needs better offloads

Fibre Channel in a Slow Decline FC being outgrown Ethernet Block & NAS Distributed Storage Big Data Hyper-converged 4K video editing FC-SANs do not go away But they don t grow New storage is Ethernet

25/50/100 GbE Rapid Growth Projected 25GbE to grow rapidly Fastest ramp in history Passes 40GbE in 2019 50 & 100GbE Also Grow Cloud/Web 2.0 customers Large enterprises $2.0 $1.5 $1.0 $0.5 $0.0 Ethernet Adapter Market ($B) $1.7 $1.5 $1.6 $1.3 $1.1 2015E 2016E 2017E 2018E 2019E 10GbE 25 GbE 40 GbE 50 GbE 100 GbE Ethernet Top of Rack Switch Market ($B) $8.0 $7.4 $6.8 $6.0 $6.0 $5.0 $4.0 $4.0 $2.0 $0.0 2015E 2016E 2017E 2018E 2019E 10 GbE 25 GbE 40 GbE 50 GbE 100 GbE Crehan Research: Q2 15 Server Class Adapter Market Share (August 2015), Long Range Forecast Server-class Adapters & LOM (July 2015), Adapter forecast includes LOM;

Ethernet Growth in 25Gb/s & Faster High Speed Adapter Market Forecast 2020 ($1.8B) By 2020, Over 50% of market Fastest ramp in history Passes 40GbE in 2019 50 & 100GbE Also Grow Cloud/Web 2.0 customers Large enterprises 25/40/50/100 GbE speeds will make up 57% of high-speed adapter sales by 2020 Crehan Research: Q3 15 Server Class Adapter Market Share (Nov 2015), Long Range Forecast Server-class Adapters & LOM (Jan 2016), Adapter forecast includes LOM;

25 Is the New 10; 50 is the new 40 25GbE = 2.5x bandwidth Only 1.5x the price 40% lower cost/bw Compatible with 10GbE ports 50GbE = 1.25x BW at same cost Upgrade from 40GbE 20% lower cost/bw Cheaper optical cables 100GbE for switch links 60% fewer links Re-use existing fiber cables! New Ethernet Drives 25Gb/s Per Lane More Data on the Same Links QSFP+ SFP+ 40GbE = 4 x 10Gb/s 10GbE QSFP28 SFP28 100GbE = 4 x 25Gb/s 25GbE

High Speed Network Cables: Quick Definition Review Direct Attach Copper (DAC) Uses pulsing electrical signals sent into shielded copper wires. At high data rates, wire acts like a radio antenna. Data signal leaves the wire and lost, so length becomes shorter. Longer the cable, thicker it becomes. Optical Transceiver Transceiver n*channels Transmit n*channels Receiver Maximum length for 10/40GbE 5m Maximum length for 25/100GbE 3m Converts electrical signals to optical. Transmits blinking laser light over tiny glass optical fibers. Key Benefits: Lowest Cost, zero power Typical usage: Inside rack, 0.5-3m reach 2 fiber types: Multi-mode up to 100m Single-mode up to 2Km Applies to both 10G/40G and 25G/100G More expensive than DAC copper Key Benefits: Long reaches Uses: Linking switches up to 2Km. Active Optical Cable 2 Transceivers with optical fiber glued in. + Lowest Cost Optical for <200m

Ceph 8

Ceph Performance Testing Ceph is a scale-out system Two logical networks 3-way replication or erasure-coding Metadata, monitoring, and client traffic Rebuild or rebalance traffic Faster network = faster performance Higher bandwidth and lower latency Increases small block IOPS Increases large block throughput Mellanox Testing 10/25/40/50 GbE

Ceph Performance Testing vs. 10GbE 4 Ceph OSD servers 3 NVMe SSDs each ConnectX-4 Lx Set network to 10, 25, 40, and 50GbE speeds Aggregate performance of 4 Ceph servers 25GbE: 67Gb/s & 242K IOPS (vs. 35Gb/s & 130K IOPS at 10GbE) 25GbE has 92% more throughput than 10GbE 25GbE has 86% more IOPS than 10GbE

Supermicro Testing: 2x10GbE vs. 40GbE 13-20 Gb/s per server with Mellanox 1x40GbE 5.6-9.6 Gb/s per server with Intel 2x 10GbE ports Key Test Results More disks = more MB/s per server, less/osd More flash is faster (usually) All-flash 2-SSDs node faster than 36 HDDs 40GbE Advantages Up to 2x read throughput per server Up to 50% decrease in latency Easier than bonding multiple 10GbE links

QCT & Red Hat Ceph Storage Testing Up to 7x more throughput per OSD server with Mellanox 1x40GbE QuantaPlex T21P-4U Dual-Node 2 OSD nodes, 70 HDD & 4 SSD per server 35x 8TB HDD + 2x PCIe SSD per node 10GbE or Mellanox 40GbE NIC Key 40GbE Test Results Up to 2700MB/s read per node (21.6Gb/s) Up to 7x faster reads than 10GbE Also faster write throughput, even when <10Gb/s

Cisco Ceph Testing: 10GbE Not Enough Cisco Test Setup UCS C3160 servers, Nexus 9396PX switch 28 or 56 6TB SAS disks; Replication or EC 4x10GbE per server, bonded (Easier if had used Mellanox 40GbE NICs) Results with 3x Replication One node read: 3700 MB/s (29.6 Gb/s) One node write: 860 MB/s rep (6.8 Gb/s) 3 nodes read: 9,700 MB/s (77.6 Gb/s) 8 nodes read: 20,000 MB/s (160 Gb/s)

Ceph Customer Network: Monash University Research University 67,000 students 9 locations 3 Ceph clusters >60 nodes >6PB storage First network: 10GbE to nodes 56GbE inter-switch Second Ceph network 25GbE to nodes 100GbE inter-switch

Hyper-Converged 15

Faster Interconnect Enables Higher ROI Faster Networking Enables Savings in VDI Deployments over vsphere* Interconnect # Virtual Desktop per Server # Servers # Switches CapEx CapEx per Virtual Desktop 10GbE 60 84 2 $ 540,545 $108 25GbE** 140 36 1 $ 279,214 $ 55 40GbE 167 30 1 $ 207,890 $ 41 50GbE,100GbE and RoCE will enable higher efficiency Significant savings when running 5000 virtual desktops * Hardware savings only (not including VMware, guest OS license) and OpEx ** Interpolation based on the 10GbE and 40GbE results Higher Performance Interconnect Enables >2X Higher Efficiency

38x Faster Live Migration on 40GbE

Higher Efficiency with VSAN on 40GbE VMware VSAN All-flash storage vcloud Suite 6.0 Network: 40/56 GbE NSX

Windows Storage on 100GbE With RDMA 2X better performance with RoCE 2X higher bandwidth & 2X better CPU efficiency RoCE achieves full Flash storage bandwidth Remote storage without compromises

Microsoft Storage Spaces Direct @ 100Gb/s Microsoft recently benchmarked 100Gb/s Storage Spaces Direct 4 Dell R730XD Servers Samsung NVMe Flash ConnectX-4 100GbE Adapters RoCE for RDMA 60 GByte/sec aggregate throughput Could transmit entire content of Wikipedia in 5 seconds 4 R730XD Servers https://blogs.technet.microsoft.com/filecab/2016/05/11/s2dthroughputtp5/

Windows Storage Spaces Faster Networks RoCE vs. TCP IOPS 10Gb/s: +58% 40Gb/s: +94% 56Gb/s: +131% RoCE vs. TCP Latency 10Gb/s: -63% 40Gb/s: -51% 56Gb/s: -43% 40GbE vs. 10GbE IOPS TCP/IP: +151% RoCE: +208% 56Gb/s vs. 40Gb/s TCP/IP: None RoCE: +20% Windows Server 2016 Storage Spaces Direct Setup * SX1012 / SX1036

Other Performance Testing 1200 MapR Comparison 40GbE up to 70% Better HGST SSDs Nexenta Edge 128KB random writes 25GbE: 14.4Gb/s Execution Time (in seconds) 1000 800 600 400 200 70% improvement 50GbE: 23 Gb/s 0 Intel 10Gb/s Mellanox 10Gb/s Mellanox 40Gb/s Ethernet Network

Thank You 23