Using FPGAs to accelerate NVMe-oF based Storage Networks

Similar documents
Challenges of High-IOPS Enterprise-level NVMeoF-based All Flash Array From the Viewpoint of Software Vendor

Hardware NVMe implementation on cache and storage systems

N V M e o v e r F a b r i c s -

THE STORAGE PERFORMANCE DEVELOPMENT KIT AND NVME-OF

Low-Overhead Flash Disaggregation via NVMe-over-Fabrics Vijay Balakrishnan Memory Solutions Lab. Samsung Semiconductor, Inc.

An NVMe-based Offload Engine for Storage Acceleration Sean Gibb, Eideticom Stephen Bates, Raithlin

Low-Overhead Flash Disaggregation via NVMe-over-Fabrics

An NVMe-based FPGA Storage Workload Accelerator

At the heart of a new generation of data center infrastructures and appliances. Sept 2017

Important new NVMe features for optimizing the data pipeline

NVM Express over Fabrics Storage Solutions for Real-time Analytics

Lenovo - Excelero NVMesh Reference Architecture

Persistent Memory over Fabrics

Application Acceleration Beyond Flash Storage

Accelerating Real-Time Big Data. Breaking the limitations of captive NVMe storage

Ziye Yang. NPG, DCG, Intel

FMS18 Invited Session 101-B1 Hardware Acceleration Techniques for NVMe-over-Fabric

2017 Storage Developer Conference. Mellanox Technologies. All Rights Reserved.

NVMe over Universal RDMA Fabrics

NVMe : Redefining the Hardware/Software Architecture

How to Network Flash Storage Efficiently at Hyperscale. Flash Memory Summit 2017 Santa Clara, CA 1

Designing Next Generation FS for NVMe and NVMe-oF

Accelerating Data Centers Using NVMe and CUDA

NVMe Over Fabrics (NVMe-oF)

Developing Extremely Low-Latency NVMe SSDs

Building dense NVMe storage

Reference Design: NVMe-oF JBOF

RAIN: Reinvention of RAID for the World of NVMe

SPDK China Summit Ziye Yang. Senior Software Engineer. Network Platforms Group, Intel Corporation

EXPERIENCES WITH NVME OVER FABRICS

S K T e l e c o m : A S h a r e a b l e D A S P o o l u s i n g a L o w L a t e n c y N V M e A r r a y. Eric Chang / Program Manager / SK Telecom

The QM2 PCIe Expansion Card Gives Your NAS a Performance Boost! QM2-2S QM2-2P QM2-2S10G1T QM2-2P10G1T

SNIA Developers Conference - Growth of the iscsi RDMA (iser) Ecosystem

Performance Prediction and Optimization using Linux/cgroups

All-NVMe Performance Deep Dive Into Ceph + Sneak Preview of QLC + NVMe Ceph

Developing Low Latency NVMe Systems for HyperscaleData Centers. Prepared by Engling Yeo Santa Clara, CA Date: 08/04/2017

InfiniBand Networked Flash Storage

Accelerate Applications Using EqualLogic Arrays with directcache

SOFTWARE-DEFINED BLOCK STORAGE FOR HYPERSCALE APPLICATIONS

Cloud Computing with FPGA-based NVMe SSDs

ARISTA: Improving Application Performance While Reducing Complexity

Application Advantages of NVMe over Fabrics RDMA and Fibre Channel

Future of datacenter STORAGE. Carol Wilder, Niels Reimers,

Scott Oaks, Oracle Sunil Raghavan, Intel Daniel Verkamp, Intel 03-Oct :45 p.m. - 4:30 p.m. Moscone West - Room 3020

Hyper-converged infrastructure with Proxmox VE virtualization platform and integrated Ceph Storage.

Multi-Host Sharing of NVMe Drives and GPUs Using PCIe Fabrics

Accelerating NVMe-oF* for VMs with the Storage Performance Development Kit

Evaluation of Chelsio Terminator 6 (T6) Unified Wire Adapter iscsi Offload

Storage Protocol Offload for Virtualized Environments Session 301-F

Evaluation of the Chelsio T580-CR iscsi Offload adapter

Unveiling the new QM2 M.2 SSD/10GbE PCIe Expansion cards

PCIe Storage Beyond SSDs

NVMe SSDs with Persistent Memory Regions

NVMe over Fabrics. High Performance SSDs networked over Ethernet. Rob Davis Vice President Storage Technology, Mellanox

genzconsortium.org Gen-Z Technology: Enabling Memory Centric Architecture

Re-Architecting Cloud Storage with Intel 3D XPoint Technology and Intel 3D NAND SSDs

Colin Cunningham, Intel Kumaran Siva, Intel Sandeep Mahajan, Oracle 03-Oct :45 p.m. - 5:30 p.m. Moscone West - Room 3020

RAIN: Reinvention of RAID for the World of NVMe. Sergey Platonov RAIDIX

From Rack Scale to Network Scale: NVMe Over Fabrics Enables Exabyte Applica>ons. Zivan Ori, CEO & Co-founder, E8 Storage

All-Flash High-Performance SAN/NAS Solutions for Virtualization & OLTP

Accelerating Ceph with Flash and High Speed Networks

Benchmarking Ceph for Real World Scenarios

Meltdown and Spectre Interconnect Performance Evaluation Jan Mellanox Technologies

Low latency and high throughput storage access

Hardened Security in the Cloud Bob Doud, Sr. Director Marketing March, 2018

NVMFS: A New File System Designed Specifically to Take Advantage of Nonvolatile Memory

Interconnect Your Future

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme

VN2VN: A new framework for L2 DAS Prafulla Deuskar Mark Wunderlich Intel

TELECOMMUNICATIONS TECHNOLOGY ASSOCIATION JET-SPEED HHS3124F / HHS2112F (10 NODES)

DataON and Intel Select Hyper-Converged Infrastructure (HCI) Maximizes IOPS Performance for Windows Server Software-Defined Storage

Flashmatrix Technology

All-Flash High-Performance SAN/NAS Solutions for Virtualization & OLTP

Highly Scalable, Non-RDMA NVMe Fabric. Bob Hansen,, VP System Architecture

Accelerating Data Center Workloads with FPGAs

Next Generation Computing Architectures for Cloud Scale Applications

Toward a Memory-centric Architecture

Learn Your Alphabet - SRIOV, NPIV, RoCE, iwarp to Pump Up Virtual Infrastructure Performance

FlashGrid Software Enables Converged and Hyper-Converged Appliances for Oracle* RAC

In-Network Computing. Sebastian Kalcher, Senior System Engineer HPC. May 2017

Networking at the Speed of Light

FPGA Implementation of Erasure Codes in NVMe based JBOFs

iscsi or iser? Asgeir Eiriksson CTO Chelsio Communications Inc

Implementing Ultra Low Latency Data Center Services with Programmable Logic

NVMe Direct. Next-Generation Offload Technology. White Paper

Implementing SQL Server 2016 with Microsoft Storage Spaces Direct on Dell EMC PowerEdge R730xd

The True Performance of Flash Storage. Flash Memory Summit 2018 Santa Clara, CA

MELLANOX EDR UPDATE & GPUDIRECT MELLANOX SR. SE 정연구

CPMD Performance Benchmark and Profiling. February 2014

Accelerating Storage with NVM Express SSDs and P2PDMA Stephen Bates, PhD Chief Technology Officer

Building a High IOPS Flash Array: A Software-Defined Approach

NexentaStor 5.x Reference Architecture

Data Processing at the Speed of 100 Gbps using Apache Crail. Patrick Stuedi IBM Research

All-Flash High-Performance SAN/NAS Solutions for Virtualization & OLTP

Enabling the NVMe CMB and PMR Ecosystem

Virtual Storage Tier and Beyond

RDMA in Embedded Fabrics

Accelerate block service built on Ceph via SPDK Ziye Yang Intel

Ethernet. High-Performance Ethernet Adapter Cards

Netronome 25GbE SmartNICs with Open vswitch Hardware Offload Drive Unmatched Cloud and Data Center Infrastructure Performance

Transcription:

Using FPGAs to accelerate NVMe-oF based Storage Networks Deboleena Sakalley IP & Solutions Architect, Xilinx Santa Clara, CA 1

Agenda NVMe-oF Offload in FPGA NVMe-oF Integrated Solution Solution Architecture Performance Conclusions Santa Clara, CA 2

NVMe-oF Offload in FPGA A given NVMe-oF Architecture Offload with FPGA Initiators NVMe PCIe Switch SW Stack NVMe-oF Bridging Initiators NVMe NVMe-oF Bridging PCIe Switch SW Stack NVMe NVMe FPGA NVMe-oF Target Driver OFED Driver PCIe Driver NVMe Driver Santa Clara, CA 3

NVMe-oF Offload in FPGA HW accelerated NVMe-oF SSD sharing across multiple hosts only in control path Doorbell exchanges to/from Discovery Connect Error handling Peer 2 peer transfer between and FPGA Initiator Initiator Initiator Ethernet NVMe NVMe NVMe-oF Bridging FPGA $? Gen3x8 PCIe Switch Bottleneck Gen3x8 q P2P Bottleneck q Too many components! q FPGA Value Add not clear Santa Clara, CA 4

NVMe-oF Integrated Solution Integrate functionality Use embedded FPGA processor Enable end-customer specific acceleration RAID Erasure coding De-dup Others Initiator Initiator Initiator Ethernet NVMe ET ROLE Processor FPGA FPGA NVMe Gen3x8 Gen3x8 PCIe Switch q P2P Bottleneck q Too many components! q FPGA Value Add not clear Santa Clara, CA 5

Solution Architecture HW implementation for Target/end point RoCE v2 Supports up to128 initiators 8 to 24 flash drives (with PCIe switch) Support for 10/25/40/100GbE and PCIe Gen3/Gen4 Datapath and control path handled in HW All recoverable errors handled in HW Ethernet Interfaces Ethernet MAC RoCE v2 (Target) ARM or Microblaze Non RoCE v2 packets Target SW Components NVMf IP PCIe RP PCIe RP NVMe Drives Embedded processor only used for Discovery Connect Error handling ET Driver NVMe-oF Driver Ethernet Driver Network Stack Management Interfaces & APIs Santa Clara, CA 6

Benchmarking Hardware Topology FPGA Based Setup Santa Clara, CA 7

Benchmarking Data: Performance FPGA Hardware Setup IOPs* SPDK vs FPGA 2500 2000 1500 1000 500 0 1 2 3 4 6 8 10 12 14 18 20 24 30 32 36 KIOPS (SPDK) KIOPS (FPGA) IOPs* Read/Write Mix FPGA Hardware Configura@on Ini$ator: Intel(R) i7-5820k @ 3.30GHz, 12 cores, 8GB RAM CNA: Mellanox CX-4 Switch: MSN2100-CB2F Model MSN2100, 16 QSFP28 ports Target (FPGA): Fidus Sidewinder Board ZU19, 2x100G QSFP+ Target (x86): Dual Socket E5-2620 (12 cores each), 128GB SSD: PCIe Gen3 x4, Capable of 740k Read IOPs each 2500 X86 (SPDK) Hardware Setup 2000 1500 1000 500 0 KIOPS (Rd) Page 8 *Data generated using fio and block_size=4k KIOPS (Wr)

Benchmarking Data: Latency FPGA Hardware Setup SPDK: read : io=2462.8mb, bw=42031kb/s, iops=10507, runt= 60000msec slat (usec): min=1, max=35, avg= 2.10, stdev= 0.51 clat (usec): min=67, max=2161, avg=92.29, stdev=11.75 lat (usec): min=75, max=2163, avg=94.47, stdev=11.7 FPGA: read : io=1219.9mb, bw=41637kb/s, iops=10409, runt= 30001msec slat (usec): min=1, max=70, avg= 3.37, stdev= 2.65 clat (usec): min=45, max=188, avg=91.47, stdev=11.49 lat (usec): min=74, max=194, avg=95.00, stdev=12.07 DAS: read : io=1283.1mb, bw=43823kb/s, iops=10955, runt= 30001msec slat (usec): min=3, max=43, avg= 3.73, stdev= 0.62 clat (usec): min=46, max=2209, avg=86.18, stdev=13.02 lat (usec): min=72, max=2213, avg=90.08, stdev=13.02 Page 9 *Data generated using fio and block_size=4k Hardware Configura@on Ini$ator: Intel(R) i7-5820k @ 3.30GHz, 12 cores, 8GB RAM CNA: Mellanox CX-4 Switch: MSN2100-CB2F Model MSN2100, 16 QSFP28 ports Target (FPGA): Fidus Sidewinder Board ZU19, 2x100G QSFP+ Target (x86): Dual Socket E5-2620 (12 cores each), 128GB SSD: PCIe Gen3 x4, Capable of 740k Read IOPs each X86 (SPDK) Hardware Setup

Conclusions FPGA Architectures today provide a highly integrated solution to NVMe over Fabric acceleration problem Allows for low latency, high throughput storage architectures with built-in accelerators FPGAs allow for workload specific optimizations enabled with partial reconfiguration Santa Clara, CA 10

Thank you! Santa Clara, CA 11

BACKUP Santa Clara, CA 12

NVMe-oF Integrated Architecture FPGA Integrated end-point HW implementation for Target/end point RoCE v2 Supports up to 128 Initiators Completely HW accelerated Doorbell exchanges through HW handshake No P2P bottleneck All end point drivers running on FPGA processor Santa Clara, CA 13

Xilinx ET + NVMf Reference Design Demo configuration Fidus Sideweinder board with ZU19EG 4 Samsung SSDs, directly attached to FPGA via PCIe Gen3x4 2 Initiators running Open NVMf 100G Mellanox on the initiators Santa Clara, CA 14

NVMf + ET : FPGA Value Add NVMf + ET solution Shell and Role Model of FPGA Shell FPGA : Always ON Hard cores : PCIeG3x8 Soft cores : NVMeOFabric, ET Role FPGA Applications can be ported : FPGA E2E datapath protection/correction End point Security Caching algorithms etc Santa Clara, CA PCIe Gen3x4 CMAC ET NVMeOFabric Role Role : Applications PCIe Gen3x4 Shell : G3x4, NVMeOF, ET 15

FPGA Value Add Attribute Xeon + External NIC Integrated NIC + FPGA Solution Cost High. Xeon adds cost Lower Lower Solution Power Higher Lower Lower Latency, Latency Outliers High Latency, large variation in latency Low latency. Shared data and control plane within the same increases latency variation Low Latency. Orthogonal data and control place minimize latency variation Custom Accelerators In Software In Software. Limited by integrated processor capability In Hardware Hardware Reconfiguration No No Yes Additional FPGA Capabilities Storage Pooling Data Sharing Storage Virtualization Custom Accelerators Workload Specific Optimizations Page 16

Xilinx Device options: MPSoC Santa Clara, CA 17

Xilinx Device options: Virtex US+ Santa Clara, CA 18