THE STORAGE PERFORMANCE DEVELOPMENT KIT AND NVME-OF

Similar documents
SPDK China Summit Ziye Yang. Senior Software Engineer. Network Platforms Group, Intel Corporation

Ziye Yang. NPG, DCG, Intel

Changpeng Liu. Senior Storage Software Engineer. Intel Data Center Group

Accelerating NVMe I/Os in Virtual Machine via SPDK vhost* Solution Ziye Yang, Changpeng Liu Senior software Engineer Intel

NVMe Over Fabrics: Scaling Up With The Storage Performance Development Kit

Changpeng Liu. Cloud Storage Software Engineer. Intel Data Center Group

Accelerating NVMe-oF* for VMs with the Storage Performance Development Kit

Jim Harris. Principal Software Engineer. Intel Data Center Group

Scott Oaks, Oracle Sunil Raghavan, Intel Daniel Verkamp, Intel 03-Oct :45 p.m. - 4:30 p.m. Moscone West - Room 3020

Accelerate block service built on Ceph via SPDK Ziye Yang Intel

Jim Harris Principal Software Engineer Intel Data Center Group

Jim Harris. Principal Software Engineer. Data Center Group

Intel Architecture 2S Server Tioga Pass Performance and Power Optimization

Changpeng Liu, Cloud Software Engineer. Piotr Pelpliński, Cloud Software Engineer

Intel SSD Data center evolution

Andreas Schneider. Markus Leberecht. Senior Cloud Solution Architect, Intel Deutschland. Distribution Sales Manager, Intel Deutschland

Ben Walker Data Center Group Intel Corporation

Future of datacenter STORAGE. Carol Wilder, Niels Reimers,

OPENSHMEM AND OFI: BETTER TOGETHER

Ed Warnicke, Cisco. Tomasz Zawadzki, Intel

Ceph BlueStore Performance on Latest Intel Server Platforms. Orlando Moreno Performance Engineer, Intel Corporation May 10, 2018

Daniel Verkamp, Software Engineer

SPDK Blobstore: A Look Inside the NVM Optimized Allocator

Munara Tolubaeva Technical Consulting Engineer. 3D XPoint is a trademark of Intel Corporation in the U.S. and/or other countries.

Fast-track Hybrid IT Transformation with Intel Data Center Blocks for Cloud

12th ANNUAL WORKSHOP 2016 NVME OVER FABRICS. Presented by Phil Cayton Intel Corporation. April 6th, 2016

April 2 nd, Bob Burroughs Director, HPC Solution Sales

Mohan J. Kumar Intel Fellow Intel Corporation

Ravindra Babu Ganapathi

Intel optane memory as platform accelerator. Vladimir Knyazkin

Re-Architecting Cloud Storage with Intel 3D XPoint Technology and Intel 3D NAND SSDs

Low-Overhead Flash Disaggregation via NVMe-over-Fabrics Vijay Balakrishnan Memory Solutions Lab. Samsung Semiconductor, Inc.

Low-Overhead Flash Disaggregation via NVMe-over-Fabrics

Storage Performance Development Kit (SPDK) Daniel Verkamp, Software Engineer

Accelerating Real-Time Big Data. Breaking the limitations of captive NVMe storage

Jim Pappas Director of Technology Initiatives, Intel Vice-Chair, Storage Networking Industry Association (SNIA) December 07, 2018

unleashed the future Intel Xeon Scalable Processors for High Performance Computing Alexey Belogortsev Field Application Engineer

FAST FORWARD TO YOUR <NEXT> CREATION

Hubert Nueckel Principal Engineer, Intel. Doug Nelson Technical Lead, Intel. September 2017

EXPERIENCES WITH NVME OVER FABRICS

Building an Open Memory-Centric Computing Architecture using Intel Optane Frank Ober Efstathios Efstathiou Oracle Open World 2017 October 3, 2017

Ultimate Workstation Performance

Intel s Architecture for NFV

Accelerating Data Center Workloads with FPGAs

Data and Intelligence in Storage Carol Wilder Intel Corporation

Välkommen. Intel Anders Huge

Andrzej Jakowski, Armoun Forghan. Apr 2017 Santa Clara, CA

Re- I m a g i n i n g D a t a C e n t e r S t o r a g e a n d M e m o r y

Out-of-band (OOB) Management of Storage Software through Baseboard Management Controller Piotr Wysocki, Kapil Karkra Intel

Extremely Fast Distributed Storage for Cloud Service Providers

Small File I/O Performance in Lustre. Mikhail Pershin, Joe Gmitter Intel HPDD April 2018

Using FPGAs to accelerate NVMe-oF based Storage Networks

2017 Storage Developer Conference. Mellanox Technologies. All Rights Reserved.

NVMe over Universal RDMA Fabrics

N V M e o v e r F a b r i c s -

NVMf based Integration of Non-volatile Memory in a Distributed System - Lessons learned

Colin Cunningham, Intel Kumaran Siva, Intel Sandeep Mahajan, Oracle 03-Oct :45 p.m. - 5:30 p.m. Moscone West - Room 3020

Intel. Rack Scale Design: A Deeper Perspective on Software Manageability for the Open Compute Project Community. Mohan J. Kumar Intel Fellow

Hardware and Software Co-Optimization for Best Cloud Experience

Accelerate Finger Printing in Data Deduplication Xiaodong Liu & Qihua Dai Intel Corporation

Ceph Optimizations for NVMe

Bei Wang, Dmitry Prohorov and Carlos Rosales

Challenges of High-IOPS Enterprise-level NVMeoF-based All Flash Array From the Viewpoint of Software Vendor

Sayantan Sur, Intel. SEA Symposium on Overlapping Computation and Communication. April 4 th, 2018

Virtuozzo Hyperconverged Platform Uses Intel Optane SSDs to Accelerate Performance for Containers and VMs

March NVM Solutions Group

High Performance Storage : A Cloud Story. Luwei He Standard Engineer, Huawei

BUILDING A BLOCK STORAGE APPLICATION ON OFED - CHALLENGES

LS-DYNA Performance on Intel Scalable Solutions

Real-Time Systems and Intel take industrial embedded systems to the next level

SNIA Developers Conference - Growth of the iscsi RDMA (iser) Ecosystem

OpenMPDK and unvme User Space Device Driver for Server and Data Center

Intel Xeon Processor E v3 Family

DPDK Performance Report Release Test Date: Nov 16 th 2016

Achieving 2.5X 1 Higher Performance for the Taboola TensorFlow* Serving Application through Targeted Software Optimization

SELINUX SUPPORT IN HFI1 AND PSM2

Meltdown and Spectre Interconnect Performance Evaluation Jan Mellanox Technologies

Fast forward. To your <next>

Intel Speed Select Technology Base Frequency - Enhancing Performance

IXPUG 16. Dmitry Durnov, Intel MPI team

NVMe Over Fabrics (NVMe-oF)

Graphics Performance Analyzer for Android

Architected for Performance. NVMe over Fabrics. September 20 th, Brandon Hoff, Broadcom.

Optimization of Lustre* performance using a mix of fabric cards

Copyright 2017 Intel Corporation

Accelerating HPC. (Nash) Dr. Avinash Palaniswamy High Performance Computing Data Center Group Marketing

DEVITO AUTOMATED HIGH-PERFORMANCE FINITE DIFFERENCES FOR GEOPHYSICAL EXPLORATION

Are You Insured Against Your Noisy Neighbor Sunku Ranganath, Intel Corporation Sridhar Rao, Spirent Communications

NVMe over Fabrics. High Performance SSDs networked over Ethernet. Rob Davis Vice President Storage Technology, Mellanox

RAIN: Reinvention of RAID for the World of NVMe

H.J. Lu, Sunil K Pandey. Intel. November, 2018

Maximize Performance and Scalability of RADIOSS* Structural Analysis Software on Intel Xeon Processor E7 v2 Family-Based Platforms

FlashGrid Software Enables Converged and Hyper-Converged Appliances for Oracle* RAC

IFS RAPS14 benchmark on 2 nd generation Intel Xeon Phi processor

POWER YOUR CREATIVITY WITH THE INTEL CORE X-SERIES PROCESSOR FAMILY

INTEL HPC DEVELOPER CONFERENCE FUEL YOUR INSIGHT

Sample for OpenCL* and DirectX* Video Acceleration Surface Sharing

A U G U S T 8, S A N T A C L A R A, C A

Intel Cluster Toolkit Compiler Edition 3.2 for Linux* or Windows HPC Server 2008*

Jomar Silva Technical Evangelist

Transcription:

14th ANNUAL WORKSHOP 2018 THE STORAGE PERFORMANCE DEVELOPMENT KIT AND NVME-OF Paul Luse Intel Corporation Apr 2018

AGENDA Storage Performance Development Kit What is SPDK? The SPDK Community Why are so many storage companies using it? How is it being used? 2 OpenFabrics Alliance Workshop 2018

WHAT IS SPDK? 3 OpenFabrics Alliance Workshop 2018

WHAT IS SPDK? Scalable and Efficient Software Ingredients Storage Performance Development Kit User space, lockless, polled-mode components Up to millions of IOPS per core Designed to extract maximum performance from nonvolatile media Storage Reference Architecture Optimized for latest generation CPUs and SSDs Open source composable building blocks (BSD licensed) http://spdk.io 4 OpenFabrics Alliance Workshop 2018

SPDK ARCHITECTURE 5 OpenFabrics Alliance Workshop 2018

THE SPDK COMMUNITY 6 OpenFabrics Alliance Workshop 2018

THE SPDK COMMUNITY Full Transparency 7 OpenFabrics Alliance Workshop 2018

WHY ARE SO MANY STORAGE COMPANIES USING SPDK? 8 OpenFabrics Alliance Workshop 2018

SOFTWARE IS BECOMING THE BOTTLENECK SPDK Unlocks New Media Potential 9 OpenFabrics Alliance Workshop 2018

SPDK BENEFITS Storage Performance Development Kit Up to 10X MORE IOPS/core for NVMe-oF* vs kernel Up to 8X MORE IOPS/core for NVMe vs kernel Up to 50% BETTER tail latency w/rocksdb workloads FASTER TTM w/less RESOURCES vs from scratch Provides FUTURE PROOFING as technologies evolve Software and workloads used in performance tests may have been optimized for performance only on Intel microprocessors. Performance tests, such as SYSmark and MobileMark, are measured using specific computer systems, components, software, operations and functions. Any change to any of those factors may cause the results to vary. You should consult other information and performance tests to assist you in fully evaluating your contemplated purchases, including the performance of that product when combined with other products. For more information go to http://www.intel.com/performance 10 OpenFabrics Alliance Workshop 2018

INDUSTRY INVOLVEMENT And many more! 11 OpenFabrics Alliance Workshop 2018

HOW IS SPDK BEING USED? 12 OpenFabrics Alliance Workshop 2018

APPLICATION ACCELERATION (LOCAL STORAGE) Provides direct access from application to media BDEV abstraction provides consistent API to various types of block devices Benefits: dramatically reduces latency and improves IO consistency 13 OpenFabrics Alliance Workshop 2018

REMOTE ACCESS TO STORAGE NVMe-oF supports different fabrics: RDMA (iwarp, RoCE) InfiniBand Fibre Channel Intel Omni-Path Architecture TCP (coming soon) Unified interface for the NVMe PCIe driver and the NVMe-oF initiator Libraries & applications 14 OpenFabrics Alliance Workshop 2018

ISCSI TARGET ACCESS As Ceph performance matures, enable higher efficiency access Potential for innovation in data services (e.g. cache, dedup...) 15 OpenFabrics Alliance Workshop 2018

VIRTUAL MACHINE ACCELERATION Provides dynamic block provisioning Increases VM density Decreases guest latency Works with KVM/QEMU 16 OpenFabrics Alliance Workshop 2018

3 RD PARTY BLOCK SERVICES Blobstore enables SSD virtualization BDEV enables stackable SW BDEV enables innovation 17 OpenFabrics Alliance Workshop 2018

NVME-OF 18 OpenFabrics Alliance Workshop 2018

SPDK NVME-OF NVMe-oF Target Emulated NVMe Controller NVMe-oF Initiator NVMe Userspace Driver Local Media Transport Transport BDEV Translation BDEV Layer NVMe Others Others Others RDMA libibverbs librdmac m FC Common Fabric Vendor Driver TCP RDMA libibverbs librdmacm Fabric FC Common Vendor Driver TCP NVMe Userspace Driver 19 OpenFabrics Alliance Workshop 2018

A WORD ON THE SPDK NVME DRIVER System Configuration: 2x Intel Xeon E5-2695v4 (HT off), Intel Speed Step enabled, Intel Turbo Boost Technology disabled, 8x 8GB DDR4 2133 MT/s, 1 DIMM per channel, CentOS* Linux* 7.2, Linux kernel 4.7.0-rc1, 1x Intel P3700 NVMe SSD (800GB), 4x per CPU socket, FW 8DV10102, I/O workload 4KB random read, Queue Depth: 1 per SSD, Performance measured by Intel using SPDK overhead tool, Linux kernel data using Linux AIO SPDK reduces NVMe software overhead up to 10x! 20 OpenFabrics Alliance Workshop 2018

A WORD ON THE SPDK NVME DRIVER Systems with multiple NVMe SSDs capable of millions of IOPS SPDK enables: more CPU cycles for storage services lower I/O latency SPDK saturates 8 NVMe SSDs with a single CPU core! System Configuration: 2x Intel Xeon E5-2695v4 (HT off), Intel Speed Step enabled, Intel Turbo Boost Technology disabled, 8x 8GB DDR4 2133 MT/s, 1 DIMM per channel, CentOS* Linux* 7.2, Linux kernel 4.7.0-rc1, 1x Intel P3700 NVMe SSD (800GB), 4x per CPU socket, FW 8DV10102, I/O workload 4KB random read, Queue Depth: 1 per SSD, Performance measured by Intel using SPDK overhead tool, Linux kernel data using Linux AIO 21 OpenFabrics Alliance Workshop 2018

NVME-OF TARGET PERFORMANCE SPDK reduces NVMe over Fabrics software overhead up to 10x! System Configuration: 2x Intel Xeon E5-2695v4 (HT off), Intel Speed Step enabled, Intel Turbo Boost Technology disabled, 8x 8GB DDR4 2133 MT/s, 1 DIMM per channel, CentOS* Linux* 7.2, Linux kernel 4.7.0-rc1, 1x Intel P3700 NVMe SSD (800GB), 4x per CPU socket, FW 8DV10102, I/O workload 4KB random read, Queue Depth: 1 per SSD, Performance measured by Intel using SPDK overhead tool, Linux kernel data using Linux AIO 22 OpenFabrics Alliance Workshop 2018

SPDK HOST+TARGET VS KERNEL HOST+TARGET SPDK reduces Optane NVMe-oF latency by 44%, write latency by 32%! System Configuration: 2x Intel Xeon E5-2695v4 (HT off), Intel Speed Step enabled, Intel Turbo Boost Technology disabled, 8x 8GB DDR4 2133 MT/s, 1 DIMM per channel, CentOS* Linux* 7.2, Linux kernel 4.7.0-rc1, 1x Intel P3700 NVMe SSD (800GB), 4x per CPU socket, FW 8DV10102, I/O workload 4KB random read, Queue Depth: 1 per SSD, Performance measured by Intel using SPDK overhead tool, Linux kernel data using Linux AIO 23 OpenFabrics Alliance Workshop 2018

NVME-OF LATENCY 4KB RANDOM READ, INTEL OPTANE SSD, SPDK TGT & KERNEL INITIATOR 20 us round trip time measured from NVMeoF initiator Out of 20usec, ~7 us spent in NVMe controller 12-13 usmeasured time in the fabric and kernel NVMe-oF initiator Disclaimer: Software and workloads used in performance tests may have been optimized for performance only on Intel microprocessors. Performance tests, such as SYSmark and MobileMark, are measured using specific computer systems, components, software, operations and functions. Any change to any of those factors may cause the results to vary. You should consult other information and performance tests to assist you in fully evaluating your contemplated purchases, including the performance of that product when combined with other products. System Configuration: 2x Intel Xeon E5-2695v4 (HT on, Intel Speed Step enabled, Intel Turbo Boost Technology enabled, 64GB DDR4 Memory, 8x 8GB DDR4 2400 MT/s, Ubuntu 16.04.1, Linux kernel 4.10.1, 1x 25GbE Mellanox 2P CX-4, CX-4 FW= 14.16.1020, mlx5_core= 3.0-1 driver, 1 ColdStream, connected to socket 0, 4KB Random Read I/0 1 initiators, each initiator connected to bx NVMe-oF subsystems using 2P 25GbE Mellanox. Performance measured by Intel using SPDK perf tool, 4KB Random Read I/O, Queue Depth: 1/NVMe-oF subsystem. numjobs 1, 300 sec runtime, direct=1, norandommap=1, FIO-2.12, SPDK commit # 42eade49 SPDK NVMf target adds just 100-200 nsto fabric overhead OpenFabrics Alliance Workshop 2018

NVME-OF LATENCY 4KB RANDOM READ, INTEL OPTANE SSD, SPDK TGT & SPDK INITIATOR 14 usround trip time measured from NVMf client Out of 14usec, ~7 usspent in NVMe controller 7 usmeasured time in the fabric and SPDK NVMe-oF initiator Disclaimer: Software and workloads used in performance tests may have been optimized for performance only on Intel microprocessors. Performance tests, such as SYSmark and MobileMark, are measured using specific computer systems, components, software, operations and functions. Any change to any of those factors may cause the results to vary. You should consult other information and performance tests to assist you in fully evaluating your contemplated purchases, including the performance of that product when combined with other products. System Configuration: 2x Intel Xeon E5-2695v4 (HT on, Intel Speed Step enabled, Intel Turbo Boost Technology enabled, 64GB DDR4 Memory, 8x 8GB DDR4 2400 MT/s, Ubuntu 16.04.1, Linux kernel 4.10.1, 1x 25GbE Mellanox 2P CX-4, CX-4 FW= 14.16.1020, mlx5_core= 3.0-1 driver, 1 ColdStream, connected to socket 0, 4KB Random Read I/0 1 initiators, each initiator connected to bx NVMe-oF subsystems using 2P 25GbE Mellanox. Performance measured by Intel using SPDK perf tool, 4KB Random Read I/O, Queue Depth: 1/NVMe-oF subsystem. numjobs 1, 300 sec runtime, direct=1, norandommap=1, FIO-2.12, SPDK commit # 42eade49 SPDK NVMf target adds just 100-200 nsec to fabric overhead OpenFabrics Alliance Workshop 2018

14th ANNUAL WORKSHOP 2018 THANK YOU Paul Luse Intel Corporation

LEGAL DISCLAIMER Intel technologies features and benefits depend on system configuration and may require enabled hardware, software or service activation. Performance varies depending on system configuration. No computer system can be absolutely secure. Tests document performance of components on a particular test, in specific systems. Differences in hardware, software, or configuration will affect actual performance. For more complete information about performance and benchmark results, visit http://www.intel.com/benchmarks. Software and workloads used in performance tests may have been optimized for performance only on Intel microprocessors. Performance tests, such as SYSmark and MobileMark, are measured using specific computer systems, components, software, operations and functions. Any change to any of those factors may cause the results to vary. You should consult other information and performance tests to assist you in fully evaluating your contemplated purchases, including the performance of that product when combined with other products. For more complete information visit http://www.intel.com/benchmarks. Benchmark results were obtained prior to implementation of recent software patches and firmware updates intended to address exploits referred to as "Spectre" and "Meltdown." Implementation of these updates may make these results inapplicable to your device or system. Intel Advanced Vector Extensions (Intel AVX)* provides higher throughput to certain processor operations. Due to varying processor power characteristics, utilizing AVX instructions may cause a) some parts to operate at less than the rated frequency and b) some parts with Intel Turbo Boost Technology 2.0 to not achieve any or maximum turbo frequencies. Performance varies depending on hardware, software, and system configuration and you can learn more at http://www.intel.com/go/turbo. Intel's compilers may or may not optimize to the same degree for non-intel microprocessors for optimizations that are not unique to Intel microprocessors. These optimizations include SSE2, SSE3, and SSSE3 instruction sets and other optimizations. Intel does not guarantee the availability, functionality, or effectiveness of any optimization on microprocessors not manufactured by Intel. Microprocessor-dependent optimizations in this product are intended for use with Intel microprocessors. Certain optimizations not specific to Intel microarchitecture are reserved for Intel microprocessors. Please refer to the applicable product User and Reference Guides for more information regarding the specific instruction sets covered by this notice. Cost reduction scenarios described are intended as examples of how a given Intel-based product, in the specified circumstances and configurations, may affect future costs and provide cost savings. Circumstances will vary. Intel does not guarantee any costs or cost reduction. Intel does not control or audit third-party benchmark data or the web sites referenced in this document. You should visit the referenced web site and confirm whether referenced data are accurate. 2018 Intel Corporation. Intel, the Intel logo, and Intel Xeon are trademarks of Intel Corporation in the U.S. and/or other countries. *Other names and brands may be claimed as property of others. OpenFabrics Alliance Workshop 2018