SRP Update. Bart Van Assche,

Similar documents
SNIA NVM Programming Model Workgroup Update. #OFADevWorkshop

BUILDING A BLOCK STORAGE APPLICATION ON OFED - CHALLENGES

The SNIA NVM Programming Model. #OFADevWorkshop

Host Redundancy, and IPoIB and SRP Redundancies

RDMA Container Support. Liran Liss Mellanox Technologies

Learn Your Alphabet - SRIOV, NPIV, RoCE, iwarp to Pump Up Virtual Infrastructure Performance

EXPERIENCES WITH NVME OVER FABRICS

iser as accelerator for Software Defined Storage Rahul Fiske, Subhojit Roy IBM (India)

InfiniBand Networked Flash Storage

OpenFabrics Interface WG A brief introduction. Paul Grun co chair OFI WG Cray, Inc.

Architected for Performance. NVMe over Fabrics. September 20 th, Brandon Hoff, Broadcom.

Past and present of the Linux NVMe driver Christoph Hellwig

Software-defined Shared Application Acceleration

SNIA Developers Conference - Growth of the iscsi RDMA (iser) Ecosystem

Linux Filesystems and Storage Chris Mason Fusion-io

OFED Storage Protocols

Introduction to High-Speed InfiniBand Interconnect

Ubuntu Linux Inbox Driver User Manual

Voltaire. Fast I/O for XEN using RDMA Technologies. The Grid Interconnect Company. April 2005 Yaron Haviv, Voltaire, CTO

NFS/RDMA over 40Gbps iwarp Wael Noureddine Chelsio Communications

iscsi Extensions for RDMA Updates and news Sagi Grimberg Mellanox Technologies

Storage Protocol Offload for Virtualized Environments Session 301-F

DB2 purescale: High Performance with High-Speed Fabrics. Author: Steve Rees Date: April 5, 2011

System Boot and RDMA. Jason Gunthorpe

Informatix Solutions INFINIBAND OVERVIEW. - Informatix Solutions, Page 1 Version 1.0

Containing RDMA and High Performance Computing

For personnal use only

PASS4TEST. IT Certification Guaranteed, The Easy Way! We offer free update service for one year

Configuration Maximums. Update 1 VMware vsphere 6.5 VMware ESXi 6.5 vcenter Server 6.5

The Host Server. Linux Configuration Guide. October 2017

SCSI and FC standards update Frederick Knight NetApp Inc

Configuring and Managing Virtual Storage

Application Acceleration Beyond Flash Storage

Open Source Storage. Ric Wheeler Architect & Senior Manager April 30, 2012

IOmark- VM. HP HP ConvergedSystem 242- HC StoreVirtual Test Report: VM- HC b Test Report Date: 27, April

Evaluating the Impact of RDMA on Storage I/O over InfiniBand

Low latency, high bandwidth communication. Infiniband and RDMA programming. Bandwidth vs latency. Knut Omang Ifi/Oracle 2 Nov, 2015

Red Hat Enterprise Linux (RHEL) 7.4-ALT Driver User Manual

RDMA Requirements for High Availability in the NVM Programming Model

InfiniBand OFED Driver for. VMware Virtual Infrastructure (VI) 3.5. Installation Guide

Application Advantages of NVMe over Fabrics RDMA and Fibre Channel

SAN Storage Array Workbook September 11, 2012

InfiniBand OFED Driver for. VMware Infrastructure 3. Installation Guide

NVMe over Fabrics support in Linux Christoph Hellwig Sagi Grimberg

Linux File Systems: Challenges and Futures Ric Wheeler Red Hat

PARAVIRTUAL RDMA DEVICE

The Exascale Architecture

NVMe over Fabrics. High Performance SSDs networked over Ethernet. Rob Davis Vice President Storage Technology, Mellanox

The Host Server. Linux Configuration Guide. August The authority on real-time data

Configuration Maximums

SCSI RDMA Protocol. Introduction. Configuring SRP CHAPTER

CTWP015: S.M.A.R.T. Features for Cactus Technologies Industrial-Grade Flash-Storage Products

Multifunction Networking Adapters

Evaluation of Chelsio Terminator 6 (T6) Unified Wire Adapter iscsi Offload

Mellanox Connect-IB Firmware (fw-connectib) Release Notes

High Performance MPI on IBM 12x InfiniBand Architecture

Simplified and Consolidated Parallel Media File System Solution

IOmark- VM. HP MSA P2000 Test Report: VM a Test Report Date: 4, March

InfiniBand Linux Operating System Software Access Layer

RoCE vs. iwarp Competitive Analysis

Infiniband and RDMA Technology. Doug Ledford

New Storage Architectures

Paving the Way to the Non-Volatile Memory Frontier. PRESENTATION TITLE GOES HERE Doug Voigt HP

InfiniPath Drivers and Software for QLogic QHT7xxx and QLE7xxx HCAs. Table of Contents

12th ANNUAL WORKSHOP 2016 NVME OVER FABRICS. Presented by Phil Cayton Intel Corporation. April 6th, 2016

ETHERNET OVER INFINIBAND

CERN openlab Summer 2006: Networking Overview

VMware vsphere 5.5 Professional Bootcamp

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme

VMware Virtual SAN. Technical Walkthrough. Massimiliano Moschini Brand Specialist VCI - vexpert VMware Inc. All rights reserved.

Linux Storage Area Network Topics & Linux Kernel Developers Summit Report

I/O Considerations for Server Blades, Backplanes, and the Datacenter

Installing and Configuring for Linux

Introducing Tegile. Company Overview. Product Overview. Solutions & Use Cases. Partnering with Tegile

Resilient and Fast Persistent Container Storage Leveraging Linux s Storage Functionalities Philipp Reisner, CEO LINBIT

IOmark- VM. IBM IBM FlashSystem V9000 Test Report: VM a Test Report Date: 5, December

Ubuntu Linux Inbox Driver User Manual

Performance Analysis and Evaluation of Mellanox ConnectX InfiniBand Architecture with Multi-Core Platforms

DAFS Storage for High Performance Computing using MPI-I/O: Design and Experience

Linux Host Utilities 6.2 Quick Start Guide

Management Scalability. Author: Todd Rimmer Date: April 2014

The Oracle Database Appliance I/O and Performance Architecture

Reliable Storage for HA, DR, Clouds and Containers Philipp Reisner, CEO LINBIT

DRBD SDS. Open Source Software defined Storage for Block IO - Appliances and Cloud Philipp Reisner. Flash Memory Summit 2016 Santa Clara, CA 1

QuickSpecs. HP InfiniBand Options for HP BladeSystems c-class. Overview

Introduction to Infiniband

Advanced Computer Networks. End Host Optimization

Overview This proposes topics and text for an InfiniBand annex for the SCSI over RDMA (SRP) standard.

FlashGrid OPEN STORAGE SOFTWARE for Oracle RAC and Oracle ASM Clusters. Deployment and Maintenance Guide rev

IBA Software Architecture SCSI RDMA Protocol (SRP) Storage Driver High Level Design. Draft 2

FC-FCoE Adapter Inbox Driver Update for Linux Kernel 2.6.x. Table of Contents

NVMe Over Fabrics (NVMe-oF)

QLogic in HPC Vendor Update IDC HPC User Forum April 16, 2008 Jeff Broughton Sr. Director Engineering Host Solutions Group

NTRDMA v0.1. An Open Source Driver for PCIe NTB and DMA. Allen Hubbe at Linux Piter 2015 NTRDMA. Messaging App. IB Verbs. dmaengine.h ntb.

iscsi or iser? Asgeir Eiriksson CTO Chelsio Communications Inc

Linux Support of NFS v4.1 and v4.2. Steve Dickson Mar Thu 23, 2017

Welcome to the IBTA Fall Webinar Series

Oracle auf HP/Violin Wirklich ein Exadata Killer? DOAG Jahreskonferenz November 2012

FC-FCoE Adapter Inbox Driver Update for Linux Kernel 2.6.x. Table of Contents

SUSE Linux Enterprise Server (SLES) 12 SP2 Driver User Manual

Transcription:

SRP Update Bart Van Assche,

Overview Involvement With SRP SRP Protocol Overview Recent SRP Driver Changes Possible Future Directions March 30 April 2, 2014 #OFADevWorkshop 2

Involvement with SRP Maintainer of the open source Linux SRP initiator and the SCST SRP target drivers. Member of the Fusion-io ION team. ION is an all-flash H.A. shared storage appliance. Flash memory provides low latency and high bandwidth. The focus of RDMA is on low latency and high bandwidth. In other words, RDMA is well suited for remote access to flash memory. March 30 April 2, 2014 #OFADevWorkshop 3

SRP Protocol Overview SRP = SCSI RDMA Protocol. Defines how to perform SCSI communication over an RDMA network. Defines how to discover InfiniBand SRP targets, how to log in, how to transfer SCSI CDB's and also how to transfer data via RDMA. Revision 16a of the SRP protocol has been approved as an official ANSI standard in 2007. 4

SRP and SCSI SRP defines a SCSI transport layer. Enables supports for e.g. these SCSI features: Reading and writing data blocks. Read capacity. Command queueing. Multiple LUNs per SCSI host. Inquire LUN information, e.g. volume identification, caching information and thin provisioning support (a.k.a. TRIM / UNMAP). Atomic (vectored) write - helps to make database software faster. VAAI (WRITE SAME, UNMAP, ATS, XCOPY). End-to-end data integrity (a.k.a. T10-PI). Persistent reservations a.k.a. cluster support. Asymmetric Logical Unit Access (ALUA). Fusion-io is actively involved in the ANSI T10 committee for standardization of new SCSI commands. 5

SRP Protocol - Login IB spec defines device management. Initiator sends device management query to subnet manager. Subnet manager reports ports with device management capabilities. Initiator sends I/O controller query to each port with device management capabilities. SRP target reports I/O controllers. Initiator sends login request to selected I/O controllers. Initiator requests SCSI LUN report and queries capacity and identification of each LUN. I/O starts. 6

Linux SRP Initiator Support Kernel driver ib_srp - implements SRP protocol. User space srptools package. srp_daemon and ibsrpdm executables. Target discovery. Target login. Interface between kernel and user space /sys/class/infiniband_srp/srp-${port}/add_target /sys/class/srp_remote_ports /sys/class/scsi_host/*/{sgid,dgid,...} /sys/class/scsi_device/*/{state,queue_depth,...} 7

SRP Login - Example # cat /etc/srp_daemon.conf a queue_size=128,max_cmd_per_lun=128 # srp_daemon -oaecd/dev/infiniband/umad1 id_ext=0002c90300fc3210,ioc_guid=0002c90300fc3210,dg id=fe800000000000000002c90300fc3211,pkey=ffff,service _id=0002c90300fc3210 id_ext=0002c90300a543b0,ioc_guid=0002c90300a543b0,d gid=fe800000000000000002c90300fc3221,pkey=ffff,servic e_id=0002c90300a543b0 [ ] # ls /sys/class/srp_remote_ports/ port-453:1 port-459:1 [ ] # lsscsi [5:0:0:0] disk FUSIONIO ION LUN 3243 /dev/sdc [5:0:0:1] disk FUSIONIO ION LUN 3243 /dev/sdd 8

Recent SRP Initiator Changes Queue size is now configurable. Optimal performance for SSDs and hard disk RAID arrays can only be achieved with a large queue size (128 instead of the default 64). Support for modifying the queue depth dynamically has been added. Path loss detection time has been reduced from about 40s to about 17s. Further reduction is possible by lowering the subnet timeout on the subnet manager. This makes a significant difference in H.A. setups. Added support for fast_io_fail_tmo and dev_loss_tmo parameters for multipath. P_Key support has been added in srp_daemon. Many smaller changes in the srptools package. 9

OFED and SRP Support Upstream Linux kernel ib_srp srptools 3.14.0 1.0.2 RHEL 6.5 2.6.32+ 0.0.4 SLES 11 SP3 3.0.101 0.0.4 MLNX OFED 2.1 3.13.0 1.0.0 OFED 3.12 3.12.0 1.0.2 Fusion-io is working with Linux distribution vendors to keep SRP support up to date. 10

SRP Initiator and SCSI Core Linux SRP initiator is a SCSI driver. Linux SCSI mid-layer builds on block layer. SRP initiator relies on SCSI core for LUN scanning, SCSI error handling,... Path removal triggers a call of scsi_remove_host(). Path removal during I/O works reliably since Linux kernel 3.8. Fusion-io contributed several patches to make the Linux SCSI core and block layer handle path removal during I/O reliably. 11

Possible Future Directions Improving Linux SCSI performance via the scsi-mq project. Higher bandwidth by using multiple RDMA channels. Latency reduction. NUMA performance improvements. FRWR support - needed e.g. for ConnectIB HCA support. End-to-end data integrity (T10-PI) support; supported by Oracle database software. Builds on FRWR support. Adding SR-IOV support. Support for Ethernet networks (RoCE and/or iwarp). Requires to switch from IB/CM to RDMA/CM. Requires modification of the target discovery software (srptools). The current target discovery software is based on InfiniBand MADs. 12

13