High Performance File Serving with SMB3 and RDMA via SMB Direct
|
|
- Brianna Allen
- 6 years ago
- Views:
Transcription
1 High Performance File Serving with SMB3 and RDMA via SMB Direct Tom Talpey, Microsoft Greg Kramer, Microsoft
2 Protocol SMB Direct New protocol supporting SMB 3.0 over RDMA Minimal CPU overhead High bandwidth, low latency Fabric agnostic iwarp, InfiniBand, RoCE IP addressing IANA port (smbdirect 5445) File Client Application SMB3 Client Network w/ RDMA support R-NIC User Kernel Network w/ RDMA support R-NIC File Server SMB3 Server NTFS SCSI Disk
3 Documented MS-SMBD MS-SMB2 Windows krdma API NDKPI Part of Windows Driver Kit Network Direct (and Verbs) heritage
4 Implemented Windows Server 2012 SMB 3.0 over SMB Direct Supports Multichannel Continuous availability All other SMB 3.0 features
5 Basics SMB Direct is a transport framing Only 3 message types 2-way full duplex transport which supports: Datagram-type send/receive exchange With fragmentation/reassembly for large Direct RDMA Read/Write SMB 3.0 binding defines transport use: Client buffer advertisement for READ and WRITE Server RDMA buffer access (push/pull)
6 Use Discovery via SMB 3.0 Multichannel RDMA attribute of interface Negotiated capabilities SMB Direct version Message and RDMA Region sizes Credits Messages RDMA Read operations (via NDK provider)
7 Three messages SMB Direct Negotiate Request Octet 0 Octet 1 Octet 2 Octet 3 MinVersion MaxVersion Once Everything else SMB Direct Data Transfer Header Octet 0 Octet 1 Octet 2 Octet 3 Reserved CreditsRequested CreditsRequested CreditsGranted PreferredSendSize MaxReceiveSize MaxFragmentedReceiveSize SMB Direct Negotiate Response Octet 0 Octet 1 Octet 2 Octet 3 MinVersion MaxVersion Flags Reserved RemainingDataLength DataOffset NegotiatedVersion CreditsRequested Status Reserved CreditsGranted DataLength Padding Data (variable) MaxReadWriteSize PreferredSendSize MaxReceiveSize MaxFragmentedReceiveSize 7
8 Transfers Send/Receive model Single logical message Possibly sent as fragmentation train Using ordering properties of RDMA Implements crediting All SMB 3.0 operations use this Direct placement model Advertises RDMA regions in scatter/gather list SMB 3.0 uses for SMB2_READ and SMB2_WRITE Only. Piggyback on existing Channel
9 Send transfers DataOffset = 24 DataLength = 1000 RemainingDataLength = 1048 SMB Direct HDR (24 bytes) SMB3 message bytes Send 0 DataOffset = 24 DataLength = 1000 RemainingDataLength = 48 SMB Direct HDR (24 bytes) SMB3 message bytes Send 1 DataOffset = 24 DataLength = 48 RemainingDataLength = 0 SMB Direct HDR (24 bytes) SMB3 message bytes Send 2 9
10 SMB3 Reads and Writes SMB3 WRITE REQUEST Octet 0 Octet 1 Octet 2 Octet 3 StructureSize DataOffset Length Offset FileId Channel RemainingBytes WriteChannelInfoOffset WriteChannelInfoLength Flags Buffer (variable) Channel array SMB3 READ REQUEST Octet 0 Octet 1 Octet 2 Octet 3 StructureSize Padding Reserved Length Offset FileId MinimumCount Channel RemainingBytes ReadChannelInfoOffset ReadChannelInfoLength Flags Buffer (variable) Previously reserved fields Octet 0 Octet 1 Octet 2 Octet 3 Address Token Length 10
11 RDMA transfers SMB Direct READ SMB Direct HDR SMB3 HDR SMB3 READ REQ MEMORY DESCRIPTORS Send RDMA Write DATA Send SMB Direct HDR SMB3 HDR SMB3 READ RESP Client Server SMB Direct WRITE SMB Direct HDR SMB3 HDR SMB3 WRITE REQ MEMORY DESCRIPTORS Send DATA RDMA Read Send SMB Direct HDR SMB3 HDR SMB3 WRITE RESP 11
12 Credits Bi-directional Count of ready receive buffers offered Dynamic can increase or decrease at any time Optional to do so Used only to control low-level SMBD message exchanges Recycled independently of SMB operations Relatively small number required (100 s even for deep random workloads)
13 Quirks Interesting corner cases Last credit Always need 1 in each endpoint to avoid deadlock (but see details in spec!) Bi-directional no requirement for same both ways Async/Cancel/Errors No reply, multiple reply, unexpected large reply NOT an RPC-like interface, much as it may resemble one
14 Efficiency True bi-directional and streaming sends Can be exposed as sockets-like interface With register/unregister/rdma rw extensions RDMA operations / completions Datamover offload to RNIC Server pull model improves performance Many options for RDMA efficiency FRMR, silent completions, coalescing, etc Resources bounded by credits and sizes
15 Performance 15
16 SDC 2011 performance results 160,000 IOPS (1KiB random reads) RAID 0 12 s RAID 0 12 s 3200 MiB/sec (512KiB sequential reads) Single 32 Gbps InfiniBand link InfiniBand switch Nehalem: 1 socket x Ghz Westmere: 2 socket x Ghz 16
17 Current performance results File Client (SMB 3.0) SQLIO RDMA NIC RDMA NIC File Server (SMB 3.0) RDMA NIC RDMA NIC NTFS Storage Spaces HBA HBA JBOD JBOD
18 Current performance results sqlio2.exe -T100 t2 s60 b512 -o4 fsequential -BN LS (1 file per volume) Avg. MB/sec* Avg. IOs/sec (512 KiB) Avg. %CPU (Client) Avg. Latency (ms) 7,340 ~14K Server fully utilized sqlio2.exe -T100 -t16 s60 -b8 -o4 frandom -BN LS (four files per volume) Avg. MB/sec* Avg. IOs/sec (8 KiB) Avg. %CPU (Client) Avg. Latency (ms) 3,711 ~453K 60 < 1 Server fully utilized * 1MB = 1,000,000 bytes 18
19 Let s take it to 11! File Client (SMB 3.0) SQLIO RDMA NIC RDMA NIC RDMA NIC File Server (SMB 3.0) RDMA NIC RDMA NIC RDMA NIC NTFS Storage Spaces HBA HBA HBA HBA HBA HBA JBOD JBOD JBOD JBOD JBOD JBOD 19
20 Let s take it to 11 16! sqlio2.exe -T100 t2 s60 b512 -o4 fsequential -BN LS (1 file per volume) Avg. MB/sec* Avg. IOs/sec (512 KiB) Avg. %CPU (client) Avg. Latency (ms) 16,253 ~31K GigaBYTES (not bits) of storage throughput! * 1MB = 1,000,000 bytes 20
21 NUMA effects on performance At these speeds, NUMA effects cannot be ignored sqlio2.exe -T100 -t16 s60 -b8 -o4 frandom -BN LS (four files per volume) Test Case NUMA aware multichannel dispatcher NUMA unaware multichannel dispatcher Avg. MB/sec* Avg. IOs/sec (8 KiB) Avg. %CPU (client) Avg. Latency (ms) 3, K 60 < 1 3, K 76 < 1 To achieve peak performance, the SMB3 / SMB Direct stack must avoid cross-numa node memory accesses whenever possible. * 1MB = 1,000,000 bytes 21
22 NUMA and SMB3 Multichannel SMB3 Multichannel can be used to improve performance on NUMA systems SMB3 session is split across multiple channels Channels affinitized to a set of NUMA nodes Client dispatches IO requests to maximize performance and minimize cross NUMA node memory accesses One example of how the Windows Server 2012 SMB3 / SMB Direct stack has been optimized for high performance on NUMA systems 22
23 That s great! Now what? Are there simple improvements we could make to the SMB Direct protocol? Goals: Ease of implementation Increase IOPS Decrease latency Decrease CPU utilization 23
24 Where can we reduce IO costs? App SMB Client Client RNIC Server RNIC ReadFile() Aggressive invalidation: Consumes CPU cycles Consumes RNIC/bus cycles Increases interrupts/sec Increases IO latency ReadFile() status Register buffer Send SMB request Register status Send status Invalidate registration Invalidate status RDMA write data Send SMB response Consumes CPU cycles 24
25 Why aggressively invalidate? Application will likely reuse same buffers for subsequent IO requests. Why not cache and reuse buffer registrations? Peer can RDMA write after IO has completed Data corruption / system crash / connection loss Peer can RDMA read after IO has completed Data leak / connection loss Registration caches are not robust enough for storage and enterprise server applications. 25
26 Why aggressively invalidate? Invalidation provides strict correctness guarantees with respect to data: Data is in a consistent state following DMA Application can safely access its data Peer no longer has access to the region No data corruption, crashes, or leaks due to peerinitiated RDMA operations Aggressive invalidation is a necessary expense, but we might be able to reduce its cost 26
27 Use Send with Invalidate? App SMB Client Client RNIC Server RNIC ReadFile() Register buffer Send SMB request Register status Send status RDMA write data ReadFile() status Send SMB response with token to invalidate RNIC invalidates registration before indicating received data Consumes CPU cycles 27
28 Benefits of send with invalidate... Reduces RNIC work requests by 1/3 rd for small IOs (IOs that require one memory descriptor) Fewer CPU cycles Fewer RNIC/bus cycles Fewer interrupts Lower IO latency Already supported by major RDMA standards iwarp InfiniBand RoCE 28
29 Benefits of send with invalidate No change to SMB Direct protocol Make send with invalidate an optional feature. Client continues to invalidate the buffer if the server does not. Minimal change to SMB3 protocol SMB3 read/write request indicates when the server is requested to invalidate a request s memory descriptor via the server s response. Not a committed plan (investigation only) Feedback? 29
30 Summary SMB3 and SMB Direct allow Windows Server 2012 to efficiently host enterprise application workloads. SMB3 / SMB Direct protocols could be enhanced in simple ways to further improve performance. Increase IOPS Decrease CPU overhead Decrease latency 30
31 Questions? 31
SMB Direct Update. Tom Talpey and Greg Kramer Microsoft Storage Developer Conference. Microsoft Corporation. All Rights Reserved.
SMB Direct Update Tom Talpey and Greg Kramer Microsoft 1 Outline Part I Ecosystem status and updates SMB 3.02 status SMB Direct applications RDMA protocols and networks Part II SMB Direct details Protocol
More information[MS-SMBD]: SMB2 Remote Direct Memory Access (RDMA) Transport Protocol
[MS-SMBD]: SMB2 Remote Direct Memory Access (RDMA) Transport Protocol Intellectual Property Rights Notice for Open Specifications Documentation Technical Documentation. Microsoft publishes Open Specifications
More informationAdvanced Computer Networks. End Host Optimization
Oriana Riva, Department of Computer Science ETH Zürich 263 3501 00 End Host Optimization Patrick Stuedi Spring Semester 2017 1 Today End-host optimizations: NUMA-aware networking Kernel-bypass Remote Direct
More informationRemote Persistent Memory With Nothing But Net Tom Talpey Microsoft
Remote Persistent Memory With Nothing But Net Tom Talpey Microsoft 1 Outline Aspiration RDMA NIC as a Persistent Memory storage adapter Steps to there: Flush Write-after-flush Integrity Privacy QoS Some
More informationHIGH-PERFORMANCE NETWORKING :: USER-LEVEL NETWORKING :: REMOTE DIRECT MEMORY ACCESS
HIGH-PERFORMANCE NETWORKING :: USER-LEVEL NETWORKING :: REMOTE DIRECT MEMORY ACCESS CS6410 Moontae Lee (Nov 20, 2014) Part 1 Overview 00 Background User-level Networking (U-Net) Remote Direct Memory Access
More informationREMOTE PERSISTENT MEMORY ACCESS WORKLOAD SCENARIOS AND RDMA SEMANTICS
13th ANNUAL WORKSHOP 2017 REMOTE PERSISTENT MEMORY ACCESS WORKLOAD SCENARIOS AND RDMA SEMANTICS Tom Talpey Microsoft [ March 31, 2017 ] OUTLINE Windows Persistent Memory Support A brief summary, for better
More informationSMB3 Extensions for Low Latency. Tom Talpey Microsoft May 12, 2016
SMB3 Extensions for Low Latency Tom Talpey Microsoft Problem Statement Storage Class Memory A new, disruptive class of storage Nonvolatile medium with RAM-like performance Low latency, high throughput,
More informationRemote Access to Ultra-Low-Latency Storage Tom Talpey Microsoft
Remote Access to Ultra-Low-Latency Storage Tom Talpey Microsoft Outline Problem Statement RDMA Storage Protocols Today Sources of Latency RDMA Storage Protocols Extended Other Protocols Needed 2 Related
More informationWINDOWS SERVER 2012 ECHOSTREAMS FLACHESAN2
WINDOWS SERVER 212 ECHOSTREAMS FLACHESAN2 Dan Lovinger Windows File Server Performance April 213 1 INTRODUCTION This article presents a summary of the performance of the Windows Server 212 File Server
More informationNFS/RDMA. Tom Talpey Network Appliance
NFS/RDMA Tom Talpey Network Appliance tmt@netapp.com 1 RDMA Remote Direct Memory Access Read and write of memory across network Hardware assisted OS bypass Application control Secure Examples: Infiniband
More informationRemote Persistent Memory SNIA Nonvolatile Memory Programming TWG
Remote Persistent Memory SNIA Nonvolatile Memory Programming TWG Tom Talpey Microsoft 2018 Storage Developer Conference. SNIA. All Rights Reserved. 1 Outline SNIA NVMP TWG activities Remote Access for
More informationAn RDMA Protocol Specification (Version 1.0)
draft-recio-iwarp-rdmap-v.0 Status of this Memo R. Recio IBM Corporation P. Culley Hewlett-Packard Company D. Garcia Hewlett-Packard Company J. Hilland Hewlett-Packard Company October 0 An RDMA Protocol
More informationEvaluating the Impact of RDMA on Storage I/O over InfiniBand
Evaluating the Impact of RDMA on Storage I/O over InfiniBand J Liu, DK Panda and M Banikazemi Computer and Information Science IBM T J Watson Research Center The Ohio State University Presentation Outline
More informationSMB3 Update David Kruse Microsoft
SMB3 Update David Kruse Microsoft Agenda Why 3.02? SMB 3.02 Changes Read/Write Flags Asymmetric Shares Relevant Windows 2012 R2 Features Diagnosibility Improvements SMB for IPC? Q/A 2 Why 3.02? Capabilities
More informationPERSISTENT MEMORY PROGRAMMING
14th ANNUAL WORKSHOP 2018 PERSISTENT MEMORY PROGRAMMING THE REMOTE ACCESS PERSPECTIVE Tom Talpey, Architect Microsoft April 10, 2018 OUTLINE SNIA NVMP Programming Model PMEM Remote Access considerations
More informationMemory Management Strategies for Data Serving with RDMA
Memory Management Strategies for Data Serving with RDMA Dennis Dalessandro and Pete Wyckoff (presenting) Ohio Supercomputer Center {dennis,pw}@osc.edu HotI'07 23 August 2007 Motivation Increasing demands
More informationApplication Acceleration Beyond Flash Storage
Application Acceleration Beyond Flash Storage Session 303C Mellanox Technologies Flash Memory Summit July 2014 Accelerating Applications, Step-by-Step First Steps Make compute fast Moore s Law Make storage
More informationFaRM: Fast Remote Memory
FaRM: Fast Remote Memory Problem Context DRAM prices have decreased significantly Cost effective to build commodity servers w/hundreds of GBs E.g. - cluster with 100 machines can hold tens of TBs of main
More informationSNIA NVM Programming Model Workgroup Update. #OFADevWorkshop
SNIA NVM Programming Model Workgroup Update #OFADevWorkshop Persistent Memory (PM) Vision Fast Like Memory PM Brings Storage PM Durable Like Storage To Memory Slots 2 Latency Thresholds Cause Disruption
More informationNFS/RDMA over 40Gbps iwarp Wael Noureddine Chelsio Communications
NFS/RDMA over 40Gbps iwarp Wael Noureddine Chelsio Communications Outline RDMA Motivating trends iwarp NFS over RDMA Overview Chelsio T5 support Performance results 2 Adoption Rate of 40GbE Source: Crehan
More information2017 Storage Developer Conference. Mellanox Technologies. All Rights Reserved.
Ethernet Storage Fabrics Using RDMA with Fast NVMe-oF Storage to Reduce Latency and Improve Efficiency Kevin Deierling & Idan Burstein Mellanox Technologies 1 Storage Media Technology Storage Media Access
More informationLearn Your Alphabet - SRIOV, NPIV, RoCE, iwarp to Pump Up Virtual Infrastructure Performance
Learn Your Alphabet - SRIOV, NPIV, RoCE, iwarp to Pump Up Virtual Infrastructure Performance TechTarget Dennis Martin 1 Agenda About Demartek I/O Virtualization Concepts RDMA Concepts Examples Demartek
More informationMicrosoft SMB Looking Forward. Tom Talpey Microsoft
Microsoft SMB Looking Forward Tom Talpey Microsoft Outline A look at SMB3 today A look at things in the works in Windows The SMB1 situation Other uses of SMB3 sambaxp 2018 Göttingen 2 SMB3 Today SMB3 is
More informationMultifunction Networking Adapters
Ethernet s Extreme Makeover: Multifunction Networking Adapters Chuck Hudson Manager, ProLiant Networking Technology Hewlett-Packard 2004 Hewlett-Packard Development Company, L.P. The information contained
More informationStorage Protocol Offload for Virtualized Environments Session 301-F
Storage Protocol Offload for Virtualized Environments Session 301-F Dennis Martin, President August 2016 1 Agenda About Demartek Offloads I/O Virtualization Concepts RDMA Concepts Overlay Networks and
More informationECE 650 Systems Programming & Engineering. Spring 2018
ECE 650 Systems Programming & Engineering Spring 2018 Networking Transport Layer Tyler Bletsch Duke University Slides are adapted from Brian Rogers (Duke) TCP/IP Model 2 Transport Layer Problem solved:
More informationSoftRDMA: Rekindling High Performance Software RDMA over Commodity Ethernet
SoftRDMA: Rekindling High Performance Software RDMA over Commodity Ethernet Mao Miao, Fengyuan Ren, Xiaohui Luo, Jing Xie, Qingkai Meng, Wenxue Cheng Dept. of Computer Science and Technology, Tsinghua
More informationPARAVIRTUAL RDMA DEVICE
12th ANNUAL WORKSHOP 2016 PARAVIRTUAL RDMA DEVICE Aditya Sarwade, Adit Ranadive, Jorgen Hansen, Bhavesh Davda, George Zhang, Shelley Gong VMware, Inc. [ April 5th, 2016 ] MOTIVATION User Kernel Socket
More informationNFS/RDMA Next Steps. Chuck Lever Oracle
NFS/RDMA Next Steps Chuck Lever Oracle What Is NFS/RDMA? Direct Memory Access (DMA) a device transfers data directly to or from host memory Remote Direct Memory Access (RDMA) a device transfers data directly
More informationSMB 3.0 Performance Dan Lovinger Principal Architect Microsoft
SMB 3.0 Performance Dan Lovinger Principal Architect Microsoft Overview Stats & Methods Scenario: OLTP Database Scenario: Cluster Motion SMB 3.0 Multi Channel Agenda: challenges during the development
More informationA Study of iscsi Extensions for RDMA (iser) Patricia Thaler (Agilent).
A Study of iscsi Extensions for RDMA (iser) Mallikarjun Chadalapaka (HP) Michael Ko (IBM) Patricia Thaler (Agilent). Uri Elzur (Broadcom) Hemal Shah (Intel) Slide 1 August 27, 2003 NICELI, ACM SIGCOMM
More informationLow latency, high bandwidth communication. Infiniband and RDMA programming. Bandwidth vs latency. Knut Omang Ifi/Oracle 2 Nov, 2015
Low latency, high bandwidth communication. Infiniband and RDMA programming Knut Omang Ifi/Oracle 2 Nov, 2015 1 Bandwidth vs latency There is an old network saying: Bandwidth problems can be cured with
More informationIsoStack Highly Efficient Network Processing on Dedicated Cores
IsoStack Highly Efficient Network Processing on Dedicated Cores Leah Shalev Eran Borovik, Julian Satran, Muli Ben-Yehuda Outline Motivation IsoStack architecture Prototype TCP/IP over 10GE on a single
More informationVoltaire. Fast I/O for XEN using RDMA Technologies. The Grid Interconnect Company. April 2005 Yaron Haviv, Voltaire, CTO
Voltaire The Grid Interconnect Company Fast I/O for XEN using RDMA Technologies April 2005 Yaron Haviv, Voltaire, CTO yaronh@voltaire.com The Enterprise Grid Model and ization VMs need to interact efficiently
More informationSpark Over RDMA: Accelerate Big Data SC Asia 2018 Ido Shamay Mellanox Technologies
Spark Over RDMA: Accelerate Big Data SC Asia 2018 Ido Shamay 1 Apache Spark - Intro Spark within the Big Data ecosystem Data Sources Data Acquisition / ETL Data Storage Data Analysis / ML Serving 3 Apache
More informationOS impact on performance
PhD student CEA, DAM, DIF, F-91297, Arpajon, France Advisor : William Jalby CEA supervisor : Marc Pérache 1 Plan Remind goal of OS Reproducibility Conclusion 2 OS : between applications and hardware 3
More informationThe Case for RDMA. Jim Pinkerton RDMA Consortium 5/29/2002
The Case for RDMA Jim Pinkerton RDMA Consortium 5/29/2002 Agenda What is the problem? CPU utilization and memory BW bottlenecks Offload technology has failed (many times) RDMA is a proven sol n to the
More informationNTRDMA v0.1. An Open Source Driver for PCIe NTB and DMA. Allen Hubbe at Linux Piter 2015 NTRDMA. Messaging App. IB Verbs. dmaengine.h ntb.
Messaging App IB Verbs NTRDMA dmaengine.h ntb.h DMA DMA DMA NTRDMA v0.1 An Open Source Driver for PCIe and DMA Allen Hubbe at Linux Piter 2015 1 INTRODUCTION Allen Hubbe Senior Software Engineer EMC Corporation
More informationOFED Storage Protocols
OFED Storage Protocols R. Pearson System Fabric Works, Inc. Agenda Why OFED Storage Introduction to OFED Storage Protocols OFED Storage Protocol Update 2 Why OFED Storage 3 Goals of I/O Consolidation Cluster
More informationFaSST: Fast, Scalable, and Simple Distributed Transactions with Two-Sided (RDMA) Datagram RPCs
FaSST: Fast, Scalable, and Simple Distributed Transactions with Two-Sided (RDMA) Datagram RPCs Anuj Kalia (CMU), Michael Kaminsky (Intel Labs), David Andersen (CMU) RDMA RDMA is a network feature that
More informationDB2 purescale: High Performance with High-Speed Fabrics. Author: Steve Rees Date: April 5, 2011
DB2 purescale: High Performance with High-Speed Fabrics Author: Steve Rees Date: April 5, 2011 www.openfabrics.org IBM 2011 Copyright 1 Agenda Quick DB2 purescale recap DB2 purescale comes to Linux DB2
More informationBrent Callaghan Sun Microsystems, Inc. Sun Microsystems, Inc
Brent Callaghan. brent@eng.sun.com Page 1 of 19 A Problem: Data Center Performance CPU 1 Gb Fibre Channel 100 MB/sec Storage Array CPU NFS 1 Gb Ethernet 50 MB/sec (via Gigaswift) NFS Server Page 2 of 19
More informationUser Datagram Protocol
Topics Transport Layer TCP s three-way handshake TCP s connection termination sequence TCP s TIME_WAIT state TCP and UDP buffering by the socket layer 2 Introduction UDP is a simple, unreliable datagram
More informationThe Design and Implementation of AQuA: An Adaptive Quality of Service Aware Object-Based Storage Device
The Design and Implementation of AQuA: An Adaptive Quality of Service Aware Object-Based Storage Device Joel Wu and Scott Brandt Department of Computer Science University of California Santa Cruz MSST2006
More informationBest Practices for Deployments using DCB and RoCE
Best Practices for Deployments using DCB and RoCE Contents Introduction... Converged Networks... RoCE... RoCE and iwarp Comparison... RoCE Benefits for the Data Center... RoCE Evaluation Design... RoCE
More informationArchitected for Performance. NVMe over Fabrics. September 20 th, Brandon Hoff, Broadcom.
Architected for Performance NVMe over Fabrics September 20 th, 2017 Brandon Hoff, Broadcom Brandon.Hoff@Broadcom.com Agenda NVMe over Fabrics Update Market Roadmap NVMe-TCP The benefits of NVMe over Fabrics
More informationThe NE010 iwarp Adapter
The NE010 iwarp Adapter Gary Montry Senior Scientist +1-512-493-3241 GMontry@NetEffect.com Today s Data Center Users Applications networking adapter LAN Ethernet NAS block storage clustering adapter adapter
More informationURDMA: RDMA VERBS OVER DPDK
13 th ANNUAL WORKSHOP 2017 URDMA: RDMA VERBS OVER DPDK Patrick MacArthur, Ph.D. Candidate University of New Hampshire March 28, 2017 ACKNOWLEDGEMENTS urdma was initially developed during an internship
More informationCERN openlab Summer 2006: Networking Overview
CERN openlab Summer 2006: Networking Overview Martin Swany, Ph.D. Assistant Professor, Computer and Information Sciences, U. Delaware, USA Visiting Helsinki Institute of Physics (HIP) at CERN swany@cis.udel.edu,
More informationTolerating Malicious Drivers in Linux. Silas Boyd-Wickizer and Nickolai Zeldovich
XXX Tolerating Malicious Drivers in Linux Silas Boyd-Wickizer and Nickolai Zeldovich How could a device driver be malicious? Today's device drivers are highly privileged Write kernel memory, allocate memory,...
More informationMaximum Performance. How to get it and how to avoid pitfalls. Christoph Lameter, PhD
Maximum Performance How to get it and how to avoid pitfalls Christoph Lameter, PhD cl@linux.com Performance Just push a button? Systems are optimized by default for good general performance in all areas.
More informationCS 537 Fall 2017 Review Session
CS 537 Fall 2017 Review Session Deadlock Conditions for deadlock: Hold and wait No preemption Circular wait Mutual exclusion QUESTION: Fix code List_insert(struct list * head, struc node * node List_move(struct
More informationApplication Advantages of NVMe over Fabrics RDMA and Fibre Channel
Application Advantages of NVMe over Fabrics RDMA and Fibre Channel Brandon Hoff Broadcom Limited Tuesday, June 14 2016 10:55 11:35 a.m. Agenda r Applications that have a need for speed r The Benefits of
More informationWindows Support for PM. Tom Talpey, Microsoft
Windows Support for PM Tom Talpey, Microsoft Agenda Windows and Windows Server PM Industry Standards Support PMDK Support Hyper-V PM Support SQL Server PM Support Storage Spaces Direct PM Support SMB3
More informationOperating Systems. 16. Networking. Paul Krzyzanowski. Rutgers University. Spring /6/ Paul Krzyzanowski
Operating Systems 16. Networking Paul Krzyzanowski Rutgers University Spring 2015 1 Local Area Network (LAN) LAN = communications network Small area (building, set of buildings) Same, sometimes shared,
More informationAdvanced Computer Networks. RDMA, Network Virtualization
Advanced Computer Networks 263 3501 00 RDMA, Network Virtualization Patrick Stuedi Spring Semester 2013 Oriana Riva, Department of Computer Science ETH Zürich Last Week Scaling Layer 2 Portland VL2 TCP
More informationWindows Support for PM. Tom Talpey, Microsoft
Windows Support for PM Tom Talpey, Microsoft Agenda Industry Standards Support PMDK Open Source Support Hyper-V Support SQL Server Support Storage Spaces Direct Support SMB3 and RDMA Support 2 Windows
More informationNFS/RDMA Draft Status
NFS/RDMA Draft Status Tom Talpey Network Appliance tmt@netapp.com 1 NFS/RDMA Internet-Drafts RDMA Transport for ONC RPC Basic ONC RPC transport definition for RDMA Transparent, or nearly so, for all ONC
More informationRevisiting Network Support for RDMA
Revisiting Network Support for RDMA Radhika Mittal 1, Alex Shpiner 3, Aurojit Panda 1, Eitan Zahavi 3, Arvind Krishnamurthy 2, Sylvia Ratnasamy 1, Scott Shenker 1 (1: UC Berkeley, 2: Univ. of Washington,
More informationEnergy-Efficient Data Transfers in Radio Astronomy with Software UDP RDMA Third Workshop on Innovating the Network for Data-Intensive Science, INDIS16
Energy-Efficient Data Transfers in Radio Astronomy with Software UDP RDMA Third Workshop on Innovating the Network for Data-Intensive Science, INDIS16 Przemek Lenkiewicz, Researcher@IBM Netherlands Bernard
More informationPAC094 Performance Tips for New Features in Workstation 5. Anne Holler Irfan Ahmad Aravind Pavuluri
PAC094 Performance Tips for New Features in Workstation 5 Anne Holler Irfan Ahmad Aravind Pavuluri Overview of Talk Virtual machine teams 64-bit guests SMP guests e1000 NIC support Fast snapshots Virtual
More informationJANUARY 28, 2014, SAN JOSE, CA. Microsoft Lead Partner Architect OS Vendors: What NVM Means to Them
JANUARY 28, 2014, SAN JOSE, CA PRESENTATION James TITLE Pinkerton GOES HERE Microsoft Lead Partner Architect OS Vendors: What NVM Means to Them Why should NVM be Interesting to OS Vendors? New levels of
More informationHow Flash-Based Storage Performs on Real Applications Session 102-C
How Flash-Based Storage Performs on Real Applications Session 102-C Dennis Martin, President August 2016 1 Agenda About Demartek Enterprise Datacenter Environments Storage Performance Metrics Synthetic
More informationExtending RDMA for Persistent Memory over Fabrics. Live Webcast October 25, 2018
Extending RDMA for Persistent Memory over Fabrics Live Webcast October 25, 2018 Today s Presenters John Kim SNIA NSF Chair Mellanox Tony Hurson Intel Rob Davis Mellanox SNIA-At-A-Glance 3 SNIA Legal Notice
More informationApplication Access to Persistent Memory The State of the Nation(s)!
Application Access to Persistent Memory The State of the Nation(s)! Stephen Bates, Paul Grun, Tom Talpey, Doug Voigt Microsemi, Cray, Microsoft, HPE The Suspects Stephen Bates Microsemi Paul Grun Cray
More informationAccelerating Real-Time Big Data. Breaking the limitations of captive NVMe storage
Accelerating Real-Time Big Data Breaking the limitations of captive NVMe storage 18M IOPs in 2u Agenda Everything related to storage is changing! The 3rd Platform NVM Express architected for solid state
More informationFast packet processing in the cloud. Dániel Géhberger Ericsson Research
Fast packet processing in the cloud Dániel Géhberger Ericsson Research Outline Motivation Service chains Hardware related topics, acceleration Virtualization basics Software performance and acceleration
More informationIntra-MIC MPI Communication using MVAPICH2: Early Experience
Intra-MIC MPI Communication using MVAPICH: Early Experience Sreeram Potluri, Karen Tomko, Devendar Bureddy, and Dhabaleswar K. Panda Department of Computer Science and Engineering Ohio State University
More informationRDMA programming concepts
RDMA programming concepts Robert D. Russell InterOperability Laboratory & Computer Science Department University of New Hampshire Durham, New Hampshire 03824, USA 2013 Open Fabrics Alliance,
More informationInfiniBand Networked Flash Storage
InfiniBand Networked Flash Storage Superior Performance, Efficiency and Scalability Motti Beck Director Enterprise Market Development, Mellanox Technologies Flash Memory Summit 2016 Santa Clara, CA 1 17PB
More informationQuickSpecs. HP Z 10GbE Dual Port Module. Models
Overview Models Part Number: 1Ql49AA Introduction The is a 10GBASE-T adapter utilizing the Intel X722 MAC and X557-AT2 PHY pairing to deliver full line-rate performance, utilizing CAT 6A UTP cabling (or
More information19: Networking. Networking Hardware. Mark Handley
19: Networking Mark Handley Networking Hardware Lots of different hardware: Modem byte at a time, FDDI, SONET packet at a time ATM (including some DSL) 53-byte cell at a time Reality is that most networking
More informationIndependent Submission Request for Comments: 7609 Category: Informational ISSN: August 2015
Independent Submission Request for Comments: 7609 Category: Informational ISSN: 2070-1721 M. Fox C. Kassimis J. Stevens IBM August 2015 Abstract IBM s Shared Memory Communications over RDMA (SMC-R) Protocol
More informationSNIA Developers Conference - Growth of the iscsi RDMA (iser) Ecosystem
SNIA Developers Conference - Growth of the iscsi RDMA (iser) Ecosystem Rob Davis Mellanox Technologies robd@mellanox.com The FASTEST Storage Protocol: iser The FASTEST Storage: Flash What it is: iscsi
More informationby Brian Hausauer, Chief Architect, NetEffect, Inc
iwarp Ethernet: Eliminating Overhead In Data Center Designs Latest extensions to Ethernet virtually eliminate the overhead associated with transport processing, intermediate buffer copies, and application
More informationImplementing Efficient and Scalable Flow Control Schemes in MPI over InfiniBand
Implementing Efficient and Scalable Flow Control Schemes in MPI over InfiniBand Jiuxing Liu and Dhabaleswar K. Panda Computer Science and Engineering The Ohio State University Presentation Outline Introduction
More informationWhat a Long Strange Trip It s Been: Moving RDMA into Broad Data Center Deployments
What a Long Strange Trip It s Been: Moving RDMA into Broad Data Center Deployments Author: Jim Pinkerton, Partner Architect, Microsoft Date: 3/25/2012 www.openfabrics.org 1 What a Long Strange Trip Who
More informationEvaluation of the Chelsio T580-CR iscsi Offload adapter
October 2016 Evaluation of the Chelsio T580-CR iscsi iscsi Offload makes a difference Executive Summary As application processing demands increase and the amount of data continues to grow, getting this
More informationAn FPGA-Based Optical IOH Architecture for Embedded System
An FPGA-Based Optical IOH Architecture for Embedded System Saravana.S Assistant Professor, Bharath University, Chennai 600073, India Abstract Data traffic has tremendously increased and is still increasing
More informationMemcached Design on High Performance RDMA Capable Interconnects
Memcached Design on High Performance RDMA Capable Interconnects Jithin Jose, Hari Subramoni, Miao Luo, Minjia Zhang, Jian Huang, Md. Wasi- ur- Rahman, Nusrat S. Islam, Xiangyong Ouyang, Hao Wang, Sayantan
More informationBirds of a Feather Presentation
Mellanox InfiniBand QDR 4Gb/s The Fabric of Choice for High Performance Computing Gilad Shainer, shainer@mellanox.com June 28 Birds of a Feather Presentation InfiniBand Technology Leadership Industry Standard
More informationMeltdown and Spectre Interconnect Performance Evaluation Jan Mellanox Technologies
Meltdown and Spectre Interconnect Evaluation Jan 2018 1 Meltdown and Spectre - Background Most modern processors perform speculative execution This speculation can be measured, disclosing information about
More informationExtremely Fast Distributed Storage for Cloud Service Providers
Solution brief Intel Storage Builders StorPool Storage Intel SSD DC S3510 Series Intel Xeon Processor E3 and E5 Families Intel Ethernet Converged Network Adapter X710 Family Extremely Fast Distributed
More informationGeneric RDMA Enablement in Linux
Generic RDMA Enablement in Linux (Why do we need it, and how) Krishna Kumar Linux Technology Center, IBM February 28, 2006 AGENDA RDMA : Definition Why RDMA, and how does it work OpenRDMA history Architectural
More informationInternet Engineering Task Force (IETF) Category: Standards Track. A. Eiriksson Chelsio Communications, Inc. R. Sharp Intel Corporation June 2014
Internet Engineering Task Force (IETF) Request for Comments: 7306 Category: Standards Track ISSN: 2070-1721 H. Shah Broadcom Corporation F. Marti W. Noureddine A. Eiriksson Chelsio Communications, Inc.
More informationW H I T E P A P E R. Comparison of Storage Protocol Performance in VMware vsphere 4
W H I T E P A P E R Comparison of Storage Protocol Performance in VMware vsphere 4 Table of Contents Introduction................................................................... 3 Executive Summary............................................................
More informationOptimizing Performance: Intel Network Adapters User Guide
Optimizing Performance: Intel Network Adapters User Guide Network Optimization Types When optimizing network adapter parameters (NIC), the user typically considers one of the following three conditions
More informationLatest Developments with NVMe/TCP Sagi Grimberg Lightbits Labs
Latest Developments with NVMe/TCP Sagi Grimberg Lightbits Labs 2018 Storage Developer Conference. Insert Your Company Name. All Rights Reserved. 1 NVMe-oF - Short Recap Early 2014: Initial NVMe/RDMA pre-standard
More informationiscsi or iser? Asgeir Eiriksson CTO Chelsio Communications Inc
iscsi or iser? Asgeir Eiriksson CTO Chelsio Communications Inc Introduction iscsi is compatible with 15 years of deployment on all OSes and preserves software investment iser and iscsi are layered on top
More informationTransport: How Applications Communicate
Transport: How Applications Communicate Week 2 Philip Levis 1 7 Layers (or 4) 7. 6. 5. 4. 3. 2. 1. Application Presentation Session Transport Network Link Physical segments packets frames bits/bytes Application
More informationA GENERAL-PURPOSE API FOR IWARP AND INFINIBAND
A GENERAL-PURPOSE API FOR IWARP AND INFINIBAND Robert D. Russell University of New Hampshire InterOperability Laboratory 2 Technology Drive, Suite 2 Durham, NH 03824-476 rdr@iol.unh.edu (603) 862-3774
More informationNetworking for Data Acquisition Systems. Fabrice Le Goff - 14/02/ ISOTDAQ
Networking for Data Acquisition Systems Fabrice Le Goff - 14/02/2018 - ISOTDAQ Outline Generalities The OSI Model Ethernet and Local Area Networks IP and Routing TCP, UDP and Transport Efficiency Networking
More informationPerformance Analysis and Evaluation of Mellanox ConnectX InfiniBand Architecture with Multi-Core Platforms
Performance Analysis and Evaluation of Mellanox ConnectX InfiniBand Architecture with Multi-Core Platforms Sayantan Sur, Matt Koop, Lei Chai Dhabaleswar K. Panda Network Based Computing Lab, The Ohio State
More informationIntroduction to High-Speed InfiniBand Interconnect
Introduction to High-Speed InfiniBand Interconnect 2 What is InfiniBand? Industry standard defined by the InfiniBand Trade Association Originated in 1999 InfiniBand specification defines an input/output
More informationCache Performance and Memory Management: From Absolute Addresses to Demand Paging. Cache Performance
6.823, L11--1 Cache Performance and Memory Management: From Absolute Addresses to Demand Paging Asanovic Laboratory for Computer Science M.I.T. http://www.csg.lcs.mit.edu/6.823 Cache Performance 6.823,
More informationIsilon Performance. Name
1 Isilon Performance Name 2 Agenda Architecture Overview Next Generation Hardware Performance Caching Performance Streaming Reads Performance Tuning OneFS Architecture Overview Copyright 2014 EMC Corporation.
More informationMuch Faster Networking
Much Faster Networking David Riddoch driddoch@solarflare.com Copyright 2016 Solarflare Communications, Inc. All rights reserved. What is kernel bypass? The standard receive path The standard receive path
More informationRoCE Update. Liran Liss, Mellanox Technologies March,
RoCE Update Liran Liss, Mellanox Technologies March, 2012 www.openfabrics.org 1 Agenda RoCE Ecosystem QoS Virtualization High availability Latest news 2 RoCE in the Data Center Lossless configuration recommended
More informationWindows Persistent Memory Support
Windows Persistent Memory Support Neal Christiansen Microsoft Agenda Review: Existing Windows PM Support What s New New PM APIs Large & Huge Page Support Dax aware Write-ahead LOG Improved Driver Model
More informationI/O Device Controllers. I/O Systems. I/O Ports & Memory-Mapped I/O. Direct Memory Access (DMA) Operating Systems 10/20/2010. CSC 256/456 Fall
I/O Device Controllers I/O Systems CS 256/456 Dept. of Computer Science, University of Rochester 10/20/2010 CSC 2/456 1 I/O devices have both mechanical component & electronic component The electronic
More information