Throughput & Latency Control in Ethernet Backplane Interconnects. Manoj Wadekar Gary McAlpine. Intel

Similar documents
Toward a Reliable Data Transport Architecture for Optical Burst-Switched Networks

Requirement Discussion of Flow-Based Flow Control(FFC)

CSE 473 Introduction to Computer Networks. Exam 2. Your name here: 11/7/2012

RDMA over Commodity Ethernet at Scale

The NE010 iwarp Adapter

Best Practices for Validating the Performance of Data Center Infrastructure. Henry He Ixia

Network Management & Monitoring

ETSF10 Internet Protocols Transport Layer Protocols

iscsi : A loss-less Ethernet fabric with DCB Jason Blosil, NetApp Gary Gumanow, Dell

Revisiting Network Support for RDMA

P802.1Qcz Congestion Isolation

TCP and QCN in a Multihop Output Generated Hot Spot Scenario

DeTail Reducing the Tail of Flow Completion Times in Datacenter Networks. David Zats, Tathagata Das, Prashanth Mohan, Dhruba Borthakur, Randy Katz

Computer Networks. ENGG st Semester, 2010 Hayden Kwok-Hay So

Linux Plumbers Conference TCP-NV Congestion Avoidance for Data Centers

Multi-class Applications for Parallel Usage of a Guaranteed Rate and a Scavenger Service

End-to-End Adaptive Packet Aggregation for High-Throughput I/O Bus Network Using Ethernet

Agenda. What are we looking at? Introduction. Aim of the project. IP Routing

Chapter 6 Connecting Device

Congestion Control in TCP

EXPERIENCES EVALUATING DCTCP. Lawrence Brakmo, Boris Burkov, Greg Leclercq and Murat Mugan Facebook

Congestion Control in TCP

Student ID: CS457: Computer Networking Date: 3/20/2007 Name:

Contents. Introduction. Methodology. Check for Output Drops. Introduction Methodology

Transport layer issues

Network Management & Monitoring Network Delay

SELECTION OF METRICS (CONT) Gaia Maselli

Presentation_ID. 2002, Cisco Systems, Inc. All rights reserved.

LANCOM Techpaper IEEE n Indoor Performance

Chapter 4. Routers with Tiny Buffers: Experiments. 4.1 Testbed experiments Setup

IEEE P802.1Qcz Proposed Project for Congestion Isolation

Network Design Considerations for Grid Computing

Managing Caching Performance and Differentiated Services

TS Open Day Data Center Fibre Channel over IP

Performance Analysis of Darwin: Transient State

Ethernet Hub. Campus Network Design. Hubs. Sending and receiving Ethernet frames via a hub

Configuring Queuing and Flow Control

Handles all kinds of traffic on a single network with one class

BITUG Big Sig December 2013 NonStop Performance Update David Sly HP UK Tech Services

Be Fast, Cheap and in Control with SwitchKV. Xiaozhou Li

Enabling Block Storage over Ethernet: The Case for Per Priority Pause

Page 1. Review: Internet Protocol Stack. Transport Layer Services. Design Issue EEC173B/ECS152C. Review: TCP

Packet Scheduling in Data Centers. Lecture 17, Computer Networks (198:552)

Configuring QoS CHAPTER

Chapter 7 CONCLUSION

Virtual Link: An Enabler of Enterprise Utility Computing

FM4000. A Scalable, Low-latency 10 GigE Switch for High-performance Data Centers

Direct Link Communication I: Basic Techniques. Data Transmission. ignore carrier frequency, coding etc.

Configuring QoS. Understanding QoS CHAPTER

Configuring Queuing and Flow Control

The Network Layer and Routers

Configuring QoS CHAPTER

Congestion. Can t sustain input rate > output rate Issues: - Avoid congestion - Control congestion - Prioritize who gets limited resources

IEEE Congestion Management Study Group November Plenary Report. San Antonio, TX November, 2004

EqualLogic Storage and Non-Stacking Switches. Sizing and Configuration

Computer Networks Principles

Attaining the Promise and Avoiding the Pitfalls of TCP in the Datacenter. Glenn Judd Morgan Stanley

Flow Control in Relay Enhanced IEEE Networks

Lab Test Report DR100401D. Cisco Nexus 5010 and Arista 7124S

Be Fast, Cheap and in Control with SwitchKV Xiaozhou Li

Configuring QoS CHAPTER

RoGUE: RDMA over Generic Unconverged Ethernet

Lecture 22: Buffering & Scheduling. CSE 123: Computer Networks Alex C. Snoeren

Appendix B. Standards-Track TCP Evaluation

The Impact of Delay Variations on TCP Performance

Page 1. Review: Internet Protocol Stack. Transport Layer Services EEC173B/ECS152C. Review: TCP. Transport Layer: Connectionless Service

Industry Standards for the Exponential Growth of Data Center Bandwidth and Management. Craig W. Carlson

WSN NETWORK ARCHITECTURES AND PROTOCOL STACK

CS 421: COMPUTER NETWORKS SPRING FINAL May 24, minutes. Name: Student No: TOT

Basic Low Level Concepts

Titan: Fair Packet Scheduling for Commodity Multiqueue NICs. Brent Stephens, Arjun Singhvi, Aditya Akella, and Mike Swift July 13 th, 2017

Top-Down Network Design

GUIDE. Optimal Network Designs with Cohesity

CSE 461 Module 10. Introduction to the Transport Layer

Networking for Data Acquisition Systems. Fabrice Le Goff - 14/02/ ISOTDAQ

Configuring QoS. Finding Feature Information. Prerequisites for QoS

CHAPTER 3 EFFECTIVE ADMISSION CONTROL MECHANISM IN WIRELESS MESH NETWORKS

Internetworking Part 1

SIMULATION FRAMEWORK MODELING

Advanced Computer Networks. Flow Control

CS Lecture 1 Review of Basic Protocols

ECEN Final Exam Fall Instructor: Srinivas Shakkottai

Messaging Overview. Introduction. Gen-Z Messaging

InfiniBand Credit-Based Link-Layer Flow-Control

Baidu s Best Practice with Low Latency Networks

Wireless USB Periodic Transfer Models. Dan Froelich Intel

What s New in VMware vsphere 4.1 Performance. VMware vsphere 4.1

Sections Describing Standard Software Features

Vorlesung Kommunikationsnetze

Reduces latency and buffer overhead. Messaging occurs at a speed close to the processors being directly connected. Less error detection

Configuring Priority Flow Control

Streaming Video over the Internet. Dr. Dapeng Wu University of Florida Department of Electrical and Computer Engineering

Problem Set Name the 7 OSI layers and give the corresponding functionalities for each layer.

NETWORK PROBLEM SET Solution

Advanced Computer Networks

NETWORK PROBLEM SET Due Date 6/28/2014

Chapter 4 Advanced Settings and Features

Quality of Service in the Internet

TCP over Wireless PROF. MICHAEL TSAI 2016/6/3

Module objectives. Integrated services. Support for real-time applications. Real-time flows and the current Internet protocols

Transcription:

Throughput & Latency Control in Ethernet Backplane Interconnects Manoj Wadekar Gary McAlpine Intel Date 3/16/04

Agenda Discuss Backplane challenges to Ethernet Simulation environment and definitions Preliminary Simulation results Illustrate current solutions/tradeoffs Summary 2

Ethernet : Backplane Concerns Frame Discard in Response to Congestion TCP timeouts & retrans = big performance hits Cluster traffic prone to frequent congestion events Unacceptable in Cluster/Backplane Interconnect LAN latencies (loaded) can be in milliseconds Voice, video, real time apps sensitive to delay, delay var. Additive over many hops Servers targeting IPC with mean latency < 10 us loaded Large switch buffers Smaller buffers enable tradeoff between smaller ports, cost Backplane/Cluster networks long links not required 3

The Challenges (The Evil Buts...) 802.3x Flow Control may be available, but... Seldom used Causes blocking, reducing throughput Frame discard avoids blocking, but... Causes timeouts & retransmissions Requires large buffers to minimize discard frequency Large buffers support long links, high throughput, but... Allows large latencies & higher cost Requires prioritization to enables low latency Prioritization can enable low latency on critical traffic, but Not supported by most Apps 4

16 Port Switch ~30 M 10 Gbs 10 Gbs 802.3 Ethernet Links 160 Gbs Max 1.5 M Shared Buff 4 Workload Gens Simulation Environment 16 x 10Gbs Switch 1.5 M Shared Buff 10 Gbs 802.3 MACs 4 X 10Gbs Workload Gens L2 WLG Source NIC w/ 2KB Buff MAC w/ 2KB Buff WLG Sink Bursty Workload Exponential Packet Size Distribution (48B to ~85KB) @ ~10 Gbs per WLG 5

Definitions Throughput Aggregate Traffic (bps) received by all End Stations during test period Latency End-to-End delay is measured per frame: First byte from Source App to last byte at Sink App Shared Memory Utilization Maximum Memory utilization at switch during test Flow Control Thresholds Memory High threshold at which switch starts sending XOFF Memory Low Threshold at which switch sends XON Frames Discarded Packets dropped at switch due to congestion 6

Modeling Scenarios Throughput, latency and packet drop for: 1.5M Shared Memory w/o flow control 32K Shared Memory w/o flow control 32K Shared Memory with 802.3x Flow Control (Hi- Threshold = 16K) 1.5M Shared Memory with various Flow Control thresholds for 802.3x 7

Frame Discard ~135K Frames Discarded w/ 32 KB ~5K Frames Discarded w/ 1.5 MB 0 Frames Discarded w/ 802.3x MAC Ctrl Frame Rate = ~165K F/S (One Link) Frame discard can be completely eliminated with Flow Control even for smaller shared memory 8

Latency Zoomed In ~350 us ~13 us ~10 us Latency can be constrained through use of Flow Control 9

Throughput ~38 Gbs (Max) ~29 Gbs (Due to Discard) ~24 Gbs (Due to Blocking) Latency reduction with 802.3x(XON/XOFF) costs Throughput 10

Throughput & Latency Summary Bursty Workload, 802.3x (XON/XOFF), 1.5MB Switch Mem, Hi-Thresh = 4KB, 16KB, 64KB, 256KB, None 900 800 Flow Control Threshold vs. Max Latency & Switch Buffer Utilization 45 818 40 38.6 700 35.6 35 Latency (us) 600 500 400 300 20.7 24.2 30.6 273 317 30 25 20 15 Throughput (Gbs) Mean Lat (us) Max Lat (us) Tput (Gbs) 200 10 100 0 81.9 48.3 54.9 7.7 20.8 9.9 18.9 4KB 16KB 64KB 256KB No FC Flow Control Threshold (bytes) 5 0 802.3x currently supports throughput/latency trade-offs 11

Issues with 802.3x Latency can be controlled, but Throughput is lost in 2 ways: Loss to XOFF blocking = 38 24 = ~ 14 Gbs or ~36% Pause Frames @ 165K F/S = ~111 Mbs or ~1.11% Initial testing of 802.1p traffic loads shows a Max latency issue on high priorities (still investigating) Latencies pushed back to end-points, but Enables higher layers to deal with it Advances in ULP stacks will significantly diminish this issue How to handle QoS requirements? Can throughput loss be avoided while keeping lower latency? 12

Summary & Next Steps 802.3x can constrain fabric latencies But creates other issues Throughput & Max latency issues remain Need to study simple enhancements to existing MAC Control Sub-layer To reclaim throughput w/o sacrificing latency or packet delivery To contain Max latency Next step to evaluate and simulate simple enhancements Will present proposals and results in next plenary meeting 13