Xen Network I/O Performance Analysis and Opportunities for Improvement
|
|
- Luke Atkinson
- 5 years ago
- Views:
Transcription
1 Xen Network I/O Performance Analysis and Opportunities for Improvement J. Renato Santos G. (John) Janakiraman Yoshio Turner HP Labs Xen Summit April 17-18, Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice
2 Motivation CPU cost for TCP connection at 1 Gbps (xen-unstable (3/16/27) ; PV Linux guest; X86-32bit) CPU utilization xen xen 2 linux linux RX TX Network I/O has high CPU cost TX: 35% cost of linux RX: 31% cost of linux page 2
3 Outline Performance Analysis for network I/O RX path (netfront/netback) Network I/O RX optimizations Network I/O TX optimization page 3
4 Performance Analysis For Network I/O RX Path page 4
5 Experimental Setup Machines: HP Proliant DL58 (client and server) P4 Xeon 2.8 Ghz, 4 CPU (MT disabled), 64 GB, 512 KB L2, 2MB L3 NIC: Intel E1 (1 Gbps) Network configuration Single switch connecting client and server Server configuration Xen unstable (c.s March 16, 27) (default xen/xenu configs) Single guest (512MB), dom also with 512MB Benchmark Simple UDP micro benchmark (1gbps, 15 bytes packets) page 5
6 CPU Profile for network RX path 14 CPU util (%) app usercopy kernel xenu grantcopy kernel xen linux xen Cost of data copy is significant both in Linux and Xen Xen has the cost of an additional data copy Xen guest kernel alone uses more CPU than linux Most cost for Xen code is in dom page 6
7 Xen Code Cost 2 CPU util (%) other mm interrupt event time schedule hypercall domain_page grant_table xen xenu Major Xen overhead is: grant table & dom page map/unmap Copy grant: several expensive atomic operations (lock instr.prefix): Atomic cmpxchg operation for updating status field (grant in use) Increment/decrement grant usage counters (multiple spinlock op) page 7
8 Dom Kernel Cost CPU util. (%) linux dom other dma+swiotlb hypercall interrupt syscall schedule mm tcp bridge network netback driver Bridge/network is large component in dom cost (Xen summit 26) Can be reduced if netfilter bridge config option is disabled Xen new code: Netback, hypercall, swiotlb Higher interrupt overhead in Xen: extra code in evtchn.c Additional high cost functions in Xen (accounted in other ) spin_unlock_irqrestore(), spin_trylock() page 8
9 Bridge netfilter cost CPU util. (%) other dma+swiotlb hypercall interrupt syscall schedule mm tcp bridge network netback driver linux nf_br no nf_br What do we need to do to disable bridge netfilter by default? Should we add a netfilter hook in netback? page 9
10 Overhead in guest CPU util. (%) linux domu other grant_table hypercall interrupt syscall schedule mm tcp network driver Netfront: 5 times more expensive than e1 driver in Linux Memory op (mm): 2 times more expensive in Xen (?) grant table: high cost of atomic cmpxchg operation to revoke grant access Other : spin_unlock_irqrestore(), spin_trylock() (same as dom) page 1
11 Source of Netfront Cost CPU util. (%) linux domu other grant_table hypercall interrupt syscall schedule mm tcp network headcopy driver Netback copy packet data into netfront page fragments Netfront copies first 2 bytes of packet from fragment into main socket buffer data area Large netfront cost is due to this extra data copy page 11
12 Opportunities for Improvement on RX path page 12
13 Reduce RX head copy size 3 CPU util. (%) linux headcopy 2 bytes headcopy 14 bytes other grant_table hypercall interrupt syscall schedule mm tcp network headcopy driver No need to have all headers in main SKB data area Copy only Ethernet header (14 bytes) Network stack copies more data as needed page 13
14 Move grant data copy into guest CPU 14 CPU util. (%) app usercopy kernel xenu grantcopy kernel xen copy in dom copy in guest Dom grant access to data and guest copy it using copy grant Cost of 2 nd copy to user buffer is reduced as data is already in the guest CPU cache (assumes cache is not evicted due to user process delaying read) Additional benefit: Improves dom (driver domain) scalability as more work is done at the guest side Data copy is more expensive in guest (alignment problem) page 14
15 Grant copy align problem Packet starts at half word copy in guest netback SKB nettfront SKB grant copy 16 bytes 2 packet netback SKB copy in dom netfront fragment page grant copy copy is expensive when destination start is not at word boundary Fix: Copy also 2 prefix bytes source and destination now aligned page 15
16 Fixing grant copy alignment 14 cpu UTIL. (%) app usercopy kernel xenu grantcopy kernel xen copy in dom copy in guest align Grant copy in guest becomes more efficient than in current Xen Grant copy in dom: destination is word aligned but Source is not word aligned Can also be improved by copying additional prefix data Either 2 or (2+16) bytes page 16
17 Fixing alignment in current Xen 14 CPU util. (%) app usercopy kernelu xenu grantcopy kernel xen original word align buffer align copy in guest Source alignment reduces copy cost Source and dest. at same buffer offset has better performance Reason (?): maybe because same offset in cache? Copy cost in dom is still more expensive than copy in guest. Different cache behavior page 17
18 Copy in guest has better cache locality grant copy in guest grant copy in dom app usercopy kernelu xenu grantcopy kernel xen CPU util (%) app usercopy kernelu xenu grantcopy kernel xen cycles L3 misses L2 misses cycles L3 misses L2 misses Dom copy has more L2 cache misses than guest copy Dom copy has lower cache locality Guest post multiple pages on IO ring. All pages in ring must be used before the same page can be reused For guest copy, pages are allocated on demand and reused more often improving cache locality page 18
19 Possible grant optimizations Define new simple copy grant: Allow only one copy operation at a time No need to keep grant usage counters (remove lock) avoid cost of atomic cmpxchg operations Separate fields used for enabling grant and usage status Avoid incrementing/decrementing page ref counters Use an RCU scheme for page deallocation (lazy deallocation) page 19
20 Potential savings in grant modifications other mm time schedule hypercall domain_page grant_table original no status update no src page pin no dst page pin Results are optimistic Still need to implement grant modifications Results are based on eliminating current operations page 2
21 Coalescing netfront RX interrupts 8 CPU util. (%) app usercopy kernelu xenu grantcopy kernel xen no batch 8 pkts 16 pkts 32 pkts 64 pkts NIC (e1) already coalescing HW interrupts (~1 packets/int) Batching packets can provide additional benefit 1% for 32 packets But adds extra latency Dynamic coalescing scheme should be beneficial page 21
22 Coalescing effect on Xen cost 12 Xen in guest context CPU util. (%) other mm time schedule hypercall domain_page grant_table no batch 8 pkts 16 pkts 32 pkts 64 pkts Except grant and domain page map/unmap, all other Xen costs are amortized by larger batches An additional reason for optimizing grant page 22
23 Combining all RX optimizations 14 CPU util (%) app usercopy kernelu xenu grantcopy kernel xen linux current optimized Cost of network I/O for RX can be significantly reduced From ~25% to ~7% ovehead (compared to linux) Largest improvement comes for moving grant copy to guest CPU page 23
24 Optimization for TX path page 24
25 Lazy page mapping on TX Dom only needs to access packet headers No need to map guest pages with packet payload NIC device access memory directly through DMA Avoid mapping guest page on packet TX Copy packet headers using I/O ring Modified grant operation returns machine address (for DMA) but does not map page in dom. Provide page fault handler to deal with cases in which dom needs to access payload Packet to dom/domu; netfilter rules page 25
26 Benefit of lazy TX page mapping 1 Gbps TCP TX (64KB msg) 1 Gbps TCP RX 9 18 CPU util. (%) app usercopy kernelu xenu grantcopy kernel xen CPU util (%) app usercopy kernelu xenu grantcopy kernel xen original TX optimization original tx optimiation Performance improvement for TX optimization ~1% for large TX ~8% for TCP RX due to ACKs Some additional improvement may be possible with grant optimizations page 26
27 Questions? page 27
Netchannel 2: Optimizing Network Performance
Netchannel 2: Optimizing Network Performance J. Renato Santos +, G. (John) Janakiraman + Yoshio Turner +, Ian Pratt * + HP Labs - * XenSource/Citrix Xen Summit Nov 14-16, 2007 2003 Hewlett-Packard Development
More informationNetwork optimizations for PV guests
Network optimizations for PV guests J. Renato Santos G. (John) Janakiraman Yoshio Turner HP Labs Summit September 7-8, 26 23 Hewlett-Packard Development Company, L.P. The information contained herein is
More informationXenoprof overview & Networking Performance Analysis
Xenoprof overview & Networking Performance Analysis J. Renato Santos G. (John) Janakiraman Yoshio Turner Aravind Menon HP Labs Xen Summit January 17-18, 2006 2003 Hewlett-Packard Development Company, L.P.
More informationBridging the Gap between Software and Hardware Techniques for I/O Virtualization
Bridging the Gap between Software and Hardware Techniques for I/O Virtualization Jose Renato Santos, Yoshio Turner, G.(John) Janakiraman, Ian Pratt HP Laboratories HPL-28-39 Keyword(s): Virtualization,
More informationEnabling Fast, Dynamic Network Processing with ClickOS
Enabling Fast, Dynamic Network Processing with ClickOS Joao Martins*, Mohamed Ahmed*, Costin Raiciu, Roberto Bifulco*, Vladimir Olteanu, Michio Honda*, Felipe Huici* * NEC Labs Europe, Heidelberg, Germany
More informationSupport for Smart NICs. Ian Pratt
Support for Smart NICs Ian Pratt Outline Xen I/O Overview Why network I/O is harder than block Smart NIC taxonomy How Xen can exploit them Enhancing Network device channel NetChannel2 proposal I/O Architecture
More informationKeeping up with the hardware
Keeping up with the hardware Challenges in scaling I/O performance Jonathan Davies XenServer System Performance Lead XenServer Engineering, Citrix Cambridge, UK 18 Aug 2015 Jonathan Davies (Citrix) Keeping
More informationOptimizing TCP Receive Performance
Optimizing TCP Receive Performance Aravind Menon and Willy Zwaenepoel School of Computer and Communication Sciences EPFL Abstract The performance of receive side TCP processing has traditionally been dominated
More informationSpeeding up Linux TCP/IP with a Fast Packet I/O Framework
Speeding up Linux TCP/IP with a Fast Packet I/O Framework Michio Honda Advanced Technology Group, NetApp michio@netapp.com With acknowledge to Kenichi Yasukata, Douglas Santry and Lars Eggert 1 Motivation
More informationTo Grant or Not to Grant
To Grant or Not to Grant (for the case of Xen network drivers) João Martins Principal Software Engineer Virtualization Team July 11, 2017 Safe Harbor Statement The following is intended to outline our
More informationAdvanced Computer Networks. End Host Optimization
Oriana Riva, Department of Computer Science ETH Zürich 263 3501 00 End Host Optimization Patrick Stuedi Spring Semester 2017 1 Today End-host optimizations: NUMA-aware networking Kernel-bypass Remote Direct
More informationSoftware Routers: NetMap
Software Routers: NetMap Hakim Weatherspoon Assistant Professor, Dept of Computer Science CS 5413: High Performance Systems and Networking October 8, 2014 Slides from the NetMap: A Novel Framework for
More informationLearning with Purpose
Network Measurement for 100Gbps Links Using Multicore Processors Xiaoban Wu, Dr. Peilong Li, Dr. Yongyi Ran, Prof. Yan Luo Department of Electrical and Computer Engineering University of Massachusetts
More informationImplementation and Analysis of Large Receive Offload in a Virtualized System
Implementation and Analysis of Large Receive Offload in a Virtualized System Takayuki Hatori and Hitoshi Oi The University of Aizu, Aizu Wakamatsu, JAPAN {s1110173,hitoshi}@u-aizu.ac.jp Abstract System
More informationVirtualization, Xen and Denali
Virtualization, Xen and Denali Susmit Shannigrahi November 9, 2011 Susmit Shannigrahi () Virtualization, Xen and Denali November 9, 2011 1 / 70 Introduction Virtualization is the technology to allow two
More informationXen and the Art of Virtualization. CSE-291 (Cloud Computing) Fall 2016
Xen and the Art of Virtualization CSE-291 (Cloud Computing) Fall 2016 Why Virtualization? Share resources among many uses Allow heterogeneity in environments Allow differences in host and guest Provide
More informationSoftRDMA: Rekindling High Performance Software RDMA over Commodity Ethernet
SoftRDMA: Rekindling High Performance Software RDMA over Commodity Ethernet Mao Miao, Fengyuan Ren, Xiaohui Luo, Jing Xie, Qingkai Meng, Wenxue Cheng Dept. of Computer Science and Technology, Tsinghua
More informationXenrelay: An Efficient Data Transmitting Approach for Tracing Guest Domain
Xenrelay: An Efficient Data Transmitting Approach for Tracing Guest Domain Hai Jin, Wenzhi Cao, Pingpeng Yuan, Xia Xie Cluster and Grid Computing Lab Services Computing Technique and System Lab Huazhong
More informationIntroduction to Oracle VM (Xen) Networking
Introduction to Oracle VM (Xen) Networking Dongli Zhang Oracle Asia Research and Development Centers (Beijing) dongli.zhang@oracle.com May 30, 2017 Dongli Zhang (Oracle) Introduction to Oracle VM (Xen)
More informationEnabling innovation in the Internet: Main Achievements of the CHANGE Project. Felipe Huici, NEC Europe
Enabling innovation in the Internet: Main Achievements of the CHANGE Project Felipe Huici, NEC Europe EU FP7 CHANGE Project Info: Start in Oct. 2010, will end in December 2013 Partner Eurescom (Prime Contractor)
More informationBackground. IBM sold expensive mainframes to large organizations. Monitor sits between one or more OSes and HW
Virtual Machines Background IBM sold expensive mainframes to large organizations Some wanted to run different OSes at the same time (because applications were developed on old OSes) Solution: IBM developed
More informationA Case for High Performance Computing with Virtual Machines
A Case for High Performance Computing with Virtual Machines Wei Huang*, Jiuxing Liu +, Bulent Abali +, and Dhabaleswar K. Panda* *The Ohio State University +IBM T. J. Waston Research Center Presentation
More informationLecture 7. Xen and the Art of Virtualization. Paul Braham, Boris Dragovic, Keir Fraser et al. 16 November, Advanced Operating Systems
Lecture 7 Xen and the Art of Virtualization Paul Braham, Boris Dragovic, Keir Fraser et al. Advanced Operating Systems 16 November, 2011 SOA/OS Lecture 7, Xen 1/38 Contents Virtualization Xen Memory CPU
More informationRedesigning Xen's Memory Sharing Mechanism for Safe and Efficient I/O Virtualization Kaushik Kumar Ram, Jose Renato Santos, Yoshio Turner
Redesigning Xen's Memory Sharing Mechanism for Safe and Efficient I/O Virtualization Kaushik Kumar Ram, Jose Renato Santos, Yoshio Turner HP Laboratories HPL-21-39 Keyword(s): No keywords available. Abstract:
More informationThe Price of Safety: Evaluating IOMMU Performance
The Price of Safety: Evaluating IOMMU Performance Muli Ben-Yehuda 1 Jimi Xenidis 2 Michal Ostrowski 2 Karl Rister 3 Alexis Bruemmer 3 Leendert Van Doorn 4 1 muli@il.ibm.com 2 {jimix,mostrows}@watson.ibm.com
More informationReducing CPU and network overhead for small I/O requests in network storage protocols over raw Ethernet
Reducing CPU and network overhead for small I/O requests in network storage protocols over raw Ethernet Pilar González-Férez and Angelos Bilas 31 th International Conference on Massive Storage Systems
More informationIsoStack Highly Efficient Network Processing on Dedicated Cores
IsoStack Highly Efficient Network Processing on Dedicated Cores Leah Shalev Eran Borovik, Julian Satran, Muli Ben-Yehuda Outline Motivation IsoStack architecture Prototype TCP/IP over 10GE on a single
More informationFast packet processing in the cloud. Dániel Géhberger Ericsson Research
Fast packet processing in the cloud Dániel Géhberger Ericsson Research Outline Motivation Service chains Hardware related topics, acceleration Virtualization basics Software performance and acceleration
More informationRe-architecting Virtualization in Heterogeneous Multicore Systems
Re-architecting Virtualization in Heterogeneous Multicore Systems Himanshu Raj, Sanjay Kumar, Vishakha Gupta, Gregory Diamos, Nawaf Alamoosa, Ada Gavrilovska, Karsten Schwan, Sudhakar Yalamanchili College
More informationTX bulking and qdisc layer
Session: Linux packet processing performance improvements TX bulking and qdisc layer Jesper Dangaard Brouer (Red Hat) John Fastabend (Intel) John Ronciak (Intel) 1/16 Linux Plumbers Conference 16 th Oct
More informationAdvanced RDMA-based Admission Control for Modern Data-Centers
Advanced RDMA-based Admission Control for Modern Data-Centers Ping Lai Sundeep Narravula Karthikeyan Vaidyanathan Dhabaleswar. K. Panda Computer Science & Engineering Department Ohio State University Outline
More informationPart 1: Introduction to device drivers Part 2: Overview of research on device driver reliability Part 3: Device drivers research at ERTOS
Some statistics 70% of OS code is in device s 3,448,000 out of 4,997,000 loc in Linux 2.6.27 A typical Linux laptop runs ~240,000 lines of kernel code, including ~72,000 loc in 36 different device s s
More informationEvaluating the Impact of RDMA on Storage I/O over InfiniBand
Evaluating the Impact of RDMA on Storage I/O over InfiniBand J Liu, DK Panda and M Banikazemi Computer and Information Science IBM T J Watson Research Center The Ohio State University Presentation Outline
More informationHigh Performance Packet Processing with FlexNIC
High Performance Packet Processing with FlexNIC Antoine Kaufmann, Naveen Kr. Sharma Thomas Anderson, Arvind Krishnamurthy University of Washington Simon Peter The University of Texas at Austin Ethernet
More informationThe latency of user-to-user, kernel-to-kernel and interrupt-to-interrupt level communication
The latency of user-to-user, kernel-to-kernel and interrupt-to-interrupt level communication John Markus Bjørndalen, Otto J. Anshus, Brian Vinter, Tore Larsen Department of Computer Science University
More informationA low-overhead networking mechanism for virtualized high-performance computing systems
J Supercomput (2012) 59:443 468 DOI 10.1007/s11227-010-0444-9 A low-overhead networking mechanism for virtualized high-performance computing systems Jae-Wan Jang Euiseong Seo Heeseung Jo Jin-Soo Kim Published
More informationTutorial: Profiling in Xen
Tutorial: Profiling in Xen J. Renato Santos HP Labs Xen Summit September 7-8, 2006 2003 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice Preparing
More informationVirtualization with XEN. Trusted Computing CS599 Spring 2007 Arun Viswanathan University of Southern California
Virtualization with XEN Trusted Computing CS599 Spring 2007 Arun Viswanathan University of Southern California A g e n d a Introduction Virtualization approaches Basic XEN Architecture Setting up XEN Bootstrapping
More informationXen and the Art of Virtualization
Xen and the Art of Virtualization Paul Barham, Boris Dragovic, Keir Fraser, Steven Hand, Tim Harris, Alex Ho, Rolf Neugebauer, Ian Pratt, Andrew Warfield Presented by Thomas DuBuisson Outline Motivation
More informationDPDK Summit China 2017
Summit China 2017 Embedded Network Architecture Optimization Based on Lin Hao T1 Networks Agenda Our History What is an embedded network device Challenge to us Requirements for device today Our solution
More informationAdvanced Operating Systems (CS 202) Virtualization
Advanced Operating Systems (CS 202) Virtualization Virtualization One of the natural consequences of the extensibility research we discussed What is virtualization and what are the benefits? 2 Virtualization
More informationVALE: a switched ethernet for virtual machines
L < > T H local VALE VALE -- Page 1/23 VALE: a switched ethernet for virtual machines Luigi Rizzo, Giuseppe Lettieri Università di Pisa http://info.iet.unipi.it/~luigi/vale/ Motivation Make sw packet processing
More informationXen on ARM. How fast is it, really? Stefano Stabellini. 18 August 2014
Xen on ARM How fast is it, really? Stefano Stabellini 18 August 2014 Status Xen Project 4.4 release: status Features: 64-bit guest support in ARMv8 stable hypercall ABI basic lifecycle operations memory
More informationTolerating Malicious Drivers in Linux. Silas Boyd-Wickizer and Nickolai Zeldovich
XXX Tolerating Malicious Drivers in Linux Silas Boyd-Wickizer and Nickolai Zeldovich How could a device driver be malicious? Today's device drivers are highly privileged Write kernel memory, allocate memory,...
More informationPacketShader: A GPU-Accelerated Software Router
PacketShader: A GPU-Accelerated Software Router Sangjin Han In collaboration with: Keon Jang, KyoungSoo Park, Sue Moon Advanced Networking Lab, CS, KAIST Networked and Distributed Computing Systems Lab,
More informationOpenFlow Software Switch & Intel DPDK. performance analysis
OpenFlow Software Switch & Intel DPDK performance analysis Agenda Background Intel DPDK OpenFlow 1.3 implementation sketch Prototype design and setup Results Future work, optimization ideas OF 1.3 prototype
More informationNetwork device virtualization: issues and solutions
Network device virtualization: issues and solutions Ph.D. Seminar Report Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy by Debadatta Mishra Roll No: 114050005
More informationMuch Faster Networking
Much Faster Networking David Riddoch driddoch@solarflare.com Copyright 2016 Solarflare Communications, Inc. All rights reserved. What is kernel bypass? The standard receive path The standard receive path
More informationWORKLOAD CHARACTERIZATION OF INTERACTIVE CLOUD SERVICES BIG AND SMALL SERVER PLATFORMS
WORKLOAD CHARACTERIZATION OF INTERACTIVE CLOUD SERVICES ON BIG AND SMALL SERVER PLATFORMS Shuang Chen*, Shay Galon**, Christina Delimitrou*, Srilatha Manne**, and José Martínez* *Cornell University **Cavium
More informationInterrupt Coalescing in Xen
Interrupt Coalescing in Xen with Scheduler Awareness Michael Peirce & Kevin Boos Outline Background Hypothesis vic-style Interrupt Coalescing Adding Scheduler Awareness Evaluation 2 Background Xen split
More informationInitial Evaluation of a User-Level Device Driver Framework
Initial Evaluation of a User-Level Device Driver Framework Stefan Götz Karlsruhe University Germany sgoetz@ira.uka.de Kevin Elphinstone National ICT Australia University of New South Wales kevine@cse.unsw.edu.au
More informationStatus Update About COLO (COLO: COarse-grain LOck-stepping Virtual Machines for Non-stop Service)
Status Update About COLO (COLO: COarse-grain LOck-stepping Virtual Machines for Non-stop Service) eddie.dong@intel.com arei.gonglei@huawei.com yanghy@cn.fujitsu.com Agenda Background Introduction Of COLO
More informationGot Loss? Get zovn! Daniel Crisan, Robert Birke, Gilles Cressier, Cyriel Minkenberg, and Mitch Gusat. ACM SIGCOMM 2013, August, Hong Kong, China
Got Loss? Get zovn! Daniel Crisan, Robert Birke, Gilles Cressier, Cyriel Minkenberg, and Mitch Gusat ACM SIGCOMM 2013, 12-16 August, Hong Kong, China Virtualized Server 1 Application Performance in Virtualized
More informationOverhead Evaluation about Kprobes and Djprobe (Direct Jump Probe)
Overhead Evaluation about Kprobes and Djprobe (Direct Jump Probe) Masami Hiramatsu Hitachi, Ltd., SDL Jul. 13. 25 1. Abstract To implement flight recorder system, the overhead
More informationArrakis: The Operating System is the Control Plane
Arrakis: The Operating System is the Control Plane Simon Peter, Jialin Li, Irene Zhang, Dan Ports, Doug Woos, Arvind Krishnamurthy, Tom Anderson University of Washington Timothy Roscoe ETH Zurich Building
More informationXen. past, present and future. Stefano Stabellini
Xen past, present and future Stefano Stabellini Xen architecture: PV domains Xen arch: driver domains Xen: advantages - small surface of attack - isolation - resilience - specialized algorithms (scheduler)
More informationDesigning High Performance DSM Systems using InfiniBand Features
Designing High Performance DSM Systems using InfiniBand Features Ranjit Noronha and Dhabaleswar K. Panda The Ohio State University NBC Outline Introduction Motivation Design and Implementation Results
More informationOperating Systems. 17. Sockets. Paul Krzyzanowski. Rutgers University. Spring /6/ Paul Krzyzanowski
Operating Systems 17. Sockets Paul Krzyzanowski Rutgers University Spring 2015 1 Sockets Dominant API for transport layer connectivity Created at UC Berkeley for 4.2BSD Unix (1983) Design goals Communication
More informationAn Energy-Efficient Asymmetric Multi-Processor for HPC Virtualization
An Energy-Efficient Asymmetric Multi-Processor for HP Virtualization hung Lee and Peter Strazdins*, omputer Systems Group, Research School of omputer Science, The Australian National University (slides
More informationMessage Passing Architecture in Intra-Cluster Communication
CS213 Message Passing Architecture in Intra-Cluster Communication Xiao Zhang Lamxi Bhuyan @cs.ucr.edu February 8, 2004 UC Riverside Slide 1 CS213 Outline 1 Kernel-based Message Passing
More informationNTRDMA v0.1. An Open Source Driver for PCIe NTB and DMA. Allen Hubbe at Linux Piter 2015 NTRDMA. Messaging App. IB Verbs. dmaengine.h ntb.
Messaging App IB Verbs NTRDMA dmaengine.h ntb.h DMA DMA DMA NTRDMA v0.1 An Open Source Driver for PCIe and DMA Allen Hubbe at Linux Piter 2015 1 INTRODUCTION Allen Hubbe Senior Software Engineer EMC Corporation
More informationebpf Offload to Hardware cls_bpf and XDP
ebpf Offload to Hardware cls_bpf and Nic Viljoen, DXDD (Based on Netdev 1.2 talk) November 10th 2016 1 What is ebpf? A universal in-kernel virtual machine 10 64-bit registers 512 byte stack Infinite size
More informationHigh-Speed Forwarding: A P4 Compiler with a Hardware Abstraction Library for Intel DPDK
High-Speed Forwarding: A P4 Compiler with a Hardware Abstraction Library for Intel DPDK Sándor Laki Eötvös Loránd University Budapest, Hungary lakis@elte.hu Motivation Programmability of network data plane
More informationIntroduction to TCP/IP Offload Engine (TOE)
Introduction to TCP/IP Offload Engine (TOE) Version 1.0, April 2002 Authored By: Eric Yeh, Hewlett Packard Herman Chao, QLogic Corp. Venu Mannem, Adaptec, Inc. Joe Gervais, Alacritech Bradley Booth, Intel
More informationRDMA-like VirtIO Network Device for Palacios Virtual Machines
RDMA-like VirtIO Network Device for Palacios Virtual Machines Kevin Pedretti UNM ID: 101511969 CS-591 Special Topics in Virtualization May 10, 2012 Abstract This project developed an RDMA-like VirtIO network
More information6.9. Communicating to the Outside World: Cluster Networking
6.9 Communicating to the Outside World: Cluster Networking This online section describes the networking hardware and software used to connect the nodes of cluster together. As there are whole books and
More informationImpact of Cache Coherence Protocols on the Processing of Network Traffic
Impact of Cache Coherence Protocols on the Processing of Network Traffic Amit Kumar and Ram Huggahalli Communication Technology Lab Corporate Technology Group Intel Corporation 12/3/2007 Outline Background
More informationKnut Omang Ifi/Oracle 6 Nov, 2017
Software and hardware support for Network Virtualization part 1 Knut Omang Ifi/Oracle 6 Nov, 2017 1 Motivation Goal: Introduction to challenges in providing fast networking to virtual machines Prerequisites:
More informationKemari: Virtual Machine Synchronization for Fault Tolerance using DomT
Kemari: Virtual Machine Synchronization for Fault Tolerance using DomT Yoshi Tamura NTT Cyber Space Labs. tamura.yoshiaki@lab.ntt.co.jp 2008/6/24 Outline Our goal Design Architecture overview Implementation
More informationLinux Network Tuning Guide for AMD EPYC Processor Based Servers
Linux Network Tuning Guide for AMD EPYC Processor Application Note Publication # 56224 Revision: 1.00 Issue Date: November 2017 Advanced Micro Devices 2017 Advanced Micro Devices, Inc. All rights reserved.
More informationIX: A Protected Dataplane Operating System for High Throughput and Low Latency
IX: A Protected Dataplane Operating System for High Throughput and Low Latency Adam Belay et al. Proc. of the 11th USENIX Symp. on OSDI, pp. 49-65, 2014. Presented by Han Zhang & Zaina Hamid Challenges
More informationA Look at Intel s Dataplane Development Kit
A Look at Intel s Dataplane Development Kit Dominik Scholz Chair for Network Architectures and Services Department for Computer Science Technische Universität München June 13, 2014 Dominik Scholz: A Look
More informationThe Convergence of Storage and Server Virtualization Solarflare Communications, Inc.
The Convergence of Storage and Server Virtualization 2007 Solarflare Communications, Inc. About Solarflare Communications Privately-held, fabless semiconductor company. Founded 2001 Top tier investors:
More informationEvaluation and improvements of I/O Scalability for Xen. Jun Kamada, Fujitsu Limited Simon Horman, VA Linux Systems Japan
Evaluation and improvements of I/O Scalability for Xen Jun Kamada, Fujitsu Limited Simon Horman, VA Linux Systems Japan Agenda Part : I/O Scalability Improvements (Bandwidth Control) Part2 : I/O Scalability
More informationAn FPGA-Based Optical IOH Architecture for Embedded System
An FPGA-Based Optical IOH Architecture for Embedded System Saravana.S Assistant Professor, Bharath University, Chennai 600073, India Abstract Data traffic has tremendously increased and is still increasing
More informationAn Extensible Message-Oriented Offload Model for High-Performance Applications
An Extensible Message-Oriented Offload Model for High-Performance Applications Patricia Gilfeather and Arthur B. Maccabe Scalable Systems Lab Department of Computer Science University of New Mexico pfeather@cs.unm.edu,
More informationXen on ARM. Stefano Stabellini
Xen on ARM Stefano Stabellini What is Xen? a type-1 hypervisor small footprint (less than 90K LOC) Xen: Open Source GPLv2 with DCO (like Linux) Diverse contributor community Xen: Open Source source: Mike
More informationCS 3516: Computer Networks
Welcome to CS 3516: Computer Networks Prof. Yanhua Li Time: 9:00am 9:50am M, T, R, and F Location: AK 219 Fall 2018 A-term 1 Some slides are originally from the course materials of the textbook Computer
More informationCS 3516: Advanced Computer Networks
Welcome to CS 3516: Advanced Computer Networks Prof. Yanhua Li Time: 9:00am 9:50am M, T, R, and F Location: Fuller 320 Fall 2017 A-term 1 Some slides are originally from the course materials of the textbook
More information2 nd Half. Memory management Disk management Network and Security Virtual machine
Final Review 1 2 nd Half Memory management Disk management Network and Security Virtual machine 2 Abstraction Virtual Memory (VM) 4GB (32bit) linear address space for each process Reality 1GB of actual
More informationCSCI-GA Operating Systems. Networking. Hubertus Franke
CSCI-GA.2250-001 Operating Systems Networking Hubertus Franke frankeh@cs.nyu.edu Source: Ganesh Sittampalam NYU TCP/IP protocol family IP : Internet Protocol UDP : User Datagram Protocol RTP, traceroute
More informationLink Virtualization based on Xen
Link Virtualization based on Xen ShinHyoung Lee, Chuck Yoo shlee@os.korea.ac.kr, hxy@os.korea.ac.kr Sponsored by: & Contents Introduction Future Internet Virtual Network Link Virtualization Related Works
More informationLatest Developments with NVMe/TCP Sagi Grimberg Lightbits Labs
Latest Developments with NVMe/TCP Sagi Grimberg Lightbits Labs 2018 Storage Developer Conference. Insert Your Company Name. All Rights Reserved. 1 NVMe-oF - Short Recap Early 2014: Initial NVMe/RDMA pre-standard
More informationDepartment of Computer Science, Institute for System Architecture, Operating Systems Group. Real-Time Systems '08 / '09. Hardware.
Department of Computer Science, Institute for System Architecture, Operating Systems Group Real-Time Systems '08 / '09 Hardware Marcus Völp Outlook Hardware is Source of Unpredictability Caches Pipeline
More informationImplemen'ng IPv6 Segment Rou'ng in the Linux Kernel
Implemen'ng IPv6 Segment Rou'ng in the Linux Kernel David Lebrun, Olivier Bonaventure ICTEAM, UCLouvain Work supported by ARC grant 12/18-054 (ARC-SDN) and a Cisco grant Agenda IPv6 Segment Rou'ng Implementa'on
More informationCSE 120 Principles of Operating Systems
CSE 120 Principles of Operating Systems Spring 2018 Lecture 15: Multicore Geoffrey M. Voelker Multicore Operating Systems We have generally discussed operating systems concepts independent of the number
More informationTCP/misc works. Eric Google
TCP/misc works Eric Dumazet @ Google 1) TCP zero copy receive 2) SO_SNDBUF model in linux TCP (aka better TCP_NOTSENT_LOWAT) 3) ACK compression 4) PSH flag set on every TSO packet Design for TCP RX ZeroCopy
More informationOn the cost of tunnel endpoint processing in overlay virtual networks
J. Weerasinghe; NVSDN2014, London; 8 th December 2014 On the cost of tunnel endpoint processing in overlay virtual networks J. Weerasinghe & F. Abel IBM Research Zurich Laboratory Outline Motivation Overlay
More informationNetwork stack specialization for performance
Network stack specialization for performance goo.gl/1la2u6 Ilias Marinos, Robert N.M. Watson, Mark Handley* University of Cambridge, * University College London Motivation Providers are scaling out rapidly.
More informationOpen Source Traffic Analyzer
Open Source Traffic Analyzer Daniel Turull June 2010 Outline 1 Introduction 2 Background study 3 Design 4 Implementation 5 Evaluation 6 Conclusions 7 Demo Outline 1 Introduction 2 Background study 3 Design
More informationFast packet processing in linux with af_xdp
Fast packet processing in linux with af_xdp Magnus Karlsson and Björn Töpel, Intel Legal Disclaimer Intel technologies may require enabled hardware, specific software, or services activation. Check with
More informationQuiz. Segment structure and fields Flow control (rwnd) Timeout interval. Phases transition ssthresh setting Cwnd setting
Quiz v 10/30/2013 (Wednesday), 20 mins v Midterm question (available on website) v TCP basics Segment structure and fields Flow control (rwnd) Timeout interval v TCP Congestion control Phases transition
More informationHKG net_mdev: Fast-path userspace I/O. Ilias Apalodimas Mykyta Iziumtsev François-Frédéric Ozog
HKG18-110 net_mdev: Fast-path userspace I/O Ilias Apalodimas Mykyta Iziumtsev François-Frédéric Ozog Why userland I/O Time sensitive networking Developed mostly for Industrial IOT, automotive and audio/video
More informationOperating Systems. 11. Memory Management Part 3 Kernel Memory Allocation. Paul Krzyzanowski Rutgers University Spring 2015
Operating Systems 11. Memory Management Part 3 Kernel Memory Allocation Paul Krzyzanowski Rutgers University Spring 2015 1 Kernel memory The kernel also needs memory User code calls malloc kernel functions
More informationReview: Hardware user/kernel boundary
Review: Hardware user/kernel boundary applic. applic. applic. user lib lib lib kernel syscall pg fault syscall FS VM sockets disk disk NIC context switch TCP retransmits,... device interrupts Processor
More informationA comparative analysis of Precision Time Protocol in native, virtual machines and container-based environments for consolidating automotive workloads
A comparative analysis of Precision Time Protocol in native, virtual machines and container-based environments for consolidating automotive workloads Speaker: Co-authors: Ong Boon Leong boon.leong.ong@intel.com
More informationData Center Virtualization: Xen and Xen-blanket
Data Center Virtualization: Xen and Xen-blanket Hakim Weatherspoon Assistant Professor, Dept of Computer Science CS 5413: High Performance Systems and Networking November 17, 2014 Slides from ACM European
More informationI/O virtualization. Jiang, Yunhong Yang, Xiaowei Software and Service Group 2009 虚拟化技术全国高校师资研讨班
I/O virtualization Jiang, Yunhong Yang, Xiaowei 1 Legal Disclaimer INFORMATION IN THIS DOCUMENT IS PROVIDED IN CONNECTION WITH INTEL PRODUCTS. NO LICENSE,
More informationCS 333 Introduction to Operating Systems. Class 11 Virtual Memory (1) Jonathan Walpole Computer Science Portland State University
CS 333 Introduction to Operating Systems Class 11 Virtual Memory (1) Jonathan Walpole Computer Science Portland State University Virtual addresses Virtual memory addresses (what the process uses) Page
More informationVirtual Memory. Patterson & Hennessey Chapter 5 ELEC 5200/6200 1
Virtual Memory Patterson & Hennessey Chapter 5 ELEC 5200/6200 1 Virtual Memory Use main memory as a cache for secondary (disk) storage Managed jointly by CPU hardware and the operating system (OS) Programs
More information