Architecting Low Latency Cloud Networks

Similar documents
Switching Architectures for Cloud Network Designs

Big Data Big Data Becoming a Common Problem

Arista 7050X, 7050X2, 7250X and 7300 Series Performance Validation

CloudVision Macro-Segmentation Service

Creating High Performance Best-In-Breed Scale- Out Network Attached Storage Solutions

Leveraging EOS and sflow for Advanced Network Visibility

Arista Networks and F5 Solution Integration

The benefits Arista s LANZ functionality will provide to network administrators: Real time visibility of congestion hotspots at the microbursts level

Arista FlexRoute TM Engine

Arista Cognitive WiFi

Networking in the Hadoop Cluster

The Impact of Virtualization on Cloud Networking

ARISTA WHITE PAPER Arista FlexRouteTM Engine

Rapid Automated Indication of Link-Loss

TAP Aggregation with DANZ

Arista 7500 Series Interface Flexibility

Five ways to optimise exchange connectivity latency

Why Big Data Needs Big Buffer Switches

Introduction: PURPOSE BUILT HARDWARE. ARISTA WHITE PAPER HPC Deployment Scenarios

NEP UK selects Arista as foundation for SMPTE ST 2110 modular OB trucks to deliver UHD content from world s largest events

10Gb Ethernet: The Foundation for Low-Latency, Real-Time Financial Services Applications and Other, Latency-Sensitive Applications

Latency Analyzer (LANZ)

Investment Protection with the Arista 7500 Series

An Overview of Arista Ethernet Capture Timestamps

Bioscience. Solution Brief. arista.com

The zettabyte era is here. Is your datacenter ready? Move to 25GbE/50GbE with confidence

Virtual Extensible LAN (VXLAN) Overview

Cloudifying Datacenter Monitoring with DANZ

Traffic Visualization with Arista sflow and Splunk

Arista AgilePorts INTRODUCTION

Arista 7060X & 7260X Performance

CHANGING DYNAMICS OF IP PEERING Arista Solution Guide

Arista Telemetry. White Paper. arista.com

Cloud Interconnect: DWDM Integrated Solution For Secure Long Haul Transmission

Arista 7160 Series Switch Architecture

Deploying IP Storage Infrastructures

Powering Next Generation Video Delivery

World Class, High Performance Cloud Scale Storage Solutions Arista and EMC ScaleIO

Arista 7050X3 Series Switch Architecture

Simplifying Network Operations through Data Center Automation

Four key trends in the networked use of FPGAs

The Arista Universal transceiver is the first of its kind 40G transceiver that aims at addressing several challenges faced by today s data centers.

ARISTA WHITE PAPER Arista 7500E Series Interface Flexibility

Routing Architecture Transformations

Solving the Virtualization Conundrum

Arista 7500E DWDM Solution and Use Cases

Software Driven Cloud Networking

Spanning Tree Protocol Interoperability With Cisco PVST+/PVRST+/MSTP

Arista 7170 Multi-function Programmable Networking

High Speed Networking for Digital Media Creation, Post Production, and Centralized Content Management

Broadcast Transition from SDI to Ethernet

Migration from Silo Security to Secure Holistic Cloud Networking

Mellanox Virtual Modular Switch

Arista Cognitive Campus Network

Chelsio Communications. Meeting Today s Datacenter Challenges. Produced by Tabor Custom Publishing in conjunction with: CUSTOM PUBLISHING

NVMe over Universal RDMA Fabrics

MOVE TO A FLEXIBLE IT MODEL ENTERPRISE DATA CENTER SOLUTIONS.

Exploring the Arista 7010T - An Overview

Arista CloudVision : Cloud Automation for Everyone

Arista EOS Precision Data Analysis with DANZ

Deploying Data Center Switching Solutions

RoCE vs. iwarp Competitive Analysis

GUIDE. Optimal Network Designs with Cohesity

By John Kim, Chair SNIA Ethernet Storage Forum. Several technology changes are collectively driving the need for faster networking speeds.

Future Routing Schemes in Petascale clusters

Arista 7050X & 7050X2 Switch Architecture ( A day in the life of a packet )

QLogic 10GbE High-Performance Adapters for Dell PowerEdge Servers

Choosing the Best Network Interface Card for Cloud Mellanox ConnectX -3 Pro EN vs. Intel XL710

Network Design Considerations for Grid Computing

DriveScale-DellEMC Reference Architecture

ARISTA: Improving Application Performance While Reducing Complexity

Exploring the 7150S Family

Hyper-Converged Rack-Level Solutions

The Case for Cloud WiFi

Iron Networks, Inc. Turnkey Converged Infrastructure-as-a-Service Platforms

Fusion iomemory PCIe Solutions from SanDisk and Sqrll make Accumulo Hypersonic

InfiniBand Networked Flash Storage

Highest Levels of Scalability Simplified Network Manageability Maximum System Productivity

IBM WebSphere MQ Low Latency Messaging Software Tested With Arista 10 Gigabit Ethernet Switch and Mellanox ConnectX

Mellanox InfiniBand Solutions Accelerate Oracle s Data Center and Cloud Solutions

Architecting the High Performance Storage Network

Enterprise Case Study

QLogic TrueScale InfiniBand and Teraflop Simulations

The Arista Advantage Cloud Networking Trends

Distributed Core Architecture Using Dell Networking Z9000 and S4810 Switches

Accelerating Real-Time Big Data. Breaking the limitations of captive NVMe storage

New Approach to Unstructured Data

Cloud Optimized Performance: I/O-Intensive Workloads Using Flash-Based Storage

Birds of a Feather Presentation

Hitachi Unified Compute Platform Pro for VMware vsphere

A Low Latency Solution Stack for High Frequency Trading. High-Frequency Trading. Solution. White Paper

Messaging Overview. Introduction. Gen-Z Messaging

ALCATEL Edge Services Router

MODERNIZE YOUR DATA CENTER. With Cisco Nexus Switches

A Scalable, Commodity Data Center Network Architecture

ATA DRIVEN GLOBAL VISION CLOUD PLATFORM STRATEG N POWERFUL RELEVANT PERFORMANCE SOLUTION CLO IRTUAL BIG DATA SOLUTION ROI FLEXIBLE DATA DRIVEN V

Header Compression Capacity Calculations for Wireless Networks

Mellanox Technologies Delivers The First InfiniBand Server Blade Reference Design

SOFTWARE DEFINED STORAGE VS. TRADITIONAL SAN AND NAS

QLogic/Lenovo 16Gb Gen 5 Fibre Channel for Database and Business Analytics

Transcription:

Architecting Low Latency Cloud Networks As data centers transition to next generation virtualized & elastic cloud architectures, high performance and resilient cloud networking has become a requirement for delivering always-on, always-available applications. Over the last decade, response time requirements have been rigorously tightened (see Figure 1), in some cases by as much as two to three orders of magnitude. In addition to the financial, education/research institutions and Web 2.0 firms, there is a growing list of customer verticals where fast response time is a top-of-mind business requirement, including rich media companies (e.g. those involved in video post-production, digital animation and broadcasting), oil & gas producers, and health care. Figure 1: Application Response Time Trend As response times gravitate towards fractions of a second, both network and compute latencies need to be reduced in parallel. In effect an architectural approach is needed so that response times are consistently and measurably enhanced across a wide set of cloud applications. Unlike prior efforts that often deploy niche and sometimes even proprietary technologies, today it is critical that low latency infrastructure fabric be built using standards-based and widely used technologies for broadbased cloud deployments. Also, because latency optimization is implemented at the infrastructure level, it needs to be future proof to meet latency demands 3 to 5 years out.

Architecting Low-Latency Cloud Networks A key attribute of latency-sensitive workloads is that they are built using distributed compute architecture. Each workload transaction spawns a large number of interactions between compute nodes, in some cases across thousands of machines and data stores, as depicted in Figure 2. For example: In algorithmic trading, machines receive market data feeds from other machines, calculate trading decisions and place trades with another set of machines. In HPC, thousands of commodity servers are clustered to behave as a super computer, connected to a clustered file system (e.g. Lustre) to crunch complex mathematical formulae and simulations. In web based cloud applications, workloads operate on thousands of dynamically clustered servers and implement multistage processing mechanisms to sift through petabytes of data. Example applications include those implemented by Web 2.0 and Internet Search companies, as well as emerging open-source initiatives such as Hadoop which provides scalable and reliable framework for distributed computing (MapReduce), distributed file system and distributed database. Figure 2: Network based machine-to-machine interactions of a single workload: A transaction To boost system-wide response time, both compute and network latencies must be minimized. Typically, commodity servers are used for computing where degrees of freedom for latency optimization are limited. One essentially selects a server with denser and faster multi-core CPUs, larger and faster memories/caches, and speedier peripheral interconnects. The other side of latency coin is the cloud network where fabric-wide optimization is called for to minimize transport latencies. In particular, network latencies need to be optimized across four key axes: Reduce latency of each network node Reduce number of network nodes needed to traverse from one stage to another Eliminate network congestion Reduce transport protocol latency The Arista Advantage Arista Networks has adopted an architectural approach to building low latency cloud networking. Its wire-rate, non-blocking and ultra low latency 1Gb/10GbE switches, along with extensible operating system (EOS), enable resilient cloud networks for transporting data, multi-media, storage and compute traffic. Architecting cloud networks based on Arista s 7xxx series switch fabric has the following advantages: Reduce network latencies across all four axes (see details below); Future-proof the network infrastructure for latency demands of today and tomorrow; Implement standard Ethernet, which is widely deployed and operationally well understood.

Network Node Latency For ultra low latency workloads, such as those in financials and HPC, shaving off every microsecond is important. Deploying the optimal 1/10Gb Ethernet switch can improve latency by 10+ microseconds for 1500B Ethernet frames (even more if the 9KB jumbo frames are used). Also, the cut-through switch architecture can reduce additional tens of microseconds of over older store-andforward designs. Arista provides dense 10Gb switches with cut-through architecture that yields switching latencies in the range of 0.6-1.2 microsecond (see Figure 3). Figure 3: Latency benefits of Arista s cut through switches over traditional store and forward switches (a) Legacy Store-and-Forward Design Intra-Rack Latency: 8µs Inter-Rack Latency: 36µs (b) Arista Cut-Through Design Intra-Rack Latency: 0.6µs Inter-Rack Latency: 2.4µs Number of Network Hops Typical data center architecture consists of a three-tier architecture, with server access, distribution and core switch layers. Often packets need to traverse all the way to the core layer when communicating across server racks. In Arista s two-tier topology with leaf and spine switches (see Figure 4), latency is significantly reduced as packets traverse fewer hops. Figure 4: Arista s 2 tier (leaf and spine) topology reduces latency over traditional 3 tier (access, distribution, core) topology

Network Congestion In most applications, IP/Ethernet packets are transported using the TCP protocol. To guarantee reliability and maintain throughput, TCP state machine uses windowing to adjust to dynamic network congestion. During congestion periods, TCP uses smaller windows, thus decreasing throughput and increasing overall transmission latency. Traditional data center architecture is oversubscribed, thus leading to congestions & dropped packets and hence large TCP latencies. Arista s 7100 switches are wire-rate switches (line rate for every 10Gb Ethernet port), thus enabling non-oversubscribed cloud network architectures with 1:1 bisectional bandwidth (as shown in Figure 4). Large, congestion-free, fabric can be architected that allows TCP to operate using large window sizes and hence at high throughput and low latency. Hundreds or thousands of machine-to- machine TCP interactions of application workloads, be they web, file/storage, multi-media or database, benefits immensely in terms of TCP throughput and latency. Transport Protocol Latency Certain low latency environments (e.g. HPC, applications leveraging Lustre file system algorithmic trading), it is desired to bypass TCP altogether and instead leverage direct machine-to-machine communication using message-passing (MPI) or remote direct memory access (RDMA). Infiniband and other proprietary interconnects have been used because of their ultra low latency, high throughput and lossless characteristics. Customers have longed for standards- based Ethernet to provide such capabilities as well as price-performance advantage so that they can benefit from the cost-efficiencies of managing a single Ethernet fabric across the public or private cloud. Arista s 7100 series of 10Gb Ethernet switches, with line rate and cut-through low latency performance, enables a single low latency cloud network. Arista s architecture also supports current and emerging Ethernet standards in IEEE and IETF on Layer-2/3 multi-pathing and congestion control to span large congestion free domains. Large Layer-2 domains, along with Arista s low latency, wire- rate and congestion-free cloud network, enable rapid migration of application workloads anywhere in the cloud. Low-Latency Application Example: Capital Markets Algorithmic trading and market data feeds are the most celebrated use cases of ultra low latency designs where it is said that 1-millisecond advantage in trading applications can be worth $100 million a year to a major brokerage firm 1. Shaving off milliseconds is so critical that brokerage firms co-locate their trading systems right on the exchange floor for direct access to market data feeds, thus eliminating distance-based latency barriers. Modern designs strive for one millisecond response time for trading transactions, with the eye towards deep sub-millisecond designs in few years. Quest for speed has often led to deployment of exotic interconnects, such as Infiniband that are operationally difficult to deploy & manage at cloud scale. Arista s line-rate Ethernet switch fabric, along with eco system partners such as Solace Systems, provides a complete algorithmic trading solution with 31-microsecond latency, as shown in Figure 5(b). Figure 5(a): Market data feed and algorithmic trading -- 5(b): Arista + Solace Systems provides ultra low latency algorithmic trading solution 1 Business at Light Speed by Richard Martin, InformationWeek, April 2007 (http://www.informationweek.com/news/infrastructure/showarticle.jhtml?articleid=199200297)

Summary The drive for scalable and efficient cloud networks has led to 10Gb Ethernet as the interface of choice for simultaneously transporting data, storage and compute information over a common Ethernet cloud. Multiple classes of applications, including those in capital markets, high performance computing and web-centric computing, rely on latency-optimized architectures for ultra fast response times. After all, it is faster response times that ensure users stickiness to a service, enhance scalability of HPC clusters and increase trading profits by speeding up algorithmic trade execution during price fluctuations. The Arista Networks 7XX series is designed solve these real word cloud- computing challenges for latency and scale. Santa Clara Corporate Headquarters 5453 Great America Parkway, Santa Clara, CA 95054 Phone: +1-408-547-5500 Fax: +1-408-538-8920 Email: info@ Ireland International Headquarters 3130 Atlantic Avenue Westpark Business Campus Shannon, Co. Clare Ireland Vancouver R&D Office 9200 Glenlyon Pkwy, Unit 300 Burnaby, British Columbia Canada V5J 5J8 San Francisco R&D and Sales Office 1390 Market Street, Suite 800 San Francisco, CA 94102 India R&D Office Global Tech Park, Tower A & B, 11th Floor Marathahalli Outer Ring Road Devarabeesanahalli Village, Varthur Hobli Bangalore, India 560103 Singapore APAC Administrative Office 9 Temasek Boulevard #29-01, Suntec Tower Two Singapore 038989 Nashua R&D Office 10 Tara Boulevard Nashua, NH 03062 Copyright 2016 Arista Networks, Inc. All rights reserved. CloudVision, and EOS are registered trademarks and Arista Networks is a trademark of Arista Networks, Inc. All other company names are trademarks of their respective holders. Information in this document is subject to change without notice. Certain features may not yet be available. Arista Networks, Inc. assumes no responsibility for any errors that may appear in this document.