ASPERA HIGH-SPEED TRANSFER. Moving the world s data at maximum speed

Similar documents
ASPERA HIGH-SPEED TRANSFER. Moving the world s data at maximum speed

TRANSFORMING CLOUD INFRASTRUCTURE TO SUPPORT BIG DATA STORAGE AND WORKFLOWS

Scaling Internet TV Content Delivery ALEX GUTARIN DIRECTOR OF ENGINEERING, NETFLIX

Application Acceleration Beyond Flash Storage

A High-Performance Storage and Ultra- High-Speed File Transfer Solution for Collaborative Life Sciences Research

End-to-End Adaptive Packet Aggregation for High-Throughput I/O Bus Network Using Ethernet

Extremely Fast Distributed Storage for Cloud Service Providers

Communication has significant impact on application performance. Interconnection networks therefore have a vital role in cluster systems.

RoCE vs. iwarp Competitive Analysis

Ultra high-speed transmission technology for wide area data movement

Mellanox InfiniBand Solutions Accelerate Oracle s Data Center and Cloud Solutions

Chelsio Communications. Meeting Today s Datacenter Challenges. Produced by Tabor Custom Publishing in conjunction with: CUSTOM PUBLISHING

IBM Aspera Direct-to- Cloud Storage

EXTENDING AN ASYNCHRONOUS MESSAGING LIBRARY USING AN RDMA-ENABLED INTERCONNECT. Konstantinos Alexopoulos ECE NTUA CSLab

Enabling High Performance Bulk Data Transfers With SSH

BlueGene/L. Computer Science, University of Warwick. Source: IBM

IBM WebSphere MQ Low Latency Messaging Software Tested With Arista 10 Gigabit Ethernet Switch and Mellanox ConnectX

Ronald van der Pol

Meltdown and Spectre Interconnect Performance Evaluation Jan Mellanox Technologies

TLDK Overview. Transport Layer Development Kit Ray Kinsella February ray.kinsella [at] intel.com IRC: mortderire

Emerging Technologies for HPC Storage

Next Steps Spring 2011 Lecture #18. Multi-hop Networks. Network Reliability. Have: digital point-to-point. Want: many interconnected points

Performance Optimizations via Connect-IB and Dynamically Connected Transport Service for Maximum Performance on LS-DYNA

IBM V7000 Unified R1.4.2 Asynchronous Replication Performance Reference Guide

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

Lecture 4: Introduction to Computer Network Design

ARISTA: Improving Application Performance While Reducing Complexity

InfiniBand Networked Flash Storage

Networking at the Speed of Light

TLDK Overview. Transport Layer Development Kit Keith Wiles April Contributions from Ray Kinsella & Konstantin Ananyev

RAIDIX Data Storage Solution. Clustered Data Storage Based on the RAIDIX Software and GPFS File System

N V M e o v e r F a b r i c s -

ClearStream. Prototyping 40 Gbps Transparent End-to-End Connectivity. Cosmin Dumitru! Ralph Koning! Cees de Laat! and many others (see posters)!

Benefits of 25, 40, and 50GbE Networks for Ceph and Hyper- Converged Infrastructure John F. Kim Mellanox Technologies

Ziye Yang. NPG, DCG, Intel

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme

Evaluating the Impact of RDMA on Storage I/O over InfiniBand

iscsi or iser? Asgeir Eiriksson CTO Chelsio Communications Inc

Data transfer over the wide area network with a large round trip time

SPDK China Summit Ziye Yang. Senior Software Engineer. Network Platforms Group, Intel Corporation

W H I T E P A P E R. Comparison of Storage Protocol Performance in VMware vsphere 4

Introducing SkyPipe link optimisation for BGAN

High bandwidth, Long distance. Where is my throughput? Robin Tasker CCLRC, Daresbury Laboratory, UK

SONAS Best Practices and options for CIFS Scalability

IBM Aspera Direct-to-Cloud Storage

SSH Bulk Transfer Performance. Allan Jude --

FUJITSU Software Interstage Information Integrator V11

2017 Storage Developer Conference. Mellanox Technologies. All Rights Reserved.

Can Parallel Replication Benefit Hadoop Distributed File System for High Performance Interconnects?

PM Support in Linux and Windows. Dr. Stephen Bates, CTO, Eideticom Neal Christiansen, Principal Development Lead, Microsoft

Advanced Computer Networks. End Host Optimization

FlashGrid Software Enables Converged and Hyper-Converged Appliances for Oracle* RAC

G-NET: Effective GPU Sharing In NFV Systems

Fast packet processing in the cloud. Dániel Géhberger Ericsson Research

Hyper-converged infrastructure with Proxmox VE virtualization platform and integrated Ceph Storage.

Be Fast, Cheap and in Control with SwitchKV. Xiaozhou Li

Aspera Direct-to-Cloud Storage WHITE PAPER

Memory-Based Cloud Architectures

What s New in VMware vsphere 4.1 Performance. VMware vsphere 4.1

A closer look at network structure:

Achieve Optimal Network Throughput on the Cisco UCS S3260 Storage Server

DDN. DDN Updates. Data DirectNeworks Japan, Inc Shuichi Ihara. DDN Storage 2017 DDN Storage

LUSTRE NETWORKING High-Performance Features and Flexible Support for a Wide Array of Networks White Paper November Abstract

Designing Next Generation Data-Centers with Advanced Communication Protocols and Systems Services

Study. Dhabaleswar. K. Panda. The Ohio State University HPIDC '09

Cluster Network Products

Analytics of Wide-Area Lustre Throughput Using LNet Routers

Lighting the Blue Touchpaper for UK e-science - Closing Conference of ESLEA Project The George Hotel, Edinburgh, UK March, 2007

An Implementation of the Homa Transport Protocol in RAMCloud. Yilong Li, Behnam Montazeri, John Ousterhout

NVMe over Fabrics. High Performance SSDs networked over Ethernet. Rob Davis Vice President Storage Technology, Mellanox

Dell PowerEdge R730xd Servers with Samsung SM1715 NVMe Drives Powers the Aerospike Fraud Prevention Benchmark

DDN. DDN Updates. DataDirect Neworks Japan, Inc Nobu Hashizume. DDN Storage 2018 DDN Storage 1

Concurrent Support of NVMe over RDMA Fabrics and Established Networked Block and File Storage

Deep Learning Performance and Cost Evaluation

Network Design Considerations for Grid Computing

Extreme Storage Performance with exflash DIMM and AMPS

The NE010 iwarp Adapter

Server Specifications

Feedback on BeeGFS. A Parallel File System for High Performance Computing

NFS/RDMA over 40Gbps iwarp Wael Noureddine Chelsio Communications

Messaging Overview. Introduction. Gen-Z Messaging

Be Fast, Cheap and in Control with SwitchKV Xiaozhou Li

Ethernet. High-Performance Ethernet Adapter Cards

Implementing SQL Server 2016 with Microsoft Storage Spaces Direct on Dell EMC PowerEdge R730xd

Internet data transfer record between CERN and California. Sylvain Ravot (Caltech) Paolo Moroni (CERN)

Department of EECS - University of California at Berkeley EECS122 - Introduction to Communication Networks - Spring 2005 Final: 5/20/2005

Future Routing Schemes in Petascale clusters

Accelerating Hadoop Applications with the MapR Distribution Using Flash Storage and High-Speed Ethernet

What s Wrong with the Operating System Interface? Collin Lee and John Ousterhout

QNAP 25GbE PCIe Expansion Card. Mellanox SN GbE/100GbE Management Switch. QNAP NAS X 25GbE card X 25GbE/100GbE switch

OpenFOAM Performance Testing and Profiling. October 2017

NVMe Takes It All, SCSI Has To Fall. Brave New Storage World. Lugano April Alexander Ruebensaal

There are 10 questions in total. Please write your SID on each page.

Birds of a Feather Presentation

BeeGFS. Parallel Cluster File System. Container Workshop ISC July Marco Merkel VP ww Sales, Consulting

Networking for Data Acquisition Systems. Fabrice Le Goff - 14/02/ ISOTDAQ

LANCOM Techpaper Routing Performance

Routers: Forwarding EECS 122: Lecture 13

Netronome 25GbE SmartNICs with Open vswitch Hardware Offload Drive Unmatched Cloud and Data Center Infrastructure Performance

An Overview of Fujitsu s Lustre Based File System

Transcription:

ASPERA HIGH-SPEED TRANSFER Moving the world s data at maximum speed

ASPERA HIGH-SPEED FILE TRANSFER Aspera FASP Data Transfer at 80 Gbps Elimina8ng tradi8onal bo<lenecks in high speed data transfer Charles Shifle< Developer bear @ us. IBM. com bear @ asperasoi.com 2

THE FASP PROTOCOL Designed to provide WAN transport at wire speed Efficiently u8lize available network bandwidth Secure Reliable Conges8on Controlled Standard UDP/IP packets Flexible I/O Layer Direct to object Storage In memory block based transport ( FASP 3+ ) Op8mized for parallel clustered storage ( FASP 4 ) Op8mized for NVMe ( FASP 4 ) DPDK based Network Stack ( FASP 4 ) 3

CHALLENGES WITH TCP AND ALTERNATIVE TECHNOLOGIES Distance degrades conditions on all networks Latency (or Round Trip Times) increase Packet losses increase Fast networks just as prone to degradation TCP performance degrades with distance Throughput bottleneck becomes more severe with increased latency and packet loss TCP does not scale with bandwidth TCP designed for low bandwidth Adding more bandwidth does not improve throughput Alternative Technologies TCP-based - Network latency and packet loss must be low UDP traffic blasters - Inefficient and waste bandwidth Data caching - Inappropriate for many large file transfer workflows Modified TCP - Improves on TCP performance but insufficient for fast networks Data compression - Time consuming and impractical for certain file types CDNs & co-lo build outs - High overhead and expensive to scale 4

FASP : HIGH-PERFORMANCE TRANSPORT Maximum transfer speed Optimal end-to-end throughput efficiency Transfer performance scales with bandwidth independent of transfer distance and resilient to packet loss Congestion Avoidance and Policy Control Automatic, full utilization of available bandwidth On-the-fly prioritization and bandwidth allocation Uncompromising security and reliability Secure, user/endpoint authentication AES-128 cryptography in transit and at-rest Scalable management, monitoring and control Real-time progress, performance and bandwidth utilization Detailed transfer history, logging, and manifest Low Overhead Less than 0.1% overhead on 30% packet loss High performance with large files or large sets of small files Resulting in Transfers up to thousands of times faster than FTP Precise and predictable transfer times Extreme scalability (concurrency and throughput) 5

TYPICAL TRANSFER APPLICATION 6

ASPERA NEXT GENERATION FASP 7

I/O INCREASING FASTER THAN MEMORY Tradi8onal Architectures assume Slow I/O and fast Memory. No Longer true as I/O interconnects Approach the speed of memory. 20%/yr Source: h<ps://www.plda.com/market- ready- conquer- pcie- 40- challenges Source: h<p://motherboard.vice.com/read/memory- is- holding- up- the- moores- law- progression- of- processing- power 8

HARDWARE CONFIGURATION 2x Intel Xeon E5-2697 v3 5x Intel DC P3700 NVMe SSD 2x Intel XL710 40 GbE Ethernet QSFP+ 9

LAN TRANSFER RESULTS IN GBIT PER SECOND 80 70 60 50 40 30 20 Disk to Disk Wire U8liza8on w/o disk 10 0 SSH AES- 128 CTR FTP iperf FASP 4x10Gbit (Intel Xeon E5-2650 v2) FASP (Intel Xeon E5-2697 v3) Results from SC 14 showing the rela8ve performance of Network Transfer Technologies 10

ENCRYPTION IS NOT A BOTTLENECK WITH HW AES-GCM AES 128 GCM Encryp0on Rate in GB/s per Core 2.5 2 1.5 1 0.5 0 Core i7 X 980 3.33GHz Xeon E5-2650 v2 2.60GHz Xeon E5-2697 v3 2.60GHz 11

DC P3700 PERFORMANCE (SINGLE DRIVE) 3000 2800 MB/s 2500 2678 MB/s 2000 1500 1000 1869 MB/s Write Synchronous Read Async Read (AIO) 500 0 128 16 4 1 Performance rela8ve to block size (SI Units in MB) 12

IMPROVING STORAGE THROUGHPUT Traditional storage is built around the idea of moving slow data from disk to memory and then from memory to application. Memory is used to cache data to improve access speeds. Cache structure quickly becomes a bottleneck as transfer speeds exceed 10 gbit/s. While individual spinning disks are slow, JBOD s of 100 s of disk have very high aggregate bandwidth Modern SSDs (especially NVMe) is very fast. Two solutions for fast data Use XFS with direct IO (and/or MMAP) Have shown performance at about 40gbit/s with hardware raid and direct attach disks. Have shown performance at about 70gbit/s with NVMe SSD. Limited by how many devices you can connect to PCIe data link. Clustered parallel storage Aspera is targeting IBM GPFS (Spectrum) or Lustre based solutions Individual nodes can be slow, but by aggregating nodes high performance is achieved Both offer direct I/O solutions 13

FUTURE DIRECTION Aspera s Goal: Transfer Solution to Petascale Datasets - 1 SI Terabyte in 2 Minutes - 1 SI Petabyte in 1⅜ Days - Performance improvements expected to scale relative to PCIe interconnect. Better integration with storage systems - Take advantage of native APIs to avoid kernel cache. Better integration with network hardware - Expected to show 100gbit/s transfers using adapters like Mellanox ConnectX -4 - Query network switch and routers? Support wider use cases - Compression, Forward Error Correcting Codecs, Berkeley Sockets API 14

THANK YOU CHARLES SHIFLETT bear @ us.ibm.com