Accelerating Hadoop Applications with the MapR Distribution Using Flash Storage and High-Speed Ethernet

Similar documents
Mellanox CloudX, Mirantis Fuel 5.1/ 5.1.1/6.0 Solution Guide

ibutils2 - InfiniBand Diagnostic Utilities Release Notes

RoCE vs. iwarp Competitive Analysis

Mellanox ConnectX-4/ ConnectX-4 Lx Plugin for RedHat OpenStack Platform 10

Mellanox ConnectX-4 NATIVE ESX Driver for VMware vsphere 6.0 Release Notes

Mellanox ConnectX-3 ESXi 6.5 Inbox Driver Release Notes. Rev 1.0

InfiniBand Switch System Family. Highest Levels of Scalability, Simplified Network Manageability, Maximum System Productivity

Mellanox NATIVE ESX Driver for VMware vsphere 6.0 Release Notes

ARISTA: Improving Application Performance While Reducing Complexity

MLNX_EN for FreeBSD Release Notes

Mellanox ConnectX-3 ESXi 6.0 Inbox Driver

PERFORMANCE ACCELERATED Mellanox InfiniBand Adapters Provide Advanced Levels of Data Center IT Performance, Productivity and Efficiency

Mellanox NATIVE ESX Driver for VMware vsphere 6.0 Release Notes

40Gb/s InfiniBand Switch Module (HSSM) for IBM BladeCenter

Mellanox SwitchX Firmware (fw-sx) Release Notes

SUSE Linux Enterprise Server (SLES) 12 SP3 Driver SLES 12 SP3

Innova-2 Flex Open for Application Acceleration EN Adapter Card. Software and Firmware Bundle Release Notes

Ubuntu Inbox Driver Release Notes. Ubuntu 16.10

Mellanox ConnectX-4 NATIVE ESX Driver for VMware vsphere 5.5/6.0 Release Notes

Red Hat Enterprise Linux (RHEL) 7.4-ALT Driver Release Notes

Mellanox NATIVE ESX Driver for VMware vsphere 6.5 Release Notes

Mellanox NATIVE ESX Driver for VMware vsphere 6.0 Release Notes

Mellanox OFED for FreeBSD for ConnectX-4/ConnectX-5 Release Note. Rev 3.4.1

Choosing the Best Network Interface Card for Cloud Mellanox ConnectX -3 Pro EN vs. Intel XL710

Red Hat Enterprise Linux (RHEL) 7.5-ALT Driver Release Notes

Highest Levels of Scalability Simplified Network Manageability Maximum System Productivity

SUSE Linux Enterprise Server (SLES) 12 SP2 Driver SLES 12 SP2

Mellanox NATIVE ESX Driver for VMware vsphere 6.0 Release Notes

Benefits of 25, 40, and 50GbE Networks for Ceph and Hyper- Converged Infrastructure John F. Kim Mellanox Technologies

Mellanox NATIVE ESX Driver for VMware vsphere 6.0 Release Notes

VPI / InfiniBand. Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability

VPI / InfiniBand. Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability

WinOF-2 Release Notes

Mellanox Virtual Modular Switch

Mellanox ConnectX-4 NATIVE ESXi Driver for VMware vsphere 5.5/6.0 Release Notes. Rev /

Mellanox GPUDirect RDMA User Manual

Application Acceleration Beyond Flash Storage

Mellanox NIC s Performance Report with DPDK Rev 1.0

Mellanox ConnectX-4 NATIVE ESX Driver for VMware vsphere 5.5/6.0 Release Notes

Mellanox GPUDirect RDMA User Manual

Mellanox Innova IPsec 4 Lx Ethernet Adapter Quick Start Guide

Spark Over RDMA: Accelerate Big Data SC Asia 2018 Ido Shamay Mellanox Technologies

SUSE Linux Enterprise Server (SLES) 15 Inbox Driver Release Notes SLES 15

Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability

N V M e o v e r F a b r i c s -

Meltdown and Spectre Interconnect Performance Evaluation Jan Mellanox Technologies

SUSE Linux Enterprise Server (SLES) 15 Inbox Driver User Manual

InfiniBand Networked Flash Storage

Building the Most Efficient Machine Learning System

BIG DATA AND HADOOP ON THE ZFS STORAGE APPLIANCE

Uncompromising Performance Elastic Network Manageability Maximum System Productivity

Cisco and Cloudera Deliver WorldClass Solutions for Powering the Enterprise Data Hub alerts, etc. Organizations need the right technology and infrastr

STAR-CCM+ Performance Benchmark and Profiling. July 2014

Can Parallel Replication Benefit Hadoop Distributed File System for High Performance Interconnects?

SUSE Linux Enterprise Server (SLES) 12 SP2 Driver User Manual

Mellanox ConnectX -3 Pro Firmware Release Notes

Altair OptiStruct 13.0 Performance Benchmark and Profiling. May 2015

Performance Optimizations via Connect-IB and Dynamically Connected Transport Service for Maximum Performance on LS-DYNA

A Plugin-based Approach to Exploit RDMA Benefits for Apache and Enterprise HDFS

Mellanox MLX4_EN Driver for VMware README

Ethernet. High-Performance Ethernet Adapter Cards

RHEL6.x Deployment over iscsi over IPoIB Interfaces

Chelsio Communications. Meeting Today s Datacenter Challenges. Produced by Tabor Custom Publishing in conjunction with: CUSTOM PUBLISHING

Building the Most Efficient Machine Learning System

Configuring Mellanox Hardware for VPI Operation Application Note

NAMD Performance Benchmark and Profiling. January 2015

Accelerating Real-Time Big Data. Breaking the limitations of captive NVMe storage

Accelerate Big Data Insights

NVMe over Universal RDMA Fabrics

Dynamix QSA TM Port Adapter Family

Apache Spark Graph Performance with Memory1. February Page 1 of 13

Low-Overhead Flash Disaggregation via NVMe-over-Fabrics Vijay Balakrishnan Memory Solutions Lab. Samsung Semiconductor, Inc.

Mellanox GPUDirect RDMA User Manual

Solutions for Scalable HPC

Low-Overhead Flash Disaggregation via NVMe-over-Fabrics

Implementing SQL Server 2016 with Microsoft Storage Spaces Direct on Dell EMC PowerEdge R730xd

Interconnect Your Future

In-Network Computing. Paving the Road to Exascale. 5th Annual MVAPICH User Group (MUG) Meeting, August 2017

FlashGrid Software Enables Converged and Hyper-Converged Appliances for Oracle* RAC

An Oracle White Paper December Accelerating Deployment of Virtualized Infrastructures with the Oracle VM Blade Cluster Reference Configuration

Uncompromising Performance. Elastic Network Manageability. Maximum System Productivity.

WHITEPAPER. Improve Hadoop Performance with Memblaze PBlaze SSD

Mellanox ConnectX-4/ConnectX-5 NATIVE ESXi Driver for VMware vsphere 6.7 Release Notes

The Future of High Performance Interconnects

Optimizing Apache Spark with Memory1. July Page 1 of 14

Service Oriented Performance Analysis

OCTOPUS Performance Benchmark and Profiling. June 2015

WinOF VPI for Windows Installation Guide

Birds of a Feather Presentation

Micron and Hortonworks Power Advanced Big Data Solutions

SNIA Developers Conference - Growth of the iscsi RDMA (iser) Ecosystem

In partnership with. VelocityAI REFERENCE ARCHITECTURE WHITE PAPER

Red Hat Enterprise Linux (RHEL) 7.3 Driver Release Notes

Mellanox Technologies Maximize Cluster Performance and Productivity. Gilad Shainer, October, 2007

OCP3. 0. ConnectX Ethernet Adapter Cards for OCP Spec 3.0

AWS & Intel: A Partnership Dedicated to fueling your Innovations. Thomas Kellerer BDM CSP, Intel Central Europe

Dell PowerEdge R730xd Servers with Samsung SM1715 NVMe Drives Powers the Aerospike Fraud Prevention Benchmark

IBM Hortonworks Design Guide 14-Sep-17 v1

DELL EMC ISILON F800 AND H600 I/O PERFORMANCE

Best Practices for Setting BIOS Parameters for Performance

Transcription:

WHITE PAPER Accelerating Hadoop Applications with the MapR Distribution Using Flash Storage and High-Speed Ethernet Contents Background... 2 The MapR Distribution... 2 Mellanox Ethernet Solution... 3 Test Setup and Results... 4 Conclusion... 6 Trademarks and special notices... 6

page 2 Background The adoption of mobile technology and e-commerce with their pervasive use in everyday life has resulted in a massive influx of data. This provides the opportunity for companies to transform the flood of data (often referred to as Big Data ) to actionable insights for competitive advantage. Organizations that successfully leverage this have increased their innovation, operational efficiency and consequently, revenue growth. However, the practical usage of the data depends on the ability to store, manage and analyze it efficiently and quickly. Apache Hadoop was born out of the need to process Big Data. Hadoop, which was initially designed for web indexing, has moved far beyond its original purpose and is increasingly becoming the go-to framework for large-scale, data-intensive deployments. Hadoop includes the MapReduce framework, that provides distributed analytics processing for mining the data; and the Hadoop Distributed File System (HDFS), which provides a scalable storage mechanism. Apache HDFS is written in Java and runs on different operating systems. However, Hadoop was designed from the beginning to accommodate multiple filesystem implementations. The MapR Distribution MapR Technologies, an enterprise software company provides a complete distribution of Apache Hadoop and continues to be the fastest Hadoop distribution in the market. It combines over twenty different open source packages from the Hadoop ecosystem along with enterprise-grade features that provide unique capabilities for data management, protection and business continuity. The MapR Distribution includes several distinct advantages including: Data ingestion via NFS as a real-time stream Analysis performed immediately on the data as it is loaded Automated response execution These advanced capabilities are largely made possible due to the innovation in the underlying file system: MapR-FS. Figure 1. The MapR Distribution including Apache Hadoop

page 3 Due to the advantages of MapR-FS and the increasing usage of flash storage in production grade Hadoop environments, network bandwidth often becomes the bottleneck during data ingestion and analytics. Investing in fast servers and flash storage for MapR does not make sense if performance is restricted by the network. In this whitepaper, we evaluate the need for 40Gb/s networks in such environments. Mellanox Ethernet Solution Mellanox offers a complete product line of end-to-end 10/25/40/50/56/100GbE Ethernet solutions tailored for Big Data applications like Hadoop and NoSQL. Mellanox deliver industry-leading Figure 2. Mellanox Ethernet Switches performance, scalability, reliability and power savings for advanced data center applications. Mellanox Ethernet switches feature consistently low latency and can support a variety of non-blocking, lossless fabric designs. Further with the Mellanox NEOTM network orchestration platform, network administrators can leverage existing data center fabric management solutions to deploy, orchestrate and monitor a large-scale cluster easily. Big Data applications utilizing TCP or UDP over IP transport can achieve the highest throughput and application density using hardware-based stateless offloads and flow steering engines in ConnectX-3 Pro network adapters. These advanced stateless offloads reduce CPU overhead in IP packet processing thereby allowing completion of heavier analytic workloads in less time in the Big Data cluster. Sockets acceleration software further increases performance for latency sensitive applications, and faster network speeds such as 40 and 56Gb/s support greater throughput. To connect the clusters, Mellanox copper, active optical cables, and transceivers offer reliable connections at speed from 10 to 100Gb/s with the highest quality, featuring error rates up to 100x lower than industry standards. Figure 3. Mellanox ConnectX-3 Pro 40Gb Ethernet Adapter Figure 4. Mellanox ConnectX-3 Pro 10Gb Ethernet Adapter Figure 5. Mellanox 40/56GbE QSFP Copper Cable Figure 6. Mellanox SFP+ Copper Cable

page 4 Test Setup and Results The test setup consists of 5 servers connected with a Mellanox SX1036 40Gb Ethernet switch. The servers run RedHat Enterprise Linux 6.5 and the MapR Distribution 5.0 with hyper-threading turned off. The data was stored on the SSDs (on flash). Table 1 shows the configuration of the setup. Figure 7. Map-R Hadoop 5-node 40GbE deployment Configuration Master Node Data Node Network Fabrics Software Stack 2x 8 Core Intel(R) Xeon(R) CPU E5-2680 @ 2.70GHz 96GB DDR3 RAM 1x HGST FlashMax II Capacity SSD (4800GB) 2x 8 Core Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.70GHz 64GB DDR3 RAM 2x HGST FlashMax II Capacity SSDs (4800GB) Mellanox ConnectX-3 40Gb Ethernet (MCX314A-BCBT) Mellanox ConnectX-3 10Gb Ethernet (MCX312A-XCBT) Intel 10Gb Ethernet Mellanox SX1036 36-port Ethernet switch Operating System: RedHat Enterprise Linux 6.5 MapR Community Edition - 5.0 (Hadoop YARN 2.7.0) MLNX_OFED_LINUX-3.0-2.0.1

page 5 For the convenience of testing, we used standard QSA module with QSFP cable to connect 10 Gb/s adapters to SX1036 switch. There are many options in terms of adapters, cables and switches. Refer to Mellanox website for more details. TeraSort Benchmark Results TeraSort is probably the best known Hadoop benchmark. The goal of TeraSort is to sort 1TB of data as quickly as possible. It stresses both the MapReduce layer as well as the underlying distributed file system. Figure 8 below shows the total execution time to sort 1TB of data in our MapR cluster. The total execution time improved by 10% by moving from Intel 10GbE to Mellanox ConnectX-3 Pro 10GbE adapters and by a further 60% by moving to higher speed 40GbE network. In addition to the 3.2X improvement in execution time, the CPU usage also decreases when migrating from Intel to Mellanox network as shown in Figure 9. This allows more Hadoop jobs to run on the server while simultaneously generating faster analytics results. Our investigation led to the finding that there were significant pagefaults while running on Intel 10Gb/s network, which explains the reason for poor execution time and CPU usage 1. 1 The test was done with out-of-box driver from RedHat Enterprise Linux 6.5. Further, no network centric tuning was done with Mellanox or Intel drivers. Figure 8. TeraSort Execution Time Figure 9. TeraSort CPU Utilization

page 6 TestDFSIO Benchmarking Results TestDFSIO is a standard example of a benchmark that measures the capacity of HDFS for reading and writing bulk data. The test measures the time taken to create a number of large files, and then uses those same files as inputs as a test to measure the read performance an HDFS instance can sustain. Figure 10 shows the aggregated throughput generated by TestDFSIO benchmark. It can be seen that the 10Gb/s network becomes the bottleneck and in order to fully utilize the cluster one needs to switch to a higher speed 40Gb/s network, in this case using the Mellanox ConnectX-3 Pro 40Gb Ethernet adapter. Figure 10. TestDFSIO Per-Node Bandwidth Conclusion Both TeraSort and TestDFSIO demonstrated that the network often becomes the bottleneck and limits the capabilities provided by MapR-FS. They further show the need to migrate to higher speed network to take the full advantage of flash storage, which is increasingly becoming common on a production grade Hadoop cluster due to their overall favorable ROI. The above benchmarks use standard TCP and UDP transport and additional performance gains are likely by enabling RDMA/RoCE (RDMA over Converged Ethernet) in the future. Our analysis also opens up the opportunity for future testing to understand the benefits of 25, 50 and 100G Ethernet speeds for Hadoop applications like Spark in such environment. 350 Oakmead Parkway, Suite 100, Sunnyvale, CA 94085 Tel: 408-970-3400 Fax: 408-970-3403 www.mellanox.com Copyright 2015. Mellanox Technologies. All rights reserved. MMellanox, BridgeX, ConnectX, CORE-Direct, InfiniBridge, InfiniHost, InfiniScale, PhyX, SwitchX, Virtual Protocol Interconnect and Voltaire are registered trademarks of Mellanox Technologies, Ltd. Connect-IB, CoolBox, FabricIT, Mellanox Federal Systems, Mellanox Software Defined Storage, MetroX, MLNX-OS, ScalableHPC, Unbreakable-Link, UFM and Unified Fabric Manager are trademarks of Mellanox Technologies, Ltd. All other trademarks are property of their respective owners. MLNX-31-417