FROM HPC TO THE CLOUD WITH AMQP AND OPEN SOURCE SOFTWARE

Similar documents
Evolving HPC Solutions Using Open Source Software & Industry-Standard Hardware

Next-Generation AMQP Messaging Performance, Architectures, and Ecosystems with Red Hat Enterprise MRG. Bryan Che MRG Product Manager Red Hat, Inc.

SR-IOV Support for Virtualization on InfiniBand Clusters: Early Experience

Red Hat Enterprise Linux MRG Red Hat Network Satellite Red Hat Enterprise Virtualization JBoss Cloud

Infiniband and RDMA Technology. Doug Ledford

QuickSpecs. HP InfiniBand Options for HP BladeSystems c-class. Overview

Improving Application Performance and Predictability using Multiple Virtual Lanes in Modern Multi-Core InfiniBand Clusters

MRG - AMQP trading system in a rack. Carl Trieloff Senior Consulting Software Engineer/ Director MRG Red Hat, Inc.

SUN CUSTOMER READY HPC CLUSTER: REFERENCE CONFIGURATIONS WITH SUN FIRE X4100, X4200, AND X4600 SERVERS Jeff Lu, Systems Group Sun BluePrints OnLine

NFS/RDMA over 40Gbps iwarp Wael Noureddine Chelsio Communications

Best Practices for Setting BIOS Parameters for Performance

Study. Dhabaleswar. K. Panda. The Ohio State University HPIDC '09

VM Migration Acceleration over 40GigE Meet SLA & Maximize ROI

Performance Analysis and Evaluation of Mellanox ConnectX InfiniBand Architecture with Multi-Core Platforms

GROMACS Performance Benchmark and Profiling. September 2012

ICON Performance Benchmark and Profiling. March 2012

DB2 purescale: High Performance with High-Speed Fabrics. Author: Steve Rees Date: April 5, 2011

CESM (Community Earth System Model) Performance Benchmark and Profiling. August 2011

TIBCO, HP and Mellanox High Performance Extreme Low Latency Messaging

GROMACS Performance Benchmark and Profiling. August 2011

CP2K Performance Benchmark and Profiling. April 2011

DESCRIPTION GHz, 1.536TB shared memory RAM, and 20.48TB RAW internal storage teraflops About ScaleMP

2-Port 40 Gb InfiniBand Expansion Card (CFFh) for IBM BladeCenter IBM BladeCenter at-a-glance guide

AcuSolve Performance Benchmark and Profiling. October 2011

AMBER 11 Performance Benchmark and Profiling. July 2011

HP InfiniBand Options for HP ProLiant and Integrity Servers Overview

SNIA Developers Conference - Growth of the iscsi RDMA (iser) Ecosystem

Mellanox Technologies Maximize Cluster Performance and Productivity. Gilad Shainer, October, 2007

CP2K Performance Benchmark and Profiling. April 2011

MM5 Modeling System Performance Research and Profiling. March 2009

Server Networking e Virtual Data Center

STAR-CCM+ Performance Benchmark and Profiling. July 2014

GW2000h w/gw175h/q F1 specifications

Reducing Network Contention with Mixed Workloads on Modern Multicore Clusters

The following InfiniBand products based on Mellanox technology are available for the HP BladeSystem c-class from HP:

The following InfiniBand products based on Mellanox technology are available for the HP BladeSystem c-class from HP:

Dell EMC Ready Bundle for HPC Digital Manufacturing Dassault Systѐmes Simulia Abaqus Performance

Accelerating Hadoop Applications with the MapR Distribution Using Flash Storage and High-Speed Ethernet

ABySS Performance Benchmark and Profiling. May 2010

IBM WebSphere MQ Low Latency Messaging Software Tested With Arista 10 Gigabit Ethernet Switch and Mellanox ConnectX

Application Acceleration Beyond Flash Storage

Sharing High-Performance Devices Across Multiple Virtual Machines

Flex System IB port FDR InfiniBand Adapter Lenovo Press Product Guide

Introduction to Infiniband

Creating an agile infrastructure with Virtualized I/O

PERFORMANCE ACCELERATED Mellanox InfiniBand Adapters Provide Advanced Levels of Data Center IT Performance, Productivity and Efficiency

Scheduling Strategies for HPC as a Service (HPCaaS) for Bio-Science Applications

Red Hat Enterprise Virtualization and KVM Roadmap. Scott M. Herold Product Management - Red Hat Virtualization Technologies

Performance Optimizations via Connect-IB and Dynamically Connected Transport Service for Maximum Performance on LS-DYNA

In the multi-core age, How do larger, faster and cheaper and more responsive memory sub-systems affect data management? Dhabaleswar K.

AcuSolve Performance Benchmark and Profiling. October 2011

HP s Performance Oriented Datacenter

Implementing Storage in Intel Omni-Path Architecture Fabrics

Altair RADIOSS Performance Benchmark and Profiling. May 2013

Data Sheet Fujitsu Server PRIMERGY CX250 S2 Dual Socket Server Node

A Low Latency Solution Stack for High Frequency Trading. High-Frequency Trading. Solution. White Paper

The rcuda middleware and applications

Z RESEARCH, Inc. Commoditizing Supercomputing and Superstorage. Massive Distributed Storage over InfiniBand RDMA

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments

HYCOM Performance Benchmark and Profiling

VPI / InfiniBand. Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability

Dell EMC Ready Bundle for HPC Digital Manufacturing ANSYS Performance

Oracle Exadata: Strategy and Roadmap

VPI / InfiniBand. Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability

To Infiniband or Not Infiniband, One Site s s Perspective. Steve Woods MCNC

HP InfiniBand Options for HP ProLiant and Integrity Servers Overview

QuickSpecs. HP ProLiant m710 Server Cartridge. Overview. HP ProLiant m710 Server Cartridge. Retired

InfiniBand Networked Flash Storage

Flex System EN port 10Gb Ethernet Adapter Product Guide

NAMD Performance Benchmark and Profiling. November 2010

INFOBrief. Dell PowerEdge Key Points

OFED Storage Protocols

10 Steps to Virtualization

Voltaire. Fast I/O for XEN using RDMA Technologies. The Grid Interconnect Company. April 2005 Yaron Haviv, Voltaire, CTO

Linux Automation.

Red Hat Enterprise Linux 6.1 High Performance Network with MRG

NAMD Performance Benchmark and Profiling. January 2015

HP BLc Intel 4X QDR InfiniBand Switch Release Notes. Firmware Version

Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability

A first look at 100 Gbps LAN technologies, with an emphasis on future DAQ applications.

Himeno Performance Benchmark and Profiling. December 2010

Unified Runtime for PGAS and MPI over OFED

Low-Overhead Flash Disaggregation via NVMe-over-Fabrics Vijay Balakrishnan Memory Solutions Lab. Samsung Semiconductor, Inc.

10 Gbit/s Challenge inside the Openlab framework

Key Measures of InfiniBand Performance in the Data Center. Driving Metrics for End User Benefits

HP 10 GbE Dual Port Mezzanine Adapter for HP BladeSystem c-class server blades

SNAP Performance Benchmark and Profiling. April 2014

Multifunction Networking Adapters

NEMO Performance Benchmark and Profiling. May 2011

ARISTA: Improving Application Performance While Reducing Complexity

HP ProLiant m300 1P C2750 CPU 32GB Configure-to-order Server Cartridge B21

Meltdown and Spectre Interconnect Performance Evaluation Jan Mellanox Technologies

Grid for Financial Services

PUBLIC AND HYBRID CLOUD: BREAKING DOWN BARRIERS

Data Sheet FUJITSU Server PRIMERGY CX2550 M4 Dual Socket Server Node

Emerging Technologies for HPC Storage

Mark Falco Oracle Coherence Development

Extending InfiniBand Globally

Oracle Exadata X7. Uwe Kirchhoff Oracle ACS - Delivery Senior Principal Service Delivery Engineer

Low-Overhead Flash Disaggregation via NVMe-over-Fabrics

Transcription:

FROM HPC TO THE CLOUD WITH AMQP AND OPEN SOURCE SOFTWARE Carl Trieloff cctrieloff@redhat.com Red Hat Lee Fisher lee.fisher@hp.com Hewlett-Packard High Performance Computing on Wall Street conference 14 September, 2009 1

From simulation to trade Scale up Grid Internal pool Messaging scheduler Messaging Scale out Another internal division trader Latency trade External resource Eg EC2 2

Red Hat Enterprise MRG Integrated platform for high performance distributed computing High speed, interoperable, open standard Messaging Deterministic, low-latency Realtime kernel High performance & throughput computing Grid scheduler for distributed workloads and Cloud computing 3

AMQP, HP Performance, scale up. 12000000 Single HP Nehalem BL460c 40G Infiniband AMQP Perftest 10000000 Messages/Sec 8000000 6000000 4000000 8 bytes 64 Bytes 256 Bytes 1024 Bytes 2000000 0 8 Broker 4 Broker 2 Broker 1 Broker Number of Brokers on the Server two Intel(R) Xeon(R) CPU X5570 @ 2.93GHz per blade (Nehalem 2.93 GHz, 8MB L3 cache, 95W) Memory 24GB(6x4GB), Memory Type DDR3-1333, HT, Turbo 2/2/3/3) Infiniband 4X QDR IB Dual-port Mezzanine HCAs(1 port connected) Infiniband Switch BLc 4X QDR IB Switch 4

AMQP Messaging on 8-node HP Nehalem Infiniband 40Gps > 11 M mes/s 7000000 3.1 3.1 3.1 3 6000000 2.5 5000000 Messages/Sec 4000000 3000000 2000000 2 1.5 1 Nehalem Harperton % Nehalem vs Harperton 1000000 0.5 0 4 Broker 2 Broker 1 Broker Number of Brokers per Server 0 5

KVM Performance AMQP Messaging Intel Nahalem 2 10Gbit Vt-D > 1 M mes/s RHEL 5.4 KVM AMQP 2-Guest 1200000 900 104 6081 1023869 800 1000000 902689 880965 700 Messages / Sec 800000 600000 400000 200000 804045 74 1297 5554 65 369145 210634 600 500 400 300 200 Msg/sec Throughput MB/sec 100 0 16 32 64 128 256 512 1024 2048 4096 Msg Size (bytes) 0 6

MRG Messaging Infiniband RDMA Latency: Under 40 Microseconds Reliably Acknowledged 0.0480 MRG Messaging Latency Test on HP BL460c G6 Infiniband 100K Message Rate 0.0460 Average Latency (ms) 0.0440 0.0420 0.0400 0.0380 32 Bytes RDMA Nehalem 256 Bytes RDMA Nehalem 1024 Bytes RDMA Nehalem 0.0360 0.0340 1 3 5 7 11 15 19 23 27 31 35 39 43 47 51 55 59 63 67 71 75 79 83 87 91 95 99 9 13 17 21 25 29 33 37 41 45 49 53 57 61 65 69 73 77 81 85 89 93 97 7

Components of the Solution Stack Solutions still matter in an industry-standard, open source world Tuning & working in labs Red Hat MRG Tuning tools Red Hat MRG Messaging / Grid Red Hat / HP Systems HP Voltaire / Red Hat RDMA Red Hat MRG - Realtime HP reduced SMI BIOS's HP compute & storage FSI-HPC Solution Stack Users Application Environment Workload Middleware Integrated Systems Server Interconnect L2 Fabric Operating System BIOS X86-64 Server Architecture Services Determinism, and performance needs to work at each layer, HP & Red Hat are partnered across the stack 8

Hardware matters Scale-Up Blades Scale-Out Rack-Optimized SL6000 Today s RFP Metrics: Performance/Watt Performance/BTU Performance/Rack HP Low Latency Lab with MRG + Red Hat MRG Lab with HP BL460/BL685 & IB 9

Dealing with SMIs HP BIOS Option for Low Latency Apps Disable frequent SMIs used for Dynamic Power Savings Mode, CPU Utilization monitoring, P-state monitoring and ECC reporting Benefits both RHEL & MRG operating environments. Latency spikes with standard BIOS settings Latencies when SMIs disabled in BIOS 10

MRG Realtime RHEL on HP systems Enables applications and transactions to run predictably, with guaranteed response times Upgrades RHEL 5 to realtime OS Provides replacement kernel for RHEL5; x86/x86_64 Preserves RHEL Application Compatibility Certified on HP hardware, see Red Hat / HP certifications Response time Time 11

MRG Realtime Scheduling Latency Vanilla Min: 1 Max: 2857 Mean: 11.47 Mode: 9.00 Median: 9.00 Std. Deviation: 54.94 MRG RT Min: 4 Max: 43 Mean: 8.34 Mode: 8.00 Median: 8.00 Std. Deviation: 1.49 12

Networking matters Voltaire DDR and QDR InfiniBand: 36 QDR QSFP ports Ethernet mngt port LEDs USB port Serial port Test Configuration: Two Nehalem-based server w/ ConnectX PCI-E HCAs, back-to-back QDR ConnectX HCA running at QDR DDR ConnectX HCA running at DDR RHEL5 UPDATE 2 Mellanox VERBs Performance Test RoEE RDMA on Enhanced Ethernet RoEE is defined to be a verbs compliant IB transport running over the emerging IEEE Converged Enhanced Ethernet standard www.openfabrics.org/archives/spring2009sonoma/monday/grun.pdf 13

MRG Grid Provides leading high performance & high throughput computing: Brings advantages of scale-out and flexible deployment to any application or workload Delivers better asset utilization, allowing applications to take advantage of all available computing resources Enables building cloud infrastructure and aggregating multiple clouds: Integrated support for virtualization as well as public clouds Seamlessly aggregates multiple cloud resources into one compute pool Provides seamless and flexible computing across: Local grids Remote grids Private and hybrid clouds ( EC2 Public clouds (Amazon Cycle-harvesting from desktop PCs 14

Based on Condor and Includes: Enterprise Supportability From Red Hat Web-Based Management Console Unified management across all of MRG for job, system, license management, and workload management/monitoring Low Latency Scheduling Enable job submission to Condor via AMQP Messaging clients Enable sub-second, low-latency scheduling for sub-second jobs Virtualization Support via libvirt Integration Support scheduling of virtual machines on Linux using libvirt API's Cloud Integration with Amazon Ec2 Enable automatic cloud provisioning, job submission, results storage, teardown via Condor scheduler Extensible, it can be a dependency for other jobs or executed based on rules (e.g. add capacity in in the cloud if local grid out of ( capacity Concurrency Limits Set limits on how much of a certain resource (e.g. software licenses, db connections) can be used at once Dynamic Slots Mark slots as partitionable and sub-divide them dynamically so that more than one job can occupy a slot at once 15

Testing and developing solutions working together...delivered in reference papers & certifications Throughput Memory Usage Red Hat / HP White Paper: 74 72 70 68 cache buff free 66 64 62 60 1-GigE 10-GigE IPoIB IB SDP IB RDMA 16

Additional Information www.redhat.com/mrg www.hp.com/go/fsi 17