Checklist for Selecting and Deploying Scalable Clusters with InfiniBand Fabrics

Size: px
Start display at page:

Download "Checklist for Selecting and Deploying Scalable Clusters with InfiniBand Fabrics"

Transcription

1 Checklist for Selecting and Deploying Scalable Clusters with InfiniBand Fabrics Lloyd Dickman, CTO InfiniBand Products Host Solutions Group QLogic Corporation November 13, SC07, Exhibitor Forum 1

2 InfiniBand is the Fabric of Choice for HPC Large and growing standards-based ecosystem Aggressive technology roadmap Low-latency, high messaging rate implementations in market Evidenced by increasing traction in Top500 2

3 3 InfiniBand Interoperability

4 Interdependent Tradeoffs Budget CapEx + OpEx Decisions Host Architecture Fabric Topology Fabric Convergence / Unification RAS, Power / Cooling Software host applications, ULPs and fabric Performance 4

5 For HPC, Application Messaging Profiles Matter! Message Rate Application Accelerators Moore s Law Multi- Moore s Law Clock Rate Fine-Grain LATENCY and Proteomics Molecular Dynamics MESSAGE RATE DOMINATE New Applications Quantum Chromodynamics Reservoir Simulation Magneto Hydrodynamics MultiPhysics Time to Solution Climate Weather Embarrassingly Parallel CFD Legacy Applications Genomics Coarse-Grain BANDWIDTH DOMINATES Message Size 5

6 Application Messaging Patterns Change with Scale 16,000,000 14,000,000 12,000,000 Number of Messages 10,000,000 8,000,000 6,000,000 4,000, ,000, ,024 2,048 Message Size (Bytes) 4,096 8,192 16,384 32,768 65, , , , Processors 32 Source: PathScale measurements, averages per node, DL_POLY_2 AWE Medium case, 19dec2005 6

7 QLogic InfiniBand Provides Superior Application Scaling Comparison Benchmark QLE7140 SDR QLE7240 DDR QLE7280 DDR vs Competitive Point-to-Point Point-to-Point Bandwidth Point-to-Point Latency Point-to-Point Message Rate (non-coalesced) OSU Bandwidth (peak) OSU Latency (0 Byte) OSU Message 4 processes/node 0.9 GB/s 1.6 µs 9 M/s 1.5 GB/s 1.5 µs 12 M/s 1.9 GB/s 1.4 µs 13 M/s +24% Tie 2.6X Scaling Scalable Latency Scalable Message Rate (non-coalesced) HPCC Random Ring 32 cores HPCC MPI Random 32 cores 1.5 µs.12 GUPs 1.4 µs.12 GUPs 1.3 µs.13 GUPs 1.8X 7.6X Scaling results are for 8 nodes / 32 cores. Latency results include a single switch crossing. Results use native MPIs: InfiniPath MPI for QLogic 7 Point-to-point microbenchmarks are no longer sufficient predictors of performance at scale.

8 137_lu SPEC MPI2007 Results Confirm Scalability QLogic SDR to 512 s _zeusmp Relative Scaling 104_milc 107_leslie3d 113_GemsFDTD 115_fds4 121_pop2 122_tachyon 126_lammps 127_wrf2 128_GAPgeofem 129_tera_tf 130_socorro

9 Bandwidth Must Account for ALL Traffic 9 50% Link Utilization

10 Host-to-Fabric Rails InfiniBand Data Center Fabric

11 Fabric Topology Considerations InfiniBand InfiniBand Data Data Center Center Fabric Fabric Multi-rail Fat Tree / FBB Oversubscription Asymmetric Topology 11

12 QLogic InfiniBand Fabric Products Fabric Switches MultiProtocol Fabric Director Modules 4X SDR 288 Ports 4X DDR 144 Ports 7 U 7 U SDR and DDR Spines 4 U Edge Fabric Switch 1 U 24 Ports 2 U 14U 96 Ports 48 Ports 24 Ports 19 rack Production quality DDR since April 2006, Over 50,000 ports to date with extensive OEM qualifications 12

13 Matching Fabric to Node Count Optimal Fabric Size for today s 24-port switch chips 1 tier 24 ports 2 tier 288 ports 3 tier 3,456 ports 288 Ports 144 Ports 96 Ports 24 Ports 48 Ports Ports Ports 9020 Switch / FC / 10 GbE Gateway 13

14 Advanced Design Capabilities Signal Integrity Internal and external bit lanes capable of 10Gb/sec Significant signal integrity challenge Leveraging established QLogic center for signal integrity analysis at headquarters Thermal Analysis Chip power nearly triple previous generation Leveraging QLogic CFD capabilities 14

15 Unified InfiniBand Fabric Reduces Cost Typical Cluster Interconnect Unified Fabric Interconnect AAAAA AAAA a a a Unified Fabric Deployments Increasing AAAAA AAAA 15

16 Cost and Power Advantages of Converged Fabric CF cost and power savings FC HBAs and switches Ethernet NICs and switches Cables $600,000 $500,000 Cost Savings of a Converged Fabric Additional cost savings Simpler management: single network $400,000 $300,000 Comparisons for a 64-node cluster CF advantage increases for larger clusters 30-38% cost savings on a 400/800-node FSC cluster Power savings from CF ~30% vs. GigE $200,000 $100,000 $0 Converged Fabric 1GbE+FC 10GbE+FC Storage $35,200 $144,000 $144,000 10GbE $200 $200 $283,400 Ethernet $0 $61,280 0 IB $113, Servers $115,200 $115,200 $115,200 Source: HP BoF, Converged Fabrics, SC07 16

17 Connectivity to ALL Storage Direct Connect InfiniBand Storage iser iscsi iser EtherNet Gateway Fibre Channel Gateway SRP 17

18 Connectivity to ALL Sockets Applications and Ethernet Networks Intra-Cluster Sockets Communications SDP VNIC TCP/IP IPoIB-UD IPoIB-CM 10 GbE Gateway EtherNet Emulation 18

19 OFED+ Provides OpenFabrics Benefits Plus Optional Value Added Capabilities OFED Stack Robust ULPs and Tools Accelerated Accelerated HPC Stack Sockets IP Transport HP-MPI Intel MPI MVAPICH OpenMPI Scali MPI Lustre QLogic SRP QLogic SDP QLogic VNIC InfiniPath Fabric Suite QLogic MPI Open MPI HP-MPI Scali MPI MVAPICH MPICH2 Standard Linux TCP/IP Stack IPoIB SDP VNIC SRP iser RDS OpenSM udapl OpenFabrics Verbs PSM API OFED Driver QLogic 7200 Series Adapters 19

20 Large Fabrics Challenges Scale of InfiniBand deployments growing rapidly 1000s of Nodes, soon 10,000s of Nodes Fabric is the cluster 1 wire vision being deployed now Requires high degree of fabric stability Fabric Diagnosis and Resiliency Many moving parts at large scale Demands of End Nodes on SA increasing 20

21 InfiniBand Fabric Suite Automates Cluster Install, Verification and Administration For both Hosts and Switches Rich set of tools and capabilities for ongoing Fabric Administration Easy to use TUI and CLI Easy integration into site specific tools Rapidly Pinpoints fabric errors Extensive verification and diagnostic capabilities Industry Leading Performance and Scalability 1024 node fabric initialization in <20 seconds Fabric Change in < 12 seconds Trap Based event notification & subscription supported Load balances LID(s) across ISL(s) Highly Resilient No single point of failure Automatic Fabric Reconfiguration Complete IBTA Support Optional SA queries (Trace Route, Link records, etc) Lowering InfiniBand OpEx Flexible Deployment Options Embedded Fabric Management Host Fabric Management 21

22 Please Visit Us to Learn More About Complete InfiniBand Booth #261 Cluster Administration using InfiniBand Fabric Suite Advanced Technology Demonstrations HPCtrack Partnership Program 9020 InfiniBand Switch and Multi-Protocol Gateway Fluent Application running multiple data sets on a converged fabric with Fibre Channel storage via a gateway InfiniBand DDR Adapter for highly scaling applications 22

23 23 Thank You

Benchmarking Clusters with High Core-Count Nodes

Benchmarking Clusters with High Core-Count Nodes Benchmarking Clusters with High Core-Count Nodes 9 th LCI International Conference on High- Performance Clustered Computing 29 April, 2008 Tom Elken Manager, Performance Engineering QLogic Corporation

More information

QLogic in HPC Vendor Update IDC HPC User Forum April 16, 2008 Jeff Broughton Sr. Director Engineering Host Solutions Group

QLogic in HPC Vendor Update IDC HPC User Forum April 16, 2008 Jeff Broughton Sr. Director Engineering Host Solutions Group QLogic in HPC Vendor Update IDC HPC User Forum April 16, 2008 Jeff Broughton Sr. Director Engineering Host Solutions Group 1 Networking for Storage and HPC Leading supplier of Fibre Channel Leading supplier

More information

Mellanox Technologies Maximize Cluster Performance and Productivity. Gilad Shainer, October, 2007

Mellanox Technologies Maximize Cluster Performance and Productivity. Gilad Shainer, October, 2007 Mellanox Technologies Maximize Cluster Performance and Productivity Gilad Shainer, shainer@mellanox.com October, 27 Mellanox Technologies Hardware OEMs Servers And Blades Applications End-Users Enterprise

More information

Infiniband and RDMA Technology. Doug Ledford

Infiniband and RDMA Technology. Doug Ledford Infiniband and RDMA Technology Doug Ledford Top 500 Supercomputers Nov 2005 #5 Sandia National Labs, 4500 machines, 9000 CPUs, 38TFlops, 1 big headache Performance great...but... Adding new machines problematic

More information

OFED Storage Protocols

OFED Storage Protocols OFED Storage Protocols R. Pearson System Fabric Works, Inc. Agenda Why OFED Storage Introduction to OFED Storage Protocols OFED Storage Protocol Update 2 Why OFED Storage 3 Goals of I/O Consolidation Cluster

More information

Performance of HPC Applications over InfiniBand, 10 Gb and 1 Gb Ethernet. Swamy N. Kandadai and Xinghong He and

Performance of HPC Applications over InfiniBand, 10 Gb and 1 Gb Ethernet. Swamy N. Kandadai and Xinghong He and Performance of HPC Applications over InfiniBand, 10 Gb and 1 Gb Ethernet Swamy N. Kandadai and Xinghong He swamy@us.ibm.com and xinghong@us.ibm.com ABSTRACT: We compare the performance of several applications

More information

Informatix Solutions INFINIBAND OVERVIEW. - Informatix Solutions, Page 1 Version 1.0

Informatix Solutions INFINIBAND OVERVIEW. - Informatix Solutions, Page 1 Version 1.0 INFINIBAND OVERVIEW -, 2010 Page 1 Version 1.0 Why InfiniBand? Open and comprehensive standard with broad vendor support Standard defined by the InfiniBand Trade Association (Sun was a founder member,

More information

HPC Customer Requirements for OpenFabrics Software

HPC Customer Requirements for OpenFabrics Software HPC Customer Requirements for OpenFabrics Software Matt Leininger, Ph.D. Sandia National Laboratories Scalable Computing R&D Livermore, CA 16 November 2006 I'll focus on software requirements (well maybe)

More information

Introduction to Infiniband

Introduction to Infiniband Introduction to Infiniband FRNOG 22, April 4 th 2014 Yael Shenhav, Sr. Director of EMEA, APAC FAE, Application Engineering The InfiniBand Architecture Industry standard defined by the InfiniBand Trade

More information

NFS/RDMA over 40Gbps iwarp Wael Noureddine Chelsio Communications

NFS/RDMA over 40Gbps iwarp Wael Noureddine Chelsio Communications NFS/RDMA over 40Gbps iwarp Wael Noureddine Chelsio Communications Outline RDMA Motivating trends iwarp NFS over RDMA Overview Chelsio T5 support Performance results 2 Adoption Rate of 40GbE Source: Crehan

More information

Study. Dhabaleswar. K. Panda. The Ohio State University HPIDC '09

Study. Dhabaleswar. K. Panda. The Ohio State University HPIDC '09 RDMA over Ethernet - A Preliminary Study Hari Subramoni, Miao Luo, Ping Lai and Dhabaleswar. K. Panda Computer Science & Engineering Department The Ohio State University Introduction Problem Statement

More information

PERFORMANCE ACCELERATED Mellanox InfiniBand Adapters Provide Advanced Levels of Data Center IT Performance, Productivity and Efficiency

PERFORMANCE ACCELERATED Mellanox InfiniBand Adapters Provide Advanced Levels of Data Center IT Performance, Productivity and Efficiency PERFORMANCE ACCELERATED Mellanox InfiniBand Adapters Provide Advanced Levels of Data Center IT Performance, Productivity and Efficiency Mellanox continues its leadership providing InfiniBand Host Channel

More information

Introduction to High-Speed InfiniBand Interconnect

Introduction to High-Speed InfiniBand Interconnect Introduction to High-Speed InfiniBand Interconnect 2 What is InfiniBand? Industry standard defined by the InfiniBand Trade Association Originated in 1999 InfiniBand specification defines an input/output

More information

Application Acceleration Beyond Flash Storage

Application Acceleration Beyond Flash Storage Application Acceleration Beyond Flash Storage Session 303C Mellanox Technologies Flash Memory Summit July 2014 Accelerating Applications, Step-by-Step First Steps Make compute fast Moore s Law Make storage

More information

Workshop on High Performance Computing (HPC) Architecture and Applications in the ICTP October High Speed Network for HPC

Workshop on High Performance Computing (HPC) Architecture and Applications in the ICTP October High Speed Network for HPC 2494-6 Workshop on High Performance Computing (HPC) Architecture and Applications in the ICTP 14-25 October 2013 High Speed Network for HPC Moreno Baricevic & Stefano Cozzini CNR-IOM DEMOCRITOS Trieste

More information

LAMMPS, LS- DYNA, HPL, and WRF on iwarp vs. InfiniBand FDR

LAMMPS, LS- DYNA, HPL, and WRF on iwarp vs. InfiniBand FDR LAMMPS, LS- DYNA, HPL, and WRF on iwarp vs. InfiniBand FDR The use of InfiniBand as interconnect technology for HPC applications has been increasing over the past few years, replacing the aging Gigabit

More information

LAMMPS and WRF on iwarp vs. InfiniBand FDR

LAMMPS and WRF on iwarp vs. InfiniBand FDR LAMMPS and WRF on iwarp vs. InfiniBand FDR The use of InfiniBand as interconnect technology for HPC applications has been increasing over the past few years, replacing the aging Gigabit Ethernet as the

More information

2008 International ANSYS Conference

2008 International ANSYS Conference 2008 International ANSYS Conference Maximizing Productivity With InfiniBand-Based Clusters Gilad Shainer Director of Technical Marketing Mellanox Technologies 2008 ANSYS, Inc. All rights reserved. 1 ANSYS,

More information

Storage System. David Southwell, Ph.D President & CEO Obsidian Strategics Inc. BB:(+1)

Storage System. David Southwell, Ph.D President & CEO Obsidian Strategics Inc. BB:(+1) Storage InfiniBand Area Networks System David Southwell, Ph.D President & CEO Obsidian Strategics Inc. BB:(+1) 780.964.3283 dsouthwell@obsidianstrategics.com Agenda System Area Networks and Storage Pertinent

More information

Ethernet: The High Bandwidth Low-Latency Data Center Switching Fabric

Ethernet: The High Bandwidth Low-Latency Data Center Switching Fabric White PAPER Ethernet: The High Bandwidth Low-Latency Data Center Switching Fabric The transition to 10 GbE is under way in the data center, and the ubiquity and reliability of Ethernet make it a desirable

More information

QuickSpecs. HP InfiniBand Options for HP BladeSystems c-class. Overview

QuickSpecs. HP InfiniBand Options for HP BladeSystems c-class. Overview Overview HP supports 40Gbps (QDR) and 20Gbps (DDR) InfiniBand products that include mezzanine Host Channel Adapters (HCA) for server blades, switch blades for c-class enclosures, and rack switches and

More information

Fibre Channel. Roadmap to your SAN Future! HP Technology Forum & Expo June, 2007, Las Vegas Skip Jones, FCIA Chair, QLogic

Fibre Channel. Roadmap to your SAN Future! HP Technology Forum & Expo June, 2007, Las Vegas Skip Jones, FCIA Chair, QLogic Fibre Channel Roadmap to your SAN Future! HP Technology Forum & Expo June, 2007, Las Vegas Skip Jones, FCIA Chair, QLogic Fibre Channel Fibre Channel is a robust and evolving technology that enables: Leading

More information

VM Migration Acceleration over 40GigE Meet SLA & Maximize ROI

VM Migration Acceleration over 40GigE Meet SLA & Maximize ROI VM Migration Acceleration over 40GigE Meet SLA & Maximize ROI Mellanox Technologies Inc. Motti Beck, Director Marketing Motti@mellanox.com Topics Introduction to Mellanox Technologies Inc. Why Cloud SLA

More information

Windows OpenFabrics (WinOF)

Windows OpenFabrics (WinOF) Windows OpenFabrics (WinOF) Gilad Shainer, Mellanox Ishai Rabinovitz, Mellanox Stan Smith, Intel April 2008 Windows OpenFabrics (WinOF) Collaborative effort to develop, test and release OFA software for

More information

InfiniBand based storage target

InfiniBand based storage target Philippe BRUIANT Business Development Manager EMEA OpenFabrics Workshop Paris June 22-23, 2006 InfiniBand based storage target Tuesday, 27 June 2006 At-A-Glance Founded in Feb. 2000 by Veterans of Cheyenne

More information

To Infiniband or Not Infiniband, One Site s s Perspective. Steve Woods MCNC

To Infiniband or Not Infiniband, One Site s s Perspective. Steve Woods MCNC To Infiniband or Not Infiniband, One Site s s Perspective Steve Woods MCNC 1 Agenda Infiniband background Current configuration Base Performance Application performance experience Future Conclusions 2

More information

Lossless 10 Gigabit Ethernet: The Unifying Infrastructure for SAN and LAN Consolidation

Lossless 10 Gigabit Ethernet: The Unifying Infrastructure for SAN and LAN Consolidation . White Paper Lossless 10 Gigabit Ethernet: The Unifying Infrastructure for SAN and LAN Consolidation Introduction As organizations increasingly rely on IT to help enable, and even change, their business

More information

Learn Your Alphabet - SRIOV, NPIV, RoCE, iwarp to Pump Up Virtual Infrastructure Performance

Learn Your Alphabet - SRIOV, NPIV, RoCE, iwarp to Pump Up Virtual Infrastructure Performance Learn Your Alphabet - SRIOV, NPIV, RoCE, iwarp to Pump Up Virtual Infrastructure Performance TechTarget Dennis Martin 1 Agenda About Demartek I/O Virtualization Concepts RDMA Concepts Examples Demartek

More information

SR-IOV Support for Virtualization on InfiniBand Clusters: Early Experience

SR-IOV Support for Virtualization on InfiniBand Clusters: Early Experience SR-IOV Support for Virtualization on InfiniBand Clusters: Early Experience Jithin Jose, Mingzhe Li, Xiaoyi Lu, Krishna Kandalla, Mark Arnold and Dhabaleswar K. (DK) Panda Network-Based Computing Laboratory

More information

Welcome to the IBTA Fall Webinar Series

Welcome to the IBTA Fall Webinar Series Welcome to the IBTA Fall Webinar Series A four-part webinar series devoted to making I/O work for you Presented by the InfiniBand Trade Association The webinar will begin shortly. 1 September 23 October

More information

Comparing Server I/O Consolidation Solutions: iscsi, InfiniBand and FCoE. Gilles Chekroun Errol Roberts

Comparing Server I/O Consolidation Solutions: iscsi, InfiniBand and FCoE. Gilles Chekroun Errol Roberts Comparing Server I/O Consolidation Solutions: iscsi, InfiniBand and FCoE Gilles Chekroun Errol Roberts SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA. Member companies

More information

Mellanox InfiniBand Training IB Professional, Expert and Engineer Certifications

Mellanox InfiniBand Training IB Professional, Expert and Engineer Certifications About Mellanox On-Site courses Mellanox Academy offers on-site customized courses for maximum content flexibility to make the learning process as efficient and as effective as possible. Content flexibility

More information

OpenFabrics Alliance Interoperability Logo Group (OFILG) Dec 2011 Logo Event Report

OpenFabrics Alliance Interoperability Logo Group (OFILG) Dec 2011 Logo Event Report OpenFabrics Alliance Interoperability Logo Group (OFILG) Dec 2011 Logo Event Report UNH-IOL 121 Technology Drive, Suite 2 Durham, NH 03824 - +1-603-862-0090 OpenFabrics Interoperability Logo Group (OFILG)

More information

Creating an agile infrastructure with Virtualized I/O

Creating an agile infrastructure with Virtualized I/O etrading & Market Data Agile infrastructure Telecoms Data Center Grid Creating an agile infrastructure with Virtualized I/O Richard Croucher May 2009 Smart Infrastructure Solutions London New York Singapore

More information

Chelsio Communications. Meeting Today s Datacenter Challenges. Produced by Tabor Custom Publishing in conjunction with: CUSTOM PUBLISHING

Chelsio Communications. Meeting Today s Datacenter Challenges. Produced by Tabor Custom Publishing in conjunction with: CUSTOM PUBLISHING Meeting Today s Datacenter Challenges Produced by Tabor Custom Publishing in conjunction with: 1 Introduction In this era of Big Data, today s HPC systems are faced with unprecedented growth in the complexity

More information

InfiniBand Networked Flash Storage

InfiniBand Networked Flash Storage InfiniBand Networked Flash Storage Superior Performance, Efficiency and Scalability Motti Beck Director Enterprise Market Development, Mellanox Technologies Flash Memory Summit 2016 Santa Clara, CA 1 17PB

More information

Gateways to Discovery: Cyberinfrastructure for the Long Tail of Science

Gateways to Discovery: Cyberinfrastructure for the Long Tail of Science Gateways to Discovery: Cyberinfrastructure for the Long Tail of Science ECSS Symposium, 12/16/14 M. L. Norman, R. L. Moore, D. Baxter, G. Fox (Indiana U), A Majumdar, P Papadopoulos, W Pfeiffer, R. S.

More information

USING OPEN FABRIC INTERFACE IN INTEL MPI LIBRARY

USING OPEN FABRIC INTERFACE IN INTEL MPI LIBRARY 14th ANNUAL WORKSHOP 2018 USING OPEN FABRIC INTERFACE IN INTEL MPI LIBRARY Michael Chuvelev, Software Architect Intel April 11, 2018 INTEL MPI LIBRARY Optimized MPI application performance Application-specific

More information

High Performance Computing

High Performance Computing High Performance Computing Dror Goldenberg, HPCAC Switzerland Conference March 2015 End-to-End Interconnect Solutions for All Platforms Highest Performance and Scalability for X86, Power, GPU, ARM and

More information

Performance Analysis and Evaluation of Mellanox ConnectX InfiniBand Architecture with Multi-Core Platforms

Performance Analysis and Evaluation of Mellanox ConnectX InfiniBand Architecture with Multi-Core Platforms Performance Analysis and Evaluation of Mellanox ConnectX InfiniBand Architecture with Multi-Core Platforms Sayantan Sur, Matt Koop, Lei Chai Dhabaleswar K. Panda Network Based Computing Lab, The Ohio State

More information

Messaging Overview. Introduction. Gen-Z Messaging

Messaging Overview. Introduction. Gen-Z Messaging Page 1 of 6 Messaging Overview Introduction Gen-Z is a new data access technology that not only enhances memory and data storage solutions, but also provides a framework for both optimized and traditional

More information

InfiniBand and Mellanox UFM Fundamentals

InfiniBand and Mellanox UFM Fundamentals InfiniBand and Mellanox UFM Fundamentals Part Number: MTR-IB-UFM-OST-A Duration: 3 Days What's in it for me? Where do I start learning about InfiniBand? How can I gain the tools to manage this fabric?

More information

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments Torben Kling-Petersen, PhD Presenter s Name Principle Field Title andengineer Division HPC &Cloud LoB SunComputing Microsystems

More information

Key Measures of InfiniBand Performance in the Data Center. Driving Metrics for End User Benefits

Key Measures of InfiniBand Performance in the Data Center. Driving Metrics for End User Benefits Key Measures of InfiniBand Performance in the Data Center Driving Metrics for End User Benefits Benchmark Subgroup Benchmark Subgroup Charter The InfiniBand Benchmarking Subgroup has been chartered by

More information

Solutions for Scalable HPC

Solutions for Scalable HPC Solutions for Scalable HPC Scot Schultz, Director HPC/Technical Computing HPC Advisory Council Stanford Conference Feb 2014 Leading Supplier of End-to-End Interconnect Solutions Comprehensive End-to-End

More information

RoCE vs. iwarp Competitive Analysis

RoCE vs. iwarp Competitive Analysis WHITE PAPER February 217 RoCE vs. iwarp Competitive Analysis Executive Summary...1 RoCE s Advantages over iwarp...1 Performance and Benchmark Examples...3 Best Performance for Virtualization...5 Summary...6

More information

Optimizing LS-DYNA Productivity in Cluster Environments

Optimizing LS-DYNA Productivity in Cluster Environments 10 th International LS-DYNA Users Conference Computing Technology Optimizing LS-DYNA Productivity in Cluster Environments Gilad Shainer and Swati Kher Mellanox Technologies Abstract Increasing demand for

More information

AcuSolve Performance Benchmark and Profiling. October 2011

AcuSolve Performance Benchmark and Profiling. October 2011 AcuSolve Performance Benchmark and Profiling October 2011 Note The following research was performed under the HPC Advisory Council activities Participating vendors: Intel, Dell, Mellanox, Altair Compute

More information

Multifunction Networking Adapters

Multifunction Networking Adapters Ethernet s Extreme Makeover: Multifunction Networking Adapters Chuck Hudson Manager, ProLiant Networking Technology Hewlett-Packard 2004 Hewlett-Packard Development Company, L.P. The information contained

More information

Birds of a Feather Presentation

Birds of a Feather Presentation Mellanox InfiniBand QDR 4Gb/s The Fabric of Choice for High Performance Computing Gilad Shainer, shainer@mellanox.com June 28 Birds of a Feather Presentation InfiniBand Technology Leadership Industry Standard

More information

VPI / InfiniBand. Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability

VPI / InfiniBand. Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability VPI / InfiniBand Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability Mellanox enables the highest data center performance with its

More information

THE OPEN DATA CENTER FABRIC FOR THE CLOUD

THE OPEN DATA CENTER FABRIC FOR THE CLOUD Product overview THE OPEN DATA CENTER FABRIC FOR THE CLOUD The Open Data Center Fabric for the Cloud The Xsigo Data Center Fabric revolutionizes data center economics by creating an agile, highly efficient

More information

Improving Application Performance and Predictability using Multiple Virtual Lanes in Modern Multi-Core InfiniBand Clusters

Improving Application Performance and Predictability using Multiple Virtual Lanes in Modern Multi-Core InfiniBand Clusters Improving Application Performance and Predictability using Multiple Virtual Lanes in Modern Multi-Core InfiniBand Clusters Hari Subramoni, Ping Lai, Sayantan Sur and Dhabhaleswar. K. Panda Department of

More information

Unified Storage and FCoE

Unified Storage and FCoE Unified Storage and FCoE Mike McNamara, NetApp February 24, 2011 Santa Clara, CA USA February 2011 1 Agenda Market Dynamics Why Unified Ethernet Storage Customer Success Stories Santa Clara, CA USA February

More information

Scaling to Petaflop. Ola Torudbakken Distinguished Engineer. Sun Microsystems, Inc

Scaling to Petaflop. Ola Torudbakken Distinguished Engineer. Sun Microsystems, Inc Scaling to Petaflop Ola Torudbakken Distinguished Engineer Sun Microsystems, Inc HPC Market growth is strong CAGR increased from 9.2% (2006) to 15.5% (2007) Market in 2007 doubled from 2003 (Source: IDC

More information

VPI / InfiniBand. Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability

VPI / InfiniBand. Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability VPI / InfiniBand Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability Mellanox enables the highest data center performance with its

More information

2-Port 40 Gb InfiniBand Expansion Card (CFFh) for IBM BladeCenter IBM BladeCenter at-a-glance guide

2-Port 40 Gb InfiniBand Expansion Card (CFFh) for IBM BladeCenter IBM BladeCenter at-a-glance guide 2-Port 40 Gb InfiniBand Expansion Card (CFFh) for IBM BladeCenter IBM BladeCenter at-a-glance guide The 2-Port 40 Gb InfiniBand Expansion Card (CFFh) for IBM BladeCenter is a dual port InfiniBand Host

More information

The Future of Interconnect Technology

The Future of Interconnect Technology The Future of Interconnect Technology Michael Kagan, CTO HPC Advisory Council Stanford, 2014 Exponential Data Growth Best Interconnect Required 44X 0.8 Zetabyte 2009 35 Zetabyte 2020 2014 Mellanox Technologies

More information

InfiniBand-based HPC Clusters

InfiniBand-based HPC Clusters Boosting Scalability of InfiniBand-based HPC Clusters Asaf Wachtel, Senior Product Manager 2010 Voltaire Inc. InfiniBand-based HPC Clusters Scalability Challenges Cluster TCO Scalability Hardware costs

More information

Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability

Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability Mellanox InfiniBand Host Channel Adapters (HCA) enable the highest data center

More information

Windows OpenFabrics (WinOF) Update

Windows OpenFabrics (WinOF) Update Windows OpenFabrics (WinOF) Update Eric Lantz, Microsoft (elantz@microsoft.com) April 2008 Agenda OpenFabrics and Microsoft Current Events HPC Server 2008 Release NetworkDirect - RDMA for Windows 2 OpenFabrics

More information

Future Routing Schemes in Petascale clusters

Future Routing Schemes in Petascale clusters Future Routing Schemes in Petascale clusters Gilad Shainer, Mellanox, USA Ola Torudbakken, Sun Microsystems, Norway Richard Graham, Oak Ridge National Laboratory, USA Birds of a Feather Presentation Abstract

More information

iwarp Learnings and Best Practices

iwarp Learnings and Best Practices iwarp Learnings and Best Practices Author: Michael Fenn, Penn State Date: March 28, 2012 www.openfabrics.org 1 Introduction Last year, the Research Computing and Cyberinfrastructure group at Penn State

More information

iser as accelerator for Software Defined Storage Rahul Fiske, Subhojit Roy IBM (India)

iser as accelerator for Software Defined Storage Rahul Fiske, Subhojit Roy IBM (India) iser as accelerator for Software Defined Storage Rahul Fiske, Subhojit Roy IBM (India) Agenda Network storage virtualization Current state of Fiber Channel iscsi seeing significant adoption Emergence of

More information

Building a Phased Plan for End-to-End FCoE. Shaun Walsh VP Corporate Marketing Emulex Corporation

Building a Phased Plan for End-to-End FCoE. Shaun Walsh VP Corporate Marketing Emulex Corporation Building a Phased Plan for End-to-End FCoE Shaun Walsh VP Corporate Marketing Emulex Corporation 1 Server Virtualization - # Driver of ELX Value #1 Driver of 10Gb Deployment Core to Cloud Computing Top

More information

Comparing Ethernet and Soft RoCE for MPI Communication

Comparing Ethernet and Soft RoCE for MPI Communication IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 7-66, p- ISSN: 7-77Volume, Issue, Ver. I (Jul-Aug. ), PP 5-5 Gurkirat Kaur, Manoj Kumar, Manju Bala Department of Computer Science & Engineering,

More information

Cisco Unified Computing System Delivering on Cisco's Unified Computing Vision

Cisco Unified Computing System Delivering on Cisco's Unified Computing Vision Cisco Unified Computing System Delivering on Cisco's Unified Computing Vision At-A-Glance Unified Computing Realized Today, IT organizations assemble their data center environments from individual components.

More information

Designing High Performance Communication Middleware with Emerging Multi-core Architectures

Designing High Performance Communication Middleware with Emerging Multi-core Architectures Designing High Performance Communication Middleware with Emerging Multi-core Architectures Dhabaleswar K. (DK) Panda Department of Computer Science and Engg. The Ohio State University E-mail: panda@cse.ohio-state.edu

More information

OpenFabrics Alliance Interoperability Working Group (OFA-IWG)

OpenFabrics Alliance Interoperability Working Group (OFA-IWG) Cover Letter Johann George Qlogic Corporation 2071 Stierlin Court Mountain View, CA 94043 OpenFabrics Alliance Interoperability Working Group (OFA-IWG) March 2008 Interoperability Event Report UNH-IOL

More information

Management Scalability. Author: Todd Rimmer Date: April 2014

Management Scalability. Author: Todd Rimmer Date: April 2014 Management Scalability Author: Todd Rimmer Date: April 2014 Agenda Projected HPC Scalability Requirements Key Challenges Path Record IPoIB Mgmt Security Partitioning Multicast Notices SA interaction Call

More information

Cisco SFS 7000D InfiniBand Server Switch

Cisco SFS 7000D InfiniBand Server Switch Data Sheet The Cisco SFS 7000D InfiniBand Server Switch sets the standard for cost-effective, low-latency, 4X DDR and SDR InfiniBand switching for building high-performance clusters. High-performance computing

More information

EXPERIENCES WITH NVME OVER FABRICS

EXPERIENCES WITH NVME OVER FABRICS 13th ANNUAL WORKSHOP 2017 EXPERIENCES WITH NVME OVER FABRICS Parav Pandit, Oren Duer, Max Gurtovoy Mellanox Technologies [ 31 March, 2017 ] BACKGROUND: NVME TECHNOLOGY Optimized for flash and next-gen

More information

All Roads Lead to Convergence

All Roads Lead to Convergence All Roads Lead to Convergence Greg Scherer VP, Server and Storage Strategy gscherer@broadcom.com Broadcom Corporation 2 Agenda The Trend Toward Convergence over Ethernet Reasons for Storage and Networking

More information

NAMD Performance Benchmark and Profiling. November 2010

NAMD Performance Benchmark and Profiling. November 2010 NAMD Performance Benchmark and Profiling November 2010 Note The following research was performed under the HPC Advisory Council activities Participating vendors: HP, Mellanox Compute resource - HPC Advisory

More information

ABySS Performance Benchmark and Profiling. May 2010

ABySS Performance Benchmark and Profiling. May 2010 ABySS Performance Benchmark and Profiling May 2010 Note The following research was performed under the HPC Advisory Council activities Participating vendors: AMD, Dell, Mellanox Compute resource - HPC

More information

Server Networking e Virtual Data Center

Server Networking e Virtual Data Center Server Networking e Virtual Data Center Roma, 8 Febbraio 2006 Luciano Pomelli Consulting Systems Engineer lpomelli@cisco.com 1 Typical Compute Profile at a Fortune 500 Enterprise Compute Infrastructure

More information

MELLANOX EDR UPDATE & GPUDIRECT MELLANOX SR. SE 정연구

MELLANOX EDR UPDATE & GPUDIRECT MELLANOX SR. SE 정연구 MELLANOX EDR UPDATE & GPUDIRECT MELLANOX SR. SE 정연구 Leading Supplier of End-to-End Interconnect Solutions Analyze Enabling the Use of Data Store ICs Comprehensive End-to-End InfiniBand and Ethernet Portfolio

More information

SCALASCA parallel performance analyses of SPEC MPI2007 applications

SCALASCA parallel performance analyses of SPEC MPI2007 applications Mitglied der Helmholtz-Gemeinschaft SCALASCA parallel performance analyses of SPEC MPI2007 applications 2008-05-22 Zoltán Szebenyi Jülich Supercomputing Centre, Forschungszentrum Jülich Aachen Institute

More information

AN INTRODUCTION TO THE PATHSCALE INFINIPATH HTX ADAPTER. LLOYD DICKMAN Distinguished Architect, Office of the CTO PathScale, Inc.

AN INTRODUCTION TO THE PATHSCALE INFINIPATH HTX ADAPTER. LLOYD DICKMAN Distinguished Architect, Office of the CTO PathScale, Inc. AN INTRODUCTION TO THE PATHSCALE INFINIPATH HTX ADAPTER LLOYD DICKMAN Distinguished Architect, Office of the CTO PathScale, Inc. 1 Executive Summary Cluster systems based on commodity processors and the

More information

Myri-10G Myrinet Converges with Ethernet

Myri-10G Myrinet Converges with Ethernet Myri-10G Myrinet Converges with Ethernet David PeGan VP, Sales dave@myri.com (Substituting for Tom Leinberger) 4 October 2006 Oklahoma Supercomputing Symposium 1 New Directions for Myricom Although Myricom

More information

ARISTA: Improving Application Performance While Reducing Complexity

ARISTA: Improving Application Performance While Reducing Complexity ARISTA: Improving Application Performance While Reducing Complexity October 2008 1.0 Problem Statement #1... 1 1.1 Problem Statement #2... 1 1.2 Previous Options: More Servers and I/O Adapters... 1 1.3

More information

SAN Virtuosity Fibre Channel over Ethernet

SAN Virtuosity Fibre Channel over Ethernet SAN VIRTUOSITY Series WHITE PAPER SAN Virtuosity Fibre Channel over Ethernet Subscribe to the SAN Virtuosity Series at www.sanvirtuosity.com Table of Contents Introduction...1 VMware and the Next Generation

More information

Cray XD1 Supercomputer Release 1.3 CRAY XD1 DATASHEET

Cray XD1 Supercomputer Release 1.3 CRAY XD1 DATASHEET CRAY XD1 DATASHEET Cray XD1 Supercomputer Release 1.3 Purpose-built for HPC delivers exceptional application performance Affordable power designed for a broad range of HPC workloads and budgets Linux,

More information

MM5 Modeling System Performance Research and Profiling. March 2009

MM5 Modeling System Performance Research and Profiling. March 2009 MM5 Modeling System Performance Research and Profiling March 2009 Note The following research was performed under the HPC Advisory Council activities AMD, Dell, Mellanox HPC Advisory Council Cluster Center

More information

Dragon Slayer Consulting

Dragon Slayer Consulting Dragon Slayer Consulting Introduction to the Value Proposition of InfiniBand Marc Staimer marcstaimer@earthlink.net (503) 579-3763 5/27/2002 Dragon Slayer Consulting 1 Introduction to InfiniBand (IB) Agenda

More information

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance 11 th International LS-DYNA Users Conference Computing Technology LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton

More information

Accelerating Real-Time Big Data. Breaking the limitations of captive NVMe storage

Accelerating Real-Time Big Data. Breaking the limitations of captive NVMe storage Accelerating Real-Time Big Data Breaking the limitations of captive NVMe storage 18M IOPs in 2u Agenda Everything related to storage is changing! The 3rd Platform NVM Express architected for solid state

More information

MILC Performance Benchmark and Profiling. April 2013

MILC Performance Benchmark and Profiling. April 2013 MILC Performance Benchmark and Profiling April 2013 Note The following research was performed under the HPC Advisory Council activities Special thanks for: HP, Mellanox For more information on the supporting

More information

THE ZADARA CLOUD. An overview of the Zadara Storage Cloud and VPSA Storage Array technology WHITE PAPER

THE ZADARA CLOUD. An overview of the Zadara Storage Cloud and VPSA Storage Array technology WHITE PAPER WHITE PAPER THE ZADARA CLOUD An overview of the Zadara Storage Cloud and VPSA Storage Array technology Zadara 6 Venture, Suite 140, Irvine, CA 92618, USA www.zadarastorage.com EXECUTIVE SUMMARY The IT

More information

HYCOM Performance Benchmark and Profiling

HYCOM Performance Benchmark and Profiling HYCOM Performance Benchmark and Profiling Jan 2011 Acknowledgment: - The DoD High Performance Computing Modernization Program Note The following research was performed under the HPC Advisory Council activities

More information

Storage Protocol Offload for Virtualized Environments Session 301-F

Storage Protocol Offload for Virtualized Environments Session 301-F Storage Protocol Offload for Virtualized Environments Session 301-F Dennis Martin, President August 2016 1 Agenda About Demartek Offloads I/O Virtualization Concepts RDMA Concepts Overlay Networks and

More information

LS-DYNA Productivity and Power-aware Simulations in Cluster Environments

LS-DYNA Productivity and Power-aware Simulations in Cluster Environments LS-DYNA Productivity and Power-aware Simulations in Cluster Environments Gilad Shainer 1, Tong Liu 1, Jacob Liberman 2, Jeff Layton 2 Onur Celebioglu 2, Scot A. Schultz 3, Joshua Mora 3, David Cownie 3,

More information

Highest Levels of Scalability Simplified Network Manageability Maximum System Productivity

Highest Levels of Scalability Simplified Network Manageability Maximum System Productivity InfiniBand Brochure Highest Levels of Scalability Simplified Network Manageability Maximum System Productivity 40/56/100/200Gb/s InfiniBand Switch System Family MELLANOX SMART INFINIBAND SWITCH SYSTEMS

More information

Communication has significant impact on application performance. Interconnection networks therefore have a vital role in cluster systems.

Communication has significant impact on application performance. Interconnection networks therefore have a vital role in cluster systems. Cluster Networks Introduction Communication has significant impact on application performance. Interconnection networks therefore have a vital role in cluster systems. As usual, the driver is performance

More information

UCX: An Open Source Framework for HPC Network APIs and Beyond

UCX: An Open Source Framework for HPC Network APIs and Beyond UCX: An Open Source Framework for HPC Network APIs and Beyond Presented by: Pavel Shamis / Pasha ORNL is managed by UT-Battelle for the US Department of Energy Co-Design Collaboration The Next Generation

More information

DELL EMC READY BUNDLE FOR VIRTUALIZATION WITH VMWARE AND FIBRE CHANNEL INFRASTRUCTURE

DELL EMC READY BUNDLE FOR VIRTUALIZATION WITH VMWARE AND FIBRE CHANNEL INFRASTRUCTURE DELL EMC READY BUNDLE FOR VIRTUALIZATION WITH VMWARE AND FIBRE CHANNEL INFRASTRUCTURE Design Guide APRIL 0 The information in this publication is provided as is. Dell Inc. makes no representations or warranties

More information

OpenFabrics Alliance Interoperability Working Group (OFA-IWG)

OpenFabrics Alliance Interoperability Working Group (OFA-IWG) OpenFabrics Alliance Interoperability Working Group (OFA-IWG) April 2007 Interoperability Event Report Cover Letter Oded Bergman Voltaire 9 Hamenofim St. Hertzeliya, Israel 46725 UNH-IOL 121 Technology

More information

Building supercomputers from embedded technologies

Building supercomputers from embedded technologies http://www.montblanc-project.eu Building supercomputers from embedded technologies Alex Ramirez Barcelona Supercomputing Center Technical Coordinator This project and the research leading to these results

More information

HP Cluster Interconnects: The Next 5 Years

HP Cluster Interconnects: The Next 5 Years HP Cluster Interconnects: The Next 5 Years Michael Krause mkrause@hp.com September 8, 2003 2003 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice

More information

V.I.B.E. Virtual. Integrated. Blade. Environment. Harveenpal Singh. System-x PLM

V.I.B.E. Virtual. Integrated. Blade. Environment. Harveenpal Singh. System-x PLM V.I.B.E. Virtual. Integrated. Blade. Environment. Harveenpal Singh System-x PLM x86 servers are taking on more demanding roles, including high-end business critical applications x86 server segment is the

More information