Mahidhar Tatineni, Director of User Services, SDSC. HPC Advisory Council China Conference October 18, 2017, Hefei, China
|
|
- Shon Stokes
- 5 years ago
- Views:
Transcription
1 Experiences with HPC and Big Data Applications on the SDSC Comet Cluster: Using Virtualization, Singularity containers, and RDMA enabled Data Analytics tools Mahidhar Tatineni, Director of User Services, SDSC HPC Advisory Council China Conference October 18, 2017, Hefei, China Acknowledgements: Trevor Cooper, Dmitry Mishin, Christopher Irving, Gregor von Laszewski (IU) Fugang Wang (IU), Rick Wagner (Globus group, U. Chicago), Phil Papadopoulos
2 This work supported by the National Science Foundation, award ACI
3 Overview Comet Hardware Compute,GPU nodes, network, flesystems MPI implementations, including MVAPICH2-GDR results Data analytics frameworks and tools on Comet RDMA-Hadoop, RDMASpark, OSU-Caffe Virtual Cluster (VC) Design layout, software VC Benchmarks: MPI, Applications NSF Award# , Gateways to Discovery: Cyberinfrastructure for the Long Tail of Science PI: Michael Norman Co-PIs: Shawn Strande, Philip Papadopoulos, Robert Sinkovits, Nancy Wilkins-Diehr SDSC Project in Collaboration with Indiana University (led by Geoffrey Fox)
4 Comet: System Characteristics Total peak fops ~2.1 PF Dell primary integrator Intel Haswell processors w/ AVX2 Mellanox FDR InfniBand 1,944 standard compute nodes (46,656 cores) Dual CPUs, each 12-core, 2.5 GHz 128 GB DDR MHz DRAM 2*160GB GB SSDs (local disk) 72 GPU nodes 36 nodes with two NVIDIA K80 cards, each with dual Kepler3 GPUs, same CPU as main partition. 36 nodes with 4 P100 GPUs each, 2 Intel Broadwell processors (14 cores each) 4 large-memory nodes 1.5 TB DDR MHz DRAM Four Haswell processors/node 64 cores/node Hybrid fat-tree topology FDR (56 Gbps) InfniBand Rack-level (72 nodes, 1,728 cores) full bisection bandwidth 4:1 oversubscription cross-rack Performance Storage (Aeon) 7.6 PB, 200 GB/s; Lustre Scratch & Persistent Storage segments Durable Storage (Aeon) 6 PB, 100 GB/s; Lustre Automatic backups of critical data Home directory storage Gateway hosting nodes Virtual image repository 100 Gbps external connectivity to Internet2 & ESNet
5 Comet Network Architecture
6 Comet Lustre Filesystems Comet features two Lustre filesystems - scratch and projects storage. Projects storage is mounted on multiple systems, including non IB connected clusters. Backend storage servers are connected via 40 Gbit ethernet fabric. Comet network design handles this aspect by using bridge switches (Mellanox). Design provides flexibility to mount filesystem on multiple machine and keeps the aggregate performance high. Each filesystem (scratch and projects) achieved 100 GB/s on and IOR bandwidth test at scale.
7 Comet: MPI options, RDMA enabled software MVAPICH2 (v2.1) is the default MPI on Comet. Intel MPI and OpenMPI also available. MVAPICH2-X v2.2a to provide unified high-performance runtime supporting both MPI and PGAS programming models. MVAPICH2-GDR (v2.2) on the GPU nodes featuring NVIDIA K80s and P100s. Benchmark and application performance results in this talk. RDMA-Hadoop (2x-1.1.0), RDMA-Spark (0.9.4) (from Dr. Panda s HiBD lab) also available.
8 Comet K80 node architecture 4 GPUs per node GPUs (0,1) and (2,3) can do P2P communication Mellanox InfniBand adapter associated with second socket (GPUs 2, 3)
9 OSU Latency (osu_latency) Benchmark Intra-node, K80 nodes Latency between GPU 2, GPU 3: 2.82 µs Latency between GPU 1, GPU 2: 3.18 µs
10 OSU Latency (osu_latency) Benchmark Inter-node, K80 nodes Latency between GPU 2, process bound to CPU 1 on both nodes: 2.27 µs Latency between GPU 2, process bound to CPU 0 on both nodes: 2.47 µs Latency between GPU 0, process bound to CPU 0 on both nodes: 2.43 µs
11 Comet P100 node architecture 4 GPUs per node GPUs (0,1) and (2,3) can do P2P communication Mellanox InfniBand adapter associated with frst socket (GPUs 0, 1)
12 OSU Latency (osu_latency) Benchmark Intra-node, P100 nodes Latency between GPU 0, GPU 1: 2.73 µs Latency between GPU 2, GPU 3: 2.95 µs Latency between GPU 1, GPU 2: 3.13 µs
13 OSU Latency (osu_latency) Benchmark Inter-node, P100 nodes Latency between GPU 0, process bound to CPU 0 on both nodes: 2.17 µs Latency between GPU 2, process bound to CPU 1 on both nodes: 2.35 µs
14 MVAPICH2-GDR Application Example:HOOMD-blue HOOMD-blue is a general-purpose particle simulation toolkit Results for the Hexagon benchmark are presented.. References: HOOMD-blue web page: HOOMD-blue Benchmarks page: glotzerlab.engin.umich.edu/hoomd-blue/benchmarks.html J. A. Anderson, C. D. Lorenz, and A. Travesset. General purpose molecular dynamics simulations fully implemented on graphics processing units Journal of Computational Physics 227(10): , May /j.jcp J. Glaser, T. D. Nguyen, J. A. Anderson, P. Liu, F. Spiga, J. A. Millan, D. C. Morse, S. C. Glotzer. Strong scaling of general-purpose molecular dynamics simulations on GPUs Computer Physics Communications 192: , July /j.cpc
15 HOOMD-Blue: Hexagon Benchmark N=1,048,576N=1,048,576 Hard particle Monte Carlo Vertices: [[0.5,0],[0.25, ],[0.25, ],[-0.5,0],[-0.25, ],[0.25, ]] d= a= nselect=4 Log fle period: time steps SDF analysis xmax==0.02 δx=10 4 period: 50 time steps navg=2000 DCD dump period:
16 HOOMD-Blue: Hexagon Benchmark Strong scaling on K80 nodes
17 RDMA-Hadoop and RDMA-Spark Network-Based Computing Lab, Ohio State University NSF funded project in collaboration with Dr. DK Panda* HDFS, MapReduce, and RPC over native InfiniBand and RDMA over Converged Ethernet (RoCE). Based on Apache distributions of Hadoop and Spark. Version RDMA-Apache-Hadoop-2.x (based on Apache Hadoop 2.6.0) available on Comet Version RDMA-Spark (based on Apache Spark 2.1.0) is available on Comet. More details on the RDMA-Hadoop and RDMA-Spark projects at: *NSF BIGDATA F: DKM: Collaborative Research: Scalable Middleware for Managing and Processing Big Data on Next Generation HPC Systems, Award #s (Ohio State), and (SDSC).
18 RDMA-Hadoop, Spark Exploit performance on modern clusters with RDMA-enabled interconnects for Big Data applications. Hybrid design with in-memory and heterogeneous storage (HDD, SSDs, Lustre). Keep compliance with standard distributions from Apache.
19
20 OSU-Caffe, CIFAR10 Quick on K80 nodes Results with K80 nodes. Current runs with data in Lustre filesystem (/oasis/scratch/comet) All Comet GPU nodes have 280GB of SSD based local scratch space. Future tests with larger test cases planned to evaluate performance advantages of using the SSDs.
21 OSU-Caffe, CIFAR10 Quick on K80 nodes
22 Virtualization on Comet Comet Virtual Clusters KVM based, SRIOV enabled full virtualization. Singularity based containerization user space only with namespaces and minimal SetUID.
23 Comet VC Use Cases Root access to nodes for custom OS and software stack. Example : CAIDA group used it for a workshop allowing attendees to modify network stack for research. Allows for isolation of tests. Simplified install for groups with existing management infrastructure. Example: Open Science Group (OSG) used their existing installation procedures to enable multiple research groups to run on Comet (including LIGO).
24 Singularity Use Cases Applications with newer library OS requirements than available on Comet e.g. Tensorflow, Torch, Caffe. Commercial application binaries with specific OS requirements. Importing docker images to enable use in a shared HPC environment.
25 Overview of Virtual Clusters on Comet Projects have persistent VM for cluster management Modest: single core, 1-2 GB of RAM Standard compute nodes will be scheduled as containers via batch system One virtual compute node per container Virtual disk images stored as ZFS datasets Migrated to and from containers at job start and end VM use allocated and tracked like regular computing
26 User Perspective Active virtual compute nodes Scheduling Storage management Coordinating network changes VM launch & shutdown Attached and synchronized Nucleus Disk images API Request nodes Console & power Persistent virtual front end Idle disk images
27 Enabling Technologies KVM Lets us run virtual machines (all processor features) SR-IOV Makes MPI go fast on VMs Rocks Systems management ZFS Disk image management VLANs Isolate virtual cluster management network pkeys Isolate virtual cluster IB network Nucleus Coordination engine (scheduling, provisioning, status, etc.) Client Cloudmesh
28 User-Customized HPC public network Virtual Frontend Hosting Frontend Virtual Frontend private physical virtual virtual Disk Image Vault Virtual Frontend private private Compute Compute Compute Compute Compute Virtual Compute Virtual Compute Virtual Compute Compute Compute Compute Compute Virtual Compute Virtual Compute Virtual Compute
29 High Performance Virtual Cluster Characteristics Comet: Providing Virtualized HPC for XSEDE Infniband Virtual Frontend private Ethernet Infniband Virtualization 8% latency overhead. Nominal bandwidth overhead Virtual Compute All nodes have Private Ethernet Infniband Local Disk Storage Virtual Compute Virtual Compute Nodes can Network boot (PXE) from its virtual frontend Virtual Compute All Disks retain state keep user confguration between boots
30 Data Storage/Filesystems Local SSD storage on each compute node Limited number of large SSD nodes (1.4TB) for large VM images Local (SDSC) network access same as compute nodes Modest (TB) storage available via NFS now Future: Secure access to Lustre?
31 Cloudmesh Developed by IU collaborators Cloudmesh client enables access to multiple cloud environments from a command shell and command line. We leverage this easy to use CLI for virtual cluster management, allowing the use of Comet as infrastructure for virtual clusters. Cloudmesh has more functionality with ability to access hybrid clouds OpenStack, (EC2, AWS, Azure); possible to extend to other systems like Jetstream, Bridges etc. Plans for customizable launchers available through command line or browser can target specifc application user communities. Reference:
32 Comet Cloudmesh Client (selected commands) cm comet cluster Attach an image cm comet boot ID vm-id-0 Power 3 nodes on for 6 hours cm comet image attach image.iso ID vm-id-0 Show the cluster details cm comet power on ID vm-id -[0-3] --walltime=6h ID Boot node 0 cm comet console vc4 Console
33 Comet Cloudmesh Usage Examples
34 Comet Cloudmesh Client : Console access (COMET)host:client$ cm comet console vc4 vm-vc4-0
35 MPI bandwidth slowdown from SR-IOV is at most 1.21 for mediumsized messages & negligible for small & large ones
36 MPI latency slowdown from SR-IOV is at most 1.32 for small messages & negligible for large ones
37 WRF Weather Modeling 96-core (4-node) calculation Nearest-neighbor communication Test Case: 3hr Forecast, 2.5km resolution of Continental US (CONUS). Scalable algorithms 2% slower w/ SR-IOV vs native IB.
38 PSDNS, 1024x1024x1024 : Strong scaling case 32 (2-node), 64 (4-node), and 128 (8node) core tests. Computational core based on FFTs. Communication intensive, mainly alltoallv bisection bandwidth limited. Cores (Nodes) Time/Step 32 (2) (4) (8) 33.99
39 Quantum ESPRESSO 48-core (3 node) calculation CG matrix inversion - irregular communication 3D FFT matrix transposes (allto-all communication) Test Case: DEISA AUSURF 112 benchmark. 8% slower w/ SR-IOV vs native IB.
40 RAxML: Code for Maximum Likelihood-based inference of large phylogenetic trees. Widely used, including by CIPRES gateway. 48-core (2 node) calculation Hybrid MPI/Pthreads Code. 12 MPI tasks, 4 threads per task. Compilers: gcc + mvapich2 v2.2, AVX options. Test Case: Comprehensive analysis, 218 taxa, 2,294 characters, 1,846 patterns, 100 bootstraps specifed. 19% slower w/ SR-IOV vs native IB.
41 MrBayes: Software for Bayesian inference of phylogeny. Widely used, including by CIPRES gateway. 32-core (2 node) calculation Hybrid MPI/OpenMP Code. 8 MPI tasks, 4 OpenMP threads per task. Compilers: gcc + mvapich2 v2.2, AVX options. Test Case: 218 taxa, 10,000 generations. 3% slower with SR-IOV vs native IB.
42 Summary Comet uses the flexibility and performance of IB network + Bridging to provide multi-machine mounted parallel filesystems enhance GPU applications using GPU Direct RDMA enhance performance of data analytics tools with RDMA enabled frameworks enable virtualized HPC clusters at scale. OSU Benchmarks and HOOMD-Blue applications show good performance using MVAPICH2-GDR. Results for OSU-Caffe with CIFAR10 benchmark show good scaling. Future tests with larger test cases planned to evaluate performance advantages of using the SSDs. Application benchmarks show good performance on virtualized cluster. PSDNS example, which is communication intensive, shows good strong scaling for a test example.
Virtualization for High Performance Computing Applications at SDSC Mahidhar Tatineni, Director of User Services, SDSC HPC Advisory Council China
Virtualization for High Performance Computing Applications at SDSC Mahidhar Tatineni, Director of User Services, SDSC HPC Advisory Council China Conference October 26, 2016, Xi an, China Comet Virtualization
More informationPerformance of Applications on Comet GPU Nodes Utilizing MVAPICH2-GDR. Mahidhar Tatineni MVAPICH User Group Meeting August 16, 2017
Performance of Applications on Comet GPU Nodes Utilizing MVAPICH2-GDR Mahidhar Tatineni MVAPICH User Group Meeting August 16, 2017 This work supported by the National Science Foundation, award ACI-1341698.
More informationGateways to Discovery: Cyberinfrastructure for the Long Tail of Science
Gateways to Discovery: Cyberinfrastructure for the Long Tail of Science ECSS Symposium, 12/16/14 M. L. Norman, R. L. Moore, D. Baxter, G. Fox (Indiana U), A Majumdar, P Papadopoulos, W Pfeiffer, R. S.
More informationComet Virtualization Code & Design Sprint
Comet Virtualization Code & Design Sprint SDSC September 23-24 Rick Wagner San Diego Supercomputer Center Meeting Goals Build personal connections between the IU and SDSC members of the Comet team working
More informationComet Virtual Clusters What s underneath? Philip Papadopoulos San Diego Supercomputer Center
Comet Virtual Clusters What s underneath? Philip Papadopoulos San Diego Supercomputer Center ppapadopoulos@ucsd.edu Overview NSF Award# 1341698, Gateways to Discovery: Cyberinfrastructure for the Long
More informationBig Data Analytics with the OSU HiBD Stack at SDSC. Mahidhar Tatineni OSU Booth Talk, SC18, Dallas
Big Data Analytics with the OSU HiBD Stack at SDSC Mahidhar Tatineni OSU Booth Talk, SC18, Dallas Comet HPC for the long tail of science iphone panorama photograph of 1 of 2 server rows Comet: System Characteristics
More informationHigh-Performance Training for Deep Learning and Computer Vision HPC
High-Performance Training for Deep Learning and Computer Vision HPC Panel at CVPR-ECV 18 by Dhabaleswar K. (DK) Panda The Ohio State University E-mail: panda@cse.ohio-state.edu http://www.cse.ohio-state.edu/~panda
More informationSR-IOV Support for Virtualization on InfiniBand Clusters: Early Experience
SR-IOV Support for Virtualization on InfiniBand Clusters: Early Experience Jithin Jose, Mingzhe Li, Xiaoyi Lu, Krishna Kandalla, Mark Arnold and Dhabaleswar K. (DK) Panda Network-Based Computing Laboratory
More informationBig Data Meets HPC: Exploiting HPC Technologies for Accelerating Big Data Processing and Management
Big Data Meets HPC: Exploiting HPC Technologies for Accelerating Big Data Processing and Management SigHPC BigData BoF (SC 17) by Dhabaleswar K. (DK) Panda The Ohio State University E-mail: panda@cse.ohio-state.edu
More informationA Virtual Comet. HTCondor Week 2017 May Edgar Fajardo On behalf of OSG Software and Technology
A Virtual Comet HTCondor Week 2017 May 3 2017 Edgar Fajardo On behalf of OSG Software and Technology 1 Working in Comet What my friends think I do What Instagram thinks I do What my boss thinks I do 2
More informationSDSC s Data Oasis Gen II: ZFS, 40GbE, and Replication
SDSC s Data Oasis Gen II: ZFS, 40GbE, and Replication Rick Wagner HPC Systems Manager San Diego Supercomputer Center Comet HPC for the long tail of science iphone panorama photograph of 1 of 2 server rows
More informationThe Impact of Inter-node Latency versus Intra-node Latency on HPC Applications The 23 rd IASTED International Conference on PDCS 2011
The Impact of Inter-node Latency versus Intra-node Latency on HPC Applications The 23 rd IASTED International Conference on PDCS 2011 HPC Scale Working Group, Dec 2011 Gilad Shainer, Pak Lui, Tong Liu,
More informationA Plugin-based Approach to Exploit RDMA Benefits for Apache and Enterprise HDFS
A Plugin-based Approach to Exploit RDMA Benefits for Apache and Enterprise HDFS Adithya Bhat, Nusrat Islam, Xiaoyi Lu, Md. Wasi- ur- Rahman, Dip: Shankar, and Dhabaleswar K. (DK) Panda Network- Based Compu2ng
More informationSharing High-Performance Devices Across Multiple Virtual Machines
Sharing High-Performance Devices Across Multiple Virtual Machines Preamble What does sharing devices across multiple virtual machines in our title mean? How is it different from virtual networking / NSX,
More informationMELLANOX EDR UPDATE & GPUDIRECT MELLANOX SR. SE 정연구
MELLANOX EDR UPDATE & GPUDIRECT MELLANOX SR. SE 정연구 Leading Supplier of End-to-End Interconnect Solutions Analyze Enabling the Use of Data Store ICs Comprehensive End-to-End InfiniBand and Ethernet Portfolio
More informationCan Parallel Replication Benefit Hadoop Distributed File System for High Performance Interconnects?
Can Parallel Replication Benefit Hadoop Distributed File System for High Performance Interconnects? N. S. Islam, X. Lu, M. W. Rahman, and D. K. Panda Network- Based Compu2ng Laboratory Department of Computer
More informationBridging Neuroscience and HPC with MPI-LiFE Shashank Gugnani
Bridging Neuroscience and HPC with MPI-LiFE Shashank Gugnani The Ohio State University E-mail: gugnani.2@osu.edu http://web.cse.ohio-state.edu/~gugnani/ Network Based Computing Laboratory SC 17 2 Neuroscience:
More informationABySS Performance Benchmark and Profiling. May 2010
ABySS Performance Benchmark and Profiling May 2010 Note The following research was performed under the HPC Advisory Council activities Participating vendors: AMD, Dell, Mellanox Compute resource - HPC
More informationNAMD Performance Benchmark and Profiling. January 2015
NAMD Performance Benchmark and Profiling January 2015 2 Note The following research was performed under the HPC Advisory Council activities Participating vendors: Intel, Dell, Mellanox Compute resource
More informationSolutions for Scalable HPC
Solutions for Scalable HPC Scot Schultz, Director HPC/Technical Computing HPC Advisory Council Stanford Conference Feb 2014 Leading Supplier of End-to-End Interconnect Solutions Comprehensive End-to-End
More informationCharacterizing and Benchmarking Deep Learning Systems on Modern Data Center Architectures
Characterizing and Benchmarking Deep Learning Systems on Modern Data Center Architectures Talk at Bench 2018 by Xiaoyi Lu The Ohio State University E-mail: luxi@cse.ohio-state.edu http://www.cse.ohio-state.edu/~luxi
More informationAcuSolve Performance Benchmark and Profiling. October 2011
AcuSolve Performance Benchmark and Profiling October 2011 Note The following research was performed under the HPC Advisory Council activities Participating vendors: Intel, Dell, Mellanox, Altair Compute
More informationHigh Performance Computing and Data Resources at SDSC
High Performance Computing and Data Resources at SDSC "! Mahidhar Tatineni (mahidhar@sdsc.edu)! SDSC Summer Institute! August 05, 2013! HPC Resources at SDSC Hardware Overview HPC Systems : Gordon, Trestles
More informationEmerging Technologies for HPC Storage
Emerging Technologies for HPC Storage Dr. Wolfgang Mertz CTO EMEA Unstructured Data Solutions June 2018 The very definition of HPC is expanding Blazing Fast Speed Accessibility and flexibility 2 Traditional
More informationDesigning High-Performance MPI Collectives in MVAPICH2 for HPC and Deep Learning
5th ANNUAL WORKSHOP 209 Designing High-Performance MPI Collectives in MVAPICH2 for HPC and Deep Learning Hari Subramoni Dhabaleswar K. (DK) Panda The Ohio State University The Ohio State University E-mail:
More informationScalability Testing of DNE2 in Lustre 2.7 and Metadata Performance using Virtual Machines Tom Crowe, Nathan Lavender, Stephen Simms
Scalability Testing of DNE2 in Lustre 2.7 and Metadata Performance using Virtual Machines Tom Crowe, Nathan Lavender, Stephen Simms Research Technologies High Performance File Systems hpfs-admin@iu.edu
More informationDesigning and Building Efficient HPC Cloud with Modern Networking Technologies on Heterogeneous HPC Clusters
Designing and Building Efficient HPC Cloud with Modern Networking Technologies on Heterogeneous HPC Clusters Jie Zhang Dr. Dhabaleswar K. Panda (Advisor) Department of Computer Science & Engineering The
More informationSuperMike-II Launch Workshop. System Overview and Allocations
: System Overview and Allocations Dr Jim Lupo CCT Computational Enablement jalupo@cct.lsu.edu SuperMike-II: Serious Heterogeneous Computing Power System Hardware SuperMike provides 442 nodes, 221TB of
More informationA Container On a Virtual Machine On an HPC? Presentation to HPC Advisory Council. Perth, July 31-Aug 01, 2017
A Container On a Virtual Machine On an HPC? Presentation to HPC Advisory Council Perth, July 31-Aug 01, 2017 http://levlafayette.com Necessary and Sufficient Definitions High Performance Computing: High
More informationOncilla - a Managed GAS Runtime for Accelerating Data Warehousing Queries
Oncilla - a Managed GAS Runtime for Accelerating Data Warehousing Queries Jeffrey Young, Alex Merritt, Se Hoon Shon Advisor: Sudhakar Yalamanchili 4/16/13 Sponsors: Intel, NVIDIA, NSF 2 The Problem Big
More informationNowlab, OSU Booth Talk SC18, Dallas
Overview - Last 10 Years of HPC Architectures and Science Enabled at SDSC Amit Majumdar Division Director, Data Enabled Scientific Computing Division San Diego Supercomputer Center University of California
More informationExploiting Full Potential of GPU Clusters with InfiniBand using MVAPICH2-GDR
Exploiting Full Potential of GPU Clusters with InfiniBand using MVAPICH2-GDR Presentation at Mellanox Theater () Dhabaleswar K. (DK) Panda - The Ohio State University panda@cse.ohio-state.edu Outline Communication
More informationHPC Capabilities at Research Intensive Universities
HPC Capabilities at Research Intensive Universities Purushotham (Puri) V. Bangalore Department of Computer and Information Sciences and UAB IT Research Computing UAB HPC Resources 24 nodes (192 cores)
More informationHigh Performance File System and I/O Middleware Design for Big Data on HPC Clusters
High Performance File System and I/O Middleware Design for Big Data on HPC Clusters by Nusrat Sharmin Islam Advisor: Dhabaleswar K. (DK) Panda Department of Computer Science and Engineering The Ohio State
More informationIn the multi-core age, How do larger, faster and cheaper and more responsive memory sub-systems affect data management? Dhabaleswar K.
In the multi-core age, How do larger, faster and cheaper and more responsive sub-systems affect data management? Panel at ADMS 211 Dhabaleswar K. (DK) Panda Network-Based Computing Laboratory Department
More informationInterconnect Your Future
Interconnect Your Future Gilad Shainer 2nd Annual MVAPICH User Group (MUG) Meeting, August 2014 Complete High-Performance Scalable Interconnect Infrastructure Comprehensive End-to-End Software Accelerators
More informationDesigning High-Performance Non-Volatile Memory-aware RDMA Communication Protocols for Big Data Processing
Designing High-Performance Non-Volatile Memory-aware RDMA Communication Protocols for Big Data Processing Talk at Storage Developer Conference SNIA 2018 by Xiaoyi Lu The Ohio State University E-mail: luxi@cse.ohio-state.edu
More informationThe Why and How of HPC-Cloud Hybrids with OpenStack
The Why and How of HPC-Cloud Hybrids with OpenStack OpenStack Australia Day Melbourne June, 2017 Lev Lafayette, HPC Support and Training Officer, University of Melbourne lev.lafayette@unimelb.edu.au 1.0
More informationPART-I (B) (TECHNICAL SPECIFICATIONS & COMPLIANCE SHEET) Supply and installation of High Performance Computing System
INSTITUTE FOR PLASMA RESEARCH (An Autonomous Institute of Department of Atomic Energy, Government of India) Near Indira Bridge; Bhat; Gandhinagar-382428; India PART-I (B) (TECHNICAL SPECIFICATIONS & COMPLIANCE
More informationEnabling Efficient Use of UPC and OpenSHMEM PGAS models on GPU Clusters
Enabling Efficient Use of UPC and OpenSHMEM PGAS models on GPU Clusters Presentation at GTC 2014 by Dhabaleswar K. (DK) Panda The Ohio State University E-mail: panda@cse.ohio-state.edu http://www.cse.ohio-state.edu/~panda
More informationDesigning Optimized MPI Broadcast and Allreduce for Many Integrated Core (MIC) InfiniBand Clusters
Designing Optimized MPI Broadcast and Allreduce for Many Integrated Core (MIC) InfiniBand Clusters K. Kandalla, A. Venkatesh, K. Hamidouche, S. Potluri, D. Bureddy and D. K. Panda Presented by Dr. Xiaoyi
More informationDeep Learning Frameworks with Spark and GPUs
Deep Learning Frameworks with Spark and GPUs Abstract Spark is a powerful, scalable, real-time data analytics engine that is fast becoming the de facto hub for data science and big data. However, in parallel,
More informationThe Stampede is Coming Welcome to Stampede Introductory Training. Dan Stanzione Texas Advanced Computing Center
The Stampede is Coming Welcome to Stampede Introductory Training Dan Stanzione Texas Advanced Computing Center dan@tacc.utexas.edu Thanks for Coming! Stampede is an exciting new system of incredible power.
More informationIBM CORAL HPC System Solution
IBM CORAL HPC System Solution HPC and HPDA towards Cognitive, AI and Deep Learning Deep Learning AI / Deep Learning Strategy for Power Power AI Platform High Performance Data Analytics Big Data Strategy
More informationAltair OptiStruct 13.0 Performance Benchmark and Profiling. May 2015
Altair OptiStruct 13.0 Performance Benchmark and Profiling May 2015 Note The following research was performed under the HPC Advisory Council activities Participating vendors: Intel, Dell, Mellanox Compute
More informationHPC Architectures. Types of resource currently in use
HPC Architectures Types of resource currently in use Reusing this material This work is licensed under a Creative Commons Attribution- NonCommercial-ShareAlike 4.0 International License. http://creativecommons.org/licenses/by-nc-sa/4.0/deed.en_us
More informationLAMMPS and WRF on iwarp vs. InfiniBand FDR
LAMMPS and WRF on iwarp vs. InfiniBand FDR The use of InfiniBand as interconnect technology for HPC applications has been increasing over the past few years, replacing the aging Gigabit Ethernet as the
More informationHigh Performance Migration Framework for MPI Applications on HPC Cloud
High Performance Migration Framework for MPI Applications on HPC Cloud Jie Zhang, Xiaoyi Lu and Dhabaleswar K. Panda {zhanjie, luxi, panda}@cse.ohio-state.edu Computer Science & Engineering Department,
More informationChelsio Communications. Meeting Today s Datacenter Challenges. Produced by Tabor Custom Publishing in conjunction with: CUSTOM PUBLISHING
Meeting Today s Datacenter Challenges Produced by Tabor Custom Publishing in conjunction with: 1 Introduction In this era of Big Data, today s HPC systems are faced with unprecedented growth in the complexity
More informationA Case for High Performance Computing with Virtual Machines
A Case for High Performance Computing with Virtual Machines Wei Huang*, Jiuxing Liu +, Bulent Abali +, and Dhabaleswar K. Panda* *The Ohio State University +IBM T. J. Waston Research Center Presentation
More informationJetstream: Adding Cloud-based Computing to the National Cyberinfrastructure
Jetstream: Adding Cloud-based Computing to the National Cyberinfrastructure funded by the National Science Foundation Award #ACI-1445604 Matthew Vaughn(@mattdotvaughn) ORCID 0000-0002-1384-4283 Director,
More informationOPEN MPI WITH RDMA SUPPORT AND CUDA. Rolf vandevaart, NVIDIA
OPEN MPI WITH RDMA SUPPORT AND CUDA Rolf vandevaart, NVIDIA OVERVIEW What is CUDA-aware History of CUDA-aware support in Open MPI GPU Direct RDMA support Tuning parameters Application example Future work
More informationHigh-Performance MPI Library with SR-IOV and SLURM for Virtualized InfiniBand Clusters
High-Performance MPI Library with SR-IOV and SLURM for Virtualized InfiniBand Clusters Talk at OpenFabrics Workshop (April 2016) by Dhabaleswar K. (DK) Panda The Ohio State University E-mail: panda@cse.ohio-state.edu
More informationAccelerating MPI Message Matching and Reduction Collectives For Multi-/Many-core Architectures Mohammadreza Bayatpour, Hari Subramoni, D. K.
Accelerating MPI Message Matching and Reduction Collectives For Multi-/Many-core Architectures Mohammadreza Bayatpour, Hari Subramoni, D. K. Panda Department of Computer Science and Engineering The Ohio
More informationApplication Performance on IME
Application Performance on IME Toine Beckers, DDN Marco Grossi, ICHEC Burst Buffer Designs Introduce fast buffer layer Layer between memory and persistent storage Pre-stage application data Buffer writes
More informationData Processing at the Speed of 100 Gbps using Apache Crail. Patrick Stuedi IBM Research
Data Processing at the Speed of 100 Gbps using Apache Crail Patrick Stuedi IBM Research The CRAIL Project: Overview Data Processing Framework (e.g., Spark, TensorFlow, λ Compute) Spark-IO Albis Pocket
More informationExploiting InfiniBand and GPUDirect Technology for High Performance Collectives on GPU Clusters
Exploiting InfiniBand and Direct Technology for High Performance Collectives on Clusters Ching-Hsiang Chu chu.368@osu.edu Department of Computer Science and Engineering The Ohio State University OSU Booth
More informationLAMMPS-KOKKOS Performance Benchmark and Profiling. September 2015
LAMMPS-KOKKOS Performance Benchmark and Profiling September 2015 2 Note The following research was performed under the HPC Advisory Council activities Participating vendors: Intel, Dell, Mellanox, NVIDIA
More informationJetstream Overview A national research and education cloud
Jetstream Overview A national research and education cloud 9th workshop on Scientific Cloud Computing (ScienceCloud) June 11, 2018 Tempe, AZ John Michael Lowe jomlowe@iu.edu Senior Cloud Engineer, UITS
More informationThe BioHPC Nucleus Cluster & Future Developments
1 The BioHPC Nucleus Cluster & Future Developments Overview Today we ll talk about the BioHPC Nucleus HPC cluster with some technical details for those interested! How is it designed? What hardware does
More informationLatest Advances in MVAPICH2 MPI Library for NVIDIA GPU Clusters with InfiniBand
Latest Advances in MVAPICH2 MPI Library for NVIDIA GPU Clusters with InfiniBand Presentation at GTC 2014 by Dhabaleswar K. (DK) Panda The Ohio State University E-mail: panda@cse.ohio-state.edu http://www.cse.ohio-state.edu/~panda
More informationLBRN - HPC systems : CCT, LSU
LBRN - HPC systems : CCT, LSU HPC systems @ CCT & LSU LSU HPC Philip SuperMike-II SuperMIC LONI HPC Eric Qeenbee2 CCT HPC Delta LSU HPC Philip 3 Compute 32 Compute Two 2.93 GHz Quad Core Nehalem Xeon 64-bit
More informationInfiniBand Networked Flash Storage
InfiniBand Networked Flash Storage Superior Performance, Efficiency and Scalability Motti Beck Director Enterprise Market Development, Mellanox Technologies Flash Memory Summit 2016 Santa Clara, CA 1 17PB
More informationThe Future of High Performance Interconnects
The Future of High Performance Interconnects Ashrut Ambastha HPC Advisory Council Perth, Australia :: August 2017 When Algorithms Go Rogue 2017 Mellanox Technologies 2 When Algorithms Go Rogue 2017 Mellanox
More informationBuilding the Most Efficient Machine Learning System
Building the Most Efficient Machine Learning System Mellanox The Artificial Intelligence Interconnect Company June 2017 Mellanox Overview Company Headquarters Yokneam, Israel Sunnyvale, California Worldwide
More informationDell EMC Ready Bundle for HPC Digital Manufacturing Dassault Systѐmes Simulia Abaqus Performance
Dell EMC Ready Bundle for HPC Digital Manufacturing Dassault Systѐmes Simulia Abaqus Performance This Dell EMC technical white paper discusses performance benchmarking results and analysis for Simulia
More informationResources Current and Future Systems. Timothy H. Kaiser, Ph.D.
Resources Current and Future Systems Timothy H. Kaiser, Ph.D. tkaiser@mines.edu 1 Most likely talk to be out of date History of Top 500 Issues with building bigger machines Current and near future academic
More informationScaling with PGAS Languages
Scaling with PGAS Languages Panel Presentation at OFA Developers Workshop (2013) by Dhabaleswar K. (DK) Panda The Ohio State University E-mail: panda@cse.ohio-state.edu http://www.cse.ohio-state.edu/~panda
More informationSTAR-CCM+ Performance Benchmark and Profiling. July 2014
STAR-CCM+ Performance Benchmark and Profiling July 2014 Note The following research was performed under the HPC Advisory Council activities Participating vendors: CD-adapco, Intel, Dell, Mellanox Compute
More informationData Processing at the Speed of 100 Gbps using Apache Crail. Patrick Stuedi IBM Research
Data Processing at the Speed of 100 Gbps using Apache Crail Patrick Stuedi IBM Research The CRAIL Project: Overview Data Processing Framework (e.g., Spark, TensorFlow, λ Compute) Spark-IO FS Albis Streaming
More informationAccelerating Hadoop Applications with the MapR Distribution Using Flash Storage and High-Speed Ethernet
WHITE PAPER Accelerating Hadoop Applications with the MapR Distribution Using Flash Storage and High-Speed Ethernet Contents Background... 2 The MapR Distribution... 2 Mellanox Ethernet Solution... 3 Test
More informationChecklist for Selecting and Deploying Scalable Clusters with InfiniBand Fabrics
Checklist for Selecting and Deploying Scalable Clusters with InfiniBand Fabrics Lloyd Dickman, CTO InfiniBand Products Host Solutions Group QLogic Corporation November 13, 2007 @ SC07, Exhibitor Forum
More informationMellanox InfiniBand Solutions Accelerate Oracle s Data Center and Cloud Solutions
Mellanox InfiniBand Solutions Accelerate Oracle s Data Center and Cloud Solutions Providing Superior Server and Storage Performance, Efficiency and Return on Investment As Announced and Demonstrated at
More informationOverview of Tianhe-2
Overview of Tianhe-2 (MilkyWay-2) Supercomputer Yutong Lu School of Computer Science, National University of Defense Technology; State Key Laboratory of High Performance Computing, China ytlu@nudt.edu.cn
More informationGROMACS Performance Benchmark and Profiling. August 2011
GROMACS Performance Benchmark and Profiling August 2011 Note The following research was performed under the HPC Advisory Council activities Participating vendors: Intel, Dell, Mellanox Compute resource
More informationGROMACS (GPU) Performance Benchmark and Profiling. February 2016
GROMACS (GPU) Performance Benchmark and Profiling February 2016 2 Note The following research was performed under the HPC Advisory Council activities Participating vendors: Dell, Mellanox, NVIDIA Compute
More informationDesigning High Performance Communication Middleware with Emerging Multi-core Architectures
Designing High Performance Communication Middleware with Emerging Multi-core Architectures Dhabaleswar K. (DK) Panda Department of Computer Science and Engg. The Ohio State University E-mail: panda@cse.ohio-state.edu
More informationAcuSolve Performance Benchmark and Profiling. October 2011
AcuSolve Performance Benchmark and Profiling October 2011 Note The following research was performed under the HPC Advisory Council activities Participating vendors: AMD, Dell, Mellanox, Altair Compute
More informationShort Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy
Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy François Tessier, Venkatram Vishwanath Argonne National Laboratory, USA July 19,
More informationSNAP Performance Benchmark and Profiling. April 2014
SNAP Performance Benchmark and Profiling April 2014 Note The following research was performed under the HPC Advisory Council activities Participating vendors: HP, Mellanox For more information on the supporting
More informationThe Future of Interconnect Technology
The Future of Interconnect Technology Michael Kagan, CTO HPC Advisory Council Stanford, 2014 Exponential Data Growth Best Interconnect Required 44X 0.8 Zetabyte 2009 35 Zetabyte 2020 2014 Mellanox Technologies
More informationAn ESS implementation in a Tier 1 HPC Centre
An ESS implementation in a Tier 1 HPC Centre Maximising Performance - the NeSI Experience José Higino (NeSI Platforms and NIWA, HPC Systems Engineer) Outline What is NeSI? The National Platforms Framework
More informationSupport for GPUs with GPUDirect RDMA in MVAPICH2 SC 13 NVIDIA Booth
Support for GPUs with GPUDirect RDMA in MVAPICH2 SC 13 NVIDIA Booth by D.K. Panda The Ohio State University E-mail: panda@cse.ohio-state.edu http://www.cse.ohio-state.edu/~panda Outline Overview of MVAPICH2-GPU
More informationFUJITSU PHI Turnkey Solution
FUJITSU PHI Turnkey Solution Integrated ready to use XEON-PHI based platform Dr. Pierre Lagier ISC2014 - Leipzig PHI Turnkey Solution challenges System performance challenges Parallel IO best architecture
More informationMILC Performance Benchmark and Profiling. April 2013
MILC Performance Benchmark and Profiling April 2013 Note The following research was performed under the HPC Advisory Council activities Special thanks for: HP, Mellanox For more information on the supporting
More informationUAntwerpen, 24 June 2016
Tier-1b Info Session UAntwerpen, 24 June 2016 VSC HPC environment Tier - 0 47 PF Tier -1 623 TF Tier -2 510 Tf 16,240 CPU cores 128/256 GB memory/node IB EDR interconnect Tier -3 HOPPER/TURING STEVIN THINKING/CEREBRO
More informationFeedback on BeeGFS. A Parallel File System for High Performance Computing
Feedback on BeeGFS A Parallel File System for High Performance Computing Philippe Dos Santos et Georges Raseev FR 2764 Fédération de Recherche LUmière MATière December 13 2016 LOGO CNRS LOGO IO December
More informationNERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber
NERSC Site Update National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory Richard Gerber NERSC Senior Science Advisor High Performance Computing Department Head Cori
More informationAccelerating HPL on Heterogeneous GPU Clusters
Accelerating HPL on Heterogeneous GPU Clusters Presentation at GTC 2014 by Dhabaleswar K. (DK) Panda The Ohio State University E-mail: panda@cse.ohio-state.edu http://www.cse.ohio-state.edu/~panda Outline
More informationHPC Systems Overview. SDSC Summer Institute August 6-10, 2012 San Diego, CA. Shawn Strande Gordon Project Manager SAN DIEGO SUPERCOMPUTER CENTER
HPC Systems Overview SDSC Summer Institute August 6-10, 2012 San Diego, CA Shawn Strande Gordon Project Manager Trestles High Productivity System Targeted at modest scale jobs and Science Gateways Appro
More informationMulti-Threaded UPC Runtime for GPU to GPU communication over InfiniBand
Multi-Threaded UPC Runtime for GPU to GPU communication over InfiniBand Miao Luo, Hao Wang, & D. K. Panda Network- Based Compu2ng Laboratory Department of Computer Science and Engineering The Ohio State
More informationLS-DYNA Performance Benchmark and Profiling. October 2017
LS-DYNA Performance Benchmark and Profiling October 2017 2 Note The following research was performed under the HPC Advisory Council activities Participating vendors: LSTC, Huawei, Mellanox Compute resource
More informationHigh Performance Computing
High Performance Computing Dror Goldenberg, HPCAC Switzerland Conference March 2015 End-to-End Interconnect Solutions for All Platforms Highest Performance and Scalability for X86, Power, GPU, ARM and
More informationTECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 13 th CALL (T ier-0)
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 13 th CALL (T ier-0) Contributing sites and the corresponding computer systems for this call are: BSC, Spain IBM System x idataplex CINECA, Italy Lenovo System
More informationNext-Generation NVMe-Native Parallel Filesystem for Accelerating HPC Workloads
Next-Generation NVMe-Native Parallel Filesystem for Accelerating HPC Workloads Liran Zvibel CEO, Co-founder WekaIO @liranzvibel 1 WekaIO Matrix: Full-featured and Flexible Public or Private S3 Compatible
More informationUK LUG 10 th July Lustre at Exascale. Eric Barton. CTO Whamcloud, Inc Whamcloud, Inc.
UK LUG 10 th July 2012 Lustre at Exascale Eric Barton CTO Whamcloud, Inc. eeb@whamcloud.com Agenda Exascale I/O requirements Exascale I/O model 3 Lustre at Exascale - UK LUG 10th July 2012 Exascale I/O
More informationCP2K Performance Benchmark and Profiling. April 2011
CP2K Performance Benchmark and Profiling April 2011 Note The following research was performed under the HPC Advisory Council activities Participating vendors: AMD, Dell, Mellanox Compute resource - HPC
More informationHPC Middle East. KFUPM HPC Workshop April Mohamed Mekias HPC Solutions Consultant. Agenda
KFUPM HPC Workshop April 29-30 2015 Mohamed Mekias HPC Solutions Consultant Agenda 1 Agenda-Day 1 HPC Overview What is a cluster? Shared v.s. Distributed Parallel v.s. Massively Parallel Interconnects
More informationLS-DYNA Performance Benchmark and Profiling. October 2017
LS-DYNA Performance Benchmark and Profiling October 2017 2 Note The following research was performed under the HPC Advisory Council activities Participating vendors: LSTC, Huawei, Mellanox Compute resource
More informationGROMACS Performance Benchmark and Profiling. September 2012
GROMACS Performance Benchmark and Profiling September 2012 Note The following research was performed under the HPC Advisory Council activities Participating vendors: AMD, Dell, Mellanox Compute resource
More information