Introduction to High-Performance Computing

Similar documents
Interconnect Your Future Enabling the Best Datacenter Return on Investment. TOP500 Supercomputers, November 2017

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

The Impact of Inter-node Latency versus Intra-node Latency on HPC Applications The 23 rd IASTED International Conference on PDCS 2011

MPI Optimizations via MXM and FCA for Maximum Performance on LS-DYNA

Optimizing LS-DYNA Productivity in Cluster Environments

Performance Optimizations via Connect-IB and Dynamically Connected Transport Service for Maximum Performance on LS-DYNA

Low Cost Supercomputing. Rajkumar Buyya, Monash University, Melbourne, Australia. Parallel Processing on Linux Clusters

Creating High Performance Clusters for Embedded Use

LS-DYNA Performance Benchmark and Profiling. October 2017

Single-Points of Performance

The Road to ExaScale. Advances in High-Performance Interconnect Infrastructure. September 2011

Birds of a Feather Presentation

HPC Considerations for Scalable Multidiscipline CAE Applications on Conventional Linux Platforms. Author: Correspondence: ABSTRACT:

Interconnect Your Future

Altair RADIOSS Performance Benchmark and Profiling. May 2013

High Performance Computing Course Notes HPC Fundamentals

High Performance Computing Course Notes Course Administration

AcuSolve Performance Benchmark and Profiling. October 2011

NEMO Performance Benchmark and Profiling. May 2011

LS-DYNA Productivity and Power-aware Simulations in Cluster Environments

Managing CAE Simulation Workloads in Cluster Environments

OpenFOAM Performance Testing and Profiling. October 2017

LS-DYNA Performance Benchmark and Profiling. April 2015

NAMD GPU Performance Benchmark. March 2011

ABySS Performance Benchmark and Profiling. May 2010

LAMMPSCUDA GPU Performance. April 2011

ANSYS Fluent 14 Performance Benchmark and Profiling. October 2012

GROMACS (GPU) Performance Benchmark and Profiling. February 2016

Outline. Course Administration /6.338/SMA5505. Parallel Machines in Applications Special Approaches Our Class Computer.

GROMACS Performance Benchmark and Profiling. August 2011

The Effect of In-Network Computing-Capable Interconnects on the Scalability of CAE Simulations

OCTOPUS Performance Benchmark and Profiling. June 2015

SNAP Performance Benchmark and Profiling. April 2014

InfiniBand Strengthens Leadership as The High-Speed Interconnect Of Choice

LAMMPS-KOKKOS Performance Benchmark and Profiling. September 2015

The Future of High Performance Interconnects

STAR-CCM+ Performance Benchmark and Profiling. July 2014

Future Routing Schemes in Petascale clusters

High Performance Computing in C and C++

Altair OptiStruct 13.0 Performance Benchmark and Profiling. May 2015

MM5 Modeling System Performance Research and Profiling. March 2009

New Features in LS-DYNA HYBRID Version

CP2K Performance Benchmark and Profiling. April 2011

MELLANOX EDR UPDATE & GPUDIRECT MELLANOX SR. SE 정연구

LS-DYNA Performance Benchmark and Profiling. October 2017

PART I - Fundamentals of Parallel Computing

CESM (Community Earth System Model) Performance Benchmark and Profiling. August 2011

AMBER 11 Performance Benchmark and Profiling. July 2011

Real Application Performance and Beyond

CPMD Performance Benchmark and Profiling. February 2014

NAMD Performance Benchmark and Profiling. November 2010

GROMACS Performance Benchmark and Profiling. September 2012

The Future of Interconnect Technology

Nehalem Hochleistungsrechnen für reale Anwendungen

CP2K Performance Benchmark and Profiling. April 2011

Introduction to High-Speed InfiniBand Interconnect

AcuSolve Performance Benchmark and Profiling. October 2011

BİL 542 Parallel Computing

Parallel Programming with MPI

QLogic in HPC Vendor Update IDC HPC User Forum April 16, 2008 Jeff Broughton Sr. Director Engineering Host Solutions Group

MILC Performance Benchmark and Profiling. April 2013

ICON Performance Benchmark and Profiling. March 2012

Real Parallel Computers

Interconnect Your Future

Considerations for LS-DYNA Workflow Efficiencies in an HPC Linux Environment

Partitioning Effects on MPI LS-DYNA Performance

HYCOM Performance Benchmark and Profiling

Cluster Scalability of Implicit and Implicit-Explicit LS-DYNA Simulations Using a Parallel File System

Considerations for LS-DYNA Efficiency in SGI IRIX and Linux Environments with a NUMA System Architecture

Manufacturing Bringing New Levels of Performance to CAE Applications

Solutions for Scalable HPC

White Paper. How the Meltdown and Spectre bugs work and what you can do to prevent a performance plummet. Contents

QLogic TrueScale InfiniBand and Teraflop Simulations

Himeno Performance Benchmark and Profiling. December 2010

Assessment of LS-DYNA Scalability Performance on Cray XD1

Performance Benefits of NVIDIA GPUs for LS-DYNA

2008 International ANSYS Conference

Paving the Road to Exascale Computing. Yossi Avni

Intel High-Performance Computing. Technologies for Engineering

Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability

Paving the Road to Exascale

Mellanox Technologies Maximize Cluster Performance and Productivity. Gilad Shainer, October, 2007

TECHNICAL OVERVIEW ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING

NAMD Performance Benchmark and Profiling. January 2015

TECHNICAL OVERVIEW ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING

VPI / InfiniBand. Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability

HIGH-PERFORMANCE STORAGE FOR DISCOVERY THAT SOARS

LAMMPS Performance Benchmark and Profiling. July 2012

represent parallel computers, so distributed systems such as Does not consider storage or I/O issues

Intel Accelerates Supercomputing. Press Briefing June 19, 2007

HPC projects. Grischa Bolls

Corporate Update. Enabling The Use of Data January Mellanox Technologies

FUSION PROCESSORS AND HPC

VPI / InfiniBand. Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Data Center Performance, Efficiency and Scalability

INFOBrief. Dell-IBRIX Cluster File System Solution. Key Points

Real Parallel Computers

Computer Architecture Crash course

Top500 Supercomputer list

BlueGene/L. Computer Science, University of Warwick. Source: IBM

Part 2: Computing and Networking Capacity (for research and instructional activities)

Transcription:

Introduction to High-Performance Computing

2 What is High Performance Computing? There is no clear definition Computing on high performance computers Solving problems / doing research using computer modeling, simulation and analysis Engineering design using computer modeling, simulation and analysis My understanding A huge number of computational and memory requirements Cannot be afforded by a PC efficiently Speeds and feeds are the keywords Who uses High-Performance Computing Research institutes, universities and government labs Weather and climate research, bioscience, energy, military etc. Engineering design: more or less every product we use Automotive, aerospace, oil and gas explorations, digital media, financial simulation Mechanical simulation, package designs, silicon manufacturing etc. Similar concepts Parallel computing: computing on parallel computers Super computing: computing on world 500 fastest supercomputers

3 When Do We Need High Performance Computing? Case1: Complete a time-consuming operation in less time I am an automotive engineer I need to design a new car that consumes less gasoline I d rather have the design completed in 6 months than in 2 years I want to test my design using computer simulations rather than building very expensive prototypes and crashing them Case 2: Complete an operation under a tight deadline I work for a weather prediction agency I am getting input from weather stations/sensors I d like to predict tomorrow s forecast today Case 3: Perform a high number of operations per seconds I am an engineer at Amazon.com My Web server gets 1,000 hits per seconds I d like my web server and databases to handle 1,000 transactions per seconds so that customers do not experience bad delays

What Does High Performance Computing Include? High-performance computing is fast computing Computations in parallel over lots of compute elements (CPU, GPU) Very fast network to connect between the compute elements Hardware Computer Architecture Vector Computers, MPP, SMP, Distributed Systems, Clusters Network Connections InfiniBand, Ethernet, Proprietary Software Programming models MPI (Message Passing Interface), SHMEM (Shared Memory), PGAS, etc. Applications Open source, commercial 4

5 Rise and Fall of HPC Computer Architectures Vector Computers (VC) - proprietary system Provided the breakthrough needed for the emergence of computational science, but they were only a partial answer Massively Parallel Processors (MPP) - proprietary systems High cost and a low performance/price ratio. Symmetric Multiprocessors (SMP) Suffers from scalability Distributed Systems Difficult to use and hard to extract parallel performance Clusters commodity and highly popular High Performance Computing - Commodity Supercomputing High Availability Computing - Mission Critical Applications

6 Top500 Supercomputers List System Architecture (June 2017) Clusters have become the most used HPC system architecture More than 86% of Top500 systems are clusters

7 Computer Food Chain: Causing the Demise of Specialized Systems Parallel Computing on a Large Number of Servers is More Efficient than using Specialized Systems

8 HPC Clusters Affordable, Efficient and Scalable HPC Solution Since the 1990s, there has been an increasing trend to move away from expensive /specialized proprietary parallel supercomputers to clusters of computers From specialized supercomputers to cost effective, general purpose systems So What s So Different about Clusters? Commodity, standard, affordable, cost effective, scalable and reliable architecture Cluster Architecture Server Server Server Server Communications Software Communications Software Communications Software Communications Software Network Interface Hardware Network Interface Hardware Network Interface Hardware Network Interface Hardware Cluster Interconnect Network

SPEED 9 Interconnect Technology: The Need for Speed and Intelligence 400G 200G 100G LS-DYNA (FEA) OpenFOAM (CFD) Homeland Security Weather Cosmological Simulations Human Genome The Large Hadron Collider (CERN) Brain Mapping 50G 40G SIZE 100 Nodes 1,000 Nodes 10,000 Nodes 100,000 Nodes 1,000,000 Nodes

10 Cluster Interconnect Commoditization/standardization are the clustering and interconnect driving forces InfiniBand and Ethernet are the most used interconnect solutions for HPC systems

11 The HPC Advantage: Reduction in Time to Market Product Development Process Concept Development Prototype/ Physical Testing Manufacturing Design Final Preparation/Delivery Without HPC With HPC Time Savings HPC-Supported Simulation Time to Market Source: IDC

12 Summary From concept to engineering, from design to test and manufacturing, from weather prediction to medical discoveries, our day to day life depends more and more on HPC simulations Safer products, accurate predictions, research, etc. High-performance compute clusters provide the most efficient, flexible, cost effective HPC environment for any HPC simulation

Thank You www.hpcadvisorycouncil.com All trademarks are property of their respective owners. All information is provided As-Is without any kind of warranty. The HPC Advisory Council makes no representation to the accuracy and completeness of the information contained herein. HPC Advisory Council undertakes no duty and assumes no obligation to update or correct any information presented herein