Synonymous with supercomputing Tightly-coupled applications Implemented using Message Passing Interface (MPI) Large of amounts of computing for short

Similar documents
Typically applied in clusters and grids Loosely-coupled applications with sequential jobs Large amounts of computing for long periods of times

Ioan Raicu Distributed Systems Laboratory Computer Science Department University of Chicago

NFS, GPFS, PVFS, Lustre Batch-scheduled systems: Clusters, Grids, and Supercomputers Programming paradigm: HPC, MTC, and HTC

Ian Foster, An Overview of Distributed Systems

Ioan Raicu Distributed Systems Laboratory Computer Science Department University of Chicago

Arguably one of the most fundamental discipline that touches all other disciplines and people

NFS, GPFS, PVFS, Lustre Batch-scheduled systems: Clusters, Grids, and Supercomputers Programming paradigm: HPC, MTC, and HTC

Managing and Executing Loosely-Coupled Large-Scale Applications on Clusters, Grids, and Supercomputers

Workflow languages and systems

Extreme scripting and other adventures in data-intensive computing

A Data Diffusion Approach to Large Scale Scientific Exploration

MOHA: Many-Task Computing Framework on Hadoop

Ioan Raicu. Distributed Systems Laboratory Computer Science Department University of Chicago

Overview Past Work Future Work. Motivation Proposal. Work-in-Progress

Introduction & Motivation Problem Statement Proposed Work Evaluation Conclusions Future Work

Grid, cloud, and science: Accelerating discovery. A View and Practice from University of Chicago

Extreme-scale scripting: Opportunities for large taskparallel applications on petascale computers

Ioan Raicu. Everyone else. More information at: Background? What do you want to get out of this course?

Case Studies in Storage Access by Loosely Coupled Petascale Applications

The Fusion Distributed File System

Harnessing Grid Resources to Enable the Dynamic Analysis of Large Astronomy Datasets

Accelerating Large Scale Scientific Exploration through Data Diffusion

Keywords: many-task computing; MTC; high-throughput computing; resource management; Falkon; Swift

Storage and Compute Resource Management via DYRE, 3DcacheGrid, and CompuStore

MATE-EC2: A Middleware for Processing Data with Amazon Web Services

Data Management in Parallel Scripting

Storage and Compute Resource Management via DYRE, 3DcacheGrid, and CompuStore Ioan Raicu, Ian Foster

Data Diffusion: Dynamic Resource Provision and Data-Aware Scheduling for Data-Intensive Applications

Introduction & Motivation Problem Statement Proposed Work Evaluation Conclusions Future Work

The Use of Cloud Computing Resources in an HPC Environment

Warehouse- Scale Computing and the BDAS Stack

Crossing the Chasm: Sneaking a parallel file system into Hadoop

System Software for Big Data and Post Petascale Computing

CS Project Report

Magellan Project. Jeff Broughton NERSC Systems Department Head October 7, 2009

Rapid Deployment of VS Workflows. Meta Scheduling Service

HETEROGENEOUS HPC, ARCHITECTURAL OPTIMIZATION, AND NVLINK STEVE OBERLIN CTO, TESLA ACCELERATED COMPUTING NVIDIA

The Quest for Scalable Support of Data-Intensive Workloads in Distributed Systems

Revealing Applications Access Pattern in Collective I/O for Cache Management

EFFICIENT ALLOCATION OF DYNAMIC RESOURCES IN A CLOUD

Adapted from: TRENDS AND ATTRIBUTES OF HORIZONTAL AND VERTICAL COMPUTING ARCHITECTURES

Introduction to Grid Computing

Write a technical report Present your results Write a workshop/conference paper (optional) Could be a real system, simulation and/or theoretical

Disk Cache-Aware Task Scheduling

INTERNATIONAL JOURNAL OF COMPUTER ENGINEERING & TECHNOLOGY (IJCET)

Introduction to High-Performance Computing

Experiences in Optimizing a $250K Cluster for High- Performance Computing Applications

Grid Scheduling Architectures with Globus

The Stampede is Coming: A New Petascale Resource for the Open Science Community

Day 9: Introduction to CHTC

Crossing the Chasm: Sneaking a parallel file system into Hadoop

Towards Jungle Computing with Ibis/Constellation

A Container On a Virtual Machine On an HPC? Presentation to HPC Advisory Council. Perth, July 31-Aug 01, 2017

Analytics in the cloud

IME (Infinite Memory Engine) Extreme Application Acceleration & Highly Efficient I/O Provisioning

Towards Data Intensive Many-Task Computing

HPC Storage Use Cases & Future Trends

FuncX: A Function Serving Platform for HPC. Ryan Chard 28 Jan 2019

High Performance Data Analytics for Numerical Simulations. Bruno Raffin DataMove

A Breakthrough in Non-Volatile Memory Technology FUJITSU LIMITED

Leveraging Software-Defined Storage to Meet Today and Tomorrow s Infrastructure Demands

High Performance Computing Course Notes HPC Fundamentals

Falkon: a Fast and Light-weight task execution framework

Apache Spark is a fast and general-purpose engine for large-scale data processing Spark aims at achieving the following goals in the Big data context

AN INTRODUCTION TO CLUSTER COMPUTING

RESEARCH DATA DEPOT AT PURDUE UNIVERSITY

On the effectiveness of application-aware self-management for scientific discovery in volunteer computing systems

Beyond Petascale. Roger Haskin Manager, Parallel File Systems IBM Almaden Research Center

Evolving To The Big Data Warehouse

Data Processing at the Speed of 100 Gbps using Apache Crail. Patrick Stuedi IBM Research

Highly Scalable, Non-RDMA NVMe Fabric. Bob Hansen,, VP System Architecture

TECHNICAL OVERVIEW ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING

Big Data Systems on Future Hardware. Bingsheng He NUS Computing

An Overview of Fujitsu s Lustre Based File System

Sharing High-Performance Devices Across Multiple Virtual Machines

High Throughput WAN Data Transfer with Hadoop-based Storage

MM5 Modeling System Performance Research and Profiling. March 2009

Pervasive DataRush TM

Performance Optimizations via Connect-IB and Dynamically Connected Transport Service for Maximum Performance on LS-DYNA

XSEDE High Throughput Computing Use Cases

Towards Scalable Data Management for Map-Reduce-based Data-Intensive Applications on Cloud and Hybrid Infrastructures

PERFORMANCE ANALYSIS AND OPTIMIZATION OF MULTI-CLOUD COMPUITNG FOR LOOSLY COUPLED MTC APPLICATIONS

Falkon: a Fast and Light-weight task execution framework

2013 AWS Worldwide Public Sector Summit Washington, D.C.

Parsl: Developing Interactive Parallel Workflows in Python using Parsl

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 6 th CALL (Tier-0)

Professor: Ioan Raicu. TA: Wei Tang. Everyone else

Data Intensive Scalable Computing

Data Analytics and Storage System (DASS) Mixing POSIX and Hadoop Architectures. 13 November 2016

DDN About Us Solving Large Enterprise and Web Scale Challenges

Ian Foster, CS554: Data-Intensive Computing

GEMTC: GPU Enabled Many-Task Computing

Advances of parallel computing. Kirill Bogachev May 2016

University at Buffalo Center for Computational Research

Topics. Big Data Analytics What is and Why Hadoop? Comparison to other technologies Hadoop architecture Hadoop ecosystem Hadoop usage examples

Data Processing at the Speed of 100 Gbps using Apache Crail. Patrick Stuedi IBM Research

Habanero Operating Committee. January

Scaling to Petaflop. Ola Torudbakken Distinguished Engineer. Sun Microsystems, Inc

Was ist dran an einer spezialisierten Data Warehousing platform?

Lustre2.5 Performance Evaluation: Performance Improvements with Large I/O Patches, Metadata Improvements, and Metadata Scaling with DNE

Transcription:

Synonymous with supercomputing Tightly-coupled applications Implemented using Message Passing Interface (MPI) Large of amounts of computing for short periods of time Usually requires low latency interconnects Measured in FLOPS 2

Typically applied in clusters and grids Loosely-coupled applications with sequential jobs Large amounts of computing for long periods of times Measured in operations per month or years 3

Bridge the gap between HPC and HTC Applied in clusters, grids, and supercomputers Loosely coupled apps with HPC orientations Many activities coupled by file system ops Many resources over short time periods Large number of tasks, large quantity of computing, and large volumes of data [MTAGS08 Workshop] Workshop on Many-Task Computing on Grids and Supercomputers 2008 [SC08] Towards Loosely-Coupled Programming on Petascale Systems [MTAGS08] Many-Task Computing for Grids and Supercomputers 4

Input Data Size Hi Med MapReduce/MTC (Data Analysis, Mining) MTC (Big Data and Many Tasks) Low HPC (Heroic MPI Tasks) HTC/MTC (Many Loosely Coupled Tasks) 1 1K 1M Number of Tasks [MTAGS08] Many-Task Computing for Grids and Supercomputers 5

Goal: enable the rapid and efficient execution of many independent jobs on large compute clusters Combines three components: a streamlined task dispatcher resource provisioning through multi-level scheduling User Task Dispatcher techniques Persistent Storage Data-Aware Scheduler data diffusion and data-aware scheduling to leverage the Wait Queue Provisioned Resources co-located computational and storage resources Dynamic Resource Executor 1 Integration into Swift to Provisioning leverage many applications Available Resources (GRAM4) Applications cover many domains: astronomy, astro-physics, Executor Executor n medicine, chemistry, economics, climate modeling, etc i text [SciDAC09] Extreme-scale scripting: Opportunities for large task-parallel applications on petascale computers [SC08] Towards Loosely-Coupled Programming on Petascale Systems [Globus07] Falkon: A Proposal for Project Globus Incubation [SC07] Falkon: a Fast and Light-weight task execution framework [SWF07] Swift: Fast, Reliable, Loosely Coupled Parallel Computation 6

Falkon is a real system Late 2005: Initial prototype, AstroPortal January 2007: Falkon v0 November 2007: Globus incubator project v0.1 Workload http://dev.globus.org/wiki/incubator/falkon 160K CPUs February 2009: Globus incubator project v0.9 1M tasks Implemented in 60 Java sec per (~20K task lines of code) and C 2 CPU years in 453 sec (~1K lines of code) Open source: svn co https://svn.globus.org/repos/falkon Source code contributors (beside myself) Yong Zhao, Zhao Zhang, Ben Clifford, Mihael Hategan [Globus07] Falkon: A Proposal for Project Globus Incubation [CLUSTER10] Middleware Support for Many-Task Computing Throughput: 2312 tasks/sec 85% efficiency 7

[TPDS09] Middleware Support for Many-Task Computing, under preparation 8

Login Nodes (x10) Client I/O Nodes (x640) Dispatcher 1 Compute Nodes (x40k) Executor 1 Executor 256 Provisioner Cobalt Dispatcher N Executor 1 Executor 256 [SC08] Towards Loosely-Coupled Programming on Petascale Systems 9

ZeptOS Falkon High-speed local disk Slower distributed storage [SC08] Towards Loosely-Coupled Programming on Petascale Systems 10

Improvement: up to 90% lower end-to-end run time Wide range of analyses Testing, interactive analysis, production runs Data mining Parameter studies [SC07] Falkon: a Fast and Light-weight task execution framework [SWF07] Swift: Fast, Reliable, Loosely Coupled Parallel Computation 11

Improvement: up to 57% lower end-to-end run time Within 4% of MPI B. Berriman, J. Good (Caltech) J. Jacob, D. Katz (JPL) [SC07] Falkon: a Fast and Light-weight task execution framework [SWF07] Swift: Fast, Reliable, Loosely Coupled Parallel Computation 12

Determination of free energies in aqueous solution Antechamber coordinates Charmm solution Charmm - free energy Improvement: up to 88% lower end-to-end run time 5X more scalable [NOVA08] Realizing Fast, Scalable and Reliable Scientific Computations in Grid Environments 13

Time (sec) Time (sec) 10000 1000 Classic benchmarks for MapReduce Word Count Sort Swift and Falkon performs similar or better than Hadoop (on 32 processors) 100 Swift+PBS Hadoop 221 863 Word Count 4688 1143 1795 7860 10000 1000 100 Swift+Falkon Hadoop 42 25 Sort 85 83 733 512 10 10 1 75MB 350MB 703MB Data Size 1 10MB 100MB 1000MB14 Data Size

Tasks Completed Number of Processors Throughput (tasks/sec) CPU Cores: 130816 Tasks: 1048576 Elapsed time: 2483 secs CPU Years: 9.3 1000000 800000 600000 Processors Active Tasks Tasks Completed Throughput (tasks/sec) Speedup: 115168X (ideal 130816) Efficiency: 88% 5000 4500 4000 3500 3000 2500 400000 200000 2000 1500 1000 500 0 0 Time (sec) [SC08] Towards Loosely-Coupled Programming on Petascale Systems 15

PDB protein descriptions ZINC 3-D structures 1 protein (1MB) 6 GB 2M structures (6 GB) FRED Manually prep DOCK6 rec file start DOCK6 Receptor (1 per protein: defines pocket to bind to) DOCK6 Manually prep FRED rec file FRED Receptor (1 per protein: defines pocket to bind to) ~4M x 60s x 1 cpu ~60K cpu-hrs NAB Script Template BuildNABScript NAB Script NAB script parameters (defines flexible residues, #MDsteps) Amber prep: 2. AmberizeReceptor 4. perl: gen nabscript report Select best ~5K Select best ~5K Amber Select best ~500 GCMC end ligands [SC08] Towards Loosely-Coupled Programming on Petascale Systems complexes ~10K x 20m x 1 cpu ~3K cpu-hrs ~500 x 10hr x 100 cpu ~500K cpu-hrs Amber Score: 1. AmberizeLigand 3. AmberizeComplex 5. RunNABScript For 1 target: 4 million tasks 500,000 cpu-hrs (50 cpu-years) 16

CPU cores: 118784 Tasks: 934803 Elapsed time: 2.01 hours Compute time: 21.43 CPU years Average task time: 667 sec Relative Efficiency: 99.7% (from 16 to 32 racks) Utilization: Sustained: 99.6% Overall: 78.3% [SC08] Towards Loosely-Coupled Programming on Petascale Systems 17

Purpose On-demand stacks of random locations within ~10TB dataset Challenge Processing Costs: O(100ms) per object Data Intensive: 40MB:1sec Rapid access to 10-10K random files Time-varying load [DADC08] Accelerating Large-scale Data Exploration through Data Diffusion [TG06] AstroPortal: A Science Gateway for Large-scale Astronomy Data Analysis AP Sloan Data Locality Number of Objects Number of Files 1 111700 111700 1.38 154345 111699 2 97999 49000 3 88857 29620 4 76575 19145 5 60590 12120 10 46480 4650 20 40460 2025 30 23695 790 18 + + + + + + + =

AstroPortal Makes it really easy for astronomers to create stackings of objects from the Sloan Digital Sky Servey (SDSS) Throughput dataset 10X higher than GPFS Reduced load 1/10 of the original GPFS load Increased scalability 8X 19

There is more to HPC than tightly coupled MPI, and more to HTC than embarrassingly parallel long jobs Data locality is critical at large-scale 20

Embarrassingly Happily parallel apps are trivial to run Logistical problems can be tremendous Loosely coupled apps do not require supercomputers Total computational requirements can be enormous Individual tasks may be tightly coupled Workloads frequently involve large amounts of I/O Make use of idle resources from supercomputers via backfilling Costs to run supercomputers per FLOP is among the best Loosely coupled apps do not require specialized system software Their requirements on the job submission and storage systems can be extremely large Shared/parallel file systems are good for all applications They don t scale proportionally with the compute resources Data intensive applications don t perform and scale well Growing compute/storage gap 21

22